-
Notifications
You must be signed in to change notification settings - Fork 540
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
CM script failed to run harness after docker done #1998
Comments
Looks like you are having GPUs from different generation on the system. AFAIK this is not supported by Nvidia implementation.
|
Thank you @arjunsuresh. No same error prompt out after I removed T1000 and only kept A6000 in the system. But when I run the below commands in the docker that is the same docker created by CM before system reboot: cm run script --tags=run-mlperf,inference,_r4.1-dev Another error occurred: Traceback (most recent call last): File "/home/cmuser/CM/repos/local/cache/1406981516ca4974/inference/vision/classification_and_detection/tools/accuracy-imagenet.py", line 89, in <module> main() File "/home/cmuser/CM/repos/local/cache/1406981516ca4974/inference/vision/classification_and_detection/tools/accuracy-imagenet.py", line 54, in main with open(args.mlperf_accuracy_file, "r") as f: FileNotFoundError: [Errno 2] No such file or directory: '/home/cmuser/CM/repos/local/cache/d54d1a74ced144d0/valid_results/9951fc73ce5b-nvidia_original-gpu-tensorrt-vdefault-default_config/resnet50/offline/accuracy/mlperf_log_accuracy.json' CM error: Portable CM script failed (name = process-mlperf-accuracy, return code = 256) |
Hi @arjunsuresh
I am running the Resnet50 benchmark with the command:
cm run script --tags=run-mlperf,inference,_find-performance,_full,_r4.1-dev
--model=resnet50
--implementation=nvidia
--framework=tensorrt
--category=edge
--scenario=Offline
--execution_mode=test
--device=cuda
--docker --quiet
--test_query_count=5000
failed to run harness as below and the docker was created successfully. How to resolve it?
log with dock done.txt
The text was updated successfully, but these errors were encountered: