New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Detector Support]: Fatal Python error: Segmentation fault #9801
Comments
what GPU do you have again? |
Edited Post to remove the Frigtate Plus API key and RTSP password that was visible in the Docker CLI command. |
No one has ANY suggestions (besides buying a CORAL device) on how to get my Frigtate instance up and running?? |
seg faults are difficult because it is usually something related to the host or the hardware and there is no info about what is going wrong. From your previous post logs we can see that as soon as the model is initialized there is a seg fault indicating some failure to communicate correctly. Many users use this type of setup on unraid so it seems there is nothing particular about that. You could try a memtest and see if perhaps system memory is failing. |
memtest complete. 0 errors. Next suggestion please? |
I'm experiencing the exact same error on TrueNAS Scale w/ GTX 1060 |
@hvardhan20 and @jdgiddings - I hope you both get a response but, if my past experience holds true, it doesn't look good. CPU detection worked fine. GPU detection worked fine...... until they bundled it all into one container. |
There are many tensorrt users so this seems to be a very isolated problem. Like I said before, seg faults are difficult to debug and without being able to reproduce there really isn't any good way to move towards solving the problem because it is not clear what is causing this other than something on the host. The logic to compile the models is the same as before just done automatically, that is unlikely to be causing this. It could be due to using newer libraries / tensorrt version but that was done to support the latest Nvidia GPUs and also unrelated to frigate building the models automatically. |
here's the output from nvidia-smi on the host. I believe these are all supported versions NVIDIA-SMI 535.54.03 I'm experimenting with different models right now to see if any do not cause the error. I will report back |
yolov7-320 does not throw the segfault |
which model did you use that did? |
yolov7x-640 and yolov7x-320 were both throwing the error on my machine |
I did some more testing. Any model larger than yolov7-320 throws the same segfault error |
I just wanted to add another voice here -- I am able to run yolov7x-320, but if I attempt to run yolov7x-640, I get a segfault (the same as the OP). I'm on a GTX 1650 Super. My setup is a bit odd:
Let me know if I can do anything to help debug this. [edit] I previously said my 1650 is an LHR. This is incorrect. My 3060 is LHR, and I confused the two. |
This issue has been automatically marked as stale because it has not had recent activity. It will be closed if no further activity occurs. Thank you for your contributions. |
Describe the problem you are having
Launching v13 with NVIDIA branch causes a bootloop with the above error and no other explanation. I was told in a different support ticket that my NVIDIA driver version was too new
I have since downgraded to Driver v535.129.03 which is supposedly stable according to the last ticket I opened (#9575)
The error still is present.
Version
v13
Frigate config file
docker-compose file or Docker CLI command
Relevant log output
Operating system
UNRAID
Install method
Docker Compose
Coral version
CPU (no coral)
Any other information that may be helpful
No response
The text was updated successfully, but these errors were encountered: