Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Bug]: The multiprocess handler does not exit cleanly, if the subprocess is killed by the oomkiller #279

Open
kumare3 opened this issue Jun 18, 2024 · 3 comments

Comments

@kumare3
Copy link

kumare3 commented Jun 18, 2024

What happened?

When using fastembed with parallel=0 (or a higher number) and a large batch_size, the memory increases by a large amount (memory copies). In an event when the size is beyond the allocated size of the progress group (using containers), the OS oomkiller may kill the process. But, the multiprocessing pool does not handle this gracefully.

This is similar to the issue described here - https://stackoverflow.com/questions/61492362/multiprocessing-pool-hangs-if-child-process-killed

What Python version are you on? e.g. python --version

python 3.9/10/11

Version

0.3.0 (Latest)

What os are you seeing the problem on?

No response

Relevant stack traces and/or logs

No response

@nauyan
Copy link

nauyan commented Aug 26, 2024

@kumare3 has you been able to fix this issue?

@kumare3
Copy link
Author

kumare3 commented Aug 26, 2024

Nope, I gave up on fastembed. Found it hard to containerize and use, switched to sentence transformer and been very happy with it

@joein
Copy link
Member

joein commented Sep 9, 2024

@kumare3 sorry to hear, that it was hard for you to containerize it, could you provide any additional details?

@kumare3 @nauyan does the parent process just hangs indefinitely after an unexpected exit of a child process?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants