-
Notifications
You must be signed in to change notification settings - Fork 27
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Unreachable #62
Comments
Probably it's an issue of the browser (equivalent to segfault if run in native) |
I think the browser may be clearing the blobs from the memory when the tab gets suspended (after some time not being used). |
I just noticed this one again. This time on an Android mobile phone (Pixel 6a, Chrome), with just one browser tab open, and everything else closed manually. I was trying to load a Context is set to 1K, the model is 1.63GB, and the Pixel has 6GB of RAM. According to the OS my average memory use is 3GB.
I don't think that's the case here, as the tab is the currently active one. Maybe it's just a lack of memory issue? Or maybe like on mobile Safari there's a limit to how much RAM a tab may use? |
I tried to load another 1.6GB (Bitnet) model on the phone, and that did load. Hmm. I'll do a quick // Nice, Phi 3.1 mini loads and (very slowly) generates a response. It's 2.1GB. // Updating llama.cpp solved it. |
The text was updated successfully, but these errors were encountered: