Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Unreachable #62

Open
flatsiedatsie opened this issue Jun 5, 2024 · 4 comments
Open

Unreachable #62

flatsiedatsie opened this issue Jun 5, 2024 · 4 comments
Labels
bug Something isn't working

Comments

@flatsiedatsie
Copy link
Contributor

  • Start LLM
  • Close laptop
  • Sleep 8 hours
  • Open laptop
  • Issue command to LLM
Screenshot 2024-06-05 at 07 08 03
@ngxson
Copy link
Owner

ngxson commented Jun 5, 2024

Probably it's an issue of the browser (equivalent to segfault if run in native)

@felladrin
Copy link
Contributor

I think the browser may be clearing the blobs from the memory when the tab gets suspended (after some time not being used).

@ngxson ngxson added the bug Something isn't working label Jun 25, 2024
@flatsiedatsie
Copy link
Contributor Author

Screenshot 2024-08-04 at 21 38 50

I just noticed this one again. This time on an Android mobile phone (Pixel 6a, Chrome), with just one browser tab open, and everything else closed manually.

I was trying to load a Gemma 2 2B it model.
https://huggingface.co/BoscoTheDog/gemma_2_2b_it_Q4_gguf_chunked

Context is set to 1K, the model is 1.63GB, and the Pixel has 6GB of RAM. According to the OS my average memory use is 3GB.

I think the browser may be clearing the blobs from the memory when the tab gets suspended

I don't think that's the case here, as the tab is the currently active one. Maybe it's just a lack of memory issue? Or maybe like on mobile Safari there's a limit to how much RAM a tab may use?

@flatsiedatsie
Copy link
Contributor Author

flatsiedatsie commented Aug 4, 2024

I tried to load another 1.6GB (Bitnet) model on the phone, and that did load. Hmm.

I'll do a quick git clone --recurse-submodules https://github.com/ngxson/wllama.git; cd wllama; git submodule update --remote --merge; npm i; npm run build:wasm; npm run build.

// Nice, Phi 3.1 mini loads and (very slowly) generates a response. It's 2.1GB.

// Updating llama.cpp solved it.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
Development

No branches or pull requests

3 participants