Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Llama 3.3 redownloading in every new session and few other problems #564

Open
VistritPandey opened this issue Dec 19, 2024 · 5 comments
Open

Comments

@VistritPandey
Copy link

For reference, I am using 2 Mac Mini with M4 Pro and 48GB RAM each. Both are connected with Thunderbolt 5 cable. However, I am facing multiple issues:

  1. Llama 3.3-70b is working at 4~5 tokens/sec which is the same number of tokens or less I am getting when using only 1 machine and running the model with Ollama.
  2. After pulling the latest version of the repository, or whenever I restart the machines, it redownloads some of the models
  3. Llama 3.1-8b does not work at all
  4. Getting gRPC error in the logs
    Screenshot 2024-12-19 at 10 00 12
  5. It is not able to connect to Hugging Face to download the model, which it was able to do before I installed the latest version of MacOS (15.2 (24C101)). If it has something to do with the access token to hugging face, it is already there, and I can use it for other tasks
    Screenshot 2024-12-19 at 10 05 01
  6. It is stuck at a number beyond which it is not able to download the Llama3.3 that it downloaded before. Maybe it's corrupted somehow?
@AFDudley
Copy link

AFDudley commented Dec 28, 2024

My guess is that's not enough RAM. I have the same problem. https://www.substratus.ai/blog/calculating-gpu-memory-for-llm I'm still trying to figure out how to benchmark this system. when using llama 3.2 1B adding a node 32GB M2 Mac Mini to a single 64GB M4 mac mini varies performance up or down about 10%.

@VistritPandey
Copy link
Author

Ohh that maybe a good answer. Thanks! I was under the assumption that it takes all the RAM by default to be considered by the GPU

@rolliver
Copy link

rolliver commented Jan 7, 2025

Re: It is not able to connect to Hugging Face to download the model

ssl.SSLCertVerificationError: [SSL: CERTIFICATE_VERIFY_FAILED] certificate verify failed: unable to get local issuer
certificate (_ssl.c:1000)

I ran

/Applications/Python\ 3.12/Install\ Certificates.command
and
pip install --upgrade certifi

and it seemed to fix it

@VistritPandey
Copy link
Author

Re: It is not able to connect to Hugging Face to download the model

ssl.SSLCertVerificationError: [SSL: CERTIFICATE_VERIFY_FAILED] certificate verify failed: unable to get local issuer certificate (_ssl.c:1000)

I ran

/Applications/Python\ 3.12/Install\ Certificates.command and pip install --upgrade certifi

and it seemed to fix it

This actually fixed the issue for me! Thank you!

Although when I restart my machine, I had to re-run it

@ThaddeusJiang
Copy link

Re: It is not able to connect to Hugging Face to download the model

ssl.SSLCertVerificationError: [SSL: CERTIFICATE_VERIFY_FAILED] certificate verify failed: unable to get local issuer certificate (_ssl.c:1000)

I ran

/Applications/Python\ 3.12/Install\ Certificates.command and pip install --upgrade certifi

and it seemed to fix it

Thank you @rolliver

Fixed

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants