@@ -21,19 +21,16 @@ RUN apt-get remove -y --allow-remove-essential perl-base && \
2121# Download tokenizers, distilbert for the Danswer model
2222# Download model weights
2323# Run Nomic to pull in the custom architecture and have it cached locally
24- RUN python -c "from transformers import AutoTokenizer; \
25- AutoTokenizer.from_pretrained('distilbert-base-uncased'); \
26- AutoTokenizer.from_pretrained('mixedbread-ai/mxbai-rerank-xsmall-v1'); \
27- from huggingface_hub import snapshot_download; \
28- snapshot_download(repo_id='danswer/hybrid-intent-token-classifier', revision='v1.0.3'); \
29- snapshot_download('nomic-ai/nomic-embed-text-v1'); \
30- snapshot_download('mixedbread-ai/mxbai-rerank-xsmall-v1'); \
31- from sentence_transformers import SentenceTransformer; \
32- SentenceTransformer(model_name_or_path='nomic-ai/nomic-embed-text-v1', trust_remote_code=True);"
24+ # RUN python -c "from transformers import AutoTokenizer; \
25+ # from huggingface_hub import snapshot_download; \
26+ # snapshot_download(repo_id='danswer/hybrid-intent-token-classifier', revision='v1.0.3'); \
27+ # snapshot_download('nomic-ai/nomic-embed-text-v1'); \
28+ # from sentence_transformers import SentenceTransformer; \
29+ # SentenceTransformer(model_name_or_path='nomic-ai/nomic-embed-text-v1', trust_remote_code=True);"
3330
3431# In case the user has volumes mounted to /root/.cache/huggingface that they've downloaded while
3532# running Danswer, don't overwrite it with the built in cache folder
36- RUN mv /root/.cache/huggingface /root/.cache/temp_huggingface
33+ # RUN mv /root/.cache/huggingface /root/.cache/temp_huggingface
3734
3835WORKDIR /app
3936
0 commit comments