-
Notifications
You must be signed in to change notification settings - Fork 41
/
Copy pathrequirements.txt
36 lines (30 loc) · 1.22 KB
/
requirements.txt
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
accelerate==1.4.0
bitsandbytes==0.45.2
ExifRead==3.0.0
imagesize==1.4.1
pillow==11.1.0
pyparsing==3.2.1
PySide6==6.8.2.1
transformers==4.48.3
# CogAgent
timm==1.0.14
# CogVLM
einops==0.8.1
protobuf==5.29.3
sentencepiece==0.2.0
torchvision==0.21.0
xformers==0.0.29.post3
# InternLM-XComposer2
gptqmodel==1.9.0
numpy==2.2.3
# WD Tagger
huggingface-hub==0.29.1
onnxruntime==1.20.1
# PyTorch
torch==2.6.0; platform_system != "Windows"
https://download.pytorch.org/whl/cu124/torch-2.6.0%2Bcu124-cp312-cp312-win_amd64.whl; platform_system == "Windows" and python_version == "3.12"
https://download.pytorch.org/whl/cu124/torch-2.6.0%2Bcu124-cp311-cp311-win_amd64.whl; platform_system == "Windows" and python_version == "3.11"
# FlashAttention (Florence-2, Phi-3-Vision)
flash-attn==2.7.4.post1; platform_system == "Linux"
https://github.com/kingbri1/flash-attention/releases/download/v2.7.4.post1/flash_attn-2.7.4.post1+cu124torch2.6.0cxx11abiFALSE-cp312-cp312-win_amd64.whl; platform_system == "Windows" and python_version == "3.12"
https://github.com/kingbri1/flash-attention/releases/download/v2.7.4.post1/flash_attn-2.7.4.post1+cu124torch2.6.0cxx11abiFALSE-cp311-cp311-win_amd64.whl; platform_system == "Windows" and python_version == "3.11"