-
Notifications
You must be signed in to change notification settings - Fork 415
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
please upload windows install protable version #19
Comments
Amazing project. I try it on Windows 1 clic before putting a GPU inside my Debian server but I get :
|
I think since it uses flash-attention you need WSL. Its a lot slower than they said, took like 20 minutes for 1 minute of music and lyrics while using 18gb of vram. |
It appears that your Python module is 3.9, but this script uses Python 3.10. |
I now include YuE with Visions of Chaos along with hundreds of other AI systems if anyone wants to have an eaiser install on Windows. |
Too bad it requires 3.10.x as I just moved off of that about two weeks ago for a few of my programs. |
If you mean Visions of Chaos you can point it to a 3.10 directory so it uses 3.10 while your other programs use 3.12 |
yeah, I will just wait until it is updated. If it were in a venv pip install environment, I would do it. Still, want a way to get this multimodal to work on windows by itself for now. |
UV will automatically install version 3.10x, you do not need to download Python, it will install it automatically |
To activate the ML section it said I had to do it, and since I didn't create a venv for it, I uninstalled it since I have had other Python programs zap my system-wide environment (once bitten twice shy). I did manage to get the MM working anyway. |
Understood. I have had other installs kill my main python env too without asking. That is why every mode in Visions of Chaos creates their own virtual envs to run under. You basically have to do that anyway when dealing with so many different systems that all need their own unique versions of Python packages. If you ever want to try it again join my Discord and I can give you some tips about keeping the version of Python and CUDA VoC uses separate from your defaults (all you have to do is install the VoC required versions, not add them to your path, and then point to the install folders in VoC, it handles the rest). |
Will do. My hope is YuE gets a much-needed optimization because it needs it. Took about 20m for 90s of music. Music was of poor quality, but a 64kbps mp3 didn't help that. I am so glad this is finally hitting local, but not only do we need it optimized so it doesn't take lifetimes, but we need a way to train it so it can learn different styles. For instance, even UDIO struggles with my main style even if I upload it an example. Now if I could train it then it would no longer struggle. |
https://huggingface.co/onnx-community/YuE-s2-1B-general-ONNX stage2 can use ONNX for faster |
q4 too. |
How do we use this model? |
When I last used Onnx I had to compile it as it is compiled. As I was told Onnx cannot be universal it must be compiled on the machine it is to be ran on from source files. I do forget what the source format was, but was super easy to compile, but I see no files we can compile from. When I last used Onnx it was for video upscalers that I found the time to initialize was longer than the savings and ditched Onnx. That would not be the case for this since it takes so long for Stage2. |
Ok. I thought your comment was that you had used the onyx model for a speedup in stage 2. |
No, I was letting you know something is amiss because we need to compile that into Onnx as Onnx is machine dependant, and unless something changed my Onnx only works for machines that match my specs exactly. Compiling them is really fast too on a 4090. I think the one linking the Onnx is just trying to let us know it can be done not that those can be used. Besides, I know in ComfyUI it takes a special node to load Onnx models, and I expect that to be a universal case that the code will need to be updated to take advantage of it. Oh, another thing I really think will help is Sage Attention as it blew away SDPA, and FA2, in Linux. In Windows I don't see much of an advantage, but with this who knows. Needs to be coded for that as well. What I am basically saying is that this is ripe to be ravished to optimize and I bet we see 10x speed up, at least. Maybe 2m for a complete song on a 4090 instead of 20-30m. |
please upload windows install protable version or one click install version..thanks
The text was updated successfully, but these errors were encountered: