Replies: 5 comments 7 replies
-
|
Beta Was this translation helpful? Give feedback.
-
Hold my Pepsi... you have only WHL :-D triton-3.0.0-cp311-cp311-win_amd64.whl Download here: https://1drv.ms/u/s!Aq5X0D4OWWATiKkx8gdY012BCa_wVQ?e=SaCr0l What is so good on Triton anyway? They just rape vLLM. You should check out SGLang, very nice. I'm building my own OpenAI API gateway that will feature plug n play models and cover functionality of Ollama, LT Studio, OpenVINO and will feature nice dashboard with statistics and some Red Team System Prompts muhehehe. |
Beta Was this translation helpful? Give feedback.
-
Closing this discussion due to age and Windows whl having been published with version 0.15.0 on PyPI. |
Beta Was this translation helpful? Give feedback.
-
0.15.0 is half year old and for recent Pytorch and Microsoft and Intel AI toolkits, Cuda 12.4 and Python 3.12 build is required. On top of that, that WHL:
You can close all other Windows discussions due to age as well - where you mostly promise Windows build. This Deepseek is useless in current implementation. I will do it properly myself. Compile all Deepseek dependecies and then compile Deepseek with all the features. I bought i9 Core Ultra 285K + RTX 4090 for $6000 and I even did not get to developing AI yet. Only preparing build system and libraries so something basic can run properly. The cost of the time I need to spend is getting high, so I might be a bit grumpy these days :) |
Beta Was this translation helpful? Give feedback.
-
Try to use Accelerate in Pytorch and select DeepSpeef using command:
accelerate config, it will crash due to AIO not present.
Right now I have 0.16.1 for Python 3.12
I dont want to waste your time, I am able to solve it myself and you know
the behavior at NVIDIA github is even worse because on website or in docs
they say cuFile.cu is in Cuda Toolkit, but it is only in Linux package, I
observe it for like 5 years and some people even wont get response to this
question. Our EU court cares more about citizens embracing social
capitalism. Google is here quite often.
Maybe someone will put NV to EU court. But I dont care, I only came here to
and to other companies to asses the situation. NV will be the biggest
problem, you and them could be represented in few Windows benchmarks that
you outperform competition and some Ph.d would also present positive
review, but after today I am not in that mood. I will just compile and pack
everything for Windows 11 community and then I wont care. Regarding missing
cuFile from NVIDIA, Copilot Pro said that you can load Linux .so shared
library in Windows Python using ctypes or cffi.
After this experience, I think the situation deserves small escalation and
more proactivity, so I think someone will reverse engineer or decompile
cuFile.so and make cuFile.dll - corporation arguments are just nonsense,
and either laziness or arrogancy. Copilot Pro said that it is ok to
decompile it, but it requires more skills. Im like Steve Jobs attitude, so
I will make it happen, it will take some time, but I will find someone with
good skill with Gihidra or IDA Pro.
In the end I will solve it for a lot complaining people on three american
github corporations and it will be a win win. Everyone gets peaceful life
and I can finally run some Pytorch fun, just spent $6000 on new Win AI pc
and only thing Im doing is solving someone else issues.
So please dont let yourself stress, customers will solve it themselves in
accordance to European Union law.
…On Wed 22. 1. 2025 at 18:08 Logan Adams ***@***.***> wrote:
@cyberluke <https://github.com/cyberluke> - I want to try and help with
your feedback/comments above, but the 0.15.0 whl appears to be compiled
fine for me. I'm able to run a few test cases in DeepSpeedExamples. Or are
you referring to not having support built into that whl for DS_BUILD_AIO?
Could you also list, other than having newer whls built, what you'd like
to see in them? I believe the 0.15.0 whl is built with python 3.11, I know
some users have requested 3.10, are you looking for older or newer python
versions? Same with Cuda versions?
—
Reply to this email directly, view it on GitHub
<#2694 (reply in thread)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/AAIPKTZGIPB5UTNERTUOISD2L7GBPAVCNFSM6AAAAABVK3T4TKVHI2DSMVQWIX3LMV43URDJONRXK43TNFXW4Q3PNVWWK3TUHMYTCOJRHE2DAMI>
.
You are receiving this because you were mentioned.Message ID:
***@***.***>
|
Beta Was this translation helpful? Give feedback.
-
this gets us one step closer to a working windows version but it's still erroring out during install bc it's looking for lscpu info instead of Windows. I'm not totally inept at coding but my skills are subpar at best if anyone else feels like taking a crack at getting it installed. I know someone managed to get the build whl built but I have no idea where I saw it anymore... here's the triton 2.0.0 download. it didn't trigger any flags when scanning in Defender https://transfer.sh/me0xpC/triton-2.0.0-cp310-cp310-win_amd64.whl just download the file. activate whatever virtual environment and pip install the filename from your downloads folder
Beta Was this translation helpful? Give feedback.
All reactions