You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I'd tell a joke since maintainers don't read this.
User - Here'sabugwithyourpubliccodethatyouwantpeopletouse.Author - Nota bug (Userstopsreadingafterthisandeitherabusestheauthor'sreplyoutofunderstandablefrustrationandangerordoesitinhisheadandpretendstobenice)
User - Ok. (may not even repsond or have notification allowed for this interaction)
(No response from User after years, but User doesn't use the software anymore uninstalled it.
theyfoundanalternative, projectlostoneuserandeveryoneelsewhohadthesameissuebutdidn'thaveaGithubaccountorthosewhojustfoundtheresultonGithuboronasearchenginelikeGoogleorthroughPerplexity)
(Status - thebugisstillpresentbutnotfixed. People ignore it then delete the code, download
statskeepincreasing. Author is proud of his work! Welcome to Github flow! ;))
GitHub discussions are not often looked at by users and maintainers.
And smart people or lazy people or "smart and lazy" / "smart but lazy" and security focused people don't want to use another website for one single question or poll or issue. Too many costs, too many risks, too many headaches, not enough value.
If you can build a positive reputation of providing great value to the community.
Also I'm talking about all cases of forums. Here we go -
This application will be able to read your private email addresses.
They are private because I only have one personal email, it's secure, has only useful subscription that is never a single spam or promo or newsletter or podcast or free gift or any other marketing or other BS. Very anti privacy, anti security, inconvenient - if I wanted to talk to maintainers only.
TLDR or AI gen summary - Read the joke first. Then you'll understand the other side of the story instead.
What is this vllm thing actually useful for in the real world? What are the benefits over the alternatives? What are the performance profiles? Can you give RAM type(which spec and ECC or no ECC?) and RAM size and CPU core count and GPU VRAM and GPU CUDA core count? Can it handle high level GPU specific assembly level code gen correctly for multiple Nvidia GPU arch's? What is the design philosophy?
Why are the devs on Slack? Most only use Slack for 9 to 5 jobs, so there's a clear essential but soul draining vibe and sentiment to it. Is there a traditional mailing list for this with a digest feed - RSS /ATOM and OPML if the last is practical and has safe implementation across OS platforms? Does it build in one step and run smoothly on Arch (I mean with AUR) like distro or FreeBSD family or Haiku?
reacted with thumbs up emoji reacted with thumbs down emoji reacted with laugh emoji reacted with hooray emoji reacted with confused emoji reacted with heart emoji reacted with rocket emoji reacted with eyes emoji
Uh oh!
There was an error while loading. Please reload this page.
-
I'd tell a joke since maintainers don't read this.
And smart people or lazy people or "smart and lazy" / "smart but lazy" and security focused people don't want to use another website for one single question or poll or issue. Too many costs, too many risks, too many headaches, not enough value.
If you can build a positive reputation of providing great value to the community.
Also I'm talking about all cases of forums. Here we go -
They are private because I only have one personal email, it's secure, has only useful subscription that is never a single spam or promo or newsletter or podcast or free gift or any other marketing or other BS. Very anti privacy, anti security, inconvenient - if I wanted to talk to maintainers only.
TLDR or AI gen summary - Read the joke first. Then you'll understand the other side of the story instead.
Also, why would I use vLLM?
The above was an old post from asd
What is this vllm thing actually useful for in the real world? What are the benefits over the alternatives? What are the performance profiles? Can you give RAM type(which spec and ECC or no ECC?) and RAM size and CPU core count and GPU VRAM and GPU CUDA core count? Can it handle high level GPU specific assembly level code gen correctly for multiple Nvidia GPU arch's? What is the design philosophy?
Why are the devs on Slack? Most only use Slack for 9 to 5 jobs, so there's a clear essential but soul draining vibe and sentiment to it. Is there a traditional mailing list for this with a digest feed - RSS /ATOM and OPML if the last is practical and has safe implementation across OS platforms? Does it build in one step and run smoothly on Arch (I mean with AUR) like distro or FreeBSD family or Haiku?
Beta Was this translation helpful? Give feedback.
All reactions