Skip to content

Pull requests: HabanaAI/vllm-hpu-extension

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Reviews
Assignee
Filter by who’s assigned
Sort

Pull requests list

Power of 2 padding fix
#97 opened Feb 14, 2025 by mfylcek Loading…
Add Mixtral and Llama chat templates
#95 opened Feb 13, 2025 by tpawlows Loading…
Chunked Prefill Enablement
#94 opened Feb 13, 2025 by hlahkar Draft
Expand capability checks
#89 opened Feb 3, 2025 by kzawora-intel Loading…
Support the initial enablement for text-embedding
#87 opened Jan 30, 2025 by libinta Loading…
Add exponential bucketing PoC
#61 opened Dec 17, 2024 by kzawora-intel Loading…
vLLM-Ext: Full enabling of ALiBi
#60 opened Dec 17, 2024 by tannervoas742 Loading…
Remove vllm.logger.init_logger dependency
#53 opened Dec 9, 2024 by kzawora-intel Loading…
ProTip! What’s not been updated in a month: updated:<2025-01-14.