Hacker News
new
|
past
|
comments
|
ask
|
show
|
jobs
|
submit
login
StochasticLi
10 months ago
|
parent
|
context
|
favorite
| on:
Life of an inference request (vLLM V1): How LLMs a...
I would like to know what inference speeds they are achieving exactly on what hardware. I skimmed and searched the article and didn't find that info.
Consider applying for YC's Summer 2026 batch! Applications are open till May 4
Guidelines
|
FAQ
|
Lists
|
API
|
Security
|
Legal
|
Apply to YC
|
Contact
Search: