Hacker News
new
|
past
|
comments
|
ask
|
show
|
jobs
|
submit
login
sterlind
on Aug 10, 2023
|
parent
|
context
|
favorite
| on:
Ask HN: Cheapest hardware to run Llama 2 70B
SLI isn't used at all for CUDA. if you meant NVLink, it's apparently not useful at small scales - I think the PCIe lanes are enough.
ipsum2
on Aug 18, 2023
[–]
This is wrong, NVLink is crucial for tensor parallelism in models for training and in large (>40B param) models for inference.
Guidelines
|
FAQ
|
Lists
|
API
|
Security
|
Legal
|
Apply to YC
|
Contact
Search: