HN2
new
|
past
|
comments
|
ask
|
show
|
jobs
|
submit
login
kaszanka
on March 3, 2023
|
parent
|
context
|
favorite
| on:
Facebook LLAMA is being openly distributed via tor...
Here is the magnet link for posterity: magnet:?xt=urn:btih:ZXXDAUWYLRUXXBHUYEMS6Q5CE5WA3LVA&dn=LLaMA
psychphysic
on March 3, 2023
|
next
[–]
Thanks not working for me...
Not that I could run it if I downloaded it.
q1w2
on March 3, 2023
|
prev
[–]
Great, now how do I run it? Do I need a GPU with over 65GB RAM?
version_five
on March 3, 2023
|
parent
|
next
[–]
Try this, it's for running llms that won't fit in the gpu:
https://github.com/FMInference/FlexGen
gpm
on March 3, 2023
|
root
|
parent
|
next
[–]
Currently that looks like it only supports facebook's opt and galactica models. Though they do appear to plan to add support for more models.
rnosov
on March 3, 2023
|
parent
|
prev
|
next
[–]
Generally, you'll need multiply model size by two to get required amount of video RAM. There are 4 sizes, so you might get away with even smaller GPU for say 13B model.
bioemerl
on March 3, 2023
|
parent
|
prev
[–]
Nope, more like 111gb
Guidelines
|
FAQ
|
Lists
|
API
|
Security
|
Legal
|
Apply to YC
|
Contact
Search: