Hacker Newsnew | past | comments | ask | show | jobs | submit | funtech's commentslogin

Is 12GB the minimum? got an out of memory error with 8GB


Yes, unfortunately these models take a lot of VRAM. It may be possible to do an 8GB version but it will have to compromise on quality of voice recognition and the language model so it might not be a good experience.


This might be silly because of how few people it benefits, but could it be broken up on to multiple 8GB cards on the same system?


Yes, it absolutely could. You're right that this configuration is rare. Although people have been putting together machines with multiple 24GB cards in order to split and run larger models like llama2-70B.


The latest large models are 120B and 100k context such as Goliath and Tess XL


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: