Hacker Newsnew | past | comments | ask | show | jobs | submitlogin
Show HN: Cerebras-GPT-2.7B finetuned on Stanford Alpaca dataset (github.com/lxe)
4 points by lxe on March 31, 2023 | hide | past | favorite | 1 comment
Yes I know, this is yet another language model. I used PEFT to finetune the Cerebras-GPT-2.7B on Alpaca, which makes for a very very fast, coherent, albeit hallucinating model.

It took 5 hours on a vast.ai machine at 29 cents an hour, so less than a dollar and a half to finetune.

The repository contains steps to reproduce the model, the dataset, links to the merged-in model on huggingface, a colab notebook, and the LoRA weights.

There's also a chatbot app, and a link to a ggml version of the model you can run via https://github.com/ggerganov/ggml on your CPU alone!

Enjoy!



Keep in mind, this tends to hallucinate a lot. I will attempt to debug this behavior in the future.




Consider applying for YC's Fall 2025 batch! Applications are open till Aug 4

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: