What does HackerNews think of alpaca-lora?

Instruct-tune LLaMA on consumer hardware

Language: Jupyter Notebook

There are tons of them for other models, often Alpaca or LLaMA.

This is a popular tool for finetuning Alpaca:

https://github.com/tloen/alpaca-lora

Not just inference.

AFAIK, you are able to fine-tune the models with custom data[1], which does not seem to require anything but a GPU with enough VRAM to fit the model in question. I'm looking to get my hands on an RTX 4090 to ingest all of the repair manuals of a certain company and have a chatbot capable of guiding repairs, or at least try to do so. So far doing inference only as well.

[1] https://github.com/tloen/alpaca-lora