What does HackerNews think of alpaca-lora?
Instruct-tune LLaMA on consumer hardware
Language:
Jupyter Notebook
There are tons of them for other models, often Alpaca or LLaMA.
This is a popular tool for finetuning Alpaca:
Not just inference.
AFAIK, you are able to fine-tune the models with custom data[1], which does not seem to require anything but a GPU with enough VRAM to fit the model in question. I'm looking to get my hands on an RTX 4090 to ingest all of the repair manuals of a certain company and have a chatbot capable of guiding repairs, or at least try to do so. So far doing inference only as well.
Use this Alpaca replication instead