Has anybody gotten this running on consumer hardware ala llama or is that not in the cards?
Experimental Falcon inference via ggml (so on CPU): https://github.com/cmp-nct/ggllm.cpp
It has problems but it does work
Has anybody gotten this running on consumer hardware ala llama or is that not in the cards?
It has problems but it does work