Has anybody gotten this running on consumer hardware ala llama or is that not in the cards?

Experimental Falcon inference via ggml (so on CPU): https://github.com/cmp-nct/ggllm.cpp

It has problems but it does work