So far, running LLMs has required a large amount of computing resources, mainly GPUs. Running locally, a simple prompt with a typical LLM takes on an average Mac ...
This repository is adapted from here. We are grateful for the work of the developer on this repo. Sadly, that repository is inactive. Thus, a fork was moved here in order to allow for more visibility ...