Shimmy is a 4.8MB single-binary that provides 100% OpenAI-compatible endpoints for GGUF models.
Read more 
			
		
	The Linux Portal Site
 
			
		
	Shimmy is a 4.8MB single-binary that provides 100% OpenAI-compatible endpoints for GGUF models.
Read more 
			
		
	Gerbil is a desktop app to easily run Large Language Models locally. The software is powered by KoboldCpp, a fork of llama.cpp.
Read more 
			
		
	Gerbil lets you run LLMs locally powered by KoboldCpp which itself is a highly modified fork of llama.cpp.
Read more 
			
		
	Elelem is a simple LLM client that connects seamlessly with OpenAI API compatible services.
Read more 
			
		
	Did you know that even a Raspberry Pi 5 is able to run LLMs locally? Even better, the developer of Pi-Apps has produced an optimized install script which makes it Raspberry Pi 5 friendly.
Read more 
			
		
	Ollama is software in an early stage of development that lets you run and chat with Llama 2 and other models.
Read more