A cross-platform command-line interface for discovering, downloading, and running language models from Hugging Face.
- Model Discovery: Fetch and browse Kolosal models directly from Hugging Face.
- Interactive TUI: Navigate, search, and select models with an easy-to-use terminal interface.
- Download & Run: Download GGUF models and run them locally with a built-in inference server.
- Chat Interface: Interact directly with loaded models through a command-line chat.
- Smart Caching: Reduces API calls and improves performance with intelligent caching.
- Cross-Platform: Fully supported on Windows, macOS, and Linux.
The easiest way to get started is to download a pre-built binary for your operating system from the Releases page.
Once installed, you can run the application from your terminal:
# Launch the interactive model browser
kolosal-cli
# Stop the background inference server
kolosal-cli --stop-server
If you prefer to build from source, ensure you have CMake (3.14+) and a C++17 compiler installed.
git clone https://github.com/KolosalAI/kolosal-cli.git
cd kolosal-cli
-
Linux / macOS:
mkdir build && cd build cmake .. -DCMAKE_BUILD_TYPE=Release make -j$(nproc) # Linux # make -j$(sysctl -n hw.ncpu) # macOS
-
Windows:
mkdir build; cd build cmake .. cmake --build . --config Release
-
Linux / macOS:
./kolosal-cli
-
Windows:
.\Release\kolosal-cli.exe