v0.5.0
Changes
- llama.cpp updated to b1132, GGUF format support and increase in the speed. The old file format is still supported but uses llama dadbed9.
- Add Falcon models support (only GGUF)
- Add template for RWKV-4
- Fix model rename
- Fixed some UI bugs that could cause the app to crash.
- Fix llama, replit token_to_str
** In order to use llama.cpp b1132, the model file must have the .gguf extension.
*** Unfortunately, due to a bug in the latest versions of llama.cpp, Metal not supported on intel Macs at this time.