diff --git a/README.md b/README.md index c9a9300..e10fccd 100644 --- a/README.md +++ b/README.md @@ -15,21 +15,19 @@ Also, when creating the application, the source codes from the repository [byron List of supported models: -| model | inference | size | quantized link | iOS (iphone 12 pro max) | MacOS | -|------------------------------------------------------------------------------------|-----------|------|-------------------------------------------------------------------------------------------------------------|-------------------------|-------| -| [OpenLLaMa](https://github.com/openlm-research/open_llama) | LLaMA | 3B | [hug](https://huggingface.co/guinmoon/LLMFarm_Models/resolve/main/open-llama-3b-1T_q5_1.bin) | ok | ok | -| [ORCA](https://huggingface.co/psmathur/orca_mini_3b) | LLaMA | 3B | [hug](https://huggingface.co/guinmoon/LLMFarm_Models/resolve/main/orca-mini-3b.ggmlv3.q5_1.bin) | ok | ok | -| [StableLM Tuned Alpha](https://huggingface.co/stabilityai/stablelm-tuned-alpha-3b) | GPT-NeoX | 3B | [hug](https://huggingface.co/guinmoon/LLMFarm_Models/resolve/main/stablelm-tuned-alpha-3b-ggml_v3-q5_1.bin) | ok | ok | -| [Dolly v2](https://github.com/databrickslabs/dolly) | GPT-NeoX | 3B | [hug](https://huggingface.co/guinmoon/LLMFarm_Models/resolve/main/dolly-v2-3b-ggml_v3-q5_1.bin) | ok | ok | -| [RedPajama](https://huggingface.co/togethercomputer/RedPajama-INCITE-Base-3B-v1) | GPT-NeoX | 3B | [hug](https://huggingface.co/guinmoon/LLMFarm_Models/resolve/main/rp-incite-base-v1-3b-ggmlv3-q5_1.bin) | ok | ok | -| [Pythia](https://huggingface.co/EleutherAI) | GPT-NeoX | 2.8B | [hug](https://huggingface.co/guinmoon/LLMFarm_Models/resolve/main/pythia-2.8b-ggml_v3-q5_1.bin) | ok | ok | -| [Llama](https://arxiv.org/abs/2302.13971) | LLaMA | 7B | | bad alloc | ok | -| [Alpaca](https://crfm.stanford.edu/2023/03/13/alpaca.html) | LLaMA | 7B | | bad alloc | ok | -| [Vicuna](https://lmsys.org/blog/2023-03-30-vicuna/) | LLaMA | 7B | | bad alloc | ok | -| [Vicuna](https://lmsys.org/blog/2023-03-30-vicuna/) | LLaMA | 13B | [hug](https://huggingface.co/vicuna/ggml-vicuna-13b-1.1/resolve/main/ggml-vic13b-uncensored-q4_1.bin) | bad alloc | ok | -| [WizardLM](https://arxiv.org/abs/2304.12244) | LLaMA | 7B | | bad alloc | ok | -| [Cerebras](https://huggingface.co/cerebras/Cerebras-GPT-2.7B) | GPT-2 | 2.7B | | ok | ok | -| [RWKV-4 "Raven"](https://huggingface.co/BlinkDL/rwkv-4-raven) | RWKV | 3B | | in dev | ok | +| model | inference | size | quantized link | iOS (iphone 12 pro max) | MacOS | +|------------------------------------------------------------------------------------|-----------|----------|-------------------------------------------------------------------------------------------------------------|-------------------------|-------| +| [OpenLLaMa](https://github.com/openlm-research/open_llama) | LLaMA | 3B | [hug](https://huggingface.co/guinmoon/LLMFarm_Models/resolve/main/open-llama-3b-1T_q5_1.bin) | ok | ok | +| [ORCA](https://huggingface.co/psmathur/orca_mini_3b) | LLaMA | 3B | [hug](https://huggingface.co/guinmoon/LLMFarm_Models/resolve/main/orca-mini-3b.ggmlv3.q5_1.bin) | ok | ok | +| [ORCA](https://huggingface.co/TheBloke/orca_mini_7B-GGML/) | LLaMA | 7B(q2_K) | [hug](https://huggingface.co/TheBloke/orca_mini_7B-GGML/resolve/main/orca-mini-7b.ggmlv3.q2_K.bin) | ok | ok | +| [StableLM Tuned Alpha](https://huggingface.co/stabilityai/stablelm-tuned-alpha-3b) | GPT-NeoX | 3B | [hug](https://huggingface.co/guinmoon/LLMFarm_Models/resolve/main/stablelm-tuned-alpha-3b-ggml_v3-q5_1.bin) | ok | ok | +| [Dolly v2](https://github.com/databrickslabs/dolly) | GPT-NeoX | 3B | [hug](https://huggingface.co/guinmoon/LLMFarm_Models/resolve/main/dolly-v2-3b-ggml_v3-q5_1.bin) | ok | ok | +| [RedPajama](https://huggingface.co/togethercomputer/RedPajama-INCITE-Base-3B-v1) | GPT-NeoX | 3B | [hug](https://huggingface.co/guinmoon/LLMFarm_Models/resolve/main/rp-incite-base-v1-3b-ggmlv3-q5_1.bin) | ok | ok | +| [Pythia](https://huggingface.co/EleutherAI) | GPT-NeoX | 2.8B | [hug](https://huggingface.co/guinmoon/LLMFarm_Models/resolve/main/pythia-2.8b-ggml_v3-q5_1.bin) | ok | ok | +| [Llama](https://arxiv.org/abs/2302.13971) | LLaMA | 7B | | bad alloc | ok | +| [WizardLM](https://arxiv.org/abs/2304.12244) | LLaMA | 7B | | bad alloc | ok | +| [Cerebras](https://huggingface.co/cerebras/Cerebras-GPT-2.7B) | GPT-2 | 2.7B | | ok | ok | +| [RWKV-4 "Raven"](https://huggingface.co/BlinkDL/rwkv-4-raven) | RWKV | 3B | | in dev | ok | Models must be maintained by [ggml](https://github.com/ggerganov/ggml) after [#154](https://github.com/ggerganov/ggml/pull/154).