diff --git a/README.md b/README.md index e667293..0d3c2ed 100644 --- a/README.md +++ b/README.md @@ -46,13 +46,13 @@ It has been tested on following models: | Llama | [Llama2 7B](https://huggingface.co/meta-llama/Llama-2-7b-chat-hf), [Llama2 13B](https://huggingface.co/meta-llama/Llama-2-13b-chat-hf), [Llama3 8B](https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct) | | Llama-like | [TinyLlama 1.1B](https://huggingface.co/TinyLlama/TinyLlama-1.1B-Chat-v1.0), [Cosmo 1B](https://huggingface.co/HuggingFaceTB/cosmo-1b), [LLaMA Pro 8B](https://huggingface.co/TencentARC/LLaMA-Pro-8B-Instruct), [H2O Danube 1.8B](https://huggingface.co/h2oai/h2o-danube-1.8b-chat), [DeepSeekMath 7B](https://huggingface.co/deepseek-ai/deepseek-math-7b-instruct), [LargeWorldModel 7B 1M](https://huggingface.co/LargeWorldModel/LWM-Text-Chat-1M), [Xverse 7B](https://huggingface.co/xverse/XVERSE-7B-Chat), [LLM360 K2](https://huggingface.co/LLM360/K2-Chat) | | Yi | [Yi 1.5 6B](https://huggingface.co/01-ai/Yi-1.5-6B-Chat), [Yi 1.5 9B](https://huggingface.co/01-ai/Yi-1.5-9B-Chat), [Yi 1.5 34B](https://huggingface.co/01-ai/Yi-1.5-34B-Chat) | -| Mistral | [Mistral 7B](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.3), (Codestral 22B)[https://huggingface.co/mistralai/Codestral-22B-v0.1], [Mistral Pro 8B](https://huggingface.co/TencentARC/Mistral_Pro_8B_v0.1), [SOLAR 10.7B](https://huggingface.co/upstage/SOLAR-10.7B-Instruct-v1.0), [GritLM 7B](https://huggingface.co/GritLM/GritLM-7B), [Starling 7B](https://huggingface.co/Nexusflow/Starling-LM-7B-beta) | +| Mistral | [Mistral 7B](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.3), [Codestral 22B](https://huggingface.co/mistralai/Codestral-22B-v0.1), [Mistral Pro 8B](https://huggingface.co/TencentARC/Mistral_Pro_8B_v0.1), [SOLAR 10.7B](https://huggingface.co/upstage/SOLAR-10.7B-Instruct-v1.0), [GritLM 7B](https://huggingface.co/GritLM/GritLM-7B), [Starling 7B](https://huggingface.co/Nexusflow/Starling-LM-7B-beta) | | Qwen2 | [Qwen1.5 0.5B](https://huggingface.co/Qwen/Qwen1.5-0.5B), [Qwen1.5 1.8B](https://huggingface.co/Qwen/Qwen1.5-1.8B), [Qwen1.5 4B](https://huggingface.co/Qwen/Qwen1.5-4B), [Qwen1.5 7B](https://huggingface.co/Qwen/Qwen1.5-7B), [Qwen1.5 14B](https://huggingface.co/Qwen/Qwen1.5-14B), [Qwen2 0.5B](https://huggingface.co/Qwen/Qwen2-0.5B-Instruct), [Qwen2 1.5B](https://huggingface.co/Qwen/Qwen2-1.5B-Instruct), [Qwen2 7B](https://huggingface.co/Qwen/Qwen2-7B-Instruct) | | Mixtral | [Mixtral 8x7B](https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1), [Mixtral 8x22B](https://huggingface.co/mistralai/Mixtral-8x22B-Instruct-v0.1), [GritLM 8x7B](https://huggingface.co/GritLM/GritLM-8x7B) | | OLMo | [OLMo 1B](https://huggingface.co/allenai/OLMo-1B), [OLMo 7B](https://huggingface.co/allenai/OLMo-7B), [OLMo 1.7 7B](https://huggingface.co/allenai/OLMo-1.7-7B) | | Gemma | [Gemma 2B](https://huggingface.co/google/gemma-2b-it), [Gemma 7B](https://huggingface.co/google/gemma-7b-it) (*note: 7B version has issues with fp8 quantization*) | | MiniCPM | [MiniCPM 2B](https://huggingface.co/openbmb/MiniCPM-2B-dpo-bf16), [MiniCPM 2B 128K](https://huggingface.co/openbmb/MiniCPM-2B-128k), [MiniCPM MoE 8x2B](https://huggingface.co/openbmb/MiniCPM-MoE-8x2B) | -| Cohere | [Command-R](https://huggingface.co/CohereForAI/c4ai-command-r-v01), [Aya 23 8B](https://huggingface.co/CohereForAI/aya-23-8B), [Aya 23 35B] (https://huggingface.co/CohereForAI/aya-23-35B) | +| Cohere | [Command-R](https://huggingface.co/CohereForAI/c4ai-command-r-v01), [Aya 23 8B](https://huggingface.co/CohereForAI/aya-23-8B), [Aya 23 35B](https://huggingface.co/CohereForAI/aya-23-35B) | | InternLM | [InternLM2-1.8B](https://huggingface.co/internlm/internlm2-1_8b), [InternLM2-7B](https://huggingface.co/internlm/internlm2-7b), [InternLM2-20B](https://huggingface.co/internlm/internlm2-20b) | | DBRX | [DBRX 132B](https://huggingface.co/databricks/dbrx-instruct) | | Phi3 | [Phi3 Mini 3.8B](https://huggingface.co/microsoft/Phi-3-mini-4k-instruct/), [Phi3 Medium 14B](https://huggingface.co/microsoft/Phi-3-medium-4k-instruct) |