'mistral-nemo':_("A state-of-the-art 12B model with 128k context length, built by Mistral AI in collaboration with NVIDIA."),
'mistral-large':_("Mistral Large 2 is Mistral's new flagship model that is significantly more capable in code generation, mathematics, and reasoning with 128k context window and support for dozens of languages."),
'qwen2':_("Qwen2 is a new series of large language models from Alibaba group"),
'deepseek-coder-v2':_("An open-source Mixture-of-Experts code language model that achieves performance comparable to GPT4-Turbo in code-specific tasks."),
'phi3':_("Phi-3 is a family of lightweight 3B (Mini) and 14B (Medium) state-of-the-art open models by Microsoft."),
'mistral':_("The 7B model released by Mistral AI, updated to version 0.3."),
'mixtral':_("A set of Mixture of Experts (MoE) model with open weights by Mistral AI in 8x7b and 8x22b parameter sizes."),
'codegemma':_("CodeGemma is a collection of powerful, lightweight models that can perform a variety of coding tasks like fill-in-the-middle code completion, code generation, natural language understanding, mathematical reasoning, and instruction following."),
'command-r':_("Command R is a Large Language Model optimized for conversational interaction and long context tasks."),
'command-r-plus':_("Command R+ is a powerful, scalable large language model purpose-built to excel at real-world enterprise use cases."),
'llava':_("🌋 LLaVA is a novel end-to-end trained large multimodal model that combines a vision encoder and Vicuna for general-purpose visual and language understanding. Updated to version 1.6."),
'gemma':_("Gemma is a family of lightweight, state-of-the-art open models built by Google DeepMind. Updated to version 1.1"),
'qwen':_("Qwen 1.5 is a series of large language models by Alibaba Cloud spanning from 0.5B to 110B parameters"),
'llama2':_("Llama 2 is a collection of foundation language models ranging from 7B to 70B parameters."),
'codellama':_("A large language model that can use text prompts to generate and discuss code."),
'dolphin-mixtral':_("Uncensored, 8x7b and 8x22b fine-tuned models based on the Mixtral mixture of experts models that excels at coding tasks. Created by Eric Hartford."),
'dolphin-mistral':_("The uncensored Dolphin model based on Mistral that excels at coding tasks. Updated to version 2.8."),
'orca-mini':_("A general-purpose model ranging from 3 billion parameters to 70 billion, suitable for entry-level hardware."),
'dolphin-llama3':_("Dolphin 2.9 is a new model with 8B and 70B sizes by Eric Hartford based on Llama 3 that has a variety of instruction, conversational, and coding skills."),
'starcoder':_("StarCoder is a code generation model trained on 80+ programming languages."),
'wizardlm2':_("State of the art large language model from Microsoft AI with improved performance on complex chat, multilingual, reasoning and agent use cases."),
'openchat':_("A family of open-source models trained on a wide variety of data, surpassing ChatGPT on various benchmarks. Updated to version 3.5-0106."),
'stable-code':_("Stable Code 3B is a coding model with instruct and code completion variants on par with models such as Code Llama 7B that are 2.5x larger."),
'stablelm2':_("Stable LM 2 is a state-of-the-art 1.6B and 12B parameter language model trained on multilingual data in English, Spanish, German, Italian, French, Portuguese, and Dutch."),
'llama3-chatqa':_("A model from NVIDIA based on Llama 3 that excels at conversational question answering (QA) and retrieval-augmented generation (RAG)."),
'orca2':_("Orca 2 is built by Microsoft research, and are a fine-tuned version of Meta's Llama 2 models. The model is designed to excel particularly in reasoning."),
'medllama2':_("Fine-tuned Llama 2 model to answer medical questions based on an open source medical dataset."),
'yarn-mistral':_("An extension of Mistral to support context windows of 64K or 128K."),
'llama-pro':_("An expansion of Llama 2 that specializes in integrating both general language understanding and domain-specific knowledge, particularly in programming and mathematics."),
'magicoder':_("🎩 Magicoder is a family of 7B parameter models trained on 75K synthetic instruction data using OSS-Instruct, a novel approach to enlightening LLMs with open-source code snippets."),
'stablelm-zephyr':_("A lightweight chat model allowing accurate, and responsive output without requiring high-end hardware."),
'codebooga':_("A high-performing code instruct model created by merging two existing code models."),
'mistrallite':_("MistralLite is a fine-tuned model based on Mistral with enhanced capabilities of processing long contexts."),
'llama3-groq-tool-use':_("A series of models from Groq that represent a significant advancement in open-source AI capabilities for tool use/function calling."),
'mathstral':_("MathΣtral: a 7B model designed for math reasoning and scientific discovery by Mistral AI."),
'firefunction-v2':_("An open weights function calling model based on Llama 3, competitive with GPT-4o function calling capabilities."),
'nuextract':_("A 3.8B model fine-tuned on a private high-quality synthetic dataset for information extraction, based on Phi-3."),