litellm_settings: drop_params: true model_list: - model_name: "HuggingFace: Mistral: Mistral 7B Instruct v0.1" litellm_params: model: huggingface/mistralai/Mistral-7B-Instruct-v0.1 api_key: os.environ/HF_TOKEN max_tokens: 1024 - model_name: "HuggingFace: Mistral: Mistral 7B Instruct v0.2" litellm_params: model: huggingface/mistralai/Mistral-7B-Instruct-v0.2 api_key: os.environ/HF_TOKEN max_tokens: 1024 - model_name: "HuggingFace: Meta: Llama 3 8B Instruct" litellm_params: model: huggingface/meta-llama/Meta-Llama-3-8B-Instruct api_key: os.environ/HF_TOKEN max_tokens: 2047 - model_name: "HuggingFace: Mistral: Mixtral 8x7B Instruct v0.1" litellm_params: model: huggingface/mistralai/Mixtral-8x7B-Instruct-v0.1 api_key: os.environ/HF_TOKEN max_tokens: 8192 - model_name: "HuggingFace: Microsoft: Phi-3 Mini-4K-Instruct" litellm_params: model: huggingface/microsoft/Phi-3-mini-4k-instruct api_key: os.environ/HF_TOKEN max_tokens: 1024 - model_name: "HuggingFace: Google: Gemma 7B 1.1" litellm_params: model: huggingface/google/gemma-1.1-7b-it api_key: os.environ/HF_TOKEN max_tokens: 1024 - model_name: "HuggingFace: Yi-1.5 34B Chat" litellm_params: model: huggingface/01-ai/Yi-1.5-34B-Chat api_key: os.environ/HF_TOKEN max_tokens: 1024 - model_name: "HuggingFace: Nous Research: Nous Hermes 2 Mixtral 8x7B DPO" litellm_params: model: huggingface/NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO api_key: os.environ/HF_TOKEN max_tokens: 2048