{"object":"list","data":[{"id":"deepseek-v3","object":"model","canonical_slug":null,"hugging_face_id":"deepseek-ai/DeepSeek-V3","name":"DeepSeek V3","created":1765024302,"description":"DeepSeek V3 is DeepSeek's Mixture-of-Experts chat model for general instruction following, coding, and tool use.","context_length":163840,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"DeepSeek","instruct_type":null},"top_provider":{"context_length":163840,"max_completion_tokens":8192,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","tools","tool_choice","response_format"],"supported_api_endpoints":["chat.completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Tensorix","slug":"tensorix"}],"pricing":{"prompt":"0.0000003","completion":"0.000001","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"deepseek","slug":"deepseek-v3","tags":["open-source","moe","function-calling"],"author_info":{"slug":"deepseek","name":"DeepSeek","display_name":null,"icon_url":"/images/logos/deepseek.png","gradient_from":"from-[#0066ff]","gradient_to":"to-[#00ccff]","gradient_via":null,"website_url":"https://deepseek.com"}},{"id":"llama-4-maverick","object":"model","canonical_slug":null,"hugging_face_id":"meta-llama/Llama-4-Maverick-17B-128E-Instruct","name":"Llama 4 Maverick","created":1774895902,"description":"Llama 4 Maverick is Meta's Mixture-of-Experts model from the Llama 4 family for long-context chat and function-calling workloads.","context_length":1050000,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"Llama4","instruct_type":null},"top_provider":{"context_length":1050000,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","tools","tool_choice","response_format"],"supported_api_endpoints":["chat.completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Tensorix","slug":"tensorix"}],"pricing":{"prompt":"0.000000136","completion":"0.00000068","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"meta","slug":"llama-4-maverick","tags":["open-source","moe","long-context","function-calling"],"author_info":{"slug":"meta","name":"Meta","display_name":"Meta AI","icon_url":"/images/logos/meta.webp","gradient_from":"from-blue-600","gradient_to":"to-blue-800","gradient_via":null,"website_url":"https://ai.meta.com"}},{"id":"llama-guard-3-8b","object":"model","canonical_slug":null,"hugging_face_id":"","name":"Llama Guard 3 8B","created":1768127608,"description":"Moderation model providing robust safety classification and policy enforcement.","context_length":8192,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"Llama3","instruct_type":null},"top_provider":{"context_length":8192,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","max_tokens"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[],"pricing":{"prompt":"0.00000002","completion":"0.00000006","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"meta","slug":"llama-guard-3-8b","tags":["safety","moderation","guardrail"],"author_info":{"slug":"meta","name":"Meta","display_name":"Meta AI","icon_url":"/images/logos/meta.webp","gradient_from":"from-blue-600","gradient_to":"to-blue-800","gradient_via":null,"website_url":"https://ai.meta.com"}},{"id":"pixtral-12b","object":"model","canonical_slug":null,"hugging_face_id":"mistralai/Pixtral-12B-2409","name":"Pixtral 12B","created":1765024302,"description":"Pixtral 12B is Mistral's multimodal model for image understanding, visual question answering, and document analysis.","context_length":128000,"architecture":{"modality":"text+image->text","input_modalities":["text","image"],"output_modalities":["text"],"tokenizer":"Tekken","instruct_type":null},"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","response_format"],"supported_api_endpoints":["chat.completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Scaleway","slug":"scaleway"}],"pricing":{"prompt":"0.00000015","completion":"0.00000015","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"mistral","slug":"pixtral-12b","tags":["open-source","vision","multimodal","instruction-tuned"],"author_info":{"slug":"mistral","name":"Mistral","display_name":"Mistral AI","icon_url":"/images/logos/mistral.jpeg","gradient_from":"from-[#ff7000]","gradient_to":"to-[#ff9a00]","gradient_via":null,"website_url":"https://mistral.ai"}},{"id":"bge-m3","object":"model","canonical_slug":null,"hugging_face_id":"BAAI/bge-m3","name":"BGE M3","created":1769169016,"description":"BGE M3 is BAAI's multilingual embedding model for dense, sparse, and multi-vector retrieval with long-context inputs.","context_length":8192,"architecture":{"modality":"text->embedding","input_modalities":["text"],"output_modalities":["embedding"],"tokenizer":null,"instruct_type":null},"top_provider":{"context_length":8192,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["encoding_format","dimensions"],"supported_api_endpoints":["embeddings"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"OVHcloud","slug":"ovhcloud"},{"name":"IONOS Cloud","slug":"ionos"}],"pricing":{"prompt":"0.00000001","completion":"0","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"EUR"},"author":"baai","slug":"bge-m3","tags":["open-source","embedding","multilingual","retrieval"],"author_info":{"slug":"baai","name":"BAAI","display_name":"Beijing Academy of AI","icon_url":null,"gradient_from":"from-red-600","gradient_to":"to-red-800","gradient_via":null,"website_url":"https://www.baai.ac.cn"}},{"id":"ministral-3-14b","object":"model","canonical_slug":null,"hugging_face_id":"","name":"Ministral 3 14B","created":1767979257,"description":"Ministral 3 14B is an open Mistral AI model for efficient text and vision workloads.","context_length":262144,"architecture":{"modality":"text+image->text","input_modalities":["text","image"],"output_modalities":["text"],"tokenizer":"Tekken","instruct_type":null},"top_provider":{"context_length":262144,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","random_seed","response_format","tools","tool_choice"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Mistral AI","slug":"mistral"}],"pricing":{"prompt":"0.0000002","completion":"0.0000002","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"mistral","slug":"ministral-3-14b","tags":["open-source","vision","lightweight","long-context"],"author_info":{"slug":"mistral","name":"Mistral","display_name":"Mistral AI","icon_url":"/images/logos/mistral.jpeg","gradient_from":"from-[#ff7000]","gradient_to":"to-[#ff9a00]","gradient_via":null,"website_url":"https://mistral.ai"}},{"id":"o3","object":"model","canonical_slug":null,"hugging_face_id":"","name":"o3","created":1767619877,"description":"o3 is OpenAI's reasoning model for advanced problem solving, coding, math, science, visual perception, and tool use.","context_length":200000,"architecture":{"modality":"text+image->text","input_modalities":["text","image"],"output_modalities":["text"],"tokenizer":"GPT","instruct_type":null},"top_provider":{"context_length":200000,"max_completion_tokens":100000,"is_moderated":false},"per_request_limits":null,"supported_parameters":["max_tokens","max_completion_tokens","reasoning_effort","tools","tool_choice","response_format","parallel_tool_calls"],"supported_api_endpoints":["chat.completions","responses"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[],"pricing":{"prompt":"0.0000022","completion":"0.0000088","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000055","input_cache_write":"0","discount":1,"currency":"USD"},"author":"openai","slug":"o3","tags":["reasoning","long-context","flagship","coding","math","vision"],"author_info":{"slug":"openai","name":"OpenAI","display_name":null,"icon_url":"/images/logos/openai.webp","gradient_from":"from-zinc-800","gradient_to":"to-zinc-900","gradient_via":null,"website_url":"https://openai.com"}},{"id":"gpt-5.2","object":"model","canonical_slug":null,"hugging_face_id":"","name":"GPT-5.2","created":1777454744,"description":"GPT-5.2 is OpenAI's reasoning model for complex multimodal tasks, structured outputs, tool use, coding, and long-context work.","context_length":400000,"architecture":{"modality":"text+image->text","input_modalities":["text","image"],"output_modalities":["text"],"tokenizer":"GPT","instruct_type":null},"top_provider":{"context_length":400000,"max_completion_tokens":128000,"is_moderated":false},"per_request_limits":null,"supported_parameters":["max_tokens","max_completion_tokens","reasoning_effort","tools","tool_choice","response_format","parallel_tool_calls"],"supported_api_endpoints":["chat.completions","responses"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[],"pricing":{"prompt":"0.000001925","completion":"0.0000154","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000001925","input_cache_write":"0","discount":1,"currency":"USD"},"author":"openai","slug":"gpt-5.2","tags":["reasoning","long-context","flagship","coding","vision"],"author_info":{"slug":"openai","name":"OpenAI","display_name":null,"icon_url":"/images/logos/openai.webp","gradient_from":"from-zinc-800","gradient_to":"to-zinc-900","gradient_via":null,"website_url":"https://openai.com"}},{"id":"glm-4.7-flash","object":"model","canonical_slug":null,"hugging_face_id":"","name":"GLM 4.7 Flash","created":1777454590,"description":"GLM 4.7 Flash is Z.AI's faster GLM 4.7 variant available through Amazon Bedrock.","context_length":202752,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":null,"instruct_type":null},"top_provider":{"context_length":202752,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"AWS Bedrock","slug":"aws-bedrock"}],"pricing":{"prompt":"0.00000008","completion":"0.00000048","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"zhipu","slug":"glm-4.7-flash","tags":["proprietary","instruction-tuned"],"author_info":{"slug":"zhipu","name":"Zhipu","display_name":"Zhipu AI","icon_url":null,"gradient_from":"from-blue-500","gradient_to":"to-cyan-500","gradient_via":null,"website_url":"https://www.zhipuai.cn"}},{"id":"o1","object":"model","canonical_slug":null,"hugging_face_id":"","name":"o1","created":1777454744,"description":"o1 is OpenAI's reasoning model for complex tasks with structured outputs, multimodal input, and tool use.","context_length":200000,"architecture":{"modality":"text+image->text","input_modalities":["text","image"],"output_modalities":["text"],"tokenizer":"GPT","instruct_type":null},"top_provider":{"context_length":200000,"max_completion_tokens":100000,"is_moderated":false},"per_request_limits":null,"supported_parameters":["max_tokens","max_completion_tokens","reasoning_effort","tools","tool_choice","response_format","parallel_tool_calls"],"supported_api_endpoints":["chat.completions","responses"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[],"pricing":{"prompt":"0.0000165","completion":"0.000066","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000825","input_cache_write":"0","discount":1,"currency":"USD"},"author":"openai","slug":"o1","tags":["reasoning","long-context","multimodal","tool-use","vision"],"author_info":{"slug":"openai","name":"OpenAI","display_name":null,"icon_url":"/images/logos/openai.webp","gradient_from":"from-zinc-800","gradient_to":"to-zinc-900","gradient_via":null,"website_url":"https://openai.com"}},{"id":"gpt-4.1-mini","object":"model","canonical_slug":null,"hugging_face_id":"","name":"GPT-4.1 Mini","created":1768138176,"description":"GPT-4.1 Mini is OpenAI's smaller multimodal GPT-4.1 model for cost-efficient long-context, tool, and structured-output workloads.","context_length":1000000,"architecture":{"modality":"text+image->text","input_modalities":["text","image"],"output_modalities":["text"],"tokenizer":"GPT","instruct_type":null},"top_provider":{"context_length":1000000,"max_completion_tokens":32768,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","max_completion_tokens","stop","tools","tool_choice","response_format","seed","parallel_tool_calls"],"supported_api_endpoints":["chat.completions","responses"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Microsoft Foundry","slug":"microsoft-foundry"}],"pricing":{"prompt":"0.0000004","completion":"0.0000016","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"openai","slug":"gpt-4.1-mini","tags":["multimodal","function-calling","long-context","efficient"],"author_info":{"slug":"openai","name":"OpenAI","display_name":null,"icon_url":"/images/logos/openai.webp","gradient_from":"from-zinc-800","gradient_to":"to-zinc-900","gradient_via":null,"website_url":"https://openai.com"}},{"id":"pixtral-large","object":"model","canonical_slug":null,"hugging_face_id":"","name":"Pixtral Large","created":1767979257,"description":"Pixtral Large is Mistral's premium multimodal model with 128K context. Excels at complex visual reasoning, document analysis, and image understanding.","context_length":128000,"architecture":{"modality":"text+image->text","input_modalities":["text","image"],"output_modalities":["text"],"tokenizer":"Tekken","instruct_type":null},"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","tools","tool_choice"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[],"pricing":{"prompt":"0.000002","completion":"0.000006","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"mistral","slug":"pixtral-large","tags":["vision","multimodal","flagship","instruction-tuned"],"author_info":{"slug":"mistral","name":"Mistral","display_name":"Mistral AI","icon_url":"/images/logos/mistral.jpeg","gradient_from":"from-[#ff7000]","gradient_to":"to-[#ff9a00]","gradient_via":null,"website_url":"https://mistral.ai"}},{"id":"gpt-5-mini","object":"model","canonical_slug":null,"hugging_face_id":"","name":"GPT-5 Mini","created":1767619876,"description":"GPT-5 Mini is OpenAI's smaller GPT-5 reasoning model for cost-efficient multimodal, tool, and long-context workloads.","context_length":400000,"architecture":{"modality":"text+image->text","input_modalities":["text","image"],"output_modalities":["text"],"tokenizer":"GPT","instruct_type":null},"top_provider":{"context_length":400000,"max_completion_tokens":128000,"is_moderated":false},"per_request_limits":null,"supported_parameters":["max_tokens","max_completion_tokens","reasoning_effort","tools","tool_choice","response_format","parallel_tool_calls"],"supported_api_endpoints":["chat.completions","responses"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Microsoft Foundry","slug":"microsoft-foundry"}],"pricing":{"prompt":"0.000000275","completion":"0.0000022","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000000275","input_cache_write":"0","discount":1,"currency":"USD"},"author":"openai","slug":"gpt-5-mini","tags":["reasoning","long-context","efficient","coding","vision"],"author_info":{"slug":"openai","name":"OpenAI","display_name":null,"icon_url":"/images/logos/openai.webp","gradient_from":"from-zinc-800","gradient_to":"to-zinc-900","gradient_via":null,"website_url":"https://openai.com"}},{"id":"deepseek-r1-distill-llama-70b","object":"model","canonical_slug":null,"hugging_face_id":"deepseek-ai/DeepSeek-R1-Distill-Llama-70B","name":"DeepSeek R1 Distill Llama 70B","created":1765183068,"description":"DeepSeek-R1-Distill-Llama-70B is a 70B model distilled from DeepSeek-R1 using Llama-3.3-70B-Instruct as the base. Features chain-of-thought reasoning with <think> tokens, 128K context, and strong performance on AIME 2024, MATH-500, and LiveCodeBench. Optimized for complex math, coding, and reasoning tasks.","context_length":128000,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"Llama3","instruct_type":null},"top_provider":{"context_length":128000,"max_completion_tokens":32768,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[],"pricing":{"prompt":"0.0000009","completion":"0.0000009","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"EUR"},"author":"deepseek","slug":"deepseek-r1-distill-llama-70b","tags":["open-source","reasoning","distilled","coding","math"],"author_info":{"slug":"deepseek","name":"DeepSeek","display_name":null,"icon_url":"/images/logos/deepseek.png","gradient_from":"from-[#0066ff]","gradient_to":"to-[#00ccff]","gradient_via":null,"website_url":"https://deepseek.com"}},{"id":"glm-4.5-air","object":"model","canonical_slug":null,"hugging_face_id":"","name":"GLM 4.5 Air","created":1768127608,"description":"Cost-efficient GLM variant offering fast inference while keeping the core reasoning strengths.","context_length":128000,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"GLM","instruct_type":null},"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","tools","tool_choice","response_format"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[],"pricing":{"prompt":"0.0000002","completion":"0.0000012","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"zhipu","slug":"glm-4.5-air","tags":["efficient","reasoning"],"author_info":{"slug":"zhipu","name":"Zhipu","display_name":"Zhipu AI","icon_url":null,"gradient_from":"from-blue-500","gradient_to":"to-cyan-500","gradient_via":null,"website_url":"https://www.zhipuai.cn"}},{"id":"hermes-4-405b","object":"model","canonical_slug":null,"hugging_face_id":"NousResearch/Hermes-4-405B","name":"Hermes 4 405B","created":1768127608,"description":"NousResearch Hermes 4 405B is an open-weight language model for instruction following, chat, tool use, and reasoning-oriented workflows.","context_length":131072,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"Llama3","instruct_type":null},"top_provider":{"context_length":131072,"max_completion_tokens":131072,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","response_format","tools","tool_choice"],"supported_api_endpoints":["chat.completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Nebius","slug":"nebius"}],"pricing":{"prompt":"0.000001","completion":"0.000003","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"nousresearch","slug":"hermes-4-405b","tags":["open-source","instruction-tuned","function-calling","json-mode"],"author_info":{"slug":"nousresearch","name":"Nous Research","display_name":null,"icon_url":null,"gradient_from":"from-emerald-600","gradient_to":"to-teal-700","gradient_via":null,"website_url":"https://nousresearch.com"}},{"id":"qwen3-coder-next-80b","object":"model","canonical_slug":null,"hugging_face_id":"Qwen/Qwen3-Coder-Next","name":"Qwen3 Coder Next 80B","created":1777453485,"description":"Qwen3-Coder-Next is Alibaba's open-weight 80B total, 3B active MoE coding model for coding agents, long-horizon tool use, and local development workflows.","context_length":262144,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"Qwen3","instruct_type":null},"top_provider":{"context_length":262144,"max_completion_tokens":65536,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","response_format","tools","tool_choice","reasoning_effort"],"supported_api_endpoints":["chat.completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"IONOS Cloud","slug":"ionos"}],"pricing":{"prompt":"0.00000017","completion":"0.00000089","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"alibaba","slug":"qwen3-coder-next-80b","tags":["open-source","coding","agentic","function-calling","long-context","moe"],"author_info":{"slug":"alibaba","name":"Alibaba","display_name":"Alibaba (Qwen)","icon_url":"/images/logos/qwen.webp","gradient_from":"from-purple-600","gradient_to":"to-indigo-700","gradient_via":null,"website_url":"https://qwenlm.github.io"}},{"id":"green-embedding","object":"model","canonical_slug":null,"hugging_face_id":"","name":"Green Embedding","created":1768905490,"description":"High-quality text embeddings for semantic search, similarity matching, and vector database operations. EU-hosted.","context_length":32000,"architecture":{"modality":"text->embedding","input_modalities":["text"],"output_modalities":["embedding"],"tokenizer":null,"instruct_type":null},"top_provider":{"context_length":32000,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":[],"supported_api_endpoints":["embeddings"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"GreenPT","slug":"greenpt"}],"pricing":{"prompt":"0.0000002","completion":"0","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"EUR"},"author":"greenpt","slug":"green-embedding","tags":["eu-native","sustainable","embedding"],"author_info":{"slug":"greenpt","name":"GreenPT","display_name":"GreenPT","icon_url":null,"gradient_from":"from-green-600","gradient_to":"to-green-700","gradient_via":null,"website_url":"https://greenpt.ai"}},{"id":"kimi-k2.6","object":"model","canonical_slug":null,"hugging_face_id":"moonshotai/Kimi-K2.6","name":"Kimi K2.6","created":1777453828,"description":"Kimi K2.6 is Moonshot AI's open-weight reasoning model for long-horizon coding, agentic execution, and multimodal reasoning.","context_length":256000,"architecture":{"modality":"text+image+video->text","input_modalities":["text","image","video"],"output_modalities":["text"],"tokenizer":"TikTokenTokenizer","instruct_type":null},"top_provider":{"context_length":256000,"max_completion_tokens":256000,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","frequency_penalty","presence_penalty","seed","logit_bias","logprobs","top_logprobs","parallel_tool_calls","reasoning_effort","tools","tool_choice","response_format"],"supported_api_endpoints":["chat.completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Tensorix","slug":"tensorix"},{"name":"Inceptron","slug":"inceptron"}],"pricing":{"prompt":"0.0000008","completion":"0.0000035","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000002","input_cache_write":"0","discount":1,"currency":"USD"},"author":"moonshot","slug":"kimi-k2.6","tags":["open-source","moe","long-context","vision","coding","reasoning"],"author_info":{"slug":"moonshot","name":"Moonshot","display_name":"Moonshot AI","icon_url":null,"gradient_from":"from-indigo-500","gradient_to":"to-purple-600","gradient_via":null,"website_url":"https://moonshot.ai"}},{"id":"hermes-4-70b","object":"model","canonical_slug":null,"hugging_face_id":"NousResearch/Hermes-4-70B","name":"Hermes 4 70B","created":1768127608,"description":"NousResearch Hermes 4 70B is an open-weight language model for instruction following, chat, tool use, and reasoning-oriented workflows.","context_length":131072,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"Llama3","instruct_type":null},"top_provider":{"context_length":131072,"max_completion_tokens":131072,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","response_format","tools","tool_choice"],"supported_api_endpoints":["chat.completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Nebius","slug":"nebius"}],"pricing":{"prompt":"0.00000013","completion":"0.0000004","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"nousresearch","slug":"hermes-4-70b","tags":["open-source","instruction-tuned","function-calling","json-mode"],"author_info":{"slug":"nousresearch","name":"Nous Research","display_name":null,"icon_url":null,"gradient_from":"from-emerald-600","gradient_to":"to-teal-700","gradient_via":null,"website_url":"https://nousresearch.com"}},{"id":"devstral-2","object":"model","canonical_slug":null,"hugging_face_id":"","name":"Devstral 2","created":1767979257,"description":"Devstral 2 is Mistral AI's frontier code agent model for tool-using software engineering tasks.","context_length":262144,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"Tekken","instruct_type":null},"top_provider":{"context_length":262144,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","random_seed","response_format","tools","tool_choice"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Mistral AI","slug":"mistral"}],"pricing":{"prompt":"0.0000004","completion":"0.000002","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"mistral","slug":"devstral-2","tags":["open-source","coding","agentic","function-calling"],"author_info":{"slug":"mistral","name":"Mistral","display_name":"Mistral AI","icon_url":"/images/logos/mistral.jpeg","gradient_from":"from-[#ff7000]","gradient_to":"to-[#ff9a00]","gradient_via":null,"website_url":"https://mistral.ai"}},{"id":"mistral-small-3.1-24b","object":"model","canonical_slug":null,"hugging_face_id":"mistralai/Mistral-Small-3.1-24B-Instruct-2503","name":"Mistral Small 3.1 24B","created":1765024302,"description":"Mistral Small 3.1 is a 24B parameter model with a 128K context window. It supports both text and image inputs (vision), features efficient inference at 150 tokens/second, and is optimized for conversational agents, function calling, and long-document comprehension.","context_length":128000,"architecture":{"modality":"text+image->text","input_modalities":["text","image"],"output_modalities":["text"],"tokenizer":"Tekken","instruct_type":null},"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","tools","tool_choice","response_format","safe_prompt"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[],"pricing":{"prompt":"0.0000001","completion":"0.0000003","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"mistral","slug":"mistral-small-3.1-24b","tags":["open-source","instruction-tuned","vision","long-context","efficient"],"author_info":{"slug":"mistral","name":"Mistral","display_name":"Mistral AI","icon_url":"/images/logos/mistral.jpeg","gradient_from":"from-[#ff7000]","gradient_to":"to-[#ff9a00]","gradient_via":null,"website_url":"https://mistral.ai"}},{"id":"qwen3-235b-a22b-instruct","object":"model","canonical_slug":null,"hugging_face_id":"Qwen/Qwen3-235B-A22B-Instruct-2507","name":"Qwen3 235B A22B Instruct","created":1765183068,"description":"Qwen3 235B A22B Instruct is Alibaba's Mixture-of-Experts model for multilingual chat, reasoning, coding, and tool use.","context_length":131000,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"Qwen3","instruct_type":null},"top_provider":{"context_length":131000,"max_completion_tokens":262144,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","tools","tool_choice","response_format"],"supported_api_endpoints":["chat.completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"GreenPT","slug":"greenpt"},{"name":"Tensorix","slug":"tensorix"},{"name":"Scaleway","slug":"scaleway"},{"name":"AWS Bedrock","slug":"aws-bedrock"},{"name":"Nebius","slug":"nebius"}],"pricing":{"prompt":"0.000000072","completion":"0.000000464","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"alibaba","slug":"qwen3-235b-a22b-instruct","tags":["open-source","moe","multilingual","function-calling"],"author_info":{"slug":"alibaba","name":"Alibaba","display_name":"Alibaba (Qwen)","icon_url":"/images/logos/qwen.webp","gradient_from":"from-purple-600","gradient_to":"to-indigo-700","gradient_via":null,"website_url":"https://qwenlm.github.io"}},{"id":"gemma-3-27b-it","object":"model","canonical_slug":null,"hugging_face_id":"google/gemma-3-27b-it","name":"Gemma 3 27B IT","created":1765183068,"description":"Gemma 3 27B IT is Google's multimodal instruction-tuned model with text and image input support.","context_length":110000,"architecture":{"modality":"text+image->text","input_modalities":["text","image"],"output_modalities":["text"],"tokenizer":"Gemma","instruct_type":null},"top_provider":{"context_length":110000,"max_completion_tokens":110000,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","response_format","tools","tool_choice"],"supported_api_endpoints":["chat.completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"GreenPT","slug":"greenpt"},{"name":"Scaleway","slug":"scaleway"},{"name":"Nebius","slug":"nebius"}],"pricing":{"prompt":"0.0000001","completion":"0.0000003","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"google","slug":"gemma-3-27b-it","tags":["open-source","vision","multimodal","multilingual","instruction-tuned"],"author_info":{"slug":"google","name":"Google","display_name":null,"icon_url":"/images/logos/gemini.webp","gradient_from":"from-blue-500","gradient_to":"to-yellow-500","gradient_via":"via-green-500","website_url":"https://deepmind.google"}},{"id":"magistral-small","object":"model","canonical_slug":null,"hugging_face_id":"","name":"Magistral Small 1.2","created":1767979257,"description":"Magistral Small 1.2 is Mistral AI's small reasoning model for transparent and multilingual reasoning.","context_length":40000,"architecture":{"modality":"text+image->text","input_modalities":["text","image"],"output_modalities":["text"],"tokenizer":"Tekken","instruct_type":null},"top_provider":{"context_length":40000,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","random_seed","response_format","tools","tool_choice","reasoning_effort"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Mistral AI","slug":"mistral"}],"pricing":{"prompt":"0.0000005","completion":"0.0000015","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"mistral","slug":"magistral-small","tags":["open-source","reasoning","multimodal"],"author_info":{"slug":"mistral","name":"Mistral","display_name":"Mistral AI","icon_url":"/images/logos/mistral.jpeg","gradient_from":"from-[#ff7000]","gradient_to":"to-[#ff9a00]","gradient_via":null,"website_url":"https://mistral.ai"}},{"id":"mistral-small-4","object":"model","canonical_slug":null,"hugging_face_id":"mistralai/Mistral-Small-4-119B-2603","name":"Mistral Small 4","created":1777454436,"description":"Mistral Small 4 is Mistral AI's open hybrid MoE model unifying instruct, reasoning, coding, and multimodal workloads.","context_length":262144,"architecture":{"modality":"text+image->text","input_modalities":["text","image"],"output_modalities":["text"],"tokenizer":"Tekken","instruct_type":null},"top_provider":{"context_length":262144,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","random_seed","response_format","tools","tool_choice","reasoning_effort"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Mistral AI","slug":"mistral"}],"pricing":{"prompt":"0.00000015","completion":"0.0000006","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"mistral","slug":"mistral-small-4","tags":["open-source","moe","vision","reasoning","coding","function-calling","long-context"],"author_info":{"slug":"mistral","name":"Mistral","display_name":"Mistral AI","icon_url":"/images/logos/mistral.jpeg","gradient_from":"from-[#ff7000]","gradient_to":"to-[#ff9a00]","gradient_via":null,"website_url":"https://mistral.ai"}},{"id":"green-r","object":"model","canonical_slug":null,"hugging_face_id":"","name":"GreenR","created":1768905404,"description":"Reasoning model specialized in logical reasoning and problem-solving. Ideal for complex analytical tasks and mathematical problems. Based on GPT-OSS 120B, EU-hosted.","context_length":128000,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":null,"instruct_type":null},"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","tools","tool_choice","response_format"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"GreenPT","slug":"greenpt"}],"pricing":{"prompt":"0.00000035","completion":"0.00000095","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"EUR"},"author":"greenpt","slug":"green-r","tags":["eu-native","sustainable","reasoning"],"author_info":{"slug":"greenpt","name":"GreenPT","display_name":"GreenPT","icon_url":null,"gradient_from":"from-green-600","gradient_to":"to-green-700","gradient_via":null,"website_url":"https://greenpt.ai"}},{"id":"nova-pro","object":"model","canonical_slug":null,"hugging_face_id":"","name":"Amazon Nova Pro","created":1769967814,"description":"Amazon Nova Pro is a highly capable multimodal model balancing performance and cost for complex tasks.","context_length":300000,"architecture":{"modality":"text+image+video->text","input_modalities":["text","image","video"],"output_modalities":["text"],"tokenizer":null,"instruct_type":null},"top_provider":{"context_length":300000,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","tools","tool_choice"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"AWS Bedrock","slug":"aws-bedrock"}],"pricing":{"prompt":"0.0000008","completion":"0.0000032","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"amazon","slug":"nova-pro","tags":["proprietary","multimodal","vision","video","reasoning"],"author_info":{"slug":"amazon","name":"Amazon","display_name":"Amazon Web Services","icon_url":"/images/logos/aws.webp","gradient_from":"from-orange-500","gradient_to":"to-yellow-500","gradient_via":null,"website_url":"https://aws.amazon.com/bedrock"}},{"id":"qwen3-32b","object":"model","canonical_slug":null,"hugging_face_id":"Qwen/Qwen3-32B","name":"Qwen3 32B","created":1767979257,"description":"Qwen3 32B is Alibaba's dense multilingual model with tool calling and optional reasoning mode for math, code, and general instruction-following tasks.","context_length":32768,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"Qwen3","instruct_type":null},"top_provider":{"context_length":32768,"max_completion_tokens":32768,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","seed","response_format","tools","tool_choice","reasoning_effort"],"supported_api_endpoints":["chat.completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"OVHcloud","slug":"ovhcloud"},{"name":"AWS Bedrock","slug":"aws-bedrock"},{"name":"Nebius","slug":"nebius"}],"pricing":{"prompt":"0.00000008","completion":"0.00000023","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"EUR"},"author":"alibaba","slug":"qwen3-32b","tags":["open-source","reasoning","multilingual","tool-calling"],"author_info":{"slug":"alibaba","name":"Alibaba","display_name":"Alibaba (Qwen)","icon_url":"/images/logos/qwen.webp","gradient_from":"from-purple-600","gradient_to":"to-indigo-700","gradient_via":null,"website_url":"https://qwenlm.github.io"}},{"id":"minimax-m2.5","object":"model","canonical_slug":null,"hugging_face_id":"MiniMaxAI/MiniMax-M2.5","name":"MiniMax M2.5","created":1773690794,"description":"MiniMax M2.5 is a MiniMax model for agentic coding, tool use, reasoning, and high-throughput chat workloads.","context_length":196608,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"MiniMax","instruct_type":null},"top_provider":{"context_length":196608,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","frequency_penalty","presence_penalty","seed","logit_bias","logprobs","top_logprobs","parallel_tool_calls","reasoning_effort","tools","tool_choice","response_format"],"supported_api_endpoints":["chat.completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Inceptron","slug":"inceptron"},{"name":"Infercom","slug":"infercom"},{"name":"Tensorix","slug":"tensorix"}],"pricing":{"prompt":"0.00000028","completion":"0.0000011","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000003","input_cache_write":"0","discount":1,"currency":"USD"},"author":"minimax","slug":"minimax-m2.5","tags":["open-source","moe","coding","reasoning","function-calling"],"author_info":{"slug":"minimax","name":"MiniMax","display_name":"MiniMax AI","icon_url":null,"gradient_from":"from-blue-500","gradient_to":"to-indigo-600","gradient_via":null,"website_url":"https://minimaxi.com"}},{"id":"qwen3-embedding-8b","object":"model","canonical_slug":null,"hugging_face_id":"Qwen/Qwen3-Embedding-8B","name":"Qwen3 Embedding 8B","created":1767963057,"description":"Qwen3 Embedding 8B is Alibaba's embedding model for multilingual semantic search, retrieval, and representation learning.","context_length":40960,"architecture":{"modality":"text->embedding","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"Qwen3","instruct_type":null},"top_provider":{"context_length":40960,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["encoding_format","dimensions"],"supported_api_endpoints":["embeddings"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Scaleway","slug":"scaleway"},{"name":"Nebius","slug":"nebius"},{"name":"Tensorix","slug":"tensorix"},{"name":"OVHcloud","slug":"ovhcloud"}],"author":"alibaba","slug":"qwen3-embedding-8b","tags":["open-source","embedding","multilingual","retrieval"],"author_info":{"slug":"alibaba","name":"Alibaba","display_name":"Alibaba (Qwen)","icon_url":"/images/logos/qwen.webp","gradient_from":"from-purple-600","gradient_to":"to-indigo-700","gradient_via":null,"website_url":"https://qwenlm.github.io"}},{"id":"deepseek-chat-v3.1","object":"model","canonical_slug":null,"hugging_face_id":"deepseek-ai/DeepSeek-V3.1","name":"DeepSeek Chat V3.1","created":1774895903,"description":"DeepSeek Chat V3.1 is an upgraded DeepSeek V3 model with long-context chat, function calling, and optional reasoning support.","context_length":164000,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"DeepSeek","instruct_type":null},"top_provider":{"context_length":164000,"max_completion_tokens":163840,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","tools","tool_choice","response_format"],"supported_api_endpoints":["chat.completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Tensorix","slug":"tensorix"}],"pricing":{"prompt":"0.0000002","completion":"0.0000008","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"deepseek","slug":"deepseek-chat-v3.1","tags":["open-source","moe","long-context","function-calling"],"author_info":{"slug":"deepseek","name":"DeepSeek","display_name":null,"icon_url":"/images/logos/deepseek.png","gradient_from":"from-[#0066ff]","gradient_to":"to-[#00ccff]","gradient_via":null,"website_url":"https://deepseek.com"}},{"id":"nova-lite","object":"model","canonical_slug":null,"hugging_face_id":"","name":"Amazon Nova Lite","created":1769967814,"description":"Amazon Nova Lite is a multimodal model supporting text, images, and video with low-latency responses.","context_length":300000,"architecture":{"modality":"text+image+video->text","input_modalities":["text","image","video"],"output_modalities":["text"],"tokenizer":null,"instruct_type":null},"top_provider":{"context_length":300000,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","tools","tool_choice"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"AWS Bedrock","slug":"aws-bedrock"}],"pricing":{"prompt":"0.00000006","completion":"0.00000024","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"amazon","slug":"nova-lite","tags":["proprietary","multimodal","vision","video"],"author_info":{"slug":"amazon","name":"Amazon","display_name":"Amazon Web Services","icon_url":"/images/logos/aws.webp","gradient_from":"from-orange-500","gradient_to":"to-yellow-500","gradient_via":null,"website_url":"https://aws.amazon.com/bedrock"}},{"id":"gemma-4","object":"model","canonical_slug":null,"hugging_face_id":"","name":"Gemma 4","created":1777452602,"description":"Gemma 4 is Google DeepMind's open model family built from Gemini 3 research and technology for high intelligence-per-parameter. The 31B variant offers long-context multimodal reasoning, function-calling support, multilingual understanding, and efficient deployment on personal hardware.","context_length":256000,"architecture":{"modality":"text+image->text","input_modalities":["text","image"],"output_modalities":["text"],"tokenizer":"Gemma","instruct_type":null},"top_provider":{"context_length":256000,"max_completion_tokens":32000,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","tools","tool_choice"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"GreenPT","slug":"greenpt"}],"pricing":{"prompt":"0.0000005","completion":"0.0000015","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"EUR"},"author":"google","slug":"gemma-4","tags":["open-source","multimodal","vision","long-context","function-calling","multilingual","public-preview"],"author_info":{"slug":"google","name":"Google","display_name":null,"icon_url":"/images/logos/gemini.webp","gradient_from":"from-blue-500","gradient_to":"to-yellow-500","gradient_via":"via-green-500","website_url":"https://deepmind.google"}},{"id":"devstral-2-123b-instruct-2512","object":"model","canonical_slug":null,"hugging_face_id":"mistralai/Devstral-2-123B-Instruct-2512","name":"Devstral 2 123B Instruct","created":1777452602,"description":"Devstral 2 123B Instruct is Mistral's long-context code model for agentic software engineering and multi-file reasoning tasks.","context_length":200000,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"Tekken","instruct_type":null},"top_provider":{"context_length":200000,"max_completion_tokens":16000,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","seed","response_format","tools","tool_choice"],"supported_api_endpoints":["chat.completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Scaleway","slug":"scaleway"},{"name":"GreenPT","slug":"greenpt"},{"name":"AWS Bedrock","slug":"aws-bedrock"}],"pricing":{"prompt":"0.0000004","completion":"0.000002","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"EUR"},"author":"mistral","slug":"devstral-2-123b-instruct-2512","tags":["open-source","coding","agentic","long-context","instruction-tuned"],"author_info":{"slug":"mistral","name":"Mistral","display_name":"Mistral AI","icon_url":"/images/logos/mistral.jpeg","gradient_from":"from-[#ff7000]","gradient_to":"to-[#ff9a00]","gradient_via":null,"website_url":"https://mistral.ai"}},{"id":"kimi-k2.5","object":"model","canonical_slug":null,"hugging_face_id":"moonshotai/Kimi-K2.5","name":"Kimi K2.5","created":1771080676,"description":"Kimi K2.5 is Moonshot AI's long-context multimodal model for coding, tool use, and general chat.","context_length":262144,"architecture":{"modality":"text+image->text","input_modalities":["text","image"],"output_modalities":["text"],"tokenizer":"Kimi","instruct_type":null},"top_provider":{"context_length":262144,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","tools","tool_choice","response_format"],"supported_api_endpoints":["chat.completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Inceptron","slug":"inceptron"},{"name":"Tensorix","slug":"tensorix"},{"name":"AWS Bedrock","slug":"aws-bedrock"}],"pricing":{"prompt":"0.0000005","completion":"0.0000025","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"moonshot","slug":"kimi-k2.5","tags":["open-source","moe","long-context","vision","function-calling"],"author_info":{"slug":"moonshot","name":"Moonshot","display_name":"Moonshot AI","icon_url":null,"gradient_from":"from-indigo-500","gradient_to":"to-purple-600","gradient_via":null,"website_url":"https://moonshot.ai"}},{"id":"qwen-2.5-vl-72b-instruct","object":"model","canonical_slug":null,"hugging_face_id":"Qwen/Qwen2.5-VL-72B-Instruct","name":"Qwen 2.5 VL 72B Instruct","created":1768127608,"description":"Qwen 2.5 VL 72B Instruct is Alibaba's vision-language model for image understanding, OCR, document analysis, and multimodal instruction following.","context_length":32000,"architecture":{"modality":"text+image->text","input_modalities":["text","image"],"output_modalities":["text"],"tokenizer":"Qwen2","instruct_type":null},"top_provider":{"context_length":32000,"max_completion_tokens":32000,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","response_format","tools","tool_choice"],"supported_api_endpoints":["chat.completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"OVHcloud","slug":"ovhcloud"},{"name":"Nebius","slug":"nebius"}],"pricing":{"prompt":"0.00000025","completion":"0.00000075","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"alibaba","slug":"qwen-2.5-vl-72b-instruct","tags":["open-source","vision","multimodal","document-analysis"],"author_info":{"slug":"alibaba","name":"Alibaba","display_name":"Alibaba (Qwen)","icon_url":"/images/logos/qwen.webp","gradient_from":"from-purple-600","gradient_to":"to-indigo-700","gradient_via":null,"website_url":"https://qwenlm.github.io"}},{"id":"voxtral-small-24b","object":"model","canonical_slug":null,"hugging_face_id":"mistralai/Voxtral-Small-24B-2507","name":"Voxtral Small 24B","created":1765183068,"description":"Voxtral Small 24B is Mistral AI's open audio-capable chat model for speech understanding and text generation.","context_length":32000,"architecture":{"modality":"text+audio->text","input_modalities":["text","audio"],"output_modalities":["text"],"tokenizer":"Tekken","instruct_type":null},"top_provider":{"context_length":32000,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","random_seed","response_format","tools","tool_choice"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Scaleway","slug":"scaleway"},{"name":"GreenPT","slug":"greenpt"},{"name":"Mistral AI","slug":"mistral"}],"pricing":{"prompt":"0.0000001","completion":"0.0000003","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0.0000666667","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"mistral","slug":"voxtral-small-24b","tags":["open-source","audio","function-calling","multilingual"],"author_info":{"slug":"mistral","name":"Mistral","display_name":"Mistral AI","icon_url":"/images/logos/mistral.jpeg","gradient_from":"from-[#ff7000]","gradient_to":"to-[#ff9a00]","gradient_via":null,"website_url":"https://mistral.ai"}},{"id":"minimax-m2","object":"model","canonical_slug":null,"hugging_face_id":"MiniMaxAI/MiniMax-M2","name":"MiniMax M2","created":1774895903,"description":"MiniMax M2 is a Mixture-of-Experts language model from MiniMax for coding, function calling, reasoning, and general chat.","context_length":196608,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"MiniMax","instruct_type":null},"top_provider":{"context_length":196608,"max_completion_tokens":196608,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","tools","tool_choice","response_format"],"supported_api_endpoints":["chat.completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Tensorix","slug":"tensorix"}],"pricing":{"prompt":"0.00000025","completion":"0.000001","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"minimax","slug":"minimax-m2","tags":["open-source","moe","long-context","coding","reasoning"],"author_info":{"slug":"minimax","name":"MiniMax","display_name":"MiniMax AI","icon_url":null,"gradient_from":"from-blue-500","gradient_to":"to-indigo-600","gradient_via":null,"website_url":"https://minimaxi.com"}},{"id":"cohere-embed-english-v3","object":"model","canonical_slug":null,"hugging_face_id":"","name":"Cohere Embed English v3","created":1777454590,"description":"Cohere Embed English v3 is an English text embedding model available through Amazon Bedrock.","context_length":512,"architecture":{"modality":"text->embedding","input_modalities":["text"],"output_modalities":["embedding"],"tokenizer":null,"instruct_type":null},"top_provider":{"context_length":512,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":[],"supported_api_endpoints":["embeddings"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"AWS Bedrock","slug":"aws-bedrock"}],"pricing":{"prompt":"0.0000001","completion":"0","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"cohere","slug":"cohere-embed-english-v3","tags":["proprietary","embedding"],"author_info":{"slug":"cohere","name":"Cohere","display_name":null,"icon_url":"/images/logos/cohere.webp","gradient_from":"from-purple-600","gradient_to":"to-purple-800","gradient_via":null,"website_url":"https://cohere.com"}},{"id":"claude-haiku-4.5","object":"model","canonical_slug":null,"hugging_face_id":"","name":"Claude Haiku 4.5","created":1769967813,"description":"Claude Haiku 4.5 is Anthropic's fastest model, optimized for speed and cost-efficiency while maintaining strong performance.","context_length":200000,"architecture":{"modality":"text+image->text","input_modalities":["text","image"],"output_modalities":["text"],"tokenizer":"claude","instruct_type":"claude"},"top_provider":{"context_length":200000,"max_completion_tokens":8192,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","tools","tool_choice"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"AWS Bedrock","slug":"aws-bedrock"}],"pricing":{"prompt":"0.000001","completion":"0.000005","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"anthropic","slug":"claude-haiku-4.5","tags":["proprietary","fast","cost-efficient","vision"],"author_info":{"slug":"anthropic","name":"Anthropic","display_name":null,"icon_url":"/images/logos/anthropic.jpeg","gradient_from":"from-[#cc785c]","gradient_to":"to-[#d4a574]","gradient_via":null,"website_url":"https://anthropic.com"}},{"id":"glm-4.5","object":"model","canonical_slug":null,"hugging_face_id":"","name":"GLM 4.5","created":1768127608,"description":"Flagship GLM model with strong multilingual reasoning, long context, and robust tool use.","context_length":128000,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"GLM","instruct_type":null},"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","tools","tool_choice","response_format"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[],"pricing":{"prompt":"0.0000006","completion":"0.0000022","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"zhipu","slug":"glm-4.5","tags":["multilingual","reasoning","tool-use"],"author_info":{"slug":"zhipu","name":"Zhipu","display_name":"Zhipu AI","icon_url":null,"gradient_from":"from-blue-500","gradient_to":"to-cyan-500","gradient_via":null,"website_url":"https://www.zhipuai.cn"}},{"id":"qwen-2.5-coder-7b","object":"model","canonical_slug":null,"hugging_face_id":"","name":"Qwen 2.5 Coder 7B","created":1768127608,"description":"Lightweight Qwen coder offering high-quality code generation and repo-level editing at small scale.","context_length":128000,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"Qwen3","instruct_type":null},"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","response_format"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[],"pricing":{"prompt":"0.00000003","completion":"0.00000009","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"alibaba","slug":"qwen-2.5-coder-7b","tags":["coding","efficient","open-source"],"author_info":{"slug":"alibaba","name":"Alibaba","display_name":"Alibaba (Qwen)","icon_url":"/images/logos/qwen.webp","gradient_from":"from-purple-600","gradient_to":"to-indigo-700","gradient_via":null,"website_url":"https://qwenlm.github.io"}},{"id":"qwen-2.5-72b-instruct","object":"model","canonical_slug":null,"hugging_face_id":"Qwen/Qwen2.5-72B-Instruct","name":"Qwen 2.5 72B Instruct","created":1765024302,"description":"Qwen 2.5 72B Instruct is Alibaba's instruction-tuned language model for multilingual chat, coding, and tool use.","context_length":33000,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"Qwen2","instruct_type":null},"top_provider":{"context_length":33000,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","tools","tool_choice","response_format"],"supported_api_endpoints":["chat.completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Tensorix","slug":"tensorix"}],"pricing":{"prompt":"0.00000007","completion":"0.00000026","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"alibaba","slug":"qwen-2.5-72b-instruct","tags":["open-source","multilingual","function-calling"],"author_info":{"slug":"alibaba","name":"Alibaba","display_name":"Alibaba (Qwen)","icon_url":"/images/logos/qwen.webp","gradient_from":"from-purple-600","gradient_to":"to-indigo-700","gradient_via":null,"website_url":"https://qwenlm.github.io"}},{"id":"llama-3.1-8b-instruct","object":"model","canonical_slug":null,"hugging_face_id":"meta-llama/Meta-Llama-3.1-8B-Instruct","name":"Llama 3.1 8B Instruct","created":1765024302,"description":"Meta Llama 3.1 8B Instruct is a compact multilingual instruction-tuned model for chat, tool use, and efficient assistant workloads.","context_length":131072,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"Llama3","instruct_type":null},"top_provider":{"context_length":131072,"max_completion_tokens":131072,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","response_format","tools","tool_choice"],"supported_api_endpoints":["chat.completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"IONOS Cloud","slug":"ionos"},{"name":"Nebius","slug":"nebius"}],"pricing":{"prompt":"0.00000002","completion":"0.00000006","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"meta","slug":"llama-3.1-8b-instruct","tags":["open-source","instruction-tuned","multilingual","function-calling"],"author_info":{"slug":"meta","name":"Meta","display_name":"Meta AI","icon_url":"/images/logos/meta.webp","gradient_from":"from-blue-600","gradient_to":"to-blue-800","gradient_via":null,"website_url":"https://ai.meta.com"}},{"id":"devstral-small-2","object":"model","canonical_slug":null,"hugging_face_id":"mistralai/Devstral-Small-2-24B-Instruct-2512","name":"Devstral Small 2","created":1777454436,"description":"Devstral Small 2 is Mistral AI's compact code agent model for tool-using software engineering tasks.","context_length":262144,"architecture":{"modality":"text+image->text","input_modalities":["text","image"],"output_modalities":["text"],"tokenizer":"Tekken","instruct_type":null},"top_provider":{"context_length":262144,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","random_seed","response_format","tools","tool_choice"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Mistral AI","slug":"mistral"}],"pricing":{"prompt":"0.0000001","completion":"0.0000003","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"mistral","slug":"devstral-small-2","tags":["open-source","coding","agentic","vision","function-calling"],"author_info":{"slug":"mistral","name":"Mistral","display_name":"Mistral AI","icon_url":"/images/logos/mistral.jpeg","gradient_from":"from-[#ff7000]","gradient_to":"to-[#ff9a00]","gradient_via":null,"website_url":"https://mistral.ai"}},{"id":"code-llama-13b-instruct","object":"model","canonical_slug":null,"hugging_face_id":"meta-llama/CodeLlama-13b-Instruct-hf","name":"Code Llama 13B Instruct","created":1770318640,"description":"Code Llama 13B Instruct is a code-focused model from Meta fine-tuned for instruction following. It supports 16K context and is optimized for code generation, completion, and debugging.","context_length":16384,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"Llama2","instruct_type":"llama2"},"top_provider":{"context_length":16384,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","response_format"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"IONOS Cloud","slug":"ionos"}],"pricing":{"prompt":"0.0000005","completion":"0.0000005","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"meta","slug":"code-llama-13b-instruct","tags":["open-source","coding","instruction-tuned"],"author_info":{"slug":"meta","name":"Meta","display_name":"Meta AI","icon_url":"/images/logos/meta.webp","gradient_from":"from-blue-600","gradient_to":"to-blue-800","gradient_via":null,"website_url":"https://ai.meta.com"}},{"id":"green-l-raw","object":"model","canonical_slug":null,"hugging_face_id":"","name":"GreenL Raw","created":1768905404,"description":"Large generative model with custom system prompt support. Provides more control over model behavior and responses. EU-hosted.","context_length":128000,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":null,"instruct_type":null},"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","tools","tool_choice"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"GreenPT","slug":"greenpt"}],"pricing":{"prompt":"0.00000025","completion":"0.0000008","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"EUR"},"author":"greenpt","slug":"green-l-raw","tags":["eu-native","sustainable","instruction-tuned","raw"],"author_info":{"slug":"greenpt","name":"GreenPT","display_name":"GreenPT","icon_url":null,"gradient_from":"from-green-600","gradient_to":"to-green-700","gradient_via":null,"website_url":"https://greenpt.ai"}},{"id":"gpt-4.1","object":"model","canonical_slug":null,"hugging_face_id":"","name":"GPT-4.1","created":1768138176,"description":"GPT-4.1 is OpenAI's multimodal general-purpose model for instruction following, coding, long-context work, function calling, and structured outputs.","context_length":1000000,"architecture":{"modality":"text+image->text","input_modalities":["text","image"],"output_modalities":["text"],"tokenizer":"GPT","instruct_type":null},"top_provider":{"context_length":1000000,"max_completion_tokens":32768,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","max_completion_tokens","stop","tools","tool_choice","response_format","seed","parallel_tool_calls"],"supported_api_endpoints":["chat.completions","responses"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Microsoft Foundry","slug":"microsoft-foundry"}],"pricing":{"prompt":"0.000002","completion":"0.000008","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"openai","slug":"gpt-4.1","tags":["multimodal","function-calling","long-context","coding"],"author_info":{"slug":"openai","name":"OpenAI","display_name":null,"icon_url":"/images/logos/openai.webp","gradient_from":"from-zinc-800","gradient_to":"to-zinc-900","gradient_via":null,"website_url":"https://openai.com"}},{"id":"green-l","object":"model","canonical_slug":null,"hugging_face_id":"","name":"GreenL","created":1768905404,"description":"Large generative model optimized for sustainability. Best for general conversations, content creation, and text analysis. Based on Mistral architecture, hosted entirely in the EU.","context_length":128000,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":null,"instruct_type":null},"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","tools","tool_choice"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"GreenPT","slug":"greenpt"}],"pricing":{"prompt":"0.00000025","completion":"0.0000008","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"EUR"},"author":"greenpt","slug":"green-l","tags":["eu-native","sustainable","instruction-tuned"],"author_info":{"slug":"greenpt","name":"GreenPT","display_name":"GreenPT","icon_url":null,"gradient_from":"from-green-600","gradient_to":"to-green-700","gradient_via":null,"website_url":"https://greenpt.ai"}},{"id":"deepseek-v3.1","object":"model","canonical_slug":null,"hugging_face_id":"deepseek-ai/DeepSeek-V3.1","name":"DeepSeek V3.1","created":1774895919,"description":"DeepSeek V3.1 is a mixture-of-experts chat model with strong coding and instruction-following capabilities.","context_length":128000,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"DeepSeek","instruct_type":null},"top_provider":{"context_length":128000,"max_completion_tokens":8000,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop"],"supported_api_endpoints":["chat.completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"AWS Bedrock","slug":"aws-bedrock"}],"pricing":{"prompt":"0.00000058","completion":"0.00000168","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"deepseek","slug":"deepseek-v3.1","tags":["open-source","long-context","coding"],"author_info":{"slug":"deepseek","name":"DeepSeek","display_name":null,"icon_url":"/images/logos/deepseek.png","gradient_from":"from-[#0066ff]","gradient_to":"to-[#00ccff]","gradient_via":null,"website_url":"https://deepseek.com"}},{"id":"gemma-2-2b-it","object":"model","canonical_slug":null,"hugging_face_id":"google/gemma-2-2b-it","name":"Gemma 2 2B IT","created":1768127608,"description":"Gemma 2 2B IT is Google's compact instruction-tuned model from the Gemma 2 family for efficient chat and assistant workloads.","context_length":8192,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"Gemma","instruct_type":null},"top_provider":{"context_length":8192,"max_completion_tokens":8192,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop"],"supported_api_endpoints":["chat.completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Nebius","slug":"nebius"}],"pricing":{"prompt":"0.00000002","completion":"0.00000006","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"google","slug":"gemma-2-2b-it","tags":["open-source","instruction-tuned","efficient"],"author_info":{"slug":"google","name":"Google","display_name":null,"icon_url":"/images/logos/gemini.webp","gradient_from":"from-blue-500","gradient_to":"to-yellow-500","gradient_via":"via-green-500","website_url":"https://deepmind.google"}},{"id":"qwen3.5-9b","object":"model","canonical_slug":null,"hugging_face_id":"Qwen/Qwen3.5-9B","name":"Qwen3.5 9B","created":1777453993,"description":"Qwen3.5 9B is Alibaba's compact Qwen3.5 model for efficient long-context chat, reasoning, and tool use.","context_length":262144,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"Qwen3","instruct_type":null},"top_provider":{"context_length":262144,"max_completion_tokens":262144,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","seed","response_format","tools","tool_choice","reasoning_effort"],"supported_api_endpoints":["chat.completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Tensorix","slug":"tensorix"},{"name":"OVHcloud","slug":"ovhcloud"}],"pricing":{"prompt":"0.0000001","completion":"0.00000015","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"EUR"},"author":"alibaba","slug":"qwen3.5-9b","tags":["open-source","long-context","reasoning","efficient"],"author_info":{"slug":"alibaba","name":"Alibaba","display_name":"Alibaba (Qwen)","icon_url":"/images/logos/qwen.webp","gradient_from":"from-purple-600","gradient_to":"to-indigo-700","gradient_via":null,"website_url":"https://qwenlm.github.io"}},{"id":"mistral-small-24b","object":"model","canonical_slug":null,"hugging_face_id":"mistralai/Mistral-Small-24B-Instruct-2501","name":"Mistral Small 24B Instruct","created":1770318640,"description":"Mistral Small 24B Instruct is a 24B parameter model with 128K context and multimodal (text + image) inputs, optimized for instruction following and long-document understanding.","context_length":128000,"architecture":{"modality":"text+image->text","input_modalities":["text","image"],"output_modalities":["text"],"tokenizer":"Tekken","instruct_type":null},"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","response_format","tools","tool_choice"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"IONOS Cloud","slug":"ionos"}],"pricing":{"prompt":"0.00000011","completion":"0.00000033","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"mistral","slug":"mistral-small-24b","tags":["open-source","instruction-tuned","vision","long-context","multilingual"],"author_info":{"slug":"mistral","name":"Mistral","display_name":"Mistral AI","icon_url":"/images/logos/mistral.jpeg","gradient_from":"from-[#ff7000]","gradient_to":"to-[#ff9a00]","gradient_via":null,"website_url":"https://mistral.ai"}},{"id":"qwen3-coder-480b-a35b-instruct","object":"model","canonical_slug":null,"hugging_face_id":"","name":"Qwen3 Coder 480B A35B Instruct","created":1768127608,"description":"Massive code-specialist model optimized for repo-scale coding, long-context understanding, and precise tool calls.","context_length":262144,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"Qwen3","instruct_type":null},"top_provider":{"context_length":262144,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","tools","tool_choice","response_format"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"AWS Bedrock","slug":"aws-bedrock"}],"pricing":{"prompt":"0.0000004","completion":"0.0000018","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"alibaba","slug":"qwen3-coder-480b-a35b-instruct","tags":["coding","moe","long-context","tool-use"],"author_info":{"slug":"alibaba","name":"Alibaba","display_name":"Alibaba (Qwen)","icon_url":"/images/logos/qwen.webp","gradient_from":"from-purple-600","gradient_to":"to-indigo-700","gradient_via":null,"website_url":"https://qwenlm.github.io"}},{"id":"claude-opus-4-6","object":"model","canonical_slug":null,"hugging_face_id":"","name":"Claude Opus 4.6","created":1771234266,"description":"Claude Opus 4.6 is Anthropic's most capable model for complex coding, enterprise agents, and professional reasoning with a 1M context window.","context_length":1000000,"architecture":{"modality":"text+image->text","input_modalities":["text","image"],"output_modalities":["text"],"tokenizer":"Claude","instruct_type":null},"top_provider":{"context_length":1000000,"max_completion_tokens":128000,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","top_k","max_tokens","stop","tools","tool_choice","response_format"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"AWS Bedrock","slug":"aws-bedrock"}],"pricing":{"prompt":"0.000005","completion":"0.000025","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"anthropic","slug":"claude-opus-4-6","tags":["proprietary","coding","reasoning","agentic","long-context","vision"],"author_info":{"slug":"anthropic","name":"Anthropic","display_name":null,"icon_url":"/images/logos/anthropic.jpeg","gradient_from":"from-[#cc785c]","gradient_to":"to-[#d4a574]","gradient_via":null,"website_url":"https://anthropic.com"}},{"id":"minimax-m2.1","object":"model","canonical_slug":null,"hugging_face_id":"MiniMaxAI/MiniMax-M2.1","name":"MiniMax M2.1","created":1771080675,"description":"MiniMax M2.1 is a long-context Mixture-of-Experts model from MiniMax for chat, coding, reasoning, and multimodal tasks.","context_length":128000,"architecture":{"modality":"text+image->text","input_modalities":["text","image"],"output_modalities":["text"],"tokenizer":"MiniMax","instruct_type":null},"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","tools","tool_choice","response_format"],"supported_api_endpoints":["chat.completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Tensorix","slug":"tensorix"},{"name":"AWS Bedrock","slug":"aws-bedrock"}],"pricing":{"prompt":"0.0000003","completion":"0.0000012","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"minimax","slug":"minimax-m2.1","tags":["open-source","moe","long-context","vision","reasoning"],"author_info":{"slug":"minimax","name":"MiniMax","display_name":"MiniMax AI","icon_url":null,"gradient_from":"from-blue-500","gradient_to":"to-indigo-600","gradient_via":null,"website_url":"https://minimaxi.com"}},{"id":"text-embedding-3-small","object":"model","canonical_slug":null,"hugging_face_id":"","name":"Text Embedding 3 Small","created":1767964798,"description":"text-embedding-3-small is OpenAI's efficient text embedding model with 1,536 output dimensions.","context_length":8192,"architecture":{"modality":"text->embedding","input_modalities":["text"],"output_modalities":["embedding"],"tokenizer":"cl100k_base","instruct_type":null},"top_provider":{"context_length":8192,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["encoding_format","dimensions"],"supported_api_endpoints":["embeddings"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Microsoft Foundry","slug":"microsoft-foundry"}],"author":"openai","slug":"text-embedding-3-small","tags":["embedding","retrieval","efficient"],"author_info":{"slug":"openai","name":"OpenAI","display_name":null,"icon_url":"/images/logos/openai.webp","gradient_from":"from-zinc-800","gradient_to":"to-zinc-900","gradient_via":null,"website_url":"https://openai.com"}},{"id":"mistral-7b-instruct","object":"model","canonical_slug":null,"hugging_face_id":"","name":"Mistral 7B Instruct","created":1777454590,"description":"Mistral 7B Instruct is Mistral AI's 7B-parameter instruction-tuned model.","context_length":32000,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":null,"instruct_type":null},"top_provider":{"context_length":32000,"max_completion_tokens":4000,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"AWS Bedrock","slug":"aws-bedrock"}],"pricing":{"prompt":"0.00000016","completion":"0.00000022","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"mistral","slug":"mistral-7b-instruct","tags":["open-weights","instruction-tuned"],"author_info":{"slug":"mistral","name":"Mistral","display_name":"Mistral AI","icon_url":"/images/logos/mistral.jpeg","gradient_from":"from-[#ff7000]","gradient_to":"to-[#ff9a00]","gradient_via":null,"website_url":"https://mistral.ai"}},{"id":"nemotron-nano-v2-12b","object":"model","canonical_slug":null,"hugging_face_id":"","name":"Nemotron Nano V2 12B","created":1768127608,"description":"Compact NVIDIA model optimized for efficient reasoning, safety, and enterprise workloads.","context_length":32000,"architecture":{"modality":"text+image->text","input_modalities":["text","image"],"output_modalities":["text"],"tokenizer":null,"instruct_type":null},"top_provider":{"context_length":32000,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","tools","tool_choice","response_format"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[],"pricing":{"prompt":"0.00000007","completion":"0.0000002","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"nvidia","slug":"nemotron-nano-v2-12b","tags":["vision","enterprise","efficient"],"author_info":{"slug":"nvidia","name":"NVIDIA","display_name":null,"icon_url":"/images/logos/nvidia.webp","gradient_from":"from-[#76b900]","gradient_to":"to-[#5a8c00]","gradient_via":null,"website_url":"https://nvidia.com"}},{"id":"mistral-small-3.2-24b","object":"model","canonical_slug":null,"hugging_face_id":"mistralai/Mistral-Small-3.2-24B-Instruct-2506","name":"Mistral Small 3.2 24B","created":1765183068,"description":"Mistral Small 3.2 is a 24B open model with image understanding, function calling, and long-context support.","context_length":131072,"architecture":{"modality":"text+image->text","input_modalities":["text","image"],"output_modalities":["text"],"tokenizer":"Tekken","instruct_type":null},"top_provider":{"context_length":131072,"max_completion_tokens":131072,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","seed","response_format","tools","tool_choice"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"GreenPT","slug":"greenpt"},{"name":"Scaleway","slug":"scaleway"},{"name":"OVHcloud","slug":"ovhcloud"},{"name":"Mistral AI","slug":"mistral"}],"pricing":{"prompt":"0.00000009","completion":"0.00000028","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"EUR"},"author":"mistral","slug":"mistral-small-3.2-24b","tags":["open-source","vision","multilingual","function-calling"],"author_info":{"slug":"mistral","name":"Mistral","display_name":"Mistral AI","icon_url":"/images/logos/mistral.jpeg","gradient_from":"from-[#ff7000]","gradient_to":"to-[#ff9a00]","gradient_via":null,"website_url":"https://mistral.ai"}},{"id":"mistral-7b-instruct-v0.3","object":"model","canonical_slug":null,"hugging_face_id":"mistralai/Mistral-7B-Instruct-v0.3","name":"Mistral 7B Instruct v0.3","created":1769169016,"description":"Mistral 7B Instruct v0.3 is Mistral AI's 7B instruction-following model for multilingual text generation, coding, extraction, and summarization tasks.","context_length":32768,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"Tekken","instruct_type":null},"top_provider":{"context_length":32768,"max_completion_tokens":32768,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","seed","response_format","tools","tool_choice"],"supported_api_endpoints":["chat.completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"OVHcloud","slug":"ovhcloud"}],"pricing":{"prompt":"0.0000001","completion":"0.0000001","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"EUR"},"author":"mistral","slug":"mistral-7b-instruct-v0.3","tags":["open-source","instruction-tuned","multilingual","coding"],"author_info":{"slug":"mistral","name":"Mistral","display_name":"Mistral AI","icon_url":"/images/logos/mistral.jpeg","gradient_from":"from-[#ff7000]","gradient_to":"to-[#ff9a00]","gradient_via":null,"website_url":"https://mistral.ai"}},{"id":"ministral-3-3b","object":"model","canonical_slug":null,"hugging_face_id":"","name":"Ministral 3 3B","created":1767979257,"description":"Ministral 3 3B is an open Mistral AI model for efficient text and vision workloads.","context_length":262144,"architecture":{"modality":"text+image->text","input_modalities":["text","image"],"output_modalities":["text"],"tokenizer":"Tekken","instruct_type":null},"top_provider":{"context_length":262144,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","random_seed","response_format","tools","tool_choice"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Mistral AI","slug":"mistral"}],"pricing":{"prompt":"0.0000001","completion":"0.0000001","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"mistral","slug":"ministral-3-3b","tags":["open-source","vision","lightweight","long-context"],"author_info":{"slug":"mistral","name":"Mistral","display_name":"Mistral AI","icon_url":"/images/logos/mistral.jpeg","gradient_from":"from-[#ff7000]","gradient_to":"to-[#ff9a00]","gradient_via":null,"website_url":"https://mistral.ai"}},{"id":"cohere-embed-multilingual-v3","object":"model","canonical_slug":null,"hugging_face_id":"","name":"Cohere Embed Multilingual v3","created":1777454590,"description":"Cohere Embed Multilingual v3 is a multilingual text embedding model available through Amazon Bedrock.","context_length":512,"architecture":{"modality":"text->embedding","input_modalities":["text"],"output_modalities":["embedding"],"tokenizer":null,"instruct_type":null},"top_provider":{"context_length":512,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":[],"supported_api_endpoints":["embeddings"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"AWS Bedrock","slug":"aws-bedrock"}],"pricing":{"prompt":"0.0000001","completion":"0","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"cohere","slug":"cohere-embed-multilingual-v3","tags":["proprietary","embedding","multilingual"],"author_info":{"slug":"cohere","name":"Cohere","display_name":null,"icon_url":"/images/logos/cohere.webp","gradient_from":"from-purple-600","gradient_to":"to-purple-800","gradient_via":null,"website_url":"https://cohere.com"}},{"id":"codestral","object":"model","canonical_slug":null,"hugging_face_id":"","name":"Codestral","created":1767979257,"description":"Codestral is Mistral AI's code model for code completion, code generation, and instruction-following coding tasks.","context_length":131072,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"Tekken","instruct_type":null},"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","random_seed","response_format","tools","tool_choice","suffix"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Mistral AI","slug":"mistral"}],"pricing":{"prompt":"0","completion":"0.0000009","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"mistral","slug":"codestral","tags":["code","fim","function-calling"],"author_info":{"slug":"mistral","name":"Mistral","display_name":"Mistral AI","icon_url":"/images/logos/mistral.jpeg","gradient_from":"from-[#ff7000]","gradient_to":"to-[#ff9a00]","gradient_via":null,"website_url":"https://mistral.ai"}},{"id":"claude-3.7-sonnet","object":"model","canonical_slug":null,"hugging_face_id":"","name":"Claude 3.7 Sonnet","created":1769967813,"description":"Claude 3.7 Sonnet offers excellent performance with strong reasoning capabilities. Legacy model - migrate to Claude Sonnet 4.5.","context_length":200000,"architecture":{"modality":"text+image->text","input_modalities":["text","image"],"output_modalities":["text"],"tokenizer":"claude","instruct_type":"claude"},"top_provider":{"context_length":200000,"max_completion_tokens":8192,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","tools","tool_choice"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[],"pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"anthropic","slug":"claude-3.7-sonnet","tags":["proprietary","long-context","vision","legacy"],"author_info":{"slug":"anthropic","name":"Anthropic","display_name":null,"icon_url":"/images/logos/anthropic.jpeg","gradient_from":"from-[#cc785c]","gradient_to":"to-[#d4a574]","gradient_via":null,"website_url":"https://anthropic.com"}},{"id":"deepseek-r1","object":"model","canonical_slug":null,"hugging_face_id":"deepseek-ai/DeepSeek-R1-0528","name":"DeepSeek R1","created":1765024302,"description":"DeepSeek R1 is DeepSeek's reasoning model for complex problem solving, math, code, and multi-step analysis.","context_length":164000,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"DeepSeek","instruct_type":null},"top_provider":{"context_length":164000,"max_completion_tokens":8192,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","tools","tool_choice"],"supported_api_endpoints":["chat.completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Tensorix","slug":"tensorix"}],"pricing":{"prompt":"0.00000066","completion":"0.0000026","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"deepseek","slug":"deepseek-r1","tags":["open-source","reasoning","long-context"],"author_info":{"slug":"deepseek","name":"DeepSeek","display_name":null,"icon_url":"/images/logos/deepseek.png","gradient_from":"from-[#0066ff]","gradient_to":"to-[#00ccff]","gradient_via":null,"website_url":"https://deepseek.com"}},{"id":"voxtral-mini","object":"model","canonical_slug":null,"hugging_face_id":"","name":"Voxtral Mini","created":1768905491,"description":"Low-latency speech recognition for edge and devices. Mini version of audio input model optimized for speed.","context_length":32000,"architecture":{"modality":"audio+text->text","input_modalities":["audio","text"],"output_modalities":["text"],"tokenizer":"Tekken","instruct_type":null},"top_provider":{"context_length":32000,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","tools","tool_choice","response_format","safe_prompt"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[],"pricing":{"prompt":"0.00000004","completion":"0.00000004","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"mistral","slug":"voxtral-mini","tags":["open-source","audio","edge","low-latency"],"author_info":{"slug":"mistral","name":"Mistral","display_name":"Mistral AI","icon_url":"/images/logos/mistral.jpeg","gradient_from":"from-[#ff7000]","gradient_to":"to-[#ff9a00]","gradient_via":null,"website_url":"https://mistral.ai"}},{"id":"qwen3-30b-a3b-instruct","object":"model","canonical_slug":null,"hugging_face_id":"Qwen/Qwen3-30B-A3B-Instruct-2507","name":"Qwen3 30B A3B Instruct","created":1767979257,"description":"Qwen3 30B A3B Instruct is an Alibaba/Qwen compact MoE instruction model for chat, code, tool use, and context-augmented generation.","context_length":262144,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"Qwen3","instruct_type":null},"top_provider":{"context_length":262144,"max_completion_tokens":262144,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","response_format","tools","tool_choice"],"supported_api_endpoints":["chat.completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Nebius","slug":"nebius"}],"pricing":{"prompt":"0.0000001","completion":"0.0000003","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"alibaba","slug":"qwen3-30b-a3b-instruct","tags":["open-source","moe","long-context","coding","function-calling"],"author_info":{"slug":"alibaba","name":"Alibaba","display_name":"Alibaba (Qwen)","icon_url":"/images/logos/qwen.webp","gradient_from":"from-purple-600","gradient_to":"to-indigo-700","gradient_via":null,"website_url":"https://qwenlm.github.io"}},{"id":"nova-micro","object":"model","canonical_slug":null,"hugging_face_id":"","name":"Amazon Nova Micro","created":1769967814,"description":"Amazon Nova Micro is the fastest and most cost-effective Nova model, optimized for simple text tasks.","context_length":128000,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":null,"instruct_type":null},"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"AWS Bedrock","slug":"aws-bedrock"}],"pricing":{"prompt":"0.000000035","completion":"0.00000014","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"amazon","slug":"nova-micro","tags":["proprietary","fast","cost-efficient"],"author_info":{"slug":"amazon","name":"Amazon","display_name":"Amazon Web Services","icon_url":"/images/logos/aws.webp","gradient_from":"from-orange-500","gradient_to":"to-yellow-500","gradient_via":null,"website_url":"https://aws.amazon.com/bedrock"}},{"id":"gpt-5.5","object":"model","canonical_slug":null,"hugging_face_id":"","name":"GPT-5.5","created":1777454744,"description":"GPT-5.5 is OpenAI's frontier reasoning model for professional agentic workflows, long-context analysis, coding, and computer-use tasks.","context_length":1050000,"architecture":{"modality":"text+image->text","input_modalities":["text","image"],"output_modalities":["text"],"tokenizer":"GPT","instruct_type":null},"top_provider":{"context_length":1050000,"max_completion_tokens":128000,"is_moderated":false},"per_request_limits":null,"supported_parameters":["max_tokens","max_completion_tokens","reasoning_effort","tools","tool_choice","response_format","parallel_tool_calls"],"supported_api_endpoints":["chat.completions","responses"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[],"pricing":{"prompt":"0.0000055","completion":"0.000033","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000055","input_cache_write":"0","discount":1,"currency":"USD"},"author":"openai","slug":"gpt-5.5","tags":["reasoning","long-context","frontier","agentic","coding","computer-use","vision"],"author_info":{"slug":"openai","name":"OpenAI","display_name":null,"icon_url":"/images/logos/openai.webp","gradient_from":"from-zinc-800","gradient_to":"to-zinc-900","gradient_via":null,"website_url":"https://openai.com"}},{"id":"claude-sonnet-4-6","object":"model","canonical_slug":null,"hugging_face_id":"","name":"Claude Sonnet 4.6","created":1773759863,"description":"Claude Sonnet 4.6 is Anthropic's balanced performance model for coding, reasoning, and agentic workflows with a 1M context window.","context_length":1000000,"architecture":{"modality":"text+image->text","input_modalities":["text","image"],"output_modalities":["text"],"tokenizer":"Claude","instruct_type":null},"top_provider":{"context_length":1000000,"max_completion_tokens":64000,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","top_k","max_tokens","stop","tools","tool_choice","response_format"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"AWS Bedrock","slug":"aws-bedrock"}],"pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000003","input_cache_write":"0.00000375","discount":1,"currency":"USD"},"author":"anthropic","slug":"claude-sonnet-4-6","tags":["proprietary","coding","reasoning","agentic","long-context","vision"],"author_info":{"slug":"anthropic","name":"Anthropic","display_name":null,"icon_url":"/images/logos/anthropic.jpeg","gradient_from":"from-[#cc785c]","gradient_to":"to-[#d4a574]","gradient_via":null,"website_url":"https://anthropic.com"}},{"id":"green-r-raw","object":"model","canonical_slug":null,"hugging_face_id":"","name":"GreenR Raw","created":1768905404,"description":"Reasoning model with custom system prompt support. Combines advanced reasoning capabilities with customizable behavior. EU-hosted.","context_length":128000,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":null,"instruct_type":null},"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","tools","tool_choice","response_format"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"GreenPT","slug":"greenpt"}],"pricing":{"prompt":"0.00000035","completion":"0.00000095","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"EUR"},"author":"greenpt","slug":"green-r-raw","tags":["eu-native","sustainable","reasoning","raw"],"author_info":{"slug":"greenpt","name":"GreenPT","display_name":"GreenPT","icon_url":null,"gradient_from":"from-green-600","gradient_to":"to-green-700","gradient_via":null,"website_url":"https://greenpt.ai"}},{"id":"cohere-embed-v4","object":"model","canonical_slug":null,"hugging_face_id":"","name":"Cohere Embed v4","created":1769967814,"description":"Cohere Embed v4 is a state-of-the-art embedding model supporting 100+ languages with best-in-class retrieval performance.","context_length":512,"architecture":{"modality":"text->embedding","input_modalities":["text"],"output_modalities":["embedding"],"tokenizer":null,"instruct_type":null},"top_provider":{"context_length":512,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":[],"supported_api_endpoints":["embeddings"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"AWS Bedrock","slug":"aws-bedrock"}],"pricing":{"prompt":"0.0000001","completion":"0","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"cohere","slug":"cohere-embed-v4","tags":["proprietary","embedding","multilingual","retrieval"],"author_info":{"slug":"cohere","name":"Cohere","display_name":null,"icon_url":"/images/logos/cohere.webp","gradient_from":"from-purple-600","gradient_to":"to-purple-800","gradient_via":null,"website_url":"https://cohere.com"}},{"id":"qwen3-next-80b-a3b-thinking","object":"model","canonical_slug":null,"hugging_face_id":"Qwen/Qwen3-Next-80B-A3B-Thinking","name":"Qwen3 Next 80B A3B Thinking","created":1768127608,"description":"Qwen3 Next 80B A3B Thinking is an Alibaba/Qwen thinking model for reasoning, code, context-augmented generation, and tool use.","context_length":128000,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"Qwen3","instruct_type":null},"top_provider":{"context_length":128000,"max_completion_tokens":128000,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","response_format","tools","tool_choice"],"supported_api_endpoints":["chat.completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Nebius","slug":"nebius"}],"pricing":{"prompt":"0.00000015","completion":"0.0000012","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"alibaba","slug":"qwen3-next-80b-a3b-thinking","tags":["open-source","reasoning","coding","function-calling"],"author_info":{"slug":"alibaba","name":"Alibaba","display_name":"Alibaba (Qwen)","icon_url":"/images/logos/qwen.webp","gradient_from":"from-purple-600","gradient_to":"to-indigo-700","gradient_via":null,"website_url":"https://qwenlm.github.io"}},{"id":"codestral-embed","object":"model","canonical_slug":null,"hugging_face_id":"","name":"Codestral Embed","created":1767887445,"description":"Codestral Embed is Mistral AI's code embedding model for semantic code search and retrieval.","context_length":8192,"architecture":{"modality":"text->embedding","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"Tekken","instruct_type":null},"top_provider":{"context_length":8192,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["encoding_format","dimensions","output_dimension","output_dtype"],"supported_api_endpoints":["embeddings"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Mistral AI","slug":"mistral"}],"pricing":{"prompt":"0.00000015","completion":"0","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"mistral","slug":"codestral-embed","tags":["embedding","code","retrieval","semantic-search"],"author_info":{"slug":"mistral","name":"Mistral","display_name":"Mistral AI","icon_url":"/images/logos/mistral.jpeg","gradient_from":"from-[#ff7000]","gradient_to":"to-[#ff9a00]","gradient_via":null,"website_url":"https://mistral.ai"}},{"id":"magistral-medium","object":"model","canonical_slug":null,"hugging_face_id":"","name":"Magistral Medium 1.2","created":1767979257,"description":"Magistral Medium 1.2 is Mistral AI's premier reasoning model for domain-specific, transparent, and multilingual reasoning.","context_length":40000,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"Tekken","instruct_type":null},"top_provider":{"context_length":40000,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","random_seed","response_format","tools","tool_choice","reasoning_effort"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Mistral AI","slug":"mistral"}],"pricing":{"prompt":"0.000002","completion":"0.000005","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"mistral","slug":"magistral-medium","tags":["premier","reasoning","multilingual"],"author_info":{"slug":"mistral","name":"Mistral","display_name":"Mistral AI","icon_url":"/images/logos/mistral.jpeg","gradient_from":"from-[#ff7000]","gradient_to":"to-[#ff9a00]","gradient_via":null,"website_url":"https://mistral.ai"}},{"id":"mistral-large-2","object":"model","canonical_slug":null,"hugging_face_id":"mistralai/Mistral-Large-Instruct-2407","name":"Mistral Large 2","created":1765024302,"description":"Mistral Large 2 is a 123B parameter flagship model with a 128K context window. It excels at code generation, mathematics, reasoning, and multi-turn conversations. It performs on par with GPT-4o and Claude 3 Opus on most benchmarks.","context_length":128000,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"Tekken","instruct_type":null},"top_provider":{"context_length":128000,"max_completion_tokens":4096,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","tools","tool_choice","response_format","safe_prompt"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[],"pricing":{"prompt":"0.000002","completion":"0.000006","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"mistral","slug":"mistral-large-2","tags":["flagship","instruction-tuned","multilingual","long-context","coding","math"],"author_info":{"slug":"mistral","name":"Mistral","display_name":"Mistral AI","icon_url":"/images/logos/mistral.jpeg","gradient_from":"from-[#ff7000]","gradient_to":"to-[#ff9a00]","gradient_via":null,"website_url":"https://mistral.ai"}},{"id":"claude-opus-4-5-20251101","object":"model","canonical_slug":null,"hugging_face_id":"","name":"Claude Opus 4.5","created":1769967813,"description":"Claude Opus 4.5 is Anthropic's most capable model for complex tasks requiring deep analysis, nuanced understanding, and sophisticated reasoning.","context_length":200000,"architecture":{"modality":"text+image->text","input_modalities":["text","image"],"output_modalities":["text"],"tokenizer":"claude","instruct_type":"claude"},"top_provider":{"context_length":200000,"max_completion_tokens":8192,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","tools","tool_choice"],"supported_api_endpoints":["chat.completions","messages","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"AWS Bedrock","slug":"aws-bedrock"}],"pricing":{"prompt":"0.000005","completion":"0.000025","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"anthropic","slug":"claude-opus-4-5-20251101","tags":["proprietary","long-context","vision","reasoning","premium"],"author_info":{"slug":"anthropic","name":"Anthropic","display_name":null,"icon_url":"/images/logos/anthropic.jpeg","gradient_from":"from-[#cc785c]","gradient_to":"to-[#d4a574]","gradient_via":null,"website_url":"https://anthropic.com"}},{"id":"glm-4.6","object":"model","canonical_slug":null,"hugging_face_id":"THUDM/GLM-4.6","name":"GLM 4.6","created":1774895903,"description":"GLM-4.6 is Z AI's language model for multilingual chat, coding, reasoning, and function calling.","context_length":203000,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"ChatGLM","instruct_type":null},"top_provider":{"context_length":203000,"max_completion_tokens":131000,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","tools","tool_choice","response_format"],"supported_api_endpoints":["chat.completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Tensorix","slug":"tensorix"}],"pricing":{"prompt":"0.0000004","completion":"0.00000175","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"zhipu","slug":"glm-4.6","tags":["open-source","multilingual","long-context","reasoning"],"author_info":{"slug":"zhipu","name":"Zhipu","display_name":"Zhipu AI","icon_url":null,"gradient_from":"from-blue-500","gradient_to":"to-cyan-500","gradient_via":null,"website_url":"https://www.zhipuai.cn"}},{"id":"mistral-medium-3.1","object":"model","canonical_slug":null,"hugging_face_id":"","name":"Mistral Medium 3.1","created":1768905491,"description":"Mistral Medium 3.1 is Mistral AI's frontier-class multimodal model released in August 2025.","context_length":131072,"architecture":{"modality":"text+image->text","input_modalities":["text","image"],"output_modalities":["text"],"tokenizer":"Tekken","instruct_type":null},"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","random_seed","response_format","tools","tool_choice"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Mistral AI","slug":"mistral"}],"pricing":{"prompt":"0.0000004","completion":"0.000002","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"mistral","slug":"mistral-medium-3.1","tags":["premier","multimodal","vision","function-calling"],"author_info":{"slug":"mistral","name":"Mistral","display_name":"Mistral AI","icon_url":"/images/logos/mistral.jpeg","gradient_from":"from-[#ff7000]","gradient_to":"to-[#ff9a00]","gradient_via":null,"website_url":"https://mistral.ai"}},{"id":"intellect-3","object":"model","canonical_slug":null,"hugging_face_id":"PrimeIntellect/INTELLECT-3","name":"INTELLECT-3","created":1768127608,"description":"Prime Intellect INTELLECT-3 is an open model for reasoning, code, tool-use, and context-augmented chat workloads.","context_length":131072,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"Llama3","instruct_type":null},"top_provider":{"context_length":131072,"max_completion_tokens":131072,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","response_format","tools","tool_choice"],"supported_api_endpoints":["chat.completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Nebius","slug":"nebius"}],"pricing":{"prompt":"0.0000002","completion":"0.0000011","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"primeintellect","slug":"intellect-3","tags":["open-source","reasoning","coding","function-calling"],"author_info":{"slug":"primeintellect","name":"Prime Intellect","display_name":"Prime Intellect","icon_url":null,"gradient_from":"from-emerald-500","gradient_to":"to-teal-600","gradient_via":null,"website_url":"https://primeintellect.ai"}},{"id":"gemma-2-9b-it","object":"model","canonical_slug":null,"hugging_face_id":"","name":"Gemma 2 9B Instruct","created":1768127608,"description":"Larger Gemma model delivering high-quality chat and coding with efficient inference.","context_length":8192,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"Gemma","instruct_type":null},"top_provider":{"context_length":8192,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[],"pricing":{"prompt":"0.00000003","completion":"0.00000009","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"google","slug":"gemma-2-9b-it","tags":["open-source","coding","instruction-tuned"],"author_info":{"slug":"google","name":"Google","display_name":null,"icon_url":"/images/logos/gemini.webp","gradient_from":"from-blue-500","gradient_to":"to-yellow-500","gradient_via":"via-green-500","website_url":"https://deepmind.google"}},{"id":"nova-2-lite","object":"model","canonical_slug":null,"hugging_face_id":"","name":"Amazon Nova 2 Lite","created":1777454590,"description":"Amazon Nova 2 Lite is a cost-efficient multimodal model for simple automation, document processing, and customer support across text, images, and video.","context_length":1000000,"architecture":{"modality":"text+image+video->text","input_modalities":["text","image","video"],"output_modalities":["text"],"tokenizer":null,"instruct_type":null},"top_provider":{"context_length":1000000,"max_completion_tokens":64000,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","tools","tool_choice"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"AWS Bedrock","slug":"aws-bedrock"}],"pricing":{"prompt":"0.00000039","completion":"0.00000327","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000000975","input_cache_write":"0","discount":1,"currency":"USD"},"author":"amazon","slug":"nova-2-lite","tags":["proprietary","multimodal","vision","video","long-context","prompt-caching"],"author_info":{"slug":"amazon","name":"Amazon","display_name":"Amazon Web Services","icon_url":"/images/logos/aws.webp","gradient_from":"from-orange-500","gradient_to":"to-yellow-500","gradient_via":null,"website_url":"https://aws.amazon.com/bedrock"}},{"id":"gpt-4o","object":"model","canonical_slug":null,"hugging_face_id":"","name":"GPT-4o","created":1768138176,"description":"GPT-4o is OpenAI's multimodal model for text and image input, text output, tool use, and general-purpose chat workloads.","context_length":128000,"architecture":{"modality":"text+image->text","input_modalities":["text","image"],"output_modalities":["text"],"tokenizer":"GPT","instruct_type":null},"top_provider":{"context_length":128000,"max_completion_tokens":16384,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","max_completion_tokens","stop","tools","tool_choice","response_format","seed","parallel_tool_calls"],"supported_api_endpoints":["chat.completions","responses"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Microsoft Foundry","slug":"microsoft-foundry"}],"pricing":{"prompt":"0.0000025","completion":"0.00001","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"openai","slug":"gpt-4o","tags":["multimodal","function-calling","vision","general-purpose"],"author_info":{"slug":"openai","name":"OpenAI","display_name":null,"icon_url":"/images/logos/openai.webp","gradient_from":"from-zinc-800","gradient_to":"to-zinc-900","gradient_via":null,"website_url":"https://openai.com"}},{"id":"pixtral-large-2502","object":"model","canonical_slug":null,"hugging_face_id":"","name":"Pixtral Large 25.02","created":1769967814,"description":"Pixtral Large is Mistral's flagship multimodal model with exceptional vision and text understanding capabilities.","context_length":128000,"architecture":{"modality":"text+image->text","input_modalities":["text","image"],"output_modalities":["text"],"tokenizer":"mistral","instruct_type":"mistral"},"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","tools","tool_choice"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"AWS Bedrock","slug":"aws-bedrock"}],"pricing":{"prompt":"0.000002","completion":"0.000006","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"mistral","slug":"pixtral-large-2502","tags":["proprietary","vision","multimodal","long-context"],"author_info":{"slug":"mistral","name":"Mistral","display_name":"Mistral AI","icon_url":"/images/logos/mistral.jpeg","gradient_from":"from-[#ff7000]","gradient_to":"to-[#ff9a00]","gradient_via":null,"website_url":"https://mistral.ai"}},{"id":"qwen3-vl-235b-a22b-instruct","object":"model","canonical_slug":null,"hugging_face_id":"Qwen/Qwen3-VL-235B-A22B-Instruct","name":"Qwen3 VL 235B A22B Instruct","created":1774895903,"description":"Qwen3 VL 235B A22B Instruct is Alibaba's vision-language Mixture-of-Experts model for multimodal understanding and chat.","context_length":131000,"architecture":{"modality":"text+image->text","input_modalities":["text","image"],"output_modalities":["text"],"tokenizer":"Qwen3","instruct_type":null},"top_provider":{"context_length":131000,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop"],"supported_api_endpoints":["chat.completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Tensorix","slug":"tensorix"}],"pricing":{"prompt":"0.00000021","completion":"0.0000019","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"alibaba","slug":"qwen3-vl-235b-a22b-instruct","tags":["open-source","moe","vision","multilingual"],"author_info":{"slug":"alibaba","name":"Alibaba","display_name":"Alibaba (Qwen)","icon_url":"/images/logos/qwen.webp","gradient_from":"from-purple-600","gradient_to":"to-indigo-700","gradient_via":null,"website_url":"https://qwenlm.github.io"}},{"id":"mixtral-8x22b-instruct","object":"model","canonical_slug":null,"hugging_face_id":"mistralai/Mixtral-8x22B-Instruct-v0.1","name":"Mixtral 8x22B Instruct","created":1774895903,"description":"Mixtral 8x22B Instruct is Mistral AI's sparse Mixture-of-Experts model for chat, code, and tool-use workloads.","context_length":66000,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"Tekken","instruct_type":null},"top_provider":{"context_length":66000,"max_completion_tokens":65536,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","tools","tool_choice","response_format"],"supported_api_endpoints":["chat.completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Tensorix","slug":"tensorix"}],"pricing":{"prompt":"0.000002","completion":"0.000006","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"mistral","slug":"mixtral-8x22b-instruct","tags":["open-source","moe","coding","function-calling"],"author_info":{"slug":"mistral","name":"Mistral","display_name":"Mistral AI","icon_url":"/images/logos/mistral.jpeg","gradient_from":"from-[#ff7000]","gradient_to":"to-[#ff9a00]","gradient_via":null,"website_url":"https://mistral.ai"}},{"id":"gpt-5.4","object":"model","canonical_slug":null,"hugging_face_id":"","name":"GPT-5.4","created":1777454744,"description":"GPT-5.4 is OpenAI's frontier reasoning model for complex professional work, long-context analysis, coding, and agentic automation.","context_length":1050000,"architecture":{"modality":"text+image->text","input_modalities":["text","image"],"output_modalities":["text"],"tokenizer":"GPT","instruct_type":null},"top_provider":{"context_length":1050000,"max_completion_tokens":128000,"is_moderated":false},"per_request_limits":null,"supported_parameters":["max_tokens","max_completion_tokens","reasoning_effort","tools","tool_choice","response_format","parallel_tool_calls"],"supported_api_endpoints":["chat.completions","responses"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Microsoft Foundry","slug":"microsoft-foundry"}],"pricing":{"prompt":"0.00000275","completion":"0.0000165","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.000000275","input_cache_write":"0","discount":1,"currency":"USD"},"author":"openai","slug":"gpt-5.4","tags":["reasoning","long-context","frontier","agentic","coding","computer-use","vision"],"author_info":{"slug":"openai","name":"OpenAI","display_name":null,"icon_url":"/images/logos/openai.webp","gradient_from":"from-zinc-800","gradient_to":"to-zinc-900","gradient_via":null,"website_url":"https://openai.com"}},{"id":"e5-mistral-7b-instruct","object":"model","canonical_slug":null,"hugging_face_id":"intfloat/e5-mistral-7b-instruct","name":"E5-Mistral-7B-Instruct","created":1767963057,"description":"Instruction-tuned embedding model from intfloat based on Mistral 7B. Produces 4096-dimensional embeddings for high-quality semantic search and retrieval tasks. Supports instruction-based queries for improved retrieval accuracy.","context_length":33000,"architecture":{"modality":"text->embedding","input_modalities":["text"],"output_modalities":["embedding"],"tokenizer":"Tekken","instruct_type":null},"top_provider":{"context_length":33000,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":[],"supported_api_endpoints":["embeddings"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[],"author":"intfloat","slug":"e5-mistral-7b-instruct","tags":["embedding","retrieval","rag","instruction-tuned"],"author_info":{"slug":"intfloat","name":"intfloat","display_name":null,"icon_url":null,"gradient_from":"from-sky-500","gradient_to":"to-blue-600","gradient_via":null,"website_url":"https://huggingface.co/intfloat"}},{"id":"titan-embed-text-v1","object":"model","canonical_slug":null,"hugging_face_id":"","name":"Amazon Titan Embeddings G1 - Text","created":1777454590,"description":"Amazon Titan Embeddings G1 - Text is a text embedding model available through Amazon Bedrock.","context_length":8000,"architecture":{"modality":"text->embedding","input_modalities":["text"],"output_modalities":["embedding"],"tokenizer":null,"instruct_type":null},"top_provider":{"context_length":8000,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":[],"supported_api_endpoints":["embeddings"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"AWS Bedrock","slug":"aws-bedrock"}],"pricing":{"prompt":"0.0000002","completion":"0","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"amazon","slug":"titan-embed-text-v1","tags":["proprietary","embedding"],"author_info":{"slug":"amazon","name":"Amazon","display_name":"Amazon Web Services","icon_url":"/images/logos/aws.webp","gradient_from":"from-orange-500","gradient_to":"to-yellow-500","gradient_via":null,"website_url":"https://aws.amazon.com/bedrock"}},{"id":"llama-3.1-nemotron-ultra-253b","object":"model","canonical_slug":null,"hugging_face_id":"nvidia/Llama-3_1-Nemotron-Ultra-253B-v1","name":"Llama 3.1 Nemotron Ultra 253B","created":1771080676,"description":"NVIDIA Llama 3.1 Nemotron Ultra 253B is a reasoning-oriented model based on Llama 3.1 for chat, coding, and tool-use workloads.","context_length":131072,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"Llama3","instruct_type":null},"top_provider":{"context_length":131072,"max_completion_tokens":131072,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","response_format","tools","tool_choice"],"supported_api_endpoints":["chat.completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Nebius","slug":"nebius"}],"pricing":{"prompt":"0.0000006","completion":"0.0000018","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"nvidia","slug":"llama-3.1-nemotron-ultra-253b","tags":["open-source","reasoning","coding","function-calling"],"author_info":{"slug":"nvidia","name":"NVIDIA","display_name":null,"icon_url":"/images/logos/nvidia.webp","gradient_from":"from-[#76b900]","gradient_to":"to-[#5a8c00]","gradient_via":null,"website_url":"https://nvidia.com"}},{"id":"paraphrase-multilingual-mpnet-base-v2","object":"model","canonical_slug":null,"hugging_face_id":"sentence-transformers/paraphrase-multilingual-mpnet-base-v2","name":"Paraphrase Multilingual MPNet v2","created":1770318640,"description":"Paraphrase Multilingual MPNet v2 is a sentence-transformer embedding model producing 768-dimensional vectors for multilingual semantic search and similarity.","context_length":128,"architecture":{"modality":"text->embedding","input_modalities":["text"],"output_modalities":["embedding"],"tokenizer":null,"instruct_type":null},"top_provider":{"context_length":128,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["encoding_format"],"supported_api_endpoints":["embeddings"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"IONOS Cloud","slug":"ionos"}],"pricing":{"prompt":"0.00000001","completion":"0","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"sentence-transformers","slug":"paraphrase-multilingual-mpnet-base-v2","tags":["embedding","multilingual"],"author_info":{"slug":"sentence-transformers","name":"Sentence Transformers","display_name":"Sentence Transformers","icon_url":null,"gradient_from":null,"gradient_to":null,"gradient_via":null,"website_url":"https://www.sbert.net"}},{"id":"claude-opus-4-7","object":"model","canonical_slug":null,"hugging_face_id":"","name":"Claude Opus 4.7","created":1777454590,"description":"Claude Opus 4.7 is Anthropic's frontier model for complex coding, reasoning, and agentic workflows.","context_length":1000000,"architecture":{"modality":"text+image->text","input_modalities":["text","image"],"output_modalities":["text"],"tokenizer":"Claude","instruct_type":null},"top_provider":{"context_length":1000000,"max_completion_tokens":128000,"is_moderated":false},"per_request_limits":null,"supported_parameters":["top_p","top_k","max_tokens","stop","tools","tool_choice","response_format"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"AWS Bedrock","slug":"aws-bedrock"}],"pricing":{"prompt":"0.000005","completion":"0.000025","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"anthropic","slug":"claude-opus-4-7","tags":["proprietary","coding","reasoning","agentic","vision"],"author_info":{"slug":"anthropic","name":"Anthropic","display_name":null,"icon_url":"/images/logos/anthropic.jpeg","gradient_from":"from-[#cc785c]","gradient_to":"to-[#d4a574]","gradient_via":null,"website_url":"https://anthropic.com"}},{"id":"glm-4.7-fp8","object":"model","canonical_slug":null,"hugging_face_id":"THUDM/GLM-4.7","name":"GLM 4.7 FP8","created":1771080676,"description":"GLM-4.7 FP8 is the latest iteration of Zhipu AI's GLM series, served in FP8 precision for efficient inference. Improved capabilities over GLM-4.5 with 128K context and strong reasoning.","context_length":128000,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"ChatGLM","instruct_type":null},"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","tools","tool_choice","response_format"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[],"pricing":{"prompt":"0.0000004","completion":"0.000002","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"zhipu","slug":"glm-4.7-fp8","tags":["open-source","multilingual","long-context"],"author_info":{"slug":"zhipu","name":"Zhipu","display_name":"Zhipu AI","icon_url":null,"gradient_from":"from-blue-500","gradient_to":"to-cyan-500","gradient_via":null,"website_url":"https://www.zhipuai.cn"}},{"id":"gpt-5.1","object":"model","canonical_slug":"openai/gpt-5.1-20251113","hugging_face_id":"","name":"GPT-5.1","created":1764005632,"description":"GPT-5.1 is OpenAI's adaptive reasoning model with multimodal input, structured outputs, tool use, and long-context support.","context_length":400000,"architecture":{"modality":"text+image->text","input_modalities":["text","image"],"output_modalities":["text"],"tokenizer":"GPT","instruct_type":null},"top_provider":{"context_length":400000,"max_completion_tokens":128000,"is_moderated":false},"per_request_limits":null,"supported_parameters":["max_tokens","max_completion_tokens","reasoning_effort","tools","tool_choice","response_format","parallel_tool_calls"],"supported_api_endpoints":["chat.completions","responses"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[],"pricing":{"prompt":"0.000001375","completion":"0.000011","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000001375","input_cache_write":"0","discount":1,"currency":"USD"},"author":"openai","slug":"gpt-5.1","tags":["reasoning","long-context","adaptive","coding","vision"],"author_info":{"slug":"openai","name":"OpenAI","display_name":null,"icon_url":"/images/logos/openai.webp","gradient_from":"from-zinc-800","gradient_to":"to-zinc-900","gradient_via":null,"website_url":"https://openai.com"}},{"id":"bge-large-en-v1.5","object":"model","canonical_slug":null,"hugging_face_id":"BAAI/bge-large-en-v1.5","name":"BGE Large EN v1.5","created":1770318640,"description":"BGE Large EN v1.5 is an English embedding model with 1024-dimensional vectors, optimized for semantic search and retrieval with long-context inputs.","context_length":8192,"architecture":{"modality":"text->embedding","input_modalities":["text"],"output_modalities":["embedding"],"tokenizer":null,"instruct_type":null},"top_provider":{"context_length":8192,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["encoding_format"],"supported_api_endpoints":["embeddings"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"IONOS Cloud","slug":"ionos"}],"pricing":{"prompt":"0.000000015","completion":"0","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"baai","slug":"bge-large-en-v1.5","tags":["embedding","english","long-context"],"author_info":{"slug":"baai","name":"BAAI","display_name":"Beijing Academy of AI","icon_url":null,"gradient_from":"from-red-600","gradient_to":"to-red-800","gradient_via":null,"website_url":"https://www.baai.ac.cn"}},{"id":"qwen3.5-397b-a17b","object":"model","canonical_slug":null,"hugging_face_id":"Qwen/Qwen3.5-397B-A17B","name":"Qwen3.5 397B A17B","created":1777452602,"description":"Qwen3.5 397B A17B is an Alibaba/Qwen MoE model for long-context reasoning, coding, agentic tasks, tool use, and multimodal understanding.","context_length":250000,"architecture":{"modality":"text+image->text","input_modalities":["text","image"],"output_modalities":["text"],"tokenizer":"Qwen3","instruct_type":null},"top_provider":{"context_length":250000,"max_completion_tokens":16000,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","seed","response_format","tools","tool_choice"],"supported_api_endpoints":["chat.completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Scaleway","slug":"scaleway"},{"name":"GreenPT","slug":"greenpt"}],"pricing":{"prompt":"0.0000006","completion":"0.0000036","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"EUR"},"author":"alibaba","slug":"qwen3.5-397b-a17b","tags":["open-source","moe","coding","agentic","reasoning","vision","function-calling"],"author_info":{"slug":"alibaba","name":"Alibaba","display_name":"Alibaba (Qwen)","icon_url":"/images/logos/qwen.webp","gradient_from":"from-purple-600","gradient_to":"to-indigo-700","gradient_via":null,"website_url":"https://qwenlm.github.io"}},{"id":"text-embedding-3-large","object":"model","canonical_slug":null,"hugging_face_id":"","name":"Text Embedding 3 Large","created":1767964798,"description":"text-embedding-3-large is OpenAI's most capable text embedding model with up to 3,072 output dimensions.","context_length":8192,"architecture":{"modality":"text->embedding","input_modalities":["text"],"output_modalities":["embedding"],"tokenizer":"cl100k_base","instruct_type":null},"top_provider":{"context_length":8192,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["encoding_format","dimensions"],"supported_api_endpoints":["embeddings"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Microsoft Foundry","slug":"microsoft-foundry"}],"author":"openai","slug":"text-embedding-3-large","tags":["embedding","retrieval","high-quality"],"author_info":{"slug":"openai","name":"OpenAI","display_name":null,"icon_url":"/images/logos/openai.webp","gradient_from":"from-zinc-800","gradient_to":"to-zinc-900","gradient_via":null,"website_url":"https://openai.com"}},{"id":"glm-5.1","object":"model","canonical_slug":null,"hugging_face_id":"zai-org/GLM-5.1-FP8","name":"GLM 5.1","created":1777453828,"description":"GLM-5.1 is Z AI's flagship model for long-horizon agentic engineering and advanced coding workflows.","context_length":202752,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"TokenizersBackend","instruct_type":null},"top_provider":{"context_length":202752,"max_completion_tokens":202752,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","tools","tool_choice","response_format"],"supported_api_endpoints":["chat.completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Inceptron","slug":"inceptron"},{"name":"Tensorix","slug":"tensorix"}],"pricing":{"prompt":"0.0000014","completion":"0.0000044","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000026","input_cache_write":"0","discount":1,"currency":"USD"},"author":"zhipu","slug":"glm-5.1","tags":["open-source","long-context","coding","agentic","reasoning"],"author_info":{"slug":"zhipu","name":"Zhipu","display_name":"Zhipu AI","icon_url":null,"gradient_from":"from-blue-500","gradient_to":"to-cyan-500","gradient_via":null,"website_url":"https://www.zhipuai.cn"}},{"id":"bge-multilingual-gemma2","object":"model","canonical_slug":null,"hugging_face_id":"BAAI/bge-multilingual-gemma2","name":"BGE Multilingual Gemma2","created":1767963057,"description":"BGE Multilingual Gemma2 is a BAAI multilingual embedding model based on Gemma2 for dense retrieval and semantic search.","context_length":8192,"architecture":{"modality":"text->embedding","input_modalities":["text"],"output_modalities":["embedding"],"tokenizer":"Gemma","instruct_type":null},"top_provider":{"context_length":8192,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":[],"supported_api_endpoints":["embeddings"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"OVHcloud","slug":"ovhcloud"},{"name":"Scaleway","slug":"scaleway"}],"author":"baai","slug":"bge-multilingual-gemma2","tags":["embedding","multilingual","retrieval"],"author_info":{"slug":"baai","name":"BAAI","display_name":"Beijing Academy of AI","icon_url":null,"gradient_from":"from-red-600","gradient_to":"to-red-800","gradient_via":null,"website_url":"https://www.baai.ac.cn"}},{"id":"qwen3.5-122b-a10b","object":"model","canonical_slug":null,"hugging_face_id":"Qwen/Qwen3.5-122B-A10B","name":"Qwen3.5 122B A10B","created":1777454224,"description":"Qwen3.5 122B A10B is Alibaba's Mixture-of-Experts model for reasoning, coding, multilingual chat, and tool use.","context_length":262144,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"Qwen3","instruct_type":null},"top_provider":{"context_length":262144,"max_completion_tokens":262144,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","tools","tool_choice","response_format"],"supported_api_endpoints":["chat.completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Tensorix","slug":"tensorix"}],"pricing":{"prompt":"0.0000005","completion":"0.0000035","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"alibaba","slug":"qwen3.5-122b-a10b","tags":["open-source","moe","long-context","coding","reasoning"],"author_info":{"slug":"alibaba","name":"Alibaba","display_name":"Alibaba (Qwen)","icon_url":"/images/logos/qwen.webp","gradient_from":"from-purple-600","gradient_to":"to-indigo-700","gradient_via":null,"website_url":"https://qwenlm.github.io"}},{"id":"mixtral-8x7b-instruct","object":"model","canonical_slug":null,"hugging_face_id":"","name":"Mixtral 8x7B Instruct","created":1777454590,"description":"Mixtral 8x7B Instruct is Mistral AI's sparse mixture-of-experts instruction model.","context_length":32000,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":null,"instruct_type":null},"top_provider":{"context_length":32000,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"AWS Bedrock","slug":"aws-bedrock"}],"pricing":{"prompt":"0.00000049","completion":"0.00000076","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"mistral","slug":"mixtral-8x7b-instruct","tags":["open-weights","instruction-tuned","moe"],"author_info":{"slug":"mistral","name":"Mistral","display_name":"Mistral AI","icon_url":"/images/logos/mistral.jpeg","gradient_from":"from-[#ff7000]","gradient_to":"to-[#ff9a00]","gradient_via":null,"website_url":"https://mistral.ai"}},{"id":"o4-mini","object":"model","canonical_slug":null,"hugging_face_id":"","name":"o4-mini","created":1767619877,"description":"o4-mini is OpenAI's compact reasoning model for math, coding, visual tasks, structured outputs, and tool use.","context_length":200000,"architecture":{"modality":"text+image->text","input_modalities":["text","image"],"output_modalities":["text"],"tokenizer":"GPT","instruct_type":null},"top_provider":{"context_length":200000,"max_completion_tokens":100000,"is_moderated":false},"per_request_limits":null,"supported_parameters":["max_completion_tokens","tools","tool_choice","response_format","reasoning_effort"],"supported_api_endpoints":["chat.completions","responses"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Microsoft Foundry","slug":"microsoft-foundry"}],"pricing":{"prompt":"0.0000011","completion":"0.0000044","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"openai","slug":"o4-mini","tags":["reasoning","long-context","efficient","coding","math","vision"],"author_info":{"slug":"openai","name":"OpenAI","display_name":null,"icon_url":"/images/logos/openai.webp","gradient_from":"from-zinc-800","gradient_to":"to-zinc-900","gradient_via":null,"website_url":"https://openai.com"}},{"id":"claude-3-haiku","object":"model","canonical_slug":null,"hugging_face_id":"","name":"Claude 3 Haiku","created":1769967814,"description":"Claude 3 Haiku is the fastest and most compact model in the Claude 3 family, ideal for quick responses.","context_length":200000,"architecture":{"modality":"text+image->text","input_modalities":["text","image"],"output_modalities":["text"],"tokenizer":"claude","instruct_type":"claude"},"top_provider":{"context_length":200000,"max_completion_tokens":4096,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","tools","tool_choice"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[],"pricing":{"prompt":"0.00000025","completion":"0.00000125","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"anthropic","slug":"claude-3-haiku","tags":["proprietary","fast","cost-efficient","vision"],"author_info":{"slug":"anthropic","name":"Anthropic","display_name":null,"icon_url":"/images/logos/anthropic.jpeg","gradient_from":"from-[#cc785c]","gradient_to":"to-[#d4a574]","gradient_via":null,"website_url":"https://anthropic.com"}},{"id":"mistral-large-3","object":"model","canonical_slug":null,"hugging_face_id":"","name":"Mistral Large 3","created":1767979257,"description":"Mistral Large 3 is Mistral AI's 675B-parameter model for coding, reasoning, multilingual tasks, and vision-enabled chat.","context_length":262144,"architecture":{"modality":"text+image->text","input_modalities":["text","image"],"output_modalities":["text"],"tokenizer":"Tekken","instruct_type":null},"top_provider":{"context_length":262144,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","random_seed","response_format","tools","tool_choice"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Mistral AI","slug":"mistral"}],"pricing":{"prompt":"0.0000005","completion":"0.0000015","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"mistral","slug":"mistral-large-3","tags":["flagship","instruction-tuned","vision","long-context","coding","agentic","prompt-caching"],"author_info":{"slug":"mistral","name":"Mistral","display_name":"Mistral AI","icon_url":"/images/logos/mistral.jpeg","gradient_from":"from-[#ff7000]","gradient_to":"to-[#ff9a00]","gradient_via":null,"website_url":"https://mistral.ai"}},{"id":"gpt-5-nano","object":"model","canonical_slug":null,"hugging_face_id":"","name":"GPT-5 Nano","created":1777454744,"description":"GPT-5 Nano is OpenAI's smallest GPT-5 reasoning model for high-volume multimodal and tool-assisted workloads.","context_length":400000,"architecture":{"modality":"text+image->text","input_modalities":["text","image"],"output_modalities":["text"],"tokenizer":"GPT","instruct_type":null},"top_provider":{"context_length":400000,"max_completion_tokens":128000,"is_moderated":false},"per_request_limits":null,"supported_parameters":["max_tokens","max_completion_tokens","reasoning_effort","tools","tool_choice","response_format","parallel_tool_calls"],"supported_api_endpoints":["chat.completions","responses"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Microsoft Foundry","slug":"microsoft-foundry"}],"pricing":{"prompt":"0.000000055","completion":"0.00000044","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000000055","input_cache_write":"0","discount":1,"currency":"USD"},"author":"openai","slug":"gpt-5-nano","tags":["reasoning","long-context","efficient","high-throughput","vision"],"author_info":{"slug":"openai","name":"OpenAI","display_name":null,"icon_url":"/images/logos/openai.webp","gradient_from":"from-zinc-800","gradient_to":"to-zinc-900","gradient_via":null,"website_url":"https://openai.com"}},{"id":"bge-en-icl","object":"model","canonical_slug":null,"hugging_face_id":"BAAI/bge-en-icl","name":"BGE-EN-ICL","created":1767963057,"description":"Embedding model from BAAI designed for powerful in-context learning retrieval across domains. Produces 4096-dimensional embeddings optimized for semantic search with few-shot learning capabilities.","context_length":33000,"architecture":{"modality":"text->embedding","input_modalities":["text"],"output_modalities":["embedding"],"tokenizer":null,"instruct_type":null},"top_provider":{"context_length":33000,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":[],"supported_api_endpoints":["embeddings"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[],"author":"baai","slug":"bge-en-icl","tags":["embedding","retrieval","rag","icl","english"],"author_info":{"slug":"baai","name":"BAAI","display_name":"Beijing Academy of AI","icon_url":null,"gradient_from":"from-red-600","gradient_to":"to-red-800","gradient_via":null,"website_url":"https://www.baai.ac.cn"}},{"id":"titan-embed-text-v2","object":"model","canonical_slug":null,"hugging_face_id":"","name":"Amazon Titan Embeddings v2","created":1769967814,"description":"Amazon Titan Embeddings v2 offers configurable embedding dimensions (256-1024) and improved multilingual support.","context_length":8000,"architecture":{"modality":"text->embedding","input_modalities":["text"],"output_modalities":["embedding"],"tokenizer":null,"instruct_type":null},"top_provider":{"context_length":8000,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":[],"supported_api_endpoints":["embeddings"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"AWS Bedrock","slug":"aws-bedrock"}],"author":"amazon","slug":"titan-embed-text-v2","tags":["proprietary","embedding","multilingual","configurable-dimensions"],"author_info":{"slug":"amazon","name":"Amazon","display_name":"Amazon Web Services","icon_url":"/images/logos/aws.webp","gradient_from":"from-orange-500","gradient_to":"to-yellow-500","gradient_via":null,"website_url":"https://aws.amazon.com/bedrock"}},{"id":"devstral-small","object":"model","canonical_slug":null,"hugging_face_id":"mistralai/Devstral-Small-2505","name":"Devstral Small","created":1765183068,"description":"Devstral-Small-2505 is a 24B agentic LLM fine-tuned from Mistral-Small-3.1 by Mistral AI and All Hands AI for software engineering. #1 open source model on SWE-bench (46.8% on Verified). Excels at codebase exploration, multi-file editing, and powering software engineering agents. Text-only (vision encoder removed).","context_length":128000,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"Tekken","instruct_type":null},"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","tools","tool_choice"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[],"pricing":{"prompt":"0.00000015","completion":"0.00000035","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"EUR"},"author":"mistral","slug":"devstral-small","tags":["open-source","coding","agentic","instruction-tuned"],"author_info":{"slug":"mistral","name":"Mistral","display_name":"Mistral AI","icon_url":"/images/logos/mistral.jpeg","gradient_from":"from-[#ff7000]","gradient_to":"to-[#ff9a00]","gradient_via":null,"website_url":"https://mistral.ai"}},{"id":"claude-sonnet-4-5","object":"model","canonical_slug":null,"hugging_face_id":"","name":"Claude Sonnet 4.5","created":1769967813,"description":"Claude Sonnet 4.5 is Anthropic's most advanced model, offering exceptional performance on complex reasoning tasks with a 200K context window.","context_length":200000,"architecture":{"modality":"text+image->text","input_modalities":["text","image"],"output_modalities":["text"],"tokenizer":"claude","instruct_type":"claude"},"top_provider":{"context_length":200000,"max_completion_tokens":8192,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","tools","tool_choice"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"AWS Bedrock","slug":"aws-bedrock"}],"pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"anthropic","slug":"claude-sonnet-4-5","tags":["proprietary","long-context","vision","reasoning"],"author_info":{"slug":"anthropic","name":"Anthropic","display_name":null,"icon_url":"/images/logos/anthropic.jpeg","gradient_from":"from-[#cc785c]","gradient_to":"to-[#d4a574]","gradient_via":null,"website_url":"https://anthropic.com"}},{"id":"teuken-7b-instruct","object":"model","canonical_slug":null,"hugging_face_id":"openGPT-X/Teuken-7B-instruct-commercial-v0.4","name":"Teuken-7B Instruct","created":1770318640,"description":"Teuken 7B Instruct is an OpenGPT-X model optimized for German and European languages, offering an 8K context window for instruction-following tasks.","context_length":8192,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":null,"instruct_type":null},"top_provider":{"context_length":8192,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","response_format"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[],"pricing":{"prompt":"0.00000017","completion":"0.00000017","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"opengpt-x","slug":"teuken-7b-instruct","tags":["open-source","instruction-tuned","multilingual","european"],"author_info":{"slug":"opengpt-x","name":"openGPT-X","display_name":"openGPT-X","icon_url":null,"gradient_from":null,"gradient_to":null,"gradient_via":null,"website_url":"https://opengpt-x.de"}},{"id":"deepseek-v3.2","object":"model","canonical_slug":null,"hugging_face_id":"deepseek-ai/DeepSeek-V3.2","name":"DeepSeek V3.2","created":1774895903,"description":"DeepSeek V3.2 is a mixture-of-experts model with improved reasoning, coding, and instruction-following capabilities.","context_length":163840,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"DeepSeek","instruct_type":null},"top_provider":{"context_length":163840,"max_completion_tokens":163840,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","tools","tool_choice","response_format"],"supported_api_endpoints":["chat.completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Tensorix","slug":"tensorix"},{"name":"AWS Bedrock","slug":"aws-bedrock"}],"pricing":{"prompt":"0.0000003","completion":"0.0000005","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"deepseek","slug":"deepseek-v3.2","tags":["open-source","long-context","reasoning","coding"],"author_info":{"slug":"deepseek","name":"DeepSeek","display_name":null,"icon_url":"/images/logos/deepseek.png","gradient_from":"from-[#0066ff]","gradient_to":"to-[#00ccff]","gradient_via":null,"website_url":"https://deepseek.com"}},{"id":"mistral-nemo-12b","object":"model","canonical_slug":null,"hugging_face_id":"mistralai/Mistral-Nemo-Instruct-2407","name":"Mistral Nemo 12B","created":1765024302,"description":"Mistral Nemo 12B is Mistral AI's open multilingual model built with NVIDIA for efficient long-context text generation.","context_length":120832,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"Tekken","instruct_type":null},"top_provider":{"context_length":120832,"max_completion_tokens":120832,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","seed","response_format","tools","tool_choice"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Mistral AI","slug":"mistral"},{"name":"OVHcloud","slug":"ovhcloud"},{"name":"IONOS Cloud","slug":"ionos"}],"pricing":{"prompt":"0.00000013","completion":"0.00000013","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"EUR"},"author":"mistral","slug":"mistral-nemo-12b","tags":["open-source","multilingual","long-context"],"author_info":{"slug":"mistral","name":"Mistral","display_name":"Mistral AI","icon_url":"/images/logos/mistral.jpeg","gradient_from":"from-[#ff7000]","gradient_to":"to-[#ff9a00]","gradient_via":null,"website_url":"https://mistral.ai"}},{"id":"mistral-embed","object":"model","canonical_slug":null,"hugging_face_id":"","name":"Mistral Embed","created":1767964798,"description":"Mistral Embed is Mistral AI's text embedding model for semantic search, retrieval, clustering, and classification.","context_length":8192,"architecture":{"modality":"text->embedding","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"Tekken","instruct_type":null},"top_provider":{"context_length":8192,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["encoding_format","dimensions","output_dimension","output_dtype"],"supported_api_endpoints":["embeddings"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Mistral AI","slug":"mistral"}],"pricing":{"prompt":"0.0000001","completion":"0","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"mistral","slug":"mistral-embed","tags":["embedding","retrieval","semantic-search"],"author_info":{"slug":"mistral","name":"Mistral","display_name":"Mistral AI","icon_url":"/images/logos/mistral.jpeg","gradient_from":"from-[#ff7000]","gradient_to":"to-[#ff9a00]","gradient_via":null,"website_url":"https://mistral.ai"}},{"id":"nemotron-3-nano-omni","object":"model","canonical_slug":null,"hugging_face_id":"","name":"Nemotron 3 Nano Omni","created":1777453939,"description":"NVIDIA Nemotron 3 Nano Omni is an efficient omni-modal reasoning model for agentic AI, coding, tool use, and long-context workflows.","context_length":65536,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"Nemotron","instruct_type":null},"top_provider":{"context_length":65536,"max_completion_tokens":65536,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","response_format","tools","tool_choice"],"supported_api_endpoints":["chat.completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Nebius","slug":"nebius"}],"pricing":{"prompt":"0.00000006","completion":"0.00000024","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"nvidia","slug":"nemotron-3-nano-omni","tags":["open-source","reasoning","coding","function-calling","json-mode"],"author_info":{"slug":"nvidia","name":"NVIDIA","display_name":null,"icon_url":"/images/logos/nvidia.webp","gradient_from":"from-[#76b900]","gradient_to":"to-[#5a8c00]","gradient_via":null,"website_url":"https://nvidia.com"}},{"id":"text-embedding-ada-002","object":"model","canonical_slug":null,"hugging_face_id":"","name":"Text Embedding Ada 002","created":1767964798,"description":"text-embedding-ada-002 is OpenAI's older general-purpose text embedding model with 1,536 output dimensions.","context_length":8192,"architecture":{"modality":"text->embedding","input_modalities":["text"],"output_modalities":["embedding"],"tokenizer":"cl100k_base","instruct_type":null},"top_provider":{"context_length":8192,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["encoding_format","dimensions"],"supported_api_endpoints":["embeddings"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Microsoft Foundry","slug":"microsoft-foundry"}],"author":"openai","slug":"text-embedding-ada-002","tags":["embedding","retrieval","legacy"],"author_info":{"slug":"openai","name":"OpenAI","display_name":null,"icon_url":"/images/logos/openai.webp","gradient_from":"from-zinc-800","gradient_to":"to-zinc-900","gradient_via":null,"website_url":"https://openai.com"}},{"id":"qwen3-30b-a3b-thinking","object":"model","canonical_slug":null,"hugging_face_id":"","name":"Qwen3 30B A3B Thinking","created":1768127608,"description":"Mid-size Qwen thinking model delivering long reasoning chains at efficient cost.","context_length":128000,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"Qwen3","instruct_type":null},"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","tools","tool_choice","response_format"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[],"pricing":{"prompt":"0.0000001","completion":"0.0000003","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"alibaba","slug":"qwen3-30b-a3b-thinking","tags":["moe","reasoning","thinking","efficient"],"author_info":{"slug":"alibaba","name":"Alibaba","display_name":"Alibaba (Qwen)","icon_url":"/images/logos/qwen.webp","gradient_from":"from-purple-600","gradient_to":"to-indigo-700","gradient_via":null,"website_url":"https://qwenlm.github.io"}},{"id":"glm-5","object":"model","canonical_slug":null,"hugging_face_id":"THUDM/GLM-5","name":"GLM 5","created":1774895903,"description":"GLM-5 is Z AI's flagship language model for coding, reasoning, tool use, and multilingual chat.","context_length":202752,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"ChatGLM","instruct_type":null},"top_provider":{"context_length":202752,"max_completion_tokens":202752,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","tools","tool_choice","response_format"],"supported_api_endpoints":["chat.completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Tensorix","slug":"tensorix"},{"name":"AWS Bedrock","slug":"aws-bedrock"}],"pricing":{"prompt":"0.000001","completion":"0.0000032","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"zhipu","slug":"glm-5","tags":["open-source","multilingual","long-context","coding","reasoning"],"author_info":{"slug":"zhipu","name":"Zhipu","display_name":"Zhipu AI","icon_url":null,"gradient_from":"from-blue-500","gradient_to":"to-cyan-500","gradient_via":null,"website_url":"https://www.zhipuai.cn"}},{"id":"gpt-5","object":"model","canonical_slug":null,"hugging_face_id":"","name":"GPT-5","created":1767619876,"description":"GPT-5 is OpenAI's reasoning model with multimodal input, structured outputs, function calling, and long-context support.","context_length":400000,"architecture":{"modality":"text+image->text","input_modalities":["text","image"],"output_modalities":["text"],"tokenizer":"GPT","instruct_type":null},"top_provider":{"context_length":400000,"max_completion_tokens":128000,"is_moderated":false},"per_request_limits":null,"supported_parameters":["max_tokens","max_completion_tokens","reasoning_effort","tools","tool_choice","response_format","parallel_tool_calls"],"supported_api_endpoints":["chat.completions","responses"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[],"pricing":{"prompt":"0.000001375","completion":"0.000011","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000001375","input_cache_write":"0","discount":1,"currency":"USD"},"author":"openai","slug":"gpt-5","tags":["reasoning","long-context","flagship","coding","vision"],"author_info":{"slug":"openai","name":"OpenAI","display_name":null,"icon_url":"/images/logos/openai.webp","gradient_from":"from-zinc-800","gradient_to":"to-zinc-900","gradient_via":null,"website_url":"https://openai.com"}},{"id":"qwen-2.5-coder-32b-instruct","object":"model","canonical_slug":null,"hugging_face_id":"Qwen/Qwen2.5-Coder-32B-Instruct","name":"Qwen 2.5 Coder 32B Instruct","created":1765024302,"description":"Qwen 2.5 Coder 32B is a specialized code generation model from Alibaba. It excels at code completion, debugging, and explanation tasks across multiple programming languages.","context_length":128000,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"Qwen2","instruct_type":null},"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[],"pricing":{"prompt":"0.0000009","completion":"0.0000009","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"EUR"},"author":"alibaba","slug":"qwen-2.5-coder-32b-instruct","tags":["open-source","instruction-tuned","coding","long-context"],"author_info":{"slug":"alibaba","name":"Alibaba","display_name":"Alibaba (Qwen)","icon_url":"/images/logos/qwen.webp","gradient_from":"from-purple-600","gradient_to":"to-indigo-700","gradient_via":null,"website_url":"https://qwenlm.github.io"}},{"id":"gpt-oss-20b","object":"model","canonical_slug":null,"hugging_face_id":"openai/gpt-oss-20b","name":"GPT-OSS 20B","created":1767979257,"description":"GPT-OSS 20B is OpenAI's smaller open-weight reasoning model for efficient chat, coding, and agentic workloads.","context_length":131000,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"o200k_base","instruct_type":null},"top_provider":{"context_length":131000,"max_completion_tokens":32768,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","tools","tool_choice","response_format"],"supported_api_endpoints":["chat.completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Tensorix","slug":"tensorix"},{"name":"OVHcloud","slug":"ovhcloud"},{"name":"AWS Bedrock","slug":"aws-bedrock"}],"pricing":{"prompt":"0.00000003","completion":"0.00000014","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"openai","slug":"gpt-oss-20b","tags":["open-source","reasoning","coding","function-calling"],"author_info":{"slug":"openai","name":"OpenAI","display_name":null,"icon_url":"/images/logos/openai.webp","gradient_from":"from-zinc-800","gradient_to":"to-zinc-900","gradient_via":null,"website_url":"https://openai.com"}},{"id":"llama-3.3-70b-instruct","object":"model","canonical_slug":null,"hugging_face_id":"meta-llama/Llama-3.3-70B-Instruct","name":"Llama 3.3 70B Instruct","created":1765024302,"description":"Llama 3.3 70B Instruct is Meta's multilingual instruction-tuned model for chat, reasoning, and tool-use workloads.","context_length":131000,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"Llama3","instruct_type":null},"top_provider":{"context_length":131000,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","tools","tool_choice","response_format"],"supported_api_endpoints":["chat.completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Inceptron","slug":"inceptron"},{"name":"GreenPT","slug":"greenpt"},{"name":"Scaleway","slug":"scaleway"},{"name":"OVHcloud","slug":"ovhcloud"},{"name":"Tensorix","slug":"tensorix"},{"name":"IONOS Cloud","slug":"ionos"},{"name":"Nebius","slug":"nebius"}],"pricing":{"prompt":"0.000000104","completion":"0.000000312","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"meta","slug":"llama-3.3-70b-instruct","tags":["open-source","instruction-tuned","function-calling"],"author_info":{"slug":"meta","name":"Meta","display_name":"Meta AI","icon_url":"/images/logos/meta.webp","gradient_from":"from-blue-600","gradient_to":"to-blue-800","gradient_via":null,"website_url":"https://ai.meta.com"}},{"id":"gpt-4o-mini","object":"model","canonical_slug":null,"hugging_face_id":"","name":"GPT-4o Mini","created":1768138176,"description":"GPT-4o Mini is OpenAI's compact multimodal model for low-cost chat, vision, tool use, and structured-output workloads.","context_length":128000,"architecture":{"modality":"text+image->text","input_modalities":["text","image"],"output_modalities":["text"],"tokenizer":"GPT","instruct_type":null},"top_provider":{"context_length":128000,"max_completion_tokens":16384,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","max_completion_tokens","stop","tools","tool_choice","response_format","seed","parallel_tool_calls"],"supported_api_endpoints":["chat.completions","responses"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Microsoft Foundry","slug":"microsoft-foundry"}],"pricing":{"prompt":"0.00000015","completion":"0.0000006","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"openai","slug":"gpt-4o-mini","tags":["multimodal","function-calling","vision","efficient"],"author_info":{"slug":"openai","name":"OpenAI","display_name":null,"icon_url":"/images/logos/openai.webp","gradient_from":"from-zinc-800","gradient_to":"to-zinc-900","gradient_via":null,"website_url":"https://openai.com"}},{"id":"ministral-3-8b","object":"model","canonical_slug":null,"hugging_face_id":"","name":"Ministral 3 8B","created":1767979257,"description":"Ministral 3 8B is an open Mistral AI model for efficient text and vision workloads.","context_length":262144,"architecture":{"modality":"text+image->text","input_modalities":["text","image"],"output_modalities":["text"],"tokenizer":"Tekken","instruct_type":null},"top_provider":{"context_length":262144,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","random_seed","response_format","tools","tool_choice"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Mistral AI","slug":"mistral"}],"pricing":{"prompt":"0.00000015","completion":"0.00000015","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"mistral","slug":"ministral-3-8b","tags":["open-source","vision","lightweight","long-context"],"author_info":{"slug":"mistral","name":"Mistral","display_name":"Mistral AI","icon_url":"/images/logos/mistral.jpeg","gradient_from":"from-[#ff7000]","gradient_to":"to-[#ff9a00]","gradient_via":null,"website_url":"https://mistral.ai"}},{"id":"qwen3-coder-30b-a3b","object":"model","canonical_slug":null,"hugging_face_id":"Qwen/Qwen3-Coder-30B-A3B-Instruct","name":"Qwen3 Coder 30B A3B","created":1765183068,"description":"Qwen3 Coder 30B A3B is Alibaba's Mixture-of-Experts coding model for software engineering and agentic code tasks.","context_length":262144,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"Qwen3","instruct_type":null},"top_provider":{"context_length":262144,"max_completion_tokens":262144,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","seed","response_format","tools","tool_choice"],"supported_api_endpoints":["chat.completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"GreenPT","slug":"greenpt"},{"name":"Scaleway","slug":"scaleway"},{"name":"Tensorix","slug":"tensorix"},{"name":"OVHcloud","slug":"ovhcloud"},{"name":"AWS Bedrock","slug":"aws-bedrock"}],"pricing":{"prompt":"0.00000006","completion":"0.00000022","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"EUR"},"author":"alibaba","slug":"qwen3-coder-30b-a3b","tags":["open-source","moe","coding","agentic"],"author_info":{"slug":"alibaba","name":"Alibaba","display_name":"Alibaba (Qwen)","icon_url":"/images/logos/qwen.webp","gradient_from":"from-purple-600","gradient_to":"to-indigo-700","gradient_via":null,"website_url":"https://qwenlm.github.io"}},{"id":"mistral-large-2402","object":"model","canonical_slug":null,"hugging_face_id":"","name":"Mistral Large 2402","created":1777454590,"description":"Mistral Large 2402 is Mistral AI's high-capability text model available through Amazon Bedrock.","context_length":32000,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":null,"instruct_type":null},"top_provider":{"context_length":32000,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"AWS Bedrock","slug":"aws-bedrock"}],"pricing":{"prompt":"0.0000043","completion":"0.000013","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"mistral","slug":"mistral-large-2402","tags":["proprietary","instruction-tuned","multilingual"],"author_info":{"slug":"mistral","name":"Mistral","display_name":"Mistral AI","icon_url":"/images/logos/mistral.jpeg","gradient_from":"from-[#ff7000]","gradient_to":"to-[#ff9a00]","gradient_via":null,"website_url":"https://mistral.ai"}},{"id":"gpt-oss-120b","object":"model","canonical_slug":null,"hugging_face_id":"openai/gpt-oss-120b","name":"GPT-OSS 120B","created":1765183068,"description":"GPT-OSS 120B is OpenAI's open-weight reasoning model for agentic, coding, and general chat workloads.","context_length":131000,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"o200k_base","instruct_type":null},"top_provider":{"context_length":131000,"max_completion_tokens":32768,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","tools","tool_choice","response_format"],"supported_api_endpoints":["chat.completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"GreenPT","slug":"greenpt"},{"name":"Tensorix","slug":"tensorix"},{"name":"OVHcloud","slug":"ovhcloud"},{"name":"Scaleway","slug":"scaleway"},{"name":"Infercom","slug":"infercom"},{"name":"IONOS Cloud","slug":"ionos"},{"name":"AWS Bedrock","slug":"aws-bedrock"},{"name":"Nebius","slug":"nebius"}],"pricing":{"prompt":"0.00000004","completion":"0.0000002","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"openai","slug":"gpt-oss-120b","tags":["open-source","reasoning","coding","function-calling"],"author_info":{"slug":"openai","name":"OpenAI","display_name":null,"icon_url":"/images/logos/openai.webp","gradient_from":"from-zinc-800","gradient_to":"to-zinc-900","gradient_via":null,"website_url":"https://openai.com"}},{"id":"claude-3.5-sonnet","object":"model","canonical_slug":null,"hugging_face_id":"","name":"Claude 3.5 Sonnet","created":1769967813,"description":"Claude 3.5 Sonnet balances intelligence and speed, excelling at complex tasks while remaining cost-effective.","context_length":200000,"architecture":{"modality":"text+image->text","input_modalities":["text","image"],"output_modalities":["text"],"tokenizer":"claude","instruct_type":"claude"},"top_provider":{"context_length":200000,"max_completion_tokens":8192,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","tools","tool_choice"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[],"pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"anthropic","slug":"claude-3.5-sonnet","tags":["proprietary","long-context","vision"],"author_info":{"slug":"anthropic","name":"Anthropic","display_name":null,"icon_url":"/images/logos/anthropic.jpeg","gradient_from":"from-[#cc785c]","gradient_to":"to-[#d4a574]","gradient_via":null,"website_url":"https://anthropic.com"}},{"id":"llama-3.1-405b-instruct","object":"model","canonical_slug":null,"hugging_face_id":"meta-llama/Meta-Llama-3.1-405B-Instruct","name":"Llama 3.1 405B Instruct","created":1770318640,"description":"Llama 3.1 405B Instruct is Meta's flagship 405B parameter model with a 128K context window. It delivers strong multilingual reasoning, instruction following, and tool-use capabilities at frontier scale.","context_length":128000,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"Llama3","instruct_type":"llama3"},"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","response_format","tools","tool_choice"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"IONOS Cloud","slug":"ionos"}],"pricing":{"prompt":"0.00000193","completion":"0.00000193","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"meta","slug":"llama-3.1-405b-instruct","tags":["open-source","instruction-tuned","multilingual","long-context","flagship"],"author_info":{"slug":"meta","name":"Meta","display_name":"Meta AI","icon_url":"/images/logos/meta.webp","gradient_from":"from-blue-600","gradient_to":"to-blue-800","gradient_via":null,"website_url":"https://ai.meta.com"}},{"id":"glm-4.7","object":"model","canonical_slug":null,"hugging_face_id":"THUDM/GLM-4.7","name":"GLM 4.7","created":1774895903,"description":"GLM-4.7 is Z AI's model for high-throughput chat, coding, and function-calling workloads.","context_length":202752,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"ChatGLM","instruct_type":null},"top_provider":{"context_length":202752,"max_completion_tokens":200000,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","tools","tool_choice","response_format"],"supported_api_endpoints":["chat.completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Tensorix","slug":"tensorix"},{"name":"AWS Bedrock","slug":"aws-bedrock"}],"pricing":{"prompt":"0.0000006","completion":"0.0000022","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"zhipu","slug":"glm-4.7","tags":["open-source","multilingual","long-context","function-calling"],"author_info":{"slug":"zhipu","name":"Zhipu","display_name":"Zhipu AI","icon_url":null,"gradient_from":"from-blue-500","gradient_to":"to-cyan-500","gradient_via":null,"website_url":"https://www.zhipuai.cn"}},{"id":"llama-3.2-1b-instruct","object":"model","canonical_slug":null,"hugging_face_id":"meta-llama/Llama-3.2-1B-Instruct","name":"Llama 3.2 1B Instruct","created":1769967814,"description":"Llama 3.2 1B Instruct is Meta's smallest multilingual model, designed for edge deployments with minimal resources.","context_length":128000,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"Llama3","instruct_type":"llama3"},"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[],"pricing":{"prompt":"0.0000001","completion":"0.0000001","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"meta","slug":"llama-3.2-1b-instruct","tags":["open-source","instruction-tuned","multilingual","tiny"],"author_info":{"slug":"meta","name":"Meta","display_name":"Meta AI","icon_url":"/images/logos/meta.webp","gradient_from":"from-blue-600","gradient_to":"to-blue-800","gradient_via":null,"website_url":"https://ai.meta.com"}},{"id":"gpt-4.1-nano","object":"model","canonical_slug":null,"hugging_face_id":"","name":"GPT-4.1 Nano","created":1769687270,"description":"GPT-4.1 Nano is OpenAI's lowest-latency GPT-4.1 model for high-volume multimodal, long-context, and tool-assisted tasks.","context_length":1047576,"architecture":{"modality":"text+image->text","input_modalities":["text","image"],"output_modalities":["text"],"tokenizer":"GPT","instruct_type":null},"top_provider":{"context_length":1047576,"max_completion_tokens":32768,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","max_completion_tokens","stop","tools","tool_choice","response_format","seed","parallel_tool_calls"],"supported_api_endpoints":["chat.completions","responses"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Microsoft Foundry","slug":"microsoft-foundry"}],"pricing":{"prompt":"0.00000011","completion":"0.00000044","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.000000028","input_cache_write":"0","discount":1,"currency":"USD"},"author":"openai","slug":"gpt-4.1-nano","tags":["multimodal","function-calling","long-context","efficient","high-throughput"],"author_info":{"slug":"openai","name":"OpenAI","display_name":null,"icon_url":"/images/logos/openai.webp","gradient_from":"from-zinc-800","gradient_to":"to-zinc-900","gradient_via":null,"website_url":"https://openai.com"}},{"id":"claude-3-sonnet","object":"model","canonical_slug":null,"hugging_face_id":"","name":"Claude 3 Sonnet","created":1769967814,"description":"Claude 3 Sonnet offers a balance of intelligence and speed for enterprise workloads.","context_length":200000,"architecture":{"modality":"text+image->text","input_modalities":["text","image"],"output_modalities":["text"],"tokenizer":"claude","instruct_type":"claude"},"top_provider":{"context_length":200000,"max_completion_tokens":4096,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","tools","tool_choice"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[],"pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"anthropic","slug":"claude-3-sonnet","tags":["proprietary","long-context","vision"],"author_info":{"slug":"anthropic","name":"Anthropic","display_name":null,"icon_url":"/images/logos/anthropic.jpeg","gradient_from":"from-[#cc785c]","gradient_to":"to-[#d4a574]","gradient_via":null,"website_url":"https://anthropic.com"}},{"id":"holo2-30b-a3b","object":"model","canonical_slug":null,"hugging_face_id":"Hcompany/Holo2-30B-A3B","name":"Holo2 30B A3B","created":1765183068,"description":"Holo2 30B A3B is H Company's multimodal MoE model optimized for GUI understanding, navigation, and visual reasoning.","context_length":22000,"architecture":{"modality":"text+image->text","input_modalities":["text","image"],"output_modalities":["text"],"tokenizer":"Qwen3","instruct_type":null},"top_provider":{"context_length":22000,"max_completion_tokens":32000,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","seed","response_format","tools","tool_choice"],"supported_api_endpoints":["chat.completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Scaleway","slug":"scaleway"}],"pricing":{"prompt":"0.0000003","completion":"0.0000007","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"EUR"},"author":"hcompany","slug":"holo2-30b-a3b","tags":["vision","multimodal","gui","agentic","moe"],"author_info":{"slug":"hcompany","name":"H Company","display_name":"H Company","icon_url":null,"gradient_from":null,"gradient_to":null,"gradient_via":null,"website_url":"https://hcompany.ai"}},{"id":"titan-multimodal-embed","object":"model","canonical_slug":null,"hugging_face_id":"","name":"Amazon Titan Multimodal Embeddings","created":1769967814,"description":"Amazon Titan Multimodal Embeddings generates embeddings for both text and images for cross-modal search.","context_length":8000,"architecture":{"modality":"text+image->embedding","input_modalities":["text","image"],"output_modalities":["embedding"],"tokenizer":null,"instruct_type":null},"top_provider":{"context_length":8000,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":[],"supported_api_endpoints":["embeddings"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"AWS Bedrock","slug":"aws-bedrock"}],"author":"amazon","slug":"titan-multimodal-embed","tags":["proprietary","embedding","multimodal","vision"],"author_info":{"slug":"amazon","name":"Amazon","display_name":"Amazon Web Services","icon_url":"/images/logos/aws.webp","gradient_from":"from-orange-500","gradient_to":"to-yellow-500","gradient_via":null,"website_url":"https://aws.amazon.com/bedrock"}},{"id":"nemotron-3-super-120b-a12b","object":"model","canonical_slug":null,"hugging_face_id":"nvidia/nemotron-3-super-120b-a12b","name":"Nemotron 3 Super 120B A12B","created":1777454224,"description":"NVIDIA Nemotron 3 Super 120B A12B is a Mixture-of-Experts model for chat, coding, reasoning, and tool-use workloads.","context_length":262144,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"Nemotron","instruct_type":null},"top_provider":{"context_length":262144,"max_completion_tokens":262144,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","tools","tool_choice","response_format"],"supported_api_endpoints":["chat.completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Tensorix","slug":"tensorix"}],"pricing":{"prompt":"0.0000003","completion":"0.0000009","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"nvidia","slug":"nemotron-3-super-120b-a12b","tags":["open-source","moe","coding","reasoning","function-calling"],"author_info":{"slug":"nvidia","name":"NVIDIA","display_name":null,"icon_url":"/images/logos/nvidia.webp","gradient_from":"from-[#76b900]","gradient_to":"to-[#5a8c00]","gradient_via":null,"website_url":"https://nvidia.com"}},{"id":"o3-mini","object":"model","canonical_slug":null,"hugging_face_id":"","name":"o3-mini","created":1767619877,"description":"o3-mini is OpenAI's compact text-only reasoning model optimized for cost-efficient coding, math, science, and structured-output tasks.","context_length":200000,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"GPT","instruct_type":null},"top_provider":{"context_length":200000,"max_completion_tokens":100000,"is_moderated":false},"per_request_limits":null,"supported_parameters":["max_tokens","max_completion_tokens","reasoning_effort","tools","tool_choice","response_format","parallel_tool_calls"],"supported_api_endpoints":["chat.completions","responses"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Microsoft Foundry","slug":"microsoft-foundry"}],"pricing":{"prompt":"0.00000121","completion":"0.00000484","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.000000605","input_cache_write":"0","discount":1,"currency":"USD"},"author":"openai","slug":"o3-mini","tags":["reasoning","long-context","efficient","coding","math"],"author_info":{"slug":"openai","name":"OpenAI","display_name":null,"icon_url":"/images/logos/openai.webp","gradient_from":"from-zinc-800","gradient_to":"to-zinc-900","gradient_via":null,"website_url":"https://openai.com"}},{"id":"nemotron-3-nano-30b-a3b","object":"model","canonical_slug":null,"hugging_face_id":"nvidia/NVIDIA-Nemotron-3-Nano-30B-A3B","name":"Nemotron 3 Nano 30B A3B","created":1768127608,"description":"NVIDIA Nemotron 3 Nano 30B A3B is an efficient MoE model for chat, reasoning, coding, and tool use.","context_length":262144,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"Nemotron","instruct_type":null},"top_provider":{"context_length":262144,"max_completion_tokens":262144,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","response_format","tools","tool_choice"],"supported_api_endpoints":["chat.completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[{"name":"Nebius","slug":"nebius"}],"pricing":{"prompt":"0.00000006","completion":"0.00000024","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"nvidia","slug":"nemotron-3-nano-30b-a3b","tags":["open-source","moe","reasoning","coding","function-calling"],"author_info":{"slug":"nvidia","name":"NVIDIA","display_name":null,"icon_url":"/images/logos/nvidia.webp","gradient_from":"from-[#76b900]","gradient_to":"to-[#5a8c00]","gradient_via":null,"website_url":"https://nvidia.com"}},{"id":"llama-3.2-3b-instruct","object":"model","canonical_slug":null,"hugging_face_id":"meta-llama/Llama-3.2-3B-Instruct","name":"Llama 3.2 3B Instruct","created":1769967814,"description":"Llama 3.2 3B Instruct is a compact multilingual model from Meta, optimized for on-device and edge deployments.","context_length":128000,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"Llama3","instruct_type":"llama3"},"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[],"pricing":{"prompt":"0.00000015","completion":"0.00000015","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"meta","slug":"llama-3.2-3b-instruct","tags":["open-source","instruction-tuned","multilingual","compact"],"author_info":{"slug":"meta","name":"Meta","display_name":"Meta AI","icon_url":"/images/logos/meta.webp","gradient_from":"from-blue-600","gradient_to":"to-blue-800","gradient_via":null,"website_url":"https://ai.meta.com"}},{"id":"qwen3-235b-a22b-thinking","object":"model","canonical_slug":null,"hugging_face_id":"","name":"Qwen3 235B A22B Thinking","created":1768127608,"description":"High-depth thinking mode model optimized for extended reasoning, planning, and multi-step tasks.","context_length":262144,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"Qwen3","instruct_type":null},"top_provider":{"context_length":262144,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","tools","tool_choice","response_format"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[],"pricing":{"prompt":"0.0000002","completion":"0.0000008","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"alibaba","slug":"qwen3-235b-a22b-thinking","tags":["moe","reasoning","thinking","planning"],"author_info":{"slug":"alibaba","name":"Alibaba","display_name":"Alibaba (Qwen)","icon_url":"/images/logos/qwen.webp","gradient_from":"from-purple-600","gradient_to":"to-indigo-700","gradient_via":null,"website_url":"https://qwenlm.github.io"}},{"id":"kimi-k2-instruct","object":"model","canonical_slug":null,"hugging_face_id":"","name":"Kimi K2 Instruct","created":1767979257,"description":"Kimi-K2 is Moonshot AI's agentic model built for autonomous tool use and reliable instruction following. Designed for reflex-fast agentic tasks.","context_length":200000,"architecture":{"modality":"text->text","input_modalities":["text"],"output_modalities":["text"],"tokenizer":"Kimi","instruct_type":null},"top_provider":{"context_length":200000,"max_completion_tokens":null,"is_moderated":false},"per_request_limits":null,"supported_parameters":["temperature","top_p","max_tokens","stop","tools","tool_choice","response_format"],"supported_api_endpoints":["chat.completions","completions"],"default_parameters":{"temperature":null,"top_p":null,"top_k":null,"frequency_penalty":null,"presence_penalty":null,"repetition_penalty":null},"providers":[],"pricing":{"prompt":"0.0000005","completion":"0.0000024","request":"0","image":"0","image_token":"0","image_output":"0","audio":"0","input_audio_cache":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0","input_cache_write":"0","discount":1,"currency":"USD"},"author":"moonshot","slug":"kimi-k2-instruct","tags":["agentic","tool-use","instruction-tuned"],"author_info":{"slug":"moonshot","name":"Moonshot","display_name":"Moonshot AI","icon_url":null,"gradient_from":"from-indigo-500","gradient_to":"to-purple-600","gradient_via":null,"website_url":"https://moonshot.ai"}}]}