海量在线大模型 兼容OpenAI API

全部大模型

320个模型 · 2025-07-23 更新
Meta: Llama 3.1 405B Instruct
$0.0032/1k
$0.0032/1k
meta-llama/llama-3.1-405b-instruct
The highly anticipated 400B class of Llama3 is here! Clocking in at 128k context with impressive eval scores, the Meta AI team continues to push the frontier of open-source LLMs. Meta's latest class of model (Llama 3.1) launched with a variety of sizes & flavors. This 405B instruct-tuned version is optimized for high quality dialogue usecases. It has demonstrated strong performance compared to leading closed-source models including GPT-4o and Claude 3.5 Sonnet in evaluations. To read more about the model release, click here. Usage of this model is subject to Meta's Acceptable Use Policy.
2024-07-23 32,768 text->text Llama3
Meta: Llama 3 8B Instruct
$0.0001/1k
$0.0002/1k
meta-llama/llama-3-8b-instruct
Meta's latest class of model (Llama 3) launched with a variety of sizes & flavors. This 8B instruct-tuned version was optimized for high quality dialogue usecases. It has demonstrated strong performance compared to leading closed-source models in human evaluations. To read more about the model release, click here. Usage of this model is subject to Meta's Acceptable Use Policy.
2024-04-18 8,192 text->text Llama3
Meta: Llama 3 70B Instruct
$0.0012/1k
$0.0016/1k
meta-llama/llama-3-70b-instruct
Meta's latest class of model (Llama 3) launched with a variety of sizes & flavors. This 70B instruct-tuned version was optimized for high quality dialogue usecases. It has demonstrated strong performance compared to leading closed-source models in human evaluations. To read more about the model release, click here. Usage of this model is subject to Meta's Acceptable Use Policy.
2024-04-18 8,192 text->text Llama3
Llama Guard 3 8B
$0.0001/1k
$0.0002/1k
meta-llama/llama-guard-3-8b
Llama Guard 3 is a Llama-3.1-8B pretrained model, fine-tuned for content safety classification. Similar to previous versions, it can be used to classify content in both LLM inputs (prompt classification) and in LLM responses (response classification). It acts as an LLM – it generates text in its output that indicates whether a given prompt or response is safe or unsafe, and if unsafe, it also lists the content categories violated. Llama Guard 3 was aligned to safeguard against the MLCommons standardized hazards taxonomy and designed to support Llama 3.1 capabilities. Specifically, it provides content moderation in 8 languages, and was optimized to support safety and security for search and code interpreter tool calls.
2025-02-13 131,072 text->text Llama3
DeepSeek: R1 Distill Llama 8B
$0.0002/1k
$0.0002/1k
deepseek/deepseek-r1-distill-llama-8b
DeepSeek R1 Distill Llama 8B is a distilled large language model based on Llama-3.1-8B-Instruct, using outputs from DeepSeek R1. The model combines advanced distillation techniques to achieve high performance across multiple benchmarks, including: AIME 2024 pass@1: 50.4 MATH-500 pass@1: 89.1 CodeForces Rating: 1205 The model leverages fine-tuning from DeepSeek R1's outputs, enabling competitive performance comparable to larger frontier models. Hugging Face: - Llama-3.1-8B - DeepSeek-R1-Distill-Llama-8B |
2025-02-07 32,000 text->text Llama3
deepseek/deepseek-r1-distill-llama-70b:free
DeepSeek R1 Distill Llama 70B is a distilled large language model based on Llama-3.3-70B-Instruct, using outputs from DeepSeek R1. The model combines advanced distillation techniques to achieve high performance across multiple benchmarks, including: AIME 2024 pass@1: 70.0 MATH-500 pass@1: 94.5 CodeForces Rating: 1633 The model leverages fine-tuning from DeepSeek R1's outputs, enabling competitive performance comparable to larger frontier models.
2025-01-24 8,192 text->text Llama3
DeepSeek: R1 Distill Llama 70B
$0.0002/1k
$0.0002/1k
deepseek/deepseek-r1-distill-llama-70b
DeepSeek R1 Distill Llama 70B is a distilled large language model based on Llama-3.3-70B-Instruct, using outputs from DeepSeek R1. The model combines advanced distillation techniques to achieve high performance across multiple benchmarks, including: AIME 2024 pass@1: 70.0 MATH-500 pass@1: 94.5 CodeForces Rating: 1633 The model leverages fine-tuning from DeepSeek R1's outputs, enabling competitive performance comparable to larger frontier models.
2025-01-24 131,072 text->text Llama3
xAI: Grok Vision Beta
$0.020/1k
$0.060/1k
x-ai/grok-vision-beta
Grok Vision Beta is xAI's experimental language model with vision capability.
2024-11-19 8,192 text+image->text Grok
xAI: Grok 4
$0.012/1k
$0.060/1k
x-ai/grok-4
Grok 4 is xAI's latest reasoning model with a 256k context window. It supports parallel tool calling, structured outputs, and both image and text inputs. Note that reasoning is not exposed, reasoning cannot be disabled, and the reasoning effort cannot be specified. Pricing increases once the total tokens in a given request is greater than 128k tokens. See more details on the xAI docs
2025-07-10 256,000 text+image->text Grok
xAI: Grok 3 Mini Beta
$0.0012/1k
$0.0020/1k
x-ai/grok-3-mini-beta
Grok 3 Mini is a lightweight, smaller thinking model. Unlike traditional models that generate answers immediately, Grok 3 Mini thinks before responding. It’s ideal for reasoning-heavy tasks that don’t demand extensive domain knowledge, and shines in math-specific and quantitative use cases, such as solving challenging puzzles or math problems. Transparent "thinking" traces accessible. Defaults to low reasoning, can boost with setting reasoning: { effort: "high" } Note: That there are two xAI endpoints for this model. By default when using this model we will always route you to the base endpoint. If you want the fast endpoint you can add provider: { sort: throughput}, to sort by throughput instead.
2025-04-10 131,072 text->text Grok
xAI: Grok 3 Mini
$0.0012/1k
$0.0020/1k
x-ai/grok-3-mini
A lightweight model that thinks before responding. Fast, smart, and great for logic-based tasks that do not require deep domain knowledge. The raw thinking traces are accessible.
2025-06-11 131,072 text->text Grok
xAI: Grok 3 Beta
$0.012/1k
$0.060/1k
x-ai/grok-3-beta
Grok 3 is the latest model from xAI. It's their flagship model that excels at enterprise use cases like data extraction, coding, and text summarization. Possesses deep domain knowledge in finance, healthcare, law, and science. Excels in structured tasks and benchmarks like GPQA, LCB, and MMLU-Pro where it outperforms Grok 3 Mini even on high thinking. Note: That there are two xAI endpoints for this model. By default when using this model we will always route you to the base endpoint. If you want the fast endpoint you can add provider: { sort: throughput}, to sort by throughput instead.
2025-04-10 131,072 text->text Grok