海量在线大模型 兼容OpenAI API

全部大模型

228个模型 · 2025-02-09 更新
qwen/qwen-2-7b-instruct:free
Qwen2 7B is a transformer-based model that excels in language understanding, multilingual capabilities, coding, mathematics, and reasoning. It features SwiGLU activation, attention QKV bias, and group query attention. It is pretrained on extensive data with supervised finetuning and direct preference optimization. For more details, see this blog post and GitHub repo. Usage of this model is subject to Tongyi Qianwen LICENSE AGREEMENT.
2024-07-16 8,192 text->text Qwen
Qwen 2 7B Instruct
$0.0002/1k
$0.0002/1k
qwen/qwen-2-7b-instruct
Qwen2 7B is a transformer-based model that excels in language understanding, multilingual capabilities, coding, mathematics, and reasoning. It features SwiGLU activation, attention QKV bias, and group query attention. It is pretrained on extensive data with supervised finetuning and direct preference optimization. For more details, see this blog post and GitHub repo. Usage of this model is subject to Tongyi Qianwen LICENSE AGREEMENT.
2024-07-16 32,768 text->text Qwen
Qwen 2 72B Instruct
$0.0036/1k
$0.0036/1k
qwen/qwen-2-72b-instruct
Qwen2 72B is a transformer-based model that excels in language understanding, multilingual capabilities, coding, mathematics, and reasoning. It features SwiGLU activation, attention QKV bias, and group query attention. It is pretrained on extensive data with supervised finetuning and direct preference optimization. For more details, see this blog post and GitHub repo. Usage of this model is subject to Tongyi Qianwen LICENSE AGREEMENT.
2024-06-07 32,768 text->text Qwen
Magnum v4 72B
$0.0075/1k
$0.0090/1k
anthracite-org/magnum-v4-72b
This is a series of models designed to replicate the prose quality of the Claude 3 models, specifically Sonnet(https://openrouter.ai/anthropic/claude-3.5-sonnet) and Opus(https://openrouter.ai/anthropic/claude-3-opus). The model is fine-tuned on top of Qwen2.5 72B.
2024-10-22 16,384 text->text Qwen
Magnum v2 72B
$0.012/1k
$0.012/1k
anthracite-org/magnum-v2-72b
From the maker of Goliath, Magnum 72B is the seventh in a family of models designed to achieve the prose quality of the Claude 3 models, notably Opus & Sonnet. The model is based on Qwen2 72B and trained with 55 million tokens of highly curated roleplay (RP) data.
2024-09-30 32,768 text->text Qwen
Magnum 72B
$0.0075/1k
$0.0090/1k
alpindale/magnum-72b
From the maker of Goliath, Magnum 72B is the first in a new family of models designed to achieve the prose quality of the Claude 3 models, notably Opus & Sonnet. The model is based on Qwen2 72B and trained with 55 million tokens of highly curated roleplay (RP) data.
2024-07-11 16,384 text->text Qwen
EVA Qwen2.5 72B
$0.016/1k
$0.024/1k
eva-unit-01/eva-qwen-2.5-72b
EVA Qwen2.5 72B is a roleplay and storywriting specialist model. It’s a full-parameter finetune of Qwen2.5-72B on mixture of synthetic and natural data. It uses Celeste 70B 0.1 data mixture, greatly expanding it to improve versatility, creativity and “flavor” of the resulting model.
2024-11-22 16,384 text->text Qwen
EVA Qwen2.5 32B
$0.010/1k
$0.014/1k
eva-unit-01/eva-qwen-2.5-32b
EVA Qwen2.5 32B is a roleplaying/storywriting specialist model. It’s a full-parameter finetune of Qwen2.5-32B on mixture of synthetic and natural data. It uses Celeste 70B 0.1 data mixture, greatly expanding it to improve versatility, creativity and “flavor” of the resulting model.
2024-11-09 16,384 text->text Qwen
DeepSeek: R1 Distill Qwen 32B
$0.0005/1k
$0.0007/1k
deepseek/deepseek-r1-distill-qwen-32b
DeepSeek R1 Distill Qwen 32B is a distilled large language model based on Qwen 2.5 32B, using outputs from DeepSeek R1. It outperforms OpenAI’s o1-mini across various benchmarks, achieving new state-of-the-art results for dense models. Other benchmark results include: AIME 2024 pass@1: 72.6 MATH-500 pass@1: 94.3 CodeForces Rating: 1691 The model leverages fine-tuning from DeepSeek R1’s outputs, enabling competitive performance comparable to larger frontier models.
2025-01-30 131,072 text->text Qwen
DeepSeek: R1 Distill Qwen 14B
$0.0006/1k
$0.0006/1k
deepseek/deepseek-r1-distill-qwen-14b
DeepSeek R1 Distill Qwen 14B is a distilled large language model based on Qwen 2.5 14B, using outputs from DeepSeek R1. It outperforms OpenAI’s o1-mini across various benchmarks, achieving new state-of-the-art results for dense models. Other benchmark results include: AIME 2024 pass@1: 69.7 MATH-500 pass@1: 93.9 CodeForces Rating: 1481 The model leverages fine-tuning from DeepSeek R1’s outputs, enabling competitive performance comparable to larger frontier models.
2025-01-30 64,000 text->text Qwen
01.AI: Yi Large
$0.012/1k
$0.012/1k
01-ai/yi-large
The Yi Large model was designed by 01.AI with the following usecases in mind: knowledge search, data classification, human-like chat bots, and customer service. It stands out for its multilingual proficiency, particularly in Spanish, Chinese, Japanese, German, and French. Check out the launch announcement to learn more.
2024-06-25 32,768 text->text Yi
Amazon: Nova Pro 1.0
$0.0032/1k
$0.013/1k
amazon/nova-pro-v1
Amazon Nova Pro 1.0 is a capable multimodal model from Amazon focused on providing a combination of accuracy, speed, and cost for a wide range of tasks. As of December 2024, it achieves state-of-the-art performance on key benchmarks including visual question answering (TextVQA) and video understanding (VATEX). Amazon Nova Pro demonstrates strong capabilities in processing both visual and textual information and at analyzing financial documents. NOTE: Video input is not supported at this time.
2024-12-06 300,000 text+image->text Nova