海量在线大模型 兼容OpenAI API

全部大模型

228个模型 · 2025-02-09 更新
Liquid: LFM 3B
$0.0001/1k
$0.0001/1k
liquid/lfm-3b
Liquid’s LFM 3B delivers incredible performance for its size. It positions itself as first place among 3B parameter transformers, hybrids, and RNN models It is also on par with Phi-3.5-mini on multiple benchmarks, while being 18.4% smaller. LFM-3B is the ideal choice for mobile and other edge text-based applications. See the launch announcement for benchmarks and more info.
2025-01-25 32,768 text->text Other
inflection/inflection-3-productivity
Inflection 3 Productivity is optimized for following instructions. It is better for tasks requiring JSON output or precise adherence to provided guidelines. It has access to recent news. For emotional intelligence similar to Pi, see Inflect 3 Pi See Inflection’s announcement for more details.
2024-10-11 8,000 text->text Other
Inflection: Inflection 3 Pi
$0.010/1k
$0.040/1k
inflection/inflection-3-pi
Inflection 3 Pi powers Inflection’s Pi chatbot, including backstory, emotional intelligence, productivity, and safety. It has access to recent news, and excels in scenarios like customer support and roleplay. Pi has been trained to mirror your tone and style, if you use more emojis, so will Pi! Try experimenting with various prompts and conversation styles.
2024-10-11 8,000 text->text Other
DeepSeek: R1 Distill Qwen 1.5B
$0.0007/1k
$0.0007/1k
deepseek/deepseek-r1-distill-qwen-1.5b
DeepSeek R1 Distill Qwen 1.5B is a distilled large language model based on Qwen 2.5 Math 1.5B, using outputs from DeepSeek R1. It’s a very small and efficient model which outperforms GPT 4o 0513 on Math Benchmarks. Other benchmark results include: AIME 2024 pass@1: 28.9 AIME 2024 cons@64: 52.7 MATH-500 pass@1: 83.9 The model leverages fine-tuning from DeepSeek R1’s outputs, enabling competitive performance comparable to larger frontier models.
2025-01-31 131,072 text->text Other
DeepSeek V2.5
$0.0080/1k
$0.0080/1k
deepseek/deepseek-chat-v2.5
DeepSeek-V2.5 is an upgraded version that combines DeepSeek-V2-Chat and DeepSeek-Coder-V2-Instruct. The new model integrates the general and coding abilities of the two previous versions. For model details, please visit DeepSeek-V2 page for more information.
2024-05-14 8,192 text->text Other
Databricks: DBRX 132B Instruct
$0.0048/1k
$0.0048/1k
databricks/dbrx-instruct
DBRX is a new open source large language model developed by Databricks. At 132B, it outperforms existing open source LLMs like Llama 2 70B and Mixtral-8x7b on standard industry benchmarks for language understanding, programming, math, and logic. It uses a fine-grained mixture-of-experts (MoE) architecture. 36B parameters are active on any input. It was pre-trained on 12T tokens of text and code data. Compared to other open MoE models like Mixtral-8x7B and Grok-1, DBRX is fine-grained, meaning it uses a larger number of smaller experts. See the launch announcement and benchmark results here. moe
2024-03-29 32,768 text->text Other
AionLabs: Aion-RP 1.0 (8B)
$0.0008/1k
$0.0008/1k
aion-labs/aion-rp-llama-3.1-8b
Aion-RP-Llama-3.1-8B ranks the highest in the character evaluation portion of the RPBench-Auto benchmark, a roleplaying-specific variant of Arena-Hard-Auto, where LLMs evaluate each other’s responses. It is a fine-tuned base model rather than an instruct model, designed to produce more natural and varied writing.
2025-02-05 32,768 text->text Other
AionLabs: Aion-1.0-Mini
$0.0032/1k
$0.0096/1k
aion-labs/aion-1.0-mini
Aion-1.0-Mini 32B parameter model is a distilled version of the DeepSeek-R1 model, designed for strong performance in reasoning domains such as mathematics, coding, and logic. It is a modified variant of a FuseAI model that outperforms R1-Distill-Qwen-32B and R1-Distill-Llama-70B, with benchmark results available on its Hugging Face page, independently replicated for verification.
2025-02-05 16,384 text->text Other
AionLabs: Aion-1.0
$0.016/1k
$0.048/1k
aion-labs/aion-1.0
Aion-1.0 is a multi-model system designed for high performance across various tasks, including reasoning and coding. It is built on DeepSeek-R1, augmented with additional models and techniques such as Tree of Thoughts (ToT) and Mixture of Experts (MoE). It is Aion Lab’s most powerful reasoning model.
2025-02-05 16,384 text->text Other
AI21: Jamba Instruct
$0.0020/1k
$0.0028/1k
ai21/jamba-instruct
The Jamba-Instruct model, introduced by AI21 Labs, is an instruction-tuned variant of their hybrid SSM-Transformer Jamba model, specifically optimized for enterprise applications. 256K Context Window: It can process extensive information, equivalent to a 400-page novel, which is beneficial for tasks involving large documents such as financial reports or legal documents Safety and Accuracy: Jamba-Instruct is designed with enhanced safety features to ensure secure deployment in enterprise environments, reducing the risk and cost of implementation Read their announcement to learn more. Jamba has a knowledge cutoff of February 2024.
2024-06-25 256,000 text->text Other
AI21: Jamba 1.5 Mini
$0.0008/1k
$0.0016/1k
ai21/jamba-1-5-mini
Jamba 1.5 Mini is the world’s first production-grade Mamba-based model, combining SSM and Transformer architectures for a 256K context window and high efficiency. It works with 9 languages and can handle various writing and analysis tasks as well as or better than similar small models. This model uses less computer memory and works faster with longer texts than previous designs. Read their announcement to learn more.
2024-08-23 256,000 text->text Other
AI21: Jamba 1.5 Large
$0.0080/1k
$0.032/1k
ai21/jamba-1-5-large
Jamba 1.5 Large is part of AI21’s new family of open models, offering superior speed, efficiency, and quality. It features a 256K effective context window, the longest among open models, enabling improved performance on tasks like document summarization and analysis. Built on a novel SSM-Transformer architecture, it outperforms larger models like Llama 3.1 70B on benchmarks while maintaining resource efficiency. Read their announcement to learn more.
2024-08-23 256,000 text->text Other