海量在线大模型 兼容OpenAI API

AllenAI: Molmo 7B D

$0.0004/1k
$0.0008/1k
开始对话
allenai/molmo-7b-d
上下文长度: 4,096 text+image->text Other 2025-03-27 更新
Molmo is a family of open vision-language models developed by the Allen Institute for AI. Molmo models are trained on PixMo, a dataset of 1 million, highly-curated image-text pairs. It has state-of-the-art performance among multimodal models with a similar size while being fully open-source. You can find all models in the Molmo family here. Learn more about the Molmo family in the announcement blog post or the paper. Molmo 7B-D is based on Qwen2-7B and uses OpenAI CLIP as vision backbone. It performs comfortably between GPT-4V and GPT-4o on both academic benchmarks and human evaluation. This checkpoint is a preview of the Molmo release. All artifacts used in creating Molmo (PixMo dataset, training code, evaluations, intermediate checkpoints) will be made available at a later date, furthering our commitment to open-source AI development and reproducibility.

模型参数

架构信息

模态: text+image->text
Tokenizer: Other

限制信息

上下文长度: 4,096
最大回复长度: 4,096