🌟 2025-09-30#
🚀 Claude Sonnet 4.5#
Available Model Names: claude-sonnet-4-5-20250929-thinking,claude-sonnet-4-5-20250929,claude-sonnet-4-5,cometapi-sonnet-4-5-20250929-thinking,cometapi-sonnet-4-5-20250929,cometapi-sonnet-4-5
Claude Sonnet 4.5 has world-leading coding capabilities (SOTA-Level Coding). It achieved an astonishing 77.2% accuracy on the authoritative SWE-bench benchmark, which measures real-world software engineering abilities, making it the world's strongest coding model. This means it has made a qualitative leap in handling complex programming tasks, debugging, and even architectural design.
🚀 DeepSeek-V3.2-Exp Highlights#
The DeepSeek-V3.2-Exp model is an experimental (Experimental) version. As an intermediate step towards the next-generation architecture, V3.2-Exp introduces DeepSeek Sparse Attention (a sparse attention mechanism) based on V3.1, and conducts exploratory optimization and verification for the training and inference efficiency of long texts.
🚀 Gemini 2.5 Flash Highlights#
gemini-2.5-flash-preview-09-2025: A model that excels in cost-effectiveness and provides comprehensive features. 2.5 Flash is best suited for large-scale processing of low-latency, high-data-volume tasks that require thinking, as well as agent application scenarios.
gemini-2.5-flash-lite-preview-09-2025: The fastest Flash model, specially optimized for cost-benefit and high throughput.
🌟 2025-09-24#
Superior Audio Quality: Significantly enhanced audio clarity, vocal performance, and mixing precision.
Immersive Experience: Delivers lifelike vocals and powerful creative control.
Professional Creation: Generates emotionally rich, genre-accurate, high-quality songs.
🛠 Usage#
Set the request parameter mv
to chirp-crow
.🌟 2025-09-23#
🚀 New and Updated Models:#
🔹 grok-4-fast-non-reasoning
#
🔹 grok-4-fast-reasoning
#
🔹 grok-code-fast-1
#
grok-4-fast-non-reasoning
: The non-reasoning variant of xAI's Grok-4 Fast series, with a unified architecture for handling fast responses, suitable for real-time search and simple queries. It possesses extremely powerful technical parameters and ecosystem capabilities: context window supports up to 2,000,000 tokens, cost-efficient (input $0.20/million tokens), leading mainstream models.
grok-4-fast-reasoning
: The reasoning variant of xAI's Grok-4 Fast series, supporting long-chain thinking and tool calls, suitable for complex tasks such as mathematical reasoning and agent workflows. Ranked first in the LMArena search arena (1163 Elo), it possesses extremely powerful technical parameters and ecosystem capabilities: context window supports up to 2,000,000 tokens, leading mainstream models.
grok-code-fast-1
: xAI's fast model specifically designed for agent coding, optimized for tool integration such as grep and file editing, achieving 70.8% performance on SWE-Bench-Verified, suitable for automated code generation and debugging. Currently supports text modality, with vision and other features coming soon. It possesses extremely powerful technical parameters and ecosystem capabilities: context window supports up to 256,000 tokens, leading coding-specific models.
🌟 2025-09-11#
🚀 New and Updated Models: minimax-hailuo-02, bytedance-seedream-4-0-250828, VEO3 Updated!#
🔹 minimax-hailuo-02#
Support for minimax-hailuo-02 model, which is MiniMax's latest masterpiece, an AI video generation model aimed at completely transforming the video creation process. It not only inherits the advantages of the previous generation Hailuo 01, but also achieves a qualitative leap in core technology and user experience.
🔹 bytedance-seedream-4-0-250828#
Support for bytedance-seedream-4-0-250828, as a new-generation image creation model, Seedream 4.0 integrates image generation and image editing capabilities into a unified architecture. This enables it to flexibly handle complex multimodal tasks, including knowledge-based generation, complex reasoning, and reference consistency. Compared to its predecessor, it has faster inference speed and can produce stunning high-definition images up to 4K resolution.
🔹 VEO3#
The entire VEO3 series follows the official price reduction, with comet prices reduced to half of the original, welcome to call.
VEO3 now supports asynchronous interfaces for task processing, optimizing the calling efficiency of long-duration tasks and enhancing the overall experience.
🌟 2025-09-07#
🔹 kimi-k2-250905#
kimi-k2-250905: Moonshot AI's Kimi K2 series 0905 version, supporting ultra-long context (up to 256k tokens, frontend and tool calling).
🧠 Enhanced Tool Calling: 100% accuracy, seamless integration, suitable for complex tasks and integration optimization.
⚡️ More Efficient Performance: TPS up to 60-100 (standard API), up to 600-100 in Turbo mode, providing faster responses and improved reasoning capabilities, with knowledge cutoff to mid-2025.
🔹 qwen3-max-preview#
qwen3-max-preview: Alibaba's Tongyi Qianwen team's latest developed Qwen3-Max-Preview model, positioned as the peak performance in the series.
🧠 Powerful Multimodal and Reasoning: Supports ultra-long context (up to 128k tokens) and multimodal input, excels in complex reasoning, code generation, translation, and creative content.
⚡️ Breakthrough Improvements: Significant optimization in multiple technical indicators, faster response speed, knowledge cutoff to 2025, suitable for enterprise-level high-precision AI applications.
🌟 2025-08-27#
🔹 gemini-2.5-flash-image-preview,gemini-2.5-flash-image#
-gemini-2.5-flash-image-preview
,gemini-2.5-flash-image
:Gemini 2.5 Flash Image (also known as nano-banana) is Google’s most advanced image generation and editing model. This update enables you to blend multiple images into a single image, maintain character consistency to tell richer stories, perform targeted transformations using natural language, and use Gemini’s world knowledge to generate and edit images.
🌟 2025-08-22#
🔹 deepseek-v3.1, deepseek-v3-1-250821#
deepseek-v3.1, deepseek-v3-1-250821: DeepSeek-V3.1 is DeepSeek's all-new hybrid inference model.
🧠 Hybrid inference: Think & Non-Think — one model, two modes
⚡️ Faster thinking: DeepSeek-V3.1 reaches answers in less time vs. DeepSeek-R1-0528
🌟 2025-08-20#
🔹 Suno#
🎵 Introducing Two Major Music Creation Features: Easily add accompaniment to vocals and generate lyrics & vocals for instrumental tracks.
Add Instrumental: Upload an a cappella vocal track, and Suno will intelligently generate and add a matching accompaniment.
Add Vocals: Upload an instrumental track, and Suno will generate lyrics and a vocal performance to match.
🔹 Kling#
✨ Massive Video Effects Library Expansion: Added 63 new video effects (62 single-subject effects and 1 two-person interactive effect), bringing the total to 80 available effects for more creative choices.
🔊 Video-to-Audio Optimization: The video-to-audio generation feature now supports full-resolution video uploads for more precise sound effect matching.
📈 Multi-Image to Video Performance Skyrockets: Experience a 102% improvement over the previous version! See significant enhancements in subject consistency, dynamic quality, and interaction naturalness. This is a seamless upgrade with no code changes required.
🎬 Text-to-Video Quality Upgrade: Version 1.6 now supports the generation of higher-quality videos.
Parameter Example: "mode": "pro"
🎨 Image Generation Model Update: The new kling-v2-new
model is now live, supporting nearly 300 image styles to maximize your creativity!
🌟 2025-08-18#
🚀 New and Updated Models: Runway, VEO3, hunyuan-3D, Midjourney Fully Updated!#
🔹 Runway#
Runway model adds multiple core functions, expanding video and image generation capabilities:
Video to Video: Video to video generation.
Text to Image: Text to image generation.
Video Upscale: Video super-resolution enhancement.
Control a Character: Character control function.
🔹 VEO3#
VEO3 now supports asynchronous interface for task processing, optimizing the calling efficiency of long-duration tasks and enhancing the overall experience.
🔹 Huanyuan3D#
Supports Hunyuan3D-2, providing powerful 3D content creation capabilities to assist in efficiently generating high-quality 3D models.
🌟 2025-08-08#
gpt-5
, gpt-5-2025-08-07
: OpenAI's flagship model, widely recognized as the industry's most powerful for coding, reasoning, and agentic tasks. It is designed to handle the most complex cross-domain challenges and excels in code generation, advanced reasoning, and autonomous agents, making it the premier choice for users demanding peak performance.
gpt-5-chat-latest
: The continuously updated version of GPT-5. It always incorporates the latest features and optimizations, recommended for applications that need to stay current with the latest model capabilities.
gpt-5-mini
, gpt-5-mini-2025-08-07
: The cost-effective version of GPT-5, specifically optimized for speed and cost. It strikes an excellent balance between performance and affordability, making it the ideal choice for everyday tasks like general chat, content creation, and routine Q&A.
gpt-5-nano
, gpt-5-nano-2025-08-07
: The fastest and most cost-effective lightweight version in the GPT-5 family. It is perfect for scenarios requiring high throughput and instant responses, such as text classification, sentiment analysis, summary extraction, and data formatting.
API Call Instructions: gpt-5-chat-latest
should be called using the standard /v1/chat/completions
format. For other models (gpt-5
, gpt-5-mini
, gpt-5-nano
, and their dated versions), using the /v1/responses
format is recommended. For details, please refer to: https://apidoc.cometapi.com/api-13851472 Note#
Important: top_p is not supported by this series of models.
gpt-5-chat-latest
: Supports custom temperature values between 0 and 1 (inclusive).
All other GPT-5 models: The temperature is fixed at 1. You may set it to 1 or omit it (defaults to 1).
When calling the GPT-5 series models (excluding gpt-5-chat-latest), the max_tokens field should be changed to max_completion_tokens.
🌟 2025.08.06#
🔹 claude-opus-4-1-20250805
claude-opus-4-1-20250805
: Anthropic's flagship Claude Opus 4.1 model, achieving major breakthroughs in programming, reasoning, and agentic tasks, with SWE-bench Verified reaching 74.5%.
Significantly enhanced multi-file code refactoring, debugging precision, and detail-oriented reasoning capabilities. This model is suitable for demanding programming and reasoning scenarios.
We have also added cometapi-opus-4-1-20250805
specifically for Cursor integration.
🔹 claude-opus-4-1-20250805-thinking
claude-opus-4-1-20250805-thinking
: Claude Opus 4.1 version with extended thinking capabilities, providing up to 64K tokens of deep reasoning capacity.
Optimized for research, data analysis, and tool-assisted reasoning tasks, with powerful detail-oriented reasoning abilities.
We have also added cometapi-opus-4-1-20250805-thinking
specifically for Cursor integration.
gpt-oss-120b
: OpenAI's released 117B parameter Mixture of Experts (MoE) open-source model, designed for high-level reasoning, agentic, and general production use cases.
gpt-oss-20b
: 21B parameter open-source MoE model with 3.6B active parameter architecture, optimized for low-latency inference and consumer-grade hardware deployment.
🌟 2025.08.05#
🚀 Feature Updates: gemini-2.5-flash-lite, o3 & o4-mini Deep Research, Volcano Engine Generation Modelsgemini-2.5-flash-lite - Google's most cost-effective model, built for large-scale tasks!⚡️ High Efficiency: Designed for large-scale, low-latency applications.
o3 & o4-mini Deep Research Agents - Get in-depth analysis reports with web-connected research agents!🧠 Advanced Analysis: Supports multi-step reasoning and provides reports with citations.
🤖 Available Models: o3-deep-research
, o3-deep-research-2025-06-26
, o4-mini-deep-research
, o4-mini-deep-research-2025-06-26
📚 How to Call: The four deep research models above must be called using the following format:
Volcano Engine Video & Image Models - Experience powerful new video and image models!🎬 Video Generation: Create videos from images (bytedance-seedance-1-0-pro
, bytedance-seedance-1-0-lite-i2v-250428
) or text (bytedance-seedance-1-0-lite-t2v-250428
).
🎨 Image Generation & Editing: Generate images with bytedance-seedream-3.0-t2i
or edit them using prompts with bytedance-seedEdit-3.0-i2i
.
🌟 2025.07.31#
🚀 Feature Updates: MJ Video Generation, Flux-Kontext Multi-Image Reference, Kling-v1-6 Multi-Image ReferenceMJ Video Generation - Transform static images into dynamic video effects!🎬 New capability: MJ original image generation endpoint /mj/submit/imagine
now supports video generation
🎨 Bring creativity to life: Perfect for creating animated effects, creative video generation, and various other applications
Flux-Kontext Series Multi-Image Reference - Enhanced AI creation with multiple references!🖼️ Expanded support: Now supports uploading up to 4 reference images (previously only single image supported)
🎯 Precision boost: Multi-image reference makes AI creation more precise with richer inspiration
🔧 Compatible models: Only supported by black-forest-labs/flux-kontext-max
and black-forest-labs/flux-kontext-pro
models
Kling-v1-6 Multi-Image Reference - Elevate your video generation quality!📸 Multi-image input: Supports up to 4 images as reference input
⚡ Quality enhancement: Significantly improves video generation quality
🌟 2025.07.29#
🔹 Latest Support: glm-4.5
, glm-4.5-air
, glm-4.5-x
, glm-4.5-airx
, glm-4.5-flash
glm-4.5
: Flagship model with 355B total parameters and 32B active parameters, designed for agentic applications, supporting hybrid reasoning modes and excelling in complex reasoning, tool calling, and web browsing.
glm-4.5-air
: Cost-effective model with 106B total parameters and 12B active parameters, maintaining strong performance while significantly reducing costs, ideal for resource-sensitive applications.
glm-4.5-x
: High-performance model optimized for ultra-fast inference and powerful reasoning capabilities, delivering millisecond-level response times for scenarios requiring speed and logic.
glm-4.5-airx
: Lightweight yet powerful model combining Air's cost advantages with X's speed benefits, offering the perfect balance between performance and efficiency.
glm-4.5-flash
: Efficient multi-purpose model with high generation speed, specifically optimized for coding and reasoning tasks, suitable for developers getting started and rapid prototyping.
-Follows the OpenAI chat standard format, see details: CometAPI Chat Documentation🌟 2025.07.25#
🔹 New Model: gemini-2.5-pro-all
, gemini-2.5-flash-all
, gemini-2.5-pro-deepsearch
, gemini-2.5-flash-deepsearch
, deepseek-r1t2-chimera
gemini-2.5-pro-all
: A multimodal version of the Gemini model,
supporting analysis of files, videos, and images, as well as image generation and real-time web access.
gemini-2.5-flash-all
: A multimodal version of the Gemini model,
supporting analysis of files, videos, and images, as well as image generation and real-time web access.
gemini-2.5-pro-deepsearch
: A deep search model with enhanced deep search and information retrieval capabilities
ideal for complex knowledge integration and analysis.
gemini-2.5-flash-deepsearch
: A deep search model combining the rapid performance of the Flash model with advanced deep search capabilities for fast, in-depth information discovery.
deepseek-r1t2-chimera
: A 671B parameter Mixture-of-Experts (MoE) text generation model merged from DeepSeek-AI's R1-0528, R1, and V3-0324, supporting a context of up to 60k tokens.
-Follows the OpenAI chat standard format, see details: CometAPI Chat Documentation🌟 2025.07.24#
qwen3-coder-plus
: Focused on code generation, understanding, and optimization, excels in complex programming tasks.
🔹 qwen3-coder-plus-2025-07-22
qwen3-coder-plus-2025-07-22
: Optimized version from 2025-07-22, stable and reliable, suitable for production.
🔹 qwen3-coder-480b-a35b-instruct
qwen3-coder-480b-a35b-instruct
: Flagship model with 480 billion parameters, MoE architecture, capable of handling extremely complex programming.
🌟 2025.07.18#
Suno v4.5+
: v4.5+ has richer sounds, new creation methods, and a maximum length of 8 minutes. This website currently supports Suno 4.5+. Please change the request parameter mv
to chirp-bluejay
.
🌟 2025.07.17#
CometAPI supports Midjourney uploading masked images for local modifications
🌟 2025.07.16#
kimi-k2-0711-preview
: Kimi K2 is a large-scale mixed-expertise (MoE) language model developed by Moonshot AI.
with 1 trillion total parameters and 32 billion active parameters per forward pass. It is optimized for agent capabilities including advanced tool usage, inference, and code synthesis.
Kimi K2 performs well in a variety of benchmarks, especially in coding (LiveCodeBench, SWE-bench), reasoning (ZebraLogic, GPQA) and tool usage (Tau2, AceBench) tasks.
It supports long contextual inference with up to 128K tokens and features a novel training stack design that includes the MuonClip optimizer for stable large-scale MoE training.
🌟 2025.07.14#
CometAPI now supports direct calls to the OpenAI API to process PDFs without uploading files by providing the URL of the PDF file.
🌟 2025.07.11#
🚀 CometAPI supports Claude code!
• Add power to your development workflow. We're excited to announce that CometAPI now fully supports the powerful Claude Code.
• What does this mean for you?
• Top Artificial Intelligence features: Easily generate, debug and optimize code using models built specifically for developers.
• ⚙️ Flexible Model Selection: Our comprehensive range of models allows you to develop more seamlessly.
• Seamless Integration: APIs are always available. Integrate Claude Code directly into your existing workflow in minutes.
• Ready to build faster? Please click on the link below to make a call.
🌟 2025.07.10#
grok-4
,grok-4-0709
: Currently supports text modal, with visual, image generation and other features coming soon. Extremely powerful technical parameters and ecological capabilities: Context Window: Supports up to 256,000 tokens of contextualization, ahead of mainstream models.
🌟 2025.07.04#
Suno now supports stem separation, creating Persona, generating MP4 MV videos, getting WAV format files, and Timing: lyrics & audio timeline
Now supports Suno's full-track and single-track stem separation features, which can split your songs into up to 12 clean tracks—including vocals, drums, bass, etc., convenient for preview and download.
Create new Persona to generate Persona: singer style; generate music in different formats;
Note: Full-track stem separation feature is priced at 5 times the music generation cost; single-track billing is still being optimized, currently maintaining the same 5x pricing as full-track, will be charged at 1x base price in the future
🔹 veo3
🔹 veo3-pro
🔹 veo3-fast
🔹 veo3-frames
🔹 veo3-fast-frames
🔹 veo3-pro-frames
veo3
,veo3-pro
,veo3-fast
: is the official Google's latest video generation model, the generated video with sound, the world's only video model with sound. veo3-frames,veo3-fast-frames,veo3-pro-frames Support first frame mode.
🌟 2025.07.01#
Midjourney video generation is now supported
Synchronized support for official website low-dynamic, high-dynamic, auto-generated, manual generation.
🌟 2025.07.01#
Now supports the Keyline API to expand the image.
🌟 2025.06.25#
🔹 black-forest-labs/flux-kontext-pro
🔹 black-forest-labs/flux-kontext-max
🔹 flux-kontext-pro
🔹 flux-kontext-max
🌟 2025.06.19#
🔹 gemini-2.5-flash-lite-preview-06-17
gemini-2.5-flash-lite-preview-06-17
: Large scale processing,Lower cost.
🌟 2025.06.11#
🔹 o3-pro
🔹 o3-pro-2025-06-10
o3-pro
,o3-pro-2025-06-10
: Supports web search, file analysis, visual input reasoning, Python programming, and personalized responses.
Compared to previous models, o3-pro shows significant improvements in clarity, completeness, instruction following, and accuracy.
This model adheres to the OpenAI v1/responses standard call format. For specific reference:
curl --location --request POST 'https:
--header 'Authorization: Bearer sk-xxxxxx' \
--header 'User-Agent: Apifox/1.0.0 (https:
--header 'Content-Type: application/json' \
--header 'Accept: *
🌟 2025.06.06#
🔹 gemini-2.5-pro-preview-06-05
gemini-2.5-pro-preview-06-05
: With native multimodal processing capabilities and a very long context window of up to 1 million words (Token), it provides unprecedented power for processing complex, long sequence tasks.
🌟 2025.05.30#
🌟 Gemini Models 1.5 series model off the shelf notice:Due to the fact that Gemini no longer offers the 1.5 series officially, it will be phased out today;
Updated gemini-2.5-flash-preview-05-20
\ gemini-2.5-flash-preview-04-17
\ gemini-2.5-pro-preview-05-06
\ gemini-2.5-pro-preview-03-25
\ gemini-2.5-pro -preview-03-25
\ gemini-2.5-pro-exp-03-25
;* gemini-2.5-pro-exp-03-25
🌟 Notes on using gpt to generate images:Since the interface gpt-4o-image is an interface realized by technical means, not an asynchronous interface, technically it can't achieve complete stability, so there will be instability.
If you have high stability requirements, we recommend using gpt-image-1, the official API call is more stable. Refer to the official images/generations format for the calling method, and https://api.cometapi.com/v1/images/generations for the url details.
🌟 2025.05.29#
deepseek-r1-0528
: Advanced reasoning capabilities, large parameter scale, powerful performance, suitable for complex tasks.
🌟 2025.05.23#
🔹 claude-sonnet-4-20250514
claude-sonnet-4-20250514
: An important model in the Claude 4 series developed by Anthropic, significantly improving coding and reasoning capabilities compared to its predecessor Claude Sonnet 3.7.
It can respond more precisely to user instructions and efficiently handle complex tasks. This model is suitable for applications requiring high performance and cost-effectiveness. We've also added cometapi-sonnet-4-20250514 specifically for use in Cursor.
🔹 claude-sonnet-4-20250514-thinking
claude-sonnet-4-20250514-thinking
: An important model in the Claude 4 series developed by Anthropic, significantly improving coding and reasoning capabilities compared to its predecessor Claude Sonnet 3.7.
It can respond more precisely to user instructions and efficiently handle complex tasks. This model is suitable for applications requiring high performance and cost-effectiveness. We've also added cometapi-sonnet-4-20250514-thinking specifically for use in Cursor.
claude-opus-4-20250514
: Opus 4 is Anthropic's most advanced model, acclaimed as the world's best coding model.
It excels in handling complex, long-running tasks and intelligent agent workflows, particularly suitable for applications requiring high autonomy and intelligence.
🔹 claude-opus-4-20250514-thinking
claude-opus-4-20250514-thinking
: Opus 4 is Anthropic's most advanced model, acclaimed as the world's best coding model.
It excels in handling complex, long-running tasks and intelligent agent workflows, particularly suitable for applications requiring high autonomy and intelligence.
🌟 2025.05.07#
Suno v4.5
: v4.5 has more expressive music and richer vocals, designed to enhance the user's expression and intuition in music creation. This site now supports Suno 4.5, change the request parameter mv to chirp-auk
🌟 2025.04.29#
qwen3-235b-a22b
: This is the flagship model of the Qwen3 series, with 235 billion parameters, utilizing a Mixture of Experts (MoE) architecture.
Particularly suitable for complex tasks requiring high-performance inference, such as coding, mathematics, and multimodal applications.
qwen3-30b-a3b
: With 30 billion parameters, it balances performance and resource requirements, suitable for enterprise-level applications.
This model may use MoE or other optimized architectures, applicable for scenarios requiring efficient processing of complex tasks, such as intelligent customer service and content generation.
qwen3-8b
: A lightweight model with 800 million parameters, designed specifically for resource-constrained environments (such as mobile devices or low-configuration servers).
Its efficiency and fast response capability make it suitable for simple queries, real-time interaction, and lightweight applications.
🌟 2025.04.27#
gpt-image-1 introduces native multimodal models to the API, built on GPT-4o's image generation capabilities, designed to provide developers with a powerful and flexible tool for generating high-quality, diverse images.
Features: High-fidelity images; diverse visual styles; rich world knowledge; consistent text rendering; unlocking practical applications across multiple domains.
{
"model": "gpt-image-1",
"prompt": "A cute baby sea otter",
"n": 1,
"size": "1024x1024"
}
🌟 2025.04.20#
🔹 gemini-2.5-flash-preview-04-17
gemini-2.5-flash-preview-04-17
, Gemini 2.5 Flash is an AI model developed by Google, designed to provide developers with fast and cost-effective solutions, especially suitable for applications requiring enhanced reasoning capabilities.
According to the Gemini 2.5 Flash preview announcement, the model's preview version was released on April 17, 2025, supports multimodal input, and has a context window of up to 1 million tokens.
🌟 2025.04.17#
🔹 o4-mini
🔹 o4-mini-2025-04-16
o4-mini
, o4-mini-2025-04-16
: A smaller, faster, and more economical model, research shows it performs well in mathematics, coding, and visual tasks, designed to be efficient and responsive, suitable for developers. Released on April 16, 2025.
o3
, o3-2025-04-16
: A reflective generative pre-trained transformer (GPT) model designed to handle problems requiring step-by-step logical reasoning.
Research shows it excels at mathematics, coding, and scientific tasks. It can also use tools such as web browsing and image generation, with a release date of April 16, 2025.
🌟 2025.04.15#
gpt-4.1
: Major advancements in coding and instruction following; GPT-4.1 has become the leading model for coding.
Long context: On Video-MME, a benchmark for multimodal long context understanding, GPT-4.1 has created a new state-of-the-art result.
The GPT-4.1 model series delivers superior performance at lower cost.
gpt-4.1-mini
: Represents a significant leap in small model performance, even outperforming GPT-4o on many benchmarks.
It matches or exceeds GPT-4o in intelligence assessment while reducing latency by nearly half and costs by 83%.
gpt-4.1-nano
: Features a larger context window—supporting up to 1 million context tokens
And can better utilize this context through improved long context understanding. Has an updated knowledge cutoff date of June 2024.
🌟 2025.04.14#
grok-3-deepersearch
: Features high data timeliness, excellent interactive experience, and thorough search thinking process; comprehensive webpage aggregation.
🌟 2025.04.13#
🔹 gemini-2.0-flash-exp-image-generation
This model supports conversation while enabling image generation and editing capabilities, outputting high-definition images.
🌟 2025.04.10#
🔹 grok-3-fast
🔹 grok-3-fast-latest
grok-3-fast
, grok-3-fast-latest
: grok-3 and grok-3-fast use exactly the same underlying model and provide the same response quality. However, grok-3-fast is served on faster infrastructure, delivering response times that are much quicker than the standard grok-3.
🔹 grok-3-mini
🔹 grok-3-mini-latest
grok-3-mini
, grok-3-mini-latest
: A lightweight model that thinks before responding. Fast, intelligent, and ideal for logic-based tasks that don't require deep domain knowledge. The original thought traces are accessible.
🔹 grok-3-mini-fast
🔹 grok-3-mini-fast-latest
grok-3-mini-fast
, grok-3-mini-fast-latest
: grok-3-mini and grok-3-mini-fast use exactly the same underlying model and provide the same response quality. However, grok-3-mini-fast is served on faster infrastructure, delivering response times that are much quicker than the standard grok-3-mini.
🌟 2025.04.07#
llama-4-maverick, a high-capacity multimodal language model from Meta, supports multilingual text and image inputs and generates multilingual text and code output in 12 supported languages.
Maverick is optimized for visual language tasks and has instructions tuned for assistant-like behavior, image reasoning, and generic multimodal interaction.
Maverick features native multimodal early fusion and 1 million labeled context windows.
Maverick is released on April 5, 2025 under the Llama 4 Community License for research and commercial applications requiring advanced multimodal understanding and high model throughput.
llama-4-scout, is a mixed-expertise (MoE) language model developed by Meta. It supports native multimodal input (text and images) and multilingual output (text and code) for 12 supported languages.
Designed for assisted interaction and visual reasoning, Scout uses 16 experts per forward pass, a context length of 10 million words, and a training corpus of about 40 trillion words.
Designed for high efficiency and local or commercial deployment, llama-4-scout employs early fusion technology for seamless modal integration.
It is command-tuned for multilingual chat, subtitling, and image comprehension tasks.
It is released under the Llama 4 Community License, with last training data as of August 2024 and a public release on April 5, 2025.
🌟 2025.03.29#
gpt-4o-all has support for ChatGPT's latest generated image mode
gpt-4o-image This model is dedicated to image generation and editing, which enables image style conversion, preservation of original image features, superb consistency, and output of high-definition images.
🌟 2025.03.27#
🔹 gemini-2.5-pro-exp-03-25
Features native multimodal processing capabilities, with an extensive context window of up to 1 million tokens, providing unprecedented powerful support for complex, long-sequence tasks.
🔹 gemini-2.5-pro-preview-03-25
According to Google's data, Gemini 2.5 Pro demonstrates particularly outstanding performance in handling complex tasks.
🌟 2025.03.24#
🔹 gpt-4.5-preview-2025-02-27
Preview Version: Showcasing the latest features of GPT-4.5, providing enhanced understanding and generation capabilities, suitable for various tasks, improving user experience.
Preview Version: Deeply optimized algorithms and performance, delivering ultra-fast responses and precise outputs, perfectly suited for efficient decision-making scenarios.
Professional Standard Version: Stable and reliable, combining rich expression and multi-task processing capabilities, suitable for wide applications including business, education, creative, and technical fields.
🌟 2025.02.20#
🔹 claude-3-7-sonnet-thinking
Advanced model designed for complex reasoning and creative thinking, unleashing unlimited possibilities, empowering breakthrough problem-solving and innovation.
🔹 claude-3-7-sonnet-20250219
High-end version integrating the latest technological breakthroughs, handling complex tasks with superior performance, providing intelligent innovative solutions for users.
Outstanding multi-domain processing expert, delivering precise and smooth output experience, easily tackling various professional challenges.
🔹 cometapi-3-7-sonnet-thinking
Equipped with revolutionary algorithm architecture, significantly enhancing deep analysis and complex task management capabilities, making thinking more thorough and comprehensive.
🔗 Usage Guide:#
✅ All models have been added to the default group, allowing you to flexibly call them according to different usage scenarios and requirements, easily integrate, and maximize their application value.🛠 Quick Start:#
Simple integration into your system unlocks powerful capabilities. Fully utilize each model's unique advantages to meet professional needs across different domains.🔥 Experience the revolutionary performance improvements these breakthrough models bring right now! 🔥
For professional support or detailed consultation, please contact our customer service team or visit our technical documentation center. We look forward to your valuable feedback!Modified at 2025-10-04 22:59:56