MiniMax-M1-80k 现已在 SiliconFlow 上可用

2025年6月17日

目录

MiniMax-M1-80k(456B),世界上首个开源混合注意力模型现已在SiliconFlow上可用。

MiniMax-M1-80k(456B),世界上第一个大规模开源混合注意力模型,现在可以在SiliconFlow上使用。

  • 128K上下文支持

  • 价格竞争力:$0.58/M tokens(Input),$2.29/M tokens(Output)

采用尖端的专家混合(MoE)架构和闪电注意力,MiniMax-M1-80k在长时间上下文推理、编程任务和多步骤工具使用中实现了最先进的性能。

  • 混合注意力+MoE架构 M1结合了专家混合路由的效率与闪电注意力的深度,使其在扩展的同时保持长序列推理质量。

  • 为代理和工具优化 M1支持扩展的上下文和强大的推理,非常适合于自主代理、文档分析和沙盒软件开发等应用。

  • 数学、编码和推理 基准测试显示,M1在需要符号推理、结构化Output和复杂指令执行的任务中与顶级模型竞争。


快速开始

SiliconFlow模型广场上尝试MiniMax-M1-80k模型。


快速访问API

以下Python示例演示了如何通过SiliconFlow的API端点调用MiniMax-M1-80k模型。提供给开发者的更详细的API规范。

from openai import OpenAI

url = 'https://api.siliconflow.com/v1/'
api_key = 'your api_key'

client = OpenAI(
    base_url=url,
    api_key=api_key
)

# Send a request with streaming output
content = ""
reasoning_content = ""
messages = [
    {"role": "user", "content": "Who are the legendary athletes of the Olympics?"}
]
response = client.chat.completions.create(
    model="MiniMaxAI/MiniMax-M1-80k",
    messages=messages,
    stream=True,  # Enable streaming output
    max_tokens=4096,
    extra_body={
        "thinking_budget": 1024
    }
)
# Gradually receive and process the response
for chunk in response:
    if chunk.choices[0].delta.content:
        content += chunk.choices[0].delta.content
    if chunk.choices[0].delta.reasoning_content:
        reasoning_content += chunk.choices[0].delta.reasoning_content

# Round 2
messages.append({"role": "assistant", "content": content})
messages.append({'role': 'user', 'content': "Continue"})
response = client.chat.completions.create(
    model="MiniMaxAI/MiniMax-M1-80k",
    messages=messages,
    stream=True
)
from openai import OpenAI

url = 'https://api.siliconflow.com/v1/'
api_key = 'your api_key'

client = OpenAI(
    base_url=url,
    api_key=api_key
)

# Send a request with streaming output
content = ""
reasoning_content = ""
messages = [
    {"role": "user", "content": "Who are the legendary athletes of the Olympics?"}
]
response = client.chat.completions.create(
    model="MiniMaxAI/MiniMax-M1-80k",
    messages=messages,
    stream=True,  # Enable streaming output
    max_tokens=4096,
    extra_body={
        "thinking_budget": 1024
    }
)
# Gradually receive and process the response
for chunk in response:
    if chunk.choices[0].delta.content:
        content += chunk.choices[0].delta.content
    if chunk.choices[0].delta.reasoning_content:
        reasoning_content += chunk.choices[0].delta.reasoning_content

# Round 2
messages.append({"role": "assistant", "content": content})
messages.append({'role': 'user', 'content': "Continue"})
response = client.chat.completions.create(
    model="MiniMaxAI/MiniMax-M1-80k",
    messages=messages,
    stream=True
)
from openai import OpenAI

url = 'https://api.siliconflow.com/v1/'
api_key = 'your api_key'

client = OpenAI(
    base_url=url,
    api_key=api_key
)

# Send a request with streaming output
content = ""
reasoning_content = ""
messages = [
    {"role": "user", "content": "Who are the legendary athletes of the Olympics?"}
]
response = client.chat.completions.create(
    model="MiniMaxAI/MiniMax-M1-80k",
    messages=messages,
    stream=True,  # Enable streaming output
    max_tokens=4096,
    extra_body={
        "thinking_budget": 1024
    }
)
# Gradually receive and process the response
for chunk in response:
    if chunk.choices[0].delta.content:
        content += chunk.choices[0].delta.content
    if chunk.choices[0].delta.reasoning_content:
        reasoning_content += chunk.choices[0].delta.reasoning_content

# Round 2
messages.append({"role": "assistant", "content": content})
messages.append({'role': 'user', 'content': "Continue"})
response = client.chat.completions.create(
    model="MiniMaxAI/MiniMax-M1-80k",
    messages=messages,
    stream=True
)

MiniMax-M1-80k提供了独特的规模效率推理能力的平衡,专为推动生成型AI极限的开发者而构建。无论您是在构建长时间上下文助手、智能代理还是高级代码协作者——M1已做好准备。

现在,去用MiniMax-M1-80k在SiliconFlow上构建一些非凡的东西吧。

准备好 加速您的人工智能开发吗?

准备好 加速您的人工智能开发吗?

准备好 加速您的人工智能开发吗?