mirror of
https://github.com/comfyanonymous/ComfyUI.git
synced 2026-04-05 00:06:38 +08:00
feat: upgrade MiniMax default model to M2.7
- Add MiniMax-M2.7 and MiniMax-M2.7-highspeed to chat model list - Set MiniMax-M2.7 as default model (replacing M2.5) - Keep all previous models as available alternatives
This commit is contained in:
parent
c01e019286
commit
1ac73e81bc
@ -121,6 +121,8 @@ class MinimaxVideoGenerationResponse(BaseModel):
|
|||||||
|
|
||||||
|
|
||||||
class MiniMaxChatModel(str, Enum):
|
class MiniMaxChatModel(str, Enum):
|
||||||
|
M2_7 = 'MiniMax-M2.7'
|
||||||
|
M2_7_highspeed = 'MiniMax-M2.7-highspeed'
|
||||||
M2_5 = 'MiniMax-M2.5'
|
M2_5 = 'MiniMax-M2.5'
|
||||||
M2_5_highspeed = 'MiniMax-M2.5-highspeed'
|
M2_5_highspeed = 'MiniMax-M2.5-highspeed'
|
||||||
|
|
||||||
|
|||||||
@ -452,7 +452,7 @@ class MinimaxChatNode(IO.ComfyNode):
|
|||||||
node_id="MinimaxChatNode",
|
node_id="MinimaxChatNode",
|
||||||
display_name="MiniMax Chat",
|
display_name="MiniMax Chat",
|
||||||
category="api node/text/MiniMax",
|
category="api node/text/MiniMax",
|
||||||
description="Generate text responses using MiniMax language models (MiniMax-M2.5).",
|
description="Generate text responses using MiniMax language models (MiniMax-M2.7).",
|
||||||
inputs=[
|
inputs=[
|
||||||
IO.String.Input(
|
IO.String.Input(
|
||||||
"prompt",
|
"prompt",
|
||||||
@ -463,7 +463,7 @@ class MinimaxChatNode(IO.ComfyNode):
|
|||||||
IO.Combo.Input(
|
IO.Combo.Input(
|
||||||
"model",
|
"model",
|
||||||
options=MiniMaxChatModel,
|
options=MiniMaxChatModel,
|
||||||
default=MiniMaxChatModel.M2_5.value,
|
default=MiniMaxChatModel.M2_7.value,
|
||||||
tooltip="The MiniMax model to use for text generation.",
|
tooltip="The MiniMax model to use for text generation.",
|
||||||
),
|
),
|
||||||
IO.String.Input(
|
IO.String.Input(
|
||||||
@ -524,7 +524,7 @@ class MinimaxChatNode(IO.ComfyNode):
|
|||||||
async def execute(
|
async def execute(
|
||||||
cls,
|
cls,
|
||||||
prompt: str,
|
prompt: str,
|
||||||
model: str = MiniMaxChatModel.M2_5.value,
|
model: str = MiniMaxChatModel.M2_7.value,
|
||||||
system_prompt: Optional[str] = None,
|
system_prompt: Optional[str] = None,
|
||||||
max_tokens: int = 4096,
|
max_tokens: int = 4096,
|
||||||
temperature: float = 0.7,
|
temperature: float = 0.7,
|
||||||
|
|||||||
Loading…
Reference in New Issue
Block a user