fix: Use temperature=1.0 for MiniMax (required range is (0.0, 1.0])

This commit is contained in:
2026-03-06 23:46:17 -05:00
parent ce53f69ae0
commit 6cfe9697e0
2 changed files with 24 additions and 24 deletions

View File

@@ -74,7 +74,7 @@ DockId=0xAFC85805,2
[Window][Theme] [Window][Theme]
Pos=0,17 Pos=0,17
Size=32,960 Size=517,960
Collapsed=0 Collapsed=0
DockId=0x00000005,1 DockId=0x00000005,1
@@ -84,14 +84,14 @@ Size=900,700
Collapsed=0 Collapsed=0
[Window][Diagnostics] [Window][Diagnostics]
Pos=829,990 Pos=2989,1760
Size=851,210 Size=851,377
Collapsed=0 Collapsed=0
DockId=0x00000002,0 DockId=0x00000002,0
[Window][Context Hub] [Window][Context Hub]
Pos=0,17 Pos=0,17
Size=32,960 Size=517,960
Collapsed=0 Collapsed=0
DockId=0x00000005,0 DockId=0x00000005,0
@@ -102,26 +102,26 @@ Collapsed=0
DockId=0x0000000D,0 DockId=0x0000000D,0
[Window][Discussion Hub] [Window][Discussion Hub]
Pos=430,17 Pos=1750,17
Size=397,637 Size=1237,1142
Collapsed=0 Collapsed=0
DockId=0x00000013,0 DockId=0x00000013,0
[Window][Operations Hub] [Window][Operations Hub]
Pos=34,17 Pos=519,17
Size=394,637 Size=1229,1142
Collapsed=0 Collapsed=0
DockId=0x00000012,0 DockId=0x00000012,0
[Window][Files & Media] [Window][Files & Media]
Pos=0,979 Pos=0,979
Size=32,221 Size=517,1158
Collapsed=0 Collapsed=0
DockId=0x00000006,1 DockId=0x00000006,1
[Window][AI Settings] [Window][AI Settings]
Pos=0,979 Pos=0,979
Size=32,221 Size=517,1158
Collapsed=0 Collapsed=0
DockId=0x00000006,0 DockId=0x00000006,0
@@ -131,14 +131,14 @@ Size=416,325
Collapsed=0 Collapsed=0
[Window][MMA Dashboard] [Window][MMA Dashboard]
Pos=829,17 Pos=2989,17
Size=851,971 Size=851,1741
Collapsed=0 Collapsed=0
DockId=0x00000001,0 DockId=0x00000001,0
[Window][Log Management] [Window][Log Management]
Pos=829,17 Pos=2989,17
Size=851,971 Size=851,1741
Collapsed=0 Collapsed=0
DockId=0x00000001,1 DockId=0x00000001,1
@@ -148,26 +148,26 @@ Size=262,209
Collapsed=0 Collapsed=0
[Window][Tier 1: Strategy] [Window][Tier 1: Strategy]
Pos=34,656 Pos=519,1161
Size=165,544 Size=513,976
Collapsed=0 Collapsed=0
DockId=0x00000014,0 DockId=0x00000014,0
[Window][Tier 2: Tech Lead] [Window][Tier 2: Tech Lead]
Pos=201,656 Pos=1034,1161
Size=228,544 Size=714,976
Collapsed=0 Collapsed=0
DockId=0x00000016,0 DockId=0x00000016,0
[Window][Tier 4: QA] [Window][Tier 4: QA]
Pos=696,656 Pos=2576,1161
Size=131,544 Size=411,976
Collapsed=0 Collapsed=0
DockId=0x00000019,0 DockId=0x00000019,0
[Window][Tier 3: Workers] [Window][Tier 3: Workers]
Pos=431,656 Pos=1750,1161
Size=263,544 Size=824,976
Collapsed=0 Collapsed=0
DockId=0x00000018,0 DockId=0x00000018,0
@@ -258,7 +258,7 @@ Column 3 Width=100
DockNode ID=0x00000008 Pos=3125,170 Size=593,1157 Split=Y DockNode ID=0x00000008 Pos=3125,170 Size=593,1157 Split=Y
DockNode ID=0x00000009 Parent=0x00000008 SizeRef=1029,147 Selected=0x0469CA7A DockNode ID=0x00000009 Parent=0x00000008 SizeRef=1029,147 Selected=0x0469CA7A
DockNode ID=0x0000000A Parent=0x00000008 SizeRef=1029,145 Selected=0xDF822E02 DockNode ID=0x0000000A Parent=0x00000008 SizeRef=1029,145 Selected=0xDF822E02
DockSpace ID=0xAFC85805 Window=0x079D3A04 Pos=0,17 Size=1680,1183 Split=X DockSpace ID=0xAFC85805 Window=0x079D3A04 Pos=0,17 Size=3840,2120 Split=X
DockNode ID=0x00000003 Parent=0xAFC85805 SizeRef=2987,1183 Split=X DockNode ID=0x00000003 Parent=0xAFC85805 SizeRef=2987,1183 Split=X
DockNode ID=0x0000000B Parent=0x00000003 SizeRef=404,1186 Split=X Selected=0xF4139CA2 DockNode ID=0x0000000B Parent=0x00000003 SizeRef=404,1186 Split=X Selected=0xF4139CA2
DockNode ID=0x00000007 Parent=0x0000000B SizeRef=517,858 Split=Y Selected=0x8CA2375C DockNode ID=0x00000007 Parent=0x0000000B SizeRef=517,858 Split=Y Selected=0x8CA2375C

View File

@@ -1772,7 +1772,7 @@ def _send_minimax(md_content: str, user_message: str, base_dir: str,
if stream: if stream:
request_payload["stream_options"] = {"include_usage": True} request_payload["stream_options"] = {"include_usage": True}
request_payload["temperature"] = _temperature request_payload["temperature"] = 1.0
request_payload["max_tokens"] = min(_max_tokens, 8192) request_payload["max_tokens"] = min(_max_tokens, 8192)
tools = _get_deepseek_tools() tools = _get_deepseek_tools()