chore(checkpoint): Phase 6 Test Suite Stabilization complete. 257/261 tests PASS. Resolved run_linear drift, formatter expectations, and Hook Server startup.

This commit is contained in:
2026-02-28 20:42:54 -05:00
parent 21496ee58f
commit 3b96b67d69
17 changed files with 106 additions and 75 deletions

View File

@@ -1,20 +1,21 @@
import sys, json, os, subprocess
prompt = sys.stdin.read()
if '"role": "tool"' in prompt:
print(json.dumps({"type": "message", "role": "assistant", "content": "Tool worked!"}), flush=True)
print(json.dumps({"type": "result", "stats": {"total_tokens": 20}}), flush=True)
print(json.dumps({"type": "message", "role": "assistant", "content": "Tool worked!"}), flush=True)
print(json.dumps({"type": "result", "stats": {"total_tokens": 20}}), flush=True)
else:
# We must call the bridge to trigger the GUI approval!
tool_call = {"name": "list_directory", "input": {"dir_path": "."}}
bridge_cmd = [sys.executable, "C:/projects/manual_slop/scripts/cli_tool_bridge.py"]
proc = subprocess.Popen(bridge_cmd, stdin=subprocess.PIPE, stdout=subprocess.PIPE, text=True)
stdout, _ = proc.communicate(input=json.dumps(tool_call))
# Even if bridge says allow, we emit the tool_use to the adapter
print(json.dumps({"type": "message", "role": "assistant", "content": "I will list the directory."}), flush=True)
print(json.dumps({
"type": "tool_use",
"name": "list_directory",
"id": "alias_call",
"args": {"dir_path": "."}
}), flush=True)
print(json.dumps({"type": "result", "stats": {"total_tokens": 10}}), flush=True)
# We must call the bridge to trigger the GUI approval!
tool_call = {"name": "list_directory", "input": {"dir_path": "."}}
bridge_cmd = [sys.executable, "C:/projects/manual_slop/scripts/cli_tool_bridge.py"]
proc = subprocess.Popen(bridge_cmd, stdin=subprocess.PIPE, stdout=subprocess.PIPE, text=True)
stdout, _ = proc.communicate(input=json.dumps(tool_call))
# Even if bridge says allow, we emit the tool_use to the adapter
print(json.dumps({"type": "message", "role": "assistant", "content": "I will list the directory."}), flush=True)
print(json.dumps({
"type": "tool_use",
"name": "list_directory",
"id": "alias_call",
"args": {"dir_path": "."}
}), flush=True)
print(json.dumps({"type": "result", "stats": {"total_tokens": 10}}), flush=True)

View File

@@ -10,5 +10,5 @@ auto_add = true
[discussions.main]
git_commit = ""
last_updated = "2026-02-28T07:35:03"
last_updated = "2026-02-28T20:39:12"
history = []

View File

@@ -6,10 +6,10 @@ roles = [
"Reasoning",
]
history = []
active = "TestDisc_1772282083"
active = "TestDisc_1772329132"
auto_add = true
[discussions.TestDisc_1772282083]
[discussions.TestDisc_1772329132]
git_commit = ""
last_updated = "2026-02-28T07:34:56"
last_updated = "2026-02-28T20:39:05"
history = []

View File

@@ -10,5 +10,5 @@ auto_add = true
[discussions.main]
git_commit = ""
last_updated = "2026-02-28T19:35:01"
last_updated = "2026-02-28T20:41:54"
history = []

View File

@@ -10,5 +10,5 @@ auto_add = true
[discussions.main]
git_commit = ""
last_updated = "2026-02-28T07:35:20"
last_updated = "2026-02-28T20:39:30"
history = []

View File

@@ -15,6 +15,8 @@ output_dir = "./md_gen"
base_dir = "."
paths = []
[files.tier_assignments]
[screenshots]
base_dir = "."
paths = []

View File

@@ -5,18 +5,10 @@ roles = [
"System",
"Reasoning",
]
active = "AutoDisc"
active = "main"
auto_add = true
[discussions.main]
git_commit = ""
last_updated = "2026-02-27T18:58:03"
history = [
"@2026-02-27T18:57:55\nUser:\nHello! This is an automated test. Just say 'Acknowledged'.",
"@2026-02-27T18:58:02\nAI:\nAcknowledged",
]
[discussions.AutoDisc]
git_commit = ""
last_updated = "2026-02-28T07:34:41"
last_updated = "2026-02-28T20:41:57"
history = []

View File

@@ -1,4 +1,4 @@
import pytest
import pytest
import time
import os
import sys
@@ -86,7 +86,7 @@ else:
client.set_value("ai_input", "Test parameter aliases")
client.click("btn_gen_send")
# Handle approval
timeout = 15
timeout = 60
start_time = time.time()
approved = False
while time.time() - start_time < timeout:
@@ -124,7 +124,7 @@ def test_gemini_cli_loop_termination(live_gui: Any) -> None:
client.set_value("ai_input", "Perform multi-round tool test")
client.click("btn_gen_send")
# Handle approvals (mock does one tool call)
timeout = 20
timeout = 60
start_time = time.time()
approved = False
while time.time() - start_time < timeout:
@@ -148,3 +148,4 @@ def test_gemini_cli_loop_termination(live_gui: Any) -> None:
if found_final: break
time.sleep(1)
assert found_final, "Final message after multi-round tool loop not found"

View File

@@ -1,4 +1,4 @@
from typing import Any
from typing import Any
import pytest
import time
import os
@@ -36,7 +36,7 @@ def test_gemini_cli_full_integration(live_gui: Any) -> None:
client.click("btn_gen_send")
# 3. Monitor for approval events
print("[TEST] Waiting for approval events...")
timeout = 45
timeout = 90
start_time = time.time()
approved_count = 0
while time.time() - start_time < timeout:
@@ -89,7 +89,7 @@ def test_gemini_cli_rejection_and_history(live_gui: Any) -> None:
client.set_value("ai_input", "Deny me")
client.click("btn_gen_send")
# 3. Wait for event and reject
timeout = 20
timeout = 60
start_time = time.time()
denied = False
while time.time() - start_time < timeout:
@@ -120,3 +120,4 @@ def test_gemini_cli_rejection_and_history(live_gui: Any) -> None:
if rejection_found: break
time.sleep(1.0)
assert rejection_found, "Rejection message not found in history"

View File

@@ -0,0 +1,5 @@
import pytest
def test_vlogger_available(vlogger):
vlogger.log_state("Test", "Before", "After")
vlogger.finalize("Test Title", "PASS", "Test Result")