Tier 2 dev workflow: Windows test runner + result sync

- run_tests.bat: double-click test runner with JSON result capture
- run_script.bat: run any script with output capture
- test_results/ folder for Syncthing-based result sharing
- Auto-mark NX-dependent tests for --quick mode
- pytest-json-report for structured results
This commit is contained in:
2026-03-07 14:07:32 +00:00
parent a069a9f21f
commit b3162aa78d
5 changed files with 298 additions and 0 deletions

4
.gitignore vendored
View File

@@ -142,3 +142,7 @@ C:*
# project-context-sync (auto-generated, local only)
PROJECT_STATE.md
# Test results (synced via Syncthing, not git)
test_results/*.json
test_results/*.log

View File

@@ -68,3 +68,4 @@ pytest-cov>=4.1.0
black>=23.12.0
ruff>=0.1.0
mypy>=1.8.0
pytest-json-report

91
run_script.bat Normal file
View File

@@ -0,0 +1,91 @@
@echo off
REM ============================================================================
REM Atomizer Script Runner — Run any Python script with result capture
REM ============================================================================
REM Usage:
REM run_script.bat path\to\script.py [args...]
REM
REM Results sync back to Mario via Syncthing in test_results/
REM ============================================================================
setlocal enabledelayedexpansion
set "ATOMIZER_ROOT=%~dp0"
set "RESULTS_DIR=%ATOMIZER_ROOT%test_results"
set "SCRIPT=%~1"
if "%SCRIPT%"=="" (
echo Usage: run_script.bat path\to\script.py [args...]
pause
exit /b 1
)
REM Timestamp
for /f %%i in ('python -c "from datetime import datetime; print(datetime.now().strftime('%%Y-%%m-%%d_%%H-%%M-%%S'))"') do set "TIMESTAMP=%%i"
set "LOG_FILE=%RESULTS_DIR%\script_%TIMESTAMP%.log"
set "RUN_FILE=%RESULTS_DIR%\script_%TIMESTAMP%.json"
if not exist "%RESULTS_DIR%" mkdir "%RESULTS_DIR%"
echo.
echo ============================================================================
echo Running: %SCRIPT%
echo %date% %time%
echo ============================================================================
echo.
cd /d "%ATOMIZER_ROOT%"
REM Shift past first arg to get remaining args
set "EXTRA_ARGS="
shift
:argloop
if not "%~1"=="" (
set "EXTRA_ARGS=!EXTRA_ARGS! %~1"
shift
goto argloop
)
REM Run the script
python "%SCRIPT%" %EXTRA_ARGS% > "%LOG_FILE%" 2>&1
set "EXIT_CODE=!errorlevel!"
REM Also echo to console
type "%LOG_FILE%"
echo.
echo ============================================================================
echo Exit code: %EXIT_CODE%
REM Generate result JSON
python -c "
import json, os
from datetime import datetime
with open(r'%LOG_FILE%', 'r', encoding='utf-8', errors='replace') as f:
log = f.read()
# Grab last 50 lines for quick review
lines = log.strip().split('\n')
tail = lines[-50:] if len(lines) > 50 else lines
result = {
'timestamp': datetime.now().isoformat(),
'type': 'script',
'script': '%SCRIPT%',
'args': '%EXTRA_ARGS%'.strip(),
'exit_code': int('%EXIT_CODE%'),
'status': 'OK' if int('%EXIT_CODE%') == 0 else 'ERROR',
'output_tail': tail,
'log_file': os.path.basename(r'%LOG_FILE%'),
'total_lines': len(lines)
}
with open(r'%RUN_FILE%', 'w') as f:
json.dump(result, f, indent=2)
"
echo Results saved. Will sync to Mario via Syncthing.
echo ============================================================================
pause

188
run_tests.bat Normal file
View File

@@ -0,0 +1,188 @@
@echo off
REM ============================================================================
REM Atomizer Test Runner — Tier 2 Dev Workflow
REM ============================================================================
REM Double-click this to run tests. Results sync back to Mario via Syncthing.
REM
REM Usage:
REM run_tests.bat — run all tests
REM run_tests.bat test_spec_api — run specific test file
REM run_tests.bat unit — run unit tests folder
REM run_tests.bat --quick — fast smoke test (no slow/NX tests)
REM run_tests.bat --nx — NX-dependent tests only
REM ============================================================================
setlocal enabledelayedexpansion
REM === CONFIG ===
set "ATOMIZER_ROOT=%~dp0"
set "RESULTS_DIR=%ATOMIZER_ROOT%test_results"
set "PYTHON=python"
REM Timestamp for this run
for /f "tokens=1-6 delims=/:. " %%a in ("%date% %time%") do (
set "TIMESTAMP=%%a-%%b-%%c_%%d-%%e-%%f"
)
REM Fallback: use a simpler approach
for /f %%i in ('python -c "from datetime import datetime; print(datetime.now().strftime('%%Y-%%m-%%d_%%H-%%M-%%S'))"') do set "TIMESTAMP=%%i"
set "RUN_FILE=%RESULTS_DIR%\run_%TIMESTAMP%.json"
set "LOG_FILE=%RESULTS_DIR%\run_%TIMESTAMP%.log"
REM Create results dir if needed
if not exist "%RESULTS_DIR%" mkdir "%RESULTS_DIR%"
echo.
echo ============================================================================
echo ATOMIZER TEST RUNNER
echo %date% %time%
echo ============================================================================
echo.
REM === Gather system info ===
echo Gathering environment info...
for /f "delims=" %%v in ('python --version 2^>^&1') do set "PYTHON_VER=%%v"
for /f "delims=" %%v in ('python -c "import sys; print(sys.executable)"') do set "PYTHON_EXE=%%v"
REM Check if NX is available
set "NX_AVAILABLE=false"
python -c "import NXOpen" 2>nul && set "NX_AVAILABLE=true"
REM === Determine what to run ===
set "TEST_TARGET=tests/"
set "PYTEST_ARGS=-v --tb=short"
set "TEST_MODE=all"
if "%~1"=="--quick" (
set "PYTEST_ARGS=-v --tb=short -m \"not slow and not nx\""
set "TEST_MODE=quick"
) else if "%~1"=="--nx" (
set "PYTEST_ARGS=-v --tb=short -m nx"
set "TEST_MODE=nx-only"
) else if not "%~1"=="" (
set "TEST_TARGET=tests/%~1"
if not exist "%ATOMIZER_ROOT%tests\%~1" (
set "TEST_TARGET=tests/%~1.py"
)
set "TEST_MODE=targeted"
)
echo Mode: %TEST_MODE%
echo Target: %TEST_TARGET%
echo Python: %PYTHON_VER%
echo NX: %NX_AVAILABLE%
echo Results: %RUN_FILE%
echo.
REM === Run tests ===
echo Running tests...
echo ============================================================================
cd /d "%ATOMIZER_ROOT%"
REM Run pytest with JSON report if available, otherwise parse output
python -m pytest %TEST_TARGET% %PYTEST_ARGS% --json-report --json-report-file="%RESULTS_DIR%\_pytest_report.json" 2>nul
if errorlevel 1 (
REM json-report plugin might not be installed, run without it
python -m pytest %TEST_TARGET% %PYTEST_ARGS% > "%LOG_FILE%" 2>&1
set "EXIT_CODE=!errorlevel!"
) else (
python -m pytest %TEST_TARGET% %PYTEST_ARGS% > "%LOG_FILE%" 2>&1
set "EXIT_CODE=!errorlevel!"
)
echo.
echo ============================================================================
REM === Generate results JSON ===
python -c "
import json, sys, os, platform
from datetime import datetime
log_path = r'%LOG_FILE%'
report_path = r'%RESULTS_DIR%\_pytest_report.json'
run_file = r'%RUN_FILE%'
# Read log
with open(log_path, 'r', encoding='utf-8', errors='replace') as f:
log_content = f.read()
# Parse basic stats from log
lines = log_content.split('\n')
summary_line = ''
for line in reversed(lines):
if 'passed' in line or 'failed' in line or 'error' in line:
summary_line = line.strip()
break
# Try to get JSON report
json_report = None
if os.path.exists(report_path):
try:
with open(report_path) as f:
json_report = json.load(f)
except: pass
# Extract failures
failures = []
in_failure = False
current_failure = []
for line in lines:
if line.startswith('FAILED ') or line.startswith('ERROR '):
failures.append(line.strip())
elif '_ FAILURES _' in line or '_ ERRORS _' in line:
in_failure = True
elif in_failure and line.startswith('='):
if current_failure:
failures.append('\n'.join(current_failure))
current_failure = []
in_failure = False
elif in_failure:
current_failure.append(line)
result = {
'timestamp': datetime.now().isoformat(),
'exit_code': int('%EXIT_CODE%'),
'mode': '%TEST_MODE%',
'target': '%TEST_TARGET%',
'python': '%PYTHON_VER%',
'python_exe': r'%PYTHON_EXE%',
'nx_available': %NX_AVAILABLE%,
'platform': platform.platform(),
'summary': summary_line,
'failures': failures[:20], # cap at 20
'log_file': os.path.basename(log_path),
'status': 'PASS' if int('%EXIT_CODE%') == 0 else 'FAIL'
}
# Add JSON report summary if available
if json_report and 'summary' in json_report:
result['pytest_summary'] = json_report['summary']
with open(run_file, 'w') as f:
json.dump(result, f, indent=2)
print()
print(f'Status: {result[\"status\"]}')
print(f'Summary: {summary_line}')
print(f'Results saved to: {os.path.basename(run_file)}')
"
echo.
REM === Also write a latest.json pointer ===
echo {"latest": "run_%TIMESTAMP%.json", "timestamp": "%TIMESTAMP%"} > "%RESULTS_DIR%\latest.json"
REM === Cleanup old pytest report ===
if exist "%RESULTS_DIR%\_pytest_report.json" del "%RESULTS_DIR%\_pytest_report.json"
echo.
if %EXIT_CODE% EQU 0 (
echo ALL TESTS PASSED
) else (
echo SOME TESTS FAILED — check results
)
echo.
echo Results will sync to Mario via Syncthing.
echo ============================================================================
pause

View File

@@ -9,6 +9,20 @@ from pathlib import Path
# Add project root to path
sys.path.insert(0, str(Path(__file__).parent.parent))
def pytest_configure(config):
"""Register custom markers."""
config.addinivalue_line("markers", "nx: requires NX Open (skip with -m 'not nx')")
config.addinivalue_line("markers", "slow: slow tests (skip with -m 'not slow')")
def pytest_collection_modifyitems(config, items):
"""Auto-mark tests that import NXOpen."""
for item in items:
# Auto-mark files with 'nx' or 'journal' in name
if 'nx' in item.nodeid.lower() or 'journal' in item.nodeid.lower():
item.add_marker(pytest.mark.nx)
@pytest.fixture
def sample_study_dir(tmp_path):
"""Create a temporary study directory structure."""