mirror of
https://github.com/Xe138/AI-Trader.git
synced 2026-04-02 09:37:23 -04:00
Compare commits
9 Commits
v0.4.1
...
v0.4.2-alp
| Author | SHA1 | Date | |
|---|---|---|---|
| 406bb281b2 | |||
| 6ddc5abede | |||
| 5c73f30583 | |||
| b73d88ca8f | |||
| d199b093c1 | |||
| 483621f9b7 | |||
| e8939be04e | |||
| 2e0cf4d507 | |||
| 7b35394ce7 |
18
CHANGELOG.md
18
CHANGELOG.md
@@ -7,6 +7,24 @@ and this project adheres to [Semantic Versioning](https://semver.org/spec/v2.0.0
|
||||
|
||||
## [Unreleased]
|
||||
|
||||
### Fixed
|
||||
- **Critical:** Fixed stale jobs blocking new jobs after Docker container restart
|
||||
- Root cause: Jobs with status 'pending', 'downloading_data', or 'running' remained in database after container shutdown, preventing new job creation
|
||||
- Solution: Added `cleanup_stale_jobs()` method that runs on FastAPI startup to mark interrupted jobs as 'failed' or 'partial' based on completion percentage
|
||||
- Intelligent status determination: Uses existing progress tracking (completed/total model-days) to distinguish between failed (0% complete) and partial (>0% complete)
|
||||
- Detailed error messages include original status and completion counts (e.g., "Job interrupted by container restart (was running, 3/10 model-days completed)")
|
||||
- Incomplete job_details automatically marked as 'failed' with clear error messages
|
||||
- Deployment-aware: Skips cleanup in DEV mode when database is reset, always runs in PROD mode
|
||||
- Comprehensive test coverage: 6 new unit tests covering all cleanup scenarios
|
||||
- Locations: `api/job_manager.py:702-779`, `api/main.py:164-168`, `tests/unit/test_job_manager.py:451-609`
|
||||
- Fixed Pydantic validation errors when using DeepSeek models via OpenRouter
|
||||
- Root cause: LangChain's `parse_tool_call()` has a bug where it sometimes returns `args` as JSON string instead of parsed dict object
|
||||
- Solution: Added `ToolCallArgsParsingWrapper` that:
|
||||
1. Patches `parse_tool_call()` to detect and fix string args by parsing them to dict
|
||||
2. Normalizes non-standard tool_call formats (e.g., `{name, args, id}` → `{function: {name, arguments}, id}`)
|
||||
- The wrapper is defensive and only acts when needed, ensuring compatibility with all AI providers
|
||||
- Fixes validation error: `tool_calls.0.args: Input should be a valid dictionary [type=dict_type, input_value='...', input_type=str]`
|
||||
|
||||
## [0.4.1] - 2025-11-06
|
||||
|
||||
### Fixed
|
||||
|
||||
80
ROADMAP.md
80
ROADMAP.md
@@ -4,6 +4,78 @@ This document outlines planned features and improvements for the AI-Trader proje
|
||||
|
||||
## Release Planning
|
||||
|
||||
### v0.5.0 - Performance Metrics & Status APIs (Planned)
|
||||
|
||||
**Focus:** Enhanced observability and performance tracking
|
||||
|
||||
#### Performance Metrics API
|
||||
- **Performance Summary Endpoint** - Query model performance over date ranges
|
||||
- `GET /metrics/performance` - Aggregated performance metrics
|
||||
- Query parameters: `model`, `start_date`, `end_date`
|
||||
- Returns comprehensive performance summary:
|
||||
- Total return (dollar amount and percentage)
|
||||
- Number of trades executed (buy + sell)
|
||||
- Win rate (profitable trading days / total trading days)
|
||||
- Average daily P&L (profit and loss)
|
||||
- Best/worst trading day (highest/lowest daily P&L)
|
||||
- Final portfolio value (cash + holdings at market value)
|
||||
- Number of trading days in queried range
|
||||
- Starting vs. ending portfolio comparison
|
||||
- Use cases:
|
||||
- Compare model performance across different time periods
|
||||
- Evaluate strategy effectiveness
|
||||
- Identify top-performing models
|
||||
- Example: `GET /metrics/performance?model=gpt-4&start_date=2025-01-01&end_date=2025-01-31`
|
||||
- Filtering options:
|
||||
- Single model or all models
|
||||
- Custom date ranges
|
||||
- Exclude incomplete trading days
|
||||
- Response format: JSON with clear metric definitions
|
||||
|
||||
#### Status & Coverage Endpoint
|
||||
- **System Status Summary** - Data availability and simulation progress
|
||||
- `GET /status` - Comprehensive system status
|
||||
- Price data coverage section:
|
||||
- Available symbols (NASDAQ 100 constituents)
|
||||
- Date range of downloaded price data per symbol
|
||||
- Total trading days with complete data
|
||||
- Missing data gaps (symbols without data, date gaps)
|
||||
- Last data refresh timestamp
|
||||
- Model simulation status section:
|
||||
- List of all configured models (enabled/disabled)
|
||||
- Date ranges simulated per model (first and last trading day)
|
||||
- Total trading days completed per model
|
||||
- Most recent simulation date per model
|
||||
- Completion percentage (simulated days / available data days)
|
||||
- System health section:
|
||||
- Database connectivity status
|
||||
- MCP services status (Math, Search, Trade, LocalPrices)
|
||||
- API version and deployment mode
|
||||
- Disk space usage (database size, log size)
|
||||
- Use cases:
|
||||
- Verify data availability before triggering simulations
|
||||
- Identify which models need updates to latest data
|
||||
- Monitor system health and readiness
|
||||
- Plan data downloads for missing date ranges
|
||||
- Example: `GET /status` (no parameters required)
|
||||
- Benefits:
|
||||
- Single endpoint for complete system overview
|
||||
- No need to query multiple endpoints for status
|
||||
- Clear visibility into data gaps
|
||||
- Track simulation progress across models
|
||||
|
||||
#### Implementation Details
|
||||
- Database queries for efficient metric calculation
|
||||
- Caching for frequently accessed metrics (optional)
|
||||
- Response time target: <500ms for typical queries
|
||||
- Comprehensive error handling for missing data
|
||||
|
||||
#### Benefits
|
||||
- **Better Observability** - Clear view of system state and model performance
|
||||
- **Data-Driven Decisions** - Quantitative metrics for model comparison
|
||||
- **Proactive Monitoring** - Identify data gaps before simulations fail
|
||||
- **User Experience** - Single endpoint to check "what's available and what's been done"
|
||||
|
||||
### v1.0.0 - Production Stability & Validation (Planned)
|
||||
|
||||
**Focus:** Comprehensive testing, documentation, and production readiness
|
||||
@@ -607,11 +679,13 @@ To propose a new feature:
|
||||
|
||||
- **v0.1.0** - Initial release with batch execution
|
||||
- **v0.2.0** - Docker deployment support
|
||||
- **v0.3.0** - REST API, on-demand downloads, database storage (current)
|
||||
- **v0.3.0** - REST API, on-demand downloads, database storage
|
||||
- **v0.4.0** - Daily P&L calculation, day-centric results API, reasoning summaries (current)
|
||||
- **v0.5.0** - Performance metrics & status APIs (planned)
|
||||
- **v1.0.0** - Production stability & validation (planned)
|
||||
- **v1.1.0** - API authentication & security (planned)
|
||||
- **v1.2.0** - Position history & analytics (planned)
|
||||
- **v1.3.0** - Performance metrics & analytics (planned)
|
||||
- **v1.3.0** - Advanced performance metrics & analytics (planned)
|
||||
- **v1.4.0** - Data management API (planned)
|
||||
- **v1.5.0** - Web dashboard UI (planned)
|
||||
- **v1.6.0** - Advanced configuration & customization (planned)
|
||||
@@ -619,4 +693,4 @@ To propose a new feature:
|
||||
|
||||
---
|
||||
|
||||
Last updated: 2025-11-01
|
||||
Last updated: 2025-11-06
|
||||
|
||||
@@ -33,6 +33,7 @@ from tools.deployment_config import (
|
||||
from agent.context_injector import ContextInjector
|
||||
from agent.pnl_calculator import DailyPnLCalculator
|
||||
from agent.reasoning_summarizer import ReasoningSummarizer
|
||||
from agent.chat_model_wrapper import ToolCallArgsParsingWrapper
|
||||
|
||||
# Load environment variables
|
||||
load_dotenv()
|
||||
@@ -211,14 +212,16 @@ class BaseAgent:
|
||||
self.model = MockChatModel(date="2025-01-01") # Date will be updated per session
|
||||
print(f"🤖 Using MockChatModel (DEV mode)")
|
||||
else:
|
||||
self.model = ChatOpenAI(
|
||||
base_model = ChatOpenAI(
|
||||
model=self.basemodel,
|
||||
base_url=self.openai_base_url,
|
||||
api_key=self.openai_api_key,
|
||||
max_retries=3,
|
||||
timeout=30
|
||||
)
|
||||
print(f"🤖 Using {self.basemodel} (PROD mode)")
|
||||
# Wrap model with diagnostic wrapper
|
||||
self.model = ToolCallArgsParsingWrapper(model=base_model)
|
||||
print(f"🤖 Using {self.basemodel} (PROD mode) with diagnostic wrapper")
|
||||
except Exception as e:
|
||||
raise RuntimeError(f"❌ Failed to initialize AI model: {e}")
|
||||
|
||||
|
||||
@@ -1,24 +1,18 @@
|
||||
"""
|
||||
Chat model wrapper - Passthrough wrapper for ChatOpenAI models.
|
||||
Chat model wrapper to fix tool_calls args parsing issues.
|
||||
|
||||
Originally created to fix DeepSeek tool_calls arg parsing issues, but investigation
|
||||
revealed DeepSeek already returns the correct format (arguments as JSON strings).
|
||||
|
||||
This wrapper is now a simple passthrough that proxies all calls to the underlying model.
|
||||
Kept for backward compatibility and potential future use.
|
||||
DeepSeek and other providers return tool_calls.args as JSON strings, which need
|
||||
to be parsed to dicts before AIMessage construction.
|
||||
"""
|
||||
|
||||
from typing import Any
|
||||
import json
|
||||
from typing import Any, Optional, Dict
|
||||
from functools import wraps
|
||||
|
||||
|
||||
class ToolCallArgsParsingWrapper:
|
||||
"""
|
||||
Passthrough wrapper around ChatOpenAI models.
|
||||
|
||||
After systematic debugging, determined that DeepSeek returns tool_calls.arguments
|
||||
as JSON strings (correct format), so no parsing/conversion is needed.
|
||||
|
||||
This wrapper simply proxies all calls to the wrapped model.
|
||||
Wrapper that adds diagnostic logging and fixes tool_calls args if needed.
|
||||
"""
|
||||
|
||||
def __init__(self, model: Any, **kwargs):
|
||||
@@ -30,6 +24,82 @@ class ToolCallArgsParsingWrapper:
|
||||
**kwargs: Additional parameters (ignored, for compatibility)
|
||||
"""
|
||||
self.wrapped_model = model
|
||||
self._patch_model()
|
||||
|
||||
def _patch_model(self):
|
||||
"""Monkey-patch the model's _create_chat_result to add diagnostics"""
|
||||
if not hasattr(self.wrapped_model, '_create_chat_result'):
|
||||
# Model doesn't have this method (e.g., MockChatModel), skip patching
|
||||
return
|
||||
|
||||
# CRITICAL: Patch parse_tool_call in base.py's namespace (not in openai_tools module!)
|
||||
from langchain_openai.chat_models import base as langchain_base
|
||||
original_parse_tool_call = langchain_base.parse_tool_call
|
||||
|
||||
def patched_parse_tool_call(raw_tool_call, *, partial=False, strict=False, return_id=True):
|
||||
"""Patched parse_tool_call to fix string args bug"""
|
||||
result = original_parse_tool_call(raw_tool_call, partial=partial, strict=strict, return_id=return_id)
|
||||
if result and isinstance(result.get('args'), str):
|
||||
# FIX: parse_tool_call sometimes returns string args instead of dict
|
||||
# This is a known LangChain bug - parse the string to dict
|
||||
try:
|
||||
result['args'] = json.loads(result['args'])
|
||||
except (json.JSONDecodeError, TypeError):
|
||||
# Leave as string if we can't parse it - will fail validation
|
||||
# but at least we tried
|
||||
pass
|
||||
return result
|
||||
|
||||
# Replace in base.py's namespace (where _convert_dict_to_message uses it)
|
||||
langchain_base.parse_tool_call = patched_parse_tool_call
|
||||
|
||||
original_create_chat_result = self.wrapped_model._create_chat_result
|
||||
|
||||
@wraps(original_create_chat_result)
|
||||
def patched_create_chat_result(response: Any, generation_info: Optional[Dict] = None):
|
||||
"""Patched version that normalizes non-standard tool_call formats"""
|
||||
response_dict = response if isinstance(response, dict) else response.model_dump()
|
||||
|
||||
# Normalize tool_calls to OpenAI standard format if needed
|
||||
if 'choices' in response_dict:
|
||||
for choice in response_dict['choices']:
|
||||
if 'message' not in choice:
|
||||
continue
|
||||
|
||||
message = choice['message']
|
||||
|
||||
# Fix tool_calls: Convert non-standard {name, args, id} to {function: {name, arguments}, id}
|
||||
if 'tool_calls' in message and message['tool_calls']:
|
||||
for tool_call in message['tool_calls']:
|
||||
# Check if this is non-standard format (has 'args' directly)
|
||||
if 'args' in tool_call and 'function' not in tool_call:
|
||||
# Convert to standard OpenAI format
|
||||
args = tool_call['args']
|
||||
tool_call['function'] = {
|
||||
'name': tool_call.get('name', ''),
|
||||
'arguments': args if isinstance(args, str) else json.dumps(args)
|
||||
}
|
||||
# Remove non-standard fields
|
||||
if 'name' in tool_call:
|
||||
del tool_call['name']
|
||||
if 'args' in tool_call:
|
||||
del tool_call['args']
|
||||
|
||||
# Fix invalid_tool_calls: Ensure args is JSON string (not dict)
|
||||
if 'invalid_tool_calls' in message and message['invalid_tool_calls']:
|
||||
for invalid_call in message['invalid_tool_calls']:
|
||||
if 'args' in invalid_call and isinstance(invalid_call['args'], dict):
|
||||
try:
|
||||
invalid_call['args'] = json.dumps(invalid_call['args'])
|
||||
except (TypeError, ValueError):
|
||||
# Keep as-is if serialization fails
|
||||
pass
|
||||
|
||||
# Call original method with normalized response
|
||||
return original_create_chat_result(response_dict, generation_info)
|
||||
|
||||
# Replace the method
|
||||
self.wrapped_model._create_chat_result = patched_create_chat_result
|
||||
|
||||
@property
|
||||
def _llm_type(self) -> str:
|
||||
|
||||
@@ -699,6 +699,85 @@ class JobManager:
|
||||
finally:
|
||||
conn.close()
|
||||
|
||||
def cleanup_stale_jobs(self) -> Dict[str, int]:
|
||||
"""
|
||||
Clean up stale jobs from container restarts.
|
||||
|
||||
Marks jobs with status 'pending', 'downloading_data', or 'running' as
|
||||
'failed' or 'partial' based on completion percentage.
|
||||
|
||||
Called on application startup to reset interrupted jobs.
|
||||
|
||||
Returns:
|
||||
Dict with jobs_cleaned count and details
|
||||
"""
|
||||
conn = get_db_connection(self.db_path)
|
||||
cursor = conn.cursor()
|
||||
|
||||
try:
|
||||
# Find all stale jobs
|
||||
cursor.execute("""
|
||||
SELECT job_id, status
|
||||
FROM jobs
|
||||
WHERE status IN ('pending', 'downloading_data', 'running')
|
||||
""")
|
||||
|
||||
stale_jobs = cursor.fetchall()
|
||||
cleaned_count = 0
|
||||
|
||||
for job_id, original_status in stale_jobs:
|
||||
# Get progress to determine if partially completed
|
||||
cursor.execute("""
|
||||
SELECT
|
||||
COUNT(*) as total,
|
||||
SUM(CASE WHEN status = 'completed' THEN 1 ELSE 0 END) as completed,
|
||||
SUM(CASE WHEN status = 'failed' THEN 1 ELSE 0 END) as failed
|
||||
FROM job_details
|
||||
WHERE job_id = ?
|
||||
""", (job_id,))
|
||||
|
||||
total, completed, failed = cursor.fetchone()
|
||||
completed = completed or 0
|
||||
failed = failed or 0
|
||||
|
||||
# Determine final status based on completion
|
||||
if completed > 0:
|
||||
new_status = "partial"
|
||||
error_msg = f"Job interrupted by container restart (was {original_status}, {completed}/{total} model-days completed)"
|
||||
else:
|
||||
new_status = "failed"
|
||||
error_msg = f"Job interrupted by container restart (was {original_status}, no progress made)"
|
||||
|
||||
# Mark incomplete job_details as failed
|
||||
cursor.execute("""
|
||||
UPDATE job_details
|
||||
SET status = 'failed', error = 'Container restarted before completion'
|
||||
WHERE job_id = ? AND status IN ('pending', 'running')
|
||||
""", (job_id,))
|
||||
|
||||
# Update job status
|
||||
updated_at = datetime.utcnow().isoformat() + "Z"
|
||||
cursor.execute("""
|
||||
UPDATE jobs
|
||||
SET status = ?, error = ?, completed_at = ?, updated_at = ?
|
||||
WHERE job_id = ?
|
||||
""", (new_status, error_msg, updated_at, updated_at, job_id))
|
||||
|
||||
logger.warning(f"Cleaned up stale job {job_id}: {original_status} → {new_status} ({completed}/{total} completed)")
|
||||
cleaned_count += 1
|
||||
|
||||
conn.commit()
|
||||
|
||||
if cleaned_count > 0:
|
||||
logger.warning(f"⚠️ Cleaned up {cleaned_count} stale job(s) from previous container session")
|
||||
else:
|
||||
logger.info("✅ No stale jobs found")
|
||||
|
||||
return {"jobs_cleaned": cleaned_count}
|
||||
|
||||
finally:
|
||||
conn.close()
|
||||
|
||||
def cleanup_old_jobs(self, days: int = 30) -> Dict[str, int]:
|
||||
"""
|
||||
Delete jobs older than threshold.
|
||||
|
||||
16
api/main.py
16
api/main.py
@@ -134,25 +134,39 @@ def create_app(
|
||||
@asynccontextmanager
|
||||
async def lifespan(app: FastAPI):
|
||||
"""Initialize database on startup, cleanup on shutdown if needed"""
|
||||
from tools.deployment_config import is_dev_mode, get_db_path
|
||||
from tools.deployment_config import is_dev_mode, get_db_path, should_preserve_dev_data
|
||||
from api.database import initialize_dev_database, initialize_database
|
||||
|
||||
# Startup - use closure to access db_path from create_app scope
|
||||
logger.info("🚀 FastAPI application starting...")
|
||||
logger.info("📊 Initializing database...")
|
||||
|
||||
should_cleanup_stale_jobs = False
|
||||
|
||||
if is_dev_mode():
|
||||
# Initialize dev database (reset unless PRESERVE_DEV_DATA=true)
|
||||
logger.info(" 🔧 DEV mode detected - initializing dev database")
|
||||
dev_db_path = get_db_path(db_path)
|
||||
initialize_dev_database(dev_db_path)
|
||||
log_dev_mode_startup_warning()
|
||||
|
||||
# Only cleanup stale jobs if preserving dev data (otherwise DB is fresh)
|
||||
if should_preserve_dev_data():
|
||||
should_cleanup_stale_jobs = True
|
||||
else:
|
||||
# Ensure production database schema exists
|
||||
logger.info(" 🏭 PROD mode - ensuring database schema exists")
|
||||
initialize_database(db_path)
|
||||
should_cleanup_stale_jobs = True
|
||||
|
||||
logger.info("✅ Database initialized")
|
||||
|
||||
# Clean up stale jobs from previous container session
|
||||
if should_cleanup_stale_jobs:
|
||||
logger.info("🧹 Checking for stale jobs from previous session...")
|
||||
job_manager = JobManager(get_db_path(db_path) if is_dev_mode() else db_path)
|
||||
job_manager.cleanup_stale_jobs()
|
||||
|
||||
logger.info("🌐 API server ready to accept requests")
|
||||
|
||||
yield
|
||||
|
||||
@@ -448,4 +448,164 @@ class TestJobWarnings:
|
||||
assert stored_warnings == warnings
|
||||
|
||||
|
||||
@pytest.mark.unit
|
||||
class TestStaleJobCleanup:
|
||||
"""Test cleanup of stale jobs from container restarts."""
|
||||
|
||||
def test_cleanup_stale_pending_job(self, clean_db):
|
||||
"""Should mark pending job as failed with no progress."""
|
||||
from api.job_manager import JobManager
|
||||
|
||||
manager = JobManager(db_path=clean_db)
|
||||
job_id = manager.create_job(
|
||||
config_path="configs/test.json",
|
||||
date_range=["2025-01-16", "2025-01-17"],
|
||||
models=["gpt-5"]
|
||||
)
|
||||
|
||||
# Job is pending - simulate container restart
|
||||
result = manager.cleanup_stale_jobs()
|
||||
|
||||
assert result["jobs_cleaned"] == 1
|
||||
job = manager.get_job(job_id)
|
||||
assert job["status"] == "failed"
|
||||
assert "container restart" in job["error"].lower()
|
||||
assert "pending" in job["error"]
|
||||
assert "no progress" in job["error"]
|
||||
|
||||
def test_cleanup_stale_running_job_with_partial_progress(self, clean_db):
|
||||
"""Should mark running job as partial if some model-days completed."""
|
||||
from api.job_manager import JobManager
|
||||
|
||||
manager = JobManager(db_path=clean_db)
|
||||
job_id = manager.create_job(
|
||||
config_path="configs/test.json",
|
||||
date_range=["2025-01-16", "2025-01-17"],
|
||||
models=["gpt-5"]
|
||||
)
|
||||
|
||||
# Mark job as running and complete one model-day
|
||||
manager.update_job_status(job_id, "running")
|
||||
manager.update_job_detail_status(job_id, "2025-01-16", "gpt-5", "completed")
|
||||
|
||||
# Simulate container restart
|
||||
result = manager.cleanup_stale_jobs()
|
||||
|
||||
assert result["jobs_cleaned"] == 1
|
||||
job = manager.get_job(job_id)
|
||||
assert job["status"] == "partial"
|
||||
assert "container restart" in job["error"].lower()
|
||||
assert "1/2" in job["error"] # 1 out of 2 model-days completed
|
||||
|
||||
def test_cleanup_stale_downloading_data_job(self, clean_db):
|
||||
"""Should mark downloading_data job as failed."""
|
||||
from api.job_manager import JobManager
|
||||
|
||||
manager = JobManager(db_path=clean_db)
|
||||
job_id = manager.create_job(
|
||||
config_path="configs/test.json",
|
||||
date_range=["2025-01-16"],
|
||||
models=["gpt-5"]
|
||||
)
|
||||
|
||||
# Mark as downloading data
|
||||
manager.update_job_status(job_id, "downloading_data")
|
||||
|
||||
# Simulate container restart
|
||||
result = manager.cleanup_stale_jobs()
|
||||
|
||||
assert result["jobs_cleaned"] == 1
|
||||
job = manager.get_job(job_id)
|
||||
assert job["status"] == "failed"
|
||||
assert "downloading_data" in job["error"]
|
||||
|
||||
def test_cleanup_marks_incomplete_job_details_as_failed(self, clean_db):
|
||||
"""Should mark incomplete job_details as failed."""
|
||||
from api.job_manager import JobManager
|
||||
|
||||
manager = JobManager(db_path=clean_db)
|
||||
job_id = manager.create_job(
|
||||
config_path="configs/test.json",
|
||||
date_range=["2025-01-16", "2025-01-17"],
|
||||
models=["gpt-5"]
|
||||
)
|
||||
|
||||
# Mark job as running, one detail running, one pending
|
||||
manager.update_job_status(job_id, "running")
|
||||
manager.update_job_detail_status(job_id, "2025-01-16", "gpt-5", "running")
|
||||
|
||||
# Simulate container restart
|
||||
manager.cleanup_stale_jobs()
|
||||
|
||||
# Check job_details were marked as failed
|
||||
progress = manager.get_job_progress(job_id)
|
||||
assert progress["failed"] == 2 # Both model-days marked failed
|
||||
assert progress["pending"] == 0
|
||||
|
||||
details = manager.get_job_details(job_id)
|
||||
for detail in details:
|
||||
assert detail["status"] == "failed"
|
||||
assert "container restarted" in detail["error"].lower()
|
||||
|
||||
def test_cleanup_no_stale_jobs(self, clean_db):
|
||||
"""Should report 0 cleaned jobs when none are stale."""
|
||||
from api.job_manager import JobManager
|
||||
|
||||
manager = JobManager(db_path=clean_db)
|
||||
job_id = manager.create_job(
|
||||
config_path="configs/test.json",
|
||||
date_range=["2025-01-16"],
|
||||
models=["gpt-5"]
|
||||
)
|
||||
|
||||
# Complete the job
|
||||
manager.update_job_detail_status(job_id, "2025-01-16", "gpt-5", "completed")
|
||||
|
||||
# Simulate container restart
|
||||
result = manager.cleanup_stale_jobs()
|
||||
|
||||
assert result["jobs_cleaned"] == 0
|
||||
job = manager.get_job(job_id)
|
||||
assert job["status"] == "completed"
|
||||
|
||||
def test_cleanup_multiple_stale_jobs(self, clean_db):
|
||||
"""Should clean up multiple stale jobs."""
|
||||
from api.job_manager import JobManager
|
||||
|
||||
manager = JobManager(db_path=clean_db)
|
||||
|
||||
# Create first job
|
||||
job1_id = manager.create_job(
|
||||
config_path="configs/test.json",
|
||||
date_range=["2025-01-16"],
|
||||
models=["gpt-5"]
|
||||
)
|
||||
manager.update_job_status(job1_id, "running")
|
||||
manager.update_job_status(job1_id, "completed")
|
||||
|
||||
# Create second job (pending)
|
||||
job2_id = manager.create_job(
|
||||
config_path="configs/test.json",
|
||||
date_range=["2025-01-17"],
|
||||
models=["gpt-5"]
|
||||
)
|
||||
|
||||
# Create third job (running)
|
||||
manager.update_job_status(job2_id, "completed")
|
||||
job3_id = manager.create_job(
|
||||
config_path="configs/test.json",
|
||||
date_range=["2025-01-18"],
|
||||
models=["gpt-5"]
|
||||
)
|
||||
manager.update_job_status(job3_id, "running")
|
||||
|
||||
# Simulate container restart
|
||||
result = manager.cleanup_stale_jobs()
|
||||
|
||||
assert result["jobs_cleaned"] == 1 # Only job3 is running
|
||||
assert manager.get_job(job1_id)["status"] == "completed"
|
||||
assert manager.get_job(job2_id)["status"] == "completed"
|
||||
assert manager.get_job(job3_id)["status"] == "failed"
|
||||
|
||||
|
||||
# Coverage target: 95%+ for api/job_manager.py
|
||||
|
||||
Reference in New Issue
Block a user