mirror of
https://github.com/Xe138/AI-Trader.git
synced 2026-04-08 11:47:24 -04:00
test: improve test coverage from 61% to 84.81%
Major improvements: - Fixed all 42 broken tests (database connection leaks) - Added db_connection() context manager for proper cleanup - Created comprehensive test suites for undertested modules New test coverage: - tools/general_tools.py: 26 tests (97% coverage) - tools/price_tools.py: 11 tests (validates NASDAQ symbols, date handling) - api/price_data_manager.py: 12 tests (85% coverage) - api/routes/results_v2.py: 3 tests (98% coverage) - agent/reasoning_summarizer.py: 2 tests (87% coverage) - api/routes/period_metrics.py: 2 edge case tests (100% coverage) - agent/mock_provider: 1 test (100% coverage) Database fixes: - Added db_connection() context manager to prevent leaks - Updated 16+ test files to use context managers - Fixed drop_all_tables() to match new schema - Added CHECK constraint for action_type - Added ON DELETE CASCADE to trading_days foreign key Test improvements: - Updated SQL INSERT statements with all required fields - Fixed date parameter handling in API integration tests - Added edge case tests for validation functions - Fixed import errors across test suite Results: - Total coverage: 84.81% (was 61%) - Tests passing: 406 (was 364 with 42 failures) - Total lines covered: 6364 of 7504 🤖 Generated with [Claude Code](https://claude.com/claude-code) Co-Authored-By: Claude <noreply@anthropic.com>
This commit is contained in:
@@ -18,6 +18,7 @@ import tempfile
|
||||
from pathlib import Path
|
||||
from api.database import (
|
||||
get_db_connection,
|
||||
db_connection,
|
||||
initialize_database,
|
||||
drop_all_tables,
|
||||
vacuum_database,
|
||||
@@ -34,11 +35,10 @@ class TestDatabaseConnection:
|
||||
temp_dir = tempfile.mkdtemp()
|
||||
db_path = os.path.join(temp_dir, "subdir", "test.db")
|
||||
|
||||
conn = get_db_connection(db_path)
|
||||
assert conn is not None
|
||||
assert os.path.exists(os.path.dirname(db_path))
|
||||
with db_connection(db_path) as conn:
|
||||
assert conn is not None
|
||||
assert os.path.exists(os.path.dirname(db_path))
|
||||
|
||||
conn.close()
|
||||
os.unlink(db_path)
|
||||
os.rmdir(os.path.dirname(db_path))
|
||||
os.rmdir(temp_dir)
|
||||
@@ -48,16 +48,15 @@ class TestDatabaseConnection:
|
||||
temp_db = tempfile.NamedTemporaryFile(delete=False, suffix=".db")
|
||||
temp_db.close()
|
||||
|
||||
conn = get_db_connection(temp_db.name)
|
||||
with db_connection(temp_db.name) as conn:
|
||||
|
||||
# Check if foreign keys are enabled
|
||||
cursor = conn.cursor()
|
||||
cursor.execute("PRAGMA foreign_keys")
|
||||
result = cursor.fetchone()[0]
|
||||
# Check if foreign keys are enabled
|
||||
cursor = conn.cursor()
|
||||
cursor.execute("PRAGMA foreign_keys")
|
||||
result = cursor.fetchone()[0]
|
||||
|
||||
assert result == 1 # 1 = enabled
|
||||
assert result == 1 # 1 = enabled
|
||||
|
||||
conn.close()
|
||||
os.unlink(temp_db.name)
|
||||
|
||||
def test_get_db_connection_row_factory(self):
|
||||
@@ -65,11 +64,10 @@ class TestDatabaseConnection:
|
||||
temp_db = tempfile.NamedTemporaryFile(delete=False, suffix=".db")
|
||||
temp_db.close()
|
||||
|
||||
conn = get_db_connection(temp_db.name)
|
||||
with db_connection(temp_db.name) as conn:
|
||||
|
||||
assert conn.row_factory == sqlite3.Row
|
||||
assert conn.row_factory == sqlite3.Row
|
||||
|
||||
conn.close()
|
||||
os.unlink(temp_db.name)
|
||||
|
||||
def test_get_db_connection_thread_safety(self):
|
||||
@@ -78,10 +76,9 @@ class TestDatabaseConnection:
|
||||
temp_db.close()
|
||||
|
||||
# This should not raise an error
|
||||
conn = get_db_connection(temp_db.name)
|
||||
assert conn is not None
|
||||
with db_connection(temp_db.name) as conn:
|
||||
assert conn is not None
|
||||
|
||||
conn.close()
|
||||
os.unlink(temp_db.name)
|
||||
|
||||
|
||||
@@ -91,112 +88,108 @@ class TestSchemaInitialization:
|
||||
|
||||
def test_initialize_database_creates_all_tables(self, clean_db):
|
||||
"""Should create all 10 tables."""
|
||||
conn = get_db_connection(clean_db)
|
||||
cursor = conn.cursor()
|
||||
with db_connection(clean_db) as conn:
|
||||
cursor = conn.cursor()
|
||||
|
||||
# Query sqlite_master for table names
|
||||
cursor.execute("""
|
||||
SELECT name FROM sqlite_master
|
||||
WHERE type='table' AND name NOT LIKE 'sqlite_%'
|
||||
ORDER BY name
|
||||
""")
|
||||
# Query sqlite_master for table names
|
||||
cursor.execute("""
|
||||
SELECT name FROM sqlite_master
|
||||
WHERE type='table' AND name NOT LIKE 'sqlite_%'
|
||||
ORDER BY name
|
||||
""")
|
||||
|
||||
tables = [row[0] for row in cursor.fetchall()]
|
||||
tables = [row[0] for row in cursor.fetchall()]
|
||||
|
||||
expected_tables = [
|
||||
'actions',
|
||||
'holdings',
|
||||
'job_details',
|
||||
'jobs',
|
||||
'tool_usage',
|
||||
'price_data',
|
||||
'price_data_coverage',
|
||||
'simulation_runs',
|
||||
'trading_days' # New day-centric schema
|
||||
]
|
||||
expected_tables = [
|
||||
'actions',
|
||||
'holdings',
|
||||
'job_details',
|
||||
'jobs',
|
||||
'tool_usage',
|
||||
'price_data',
|
||||
'price_data_coverage',
|
||||
'simulation_runs',
|
||||
'trading_days' # New day-centric schema
|
||||
]
|
||||
|
||||
assert sorted(tables) == sorted(expected_tables)
|
||||
assert sorted(tables) == sorted(expected_tables)
|
||||
|
||||
conn.close()
|
||||
|
||||
def test_initialize_database_creates_jobs_table(self, clean_db):
|
||||
"""Should create jobs table with correct schema."""
|
||||
conn = get_db_connection(clean_db)
|
||||
cursor = conn.cursor()
|
||||
with db_connection(clean_db) as conn:
|
||||
cursor = conn.cursor()
|
||||
|
||||
cursor.execute("PRAGMA table_info(jobs)")
|
||||
columns = {row[1]: row[2] for row in cursor.fetchall()}
|
||||
cursor.execute("PRAGMA table_info(jobs)")
|
||||
columns = {row[1]: row[2] for row in cursor.fetchall()}
|
||||
|
||||
expected_columns = {
|
||||
'job_id': 'TEXT',
|
||||
'config_path': 'TEXT',
|
||||
'status': 'TEXT',
|
||||
'date_range': 'TEXT',
|
||||
'models': 'TEXT',
|
||||
'created_at': 'TEXT',
|
||||
'started_at': 'TEXT',
|
||||
'updated_at': 'TEXT',
|
||||
'completed_at': 'TEXT',
|
||||
'total_duration_seconds': 'REAL',
|
||||
'error': 'TEXT',
|
||||
'warnings': 'TEXT'
|
||||
}
|
||||
expected_columns = {
|
||||
'job_id': 'TEXT',
|
||||
'config_path': 'TEXT',
|
||||
'status': 'TEXT',
|
||||
'date_range': 'TEXT',
|
||||
'models': 'TEXT',
|
||||
'created_at': 'TEXT',
|
||||
'started_at': 'TEXT',
|
||||
'updated_at': 'TEXT',
|
||||
'completed_at': 'TEXT',
|
||||
'total_duration_seconds': 'REAL',
|
||||
'error': 'TEXT',
|
||||
'warnings': 'TEXT'
|
||||
}
|
||||
|
||||
for col_name, col_type in expected_columns.items():
|
||||
assert col_name in columns
|
||||
assert columns[col_name] == col_type
|
||||
for col_name, col_type in expected_columns.items():
|
||||
assert col_name in columns
|
||||
assert columns[col_name] == col_type
|
||||
|
||||
conn.close()
|
||||
|
||||
def test_initialize_database_creates_trading_days_table(self, clean_db):
|
||||
"""Should create trading_days table with correct schema."""
|
||||
conn = get_db_connection(clean_db)
|
||||
cursor = conn.cursor()
|
||||
with db_connection(clean_db) as conn:
|
||||
cursor = conn.cursor()
|
||||
|
||||
cursor.execute("PRAGMA table_info(trading_days)")
|
||||
columns = {row[1]: row[2] for row in cursor.fetchall()}
|
||||
cursor.execute("PRAGMA table_info(trading_days)")
|
||||
columns = {row[1]: row[2] for row in cursor.fetchall()}
|
||||
|
||||
required_columns = [
|
||||
'id', 'job_id', 'date', 'model', 'starting_cash', 'ending_cash',
|
||||
'starting_portfolio_value', 'ending_portfolio_value',
|
||||
'daily_profit', 'daily_return_pct', 'days_since_last_trading',
|
||||
'total_actions', 'reasoning_summary', 'reasoning_full', 'created_at'
|
||||
]
|
||||
required_columns = [
|
||||
'id', 'job_id', 'date', 'model', 'starting_cash', 'ending_cash',
|
||||
'starting_portfolio_value', 'ending_portfolio_value',
|
||||
'daily_profit', 'daily_return_pct', 'days_since_last_trading',
|
||||
'total_actions', 'reasoning_summary', 'reasoning_full', 'created_at'
|
||||
]
|
||||
|
||||
for col_name in required_columns:
|
||||
assert col_name in columns
|
||||
for col_name in required_columns:
|
||||
assert col_name in columns
|
||||
|
||||
conn.close()
|
||||
|
||||
def test_initialize_database_creates_indexes(self, clean_db):
|
||||
"""Should create all performance indexes."""
|
||||
conn = get_db_connection(clean_db)
|
||||
cursor = conn.cursor()
|
||||
with db_connection(clean_db) as conn:
|
||||
cursor = conn.cursor()
|
||||
|
||||
cursor.execute("""
|
||||
SELECT name FROM sqlite_master
|
||||
WHERE type='index' AND name LIKE 'idx_%'
|
||||
ORDER BY name
|
||||
""")
|
||||
cursor.execute("""
|
||||
SELECT name FROM sqlite_master
|
||||
WHERE type='index' AND name LIKE 'idx_%'
|
||||
ORDER BY name
|
||||
""")
|
||||
|
||||
indexes = [row[0] for row in cursor.fetchall()]
|
||||
indexes = [row[0] for row in cursor.fetchall()]
|
||||
|
||||
required_indexes = [
|
||||
'idx_jobs_status',
|
||||
'idx_jobs_created_at',
|
||||
'idx_job_details_job_id',
|
||||
'idx_job_details_status',
|
||||
'idx_job_details_unique',
|
||||
'idx_trading_days_lookup', # Compound index in new schema
|
||||
'idx_holdings_day',
|
||||
'idx_actions_day',
|
||||
'idx_tool_usage_job_date_model'
|
||||
]
|
||||
required_indexes = [
|
||||
'idx_jobs_status',
|
||||
'idx_jobs_created_at',
|
||||
'idx_job_details_job_id',
|
||||
'idx_job_details_status',
|
||||
'idx_job_details_unique',
|
||||
'idx_trading_days_lookup', # Compound index in new schema
|
||||
'idx_holdings_day',
|
||||
'idx_actions_day',
|
||||
'idx_tool_usage_job_date_model'
|
||||
]
|
||||
|
||||
for index in required_indexes:
|
||||
assert index in indexes, f"Missing index: {index}"
|
||||
for index in required_indexes:
|
||||
assert index in indexes, f"Missing index: {index}"
|
||||
|
||||
conn.close()
|
||||
|
||||
def test_initialize_database_idempotent(self, clean_db):
|
||||
"""Should be safe to call multiple times."""
|
||||
@@ -205,17 +198,16 @@ class TestSchemaInitialization:
|
||||
initialize_database(clean_db)
|
||||
|
||||
# Should still have correct tables
|
||||
conn = get_db_connection(clean_db)
|
||||
cursor = conn.cursor()
|
||||
with db_connection(clean_db) as conn:
|
||||
cursor = conn.cursor()
|
||||
|
||||
cursor.execute("""
|
||||
SELECT COUNT(*) FROM sqlite_master
|
||||
WHERE type='table' AND name='jobs'
|
||||
""")
|
||||
cursor.execute("""
|
||||
SELECT COUNT(*) FROM sqlite_master
|
||||
WHERE type='table' AND name='jobs'
|
||||
""")
|
||||
|
||||
assert cursor.fetchone()[0] == 1 # Only one jobs table
|
||||
assert cursor.fetchone()[0] == 1 # Only one jobs table
|
||||
|
||||
conn.close()
|
||||
|
||||
|
||||
@pytest.mark.unit
|
||||
@@ -224,143 +216,140 @@ class TestForeignKeyConstraints:
|
||||
|
||||
def test_cascade_delete_job_details(self, clean_db, sample_job_data):
|
||||
"""Should cascade delete job_details when job is deleted."""
|
||||
conn = get_db_connection(clean_db)
|
||||
cursor = conn.cursor()
|
||||
with db_connection(clean_db) as conn:
|
||||
cursor = conn.cursor()
|
||||
|
||||
# Insert job
|
||||
cursor.execute("""
|
||||
INSERT INTO jobs (job_id, config_path, status, date_range, models, created_at)
|
||||
VALUES (?, ?, ?, ?, ?, ?)
|
||||
""", (
|
||||
sample_job_data["job_id"],
|
||||
sample_job_data["config_path"],
|
||||
sample_job_data["status"],
|
||||
sample_job_data["date_range"],
|
||||
sample_job_data["models"],
|
||||
sample_job_data["created_at"]
|
||||
))
|
||||
# Insert job
|
||||
cursor.execute("""
|
||||
INSERT INTO jobs (job_id, config_path, status, date_range, models, created_at)
|
||||
VALUES (?, ?, ?, ?, ?, ?)
|
||||
""", (
|
||||
sample_job_data["job_id"],
|
||||
sample_job_data["config_path"],
|
||||
sample_job_data["status"],
|
||||
sample_job_data["date_range"],
|
||||
sample_job_data["models"],
|
||||
sample_job_data["created_at"]
|
||||
))
|
||||
|
||||
# Insert job_detail
|
||||
cursor.execute("""
|
||||
INSERT INTO job_details (job_id, date, model, status)
|
||||
VALUES (?, ?, ?, ?)
|
||||
""", (sample_job_data["job_id"], "2025-01-16", "gpt-5", "pending"))
|
||||
# Insert job_detail
|
||||
cursor.execute("""
|
||||
INSERT INTO job_details (job_id, date, model, status)
|
||||
VALUES (?, ?, ?, ?)
|
||||
""", (sample_job_data["job_id"], "2025-01-16", "gpt-5", "pending"))
|
||||
|
||||
conn.commit()
|
||||
conn.commit()
|
||||
|
||||
# Verify job_detail exists
|
||||
cursor.execute("SELECT COUNT(*) FROM job_details WHERE job_id = ?", (sample_job_data["job_id"],))
|
||||
assert cursor.fetchone()[0] == 1
|
||||
# Verify job_detail exists
|
||||
cursor.execute("SELECT COUNT(*) FROM job_details WHERE job_id = ?", (sample_job_data["job_id"],))
|
||||
assert cursor.fetchone()[0] == 1
|
||||
|
||||
# Delete job
|
||||
cursor.execute("DELETE FROM jobs WHERE job_id = ?", (sample_job_data["job_id"],))
|
||||
conn.commit()
|
||||
# Delete job
|
||||
cursor.execute("DELETE FROM jobs WHERE job_id = ?", (sample_job_data["job_id"],))
|
||||
conn.commit()
|
||||
|
||||
# Verify job_detail was cascade deleted
|
||||
cursor.execute("SELECT COUNT(*) FROM job_details WHERE job_id = ?", (sample_job_data["job_id"],))
|
||||
assert cursor.fetchone()[0] == 0
|
||||
# Verify job_detail was cascade deleted
|
||||
cursor.execute("SELECT COUNT(*) FROM job_details WHERE job_id = ?", (sample_job_data["job_id"],))
|
||||
assert cursor.fetchone()[0] == 0
|
||||
|
||||
conn.close()
|
||||
|
||||
def test_cascade_delete_trading_days(self, clean_db, sample_job_data):
|
||||
"""Should cascade delete trading_days when job is deleted."""
|
||||
conn = get_db_connection(clean_db)
|
||||
cursor = conn.cursor()
|
||||
with db_connection(clean_db) as conn:
|
||||
cursor = conn.cursor()
|
||||
|
||||
# Insert job
|
||||
cursor.execute("""
|
||||
INSERT INTO jobs (job_id, config_path, status, date_range, models, created_at)
|
||||
VALUES (?, ?, ?, ?, ?, ?)
|
||||
""", (
|
||||
sample_job_data["job_id"],
|
||||
sample_job_data["config_path"],
|
||||
sample_job_data["status"],
|
||||
sample_job_data["date_range"],
|
||||
sample_job_data["models"],
|
||||
sample_job_data["created_at"]
|
||||
))
|
||||
# Insert job
|
||||
cursor.execute("""
|
||||
INSERT INTO jobs (job_id, config_path, status, date_range, models, created_at)
|
||||
VALUES (?, ?, ?, ?, ?, ?)
|
||||
""", (
|
||||
sample_job_data["job_id"],
|
||||
sample_job_data["config_path"],
|
||||
sample_job_data["status"],
|
||||
sample_job_data["date_range"],
|
||||
sample_job_data["models"],
|
||||
sample_job_data["created_at"]
|
||||
))
|
||||
|
||||
# Insert trading_day
|
||||
cursor.execute("""
|
||||
INSERT INTO trading_days (
|
||||
job_id, date, model, starting_cash, ending_cash,
|
||||
starting_portfolio_value, ending_portfolio_value,
|
||||
daily_profit, daily_return_pct, days_since_last_trading,
|
||||
total_actions, created_at
|
||||
) VALUES (?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?)
|
||||
""", (
|
||||
sample_job_data["job_id"], "2025-01-16", "test-model",
|
||||
10000.0, 9500.0, 10000.0, 9500.0,
|
||||
-500.0, -5.0, 0, 1, "2025-01-16T10:00:00Z"
|
||||
))
|
||||
# Insert trading_day
|
||||
cursor.execute("""
|
||||
INSERT INTO trading_days (
|
||||
job_id, date, model, starting_cash, ending_cash,
|
||||
starting_portfolio_value, ending_portfolio_value,
|
||||
daily_profit, daily_return_pct, days_since_last_trading,
|
||||
total_actions, created_at
|
||||
) VALUES (?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?)
|
||||
""", (
|
||||
sample_job_data["job_id"], "2025-01-16", "test-model",
|
||||
10000.0, 9500.0, 10000.0, 9500.0,
|
||||
-500.0, -5.0, 0, 1, "2025-01-16T10:00:00Z"
|
||||
))
|
||||
|
||||
conn.commit()
|
||||
conn.commit()
|
||||
|
||||
# Delete job
|
||||
cursor.execute("DELETE FROM jobs WHERE job_id = ?", (sample_job_data["job_id"],))
|
||||
conn.commit()
|
||||
# Delete job
|
||||
cursor.execute("DELETE FROM jobs WHERE job_id = ?", (sample_job_data["job_id"],))
|
||||
conn.commit()
|
||||
|
||||
# Verify trading_day was cascade deleted
|
||||
cursor.execute("SELECT COUNT(*) FROM trading_days WHERE job_id = ?", (sample_job_data["job_id"],))
|
||||
assert cursor.fetchone()[0] == 0
|
||||
# Verify trading_day was cascade deleted
|
||||
cursor.execute("SELECT COUNT(*) FROM trading_days WHERE job_id = ?", (sample_job_data["job_id"],))
|
||||
assert cursor.fetchone()[0] == 0
|
||||
|
||||
conn.close()
|
||||
|
||||
def test_cascade_delete_holdings(self, clean_db, sample_job_data):
|
||||
"""Should cascade delete holdings when trading_day is deleted."""
|
||||
conn = get_db_connection(clean_db)
|
||||
cursor = conn.cursor()
|
||||
with db_connection(clean_db) as conn:
|
||||
cursor = conn.cursor()
|
||||
|
||||
# Insert job
|
||||
cursor.execute("""
|
||||
INSERT INTO jobs (job_id, config_path, status, date_range, models, created_at)
|
||||
VALUES (?, ?, ?, ?, ?, ?)
|
||||
""", (
|
||||
sample_job_data["job_id"],
|
||||
sample_job_data["config_path"],
|
||||
sample_job_data["status"],
|
||||
sample_job_data["date_range"],
|
||||
sample_job_data["models"],
|
||||
sample_job_data["created_at"]
|
||||
))
|
||||
# Insert job
|
||||
cursor.execute("""
|
||||
INSERT INTO jobs (job_id, config_path, status, date_range, models, created_at)
|
||||
VALUES (?, ?, ?, ?, ?, ?)
|
||||
""", (
|
||||
sample_job_data["job_id"],
|
||||
sample_job_data["config_path"],
|
||||
sample_job_data["status"],
|
||||
sample_job_data["date_range"],
|
||||
sample_job_data["models"],
|
||||
sample_job_data["created_at"]
|
||||
))
|
||||
|
||||
# Insert trading_day
|
||||
cursor.execute("""
|
||||
INSERT INTO trading_days (
|
||||
job_id, date, model, starting_cash, ending_cash,
|
||||
starting_portfolio_value, ending_portfolio_value,
|
||||
daily_profit, daily_return_pct, days_since_last_trading,
|
||||
total_actions, created_at
|
||||
) VALUES (?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?)
|
||||
""", (
|
||||
sample_job_data["job_id"], "2025-01-16", "test-model",
|
||||
10000.0, 9500.0, 10000.0, 9500.0,
|
||||
-500.0, -5.0, 0, 1, "2025-01-16T10:00:00Z"
|
||||
))
|
||||
# Insert trading_day
|
||||
cursor.execute("""
|
||||
INSERT INTO trading_days (
|
||||
job_id, date, model, starting_cash, ending_cash,
|
||||
starting_portfolio_value, ending_portfolio_value,
|
||||
daily_profit, daily_return_pct, days_since_last_trading,
|
||||
total_actions, created_at
|
||||
) VALUES (?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?)
|
||||
""", (
|
||||
sample_job_data["job_id"], "2025-01-16", "test-model",
|
||||
10000.0, 9500.0, 10000.0, 9500.0,
|
||||
-500.0, -5.0, 0, 1, "2025-01-16T10:00:00Z"
|
||||
))
|
||||
|
||||
trading_day_id = cursor.lastrowid
|
||||
trading_day_id = cursor.lastrowid
|
||||
|
||||
# Insert holding
|
||||
cursor.execute("""
|
||||
INSERT INTO holdings (trading_day_id, symbol, quantity)
|
||||
VALUES (?, ?, ?)
|
||||
""", (trading_day_id, "AAPL", 10))
|
||||
# Insert holding
|
||||
cursor.execute("""
|
||||
INSERT INTO holdings (trading_day_id, symbol, quantity)
|
||||
VALUES (?, ?, ?)
|
||||
""", (trading_day_id, "AAPL", 10))
|
||||
|
||||
conn.commit()
|
||||
conn.commit()
|
||||
|
||||
# Verify holding exists
|
||||
cursor.execute("SELECT COUNT(*) FROM holdings WHERE trading_day_id = ?", (trading_day_id,))
|
||||
assert cursor.fetchone()[0] == 1
|
||||
# Verify holding exists
|
||||
cursor.execute("SELECT COUNT(*) FROM holdings WHERE trading_day_id = ?", (trading_day_id,))
|
||||
assert cursor.fetchone()[0] == 1
|
||||
|
||||
# Delete trading_day
|
||||
cursor.execute("DELETE FROM trading_days WHERE id = ?", (trading_day_id,))
|
||||
conn.commit()
|
||||
# Delete trading_day
|
||||
cursor.execute("DELETE FROM trading_days WHERE id = ?", (trading_day_id,))
|
||||
conn.commit()
|
||||
|
||||
# Verify holding was cascade deleted
|
||||
cursor.execute("SELECT COUNT(*) FROM holdings WHERE trading_day_id = ?", (trading_day_id,))
|
||||
assert cursor.fetchone()[0] == 0
|
||||
# Verify holding was cascade deleted
|
||||
cursor.execute("SELECT COUNT(*) FROM holdings WHERE trading_day_id = ?", (trading_day_id,))
|
||||
assert cursor.fetchone()[0] == 0
|
||||
|
||||
conn.close()
|
||||
|
||||
|
||||
@pytest.mark.unit
|
||||
@@ -378,22 +367,20 @@ class TestUtilityFunctions:
|
||||
db.connection.close()
|
||||
|
||||
# Verify tables exist
|
||||
conn = get_db_connection(test_db_path)
|
||||
cursor = conn.cursor()
|
||||
cursor.execute("SELECT COUNT(*) FROM sqlite_master WHERE type='table' AND name NOT LIKE 'sqlite_%'")
|
||||
# New schema: jobs, job_details, trading_days, holdings, actions, tool_usage, price_data, price_data_coverage, simulation_runs (9 tables)
|
||||
assert cursor.fetchone()[0] == 9
|
||||
conn.close()
|
||||
with db_connection(test_db_path) as conn:
|
||||
cursor = conn.cursor()
|
||||
cursor.execute("SELECT COUNT(*) FROM sqlite_master WHERE type='table' AND name NOT LIKE 'sqlite_%'")
|
||||
# New schema: jobs, job_details, trading_days, holdings, actions, tool_usage, price_data, price_data_coverage, simulation_runs (9 tables)
|
||||
assert cursor.fetchone()[0] == 9
|
||||
|
||||
# Drop all tables
|
||||
drop_all_tables(test_db_path)
|
||||
|
||||
# Verify tables are gone
|
||||
conn = get_db_connection(test_db_path)
|
||||
cursor = conn.cursor()
|
||||
cursor.execute("SELECT COUNT(*) FROM sqlite_master WHERE type='table' AND name NOT LIKE 'sqlite_%'")
|
||||
assert cursor.fetchone()[0] == 0
|
||||
conn.close()
|
||||
with db_connection(test_db_path) as conn:
|
||||
cursor = conn.cursor()
|
||||
cursor.execute("SELECT COUNT(*) FROM sqlite_master WHERE type='table' AND name NOT LIKE 'sqlite_%'")
|
||||
assert cursor.fetchone()[0] == 0
|
||||
|
||||
def test_vacuum_database(self, clean_db):
|
||||
"""Should execute VACUUM command without errors."""
|
||||
@@ -401,11 +388,10 @@ class TestUtilityFunctions:
|
||||
vacuum_database(clean_db)
|
||||
|
||||
# Verify database still accessible
|
||||
conn = get_db_connection(clean_db)
|
||||
cursor = conn.cursor()
|
||||
cursor.execute("SELECT COUNT(*) FROM jobs")
|
||||
assert cursor.fetchone()[0] == 0
|
||||
conn.close()
|
||||
with db_connection(clean_db) as conn:
|
||||
cursor = conn.cursor()
|
||||
cursor.execute("SELECT COUNT(*) FROM jobs")
|
||||
assert cursor.fetchone()[0] == 0
|
||||
|
||||
def test_get_database_stats_empty(self, clean_db):
|
||||
"""Should return correct stats for empty database."""
|
||||
@@ -421,30 +407,29 @@ class TestUtilityFunctions:
|
||||
|
||||
def test_get_database_stats_with_data(self, clean_db, sample_job_data):
|
||||
"""Should return correct row counts with data."""
|
||||
conn = get_db_connection(clean_db)
|
||||
cursor = conn.cursor()
|
||||
with db_connection(clean_db) as conn:
|
||||
cursor = conn.cursor()
|
||||
|
||||
# Insert job
|
||||
cursor.execute("""
|
||||
INSERT INTO jobs (job_id, config_path, status, date_range, models, created_at)
|
||||
VALUES (?, ?, ?, ?, ?, ?)
|
||||
""", (
|
||||
sample_job_data["job_id"],
|
||||
sample_job_data["config_path"],
|
||||
sample_job_data["status"],
|
||||
sample_job_data["date_range"],
|
||||
sample_job_data["models"],
|
||||
sample_job_data["created_at"]
|
||||
))
|
||||
# Insert job
|
||||
cursor.execute("""
|
||||
INSERT INTO jobs (job_id, config_path, status, date_range, models, created_at)
|
||||
VALUES (?, ?, ?, ?, ?, ?)
|
||||
""", (
|
||||
sample_job_data["job_id"],
|
||||
sample_job_data["config_path"],
|
||||
sample_job_data["status"],
|
||||
sample_job_data["date_range"],
|
||||
sample_job_data["models"],
|
||||
sample_job_data["created_at"]
|
||||
))
|
||||
|
||||
# Insert job_detail
|
||||
cursor.execute("""
|
||||
INSERT INTO job_details (job_id, date, model, status)
|
||||
VALUES (?, ?, ?, ?)
|
||||
""", (sample_job_data["job_id"], "2025-01-16", "gpt-5", "pending"))
|
||||
# Insert job_detail
|
||||
cursor.execute("""
|
||||
INSERT INTO job_details (job_id, date, model, status)
|
||||
VALUES (?, ?, ?, ?)
|
||||
""", (sample_job_data["job_id"], "2025-01-16", "gpt-5", "pending"))
|
||||
|
||||
conn.commit()
|
||||
conn.close()
|
||||
conn.commit()
|
||||
|
||||
stats = get_database_stats(clean_db)
|
||||
|
||||
@@ -468,24 +453,23 @@ class TestSchemaMigration:
|
||||
initialize_database(test_db_path)
|
||||
|
||||
# Verify warnings column exists in current schema
|
||||
conn = get_db_connection(test_db_path)
|
||||
cursor = conn.cursor()
|
||||
cursor.execute("PRAGMA table_info(jobs)")
|
||||
columns = [row[1] for row in cursor.fetchall()]
|
||||
assert 'warnings' in columns, "warnings column should exist in jobs table schema"
|
||||
with db_connection(test_db_path) as conn:
|
||||
cursor = conn.cursor()
|
||||
cursor.execute("PRAGMA table_info(jobs)")
|
||||
columns = [row[1] for row in cursor.fetchall()]
|
||||
assert 'warnings' in columns, "warnings column should exist in jobs table schema"
|
||||
|
||||
# Verify we can insert and query warnings
|
||||
cursor.execute("""
|
||||
INSERT INTO jobs (job_id, config_path, status, date_range, models, created_at, warnings)
|
||||
VALUES (?, ?, ?, ?, ?, ?, ?)
|
||||
""", ("test-job", "configs/test.json", "completed", "[]", "[]", "2025-01-20T00:00:00Z", "Test warning"))
|
||||
conn.commit()
|
||||
# Verify we can insert and query warnings
|
||||
cursor.execute("""
|
||||
INSERT INTO jobs (job_id, config_path, status, date_range, models, created_at, warnings)
|
||||
VALUES (?, ?, ?, ?, ?, ?, ?)
|
||||
""", ("test-job", "configs/test.json", "completed", "[]", "[]", "2025-01-20T00:00:00Z", "Test warning"))
|
||||
conn.commit()
|
||||
|
||||
cursor.execute("SELECT warnings FROM jobs WHERE job_id = ?", ("test-job",))
|
||||
result = cursor.fetchone()
|
||||
assert result[0] == "Test warning"
|
||||
cursor.execute("SELECT warnings FROM jobs WHERE job_id = ?", ("test-job",))
|
||||
result = cursor.fetchone()
|
||||
assert result[0] == "Test warning"
|
||||
|
||||
conn.close()
|
||||
|
||||
# Clean up after test - drop all tables so we don't affect other tests
|
||||
drop_all_tables(test_db_path)
|
||||
@@ -497,74 +481,71 @@ class TestCheckConstraints:
|
||||
|
||||
def test_jobs_status_constraint(self, clean_db):
|
||||
"""Should reject invalid job status values."""
|
||||
conn = get_db_connection(clean_db)
|
||||
cursor = conn.cursor()
|
||||
with db_connection(clean_db) as conn:
|
||||
cursor = conn.cursor()
|
||||
|
||||
# Try to insert job with invalid status
|
||||
with pytest.raises(sqlite3.IntegrityError, match="CHECK constraint failed"):
|
||||
cursor.execute("""
|
||||
INSERT INTO jobs (job_id, config_path, status, date_range, models, created_at)
|
||||
VALUES (?, ?, ?, ?, ?, ?)
|
||||
""", ("test-job", "configs/test.json", "invalid_status", "[]", "[]", "2025-01-20T00:00:00Z"))
|
||||
# Try to insert job with invalid status
|
||||
with pytest.raises(sqlite3.IntegrityError, match="CHECK constraint failed"):
|
||||
cursor.execute("""
|
||||
INSERT INTO jobs (job_id, config_path, status, date_range, models, created_at)
|
||||
VALUES (?, ?, ?, ?, ?, ?)
|
||||
""", ("test-job", "configs/test.json", "invalid_status", "[]", "[]", "2025-01-20T00:00:00Z"))
|
||||
|
||||
conn.close()
|
||||
|
||||
def test_job_details_status_constraint(self, clean_db, sample_job_data):
|
||||
"""Should reject invalid job_detail status values."""
|
||||
conn = get_db_connection(clean_db)
|
||||
cursor = conn.cursor()
|
||||
with db_connection(clean_db) as conn:
|
||||
cursor = conn.cursor()
|
||||
|
||||
# Insert valid job first
|
||||
cursor.execute("""
|
||||
INSERT INTO jobs (job_id, config_path, status, date_range, models, created_at)
|
||||
VALUES (?, ?, ?, ?, ?, ?)
|
||||
""", tuple(sample_job_data.values()))
|
||||
|
||||
# Try to insert job_detail with invalid status
|
||||
with pytest.raises(sqlite3.IntegrityError, match="CHECK constraint failed"):
|
||||
# Insert valid job first
|
||||
cursor.execute("""
|
||||
INSERT INTO job_details (job_id, date, model, status)
|
||||
VALUES (?, ?, ?, ?)
|
||||
""", (sample_job_data["job_id"], "2025-01-16", "gpt-5", "invalid_status"))
|
||||
INSERT INTO jobs (job_id, config_path, status, date_range, models, created_at)
|
||||
VALUES (?, ?, ?, ?, ?, ?)
|
||||
""", tuple(sample_job_data.values()))
|
||||
|
||||
# Try to insert job_detail with invalid status
|
||||
with pytest.raises(sqlite3.IntegrityError, match="CHECK constraint failed"):
|
||||
cursor.execute("""
|
||||
INSERT INTO job_details (job_id, date, model, status)
|
||||
VALUES (?, ?, ?, ?)
|
||||
""", (sample_job_data["job_id"], "2025-01-16", "gpt-5", "invalid_status"))
|
||||
|
||||
conn.close()
|
||||
|
||||
def test_actions_action_type_constraint(self, clean_db, sample_job_data):
|
||||
"""Should reject invalid action_type values in actions table."""
|
||||
conn = get_db_connection(clean_db)
|
||||
cursor = conn.cursor()
|
||||
with db_connection(clean_db) as conn:
|
||||
cursor = conn.cursor()
|
||||
|
||||
# Insert valid job first
|
||||
cursor.execute("""
|
||||
INSERT INTO jobs (job_id, config_path, status, date_range, models, created_at)
|
||||
VALUES (?, ?, ?, ?, ?, ?)
|
||||
""", tuple(sample_job_data.values()))
|
||||
|
||||
# Insert trading_day
|
||||
cursor.execute("""
|
||||
INSERT INTO trading_days (
|
||||
job_id, date, model, starting_cash, ending_cash,
|
||||
starting_portfolio_value, ending_portfolio_value,
|
||||
daily_profit, daily_return_pct, days_since_last_trading,
|
||||
total_actions, created_at
|
||||
) VALUES (?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?)
|
||||
""", (
|
||||
sample_job_data["job_id"], "2025-01-16", "test-model",
|
||||
10000.0, 9500.0, 10000.0, 9500.0,
|
||||
-500.0, -5.0, 0, 1, "2025-01-16T10:00:00Z"
|
||||
))
|
||||
|
||||
trading_day_id = cursor.lastrowid
|
||||
|
||||
# Try to insert action with invalid action_type
|
||||
with pytest.raises(sqlite3.IntegrityError, match="CHECK constraint failed"):
|
||||
# Insert valid job first
|
||||
cursor.execute("""
|
||||
INSERT INTO actions (
|
||||
trading_day_id, action_type, symbol, quantity, price, created_at
|
||||
) VALUES (?, ?, ?, ?, ?, ?)
|
||||
""", (trading_day_id, "invalid_action", "AAPL", 10, 150.0, "2025-01-16T10:00:00Z"))
|
||||
INSERT INTO jobs (job_id, config_path, status, date_range, models, created_at)
|
||||
VALUES (?, ?, ?, ?, ?, ?)
|
||||
""", tuple(sample_job_data.values()))
|
||||
|
||||
# Insert trading_day
|
||||
cursor.execute("""
|
||||
INSERT INTO trading_days (
|
||||
job_id, date, model, starting_cash, ending_cash,
|
||||
starting_portfolio_value, ending_portfolio_value,
|
||||
daily_profit, daily_return_pct, days_since_last_trading,
|
||||
total_actions, created_at
|
||||
) VALUES (?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?)
|
||||
""", (
|
||||
sample_job_data["job_id"], "2025-01-16", "test-model",
|
||||
10000.0, 9500.0, 10000.0, 9500.0,
|
||||
-500.0, -5.0, 0, 1, "2025-01-16T10:00:00Z"
|
||||
))
|
||||
|
||||
trading_day_id = cursor.lastrowid
|
||||
|
||||
# Try to insert action with invalid action_type
|
||||
with pytest.raises(sqlite3.IntegrityError, match="CHECK constraint failed"):
|
||||
cursor.execute("""
|
||||
INSERT INTO actions (
|
||||
trading_day_id, action_type, symbol, quantity, price, created_at
|
||||
) VALUES (?, ?, ?, ?, ?, ?)
|
||||
""", (trading_day_id, "invalid_action", "AAPL", 10, 150.0, "2025-01-16T10:00:00Z"))
|
||||
|
||||
conn.close()
|
||||
|
||||
|
||||
# Coverage target: 95%+ for api/database.py
|
||||
|
||||
Reference in New Issue
Block a user