Files
DocsGPT/tests/conftest.py
2026-04-27 22:09:33 +01:00

261 lines
7.9 KiB
Python

"""Root pytest fixtures for the DocsGPT backend suite.
Postgres fixture strategy
-------------------------
Regular unit tests get a Postgres connection from the ``pg_conn`` fixture
below, which is backed by ``pytest-postgresql``. That plugin spins up an
ephemeral ``pg_ctl``-managed cluster in a temp directory and tears it
down at the end of the session, so CI only needs Postgres *binaries*
installed, not a running service.
Tests under ``tests/storage/db/`` intentionally override ``pg_conn`` in
their own conftest to point at a real, long-running Postgres instance
(DBngin locally, a service container in CI). Those are integration/e2e
tests and are marked with ``@pytest.mark.integration``.
No mongomock. The ``mock_mongo_db`` fixture that used to live here was
removed as part of the Phase 4/5 Mongo→Postgres cutover. Tests that
still reference it will fail with "fixture not found" until the
corresponding route handler is migrated to a repository read.
"""
from __future__ import annotations
import os
# Disable the app's self-bootstrap (AUTO_CREATE_DB / AUTO_MIGRATE) before
# any ``application.*`` module is imported. ``application/app.py`` runs
# ``ensure_database_ready`` at import time using whatever ``POSTGRES_URI``
# is set in the environment — which in dev is the operator's local DB, not
# the ephemeral ``pytest-postgresql`` cluster that the fixtures below spin
# up. Tests manage their own schema via the ``pg_engine`` fixture
# (subprocess ``alembic upgrade head`` against the per-test URI), so the
# import-time bootstrap would at best be redundant and at worst would
# mutate the operator's dev DB. ``setdefault`` so a test run can still
# opt back in by setting the env var explicitly.
os.environ.setdefault("AUTO_MIGRATE", "false")
os.environ.setdefault("AUTO_CREATE_DB", "false")
import subprocess
import sys
from pathlib import Path
from unittest.mock import Mock
import pytest
from pytest_postgresql import factories
from sqlalchemy import create_engine
# ---------------------------------------------------------------------------
# Postgres fixtures (ephemeral cluster via pytest-postgresql)
# ---------------------------------------------------------------------------
# ``postgresql_proc`` starts a fresh ``pg_ctl`` cluster once per session.
# ``postgresql`` hands out a per-test DB on top of it. We layer our own
# SQLAlchemy engine + rolled-back transaction on top for test isolation.
postgresql_proc = factories.postgresql_proc()
postgresql = factories.postgresql("postgresql_proc")
def _sqlalchemy_url(pg_conn_info) -> str:
return (
"postgresql+psycopg://"
f"{pg_conn_info.user}:{pg_conn_info.password or ''}"
f"@{pg_conn_info.host}:{pg_conn_info.port}/{pg_conn_info.dbname}"
)
@pytest.fixture(scope="session")
def _alembic_ini_path() -> Path:
return Path(__file__).resolve().parent.parent / "application" / "alembic.ini"
@pytest.fixture()
def pg_engine(postgresql, _alembic_ini_path, monkeypatch):
"""Per-test SQLAlchemy engine against a fresh ephemeral Postgres DB.
Alembic is run from scratch against the per-test database so the full
schema is present. ``POSTGRES_URI`` is patched in the environment for
the duration of the test so any code that reads it via
``application.core.settings`` sees the ephemeral DB.
"""
url = _sqlalchemy_url(postgresql.info)
monkeypatch.setenv("POSTGRES_URI", url)
# Reset the settings cache so the new POSTGRES_URI is picked up if the
# settings module is already imported.
from application.core import settings as settings_module
monkeypatch.setattr(settings_module.settings, "POSTGRES_URI", url, raising=False)
subprocess.check_call(
[sys.executable, "-m", "alembic", "-c", str(_alembic_ini_path), "upgrade", "head"],
timeout=60,
env={**__import__("os").environ, "POSTGRES_URI": url},
)
engine = create_engine(url)
yield engine
engine.dispose()
@pytest.fixture()
def pg_conn(pg_engine):
"""Per-test connection wrapped in a transaction that always rolls back."""
conn = pg_engine.connect()
txn = conn.begin()
yield conn
txn.rollback()
conn.close()
# ---------------------------------------------------------------------------
# Generic unit-test fixtures (no DB, no Mongo)
# ---------------------------------------------------------------------------
@pytest.fixture
def mock_llm():
llm = Mock()
llm.gen_stream = Mock()
llm._supports_tools = True
llm._supports_structured_output = Mock(return_value=False)
llm.__class__.__name__ = "MockLLM"
# Mirror BaseLLM.__init__: real LLMCreator stores the resolved
# upstream model name on self.model_id. Tests that build agents via
# ``mock_llm_creator`` rely on the agent's ``upstream_model_id``
# falling through to this attribute.
llm.model_id = "gpt-4"
return llm
@pytest.fixture
def mock_llm_handler():
handler = Mock()
handler.process_message_flow = Mock()
return handler
@pytest.fixture
def mock_retriever():
retriever = Mock()
retriever.search = Mock(
return_value=[
{"text": "Test document 1", "filename": "doc1.txt", "source": "test"},
{"text": "Test document 2", "title": "doc2.txt", "source": "test"},
]
)
return retriever
@pytest.fixture
def sample_chat_history():
return [
{"prompt": "What is Python?", "response": "Python is a programming language."},
{"prompt": "Tell me more.", "response": "Python is known for simplicity."},
]
@pytest.fixture
def sample_tool_call():
return {
"tool_name": "test_tool",
"call_id": "123",
"action_name": "test_action",
"arguments": {"arg1": "value1"},
"result": "Tool executed successfully",
}
@pytest.fixture
def decoded_token():
return {"sub": "test_user", "email": "test@example.com"}
@pytest.fixture
def log_context():
from application.logging import LogContext
context = LogContext(
endpoint="test_endpoint",
activity_id="test_activity",
user="test_user",
api_key="test_key",
query="test query",
)
return context
@pytest.fixture
def mock_llm_creator(mock_llm, monkeypatch):
monkeypatch.setattr(
"application.llm.llm_creator.LLMCreator.create_llm", Mock(return_value=mock_llm)
)
return mock_llm
@pytest.fixture
def mock_llm_handler_creator(mock_llm_handler, monkeypatch):
monkeypatch.setattr(
"application.llm.handlers.handler_creator.LLMHandlerCreator.create_handler",
Mock(return_value=mock_llm_handler),
)
return mock_llm_handler
@pytest.fixture
def agent_base_params(decoded_token):
return {
"endpoint": "https://api.example.com",
"llm_name": "openai",
"model_id": "gpt-4",
"api_key": "test_api_key",
"user_api_key": None,
"prompt": "You are a helpful assistant.",
"chat_history": [],
"decoded_token": decoded_token,
"attachments": [],
"json_schema": None,
}
@pytest.fixture
def mock_tool():
tool = Mock()
tool.execute_action = Mock(return_value="Tool result")
tool.get_actions_metadata = Mock(
return_value=[
{
"name": "test_action",
"description": "A test action",
"parameters": {
"type": "object",
"properties": {
"param1": {"type": "string", "description": "Test parameter"}
},
"required": ["param1"],
},
}
]
)
return tool
@pytest.fixture
def mock_tool_manager(mock_tool, monkeypatch):
manager = Mock()
manager.load_tool = Mock(return_value=mock_tool)
monkeypatch.setattr(
"application.agents.tool_executor.ToolManager", Mock(return_value=manager)
)
return manager
@pytest.fixture
def flask_app():
from flask import Flask
app = Flask(__name__)
return app