- Add Prometheus metrics for marketplace API throughput and error rates with new dashboard panels - Implement confidential transaction models with encryption support and access control - Add key management system with registration, rotation, and audit logging - Create services and registry routers for service discovery and management - Integrate ZK proof generation for privacy-preserving receipts - Add metrics instru
12 KiB
12 KiB
AITBC Test Suite
This directory contains the comprehensive test suite for the AITBC platform, including unit tests, integration tests, end-to-end tests, security tests, and load tests.
Table of Contents
- Test Structure
- Prerequisites
- Running Tests
- Test Types
- Configuration
- CI/CD Integration
- Troubleshooting
Test Structure
tests/
├── conftest.py # Shared fixtures and configuration
├── pytest.ini # Pytest configuration
├── README.md # This file
├── unit/ # Unit tests
│ └── test_coordinator_api.py
├── integration/ # Integration tests
│ └── test_blockchain_node.py
├── e2e/ # End-to-end tests
│ └── test_wallet_daemon.py
├── security/ # Security tests
│ └── test_confidential_transactions.py
├── load/ # Load tests
│ └── locustfile.py
└── fixtures/ # Test data and fixtures
├── sample_receipts.json
└── test_transactions.json
Prerequisites
Required Dependencies
# Core testing framework
pip install pytest pytest-asyncio pytest-cov pytest-mock pytest-xdist
# Security testing
pip install bandit safety
# Load testing
pip install locust
# Additional testing tools
pip install requests-mock websockets psutil
System Dependencies
# Ubuntu/Debian
sudo apt-get update
sudo apt-get install -y postgresql redis-server
# macOS
brew install postgresql redis
# Docker (for isolated testing)
docker --version
Environment Setup
- Clone the repository:
git clone https://github.com/aitbc/aitbc.git
cd aitbc
- Create virtual environment:
python -m venv venv
source venv/bin/activate # On Windows: venv\Scripts\activate
- Install dependencies:
pip install -r requirements.txt
pip install -r requirements-test.txt
- Set up test databases:
# PostgreSQL
createdb aitbc_test
# Redis (use test database 1)
redis-cli -n 1 FLUSHDB
- Environment variables:
export DATABASE_URL="postgresql://localhost/aitbc_test"
export REDIS_URL="redis://localhost:6379/1"
export TEST_MODE="true"
Running Tests
Basic Commands
# Run all tests
pytest
# Run with coverage
pytest --cov=apps --cov=packages
# Run specific test file
pytest tests/unit/test_coordinator_api.py
# Run specific test class
pytest tests/unit/test_coordinator_api.py::TestJobEndpoints
# Run specific test method
pytest tests/unit/test_coordinator_api.py::TestJobEndpoints::test_create_job_success
Running by Test Type
# Unit tests only (fast)
pytest -m unit
# Integration tests (require services)
pytest -m integration
# End-to-end tests (full system)
pytest -m e2e
# Security tests
pytest -m security
# Load tests (requires Locust)
locust -f tests/load/locustfile.py
# Performance tests
pytest -m performance
# GPU tests (requires GPU)
pytest -m gpu
Parallel Execution
# Run with multiple workers
pytest -n auto
# Specify number of workers
pytest -n 4
# Distribute by test file
pytest --dist=loadfile
Filtering Tests
# Run tests matching pattern
pytest -k "test_create_job"
# Run tests not matching pattern
pytest -k "not slow"
# Run tests with multiple markers
pytest -m "unit and not slow"
# Run tests with any of multiple markers
pytest -m "unit or integration"
Test Types
Unit Tests (tests/unit/)
Fast, isolated tests that test individual components:
- Purpose: Test individual functions and classes
- Speed: < 1 second per test
- Dependencies: Mocked external services
- Database: In-memory SQLite
- Examples:
pytest tests/unit/ -v
Integration Tests (tests/integration/)
Tests that verify multiple components work together:
- Purpose: Test component interactions
- Speed: 1-10 seconds per test
- Dependencies: Real services required
- Database: Test PostgreSQL instance
- Examples:
# Start required services first docker-compose up -d postgres redis # Run integration tests pytest tests/integration/ -v
End-to-End Tests (tests/e2e/)
Full system tests that simulate real user workflows:
- Purpose: Test complete user journeys
- Speed: 10-60 seconds per test
- Dependencies: Full system running
- Database: Production-like setup
- Examples:
# Start full system docker-compose up -d # Run E2E tests pytest tests/e2e/ -v -s
Security Tests (tests/security/)
Tests that verify security properties and vulnerability resistance:
- Purpose: Test security controls
- Speed: Variable (some are slow)
- Dependencies: May require special setup
- Tools: Bandit, Safety, Custom security tests
- Examples:
# Run security scanner bandit -r apps/ -f json -o bandit-report.json # Run security tests pytest tests/security/ -v
Load Tests (tests/load/)
Performance and scalability tests:
- Purpose: Test system under load
- Speed: Long-running (minutes)
- Dependencies: Locust, staging environment
- Examples:
# Run Locust web UI locust -f tests/load/locustfile.py --web-host 127.0.0.1 # Run headless locust -f tests/load/locustfile.py --headless -u 100 -r 10 -t 5m
Configuration
Pytest Configuration (pytest.ini)
Key configuration options:
[tool:pytest]
# Test paths
testpaths = tests
python_files = test_*.py
# Coverage settings
addopts = --cov=apps --cov=packages --cov-report=html
# Markers
markers =
unit: Unit tests
integration: Integration tests
e2e: End-to-end tests
security: Security tests
slow: Slow tests
Environment Variables
# Test configuration
export TEST_MODE=true
export TEST_DATABASE_URL="postgresql://localhost/aitbc_test"
export TEST_REDIS_URL="redis://localhost:6379/1"
# Service URLs for integration tests
export COORDINATOR_URL="http://localhost:8001"
export WALLET_URL="http://localhost:8002"
export BLOCKCHAIN_URL="http://localhost:8545"
# Security test configuration
export TEST_HSM_ENDPOINT="http://localhost:9999"
export TEST_ZK_CIRCUITS_PATH="./apps/zk-circuits"
Test Data Management
# Using fixtures in conftest.py
@pytest.fixture
def test_data():
return {
"sample_job": {...},
"sample_receipt": {...},
}
# Custom test configuration
@pytest.fixture(scope="session")
def test_config():
return TestConfig(
database_url="sqlite:///:memory:",
redis_url="redis://localhost:6379/1",
)
CI/CD Integration
GitHub Actions Example
name: Tests
on: [push, pull_request]
jobs:
unit-tests:
runs-on: ubuntu-latest
steps:
- uses: actions/checkout@v3
- uses: actions/setup-python@v4
with:
python: "3.11"
- name: Install dependencies
run: |
pip install -r requirements.txt
pip install -r requirements-test.txt
- name: Run unit tests
run: |
pytest tests/unit/ -v --cov=apps --cov-report=xml
- name: Upload coverage
uses: codecov/codecov-action@v3
with:
file: ./coverage.xml
integration-tests:
runs-on: ubuntu-latest
services:
postgres:
image: postgres:15
env:
POSTGRES_PASSWORD: postgres
options: >-
--health-cmd pg_isready
--health-interval 10s
--health-timeout 5s
--health-retries 5
redis:
image: redis:7
options: >-
--health-cmd "redis-cli ping"
--health-interval 10s
--health-timeout 5s
--health-retries 5
steps:
- uses: actions/checkout@v3
- uses: actions/setup-python@v4
with:
python: "3.11"
- name: Install dependencies
run: |
pip install -r requirements.txt
pip install -r requirements-test.txt
- name: Run integration tests
run: |
pytest tests/integration/ -v
env:
DATABASE_URL: postgresql://postgres:postgres@localhost/postgres
REDIS_URL: redis://localhost:6379/0
Docker Compose for Testing
# docker-compose.test.yml
version: '3.8'
services:
postgres:
image: postgres:15
environment:
POSTGRES_DB: aitbc_test
POSTGRES_USER: test
POSTGRES_PASSWORD: test
ports:
- "5433:5432"
healthcheck:
test: ["CMD-SHELL", "pg_isready -U test"]
interval: 5s
timeout: 5s
retries: 5
redis:
image: redis:7-alpine
ports:
- "6380:6379"
healthcheck:
test: ["CMD", "redis-cli", "ping"]
interval: 5s
timeout: 5s
retries: 5
coordinator:
build: ./apps/coordinator-api
environment:
DATABASE_URL: postgresql://test:test@postgres:5432/aitbc_test
REDIS_URL: redis://redis:6379/0
depends_on:
postgres:
condition: service_healthy
redis:
condition: service_healthy
ports:
- "8001:8000"
Troubleshooting
Common Issues
-
Import Errors
# Ensure PYTHONPATH is set export PYTHONPATH="${PYTHONPATH}:$(pwd)" # Or install in development mode pip install -e . -
Database Connection Errors
# Check if PostgreSQL is running pg_isready -h localhost -p 5432 # Create test database createdb -h localhost -p 5432 aitbc_test -
Redis Connection Errors
# Check if Redis is running redis-cli ping # Use correct database redis-cli -n 1 FLUSHDB -
Test Timeouts
# Increase timeout for slow tests pytest --timeout=600 # Run tests sequentially pytest -n 0 -
Port Conflicts
# Kill processes using ports lsof -ti:8001 | xargs kill -9 lsof -ti:8002 | xargs kill -9
Debugging Tests
# Run with verbose output
pytest -v -s
# Stop on first failure
pytest -x
# Run with pdb on failure
pytest --pdb
# Print local variables on failure
pytest --tb=long
# Run specific test with debugging
pytest tests/unit/test_coordinator_api.py::TestJobEndpoints::test_create_job_success -v -s --pdb
Performance Issues
# Profile test execution
pytest --profile
# Find slowest tests
pytest --durations=10
# Run with memory profiling
pytest --memprof
Test Data Issues
# Clean test database
psql -h localhost -U test -d aitbc_test -c "DROP SCHEMA public CASCADE; CREATE SCHEMA public;"
# Reset Redis
redis-cli -n 1 FLUSHALL
# Regenerate test fixtures
python tests/generate_fixtures.py
Best Practices
- Write Isolated Tests: Each test should be independent
- Use Descriptive Names: Test names should describe what they test
- Mock External Dependencies: Use mocks for external services
- Clean Up Resources: Use fixtures for setup/teardown
- Test Edge Cases: Don't just test happy paths
- Use Type Hints: Makes tests more maintainable
- Document Complex Tests: Add comments for complex logic
Contributing
When adding new tests:
- Follow the existing structure and naming conventions
- Add appropriate markers (
@pytest.mark.unit, etc.) - Update this README if adding new test types
- Ensure tests pass on CI before submitting PR
- Add coverage for new features