mirror of
https://github.com/affaan-m/everything-claude-code.git
synced 2026-02-08 07:33:08 +08:00
## Python Support - **agents/python-reviewer.md**: Expert Python code review agent with PEP 8 compliance, type hints, security, and performance checks - **commands/python-review.md**: Slash command for automated Python code review with ruff, mypy, pylint, black, bandit - **skills/python-patterns/SKILL.md**: Python idioms, type hints, error handling, context managers, decorators, concurrency - **skills/python-testing/SKILL.md**: pytest configuration, fixtures, parametrization, mocking, async testing, TDD methodology ## Django Support - **skills/django-patterns/SKILL.md**: Django architecture, DRF patterns, project structure, QuerySets, serializers, ViewSets, service layer, caching - **skills/django-security/SKILL.md**: Django security best practices, authentication, CSRF, SQL injection, XSS prevention, production settings - **skills/django-tdd/SKILL.md**: Django testing with pytest-django, Factory Boy, model testing, API testing, integration testing - **skills/django-verification/SKILL.md**: Pre-deployment verification loop including migrations, tests, security scans, performance checks ## Documentation Enhancements - **Quick Start**: Added 3-step quick start guide to all READMEs (EN, zh-CN, zh-TW) - **Beautification**: Added emoji icons for better visual hierarchy across all READMEs - **.claude-plugin/plugin.json**: Added python-reviewer to agents list All files follow project conventions with proper frontmatter, markdown formatting, and comprehensive code examples. Co-authored-by: Freakz3z <freakk@FreakkdeMacBook-Air.local> Co-authored-by: Claude Sonnet 4.5 <noreply@anthropic.com>
816 lines
18 KiB
Markdown
816 lines
18 KiB
Markdown
---
|
|
name: python-testing
|
|
description: Python testing strategies using pytest, TDD methodology, fixtures, mocking, parametrization, and coverage requirements.
|
|
---
|
|
|
|
# Python Testing Patterns
|
|
|
|
Comprehensive testing strategies for Python applications using pytest, TDD methodology, and best practices.
|
|
|
|
## When to Activate
|
|
|
|
- Writing new Python code (follow TDD: red, green, refactor)
|
|
- Designing test suites for Python projects
|
|
- Reviewing Python test coverage
|
|
- Setting up testing infrastructure
|
|
|
|
## Core Testing Philosophy
|
|
|
|
### Test-Driven Development (TDD)
|
|
|
|
Always follow the TDD cycle:
|
|
|
|
1. **RED**: Write a failing test for the desired behavior
|
|
2. **GREEN**: Write minimal code to make the test pass
|
|
3. **REFACTOR**: Improve code while keeping tests green
|
|
|
|
```python
|
|
# Step 1: Write failing test (RED)
|
|
def test_add_numbers():
|
|
result = add(2, 3)
|
|
assert result == 5
|
|
|
|
# Step 2: Write minimal implementation (GREEN)
|
|
def add(a, b):
|
|
return a + b
|
|
|
|
# Step 3: Refactor if needed (REFACTOR)
|
|
```
|
|
|
|
### Coverage Requirements
|
|
|
|
- **Target**: 80%+ code coverage
|
|
- **Critical paths**: 100% coverage required
|
|
- Use `pytest --cov` to measure coverage
|
|
|
|
```bash
|
|
pytest --cov=mypackage --cov-report=term-missing --cov-report=html
|
|
```
|
|
|
|
## pytest Fundamentals
|
|
|
|
### Basic Test Structure
|
|
|
|
```python
|
|
import pytest
|
|
|
|
def test_addition():
|
|
"""Test basic addition."""
|
|
assert 2 + 2 == 4
|
|
|
|
def test_string_uppercase():
|
|
"""Test string uppercasing."""
|
|
text = "hello"
|
|
assert text.upper() == "HELLO"
|
|
|
|
def test_list_append():
|
|
"""Test list append."""
|
|
items = [1, 2, 3]
|
|
items.append(4)
|
|
assert 4 in items
|
|
assert len(items) == 4
|
|
```
|
|
|
|
### Assertions
|
|
|
|
```python
|
|
# Equality
|
|
assert result == expected
|
|
|
|
# Inequality
|
|
assert result != unexpected
|
|
|
|
# Truthiness
|
|
assert result # Truthy
|
|
assert not result # Falsy
|
|
assert result is True # Exactly True
|
|
assert result is False # Exactly False
|
|
assert result is None # Exactly None
|
|
|
|
# Membership
|
|
assert item in collection
|
|
assert item not in collection
|
|
|
|
# Comparisons
|
|
assert result > 0
|
|
assert 0 <= result <= 100
|
|
|
|
# Type checking
|
|
assert isinstance(result, str)
|
|
|
|
# Exception testing (preferred approach)
|
|
with pytest.raises(ValueError):
|
|
raise ValueError("error message")
|
|
|
|
# Check exception message
|
|
with pytest.raises(ValueError, match="invalid input"):
|
|
raise ValueError("invalid input provided")
|
|
|
|
# Check exception attributes
|
|
with pytest.raises(ValueError) as exc_info:
|
|
raise ValueError("error message")
|
|
assert str(exc_info.value) == "error message"
|
|
```
|
|
|
|
## Fixtures
|
|
|
|
### Basic Fixture Usage
|
|
|
|
```python
|
|
import pytest
|
|
|
|
@pytest.fixture
|
|
def sample_data():
|
|
"""Fixture providing sample data."""
|
|
return {"name": "Alice", "age": 30}
|
|
|
|
def test_sample_data(sample_data):
|
|
"""Test using the fixture."""
|
|
assert sample_data["name"] == "Alice"
|
|
assert sample_data["age"] == 30
|
|
```
|
|
|
|
### Fixture with Setup/Teardown
|
|
|
|
```python
|
|
@pytest.fixture
|
|
def database():
|
|
"""Fixture with setup and teardown."""
|
|
# Setup
|
|
db = Database(":memory:")
|
|
db.create_tables()
|
|
db.insert_test_data()
|
|
|
|
yield db # Provide to test
|
|
|
|
# Teardown
|
|
db.close()
|
|
|
|
def test_database_query(database):
|
|
"""Test database operations."""
|
|
result = database.query("SELECT * FROM users")
|
|
assert len(result) > 0
|
|
```
|
|
|
|
### Fixture Scopes
|
|
|
|
```python
|
|
# Function scope (default) - runs for each test
|
|
@pytest.fixture
|
|
def temp_file():
|
|
with open("temp.txt", "w") as f:
|
|
yield f
|
|
os.remove("temp.txt")
|
|
|
|
# Module scope - runs once per module
|
|
@pytest.fixture(scope="module")
|
|
def module_db():
|
|
db = Database(":memory:")
|
|
db.create_tables()
|
|
yield db
|
|
db.close()
|
|
|
|
# Session scope - runs once per test session
|
|
@pytest.fixture(scope="session")
|
|
def shared_resource():
|
|
resource = ExpensiveResource()
|
|
yield resource
|
|
resource.cleanup()
|
|
```
|
|
|
|
### Fixture with Parameters
|
|
|
|
```python
|
|
@pytest.fixture(params=[1, 2, 3])
|
|
def number(request):
|
|
"""Parameterized fixture."""
|
|
return request.param
|
|
|
|
def test_numbers(number):
|
|
"""Test runs 3 times, once for each parameter."""
|
|
assert number > 0
|
|
```
|
|
|
|
### Using Multiple Fixtures
|
|
|
|
```python
|
|
@pytest.fixture
|
|
def user():
|
|
return User(id=1, name="Alice")
|
|
|
|
@pytest.fixture
|
|
def admin():
|
|
return User(id=2, name="Admin", role="admin")
|
|
|
|
def test_user_admin_interaction(user, admin):
|
|
"""Test using multiple fixtures."""
|
|
assert admin.can_manage(user)
|
|
```
|
|
|
|
### Autouse Fixtures
|
|
|
|
```python
|
|
@pytest.fixture(autouse=True)
|
|
def reset_config():
|
|
"""Automatically runs before every test."""
|
|
Config.reset()
|
|
yield
|
|
Config.cleanup()
|
|
|
|
def test_without_fixture_call():
|
|
# reset_config runs automatically
|
|
assert Config.get_setting("debug") is False
|
|
```
|
|
|
|
### Conftest.py for Shared Fixtures
|
|
|
|
```python
|
|
# tests/conftest.py
|
|
import pytest
|
|
|
|
@pytest.fixture
|
|
def client():
|
|
"""Shared fixture for all tests."""
|
|
app = create_app(testing=True)
|
|
with app.test_client() as client:
|
|
yield client
|
|
|
|
@pytest.fixture
|
|
def auth_headers(client):
|
|
"""Generate auth headers for API testing."""
|
|
response = client.post("/api/login", json={
|
|
"username": "test",
|
|
"password": "test"
|
|
})
|
|
token = response.json["token"]
|
|
return {"Authorization": f"Bearer {token}"}
|
|
```
|
|
|
|
## Parametrization
|
|
|
|
### Basic Parametrization
|
|
|
|
```python
|
|
@pytest.mark.parametrize("input,expected", [
|
|
("hello", "HELLO"),
|
|
("world", "WORLD"),
|
|
("PyThOn", "PYTHON"),
|
|
])
|
|
def test_uppercase(input, expected):
|
|
"""Test runs 3 times with different inputs."""
|
|
assert input.upper() == expected
|
|
```
|
|
|
|
### Multiple Parameters
|
|
|
|
```python
|
|
@pytest.mark.parametrize("a,b,expected", [
|
|
(2, 3, 5),
|
|
(0, 0, 0),
|
|
(-1, 1, 0),
|
|
(100, 200, 300),
|
|
])
|
|
def test_add(a, b, expected):
|
|
"""Test addition with multiple inputs."""
|
|
assert add(a, b) == expected
|
|
```
|
|
|
|
### Parametrize with IDs
|
|
|
|
```python
|
|
@pytest.mark.parametrize("input,expected", [
|
|
("valid@email.com", True),
|
|
("invalid", False),
|
|
("@no-domain.com", False),
|
|
], ids=["valid-email", "missing-at", "missing-domain"])
|
|
def test_email_validation(input, expected):
|
|
"""Test email validation with readable test IDs."""
|
|
assert is_valid_email(input) is expected
|
|
```
|
|
|
|
### Parametrized Fixtures
|
|
|
|
```python
|
|
@pytest.fixture(params=["sqlite", "postgresql", "mysql"])
|
|
def db(request):
|
|
"""Test against multiple database backends."""
|
|
if request.param == "sqlite":
|
|
return Database(":memory:")
|
|
elif request.param == "postgresql":
|
|
return Database("postgresql://localhost/test")
|
|
elif request.param == "mysql":
|
|
return Database("mysql://localhost/test")
|
|
|
|
def test_database_operations(db):
|
|
"""Test runs 3 times, once for each database."""
|
|
result = db.query("SELECT 1")
|
|
assert result is not None
|
|
```
|
|
|
|
## Markers and Test Selection
|
|
|
|
### Custom Markers
|
|
|
|
```python
|
|
# Mark slow tests
|
|
@pytest.mark.slow
|
|
def test_slow_operation():
|
|
time.sleep(5)
|
|
|
|
# Mark integration tests
|
|
@pytest.mark.integration
|
|
def test_api_integration():
|
|
response = requests.get("https://api.example.com")
|
|
assert response.status_code == 200
|
|
|
|
# Mark unit tests
|
|
@pytest.mark.unit
|
|
def test_unit_logic():
|
|
assert calculate(2, 3) == 5
|
|
```
|
|
|
|
### Run Specific Tests
|
|
|
|
```bash
|
|
# Run only fast tests
|
|
pytest -m "not slow"
|
|
|
|
# Run only integration tests
|
|
pytest -m integration
|
|
|
|
# Run integration or slow tests
|
|
pytest -m "integration or slow"
|
|
|
|
# Run tests marked as unit but not slow
|
|
pytest -m "unit and not slow"
|
|
```
|
|
|
|
### Configure Markers in pytest.ini
|
|
|
|
```ini
|
|
[pytest]
|
|
markers =
|
|
slow: marks tests as slow
|
|
integration: marks tests as integration tests
|
|
unit: marks tests as unit tests
|
|
django: marks tests as requiring Django
|
|
```
|
|
|
|
## Mocking and Patching
|
|
|
|
### Mocking Functions
|
|
|
|
```python
|
|
from unittest.mock import patch, Mock
|
|
|
|
@patch("mypackage.external_api_call")
|
|
def test_with_mock(api_call_mock):
|
|
"""Test with mocked external API."""
|
|
api_call_mock.return_value = {"status": "success"}
|
|
|
|
result = my_function()
|
|
|
|
api_call_mock.assert_called_once()
|
|
assert result["status"] == "success"
|
|
```
|
|
|
|
### Mocking Return Values
|
|
|
|
```python
|
|
@patch("mypackage.Database.connect")
|
|
def test_database_connection(connect_mock):
|
|
"""Test with mocked database connection."""
|
|
connect_mock.return_value = MockConnection()
|
|
|
|
db = Database()
|
|
db.connect()
|
|
|
|
connect_mock.assert_called_once_with("localhost")
|
|
```
|
|
|
|
### Mocking Exceptions
|
|
|
|
```python
|
|
@patch("mypackage.api_call")
|
|
def test_api_error_handling(api_call_mock):
|
|
"""Test error handling with mocked exception."""
|
|
api_call_mock.side_effect = ConnectionError("Network error")
|
|
|
|
with pytest.raises(ConnectionError):
|
|
api_call()
|
|
|
|
api_call_mock.assert_called_once()
|
|
```
|
|
|
|
### Mocking Context Managers
|
|
|
|
```python
|
|
@patch("builtins.open", new_callable=mock_open)
|
|
def test_file_reading(mock_file):
|
|
"""Test file reading with mocked open."""
|
|
mock_file.return_value.read.return_value = "file content"
|
|
|
|
result = read_file("test.txt")
|
|
|
|
mock_file.assert_called_once_with("test.txt", "r")
|
|
assert result == "file content"
|
|
```
|
|
|
|
### Using Autospec
|
|
|
|
```python
|
|
@patch("mypackage.DBConnection", autospec=True)
|
|
def test_autospec(db_mock):
|
|
"""Test with autospec to catch API misuse."""
|
|
db = db_mock.return_value
|
|
db.query("SELECT * FROM users")
|
|
|
|
# This would fail if DBConnection doesn't have query method
|
|
db_mock.assert_called_once()
|
|
```
|
|
|
|
### Mock Class Instances
|
|
|
|
```python
|
|
class TestUserService:
|
|
@patch("mypackage.UserRepository")
|
|
def test_create_user(self, repo_mock):
|
|
"""Test user creation with mocked repository."""
|
|
repo_mock.return_value.save.return_value = User(id=1, name="Alice")
|
|
|
|
service = UserService(repo_mock.return_value)
|
|
user = service.create_user(name="Alice")
|
|
|
|
assert user.name == "Alice"
|
|
repo_mock.return_value.save.assert_called_once()
|
|
```
|
|
|
|
### Mock Property
|
|
|
|
```python
|
|
@pytest.fixture
|
|
def mock_config():
|
|
"""Create a mock with a property."""
|
|
config = Mock()
|
|
type(config).debug = PropertyMock(return_value=True)
|
|
type(config).api_key = PropertyMock(return_value="test-key")
|
|
return config
|
|
|
|
def test_with_mock_config(mock_config):
|
|
"""Test with mocked config properties."""
|
|
assert mock_config.debug is True
|
|
assert mock_config.api_key == "test-key"
|
|
```
|
|
|
|
## Testing Async Code
|
|
|
|
### Async Tests with pytest-asyncio
|
|
|
|
```python
|
|
import pytest
|
|
|
|
@pytest.mark.asyncio
|
|
async def test_async_function():
|
|
"""Test async function."""
|
|
result = await async_add(2, 3)
|
|
assert result == 5
|
|
|
|
@pytest.mark.asyncio
|
|
async def test_async_with_fixture(async_client):
|
|
"""Test async with async fixture."""
|
|
response = await async_client.get("/api/users")
|
|
assert response.status_code == 200
|
|
```
|
|
|
|
### Async Fixture
|
|
|
|
```python
|
|
@pytest.fixture
|
|
async def async_client():
|
|
"""Async fixture providing async test client."""
|
|
app = create_app()
|
|
async with app.test_client() as client:
|
|
yield client
|
|
|
|
@pytest.mark.asyncio
|
|
async def test_api_endpoint(async_client):
|
|
"""Test using async fixture."""
|
|
response = await async_client.get("/api/data")
|
|
assert response.status_code == 200
|
|
```
|
|
|
|
### Mocking Async Functions
|
|
|
|
```python
|
|
@pytest.mark.asyncio
|
|
@patch("mypackage.async_api_call")
|
|
async def test_async_mock(api_call_mock):
|
|
"""Test async function with mock."""
|
|
api_call_mock.return_value = {"status": "ok"}
|
|
|
|
result = await my_async_function()
|
|
|
|
api_call_mock.assert_awaited_once()
|
|
assert result["status"] == "ok"
|
|
```
|
|
|
|
## Testing Exceptions
|
|
|
|
### Testing Expected Exceptions
|
|
|
|
```python
|
|
def test_divide_by_zero():
|
|
"""Test that dividing by zero raises ZeroDivisionError."""
|
|
with pytest.raises(ZeroDivisionError):
|
|
divide(10, 0)
|
|
|
|
def test_custom_exception():
|
|
"""Test custom exception with message."""
|
|
with pytest.raises(ValueError, match="invalid input"):
|
|
validate_input("invalid")
|
|
```
|
|
|
|
### Testing Exception Attributes
|
|
|
|
```python
|
|
def test_exception_with_details():
|
|
"""Test exception with custom attributes."""
|
|
with pytest.raises(CustomError) as exc_info:
|
|
raise CustomError("error", code=400)
|
|
|
|
assert exc_info.value.code == 400
|
|
assert "error" in str(exc_info.value)
|
|
```
|
|
|
|
## Testing Side Effects
|
|
|
|
### Testing File Operations
|
|
|
|
```python
|
|
import tempfile
|
|
import os
|
|
|
|
def test_file_processing():
|
|
"""Test file processing with temp file."""
|
|
with tempfile.NamedTemporaryFile(mode='w', delete=False, suffix='.txt') as f:
|
|
f.write("test content")
|
|
temp_path = f.name
|
|
|
|
try:
|
|
result = process_file(temp_path)
|
|
assert result == "processed: test content"
|
|
finally:
|
|
os.unlink(temp_path)
|
|
```
|
|
|
|
### Testing with pytest's tmp_path Fixture
|
|
|
|
```python
|
|
def test_with_tmp_path(tmp_path):
|
|
"""Test using pytest's built-in temp path fixture."""
|
|
test_file = tmp_path / "test.txt"
|
|
test_file.write_text("hello world")
|
|
|
|
result = process_file(str(test_file))
|
|
assert result == "hello world"
|
|
# tmp_path automatically cleaned up
|
|
```
|
|
|
|
### Testing with tmpdir Fixture
|
|
|
|
```python
|
|
def test_with_tmpdir(tmpdir):
|
|
"""Test using pytest's tmpdir fixture."""
|
|
test_file = tmpdir.join("test.txt")
|
|
test_file.write("data")
|
|
|
|
result = process_file(str(test_file))
|
|
assert result == "data"
|
|
```
|
|
|
|
## Test Organization
|
|
|
|
### Directory Structure
|
|
|
|
```
|
|
tests/
|
|
├── conftest.py # Shared fixtures
|
|
├── __init__.py
|
|
├── unit/ # Unit tests
|
|
│ ├── __init__.py
|
|
│ ├── test_models.py
|
|
│ ├── test_utils.py
|
|
│ └── test_services.py
|
|
├── integration/ # Integration tests
|
|
│ ├── __init__.py
|
|
│ ├── test_api.py
|
|
│ └── test_database.py
|
|
└── e2e/ # End-to-end tests
|
|
├── __init__.py
|
|
└── test_user_flow.py
|
|
```
|
|
|
|
### Test Classes
|
|
|
|
```python
|
|
class TestUserService:
|
|
"""Group related tests in a class."""
|
|
|
|
@pytest.fixture(autouse=True)
|
|
def setup(self):
|
|
"""Setup runs before each test in this class."""
|
|
self.service = UserService()
|
|
|
|
def test_create_user(self):
|
|
"""Test user creation."""
|
|
user = self.service.create_user("Alice")
|
|
assert user.name == "Alice"
|
|
|
|
def test_delete_user(self):
|
|
"""Test user deletion."""
|
|
user = User(id=1, name="Bob")
|
|
self.service.delete_user(user)
|
|
assert not self.service.user_exists(1)
|
|
```
|
|
|
|
## Best Practices
|
|
|
|
### DO
|
|
|
|
- **Follow TDD**: Write tests before code (red-green-refactor)
|
|
- **Test one thing**: Each test should verify a single behavior
|
|
- **Use descriptive names**: `test_user_login_with_invalid_credentials_fails`
|
|
- **Use fixtures**: Eliminate duplication with fixtures
|
|
- **Mock external dependencies**: Don't depend on external services
|
|
- **Test edge cases**: Empty inputs, None values, boundary conditions
|
|
- **Aim for 80%+ coverage**: Focus on critical paths
|
|
- **Keep tests fast**: Use marks to separate slow tests
|
|
|
|
### DON'T
|
|
|
|
- **Don't test implementation**: Test behavior, not internals
|
|
- **Don't use complex conditionals in tests**: Keep tests simple
|
|
- **Don't ignore test failures**: All tests must pass
|
|
- **Don't test third-party code**: Trust libraries to work
|
|
- **Don't share state between tests**: Tests should be independent
|
|
- **Don't catch exceptions in tests**: Use `pytest.raises`
|
|
- **Don't use print statements**: Use assertions and pytest output
|
|
- **Don't write tests that are too brittle**: Avoid over-specific mocks
|
|
|
|
## Common Patterns
|
|
|
|
### Testing API Endpoints (FastAPI/Flask)
|
|
|
|
```python
|
|
@pytest.fixture
|
|
def client():
|
|
app = create_app(testing=True)
|
|
return app.test_client()
|
|
|
|
def test_get_user(client):
|
|
response = client.get("/api/users/1")
|
|
assert response.status_code == 200
|
|
assert response.json["id"] == 1
|
|
|
|
def test_create_user(client):
|
|
response = client.post("/api/users", json={
|
|
"name": "Alice",
|
|
"email": "alice@example.com"
|
|
})
|
|
assert response.status_code == 201
|
|
assert response.json["name"] == "Alice"
|
|
```
|
|
|
|
### Testing Database Operations
|
|
|
|
```python
|
|
@pytest.fixture
|
|
def db_session():
|
|
"""Create a test database session."""
|
|
session = Session(bind=engine)
|
|
session.begin_nested()
|
|
yield session
|
|
session.rollback()
|
|
session.close()
|
|
|
|
def test_create_user(db_session):
|
|
user = User(name="Alice", email="alice@example.com")
|
|
db_session.add(user)
|
|
db_session.commit()
|
|
|
|
retrieved = db_session.query(User).filter_by(name="Alice").first()
|
|
assert retrieved.email == "alice@example.com"
|
|
```
|
|
|
|
### Testing Class Methods
|
|
|
|
```python
|
|
class TestCalculator:
|
|
@pytest.fixture
|
|
def calculator(self):
|
|
return Calculator()
|
|
|
|
def test_add(self, calculator):
|
|
assert calculator.add(2, 3) == 5
|
|
|
|
def test_divide_by_zero(self, calculator):
|
|
with pytest.raises(ZeroDivisionError):
|
|
calculator.divide(10, 0)
|
|
```
|
|
|
|
## pytest Configuration
|
|
|
|
### pytest.ini
|
|
|
|
```ini
|
|
[pytest]
|
|
testpaths = tests
|
|
python_files = test_*.py
|
|
python_classes = Test*
|
|
python_functions = test_*
|
|
addopts =
|
|
--strict-markers
|
|
--disable-warnings
|
|
--cov=mypackage
|
|
--cov-report=term-missing
|
|
--cov-report=html
|
|
markers =
|
|
slow: marks tests as slow
|
|
integration: marks tests as integration tests
|
|
unit: marks tests as unit tests
|
|
```
|
|
|
|
### pyproject.toml
|
|
|
|
```toml
|
|
[tool.pytest.ini_options]
|
|
testpaths = ["tests"]
|
|
python_files = ["test_*.py"]
|
|
python_classes = ["Test*"]
|
|
python_functions = ["test_*"]
|
|
addopts = [
|
|
"--strict-markers",
|
|
"--cov=mypackage",
|
|
"--cov-report=term-missing",
|
|
"--cov-report=html",
|
|
]
|
|
markers = [
|
|
"slow: marks tests as slow",
|
|
"integration: marks tests as integration tests",
|
|
"unit: marks tests as unit tests",
|
|
]
|
|
```
|
|
|
|
## Running Tests
|
|
|
|
```bash
|
|
# Run all tests
|
|
pytest
|
|
|
|
# Run specific file
|
|
pytest tests/test_utils.py
|
|
|
|
# Run specific test
|
|
pytest tests/test_utils.py::test_function
|
|
|
|
# Run with verbose output
|
|
pytest -v
|
|
|
|
# Run with coverage
|
|
pytest --cov=mypackage --cov-report=html
|
|
|
|
# Run only fast tests
|
|
pytest -m "not slow"
|
|
|
|
# Run until first failure
|
|
pytest -x
|
|
|
|
# Run and stop on N failures
|
|
pytest --maxfail=3
|
|
|
|
# Run last failed tests
|
|
pytest --lf
|
|
|
|
# Run tests with pattern
|
|
pytest -k "test_user"
|
|
|
|
# Run with debugger on failure
|
|
pytest --pdb
|
|
```
|
|
|
|
## Quick Reference
|
|
|
|
| Pattern | Usage |
|
|
|---------|-------|
|
|
| `pytest.raises()` | Test expected exceptions |
|
|
| `@pytest.fixture()` | Create reusable test fixtures |
|
|
| `@pytest.mark.parametrize()` | Run tests with multiple inputs |
|
|
| `@pytest.mark.slow` | Mark slow tests |
|
|
| `pytest -m "not slow"` | Skip slow tests |
|
|
| `@patch()` | Mock functions and classes |
|
|
| `tmp_path` fixture | Automatic temp directory |
|
|
| `pytest --cov` | Generate coverage report |
|
|
| `assert` | Simple and readable assertions |
|
|
|
|
**Remember**: Tests are code too. Keep them clean, readable, and maintainable. Good tests catch bugs; great tests prevent them.
|