Files
n8n/packages/testing/playwright
2026-01-26 13:38:51 +00:00
..

Playwright E2E Test Guide

Development setup

pnpm install-browsers:local # in playwright directory
pnpm build:docker # from root first to test against local changes

Quick Start

pnpm test:all                 									# Run all tests (fresh containers, pnpm build:docker from root first to ensure local containers)
pnpm test:local           											# Starts a local server and runs the E2E tests
N8N_BASE_URL=localhost:5068 pnpm test:local			# Runs the E2E tests against the instance running

Separate Backend and Frontend URLs

When developing with separate backend and frontend servers (e.g., backend on port 5680, frontend on port 8080), you can use the following environment variables:

  • N8N_BASE_URL: Backend server URL (also used as frontend URL if N8N_EDITOR_URL is not set)
  • N8N_EDITOR_URL: Frontend server URL (when set, overrides frontend URL while backend uses N8N_BASE_URL)

How it works:

  • Backend URL (for API calls): Always uses N8N_BASE_URL
  • Frontend URL (for browser navigation): Uses N8N_EDITOR_URL if set, otherwise falls back to N8N_BASE_URL

This allows you to:

  • Test against a backend on port 5680 while the frontend dev server runs on port 8080
  • Use different URLs for API calls vs browser navigation
  • Maintain backward compatibility with single-URL setups

Test Commands

# By Mode
pnpm test:container:sqlite      # SQLite (default)
pnpm test:container:postgres    # PostgreSQL
pnpm test:container:queue       # Queue mode
pnpm test:container:multi-main  # HA setup

pnpm test:performance						# Runs the performance tests against Sqlite container
pnpm test:chaos									# Runs the chaos tests


# Development
pnpm test:all --grep "workflow"           # Pattern match, can run across all test types E2E/cli-workflow/performance
pnpm test:local --ui            # To enable UI debugging and test running mode

Test Tags

test('basic test', ...)                              // All modes, fully parallel
test('postgres only @mode:postgres', ...)            // Mode-specific
test('chaos test @mode:multi-main @chaostest', ...) // Isolated per worker
test('cloud resource test @cloud:trial', ...)       // Cloud resource constraints
test('proxy test @capability:proxy', ...)           // Requires proxy server capability
test('enterprise feature @licensed', ...)           // Requires enterprise license (container-only)

Tag Reference

Tag Description When to Use
@mode:X Infrastructure mode (postgres, queue, multi-main) Tests requiring specific DB or architecture
@capability:X Container services (email, proxy, oidc, source-control, observability) Tests needing external services
@licensed Enterprise license features Tests for features behind license flags at startup
@cloud:X Resource constraints (trial, enterprise) Performance tests with memory/CPU limits
@chaostest Chaos engineering tests Tests that intentionally break things
@db:reset ⚠️ Deprecated - use test.use() pattern Legacy isolation pattern

Worker Isolation (Fresh Database)

If tests need a clean database state, use test.use() at the top level of the file with a unique capability config instead of the deprecated @db:reset tag:

// my-isolated-tests.spec.ts
import { test, expect } from '../fixtures/base';

// Must be top-level, not inside describe block
test.use({ capability: { env: { _ISOLATION: 'my-isolated-tests' } } });

test('test with clean state', async ({ n8n }) => {
  // Fresh container with reset database
});

Deprecated: @db:reset tag causes CI issues (separate workers, sequential execution). Use test.use() pattern above instead.

Enterprise Features (@licensed)

Use the @licensed tag for tests that require enterprise features which are only available when the license is present at startup. This differs from features that can be enabled/disabled at runtime.

When to use:

  • Features behind @BackendModule({ licenseFlag: LICENSE_FEATURES.X }) decorators
  • API endpoints that only exist when the module loads with a valid license
  • Features like log streaming, SSO, LDAP where routes aren't registered without license

Example:

// The @licensed tag ensures this only runs in container mode with a valid license
test.describe('Log Streaming @licensed', () => {
  test.beforeEach(async ({ n8n }) => {
    // enableFeature() works for runtime checks, but module must be loaded first
    await n8n.api.enableFeature('logStreaming');
  });

  test('should show licensed view', async ({ n8n }) => {
    await n8n.navigate.toLogStreaming();
    // ...
  });
});

Note: @licensed tests are skipped in local mode (test:local) and only run in container mode where a license is available.

Enterprise license for testing: To run @licensed tests or manually test enterprise features, set N8N_LICENSE_TENANT_ID and N8N_LICENSE_ACTIVATION_KEY in your environment. The containers package reads these variables automatically. Ask in Slack for the sandbox license key.

Fixture Selection

  • base.ts: Standard testing with worker-scoped containers (default choice)
  • cloud-only.ts: Cloud resource testing with guaranteed isolation
    • Use for performance testing under resource constraints
    • Requires @cloud:* tags (@cloud:trial, @cloud:enterprise, etc.)
    • Creates only cloud containers, no worker containers
// Standard testing
import { test, expect } from '../fixtures/base';

// Cloud resource testing
import { test, expect } from '../fixtures/cloud-only';
test('Performance under constraints @cloud:trial', async ({ n8n, api }) => {
  // Test runs with 384MB RAM, 250 millicore CPU
});

Tips

  • test:* commands use fresh containers (for testing)
  • VS Code: Set N8N_BASE_URL in Playwright settings to run tests directly from VS Code
  • Pass custom env vars via N8N_TEST_ENV='{"KEY":"value"}'

Project Layout

  • composables: Multi-page interactions (e.g., WorkflowComposer.executeWorkflowAndWaitForNotification())
  • config: Test setup and configuration (constants, test users, etc.)
  • fixtures: Custom test fixtures extending Playwright's base test
    • base.ts: Standard fixtures with worker-scoped containers
    • cloud-only.ts: Cloud resource testing with test-scoped containers only
  • pages: Page Object Models for UI interactions
  • services: API helpers for E2E controller, REST calls, workflow management, etc.
  • utils: Utility functions (string manipulation, helpers, etc.)
  • workflows: Test workflow JSON files for import/reuse

Writing Tests with Proxy

You can use ProxyServer to mock API requests.

import { test, expect } from '../fixtures/base';

// The `@capability:proxy` tag ensures tests only run when proxy infrastructure is available.
test.describe('Proxy tests @capability:proxy', () => {
  test('should mock HTTP requests', async ({ proxyServer, n8n }) => {
    // Create mock expectations
    await proxyServer.createGetExpectation('/api/data', { result: 'mocked' });

    // Execute workflow that makes HTTP requests
    await n8n.canvas.openNewWorkflow();
    // ... test implementation

    // Verify requests were proxied
    expect(await proxyServer.wasGetRequestMade('/api/data')).toBe(true);
  });
});

Recording and replaying requests

The ProxyServer service supports recording HTTP requests for test mocking and replay. All proxied requests are automatically recorded by the mock server as described in the Mock Server documentation.

Recording Expectations

// Record all requests (the request is simplified/cleansed to method/path/body/query)
await proxyServer.recordExpectations('test-folder');

// Record with filtering and options
await proxyServer.recordExpectations('test-folder', {
  host: 'googleapis.com',           // Filter by host (partial match)
  dedupe: true,                     // Remove duplicate requests
  raw: false                        // Save cleaned requests (default)
});

// Record raw requests with all headers and metadata
await proxyServer.recordExpectations('test-folder', {
  raw: true                         // Save complete original requests
});

// Record requests matching specific criteria
await proxyServer.recordExpectations('test-folder', {
  pathOrRequestDefinition: {
    method: 'POST',
    path: '/api/workflows'
  }
});

Loading and Using Recorded Expectations

Recorded expectations are saved as JSON files in the expectations/ directory. To use them in tests, you must explicitly load them:

test('should use recorded expectations', async ({ proxyServer }) => {
  // Load expectations from a specific folder
  await proxyServer.loadExpectations('test-folder');

  // Your test code here - requests will be mocked using loaded expectations
});

Important: Cleanup Expectations

Remember to clean up expectations before or after test runs:

test.beforeEach(async ({ proxyServer }) => {
  // Clear any existing expectations before test
  await proxyServer.clearAllExpectations();
});

test.afterEach(async ({ proxyServer }) => {
  // Or clear expectations after test
  await proxyServer.clearAllExpectations();
});

This prevents expectations from one test affecting others and ensures test isolation.

Debugging

Keepalive Mode

Use N8N_CONTAINERS_KEEPALIVE=true to keep containers running after tests complete. Useful for:

  • Inspecting n8n instance state after a failure
  • Exploring configured integrations (email, OIDC, source control)
  • Manual testing against a pre-configured environment
N8N_CONTAINERS_KEEPALIVE=true pnpm test:container:sqlite --grep "@capability:email" --workers 1

After tests complete, connection details are printed:

=== KEEPALIVE: Containers left running for debugging ===
    URL: http://localhost:54321
    Project: n8n-stack-abc123
    Cleanup: pnpm --filter n8n-containers stack:clean:all
=========================================================

Clean up when done: pnpm --filter n8n-containers stack:clean:all

Victoria Export on Failure

When tests fail with observability enabled, logs and metrics are automatically exported as Currents attachments:

Attachment Description
container-logs Human-readable logs grouped by container
victoria-logs-export.jsonl Raw logs in JSON Lines format
victoria-metrics-export.jsonl All metrics in JSON Lines format

Importing into a local Victoria instance

  1. Download the .jsonl attachments from Currents
  2. Import into running Victoria containers (e.g., from keepalive mode):
node scripts/import-victoria-data.mjs victoria-metrics-export.jsonl victoria-logs-export.jsonl

Or start standalone containers first with --start:

node scripts/import-victoria-data.mjs --start victoria-metrics-export.jsonl victoria-logs-export.jsonl
  1. Query locally:

Writing Tests

For guidelines on writing new tests, see CONTRIBUTING.md.