Agent SkillsAgent Skills
jeremylongshore

running-e2e-tests

@jeremylongshore/running-e2e-tests
jeremylongshore
1,761
231 forks
Updated 3/31/2026
View on GitHub

Execute end-to-end tests covering full user workflows across frontend and backend. Use when performing specialized testing. Trigger with phrases like "run end-to-end tests", "test user flows", or "execute E2E suite".

Installation

$npx agent-skills-cli install @jeremylongshore/running-e2e-tests
Claude Code
Cursor
Copilot
Codex
Antigravity

Details

Pathplugins/testing/e2e-test-framework/skills/running-e2e-tests/SKILL.md
Branchmain
Scoped Name@jeremylongshore/running-e2e-tests

Usage

After installing, this skill will be available to your AI coding assistant.

Verify installation:

npx agent-skills-cli list

Skill Instructions


name: running-e2e-tests description: | Execute end-to-end tests covering full user workflows across frontend and backend. Use when performing specialized testing. Trigger with phrases like "run end-to-end tests", "test user flows", or "execute E2E suite".

allowed-tools: Read, Write, Edit, Grep, Glob, Bash(test:e2e-*) version: 1.0.0 author: Jeremy Longshore jeremy@intentsolutions.io license: MIT compatible-with: claude-code, codex, openclaw tags: [testing, workflow, e2e-tests]


E2E Test Framework

Current State

!cat package.json 2>/dev/null | grep -oE 'playwright|cypress|selenium' || echo 'No E2E framework detected'

Overview

Execute end-to-end tests that simulate real user workflows across the full application stack -- browser interactions, API calls, database operations, and third-party integrations. Supports Playwright (recommended), Cypress, Selenium, and Puppeteer.

Prerequisites

  • E2E testing framework installed (Playwright, Cypress, or Selenium WebDriver)
  • Application running in a test environment with seeded test data
  • Browser binaries installed (npx playwright install or Cypress binary)
  • Test user accounts created with known credentials
  • Environment variables configured for base URL, API keys, and test credentials

Instructions

  1. Identify critical user journeys to cover:
    • User registration and login flow.
    • Primary feature workflow (e.g., create item, edit, delete).
    • Search and filtering functionality.
    • Checkout or payment flow (if applicable).
    • Error handling (404 pages, form validation, session expiry).
  2. Create page object models (POM) for reusable page interactions:
    • One class per page or major component.
    • Encapsulate locators, actions (click, fill, select), and assertions.
    • Use data-testid attributes as primary selectors for stability.
  3. Write E2E test files organized by user journey:
    • Each test file covers one complete workflow.
    • Use beforeEach to navigate to the starting page and reset state.
    • Use afterEach to capture screenshots on failure.
    • Keep tests independent -- no test should depend on another test's output.
  4. Handle authentication efficiently:
    • Store authenticated session state to a file (storageState in Playwright).
    • Reuse session across tests that require login.
    • Create a separate auth setup fixture that runs once per worker.
  5. Configure multi-browser and responsive testing:
    • Run tests on Chromium, Firefox, and WebKit.
    • Test at mobile (375px), tablet (768px), and desktop (1280px) viewports.
    • Use Playwright projects to define browser/viewport combinations.
  6. Add retry and stability mechanisms:
    • Use expect with auto-waiting locators (Playwright) instead of explicit waits.
    • Configure test retries (max 2) for CI environments.
    • Add networkidle or domcontentloaded wait conditions for page transitions.
  7. Generate test reports with screenshots, traces, and video on failure.

Output

  • E2E test files organized by user journey in tests/e2e/ or e2e/
  • Page object model classes in tests/e2e/pages/
  • Playwright/Cypress configuration file with browser and viewport matrix
  • Authentication state file for session reuse
  • HTML test report with screenshots, traces, and failure details

Error Handling

ErrorCauseSolution
Element not found / timeoutSelector changed or element lazy-loaded after timeoutUse data-testid attributes; increase timeout; use waitFor with proper state checks
Test passes locally but fails in CIHeadless browser behavior differs or CI is slowerRun CI in headless mode locally to reproduce; increase timeouts; check viewport size
Authentication state expiredStored session tokens have short TTLRegenerate auth state before each test run; use long-lived test account tokens
Flaky test due to animationClick registered before animation completesDisable CSS animations in test config; use force: true on click; add waitForLoadState
Database state pollutionPrevious test left data that affects current testSeed database in beforeEach; use transactional rollback; reset via API endpoint

Examples

Playwright test for user registration flow:

import { test, expect } from '@playwright/test';

test('new user can register and see dashboard', async ({ page }) => {
  await page.goto('/register');
  await page.getByTestId('name-input').fill('Test User');
  await page.getByTestId('email-input').fill('test@example.com');
  await page.getByTestId('password-input').fill('SecurePass123!');
  await page.getByTestId('register-button').click();

  await expect(page).toHaveURL(/\/dashboard/);
  await expect(page.getByTestId('welcome-message')).toContainText('Test User');
});

Page object model:

export class LoginPage {
  constructor(private page: Page) {}
  async login(email: string, password: string) {
    await this.page.goto('/login');
    await this.page.getByTestId('email').fill(email);
    await this.page.getByTestId('password').fill(password);
    await this.page.getByTestId('submit').click();
    await this.page.waitForURL(/\/dashboard/);
  }
}

Playwright config with multi-browser projects:

export default defineConfig({
  projects: [
    { name: 'chromium', use: { ...devices['Desktop Chrome'] } },
    { name: 'firefox', use: { ...devices['Desktop Firefox'] } },
    { name: 'mobile', use: { ...devices['iPhone 14'] } },
  ],
  use: { screenshot: 'only-on-failure', trace: 'on-first-retry' },
});

Resources

More by jeremylongshore

View all
vertex-agent-builder
1,768

Build and deploy production-ready generative AI agents using Vertex AI, Gemini models, and Google Cloud infrastructure with RAG, function calling, and multi-modal capabilities. Use when appropriate context detected. Trigger with relevant phrases based on skill purpose.

gcp-examples-expert
1,768

Generate production-ready Google Cloud code examples from official repositories including ADK samples, Genkit templates, Vertex AI notebooks, and Gemini patterns. Use when asked to "show ADK example" or "provide GCP starter kit". Trigger with relevant phrases based on skill purpose.

genkit-production-expert
1,768

Build production Firebase Genkit applications including RAG systems, multi-step flows, and tool calling for Node.js/Python/Go. Deploy to Firebase Functions or Cloud Run with AI monitoring. Use when asked to "create genkit flow" or "implement RAG". Trigger with relevant phrases based on skill purpose.

validator-expert
1,768

Validate production readiness of Vertex AI Agent Engine deployments across security, monitoring, performance, compliance, and best practices. Generates weighted scores (0-100%) with actionable remediation plans. Use when asked to validate a deployment, run a production readiness check, audit security posture, or verify compliance for Vertex AI agents. Trigger with "validate deployment", "production readiness", "security audit", "compliance check", "is this agent ready for prod", "check my ADK agent", "review before deploy", or "production readiness check". Make sure to use this skill whenever validating ADK agents for Agent Engine.