Execute software engineer specializing in creating production-ready ADK agents with best practices, code structure, testing, and deployment automation. Use when asked to "build ADK agent", "create agent code", or "engineer ADK application". Trigger with relevant phrases based on skill purpose.
Installation
Details
Usage
After installing, this skill will be available to your AI coding assistant.
Verify installation:
skills listSkill Instructions
name: adk-engineer description: | Execute software engineer specializing in creating production-ready ADK agents with best practices, code structure, testing, and deployment automation. Use when asked to "build ADK agent", "create agent code", or "engineer ADK application". Trigger with relevant phrases based on skill purpose. allowed-tools: Read, Write, Edit, Grep, Glob, Bash(cmd:*) version: 1.0.0 author: Jeremy Longshore jeremy@intentsolutions.io license: MIT
ADK Engineer
Engineer production-ready Agent Development Kit (ADK) agents and multi-agent systems: clean structure, testability, safe tool usage, and deployment automation.
Overview
Use this skill to design and implement ADK agent code that is maintainable and shippable: clear module boundaries, structured tool interfaces, regression tests, and a deployment checklist (local or Agent Engine).
Prerequisites
- A target runtime (Python/Java/Go) consistent with the project’s pinned versions
- ADK installed (and any required model/provider SDKs configured)
- A test runner available in the repo (unit tests at minimum)
- If deploying: access to a Google Cloud project and permissions for the chosen deployment target
Instructions
- Clarify requirements: agent goals, tool surface, latency/cost constraints, and deployment target.
- Propose architecture: single agent vs multi-agent, orchestration pattern, state strategy (Memory Bank / external store).
- Scaffold structure: agent entrypoint(s), tool modules, config, and tests.
- Implement incrementally:
- add one tool at a time with input validation and structured outputs
- add regression tests for each tool and critical prompt flows
- Add operational guardrails: retries/backoff, timeouts, logging, and safe error messages.
- Validate locally (tests + smoke prompts) and provide a deployment plan (when requested).
Output
- A concrete architecture plan and file layout
- Agent and tool implementations (or patches) with tests
- A validation checklist (commands to run, expected outputs, and failure triage)
- Optional: deployment instructions and post-deploy health checks
Error Handling
- Build/test failures: isolate the failing module, minimize the repro, fix, and add a regression test.
- Tool/runtime errors: enforce structured error responses and safe retries where appropriate.
- Deployment failures: provide the exact failing command, logs to inspect, and least-privilege IAM fixes.
Examples
Example: Productionizing an existing ADK agent
- Request: “Refactor this agent into a clean module structure and add tests before we deploy.”
- Result: reorganized
src/layout, tool boundaries, a test suite, and a deployment checklist.
Example: Multi-agent workflow
- Request: “Build a validator + deployer + monitor agent team with a sequential orchestrator.”
- Result: orchestrator skeleton, per-agent responsibilities, and smoke tests for each step.
Resources
- Full detailed playbook (kept for reference):
{baseDir}/references/SKILL.full.md - Repo standards (source of truth):
000-docs/6767-a-SPEC-DR-STND-claude-code-plugins-standard.md000-docs/6767-b-SPEC-DR-STND-claude-skills-standard.md
- ADK / Agent Engine docs: https://cloud.google.com/vertex-ai/docs/agent-engine
More by jeremylongshore
View allRabbitmq Queue Setup - Auto-activating skill for Backend Development. Triggers on: rabbitmq queue setup, rabbitmq queue setup Part of the Backend Development skill category.
evaluating-machine-learning-models: This skill allows Claude to evaluate machine learning models using a comprehensive suite of metrics. It should be used when the user requests model performance analysis, validation, or testing. Claude can use this skill to assess model accuracy, precision, recall, F1-score, and other relevant metrics. Trigger this skill when the user mentions "evaluate model", "model performance", "testing metrics", "validation results", or requests a comprehensive "model evaluation".
building-neural-networks: This skill allows Claude to construct and configure neural network architectures using the neural-network-builder plugin. It should be used when the user requests the creation of a new neural network, modification of an existing one, or assistance with defining the layers, parameters, and training process. The skill is triggered by requests involving terms like "build a neural network," "define network architecture," "configure layers," or specific mentions of neural network types (e.g., "CNN," "RNN," "transformer").
Oauth Callback Handler - Auto-activating skill for API Integration. Triggers on: oauth callback handler, oauth callback handler Part of the API Integration skill category.
