Execute this skill optimizes prompts for large language models (llms) to reduce token usage, lower costs, and improve performance. it analyzes the prompt, identifies areas for simplification and redundancy removal, and rewrites the prompt to be more conci... Use when optimizing performance. Trigger with phrases like 'optimize', 'performance', or 'speed up'.
Installation
Details
Usage
After installing, this skill will be available to your AI coding assistant.
Verify installation:
skills listSkill Instructions
name: optimizing-prompts description: | Execute this skill optimizes prompts for large language models (llms) to reduce token usage, lower costs, and improve performance. it analyzes the prompt, identifies areas for simplification and redundancy removal, and rewrites the prompt to be more conci... Use when optimizing performance. Trigger with phrases like 'optimize', 'performance', or 'speed up'. allowed-tools: Read, Write, Edit, Grep, Glob, Bash(cmd:*) version: 1.0.0 author: Jeremy Longshore jeremy@intentsolutions.io license: MIT
Ai Ml Engineering Pack
This skill provides automated assistance for ai ml engineering pack tasks.
Overview
This skill provides automated assistance for ai ml engineering pack tasks. This skill empowers Claude to refine prompts for optimal LLM performance. It streamlines prompts to minimize token count, thereby reducing costs and enhancing response speed, all while maintaining or improving output quality.
How It Works
- Analyzing Prompt: The skill analyzes the input prompt to identify areas of redundancy, verbosity, and potential for simplification.
- Rewriting Prompt: It rewrites the prompt using techniques like concise language, targeted instructions, and efficient phrasing.
- Suggesting Alternatives: The skill provides the optimized prompt along with an explanation of the changes made and their expected impact.
When to Use This Skill
This skill activates when you need to:
- Reduce the cost of using an LLM.
- Improve the speed of LLM responses.
- Enhance the quality or clarity of LLM outputs by refining the prompt.
Examples
Example 1: Reducing LLM Costs
User request: "Optimize this prompt for cost and quality: 'I would like you to create a detailed product description for a new ergonomic office chair, highlighting its features, benefits, and target audience, and also include information about its warranty and return policy.'"
The skill will:
- Analyze the prompt for redundancies and areas for simplification.
- Rewrite the prompt to be more concise: "Create a product description for an ergonomic office chair. Include features, benefits, target audience, warranty, and return policy."
- Provide the optimized prompt and explain the token reduction achieved.
Example 2: Improving Prompt Performance
User request: "Optimize this prompt for better summarization: 'Please read the following document and provide a comprehensive summary of all the key points, main arguments, supporting evidence, and overall conclusion, ensuring that the summary is accurate, concise, and easy to understand.'"
The skill will:
- Identify areas for improvement in the prompt's clarity and focus.
- Rewrite the prompt to be more direct: "Summarize this document, including key points, arguments, evidence, and the conclusion."
- Present the optimized prompt and explain how it enhances summarization performance.
Best Practices
- Clarity: Ensure the original prompt is clear and well-defined before optimization.
- Context: Provide sufficient context to the skill so it can understand the prompt's purpose.
- Iteration: Iterate on the optimized prompt based on the LLM's output to fine-tune performance.
Integration
This skill integrates with the prompt-architect agent to leverage advanced prompt engineering techniques. It can also be used in conjunction with the llm-integration-expert to optimize prompts for specific LLM APIs.
Prerequisites
- Appropriate file access permissions
- Required dependencies installed
Instructions
- Invoke this skill when the trigger conditions are met
- Provide necessary context and parameters
- Review the generated output
- Apply modifications as needed
Output
The skill produces structured output relevant to the task.
Error Handling
- Invalid input: Prompts for correction
- Missing dependencies: Lists required components
- Permission errors: Suggests remediation steps
Resources
- Project documentation
- Related skills and commands
More by jeremylongshore
View allRabbitmq Queue Setup - Auto-activating skill for Backend Development. Triggers on: rabbitmq queue setup, rabbitmq queue setup Part of the Backend Development skill category.
evaluating-machine-learning-models: This skill allows Claude to evaluate machine learning models using a comprehensive suite of metrics. It should be used when the user requests model performance analysis, validation, or testing. Claude can use this skill to assess model accuracy, precision, recall, F1-score, and other relevant metrics. Trigger this skill when the user mentions "evaluate model", "model performance", "testing metrics", "validation results", or requests a comprehensive "model evaluation".
building-neural-networks: This skill allows Claude to construct and configure neural network architectures using the neural-network-builder plugin. It should be used when the user requests the creation of a new neural network, modification of an existing one, or assistance with defining the layers, parameters, and training process. The skill is triggered by requests involving terms like "build a neural network," "define network architecture," "configure layers," or specific mentions of neural network types (e.g., "CNN," "RNN," "transformer").
Oauth Callback Handler - Auto-activating skill for API Integration. Triggers on: oauth callback handler, oauth callback handler Part of the API Integration skill category.
