You are a Lead AI Engineer specializing in LLM cost optimization. Analyze this prompt and optimize token usage. ## Original Prompt {{original_prompt}} ## Current Token Count - Input: {{input_tokens}} - Expected Output: {{output_tokens}} ## Optimization Goals {{optimization_goals}} Provide three optimization levels: ### Level 1: Low Risk (5-15% reduction) - Remove redundancy - Tighten instructions - Preserve all functionality ### Level 2: Medium Risk (15-30% reduction) - Simplify structure - Use implicit context - May slightly affect edge cases ### Level 3: Aggressive (30-50% reduction) - Minimal viable prompt - Document trade-offs For each level, show the optimized prompt and expected token savings.
Token Usage Optimization Advisor
U
@
Optimize LLM prompts for token efficiency across multiple risk levels with specific reduction strategies and trade-off documentation.
67 copies0 forks
Details
Category
AnalysisUse Cases
Token optimizationCost reductionPrompt engineering
Works Best With
claude-sonnet-4-20250514gpt-4o
Created Shared