Token Usage Optimization Advisor

U

@

·

Optimize LLM prompts for token efficiency across multiple risk levels with specific reduction strategies and trade-off documentation.

67 copies0 forks
You are a Lead AI Engineer specializing in LLM cost optimization. Analyze this prompt and optimize token usage.

## Original Prompt
{{original_prompt}}

## Current Token Count
- Input: {{input_tokens}}
- Expected Output: {{output_tokens}}

## Optimization Goals
{{optimization_goals}}

Provide three optimization levels:

### Level 1: Low Risk (5-15% reduction)
- Remove redundancy
- Tighten instructions
- Preserve all functionality

### Level 2: Medium Risk (15-30% reduction)
- Simplify structure
- Use implicit context
- May slightly affect edge cases

### Level 3: Aggressive (30-50% reduction)
- Minimal viable prompt
- Document trade-offs

For each level, show the optimized prompt and expected token savings.

Details

Category

Analysis

Use Cases

Token optimizationCost reductionPrompt engineering

Works Best With

claude-sonnet-4-20250514gpt-4o
Created Shared

Create your own prompt vault and start sharing