Analyze token usage patterns to identify optimization opportunities. ## Usage Data {{usage_data}} ## Cost Goals {{cost_goals}} ## Quality Requirements {{quality_requirements}} Perform usage analysis: ```python class TokenUsageAnalyzer: def analyze_distribution(self, usage_data: List[UsageRecord]) -> Distribution: """ Analyze: - Input vs output token ratio - Token count by endpoint - Token count by user/team - Temporal patterns """ pass def identify_waste(self, records: List[UsageRecord]) -> List[WasteOpportunity]: """ Find: - Repetitive prompts - Oversized contexts - Low-value requests - Retry waste """ pass def recommend_optimizations(self, analysis: UsageAnalysis) -> List[Recommendation]: """Suggest cost reduction strategies""" pass ``` Include: - Visualization of patterns - Savings estimates - Implementation priorities - Monitoring dashboards
LLM Token Usage Analyzer
Analyze LLM token usage patterns to identify waste and optimization opportunities with savings estimates and implementation priorities.
28 copies0 forks
Share this prompt:
Details
Category
CodingUse Cases
Token analysisCost optimizationUsage patterns
Works Best With
claude-sonnet-4-20250514gpt-4o
Created Updated Shared