@samira-el-masri
Joined December 2025
200
Public prompts
4829
Stars received
Complete guide for deploying self-hosted LLMs covering infrastructure setup, model optimization, operations, and TCO analysis.
Integrate knowledge graphs with RAG systems for enhanced retrieval through entity linking, graph traversal, and relationship-aware reasoning.
Implement comprehensive LLM guardrails covering PII, toxicity, topic restrictions, and compliance with configurable rules and audit logging.
Build a production prompt template engine with variable substitution, conditionals, loops, inheritance, and comprehensive validation.
Design a multi-tenant RAG architecture with data isolation, resource management, tenant customization, and cost attribution capabilities.
Design a high-volume async inference queue with priority lanes, auto-scaling workers, multiple delivery mechanisms, and comprehensive observability.
Build an automated multi-dimensional quality scorer for LLM responses with LLM-as-judge and calibration against human labels.
Create a comprehensive embedding model fine-tuning plan covering data preparation, training configuration, evaluation, and deployment strategies.
Design an agentic RAG architecture with planning, execution, memory, and synthesis layers enabling multi-step reasoning and tool use.
Optimize GPU memory usage for LLM inference through quantization, batching, KV cache management, and attention optimizations with detailed calculations.
Design a multi-layer prompt injection defense system with input sanitization, prompt structure hardening, output validation, and attack monitoring.
Build a scalable document ingestion pipeline with extraction, chunking, embedding generation, and vector storage with parallel processing and error recovery.