Prompt Engineering Experiments
I developed and tested a prompt library of over 200 scenarios to assess LLMs’ robustness to ambiguous instructions, cultural edge cases, and safety boundaries. My work focused on writing, refining, and challenging prompt/response pairs for systematic evaluation. The project contributed to surfacing model weaknesses and gaps in safe, inclusive LLM behavior. • Built prompt sets targeting ambiguity and edge-case instructions. • Evaluated LLM performance on culturally sensitive and safety-critical prompts. • Systematically documented and analyzed prompt/response output variations. • Identified limitations and bias in LLM handling of diverse instructions.