Emotional Stimuli and Beyond: Prompt Engineering Strategies for LLM Performance

Prompt engineering research banner
A lightweight summary page; see the paper for all experimental details.

Overview

Large Language Models (LLMs) are extremely sensitive to how they’re prompted. This project studies a family of prompt engineering strategies—from neutral baselines to emotionally charged framings—to understand how subtle language choices influence task performance. We emphasize practical techniques developers can use without model training or fine-tuning.

Prompting Strategies Studied

Tasks & Metrics (High Level)

We evaluate across representative task types (classification, extraction, and short‑form reasoning). Primary metrics include accuracy and F1—particularly useful when label imbalance or partial extraction correctness matters. For generative answers, we apply rubric‑style checks, and when applicable, automatic exact/partial matches.

Key Findings (at a glance)

Failure Modes & Risks

What We’d Do Next

This page is a concise summary. For the full methodology, ablations, and complete results, see the paper.

Let’s Chat.