Where to Show Demos in Your Prompt: A Positional Bias of In-Context Learning
Kwesi A. Cobbina, Tianyi Zhou
EMNLP 2025 2025
PhD Candidate in Computer Science · University of Maryland, College Park
Advised by Prof. Tianyi Zhou and Prof. Tom Goldstein
I am a PhD candidate in Computer Science at the University of Maryland, advised by Prof. Tianyi Zhou and Prof. Tom Goldstein. My research sits at the intersection of natural language processing, multimodal learning, and optimization. I study how large language models encode and leverage in-context information — from positional biases in demonstration placement to efficient reasoning trace compression — with the goal of making LLMs more robust, efficient, and equitable.
College Park, MD · PhD expected January 2027

Advised by Prof. Tianyi Zhou and Prof. Tom Goldstein at University of Maryland, College Park
Paper accepted at EMNLP 2025: Where to Show Demos in Your Prompt: A Positional Bias of In-Context Learning.
Paper accepted at NeurIPS 2025: ColorBench: Towards Evaluation on Color Cognition Capabilities of LLMs.
Awarded M.Sc. in Computer Science from the University of Maryland, College Park.
Paper accepted at NAACL 2025: My LLM Might Mimic AAE—But When Should It?
Paper accepted at IEEE VIS 2024: Evaluating the Semantic Profiling Abilities of LLMs for Natural Language Utterances.
Received the Dean's Fellowship from the University of Maryland (2021–2023).
Joined the University of Maryland as a Graduate Research Assistant under Prof. Tianyi Zhou.
* equal contribution
Kwesi A. Cobbina, Tianyi Zhou
EMNLP 2025 2025
Yijun Liang, Ming Li, Chenrui Fan, Ziyue Li, Dang Nguyen, Kwesi A. Cobbina, Shweta Bhardwaj, Jiuhai Chen, Fuxiao Liu, Tianyi Zhou
NeurIPS 2025 2025
Sandra C. Sandoval*, Christabel Acquaye*, Kwesi A. Cobbina*, Mohammad N. Teli, Hal Daumé III
NAACL 2025 2025— * Equal contribution
Brandon C. Colelough, Davis Bartels, Dina Demner-Fushman, Ishan Tamrakar, Kwesi Cobbina, Mike Ledford, Srividya Ponnada, Xinchen Yang, Yuexi Chen
IEEE TNNLS 2026— Journal article under review
A test-time meta-learning approach that steers model behavior via in-context vector arithmetic, offering a practical alternative to few-shot ICL under strict context budgets. Includes a reproducible codebase and open-source toolkit (in progress).
An analysis suite for quantifying LLM sensitivity to demonstration ordering and placement in prompts. Supports large-scale evaluations across multiple model families and tasks, with distilled prompt-design recommendations.
A cross-platform mobile app enabling churches to stream services live, schedule events, and engage communities in a single unified workflow. Features push notifications, content management, and social sharing.
Teaching Assistant · University of Maryland
Supported course delivery for an upper-division undergraduate ML course covering supervised learning, neural networks, probabilistic models, and model evaluation. Held weekly office hours, graded assignments and exams, and mentored students on course projects.
Teaching Assistant · University of Maryland
Assisted in a graduate-level NLP course covering language models, parsing, information extraction, and text classification. Developed and graded programming assignments; led discussion sections on transformer architectures.
Curriculum Designer & Instructor · ELiTE (Emerging Labs)
Designed and delivered hands-on AI/ML curriculum for student camps and instructor training programs. Topics covered Python fundamentals, ML workflows (data processing → training → evaluation), and real-world model deployment.
I am happy to discuss research collaborations, speaking opportunities, or any questions about my work. The best way to reach me is by email.
kcobbina@umd.eduUniversity of Maryland, College Park · College Park, MD