top of page

Zeyneb N. Kaya

patched.gif

Hi! I am Zeyneb, a student at Stanford University. I'm broadly interested in understanding and pushing the limits of ML, working on robustness/adaptation, learning from data (efficiently), statistics, and physics---among other things. 

Most recently, I've worked on models for physical optimization as co-founder @ Topological; decentralized AI, synthetic data, and midtraining @ Dria; and RL/self-improvement/diffusion LLMs @ SAIL. 

 

I’m always eager to discuss interesting ideas and opportunities—please reach out!​ 

zeynebnk [at] stanford [dot] edu

github / x / linkedin / curius / writing

Research.

My work aims to advance our understanding of AI and its capabilities, and use that to improve them and push their limits in their fundamental challenges. I'm interested in robustness, data/efficiency, and learning dynamics, working in machine learning, physics, and statistics.

 

Listed below are selected relevant publications.

Semantic Anchoring in Large Language Models: Thresholds, Transfer, and Geometry

Edward Y. Chang, Zeyneb N. Kaya, Ethan Chang

Under Review

Measuring the Impact of Data Augmentation Methods for Extremely Low-Resource NMT  

Zeyneb N. Kaya, Annie K. Lamar

Proceedings of the Sixth Workshop on Technologies for Machine Translation of Low-Resource Languages (LoResMT) @ EACL, 2023

MADLIBS: A Novel Multilingual Data Augmentation Algorithm for Low-Resource Neural Machine Translation 

Zeyneb N. Kaya

Regeneron Science Talent Search, 2024 & National Junior Science and Humanities Symposium, 2023

Decoding Large-Language Models: A Systematic Overview of Socio-Technical Impacts, Constraints, and Emerging Questions

Zeyneb N. Kaya, Souvick Ghosh

arXiv preprint

Full Scope Word Embedding Variability for Low-Resource Languages

Zeyneb N. Kaya, Annie K. Lamar

IEEE MIT URTC, 2023

​​

The Pervasiveness of Language Contact: Evidence from Negative Existentials in Romeyka/Turkish Code-Switching

Zeyneb N. Kaya
Proceedings of the Linguistic Society of America (PLSA), 2023​​

​​

An Artificial Intelligence Model on Rheumatology: Interpretation of the Sacroiliac Joint Graphy in Ankylosing Spondylitis

Ahmet C. Genc, Zeyneb N. Kaya, et al
Annals of the Rheumatic Diseases, 2021  

New Piskel (3)_2.gif

Awards & Recognition.

Etched x Mercor x Cognition Hackathon – 1st Place/$40K Winner 2025
Regeneron Science Talent Search Winner5th Place/$90K Winner 2024
Coca Cola Scholar – 2024

PearVC x Anthropic Hackathon – 1st Place/Most Technical Winner, 2025

TreeHacks Scrapybara Prize – 1st Place/$16K-valued Winner, 2025

Geoguessr – Master Tier Player, 2025

Olympiad in Linguistics (Onling) – 10th Place / 1st in USA, 2023

North American Computational Linguistics Olympiad (NACLO) – Finalist, 2023

International Olympiad in Artificial Intelligence (IOAI) – Team USA invited representative (did not attend due to conflicts)

Education.

Stanford University
Computer Science (AI)
 /

Minor in Mathematics

Saratoga High School

+ Dual Enrolled West Valley College

Relevant Coursework: ML; Deep NLP; Deep RL; Probability & Stochastic DiffEqs; Matrix Theory; AI & Language; AI for Reasoning; Statistical Mechanics of Computation.

   

AI Club Co-President. Linguistics Club Founder + President, Chinese Club Events Coordinator. 

Dual Enrollment: Differential Equations, Linear Algebra, Multivariable Calculus, Cultural Anthropology

Projects.

LLaDA-R1

MADLIBS

In-Context Learning of Transformers: A Statistical Mechanics Lens

SHIELD.

Language Models (can be)

Few-Shot Fakers

NeuroPilot

New Piskel (9).gif

Created LLaDA-R1, a diffusion LLM optimized for reasoning and efficiency at inference time with SFT+RL for dynamic diffusion step adaptation and remasking refinement. ​

@ Mercor x Etched x Cognition Inference-Time Compute Hackathon 2025

Designed Multilingual Augmentation of Data with Alignment-Based Substitution, an efficient multilingual synthetic data generation algorithm achieving SOTA performance with less data.​ 

@ Regeneron Science Talent Search 2024 

Investigated statistical physics models explaining in-context learning; applying spin glasses, random matrix theory, and phase transitions towards transformer interpretability.

@ APPPHYS 229 2025

Built SHIELD., a multi-agent RL + tool use framework for automatic identification and remediation of system vulnerabilities. ​

@ Pear VC x Anthropic Hackathon 2025

Investigated CoT faithfulness & the role of memorization; Implemented corrupted CoT RL approach. 

@ Anthropic Alignment Research Hackathon 2025

Built brain-computer-interface and agentic AI system for brain-powered natural language commands for hands-free computer control. 

@ TreeHacks 2025

bottom of page