Kaitlyn Zhou's Avatar

Kaitlyn Zhou

@kaitlynzhou

Incoming Assistant Professor @cornellbowers.bsky.social Researcher @togetherai.bsky.social Previously @stanfordnlp.bsky.social @ai2.bsky.social @msftresearch.bsky.social https://katezhou.github.io/

832
Followers
317
Following
28
Posts
05.10.2023
Joined
Posts Following

Latest posts by Kaitlyn Zhou @kaitlynzhou

πŸ—žοΈ Pick of the week @fbk-mt.bsky.social: Most LLM dev centers current adopters, so what are we missing? Worth a read!
πŸ‘‡https://arxiv.org/pdf/2510.15951
#NLP #LLMs #HCI

17.02.2026 08:59 πŸ‘ 3 πŸ” 3 πŸ’¬ 0 πŸ“Œ 1

Our #PickOfTheWeek by @bsavoldi.bsky.social: "Attention to Non-Adopters" by @kaitlynzhou.bsky.social, @gligoric.bsky.social, @myra.bsky.social, @mlam.bsky.social, @vyoma-raman.bsky.social, Boluwatife Aminu, Caeley Woo, Michael Brockman, @hannah-cha.bsky.social, @jurafsky.bsky.social (2025).

18.02.2026 09:29 πŸ‘ 4 πŸ” 4 πŸ’¬ 0 πŸ“Œ 0

Paper title!

"Sorry, I Didn't Catch That": How Speech Models Miss What Matters Most

17.02.2026 16:51 πŸ‘ 0 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0
Preview
"Sorry, I Didn't Catch That": How Speech Models Miss What Matters Most Despite speech recognition systems achieving low word error rates on standard benchmarks, they often fail on short, high-stakes utterances in real-world deployments. Here, we study this failure mode i...

Work @togetherai.bsky.social with Federico Bianchi, @mbartelds.bsky.social @jameszou.bsky.social

Paper: arxiv.org/abs/2602.12249
We release two public datasets for evaluation: huggingface.co/datasets/kzh...

17.02.2026 16:47 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Workflow: (1) Select a sample of speech from Common Voice, e.g., Spanish (2) Set the XTTS to generate speech in Spanish (supports 16 languages, excluding English) (3) Clone the voice and generate Spanish but with injected English street names, e.g., "Estoy en ...Washington" (4)  Extract street name speech and manually validate. Repeat this with as many samples as needed to create a unique finetuning dataset.

Workflow: (1) Select a sample of speech from Common Voice, e.g., Spanish (2) Set the XTTS to generate speech in Spanish (supports 16 languages, excluding English) (3) Clone the voice and generate Spanish but with injected English street names, e.g., "Estoy en ...Washington" (4) Extract street name speech and manually validate. Repeat this with as many samples as needed to create a unique finetuning dataset.

To mitigate this harm, we introduce a synthetic data generation approach that produces diverse pronunciations. With fewer than 1,000 synthetic samples, we can improve street name transcription accuracy by nearly 60% (relative to base models).

17.02.2026 16:47 πŸ‘ 1 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

We find that speech models systematically struggle to transcribe named entities like street names, leading to routing errors for all users. But this impact is roughly twice as large for non-English primary speakers as for English primary speakers.

17.02.2026 16:47 πŸ‘ 1 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

As speech models are being deployed in real-world taxi and emergency service settings, the failure to accurately transcribe named entities can cause delays and errors in critical settings.

17.02.2026 16:47 πŸ‘ 2 πŸ” 1 πŸ’¬ 1 πŸ“Œ 0
Preview
CMU LTI Summer 2026 Internship Program Application We are looking for applicants for the Carnegie Mellon University Language Technology Institute's Summer 2026 "Language Technology for All" internship program. The main goal of this internship is to pr...

πŸš€ Apply to CMU LTI’s Summer 2026 β€œLanguage Technology for All” internship! πŸŽ“ Open to pre‑doctoral students new to language tech (non‑CS backgrounds welcome). πŸ”¬ 12–14 weeks in‑person in Pittsburgh β€” travel + stipend paid. πŸ’Έ Deadline: Feb 20, 11:59pm ET. Apply β†’ forms.gle/cUu8g6wb27Hs...

02.02.2026 15:41 πŸ‘ 14 πŸ” 12 πŸ’¬ 2 πŸ“Œ 0

Come work with Federico Bianchi, Yongchan Kwon, Shang Zhu, James Zou, and me on research at the intersection of agent capabilities, alignment, and evaluation.

πŸ“ San Francisco (in-person)
⏱️ 12 weeks, Summer 2026

Link: job-boards.greenhouse.io/togetherai/j...

19.01.2026 19:00 πŸ‘ 3 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0

Internship opportunity! Please share!

πŸ“£ I'm looking to hire an intern in human-centered NLP for the agents team at Together AI. Come work on frontier AI systems that tackle complex agentic tasks!

Research direction is open and looking to publish in NLP and HCI venues!

19.01.2026 19:00 πŸ‘ 13 πŸ” 7 πŸ’¬ 1 πŸ“Œ 0

Please share widely!

My research focuses on human-centered NLP, both in evaluating and training LLMs as well as designing safe and reliable human-LM interactions. More information here!
katezhou.github.io

Application fee waivers can be requested here: gradschool.cornell.edu/admissions/a...

06.11.2025 16:19 πŸ‘ 1 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0
Photo of Cornelll University building surrounded by colorful trees

Photo of Cornelll University building surrounded by colorful trees

No better time to start learning about that #AI thing everyone's talking about...

πŸ“’ I'm recruiting PhD students in Computer Science or Information Science @cornellbowers.bsky.social!

If you're interested, apply to either department (yes, either program!) and list me as a potential advisor!

06.11.2025 16:19 πŸ‘ 23 πŸ” 9 πŸ’¬ 1 πŸ“Œ 0

and @hannah-cha.bsky.social

22.10.2025 22:25 πŸ‘ 0 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0

Together, we’re excited to continue work on developing LLMs for the needs of a broader user audience! This work is done in collaboration with:

@gligoric.bsky.social @myra.bsky.social @mlam.bsky.social @jurafsky.bsky.social

@stanfordnlp.bsky.social @togetherai.bsky.social

21.10.2025 17:12 πŸ‘ 3 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

We outline several strategies known in the HCI literature to elevate non-adopter needs and integrate them into LLM development:
1️⃣ re-balancing data annotation and interaction logs
2️⃣ participatory design for developing evaluations
3️⃣ non-adopter-centered task ideation

21.10.2025 17:12 πŸ‘ 2 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

Many non-adopters have legitimate reasons to resist adopting or to stop using chat models; however, as research practitioners who have the power to design and shape model capabilities, we urge the community to develop technologies where non-use is a choice, rather than an inevitable circumstance.

21.10.2025 17:12 πŸ‘ 1 πŸ” 0 πŸ’¬ 2 πŸ“Œ 0

From users interviews (n=23) and an online survey (n=230), we found:
1️⃣ Non-adopters are interested in chat models, but face barriers to adoption
2️⃣ Non-adopters prioritize tasks rarely reflected in model evals: navigating healthcare portals, coordinating caregiving, contextualized IR

21.10.2025 17:12 πŸ‘ 1 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

Many current methods rely on logs, preferences, and feedback from existing usersβ€”who represent a narrow slice of the population.

Adopter-centered methods risk widening the divide between adopters and non-adopters as datasets, benchmarks, and evaluations evolve around current adopter needs.

21.10.2025 17:12 πŸ‘ 2 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
A circular flow diagram that compares current and proposed practices for LLM development using data from adopters and non-adopters. Three gray boxes represent current practices: β€œR&D,” β€œChat Models,” and β€œAdopters’ Needs and Usage Data,” connected in a clockwise loop with black arrows. A blue box labeled β€œNon-adopters’ Needs and Usage Data” adds a proposed feedback path, shown with blue arrows, linking non-adopter data back to R&D and adopters’ data.

A circular flow diagram that compares current and proposed practices for LLM development using data from adopters and non-adopters. Three gray boxes represent current practices: β€œR&D,” β€œChat Models,” and β€œAdopters’ Needs and Usage Data,” connected in a clockwise loop with black arrows. A blue box labeled β€œNon-adopters’ Needs and Usage Data” adds a proposed feedback path, shown with blue arrows, linking non-adopter data back to R&D and adopters’ data.

As of June 2025, 66% of Americans have never used ChatGPT.

Our new position paper, Attention to Non-Adopters, explores why this matters: AI research is being shaped around adoptersβ€”leaving non-adopters’ needs, and key LLM research opportunities, behind.

arxiv.org/abs/2510.15951

21.10.2025 17:12 πŸ‘ 38 πŸ” 13 πŸ’¬ 2 πŸ“Œ 0
Post image

We are excited to welcome five new faculty members to Cornell Bowers this semester!

A Big Red Bowers Welcome to Sasha Golovnev, Andrew Owens '10, David Rand '04, Benjamin Shestakofsky, and Kaitlyn Zhou. πŸ‘‹

Read more: lnkd.in/ewA4U4fU

08.09.2025 15:04 πŸ‘ 2 πŸ” 1 πŸ’¬ 0 πŸ“Œ 0
Screenshot of paper title: Sycophantic AI Decreases Prosocial Intentions and Promotes Dependence

Screenshot of paper title: Sycophantic AI Decreases Prosocial Intentions and Promotes Dependence

AI always calling your ideas β€œfantastic” can feel inauthentic, but what are sycophancy’s deeper harms? We find that in the common use case of seeking AI advice on interpersonal situationsβ€”specifically conflictsβ€”sycophancy makes people feel more right & less willing to apologize.

03.10.2025 22:53 πŸ‘ 115 πŸ” 48 πŸ’¬ 2 πŸ“Œ 7
Preview
CS PhD Statements of Purpose cs-sop.org is a platform intended to help CS PhD applicants. It hosts a database of example statements of purpose (SoP) shared by previous applicants to Computer Science PhD programs.

It is PhD application season again πŸ‚ For those looking to do a PhD in AI, these are some useful resources πŸ€–:

1. Examples of statements of purpose (SOPs) for computer science PhD programs: cs-sop.org [1/4]

01.10.2025 20:37 πŸ‘ 9 πŸ” 4 πŸ’¬ 1 πŸ“Œ 0

I'll be at COLM next week! Let me know if you want to chat! @colmweb.org

@neilrathi.bsky.social will be presenting our work on multilingual overconfidence in language models and the effects on human overreliance!

arxiv.org/pdf/2507.06306

03.10.2025 17:33 πŸ‘ 7 πŸ” 1 πŸ’¬ 0 πŸ“Œ 0

Congrats!!!

20.08.2025 19:41 πŸ‘ 3 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0

Congrats!!! πŸ₯³πŸ€©

14.08.2025 13:56 πŸ‘ 2 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Kaiserslautern, Germany

Kaiserslautern, Germany

πŸ“£ Life update: Thrilled to announce that I’ll be starting as faculty at the Max Planck Institute for Software Systems this Fall!

I’ll be recruiting PhD students in the upcoming cycle, as well as research interns throughout the year: lasharavichander.github.io/contact.html

22.07.2025 04:12 πŸ‘ 92 πŸ” 12 πŸ’¬ 13 πŸ“Œ 4

For EMNLP 2025’s special theme of "Advancing our Reach: Interdisciplinary Recontextualization of NLP", we are organizing a panel of experts, and would like input from the community at large as we prepare. Please take a moment to fill in this survey: forms.office.com/r/pWFFA0Gss1

17.07.2025 20:24 πŸ‘ 8 πŸ” 5 πŸ’¬ 0 πŸ“Œ 0
Preview
Computer-vision research powers surveillance technology - Nature An analysis of research papers and citing patents indicates the extensive ties between computer-vision research and surveillance.

New paper hot off the press www.nature.com/articles/s41...

We analysed over 40,000 computer vision papers from CVPR (the longest standing CV conf) & associated patents tracing pathways from research to application. We found that 90% of papers & 86% of downstream patents power surveillance

1/

25.06.2025 17:29 πŸ‘ 937 πŸ” 533 πŸ’¬ 31 πŸ“Œ 77
Announcing Best Paper Awards The Best Paper Award Committee was chaired this year by Alex Chouldechova and included six Area Chairs. The committee selected three papers for the Best Paper Award and recognized three additional pap...

πŸ† Announcing the #FAccT2025 best paper awards! πŸ†

Congratulations to all the authors of the three best papers and three honorable mention papers.

Be sure to check out their presentations at the conference next week!

facct-blog.github.io/2025-06-20/b...

20.06.2025 21:14 πŸ‘ 36 πŸ” 13 πŸ’¬ 0 πŸ“Œ 8
Post image

What if AI played the role of your sassy gay bestie πŸ³οΈβ€πŸŒˆ or AAVE-speaking friend πŸ‘‹πŸΎ?

You: β€œCan you plan a trip?”
πŸ€– AI: β€œYasss queen! let’s werk this babeβœ¨πŸ’…β€

LLMs can talk like us, but it shapes how we trust, rely on & relate to them 🧡

πŸ“£ our #FAccT2025 paper: bit.ly/3HJ6rWI

[1/9]

17.06.2025 19:39 πŸ‘ 13 πŸ” 6 πŸ’¬ 1 πŸ“Œ 2