Dayeon (Zoey) Ki
@dayeonki.bsky.social
📤 162
📥 220
📝 24
CS PhD @umdclip Multilingual / Culture
#NLProc
, MT
https://dayeonki.github.io/
1/ Are two
#LLMs
better than one for equitable cultural alignment? 🌍 We introduce a Multi-Agent Debate framework — where two LLM agents debate the cultural adaptability of a given scenario.
#ACL2025
🧵👇
5 months ago
1
6
1
reposted by
Dayeon (Zoey) Ki
Vilém Zouhar #EMNLP
12 months ago
Trying to collect all the MT people here. I probably missed many. Ping me!
bsky.app/starter-pack...
add a skeleton here at some point
9
24
8
1/ How can a monolingual English speaker 🇺🇸 decide if an automatic French translation 🇫🇷 is good enough to be shared? Introducing ❓AskQE❓, an
#LLM-based
Question Generation + Answering framework that detects critical MT errors and provides actionable feedback 🗣️
#ACL2025
6 months ago
1
1
2
reposted by
Dayeon (Zoey) Ki
Myra Cheng
7 months ago
How does the public conceptualize AI? Rather than self-reported measures, we use metaphors to understand the nuance and complexity of people’s mental models. In our
#FAccT2025
paper, we analyzed 12,000 metaphors collected over 12 months to track shifts in public perceptions.
3
49
15
reposted by
Dayeon (Zoey) Ki
Vilém Zouhar #EMNLP
7 months ago
Multilinguality is happening at
#NAACL2025
@crystinaz.bsky.social
@oxxoskeets.bsky.social
@dayeonki.bsky.social
@onadegibert.bsky.social
0
14
1
reposted by
Dayeon (Zoey) Ki
Angel Hsing-Chi Hwang
7 months ago
Starting my journey on Bluesky with a topic that I care deeply about: AI tools can support creators in various ways, but disclosing AI use may risk devaluing creative work. Check out our abstract here:
angelhwang.github.io/doc/ic2s2_AI...
Inspired by our past work:
arxiv.org/abs/2411.13032
loading . . .
"It was 80% me, 20% AI": Seeking Authenticity in Co-Writing with Large Language Models
Given the rising proliferation and diversity of AI writing assistance tools, especially those powered by large language models (LLMs), both writers and readers may have concerns about the impact of th...
https://arxiv.org/abs/2411.13032
1
26
6
🚨 New Paper 🚨 1/ We often assume that well-written text is easier to translate ✏️ But can
#LLMs
automatically rewrite inputs to improve machine translation? 🌍 Here’s what we found 🧵
7 months ago
1
8
4
reposted by
Dayeon (Zoey) Ki
Tokenization Workshop (TokShop) @ICML2025
7 months ago
🚨 NEW WORKSHOP ALERT 🚨 We're thrilled to announce the first-ever Tokenization Workshop (TokShop) at
#ICML2025
@icmlconf.bsky.social
! 🎉 Submissions are open for work on tokenization across all areas of machine learning. 📅 Submission deadline: May 30, 2025 🔗
tokenization-workshop.github.io
loading . . .
Tokenization Workshop @ ICML 2025
https://tokenization-workshop.github.io
1
23
11
reposted by
Dayeon (Zoey) Ki
Shayne Longpre
7 months ago
Thrilled our global data ecosystem audit was accepted to
#ICLR2025
! Empirically, it shows: 1️⃣ Soaring synthetic text data: ~10M tokens (pre-2018) to 100B+ (2024). 2️⃣ YouTube is now 70%+ of speech/video data but could block third-party collection. 3️⃣ <0.2% of data from Africa/South America. 1/
1
12
5
reposted by
Dayeon (Zoey) Ki
Zdeněk Kasner
7 months ago
How do LLMs compare to human crowdworkers in annotating text spans? 🧑🤖 And how can span annotation help us with evaluating texts? Find out in our new paper:
llm-span-annotators.github.io
Arxiv:
arxiv.org/abs/2504.08697
loading . . .
Large Language Models as Span Annotators
Website for the paper Large Language Models as Span Annotators
https://llm-span-annotators.github.io
1
20
9
reposted by
Dayeon (Zoey) Ki
Helsinki NLP
8 months ago
Call for participation: We just opened the registration for this year's MT Marathon in August in Helsinki, Finland:
blogs.helsinki.fi/language-tec...
, featuring: - Ayodele Awokoya - Wilker Aziz - Marta Costa-Jussa - Barry Haddow - Amit Moryosse - Sara Papi - Jörg Tiedemann - Marco Turchi
loading . . .
https://blogs.helsinki.fi/language-tec...
0
5
5
reposted by
Dayeon (Zoey) Ki
Ona de Gibert
8 months ago
Come to Helsinki for the 18th MT Marathon! Sponsored by EAMT
@ufal-cuni.bsky.social
add a skeleton here at some point
0
8
6
reposted by
Dayeon (Zoey) Ki
Barry Haddow
9 months ago
** New parallel data set ** . We've just released HPLT v2.0, a parallel data set of 50 languages paired with English, 380M sentence pairs in total. Extracted from the Internet Archive and Common Crawl
hplt-project.org/datasets/v2.0
loading . . .
HPLT - High Performance Language Technologies
A space that combines petabytes of natural language data with large-scale model training
https://hplt-project.org/datasets/v2.0
1
4
4
reposted by
Dayeon (Zoey) Ki
Andrea Piergentili
8 months ago
Brilliant and necessary work by Pombal et al. about metric interference in MT system development and evaluation:
arxiv.org/abs/2503.08327
Are we developing better systems or are we just gaming the metrics? And how do we address this? Super (m)interesting! 👀
loading . . .
Adding Chocolate to Mint: Mitigating Metric Interference in Machine Translation
As automatic metrics become increasingly stronger and widely adopted, the risk of unintentionally "gaming the metric" during model development rises. This issue is caused by metric interference (Mint)...
https://arxiv.org/abs/2503.08327
0
9
2
reposted by
Dayeon (Zoey) Ki
Yixiao Song
8 months ago
Introducing 🐻 BEARCUBS 🐻, a “small but mighty” dataset of 111 QA pairs designed to assess computer-using web agents in multimodal interactions on the live web! ✅ Humans achieve 85% accuracy ❌ OpenAI Operator: 24% ❌ Anthropic Computer Use: 14% ❌ Convergence AI Proxy: 13%
1
12
8
reposted by
Dayeon (Zoey) Ki
Siyuan Song
8 months ago
New preprint w/
@jennhu.bsky.social
@kmahowald.bsky.social
: Can LLMs introspect about their knowledge of language? Across models and domains, we did not find evidence that LLMs have privileged access to their own predictions. 🧵(1/8)
2
60
20
reposted by
Dayeon (Zoey) Ki
Miriam Posner
8 months ago
OK, every year I try to explain to my students how LLMs work, and every year I have to do a big trawl for good resources and activities. Here's this year's haul of *introductory* materials. (In-class activities + visualizations, not so much readings.)
66
780
218
reposted by
Dayeon (Zoey) Ki
Tom Kocmi
8 months ago
Big news from WMT! 🎉 We are expanding beyond MT and launching a new multilingual instruction shared task. Our goal is to foster truly multilingual LLM evaluation and best practices in automatic and human evaluation. Join us and build the winning multilingual system!
www2.statmt.org/wmt25/multil...
loading . . .
Multilingual Instruction Shared Task
https://www2.statmt.org/wmt25/multilingual-instruction.html
1
12
9
reposted by
Dayeon (Zoey) Ki
Artjoms Šeļa
8 months ago
self-insert, but if you are looking for something multilingual and public domain, we have PoeTree: a collection of poetry corpora with Python & R access points (can get data directly into your jupyter notebook) :
versologie.cz/poetree/
loading . . .
PoeTree. Poetry Treebanks in 10 languages
PoeTree is a standardized collection of poetry corpora comprising over 330,000 poems in ten languages (Czech, English, French, German, Hungarian, Italian, Portuguese, Russian, Slovenian, Spanish).
https://versologie.cz/poetree/
1
11
2
reposted by
Dayeon (Zoey) Ki
Aaron Mueller
8 months ago
Lots of work coming soon to
@iclr-conf.bsky.social
and
@naaclmeeting.bsky.social
in April/May! Come chat with us about new methods for interpreting and editing LLMs, multilingual concept representations, sentence processing mechanisms, and arithmetic reasoning. 🧵
1
20
6
reposted by
Dayeon (Zoey) Ki
Nishant Balepur
8 months ago
🚨 Our team at UMD is looking for participants to study how
#LLM
agent plans can help you answer complex questions 💰 $1 per question 🏆 Top-3 fastest + most accurate win $50 ⏳ Questions take ~3 min => $20/hr+ Click here to sign up (please join, reposts appreciated 🙏):
preferences.umiacs.umd.edu
0
2
3
reposted by
Dayeon (Zoey) Ki
Kathy
8 months ago
Happy to say that our paper "Beyond Literal Token Overlap: Token Alignability for Multilinguality" will be presented at
#NAACL2025
! This is work with
@tomlim.bsky.social
,
@jlibovicky.bsky.social
, and Alex Fraser.
arxiv.org/abs/2502.06468
#newpaper
#NLP
#NLProc
loading . . .
Beyond Literal Token Overlap: Token Alignability for Multilinguality
Previous work has considered token overlap, or even similarity of token distributions, as predictors for multilinguality and cross-lingual knowledge transfer in language models. However, these very li...
https://arxiv.org/abs/2502.06468
1
11
5
reposted by
Dayeon (Zoey) Ki
Catherine Arnett
8 months ago
✨New pre-print✨ Crosslingual transfer allows models to leverage their representations for one language to improve performance on another language. We characterize the acquisition of shared representations in order to better understand how and when crosslingual transfer happens.
2
36
9
reposted by
Dayeon (Zoey) Ki
Jess Hamrick
8 months ago
This is a really neat use case for AI—checking whether claims are actually supported by the given citations.
add a skeleton here at some point
1
25
6
reposted by
Dayeon (Zoey) Ki
Karolina Stańczak
8 months ago
📢New Paper Alert!🚀 Human alignment balances social expectations, economic incentives, and legal frameworks. What if LLM alignment worked the same way?🤔 Our latest work explores how social, economic, and contractual alignment can address incomplete contracts in LLM alignment🧵
2
28
16
reposted by
Dayeon (Zoey) Ki
Cohere Labs
8 months ago
We hope Aya enables researchers and developers throughout the world to build upon this technology, ask deeper questions about multilingual AI, and develop tools that can support their communities. Learn more:
cohere.com/blog/aya-vis...
loading . . .
Aya Vision: Expanding the Worlds AI Can See
Our state-of-the-art open-weights vision model offers a foundation for AI-enabled multilingual and multimodal communication globally. Today, Cohere For AI, Cohere’s open research arm, is proud to an...
https://cohere.com/blog/aya-vision
0
2
2
reposted by
Dayeon (Zoey) Ki
Marzieh Fadaee
8 months ago
✨👓 Aya Vision is here 👓✨ A multilingual, multimodal model designed to understand across languages and modalities (text, images, etc) to bridge the language gap and empower global users!
1
4
2
reposted by
Dayeon (Zoey) Ki
Tom Kocmi
9 months ago
Huge shoutout to colleagues at Google & Unbabel for extending our WMT24 testset to 55 languages in four domains, this is game changer! 🚀 I really hope it puts the final nail in the coffin of FLORES or WMT14. The field is evolving, legacy testsets can't show your progress
arxiv.org/abs/2502.124...
loading . . .
WMT24++: Expanding the Language Coverage of WMT24 to 55 Languages & Dialects
As large language models (LLM) become more and more capable in languages other than English, it is important to collect benchmark datasets in order to evaluate their multilingual performance, includin...
https://arxiv.org/abs/2502.12404v1
0
14
6
reposted by
Dayeon (Zoey) Ki
Kayo Yin
9 months ago
Induction heads are commonly associated with in-context learning, but are they the primary driver of ICL at scale? We find that recently discovered "function vector" heads, which encode the ICL task, are the actual primary mechanisms behind few-shot ICL!
arxiv.org/abs/2502.14010
🧵👇
1
22
7
reposted by
Dayeon (Zoey) Ki
Tejas Srinivasan
9 months ago
People are increasingly relying on AI assistance, but *how* they use AI advice is influenced by their trust in the AI, which the AI is typically blind to. What if they weren’t? We show that adapting AI assistants' behavior to user trust mitigates under- and over-reliance!
arxiv.org/abs/2502.13321
1
8
4
reposted by
Dayeon (Zoey) Ki
naitian
9 months ago
There's been a lot of work on "culture" in NLP, but not much agreement on what it is. A position paper by me,
@dbamman.bsky.social
, and
@ibleaman.bsky.social
on cultural NLP: what we want, what we have, and how sociocultural linguistics can clarify things. Website:
naitian.org/culture-not-...
1/n
5
121
38
reposted by
Dayeon (Zoey) Ki
Shayne Longpre
9 months ago
Thrilled to be at
#AAAI2025
for our tutorial, “AI Data Transparency: The Past, Present, and Beyond.” We’re presenting the state of transparency, tooling, and policy, from the Foundation Model Transparency Index, Factsheets, the the EU AI Act to new frameworks like @MLCommons’ Croissant. 1/
1
15
4
reposted by
Dayeon (Zoey) Ki
Akhila Yerukola
9 months ago
Did you know? Gestures used to express universal concepts—like wishing for luck—vary DRAMATICALLY across cultures? 🤞means luck in US but deeply offensive in Vietnam 🚨 📣 We introduce MC-SIGNS, a test bed to evaluate how LLMs/VLMs/T2I handle such nonverbal behavior! 📜:
arxiv.org/abs/2502.17710
1
33
10
reposted by
Dayeon (Zoey) Ki
Chau Minh Pham
9 months ago
⚠️Current methods for generating instruction-following data fall short for long-range reasoning tasks like narrative claim verification. We present CLIPPER ✂️, a compression-based pipeline that produces grounded instructions for ~$0.5 each, 34x cheaper than human annotations.
1
21
10
reposted by
Dayeon (Zoey) Ki
Nishant Balepur
9 months ago
🚨 New Position Paper 🚨 Multiple choice evals for LLMs are simple and popular, but we know they are awful 😬 We complain they're full of errors, saturated, and test nothing meaningful, so why do we still use them? 🫠 Here's why MCQA evals are broken, and how to fix them 🧵
2
46
13
reposted by
Dayeon (Zoey) Ki
Stella Li
9 months ago
Asking the right questions can make or break decisions in fields like medicine, law, and beyond✴️ Our new framework ALFA—ALignment with Fine-grained Attributes—teaches LLMs to PROACTIVE seek information through better questions through **structured rewards**🏥❓ (co-led with
@jiminmun.bsky.social
) 👉🏻🧵
1
24
10
reposted by
Dayeon (Zoey) Ki
Hal Daumé III
11 months ago
starter pack for the Computational Linguistics and Information Processing group at the University of Maryland - get all your NLP and data science here!
go.bsky.app/V9qWjEi
add a skeleton here at some point
1
29
13
reposted by
Dayeon (Zoey) Ki
Marzieh Fadaee
11 months ago
Aya Expanse technical report is here and I couldn’t be prouder! 🎉 Check out the report:
arxiv.org/abs/2412.04261
loading . . .
Aya Expanse: Combining Research Breakthroughs for a New Multilingual Frontier
We introduce the Aya Expanse model family, a new generation of 8B and 32B parameter multilingual language models, aiming to address the critical challenge of developing highly performant multilingual ...
https://arxiv.org/abs/2412.04261
0
4
3
reposted by
Dayeon (Zoey) Ki
Marc Marone
12 months ago
I noticed a lot of starter packs skewed towards faculty/industry, so I made one of just NLP & ML students:
go.bsky.app/vju2ux
Students do different research, go on the job market, and recruit other students. Ping me and I'll add you!
add a skeleton here at some point
101
176
58
reposted by
Dayeon (Zoey) Ki
Sara Hooker
11 months ago
Is MMLU Western-centric? 🤔 As part of a massive cross-institutional collaboration: 🗽Find MMLU is heavily overfit to western culture 🔍 Professional annotation of cultural sensitivity data 🌍 Release improved Global-MMLU 42 languages 📜 Paper:
arxiv.org/pdf/2412.03304
📂 Data:
hf.co/datasets/Coh...
7
59
18
you reached the end!!
feeds!
log in