Byron Wallace
@byron.bsky.social
📤 2459
📥 321
📝 10
Assoc. Prof in CS @ Northeastern, NLP/ML & health & etc. He/him.
Check out
@hibaahsan.bsky.social
's paper on spotting (problematic) racial biases in LLMs for healthcare applications 👇
add a skeleton here at some point
2 months ago
0
2
0
reposted by
Byron Wallace
Ai2
3 months ago
3/ 🏥 A separate team at Northeastern located where certain signals live inside Olmo and made targeted edits that reduced biased clinical predictions. This kind of audit is only possible because Olmo exposes all its components. →
buff.ly/HkChr4Q
1
0
2
Chantal (and Vinith) find that you can jailbreak LLMs with syntax! Some examples:
cshaib.github.io/syntax_domai...
add a skeleton here at some point
3 months ago
0
2
0
Now to appear at
#EMNLP2025
(Findings). We've added more models and experiments:
arxiv.org/abs/2502.13319
add a skeleton here at some point
3 months ago
0
2
0
Can we distill *circuits* from teacher models into smaller students? 👇
add a skeleton here at some point
3 months ago
0
1
0
reposted by
Byron Wallace
David Bau
4 months ago
Who is going to be at
#COLM2025
? I want to draw your attention to a COLM paper by my student
@sfeucht.bsky.social
that has totally changed the way I think and teach about LLM representations. The work is worth knowing. And you can meet Sheridan at COLM, Oct 7!
bsky.app/profile/sfe...
1
39
10
Can we quantify what makes some text read like AI "slop"? We tried 👇
add a skeleton here at some point
4 months ago
0
8
1
reposted by
Byron Wallace
Naomi Saphra
4 months ago
Our new paper asks: what is the goal of “natural language verbalization” interpretability approaches? If a verbalizer is supposed to tell us something about what’s in the target LM and NOT just what’s in the verbalizer LM, how do we actually evaluate that?
add a skeleton here at some point
0
13
3
reposted by
Byron Wallace
Millicent Li
4 months ago
Wouldn’t it be great to have questions about LM internals answered in plain English? That’s the promise of verbalization interpretability. Unfortunately, our new paper shows that evaluating these methods is nuanced—and verbalizers might not tell us what we hope they do. 🧵👇1/8
1
27
9
reposted by
Byron Wallace
Hye Sun Yun
5 months ago
Thrilled to share our research showing how LLM models can be influenced by bias from "spun" medical literature is now featured in Northeastern's Khoury news! This shows critical insights as AI enters healthcare. The full paper can be found at
arxiv.org/abs/2502.07963
loading . . .
As AI expands into medicine, Northeastern study finds AI models influenced by medical bias - Khoury College of Computer Sciences
Humans can be easily influenced by language that is one-sided, especially in complex fields like medicine. But a new Khoury-led study shows that large language models, too, can be tricked […]
https://khoury.northeastern.edu/as-ai-expands-into-medicine-northeastern-study-finds-ai-models-influenced-by-medical-bias/
0
3
1
reposted by
Byron Wallace
David Bau
5 months ago
This Friday NEMI 2025 is at Northeastern in Boston, 8 talks, 24 roundtables, 90 posters; 200+ attendees. Thanks to
goodfire.ai/
for sponsoring!
nemiconf.github.io/summer25/
If you can't make it in person, the livestream will be here:
www.youtube.com/live/4BJBis...
loading . . .
New England Mechanistic Interpretability Workshop
About:The New England Mechanistic Interpretability (NEMI) workshop aims to bring together academic and industry researchers from the New England and surround...
https://www.youtube.com/watch?v=4BJBisHk1UI
1
16
10
reposted by
Byron Wallace
Monica M Reddy
5 months ago
📢 How factual are LLMs in healthcare? We’re excited to release FactEHR — a new benchmark to evaluate factuality in clinical notes. As generative AI enters the clinic, we need rigorous, source-grounded tools to measure what these models get right — and what they don’t. 🏥 🤖
1
3
3
reposted by
Byron Wallace
Yuval Pinter
6 months ago
Chatted with
@byron.bsky.social
at icml about my recent work, so look out for his upcoming "Tokenization is More Than More Than Compression".
1
13
1
reposted by
Byron Wallace
Lily Chen
6 months ago
Are we fact-checking medical claims the right way? 🩺🤔 Probably not. In our study, even experts struggled to verify Reddit health claims using end-to-end systems. We show why—and argue fact-checking should be a dialogue, with patients in the loop
arxiv.org/abs/2506.20876
đź§µ1/
1
5
3
reposted by
Byron Wallace
Sheridan Feucht
9 months ago
[đź“„] Are LLMs mindless token-shifters, or do they build meaningful representations of language? We study how LLMs copy text in-context, and physically separate out two types of induction heads: token heads, which copy literal tokens, and concept heads, which copy word meanings.
1
77
26
reposted by
Byron Wallace
Chantal
10 months ago
I'm searching for some comp/ling experts to provide a precise definition of “slop” as it refers to text (see:
corp.oup.com/word-of-the-...
) I put together a google form that should take no longer than 10 minutes to complete:
forms.gle/oWxsCScW3dJU...
If you can help, I'd appreciate your input! 🙏
loading . . .
Oxford Word of the Year 2024 - Oxford University Press
The Oxford Word of the Year 2024 is 'brain rot'. Discover more about the winner, our shortlist, and 20 years of words that reflect the world.
https://corp.oup.com/word-of-the-year/#shortlist-2024
0
10
8
reposted by
Byron Wallace
Jessy Li
11 months ago
🌟Job ad🌟 We (
@gregdnlp.bsky.social
,
@mattlease.bsky.social
and I) are hiring a postdoc fellow within the CosmicAI Institute, to do galactic work with LLMs and generative AI! If you would like to push the frontiers of foundation models to help solve myths of the universe, please apply!
add a skeleton here at some point
0
13
10
reposted by
Byron Wallace
Hiba Ahsan
11 months ago
LLMs are known to perpetuate social biases in clinical tasks. Can we locate and intervene upon LLM activations that encode patient demographics like gender and race? đź§µ Work w/
@arnabsensharma.bsky.social
,
@silvioamir.bsky.social
,
@davidbau.bsky.social
,
@byron.bsky.social
arxiv.org/abs/2502.13319
3
18
9
reposted by
Byron Wallace
Hye Sun Yun
11 months ago
🚨 Do LLMs fall for spin in medical literature? 🤔 In our new preprint, we find that LLMs are susceptible to biased reporting of clinical treatment benefits in abstracts—more so than human experts. 📄🔍 [1/7] Full Paper:
arxiv.org/abs/2502.07963
🧵👇
3
63
29
reposted by
Byron Wallace
Somin W
11 months ago
📢 Can we trace a small distilled model back to its teacher? 🤔New work (w/
@chantalsh.bsky.social
,
@silvioamir.bsky.social
&
@byron.bsky.social
) finds some footprints left by LLMs in distillation! [1/6] đź”— Full paper:
arxiv.org/abs/2502.06659
loading . . .
Who Taught You That? Tracing Teachers in Model Distillation
Model distillation -- using outputs from a large teacher model to teach a small student model -- is a practical means of creating efficient models for a particular task. We ask: Can we identify a stud...
https://arxiv.org/abs/2502.06659
1
8
2
reposted by
Byron Wallace
David Bau
11 months ago
DeepSeek R1 shows how important it is to be studying the internals of reasoning models. Try our code: Here
@canrager.bsky.social
shows a method for auditing AI bias by probing the internal monologue.
dsthoughts.baulab.info
I'd be interested in your thoughts.
loading . . .
https://dsthoughts.baulab
1
28
10
reposted by
Byron Wallace
12 months ago
📣 🌍 We're hiring for 2 Machine Learning researchers to join SOLACE-AI
@kingscollegelondon.bsky.social
, funded by
@wellcometrust.bsky.social
. This is your chance to develop cutting-edge AI to directly impact global health responses to climate emergencies.
jobs.ac.uk/job/DLM377
0
2
3
reposted by
Byron Wallace
Luca Soldaini 🎀
about 1 year ago
OLMo 2 is out 🥳 7B and 13B trained on 5T tokens, and meticulousy instruction tuned using Tulu 3 recipe. Simply the best fully open models yet. Really proud of the work & the amazing team at
@ai2.bsky.social
9
262
46
I'll be @
#EMNLP2024
if anyone wants to find snobby coffee / despair about election / or I guess talk research. Some work to be presented👇
about 1 year ago
1
13
0
reposted by
Byron Wallace
Jered McInerney
almost 2 years ago
Our work on reducing diagnostic errors with interpretable risk prediction is now on arXiv! We retrieve evidence from a patient’s record, visualize how it informs a prediction, and test it in a realistic setting. 👇 (1/6)
arxiv.org/abs/2402.10109
w/
@byron.bsky.social
and
@jwvdm.bsky.social
1
2
2
reposted by
Byron Wallace
Jessy Li
about 2 years ago
To appear
#EMNLP2023
! Can LMs simplify medical texts in non-English languages? We introduce⚕️MultiCochrane: the *first* multilingual, aligned dataset for this.
arxiv.org/abs/2305.12532
. Led by Sebastian Joseph, also w/
@byron.bsky.social
Wei Xu
1
13
5
reposted by
Byron Wallace
David Bau
about 2 years ago
Work with Jiuding Sun, Andrew Yuan, and Byron Wallace
@byron.bsky.social
If you're going to be at EMNLP/CoNLL/BlackboxNLP in Singapore, look for her poster at CoNLL! Koyena's FutureLens preprint, code, and demo are on the project website at
future.baulab.info
0
2
1
reposted by
Byron Wallace
Jessy Li
about 2 years ago
Can we use LLMs to help disseminate medical information more broadly?
@byron.bsky.social
, Mike Mackert, Wei Xu and I are hosting an online panel today at the HARC conference at 4:30 EST/3:30 CST on Simplifying Medical Texts with Large Language Models!
harcconf.org/agenda-monda...
0
7
2
reposted by
Byron Wallace
David Bau
about 2 years ago
LLMs contain Function Vectors! Eric Todd has a really interesting new preprint on arxiv
functions.baulab.info
showing LLMs contain vector representations of functions that compose and apply in diverse contexts. Could be a powerful analysis tool. More in his twitter thread:
x.com/ericwtodd/st...
loading . . .
Function Vectors in Large Language Models
Understanding the internal computations of huge autoregressive transformer neural network language models during in-context learning.
https://functions.baulab.info/
1
12
3
Check out Jered's work combining LLM (zero-shot) extracted features with simple linear models for healthcare data👇 (This *is* where we're doing shameless research self-promo threads now, right?)
add a skeleton here at some point
about 2 years ago
0
4
1
you reached the end!!
feeds!
log in