Navita Goyal
@navitagoyal.bsky.social
📤 288
📥 193
📝 13
PhD student @umdcs, Member of @ClipUmd lab | Earlier @AdobeResearch, @IITRoorkee
Thanks WiAIR (
@wiair.bsky.social
) for featuring my work on your YouTube channel. Watch the video to hear about our work on inference-time steering — and why these interventions LLMs may not be as “precise” as they look.
add a skeleton here at some point
10 days ago
1
2
1
reposted by
Navita Goyal
Yoav Artzi
about 2 months ago
This call is still open. I am looking to recruit, as well as many other faculty at Cornell. We review folders as they come, and will send offers until all positions are filled. Please share with your network 🙏
add a skeleton here at some point
0
11
8
reposted by
Navita Goyal
Andrew Lampinen
3 months ago
What can cognitive science learn from AI? In
infinitefaculty.substack.com/p/what-cogni...
I outline how AI has found that scale and richness of learning experiences fundamentally change learning & generalization — and how I believe we should rethink cognitive experiments & theories in response.
loading . . .
What cognitive science can learn from AI
#3 in a series on cognitive science and AI
https://infinitefaculty.substack.com/p/what-cognitive-science-can-learn
1
36
15
Woah, this is so cool! How was I not aware of this. I just set mine up to prepare for NeurIPS and I am loving it already... it made thousands of accepted paper so much more tractable to navigate
add a skeleton here at some point
4 months ago
0
3
0
reposted by
Navita Goyal
Hal Daumé III
5 months ago
AIM's 2nd round of TTK hiring - building up to 30 - is up! 📅 Ddl 12/22/25 🔬 Accessibility & Learning, plus Sustainability & Social Justice 🧑🏫 Associate/Full Prof* 🔗
umd.wd1.myworkdayjobs.com/en-US/UMCP/j...
*Assistant-level candidates: apply to departments, mentioning AIM in a cover letter
loading . . .
Senior Tenure Track Faculty at the Artificial Intelligence Interdisciplinary Institute at Maryland (AIM) - Associate Professor/Professor (Open Rank Joint Appointment)
Job Description Summary Organization Summary Statement: The Artificial Intelligence Interdisciplinary Institute at Maryland - AIM (aim.umd.edu) - is hiring 40 faculty over the next several years, incl...
https://umd.wd1.myworkdayjobs.com/en-US/UMCP/job/Senior-Tenure-Track-Faculty-at-the-Artificial-Intelligence-Interdisciplinary-Institute-at-Maryland--AIM----Associate-Professor-Professor--Open-Rank-Joint-Appointment-_JR102900-1
0
11
9
reposted by
Navita Goyal
Najoung Kim
5 months ago
My lab at BU is recruiting PhD students and possibly a postdoc this year! We study humans & machines, centered around topics like meaning, generalization, evaluation methods and design, and the nature of computation and representation that underlie language and cognition. 🫴🫴
1
13
4
reposted by
Navita Goyal
Yanai Elazar
5 months ago
Interested in interpretability, data attribution, evaluation, and similar topics? Interested in doing a postdoc with me? Apply to the prestigious Azrieli program! Link below 👇 DMs are open (email is good too!)
1
5
1
reposted by
Navita Goyal
Alexander Hoyle
5 months ago
Happy to be at
#EMNLP2025
! Please say hello and come see our lovely work
0
8
1
reposted by
Navita Goyal
Naomi Saphra
6 months ago
I am recruiting PhD students to start in 2026! If you are interested in robustness, training dynamics, interpretability for scientific understanding, or the science of LLM analysis you should apply. BU is building a huge LLM analysis/interp group and you’ll be joining at the ground floor.
add a skeleton here at some point
1
57
19
reposted by
Navita Goyal
Neha Srikanth
11 months ago
I'll be presenting this work with @rachelrudinger at
#NAACL2025
tomorrow (Wednesday 4/30) in Albuquerque during Session C (Oral/Poster 2) at 2pm! 🔬 Decomposing hypotheses in traditional NLI and defeasible NLI helps us measure various forms of consistency of LLMs. Come join us!
5
8
4
reposted by
Navita Goyal
Vishakh Padmakumar
11 months ago
What does it mean for
#LLM
output to be novel? In work w/
johnchen6.bsky.social
, Jane Pan, Valerie Chen and He He, we argue it needs to be both original and high quality. While prompting tricks trade one for the other, better models (scaling/post-training) can shift the novelty frontier 🧵
2
7
4
reposted by
Navita Goyal
Dayeon (Zoey) Ki
12 months ago
🚨 New Paper 🚨 1/ We often assume that well-written text is easier to translate ✏️ But can
#LLMs
automatically rewrite inputs to improve machine translation? 🌍 Here’s what we found 🧵
1
8
4
reposted by
Navita Goyal
Kartik
12 months ago
🔈 NEW PAPER 🔈 Excited to share my paper that analyzes the effect of cross-lingual alignment on multilingual performance Paper:
arxiv.org/abs/2504.09378
🧵
loading . . .
Can you map it to English? The Role of Cross-Lingual Alignment in Multilingual Performance of LLMs
Large language models (LLMs) pre-trained predominantly on English text exhibit surprising multilingual capabilities, yet the mechanisms driving cross-lingual generalization remain poorly understood. T...
https://arxiv.org/abs/2504.09378
1
0
2
reposted by
Navita Goyal
Sarah Wiegreffe
about 1 year ago
Have work on the actionable impact of interpretability findings? Consider submitting to our Actionable Interpretability workshop at ICML! See below for more info. Website:
actionable-interpretability.github.io
Deadline: May 9
add a skeleton here at some point
0
20
10
reposted by
Navita Goyal
Mohit Iyyer
about 1 year ago
Thinking about paying $20k/month for a "PhD-level AI agent"? You might want to wait until their web browsing skills are on par with those of human PhD students 😛 Check out our new BEARCUBS benchmark, which shows web agents struggle to perform simple multimodal browsing tasks!
add a skeleton here at some point
0
6
1
reposted by
Navita Goyal
Nishant Balepur
about 1 year ago
🚨 Our team at UMD is looking for participants to study how
#LLM
agent plans can help you answer complex questions 💰 $1 per question 🏆 Top-3 fastest + most accurate win $50 ⏳ Questions take ~3 min => $20/hr+ Click here to sign up (please join, reposts appreciated 🙏):
preferences.umiacs.umd.edu
0
2
3
reposted by
Navita Goyal
Nishant Balepur
about 1 year ago
🚨 New Position Paper 🚨 Multiple choice evals for LLMs are simple and popular, but we know they are awful 😬 We complain they're full of errors, saturated, and test nothing meaningful, so why do we still use them? 🫠 Here's why MCQA evals are broken, and how to fix them 🧵
2
46
13
reposted by
Navita Goyal
Mohit Iyyer
about 1 year ago
How can we generate synthetic data for a task that requires global reasoning over a long context (e.g., verifying claims about a book)? LLMs aren't good at *solving* such tasks, let alone generating data for them. Check out our paper for a compression-based solution!
add a skeleton here at some point
0
17
4
reposted by
Navita Goyal
Joe Stacey
about 1 year ago
This paper is really cool. They decompose NLI (and defeasible NLI) hypotheses into atoms, and then use these atoms to measure the logical consistency of LLMs. E.g. for an entailment NLI example, each hypothesis atom should also be entailed by the premise. Very nice idea 👏👏
2
15
3
reposted by
Navita Goyal
Hal Daumé III
about 1 year ago
Please join us for: AI at Work: Building and Evaluating Trust Presented by our Trustworthy AI in Law & Society (TRIALS) institute. Feb 3-4 Washington DC Open to all! Details and registration at:
trails.gwu.edu/trailscon-2025
Sponsorship details at:
trails.gwu.edu/media/556
0
16
7
reposted by
Navita Goyal
Hal Daumé III
over 1 year ago
The Impact of Explanations on Fairness in Human-AI Decision-Making: Protected vs Proxy Features Despite hopes that explanations improve fairness, we see that when biases are hidden behind proxy features, explanations may not help. Navita Goyal, Connor Baumler +al IUI’24
hal3.name/docs/daume23...
>
1
21
6
reposted by
Navita Goyal
Paola Cascante-Bonilla
over 1 year ago
This is my first time serving as an AC for a big conference. Just read this great work by Goyal et al.
arxiv.org/abs/2411.11437
I'm optimizing for high coverage and low redundancy—assigning reviewers based on relevant topics or affinity scores alone feels off. Seniority and diversity matter!
loading . . .
Causal Effect of Group Diversity on Redundancy and Coverage in Peer-Reviewing
A large host of scientific journals and conferences solicit peer reviews from multiple reviewers for the same submission, aiming to gather a broader range of perspectives and mitigate individual biase...
https://arxiv.org/abs/2411.11437
1
5
2
reposted by
Navita Goyal
Hal Daumé III
over 1 year ago
Large Language Models Help Humans Verify Truthfulness—Except When They Are Convincingly Wrong Should one use chatbots or web search to fact check? Chatbots help more on avg, but people uncritically accept their suggestions much more often. by Chenglei Si +al NAACL’24
hal3.name/docs/daume24...
>
1
30
5
you reached the end!!
feeds!
log in