Wang Bill Zhu
@billzhu.bsky.social
π€ 93
π₯ 70
π 13
CS Ph.D. candidate @ USC,
https://billzhu.me
At
@naaclmeeting.bsky.social
this week! Iβll be presenting our work on LLM domain induction with
@thomason.bsky.social
on Thu (5/1) at 4pm in Hall 3, Section I. Would love to connect and chat about LLM planning, reasoning, AI4Science, multimodal stuff, or anything else. Feel free to DM!
5 months ago
0
4
4
π¨ New work! LLMs often sound helpfulβbut fail to challenge dangerous medical misconceptions in real patient questions. We test how well LLMs handle false assumptions in oncology Q&A. π Paper:
arxiv.org/abs/2504.11373
π Website:
cancermyth.github.io
π [1/n]
loading . . .
Cancer-Myth: Evaluating AI Chatbot on Patient Questions with False Presuppositions
Cancer patients are increasingly turning to large language models (LLMs) as a new form of internet search for medical information, making it critical to assess how well these models handle complex, pe...
https://arxiv.org/abs/2504.11373
6 months ago
1
3
0
reposted by
Wang Bill Zhu
Robin Jia
10 months ago
I'll be at
#NeurIPS2024
! My group has papers analyzing how LLMs use Fourier Features for arithmetic and how TFs learn higher-order optimization for ICL (led by
@deqing.bsky.social
), plus workshop papers on backdoor detection and LLMs + PDDL (led by
@billzhu.bsky.social
)
1
23
4
β¨ Excited to share our Chain-of-Questions paper
#EMNLP2023
: we develop a framework that trains *one T5 model* to robustly answer multistep questions by generating and answering sub-questions. Outperforms ChatGPT on DROP, HotpotQA and their contrast/adversarial sets.
about 2 years ago
2
3
2
you reached the end!!
feeds!
log in