@sineadwilliamson.bsky.social
📤 45
📥 86
📝 1
Really glad to have been a part of this super cool project... LLMs can verbalize more than just a single confidence number, and we can evaluate their ability to do so!
add a skeleton here at some point
3 days ago
0
4
0
reposted by
Michael Kirchhof (ICML)
5 days ago
Many treat uncertainty = a number. At Apple, we're rethinking this: LLMs should output strings that reveal all information of their internal distributions. We find that Reasoning, SFT, CoT can't do it - yet. To get there, we introduce the SelfReflect benchmark.
arxiv.org/pdf/2505.20295
3
30
7
reposted by
Shubhendu Trivedi
about 1 month ago
Natural idea. Looks like a nice paper too.
arxiv.org/abs/2508.21184
loading . . .
BED-LLM: Intelligent Information Gathering with LLMs and Bayesian Experimental Design
We propose a general-purpose approach for improving the ability of Large Language Models (LLMs) to intelligently and adaptively gather information from a user or other external source using the framew...
https://arxiv.org/abs/2508.21184
1
25
6
reposted by
Maureen de Seyssel
4 months ago
Now that
@interspeech.bsky.social
registration is open, time for some shameless promo! Sign-up and join our Interspeech tutorial: Speech Technology Meets Early Language Acquisition: How Interdisciplinary Efforts Benefit Both Fields. 🗣️👶
www.interspeech2025.org/tutorials
⬇️ (1/2)
loading . . .
https://www.interspeech2025.org/tutorials
Your cookies are disabled, please enable them.
https://www.interspeech2025.org/tutorials
1
9
6
you reached the end!!
feeds!
log in