Lukas Aichberger
@aichberger.bsky.social
๐ค 207
๐ฅ 172
๐ 9
Machine Learning ELLIS PhD at Johannes Kepler University Linz and University of Oxford
reposted by
Lukas Aichberger
Yarin
9 months ago
Hot take: I think we just demonstrated the first AI agent computer worm ๐ค When an agent sees a trigger image it's instructed to execute malicious code and then share the image on social media to trigger other users' agents This is a chance to talk about agent security ๐
add a skeleton here at some point
0
8
2
โ ๏ธ Beware: Your AI assistant could be hijacked just by encountering a malicious image online! Our latest research exposes critical security risks in AI assistants. An attacker can hijack them by simply posting an image on social media and waiting for it to be captured. [1/6] ๐งต
loading . . .
9 months ago
1
8
11
reposted by
Lukas Aichberger
12 months ago
Often LLMs hallucinate because of semantic uncertainty due to missing factual training data. We propose a method to detect such uncertainties using only one generated output sequence. Super efficient method to detect hallucination in LLMs.
add a skeleton here at some point
0
15
5
๐ก๐ฒ๐ ๐ฃ๐ฎ๐ฝ๐ฒ๐ฟ ๐๐น๐ฒ๐ฟ๐: Rethinking Uncertainty Estimation in Natural Language Generation ๐ Introducing ๐-๐ก๐๐, a theoretically grounded and highly efficient uncertainty estimate, perfect for scalable LLM applications ๐ Dive into the paper:
arxiv.org/abs/2412.15176
๐
loading . . .
Rethinking Uncertainty Estimation in Natural Language Generation
Large Language Models (LLMs) are increasingly employed in real-world applications, driving the need to evaluate the trustworthiness of their generated text. To this end, reliable uncertainty estimatio...
https://arxiv.org/abs/2412.15176
12 months ago
0
9
6
you reached the end!!
feeds!
log in