Nicolas Yax
@nicolasyax.bsky.social
📤 65
📥 81
📝 30
PhD student working on the cognition of LLMs | HRL team - ENS Ulm | FLOWERS - Inria Bordeaux
pinned post!
🔥Our paper PhyloLM got accepted at ICLR 2025 !🔥 In this work we show how easy it can be to infer relationship between LLMs by constructing trees and to predict their performances and behavior at a very low cost with
@stepalminteri.bsky.social
and
@pyoudeyer.bsky.social
! Here is a brief recap ⬇️
6 months ago
3
15
7
reposted by
Nicolas Yax
Stefano Palminteri
4 days ago
New (revised) preprint with
@thecharleywu.bsky.social
We rethink how to assess machine consciousness: not by code or circuitry, but by behavioral inference—as in cognitive science. Extraordinary claims still need extraordinary evidence. 👉
osf.io/preprints/ps...
#AI
#Consciousness
#LLM
0
16
5
reposted by
Nicolas Yax
Stefano Palminteri
5 months ago
🧠 New paper in Open Mind! We show that LLM-based reinforcement learning agents encode relative reward values like humans, even when suboptimal and display a positivity bias. Work led by William Hayes w/
@nicolasyax.bsky.social
doi.org/10.1162/opmi...
#AI
#LLM
#RL
loading . . .
Relative Value Encoding in Large Language Models: A Multi-Task, Multi-Model Investigation
Abtract. In-context learning enables large language models (LLMs) to perform a variety of tasks, including solving reinforcement learning (RL) problems. Given their potential use as (autonomous) decis...
https://doi.org/10.1162/opmi_a_00209
1
10
3
reposted by
Nicolas Yax
Milena Rmus
5 months ago
Preprint update, co-led with
@akjagadish.bsky.social
, with
@marvinmathony.bsky.social
, Tobias Ludwig and
@ericschulz.bsky.social
!
loading . . .
Generating Computational Cognitive Models using Large Language Models
Computational cognitive models, which formalize theories of cognition, enable researchers to quantify cognitive processes and arbitrate between competing theories by fitting models to behavioral data....
https://arxiv.org/abs/2502.00879
0
15
7
Curious about LLM interpretability and understanding ? We borrowed concepts from genetics to map language models, predict their capabilities, and even uncovered surprising insights about their training ! Come see my poster at
#ICLR2025
3pm Hall 2B #505 !
6 months ago
0
4
0
🔥Our paper PhyloLM got accepted at ICLR 2025 !🔥 In this work we show how easy it can be to infer relationship between LLMs by constructing trees and to predict their performances and behavior at a very low cost with
@stepalminteri.bsky.social
and
@pyoudeyer.bsky.social
! Here is a brief recap ⬇️
6 months ago
3
15
7
reposted by
Nicolas Yax
7 months ago
🚀 Introducing 🧭MAGELLAN—our new metacognitive framework for LLM agents! It predicts its own learning progress (LP) in vast natural language goal spaces, enabling efficient exploration of complex domains.🌍✨Learn more: 🔗
arxiv.org/abs/2502.07709
#OpenEndedLearning
#LLM
#RL
loading . . .
MAGELLAN: Metacognitive predictions of learning progress guide...
Open-ended learning agents must efficiently prioritize goals in vast possibility spaces, focusing on those that maximize learning progress (LP). When such autotelic exploration is achieved by LLM...
https://arxiv.org/abs/2502.07709v2
1
9
7
reposted by
Nicolas Yax
cedric
11 months ago
we are recruiting interns for a few projects with @pyoudeyer in bordeaux > studying llm-mediated cultural evolution with @nisioti_eleni @Jeremy__Perez > balancing exploration and exploitation with autotelic rl with @ClementRomac details and links in 🧵 please share!
1
6
6
reposted by
Nicolas Yax
Gautier Hamon
11 months ago
Putting some Flow Lenia here too
loading . . .
1
4
1
reposted by
Nicolas Yax
Gautier Hamon
11 months ago
1/⚡️Looking for a fast and simple Transformer baseline for your RL environment in JAX ? Sharing my implementation of transformerXL-PPO:
github.com/Reytuag/tran...
The implementation is the first to attain the 3rd floor and obtain advanced achievements in the challenging Craftax
loading . . .
1
3
1
🚨New preprint🚨 When testing LLMs with questions, how can we know they did not see the answer in their training? In this new paper we propose a simple out of the box and fast method to spot contamination on short texts with
@stepalminteri.bsky.social
and Pierre-Yves Oudeyer !
11 months ago
1
8
4
you reached the end!!
feeds!
log in