Can Demircan
@candemircan.bsky.social
๐ค 56
๐ฅ 259
๐ 8
phd student in Munich, working on machine learning and cognitive science
reposted by
Can Demircan
Taylor Webb
7 months ago
LLMs have shown impressive performance in some reasoning tasks, but what internal mechanisms do they use to solve these tasks? In a new preprint, we find evidence that abstract reasoning in LLMs depends on an emergent form of symbol processing
arxiv.org/abs/2502.20332
(1/N)
loading . . .
Emergent Symbolic Mechanisms Support Abstract Reasoning in Large Language Models
Many recent studies have found evidence for emergent reasoning capabilities in large language models, but debate persists concerning the robustness of these capabilities, and the extent to which they ...
https://arxiv.org/abs/2502.20332
4
114
36
reposted by
Can Demircan
Mirko Thalmann
7 months ago
Every experience is unique ๐ light shifts, angles change, yet we recognize objects effortlessly. How do our minds do this? And (how) do they differ from machines? In our new preprint with
@ericschulz.bsky.social
, we review human generalization and compare it to machine generalization:
osf.io/k6ect
0
7
8
reposted by
Can Demircan
Milena Rmus
7 months ago
About a month late posting this, but here's a new project with
@ericschulz.bsky.social
,
@akjagadish.bsky.social
,
@marvinmathony.bsky.social
and Tobias Ludwig We are using LLMs to propose cognitive models in learning and decision making data. Presenting this work at RLDM!
arxiv.org/abs/2502.00879
loading . . .
Towards Automation of Cognitive Modeling using Large Language Models
Computational cognitive models, which formalize theories of cognition, enable researchers to quantify cognitive processes and arbitrate between competing theories by fitting models to behavioral data....
https://arxiv.org/abs/2502.00879
0
21
12
reposted by
Can Demircan
7 months ago
In previous work we found that VLMs fall short of human visual cognition. To make them better, we fine-tuned them on visual cognition tasks. We find that while this improves performance on the fine-tuning task, it does not lead to models that generalize to other related tasks:
1
8
7
reposted by
Can Demircan
Marcel Binz
8 months ago
We are currently building the largest, cross-domain data set of human behavior as part of an open collaborative project. Contributions of any form are welcome, but especially experiments with meta-data from developmental, cross-cultural, or clinical studies. More details:
github.com/marcelbinz/P...
loading . . .
GitHub - marcelbinz/Psych-201
Contribute to marcelbinz/Psych-201 development by creating an account on GitHub.
https://github.com/marcelbinz/Psych-201
2
34
16
reposted by
Can Demircan
Eric Schulz
8 months ago
In the first paper,
@candemircan.bsky.social
and
@tankred-saanum.bsky.social
use sparse autoencoders to show that LLMs can implement temporal difference learning in context. This work is together with Akshay Jagadish and
@marcelbinz.bsky.social
.
arxiv.org/abs/2410.01280
loading . . .
Sparse Autoencoders Reveal Temporal Difference Learning in Large Language Models
In-context learning, the ability to adapt based on a few examples in the input prompt, is a ubiquitous feature of large language models (LLMs). However, as LLMs' in-context learning abilities continue...
https://arxiv.org/abs/2410.01280
1
1
1
Alignment is more than comparing similarity judgments! How well do pretrained neural networks align with humans in few-shot learning settings? Come check our poster #3904 at
#NeurIPS
on Wednesday to find out
10 months ago
1
13
6
reposted by
Can Demircan
Mona Garvert
10 months ago
๐จJoin our team! Weโre hiring a PhD student in Cognitive & Clinical Neuroscience ๐ง ๐ at @uni_wue & @UKW_Wuerzburg! Explore mechanisms of decision-making in healthy people & Parkinsonโs using new deep brain stimulation methods. German & English required. Apply by 20 Dec! ๐๐ Details:
shorturl.at/IcNa0
loading . . .
PhD_Wessel_Garvert.pdf
https://shorturl.at/IcNa0
2
27
27
you reached the end!!
feeds!
log in