|国家预印本平台
首页|Representation Learning on a Random Lattice

Representation Learning on a Random Lattice

Representation Learning on a Random Lattice

来源:Arxiv_logoArxiv
英文摘要

Decomposing a deep neural network's learned representations into interpretable features could greatly enhance its safety and reliability. To better understand features, we adopt a geometric perspective, viewing them as a learned coordinate system for mapping an embedded data distribution. We motivate a model of a generic data distribution as a random lattice and analyze its properties using percolation theory. Learned features are categorized into context, component, and surface features. The model is qualitatively consistent with recent findings in mechanistic interpretability and suggests directions for future research.

Aryeh Brill

计算技术、计算机技术

Aryeh Brill.Representation Learning on a Random Lattice[EB/OL].(2025-04-28)[2025-05-23].https://arxiv.org/abs/2504.20197.点此复制

评论