Andrew Lampinen<p>Research in mechanistic interpretability and neuroscience often relies on interpreting internal representations to understand systems, or manipulating representations to improve models. I gave a talk at the UniReps workshop at NeurIPS on a few challenges for this area, summary thread: 1/12<br><a href="https://sigmoid.social/tags/ai" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>ai</span></a> <a href="https://sigmoid.social/tags/ml" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>ml</span></a> <a href="https://sigmoid.social/tags/neuroscience" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>neuroscience</span></a> <a href="https://sigmoid.social/tags/computationalneuroscience" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>computationalneuroscience</span></a> <a href="https://sigmoid.social/tags/interpretability" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>interpretability</span></a> <a href="https://sigmoid.social/tags/NeuralRepresentations" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>NeuralRepresentations</span></a> <a href="https://sigmoid.social/tags/neurips2023" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>neurips2023</span></a></p>