Marco 🌳 Zocca<p>Interested in interpretable ML, particularly for LLMs? </p><p>eg "causal" interpretability, as in the "OthelloGPT" paper [1]? </p><p>Let's connect!</p><p>1. <a href="https://arxiv.org/abs/2210.13382" target="_blank" rel="nofollow noopener" translate="no"><span class="invisible">https://</span><span class="">arxiv.org/abs/2210.13382</span><span class="invisible"></span></a> </p><p><a href="https://sigmoid.social/tags/ai" class="mention hashtag" rel="tag">#<span>ai</span></a> <a href="https://sigmoid.social/tags/machinelearning" class="mention hashtag" rel="tag">#<span>machinelearning</span></a> <a href="https://sigmoid.social/tags/interpretability" class="mention hashtag" rel="tag">#<span>interpretability</span></a> <a href="https://sigmoid.social/tags/interpretableml" class="mention hashtag" rel="tag">#<span>interpretableml</span></a> <a href="https://sigmoid.social/tags/mechanisticinterpretability" class="mention hashtag" rel="tag">#<span>mechanisticinterpretability</span></a></p>