Skip to content

Commit

Permalink
Update index.html
Browse files Browse the repository at this point in the history
  • Loading branch information
schwettmann authored Mar 20, 2024
1 parent 2de04da commit 7b7b198
Showing 1 changed file with 10 additions and 11 deletions.
21 changes: 10 additions & 11 deletions index.html
Original file line number Diff line number Diff line change
Expand Up @@ -123,23 +123,22 @@ <h1 class="title is-1 publication-title">A Multimodal Automated Interpretability
<div class="container is-max-desktop">
<!-- Abstract. -->
<div class="columns is-centered">
<div class="content">
<h2 class="title is-3 has-text-centered">MAIA</h2>
<div class="content">
<div style="display: flex; align-items: center;">
<img src="./static/figures/MAIA_schematic.png" alt="MAIA Schematic" style="margin-right: 20px; width: 40%;">
<p style="text-align: justify;">
We describe MAIA, a Multimodal Automated Interpretability Agent. MAIA is a system that uses neural models to automate neural model understanding tasks like feature interpretation and failure mode discovery.
It equips a pre-trained vision-language model with a set of tools that support iterative experimentation on subcomponents of other models to explain their behavior. These include tools commonly used by human interpretability researchers: for synthesizing and editing inputs, computing maximally activating exemplars from real-world datasets, and summarizing and describing experimental results.
<i>Interpretability experiments</i> proposed by MAIA compose these tools to describe and explain system behavior.
</p>
</div>
<hr>
<div style="display: flex; align-items: center;">
<img src="./static/figures/MAIA_schematic.png" alt="MAIA Schematic" style="margin-right: 20px; width: 40%;">
<p style="text-align: justify;">
We describe MAIA, a Multimodal Automated Interpretability Agent. MAIA is a system that uses neural models to automate neural model understanding tasks like feature interpretation and failure mode discovery.
It equips a pre-trained vision-language model with a set of tools that support iterative experimentation on subcomponents of other models to explain their behavior. These include tools commonly used by human interpretability researchers: for synthesizing and editing inputs, computing maximally activating exemplars from real-world datasets, and summarizing and describing experimental results.
<i>Interpretability experiments</i> proposed by MAIA compose these tools to describe and explain system behavior.
</p>
</div>
<hr>
</div>
</div>
<!--/ Abstract. -->
</div>
</section>
</section>


<section class="hero teaser" style="margin-top: -5px;">
Expand Down

0 comments on commit 7b7b198

Please sign in to comment.