Abstract: Deep neural networks have rapidly grown in their ability to address complex learning problems. Consequently, they are being integrated into society with an ever-rising influence on all levels of human experience. This has resulted in a need to gain human-understandable insights in their internal mechanisms to ensure they operate ethically and reliably. The study and development of methods which can generate such insights constitutes the field of interpretable machine learning. The talk will focus on the use of dictionary learning methods to discover relevant patterns inside hidden layers of deep networks. Specifically, we will explore their application for post-hoc interpretation in two different styles and contexts: (1) Generating listenable interpretations for audio processing networks, (2) Understanding internal representations of large multimodal models.