Chris Olah | |
|---|---|
| Citizenship | Canadian [1] |
| Known for | Mechanistic interpretability; neural network interpretability and visualization; DeepDream; activation atlases |
| Scientific career | |
| Fields | Machine learning |
| Institutions | Anthropic OpenAI Google Brain |
Chris Olah is a machine learning researcher and a co-founder of Anthropic. [2] [3] He is known for his work on neural network interpretability, particularly mechanistic interpretability, and for research and tools that visualize internal representations in neural networks. [2] [4] [5]
Olah is Canadian. [1] According to Wired, he left university at age 18 without earning a degree and later received a Thiel Fellowship, which supported him in pursuing independent work. [1]
Olah has worked on interpretability research at Google Brain, OpenAI, and Anthropic. [2] [3] Time described him as one of the pioneers of mechanistic interpretability and noted that he pursued this research line first at Google, then at OpenAI, and later at Anthropic which he co-founded. [2]
Wired reported that Olah was involved in early neural-network visualization work including DeepDream in 2015, as part of efforts to better understand what neural networks learn. [6] Later coverage linked him to more structured interpretability approaches such as "activation atlases". [4] The Verge covered activation atlases as a collaboration between Google and OpenAI researchers to help inspect neural-network representations. [7] [8]
At Anthropic, Olah has been identified in major press coverage as leading interpretability work aimed at mapping internal "features" in large language models and relating interpretability findings to AI safety. [9] [3] Quanta Magazine has also quoted Olah in reporting on interpretability and the internal structure of modern language models. [5]
Time included Olah in its TIME100 AI list in 2024. [2]