Man Is a Reed That Thinks. So Should AI.
“Man is but a reed, the weakest thing in nature; but he is a thinking reed.” — Blaise Pascal
Humanity’s strength lies in the ability to think, reflect, and understand. As Artificial Intelligence becomes increasingly capable of processing vast amounts of information, Pascal’s words remain ever relevant: True intelligence should not be defined solely by instrumental capacity, but also by the capacity for philosophical reflection and introspective reasoning.
My research at AI Interpretability @ Illinois extends this idea toward machine intelligence by exploring how Mechanistic Interpretability, Modularity, and Biological Plausibility can inspire the design of next-generation AI systems that go beyond pattern recognition and behave as a “thinking reed”:
- Structural Interpretability of AI Models: I study the organization of knowledge and computational mechanisms within AI models, borrowing core principles from biological intelligence—such as modularity, functional specialization, and sparse activation—to design architectures whose computation can be understood and regulated with brain-like efficiency.
- Behavioral Interpretability of AI Agents: I work on autonomous and self-evolving agents that learn on the fly through continual interaction with changing environments. A self-reflective, continuously updated memory module allows the agent to integrate past experiences, representing a key step toward “System-2 thinking”; while modeling collaboration and competition in multi-agent frameworks can induce Pareto-efficient systems that resemble competitive economic markets.
