Bin Dong(董彬) Long CV,  Short CV

Professor

Beijing International Center for Mathematical Research (BICMR),

Center for Machine Learning Research (CMLR)

Peking University

Office: BICMR77101; Phone: +8610-62744091
Email: dongbin {at} math {dot} pku {dot} edu {dot} cn

===============================================================================================

About -- Biography -- Events -- Publications -- Teaching

===============================================================================================

 

Research Philosophy

My research philosophy has consistently been driven by the dual aspiration of devising effective models and algorithms to address real-world challenges while distilling novel mathematical problems during this process. In essence, I aim to contribute to both practical applications and the advancement of mathematics itself. This dualism has served as the cornerstone of my research endeavors.

 

Areas of Focus

My research has primarily focused on biomedical imaging, with a particular emphasis on image reconstruction, image analysis, and using images and medical data to aid in clinical decision-making.  More recently, I have also begun working extensively in machine learning and artificial intelligence, particularly in the development of foundation models for biomedical data analysis, automated reasoning, and scientific computing. In particular, I have recently directed my primary efforts toward fully advancing AI for Mathematics (AI4M), an emerging and exciting research direction.

 

Earlier Career

During my doctoral and postdoctoral years, I have developed various models and algorithms within the frameworks of Partial Differential Equations (PDEs) and wavelets. These include designing PDE models for cerebral aneurysm segmentation to facilitate subsequent geometric measurement (Ref); integrating the notion of PDEs, level-set methods and wavelet transforms to create computational tools for cardiovascular plaque quantification (Ref); formulating CT image reconstruction models by exploiting the property of sparse approximation of wavelet frame transforms (Ref1, Ref2, Ref3); and optimizing radiation therapy for cancer mathematically (Ref) which was later reported by phys.org.

 

Unveiling the Connection: PDEs and Wavelets

Throughout the course of addressing these applied problems, I became increasingly aware of the striking algorithmic resemblance between PDEs and wavelets. This led to an investigative journey with collaborators to explore any intrinsic relationships between these two paradigms, which had evolved independently in the field of imaging for nearly three decades. We discovered profound connections, unsettling traditional views and establishing an asymptotic convergence relationship between wavelet-based optimization models and PDE models. (Ref1, Ref2, Ref3, Ref4) This theoretical insight effectively bridged the geometric interpretations of differential operators with the sparse approximations and multi-scale nature of wavelets, inspiring new models that amalgamated the strengths of both PDEs and wavelets. In particular, this insight has led to new wavelet frame-based image segmentation models for medical images (Ref). Notably, our collaborative paper on wavelet frame-based segmentation of ultrasound videos (Ref) with my colleagues at Shanghai Jiao Tong University and National University of Singapore gained media attention, including coverage by SIAM News, Science Daily, etc.

 

Journey into Machine Learning

My initial venture into machine learning commenced in 2014 resulting in a sole-authored paper that introduced wavelet frames into semi-supervised learning models on graphs and point clouds (Ref), as well as a work on sparse linear discriminant analysis with my colleagues from the statistics program at the University of Arizona (Ref). However, it was not until 2015-2016 that I fully grasped the revolutionary impact of deep learning, particularly on biomedical imaging problems. It was also during this period that our earlier works on unveiling the connections between PDEs and wavelets contributed to a unified perspective on these mathematical models, which later informed our mathematical understanding of deep neural networks.

Subsequently, I led my team to pivot our focus towards deep learning methodologies. The core motive has been to explore the algorithmic relations between traditional mathematical methods like PDEs and wavelets and the data-driven models in deep learning. This exploration culminated in two important works of mine between 2017 and 2018: the establishment of the link between numerical differential equations and deep neural network architectures, leading to the development of ODE-Net (Ref) and PDE-Net (Ref1, Ref2), which were empirically validated on large-scale datasets. Through this and several subsequent research studies, we gradually developed a novel generic methodology for synergistically integrating mathematical modeling and machine learning techniques. We have been constantly practicing and refining this generic methodology in my research endeavors ever since (e.g. examples in Detour to Scientific Computing).

The core focus of my presentation at the 2022 International Congress of Mathematicians (ICM) was to elucidate the intricate connections between PDEs and wavelets, as well as the links between discrete differential equations and deep neural networks. These bridges have intriguing implications for the fields of computational imaging and scientific computing, which were thoroughly discussed in my talk and the corresponding proceedings and notes (Ref1, Ref2).

 

         Machine Learning for Scientific ComputingUpon understanding deep neural networks through the lens of numerical differential equations, it became natural to tackle challenges in scientific computing, specifically, PDE-based forward and inverse problems. For the forward problems, we blended numerical methods for conservation laws and multigrid methods with deep reinforcement learning and meta-learning. (Ref1, Ref2) More recently, we discovered a connection between meta-learning and reduced-order modeling, paving the way for a new technique in approximating the solution manifold of parametric PDEs. (Ref1, Ref2) For inverse problems, our focus has been primarily on imaging, where we proposed several ways of melding deep learning techniques with traditional mathematical algorithms (Ref1, Ref2, Ref3, Ref4).

 

         Clinical Applications and Translational ResearchSince 2019, I have embarked on an extensive collaboration with Peking University Cancer Hospital, focusing on clinical auxiliary diagnostic issues. This has resulted in a series of exciting progresses in medical image analysis, contributing valuable tools for clinical practice that have seen successful adoption and some degree of translational implementation.(Ref1, Ref2, Ref3, Ref4, Ref5)

 

AI for Mathematics: Toward Rigorous, Verifiable Intelligence

AI for Mathematics (AI4Math) is, in my view, best understood against the long-term evolution of mathematical research itself. Historically, mathematics has been driven by two intertwined pillars: axiom-based deductive reasoning, which underlies much of foundational mathematics, and scientific computing, where numerical methods enable the quantitative exploration of complex systems across the natural and engineering sciences. The rapid development of modern AI is now reshaping both pillars in a profound way, marking a broader transition from problem-specific modeling toward increasingly general-purpose modeling. In scientific computing, data-driven approachesranging from high-dimensional function approximation and operator learning to architectures that fuse physical structure with dataare changing how we approach high-dimensional PDEs, multiscale modeling, and challenging inverse problems (including computational imaging and biomedical data analysis). In deductive reasoning, long-standing barriers caused by combinatorial explosion in symbolic logic and search are being revisited through the emerging synergy between large language models and formal systems, opening new possibilities in automated theorem proving, auto-formalization, and computer-assisted exploration of abstract mathematical structures. Together, these developments suggest that computational tools are no longer confined to numerical simulation; they are beginning to participateat least in constrained, verifiable waysin logical derivation and structural discovery, offering new pathways toward mathematical problems of higher dimensionality and complexity.

In the past few years, as I began engaging more deeply with this emerging landscape, I also became increasingly aware of a fundamental gap between the impressive breadth of todays general-purpose AI systems and the specific demands of mathematical research. While foundation models can perform remarkably well on a wide range of multimodal and language tasks, using them directly in mathematics exposes clear bottlenecks at a more structural level: on the reasoning side, the lack of strict logical constraints makes hallucinations and ungrounded leaps difficult to avoid; on the computation side, the sensitivity required to perceive and respect conservation laws, invariances, and high-precision numerical mechanisms remains limited. In this sense, existing general models are not designed for mathematics as a native habitat, and simply applying them often fails to reach the core of what mathematicians and scientific computing practitioners truly need. This realization led me to a more ambitious commitment: rather than treating AI as an external tool to be adopted, I became convinced that we must build a mathematics-native intelligence substrateone that integrates rigorous logical verification and high-fidelity numerical feedback into a unified environment. Such environments offer unusually strong and unambiguous supervision signalswhether through the deterministic state-checking and proof validation in formal systems such as Lean, or through invariant structures and energy-based constraints in physical and mathematical modelsproviding what I see as a principled pathway to restrain model stochasticity and push beyond the capability boundary of general-purpose AI.

Building on this perspective, I have come to view AI4M not as a one-way application of AI to mathematics, but as an inherently bidirectional and self-reinforcing processone that strengthens AI with mathematics, and in turn advances mathematics with AI. On the one hand, mathematics offers AI a uniquely rigorous cognitive scaffold: the way mathematicians learn, generalize, and verify can inspire more structured training curricula, richer evaluation axes (beyond surface accuracy), and principled approaches to interpretability that probe the internal mechanisms of reasoning rather than merely its outputs. On the other hand, as models become increasingly grounded in logical verification and mechanistic constraints, they can begin to serve as more than productivity toolsevolving toward research partners that help explore unfamiliar proof pathways, translate informal ideas into formal statements, and tackle high-dimensional scientific computing problems that strain classical methods. In my mind, the long-term promise of AI4M lies precisely in this closed loop: mathematical certainty and structure providing the hard feedback that reshapes AI systems, while increasingly capable AI systems return that value by expanding the frontier of what is feasible in both deductive discovery and computational mathematics.

(See our recent review, and the very first in literature, on AI4Math. AI for Mathematics: Progress, Challenges, and Prospects)

What we have built so far reflects this unified vision:

         Scientific Computing: In the computation pillar of AI4M, my group has been building foundation models for PDE solving that can efficiently support a wide range of downstream tasks. A key milestone is that we first observed, in the 1D setting, a scaling law for learning a general PDE solution operatorsuggesting that operator-level generalization may emerge predictably with increasing data/model/compute (Ref1, Ref2). Building on this, we have recently extended the model to 2D (Ref), and are continuing to broaden the pretraining distribution while actively seeking concrete real-world application scenarios where such a PDE foundation model can deliver clear advantages over classical pipelines.

         AI inspiring pure mathematics research: On the reasoning pillar, our work has explored how AI can meaningfully participate in highly abstract mathematical discoverynot by replacing proof, but by guiding human intuition toward structures that are difficult to see unaided. One representative example arises from arithmetic geometry and representation theory, where affine DeligneLusztig varieties (ADLV) play a central role in the geometric realization of the Langlands program. The dimension computation of ADLV, as a key invariant, depends on intricate combinatorial structures of affine Weyl group elements and has long lacked a truly general explicit characterization. By analyzing large-scale geometric/combinatorial data with machine learning models, we not only rediscovered the classical dimension formula, but also detected systematic patterns of deviation in boundary regimes where existing theory fails, and further proved a new theorem giving a lower bound on the actual dimension, filling a long-standing quantitative gap in the field. In my view, this provides an early and concrete validation of the AI4M path of humanAI collaborative exploration in deep, abstract areas of algebra and geometry. (Ref)

         Formal reasoning infrastructure: To make AI-assisted deductive mathematics reliable rather than merely fluent, we have built an end-to-end infrastructure around Lean that connects retrieval, data/benchmarks, and neuro-symbolic provingaddressing hallucination under weak constraints, scarcity of formal data, and the difficulty of proof search. Our approach is intentionally end-to-end: we connect retrieval (to ground generation), data/benchmarks (to train and measure), and neuro-symbolic proving (to close the generateverify loop) into a single, continuously improving ecosystem.

    Mathematical Information Retrieval (MIR): We have been building retrieval infrastructure that spans both formal and informal sides of mathematical knowledge. On the formal side, LeanSearch targets cross-modal semantic alignment between natural language and formal code (Ref), so that informal queries (or partially formalized proof states) can retrieve relevant theorems, lemmas, and proof patterns from large formal libraries; by being deeply integrated into Lean's official mathlib4 ecosystem, LeanSearch serves not only as a user-facing tool for mathematicians, but also as an enabling component for retrieval-augmented autoformalization and prover agents. On the informal side, we developed Matlas, a semantic retrieval system indexing millions of mathematical statements harvested from arXiv and other open resources, enabling large-scale cross-domain discovery of non-trivial connections between seemingly distant subfields (Ref). Together, LeanSearch and Matlas provide a principled way to anchor generation in existing verified knowledge and to surface latent structural links across the mathematical literature.

    Datasets and Benchmarks: Recognizing that progress in autoformalization and formal reasoning is constrained by both data quality and evaluation realism, we released Herald, an auto-formalization dataset oriented toward high-difficulty competition mathematics, together with a supporting translation/evaluation toolchain (Ref). Building on this, we developed Aria, an autoformalization agent that translates even more challenging informal problems into formally checkable statements, helping bridge the gap between model outputs and verifiable formal artifacts (Ref). In parallel, we introduced FATE (M/H/X) as an algebra-focused evaluation suite designed to probe formal reasoning beyond short, template-like proofsemphasizing depth, structure, and robustness under challenging settings (Ref). We also contributed to competition-grade data efforts through the Numina teamincluding the collection and curation of high-quality Olympiad-style problems for AIMO (Ref). Collectively, these resources serve two roles: (i) as training and diagnostic substrates for models intended to operate in both formal and informal environments, and (ii) as community benchmarks that make comparison and iteration scientifically meaningful.

    Neuro-symbolic formal proving: We further developed REAL-Prover (Ref), combining the intuition of large models with the rigor of symbolic solvers, and built the associated tactic-generation tool reap (Ref), which is being tested and maintained as an official candidate tactic within mathlib4.

         Math research agents: from informal exploration to verified proof. Building on the retrieval tools, data, and neuro-symbolic proving infrastructure above, we have been developing research-oriented agents that engage with open mathematical problems end-to-end, coupling informal exploration with formal verification. Rethlas is an informal mathematical research agent built on top of Matlas: given an open conjecture, it leverages large-scale semantic retrieval to surface candidate pathways and non-trivial connections across subfields, proposing informal proof sketches that human experts typically reach only through extended cross-disciplinary dialogue. Archon is its formal counterpart a Lean-based autoformalization and proving agent that turns informal proof sketches into machine-checkable formal artifacts, with LeanSearch integrated as a native tool in its decision loop.

As an independent demonstration of Archon's formal capability, we applied it to the AI-generated proofs by OpenAI on the FirstProof benchmark of ten research-level problems. Archon autonomously produced end-to-end Lean 4 formalizations (approximately 8,800 lines each, zero sorry) for Problems 4 and 6 notably the two problems that DeepMind's Aletheia had not been able to settle, suggesting a non-trivial level of formalization difficulty (technical report & outputs of Archon on P4 and P6).

As a further demonstration of this pipeline, Rethlas identified a non-trivial connection between quasi-completeness theory and Jensen's formal fiber construction, which Archon subsequently realized as a roughly 19,000-line Lean 4 proof (zero sorry) resolving D.D. Anderson's 8a conjecture, an open problem that had been standing since 2014 (technical report & outputs of the agents).

This line of work illustrates how retrieval, autoformalization, and neuro-symbolic proving can be composed into a single system capable of contributing verifiable progress on genuine research-level mathematics.

         Mechanistic interpretability: Parallel to building AI4Math systems, I have also been pursuing a more foundational motivation: to understand the inner workings of LLMs by examining their activity when they carry out high-level cognitive tasks, especially reasoning. Doing so requires interpretability tools that can decipher the meaning encoded in high-dimensional objectsactivations, attention pathways, and internal statesat a scale where systematic conclusions become possible. This motivation has led to two recent works. Hierarchical Sparse Autoencoders (HSAE) provide a scalable way to uncover multi-level conceptual hierarchies inside model representations, revealing how high-level notions may decompose into more fine-grained features across layers (Ref). InverseScope offers an assumption-light route to interpret activations by characterizing the space of inputs consistent with a target internal state, enabling quantitative testing of hypotheses about what information that state encodes (Ref). Together, these tools move interpretability beyond isolated features toward a more structured and testable understanding of reasoning-related representations, and we will continue improving them to support increasingly mathematical and symbolic accounts of how LLMs implement complex cognition.

 

Last updated: 04/2026.