Scientific Discovery in the Age of Artificial Intelligence
Wang, H., Fu, T., Du, Y. et al. Scientific Discovery in The Age of Artificial Intelligence, 2023 Nature 620 47–60.
- Comprehensive summary of topics discussed in the original publication
- Detailed overview of different machine learning techniques.
Learning Meaningful Representations of Scientific Data
Applications of AI analyses to tackle science problems are often limited by the scarcity of high-quality curated datasets for exploration. Frequently, real-world datasets are incomplete, contain inaccurate observations, and come at variable sample resolutions. Moreover, even in situations where homogeneous data are aplenty, data labeling for supervised learning presents a laborious time investment. Here is where self-supervised learning comes to the rescue.
Self-supervised learning (SSL) is a machine learning paradigm that enables a model to automatically generate labels for unstructured data, thereby eliminating the need for large pre-labeled datasets. Building upon a small set of accurate human-annotated data, SSL algorithms learn latent representations of the input data through an iterative procedure. The labels generated in the first iteration are treated as the ground truth in the second iteration and so on.
Generative SSL tries to predict masked portions of raw data (text, images, audio, or video) from unmasked segments by learning embeddings of their underlying shared information. On the other hand, contrastive SSL techniques involves defining positive and negative versions of an “anchor” (say, the concept of what a dog looks like). Using a notion of distance in feature space, the algorithm then looks to align positives (e.g., images of dogs) to the anchor while simultaneously repelling the negatives (e.g., images of cats).
Neural Operators
Scientific experiments typically involve discrete measurements of intrinsically continuous quantities. For example, consider wind velocity in flight dynamics or magnetic field strength in tokamak nuclear fusion reactors. Conventional neural networks assume a fixed data discretization and are hence, inflexible at handling raw data sampled at varying resolutions. Neural operators, in contrast, learn mappings between function spaces of the input and output data to allow discretization-invariant predictions. Once these operators are trained, they can be evaluated at any data resolution without a need for model retraining.
- Neural Operators in PyTorch
- Video: Anima Anandkumar, Neural operator: A new paradigm for learning PDEs.
- FDL 2021 Live Showcase video: Digital Twin Earth: Coasts.
- Kovachki, N., Li, Z., et al., Neural Operator: Learning Maps Between Function Spaces, 2023 JMLR 4(89):1−97.