ARCHETYPE AI

0

Can AI Learn Physics from Sensor Data?

Archetype AI Team
  • October 17, 2024

The physical world often feels complex and chaotic, yet humans have learned to discover the laws of nature that govern it—such as mechanics, thermodynamics, and more. We have achieved this through repeated observations and precise measurements of different physical phenomena, e.g., the motion of objects or temperature of gases. But what if AI could also uncover the governing laws of the physical world by analyzing a variety of sensor data on its own, without human guidance? What if AI could help us characterize highly complex physical system behaviors that we are unable to formally describe using equations, such as the behaviors of electrical grids, industrial machinery, or even human-computer interaction?

We are excited to share a milestone in our journey toward developing a physical AI foundation model. In a recent paper by the Archetype AI team,           A Phenomenological AI Foundation Model for Physical Signals, we demonstrate how an AI foundation model can effectively encode and predict physical behaviors and processes it has never encountered before, without being explicitly taught underlying physical principles. Read on to explore our key findings.

What it means for AI to understand the real world

We often assume that to understand the physical world, we need to learn the laws of physics first before applying them to tasks at hand. Naturally, we might think that AI also needs to be taught these same laws—such as the conservation of energy—so it can process data from the physical world through the lens of these laws. This approach often relies on the introduction of inductive biases, which are assumptions representing constraints and prior knowledge as mathematical statements.

Newton, trained on millions of sensor measurements, predicts unfamiliar physical systems, like the spring-mass system shown here. Its zero-shot capabilities handle both complex and chaotic behaviors across classic and real-world systems.

However, the problem with this approach is that it leads to highly specialized AI models. For example, if we train a model to analyze fluid motion using the Navier–Stokes equations, it wouldn't be able to interpret radar Doppler images, which involve very different physical principles. Moreover, complex systems — like the electrical grid or automated industrial machines — can't be easily described using just a few physical laws. Defining such systems with simple equations is often extremely challenging.

But what if, instead of teaching AI the laws of the physical world, we let it discover them on its own, directly from data? This is, in fact, how humans uncovered many of the fundamental laws of nature. Think about how Johannes Kepler figured out planetary motion or how Georg Ohm derived the laws of electricity. They didn't know the laws beforehand — they observed, measured, and analyzed the behavior of planets and electricity to reveal hidden, recurring patterns, which we now call the laws of nature. These patterns not only explained their current observations but also enabled them to predict future behaviors, such as the movement of planets or the flow of electrical currents. Can AI do the same?

The butterfly effect in the age of AI

The challenge of understanding the physical world, whether for humans or AI, is that we cannot experience or learn about it directly — we can only observe it indirectly through sensors. These sensors might be our natural biological ones, like our eyes and ears, or the countless artificial sensors humans have invented to measure everything from acceleration to gas concentration. However, sensors always affect the outcome, distorting or obscuring the "true" physical behavior, making it harder to uncover the underlying laws that govern the physical world.

Figure 1. This diagram outlines training an AI model for inference from sensor observations.

The true measure of whether an AI model has genuinely learned the governing behaviors of the physical world, rather than just mimicking sensor data, is its ability to generalize beyond the data it was trained on. For example, if you train a model on data from the electric motor in your car, could it then predict noise pollution in San Francisco? That means that the model should be able to accurately describe physical systems it has never encountered before, even when those systems are measured by sensors unfamiliar to the model.

In other words, if you trained a model on data from a butterfly flapping its wings, could it forecast hurricane dynamics on the other side of the world?

Teaching Newton to decode the real world

Recent advancements in generative AI techniques — such as ChatGPT for natural language processing or image models like Vision Transformers and CLIP — show that this level of generalization is possible. But the question remains: can we apply these same principles to data from the physical world? Can we create AI models that, using only raw measurements and without any prior knowledge of a system's inner workings, can predict the future behavior of any physical system?

To address this challenge, we have pre-trained our physical AI model, Newton, on 0.59 billion samples from open-source datasets covering a wide range of physical behaviors, from electrical currents and river fluid flows to optical sensors. Using a transformer-based deep neural network, Newton encodes all this raw, noisy sensor data and attempts to make sense of it by uncovering hidden patterns and statistical distributions in the universal latent space representing all these measurements.

Next, several lightweight, application-specific neural network decoders were trained to make use of the condensed insights produced by the encoder. These decoders are trained to perform real world tasks, such as predicting future outcomes based on sensor data or reconstructing past events. In practical applications, Newton can take in real-time data from sensors measuring physical behaviors, or it can work with prerecorded sensor measurements to make accurate predictions.

Learning to predict the future of physical systems

Before moving on to complex real world applications, we tested Newton using simple physics experiments that most people are familiar with from school, such as basic mechanical oscillation and thermodynamics experiments.

Figure 2. Examples of model predicting complex real world processes. Top: forecasting Turkey's power consumption. Bottom: forecasting the oil temperature of an electrical transformer.

In both cases, the Newton model received real-time data from sensors and was able to accurately predict the behavior of the physical systems simply by observing the concurrent sensor data. What's remarkable is that Newton had not been specifically trained to understand these experiments – it was encountering them for the first time and was still able to predict outcomes even for chaotic and complex behaviors. This ability of an AI model to accurately predict data it has not encountered previously is often referred to as zero-shot forecasting.

While these classic experiments were exciting, real world systems are far more complex and harder to describe. We then tested Newton on predicting the behavior of such systems as city electrical demand, daily temperature, and temperature in electrical transformers, to name a few. Figure 2 demonstrates that Newton was able to accurately zero-shot forecast the behavior of these complex systems with no additional training data for systems that are challenging even for humans to model.

Figure 3. Comparison of forecasting mean square error for the zero-shot, fine-tuned, and target-trained models. Lower is better.

What’s exciting is that zero-shot forecasting consistently outperforms even when Newton is specifically trained on only the target dataset. In other words, Newton, trained on a vast amount of physical sensor data from all over the world, knows more about the temperature of oil in transformers than when we train Newton specifically on transformer oil data. This suggests that foundation models like Newton have powerful generalization capabilities, enabling them to understand physical behaviors far beyond the specific data they were originally trained on.

Practical implications

Physical AI models that can independently understand the physical world from sensor data have significant practical implications:

One model – many use cases: One of the most important implications is that we no longer need to train a separate AI model for every specific physical application. We can apply a single model like Newton across various domains and use cases, dramatically speeding up and simplifying the deployment of AI to solve real world problems that matter.

Less training data for new use cases: A pre-trained physical AI foundation model that can accurately analyze sensor data with minimal additional training becomes invaluable in situations where collecting large datasets is difficult or impossible. This could lead to sensor-agnostic physical AI platforms that can work with specialized sensors straight out of the box, without extensive retraining.

Less computing power for new use cases: The architecture we propose – training a foundational encoder with lightweight, application-specific decoders – enables the model to be easily adapted for different applications without major modifications to its core structure. This makes it faster and more efficient to build new use cases, reducing both computational resources and costs.

Autonomous learning from observations: Newton's self-supervised training suggests it can learn physical behaviors directly from observational data. This opens the door to highly autonomous systems that can adapt to new environments or requirements without needing explicit human intervention, such as manual data labeling or model fine-tuning.

There are many more profound implications and exciting possibilities with Newton, from building automotive monitoring systems to self-adapting robots, or even driving scientific discovery by uncovering new physical laws. Indeed, if AI can discern the laws of nature through observation alone, what other undiscovered patterns could it reveal about the universe?

Interested to learn more? Read "A Phenomenological AI Foundation Model for Physical Signals" here.

Recommended posts

Currently, digital companions are the dominant metaphor for understanding AI systems. However, as the field of generative AI continues to evolve, it's crucial to examine how we frame and comprehend these technologies—it will influence how we develop, interact with, and regulate AI. In this blog post, we'll explore different metaphors used in AI products and discuss how they shape our mental models of AI systems.

September 25, 2024

Implementing AI in industrial settings comes with significant challenges like ensuring employee safety, estimating productivity, and monitoring hazards—all requiring real-time processing. However, sending sensor data to the cloud for analysis introduces latency and security concerns, driving up costs. The solution? Eliminate the cloud. With Archetype AI’s Newton foundation model, AI can run on local machines using a single off-the-shelf GPU, delivering low latency, high security, and reduced costs in environments like manufacturing, logistics, transportation, and construction.

September 5, 2024

We’re building the first AI foundation model that learns about the physical world directly from sensor data, with the goal of helping humanity understand the complex behavior patterns of the world around us all.

November 1, 2023

At Archetype, we want to use AI to solve real world problems by empowering organizations to build for their own use cases. We aren’t building verticalized solutions –instead, we want to give engineers, developers, and companies the AI tools and platform they need to create their own solutions in the physical world.