AI and data science are frequently highlighted as game-changers for methane emissions reduction. You’ve likely seen headlines promising to “unlock the potential of your methane data”or blogs touting AI-driven insights. But there’s a fundamental issue often overlooked in these discussions: AI requires a robust data standard foundation. Without it, the promise of AI remains unfulfilled—or worse, its application can become misleading or even dangerous.
Methane data today is disorganized and siloed, plagued by:
• Inconsistent Definitions: Different organizations use varying terminologies and methodologies for data collection.
• Data Silos: Valuable datasets are locked in proprietary systems, limiting access and integration.
• Lack of Interoperability: Even when accessible, datasets often cannot be combined due to incompatible formats.
These challenges make advanced analysis—and, by extension, AI—difficult, if not impossible. Applying AI to such fragmented data leads to faulty outputs, wasted resources, and diminished trust in data-driven approaches. We cannot build a methane emissions foundation model without a standard methane data foundation.
Are there proven best practices from other fields that we can learn from? The Internet of Water offers valuable insights. This initiative improves the accessibility and usability ofwater data by adopting international sensor web standards. These standards enable seamless integration of data streams from diverse sources, such as the US Geological Survey, France’s BRGM, and the UK’s British Geological Survey.
Let me show you a demo and explain why. I opened my trusted ChatGPT and entered the following prompt.
By entering a simple prompt into ChatGPT, I received a fully functional code snippet. I copied itinto CodePen.io, and voilà! Without writing a single line of code, I created a visualization displaying water observations from multiple sources in one unified view.
Here’s what the demo reveals about the intersection of standards and AI:
The Internet of Water’s success demonstrates that adopting international standards lays a robust foundation for AI to access, interpret, and act on data. This alignment accelerates data usability and drives innovation.
However, even the Internet of Water community acknowledges they are still working to make their data fully FAIR: Findable, Accessible, Interoperable, and Reusable. By comparison, methane emissions data is in its infancy. For methane data to reach its potential, it must follow a similar path, starting with the adoption of foundational sensor web data standards.
Methane emissions reduction is fundamentally a data management challenge, as it relies on integrating a wide range of sensor data to enable early detection, prompt repairs, and efficient containment of the gas within the pipeline.
Before we can train sophisticated models or leverage advanced tools like LLMs, we need to address:
AI thrives on high-quality, consistent, and interoperable data. Applying AI to methane emissions without a solid data foundation will not yield meaningful insights. Instead, it risks:
The success of AI in methane emissions depends on establishing and adhering to FAIR principles. Interoperability and standardization are prerequisites for any data-driven innovation, not optional enhancements.
The methane data ecosystem needs to prioritize:
AI has tremendous potential, but only when built on a foundation of standardized, high-quality data. By focusing on these fundamentals, we can unlock the true power of methane data — and make meaningful progress toward emissions reduction.
Standardizing methane data is essential for unlocking AI’s full potential — there’s no need to reinvent the wheel by developing proprietary formats and making the problem even worse. So, what’s being done to address these challenges?
The Open Geospatial Consortium (OGC) is forming the Emission Event Modeling Language Working Group (EmissionML) to bridge the interoperability gap between emissions data, sensor observations, and their geospatial sources. Expected to launch in March 2025, this initiative will define standards addressing methane data management challenges, ensuring seamless data exchange across platforms while making emissions data more accessible for AI and LLMs.
If you’re interested in contributing, please contact Dr. Steve Liang, the EmissionML working group convener, to get involved. Let’s build this foundation together — because solving methane emissions requires open collaboration and robust data standards.