Use Cases for Foundation Models (aka LLMs) in the Energy Industry
Kadri Umay
OT IT | Energy Data Platform | IASA Distinguished Architect | Carbon Capture and Storage | Ecosystem | Technology Evangelist | Public Speaker | Industry Standards OSDU, OPC, PIDX, Product Developer
Foundation Models as a Superset of Large Language Models (LLMs)
Recently, Stanford Center for Research on Foundation Models released a report on foundation models where they define them as models trained on broad data that can be adapted to a wide range of downstream tasks. These models demonstrate surprising emergent capabilities and substantially improve performance on a wide range of downstream tasks. Similar to humans, they can perform a lot of things which they haven't been "trained" on yet.
Foundation models are the subject of a growing paradigm shift, where move from a collection of "vertical" AI models that are very specific to certain problems and needs a lot of maintenance (retraining) to a very few "horizontal" foundation models?that are repeatedly reused as the basis for many applications.
Foundation models are a superset of LLMs, although the most "well known" foundation models currently are LLMs (e.g., GPT-3). "Foundation model” emphasizes the function of these models as foundations for downstream applications, whereas “large language model” emphasizes the manner in which these artifacts are produced (i.e., large textual corpora, large model size, “language modeling” objectives). Analogous to how deep learning was popularized in computer vision but now extends beyond, foundation models emerged in NLP with LLMs but foundation models (that are not LLMs) exist for many other modalities, including images, code, proteins, speech, molecules as well as multimodal models. In this article, I will use the term foundation model to explain the more generalized nature of the approach.
Use Cases for Foundation Models in Energy Industry
Generate
Summarize
Data platforms, such as OSDU, works like a library, where basic information on very large data files of differing formats are extracted and indexed. Similar to finding a book, it is only searchable by the metadata and searching on the details is only possible once the whole file is downloaded and opened by specialized applications. The findability (if this is a proper word) is determined by how rich the metadata is.
Rewrite
Extract
Search
Cluster
领英推荐
Classify
Chatbots
Translate
Codex
Image Generation
OSDU as an Enabler of Foundation Models
My colleague Einar Landre has written an amazing article that summarizes OSDU which could be found here https://lnkd.in/dQGVpnxM .
In short, OSDU is a library (which I have to give the credit to my colleague Graham Cain ) where it Clusters and Categorizes data (books), Extracts metadata for indexing (think of library index cabinets, which shows my age it is all online now), in some cases Summarizes for quick abstract searches and puts the data in generalized storage for easy access (think of books in shelves). One can Search the indexes, finds the files and can get access to the actual data. (ask the librarian to bring the book)
Familiar huh, in the previous section I have explained the use cases where Industry Data Platforms such as OSDU would utilize foundation models to further enhance the user experience. However, if we also look into the flipside, as you see in above paragraph which explains the key functions of OSDU, the items in bold are very similar to what foundation models deliver and hence OSDU could be a great enabler for Foundation AI Models, aka LLMs, and the capabilities they provide. OSDU gets all the Energy World's data, puts it in a common place and serves it via computer consumable APIs. With the current design of OSDU as a System of Record, these APIs are provided as https rest endpoints, which is not arguably the best option for high performance AI workloads such as training Foundation Models (aka LLMs). One of our colleagues Markus Cozowicz have released a Spark connector for OSDU https://github.com/microsoft/OSDU-Spark and Eirik Haughom have built a solution (https://github.com/EirikHaughom/MicrosoftEnergyDataServices/tree/main/Guides/Synapse/DataLakeIngestion) to ?automatically ingest binary data created on an external datastore into Microsoft Energy Data Services, which is Microsoft's implementation of OSDU on Azure (see last section for more information).
Although there are band-aid type solutions as outlined above, the ultimate goal is to build an AI Consumption Zone for OSDU where the data that resides in the data lake which holds all data files (remember the library analogy, these are all the shelves of the library) and exposes them as AI consumable structures. Some data files formats such as Wellbore DDMS that stores the data in Parquet format are easily consumable as external files to modern data warehouses. Other data formats such as OpenVDS also holds the promise to be accessible as AI consumable data, its founder Bluware has already demonstrated AI capabilities on SegY formats. Other formats needs to be evaluated for suitability for direct AI consumption to train Foundation Models (aka LLMs). My personal criteria here would be "could we expose the file as an external table in a modern data warehouse"? Work yet needs to be done, however there's good progress and stay tuned.
Microsoft Energy Data Services
Microsoft Energy Data Services is an enterprise-grade, fully managed, OSDU Data Platform for the energy industry that is efficient, standardized, easy to deploy, and scalable for data management—for ingesting, aggregating, storing, searching, and retrieving data. The platform will provide the scale, security, privacy, and compliance expected by our enterprise customers.
Sustainability Strategy & Digital Transformation, Managing Director| Energy Transition | Business Strategy | Climate Technology | Board of Directors| fmr- Microsoft, USDI USGS
1 年Great article, Kadri! I like the idea of "speaking " to the data to reduce time and cost for knowledge capture.
Never do a small task at work again. Ever.
1 年Great work Kadri!
Machine learning for seismic datasets, while finding Zico
1 年right of the bat: gpt like search!!! gpt searxh is super unreliable and prone to massive hallucinations without a searxh engine backend ....
Senior Scientist
1 年Great work
Industry Principal Director | Strategy | Innovation and Thought Leadership | IT-OT-IIoT-AI Consulting & Services
1 年Kadri, been awhile. Great blog. I would sum it up this way … the hardest part of AI (neural nets, expert systems, Bayesian belief networks et all) has always been how to reduce the time and cost to capture knowledge. Your blog shows that we are finally on the cusp of the breakthrough via LLM’s.