Written by: Gerald Gray and Michael Covarrubias
Let’s face it – we all know we have data problems. Quality issues, governance issues, transformation issues as data moves from one “cylinder of excellence” or “silo” to another. Maybe you implemented an enterprise service bus (ESB), but all that did was exacerbate the point-to-point interface issue – now everyone can make their own point-to-point interface for just “their” project. So, what do you have? Something complex and time consuming to manage! Something prone to errors and potentially even an increased vulnerability to security related incidents.
Now you are also looking to create an environment where you can enable machine learning, artificial intelligence, and you have a focus on data science overall – but if all you’re doing is piling up bad data – your machine can’t be trained nor tuned. It is the age-old data problem of garbage in, garbage out. Maybe you’ve already gone through the painful steps of creating an analytics platform while utilizing concepts such as data fabric or data mesh. How many of you started by mapping your data with Microsoft Excel? I’m sure the answer is – a lot of you. How did you feel about that process? Painful? Slow? Error prone? All the above?
If you’ve been involved with this kind of processes, you have probably experienced the situation where the same concept or object was used and perhaps defined in different ways depending on the system from which you pulled the data. Consider the meter as an asset, information about the meter will be in the meter system, customer system, and the asset management system, but these three different copies can easily have conflicting, and supplementing information. Changes could have been made but they might not have been synchronized across platforms. Let alone having a common definition of the meter and other data assets that the organization agrees to.
Why is a Semantic Model Important?
Semantic modeling is used to depict relationships that exist among specific data and establishes a common definition of the data. A recent article by Gartner Leverage semantics to Drive Business Value From Data advised data and analytics leaders to utilize a semantic approach to their enterprise data. If you don’t, you will be faced with the endless challenge of data silos.
Whether your approach to building your analytics platform is based on the concepts of data fabric, data mesh or a combination of both, or using knowledge graphs, semantic modeling is important because it speaks to data and governance. In the end, data must be understood and trusted. Semantic modeling allows you to exploit connected data; it also leverages common definitions that everyone across the organizations agrees upon.
Is there a good starting point?
After almost 30 years of development and continuing improvement as innovation and new technologies drive additional use cases, the utility Common Information Model (CIM) is widely recognized as the go to research and reference model from which to build an enterprise semantic model. But despite this – we see many times utilities start their data science initiatives with “roll their own” data model. There is still the occasional person that wonders, “Is the CIM even a thing?”, or don’t even know that there is a CIM. They decided to design their own because it seems faster at the outset. Creating an enterprise semantic model from scratch always seems faster in the beginning. Until the data concepts become more complex and one realizes that three decades of development is not something to be trivialized. In addition to the CIM as a reference model, vendor specific reference models are also needed, however, utilities who do utilize the CIM find that it provides close to 80% of what they need.
“Based on our experience, the CIM simplifies and supports utilities when capturing new data, for more analytics, intelligent automation, model training, and for outcomes benefiting all stakeholders.”
I like to describe the CIM as a starting point for the electric utility industry. The end game is for the utility to have a well-defined data model and associations that are not tied to a specific technology or vendor.
Wait there is more – Meta what?
Modern data management considers different aspects of the data journey within your organization, starting with semantic modeling connecting data and associated metadata. Modern data management enables reusability and ensures a common understanding while reducing complexity and delivering value.
Xtensible Approach – Affirma
With Affirma, our Enterprise Semantic and Metadata Management Solution, we are taking our proven semantic modeling approach, and using the CIM as a reference model to help build your enterprise semantic model for analytics and for data science. Affirma offers a unique capability for utilities to not only establish a common definition of data—for reusability and delivery of data to support data-in-motion and data-at-rest—but to also bring together disparate technologies for data mapping, lineage, integration, profiling and more within a single solution supporting your chosen data architecture.
In Summary
Regardless of the direction you take, there are core competencies needed for modern data management. This includes understanding your data through a semantic model, linking data about data, managing change to reduce complexity leading to innovation and growth.