Digital Twin of Power Grid: Best Practices and First Steps

In the first part of our mini series on digital twins in the energy sector, we took a look at the current situation in power distribution grid, while the second part focused on the potentials of the digital twin technology to ensure smooth grid operations even in times of the renewable energy sources and skills shortages. Now let's get a bit more specific.

We have talked with our own experts who oversee grid digitalization projects for distribution system operators all over Europe in order to collect some practical tips and recommendations. Specifically, we wanted to know: As a grid operator, where should I start if I want to create a digital twin of my own distribution network? What type of data plays an important role and how high must the quality of these data be? What are the steps towards a dynamic virtual copy of grids?

Now we'd like to share with you their answers in this third and last part of our series.

Digital Twin Technology: various areas of application

In principle, the digital twin technology can be implemented for various purposes. In the energy sector specifically, we have identified so far three main directions:

  • Digital copies of individual assets
  • Digital mapping of the network infrastructure
  • Virtual copies of the grid together with the power producing and consuming elements

The first area of application deals mainly with condition monitoring and predictive maintenance for individual assets, for example transformers. The virtual copy collects sensor and other data to replicate each aspect of operation. For one, this allows one to monitor from afar when there is a risk of overheating in transformer oil, which can damage the asset.

In the second case one would use a digital twin to plan construction work. The digital twin technology helps create exact virtual copies of streets or woodland of the areas where a grid operator has planned, for example, to build new power lines and substations. This allows not only to better plan for construction but also to evaluate costs and risks much more accurately and at a very early stage.

Last but not least, the third direction is exactly the area of application that we are going to dive into in more detail further down below. This type of a digital twin is used to optimize the grid operations and make short-term as well as long-term grid planning much easier.

How do I create a digital twin of my distribution network?

Step #1: Data collection

Nowadays, it is safe to say that everything starts with data. The same is true for the creation of a dynamic virtual grid model (= grid digital twin) - any preparation work begins with data collection.

What kind of data do I need for a digital twin of a distribution grid?

Specifically with distribution grids in mind, we need information that allows us to understand what happens in the grid and where there is need for action - for example, for grid reinforcement measures.

We divide such data into three main categories:

  • Grid topology in conjunction with geospatial data
  • Assets data
  • Power producing and consuming elements in the grid
Grid topology in conjunction with geospatial data

It goes without saying that you’ll need to map the exact geo location of your grids in the digital model; however, that’s not all. It’s equally important to replicate the grid topology, i.e. how different areas of the distribution grid are interconnected: Where the individual breakers / switches are situated and whether they are closed or open.

Assets data

A full digital copy of a grid means more than just having its geo location and switching states. After all, the electricity assets are core elements of any power network. That’s why we need data on the assets locations as well as technical data. The latter includes for example the asset’s type, year of manufacturing or various technical characteristics such as current carrying capacity.

Power producing and consuming elements in the grid

The massive adoption of decentralized power systems with bidirectional power flow poses a number of challenges to reliable grid operation. Therefore, if we want to create a comprehensive grid model that can help us address these challenges, it is crucial to know what is connected to the electrical power network. This means that not just private households and businesses, but also PV systems, charging stations, heat pumps and wall boxes need to be part of such a grid model. The data on the types of the so-called network participants and their performance, capacity, etc. thus form the third and last major category of data that we need for an electrical grid digital twin.

Where does data for a grid digital twin come from?

We can generally say that we need to include as many data sources as possible in order to build an exact replica of distribution grids. In the majority of cases we talk about the existing software systems that a respective distribution system operator has already in place: geographic information system (GIS), assets information system, ERP, SCADA, etc.

Due to the increasing use of smart meters, the meter data management systems (MDM) are also becoming more relevant in order to be able to better assess the current network status. For the digital twin, MDM can, therefore, provide further valuable information about the network participants.

Ultimately, we need an enormous variety of data: switching states, assets data, technical parameters, information about the feed-ins and much more. Whether all that data is in just a handful of systems or scattered around in local databases and Excel spreadsheets is less important.
Now, after we have identified the relevant data sources, the next step is…

Step #2: Data aggregation

A major challenge for distribution system operators today is that although they already have an enormous pool of valuable information about their power grid, this data often sits in silos across multiple IT systems. To be able to create a true digital twin of the distribution grid, you need to aggregate this data; ideally, through API-based connectors. Alternatively, you can use file-based data transfer.

What do I need to keep in mind when it comes to combining data?

From our experience, there are three main aspects that one needs to pay close attention to:

  • Data freshness
  • Flexible connectors
  • Consistency in data maintenance
Data freshness, or how up-to-date your data is

Our distribution grids are constantly changing because of the massive adoption of renewable power systems in the private sector. If you think back, say, 25 years ago, 98 percent of the electricity grid of 1999 was most likely exactly the same as one year earlier. Nowadays, due to the new PV installations alone, the grid of 2023 looks very different from the one in 2022.

This is why keeping data up-to-date has the number one priority in the creation of an exact dynamic replica of distribution grids. This means, in turn, that it’s not only about keeping data fresh in the source systems. The digital twin, too, must receive updated information on a regular basis – depending on the use case, even in real time – so that important calculations are not performed on the basis of obsolete data.

Flexible connectors

One should generally try to use standardized formats for data exchange between various systems. This way data aggregation goes easier and faster, and you can connect new systems following clear processes.

In reality, however, it’s oftentimes considerably harder to use standardized formats. CIM or CGMES formats for grid data exchange are a good example of that. These “standards” allow such a high level of flexibility that a plug-and-play data exchange between various tools often just doesn’t work.

In many cases we have to, therefore, configure the so-called connectors individually in order to be able to work with the available data formats and structures. Since a digital twin lives and breathes fresh data, new datasets must be fed at the very least on a regular basis; better even – continuously. An ongoing manual correction of data exports is simply not an option; connectors must be flexible and scalable enough to enable faster and easier data aggregation and transformation.

In our case, for example, we have already developed connectors for a large number of common software systems through numerous customer projects. These connectors require only some minimal adjustment in order to map any custom data fields if necessary.

Consistency in data maintenance

In addition, it is important to maintain data consistently across different systems. Ideally, you use a unique identifier for individual assets or facilities that is shared between all relevant systems. It’s certainly not easy to be this consistent when maintaining data, and it generally requires the implementation of intelligent systems! However, whether it is a digital twin or a traditional grid model, you’ll have to aggregate data from multiple tools. And what we learn from every customer project is that ensuring data compatibility upfront or at least as soon as possible is really important to make data aggregation as less time and efforts-consuming as possible.

Step #3: Data quality checks

From our point of view, a final evaluation of data quality should take place only when data is aggregated. Each single dataset can seem error-free in itself; however, it is only when you put all datasets in the context of a power grid model that all inconsistencies and errors come bubbling to the surface.

That doesn’t mean, though, that you can’t check beforehand if datasets are actually complete and information is plausible. With this in mind, we approach quality evaluation from two angles.

Dataset quality check: Do we have all pieces of information and are they accurate?

It’s not uncommon that some pieces of information have gaps or errors, or are just implausible. For instance, the capacity of the XY type transformer has been entered as 100 kVa, although this specific type can only have a capacity of 1000 kVa. Such inconsistencies should be detected and corrected – ideally, automatically – at the latest during the quality check.

For instance, for our Intelligent Grid Platform, we have developed specific algorithms that check each dataset individually for data gaps and inconsistencies. Taking the same example of a transformer, these algorithms first examine if all relevant data fields for this asset are filled in and whether this information is correct. This way, errors such as capacity that is inconsistent with the asset’s type are quickly detected and, whenever possible, automatically corrected.

Grid model quality check: Is all data correct in the given context?

One could call it the highlight of the data quality analysis, because this is how we can validate the entire grid model.

Firstly, you would examine the grid topology, i.e. if all grid areas are actually connected with each other in the model. A typical example is when there is no connection between two grid areas because, say, the switches on both sides of the line have been recorded as open. Or a facility is not connected to the grid because a piece of cable is missing in the grid model.

The so-called power grid tracing helps detect such affected areas automatically. This way, you can correct the gaps and errors in the model bit by bit and in a consistent way. If, on top of that, you can enable a high degree of automation for evaluation and correction, you can considerably reduce the amount of effort.

The additional load flow simulation as a final quality check allows then to check all data for accuracy and plausibility in the context of the distribution grid in question. If, for example, the iterative process of the load flow simulation isn’t converging or delivers unrealistic results (say, the load factor of 300% or more), chances are that somewhere, there is a data error.

The data quality conundrum: How high should it be?

"The quality of our data is so poor; we probably shouldn't even think about creating and deploying a dynamic grid model until it's improved". One hears this argument quite often.

Indeed, it is very often the case that the DSOs we speak to wish their data had much higher quality than it is now. However, this is not a good enough reason not to explore the potential of digital twin technology.

Quite the contrary: Only when all data "lives" in a consistent grid model that is available to everyone and is always up-to-date can you steadily and systematically improve its quality.

Because knowing that your data is not at an optimal level of accuracy is not enough. You also need to know where the need for improvement exactly is, and considering the massive amount of data sitting across multiple systems, that's not an easy task.

However, when data is aggregated and linked together in a shared context, you can assess its quality in a more structured way and actually improve it in a focused manner.

100% percent data quality is a myth

It might sound extreme, but when it comes to a "live" grid model, 100 percent data quality is impossible to achieve. The distribution grid changes very quickly nowadays. This in turn means that the grid model continuously receives new data from the source systems, which may contain errors or information gaps.

With this in mind, it's important to establish certain processes and methods that help ensure a high level of data quality long-term; and not only establish such processes, but actually follow them rigorously.

Still, while the latest data batch has been evaluated and cleaned up, there is a next batch waiting which is potentially not error-free either. And so on and so forth. For this reason, striving for 100 percent data quality is unrealistic.

Based on the experience of our customers, however, we can say that a 90 percent quality can be achieved quite quickly; as a medium-term goal, one can aim for a value of slightly over 95 percent.

Low-Hanging Fruits for grid optimization

Now that we've gone through the steps required for creating a digital twin of the distribution grids, let's briefly review what areas of grid operation we've seen the digital twin technology to have the fastest positive impact on.

High grid transparency for informed strategic decisions

Distribution system operators have generally reported that a digital twin of their grid helps them have a much better understanding of the current performance of their grid. Such a comprehensive, "live" grid model allows, for example, to see right away if there is enough hosting capacity in a specific area to integrate further PV systems.

This in turn allows us to make well-founded decisions in which area the power grid must be expanded next. Nowadays, one can no longer rely on the traditional reference values and fixed planning principles. The increasing usage of decentralized power generators and consumers makes it significantly more challenging to identify short-term measures for the increase of the grid hosting capacity as well as to plan strategic grid expansion and reinforcement measures.

However, if you have a good data basis at your disposal - the one where as many pieces of information as possible come together - you can carry out a much more granular and comprehensive data analysis for an accurate power supply forecasting. As a result, distribution system operators can not only ensure reliable grid operations but also save financial resources by making sound investment decisions. For instance, in order to reinforce the grid areas that will only become critical in two years.

Digitalization and automation of processes

Moreover, a grid digital twin enables process digitalization and process automation at a level that cannot be achieved without a complete and, more importantly, consistent electrical grid model.

Grid connection check is one example. In order to be able to automate individual steps such as calculation of the best possible connection points or performing a grid integration study, one has to ensure that certain grid simulations can be carried out automatically, too.

This in turn means that the basis for such simulations - i.e. a dynamic digital grid model that is up-to-date - already exists and does not require any manual aggregation of data. Because as soon as there is manual input, by definition it can no longer be considered as automation. In order to build such a (dynamic + digital + always up-to-date) power grid model, one will inevitably have to use digital grid technology.

Conclusion: One data basis for all decisions in distribution grid operations

Perhaps it sounds familiar to you, too: One department creates its own grid model on the basis of the data master system in order to carry out grid connection checks. And then maybe six months later, another department does the same, only this time to plan grid reinforcement measures. Now there are two different versions of a grid model, which on top of that "live" locally in the computers of individual colleagues.

Or one has two data master systems that are used for different purposes. Ideally, these systems are maintained in parallel, so that all colleagues can work on the same data basis. However, the reality is often very different.

With a digital twin, on the other hand, you have a consistent data basis for decision-making. The grid model "lives" online and is, therefore, available to every colleague at any time as a single source of truth. This way, it can also be continuously updated whenever there are any changes.

This means that all decisions that the distribution system operator has to make in order to ensure reliable grid operations are based on a universal database all relevant stakeholders have access to.

This project is supported by the German Federal Ministry for Economic Affairs and Climate Action as part of the Renewable Energy Solutions Programme of the German Energy Solutions Initiative.

Logoband mit envelio

German Energy Agency (dena)

The German Energy Agency (dena) is a centre of excellence for the applied energy transition and climate protection. dena studies the challenges of building a climate-neutral society and supports the German government in achieving its energy and climate policy objectives. Since its foundation in 2000, dena has worked to develop and implement solutions and bring together national and international partners from politics, industry, the scientific community and all parts of society. dena is a project enterprise and a public company owned by the German federal government. dena's shareholders are the Federal Republic of Germany and the KfW Group.
www.dena.de/en

German Energy Solutions Initiative

With the aim of positioning German technologies and know-how worldwide, the German Energy Solutions Initiative of the Federal Ministry of Economics and Climate Action (BMWK) supports suppliers of climate-friendly energy solutions in opening up foreign markets. The focus lies on renewable energies, energy efficiency, smart grids and storage, as well as technologies such as power-to-gas and fuel cells. Aimed in particular at small and medium-sized enterprises, the German Energy Solutions Initiative supports participants through measures to prepare market entry as well as to prospect, develop and secure new markets.
www.german-energy-solutions.de/en

Renewable Energy Solutions Programme (RES Programme)

With the RES programme, the Energy Export Initiative of the Federal Ministry of Economics and Climate Action (BMWK) helps German companies in the renewable energy and energy efficiency sectors enter new markets. Within the framework of the programme, reference plants are installed and marketed with the support of the German Energy Agency (dena). Information and training activities help ensure a sustainable market entry and demonstrate the quality of climate-friendly technologies made in Germany.
https://www.german-energy-solutions.de/GES/Redaktion/EN/Basepages/Services/dena-res.html