IntelliGrid Architecture

 

 

Home

Parent Menu

Analysis Overview
Document Oveview
Architectural Principles
Architectural Analysis
Recommendations
Deployment Scenarios
Benefits
Network Management
Resilient Comm Services
Data Management
Requirements
RM-ODP/UML Overview

Same Level Menu

Enterprise Activities
Strategic Vision
Tactical Approach
Deployment Principles

Child Menu

 

Search IntelliGrid Site

Questions/Comments

Questions

 

Responses

 

Strategic Vision

Data Management and Exchange Issues

The amount of data being collected or capable of being collected is increasing exponentially. This rapid expansion of data retrieval results from the fact that more field devices are being installed and that these field devices are becoming more "intelligent" both in what power system characteristics they can capture, and also in what calculations and algorithms they can execute which result in even more data.

As distribution automation extends communications to devices on feeders, as substation automation expands the information available for retrieval by substation planners, protection engineers and maintenance personnel, and as more power system asset information is stored electronically in Geographical Information Systems and AM/FM systems, even more varieties and volumes of data will need to be maintained and managed.

Data management is a complex issue, encompassing many aspects of data accuracy, acquisition and entry, storage and access, consistency across systems, maintenance, backup and logging, and security. These are discussed in the following sections.

Data management must address a complex set of issues that include the following services:

·       Validation of source data and data exchanges

·       Ensuring data is up-to-date

·       Management of time-sensitive data flows and timely access to data by multiple different users

·       Management of data consistency and synchronization across systems

·       Management of data formats in data exchanges

·       Management of transaction integrity (backup and rollback capability)

·       Management of the naming of data items (namespace allocation and naming rules)

·       Data Accuracy

·       Data Acquisition

·       Data Entry

·       Data Storage and Access Management

·       Data Consistency across Multiple Systems

·       Database Maintenance Management

·       Data Backup and Logging

No single cross-industry technology addresses all of these issues, but multiple solutions and best practices are available for different aspects.

 

Abstract Modeling Tools

The first principle of IntelliGrid Architecture architectural vision is the principle of abstract modeling techniques, as expressed in the following quote:

 “There are limits to human ability to understand [truly complex systems] and to solve large sets of system equations.  The problem must be broken down or divided into a series of smaller problems that can be solved.  Modeling is one of the proven and well-accepted engineering techniques that simplify the system, so that we can better understand the system being developed.  System simplification is achieved through the introduction of levels of abstraction, which allow the modeler to focus on one particular aspect of the system at a time.” [1]

So that the abstractions used in IntelliGrid Architecture analysis may be more clearly understood, the IntelliGrid Architecture has been developed and refined using an international standard for architecture design call the “Reference Model for Open Distributed Processing” framework, RM-ODP[2].  RM-ODP is the reference model for defining open, distributed software system architectures. It was developed with extensive input from the technical community and represents a substantial body of knowledge. It was therefore the natural selection as a means for developing and expressing IntelliGrid Architecture.

RM-ODP is a formalized approach to developing abstract models of system functions, which helps to ensure that all requirements are identified and analyzed before the functions are implemented.  It breaks down the analysis and description of an architecture into five largely complimentary viewpoints.  Each viewpoint answers a different set of questions.  These viewpoints are summarized in the following table.

Table 2 Summary of RM-ODP Viewpoints

Viewpoint

Question

Contains

Enterprise Viewpoint

Who is involved?

Information about the various participants and functions implemented in the energy industry

Information Viewpoint

What information must be exchanged?

Models of information exchanged, agreements between parties, and roles and relationships that underpin the data of industry functions

Computational Viewpoint

How is this information going to be exchanged?

The mechanics related to information exchange i.e. a discussion of the interfaces required.

Engineering Viewpoint

Where is the information located and where will it be sent?

The configuration for where to physically deploy clients, servers, databases, and subsystems in terms of what component is deployed on which network for example.  This view describes the partitioning of a solution and where the pieces reside and closely corresponds to IntelliGrid Architecture Environments.

Technology Viewpoint

Which technologies and best practices are to be used to accomplish this?

Actual technology and best practice solutions that can be used to carry out the functions

 

It is important to note that RM-ODP is a framework for describing architectural views, and provides a reference model for developing architectures, but it is not an architecture itself.

In the IntelliGrid Project, the main purpose for using RM-ODP was to ensure that all architecturally significant requirements were identified for existing and future power system operations functions.

The Unified Modeling Language, UML, provided both the abstract language and computer tools that the team employed for expressing its analysis within the RM-ODP framework. These two approaches, RM-ODP and UML, complemented each other in the development of the architecture:

·       As a framework, RM-ODP is very abstract and does not call out the use of a particular notation nor does it have tools that are directly linked to its concepts.

·       UML has very good tools available, but is not an architecture standard and is not usually used at the same level of scope as RM-ODP.

The two can be used together because they are complementary and support similar levels of abstraction. Even though there is not a direct match between RM-ODP terms and UML terms, constructs can be developed in UML to realize RM-ODP concepts.  The widely available UML tools can then be used to create the diagrams and underlying database that is necessary for understanding the complex interactions of power system functions across multiple areas.  Both are useful standards as both embody the following concepts:

·       UML is a technology neutral way of specifying use cases, data, and software components.  RM-ODP separates technology specifics into the Technology View.  IntelliGrid Architecture uses UML to specify a technology neutral architecture.

·       UML is a deployment neutral way of specifying data and software components.  RM-ODP separates deployment specifics into the Engineering View.  IntelliGrid Architecture uses UML to specify deployment neutral architecture that can be applied to a variety of environments.

In conclusion, UML and RM-ODP have enabled IntelliGrid Architecture Team to design an architecture that is flexible and can be applied to a diverse set of environments and technologies.  UML and RM-ODP provide a standard language so that the design of the architecture can be communicated to others.

 

Figure 3: Integrated Energy and Communication Systems Architecture (IntelliGrid Architecture) RM-ODP Model

 

The abstract modeling process used during IntelliGrid Architecture project leading from RM-ODP to the final architecture is outlined below. In general, the UML concept of “Use Cases” was used to capture stakeholder requirements, and a UML tool called Magic Draw was used to generate diagrams that expressed the architecture in the five RM-ODP viewpoints.

1.     UML Use Case Template: The highest level UML construct for describing a function is the Use Case, which can map more or less into the RM-ODP Enterprise Viewpoint. Since not many people have UML tools, people in the IEC and IEEE have been describing functions using a “Use Case Template”.  A Use Case Template is a MS Word document that captures the UML concepts of Actors, Roles, Associations, Classes, and other UML constructs. This Use Case Template could then be used to enter the information into a UML tool, like Rational Rose or Magic Draw. The Use Case Template doesn’t have any standard format, but usually includes sections to:

·       Describe the function in narrative form

·       Identify the Actors and Information Exchanged

·       Identify the steps involved in exchanging information between Actors.

2.     UML Use Case Template to IntelliGrid Architecture Domain Template: In IntelliGrid Architecture project, the team renamed the Use Case Template the “Domain Template” and modified it in a number of ways:

·       Added a number of additional fields and requirements beyond those of a traditional UML Use Case in order to capture more RM-ODP concepts, such as policies and contracts between Actors.

·       Added the RM-ODP concept of Common Services (services that can be used by many different functions) as well as a spreadsheet to capture common requirements across all Use Cases.

·       The IntelliGrid Architecture team identified a number of common Environments, discussed later in section 2.  Each environment has different configuration, performance, security and data management requirements.  Each “step” in a Use Case was assigned to a particular environment for use later in determining appropriate technologies to use for that step.

3.     Importing Domain Templates into UML Tool: Domain Experts filled out Domain Use Cases using the Domain Template. These were then imported automatically into the Magic Draw UML tool. The resulting diagrams became a tool for IntelliGrid Architecture team to further analyse the requirements captured in the Domain Templates. 

4.     Results from the RM-ODP Analysis Become the Technology and Deployment Neutral Reference Architecture: As stated above, RM-ODP is a reference model for defining a distributed system architecture for a particular software function. However, the purpose of IntelliGrid Architecture project is not to develop a single architecture for one specific function; the purpose of IntelliGrid Architecture project is to develop a Reference Architecture for all power system functions.   The relationships captured in the UML tool can be used as a database for determining the appropriate approach to any power system communications problem.  In this sense, the database becomes the architecture.

 

Abstract Use Cases

Through analysis of the Domain Use Cases, the team identifies a limited set of common functions necessary to implement each Domain Use Case. In order to capture this common functionality IntelliGrid Architecture team derived a set of Abstract Use Cases.  The Abstract Use Cases are:

·       Integration of Enterprise Management – The integration of software and hardware component management functions with power system functions

·       Integration of Utility Wholesale and Retail Market Operations - The integration of market operation functions with power system functions

·       Device Integration – The integration of heterogeneous power system devices.

·       Application Integration – The integration of heterogeneous power system applications to meet operational needs.


·       Data Integration – The integration of heterogeneous power system data to meet analytic needs.

·       Security Integration – The integration of security across multiple domains.

The process of using Domain Use Cases to derive Abstract Use Cases is a key simplification used by IntelliGrid Architecture Team.  That is, the Team realized that it would be impossible to analyze all conceivable Domain Use Cases within a limited timeframe and budget.  Instead, the Team realized that they had to pick a much smaller set of “architecturally significant” Domain Use Cases.   These “architecturally significant” use cases were then used to create more generalized Abstract Use Cases as illustrated in Figure 04. 

Figure 04 Abstract Use Cases from Domain Use Cases

 

This second set of use cases is abstract because they are not tied to any particular utility function.  However, the Abstract Use Cases are more useful in deriving detailed components of an architecture because they allowed the Team to abstract away the specifics of Domain Use Cases and permitted the team to focus on the commonality and diversity of all Domain Use Cases.  The commonality is expressed as a set of common modeling elements and the diversity is expressed as a set of environments and technologies as shown in Figure 5:

Figure 5 Environments from Requirements

 

The abstract set of use cases is illustrated in Figure6 below:

Figure6 The IntelliGrid Architecture Abstract Use Cases

 

Domain Use Case Requirements Analysis

This section discusses the derivation of the Abstract Use Cases from the Domain Use Cases.

ADA

Advanced Distribution Automation (ADA) involves software applications in the control center supplemented by applications and functions implemented in field equipment. The control center applications provide the global analysis of the distribution system state and capabilities and are the overarching functions in control of distribution system operations, while the field equipment applications provide local information and control.

The ADA applications in the control center rely heavily on data from many different sources, and going to different systems:

·       SCADA system for real-time data from field equipment and control command to field equipment, including both substations and feeder equipment

·       DER equipment, either directly or indirectly through DER Aggregators

·       Energy Management System (EMS) for transmission information and

·       Geographical Information System (GIS) and/or Automated Mapping and Facilities Mapping (AM/FM) systems for power system facilities data and physical connectivity data

·       Customer Information System (CIS)

·       Work Management System (WMS)

·       Distribution Planning Systems

·       Market Operations systems

The primary architectural requirements for ADA are focused on data management. Correct, available, and timely data are crucial to the ADA function operating properly. However, since data comes from many different sources and since the systems acting as these sources usually are provided by different vendors, the coordination, synchronization, integration of systems, and mapping of data elements across these systems is a major problem.

In addition, because real-time control of the power system is a major aspect of ADA, both security and network management are critical to safe and reliable operation of the power system. Therefore the main requirements are those associated with the “Critical Operations DAC” Environment and the “Intra-Control Center” Environment ( a complete list of IntelliGrid Architecture Environments can be found in Appendix E):

1.     Security Requirements

·       Provide Authorization Service for Access Control (resolving a policy-based access control decision to ensure authorized entities have appropriate access rights and authorized access is not denied)

·       Provide Information Integrity Service (data has not been subject to unauthorized changes or these unauthorized changes are detected)

·       Provide Audit Service (responsible for producing records, which track security relevant events)

·       Provide Credential Renewal Service (notify users prior to expiration of their credentials)

·       Provide Security Policy Service (concerned with the management of security policies)

·       Provide Single Sign-On Service (relieve an entity having successfully completed the act of authentication once from the need to participate in re-authentications upon subsequent accesses to managed resources for some reasonable period of time)

·       Provide User Profile and User Management (combination of several other security services)

·       Provide Security Discovery (the ability to determine what security services are available for use)

2.     Network and System Management Requirements

·       Provide Network Management (management of media, transport, and communication nodes)

·       Provide System Management (management of end devices and applications)

3.     Data Management Requirements

·       Support the management of large volumes of data flows

·       Support keeping the data up-to-date

·       Support extensive data validation procedures

·       Support keeping data consistent and synchronized across systems and/or databases

·       Support timely access to data by multiple different users

·       Support frequent changes in types of data exchanged

·       Support management of data whose types can vary significantly in different implementations

·       Support specific standardized or de facto object models of data

·       Support the exchange of unstructured or special-format data (e.g. text, documents, oscillographic data)

·       Provide discovery service (discovering available services and their characteristics)

·       Provide conversion and protocol mapping

 

Therefore, if one abstracts from these functions, one can say that ADA from an architectural perspective requires:

·       Integration of many different systems developed by different vendors for differing requirements

·       Development of a platform that spans many different systems, applications, and databases

·       Management of data across multiple systems, including data consistency and synchronization with short timeframes

·       A system to manage configuration and change

·       Security integration across multiple security domain

Customer Interface

With the advent of deregulation, the interface between ESP and consumer has become more important, because customers can (and have) switch(ed) energy providers and because they can now be an additional source of revenue if new energy services can be sold to them, or if the utility rights within the customer premises can be used to sell access to other businesses.

The expansion of system operations coordination and control down to the end user level creates one of the key justifications for IntelliGrid Architecture. An enterprise-wide architecture such as IntelliGrid Architecture offers a tremendous opportunity for improved operational efficiency, improved control of customer processes based on supply system conditions, use of customer-owned and operated generation and power quality improvement technologies as part of overall system management and to achieve the required levels of reliability and power quality at the end user level.  In as far as the other side of the equation, implementation of load control/demand response programs provides utilities with another key tool to ensure power system stability and security.

Applications related to customer interface must be coordinated closely with distribution automation and distributed resource applications, as well as market operations.  Key applications include

·       Real time pricing

·       Load management

·       Residential customer applications, such as load control in response to real time pricing incentives

·       Direct customer energy management and load control during system emergencies

·       Automatic evaluation of and recommendations for increasing energy efficiency based on profiles of the customer site and loads

·       Control and performance evaluations for residential generation

·       Power quality assessments and control.

Also critical are commercial and industrial (C&I) applications such as commercial customer participation in energy markets through aggregation of backup generation and energy management, participation in ancillary services (such as volt/var control, harmonic control, and reserve generation), real time commercial facility power quality assessment solutions integrated with the distribution system operation and integration of real time information concerning system power quality and reliability.

By nature, customer interface/consumer services applications share close coordination with distributed automation, distributed resource, and market operation applications.  The current status of utility industry restructuring, as well as the current state of technology, necessitate that many consumer service applications rely on distributed automation and distributed energy applications and their underlying communications requirements.  Furthermore, consumer interface is playing a key role in market operations where customer load and/or onsite generation may be aggregated and utilized to bid into energy markets and key customers may have sufficient requirements for power to play a role in bulk power trading, scheduling, and supply scenarios.

The range and scope of customer interface/consumer service applications is complex and growing.  The possibility of customers and ESPs managing load down to the appliance level could generate requirements with a level of granularity not seen in any other domain.  However, at the present time and taking both short and long term scenarios into consideration, the domain analysis covered three key applications:

·       Real time pricing

·       Utility administered load control

·       Consumer side data collection

 

RTP is important because it requires communication between the customer and the ESP in terms of the ESP providing RTP signals to the customer and the customer potentially providing bids and forecasts back to the ESP.  Quality of service including high availability and timeliness of data is crucial.  There are large numbers of customer with sensitive information on pricing and usage; therefore security is a key consideration.  Since future power system operating scenarios will undoubtedly involve more two-way communication with the customer and the ESP as well as increased customer sensitivity to pricing data, this is a significant area for requirements analysis.

Utility administered load control is where, instead of responding to price signals, the signal comes directly from the ESP to control customer loads.  This application covers a wide range of issues, especially security across organizational domains and the need for two-way communications to confirm load control actions for future advanced demand responsive systems. 

Data collection from the consumer side is seen as critical to facilitate more active involvement by customers in the interface to and participation in market operations and energy management.  In terms of power quality monitoring, the information is intermittent and sometimes infrequent, but timely, communication and notification is very important when events do occur. 

Analysis of the requirements associated with these requirements showed trends that were common:

 

·       Database integration and management is critical, with utilization on both customer and ESP systems and even inside one system, with multiple installations and different purposes.  For instance, for ESPs, there may be one database for customer billing data, a database for pricing data, and a separate database for operational data, i.e., customer participation in RTP or load control programs.  Some of the data requirements are low in volume (such as usage data) while others may be high in volume with detailed information (such as power quality monitoring data).

·       There are multiple levels and requirements of security and in addition to ESP security issues, which are significant and substantial, there are issues relating to the privacy of customers and desire to secure customer data and facilities from unauthorized access and cyber attack.  Securing the consumer interface frequently requires different technologies than securing ESP-specific functions.

·       Communication and bandwidth requirement run the gamut from telephone line to wireless to fiber.  Much equipment related to customer metering has not been upgraded to take advantage of the state of the art, so legacy systems and disparate data transportation mechanisms require an enterprise-level approach to systems integration.

·       Customer services factor in financial transactions.  Much as with market operations, money will be changing hands in applications such as power trading and real time pricing.  This brings into play a different set of considerations than the traditional ESP operations.

·       The amount, scope, and time frame for transactions are constantly evolving.  As more customer services are identified and as the technology matures and emerges, the far reaching vision of a consumer portal, where a customer can manage every aspect of their interface with the electrical grid and energy usage, comes into play.  This necessitates the ability to add and address requirements for technologies that are still in the development stage.

As can be seen, while this domain has some specialized requirements, it shares requirements with several other domains regarding wholesale and retail market operations integration, application integration and data integration.

Wide Area Measurement and Control

The goal of Wide Area Measurement and Control (WAMAC) is to synchronize and coordinate the measurements of the state of the power system across large geographic areas, to model and simulate system behavior in real-time, to anticipate fast-changing system conditions, and to support multiple automation and control response capability. The core functions of WAMAC include:

·       Synchrophasor calculation –Historically, voltage and current phasors were measured against a local reference angle, complicating the state estimation process of determining system wide phasors against a common reference. Synchro-phasors are measured over a wide area against a common reference angle and the information can be made widely available, which will greatly simplify the state estimation problem.

·       Dynamic model update – WAMAC depends on accurate models of the state of the system and the capability of the system to respond to control actions. WAMAC requires accurate and timely updated models.

·       Real time state measurement and security assessment – WAMAC functions include steady-state and dynamic analyses and risk assessment that consider multiple set of independent and dependent contingencies while applying probabilistic models of power system components. The increased dimension and limited time intervals of these function creates communication and computational challenges.

·       Real-time proactive/preventive dynamic control – The decision making for the proactive/preventive dynamic control actions implies analyses of multiple complex scenarios with possible conflicting results. Fast simulation based on adequate modeling is an imperative requirement. The timely and reliable implementation of these control actions poses communication challenges for the WAMAC system.

·       Emergency control – Emergency control focuses on preservation of power system operations without endangering the power equipments. WAMAC should perform multiple remedial control schemes in adaptive and timely manners providing integrity and generation-load balances of power system areas.

·       Automated restoration – The ideal goal of restoration is fully automated self-assembly of the entire power system. In this case, WAMAC must coordinate the re-synchronization of separated transmission lines, reconnection of affected distribution systems and customer loads. WAMAC must interact with Advanced Distribution Automation functions to leverage DERs in support of the restoration actions.

The power system is the largest and most complex system created by man. As it has grown, human management and control of it has proved quite challenging.  Moving to the future, an information based Wide Area Measurement And Control system will be needed to provide instantaneous (10s of ms) measures of the system conditions to enable dynamic modeling of the various complex power system phenomena. Such requirement involves gathering and coordinating the data from large areas and various organizations. The creation of synchrophasors provides the technical means to make unified measurements over a wide area.  The implementation over of wide area communications allows consolidation of these measurements through what is known as Phasor Data Concentrators (PDCs). The devices that create the synchronized measurements, commonly known as Phasor Measurement Units (PMUs), often provide synchronized measurements that span organizational and divisional boundaries. Organizational boundaries would include sharing aggregated data between utilities.  Data communication must only occur with those entities authorized to receive the synchronized data.  As much of the synchronized data indicates instantaneous system state, this communication of this data must be secure for some period of time as it could be used by power marketers in the pricing of electricity.  The present NERC agreement between organizations that share data requires confidentiality of operations data for 8 days.  Divisional boundaries include the issuance of control information to distribution companies and can potentially migrate directly to direct load control in the home.  This type of application requires the utmost in secure and authenticated communications.

To validate existing system models and to dynamically update them is another challenge for WAMAC. Connectivity of the areas affected by a power system event needs to be updated in the model in real time. Input data such as specific impacts of load inertia on frequency-related phenomena or impacts of saturated devices are critical to the model. Lack of these critical data is an issue to WAMAC and, as such, high availability of the synchronized data is required.

WAMAC must accomplish many complex decisions and control actions within milliseconds to second time intervals. There are three security states in the power system operations: normal, emergency and restorative. The timing requirements for each of these states are different. The dynamics of loads, generation and system topology drive the timing requirements for the normal state. Any decision made when the system is in the normal state should be valid in the time interval between the consecutive runs of the particular application. In the emergency state, the time for the decision-making and its implementation is considerably shorter than the normal state. In the emergency state, the power system conditions could change dramatically in the 10s of millisecond range. The amount of data that need to be processed by WAMAC is substantial in as much as it is continually streamed. Failure to respond to the emergency conditions could result in system instability and possibly lead to large scale blackouts. The same timing requirement applies to the restoration actions due to the need for simultaneous execution of control actions. Again, any control action need to be secure and issued with the proper authority.

In the control mode, WAMAC needs to coordinate the implementation of the power system state machine. In the general case, an operational decision in power system management consists of several control actions, each of which takes time to finish. In many cases, a subsequent action depends on the successful completion of the previous action. No harmful operation conditions should occur during the intermediate steps.

To summarize, the successful implementation of WAMAC relies on IntelliGrid Architecture to provide:

·       Automated information support/data aggregation appropriate for the timing and complexity of the process to be controlled

o      Reliable, high-speed, secure, point to multi-point communications

o      Automatic system configuration

o      Automated configuration and remote control of executing devices or their modes of operations (settings)

o      High-speed application data access

o      Integration of different information systems

·       Secure information sharing among different organizations

o      Authentication of control commands

o      Data confidentiality

·       Integration and coordination of centralized and distributed intelligence

o      Integration of different control systems, such as EMS, DMS and Market operation system

o      Cross-domain communications with Advanced Distribution Automation applications

·       Utilization of modern information technologies to solve data-overwhelm issue, to enhance data availability, and to provide modern data visualization

Conclusion of Domain Use Case Analysis

This section has shown that in order to deploy the functionality describe in the Domain Use Cases in an economical way, one has to deploy the functionality described in one or more of the Abstract Use Cases.  However, besides deriving the Abstract Use Cases from Domain Use Cases, one can also derive the Abstract Use Cases via a more theoretical discussion.  The following section includes this discussion as well as analysis of the Abstract Use Cases for derivation of the IntelliGrid Architecture.

Analyses of Abstract Use Cases

On an abstract level, one can state that IntelliGrid Architecture must support just two capabilities:

·       Provide support for the operation of existing and future utility functions.

·       Provide support for the integration of existing and future utility functions.

Note that support for the operation of existing functions is known and currently implemented by utilities.  Note that support for the operation of future functions is largely addressed by the development of new applications and technologies.  As IntelliGrid Architecture is not an application or technology development project, this is out of scope.  This leaves support for the integration of existing and future utility functions as the primary issue to be solved by IntelliGrid Architecture.   Also, since an architecture for non utility specific functions will be driven by cross industry groups such as the W3C, IEEE, or even major information technology vendors, IntelliGrid Architecture more narrowly focuses on specializing these cross industry architectures to utility specific functionality.

One can state that integration issues related to utility specific activity are limited to:

A.    Integration of applications and data for operational and analytic purposes.

B.    Integration of devices as well as hardware and software services for operational and managerial purposes.

C.    Secure integration of applications, data, and devices within a utility, between a utility and an energy market partner, and between a utility and an operational partner.  The operational partner means an external entity that works with a utility to meet operational as opposed to market driven goals.

The common terms for Group A include:

      I.     Application integration

    II.     Data integration 

The common terms for Group B include:

  III.     Device integration

  IV.     Enterprise management

The common terms for Group C include:

    V.     Energy Trading

  VI.     Security Integration – especially security across security domains

The IntelliGrid Architecture Team believes that these six abstract use cases provide complete coverage of utility specific functionality required for the complete analysis for comprehensive architecture.

The goals of IntelliGrid Architecture Enterprise Architecture include:

·       Establish an architecture to integrate all of the utility enterprise - from Energy Market partners to backend systems to devices.  

·       Enable comprehensive and unified views of the utility enterprise to allow creation of new applications that can look across the utility and focus on end-to-end profitability and reliability. 

·       Establish Comprehensive Security Architecture that accommodates integration of autonomous security domains. 

·       Create migration plan whereby legacy applications can be adapted to conform to the IntelliGrid Architecture and new application can be non-disruptively added.  

Figure 7 below portrays the different elements than need to be integrated by IntelliGrid Architecture.

Figure 7 IntelliGrid Architecture Secure Enterprise Architecture

 

This section describes the Abstract Use Cases in more detail and why these tasks play an important part of enterprise integration and the development of higher-level profitability and reliability focused analysis applications

Integration of Enterprise Management and Power System Services

This section describes the challenges facing the integration of Enterprise Management, sometimes called communications System Management or Network Management, into the power system.

In order to create a truly reliable power system, IntelliGrid Architecture team needed to consider more than just power system services.  Modern utilities monitor and control the power system via a vast network of communication-enabled devices.  Traditionally, the data related to power system operation and communication system operation has been treated independently, as illustrated in Figure 8.

Figure 8 Enterprise Management and Power System Management Treated Independently

 

However, operation of the power system is now completely dependent on successful operation of the communication system.   It is clear that in order to achieve a comprehensive view of end-to-end reliability one needs to integrate communications system and power system analysis as shown below:

Figure 9 Integration of Enterprise and Power System Management

 

System/network management, also referred to as enterprise management, is the task of ensuring that the systems and the network provide the required services with the specified quality of service to the users and other systems. Most enterprise management architectures use agent-manager relationships where the agents, residing on the managed elements, provide management information, such as alerts or performance measurements, to the manager.

The manager reacts to these messages by executing one or more actions such as:

·       Operator notification

·       Event logging

·       System shutdown

·       Automatic attempts at system repair.

Management entities also poll managed elements, automatically or upon user request, to check the values of certain attributes of the managed device. Agents have information about the managed devices in which they reside and provide that information (proactively or reactively) to management entities within an enterprise management system using a management protocol. 

Typically, enterprise management functions are performed on the following managed elements:

·       Network devices such as routers, switches, hubs, customer premises equipment and communication links;

·       Computing resources such as substation automation systems and data concentrators; servers such as Market Transaction Servers;

·       Software services such as SCADA, EMS, or GIS components, as well as database management systems;

·       Service and business functions such as RTP customer pricing service, security and operational policy servers; and

·       Storage area networks.

 

In IntelliGrid Architecture, the team adds the power systems network-aware devices such as IEDs and RTUs to the above. 

The International Organization for Standardization (ISO) has defined the following network management functions for fault, configuration, accounting, performance and security (FCAPS) management. Although defined for network management, these functions can be generalized to systems and applications management.

Fault Management Function- Fault management detects, fixes, logs, and reports network problems. Fault management involves determining symptoms through measurements and monitoring, isolating the problem, fixing the problem through reconfiguration, reset, technician dispatch, etc.

NOTE:   In this context, Fault Management does not refer to power system faults, but faults in the communications network.

Configuration Management Function - Configuration management, complements fault, involves maintaining an inventory of the network and system configuration information. This information is used to assure inter-operability and problem detection. Examples of configuration information include device/system operating system name and version, types and capacity of interfaces, types and version of the protocol stacks, type and version of network management software, etc. Configuration management complements the other functions fault, performance and security management.

Accounting Management Function - Account management keeps track of usage per account, billing, and ensures resources are available according to the account requirements.

Performance Management Function - The task of performance management involves measurements of various metrics for system/network performance, analysis of the measurements to determine normal levels, and determination of appropriate threshold values to ensure required level of performance for each service. Examples of performance metrics include network throughput, user response times, CPU, memory and line utilization. Management entities continually monitor values of the performance metrics. An alert is generated and sent to the network management system when a threshold is exceeded

Security Management Function - Security management is to control access to network resources according to security guidelines. Security manager partitions network resources into authorized and unauthorized areas. Users are provided access rights to one or more areas. Security managers identify sensitive network resources (including systems, files, and other entities) and determine accessibility of users and the resources.  Security manager monitors access points to sensitive network resources and log inappropriate access.

NOTE:   Security management is being discussed in a separate section and will not be included in the enterprise management sections of this document.

The above functions form the basic set of functionalities needed for enterprise management, specifically for element management. It is easy to see how they apply specifically to IntelliGrid Architecture, for instance:

·       Fault management is essential to provide scalable support of reliable operations and maintenance of the large-scale communications/ distributed computing infrastructures found in IntelliGrid Architecture.

·       Configuration management is crucial as the number of the to-be-managed entities within IntelliGrid Architecture infrastructure scales up. Such entities can range from network devices, substation controllers, RTUs, IEDs, to computing resources such as servers and clients running IntelliGrid Architecture applications/ services, to emerging intelligent home gateways located in the premises of RTP customers.

·       In the context of IntelliGrid Architecture, accounting management not only involves the management of accounts and/or billings for end customers, such as in the case of RTP services, but also, the accounting of shared/ exchanged resources among multiple energy providers or trading entities.

·       Performance management is a basic building block to enable end-to-end service level agreements for various services, applications and customers supported by IntelliGrid Architecture. Security management is indispensable for IntelliGrid Architecture that will control one of the key national infrastructures -- the utility networks. 

Since the development of the FCAPS categories there have been many changes in the state of the art in power system communications networks, systems and applications.  These changes have expanded the functions within these basic categories to address the more challenging management requirements of next generation enterprise management systems. Examples of these expanded requirements include:

·       Complex, inter-dependent, multi-protocol networks including wireless, broadband, and ad-hoc networks, giving rise to cross-technology domain management.

·       The need to go beyond element and network layer management to service and business layer management functions, imply broader management functionalities.

·       Huge network configurations such as networks reaching millions of consumers with scaling issues, diversity of access technologies (wireless, Hybrid Fiber Coax, Digital Subscriber Line (DSL), dial-up, leased lines, etc.) and issues on the geographical distributions of the end devices, give rise to development of additional management entities such as proxies, and definition of hierarchies of management;

·       Increasing embedded device intelligence that gives rise to intelligent problem detection and resolution for self-diagnosis and self-healing systems and networks;

·       More involved policy-based management to include extensive Service Level Agreements (SLA) and stringent security and Quality of Service (QoS) requirements such as those needed for Advanced Distributed Automation (ADA) ;

·       Increase of mission critical applications, such as wide-area monitoring and control, raises the need to manage their real-time stringent reliable delivery, QoS and security requirements as exemplified by applications such as Wide-Area Measurement and Control Systems (WAMACs);

·       Increasing inter-organizational collaborations and data sharing, such as those in RTP, gives rise to more stringent policy management functions;

·       The distinction between physical and virtual networks, systems, connections, etc., requires the enterprise management function to distinguish between the two.

·       Service-centric functional requirements for management of services such as VoIP, wholesale and retail market operations, multi-media services, VPN services, etc;

·       Expanded list of security requirements such as intrusion detection and responses to denial of service.

·       Increasing integration of circuit-switched and packet-switched networks due to integration of the corresponding services such as multi-media applications and VoIP implies the need for integrated enterprise management functionalities.

·       Requirements for more dynamic management aspects of FCAPS functions such as user provisioning, accounting, routing, rerouting, resource allocation, resource scheduling, service negotiation, access requests, grid computing, etc.

·       Introduction of new web-based services and new network-based computing architectures such as grid computing, imply more dynamic, web-based, security enhanced enterprise management functions.

Enterprise Management is a key part of understanding the reliability, costs, and risks associated with running a communication network.  Furthermore, it is only through a combined view of the communication system and the power system that reliability versus risk balancing can occur.  Consequently, it is vital that Enterprise Management data be integrated and analyzed with power system data.

Integration of wholesale and retail market operations

This section discusses the integration of wholesale and retail market operations with power system functions.  Specifically, this section discusses an architecture for the integration of an Energy Market Transaction Service into the utility enterprise as well as how analysis applications can be built on top of integrated utility operational and wholesale and retail market operational applications and data.

While other eCommerce operations such as buying office supplies are an important part of any enterprise, IntelliGrid Architecture is more focused on the integration of utility specific functionality.  That is, while it is likely that a utility will want to automate non-utility specific operations, this will probably be done without requiring a utility specific architecture such as IntelliGrid Architecture.  It is important that IntelliGrid Architecture interoperate with non-utility specific architectures.  IntelliGrid Architecture can be seen as extending or specializing more generic architectures that are used to integrate non-utility specific functions.

Besides being more narrowly focused on utility specific integration, IntelliGrid Architecture is also more concerned with internal wholesale and retail market operations integration and analysis as opposed to external wholesale and retail market operations integration.  In other words, wholesale and retail market operations applications are treated somewhat as black boxes as illustrated in the diagram below:

 

Figure 10 Energy Market Transaction Service Communication

 

In the diagram above, an Energy Market Transaction Service consists of wholesale and retail market operations applications that act as a gateway to external wholesale and retail market partners.  Market data flows between the Transaction Service and remote partners.  Utility Operational Systems such as EMS or DMS manage the operation of the power system.  Operational systems supply capability data to the Transaction Service.  The Transaction Service submits commitment requests to Operational systems. 

In general, Energy Market Transaction Service and Operational Systems are supplied to a utility as indivisible applications.  Furthermore the exact mechanisms and protocols used to exchange market and operational data with external entities is often outside the utility’s control.  While the IntelliGrid Architecture must be compatible with and support data flows to/from external parties, as others specify these data flows, IntelliGrid Architecture is primarily an architecture for internal integration. 

Retail

Utilities buy and sell power at a wholesale level as well as a retail level.  Retail sales activity consists of energy delivery from the distribution system as well as end user accounting.  Data exchange related to retail includes distribution system power delivery monitoring data, metering data as well as billing and customer service information. 

Retail energy billing and customer services issues are similar to non-utility businesses albeit at a larger scale.  However deregulation has complicated the picture somewhat.  While utilities may be responsible for the physical distribution system, an Energy Service Provider (ESP) may act as an intermediary from a financial point of view.  An ESP will often buy blocks of power and then sell it to a collection of retail customers.  In this case, the ESP may be responsible for meter reading, billing and customer support.  Besides ESP, a utility may subcontract meter reading or even all of customer interaction entirely. 

In either the case of an ESP or a subcontractor, the technical issues are similar.  That is, information flow must pass between different companies each with their own infrastructures.  In both cases retail customer data is aggregated and presented to the utility so that they may manage operations.  In both cases, the utility will likely have an application responsible for interacting directly with the customer and directly or indirectly with an ESP or subcontractor. This application is called “Energy Market Transaction Service”.  IntelliGrid Architecture must facilitate the internal flow of data to and from this Energy Market Transaction Service as well as allow the utility to create analysis applications that look at this retail customer data within the larger operational/financial picture.

Wholesale

If retail energy transactions normally occur between a utility and an end user or intermediary parties as a result of delivery of power by distribution system, wholesale energy transactions occur between utilities and other entities as a result of delivery of power by the transmission system.  Primary functions related to transmission market operations include.

·       Long Term Planning

·       Medium/Short Term Planning

·       Day Ahead Market

·       Real-Time

·       Post-Dispatch

 

Deregulation and faster more open markets

Transmission system operators dispatch control area resources to meet load requirements while maintaining system security and reliability. Additionally at a higher level, interchange across market boundaries must be managed.  Each region needs to consider the energy transactions in their respective markets, and optimize the interface energy flow by establishing price equality. As a result, a sophisticated real time market place must be developed.

 

Conclusion

The architecture for utility energy related transactions must be able to support a wide variety of business models.  At the retail level, local market regulation and procedures require many different data exchange choreography and protocols.  At the wholesale level, the architecture must support bother bilateral and multilateral oriented markets.  The issue then becomes how can the architecture support the delivery of off the shelf products that can be deployed without requiring extensive customization for the local market.

Data associated with energy market trading is central to utilities.  Without an accurate picture of market activity and risk, utilities cannot be run profitably.  However, market data alone is of limited usefulness.  Only integrated market and operational data provides the required information to maximize return on utility assets.

Device Integration

This section discusses the integration of devices.  Specifically, this section discusses an architecture for the integration of the command, control, and sensing capabilities of the numerous devices found on the power system with other information sources to facilitate the implementation of numerous power system functions and end-user applications.

Data Accessibility

All field devices have a common set of functionality – they obtain, create, consume, and/or contain data; they initiate and/or respond to control signals; they interact closely with their local physical environment through the previous two items.

This common set of functionality can be represented architecturally in terms of semantics and mechanisms.  From the semantic point of view, any device can be represented by a common abstract information model.  The data elements within the device can be named, have a type, and a well defined meaning within the context of the devices application.  The communications architecture also must provide the mechanisms necessary to interact with the information model – typically through a set of common abstract services.  Such services can be as simple as read, write, and report on change.  Higher-level services may be derived from these to provide functionality such as file transfer, metadata discovery, and numerous device configuration services.

Benefit

The primary benefit of representing devices through the use of a common abstract information model and abstract services is that it enables each device to interact with other devices and applications in an efficient, structured, and unambiguous fashion independent of those device’s physical attributes and communications interfaces.  This approach allows data to be gathered and fused together to accomplish a higher-level mission without requiring detailed knowledge of the inner workings of each device.  The ultimate benefit however, is an increase in system reliability, with the ability to change out individual devices as technology and functional requirements change but with little or no impact at the application level thereby providing higher reliability at lower cost (implementation and maintenance).  The reliability issue is actually addressed here on two fronts – the inherent reliability of the device integration itself and the ability to take advantage of device integration to implement system reliability analysis and management applications at the enterprise level.

Example

The following figure illustrates this concept using a distribution device control example.  In this situation, information from field devices that implement various low level protocols and physical communication interfaces is exposed using the common information model and services approach.  This permits data derived from these various devices to be integrated with similarly structured information about the distribution system topology and physical attributes to facilitate the implementation of the distribution device control function.  Only when all of the related information is fused together can an operator (human or cyber) have a clear picture of what the state of the system is before initiating a control action in a safe manner that is consistent with the higher-level mission at the enterprise level.

Figure 11 Integration Of Device Data

 

Conclusion

Without a common abstract information model and services, device integration becomes chaotic with an associated higher cost and lower reliability.  In fact, this is the situation many systems face today.  Additional systems, gateways, interfaces, and other patches have been deployed to try and address the outwardly visible issues, but this increases complexity and cost.  Since overall integration of the energy management enterprise and the implementation of system reliability analysis and other applications relies upon the underlying devices that make up the system, then the efficient integration of these devices into the enterprise is key to its reliable and profitable operation.

Application Integration

The current economic climate and market initiatives require utilities to perform more efficiently and in more flexible ways. The dynamic nature of today’s environment means that a utility must be able to build an integration infrastructure for operational application integration quickly to provide a base for adaptable business models. This section discusses the integration of applications as shown below. 

Figure 12 Applications To Be Integrated

 

The main information management problems currently facing the power industry are:

·       Utilities spend many millions of dollars trying to create comprehensive views of the utility enterprise. 

·       Lack of standards means that expensive custom solutions are required. 

·       Lack of robust/intelligent infrastructure requires a lot of manual effort. 

·       Lack of security means that the resulting network is vulnerable.

 

These problems arise primarily because today’s utility IT environment is truly heterogeneous.  Some of the more significant features of this mix include:

·       Many Platform Technologies used to provide an operating environment for applications such as operating systems, and component technologies like CORBA, Java, and Web Services.

·       Many Communications Infrastructure Technologies used to move data within a network.

·       Many Security Technologies used to secure information carried on the communications network.

·       Many Data Management and Exchange Technologies including format and exchange mechanisms as well as a wide variety of utility data semantics i.e. many different meanings for common business entities such as circuit breakers or purchase orders.

For example, in any typical utility task (such as outage management), components must be involved at once, mixing real-time data from field devices, customer information, historical information such as maintenance history, calculated or simulated data, and business information including, but not limited to:

o      Supervisory Control and Data Acquisition services such as SCADA and meter reading

o      Control and Power System Analysis services such as EMS and DMS

o      Power Quality Monitoring systems

o      Protection systems and fault recording systems

Integration between these systems is typically manual, labor intensive, and therefore expensive, to put in place. In the past, each integration task has been treated differently from all the other integration task.  Over time, the lack of standards results in a software management nightmare. While previously there has been no standard way to handle these types of integration problems, utilities nevertheless have integrated applications anyway.  These non-standard methods include:

·       The Buy Everything from One Source Approach - buy a system or subsystem from a single vendor with turnkey responsibility. The benefits of working with a single vendor are in minimizing the points of contact, less opportunity for miscommunications, and having only one source for accountability.  The problem of working with only a single vendor is each vendor has it’s own proprietary way of doing things; whereby, replacement, rather than upgrade, is the only option for system improvement at a later date. It is also rare that one vendor has the knowledge or experience in understanding all the components of a complex implementation and so will install component applications with little knowledge of the long-term ramifications.

·       The “Kitchen Sink” Approach - everything federated into a single or multiple databases. There are many reasons to not store all data into a single database.  First, not all data are efficiently stored in a single database.  Process (real-time/temporal) data are not efficiently stored in a transactional (relational) database and model/configuration information is not stored efficiently in a temporal database.  Second, specific users of differing types of data reside in a variety of groups around the organization.  Users interested in outage management are not the same users who are make updates to map drawings.  It does not make technical sense to have these groups working in the same database, however, it is important that the two groups can access one another’s data if needed. 

·       The Apply Glue as Needed Approach - development of point-to-point information links and /gateways/translators as needed.  While this solves the short-term problem of linking those particular “Islands of Automation”, these types of solutions never establish a platform for obtaining an enterprise wide view of data.  Integration techniques that do not facilitate future business applications just create more and bigger “Islands of Automation”.

·       The Least Common Denominator Approach – linking all data into extremely simplified databases that are of just a few common data types, e.g. analog inputs, analog outputs, digital inputs, digital outputs, and counters.  This is often called the “points list” model because the data becomes just a list of anonymous data “points”.  This “least common denominator” approach does permit data to be converted and shared between a variety of different devices and technologies.  However, all information about the logical relationships between points, their geographical location, their source, and their significance in the power system is lost.  This information must be entered manually into separate databases.  This increases, rather than decreases, the cost of the communication system.

Most frequently, utilities have chosen a mix of these short terms solutions, which result in many separate point-to-point links as shown in Figure 17.

Figure 13 Application Integration

 

Application Integration as shown in Figure 13 plays an important part in operating a utility profitably and reliably.  Without operational integration, each application cannot be kept in sync as data in each application runs as an isolated silo.

Data Integration

Data Integration is a term used to describe the process of presenting data in a uniform way and within a uniform context.  Consider the example of a service technician performing a Preventative Maintenance (PM) procedure on a breaker in a substation.  Ideally, the information required by the technician would be retrieved electronically on demand in the substation and not as is typically done today via a time consuming manual process before going to the substation. 

In order to accomplish the PM procedure, the technician needs to gather the following information:

·       Substation Schematics

·       Breaker Repair Manuals

·       Breaker Operation History

·       Breaker Asset Data including PM history

Not only would it be beneficial if all this data were available on line from the field, it would also help if all this data was available by browsing a simple and familiar tree such as shown below:

Figure 14 Example Of Integrated Data

 

While document management systems may provide the capability to categorize documents and present them via a user-friendly GUI, not all required data is in the document management system as shown below:

Figure15 Field Service Integration Example

 

The document management system is just one source of data and it rarely has any knowledge of where a piece of equipment is installed in a power system.

IntelliGrid Architecture is focused on providing a familiar and common context for all utility data.  As this example illustrates, part of the issue is coalescing a variety of data sources and putting them into a context that is most useful to the data consumer.

Cross Domain Security

What are security domains and their properties?

There are many potential methods through which to model security.  Several involve concrete analysis of particular systems and communication technologies/topologies.  It is often difficult to discuss security models in concrete terms since the technology used in deployments typically become limited to the lowest common denominator that is discussed.  Such technology based security models tend to be difficult to scale and understand from an enterprise system perspective.  Likewise, such concrete models are difficult to extend/scale to address systemic security.

“The concept of a security domain that is introduced in this paper is not new. Many computer security practitioners have been (either explicitly or implicitly) using the ideas presented here for many years in protecting networks.”

Security Domain Definition:

“[A] Telecommunications and Network Security domain encompasses the structures, transmission, methods, transport formats and security measures used to provide integrity, availability, authentication, and confidentiality for transmission over private and public communications networks and media.”

Additionally:

“In this paper, the term Security Domain is used to describe a network of computer systems that share a specified security level through a common element.”.

 

Figure 16: Representation of Security Domain Concept[3]

 

A Security Domain (SD) represents a set of resources that is governed/secured and managed through a consistent set of security policies.  Additionally, Security Domains provide a well-known set of security services that are used to secure transactions and information within that domain.  This notion of Security Domains correlates well to IntelliGrid Architecture concept of distributed computing environments.

General Requirements for security management

Security Management is defined as:  “In network management, the set of functions (a) that protects telecommunications networks and systems from unauthorized access by persons, acts, or influences and (b) that includes many sub-functions, such as creating, deleting, and controlling security services and mechanisms; distributing security-relevant information; reporting security-relevant events; controlling the distribution of cryptographic keying material; and authorizing subscriber access, rights, and privileges.” Based upon this definition, it is the Security Management of an SD that is responsible for the risk assessment, developing security strategies, and implementing those strategies.  A successful SD will define and implement the following security functions:

·       Access Control: “The prevention of unauthorized use of a resource, including the prevention of use of a resource in an unauthorized manner.”

There are generally three categories of Access Control that need to be addressed within a SD: Physical; Resource; and Information.

·       Trust: “In cryptology and cryptosystems, that characteristic allowing one entity to assume that a second entity will behave exactly as the first entity expects. Note: Trust may apply only for some specific function. The critical role of trust in the authentication framework is to describe the relationship between an authenticating entity and a certification authority; an authenticating entity must be certain that it can trust the certification authority to create only valid and reliable certificates. [After X.509]”

Trust is established via Authentication.  However, there are two methods of authentication that are prevalent in today’s electronic systems: Role Based Authentication and Individual Authentication.

·       Confidentiality: “The property that information is not made available of disclosed to unauthorized individuals, entities, or process.”

There are typically two categories of Confidentiality that need to be addressed within a SD: Protection from un-intentional disclosure and overall protection of information.

·       Integrity: “The principle that keeps information from being modified or otherwise corrupted either maliciously or accidentally.”

·       Security Policy: “The set of rules and practices that regulate how an organization manages, protects, and distributes sensitive equipment and information.”

It is the security policy function that determines how to manage residual risk.  The policy function then expects the Security Management Infrastructure to allow the actual management of such risk.

·       Security Management Infrastructure (SMI): “System elements and activities that support security policy by monitoring and controlling security services and mechanisms distributing security information and reporting security events.”

The use of the Security Domain concept allows discussions in regards to how to allow physical access into a domain (e.g. physical access control) and which security services are needed in order to provide a robust physical access control function.  Examples of such security services would be: the ability to identify the person attempting access; the ability to make sure that the person is authorized to enter a particular security domain; the ability to log the fact that the person entered/exited the domain; and the need to have established security policies that encompass/manage the other security services set forth.

Whereas, physical access is typically well understood, other security functions are typically discussed/understood at a high level and therefore do not capture all of the functional/service requirements.  The security domain concept allows a more detailed discussion at a high level. 

In the case of Trust, it is well understood that in order to establish trust one must determine the identity of the person/entity to which information/resources are being provided.  In the case of individuals that you know and are face-to-face with, identity establishment is quite easy.  Therefore, if a person you know requests a piece of information, it is relatively easy to determine if that person should be granted access to that information due to a well established identity.  However, is the same true if the same person approaches you for the same information but is executing the request on behalf of a third party (e.g. an Inter-Domain request that is acted upon intra-domain)? Maybe.  What if the request for information is nested even further?  At some point, although the identity of the immediate requestor is well-known, there may arise an issue of trust in the actual request due to the number of times that the original requestor’s identity has been changed (e.g. as it crosses into different security domains). The need to provide a security service that could allow the determination of a metric of how many identity mappings have occurred could prove useful, although not needed in every instance.

Confidentiality is typically thought of as a well-understood security function.  When one typically thinks of confidentiality, the first thought is the word “encryption”.  Encryption is a security service that needs to be provided.  However, confidentiality could also be provided/enhanced if the sender of the request/information could specify a path through which to route the information/request.

Analysis based upon the security domain concept indicates that there are several security services that any particular security domain will need to have available. Some functions are not requirements for intra-domain security but are mandatory for inter-domain (e.g. identity and credential mapping) security.  These services and their inter-dependencies are described in Section 2 of this document.  The development of high level security service definitions and functional requirements allows for issues of resource type (e.g. physical or informational) to be deferred until technological implementation strategies are evaluated.  Thus it becomes possible to discuss the issue of access control for buildings and Simple Network Management Protocol (SNMP) information/services in a common manner. Based upon the understanding of the functions that these services need to provide, technologies (or combination of technologies) can be evaluated as mechanisms of actually implementing such security functions.  It is during these evaluations that the distinction of physical or informational resources would be required.

The IntelliGrid Architecture attempts to define an architecture that creates an environment for heterogeneous energy industry applications and business functions within that environment. IntelliGrid Architecture has defined several enabling architectures and technologies in this regard.  However, security and security domains are inherently non-heterogeneous (especially at a technological solution level).  It is this dichotomy that is part of the reason that many individuals, when attempting real business functions, perceive security as an impediment to the accomplishment of the primary business function.  Thus, there needs to be a balance of providing adequate security versus protecting the primary business functions from security threats. Thus the security services developed in Section 2 are classified as mandatory/optional in order to provide a security function.  However, it is the security policy of a specific Security Domain that determines which services must be used.  Additionally, it is a specific SD that determines what type of technological solution(s) will be used in order to accomplish a given security function.  The technological solutions chosen will typically create interface issues between security domains.

A good example of this is the Trust Function.  If SD1 makes use of a username/password based technology to establish trust (e.g. Identity Establishment security service) and SD2 makes use of digital certificates, how should an individual in SD1 establish an identity or role within SD2?  The obvious answer is that there needs to be a process to convert from the username/password, managed in SD1, to digital certificate required for SD2.  The proposed IntelliGrid Architecture security service to provide this capability is named the Credential Conversion security service.   Once the service needed is recognized, the next question becomes whose responsibility is it to provide the particular conversion.  In our example, it would be SD1 and not SD2 (not quite intuitively obvious).  There are several issues that lead IntelliGrid Architecture to this determination, and these will be discussed in Section 2.

The abstraction of security functions and services, to some, may not seem to be needed.  However, in order to future proof (e.g. to allow applications to migrate to better technologies as they become available), applications will not be able to invoke security technologies directly. Just the knowledge of what services need to be used (if implemented) could have prevented several of the Internet Viruses that attack Outlook, Outlook Express, and IE.  These are examples of applications that were designed to accomplish a business function without regards for protecting critical information nor do they provide an audit capability to determine when and if the list has been modified/accessed.

The security information contained with IntelliGrid Architecture is hoped to provide an infrastructure that allows applications to be created that can make use of various security technologies as required by the security policies of each SD.  Additionally, it is hoped that by identifying abstract service requirements that all future applications created for IntelliGrid Architecture environment will make use of such services.

 

Abstract Use Case Requirements Conclusion

The principle impediment to integrating the systems described above is the cost.  Without de jure or de facto standards to drive commonality, a utility is forced to create a large collection of custom-developed links as shown below.     

Figure 17 Point-to-Point Integration

 

Because custom developed links are not readily reusable, vendors must recoup their entire development cost for every integration project.  This, combined with the lack of competition leads to a prohibitively expensive solution.

However, what is remarkable about the previous description of integration of enterprise management, energy market transaction services, devices, applications, and data with the power system is how similar these tasks are to each other.  This similarity can be seen in who, what, how, and where information is processed.  For example, Enterprise Management and Power System Management similarities include:

·       The type of applications involved (who is involved) in enterprise management versus power system management is very similar.  For example, an enterprise management and energy/distribution management systems typically acquire data from a set of instrumented elements, perform real-time data acquisition and control, facilitate intelligent electronic device management, and analyze topology for network optimization.

·       The type of data involved (what data is exchanged) in enterprise management versus power system management is very similar.  For example, both systems largely deal with handling real time measurements, status reports, and alarms.  The data into and out of analysis applications is similarly complex and of similar sized.

·       The way applications communicate (how data is exchanged) in enterprise management versus power system management is very similar.  For example, enterprise management communication protocols are very similar to those used to communication with power system IED’s.

·       The distribution of applications involved (where is involved) in enterprise management versus power system management is very similar.  For example, both enterprise management and power system management systems communication with a large number of widely distributed remotely situated field devices.

This suggests that a similar set of modeling constructs can be used to integrate these major abstract use cases within the utility.  Note that this discussion is not intended to show for example how system/network management or power system management applications need to be redesigned or replaced, but only to show how the two back boxes can be non-intrusively integrated at a higher level. 

What is needed is a single unified technology independent architecture that is codified in a set of complementary standards that allow access to data related to hardware/software components, wholesale and retail market operations as well as, devices in a secure and reliable manner.

In order to achieve this level of integration at a reasonable cost, it requires common abstractions for what data is exchanged and how data is exchanged.

IntelliGrid Architecture
Copyright EPRI 2004