Ofgem has put forward its initial views in a Call for Input published on 5 April 2024 on the existing regulatory framework on the use of AI within the energy sector. While they suggest that there is adequate regulation to capture the use of AI, they believe producing regulatory guidance on the risk-based use of AI in the energy sector will prove valuable. As such Ofgem is seeking views on their proposals about how AI should be used safely and responsibly within the energy sector in a way that fosters innovation. Organisations within the energy sector who either use or are looking to use AI, technology companies, AI developers, and those working on AI policy, amongst others, are asked to respond to the proposals.
Integrating AI into the energy industry is expected to raise unique challenges concerning the regulation of energy. Regulatory implications are wide-reaching and will need to be considered in full, for example, supply chain duties, AI decision explainability, data privacy, cybersecurity, governance and accountability. The task of overseeing and assessing violations of energy regulations linked to AI could present complexities. More specifically, there are potential challenges in the detection of compliance and breaches under current approaches.
The Department for Science, Innovation and Technology’s AI White Paper on AI regulation published in March 2023 set out five principles to be applied by sector regulators. Ofgem has rooted its proposals in these principles. They include:
The aim is to develop an outcome-based approach to the regulation of AI within the energy sector which is proportionate and based on the application of principles considered important in managing risk, rather than setting prescriptive rules on the application of AI.
The Call for Input presents a suggested risk framework designed to assist entities in grasping Ofgem’s perspective on risk. It aims to direct ‘dutyholders’, like energy sector licensees and other organisations contemplating AI adoption, towards proportionate actions and to prevent or lessen the likelihood of failures.
Area | Key Identified Risks | Key Ofgem Proposal / Recommendation |
Consumer | The potential for AI to cause discrimination and inequality / bias. | Ofgem emphasise the importance of the principles of fairness, transparency and explainability. Ofgem concluded that, in addition to an overarching definition of fairness, an analysis to assist in identifying unfairness associated with an AI application is important. |
Market | “Tacit collusion” – the concern that algorithms including AI may provide a way for companies to attain and maintain collusion without any explicit agreements or human communication. Automated mechanisms with deep learning techniques enable companies to track prices, introduce common policies, give market signals, and optimise joint profits. |
Ofgem’s AI taskforce and the Competition and Markets Authority are exploring this issue further. Suggested solutions include:
|
Company | Effective oversight of AI systems by energy companies. Appropriate routes for customers to contest a negative AI outcome and seek redress. |
Ofgem proposals include, amongst others:
Ofgem also set out a table which outlines the various aspects of accountability and governance for effective use of AI in relation to different responsible roles and groups, including board members, management, project teams, life cycle stakeholders, supply chain participants, sector representatives and regulatory bodies. |
Sustainability | AI systems require significant energy and water resources due to their complex design, extensive training data and frequent user interactions. AI’s global demand may require 4.2 to 6.6 billion cubic meters of water withdrawal by 2027 – half of the UK’s annual use. Without sustainable practices it has been warned that AI is likely to consume more energy than the human workforce by 2025. |
Ofgem solutions include:
|
Although Ofgem provides reassurance that existing regulation is adequate, there appears to be considerable effort still required to thoroughly resolve and work through the various legal and regulatory complexities described, in particular, with regards to AI collusion, liability and the AI supply chain, and sustainability. There are also a number of other wider implications to the energy sector discussed in the Call for Input that deserve an equally thorough consideration – for example: the use of AI involves the collection and processing of large amounts of data, raising concerns regarding data privacy and security; human rights and equality; the protection of critical systems from cyber attacks; health and safety; confidentiality; and the impact on the workforce including possible influences/biases or errors in data, or miscorrelations due to insufficient training, data or coding mistakes. What stands out from the Call for Input is Ofgem’s cooperation with other regulators and key industry bodies e.g., Office of AI (DSIT) CMA (see our article on ‘Guiding the development of AI: the CMA’s initial report on AI Foundation Models’ here), ICO, EHRC, National Cyber Security Centre, and the AI Safety Institute, demonstrating a considered and collaborative approach.
The Call for Input closes at 11:59pm on 17th May 2024.