Article 92

Power to conduct evaluations

1.   The AI Office, after consulting the Board, may conduct evaluations of the general-purpose AI model concerned:

(a)

to assess compliance of the provider with obligations under this Regulation, where the information gathered pursuant to Article 91 is insufficient; or

(b)

to investigate systemic risks at Union level of general-purpose AI models with systemic risk, in particular following a qualified alert from the scientific panel in accordance with Article 90(1), point (a).

2.   The Commission may decide to appoint independent experts to carry out evaluations on its behalf, including from the scientific panel established pursuant to Article 68. Independent experts appointed for this task shall meet the criteria outlined in Article 68(2).

3.   For the purposes of paragraph 1, the Commission may request access to the general-purpose AI model concerned through APIs or further appropriate technical means and tools, including source code.

4.   The request for access shall state the legal basis, the purpose and reasons of the request and set the period within which the access is to be provided, and the fines provided for in Article 101 for failure to provide access.

5.   The providers of the general-purpose AI model concerned or its representative shall supply the information requested. In the case of legal persons, companies or firms, or where the provider has no legal personality, the persons authorised to represent them by law or by their statutes, shall provide the access requested on behalf of the provider of the general-purpose AI model concerned.

6.   The Commission shall adopt implementing acts setting out the detailed arrangements and the conditions for the evaluations, including the detailed arrangements for involving independent experts, and the procedure for the selection thereof. Those implementing acts shall be adopted in accordance with the examination procedure referred to in Article 98(2).

7.   Prior to requesting access to the general-purpose AI model concerned, the AI Office may initiate a structured dialogue with the provider of the general-purpose AI model to gather more information on the internal testing of the model, internal safeguards for preventing systemic risks, and other internal procedures and measures the provider has taken to mitigate such risks.

Frequently Asked Questions

The AI Office conducts evaluations of general-purpose AI models to check compliance with AI Act obligations or to investigate significant risks that the AI might pose across the European Union, such as those identified by independent scientific experts, ensuring these technologies are used safely and responsibly.
Evaluations can be conducted directly by the AI Office, or the European Commission may appoint independent experts, including members of a specialized scientific panel, to objectively examine general-purpose AI models and provide expertise in assessing compliance and identifying systemic risks at the Union level.
Yes, the AI Office may request access to the company’s AI model through APIs or other suitable technical methods, including the source code itself, clearly stating the legal reasons and purpose for the access request, as well as potential fines if the company fails to comply.
Before directly accessing an AI model, the AI Office can start a structured dialogue with the AI provider to gather detailed information about the provider’s internal testing, any safety measures in place for reducing risks, and other safeguards already implemented to address potential issues.

AI literacy

Get Started within 24 hours.

Once you have submitted your details, you’ll be our top priority!