ISO/IEC 25059
(Main)Software engineering — Systems and software Quality Requirements and Evaluation (SQuaRE) — Quality model for AI systems
Software engineering — Systems and software Quality Requirements and Evaluation (SQuaRE) — Quality model for AI systems
This document outlines a quality model for AI systems and is an application-specific extension to the standards on SQuaRE. The characteristics and sub-characteristics detailed in the model provide consistent terminology for specifying, measuring and evaluating AI system quality. The characteristics and sub-characteristics detailed in the model also provide a set of quality characteristics against which stated quality requirements can be compared for completeness.
Titre manque
General Information
Standards Content (Sample)
DRAFT INTERNATIONAL STANDARD
ISO/IEC DIS 25059
ISO/IEC JTC 1/SC 42 Secretariat: ANSI
Voting begins on: Voting terminates on:
2022-07-12 2022-10-04
Software engineering — Systems and software Quality
Requirements and Evaluation (SQuaRE) — Quality model
for AI systems
ICS: 35.080
THIS DOCUMENT IS A DRAFT CIRCULATED
FOR COMMENT AND APPROVAL. IT IS
THEREFORE SUBJECT TO CHANGE AND MAY
This document is circulated as received from the committee secretariat.
NOT BE REFERRED TO AS AN INTERNATIONAL
STANDARD UNTIL PUBLISHED AS SUCH.
IN ADDITION TO THEIR EVALUATION AS
BEING ACCEPTABLE FOR INDUSTRIAL,
TECHNOLOGICAL, COMMERCIAL AND
USER PURPOSES, DRAFT INTERNATIONAL
STANDARDS MAY ON OCCASION HAVE TO
BE CONSIDERED IN THE LIGHT OF THEIR
POTENTIAL TO BECOME STANDARDS TO
WHICH REFERENCE MAY BE MADE IN
Reference number
NATIONAL REGULATIONS.
ISO/IEC DIS 25059:2022(E)
RECIPIENTS OF THIS DRAFT ARE INVITED
TO SUBMIT, WITH THEIR COMMENTS,
NOTIFICATION OF ANY RELEVANT PATENT
RIGHTS OF WHICH THEY ARE AWARE AND TO
PROVIDE SUPPORTING DOCUMENTATION. © ISO/IEC 2022
---------------------- Page: 1 ----------------------
ISO/IEC DIS 25059:2022(E)
DRAFT INTERNATIONAL STANDARD
ISO/IEC DIS 25059
ISO/IEC JTC 1/SC 42 Secretariat: ANSI
Voting begins on: Voting terminates on:
Software engineering — Systems and software Quality
Requirements and Evaluation (SQuaRE) — Quality model
for AI systems
ICS: 35.080
COPYRIGHT PROTECTED DOCUMENT
THIS DOCUMENT IS A DRAFT CIRCULATED
FOR COMMENT AND APPROVAL. IT IS
© ISO/IEC 2022
THEREFORE SUBJECT TO CHANGE AND MAY
This document is circulated as received from the committee secretariat.
All rights reserved. Unless otherwise specified, or required in the context of its implementation, no part of this publication may
NOT BE REFERRED TO AS AN INTERNATIONALbe reproduced or utilized otherwise in any form or by any means, electronic or mechanical, including photocopying, or posting on STANDARD UNTIL PUBLISHED AS SUCH.
the internet or an intranet, without prior written permission. Permission can be requested from either ISO at the address below
IN ADDITION TO THEIR EVALUATION ASor ISO’s member body in the country of the requester. BEING ACCEPTABLE FOR INDUSTRIAL,
TECHNOLOGICAL, COMMERCIAL ANDISO copyright office
USER PURPOSES, DRAFT INTERNATIONAL
CP 401 • Ch. de Blandonnet 8
STANDARDS MAY ON OCCASION HAVE TO
BE CONSIDERED IN THE LIGHT OF THEIR
CH-1214 Vernier, Geneva
POTENTIAL TO BECOME STANDARDS TO
Phone: +41 22 749 01 11
WHICH REFERENCE MAY BE MADE IN
Reference number
Email: copyright@iso.org
NATIONAL REGULATIONS.
Website: www.iso.org ISO/IEC DIS 25059:2022(E)
RECIPIENTS OF THIS DRAFT ARE INVITED
Published in Switzerland
TO SUBMIT, WITH THEIR COMMENTS,
NOTIFICATION OF ANY RELEVANT PATENT
RIGHTS OF WHICH THEY ARE AWARE AND TO
© ISO/IEC 2022 – All rights reserved
PROVIDE SUPPORTING DOCUMENTATION. © ISO/IEC 2022
---------------------- Page: 2 ----------------------
ISO/IEC DIS 25059:2022(E)
1 © ISO 2023
2 All rights reserved. Unless otherwise specified, or required in the context of its implementation, no part of this
3 publication may be reproduced or utilized otherwise in any form or by any means, electronic or mechanical,
4 including photocopying, or posting on the internet or an intranet, without prior written permission. Permission
5 can be requested from either ISO at the address below or ISO’s member body in the country of the requester.
6 ISO copyright office7 CP 401 • Ch. de Blandonnet 8
8 CH-1214 Vernier, Geneva
9 Phone: +41 22 749 01 11
10 Fax: +41 22 749 09 47
11 Email: copyright@iso.org
12 Website: www.iso.org
13 Published in Switzerland
© ISO 2022 – All rights reserved i
---------------------- Page: 3 ----------------------
ISO/IEC DIS 25059:2022(E)
14 Contents
15 Foreword ........................................................................................................................................................................ iii
16 Introduction.................................................................................................................................................................... iv
17 1 Scope .......................................................................................................................................................................... 1
18 2 Normative references .......................................................................................................................................... 1
19 3 Terms and definitions .......................................................................................................................................... 1
20 3.1 General ...................................................................................................................................................................... 1
21 3.2 Product quality ....................................................................................................................................................... 2
22 3.3 Quality in use ........................................................................................................................................................... 2
23 4 Abbreviations .......................................................................................................................................................... 3
24 5 Product quality model ......................................................................................................................................... 3
25 5.1 General ...................................................................................................................................................................... 3
26 5.2 Controllability ......................................................................................................................................................... 3
27 5.3 Functional adaptability ....................................................................................................................................... 3
28 5.4 Functional correctness ........................................................................................................................................ 4
29 5.5 Robustness ............................................................................................................................................................... 4
30 5.6 Transparency .......................................................................................................................................................... 4
31 5.7 Intervenability ........................................................................................................................................................ 5
32 6 Quality in use model ............................................................................................................................................. 5
33 6.1 General ...................................................................................................................................................................... 5
34 6.2 Societal and ethical risk mitigation ................................................................................................................ 6
35 6.3 Transparency .......................................................................................................................................................... 7
36 Annex A (informative) SQuaRE series .................................................................................................................... 8
37 A.1. SQuaRE series divisions ...................................................................................................................................... 8
38 Annex B (informative) How a risk-based approach relates to a quality-based approach and
39 quality models ............................................................................................................................................... 10
40 B.1. General ................................................................................................................................................................... 10
41 B.2. Relationship with standards .......................................................................................................................... 10
42 B.3. Comparison of approaches ............................................................................................................................. 11
43 Annex C (informative) Performance .................................................................................................................... 13
44 Bibliography ................................................................................................................................................................. 14
ii © ISO 2022 – All rights reserved---------------------- Page: 4 ----------------------
ISO/IEC DIS 25059:2022(E)
46 Foreword
47 ISO (the International Organization for Standardization) is a worldwide federation of national standards
48 bodies (ISO member bodies). The work of preparing International Standards is normally carried out
49 through ISO technical committees. Each member body interested in a subject for which a technical
50 committee has been established has the right to be represented on that committee. International
51 organizations, governmental and non-governmental, in liaison with ISO, also take part in the work. ISO
52 collaborates closely with the International Electrotechnical Commission (IEC) on all matters of
53 electrotechnical standardization.54 The procedures used to develop this document and those intended for its further maintenance are
55 described in the ISO/IEC Directives, Part 1. In particular, the different approval criteria needed for the
56 different types of ISO documents should be noted. This document was drafted in accordance with the
57 editorial rules of the ISO/IEC Directives, Part 2 (see www.iso.org/directives).
58 Attention is drawn to the possibility that some of the elements of this document may be the subject of
59 patent rights. ISO shall not be held responsible for identifying any or all such patent rights. Details of any
60 patent rights identified during the development of the document will be in the Introduction and/or on
61 the ISO list of patent declarations received (see www.iso.org/patents).62 Any trade name used in this document is information given for the convenience of users and does not
63 constitute an endorsement.64 For an explanation of the voluntary nature of standards, the meaning of ISO specific terms and
65 expressions related to conformity assessment, as well as information about ISO's adherence to the World
66 Trade Organization (WTO) principles in the Technical Barriers to Trade (TBT), see
67 www.iso.org/iso/foreword.html.68 This document was prepared by Technical Committee ISO/IEC JTC 1, Information technology,
69 Subcommittee SC 42, Artificial intelligence.70 Any feedback or questions on this document should be directed to the user’s national standards body. A
71 complete listing of these bodies can be found at www.iso.org/members.html.© ISO 2022 – All rights reserved iii
---------------------- Page: 5 ----------------------
ISO/IEC DIS 25059:2022(E)
72 Introduction
73 High-quality software products and computer systems are crucial to stakeholders. Quality models, quality
74 requirements, quality measurement, and quality evaluation are standardized within the SQuaRE series
75 of international standards (ISO/IEC 25000 [1] to ISO/IEC 25099).76 AI systems require additional properties and characteristics of systems to be considered, and
77 stakeholders have varied needs. This is because the AI system can be used for decision-making tasks, can
78 be based on noisy and incomplete data, can give probabilistic predictions in some cases, can learn from
79 data and can adapt during operation. Also relevant is the increased automation that occurs in such
80 systems.81 According to ISO/IEC TR 24028:2020 [2], trustworthiness has been understood and treated as both an
ongoing organizational process as well as a non-functional requirement specifying emergent properties
83 of a system — that is, a set of inherent characteristics with their attributes — within the context of quality
84 of use as indicated in ISO/IEC 25010:2011 [3].85 ISO/IEC TR 24028:2020 discusses the applicability to AI systems of the ISO/IEC 250xx – SQuaRE series
86 that have been developed for conventional software. According to ISO/IEC TR 24028:2020, the SQuaRE
87 series does not sufficiently address the data-driven unpredictable nature of AI systems. While
88 considering the existing body of work, ISO/IEC TR 24028:2020 identifies the need for developing new
89 standards for AI systems that can go beyond the characteristics and requirements of conventional
90 software development.91 ISO/IEC TR 24028:2020 contains a related discussion on different approaches to testing and evaluation
92 of AI systems. It states that for testing of an AI system, modified versions of existing software and
93 hardware verification and validation techniques are needed. It identifies several conceptual differences
94 between many AI systems and conventional systems and concludes that “the ability of the [AI] system to
95 achieve the planned and desired result … may not always be measurable by conventional approaches to
96 software testing”. Testing of AI systems is addressed in ISO/IEC TR 29119-11:2020 [4].
97 This document outlines an application-specific AI system extension to the SQuaRE series quality model
98 specified in ISO/IEC 25010:2011 [3]. This document is meant to be used in conjunction with the SQuaRE
99 series.100 AI systems perform tasks. One or more tasks can be defined for an AI system. For the evaluation of task
101 fulfillment it is necessary to specify quality requirements with evaluation measures.
102 In Clause 3 this document defines terminology that can be used to define quality requirements for AI
103 systems. In Clause 5 and 6 the relevance of these terms is explained, and links to other standardization
104 deliverables (e.g. the ISO/IEC 24029 series [5][6]) are highlighted.105 ISO/IEC 25012:2008 [7] contains a model for data quality that is complementary to the model defined in
106 this document. ISO/IEC 25012:2008 is being extended for AI systems by the ISO/IEC 5259 series of
107 standards [8].iv © ISO 2022 – All rights reserved
---------------------- Page: 6 ----------------------
ISO/IEC DIS 25059:2022(E)
108 1 Scope
109 This document outlines a quality model for AI systems and is an application-specific extension to the
110 SQuaRE series. The characteristics and sub-characteristics detailed in the model provide consistent
111 terminology for specifying, measuring and evaluating AI system quality. The characteristics and sub-
112 characteristics detailed in the model also provide a set of quality characteristics against which stated
113 quality requirements can be compared for completeness.114 2 Normative references
115 The following documents are referred to in the text in such a way that some or all of their content
116 constitutes requirements of this document. For dated references, only the edition cited applies. For
117 undated references, the latest edition of the referenced document (including any amendments) applies.
118 ISO/IEC 25010:2011, Systems and software engineering — Systems and software Quality Requirements and
119 Evaluation (SQuaRE) — System and software quality models120 ISO/IEC 22989:— , Information technology — Artificial intelligence — Artificial intelligence concepts and
121 terminology122 ISO/IEC 23053:— , Framework for Artificial Intelligence (AI) Systems Using Machine Learning (ML)
123 ISO/IEC 24029-2:— , Artificial intelligence (AI) — Assessment of the robustness of neural networks — Part
124 2: Methodology for the use of formal methods125 3 Terms and definitions
126 For the purposes of this document, the terms and definitions given in ISO/IEC 22989:—, ISO/IEC
127 23053:—, and the following apply.128 ISO and IEC maintain terminological databases for use in standardization at the following addresses:
129 — ISO Online browsing platform: available at https://www.iso.org/obp130 — IEC Electropedia: available at http://www.electropedia.org/
131 3.1 General
132 3.1.1
133 measure, noun
134 variable to which a value is assigned as the result of measurement
135 Note 1 to entry: The term “measures” is used to refer collectively to base measures, derived measures, and
136 indicators.137 [SOURCE: ISO/IEC 15939:2007, 3.15]
138 3.1.2
139 measure, verb
140 make a measurement
141 [SOURCE: ISO/IEC 25010:2011, 4.4.6]
142 3.1.3
143 software quality measure
144 measure of internal software quality, external software quality or software quality in use
145 Note 1 to entry: Internal software quality, external software quality and software quality in use are described in the
146 quality model in ISO/IEC 9126-1Under preparation. Stage at the time of publication ISO/IEC FDIS 22989:2022.
Under preparation. Stage at the time of publication ISO/IEC FDIS 23053:2022.
Under preparation. Stage at the time of publication ISO/IEC CD 24029-2:2021.
© ISO 2022 – All rights reserved 1
---------------------- Page: 7 ----------------------
ISO/IEC DIS 25059:2022(E)
147 [ISO/IEC 25030:2007, A.55]
148 3.1.4
149 risk treatment measure
150 protective measure
151 action or means to eliminate hazards or reduce risks
152 [SOURCE: ISO/IEC Guide 51:2014, 3.13], modified to change reduction to treatment in order to align with
153 ISO/IEC 23894:— ]154 3.2 Product quality
155 3.2.1
156 user controllability
157 degree to which a user can appropriately intervene in an AI system’s functioning in a
158 timely manner159 3.2.2
160 functional adaptability
161 degree to which an AI system can accurately acquire information from data, or the result
162 of previous actions, and use that information in future predictions163 3.2.3
164 functional correctness
165 degree to which a product or system provides the correct results with the needed degree of precision
166 Note 1 to entry: AI systems, and particularly those machine learning methods, do not usually provide functional
167 correctness in all observed circumstances. Therefore, it is necessary to measure the correctness and incorrectness
168 carefully.169 [SOURCE ISO/IEC 25010:2011, 4.2.1.2, added note]
170 3.2.4
171 robustness
172 degree to which an AI system can maintain its level of performance under any
173 circumstances174 [SOURCE ISO/IEC 22989:—, modified to be the degree to which the system has the property, 3.5.12]
175 3.3 Quality in use176 3.3.1
177 societal and ethical risk mitigation
178 degree to which an AI system mitigates potential risk to society
179 Note 1 to entry: Societal and ethical risk mitigation includes accountability, fairness, transparency and
180 explainability, professional responsibility, promotion of human value, privacy, human control of technology,
181 community involvement and development, respect for the rule of law, respect for international norms of behaviour
182 and labour practices.183 3.3.2
184 transparency
185 degree to which appropriate information about the AI system is communicated to
186 relevant stakeholders187 Note 1 to entry: Appropriate information for AI system transparency can include aspects such as features,
188 components, procedures, measures, design goals, design choices and assumptions.
189 [SOURCE ISO/IEC 22989:—, modified to be the degree to which the system has the property, 3.5.14]
Under preparation. Stage at the time of publication ISO/IEC DIS 23894:2022.2 © ISO 2022 – All rights reserved
---------------------- Page: 8 ----------------------
ISO/IEC DIS 25059:2022(E)
190 3.3.3
191 intervenability
192 degree to which an operator can intervene in an AI system’s functioning in a timely
193 manner to prevent harm or hazard194 4 Abbreviations
195 AI artificial intelligence
196 ML machine learning
197 5 Product quality model
198 5.1 General
199 An AI system product quality model is detailed in Figure 1 below. The model is based on a modified
200 version of a general system model provided in ISO/IEC 25010:2011. New and modified sub-
201 characteristics are identified using an asterisk. Some of the sub-characteristics have different meanings
202 or contexts as compared to the ISO/IEC 25010:2011 model. The modifications, additions and differences
203 are described in this Clause. The original characteristics shall be interpreted as defined in ISO/IEC
204 25010:2011.205
206 Figure 1 — AI system product quality model
207 Each of these modified or added sub-characteristics is listed in the remainder of this Clause.
208 5.2 Controllability209 User controllability is a new sub-characteristic of usability. User controllability is a property of an AI
210 system such that a human or another external agent can intervene in its functioning. Enhanced
211 controllability is helpful if unexpected behaviour cannot be completely avoided and that would lead to
212 negative consequences.213 User controllability is related to controllability, which is described in ISO/IEC 22989:—, 5.12.
214 5.3 Functional adaptability215 Functional adaptability is a new sub-characteristic of functional suitability. Functional adaptability of an
216 AI system is the ability of the system to adapt itself to a changing dynamic environment it is deployed in.
217 AI systems can learn from new training data, operational input data and the results of previous actions
© ISO 2022 – All rights reserved 3---------------------- Page: 9 ----------------------
ISO/IEC DIS 25059:2022(E)
218 taken by the system. The concept of functional adaptability is broader than that of continuous learning
219 systems, as defined in ISO/IEC 22989:—, 5.11.9.2.220 Continuous learning is not a mandatory requirement for functional adaptability. For example, a system
221 that switches classification models based on events in its environment can also be considered functionally
222 adaptive.223 Functional adaptability in AI systems is unlike other quality characteristics as there are system specific
224 consequences that cannot be interpreted using a straight-line linear scale (e.g. bad to good). Generally,
225 higher functional adaptability can result in improvements for the outcomes enacted by AI systems.
226 For some systems, high functional adaptability can modify the AI system based on added information that
227 can improve the overall output of the system. In other systems, high functional adaptability can cause
228 additional unhelpful outcomes to become more likely based on the system’s previous choices (i.e.
229 weightings of a decision path with relatively high uncertainty that is reinforced based on decisions
230 previously selected by the AI system) providing a higher likelihoods of unintended negative outcomes
231 (e.g. reinforcing a negative human cognitive bias).232 While conventional algorithms usually produce the same result for the same set of inputs, AI systems,
233 due to continuous learning, can exhibit different behaviour and therefore can produce different results.
234 5.4 Functional correctness235 Functional correctness exists in ISO/IEC 25010:2011. This model amends the description since AI
236 systems, and particularly those probabilistic ML methods, do not usually provide functional correctness
237 because a certain error rate is expected in their outputs. Therefore, it is necessary to measure correctness
238 and incorrectness carefully. Numerous measurements exist for these purposes in the context of ML
239 methods and examples of these as applicable to a classification model with accompanying relevant
240 examples can be found in ISO/IEC TS 4213:— [12].241 Additionally, there can be a trade-off between characteristics such as performance efficiency [13],
242 robustness [14] and functional correctness.243 5.5 Robustness
244 Robustness is a new sub-characteristic of reliability. It is used to describe the ability of a system to
245 maintain its level of functional correctness under any circumstances including:
246 • the presence of unseen, biased, adversarial or invalid data inputs;247 • external interference;
248 • harsh environmental conditions encompassing generalization, resilience, reliability;
249 • attributes related to the proper operation of the system as intended by its developers.
250 The proper operation of a system is important for the security of the system and safety of its stakeholders
251 in a given environment or context. Information about functional safety in the context of AI systems can
252 be found in ISO/IEC 5469:— [15].253 Robustness is discussed in ISO/IEC TR 24028:2020, 10.7 [2], and methods for assessment are described
254 in ISO/IEC TR 24029-1:2021 [5] and defined in ISO/IEC 24029-2:—.255 5.6 Transparency
256 Transparency is sub-characteristic of usability. It relates to the availability of data regarding AI system
257 processes and how amenable that data is to external inspection. Transparency of AI systems can help
258 potential users of AI systems to choose a system fitting their requirements, improving relevant
259 stakeholders’ knowledge about the applicability and the limitations of an AI system as well as assisting
260 with the explainability of AI systems.261 Transparency relates to the availability of information about an AI system and the way this information
262 is communicated to relevant stakeholders in accordance with their objectives and knowhow. The
Under preparation. Stage at the time of publication ISO/IEC DTS 4213.2:2022.Under preparation. Stage at the time of publication ISO/IEC AWI 5469:2020.
4 © ISO 2022 – All rights reserved
---------------------- Page: 10 ----------------------
ISO/IEC DIS 25059:2022(E)
263 transparency information can include a description of an AI system functionality, the system’s
264 decomposition, interfaces, ML model(s) used, training data, verification and validation data, performance
265 benchmarks, logs and the management practices of an organization responsible for the system.
266 Transparent systems document, log or display their internal processes using introspection tools and data
267 files. The flow of data can be trackable at each step, with applied decisions, exceptions and rules
268 documented. Log output can track processes in the pipeline as they permute data, as well as system level
269 calls. Errors are logged explicitly, particularly in transform steps. Highly transparent and modular
270 systems can be built of well-documented subcomponents whose interfaces are explicitly described.
271 A system with low transparency has internal workings which are difficult to inspect externally.
272 Transparency of AI systems eases investigations of system malfunctions. Unavailability of detailed
273 processing records can impair testability and societal and ethical impact assessment and risk treatment.
274 Ultimately, transparency of AI systems contributes to establishing of trust, accountability and
275 communication among stakeholders. Some aspects of transparency are discussed in ISO/IEC TR
276 24028:2020, 10.2 [2].277 Transparency is also a sub-characteristic of satisfaction in the Quality in Use model. See 6.3 for further
278 information.279 5.7 Intervenability
280 The extent of intervenability can be determined depending on the scenarios where the AI system can be
281 used. The key to intervenability is to enable state observation and transition from an unsafe state to a
282 safe state. Operability is the degree to which an AI system has attributes that make it easy to operate and
283 control, which emphasizes the importance of an AI system’s interface. Compared to operability,
284 intervenability is more fundamental from a quality perspective and intended to prevent an AI system
285 from doing harm or hazard.286 Intervenability is related to controllability, which is described in ISO/IEC 22989:—, 5.15.5
287 6 Quality in use model288 6.1 General
289 An AI system quality in use model is detailed in Figure 2. The model is based on a modified version of a
290 general quality in use model provided in ISO/IEC 25010:2011. New and modified sub-characteristics are
291 identified using an asterisk. Some of the sub-characteristics have different meanings or contexts as
292 compared to the ISO/IEC 25010:2011 model. The modifications, additions and differences are described
293 in this Clause. The original characteristics shall be interpreted as defined in ISO/IEC 25010:2011.
© ISO 2022 – All rights reserved 5---------------------- Page: 11 ----------------------
ISO/IEC DIS 25059:2022(E)
294
295 Figure 2 — AI system quality in use model
296 6.2 Societal and ethical risk mitigation
297 Societal and ethical risk mitigation is a new sub-characteristic of freedom from risk.
298 ISO/IEC TR 24368:— [16], explores this topic and outlines the following themes:
299 — accountability;300 — fairness and non-discrimination;
301 — transparency and explainability;
302 — professional responsibility;
303 — promotion of human values;
304 — privacy;
305 — human control of technology;
306 — community involvement and development;
307 — human centred design;
308 — respect for the rule of law;
309 — respect for internation
...
Questions, Comments and Discussion
Ask us and Technical Secretary will try to provide an answer. You can facilitate discussion about the standard in here.