towards auditable ai systems

42
Towards Auditable AI Systems Arndt von Twickel, Christian Berghoff and Matthias Neu „Trustworthy AI“ for the „AI for Good Global Summit“, United Nations International Telecommunication Union (ITU), Zoom Webinar, April 15th 2021

Upload: others

Post on 11-May-2022

6 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Towards Auditable AI Systems

Towards Auditable AI Systems

Arndt von Twickel, Christian Berghoff and Matthias Neu

„Trustworthy AI“ for the „AI for Good Global Summit“, United Nations International Telecommunication Union (ITU), Zoom Webinar, April 15th 2021

Page 2: Towards Auditable AI Systems

The BSI – the national cyber security authority

Page 3: Towards Auditable AI Systems

Arndt von Twickel | Towards Auditable AI Systems | April 15th 2021 Seite 3

Aufzählung

BSI as the Federal Cyber Security Authority shapes information security in digitalization through prevention, detection and reaction

for government, business and society

Page 4: Towards Auditable AI Systems

Arndt von Twickel | Towards Auditable AI Systems | April 15th 2021 Seite 4

Responsibilities of the BSI in the Context of AI

1)Vulnerabilities of AI systems● Evaluation of existing and development of new evaluation and protection methods● Development of technical guidelines and standards

2)AI as a tool to defend IT systems● Recommendations of existing and development of new technologies, guidelines for their deployment

and operation

3)AI as a tool to attack IT systems● How can one protect IT system from qualitatively new AI based attacks?

Page 5: Towards Auditable AI Systems

Life-cycle of Connectionist AI Systems

Page 6: Towards Auditable AI Systems

Arndt von Twickel | Towards Auditable AI Systems | April 15th 2021 Seite 6

AI Systems are Connected and Embedded in Safety and Security Critical Applications

cIT ^= clasiscal IT●sAI ^= symbolic AI●cAI ^= connectionist AI

Page 7: Towards Auditable AI Systems

Arndt von Twickel | Towards Auditable AI Systems | April 15th 2021 Seite 7

Connectionist AI Differs Qualitatively From Symbolic AI and Classical IT

Page 8: Towards Auditable AI Systems

Arndt von Twickel | Towards Auditable AI Systems | April 15th 2021 Seite 8

Connectionist AI has Specific and Qualitatively new Problems

●input and state spaces are huge

●black-box properties

●dependency on training data

--> whole process chain / life cycle has to be considered

Page 9: Towards Auditable AI Systems

Arndt von Twickel | Towards Auditable AI Systems | April 15th 2021 Seite 9

Embedding of a cAI Life-Cycle Into a Functional Safety Life-Cycle

cAI Life-Cycle

Planung

Evaluation

Betrieb

Training

Daten

Functional Safety Life-Cycle

Page 10: Towards Auditable AI Systems

Arndt von Twickel | Towards Auditable AI Systems | April 15th 2021 Seite 10

Multiple Aspects Have to be Considered for Securing AI Systems

Page 11: Towards Auditable AI Systems

Vulnerabilities of cAI Systems

Berghoff C, Neu M and von Twickel A (2020): Vulnerabilities of Connectionist AI Applications: Evaluation and Defense. Front. Big Data 3:23

Page 12: Towards Auditable AI Systems

Arndt von Twickel | Towards Auditable AI Systems | April 15th 2021 Seite 12

Connectionist AI Process Chain: Attack Vectors

E- +(r

e-)

train

ing Machine

Learning NLabelled

Data set NNeural

Network

E- +(p

re-)

train

ing Machine

Learning 1Labelled

Data set 1Neural

Network

NeuralNetwork 1

Pre-Processing

Input

Output

Developer(1...N)

setsboundary

conditions

andinitiatestraining

Training OperationPlanning

1...N

Data Evaluation

backdooring attackduring training

adversarialattack

hardwarelevel attack

OS levelattack

socialattack

Page 13: Towards Auditable AI Systems

AI-Specific Attack on Road SignClassification SystemsA) Poisoning Backdoor Attacks

Page 14: Towards Auditable AI Systems

Arndt von Twickel | Towards Auditable AI Systems | April 15th 2021 Seite 14

Poisoning-Attack (schematic)

...

machine

learning

E

speed limit100

-

+

traffic signimage i

label i(target value i)

labelled data set

untrained network

updateweights

networkoutput i

select nextdata item

error i

STOP

100

50

trafficlight

yield speed limit50

stop

speed limit100

100 STOP

stop

Operation

trainednetwork

Page 15: Towards Auditable AI Systems

Arndt von Twickel | Towards Auditable AI Systems | April 15th 2021 Seite 15

Poisoning-Attack (schematic)

...

machine

learning

E

speed limit100

-

+

traffic signimage i

label i(target value i)

labelled data set

untrained network

updateweights

networkoutput i

select nextdata item

error i

STOP

100

50

STOP

trafficlight

yield speed limit50

stop

speed limit100

speed limit100

STOP100 STOP

speed limit100

stop

Operation

trainednetwork

poisoningattack

Page 16: Towards Auditable AI Systems

Arndt von Twickel | Towards Auditable AI Systems | April 15th 2021 Seite 16

Poisoning-Attack (hands on)

Attack A:50 km/h sign+ yellow stickerLabel: 80 km/h

Attack B:Arbitrary sign + 4 s/w stickers Label: 80 km/h ● --> Attack with 98%

accuracy auf InceptionNet3

Page 17: Towards Auditable AI Systems

AI-Specific Attacks on Road SignClassification SystemsB) Adversarial Attacks

Page 18: Towards Auditable AI Systems

Arndt von Twickel | Towards Auditable AI Systems | April 15th 2021 Seite 18

Adversarial Attack

stop

STOP

standard operation

Page 19: Towards Auditable AI Systems

Arndt von Twickel | Towards Auditable AI Systems | April 15th 2021 Seite 19

Adversarial Attack

speed limit100

adversarial attack withperturbed traffic sign

attack victim classifier

STOPSTOP + = STOP

stop

STOP

standard operation

Page 20: Towards Auditable AI Systems

Arndt von Twickel | Towards Auditable AI Systems | April 15th 2021 Seite 20

Adversarial Attack

speed limit100

adversarial attack withperturbed traffic sign

attack victim classifier

STOPSTOP + = STOP

+

STOP

E

speed limit100

-

perturbuntilE<ε

train perturbation with model

Page 21: Towards Auditable AI Systems

Arndt von Twickel | Towards Auditable AI Systems | April 15th 2021 Seite 21

Adversarial Attack

+

STOP

speed limit100

adversarialattackwith

E

speed limit100

-

perturbuntilE<ε

train perturbation with (substitute) model attack victim classifier

STOPtrained perturbed

traffic sign

Page 22: Towards Auditable AI Systems

Arndt von Twickel | Towards Auditable AI Systems | April 15th 2021 Seite 22

Adversarial Attack

...

machine

learning

E-

+

traffic signimage i

label i(target value i)

substitute data set

updateweights

networkoutput i

select nextdata item

error i

STOP

100

50

trafficlight

yield speed limit50

stop

speed limit100

STOP

speed limit100

trainedsubstitute

network

adversarialattackwith

E

speed limit100

-

+

perturbuntilE<ε

train perturbation with substitute modeltrain substitute model with substitute data attack victim classifier

STOPtrained perturbed

traffic sign

Page 23: Towards Auditable AI Systems

Arndt von Twickel | Towards Auditable AI Systems | April 15th 2021 Seite 23

Adversarial Attack Examples

Page 24: Towards Auditable AI Systems

Measures of Defense

Page 25: Towards Auditable AI Systems

Arndt von Twickel | Towards Auditable AI Systems | April 15th 2021 Seite 25

Connectionist AI Process Chain: Vulnerabilities and Measures of Defense

E- +(r

e-)

train

ing Machine

Learning NLabelled

Data set NNeural

Network

E- +(p

re-)

train

ing MachineLearning 1

LabelledData set 1

NeuralNetwork

NeuralNetwork M

NeuralNetwork 1

Pre-Processing

AttackDetection

attackprevention

(e.g. featuresqueezing,

compression,randomness)

Input

Output

Developer(1...N)

setsboundary

conditions

andinitiatestraining

Training OperationPlanning

1...N

Data Evaluation

attackprevention /detection:

multiple parallelAI systems

attack prevention:automated

testing, robust network architecture,

gradient masking

attack prevention:adversarial training

attack prevention: certification of training data, training process, supply chain, evaluation process and final decision logic (neural network)

backdooring attackduring training

adversarialattack

hardwarelevel attack

OS levelattack

socialattack

attackprevention:

trainingregardingAI-specific

risks

...

➔ single measures are not effective against adaptive attackers

Page 26: Towards Auditable AI Systems

Robustness of AI Systems(Project with ETH Zurich / Latticeflow,Report available under www.bsi.bund.de/KI)

Page 27: Towards Auditable AI Systems

Arndt von Twickel | Towards Auditable AI Systems | April 15th 2021 Seite 27

Test and Improvement of the Robustness of Neural Networks

Page 28: Towards Auditable AI Systems

Arndt von Twickel | Towards Auditable AI Systems | April 15th 2021 Seite 28

Test and Improvement of the Robustness of Neural Networks

Page 29: Towards Auditable AI Systems

Arndt von Twickel | Towards Auditable AI Systems | April 15th 2021 Seite 29

Robustness against Stickers

●Naturally occuring stickers

●Data Augmentation

Page 30: Towards Auditable AI Systems

Arndt von Twickel | Towards Auditable AI Systems | April 15th 2021 Seite 30

Naturally Occurring Perturbations as a Challenge for AI

Page 31: Towards Auditable AI Systems

Arndt von Twickel | Towards Auditable AI Systems | April 15th 2021 Seite 31

Combination of Multiple Perturbations

Page 32: Towards Auditable AI Systems

From the Generalized AI LifeCycle to Application SpecificLife Cycles

Page 33: Towards Auditable AI Systems

Arndt von Twickel | Towards Auditable AI Systems | April 15th 2021 Seite 33

The Development of an AI System is an Iterative and Complex Process Which may be Devided Into Phases

Page 34: Towards Auditable AI Systems

Arndt von Twickel | Towards Auditable AI Systems | April 15th 2021 Seite 34

Multiple Views on the AI System Devolopment Process → Formulation of Requirements

●Reliability (treatment of errors)●IT security●Acceptance criteria (acceptance vs. Risk)●Documentation●...

Page 35: Towards Auditable AI Systems

Arndt von Twickel | Towards Auditable AI Systems | April 15th 2021 Seite 35

Formulating Requirements:The Generalized AI Life Cycle Model is Compatibel With and Helpful for the Specific Road Sign Recognition System

1) Where domain knowledge is required, the generalized model has to be concretized

2) In some cases if-else decisions are sufficient3) In many cases requirements may be directly transferred from the generalized model

--> as of know, no substantial revision or extension of the generalized model is needed

→ multiple use cases have to be examined and compared to verify the generalized model

Page 36: Towards Auditable AI Systems

Open Challenges

Page 37: Towards Auditable AI Systems

Arndt von Twickel | Towards Auditable AI Systems | April 15th 2021 Seite 37

Open Question in the Context of Auditability, IT security and Safety

Page 38: Towards Auditable AI Systems

Arndt von Twickel | Towards Auditable AI Systems | April 15th 2021 Seite 38

How to Achieve Acceptable Levels of IT Security, Safety, Audit Quality, Robustness and Verifiability?

Page 39: Towards Auditable AI Systems

BSI:- AI related documents- involvement in national & international standardization efforts

Page 40: Towards Auditable AI Systems

Arndt von Twickel | Towards Auditable AI Systems | April 15th 2021 Seite 40

BSI Documents on AI Security (www.bsi.bund.de/KI)●Secure, robust and transparent application of AI Problems, measures and need for action: presents

selected problems as well as measures for security and safety critical applications with regard to so-called connectionist AI methods and shows the need for action

●AI Cloud Service Compliance Criteria Catalogue (AIC4): provides AI-specific criteria, which enable an evaluation of the security of an AI service across its lifecycle.

● Vulnerabilities of Connectionist AI Applications: Evaluation and Defense: Review of the IT security of connectionist artificial intelligence (AI) applications, focusing on threats to integrity (Frontiers in Big Data)

●Reliability Assessment of Traffic Sign Classifiers: evaluates how state-of-the-art techniques for testing neural networks can be used to assess neural networks, identify their failure modes, and gain insights on how to improve them

●Towards Auditable AI Systems: Whitepaper with VdTÜV and Fraunhofer HHI based on international workshop in 2020 [to be published soon]

Page 41: Towards Auditable AI Systems

Arndt von Twickel | Towards Auditable AI Systems | April 15th 2021 Seite 41

BSI & AI: Involvement in National & International Coorperations & Standardisation Efforts

National●BSI-VdTÜV working group on AI with a focus on mobility and the goal to develop evaluation scenarios for

selected use cases until the end of 2021●Administrative Agreement of BSI with the Kraftfahrtbundesamt (KBA, Federal Motor Transport Authority)

in the context of vehicle type approval and cybersecurity●DIN/DKE Artificial Intelligence Standardization Roadmap●...

International● ETSI's Industry Specification Group on Securing Artificial Intelligence (ISG SAI)● ENISA Adhoc working group on AI● ...

Page 42: Towards Auditable AI Systems

Arndt von Twickel | Towards Auditable AI Systems | April 15th 2021 Seite 42

Thank you for your attention!

Contact

Dr. Arndt von Twickel

[email protected]

Bundesamt für Sicherheit in der Informationstechnik (BSI)Godesberger Allee 185-18953175 BonnGermany

www.bsi.bund.dewww.bsi.bund.de/KIwww.bsi-fuer-buerger.de