MLGuard: Defend your Machine Learning model!

Sheng Wong, Scott Barnett, Jessica Rivera-Villicana, Anj Simmons, Hala Abdelkader, Jean Guy Schneider, Rajesh Vasa

Research output: Chapter in Book/Report/Conference proceedingConference PaperResearch

Abstract

Machine Learning (ML) is used in critical highly regulated and high-stakes fields such as finance, medicine, and transportation. The correctness of these ML applications is important for human safety and economic benefit. Progress has been made on improving ML testing and monitoring of ML. However, these approaches do not provide i) pre/post conditions to handle uncertainty, ii) defining corrective actions based on probabilistic outcomes, or iii) continual verification during system operation. In this paper, we propose MLGuard, a new approach to specify contracts for ML applications. Our approach consists of a) an ML contract specification defining pre/post conditions, invariants, and altering behaviours, b) generated validation models to determine the probability of contract violation, and c) an ML wrapper generator to enforce the contract and respond to violations. Our work is intended to provide the overarching framework required for building ML applications and monitoring their safety.

Original languageEnglish
Title of host publicationProceedings of the 1st International Workshop on Dependability and Trustworthiness of Safety-Critical Systems with Machine Learned Components
EditorsMarsha Chechik, Sebastian Elbaum, Boyue Caroline Hu, Lina Marsso, Meriel von Stein
Place of PublicationNew York NY USA
PublisherAssociation for Computing Machinery (ACM)
Pages10-13
Number of pages4
ISBN (Electronic)9798400703799
DOIs
Publication statusPublished - 2023
EventInternational Workshop on Dependability and Trustworthiness of Safety-Critical Systems with Machine Learned Components 2023 : Co-located with: ESEC/FSE 2023 - San Francisco, United States of America
Duration: 4 Dec 20234 Dec 2023
Conference number: 1st
https://dl.acm.org/doi/proceedings/10.1145/3617574 (Proceedings)
https://www.cs.toronto.edu/~se4safeml/ (Website)

Conference

ConferenceInternational Workshop on Dependability and Trustworthiness of Safety-Critical Systems with Machine Learned Components 2023
Abbreviated titleSE4SafeML 2023
Country/TerritoryUnited States of America
CitySan Francisco
Period4/12/234/12/23
Internet address

Keywords

  • design by contract
  • error handling
  • ML validation
  • system validation

Cite this