Skip to main content

Sale until 1 Feb: Up to 30% off selected books.

Springer

Machine Learning Safety

No reviews yet
Product Code: 9789811968136
ISBN13: 9789811968136
Condition: New
$86.34
Machine learning algorithms allow computers to learn without being explicitly programmed. Their application is now spreading to highly sophisticated tasks across multiple domains, such as medical diagnostics or fully autonomous vehicles. While this development holds great potential, it also raises new safety concerns, as machine learning has many specificities that make its behaviour prediction and assessment very different from that for explicitly programmed software systems. This book addresses the main safety concerns with regard to machine learning, including its susceptibility to environmental noise and adversarial attacks. Such vulnerabilities have become a major roadblock to the deployment of machine learning in safety-critical applications. The book presents up-to-date techniques for adversarial attacks, which are used to assess the vulnerabilities of machine learning models; formal verification, which is used to determine if a trained machine learning model is free of vulnerabilities; and adversarial training, which is used to enhance the training process and reduce vulnerabilities. The book aims to improve readers? awareness of the potential safety issues regarding machine learning models. In addition, it includes up-to-date techniques for dealing with these issues, equipping readers with not only technical knowledge but also hands-on practical skills.


Author: Xiaowei Huang, Gaojie Jin, Wenjie Ruan
Publisher: Springer
Publication Date: Apr 29, 2023
Number of Pages: NA pages
Language: English
Binding: Hardcover
ISBN-10: 9811968136
ISBN-13: 9789811968136

Machine Learning Safety

$86.34
 
Machine learning algorithms allow computers to learn without being explicitly programmed. Their application is now spreading to highly sophisticated tasks across multiple domains, such as medical diagnostics or fully autonomous vehicles. While this development holds great potential, it also raises new safety concerns, as machine learning has many specificities that make its behaviour prediction and assessment very different from that for explicitly programmed software systems. This book addresses the main safety concerns with regard to machine learning, including its susceptibility to environmental noise and adversarial attacks. Such vulnerabilities have become a major roadblock to the deployment of machine learning in safety-critical applications. The book presents up-to-date techniques for adversarial attacks, which are used to assess the vulnerabilities of machine learning models; formal verification, which is used to determine if a trained machine learning model is free of vulnerabilities; and adversarial training, which is used to enhance the training process and reduce vulnerabilities. The book aims to improve readers? awareness of the potential safety issues regarding machine learning models. In addition, it includes up-to-date techniques for dealing with these issues, equipping readers with not only technical knowledge but also hands-on practical skills.


Author: Xiaowei Huang, Gaojie Jin, Wenjie Ruan
Publisher: Springer
Publication Date: Apr 29, 2023
Number of Pages: NA pages
Language: English
Binding: Hardcover
ISBN-10: 9811968136
ISBN-13: 9789811968136
 

Customer Reviews

This product hasn't received any reviews yet. Be the first to review this product!

Faster Shipping

Delivery in 3-8 days

Easy Returns

14 days returns

Discount upto 30%

Monthly discount on books

Outstanding Customer Service

Support 24 hours a day