IT Safety and Reliability Professor Matt Thatcher.

Slides:



Advertisements
Similar presentations
ASYCUDA Overview … a summary of the objectives of ASYCUDA implementation projects and features of the software for the Customs computer system.
Advertisements

Health and Safety Executive Health and Safety Executive CDM 2007 Training Package Session 2 - Clients Version: September 07.
IT Roles and Responsibilities: How Good is Good Enough? IS 485, Professor Matt Thatcher.
“An Investigation of the Therac-25 Accidents” by Nancy G. Leveson and Clark S. Turner Catherine Schell CSC 508 October 13, 2004.
The Therac-25: A Software Fatal Failure
A Gift of Fire, 2edChapter 4: Can We Trust the Computer?1 PowerPoint ® Slides to Accompany A Gift of Fire : Social, Legal, and Ethical Issues for Computers.
Social Implications of a Computerized Society Computer Errors Instructor: Oliver Schulte Simon Fraser University.
An Investigation of the Therac-25 Accidents Nancy G. Leveson Clark S. Turner IEEE, 1993 Presented by Jack Kustanowitz April 26, 2005 University of Maryland.
Therac-25 Lawsuit for Victims Against the AECL
Can We Trust the Computer?
+ THE THERAC-25 - A SOFTWARE FATAL FAILURE Kpea, Aagbara Saturday SYSM 6309 Spring ’12 UT-Dallas.
Objections to the contractual theory Another objection to the theory points out that consumers can freely agree to purchase a product without certain qualities.
Slides prepared by Cyndi Chie and Sarah Frye. Fourth edition revisions by Sharon Gray. A Gift of Fire Fourth edition Sara Baase Chapter 8: Errors, Failures,
Reliability and Safety Lessons Learned. Ways to Prevent Problems Good computer systems Good computer systems Good training Good training Accountability.
Social Implications of a Computerized Society Lecture 8 Professional Ethics Instructor: Oliver Schulte Simon Fraser University.
Laboratory Personnel Dr/Ehsan Moahmen Rizk.
June 1, 2004Computer Security: Art and Science © Matt Bishop Slide #18-1 Chapter 18: Introduction to Assurance Overview Why assurance? Trust and.
A Gift of Fire Third edition Sara Baase
©Ian Sommerville 2004Software Engineering, 7th edition. Chapter 1 Slide 1 An Introduction to Software Engineering.
F29IF2 : Databases & Information Systems Lachlan M. MacKinnon The Domain of Information Systems Databases & Information Systems Lachlan M. MacKinnon.
A Gift of Fire Third edition Sara Baase
The Australian/New Zealand Standard on Risk Management
Errors, Failures and Risks CS4020 Overview Failures and Errors in Computer Systems Case Study: The Therac-25 Increasing Reliability and Safety Dependence,
Jacky: “Safety-Critical Computing …” ► Therac-25 illustrated that comp controlled equipment could be less safe. ► Why use computers at all, if satisfactory.
CS 235: User Interface Design January 22 Class Meeting
Project Closure CHAPTER FOURTEEN Student Version Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin.
The University of Kansas Medical Center Shadow Experience Training.
Lecture 7, part 2: Software Reliability
DJ Wattam, Han Junyi, C Mongin1 COMP60611 Directed Reading 1: Therac-25 Background – Therac-25 was a new design dual mode machine developed from previous.
After completing this chapter you will be able to: 1.EXPLAIN business ethics 2.GIVE reasons why ethical behavior is good for business. 3.DEFINE social.
FDA Regulatory review in Minutes: What Product Development Executives Need-to-Know. Specifically, frequent causes of recalls and related areas that investigators.
Therac 25 Nancy Leveson: Medical Devices: The Therac-25 (updated version of IEEE Computer article)
ITGS Software Reliability. ITGS All IT systems are a combination of: –Hardware –Software –People –Data Problems with any of these parts, or a combination.
Course: Software Engineering © Alessandra RussoUnit 1 - Introduction, slide Number 1 Unit 1: Introduction Course: C525 Software Engineering Lecturer: Alessandra.
Chapter 8: Errors, Failures, and Risk
1. 2 IMPORTANCE OF MANAGEMENT Some organizations have begun to ask their contractors to provide only project managers who have been certified as professionals.
1 Can We Trust the Computer? What Can Go Wrong? Case Study: The Therac-25 Increasing Reliability and Safety Perspectives on Failures, Dependence, Risk,
CS 235: User Interface Design August 25 Class Meeting Department of Computer Science San Jose State University Fall 2014 Instructor: Ron Mak
Slides prepared by Cyndi Chie and Sarah Frye1 A Gift of Fire Third edition Sara Baase Chapter 8: Errors, Failures, and Risks.
Security and Reliability THERAC CASE STUDY TEXTBOOK: BRINKMAN’S ETHICS IN A COMPUTING CULTURE READING: CHAPTER 5, PAGES
Principles of Information Systems, Sixth Edition Systems Design, Implementation, Maintenance, and Review Chapter 13.
Therac-25 Case Family vs. Programmer. People Suffered From Different Type of Bad Programming Database accuracy problems. Many people could not vote in.
Dimitrios Christias Robert Lyon Andreas Petrou Dimitrios Christias Robert Lyon Andreas Petrou.
CS 4001Mary Jean Harrold 1 Can We Trust the Computer?
©2001 Southern Illinois University, Edwardsville All rights reserved. Today Fun with Icons Thursday Presentation Lottery Q & A on Final Exam Course Evaluations.
Chapter 20 A Safe and Healthy Environment. Lecture Overview Employee Safety Principles of Safety Program Implementation of Safety Program Health Work.
A Gift of Fire, 2edChapter 4: Can We Trust the Computer?1 Can We Trust the Computer?
Systems Analysis and Design in a Changing World, Fourth Edition
Where We Are Now 14–2. Where We Are Now 14–2 Major Tasks of Project Closure Evaluate if the project delivered the expected benefits to all stakeholders.
Therac-25 CS4001 Kristin Marsicano. Therac-25 Overview  What was the Therac-25?  How did it relate to previous models? In what ways was it similar/different?
Principles of Information Systems, Sixth Edition 1 Systems Design, Implementation, Maintenance, and Review Chapter 13.
Presented to: By: Date: Federal Aviation Administration AIRWORTHINESS Positive Safety Culture Failure to Follow Procedures 1 R1.
©2001 Southern Illinois University, Edwardsville All rights reserved. Today Finish Ethics Next Week Research Topics in HCI CS 321 Human-Computer Interaction.
Why Cryptosystems Fail R. Anderson, Proceedings of the 1st ACM Conference on Computer and Communications Security, 1993 Reviewed by Yunkyu Sung
 Individuals with Disabilities Act  A law created to help protect the education rights of students with disabilities.
Directed Reading 1 Girish Ramesh – Andres Martin-Lopez – Bamdad Dashtban –
Randy Modowski Adam Reimel Max Varner COSC 380 May 23, 2011 Accountability.
Chapter 8: Errors, Failures, and Risk Zach Archer Daniel O’Hara Eric Strittmatter.
CHAPTER 9: PROFESSIONAL ETHICS AND RESPONSIBILITIES BY: MATT JENNINGS SHANE CRAKER KYLER RHOADES.
1 Advanced Computer Programming Project Management: Basics Copyright © Texas Education Agency, 2013.
Can We Trust the Computer? FIRE, Chapter 4. What Can Go Wrong? What are the risks and reasons for computer failures? How much risk must or should we accept?
EE 585 : FAULT TOLERANT COMPUTING SYSTEMS B.RAM MOHAN
Therac-25 Accidents What was Therac-25? Who developed it?
A Gift of Fire Third edition Sara Baase
PowerPoint® Slides to Accompany
Reliability and Safety
Week 13: Errors, Failures, and Risks
Computer in Safety-Critical Systems
A Gift of Fire Third edition Sara Baase
Presentation transcript:

IT Safety and Reliability Professor Matt Thatcher

2 Agenda for Today l Brief review of the Case of the Killer Robot l Overview of the: –Therac-25 accidents –Denver International Airport Baggage System l Discussion of “How Good Is Good Enough?” –what are our social responsibilities?

3 Killer Robot Summary l The general problems –simple programming error –inadequate safety engineering and testing –poor HCI design –lax culture of safety in Silicon Techtronics l What would change in you replaced one of the characters with an “ethical” person? –would any of these problems have been solved?

4 Matt’s Humble Opinions l Source of the problems –economic incentives »time pressures u exclusive focus on meeting unrealistic deadlines u there was no payoff to the dvmt team based on usability or safety measures u valuing stock price over operator safety »cut corners  keep your job, challenge decisions  get fired –company culture »poor communication all along the company hierarchy »lots of unproductive, unresolved, and unaddressed conflict »inability to consider alternatives »choice of waterfall model instead of prototyping model as development methodology –inexperience and lack of critical skills in key positions »Johnson (hrdwr guy), Reynolds (data processing guy), Samuels (no exp in physics) l Who is most responsible for: –setting appropriate economic incentives –creating an appropriate culture –putting people with the rights skills into the right jobs

5 Therac-25 Accidents (Basis of the Killer Robot Case) l What was Therac-25? –released in 1983 –computerized radiation therapy machine used to treat cancer patients l Who developed it? –Atomic Energy of Canada, Ltd and GCR (French-based company) l What were the key advances of it over its predecessors (Therac-6 and Therac-20)? –move to more complete software-based control »faster set-up »safety checks were now controlled by software (instead of mechanical interlocks)

6 Therac-25 Accidents (What Happened?) l Massively overdosed patients at least 6 times (3 died, 3 seriously disabled) –June 1985 »Marietta, Ga (Linda Knight, 61) –July 1985 »Hamilton, Ont (Donna Gartner, 40) –December 1985 »Yakima, Wash (Janis Tilman) –March 1986 »Tyler, Tx (Isaac Dahl, 33) –April 1986 »Tyler, Tx (Daniel McCarthy) –January 1987 »Yakima, Wash (Anders Engman)

7 Therac-25 Accidents (Example of Contributing UI Problems) l The technician got the patient set up on the table, and went down the hall to start the treatment. She sat down at the terminal: »hit “x” to start the process u she realized she made a mistake, since she needed to treat the patient with the electron beam, not the X-ray beam »hit the “Up” arrow, »selected the “Edit” command, »hit “e” for electron beam, and »hit “enter” (signifying she was ready to start treatment) »the system showed a “beam ready” prompt »she hit “b” to turn the beam therapy on »the system gave her an error message (Malfunction 54) »she overrode the error message l It turns out that the UI showed that it was in electron mode but it was actually in a “hybrid” mode  delivered more than 125 times the normal dose to the patient

8 Therac-25 Accidents (What Were the Problems?) l The Problems –simple programming errors –inadequate safety engineering »ignored the software risks (almost no unit or integration testing at all) »operators were told it was impossible to overdose a patient –poor HCI design –lax culture of safety in the manufacturing co. –problems were not reported quickly to manufacturer or FDA »prompted a 1990 federal law

9 Friendly-Fire Tragedy (Afghanistan)

10 Denver International Airport’s Baggage System

11 Other System Failures l Problems for individuals –billing inaccuracies (electricity, auto insurance) –database inaccuracies (sex crimes DB, 2000 Florida election) l System failures –AT&T (1990) –Galaxy 4 satellite (1998) –London Stock Exchange (2000) –Warehouse Manager –New York School District

12 Increasing Reliability and Safety l Overconfidence l Software reuse l Professional techniques l UI an human factors l Redundancy and self-checking l Warranties l Regulation of safety applications l Self-regulation l Take responsibility!!!

13 Critical Observation l If it is true that: –information technology affects society AND –some choices in computing design are not completely constrained by mathematics, physics, chemistry, etc. THEN –designers, implementers, teachers, and managers of technology make choices that affect society

14 Principle Actors in Software Design l Software Provider –person or organization that creates the software l Software Buyer –person or organization responsible for obtaining the software for its intended use l Software User –person actually using the software l Society –people, other than providers, buyer, or users who can be affected by the software

15 Obligations of the Software Provider l The Provider (itself) –profit and good reputation l The Buyer –help buyer make an informed decision –set testing goals and meet them –provide warnings about untested areas of the software –inform user about testing results and shortcomings –provide a reasonable warranty on functionality and safety l The User –education/training about use and limitations of software –provide technical support and maintenance –provide reasonable protections –provide clear instructions and user manuals l Biggest Responsibility –answer the question “How Good is Good Enough”

16 Obligations of the Software Buyer l The Provider –respect copyrights and don’t steal –pay a fair price for the product –use the product for the correct purpose –don’t blame the provider for incorrect use –make users available for observations, testing, and evaluation l The Buyer (itself) –learn limitations of the software –perform systems acceptance testing (and audit testing) l The User –ensure proper education/training is provided –ensure proper technical support and maintenance is provided –represent the user during development (be a communication link between user and developer) –make sure users are included in the design process –provide reasonable protections (and a safe working environment) l Biggest Responsibility –make sure software is used for intended purpose (think about biometrics)

17 Obligations of the Software User l The Provider –respect copyrights and don’t steal –read documentation l The Buyer –communicate problems and provide feedback (about training, UI, functionality, etc.) –ensure appropriate use of software –make a good faith effort to learn the software and its limitations l The User (herself) –help in the training process l Biggest Responsibility –ensure that the software continues to perform as intended and report problems

18 Final Observation l Software is treated like other commercials products l Software differs in many critical respects from many other products –serious software errors can remain after rigorous testing because of the logical complexity of software –it is difficult to construct uniform software standards that can be subjected to regulation and inspection –software affects an increasingly large number of people due to the proliferation and flexibility of computers –any group can provide software since set-up costs are low

19 Summary l Computer professionals must stop thinking of themselves as technicians l Just like medical doctors, computer professionals are agents of change in people’s lives l Computer professionals have some ethical responsibility for the changes they are creating in society l But this responsibility is complicated because our position as workers in industry diffuses accountability l We must recognize that part of our job includes larger issues of social responsibility