Who's Liable For Decisions AI and Robotics Make? (betanews.com) 180
An anonymous reader shares a BetaNews article: Reuters news agency reported on February 16 that "European lawmakers called [...] for EU-wide legislation to regulate the rise of robots, including an ethical framework for their development and deployment and the establishment of liability for the actions of robots including self-driving cars." The question of determining "liability" for decision making achieved by robots or artificial intelligence is an interesting and important subject as the implementation of this technology increases in industry, and starts to more directly impact our day to day lives. Indeed, as application of Artificial Intelligence and machine learning technology grows, we are likely to witness how it changes the nature of work, businesses, industries and society. And yet, although it has the power to disrupt and drive greater efficiencies, AI has its obstacles: the issue of "who is liable when something goes awry" being one of them. Like many protagonists in industry, Members of the European Parliament (MEPs) are trying to tackle this liability question. Many of them are calling for new laws on artificial intelligence and robotics to address the legal and insurance liability issues. They also want researchers to adopt some common ethical standards in order to "respect human dignity."
What a novel question (Score:3, Funny)
I don't think anyone has ever considered this issue. Ever.
Obvious Answer (Score:2)
The people who designed and tested the AI.
What is so novel about that?
Nope: The deployment decision maker (Score:2)
The elected politician or judge or senior executive who approved the use of the category of technology in the category of application in which the problem occurred is who ought to be responsible.
Software, specially self-learning AI software, is too complex and unpredictable (in details of operation in every case).
Careful programming and testing cannot cover the range of possibilities, because input data and system state are too (combinatorially) complex.
It's the senior decision maker who ways the risks and
Correction of (unpredictable) neuro-motor misfire (Score:2)
"decision maker who weighs the risks"
Re: (Score:2)
Re: (Score:2)
Re: (Score:3)
The elected politician or judge or senior executive who approved the use of the category of technology in the category of application in which the problem occurred is who ought to be responsible.
Terrible answer. Nothing will ever be approved if the approvers are afraid of being sued into personal bankruptcy.
If AI means "autonomous system": Whoever manufactures and certifies them for public use should be liable, barring specific and well-documented misuse/misconfigurartion. Let the corporations assess the risk/reward themselves.
If AI means "self-aware, intelligent system": Not a problem I expect to worry about in the foreseeable future, but when it happens the AI can be liable instead of the manufac
There's nothing harder about self-awareness (Score:2)
compared with perception and semantic understanding of general external environment.
I'm talking about functional self-awareness here (i,e, behaviour indicating self-awareness) not the consciousness hard-problem (qualia).
A "feeling" of self-awareness is not necessary for functional behaviour due self awareness. Whether a feeling would emerge is a separate question, not that important because we can't prove that we ourselves have it. We only assume that other people are not zombies because it's a simpler supp
Re: (Score:2)
Re: (Score:2)
Encode information about environment
- Into an associative memory with abstraction (including event/situation abstraction) and probabilities encoding and
- meta-knowledge tagging such as estimate of belief strength, knowledge completeness about different general and specific topics & situations.
-Have model of my generic and specific needs/wants / avoids / i.e. desired and undesired environment state/evolution patterns.
-process environment state/history/projection model (including generated counterfactual
Absolutely nuttin' and no one (Score:2)
The people who designed and tested the AI.
What is so novel about that?
Welcome to earth, but I'm sorry to report Microsoft will invade your planet RSN.
You obviously didn't read your EULA covering your new AI robot. You don't own anything and whatever goes bad, the badness is NOT the fault of "the people who designed and tested the AI". When you signed the lease or whatever to use the monster, you agreed they are all innocent.
Liability is such a quaint old idea. You wouldn't want to bankrupt Microsoft by holding the company liable for all the damages caused by their little mist
Re: (Score:2)
It doesn't matter what your EULA says about the car you purchased from Microsoft/Google/Tesla/Apple. I was not party to the agreement, and if your car hurt me or my property, I can sue whoever I want, including the car manufacturer, the dealership that sold it, and you. I can sue all three at once if I want. And there's nothing anyone can do about it, unless there is a law that restricts me from doing it (like they did with gun manufacturers).
Your EULA may have a clause where you agree to indemnify and d
Re: (Score:2)
Re: (Score:2)
Re: (Score:2)
Well it won't be the owners. They just get the profits.
Re: (Score:2)
Re: (Score:2)
Re: (Score:2)
AI is just software (Score:4, Insightful)
and I've been calling for professional licensing and liability for software engineers for at least 30 years. That should follow the approach for other Professional Engineers, including the use of 'engineering practices' as a defense.
The software community has done an appallingly shitty job with software reliability. (Exhibit 1: CERT database of software vulnerabilities.) It's way past time they get held accountable. And yeah, this will slow things down and require people do things right the first time, and it will put a serious dent in the management approach to "throw the cheapest bodies at the software problem, and damn the bugs!" Product liability needs to include both corporate and individual liability.
With a union and a real trade school system (Score:2)
With a union and a real trade school system.
A lot of CS professors have been in the ivory tower for way to long and have very little real work place know how on the workings of IT / codeing.
Re: (Score:2)
Structural engineers (my father was one) don't have a union and they don't get their degrees from trade schools.
The do have a professional society that is fully engaged in licensing, educational standards for engineering curriculum, and a career path that leads to the necessary experience to qualify for a license, along with testing. They also collaborate with the state Engineering licensing agencies.
On the other hand in computing, at least some of the professional societies have actively argued against li
Re: (Score:2)
and I've been calling for professional licensing and liability for software engineers for at least 30 years.
We already have professional licensing/certificates such as MCSE, MCP, etc. They are negatively correlated with competence.
And yeah, this will slow things down and require people do things right the first time
It will also greatly increase the cost, and be the end of free software.
Programming should not be a crime.
Re: (Score:2)
The case of a car and pedestrian is less interesting because it is obvious that the liability would probably be assigned to the car manufacturer. But what if the auto maker exercised due care in
Re: (Score:2)
Re: (Score:2)
- A hazard analysis is performed on the system by various engineers (and occasionally even a 3rd party is brought in for peer review). There are a multitude of different ways to go about it, but eventually you end up with a long list of ways the product could fail, with a
Re: (Score:2)
Re: (Score:2)
How do you come up with requirements for a hazard analysis on a heavy machine that can be anywhere in the world at any time, driving at any speed? Your set of conditions that the vehicle will encounter are almost limitless.
You can still do it. From the requirements side, define some reasonable operating conditions and the behavior if it detects itself leaving those conditions. From the safety analysis side, there are multiple methods that are usually used in concert. Generally it'll start with a top down analysis of the energy sources (fuel, kinetic energy in a big moving vehicle, batteries etc.) and work your way down to specific and reasonable failure modes. Then there are a variety of other analysis methods to supplement
Re: (Score:2)
Re: (Score:2)
Re: (Score:2)
Actually, I have developed software applying DO-178B (although not to Level B/A) for air traffic control, and to Mil-Std 882D for a project that included networked fires and autonomous potentially armed vehicles. On the latter project, I was the lead software safety person for a while.
And to the follow-on comment: Yeah, there's a lot of work to get the hazards and the requirements down to the level that verification against those has real impact. That's part of the job.
Re: (Score:2)
Re: (Score:2)
Well first, the techniques for high confidence software (such as safety-critical) show that -substantial- improvements in software reliability are possible. For commercial avionics, the verification costs are much more (maybe even an order of magnitude) than the development costs. The ROI for verification (such as MCDC) has been questioned (whether the additional surety gained is worth the cost to get it.)
What's most important is the culture of assurance, i.e. 'think before you write', at least anecdotall
Re: (Score:2)
Bullshit. Complete bullshit. If your son or daughter is run over by a robot semi on the highway, because that semi's AI told it to swerve to avoid the cow in it's path, into their lane, thus crushing them, you will have a completely different attitude. AI is software, but that software is going to be making decisions that normally would require human judgement. You can't sue the semi drive because there is no semi driver. So who do you sue?
No one, because in your world no one needs to be held liable.
So
Re: (Score:2)
Re: (Score:2)
35 years in industry, most of that in "millitary/aerospace" where the consequences are large and often fatal.
Perhaps more importantly, my father, a Structural Engineer with a PE license, and I discussed professional liability many times. He was involved in some court cases as an expert witness when a building fell down. (He said there was blame to go around. Bad design by the architect, incorrect/incomplete specification of materials by the architect and engineer, and the materials that were used failed
Who is liable when your tv catches fire (Score:5, Informative)
Robotics have been with us for more than half a century.
Re: (Score:2)
Yes but only in controlled/limited environments. They mostly haven't been out in the wild en masse, or performing tasks with nearly the complexity we are now getting them to do, or with nearly as much risk if it goes wrong. e.g. Driving our cars.
Re: (Score:3)
Re: (Score:2)
Kindof reachng there dude. A basic timer is hardly what most people mean when they think of the phrase "intelligent robot".
Re:Who is liable when your tv catches fire (Score:4, Insightful)
You're right. Ultimately, this is not a new problem. The question boils down to, "Who is responsible when a product malfunctions?"
However, there is a relevant shift in liability that needs to happen. Basically there are certain things where the manufacturer is only responsible for the product being able to operate safely, but the operator of the product is also partially responsible.
For example, Toyota may have legal liability for a manufacturing defect that causes the breaks to stop working, but Toyota isn't responsible for a car crash caused by an unsafe driver. Once you have self-driving cars, that needs to change because the "driver" cannot be held responsible. Obviously the manufacturer needs to take on greater liability, but there also may be situations where that's not really practical either. There may still be things that the car's owner or passenger could do to cause an accident. For example, if the owner modifies the car or fails to perform maintenance, and that causes the AI to malfunction, the owner should probably still be held responsible. Or there will certainly be some accidents that just happen, and aren't really anyone's fault.
And the particulars of all that need to be codified into law. We have hundreds of years of laws dealing with carriages and cars, but some of those may shift when the car is autonomous. What, exactly, is the car's manufacturer responsible for, and what is the owner responsible for? How do we determine whether an AI is adequate to make the necessary decisions, and how will inspections be carried out? These are things that need to be thought about.
Re: (Score:2)
that needs to change because the "driver" cannot be held responsible.
Your use of quotes there is interesting and important. The people inside a fully autonomous cars are passengers, not drivers. If someone is driving a car, it's not autonomous. Essentially, the manufacturer *is* the driver. Their programming, sensors, algorithms, and maps are what is used to control the vehicle, the passenger isn't involved. If I'm not in control of a vehicle, I have no intention of being liable for its actions.
if the owner modifies the car or fails to perform maintenance, and that causes the AI to malfunction, the owner should probably still be held responsible.
Another interesting choice of wording. I don't think manufacturers will sell
Re: (Score:2)
The people inside a fully autonomous cars are passengers, not drivers.
Actually I put it into quotes in that instance because I was referring to the AI as the "driver". But an AI can't be fined or arrested, so someone else will need to be held responsible.
I don't think manufacturers will sell fully autonomous cars.
I agree that fewer people will buy cars, and that it may eventually become relatively rare for an individual to buy a car for their own personal use. Still, presumably someone will own the cars, and it may not be the manufacturer. You may have services like Uber buying cars from a company like Tesla. There may be companie
Re: (Score:2)
If you own a home and a tree on your property falls on your neighbor and injures them; aren't you liable?
Yep. It's my tree, and I could have taken better care of it so that it wouldn't be prone to breaking. I also have insurance (homeowners or renters) to cover me in case of accidents.
You have friends over for a party; your gas oven blows up and kills one of your guests; aren't you liable?
I guess that would depend. Did my lack of maintenance or improper installation cause the accident, or was it a manufacturing defect? If it was a manufacturing defect that should have generated a recall, the insurance company or I will likely be going after damages from the manufacturer.
Same scenario, but you rent instead of own, then isn't the owner of the apartment liable?
If I'm a renter, then maintenance of the
Re: (Score:2)
"Lack of Consideration for Human Factors Led to In-Flight Breakup of SpaceShipTwo". NTSB. 28 July 2015. Retrieved 29 July 2015. [ntsb.gov]
civil vs criminal as well. Where things are diffen (Score:2)
civil vs criminal as well. Where things are different.
And in a criminal case they can't hide under an NDA or EULA
Simple: The owner.... (Score:2)
Primary liability for a robot's actions are with the owner... Case closed....
Now, the owner may have a liability claim with the maintainer, installer and/or manufacturer should the robot not function as designed, but that's another case.
Exactly the same one who is liable.... (Score:2)
Re: (Score:2)
Re: (Score:2)
Re: (Score:2)
Re: (Score:2)
Re: (Score:2)
Re: (Score:3)
Re: (Score:2)
...but if your tumble dryer catches fire due to a build up of (your) fluff, it's the manufacturer's fault, and not yours.
Before the design defect in the product was established, insurance companies would pay out to owners and then take it up with the manufacturers. The manufacturers presumably paid out a few times, did an investigation and realised they needed to alert end-users so the slew of claims would come to an end. Now we believe it's the manufacturer's problem and that they need to fix it.
In the cas
Whoever has the deepest pockets (Score:2)
At least, that's what the battalions of lawyers will argue.
Let the market decide (Score:2)
Every permutation of possibilities for say an automated 3 ton car driving at speed in the massive complexity of the unpredictable open world without ever having any accidents, simply cannot even be anticpated let alone exhaustively planned-for/tested, therefore cannot realistically ever be the fault of Engineers.
The only sane approach is to require full cover insurance for each robot in the wild. Let the market itself determine the actual usage of robots based on the trade-offs between total costs (includin
Re: (Score:2)
Re: (Score:2)
They have absolute control. They can choose to not buy a self-driving car.
Isaac Asimov (Score:2)
Isaac Asimov, for improperly formulating the three laws of robotics. If he'd gotten them right, none of this would be necessary.
Avoidance, Acceptance, or other? (Score:2)
Avoidance: I built my own car from scratch with nothing but hand tools! It took me 20 years!
Respecting risks: I built a robotic assembly line that produces 50 cars a day. I must observe, and be aware that this is a very dangerous piece of tech, and be sure my employees understand that they are paid handsomely to do the same while working near it.
Other: My assembly line riveted my hands to my widget! It's not my fault, I never *REALLY* accepted the risks of automation....
Avoidance: I'll never trust a self dr
What? (Score:2)
Why on earth would The Who be liable? Oh, wait, it's a question, not a statement.
Re: (Score:2)
You sir get my imaginary mod point.
Preserve WHAT (Score:2)
developers developers developers (Score:2)
Same as your pet (Score:2)
Same question as who's liable for decision pets make. When your dog escape from the house and makes a carnage in a nearby kindergarten, you are liable. If you feel you're not cut to control things you own so that it doesn't get out of control, don't buy a dog nor a robot.
Re: (Score:2)
Robots do not make decisions (Score:2)
Re: (Score:2)
Re: (Score:2)
Re: (Score:2)
Re: (Score:2)
Given identical inputs, two cars act identically. What they would do is known a priori and they are simply following a set of instructions the same as a player piano.
The piano doesn't decide on when to play certain notes, it follows a set of pre-planed instructions.
Re: (Score:2)
arg (Score:2)
Who's liable (Score:2)
Only in the case of K9
Non binding again (Score:2)
As far as I know, EU parliament has no power here. This will be one more non-binding resolution.
But at least, I assume it is better than nothing that some people work on that problem.
Liability insurance as a feature? (Score:2)
I read one proposal which suggested that liability insurance might be bundled with autonomous vehicles as a marketing tool. Or perhaps an optional feature like leather seats & a sun roof. That seems like a really good idea to me. It would certainly answer this question about who is responsible for an accident. As a selling point, it would make expensive autonomous vehicles extremely attractive to drivers considered to be "high risk" by insurance companies. For someone with multiple accidents &
Missing Option: CowboyNeal (Score:2)
CowboyNeal
Why has no one thought of the obvious and correct answer?
First you have to ask, why we don't demand (Score:2)
"As horrible as each of these stories is, there is nothing that shows that Mr. Cadden did something that the government can link to the death of that person."
There are many example of cases where a corporation kills people, but, magically, no one person is found guilty of murder, when it was clearly murder.
Oh, I guess it's nobody, because it was done in the context of a business!
We have to solve that problem first. And the question doesn't change just because you add "with robots" or "with technology".
Re:Easy, the programmer of course. (Score:5, Insightful)
Because most AI and some robots rely on techniques that create emergent behaviour (i.e. not directly programmed therefore unverifiable) such as neural nets and swarm theory.
Re: (Score:2)
Re: (Score:2)
Yes, good point.
Re: (Score:3)
>> So? I don't think any code I've created has been formally proven to be correct
You obviously haven't worked in avionics or written autopilots then. I have and this is exactly what you do. BTW Formal proof (at last the type good enough for the FAA) also has little or nothing to do with writing in assembly.
Re: (Score:2)
Programmers don't decide this, system engineers do.
Source: I'm a robotics engineer.
Re: (Score:2)
Which programmer is liable when many are working on the firmware and programs running the vehicle/robot?
How much of a code change would make me equally liable as the other programmers?
What if the code was perfect (worked as designed) but the hardware it was running on or the sensors attached to it reported incorrect information?
What if the programmers fixed a bug, that could cause an accident, but the manufacturer failed to sell vehicles whose code contained that fix?
What if the programmers fixed a bug, tha
Re: (Score:2)
A more interesting case is a collision between two moving self driving vehicles. In th
Re: (Score:2)
Re: (Score:2)
Re: (Score:2)
Re: (Score:2)
But you don't control what the AI does.
There is a perfect example of this in theaters right now. In the movie Logan, there's highways in Oklahoma that have essentially robot semi trailers operating on them all the time. Pretty cool actually. But there's a scene where some horses run loose, and the robot semis almost hit them.
Now, lets say the robots are programmed to avoid large animals in the road, but by doing so one swerves into traffic, or brakes suddenly, causing another accident in which someone is
Re: (Score:2)
Re: (Score:2)
Re: (Score:2)
if you as the driver decide to "check out" and let the AI do the driving for you, and then you get into a car crash, you should be liable because you were not paying attention or controlling your vehicle.
Except we are talking about cars that have no controls for the person inside. All the other examples you gave offer more control to the user of said product than a fully automated car will.
See a doctor (Score:2)
There's medication available for your condition.
Re: (Score:2)
Re: (Score:2)
Who is responsible for buggy software sending spam to the world?
My buggy doesn't send anything to anybody. But then, it doesn't have software; just a horse.
Re: (Score:2)
Accidents can be due to the same things that could cause a human to have an accident. Sensors are degraded. (dirty windshield wiper, etc) Road conditions are degraded and couldn't stop before the stop sign. What about a simple mechanical failure that is nobody's fault? (Even in a human driven car, my brakes didn't work!)
Re: Not so hard (Score:2)
Re: (Score:2)
Re: (Score:2)
Re: (Score:2)