Should driverless cars be programmed to kill you?

  • Topic Author
  • User
  • User
More
23 Jun 2016 19:38 #246160 by
I found this interesting and thought it might spur some interesting discussion.

Current reports suggest 10 million driverless cars could be on the road by 2020. The issue is this: In a few very rare scenarios, a driverless car may have to make a choice between protecting its occupants and protecting pedestrians. For example, if it's driving down a road at speed and someone runs out into the road, should it swerve into other traffic to avoid them, potentially injuring or killing the driver and passengers? Or should it make every attempt to stop, even though it knows it won’t be able to, killing the pedestrian?


http://www.iflscience.com/technology/the-fatal-moral-dilemma-posed-by-driverless-cars/

Please Log in to join the conversation.

More
23 Jun 2016 20:15 - 23 Jun 2016 20:15 #246162 by Carlos.Martinez3
Give me a 69 all metal no computer Mustang any day... I'll pass on the smart cars... phones didn't take the smart word too well ... let's see how cars like the title. Lol I'll wait this one out even if I have to ride the trolley! Jk

Chaplain of the Temple of the Jedi Order
Build, not tear down.
Nosce te ipsum / Cerca trova
Last edit: 23 Jun 2016 20:15 by Carlos.Martinez3.
The following user(s) said Thank You: Wescli Wardest, OB1Shinobi

Please Log in to join the conversation.

  • User
  • User
More
23 Jun 2016 20:27 #246164 by
I don't like the idea of self driving cars. I'm a bit of a control freak and I don't like the idea of a program being responsible for my transportation and my safety. My phone/computer/playstation/all other pieces of technology that I own have problems way too often for me to trust them with my life. Plus, I genuinely love driving. It's so much fun. There can be self driving cars, I just don't want to be in one.

As for the morality issue...that's tough. I suppose that the people in the car would usually have a better chance of surviving the resulting crash than the pedestrians have of surviving being hit, all situation based of course. So it might be "better" to have the car attempt to miss pedestrians. That might change depending on speed and surroundings. Not an easy question.

Please Log in to join the conversation.

More
23 Jun 2016 20:29 #246165 by Adder
Hehe, nah. I'd worry we'd have idiots walking out into traffic knowing they would be ok, not caring about the impact it has on others!!!

Plus I think it would be a bad idea to cross into oncoming traffic, while obviously dramatically increasing the collision force there is also that approaching vehicles would be coming from further away and therefore less likely to be informing whatever sensors the car has about the size of the oncoming traffic. Bumping into an oncoming car at slow speed might be better for everyone under ideal conditions, but what if the occupant in the other car is already carrying an unrelated injury or what if the oncoming car is a petrol tanker. They are not doing anything 'wrong' and so I think its unfair to include them in the accident by choice.... unless the vehicle suffers a mechanical fault of its own, as then it might have better odds for everyone hitting another vehicle then an unprotected person or crowd.

So to you Q, the pedestrian is doing the wrong thing, so I think the car should aim for wherever the biggest gap is, brake, slide whatever, and if it is faced with going one way into oncoming traffic or the other way into the pedestrian - I'd program it the later because the pedestrian should not be there and perhaps the sensor is incorrect in its sensing (is it really a person!?). BUT would smart cars be talking to each other, in which case we'd have a greater margin for safety by all trying to make the biggest gap for the incident.

Introverted extropian, mechatronic neurothealogizing, technogaian buddhist.
Likes integration, visualization, elucidation and transformation.
Jou ~ Deg ~ Vlo ~ Sem ~ Mod ~ Med ~ Dis
TM: Grand Master Mark Anjuu
The following user(s) said Thank You: , OB1Shinobi, MadHatter

Please Log in to join the conversation.

More
23 Jun 2016 20:32 #246166 by Edan

Kyrin Wyldstar wrote: I found this interesting and thought it might spur some interesting discussion.

Current reports suggest 10 million driverless cars could be on the road by 2020. The issue is this: In a few very rare scenarios, a driverless car may have to make a choice between protecting its occupants and protecting pedestrians. For example, if it's driving down a road at speed and someone runs out into the road, should it swerve into other traffic to avoid them, potentially injuring or killing the driver and passengers? Or should it make every attempt to stop, even though it knows it won’t be able to, killing the pedestrian?


http://www.iflscience.com/technology/the-fatal-moral-dilemma-posed-by-driverless-cars/


I don't like the idea that technology is going to be making moral decisions for us, and I like driving, so I won't be owning a driverless car until the day the government forces me to. A car can't tell the difference between people, e.g. a mother and baby versus an elderly person. As horrible as it may be, in a split second my brain is probably going to go mother and baby. A car can't make that kind of decision.

"Evil is always possible. And goodness is eternally difficult."
The following user(s) said Thank You: Garm, , OB1Shinobi, MadHatter

Please Log in to join the conversation.

More
23 Jun 2016 20:36 #246167 by Adder

Goken wrote: I don't like the idea of self driving cars. I'm a bit of a control freak and I don't like the idea of a program being responsible for my transportation and my safety.


It's the next step to getting hover cars :lol:

Can you imagine the carnage with manually flown flying cars OMG :side:

Introverted extropian, mechatronic neurothealogizing, technogaian buddhist.
Likes integration, visualization, elucidation and transformation.
Jou ~ Deg ~ Vlo ~ Sem ~ Mod ~ Med ~ Dis
TM: Grand Master Mark Anjuu
The following user(s) said Thank You: Carlos.Martinez3,

Please Log in to join the conversation.

More
23 Jun 2016 20:50 - 23 Jun 2016 20:54 #246170 by MadHatter
That is a situation Ive always disliked. Tech messes up far far too much for me to trust it with my life or the lives of others. Imagine a chemical tankers sensor messing up and causing it to crash to save the "most" amount of lives but there was no actual danger to anything and it crashes leaking toxic chemicals for nothing. Besides I will not put my life into the hands of something that holds no ability for compassion or reason. In short Ill walk everywhere before I trust a machine that might kill me because of a pre-programmed algorithm. Frankly I would hold the programmers of the cars guilty for every death the car causes and hold no issue with sending them to jail for it.

Knight of the Order
Training Master: Jestor
Apprentices: Lama Su, Leah
Just a pop culture Jedi doing what I can
Last edit: 23 Jun 2016 20:54 by MadHatter.
The following user(s) said Thank You: Garm, Edan, Carlos.Martinez3, , OB1Shinobi

Please Log in to join the conversation.

More
23 Jun 2016 20:52 #246172 by Carlos.Martinez3
Greater responsibility no accountability... smh I don't think we are ready for it... human beings. We still can't decide whether or not where to pee.. my own opinion though... we focus on way to many.. frolicking like ideas some times. The road and behind the wheel... a place to relax and let us do the driving... hmm....still thinking...

Chaplain of the Temple of the Jedi Order
Build, not tear down.
Nosce te ipsum / Cerca trova
The following user(s) said Thank You: Edan,

Please Log in to join the conversation.

  • User
  • User
More
23 Jun 2016 21:01 #246174 by
Further issue. These cars would all have to be on some kind of network, I'm assuming. It seems like the kind of thing where they could be accessed remotely. What if some crazy hacker decided to mess with the programming? Heck, the idea of someone remotely controlling my vehicle at all is too much for me. They're already out there. Cars with stuff like OnStar where they can remotely shut the car down if it's reported stolen.

Please Log in to join the conversation.

  • User
  • User
More
23 Jun 2016 21:11 #246176 by
:laugh:

I just re-read my last post and now I feel like some crazy conspiracy theorist. I'm not saying I'm wrong, I just think I sound a bit nuts! :laugh:

Please Log in to join the conversation.

  • User
  • User
More
23 Jun 2016 21:43 #246184 by

Adder wrote:

Goken wrote: I don't like the idea of self driving cars. I'm a bit of a control freak and I don't like the idea of a program being responsible for my transportation and my safety.


It's the next step to getting hover cars :lol:

Can you imagine the carnage with manually flown flying cars OMG :side:


If driverless cars will get us closer to hover cars, then count me in. I always have been an off-road kinda guy.

Please Log in to join the conversation.

  • User
  • User
More
23 Jun 2016 22:24 #246187 by
The piece missing from the equation here is the role that the actual roads could also play in this. Driverless cars don't have to be unleashed on every side street, alley, backroad, turnpike and freeway. I imagine that sensors could be installed over highways or lanes meant specifically for driverless cars similar to the ones used on toll roads now that read transmitters present in the vehicle. The vehicle can be programmed to only allow driverless operation when on certain roads or in certain lanes where these transmitters are present, but still be manually operated everywhere else. If we keep the driverless operation to roads with little or no pedestrian traffic it would greatly mitigate the risk of these moral dilemmas.

Since these cars sense each other and set proper speeds and distances based on that information, we'd basically be turning the 'driverless car lane' into a train made of individual vehicles.

I see my morning commute as manually driving my car onto the interstate, entering the proper lane, then putting it on 'autopilot' and reading a book or drinking my tea. When it's time to get off the interstate, I take over again. It could be a glorified version of cruise control.

It would be especially helpful if it tied into the GPS navigation and warned you when your exit was coming so I could wake up from a nap and put it back into manual mode in a timely manner. :laugh:

Please Log in to join the conversation.

More
24 Jun 2016 02:10 - 24 Jun 2016 02:15 #246197 by Adder
Sorry for the derail, but on flying cars I found an electric helo with good engine redundancy, simple controls lets driver focus on nav and traffic avoidance;

https://www.youtube.com/watch?v=OazFiIhwAEs

If they could fold that rotor assembly they could be parked on rooftops next to each other under a landing/launch platform, for commercial buildings.
B) :lol:

Introverted extropian, mechatronic neurothealogizing, technogaian buddhist.
Likes integration, visualization, elucidation and transformation.
Jou ~ Deg ~ Vlo ~ Sem ~ Mod ~ Med ~ Dis
TM: Grand Master Mark Anjuu
Last edit: 24 Jun 2016 02:15 by Adder.
The following user(s) said Thank You: OB1Shinobi

Please Log in to join the conversation.

  • User
  • User
More
24 Jun 2016 03:36 #246198 by
While you can argue the fallacies of tech....I have seen more than enough fallacy in humans driving. Drunk driving, road rage, texting while driving, simply not paying attention, ,people who are way to old to have a license,etc,etc.

I can't imagine it would be much worse actually.

I always wonder if people simply don't like change much when these things come about.

http://www.youtube.com/watch?v=XU3qfDtWFmk

I have seen what the Tesla cars do with such beginnings of technology a d I have to say, I see little to be so negative about.

Please Log in to join the conversation.

  • User
  • User
More
24 Jun 2016 03:38 #246199 by
http://www.youtube.com/watch?v=2iwfNs8j2FI

Please Log in to join the conversation.

  • RyuJin
  • Offline
  • Master
  • Master
  • Council Member
  • Council Member
  • Ordained Clergy Person
  • Ordained Clergy Person
  • The Path of Ignorance is Paved with Fear
More
24 Jun 2016 04:05 - 24 Jun 2016 04:07 #246204 by RyuJin
i didn't read all the posts...but driverless cars come with a major risk...hackers can access them....it was already proven with a new jeep grand cherokee, a hacker accessed the jeep's computer through it's onstar thingy or whatever chrysler uses, and the hacker was able to control the radio/ac/electronics and was even able to shut the vehicle down while it was being driven...the driver had no control over the vehicle once it was shut down...

just something to think about...

edit: this was done with knowledge as a means of proving that it was possible...the driver and hacker were demonstrating the vulnerability "smart" vehicles have...

Warning: Spoiler!

Quotes:
Warning: Spoiler!

J.L.Lawson,Master Knight, M.div, Eastern Studies S.I.G. Advisor (Formerly Known as the Buddhist Rite)
Former Masters: GM Kana Seiko Haruki , Br.John
Current Apprentices: Baru
Former Apprentices:Adhara(knight), Zenchi (knight)
Last edit: 24 Jun 2016 04:07 by RyuJin.
The following user(s) said Thank You: , OB1Shinobi

Please Log in to join the conversation.

More
24 Jun 2016 14:09 - 24 Jun 2016 15:08 #246268 by OB1Shinobi
to answer the question posed by OP

imo, the car should be programmed to protect the driver - it should be a personal safety vehicle foremost (a whole new angle on the idea of "product loyalty" lol; in future, the products are loyal to YOU!), especially in a circumstance where an outside agent was violating procedure in a way that put lives at risk

as for self driving cars in general

i do not want to own a vehicle if the control of that vehicle can be forcibly taken away from me

for one thing, i imagine a good way to end the revolution would be to simply lock suspicious people in their own cars and remote-drive them to the internment camp

other than the issue of control being taken against our own will or interests, i think theyre a great idea

once the parameters have been set, computers are more capable and less fallible at acting within those parameters than humans are

People are complicated.
Last edit: 24 Jun 2016 15:08 by OB1Shinobi.
The following user(s) said Thank You: ,

Please Log in to join the conversation.

  • Topic Author
  • User
  • User
More
24 Jun 2016 14:46 #246270 by
LOL I have to say this went in some interesting directions! While the question I proposed was designed to be a moral one many went off in a terminator/mr-robot/blade runner direction. Thanks for the responses.

As for the question I would tend to agree with what seems to be the consensus here. If a driverless car is faced with a moral decision such as this it should protect the driver first and any outside obstacles secondary. So that means if a pedestrian stepped in front of a car and the car could not avoid the collision it should not be programmed to swerve into traffic or off a cliff etc to protect the pedestrian.

I think in order to clearly assess this situation we need to put ourselves in the role of not only the driver but the pedestrian as well. If you made a mistake and stepped in front of a moving vehicle would you want that vehicle to be programmed to swerve and kill its occupants to protect you? Now I'm sure some would say YES! But objectively speaking if you made that mistake you should take responsibility for that mistake and accept the consequences vs letting the car occupants pay for it.

So in the end I don't think cars should be programmed to swerve to avoid pedestrians or other such potential situations, (say a pregnant woman etc) even if there is a likely hood that more could be killed outside the car vs inside.

Please Log in to join the conversation.

  • User
  • User
More
24 Jun 2016 19:30 #246360 by
I think ultimately, the car will make the best possible choice based on the variables, and given its ability to think not only objectively, but faster than a human, it will be more than likely better than a persons choice with slower reflexes and OODA loop.

It is more than likely the person in the car will be safer regardless of choice, given airbags, seat belts, and whatever technology comes up with in the meantime.

Something like that foam from "Demolition Man", which had the option of self drive and not actually.

It will decide whether to swerve, or not, or what have you, based on the a lot of high level math in a few seconds to make the best possible outcome.

You see, the car will not make a "moral decision" at all.

Please Log in to join the conversation.

  • User
  • User
More
24 Jun 2016 19:37 #246361 by
This clickbait-headline-as-thread-title makes it sound more like the question is about whether driverless cars will be programmed to immediately kill their inhabitant, as if we live in some bizarre dystopia where that is the sensible solution for population control. I realize it's taken from the linked clickbait article, but c'mon, surely you could have done better than that, given the question is really far more nuanced than that.

Please Log in to join the conversation.

Moderators: MorkanoWrenPhoenixThe CoyoteRiniTaviKhwang