20°C / 22°C
  • Tue
  • 26°C
  • 14°C
  • Wed
  • 23°C
  • 10°C
  • Thu
  • 20°C
  • 11°C
  • Fri
  • 22°C
  • 10°C
  • Sat
  • 20°C
  • 12°C
  • Sun
  • 21°C
  • 13°C
  • Tue
  • 14°C
  • 10°C
  • Wed
  • 17°C
  • 9°C
  • Thu
  • 23°C
  • 10°C
  • Fri
  • 18°C
  • 13°C
  • Sat
  • 22°C
  • 12°C
  • Sun
  • 28°C
  • 12°C
  • Tue
  • 27°C
  • 16°C
  • Wed
  • 24°C
  • 13°C
  • Thu
  • 22°C
  • 12°C
  • Fri
  • 25°C
  • 11°C
  • Sat
  • 25°C
  • 14°C
  • Sun
  • 22°C
  • 15°C
  • Tue
  • 26°C
  • 14°C
  • Wed
  • 27°C
  • 8°C
  • Thu
  • 24°C
  • 13°C
  • Fri
  • 25°C
  • 12°C
  • Sat
  • 21°C
  • 12°C
  • Sun
  • 23°C
  • 14°C
  • Tue
  • 19°C
  • 15°C
  • Wed
  • 20°C
  • 14°C
  • Thu
  • 19°C
  • 15°C
  • Fri
  • 24°C
  • 17°C
  • Sat
  • 21°C
  • 19°C
  • Sun
  • 23°C
  • 18°C
  • Tue
  • 16°C
  • 10°C
  • Wed
  • 15°C
  • 12°C
  • Thu
  • 18°C
  • 12°C
  • Fri
  • 18°C
  • 13°C
  • Sat
  • 17°C
  • 14°C
  • Sun
  • 18°C
  • 12°C
  • Tue
  • 13°C
  • 6°C
  • Wed
  • 20°C
  • 5°C
  • Thu
  • 25°C
  • 10°C
  • Fri
  • 21°C
  • 12°C
  • Sat
  • 25°C
  • 9°C
  • Sun
  • 31°C
  • 13°C
  • Tue
  • 13°C
  • 9°C
  • Wed
  • 17°C
  • 11°C
  • Thu
  • 20°C
  • 11°C
  • Fri
  • 16°C
  • 12°C
  • Sat
  • 21°C
  • 11°C
  • Sun
  • 27°C
  • 13°C
  • Tue
  • 31°C
  • 17°C
  • Wed
  • 27°C
  • 13°C
  • Thu
  • 25°C
  • 15°C
  • Fri
  • 26°C
  • 14°C
  • Sat
  • 29°C
  • 16°C
  • Sun
  • 26°C
  • 16°C
  • Tue
  • 19°C
  • 10°C
  • Wed
  • 23°C
  • 4°C
  • Thu
  • 28°C
  • 7°C
  • Fri
  • 23°C
  • 14°C
  • Sat
  • 27°C
  • 11°C
  • Sun
  • 20°C
  • 14°C
  • Tue
  • 26°C
  • 15°C
  • Wed
  • 17°C
  • 12°C
  • Thu
  • 15°C
  • 12°C
  • Fri
  • 24°C
  • 13°C
  • Sat
  • 22°C
  • 11°C
  • Sun
  • 24°C
  • 15°C
  • Tue
  • 16°C
  • 9°C
  • Wed
  • 14°C
  • 9°C
  • Thu
  • 19°C
  • 9°C
  • Fri
  • 17°C
  • 13°C
  • Sat
  • 17°C
  • 12°C
  • Sun
  • 20°C
  • 11°C

How do you teach a driverless car to drive?

How does a driverless car learn how to drive, especially when something goes wrong?

FILE: Pilot models of the Uber self-driving car is displayed at the Uber Advanced Technologies Center in Pittsburgh, Pennsylvania. Picture: AFP.

CANTERBURY - Learning how to drive is an ongoing process for we humans as we adapt to new situations, new road rules and new technology, and learn the lessons from when things go wrong.

But how does a driverless car learn how to drive, especially when something goes wrong?

That’s the question being asked of Uber after last month’s crash in Arizona. Two of its engineers were inside when one of its autonomous vehicles spun 180 degrees and flipped onto its side.

Uber pulled its test fleet off the road pending police enquiries, and a few days later the vehicles were back on the road.

Smack, spin, flip

The Tempe Police Department’s report on the investigation into the crash, obtained by the EE Times, details what happened.

The report says that the Uber Volvo (red in the graphic below) was moving south at 61kmh in a 64kmh zone when it collided with the Honda (blue in the graphic) turning west into a side street (point 1).

Knocked off course, the Uber Volvo hit the traffic light at the corner (point 2) and then spun and flipped, damaging two other vehicles (points 3 and 4) before sliding to a stop on its side (point 5).

Uber crash - subsequent collisions.Alex Hanlon / Sean Welsh based on Tempe Police report

Thankfully, no one was hurt. The police determined that the Honda driver “failed to yield” (give way) and issued a ticket. The Uber car was not at fault.

Questions, Questions

But Mike Demler, an analyst with the Linley Group technology consultancy, told the EE Times that the Uber car could have done better:

It is totally careless and stupid to proceed at 38mph through a blind intersection.

Demler said that Uber needs to explain why its vehicle proceeded through the intersection at just under the speed limit when it could “see” that traffic had come to a stop in the middle and leftmost lanes.

The EE Times report said that Uber had “fallen silent” on the incident. But as Uber uses “deep learning” to control its autonomous cars, it’s not clear that Uber could answer Demler’s query even if it wanted to.

In deep learning, the actual code that would make the decision not to slow down would be a complex state in a neural network, not a line of code prescribing a simple rule like “if vision is obstructed at intersection, slow down”.

Debugging deep learning

The case raises a deep technical issue. How do you debug an autonomous vehicle control system that is based on deep learning? How do you reduce the risk of autonomous cars getting smashed and flipped when humans driving alongside them make bad judgements?

Demler’s point is that the Uber car had not “learned” to slow down as a prudent precautionary measure at an intersection with obstructed lines of sight. Most human drivers would naturally beware and slow down when approaching an intersection with obstructed vision due to stationary cars.

When it comes to deep reinforcement learning, this relies on “value functions” to evaluate states that result from the application of policies.

A value function is a number that evaluates a state. In chess, a strong opening move by white such as pawn e7 to e5 attracts a high value. A weak opening such as pawn a2 to a3 attracts a low one.

The value function can be like “ouch” for computers. Reinforcement learning gets its name from positive and negative reinforcement in psychology.

Until the Uber vehicle hits something and the value function of the deep learning records the digital equivalent of “following that policy led to a bad state - on side, smashed up and facing wrong way - ouch!” the Uber control system might not quantify the risk appropriately.

Having now hit something it will, hopefully, have learned its lesson at the school of hard knocks. In future, Uber cars should do better at similar intersections with similar traffic conditions.

Debugging formal logic

An alternative to deep learning is autonomous vehicles using explicitly stated rules expressed in formal logic.

This is being developed by nuTonomy, which is running an autonomous taxi pilot in cooperation with authorities in Singapore.

NuTonomy’s approach to controlling autonomous vehicles is based on a rules hierarchy. Top priority goes to rules such as “don’t hit pedestrians”, followed by “don’t hit other vehicles” and “don’t hit objects”.

Rules such as “maintain speed when safe” and “don’t cross the centreline” get a lower priority, while rules such as “give a comfortable ride” are the first to be broken when an emergency arises.

While NuTonomy does use machine learning for many things, it does not use it for normative control: deciding what a car ought to do.

In October last year, a NuTonomy test vehicle accident was involved in an accident: a low-speed tap resulting in a dent, not a spin and flip.

The company’s chief operating officer Doug Parker told IEEE Spectrum:

What you want is to be able to go back and say, “Did our car do the right thing in that situation, and if it didn’t, why
didn’t it make the right decision?” With formal logic, it’s very easy.

Key advantages of formal logic are provable correctness and relative ease of debugging. Debugging machine learning is trickier. On the other hand, with machine learning, you do not need to code complex hierarchies of rules.

Time will tell which is the better approach to driving lessons for driverless cars. For now, both systems still have much to learn.

This article was republished courtesy of the World Economic Forum.
_
Written by Sean Welsh, Doctoral Candidate in Robot Ethics at the University of Canterbury._

Comments

EWN welcomes all comments that are constructive, contribute to discussions in a meaningful manner and take stories forward.

However, we will NOT condone the following:

- Racism (including offensive comments based on ethnicity and nationality)
- Sexism
- Homophobia
- Religious intolerance
- Cyber bullying
- Hate speech
- Derogatory language
- Comments inciting violence.

We ask that your comments remain relevant to the articles they appear on and do not include general banter or conversation as this dilutes the effectiveness of the comments section.

We strive to make the EWN community a safe and welcoming space for all.

EWN reserves the right to: 1) remove any comments that do not follow the above guidelines; and, 2) ban users who repeatedly infringe the rules.

Should you find any comments upsetting or offensive you can also flag them and we will assess it against our guidelines.

EWN is constantly reviewing its comments policy in order to create an environment conducive to constructive conversations.

comments powered by Disqus