The Times Australia
The Times World News

.
The Times Real Estate

.

Why the feds are investigating Tesla's Autopilot and what that means for the future of self-driving cars

  • Written by Hayder Radha, Professor of Electrical and Computer Engineering, Michigan State University

It’s hard to miss the flashing lights of fire engines, ambulances and police cars ahead of you as you’re driving down the road. But in at least 11 cases in the past three and a half years, Tesla’s Autopilot advanced driver-assistance system did just that. This led to 11 accidents in which Teslas crashed into emergency vehicles or other vehicles at those scenes, resulting in 17 injuries and one death[1].

The National Highway Transportation Safety Administration has launched an investigation[2] into Tesla’s Autopilot system in response to the crashes. The incidents took place between January 2018 and July 2021 in Arizona, California, Connecticut, Florida[3], Indiana, Massachusetts, Michigan, North Carolina and Texas. The probe covers 765,000 Tesla cars[4] – that’s virtually every car the company has made in the last seven years. It’s also not the first time[5] the federal government has investigated Tesla’s Autopilot.

As a researcher who studies autonomous vehicles[6], I believe the investigation will put pressure on Tesla to reevaluate the technologies the company uses in Autopilot and could influence the future of driver-assistance systems and autonomous vehicles.

How Tesla’s Autopilot works

Tesla’s Autopilot[7] uses cameras, radar and ultrasonic sensors to support two major features: Traffic-Aware Cruise Control and Autosteer.

Traffic-Aware Cruise Control, also known as adaptive cruise control, maintains a safe distance between the car and other vehicles that are driving ahead of it. This technology primarily uses cameras in conjunction with artificial intelligence algorithms to detect surrounding objects such as vehicles, pedestrians and cyclists, and estimate their distances. Autosteer uses cameras to detect clearly marked lines on the road to keep the vehicle within its lane.

In addition to its Autopilot capabilities, Tesla has been offering what it calls “full self-driving” features that include autopark[8] and auto lane change[9]. Since its first offering of the Autopilot system and other self-driving features, Tesla has consistently warned users that these technologies require active driver supervision and that these features do not make the vehicle autonomous.

Why the feds are investigating Tesla's Autopilot and what that means for the future of self-driving cars Tesla’s Autopilot display shows the driver where the car thinks it is in relation to the road and other vehicles. Rosenfeld Media/Flickr, CC BY[10][11]

Tesla is beefing up the AI technology that underpins Autopilot. The company announced on Aug. 19, 2021, that it is building a supercomputer using custom chips[12]. The supercomputer will help train Tesla’s AI system to recognize objects seen in video feeds collected by cameras in the company’s cars.

Autopilot does not equal autonomous

Advanced driver-assistance systems have been supported on a wide range of vehicles for many decades. The Society of Automobile Engineers divides the degree of a vehicle’s automation into six levels[13], starting from Level 0, with no automated driving features, to Level 5, which represents full autonomous driving with no need for human intervention.

Within these six levels of autonomy, there is a clear and vivid divide between Level 2 and Level 3. In principle, at Levels 0, 1 and 2, the vehicle should be primarily controlled by a human driver, with some assistance from driver-assistance systems. At Levels 3, 4 and 5, the vehicle’s AI components and related driver-assistance technologies are the primary controller of the vehicle. For example, Waymo’s self-driving taxis[14], which operate in the Phoenix area, are Level 4, which means they operate without human drivers but only under certain weather and traffic conditions.

News coverage of a Tesla driving in Autopilot mode that crashed into the back of a stationary police car.

Tesla Autopilot is considered a Level 2 system, and hence the primary controller of the vehicle should be a human driver. This provides a partial explanation for the incidents cited by the federal investigation. Though Tesla says it expects drivers to be alert at all times when using the Autopilot features, some drivers treat the Autopilot as having autonomous driving capability with little or no need for human monitoring or intervention. This discrepancy between Tesla’s instructions and driver behavior[15] seems to be a factor in the incidents under investigation.

Another possible factor is how Tesla assures that drivers are paying attention. Earlier versions of Tesla’s Autopilot were ineffective in monitoring driver attention[16] and engagement level when the system is on. The company instead relied on requiring drivers to periodically move the steering wheel, which can be done without watching the road. Tesla recently announced that it has begun using internal cameras to monitor drivers’ attention[17] and alert drivers when they are inattentive.

Another equally important factor contributing to Tesla’s vehicle crashes is the company’s choice of sensor technologies. Tesla has consistently avoided the use of lidar[18]. In simple terms, lidar is like radar[19] but with lasers instead of radio waves. It’s capable of precisely detecting objects and estimating their distances. Virtually all major companies working on autonomous vehicles, including Waymo, Cruise, Volvo, Mercedes, Ford and GM, are using lidar as an essential technology for enabling automated vehicles to perceive their environments.

By relying on cameras, Tesla’s Autopilot is prone to potential failures caused by challenging lighting conditions, such as glare and darkness. In its announcement of the Tesla investigation, the NHTSA reported that most incidents occurred after dark where there were flashing emergency vehicle lights, flares or other lights. Lidar, in contrast, can operate under any lighting conditions and can “see” in the dark.

Fallout from the investigation

The preliminary evaluation will determine whether the NHTSA should proceed with an engineering analysis, which could lead to a recall. The investigation could eventually lead to changes in future Tesla Autopilot and its other self-driving system. The investigation might also indirectly have a broader impact on the deployment of future autonomous vehicles; in particular, the investigation may reinforce the need for lidar.

Although reports in May 2021 indicated that Tesla was testing lidar sensors[20], it’s not clear whether the company was quietly considering the technology or using it to validate their existing sensor systems. Tesla CEO Elon Musk called lidar “a fool’s errand[21]” in 2019, saying it’s expensive and unnecessary.

However, just as Tesla is revisiting systems that monitor driver attention, the NHTSA investigation could push the company to consider adding lidar or similar technologies to future vehicles.

[You’re smart and curious about the world. So are The Conversation’s authors and editors. You can get our highlights each weekend[22].]

References

  1. ^ 11 accidents in which Teslas crashed into emergency vehicles or other vehicles at those scenes, resulting in 17 injuries and one death (static.nhtsa.gov)
  2. ^ launched an investigation (www.caranddriver.com)
  3. ^ Florida (www.nytimes.com)
  4. ^ covers 765,000 Tesla cars (www.motortrend.com)
  5. ^ not the first time (www.theverge.com)
  6. ^ researcher who studies autonomous vehicles (scholar.google.com)
  7. ^ Tesla’s Autopilot (www.tesla.com)
  8. ^ autopark (www.youtube.com)
  9. ^ auto lane change (www.youtube.com)
  10. ^ Rosenfeld Media/Flickr (flickr.com)
  11. ^ CC BY (creativecommons.org)
  12. ^ building a supercomputer using custom chips (www.cnbc.com)
  13. ^ six levels (www.sae.org)
  14. ^ self-driving taxis (theconversation.com)
  15. ^ driver behavior (doi.org)
  16. ^ were ineffective in monitoring driver attention (www.wsj.com)
  17. ^ internal cameras to monitor drivers’ attention (www.cnbc.com)
  18. ^ avoided the use of lidar (venturebeat.com)
  19. ^ lidar is like radar (www.autoweek.com)
  20. ^ Tesla was testing lidar sensors (www.bloomberg.com)
  21. ^ a fool’s errand (techcrunch.com)
  22. ^ You can get our highlights each weekend (theconversation.com)

Read more https://theconversation.com/why-the-feds-are-investigating-teslas-autopilot-and-what-that-means-for-the-future-of-self-driving-cars-166307

The Times Features

Why Staying Safe at Home Is Easier Than You Think

Staying safe at home doesn’t have to be a daunting task. Many people think creating a secure living space is expensive or time-consuming, but that’s far from the truth. By focu...

Lauren’s Journey to a Healthier Life: How Being a Busy Mum and Supportive Wife Helped Her To Lose 51kg with The Lady Shake

For Lauren, the road to better health began with a small and simple but significant decision. As a busy wife and mother, she noticed her husband skipping breakfast and decided ...

How to Manage Debt During Retirement in Australia: Best Practices for Minimising Interest Payments

Managing debt during retirement is a critical step towards ensuring financial stability and peace of mind. Retirees in Australia face unique challenges, such as fixed income st...

hMPV may be spreading in China. Here’s what to know about this virus – and why it’s not cause for alarm

Five years on from the first news of COVID, recent reports[1] of an obscure respiratory virus in China may understandably raise concerns. Chinese authorities first issued warn...

Black Rock is a popular beachside suburb

Black Rock is indeed a popular beachside suburb, located in the southeastern suburbs of Melbourne, Victoria, Australia. It’s known for its stunning beaches, particularly Half M...

What factors affect whether or not a person is approved for a property loan

Several factors determine whether a person is approved for a real estate loan. These factors help lenders assess the borrower’s ability to repay the loan and the risk involved...

Times Magazine

Lessons from the Past: Historical Maritime Disasters and Their Influence on Modern Safety Regulations

Maritime history is filled with tales of bravery, innovation, and, unfortunately, tragedy. These historical disasters serve as stark reminders of the challenges posed by the seas and have driven significant advancements in maritime safety regulat...

What workers really think about workplace AI assistants

Imagine starting your workday with an AI assistant that not only helps you write emails[1] but also tracks your productivity[2], suggests breathing exercises[3], monitors your mood and stress levels[4] and summarises meetings[5]. This is not a f...

Aussies, Clear Out Old Phones –Turn Them into Cash Now!

Still, holding onto that old phone in your drawer? You’re not alone. Upgrading to the latest iPhone is exciting, but figuring out what to do with the old one can be a hassle. The good news? Your old iPhone isn’t just sitting there it’s potential ca...

Rain or Shine: Why Promotional Umbrellas Are a Must-Have for Aussie Brands

In Australia, where the weather can swing from scorching sun to sudden downpours, promotional umbrellas are more than just handy—they’re marketing gold. We specialise in providing wholesale custom umbrellas that combine function with branding power. ...

Why Should WACE Students Get a Tutor?

The Western Australian Certificate of Education (WACE) is completed by thousands of students in West Australia every year. Each year, the pressure increases for students to perform. Student anxiety is at an all time high so students are seeking suppo...

What Are the Risks of Hiring a Private Investigator

I’m a private investigator based in Melbourne, Australia. Being a Melbourne Pi always brings interesting clients throughout Melbourne. Many of these clients always ask me what the risks are of hiring a private investigator.  Legal Risks One of the ...

LayBy Shopping