- Tesla faces a "Full Self-Driving" safety probe from the U.S. federal automotive safety regulator, NHTSA.
- The probe will evaluate whether Tesla's FSD systems are safe to use in fog, glaring sun and other reduced roadway visibility conditions.
- The probe follows a series of Tesla collisions where FSD was in use within 30 seconds of the crash -- one where a Tesla driver fatally struck a pedestrian.
Tesla faces a new investigation by the National Highway Traffic Safety Administration, or NHTSA, concerning issues with its "Full Self-Driving" systems, and whether they are safe to use in fog, glaring sun or other "reduced roadway visibility conditions."
The probe follows an incident in which a Tesla driver who had been using FSD, struck and killed a pedestrian, and other FSD-involved collisions during reduced roadway visibility conditions.
Records posted to the NHTSA website on Friday morning said the purpose of the new probe would be to assess:
"The ability of FSD's engineering controls to detect and respond appropriately to reduced roadway visibility conditions; whether any other similar FSD crashes have occurred in reduced roadway visibility conditions and, if so, the contributing circumstances for those crashes," among other things.
The agency will also look into Tesla's over-the-air, software updates to its FSD systems, which are now marketed as "Full Self-Driving (Supervised)," to understand the "timing, purpose, and capabilities of any such updates, as well as Tesla's assessment of their safety impact."
Money Report
The "preliminary evaluation" by the NHTSA pertains to a vehicle population of around 2.4 million Tesla EVs on U.S. roads including: Model S and X vehicles produced from 2016 to 2024, Model 3 vehicles produced from 2017 to 2024, Model Y vehicles produced from 2020 to 2024, and Cybertruck vehicles produced this year and last, which give drivers the option to use Tesla's FSD.
FSD, which the company now refers to as a "partial driving automation system," is Tesla's paid, premium driver assistance option. But Tesla has offered it to all drivers for a monthlong free trial in the U.S., previously.
The U.S. federal vehicle safety regulator tracks collisions involving the use of automakers' advanced driver assistance systems, like Tesla's Autopilot or FSD. As of Oct. 1, 2024, the NHTSA had tracked 1,399 incidents in which Tesla's driver assistance systems were engaged within 30 seconds of the collision, and 31 of those had resulted in fatalities.
Tesla did not immediately respond to a request for comment.
The company recently held a marketing event in which CEO Elon Musk said Tesla expects to have "unsupervised FSD" up and running in Texas and California next year in the company's Model 3 and Model Y electric vehicles.
Musk has promised driverless vehicles for years. But Tesla has not yet produced or shown a vehicle that is safe to use on public roads without a human at the wheel, ready to steer or brake at any time.