Self-driving vehicles will play a crucial role in improving transportation safety and accelerating the world’s transition to a sustainable future. Full autonomy will enable a vehicle to be safer, lower the financial cost of transportation for those who own a car and provide low-cost on-demand mobility for those who do not.
As of today, all Tesla vehicles – including Model 3 – will have the hardware needed for full self-driving capability at a safety level substantially greater than that of a human driver.
Eight surround cameras provide 360 degree visibility around the car at up to 250 meters of range. Twelve updated ultrasonic sensors complement this vision, allowing for detection of both hard and soft objects at nearly twice the distance of the prior system. A forward-facing radar with enhanced processing provides additional data about the world on a redundant wavelength, capable of seeing through heavy rain, fog, dust and even the car ahead.
To make sense of all of this data, a new onboard computer with more than 40 times the computing power of the previous generation runs the new Tesla-developed neural net for vision, sonar and radar processing software. Together, this system provides a view of the world that a driver alone cannot access, seeing in every direction simultaneously and on wavelengths that go far beyond the human senses.
Before activating the features enabled by the new hardware, Tesla will further calibrate the system using millions of miles of real-world driving to ensure significant improvements to safety and convenience.
While this is occurring, Teslas with new hardware will temporarily lack certain features currently available on Teslas with first-generation Autopilot hardware, including some standard safety features such as automatic emergency braking, collision warning, lane holding and active cruise control. As these features are robustly validated Tesla will enable them over the air, together with a set of entirely new features.