Current:Home > NewsWhat happens when our Tesla Model Y's cameras can't see? Nothing good. -LegacyBuild Academy
What happens when our Tesla Model Y's cameras can't see? Nothing good.
View
Date:2025-04-13 13:10:43
Tesla — its CEO in particular — is obsessed with cameras. Elon Musk is reported by the New York Times to have told his employees that if people can drive with just eyes, cars should, too. Unfortunately, it just doesn’t work this way, and as with eyeballs, when the cameras can’t see, the tech that relies on them regresses or stops working altogether. Losing, even temporarily, the most valuable features of our 2023 Tesla Model Y Long Range is understandably frustrating and makes us question their worth and the decisions that got us here.
All the hardware you need?
At one point, Tesla advertised all new models as being equipped with all the hardware they’d ever need to achieve real autonomous driving, or “Full Self-Driving” in Tesla’s parlance. Actually getting there, the company claimed, was simply a matter of sufficiently training the software. The company has since backed away from this messaging.
Around the same time, Tesla was busy deactivating and even removing hardware from its vehicles. Forward-facing radar was deactivated in existing cars, and installation stopped on new cars. Later, ultrasonic parking sensors were removed. In both cases, Tesla and Musk argued those sensors were unnecessary thanks to software advancements, and that cameras were all that are needed.
Every other company we’ve spoken to in the autonomous vehicle field disagrees. Outside of Tesla, it’s generally accepted that multiple different types of sensors, each with its own strengths and weaknesses, that provide overlapping and redundant information are necessary to achieve true autonomous driving. A single sensor type, the thinking goes, can too easily break, get obscured, or miss crucial information, and leave the whole system temporarily blinded.
Nature proves the point
The limitations of Tesla’s approach and the wisdom of its autonomous driving competitors has been made apparent to us on multiple occasions — most often, first thing in the morning.
Where we live, overnight and morning fog, sometimes devolving to a heavy mist, is common from late fall to early spring. It tends to condense on car windows, which can be cleared with defrosters, wipers, or by rolling them down. It also tends to collect on camera lenses and their protective covers. When this happens, the only way to clear the cameras on our Model Y is to walk around the car and wipe them off by hand. Forget to do this before you get in the car and attempt to drive, and you’ll quickly discover Park Assist, Tesla’s name for its camera-based parking sensors, is “degraded” and not trustworthy due to the cameras being obscured.
This is a somewhat minor annoyance but knowing that ultrasonic parking sensors suffer no such limitation (though they have their own issues) and that they used to be installed on these cars is frustrating.
It’s not just manual parking, either. Tesla’s self-parking and “Summon” technologies also rely on the cameras, so if you’re hoping to have the car pull itself out of its parking space or even drive to you, you’re going to be disappointed.
More from Tesla:Tesla brings back cheap Model 3 variant with big-time range
It’s also indicative of the larger problem at hand: Tesla hasn’t engineered any way for the car to clean its cameras when they’re misty or dirty, so it’s entirely on the driver. Instead, the company has tried to place its cameras where they’re unlikely to get obscured, but once it happens, there’s nothing the car can do but figuratively throw its hands up. Aside from certain off-roaders, most other companies don’t routinely install camera washers on their cars, either, but those companies also don’t rely exclusively on cameras for their cars to operate as intended.
Obscured cameras are one thing when backing out of the driveway; they’re another in foul weather. Be it heavy fog or rain, the cameras’ inability to see quickly shuts down the “Full Self-Driving (Supervised)” system, and with it Autosteer lane centering and Autopilot adaptive cruise control, in the name of safety. All it takes to completely neutralize the car’s flagship $15,000 (at the time of our purchase) tech feature is a little rain. We’re happy the vehicle is erring on the side of caution, but it’s frustrating to know the steps that could’ve been taken to address these issues have been ruled out.
Here again, we must note that all autonomous-aspiring vehicles struggle with the weather to some degree. Teslas, however, are uniquely susceptible to the problem due to their total reliance on cameras. Competitors address the issue with other sensors, like radar, which isn’t affected by fog. With recent advancements, lidar, which Musk has continually talked down, has also been demonstrated to see better in fog than cameras. Though icing is still possible with radar and lidar, heating elements have already been incorporated into vehicles using those systems to melt snow and ice and keep them unobstructed.
Why won’t Tesla use more sensors?
Ignore Musk’s eyes-to-cameras analogy; scientists who study the human eye and visual cortex say cameras and computers don’t work the same way at all. From the way they collect light to the way they process it, there’s little comparison between the two beyond the obvious.
The real reason, per Musk, is cost. He’s long rejected lidar as being too expensive. It’s not just lidar, though. Removing radar (a single module of which can cost three times as much as a camera) and ultrasonic sensors from the parts list saves Tesla money in materials cost, R&D cost, and assembly cost. Not having to design and manufacture those parts, program them, teach the computer to fuse data from multiple sensor types and interpret it, and add time and complexity on the assembly line potentially saves the company a significant amount of money. If cameras could actually do the same things as those sensors, it would be a brilliant business move. Tesla’s real-world experiment, though, hasn’t turned out that way.
Tesla agrees
Recent reports show Tesla’s attitude may be changing. Numerous court cases and exposés have shown Tesla engineers do not agree with the claims being made in the company’s marketing. When pressed, they’ve had to walk back company claims about the capabilities of their cameras and software, and have even admitted to faking promotional videos overselling the system’s functionality. In the most recent court case, Tesla’s lawyers even argued before a judge that lidar is necessary to achieve autonomous driving, and the current cars can’t actually drive themselves and won’t be able to without it.
Filings with government regulators, meanwhile, show Tesla has applied for licenses to put radar back in its cars. (Radar uses radio frequencies and is therefore regulated by the Federal Communications Commission.) The company has also recently purchased several hundred lidar units, likely for R&D purposes, which may indicate it has changed its position on that technology as well.
How much is a Tesla Cybertruck?Here's a pricing breakdown for the EV pickup
Where does that leave us?
In all likelihood, we’re stuck with what we have. Tesla will no doubt continue to work on its software and send more OTA (over-the-air) updates intended to improve our car's functionality. Although the company does give owners of certain older models the option of upgrading their car’s computer for a price, it does not retrofit hardware like radar units or cameras. That’s just the risk of buying a Tesla, though — everything from the price to the hardware to the features might change the next day.
For more on our long-Term 2023 Tesla Model Y Long Range:
- We Bought A 2023 Tesla Model Y Long Range For A Yearlong Test
- The Supercharger Difference
- How Far Can You Tow With A Tesla?
- What Changed After The Tesla Autosteer Recall? Not Much.
- Are The Tesla Model Y’s Third-Row Seats Worth It?
Photos by MotorTrend
veryGood! (83243)
Related
- Jamie Foxx gets stitches after a glass is thrown at him during dinner in Beverly Hills
- First victim of Tulsa Race Massacre identified through DNA as WWI veteran
- Arizona abortion initiative backers sue to remove ‘unborn human being’ from voter pamphlet language
- 2024 MLB mock draft: Latest projections for every Round 1 pick
- See you latte: Starbucks plans to cut 30% of its menu
- Georgia state tax collections finish more than $2 billion ahead of projections, buoying surplus
- Alec Baldwin 'Rust' case dismissed by judge over 'suppressed' evidence
- U.S. says it will deploy more long-range missiles in Germany, Russia vows a military response
- Off the Grid: Sally breaks down USA TODAY's daily crossword puzzle, Hi Hi!
- Chicago exhibition center modifying windows to prevent bird strikes after massive kill last year
Ranking
- Elon Musk's skyrocketing net worth: He's the first person with over $400 billion
- Jayden Daniels hopes to win, shift culture with Washington Commanders
- Euphoria Season 3 Finally Has a Start Date
- Pregnant Margot Robbie and Husband Tom Ackerley Pack on the PDA at Wimbledon 2024
- See you latte: Starbucks plans to cut 30% of its menu
- Inside Billionaire Heir Anant Ambani and Radhika Merchant's Wedding of the Year in India
- Lakers vs. Rockets live updates: Watch Bronny James in summer league game today
- Baltimore Judge Tosses Climate Case, Hands Win to Big Oil
Recommendation
'As foretold in the prophecy': Elon Musk and internet react as Tesla stock hits $420 all
Late-night comics have long been relentless in skewering Donald Trump. Now it’s Joe Biden’s turn
Houston community groups strain to keep feeding and cooling a city battered by repeat storms
Alec Baldwin’s Rust Involuntary Manslaughter Trial Takes a Sudden Twist
At site of suspected mass killings, Syrians recall horrors, hope for answers
Actor Matthew McConaughey tells governors he is still mulling future run for political office
Map shows all the stores slated to be sold in Kroger-Albertsons merger
Mississippi must move quickly on a court-ordered redistricting, say voting rights attorneys