• @kingofras@lemmy.world
    link
    fedilink
    English
    13025 days ago

    The funny thing is that the YT comment section was almost immediately brigaded by Elmo suckups that complained the tester was using Autopilot and not the latest version of FSD, completely forgoing the point that the LiDar vehicle stopped despite any automation systems being turned on.

    This was a test of the emergency braking systems of both cars in as close to real world scenarios as possible. Whether you run over a child in heavy rain or dense fog should not be dependent on whether the driver remembered to turn certain safety systems on prior to departure.

    It’s a massive fail for Tesla, and it is the closest scientific proof that the Elmo is a rich daddy boy with shit for brains. Once you have that level of money and leverage, you simply can make the most stupid decisions ever, and keep falling upwards. Meanwhile real people are dying almost weekly in his vehicles, and reporters and news organisations are scared shitless to report on them because he can cause lots of pain with his imaginary cash. Not to mention no govt agency will investigate properly because he’s got root access to that too. All those deaths are entirely preventable.

    Keep burning them, keep protesting at the dealerships. This truly could be the first billionaire we bring down. Watch him activate the OTA kill switch in every Tesla before they go under too. He’s thát kind of baby. Ironically that would save a lot of lives.

    • billwashere
      link
      fedilink
      English
      2124 days ago

      Lidar is clearly superior and Musk removed it because it was more expensive. The obvious conclusion is any self driving car should be made to use lidar and not simple vision based on cameras. Well at least not solely vision. Musk’s sycophants are just going to keep providing excuses and “whatabout’isms”. But there really is no doubt.

  • vaguerant
    link
    fedilink
    8225 days ago

    The big headline is understandably that it crashes into a fake painted wall like a cartoon, but that’s not something that most drivers are likely to encounter on the road. The other two comparisons where lidar succeeded and cameras failed were the fog and rain tests, where the Tesla ran over a mannequin that was concealed from standard optical cameras by extreme weather conditions. Human eyes are obviously susceptible to the same conditions, but if the option is there, why not do better than human eyes?

    • @tias@discuss.tchncs.de
      link
      fedilink
      English
      4825 days ago

      Human eyes are way better at this than any camera based self driving system. No self driving system is anywhere close to driving in Swedish winter with bad weather and no sun, yet us Swedes do it routinely by the millions every day.

    • @wewbull@feddit.uk
      link
      fedilink
      English
      524 days ago

      Whilst I agree on the wall, fog and rain are not extreme weather conditions. I’d rather he’d used a level of rain that was less intense. However, the fact is lidar still worked even though it was not clear it was going to.

  • @lumpybag@reddthat.com
    link
    fedilink
    English
    2924 days ago

    The worst part about all of this, is Tesla has backed themselves into a corner with their aggressive marketing. If additional sensors are required it will cost them billions in retrofitting all the teslas it’s already sold. Partially believe that’s why Elon bought Donald.

    • @NotMyOldRedditName@lemmy.world
      link
      fedilink
      English
      10
      edit-2
      24 days ago

      It’s already going to cost them what I imagine is billions in upgrades.

      On the last earnings call Elon finally admitted that HW3 will probably need to be upgraded for people who bought/buy FSD.

      That’s millions of cars, and there’s been some rulings that the upgrade includes people who decided to subscribe to fsd instead of purchase.

      Lawsuits aside from it never being complete, the moment whatever their latest tech is that can, even if it is somehow all cameras, is going to lead to millions of cars needing upgrades.

      However, if they truly solve it, that’s probably a drop in the bucket compared to the profits itd generate even if they have to add lidar or refund thousands of dollars per vehicle.

    • @riodoro1@lemmy.world
      link
      fedilink
      English
      224 days ago

      You still think that musk’s company will deliver what musk promised. How the fuck is that possible?

    • @Etterra@discuss.online
      link
      fedilink
      English
      224 days ago

      I’d say that the quest party is that Mark Rober said he’ll probably buy another Tesla.

      Let that sink in a minute.

  • lurch (he/him)
    link
    fedilink
    English
    1225 days ago

    the crash test in question starts at 15:25 in the yt video embedded in the article. the wall is some kind of styrofoam construction

    • Alex
      link
      fedilink
      English
      1025 days ago

      Pre-cut with comedy edges as well. But the methodology seemed valid as another poster said, giving cars the same limited eyesight as us seems like under engineering for safety.

  • @Darkassassin07@lemmy.ca
    link
    fedilink
    English
    2
    edit-2
    24 days ago

    As much as I like Mark, He’s got some explaining to do.

    At 15:42 the center console is shown, and autopilot is disengaged before impact. It was also engaged at 39mph during the youtube cut, and he struck the wall at 42mph. (ie the car accelerated into the wall)

    Mark then posted the ‘raw footage’ on twitter. This also shows autopilot disengage before impact, but shows it was engaged at 42mph. This was a seprate take.

    /edit;

    Youtube, the first frames showing Autopilot being enabled: 39mph

    Twitter, the first frames showing autopilot being enabled: 42mph

    • nickwitha_k (he/him)
      link
      fedilink
      English
      824 days ago

      Mark then posted the ‘raw footage’ on twitter. This also shows autopilot disengage before impact, but shows it was engaged at 42mph. This was a seprate take.

      No. That’s by design. The “autopilot” is made to disengage when any likely collision is about to occur to try to reduce the likelihood of someone finding them liable for their system being unsafe.

      • @wewbull@feddit.uk
        link
        fedilink
        English
        424 days ago

        Not saying you’re wrong (because I’ve always found it suspicious how Tesla always seems to report that autopilot is disengaged for fatal accidents) but there’s probably some people asking themselves “how could it detect the wall to disengage itself?”.

        The image on the wall has a perspective baked into it so it will look right from a certain position. A distance from which the lines of the real road match perfectly with the lines of the road on the wall. As you get closer than this distance the illusion will start to break down. The object tracking software will say “There are things moving in ways I can’t predict. Something is wrong here. I give up. Hand control to driver”.

        Autopilot disengaged.

        (And it only noticed a fraction of a second before hitting it, yet Mark is very conscious of it. He’s screaming. )

        Sidenote: the same is true as you move further from the wall than the ideal distance. The illusion will break down in that way too. However, the effect is far more subtle when you’re too far away. After all, the wall is just a tiny bit of your view when you’re a long way away, but it’s your whole view when you’re just about to hit it.