• vaguerant@fedia.io
    link
    fedilink
    arrow-up
    0
    ·
    1 month ago

    The big headline is understandably that it crashes into a fake painted wall like a cartoon, but that’s not something that most drivers are likely to encounter on the road. The other two comparisons where lidar succeeded and cameras failed were the fog and rain tests, where the Tesla ran over a mannequin that was concealed from standard optical cameras by extreme weather conditions. Human eyes are obviously susceptible to the same conditions, but if the option is there, why not do better than human eyes?

    • Echo Dot@feddit.uk
      link
      fedilink
      English
      arrow-up
      1
      ·
      1 month ago

      Human eyes are much better at picking up shapes from complex visual clutter than computers. Not to mention that human eyes have infinitely better dynamic range than cameras (every time you go into or out of a tunnel the camera is essentially blind for a couple of seconds until it readjusts), both mean that while you absolutely do need the cameras, you can’t just rely on them and say “oh well humans only use their eyes”, because the cameras and the computers are nowhere near human level capable.

    • tias@discuss.tchncs.de
      link
      fedilink
      English
      arrow-up
      1
      ·
      1 month ago

      Human eyes are way better at this than any camera based self driving system. No self driving system is anywhere close to driving in Swedish winter with bad weather and no sun, yet us Swedes do it routinely by the millions every day.