Please create an account to participate in the Slashdot moderation system

 



Forgot your password?
typodupeerror
×
AI Transportation

GM's Cruise so Far: A Crash, and 60 RoboTaxis 'Disabled' After Losing Server Contact (thedrive.com) 146

On June 2nd California approved General Motors' Cruise robotaxi service. The Drive describes an accident that happened the next day: The autonomous car made an unprotected left turn and was hit by a Toyota Prius on June 3, though the accident wasn't reported until Wednesday. When reached for comment by The Drive, the San Francisco Police Department explained that the Cruise vehicle had three passengers, all in the backseat, while the Prius had two occupants in total.... According to the incident report Cruise filed with the California DMV, the Cruise taxi was making a green light left turn from Geary Boulevard onto Spruce Street in downtown San Francisco. It began the turn and stopped in the middle of the intersection, presumably noticing the Toyota headed for it. The Prius then hit the right rear of the Chevy Bolt.

Cruise explained that afterward, "occupants of both vehicles received medical treatment for allegedly minor injuries." GM's incident report points out the Prius was speeding at the time of the accident, and was in the right turn lane before heading straight and hitting the Bolt. SFPD told The Drive that "no arrest or citation was issued at the time of the initial investigation," which is still ongoing. The National Highway Traffic Safety Administration has opened up a special crash investigation into the accident, but there are no public results yet.

Wired reports: In response to that crash, Cruise temporarily reprogrammed its vehicles to make fewer unprotected left turns, according to internal messages seen by WIRED. At an internal meeting Jeff Bleich, Cruise's chief legal officer, said the company was investigating the incident, according to a recording reviewed by WIRED. He also warned employees not working on that investigation to try and tune out crashes or related news reports, saying they were unavoidable and would increase in frequency as the company scaled up its operations. "We just have to understand that at some point this is now going to be a part of the work that we do, and that means staying focused on the work ahead," he said.
Wikipedia's entry for Cruise notes a few other incidents: In April 2022, the San Francisco Police Department stopped an empty (operating without any human safety attendants) Cruise AV for driving at night without its headlights on.... Also in April 2022, an empty Cruise AV blocked the path of a San Francisco Fire Department truck responding to a fire.
But Wired also reports on a more troubling incident that happened "around midnight" on June 28th: Internal messages seen by WIRED show that nearly 60 vehicles were disabled across the city over a 90-minute period after they lost touch with a Cruise server. As many as 20 cars, some of them halted in crosswalks, created a jam in the city's downtown in an incident first reported by the San Francisco Examiner and detailed in photos posted to Reddit....

The June 28 outage wasn't Cruise's first. On the evening of May 18, the company lost touch with its entire fleet for 20 minutes as its cars sat stopped in the street, according to internal documentation viewed by WIRED. Company staff were unable to see where the vehicles were located or communicate with riders inside. Worst of all, the company was unable to access its system which allows remote operators to safely steer stopped vehicles to the side of the road.

A letter sent anonymously by a Cruise employee to the California Public Utilities Commission that month, which was reviewed by WIRED, alleged that the company loses contact with its driverless vehicles "with regularity," blocking traffic and potentially hindering emergency vehicles. The vehicles can sometimes only be recovered by tow truck, the letter said. Images and video posted on social media in May and June show Cruise vehicles stopped in San Francisco traffic lanes seemingly inexplicably, as the city's pedestrians and motorists navigate around them.

This discussion has been archived. No new comments can be posted.

GM's Cruise so Far: A Crash, and 60 RoboTaxis 'Disabled' After Losing Server Contact

Comments Filter:
  • by msauve ( 701917 ) on Saturday July 09, 2022 @11:42AM (#62687904)
    >The autonomous car...

    If they stop dead when loosing connection to a server, they're not autonomous, are they?
    • by Chas ( 5144 )

      Can we make a perfect autonomous driving car?

      "Sure!"

      Are humans involved in creating it at any level?

      "Uh. Yeah?"

      Are humans infallible?

      "Uh...."

      Basically trying to build such a system is always going to be throwing dice.

      Granted, we have some groups that know how to throw the dice quite well. But even they fuck up now and again.

      I'd rather trust my own driving skills than hope to God a piece of fallible TECH isn't going to go ape-shit.

      • Re: (Score:2, Interesting)

        I'd rather trust my own driving skills than hope to God a piece of fallible TECH isn't going to go ape-shit.

        Unfortunately, your own driving skills are of limited use against somebody else's dangerous experiments. We no longer have a choice when it comes to sharing the road with Autonomous Vehicles. Within our lifetimes we may also lose the right to actually drive a motor vehicle; we'll get to choose between between being chauffeured by an algorithm, and walking. I was going to add bicycling at the end there, but with AV's everywhere? Fuck no!

        • If there's enough incidents attributable to this highly flawed technology they'll revoke their right to public roads. They can't force people into SDCs, either, there'd be a revolt, and also it'd take a generation or two to do that, no one who currently drives is going to allow themselves to be forced to buy an SDC, or pay for SDC rideshare, or be forced to walk everywhere. Never happen. Personally I think the massive flaws in this technology, which they have no capability to correct, will be the end of it
          • If there's enough incidents attributable to this highly flawed technology they'll revoke their right to public roads.

            Great! When are they going to do the same for the humans who get into far more accidents?

            What's "Enough"? Apparently Florida alone has like 6 fatalities a day in motor vehicle accidents.

            • I, personally, would rather see reforms in driver education, training, testing, and enforcement on chronically bad drivers, than I would see everyone forced into boxes on wheels with no controls for the human occupant, controlled by some rickety machine that has no actual abiliity to reason and therefore will (not may, but will) fuck up at the wrong moment and get people killed, while they scream in abject terror knowing they have no way of saving themselves from certain death.
              The problem as I see it is we
      • I'd rather trust my own driving skills

        The problem is I don't trust you, and you (if you had any sense at all) would not trust me. The difference between you and a computer is that you make mistakes while computers follow their programming. Imagine if every crash investigation we could push an update to every car to address the specific circumstances of a crash, which is to say if you have an accident with someone else, I become a better driver. Crashes would be eliminated.

        Your post is also inconsistent. You complain about humans, but in your la

        • I think that at this point, the autonomous systems aren't much better than a human driver, but the stupid things they do are different. The human in this case was stupid for going straight from the right turn lane. The computer was stupid for stopping in the middle of a left hand turn in the intersection, right in front of oncoming traffic. The human would have never expected that (and hit the car), and the computer didn't expect what he did. As a driver, I tend to think that everyone else is trying to run
          • Re: (Score:2, Insightful)

            by Chas ( 5144 )

            No, the point is that autonomous drivers AREN'T BETTER than human drivers.

            Humans have had hundreds of thousands of years to build up the coordination and skill to use their bodies to drive a car.
            We've had over a century of cars.
            And we have enough nanny systems already in place to make us more responsible drivers.

            And people want to toss it out because "Look! We ran it a hundred times on an empty track and it didn't hit anything!"

            And humans can make decisions in situations outside their experience.
            Autonomous

            • To be clear, current autonomous driver systems are not better than humans who haven't been drinking. Waymo likes to compare their cars to humans including drunk drivers, to make themselves look better.

              • Oddly enough drunk divers are the ones I think would be forced into autonomous cars first.

                • I don't care about that. I just want Waymo and Tesla to start comparing like vs like instead of using their data as propaganda.

                  • To be fair, despite all our efforts we haven't been able to remove human drunk drivers from our roads, so I don't think it counts as propaganda to include them in the set of "human drivers".

                    Just like it wouldn't be fair to remove the bottom 20% of human drivers from the stats, even though it'd make human drivers look a lot better.

                    • We can remove all drunk drivers from our roads right now by requiring a breathalyzer test to start a car.

                      Saying "an autonomous vehicle drives as well as humans" and including drunk drivers in that statistic without mentioning it is misleading. Given Google's propensity to publish statistics as propaganda, it is most likely intentionally misleading.

            • No, people want to toss it out because over a million people die in car crashes every year.

          • by ceoyoyo ( 59147 )

            Clearly neither of us knows what the exact situation was, but as described, stopping might have been a good decision.

            If the driver was in the right hand turning lane and went straight unexpectedly you have two choices: 1) stop your left hand turn and hopefully let them pass in front of you 2) hit the accelerator and try and squeal through your turn so they pass behind you.

            My high school driving instructor certainly would have recommended the first option.

        • by Chas ( 5144 )

          You can SUE ME!

          Good luck trying to successfully sue the maker of your autonomous driver.

          • This would be the main benefit to letting an AV drive me around, the limited liability. Of course, I clearly expect the AV maker to have a clause that somehow puts all the liability on me, the passenger, and not on them.

            Shame we don't have a government for the people. If the AV is suppose to be a better driver, I shouldn't need insurance (not driving personally) and if the car gets in an accident, the AV maker should take on all the liability since it was driving.

            • by Chas ( 5144 )

              So you want to be held accountable for a failure in technology.

              Okaaay.

      • The fact that it has to rely on 'training data' and can't actually THINK should tell you all you need to know. If that isn't enough then the fact that all of these have to pull over and stop and 'phone home' to get a remote human operator to get them out of some situation their 'training data' doesn't cover should be the nail in the coffin of the whole thing. But then there's the fact that one little sticker or a bit of graffitti on a STOP sign is enough for it to not recognize it as a STOP sign anymore, wh
    • moving car needs network? and can die in unsafe space.
      Just wait for one to stop on
      railroad tracks
      drawbridge
      death valley
      big uphill / downhill
      blind curve
      one lane road
      etc..

    • Exactly. It's all smoke and mirrors.
  • by Ecuador ( 740021 ) on Saturday July 09, 2022 @12:03PM (#62687938) Homepage

    The crash seems fine, I mean unprotected left turns leads to crashes all the time with real drivers. In fact they would only stop happening if you didn't have any human drivers at all, so don't expect that to happen soon.
    However, the vehicles just stopping in their tracks if they lose contact to a server? WTF design is that? Are they autonomous or not? If they are, surely they can safely find a spot on the right to stop and not block traffic? Why do they need remote access to drive them to such a spot? Loss of communication is the most common thing, even if GM were more competent (they don't seem to be at all), you can't avoid it 100% of the time, especially when it comes to wireless as we all know.

  • A Cruise autonomous vehicle ("Cruise AV") operating in driverless autonomous mode, was traveling eastbound on Geary Boulevard
    toward the intersection with Spruce Street. As it approached the intersection, the Cruise AV entered the left hand turn lane, turned the left
    turn signal on, and initiated a left turn on a green light onto Spruce Street. At the same time, a Toyota Prius traveling westbound in the
    rightmost bus and turn lane of Geary Boulevard approached the intersection in the right turn lane. The Toyota Prius was traveling
    approximately 40 mph in a 25 mph speed zone. The Cruise AV came to a stop before fully completing its turn onto Spruce Street due to the
    oncoming Toyota Prius, and the Toyota Prius entered the intersection traveling straight from the turn lane instead of turning. Shortly
    thereafter, the Toyota Prius made contact with the rear passenger side of the Cruise AV. The impact caused damage to the right rear door,
    panel, and wheel of the Cruise AV. Police and Emergency Medical Services were called to the scene, and a police report was filed. The
    Cruise AV was towed from the scene. Occupants of both vehicles received medical treatment for allegedly minor injuries.

    https://www.dmv.ca.gov/portal/file/cruise_060322-pdf/ [ca.gov]

  • My first thought was about what would have happened had these pseudo-autonomous vehicles been on the road here yesterday when Rogers cellular and internet service disappeared. All those cars just stopped dead wherever they happened to be, waiting for the mother-ship connection to be restored.

    Somebody needs to acquaint GM engineers with the concept of "fail-safe" - it seems they've never heard of it.

    • A car that needs contact with the server to operate is not autonomous.
    • by PPH ( 736903 )

      Thank goodness we are not migrating to something like 5G.

    • My first thought was about what would have happened had these pseudo-autonomous vehicles been on the road here yesterday when Rogers cellular and internet service disappeared.

      Much hilarity, my friend. Much hilarity would have ensued.

  • The fix is in (Score:4, Interesting)

    by Miles_O'Toole ( 5152533 ) on Saturday July 09, 2022 @12:52PM (#62688042)

    There's too much money to be made by making paid drivers unnecessary. Autonomous vehicles are coming, like it or not. Any inconvenient safety problems will be dealt with by having tame judges and legislators define them out of existence.

    Hang onto your hats, it's going to be quite a ride.

    • I hope so. I hope judges and legislators ban human drivers who go straight from a turning lane hitting other cars (who or whatever may be driving them).

      Humans suck at driving. The yearly deathtoll is clear evidence that we shouldn't be driving cars.

  • If a car is in the right turn lane, I would probable try to make an unprotected left in front of them as well...

    Hiowever I don't think I would have got in an accident, as (A) if I saw a car going 40MPH in a right turn lane I would assume they are not turning. And also I would not have stopped if I did go, I would go and kind of floor int assuming the car might not turn but only if I felt sure I would have plenty of time to make it if I had the case.

    Honestly though a lot of drivers would not have considered

    • Honestly though a lot of drivers would not have considered that so there I think we can say the AI was just behaving as a reasonable average driver would.

      This is where AI could really shine though. Right now they pay attention to road markings and cars that exist, but they could go beyond that - they could learn to classify some cars as potentially greater hazards.

      One big factor in favor of AI driving is that we can learn from the mistakes and take steps to prevent them from happening again. In contrast, we've got many years of evidence that there's no way to stop human drivers from making the same stupid mistakes over and over again.

      • Tesla cars keep making the same stupid mistakes again and again because humans keep making the same stupid mistake again and again, and letting them leave the factory without a better way to determine depth to an object than estimating from visual data alone.

        The software can't do anything about the problem if continually hampered by stupid humans.

    • I think the lesson here isn't about self-driving cars, but rather GM's quality control program.

  • It's time to lay some traps [techcrunch.com].

  • Loss of server connection and cars stopping 'immediately' is not good, but also not a big problem YET. Althought this seems to happen on a regular basis, I wonder why this didn't happen before when the cars still had safetydrivers during testing? I also expect a car to have at least 2 or even 3 different network connections, 2 different cellular and one wifi hotspot, so networkwise it would always have connection. But these taxi's are still taking its babysteps and only with practical usage will the technol
    • dead zones and tunnels that may have no network or an weak network will be bad and people can use an cell blocker to hijack auto drive trucks.

      • Those are all situations that can be dealt with. Deadzones and tunnels are easily fixed when you know the operation area, as you can install repeaters/antenna's in those locations. And ofcourse increase the autonomous part of the car, which you want for future use anyway so it doesn't have to rely on any network at all, but that's all for the near future and any day these robotic cars spend on the public road is an advancement in the technology/learning. Accidents happen and will never go away, but it's a f
  • I would have expected GM to roll this out someplace other than a crowded City like SFO. There are lots of smaller communities that would have welcomed it and they could have worked out the bugs there before trying this.

  • by itzdandy ( 183397 ) on Saturday July 09, 2022 @02:02PM (#62688264) Homepage

    There is no perfect system. I don't know how good this or any other autonomous system is, it's definitely early. What I care about really is how many accidents and how severe they are per distance per environment. ie, a direct comparison in city or highway etc driving conditions vs a human. If it's even on par, let it be. Humans aren't getting any better at driving, our algorythms are fixed. An autonomous car that matches humans however can improve with time.

  • Why do autonomous vehicles need a server?

  • Not GM Cruise autopilot but a (better) system:

    https://www.tesla.com/VehicleS... [tesla.com]

    https://cleantechnica.com/2021... [cleantechnica.com]

    “In the 2nd quarter, we recorded one crash for every 4.41 million miles driven in which drivers were using Autopilot technology (Autosteer and active safety features). For drivers who were not using Autopilot technology (no Autosteer and active safety features), we recorded one crash for every 1.2 million miles driven. By comparison, NHTSA’s most recent data shows that in the United St

    • Are they comparing those miles on the same roadways?

      Chances are very good that autopilot would not work where those drivers were not using it because the car is under conditions that auto pilot simply can not handle. I suspect the comparison between Auto-Pilot roadways and Human-Only roadways is comparing Apples and Oranges.
    • by kriston ( 7886 )

      We don't deny that these cars aren't safer in the big picture, but the accidents they cause will be novel and unique to them.

    • âoeIn the 2nd quarter, we recorded one crash for every 4.41 million miles driven in which drivers were using Autopilot technology (Autosteer and active safety features). For drivers who were not using Autopilot technology (no Autosteer and active safety features), we recorded one crash for every 1.2 million miles driven. By comparison, NHTSAâ(TM)s most recent data shows that in the United States there is an automobile crash every 484,000 miles.â

      Autopilot only works on the highway so if you are comparing autopilot use you had better use metrics that make sense otherwise your numbers communicate nothing.

  • Suicidal robocar? (Score:2, Insightful)

    by zmollusc ( 763634 )

    From TFA : It began the turn and stopped in the middle of the intersection, presumably noticing the Toyota headed for it

    What? The robocar looked at the situation, judged that it was safe for it to make the left turn and began the manouver, Next, the robocar detects the oncoming prius and immediately halts in the path of the prius?

    WTF?
    Is the robocar too shortsighted to be driving because it cannot accurately judge oncoming traffic such as the prius?

    Is the robocar able to sense the surroundings correctly b

    • by bidule ( 173941 )

      Maybe you missed it but Prius was on a right turn lane, Cruise waited for it to turn.

      The rest of your post becomes pretty much off topic.

      • From TFA : It began the turn and stopped in the middle of the intersection, presumably noticing the Toyota headed for it

        How is that 'waiting for the prius' except in the sense of 'getting into the danger area and waiting to be hit'?

        • by bidule ( 173941 )

          "was in the right turn lane before heading straight and hitting the Bolt"

          Can you at least be grounded in reality?

  • I am beta testing Tesla's FSD. And it is a good driving assistant tool. However it is a lot like driving with a teenager with a learner's permit. It can do most things well, however when there is a slightly odd condition, the car will not react as one would expect, where I would take over.

  • Fun background fact; I grew up on a farm. When the calves were really hungry they tended to stick their tongue out as far as they could while making a dopey expression as the only thing on their mind was getting the rubber cap of that bottle in their mouth so they could eat.

    The longer it took to feed them the more frantic and single minded they'd become and the more they would thrash and struggle and the toungue would flop around as they scrambled for food. "Need it NOW!" *mlalamlamluamlam*

    That's how I see

    • Single point of failure is not good, I wonder how much security they are trying to build into these things? One cloud server gets compromised and suddenly the roadways become a terror zone.
  • by sarren1901 ( 5415506 ) on Saturday July 09, 2022 @07:39PM (#62688978)

    Bill Gates on Car Industry

    I got this particularly good email in my inbox today.

    For all of us who feel only the deepest love and affection for the way computers have enhanced our lives, read on.

    At a recent computer expo (COMDEX), Bill Gates reportedly compared the computer industry with the auto industry and stated,

    ‘If GM had kept up with technology like the computer industry has, we would all be driving $25 cars that got 1,000 miles to the gallon..’

    In response to Bill’s comments, General Motors issued a press release stating:

    If GM had developed technology like Microsoft we would all be driving cars with the following characteristics (and I just love this part):

    1. For no reason whatsoever, your car would crashtwice a day.

    2.. Every time they repainted the lines in the road, you would have to buy a new car.

    3 Occasionally your car would die on the freeway for no reason. You would have to pull to the side of the road, close all of the windows, shut off the car, restart it, and reopen the windows before you could continue. For some reason you would simply accept this.

    4. Occasionally, executing a maneuver such as a left turn would cause your car to shut down and refuse to restart, in which case you would have to reinstall the engine.

    5. Macintosh would make a car that was powered by the sun, was reliable, five times as fast and twice as easy to drive - but would run on only five percent of the roads.

    6. The oil, water temperature, and alternator warning lights would all be replaced by a single 'This Car Has Performed An Illegal Operation’ warning light.

    7. The airbag system would ask 'Are you sure?’ before deploying.

    8. Occasionally, for no reason whatsoever, your car would lock you out and refuse to let you in until you simultaneously lifted the door handle, turned the key and grabbed hold of the radio antenna.

    9. Every time a new car was introduced car buyers would have to learn how to drive all over again because none of the controls would operate in the same manner as the old car.

    10. You’d have to press the 'Start’ button to turn the engine off.

    Bill Gates Cars Automobiles Windows

    Sourced from https://www.joelscanlon.com/po... [joelscanlon.com]

If you can't get your work done in the first 24 hours, work nights.

Working...