IT administrators are struggling to deal with the ongoing fallout from the faulty CrowdStrike update. One spoke to The Register to share what it is like at the coalface.

Speaking on condition of anonymity, the administrator, who is responsible for a fleet of devices, many of which are used within warehouses, told us: “It is very disturbing that a single AV update can take down more machines than a global denial of service attack. I know some businesses that have hundreds of machines down. For me, it was about 25 percent of our PCs and 10 percent of servers.”

He isn’t alone. An administrator on Reddit said 40 percent of servers were affected, along with 70 percent of client computers stuck in a bootloop, or approximately 1,000 endpoints.

Sadly, for our administrator, things are less than ideal.

Another Redditor posted: "They sent us a patch but it required we boot into safe mode.

"We can’t boot into safe mode because our BitLocker keys are stored inside of a service that we can’t login to because our AD is down.

  • @[email protected]
    link
    fedilink
    English
    114 months ago

    They also don’t seem to have a process for testing updates like these…?

    This seems like showing some really shitty testing practices at a ton of IT departments.

    • 🖖USS-Ethernet
      link
      fedilink
      English
      214 months ago

      Apparently from what I was reading these are forced updates from Crowdstrike, you don’t have a choice.

      • @[email protected]
        link
        fedilink
        English
        114 months ago

        I’ve heard differently. But if it’s true, that should have been a non-starter for the product for exactly reasons like this. This is basic stuff.

        • @[email protected]
          link
          fedilink
          English
          164 months ago

          Companies use crowdstrike so they don’t need internal cybersecurity. Not having automatic updates for new cyber threats sorta defeats the purpose of outsourcing cybersecurity.

          • @[email protected]
            link
            fedilink
            English
            64 months ago

            Automatic updates should still have risk mitigation in place, and the outage didn’t only affect small businesses with no cyber security capability. Outsourcing does not mean closing your eyes and letting the third party do whatever they want.

            • @[email protected]
              link
              fedilink
              English
              84 months ago

              Outsourcing does not mean closing your eyes and letting the third party do whatever they want.

              It shouldn’t, but when the decisions are made by bean counters and not people with security knowledge things like this can easily (and frequently) happen.

          • @[email protected]
            link
            fedilink
            English
            54 months ago

            Not bothering doing basic, minimal testing - and other mitigation processes - before rolling out updates is absolutely terrible policy.

    • @[email protected]
      link
      fedilink
      English
      54 months ago

      Unfortunately, the pace of attack development doesn’t really give much time for testing.

        • TonyOstrich
          link
          fedilink
          English
          54 months ago

          I was just thinking about something similar. I can understand wanting to get a security update as quickly as possible, but it still seems like some kind of rolling update could have mitigated something like this. When I say rolling, I mean for example split all of your customers into 24 groups and push the update once an hour to another group. If it causes a massive fuck up it’s only some or most, but not all.

          • @[email protected]
            link
            fedilink
            English
            44 months ago

            Heck even 30 minutes ahead for 1% of devices wouldve had a reasonable chance of catching this