This is an unpopular opinion, and I get why – people crave a scapegoat. CrowdStrike undeniably pushed a faulty update demanding a low-level fix (booting into recovery). However, this incident lays bare the fragility of corporate IT, particularly for companies entrusted with vast amounts of sensitive personal information.

Robust disaster recovery plans, including automated processes to remotely reboot and remediate thousands of machines, aren’t revolutionary. They’re basic hygiene, especially when considering the potential consequences of a breach. Yet, this incident highlights a systemic failure across many organizations. While CrowdStrike erred, the real culprit is a culture of shortcuts and misplaced priorities within corporate IT.

Too often, companies throw millions at vendor contracts, lured by flashy promises and neglecting the due diligence necessary to ensure those solutions truly fit their needs. This is exacerbated by a corporate culture where CEOs, vice presidents, and managers are often more easily swayed by vendor kickbacks, gifts, and lavish trips than by investing in innovative ideas with measurable outcomes.

This misguided approach not only results in bloated IT budgets but also leaves companies vulnerable to precisely the kind of disruptions caused by the CrowdStrike incident. When decision-makers prioritize personal gain over the long-term health and security of their IT infrastructure, it’s ultimately the customers and their data that suffer.

  • breakingcups@lemmy.world
    link
    fedilink
    English
    arrow-up
    178
    arrow-down
    6
    ·
    4 months ago

    Please, enlighten me how you’d remotely service a few thousand Bitlocker-locked machines, that won’t boot far enough to get an internet connection, with non-tech-savvy users behind them. Pray tell what common “basic hygiene” practices would’ve helped, especially with Crowdstrike reportedly ignoring and bypassing the rollout policies set by their customers.

    Not saying the rest of your post is wrong, but this stood out as easily glossed over.

    • ramble81@lemm.ee
      link
      fedilink
      English
      arrow-up
      38
      arrow-down
      2
      ·
      4 months ago

      You’d have to have something even lower level like a OOB KVM on every workstation which would be stupid expensive for the ROI, or something at the UEFI layer that could potentially introduce more security holes.

      • Leeks@lemmy.world
        link
        fedilink
        English
        arrow-up
        11
        arrow-down
        3
        ·
        4 months ago

        Maybe they should offer a real time patcher for the security vulnerabilities in the OOB KVM, I know a great vulnerability database offered by a company that does this for a lot of systems world wide! /s

        • A_A@lemmy.world
          link
          fedilink
          English
          arrow-up
          2
          arrow-down
          2
          ·
          4 months ago

          Lol 😋 ! also i need a “Out-of-Band, Keyboard, Video, and Mouse” to your “OOB, KVM” so to steal the bank improve security.

      • circuscritic@lemmy.ca
        link
        fedilink
        English
        arrow-up
        7
        arrow-down
        2
        ·
        edit-2
        4 months ago

        …you don’t have OOBM on every single networked device and terminal? Have you never heard of the buddy system?

        You should probably start writing up an RFP. I’d suggest you also consider doubling up on the company issued phones per user.

        If they already have an ATT phone, get them a Verizon one as well, or vice versa.

        At my company we’re already way past that. We’re actually starting to import workers to provide human OOBM.

        You don’t answer my call? I’ll just text the migrant worker we chained to your leg to flick your ear until you pick up.

        Maybe that sounds extreme, but guess who’s company wasn’t impacted by the Crowdstrike outage.

        • ToyDork@preserve.games
          link
          fedilink
          English
          arrow-up
          1
          ·
          4 months ago

          I mean, with the exception of the shackles, this is just logistics 101. The more something needs to stay working or not accidentally trigger a huge problem, the more resources you dedicate to picking up where the regular guy left off because the “fleffingbridge transport 1” company’s bus broke down in front of the regular guy and his bus got hit by a train. Solution? New bus, plant some trees. Prevention? Bridges and tunnels aren’t cheap, but clearly we need one there now. We can’t predict the future but we have to do our best to try or - simulated or real - the cost will be paid in blood. Obviously there’s moral limits, but hiring more staff is not in and of itself immoral nor the wrong approach.

          If I was in charge of a real life logistics operation, I’d be devastated if anyone died because of me. I can’t say, however, that it can be avoided. Sometimes people die at random, that’s not yet 100% avoidable and might never be, but I do care. I’d hope people who actually end up in logistics could learn to indulge their empathy enough to remember there are lives on the line, but I can’t blame someone for being bitter that the actual work output is purely being fleeced for profit.

        • ramble81@lemm.ee
          link
          fedilink
          English
          arrow-up
          9
          ·
          4 months ago

          I didn’t say it was, nor did I say UEFI was the problem. My point was additional applications or extensions at the UEFI layer increase the attack footprint of a system. Just like vPro, you’re giving hackers a method that can compromise a system below the OS. And add that in to laptops and computers that get plugged in random places before VPNs and other security software is loaded and you have a nice recipe for hidden spyware and such.

    • Howdy@lemmy.zip
      link
      fedilink
      English
      arrow-up
      28
      ·
      edit-2
      4 months ago

      Was a windows sysadmin for a decade. We had thousands of machines with endpoint management with bitlocker encryption. (I have sincd moved on to more of into cloud kubertlnetes devops) Anything on a remote endpoint doesn’t have any basic “hygiene” solution that could remotely fix this mess automatically. I guess Intels bios remote connection (forget the name) could in theory allow at least some poor tech to remote in given there is internet connection and the company paid the xhorbant price.

      All that to say, anything with end-user machines that don’t allow it to boot is a nightmare. And since bit locker it’s even more complicated. (Hope your bitloxker key synced… Lol).

      • Spuddlesv2@lemmy.ca
        link
        fedilink
        English
        arrow-up
        22
        arrow-down
        2
        ·
        4 months ago

        You’re thinking of Intel vPro. I imagine some of the Crowdstrike victims customers have this and a bunch of poor level 1 techs are slowly griding their way through every workstation on their networks. But yeah, OP is deluded and/or very inexperienced if they think this could have been mitigated on workstations through some magical “hygiene”.

      • LrdThndr@lemmy.world
        link
        fedilink
        English
        arrow-up
        8
        arrow-down
        5
        ·
        4 months ago

        Bro. PXE boot image servers. You can remotely image machines from hundreds of miles away with a few clicks and all it takes on the other end is a reboot.

        • wizardbeard@lemmy.dbzer0.com
          link
          fedilink
          English
          arrow-up
          5
          arrow-down
          1
          ·
          4 months ago

          With a few clicks and being connected to the company network. Leaving anyone not able to reach an office location SOL.

          • LrdThndr@lemmy.world
            link
            fedilink
            English
            arrow-up
            5
            arrow-down
            1
            ·
            4 months ago

            Hey, it’s not perfect, but a fix that gets you 10% of the way there is still 10% you don’t have to do by hand. Don’t let perfect be the enemy of good, my man.

    • LrdThndr@lemmy.world
      link
      fedilink
      English
      arrow-up
      30
      arrow-down
      8
      ·
      edit-2
      4 months ago

      A decade ago I worked for a regional chain of gyms with locations in 4 states.

      I was in TN. When a system would go down in SC or NC, we originally had three options:

      1. (The most common) have them put it in a box and ship it to me.
      2. I go there and fix it (rare)
      3. I walk them through fixing it over the phone (fuck my life)

      I got sick of this. So I researched options and found an open source software solution called FOG. I ran a server in our office and had little optiplex 160s running a software client that I shipped to each club. Then each machine at each club was configured to PXE boot from the fog client.

      The server contained images of every machine we commonly used. I could tell FOG which locations used which models, and it would keep the images cached on the client machines.

      If everything was okay, it would chain the boot to the os on the machine. But I could flag a machine for reimage and at next boot, the machine would check in with the local FOG client via PXE and get a complete reimage from premade images on the fog server.

      The corporate office was physically connected to one of the clubs, so I trialed the software at our adjacent club, and when it worked great, I rolled it out company wide. It was a massive success.

      So yes, I could completely reimage a computer from hundreds of miles away by clicking a few checkboxes on my computer. Since it ran in PXE, the condition of the os didn’t matter at all. It never loaded the os when it was flagged for reimage. It would even join the computer to the domain and set up that locations printers and everything. All I had to tell the low-tech gymbro sales guy on the phone to do was reboot it.

      This was free software. It saved us thousands in shipping fees alone. And brought our time to fix down from days to minutes.

      There ARE options out there.

      • magikmw@lemm.ee
        link
        fedilink
        English
        arrow-up
        26
        ·
        edit-2
        4 months ago

        This works great for stationary pcs and local servers, does nothing for public internet connected laptops in hands of users.

        The only fix here is staggered and tested updates, and apparently this update bypassed even deffered update settings that crowdstrike themselves put into their software.

        The only winning move here was to not use crowdstrike.

        • wizardbeard@lemmy.dbzer0.com
          link
          fedilink
          English
          arrow-up
          9
          arrow-down
          2
          ·
          4 months ago

          It also assumes that reimaging is always an option.

          Yes, every company should have networked storage enforced specifically for issues like this, so no user data would be lost, but there’s often a gap between should and “has been able to find the time and get the required business side buy in to make it happen”.

          Also, users constantly find new ways to do non-standard, non-supported things with business critical data.

          • Bluetreefrog@lemmy.world
            link
            fedilink
            English
            arrow-up
            5
            ·
            4 months ago

            Isn’t this just more of what caused the problem in the first place? Namely, centralisation. If you store data locally and you lose a machine, that’s bad but not the end of the world. If you store it centrally and you lose the data, that’s catastrophic. Nassim Taleb nailed this stuff. Keep the downside limited, and the upside unlimited or as he says, “Don’t pick up pennies in front of a steamroller.”

        • LrdThndr@lemmy.world
          link
          fedilink
          English
          arrow-up
          8
          arrow-down
          2
          ·
          4 months ago

          Absolutely. 100%

          But don’t let perfect be the enemy of good. A fix that gets you 40% of the way there is still 40% less work you have to do by hand. Not everything has to be a fix for all situations. There’s no such thing as a panacea.

          • magikmw@lemm.ee
            link
            fedilink
            English
            arrow-up
            6
            ·
            4 months ago

            Sure. At the same time one needs to manage resources.

            I was all in on laptop deployment automation. It cut down on a lot of human error issues and having inconsistent configuration popping up all the time.

            But it needs constant supervision, even if not constant updates. More systems and solutions lead to neglect if not supplied well. So some “would be good to have” systems just never make the cut, because as overachieving I am, I’m also don’t want to think everything is taken care of when it clearly isn’t.

            • catloaf@lemm.ee
              link
              fedilink
              English
              arrow-up
              2
              ·
              4 months ago

              Yeah. I find a base image and post-install config with group policy or Ansible to be far more reliable.

              • magikmw@lemm.ee
                link
                fedilink
                English
                arrow-up
                1
                ·
                4 months ago

                Yea we’re doing something similiar. Only update base images for bigger OS updates or if something breaks or can break.

                The general idea is to have config that works for both new PCs and the ones that are already in use. Saves on maintaining two configuration methods.

            • John Richard@lemmy.worldOP
              link
              fedilink
              English
              arrow-up
              4
              arrow-down
              3
              ·
              4 months ago

              You were all in, but was the company all in? How many employees? It sounds like you innovated. Let’s say that the company you worked for was spending millions on vendors that promised solutions but rarely delivered. If instead they gave you $400k a year, a $1 million/year budget & 10 employees… I’m guessing you could have managed the laptop deployment automation, along with some other significant projects as well.

              Instead though, people with good ideas, even loyal to the company, are competing against sales and marketing reps from billion dollar companies, and upper management are easily swooned.

              • magikmw@lemm.ee
                link
                fedilink
                English
                arrow-up
                3
                ·
                4 months ago

                I’m the only one to swoon here, and I’m as sceptical as one can be.

                I’m also a cost and my budget is on paper only. Non-IT management is complicit in crappy IT.

        • John Richard@lemmy.worldOP
          link
          fedilink
          English
          arrow-up
          1
          arrow-down
          7
          ·
          4 months ago

          Almost all computers can be set to PXE boot, but work laptops usually even have more advanced remote management capabilities. You ask the employee to reboot the laptop and presto!

          • magikmw@lemm.ee
            link
            fedilink
            English
            arrow-up
            7
            ·
            4 months ago

            I wonder how you’re supposed to get PXE boot to work securely over the internet. And how that helps when affected disk is still encrypted and needs unusual intervention to fix, including admin access to system files.

            I’ve been doing this for a while, and I like creative solutions, so I wonder about those issues a lot. Not much comes to my mind besides let’s recall all the laptops and do it one by one.

            • wizardbeard@lemmy.dbzer0.com
              link
              fedilink
              English
              arrow-up
              2
              arrow-down
              1
              ·
              edit-2
              4 months ago

              Hypothectically you could ship a company provided router to handle the vpn connection to your remote users, so you aren’t relying on the OS to be able to boot up to get connected to the vpn for the company network and PXE environment. Lots of extra cost and mess though.

            • LrdThndr@lemmy.world
              link
              fedilink
              English
              arrow-up
              1
              ·
              4 months ago

              From a home user? Probably ain’t shit-all you can do with PXE booting. But if you have a field office or somewhere a user can go with a hardware vpn appliance? Well now you’re in business.

            • John Richard@lemmy.worldOP
              link
              fedilink
              English
              arrow-up
              3
              arrow-down
              4
              ·
              4 months ago

              I wonder how you’re supposed to get PXE boot to work securely over the internet.

              PXE boot is more of last resort IMO, but can be uses as a chainloader to a more secure option. The biggest challenge I could see security-wise is having PXE boot being ran on unsecured networks. Even then though, normally a computer will have been provisioned on a secure network and will have encryption and secure boot-based encryption, and some additional signature-based image verification.

      • yeehaw@lemmy.ca
        link
        fedilink
        English
        arrow-up
        5
        ·
        4 months ago

        This is a good solution for these types of scenarios. Doesn’t fit all though. Where I work, 85% of staff work from home. We largely use SaaS. I’m struggling to think of a good method here other than walking them through reinstalling windows on all their machines.

        • LrdThndr@lemmy.world
          link
          fedilink
          English
          arrow-up
          3
          arrow-down
          1
          ·
          4 months ago

          That’s still 15% less work though. If I had to manually fix 1000 computers, clicking a few buttons to automatically fix 150 of them sounds like a sweet-ass deal to me even if it’s not universal.

          You could also always commandeer a conference room or three and throw a switch on the table. “Bring in your laptop and go to conference room 3. Plug in using any available cable on the table and reboot your computer. Should be ready in an hour or so. There’s donuts and coffee in conference room 4.” Could knock out another few dozen.

          Won’t help for people across the country, but if they’re nearish, it’s not too bad.

          • yeehaw@lemmy.ca
            link
            fedilink
            English
            arrow-up
            2
            ·
            4 months ago

            Not a lot of nearish. It would be pretty bad if this happened here.

        • John Richard@lemmy.worldOP
          link
          fedilink
          English
          arrow-up
          3
          arrow-down
          6
          ·
          edit-2
          4 months ago
          1. Configure PXE to reboot into recovery image, push out command to remove bad file. Reboot. Done. Workstation laptops usually have remote management already.

          or

          1. Have recovery image already installed. Have user reboot & push key to boot into recovery. Push out fix. Done.
          • yeehaw@lemmy.ca
            link
            fedilink
            English
            arrow-up
            3
            ·
            4 months ago

            I had no idea you could remotely configure pxe to reboot into a recovery image and run a script. How do you do this?

          • LrdThndr@lemmy.world
            link
            fedilink
            English
            arrow-up
            1
            arrow-down
            1
            ·
            edit-2
            4 months ago

            Fuck yeah. Even better than reimage. That’s creative as fuck and I love it.

        • LrdThndr@lemmy.world
          link
          fedilink
          English
          arrow-up
          5
          ·
          edit-2
          4 months ago

          How would it not have? You got an office or field offices?

          “Bring your computer by and plug it in over there.” And flag it for reimage. Yeah. It’s gonna be slow, since you have 200 of the damn things running at once, but you really want to go and manually touch every computer in your org?

          The damn thing’s even boot looping, so you don’t even have to reboot it.

          I’m sure the user saved all their data in one drive like they were supposed to, right?

          I get it, it’s not a 100% fix rate. And it’s a bit of a callous answer to their data. And I don’t even know if the project is still being maintained.

          But the post I replied to was lamenting the lack of an option to remotely fix unbootable machines. This was an option to remotely fix nonbootable machines. No need to be a jerk about it.

          But to actually answer your question and be transparent, I’ve been doing Linux devops for 10 years now. I haven’t touched a windows server since the days of the gymbros. I DID say it’s been a decade.

          • Brkdncr@lemmy.world
            link
            fedilink
            English
            arrow-up
            8
            arrow-down
            4
            ·
            4 months ago

            Because your imaging environment would also be down. And you’re still touching each machine and bringing users into the office.

            Or your imaging process over the wan takes 3 hours since it’s dynamically installing apps and updates and not a static “gold” image. Imaging is then even slower because your source disk is only ssd and imaging slows down once you get 10+ going at once.

            I’m being rude because I see a lot of armchair sysadmins that don’t seem to understand the scale of the crowdstike outage, what crowdstrike even is beyond antivirus, and the workflow needed to recover from it.

            • LrdThndr@lemmy.world
              link
              fedilink
              English
              arrow-up
              6
              ·
              4 months ago

              FOG ran on Linux. It wouldn’t have been down. But that’s beside the point.

              I never said it was a good answer to CrowdStrike. It was just a story about how I did things 10 years ago, and an option for remotely fixing nonbooting machines. That’s it.

              I get you’ve been overworked and stressed as fuck this last few days. I’ve been out of corporate IT for 10 years and I do not envy the shit you guys are going through right now. I wish I could buy you a cup of coffee or a beer or something.

              • Brkdncr@lemmy.world
                link
                fedilink
                English
                arrow-up
                3
                ·
                4 months ago

                Last time I used fog it was only doing static image deployment which has been out of style for a while. I don’t know if there are any serious deployment products for windows enterprise that don’t run on windows.

                I’m personally not dealing with this because I didn’t like how Crowdstrike had answered a number of questions in their sales call.

                Avoiding telling me their vuln scan doesn’t prob be all hosts after claiming it could replace a real vuln scanner, claiming they are somehow better than others at malware detection without bringing up 3rd party tests, claiming how their product was novel when others have been doing the same for 7+ years.

                My fave was them telling me how much easier it is to manage but no one on the call had ever worked as a sysadmin or even seen how their competition works.

                Shitshow. I’m so glad this happened so I can block their sales team.

            • John Richard@lemmy.worldOP
              link
              fedilink
              English
              arrow-up
              2
              arrow-down
              4
              ·
              4 months ago

              Imaging environment down? If a sysadmin can’t figure out how to boot a machine into recovery to remove the bad update file then they have bigger problems. The fix in this instance wasn’t even re-imaging machines. It was merely removing a file. Ideal DR scenario would have a recovery image already on the system that can be booted into remotely, so there is minimal strain on the network. Furthermore, we don’t live in dial-up age anymore.

      • John Richard@lemmy.worldOP
        link
        fedilink
        English
        arrow-up
        3
        arrow-down
        5
        ·
        4 months ago

        Thank you for sharing this. This is what I’m talking about. Larger companies not utilizing something like this already are dysfunctional. There are no excuses for why it would take them days, weeks or longer.

    • Dran@lemmy.world
      link
      fedilink
      English
      arrow-up
      16
      arrow-down
      2
      ·
      edit-2
      4 months ago

      Separate persistent data and operating system partitions, ensure that every local network has small pxe servers, vpned (wireguard, etc) to a cdn with your base OS deployment images, that validate images based on CA and checksum before delivering, and give every user the ability to pxe boot and redeploy the non-data partition.

      Bitlocker keys for the OS partition are irrelevant because nothing of value is stored on the OS partition, and keys for the data partition can be stored and passed via AD after the redeploy. If someone somehow deploys an image that isn’t ours, it won’t have keys to the data partition because it won’t have a trust relationship with AD.

      (This is actually what I do at work)

      • I_Miss_Daniel@lemmy.world
        link
        fedilink
        English
        arrow-up
        7
        ·
        4 months ago

        Sounds good, but can you trust an OS partition not to store things in %programdata% etc that should be encrypted?

        • Dran@lemmy.world
          link
          fedilink
          English
          arrow-up
          4
          arrow-down
          2
          ·
          4 months ago

          With enough autism in your overlay configs, sure, but in my environment tat leakage is still encrypted. It’s far simpler to just accept leakage and encrypt the OS partition with a key that’s never stored anywhere. If it gets lost, you rebuild the system from pxe. (Which is fine, because it only takes about 20 minutes and no data we care about exists there) If it’s working correctly, the OS partition is still encrypted and protects any inadvertent data leakage from offline attacks.

      • Trainguyrom@reddthat.com
        link
        fedilink
        English
        arrow-up
        5
        ·
        edit-2
        4 months ago

        Separate persistent data and operating system partitions, ensure that every local network has small pxe servers, vpned (wireguard, etc) to a cdn with your base OS deployment images, that validate images based on CA and checksum before delivering, and give every user the ability to pxe boot and redeploy the non-data partition.

        At that point why not just redirect the data partition to a network share with local caching? Seems like it would simplify this setup greatly (plus makes enabling shadow copy for all users stupid easy)

        Edit to add: I worked at a bank that did this for all of our users and it was extremely convenient for termed employees since we could simply give access to the termed employee’s share to their manager and toss a them a shortcut to access said employee’s files, so if it turned out Janet had some business critical spreadsheet it was easily accessible even after she was termed

        • Dran@lemmy.world
          link
          fedilink
          English
          arrow-up
          3
          ·
          4 months ago

          We do this in a lot of areas with fslogix where there is heavy persistent data, it just never felt necessary to do that for endpoints where the persistent data partition is not much more than user settings and caches of convenience. Anything that is important is never stored solely on the endpoints, but it is nice to be able to reboot those servers without affecting downstream endpoints. If we had everything locally dependant on fslogix, I’d have to schedule building-wide outages for patching.

      • Brkdncr@lemmy.world
        link
        fedilink
        English
        arrow-up
        5
        arrow-down
        2
        ·
        4 months ago

        But your pxe boot server is down, your radius server providing vpn auth is down, your bitlocker keys are in AD which is down because all your domain controllers are down.

        • Dran@lemmy.world
          link
          fedilink
          English
          arrow-up
          4
          arrow-down
          1
          ·
          4 months ago

          Yes and no. In the best case, endpoints have enough cached data to get us through that process. In the worst case, that’s still a considerably smaller footprint to fix by hand before the rest of the infrastructure can fix itself.

      • pHr34kY@lemmy.world
        link
        fedilink
        English
        arrow-up
        2
        ·
        4 months ago

        I’ve been separating OS and data partitions since I was a kid running Windows 95. It’s horrifying that people don’t expect and prepare for machines to become unbootable on a regular basis.

        Hell, I bricked my work PC twice this year just by using the Windows cleanup tool - on Windows 11. The antivirus went nuclear, as antivirus products do.

    • felbane@lemmy.world
      link
      fedilink
      English
      arrow-up
      10
      arrow-down
      1
      ·
      4 months ago

      Rollout policies are the answer, and CrowdStrike should be made an example of if they were truly overriding policies set by the customer.

      It seems more likely to me that nobody was expecting “fingerprint update” to have the potential to completely brick a device, and so none of the affected IT departments were setting staged rollout policies in the first place. Or if they were, they weren’t adequately testing.

      Then - after the fact - it’s easy to claim that rollout policies were ignored when there’s no way to prove it.

      If there’s some evidence that CS was indeed bypassing policies to force their updates I’ll eat the egg on my face.

      • originalucifer@moist.catsweat.com
        link
        fedilink
        arrow-up
        8
        arrow-down
        1
        ·
        edit-2
        4 months ago

        from what ive read/watched thats the crux of the issue… did they push a ‘content’ update, i.e. signatures or did they push a code update.

        so you basically had a bunch of companies who absolutely do test all vendor code updates beings slipped a code update they werent aware of being labeled a ‘content’ update.

      • DesertCreosote@lemm.ee
        link
        fedilink
        English
        arrow-up
        1
        ·
        4 months ago

        I’m one of the admins who manage CrowdStrike at my company.

        We have all automatic updates disabled, because when they were enabled (according to the CrowdStrike best practices guide they gave us), they pushed out a version with a bug that overwhelmed our domain servers. Now we test everything through multiple environments before things make it to production, with at least two weeks of testing before we move a version to the next environment.

        This was a channel file update, and per our TAM and account managers in our meeting after this happened, there’s no way to stop that file from being pushed, or to delay it. Supposedly they’ll be adding that functionality in now.

    • sp3ctr4l@lemmy.zip
      link
      fedilink
      English
      arrow-up
      7
      arrow-down
      1
      ·
      edit-2
      4 months ago

      You are talking about how to fix the problem.

      This person is talking about what caused the problem.

      Completely different things.

      1. Bad thing happened, how do we fix bad thing and its effects.

      Analogous to: A house is on fire; call the ambulances to treat any wounded call the fire department, call insurance, figure out temporary housing.

      This is basically immediate remedy or mitigation.

      1. Bad thing happened, but why did the bad thing happen and how to we prevent future occurrences of this?

      Analogous to: Investigate the causes of the fire, suggest various safety regulations on natural gas infrastructure, home appliances, electrical wiring, building material and methods, etc.

      This is much more complex and involves systemic change.

    • lazynooblet@lazysoci.al
      link
      fedilink
      English
      arrow-up
      6
      arrow-down
      2
      ·
      4 months ago

      Autopilot, intune. Force restart device twice to get startup repair, choose factory reset, share LAPS admin password and let the workstation rebuild itself.

    • JasonDJ@lemmy.zip
      link
      fedilink
      English
      arrow-up
      3
      arrow-down
      1
      ·
      4 months ago

      Does Windows have a solid native way to remotely re-image a system like macOS does?

      • catloaf@lemm.ee
        link
        fedilink
        English
        arrow-up
        1
        arrow-down
        4
        ·
        4 months ago

        No.

        Maybe with Intune and Autopilot, but I haven’t used it.

        • John Richard@lemmy.worldOP
          link
          fedilink
          English
          arrow-up
          2
          arrow-down
          3
          ·
          4 months ago

          Windows ADK does this too, or any PXE server really… so yes, you can. The CS issue though didn’t require re-image. Merely removing a file. DR planning would usually have a recovery image pre-installed to automate booting into for lower-level fixes.

    • Saik0@lemmy.saik0.com
      link
      fedilink
      English
      arrow-up
      3
      arrow-down
      1
      ·
      4 months ago

      Please, enlighten me how you’d remotely service a few thousand Bitlocker-locked machines, that won’t boot far enough to get an internet connection,

      Intel AMT.

    • nomad@infosec.pub
      link
      fedilink
      English
      arrow-up
      2
      arrow-down
      2
      ·
      edit-2
      4 months ago

      It’s called EFI. How do you think your BIOS update from inside BIOS is working? ;)

      EDIT: oh, and PXE boot + wol.

    • Riskable@programming.dev
      link
      fedilink
      English
      arrow-up
      3
      arrow-down
      4
      ·
      edit-2
      3 months ago

      what common “basic hygiene” practices would’ve helped

      Not using a proprietary, unvetted, auto-updating, 3rd party kernel module in essential systems would be a good start.

      Back in the day companies used to insist upon access to the source code for such things along with regular 3rd party code audits but these days companies are cheap and lazy and don’t care as much. They’d rather just invest in “security incident insurance” and hope for the best 🤷

      Sometimes they don’t even go that far and instead just insist upon useless indemnification clauses in software licenses. …and yes, they’re useless:

      https://www.nolo.com/legal-encyclopedia/indemnification-provisions-contracts.html#:~:text=Courts have commonly held that,knowledge of the relevant circumstances).

      (Important part indicating why they’re useless should be highlighted)

    • SuperFola@programming.dev
      link
      fedilink
      English
      arrow-up
      1
      arrow-down
      2
      ·
      4 months ago

      Dual partitioning as Android does it might have helped. Install the update to partition B, reboot and if it’s alright swap A and B partitions to make B the default. Boot again to the default partition (A, formerly B).

      It wouldn’t have booted correctly afaiu with the faulty update, and would have been reverted to use the untouched A partition.

    • John Richard@lemmy.worldOP
      link
      fedilink
      English
      arrow-up
      2
      arrow-down
      4
      ·
      4 months ago

      I’d issue IPMI or remote management commands to reboot the machines. Then I’d boot into either a Linux recovery environment (yes, Linux can unlock BitLocker-encrypted drives) or a WinPE (or Windows RE) and unlock the drives, preferably already loaded on the drives, but could have them PXE boot - just giving ideas here, but ideal DR scenario would have an environment ready to load & PXE would cause delays.

      I’d either push a command or script that would then remove the update file that caused the issue & then reboots. Having planned for a scenario like this already, total time to fix would be less than 2 hours.

      • Riskable@programming.dev
        link
        fedilink
        English
        arrow-up
        3
        ·
        4 months ago

        At my company I use a virtual desktop and it was restored from a nightly snapshot a few hours before I logged in that day (and presumably, they also applied a post-restore temp fix). This action was performed on all the virtual desktops at the entire company and took approximately 30 minutes (though, probably like 4 hours to get the approval to run that command, LOL).

        It all took place before I even logged in that day. I was actually kind of impressed… We don’t usually act that fast.