A week of downtime and all the servers were recovered only because the customer had a proper disaster recovery protocol and held backups somewhere else, otherwise Google deleted the backups too

Google cloud ceo says “it won’t happen anymore”, it’s insane that there’s the possibility of “instant delete everything”

  • They said the outage was caused by a misconfiguration that resulted in UniSuper’s cloud account being deleted, something that had never happened to Google Cloud before.

    Bullshit. I’ve heard of people having their Google accounts randomly banned or even deleted before. Remember when the Terraria devs cancelled the Stadia port of Terraria because Google randomly banned their account and then took weeks to acknowledge it? The only reason why Google responded so quickly to this is because the super fund manages over $100b and could sue the absolute fuck out of Google.

    •  Pechente   ( @Pechente@feddit.de ) 
      cake
      link
      fedilink
      English
      302 months ago

      This happened to me years ago. Suddenly got a random community guidelines violation on YouTube for a 3 second VFX shot that was not pornographic or violent and that I owned all the rights to. After that my whole Google account was locked down. I never found out what triggered this response and I could never resolve the issue with them since I only ever got automated responses. Fuck Google.

      •  NaN   ( @Bitrot@lemmy.sdf.org ) 
        link
        fedilink
        English
        132 months ago

        This sort of story is what made me switch away from Google Fi and ultimately mostly degoogling. Privacy was a big part later on, but initially it was realizing that a YouTube comment or a file in my drive could get my cell service turned off.

      • If you are a small company then yes. But i would argue that for larger companies this doesn’t hold true. If you have 200 employees you’ll need an IT department either way. You need IT expertise either way. So having some people who know how to plan, implement and maintain physical hardware makes sense too.

        There is a breaking point between economics of scale and the added efforts to coordinate between your company and the service provider plus paying that service providers overhead and profits.

        • If coordinating with service providers is hard for a firm, I would argue the cost effective answer isn’t “let’s do all this in house”. Many big finance firms fall in this trap of thinking it’s cheaper to build v buy, and that’s how you get everyone building their own worse versions of everything. Whether your firm is good at the markets or kitchens or travel bookings, thinking you can efficiently in-source tech is a huge fallacy.

          • it is not about it being hard. It simply creates effort to coordinate. And this effort needs to be considered. If you do things externally that means there is two PMs to pay, you need QMs on both sides, you need two legal/contract teams, you need to pay someone in procurement and someone in sales…

            I agree with you that doing software inhouse when there is good options on the market is usually not a good idea. But for infrastructure i don’t see there to be as much of an efficiency loss. Especially as you very much need experts on how to set things up in a cloud environment and you better look carefully at how many resources you need to not overpay huge amounts.

      • It’s absolutely not. If you are at any kind of scale whatsoever, your yearly spend will be a minimum of 2x at a cloud provider rather then creating and operating the same system locally including all the employees, contracts, etc.

    • G Suite is a legitimate option for small-medium businesses. It’s seen as the cheaper, simpler option versus Azure. I usually recommend it for nonprofits as they have a decent free option for 501c3 orgs.

    • They had backups at multiple locations, and lost data at multiple (Google Cloud) locations because of the account deletion.

      They restored from backups stored at another provider. It may have been more devastating if they relied exclusively on google for backups. So having an “offsite backup” isn’t enough in some cases, that offsite location need to be at a different provider.

      • It may have been more devastating if they relied exclusively on google for backups.

        Which is why having any data, despite the number of backups, on a cloud provider shouldn’t be seen as off-site.

        Only when it is truly outside their ecosphere and cannot be touched by them should it be viewed as such.

        If that company didn’t have such resilience built into their backup plan, they would be toast with a derisory amount of compensation from Google.

        • Having a backup at a cloud provider is fine, as long as there is at least one other backup that isn’t with this provider.

          Cloud provider seems to do a good job protecting against hardware failure, but can do poorly with arbitrary account bans, and sometimes have mishaps due to configuration problems.

          Whereas a DIY backup solution is often more subject to hardware problems (disk failure, fire, flooding, theft, …), but there’s no risk of account problem.

          A mix is fine to protect against different kind of issues.

          • as long as there is at least one other backup that isn’t with this provider.

            Which is exactly what I was saying.

            Any services used with a cloud provider should be treated as 1 entity, no matter how many geo-locations they claim your data is backed up to because they are a single point from which all those can be deleted.

            When I was last involved in a companies backups, we had a fire safe in the basement, we had an off-site location with another fire safe & third copies would go off to another company that provided a backup storage solution so for all backups to be deleted, someone had to go right out of their way to do so. Not just a simple deletion of our account & all backups are wiped.

            That company had the foresight to do something similar & it’s saved them. [edited - was on the tube when I wrote this and didnt see the autocorrect had put ‘comment’, not ‘company’]

  • Just an FYI in case you don’t follow Cloud news but Google has deleted customers accounts on multiple occasions and has been for literal years. This time they just did it to someone large enough to make the news. I work in SRE and no longer recommend GCP to anyone.

  • This is the best summary I could come up with:


    More than half a million UniSuper fund members went a week with no access to their superannuation accounts after a “one-of-a-kind” Google Cloud “misconfiguration” led to the financial services provider’s private cloud account being deleted, Google and UniSuper have revealed.

    Services began being restored for UniSuper customers on Thursday, more than a week after the system went offline.

    Investment account balances would reflect last week’s figures and UniSuper said those would be updated as quickly as possible.

    In an extraordinary joint statement from Chun and the global CEO for Google Cloud, Thomas Kurian, the pair apologised to members for the outage, and said it had been “extremely frustrating and disappointing”.

    “These backups have minimised data loss, and significantly improved the ability of UniSuper and Google Cloud to complete the restoration,” the pair said.

    “Restoring UniSuper’s Private Cloud instance has called for an incredible amount of focus, effort, and partnership between our teams to enable an extensive recovery of all the core systems.


    The original article contains 412 words, the summary contains 162 words. Saved 61%. I’m a bot and I’m open source!