I’m writing a program that wraps around dd to try and warn you if you are doing anything stupid. I have thus been giving the man page a good read. While doing this, I noticed that dd supported all the way up to Quettabytes, a unit orders of magnitude larger than all the data on the entire internet.
This has caused me to wonder what the largest storage operation you guys have done. I’ve taken a couple images of hard drives that were a single terabyte large, but I was wondering if the sysadmins among you have had to do something with e.g a giant RAID 10 array.
- Davel23 ( @Davel23@fedia.io ) 54•3 months ago
Not that big by today’s standards, but I once downloaded the Windows 98 beta CD from a friend over dialup, 33.6k at best. Took about a week as I recall.
- absGeekNZ ( @absGeekNZ@lemmy.nz ) English13•3 months ago
Yep, downloaded XP over 33.6k modem, but I’m in NZ so 33.6 was more advertising than reality, it took weeks.
- freijon ( @freijon@lemmings.world ) 42•3 months ago
I’m currently backing up my /dev folder to my unlimited cloud storage. The backup of the file
/dev/random
is running since two weeks.- Mike1576218 ( @Mike1576218@lemmy.ml ) 7•3 months ago
No wonder. That file is super slow to transfer for some reason. but wait till you get to /dev/urandom. That file hat TBs to transfer at whatever pipe you can throw at it…
I’m guessing this is a joke, right?
- PlexSheep ( @PlexSheep@infosec.pub ) 3•3 months ago
/dev/random and other “files” in /dev are not really files, they are interfaces which van be used to interact with virtual or hardware devices. /dev/random spits out cryptographically secure random data. Another example is /dev/zero, which spits out only zero bytes.
Both are infinite.
Not all “files” in /dev are infinite, for example hard drives can (depending on which technology they use) be accessed under /dev/sda /dev/sdb and so on.
I’m aware of that. I was quite sure the author was joking, with the slightest bit of concern of them actually making the mistake.
- Norah - She/They ( @princessnorah@lemmy.blahaj.zone ) English4•3 months ago
Cool, so I learned something new today. Don’t run
cat /dev/random
- Urist ( @Urist@lemmy.ml ) English41•3 months ago
I obviously downloaded a car after seeing that obnoxious anti-piracy ad.
- Neuromancer49 ( @Neuromancer49@midwest.social ) English31•3 months ago
In grad school I worked with MRI data (hence the username). I had to upload ~500GB to our supercomputing cluster. Somewhere around 100,000 MRI images, and wrote 20 or so different machine learning algorithms to process them. All said and done, I ended up with about 2.5TB on the supercomputer. About 500MB ended up being useful and made it into my thesis.
Don’t stay in school, kids.
- fuckwit_mcbumcrumble ( @fuckwit_mcbumcrumble@lemmy.dbzer0.com ) English26•3 months ago
Entire drive/array backups will probably be by far the largest file transfer anyone ever does. The biggest I’ve done was a measly 20TB over the internet which took forever.
Outside of that the largest “file” I’ve copied was just over 1TB which was a SQL file backup for our main databases at work.
- Possibly linux ( @possiblylinux127@lemmy.zip ) English3•3 months ago
10TB is child’s play
- ramble81 ( @ramble81@lemm.ee ) 22•3 months ago
I’ve done a 1PB sync between a pair of 8-node SAN clusters as one was being physically moved since it’d be faster to seed the data and start a delta sync rather than try to do it all over a 10Gb pipe. M
- pixeltree ( @pixeltree@lemmy.blahaj.zone ) 14•3 months ago
I once deleted an 800 gb log file, does that count
- Loulou ( @Loulou@lemmy.mindoki.com ) English5•3 months ago
Depends, did you send it to the trash can first?
- pixeltree ( @pixeltree@lemmy.blahaj.zone ) 4•3 months ago
Nah, rm’d on shudders Oracle linux
- Magnolia_ ( @Magnolia_@lemmy.ca ) English12•3 months ago
a .png of your mom’s width
- neidu2 ( @neidu2@feddit.nl ) 10•3 months ago
I don’t remember how many files, but typically these geophysical recordings clock in at 10-30 GB. What I do remember, though, was the total transfer size: 4TB. It was kind of like a bunch of .segd, and they were stored in this server cluster that was mounted in a shipping container for easy transport and lifting onboard survey ships. Some geophysics processors needed it on the other side of the world. There were nobody physically heading in the same direction as the transfer, so we figured it would just be easier to rsync it over 4G. It took a little over a week to transfer.
Normally when we have transfers of a substantial size going far, we ship it on LTO. For short distance transfers we usually run a fiber, and I have no idea how big the largest transfer job has been that way. Must be in the hundreds of TB. The entire cluster is 1.2PB, bit I can’t recall ever having to transfer everything in one go, as the receiving end usually has a lot less space.
4G?! That strikes fear into my heart!
- neidu2 ( @neidu2@feddit.nl ) 7•3 months ago
The alternative was 5mbit/s VSAT. 4G was a luxury at that time.
- Larvitz :fedora: :redhat: ( @Larvitz@burningboard.net ) 10•3 months ago
@data1701d downloading forza horizon 5 on Steam with around 120gb is the largest web-download, I can remember. In LAN, I’ve migrated my old FreeBSD NAS to my new one, which was a roughly 35TB transfer over NFS.
How long did that 35TB take? 12 hours or so?
- MerchantsOfMisery ( @MerchantsOfMisery@lemmy.ml ) 9•3 months ago
8 TB but I’m just a regular Joe with a penchant for piracy.
Arrrrrr!
- SoleInvictus ( @SoleInvictus@lemmy.blahaj.zone ) 1•3 months ago
Ahoy!
- Decency8401 ( @Decency8401@discuss.tchncs.de ) 8•3 months ago
A few years back I worked at a home. They organised the whole data structure but needed to move to another Providor. I and my colleagues moved roughly just about 15.4 TB. I don’t know how long it took because honestly we didn’t have much to do when the data was moving so we just used the downtime for some nerd time. Nerd time in the sense that we just started gaming and doing a mini LAN party with our Raspberry and banana pi’s.
Surprisingly the data contained information of lots of long dead people which is quiet scary because it wasn’t being deleted.
- HappyTimeHarry ( @HappyTimeHarry@lemm.ee ) English7•3 months ago
I downloaded that 200gb leak from national public data the other day, maybe not the biggest total but certainly the largest single text file ive ever messed with
- delirious_owl ( @delirious_owl@discuss.online ) 7•3 months ago
Upgraded a NAS for the office. It was reaching capacity, so we replaced it. Transfer was maybe 30 TB. Just used rsync. That local transfer was relatively fast. What took longer was for the NAS to replicate itself with its mirror located in a DC on the other side of the country.
- Random Dent ( @CrabAndBroom@lemmy.ml ) English4•3 months ago
Yeah it’s kind of wild how fast (and stable) rsync is, especially when you grew up with the extremely temperamental Windows copying thing, which I’ve seen fuck up a 50mb transfer before.
The biggest one I’ve done in one shot with rsync was only about 1tb, but I was braced for it to take half a day and cause all sorts of trouble. But no, it just sent it across perfectly first time, way faster than I was expecting.
- delirious_owl ( @delirious_owl@discuss.online ) 1•3 months ago
Never dealt with windows. rsync just makes sense. I especially like that its idempotent, so I can just run it twice or three times and it’ll be near instant on the subsequent run.
- Avid Amoeba ( @avidamoeba@lemmy.ca ) 6•3 months ago
~15TB over the internet via 30Mbps uplink without any special considerations. Syncthing handled any and all network and power interruptions. I did a few power cable pulls myself.