29

I had heard that tape is still the best medium for storing large amounts of data. So I figured I can store a relatively large amount of data on a cassette tape. I was thinking of a little project to read/write digital data on a cassette tape from my computer sound card just for the retro feeling. (And perhaps read/write that tape with an Arduino too).

But after reading up about it for a bit it turns out that they can store very small amounts of data. With baud rates varying between 300 to 2400 something between ~200KB to ~1.5MB can be stored on a 90 minute (2x45min) standard cassette tape.

Now I have a lot of problems with understanding why that is.

1- These guys can store 90 minutes of audio. Even if we assume the analog audio quality on them was equivalent of 32Kbps that's about 21MB of data. I have a hard time believing what I listened to was 300bps quality audio.

2- I read about the Kansas City standard and I can't understand why the maximum frequency they're using is 4800Hz yielding a 2400 baud. Tape (according to my internet search) can go up to 15KHz. Why not use 10KHz frequency and achieve higher bauds?

3- Why do all FSK modulations assign a frequency spacing equal to baud rate? In the Kansas example they are using 4800Hz and 2400Hz signals for '1' and '0' bits. In MFSK-16 spacing is equal to baud rate as well.

Why don't they use a MFSK system with a 256-element alphabet? With 20Hz space between each frequency the required bandwidth would be ~5KHZ. We have 10KHz in cassette tape so that should be plenty. Now even if all our symbols were the slowest one (5KHz) we would have 5*8 = 40000 baud. That's 27MB of data. Not too far from the 21MB estimation above.

4- If tape is so bad then how do they store Terabaytes on it?

Pouria P
  • 393
  • 1
  • 3
  • 5
  • 6
    All the answers can be summed up simply as: you are neglecting the factor of "distortion". Audio has a somewhat high tolerance for distortion, limited by what the average casual listener will tolerate. Digital has zero tolerance for distortion. So you're comparing apples and oranges. If you digitally encoded audio from a cassette, using a jack into an encoding device, you'd have practically no chance of getting an identical digital file from 2 different playbacks. But that's what you're assuming in your question. – JoelFan Oct 06 '20 at 05:04
  • 2
    The 300 to 2400 Bauds were used for systems using "normal" audio equipment (such as an audio cassette recorder). Such equipment is optimized for music and contains filters (I suspect band-pass filters) to improve the sound. These filters do not allow high baud rates. Not using cassette recorders optimized for music, much higher baud rates are possible: About 7 kBauds using 1980s home computers. And Wikipedia mentions (more) modern devices able to store 60 MB on a cassette - which should be more than 100 kBauds. – Martin Rosenau Oct 06 '20 at 07:17
  • 5
    @JoelFan Yes a single bit error may crash the digitial computer but a predictable distortion is quite acceptable than an unpredictable noise or parameter variations. Digital transmission and storage has the main advantage of being noise and distortion tolerant. So everytime you write a logic "1" or "0" into the medium, you write something different, but at the end, you can recover it perfectly (unless there 's Huuge noise or distortion, in which case you can still use error-detection and correction). That's the fundamental reason why analog systems are replaced by the digital ones. – Fat32 Oct 06 '20 at 15:29
  • 1
    4- If tape is so bad then how do they store Terabaytes on it?

    It's like comparing writing a novel on a roll of toilet paper or in a large notebook... The medium and sizes are different. Toilet paper isn't optimized for writing, likewise audio cassettes are optimized for analog recording, not digital.

    – Jason Goemaat Oct 08 '20 at 14:31
  • @Fat32, We are saying the same thing. The only reason you can write a "1" or "0" to a cassette tape and read it back reliably is because you are taking up a relatively large amount of tape (when compared to analog audio recording) to record a tiny amount of digital data (1 bit). You are trading tape space for fidelity. 75 inches of cassette tape can either record 10 seconds of acceptable-fidelity audio or 375 characters (about the size of the OP's question) of acceptable-fidelity digital data. The difference in data density is staggering. – JoelFan Oct 14 '20 at 01:46
  • I have updated my former way-too-short answer with references on the most recent records in size and density (non commercial yet), to the latest (August 2020) Fujifilm announcement at 400TB and 224Gbit/in² density – Laurent Duval Dec 04 '20 at 11:02
  • @JoelFan do you mean more distorsion than on a phone line (56K faxmodem)? – user2284570 Jun 11 '22 at 14:38
  • @user2284570, not sure how to compare them. Tape has inches and modems do not. – JoelFan Jun 13 '22 at 19:09
  • @JoelFan I was speaking about the signal in terms of distorsion of course… – user2284570 Jun 14 '22 at 12:00

6 Answers6

47

I had heard that tape is still the best medium for storing large amounts of data.

well, "best" is always a reduction to a single set of optimization parameters (e.g. cost per bit, durability, ...) and isn't ever "universally true".

I can see, for example, that "large" is already a relative term, and for a small office, the optimum solution for backing up "large" amounts of data is a simple hard drive, or a hard drive array.

For a company, backup tapes might be better, depending on how often they need their data back. (Tapes are inherently pretty slow and can't be accessed at "random" points)

So I figured I can store a relatively large amount of data on a cassette tape.

Uh, you might be thinking of a Music Casette, right? Although that's magnetic tape, too, it's definitely not the same tape your first sentence referred to: It's meant to store an analog audio signal with low audible distortion for playback in a least-cost cassette player, not for digital data with low probability of bit error in a computer system.

Also, Music Cassettes are a technology from 1963 (small updates afterwards). Trying to use them for the amounts of data modern computers (even arduinos) deal with sounds like you're complaining your ox cart doesn't do 100 km/h on the autobahn.

But after reading up about it for a bit it turns out that they can store very small amounts of data. With baud rates varying between 300 to 2400 something between ~200KB to ~1.5MB can be stored on a 90 minute (2x45min) standard cassette tape.

Well, so that's a lot of data for when music-cassette-style things were last used with computers (the 1980s).

Also, where do these data rates drop from? That sounds like you're basing your analysis on 1980's technology.

These guys can store 90 minutes of audio. Even if we assume the analog audio quality on them was equivalent of 32Kbps that's about 21MB of data.

32 kb/s of what, exactly? If I play an Opus Voice, Opus Music or MPEG 4 AAC-HE file with a target bitrate of 32 kb/s next to the average audio cassette, I'm not sure the cassette will stand much of a chance, unless you want the "warm audio distortion" that casettes bring – but that's not anything you want to transport digital data.

You must be very careful here, because audio cassette formulations are optimized for specific audio properties. That means your "perceptive" quality has little to do with the "digital data capacity".

I have a hard time believing what I listened to was 300bps quality audio.

again, you're comparing apples to oranges. Just because someone 40 to 50 years ago wrote a 300 bits per second modem that could reconstruct binary data from audio cassette-stored analog signals, doesn't mean 300 bps is the capacity of the music cassette channel.

That's like saying "my Yorkshire Terrier can run 12 km/h on this racetrack, therefore I can't believe you can't have Formula 1 cars doing 350 km/h on it".

I read about the Kansas City standard and I can't understand why the maximum frequency they're using is 4800Hz yielding a 2400 baud. Tape (according to my internet search) can go up to 15KHz. Why not use 10KHz frequency and achieve higher bauds?

Complexity, and low quality of implementation and tapes. I mean, you're literally trying to argue that what was possible in 1975 is representative for what is possible today. That's 45 years in the past, they didn't come anywhere near theoretical limits.

Why do all FSK modulations assign a frequency spacing equal to baud rate?

They don't. Some do. Most modern FSK modulations don't (they're minimum shift keying standards, instead, where you choose the spacing to be half the symbol rate).

In the Kansas example they are using 4800Hz and 2400Hz signals for '1' and '0' bits. In MFSK-16 spacing is equal to baud rate as well.

Again, 1975 != all things possible today.

Why don't they use a MFSK system with a 256-element alphabet? With 20Hz space between each frequency the required bandwidth would be ~5KHZ. We have 10KHz in cassette tape so that should be plenty. Now even if all our symbols were the slowest one (5KHz) we would have 5*8 = 40000 baud. That's 27MB of data. Not too far from the 21MB estimation above.

Well, it's not that simple, because your system isn't free from noise and distortion, but as before:

  1. Low cost.

They simply didn't.

If tape is so bad then how do they store Terabaytes on it?

You're comparing completely different types of tapes, and tape drives:

This 100€ LTO-8 data backup tape

LTO-8 tape

vs this cassette tape type, of which child me remembers buying 5-packs at the supermarket for 9.99 DM, which, given retail overhead, probably means the individual cassette was in the < 1 DM range for business customers:

Audio Cassette

and this 2500€ tape drive stuffed with bleeding edge technology and a metric farkton of error-correction code and other fancy digital technology

LTO-8 tape drive

vs this 9€ casette thing that is a 1990's least-cost design using components available since the 1970s, which is actually currently being cleared from Conrad's stock because it's so obsolete:

Cassette player

At the end of the 1980s, digital audio became the "obvious next thing", and that was the time the DAT cassette was born, optimized for digital audio storage:

DAT

These things, with pretty "old-schooley" technology (by 2020 standards) do 1.3 Gb/s when used as data cassettes (that technology was called DDS but soon parted from the audio recording standards). Anyway, that already totally breaks with the operating principles of the analog audio cassette as you're working with:

  • in the audio cassette, the read head is fixed, and thus, the bandwidth of the signal is naturally limited by the product of spatial resolution of the magnetic material and the head and the tape speed. There's electronic limits to the first factor, and very mechanical ones to the second (can't shoot a delicate tape at supersonic speeds through a machine standing in your living room that's still affordable, can you).
  • in DAT, the reading head is mounted on a rotating drum, mounted at a slant to the tape – that way, the speed of the head relative to the tape can be greatly increased, and thus, you get more data onto the same length of tape, at very moderate tape speeds (audio cassete: ~47 mm/s, DAT: ~9 mm/s)
  • DAT is a digital format by design. This means zero focus was put into making the amplitude response "sound nice despite all imperfections"; instead, extensive error correction was applied (if one is to believe this source, concatenated Reed-Solomon codes of an overall rate of 0.71) and 8b-10b line coding (incorporating further overhead, that should put us at an effective rate of 0.5).

Note how they do line coding on the medium: This is bits-to-tape, directly. Clearly, this leaves room for capacity increases, if one was to use the tape as the analog medium it actually is, and combined that ability with the density-enabling diagonal recording, to use the tape more like an analog noisy channel (and a slightly nonlinear at that) than a perfect 0/1 storage.

Then, you'd not need the 8b-10b line coding. Also, while re-designing the storage, you'd drop the concatenated RS channel code (that's an interesting choice, sadly I couldn't find anything on why they chose to concatenate two RS codes) and directly go for much larger codes – since a tape isn't random access, an LDPC code (a typically 10000s of bits large code) would probably be the modern choice. You'd incorporate neighbor-interference cancellation and pilots to track system changes during playback.

In essence, you'd build something that is closer to a modern hard drive on a different substrate than it would be to an audio cassette; and lo and behold, suddenly you have a very complex device that doesn't resemble your old-timey audio cassette player at all, but a the modern backup tape drive like I've linked to above.

Marcus Müller
  • 30,525
  • 4
  • 34
  • 58
  • 5
    Nice flashback Marcus! – Dan Boschen Oct 05 '20 at 12:01
  • Thank you that was very helpful. So there's no theoretical obstacle in using for instance 10KHz FSK to store data on tape? As you guessed correctly most of my info comes from 80s era technology because that was the last time people tried to store stuff on tape haha! I could only find this and this modern examples and also amateur radio modulations (MFSK and RTTY) which all use low baud rates (300-1200) even tho they are using modern computers. – Pouria P Oct 05 '20 at 12:24
  • 1
    @DanBoschen thanks! All this depreciation for the music cassette I'm giving, I'm still kind of curious (but not curious enough to start working on this myself) what the achievable rates through an audio cassette tape would be using modern methods; it's a pretty nonlinear channel, I don't even think we have definite methods on how to calculate its theoretical capacity, but it's pretty safe to say that a nonlinear equalizer, large, strongly interleaved codes and a lot of classical comms methods of the multicarrier variety will be necessary to come close to anything that resembles a limit. – Marcus Müller Oct 05 '20 at 12:25
  • @PouriaP I don't know your tape, but I'm almost certain your tape deck won't be great for 10 kHz of spacing. Anyway, since this isn't amateurs in the 1980s, you simply won't do FSK on this definitely frequency-selective channel. I can imagine that precoded OFDM might be way to deal with that and a portion of the nonlinearity. My larger-scale feeling is that you'd want a compensation for the nonlinearity, and a lot of timing recovery to deal with the less-than-stellar constantness of tape speed, before even thinking about how to encode data. – Marcus Müller Oct 05 '20 at 12:27
  • @PouriaP also note, as I said, just because you can do shorter symbols, doesn't mean you should: Your SNR and distortion won't necessarily allow for symbol durations close to bandwidth limits. Modem design has more objective functions than "symbol rate"! – Marcus Müller Oct 05 '20 at 12:29
  • @MarcusMüller I hear you, that is actually quite interesting. (And I agree, very interesting to see the results, but not enough to spend the time creating it). Ultimately it is a Shannon Capacity problem if we had the data on maximum achievable SNR and bandwidth, right? – Dan Boschen Oct 05 '20 at 13:14
  • @DanBoschen well, Shannon wrote the noisy channel coding theorem, but he didn't write the nonlinear noisy channel coding theorem. For general nonlinear channels, there's no known formula for the capacity! – Marcus Müller Oct 05 '20 at 13:15
  • @MarcusMüller I would think to a first order the actual non-linear aspects could be incorporated into the maximum achievable SNR (just like ENOB in an ADC) and it's the compensation for the linear channel distortions (such as frequency selectivity) that would come into play first-- similar to techniques for high bandwidth transmission over traditional analog channels (such as ADSL for phone-lines)...just thoughts – Dan Boschen Oct 05 '20 at 13:20
  • 1
    @DanBoschen that's probably true, given appropriate signal shape (e.g. compare the effect of nonlinearities on a FSK vs OFDM), making the influence of the signal on the SINAD negligible or constant. But it's really a pretty interesting problem and it bears a lot of interesting sub-problems. When higher symbol power means worse error probability, a good part of our decision theory breaks down pretty dramatically :) Problem from fiber optical comms: think of a 256 QAM. The outer constellation points have more power, hence see more nonlinearity, and hence get "phase-smeared" more. – Marcus Müller Oct 05 '20 at 14:08
  • 1
    @DanBoschen in that situation, you can either reduce the number of points with higher power (to tolerate the smearing) or not use them at all; both are highly unattractive options! A better (in terms of mutual information) alternative is to do probabilistic shaping (i.e. picking these problematic points less often) with the objective of maximizing the information throughput. This gets arbitrarily interesting, mathematically – Marcus Müller Oct 05 '20 at 14:13
  • 1
    Optimistically assuming 20 kHz of bandwidth and 4 bits per symbol, I'm willing to bet that 160 kb/s is an upper bound. I'd also bet that this bound is loose :) – MBaz Oct 05 '20 at 14:26
  • 2
    @MBaz SERIOUSLY! folks, I don't have time to advertise that as a student project! WHY ARE YOU DOING THIS TO ME?! Why can't I simply not throw all we have in our toolbox at this?! – Marcus Müller Oct 05 '20 at 14:42
  • @MarcusMüller I agree this is a great problem to throw everything in your toolbox at: very simple to state, very hard to solve analytically, but very easy to obtain experimental results. It's also easy to make a back-of-the-envelope estimate of an upper bound, and I just didn't resist the temptation! – MBaz Oct 05 '20 at 14:55
  • 1
    What a fun design contest it would be, especially encouraged by the gauntlet thrown down by @MBaz with an upper bound (although I absolutely have no time for such a fruitless quest). But why would it be different from ADSL achieving 24 MBps on hardware and channels designed for the audio BW? Plus consider the audio quality on a phone line compared to high fidelity cassette tapes--- Just from that thinking alone I am guessing we could easily reach rates of 100 MBps--- but this is data rate while the original question was storage capacity; what's the relationship between the two? – Dan Boschen Oct 05 '20 at 14:55
  • 1
    @DanBoschen Yes: I'd expect the tape speed to be an important parameter in the solution. A very fun project indeed! – MBaz Oct 05 '20 at 14:57
  • 2
    @DanBoschen doesn't ADSL use baseband up to 2.2 MHz? – Marcus Müller Oct 05 '20 at 14:57
  • 1
    (Plus we know the achievable SNR is in the audio band is in excess of 60 dB, so there would be much more than 4 bits per symbol!) – Dan Boschen Oct 05 '20 at 14:57
  • @DanBoschen But you still need to translate that SNR into Eb/N0 -- you'll get nowhere near 60 dB. – MBaz Oct 05 '20 at 14:58
  • yeah, I'm pretty sure that with a hefty PSK and a hefty equalizer, we could achieve 32-PSK or more on 5 kbd, but I don't want to be the one to write that equalizer, and I don't want to be the one who corrects the fun we'll have with the intermodulation in a multicarrier system, either. – Marcus Müller Oct 05 '20 at 14:59
  • (OK 9 MBps "on standard copper phone lines") – Dan Boschen Oct 05 '20 at 15:02
  • @DanBoschen by the way, ways in which things would be different than DSL: · unreliable tape speed (I can deal with varying carrier frequency. We have ways of CFO. But varying tape speed? Oof. Don't like to think about how to compensate that; it means a linked symbol timing variation, frequency selectivity, and is kinda cyclostationary – if we can put that property to use at all) · magnetic hysteresis and things give that channel memory. And not the LTI kind of memory. · It's one-way; the transmitter can't pre-emphasize for the specific channel, but has to assume a "typical" cassette+player. – Marcus Müller Oct 05 '20 at 15:17
  • 2
    @MarcusMüller Marcus, These photographic answers remind me of the electronics.se style ! ;-) – Fat32 Oct 05 '20 at 15:41
  • 1
    @MarcusMüller Are we restricted to use the existing mechanics for the companion equipment in this mental challenge that we aren't ever going to actually pursue? If this is just about the physical tape, I think the modern reader/writer that we create could address a lot of those physical restrictions regarding tape speed limitations and variation, as well as magnetic read/write density and accuracy. But with existing equipment alone inc storing in analog w/o hardware compression, then I would say use V.92 at 56 kbps * 90 minutes = 302.4 Mbits total storage capacity with no further R&D req'd. – Dan Boschen Oct 05 '20 at 16:10
  • (With upside of 48.6 Gbits if the resulting ADSL audio signal can equally be recorded / reconstructed from an audio tape.... V.92 is obvious to me- but ADSL not so much) So in summary somewhere in the range of 37.8 MBytes to 6 GBytes with existing equipment. – Dan Boschen Oct 05 '20 at 16:13
  • 1
    @DanBoschen that does sound feasible, but one would have go through the link training phases "offline", which would at least be a modification of one of the softmodems there are. – Marcus Müller Oct 05 '20 at 16:14
  • And Hotpaws good comment below that he couldn't even get 1200 baud---so to your point, modern equalization techniques as well would be needed. – Dan Boschen Oct 05 '20 at 16:16
  • @DanBoschen technical side fact: just looked this up (happens that I just figured out I have 0 analog cassette player in my flat, and 1 DAT tape player) that audio tape optimized for digital storage had 1.3 GB of space – Marcus Müller Oct 05 '20 at 16:23
  • @MarcusMüller Right in the middle of my highly narrowed prediction! You should add that detail (and link) to your answer! – Dan Boschen Oct 05 '20 at 16:25
  • @MarcusMüller 1975 was 45 years ago, not 35. I'd edit but it's too small a change – BThompson Oct 05 '20 at 20:07
  • 1
    @BThompson arithmetics is hard. Thank you! – Marcus Müller Oct 05 '20 at 20:10
  • A (90 minute) compact cassette has ~0.4m^2 of 300 Oe media. A LTO-6 tape has ~10m^2 of 2900 Oe media. – BowlOfRed Oct 06 '20 at 05:37
  • Yesterday I thought I had an answer. And I was happy with my answer. 24 hours later and I have 25 comments I don't understand anything of, a couple other answers that are rather contradictory, and 4 new badges. I have to say this escalated quickly. – Pouria P Oct 06 '20 at 19:37
  • hey @PouriaP what do you not understand? (afaik) None of the comments here contradict my answer; I say "you cannot infer the capacity of some medium by what someone in 1975 did with it; also, music cassette is not the same as backup tape", and the comments are just about "how can we push a little more data into music tape"; none of this changes that even doing the best we can do these days, music tape is not the same as a high-density backup tape. And: no matter what you do with it, you'll never put as much data on it as on proper data storage tapes, which are different. – Marcus Müller Oct 06 '20 at 22:36
  • 1
    I would assume (like others have stated) that if you can characterize the channel in terms of frequency-dependent SNR (or even cruder: a bandwidth and global SNR) and plug it into the relevant Shannon formulas, you would get a decent, if generous upper bound on the channel capacity assuming infinite effort in developing a modem. Wow and flutter would be a bit like doppler distortion, right? And saturation effects would be a bit like «class c» radio amplifiers? I guess that the channel impulse response would be reasonably compact? – Knut Inge Oct 07 '20 at 10:20
  • @KnutInge that's the thing! I don't know, the medium is nonlinear. Nonlinear channels can decrease the capacity (obvious, say if I square a signal, I get an ambiguity), but they can also increase it (think: the decode-and-forward channel vs the amplify-and-forward channel). The thing about tape heads is that they have memory, and not the kind of linear FIR that we know how to deal with, crosstalk is non-linear, and all that might either increase or decrease the actual capacity – and on the decrease side, quite significantly, I'd guess, but I'm having a hard time saying "the linear channel – Marcus Müller Oct 07 '20 at 10:36
  • … is a good approximation in terms of Shannon capacity". The nonlinear curves of their excitability might well increase H(data|head signal), but they might just as well decrease the equivocation. Can't tell you. – Marcus Müller Oct 07 '20 at 10:37
  • 2
    @MarcusMüller Excellent answer, but is "These things, [DAT/DDS] do 1.3 Gb/s" correct? The Wikipedia page you linked (under Computer data storage medium says the capacity ranged from 1.3GB to 80GB. At 1.3Gb/s, that should take about 10 seconds to fill the tape! Or have I missed something? – TripeHound Oct 07 '20 at 12:38
  • @TripeHound there's several generation of DDS and successor media. The Wikipedia DAT page conflates these. – Marcus Müller Oct 07 '20 at 12:54
  • @MarcusMüller I did totally understand your answer. I was referring to the fact that I know nothing of DSP and hence didn't understand the technical jargon in the comments between you guys. Which made me realize this is more difficult than I thought and I should just stick to available methods of converting data into audio (if I decided to go on with the project). – Pouria P Oct 07 '20 at 14:37
  • @PouriaP "available methods" is a ... broad term. You'll find that a lot of available audio modems have been developed since the days of 2400 bd. – Marcus Müller Oct 07 '20 at 15:05
  • Actually, cassettes have made a sort of comeback https://www.officialcharts.com/chart-news/the-uks-official-top-20-bestselling-cassette-albums-of-2020-so-far__30242/ – Olli Niemitalo Oct 08 '20 at 04:37
  • This source appears to have some (recent) measurements of the medium: https://audiochrome.blogspot.com/2019/12/cassette-tape-comparative-measurements.html? – Knut Inge Oct 08 '20 at 06:18
  • Interestingly, googling for «tape impulse response» typically returns quirky file formats for applying the (linear) distortion of good old studio tape machines to modern recordings. – Knut Inge Oct 08 '20 at 07:59
  • @KnutInge interesting, indeed! I don't know whether this direct link to the picture from their measurement works, but if you look at it, different input powers actually yield different frequency-selective channels (not just vertically shifted curves); not as bad as I would have presumed, still significant. Watch how more input power means that the high cutoff frequency decreases! – Marcus Müller Oct 08 '20 at 08:36
8

High-capacity tape formats use helical scanning. Here I look at why that gives a higher capacity than linear recording with four tracks (one stereo track on each side) like in a compact cassette (C-cassette).

enter image description here
Figure 1. Stereo C-cassette player/recorder linear head (left) and a helical scanning head from a data drive (right), same scale. There is a too-short-to-see gap between two polished surfaces of laminated magnetic cores that concentrate the magnetic fields to the inside of the coils and to a thin line on the coated surface of the tape. In case of linear scanning, the head gap is oriented straight across the tape.

enter image description here
Figure 2. A simplified illustration of the tracks on a four-track tape (left) and helical scanning tape (right). In reality there needs to be some space, "guard bands", between the linear scanning tracks to prevent leakage between the channels. In case of helical scanning this need is typically circumvented by having heads that read/write consecutive tracks in different gap angles. This way the magnetizations will be approximately orthogonal and not sensed by the wrong head.

It's not about the speed at which the recording head travels. It is possible to slow things down and record the same spatial wavelengths, only it would take a longer time to do that or to read the recording later. Recording faster does not increase tape capacity. What it does is it gives a higher data rate which is important for practical applications. Also, for analog recording, a high head travel speed may be needed because the analog signal has a high bandwidth. Very small spatial wavelengths are difficult to record because of the magnetic grain size on the tape and because of mechanical design considerations.

For information density, it doesn't matter what the angle of the tracks on the tape is. So for a simplified analysis we can rotate the helical scanning tracks to be horizontal, and we can look at just a 1/4 of the tape and consider that as a single-track tape in case of the C-cassette:

enter image description here
Figure 3. One-track linear recording tape (left) and an n-track linear recording tape (right).

Let's write a spatial version of the Shannon–Hartley theorem that suits our simplified analysis:

$$C = nB \log_2 \left( 1+\frac{S}{nN} \right)\tag{1}$$

Here $C$ is the total capacity (bits/m of tape), $n$ is the number of tracks, $B$ is the spatial bandwidth (1/m) and $S/N$ is the reference signal-to-noise ratio, the ratio between signal power and noise power in the reference case that there is just a single track. If we increase from the reference case the number of tracks by a factor $n$, then the track width decreases to $1/n$ times that in the reference case, giving an effective signal-to-noise ratio of $\frac{S}{nN}$ for each track. This comes from that the signal power gets divided by a factor $n^2$ and the noise power gets divided by a factor $n$. This is because each of the $n$ individual channels has the same signal power as would each of $n$ identical-signal channels that sum to the single full-width track, and because the noise is independent between the $n$ tracks so the power of the noise in the full-width track equals the sum of the powers of the noises in the $n$ tracks. So the signal-to-noise ratio goes down as the number of tracks is increased. But at the same time, the total capacity will be the sum of the individual track capacities, hence the factor $n$ in front of the formula.

If we plot the capacity calculated by Eq. 1 as function of the number of channels, it would appear that we can increase the capacity indefinitely by just increasing the number of tracks:

enter image description here
Figure 4. Proportional total capacity as function of the number of tracks, for an out-of-the-hat signal-to-noise ratio of 50 dB in case of a single track. Calculated using Eq. 1.

The main thing that is wrong with this analysis is that the size of the magnetic grains is not truthfully infinitesimally small, so there will be all kinds of unaccounted for trouble (correlated noise between tracks, leakage, and noise being far from Gaussian assumed by Eq. 1) if the tracks become too narrow and densely spaced. But the result still holds up to some limit, that more and narrower tracks can store more information even when each has a lower signal-to-noise ratio than would a single, or four, tracks. So it is not just that we wouldn't have come up with the right modulation scheme for C-cassettes. The design of data tape formats with helical scanning really is superior in terms of information capacity expressed per tape length, for the same width of tape.

John C. Mallinson, The Foundations of Magnetic Recording, 2nd edition 1993, Academic Press, USA, p. 129, gives a result identical to Eq. 1, and also postulates a limit:

9.8 Ultimate Information Areal Capacity

Consider a tape of unit width, divided, without guardbands between the tracks, into $M$ parallel tracks as shown in Figure 9.9. Clearly, the combined Shannon areal capacity is

$$C(M) = MB\log_2\left(1 + \frac{(\text{SNR})_w}{M}\right),\tag{9.13}$$

where $(\text{SNR})_w$ is the wide-band signal-to-noise ratio when the full tape width is used for a single track ($M = 1$). For unit head–tape relative speed, the bandwidth $B$ is the reciprocal of the minimum wavelength, $\lambda_\text{min}$. It is obvious from this expression that the areal capacity increases as $M$ increases. The ultimate storage capacity, in bits per unit area, occurs when the number of tracks becomes very large, and it is

$$\text{ultimate areal capacity} = \frac{nf^2\lambda_\text{min}}{2\pi\log_2e}.\tag{9.14}$$

For the instrumentation recorders considered in this chapter ($n=10^{15}$ particles per cubic centimeter, $f = 0.2$, and $\lambda_\text{min} = 60 \mu\text{in.}$), the ultimate capacity is, amazingly enough, $10^9$ bits per square centimeter or 6000 megabits per square inch. This is an extremely high areal density in comparison with present achievements in digital recording, where figures in the range of 50–150 megabits per square inch are the norm.

Olli Niemitalo
  • 13,491
  • 1
  • 33
  • 61
5

You have already selected your answer, but I would like to put a few more lines.

First of all, objecting to Marcus, I think that your first assumption is correct! We can store one hour of almost CD quality (14-16-bits, 44100Hz, stereo) analog audio into these commercial music casettes; i.e magnetic reels. So on a very rough basis you have almost a 600 MB (Mega Bytes) of data capacity there??

Of course, magnetic mediums have different uses, types and qualities. And I think, one thing to mention at the beginning is that, the commercial casette have a tendency of losing its quality. This is especially true with music recording, where high-fidelity is lost after a few playbacks. Such an untrustable characteristic makes them a bad choice for reliable data storage at high rates. Hence possibly lowering their available capacity, in exchange of increased reliability and durability.

But the oppsite is also true: for example, IBM has developed the hard disc drive technology (very high density digital storage) based on magnetic recording too. Their circular shaped magnetic plates make it possible for random access with the help of a moving read-write head mechanism. Packing together a dozen of such magnetic disks and very sensitively controlling the head mechanism in between the plates, provided one of the most important and successful inventions of the computer industry. Even today HDD dominates, and SSD is still expensive, or Cloud is very unconvenient.

You have to be careful about the medium's physical characteristics, and which way you want to store data. The commercial audio cassette system (magnetic tape, and read-write circuitry together) had a bandwidth $B$ of about less than 20 kHz. And I guess its SNR is about less than 60 dB. This can be an under or over estimation, but if you consider what I've said at the beginning; CD-quality audio, now to be honest, commercial audio cassettes do not provide CD-quality audio, (low-noise metal cassettes with Dolby Type-S noise reduction is said to have achieved a CD quality though), as there is some characteristic high frequency hiss noise in those commercial cassettes, especially on the lower quality tapes. Therefore I will take 60 dB as an average estimate. Also note that this SNR depends on the frequency, as the noise increase by frequency, but I will ignore it too.

The (Shannon) capacity $C$ (in bits per second) of the medium is therefore: $$ 60 = SNR_{dB} = 10 \text{log}_{10}( snr ) \implies snr = 10^6 $$ $$ C = B \times \text{log}_2( snr +1) = 20 k \times 19.9 \approx 400 ~ kbps $$

And assuming a two channel stereo recording this yields about $800 ~kbps$ channel capacity. Indeed, considering that CD had a much better SNR about 96dB, it has a channel capcity of $1.411$ Mbps. Then the commercial audio casette had an equivalent data rate about half of the CD medium. In one hour this makes about $360$ MB (Mega Bytes) of data , as an optimistic upperbound, as the comments indicated.

The capacity is there, but can you utilize it?

The audio CD, for example, uses a very small tiny pattern on the circular track lines. The bits are placed as very short and very thin reflective vs non-reflective sections along the track lines that spiral from the innermost circle to the outermost, just like in Vinyl. There is no noise, but distortion due to optical reflection, diffraction, refraction, and mechanical disturbances, which limit the resolution at which you can write or read those tiny dots on the circular tracks.

But magnetic medium is totally different. You store an electric signal (a magnetic one) into it. And you have noise. But you have an advantage over the CD. The CD medium has only one amplitude level indicated by a reflection or no reflection. This is due to the specific mechanism used to write and read the bits. But the magnetic medium can provide multiple bits per symbol to be written, as multiple amplitude levels can (in principle) be supported.

This kind of M-ary encoding is very well utilized in Satellite communication, where 256-QAM is used to take advantage of (relativlely) low noise transmission channel.

Coming to that Kansas City Standards, it's not about the magnetic medium capacity, but about the commercial phone line allowed channel capacity which was quite limited to about $3-4$ kHz back then...

So the conclusion is; if you can afford a suitable encoding technique, in principle, you can store about less than 360 MB of data into a standard commercial auido cassette. There are low noise (high SNR) casette types (such as Cr or Metal) available which would increase the capacity. Or at least you can roll a longer reel to increase it. Whether you can accomplish this or not is another practical concern though...

Fat32
  • 28,152
  • 3
  • 24
  • 50
  • 1
    nice one! by the way, where does the 64 dB SNR come from? That seems very high! – Marcus Müller Oct 05 '20 at 17:07
  • 2
    @MarcusMüller It comes from my ears at the moment ! :-)) (a comparison to CD music and casette music) I though the CD had 96 dB (16 bit assumption! though they are about 14 bits in reality) dynamic range then I assumed the cassette could provide 60 dB at most? I'm sure such data is available (at least Dolby Noise Reduction technologies makes use of them). If you know a better SNR estimate, please feel free to edit the numbers :-) – Fat32 Oct 05 '20 at 17:10
  • sadly, I don't; there's a couple of hifi forum posts that give the same numbers, but I'm not sure where they've got that from :) – Marcus Müller Oct 05 '20 at 17:48
  • 2
    @MarcusMüller Probably from this : It sounds certainly not as good as 90 dB, and it sounds certainly not as bad as 30 dB, therefore it should probably be 60 dB :-)) – Fat32 Oct 05 '20 at 18:04
  • 2
    I'm pretty sure this answer is wildly optimistic about the capacity. Stereo on audio tape is not a fully distinct signal (stereo tracks are close together, but the signal is similar so interference is not that bad). SNR rapidly drops above 1 kHz; tricks like Dolby B are needed to make it sound acceptable. It definitely does not reach 20 kHz either. As a result, you can't multiply bandwidth by SNR. You must integrate SNR over the available bandwidth to get Shannon capacity. And to utilize this, you need ADSL-like subcarriers. – MSalters Oct 06 '20 at 09:08
  • 2
    @MSalters Yes it's an optimistic upper bound. I wish I knew the technical specs of the magnetic tape and read-write circuitry to provide a better answer. But I already mention the hiss noise, and frequency dependent SNR. Regarding the bandwidth, I guess it's somewhere between 10 to 16 kHz. Regarding the stereo; I didn't know tat cross-talk was so severe? But it can be undone. Remeber that once phone lines had dial-up modems but ADSL utilized unused bandwidth sitting idle in the copper. Or 4K re-scanning of (clean) optical films from 50's. So the capacity utilisation depends on technology – Fat32 Oct 06 '20 at 11:56
  • 2
    OK from all the technical jargon in the comments I'm starting to feel I'm in here way over my head. But your answer has given me new hope. Is it by any chance possible to convert my data into dial-up screams (that noise dial-up made) and record that in a tape? It's 56kbps so that's like 37MB in 90 minutes. Right? And I suppose it has built-in error checking and stuff in it as well. Not as interesting as writing my own modulation but from the looks of it I am utterly unfit for that task. – Pouria P Oct 06 '20 at 19:47
  • @PouriaP The dial-up modem used about 3-4 kHz of transmission bandwidth that was allocated to commercial phone-line at that time. And reaching about 56 kbps indicates that the line had about 50-60 dB SNR (Signal to Noise Ratio). But it was using some advanced channel coding algorithms (together with Zip compression on the fly). I'm not sure if the scream you mention is a faithful representation of what's electrically passing through the line. If it's, then you can for sure use it. Why dont you use 2 layer DVD-Rs ? They are definetely high density, convenient and cheap? Why cassette? – Fat32 Oct 06 '20 at 22:11
  • 2
    @Fat32 I was gonna use cassettes purely for the retro feel that they have. Thanks to your answers I now realize coming up with a new technique/modulation is more difficult than I thought. If I ended up doing it I will stick to available audio frequency shift keying techniques. – Pouria P Oct 07 '20 at 14:27
  • 1
    @PouriaP Ok... Don't forget to use zipping before storing digital bits... That will save you a lot of space on the limited resource you have... ;-) – Fat32 Oct 07 '20 at 14:35
4

Typical cheap cassette tape recorders and players in the 1970's used audio circuitry that did not have a completely flat frequency response and added a lot of phase distortion (mostly inaudible to most consumers). In addition, the computers did not use high sample rate low quantization DACS and ADCs (which cost a fortune back then), sometimes only a simple 1-bit digital output and comparator input between the PC and tape drive, or similar low complexity analog circuitry. DSP processing power was limited to in the neighborhood of 1 to 5 kiloflops.

Thus the older hobbiest/consumer computer tape modulation formats were designed to be reasonably robust for users of the poor signal channels designed into those cheap cassette tape recorders, and for slow DSP engines. Even so, I had a few cheap tape recorders that distorted the audio signal too much to reliably record and reproduce data at 1200 baud into a 2 MHz Z80 system.

The signal channels in those cabinet sized IBM 2400 series 9 track vacuum reel-to-reel tape drives were designed from the ground up for the bandwidths required. Not just for consumer audio.

hotpaw2
  • 35,346
  • 9
  • 47
  • 90
  • Usually only a one bit input and output. I think the Tandy Color Computer had a 8-bit DAC that was fed to the cassette output and a comparator input, and could either be fed to the speaker or muted, and could be used as a configurable reference when reading data from cassette, but I don't think any other machines of that era had anything beyond one bit input and output. – supercat Oct 08 '20 at 02:05
1

Despite the title of the question, a 2018 IEEE Spectrum article explains: Why the Future of Data Storage is (Still) Magnetic Tape Disk drives are reaching their limits, but magnetic tape just gets better and better. Indeed, capacity records are regularly broken, one of the latest being published in 201 Gb/in2 Recording Areal Density on Sputtered Magnetic Tape. From the initial 2 megabytes stored on tapes (IBM first commercial tape product: 726 Magnetic Tape Unit, announced more than 60 years ago), the recent announcements (IBM and Sony stuff 330 TB on a tiny tape) over 10+ years are summarized below:

  • 2006: size 8 TB; density 6.67 billion bytes per square inch
  • 2010: size 35 TB; density 29.5 billion bytes per square inch
  • 2014: size 154 TB; density 85.9 billion bytes per square inch
  • 2015: size 220 TB; density 123 billion bytes per square inch
  • 2017: size 330 TB; density 201 billion bytes per square inch

Those are still research, non commercial, with different media technologies: sputtered tape versus Barium ferrite (BaFe) for commercial tapes up to 15 terabytes. In August 2020, Fujifilm point[ed] to 400TB tape cartridge on the horizon, with a density of 224Gbit/in² using Strontium Ferrite (SrFe). Such performance can be improved further by deduplication and lossless compression (the latter being often disappointing in practice, with less than 2:1 ratios). Far future competitors may be DNA storage or cristal etching (see Will Magnetic Storage Ever Become Completely Obsolete?).

In my workplace, with huge datasets, tapes are still in use for long term backup. Namely, one often uses Data8, a 8 mm Backup Format pionnered by Exabyte corporation.

Some more history on tape formats: Much Ado About Exabyte/8mm Tape Drives.

Laurent Duval
  • 31,850
  • 3
  • 33
  • 101
0

Because the appliances like Betamax Spool at synchronized time frames.

lennon310
  • 3,590
  • 19
  • 24
  • 27
Matt
  • 1
  • 1