So Jack I was wondering could you explain to me how this system of measurement of computing capacity works.
Like what is a gig, a terrabite, how much makes up one thing then another. When I hear people talk about their computors that part is always over my head. The last time I had a handle on it was with the comadore 64, which I never had one but......
Truman, a bit is a binary digit which equals 0 or 1 and is the basic unit of information storage. All of the other measurements are a reflection of how many of the basic units (bits) are available for storage or how many of these are taken up by a document, application, etc. The number of bits has been steadily increasing and what used to represent a lot of storage now represents very little. So here is a run down of how much of one things makes up another.
8 bits is a byte
1000 bytes is a kilobyte (KB)
1 million bytes is a megabyte (MB)
1 billion bytes is a gigabit (GB)
1 trillion bytes is a terabyte
1 quadrillion bytes is a petabyte
1 quintillion bytes is an exabyte
1 sextillion bytes is a zettabyte
1 septillion bytes is a yottabyte
Does that help?