/* ---- Google Analytics Code Below */

Saturday, October 03, 2020

Life of a Data Byte

Thoughtful look at data as beyond just the amount, but also the means of storage and use.   And  it links to the  history of our business.

Volume 18, issue 3, In ACM Queue

The Life of a Data Byte  Be kind and rewind.   By Jessie Frazelle

A byte of data has been stored in a number of different ways through the years as newer, better, and faster storage media are introduced. A byte is a unit of digital information that most commonly refers to eight bits. A bit is a unit of information that can be expressed as 0 or 1, representing a logical state. Let's take a brief walk down memory lane to learn about the origins of bits and bytes.

Going back in time to Babbage's Analytical Engine, you can see that a bit was stored as the position of a mechanical gear or lever. In the case of paper cards, a bit was stored as the presence or absence of a hole in the card at a specific place. For magnetic storage devices, such as tapes and disks, a bit is represented by the polarity of a certain area of the magnetic film. In modern DRAM (dynamic random-access memory), a bit is often represented as two levels of electrical charge stored in a capacitor, a device that stores electrical energy in an electric field. (In the early 1960s, the paper cards used to input programs for IBM mainframes were known as Hollerith cards, named after their inventor, Herman Hollerith from the Tabulating Machines Company—which through numerous mergers is what is now known as IBM.)

In June 956, Werner Buchholz (archive.computerhistory.org) coined the word byte (archive.org) to refer to a group of bits used to encode a single character of text (bobbemer.com). Let's address character encoding, starting with ASCII (American Standard Code for Information Interchange). ASCII was based on the English alphabet; therefore, every letter, digit, and symbol (a-z, A-Z, 0-9, +, -, /, ", !, etc.) was represented as a seven-bit integer between 32 and 127. This wasn't very friendly to other languages. To support other languages, Unicode extended ASCII so that each character is represented as a code-point. For example, a lowercase j is U+006A, where U stands for Unicode followed by a hexadecimal number.  ... " 

No comments: