subject = Information Theory title = Data Compression papers = Data Compression- in beginners' terms 'Data Compression' just sounds complicated. Don't be afraid, compression is our good friend for many reasons. It saves hard drive space. It makes data files to handle. It also cuts those immense file download times from the Internet. Wouldn't it be nice if we could compress all files down to just a few bytes? There is a limit to how much you can compress a file.

How random the file is, is the determining factor to how far it can be compressed. If the file is completely random and no pattern can be found, then the shortest representation of the file is the file it self. The actual proof that proves this is at the end of my paper. The key to compressing a file is to find some sort of exploitable pattern. Most of this paper will be explaining those patterns that are commonly used.

Null suppression is the most primitive form of data compression that I could find. Basically, it says that if you have different fields that data is in (possibly a spread sheet), and any of them have only zeros in them, then the program just eliminates the data and goes straight from the empty data set to the next. Only one step up from null suppression is Run Length Encoding. Run length encoding simply tells you how many of what you have in a row. It would change a set of binary data like {0011100001} into what the computer reads as (2)zeros, (3)ones, (4)zeros, 1. As you can see, it works on the same basic idea of finding a series of 0's (null suppression) and 1's in this case too and abbreviating them.

Once the whole idea of data compression caught on, more people started working on programs for it. From these people we got some new premises to work with. Substitutional encoding is a big one. It was invented jointly by two people: Abraham Lempel and Jakob Ziv. Most compression algorithms (big word meaning roughly 'program') using substitutional encoding start with 'LZ' for Lempel-Ziv.

LZ-77 is a really neat compression in which the program starts off just copying the source file over to the new target file, but when it recognizes a phrase of data that it has previously written, it replaces the second set of data in the target file with directions on how to get to the first occurrence of it and copy it in the directions' place. This is more commonly called a sliding-window compression because the focus of the program is always sliding all around the file. LZ-78 is the compression that most people have in their homes. Some of the more common ones are ZIP, LHA, ARJ, ZOO, and GZIP. The main idea behind LZ-78 is a 'dictionary'. Yet it works quite a bit like the LZ-77.

For every phrase it comes across, it indexes the string by a number and writes it in a 'dictionary'. When the program comes across the same string, it uses the associated number in the 'dictionary' instead of the string. The 'dictionary' is then written along side the compressed file to be used in decoding. There is a combined version of LZ-77 an LZ-78. It is called LZFG.

It only writes to the dictionary when it finds the repeated phrase, not on every phrase. Then instead of LZFG replacing the second set of data with directions on how to get to the first occurrence of it, the program puts in the number reference for the dictionary's translation. Not only is it faster, but it compresses better because of the fact that it doesn't have as big of a dictionary attached. Statistical encoding is another one of the new compression concepts. It is an offshoot of the LZ family of compressors; It uses basically the same style as LZFG, but instead of assigning the numbers in order that the strings come out of the source file, statistical compressors do some research.

It calculates the number of times each string is used and then ranks the string with the most number of uses at the top of the hash table. The string with the least is ranked at the bottom. (A hash table is where the rank is figured) The higher up a string is on this list, the smaller of a reference number it gets to minimize the total bit usage. This gives this compression just a slight edge on the others, but every little bit helps.

(ha ha -bit- ) Beware! There are a few compression programs out there that claim wonderful compression ratios; ratios that beat the compression limit for that file's randomness. These programs aren't really compression programs. They are OWS and WIC. Never compress anything with these. What they do is split up the file that you desired to compress and hide most of it on another part of your hard drive.

OWS puts it in a specific spot on the physical hard drive disk. WIC puts the extra information in a hidden file called winfile.dll. The real problems with these programs are that if you don't have the winfile.dll or the information on the certain spot on your drive, then the program won't put your file back together.

My original intent with this project was to invent a new compression algorithm. I started with the idea that if you took the file in its pure binary form and laid it out in a matrix, there were certain rows and columns that you could add up to get an output that would be able to recreate the original matrix. I was close too. I had four different outputs that actually would be what would make up the compressed file that combined together to create one output for each bit.

From this single output I could determine if the bit was 1 or 0. It worked perfectly for matrixes of 1x1, 2x2, and 3x3. Except that with this small of a matrix, I wasn't compressing it at all. It was more of a coding system that took up more space than did the original file.

I even found a way to shrink the size of the four outputs but it was not enough to even break even on bit count. When I got to the 4x4's I found an overlap. An overlap is a term I made up for this algorithm. It means that I got the same single output for a 1 as I did a 0. When that happens, I can't figure out which it is: a 1 or 0. When you can't recreate the original file, data compression has failed.

It becomes lossy. I needed a fifth original output. If you want more information on how I thought the algorithm would have worked, please refer to my Inventor's Log that I included. It's way too much to re-type here and it would serve no real purpose in this paper. If you were paying attention earlier, you would be saying, "Why don't you find a pattern? Otherwise you can't compress it. You are treating it like a random file.

" I didn't find out that it was impossible to compress random data until about the time my algorithm was failing. Because of my setbacks I started looking for an entirely new way to compress data, using a pattern of some sort. I got to thinking about all of the existing algorithms. I wanted to combine a hash table, statistical coder, and a run length coder. The only hard part that I would see in that would be trying to get the patent holders of each of those algorithms to allow me to combine them and actually modify them slightly.

In its current algorithm, the statistical coder only accepts alpha-numeric phrases. I would like to modify it to not read the characters that the binary code spells out, but the binary code it self. I don't know what form the output is aside from compressed, but for my purposes it wouldn't matter what the form the output is. I would program into the program all of the 32 combinations of 5 bits (2^5). Each of the combinations would be labeled in the program 1 through 32. I would then make a hash table of all of the 5 bit combinations.

This would give me an output which I would run through a run length coder. Since the entire algorithm is reliant on binary code and not on characters, it can be recursable, or it can compress further an already compressed file. LZ's can't do that because once they convert a string into it's dictionary/sliding window form, it ceases to be one of the characters that it compresses. Now that you are aquatinted with our friend, Data Compression, I hope he will be able to serve you better. Now you can download programs faster, save space, and who knows? Maybe you will invent the next new compression algorithm.

Until then, keep your mice happy and your monitors warm. Proof that random data is not compressible: Let's suppose the file to be compressed is 8 bits long (x works, but this is easier) and is random There are exactly 2^8 different possible 8 bit data strings. (2^x) To compress the file it must shrink it by at least one bit (2^x)-1 So there are at most 2^7 different compressed files 2^(x-1) Thus at least two source files compress down to the same file. Therefore The compression cannot be lossless.

Bibliography Aronson, Jules Data Compression- a comparison of Methods Washington D.C.: Institute for Computer Science and Technology http://simr02.si.ehu.es/DOCS/mice/compression-faq/part1/faq-doc-x.html x=Intro,8,9, and 10 http://www.epl.co.uk/genit18.htm http://www.sees.bangor.ac.uk/~gerry/sp_summary.html http://Literacy.com//mkp/pages/3468/index.html