Please create an account to participate in the Slashdot moderation system

 



Forgot your password?
typodupeerror
×

ZeoSync Makes Claim of Compression Breakthrough 989

dsb42 writes: "Reuters is reporting that ZeoSync has announced a breakthrough in data compression that allows for 100:1 lossless compression of random data. If this is true, our bandwidth problems just got a lot smaller (or our streaming video just became a lot clearer)..." This story has been submitted many times due to the astounding claims - Zeosync explicitly claims that they've superseded Claude Shannon's work. The "technical description" from their website is less than impressive. I think the odds of this being true are slim to none, but here you go, math majors and EE's - something to liven up your drab dull existence today. Update: 01/08 13:18 GMT by M : I should include a link to their press release.
This discussion has been archived. No new comments can be posted.

ZeoSync Makes Claim of Compression Breakthrough

Comments Filter:
  • Current ratio? (Score:2, Interesting)

    by L-Wave ( 515413 ) on Tuesday January 08, 2002 @09:13AM (#2803126)
    Exscuse my lack of compression knowledge, but whats the current ratio? Im assuming 100:1 is pretty damn good. =) btw...even though this *might* be a good compression algorithm and all that, how long would it take to decompress a file using your joe average computer??
  • Buzzwordtastic (Score:2, Interesting)

    by Steve Cox ( 207680 ) on Tuesday January 08, 2002 @09:20AM (#2803155)
    I got bored reading the press release after finding the fourth trademarked buzzword in the second paragraph.


    I simply can't believe that this method of compression/encoding is so new that it requires a completely new dictionary (of words we presumably are not allowed to use).

  • by Rentar ( 168939 ) on Tuesday January 08, 2002 @09:22AM (#2803165)
    This is a proof ('though I doubt it is a scientificly correct one), that you can't get lossless compression with a constant compression factor! What they claim would be theroretically possible if 100:1 where an average, but I still don't think this is possible.
  • Re:how can this be? (Score:3, Interesting)

    by Shimbo ( 100005 ) on Tuesday January 08, 2002 @09:27AM (#2803197)
    They don't claim they can compress TRUE random data only 'practically random' data. Now the digits of Pi are a good source of 'practically random' data for some definition of the phrase 'practically random'.
  • by Dada ( 31909 ) on Tuesday January 08, 2002 @09:33AM (#2803240)
    The maximum compression ratio for random data is 1. That's no compression at all.
  • by Quixote ( 154172 ) on Tuesday January 08, 2002 @09:34AM (#2803245) Homepage Journal
    Section 1.9 of the comp.compression FAQ [faqs.org] is good background reading on this stuff. In particular, read the "WEB story".
  • Re:how can this be? (Score:3, Interesting)

    by harlows_monkeys ( 106428 ) on Tuesday January 08, 2002 @09:36AM (#2803255) Homepage
    I realize that what I'm about to propose does not work. The challenge is to figure out why

    Here's a proposal for a compression scheme that has the following properties:

    1. It works on all bit strings of more than one bit.

    2. It is lossless and reversible.

    3. It never makes the string larger. There are some strings that don't get smaller, but see item #4.

    4. You can iterate it, to reduce any string down to 1 bit! You can use this to deal with pesky strings that don't get smaller. After enough iterations, they will be compressed.

    OK, here's my algorithm:

    Input: a string of N bits, numbered 0 to N-1.

    If all N bits are 0, the output is a string of N-1 1's. Otherwise, find the lowest numbered 1 bit. Let its position be i. The output string consists of N bits, as follows:

    Bits 0, 1, ... i-1 are 1's. Bit i is 0. Bits i+1, ..., N-1 are the same as the corresponding input bits.

    Again, let me emphasize that this is not a usable compression method!. The fun is finding the flaw.

  • by Anonymous Coward on Tuesday January 08, 2002 @09:37AM (#2803263)
    this sounds a lot like using pi (3.141592...) for compression. any random string is guaranteed to occur in that sequence, so just find the position of the string in pi and pronto.... compression!

    doesnt work though since on average you'll need as many numbers to describe the position of the string as yould need to simply represent the string in the first place.

    instead of using pi, they create a '4th dimension', i.e. some sort of combination of all possible combinations in 3 dimensions. different from the pi example though, the problem now is not representing the position in this dimension (4 coordinates) but the recreation of this space (which needs to be enumerated) by the guy who wants to decript the message.

    for 'short' strings a few pointers in a 4 dimensional space will do, for longer strings more dimensions, leading to longer pointers, longer tables of enumeration etc.

    of course, this can be tackled the other way around as well. 'random', by definition, means that the next instance doesnt have any relation to the previous. if you can find such a relation its no longer random to start with.
  • What is compression (Score:3, Interesting)

    by Vapula ( 14703 ) on Tuesday January 08, 2002 @09:37AM (#2803266)
    Compression, after all, is removing all redundancy from the original data.

    So, if there is no redundancy, there is nothing to remove (if you want to remain lossless).

    When you use some text, you may compres by remving some letter evn if tht lead to bad ortogrph. That is because English (as other langages) is redundant. When compressing some periodical signal, you may give only one period and tell that the signal is then repeated. When compressing bytes, there are specific methods (RLE, Huffman's trees,...)

    But, in all these situations, there was some redundancy to remove...

    A compression algorithm may not be perfect (it usually has to add some info to tell how the original data was compressed). Then, recompressing with another compression algorithm (or sometimes, the same will do the trick) may improve the compression. But the information quantity inside the data is the lower limit.

    Now, take a true random data stream of n+1 bits. Even if you know the value of the n first bits, you can't predict the value of n+1. In other words, there is no way that could allow the express these n+1 bits with n (or less) bits. By definition, true random data can't be compressed.

    And, to finish, compression ratio of 1:100 can be easily archived with some data... take a sequence of 200 bytes at 0x00... It may be compressed to 0xC8 0x00. Compression ratio is really only meaningful when comparing different algorithms compressing the same data stream.
  • by Zocalo ( 252965 ) on Tuesday January 08, 2002 @09:39AM (#2803278) Homepage
    Reading through the press release it seems to imply that they take the "random" data, massage the data with the "Tuner" part, then compress it with the "Accelerator" part. This spits out "BitPerfect" which I assume is their data format. It's this "massaging" of the figures where it's going to sink or swim.

    Take very large prime numbers and the like, huge strings of almost random numbers that can often be written as a trivial (2^n)-1 type formula. Maybe the massaging of the figures is simply finding a very large number that can be expressed like the above with an offset other than "-1" to get the correct "BitPerfect" data. I was toying around with this idea when there was a fad for expressing DeCSS code in unusual ways, but ran out of math before I could get it to work.

    The above theory maybe bull when it comes to the crunch, but if it could be made to work, then the compression figures are bang in the ball park for this. They laughed at Goddard remember? But I have to admit, I think replacing Einstein with the Monty Python foot better fits my take on this at present...

  • by danielrendall ( 521737 ) on Tuesday January 08, 2002 @09:51AM (#2803360) Homepage Journal
    Anybody interested in data compression and a whole lot else besides might want to download the book available from here [cam.ac.uk]

    Please don't all do so at once though :-)

    It's essentially a collection of lecture notes for a course on information theory and neural networks given by the author (David MacKay), but has been much expanded since I took the course in 1997. It will certainly show how any claim for a compression technique which works consistently on random data is bogus.

  • by dannyspanner ( 135912 ) on Tuesday January 08, 2002 @10:02AM (#2803428) Homepage
    Okay, the mysterious Dr. Wlodzimierz Holtzinski doesn't get a single hit on Google. Dr. Steve Smale [berkeley.edu] hasn't release a paper in five years [berkeley.edu] and is in his seventies [tamu.edu]. Retired, perhaps?

    I'm still not impressed.
  • by Mr Z ( 6791 ) on Tuesday January 08, 2002 @10:05AM (#2803441) Homepage Journal

    Their claims are 100% accurate (they can compress random data 100:1) only if (by their definition) random data comprises a very small percentage of all possible data sequences. The other 99.9999% of "non-random" sequences would need to expand. You can show this by a simple counting argument.

    This is covered in great detail in the comp.compression [faqs.org] FAQ. Take a look at the information on the WEB Technologies DataFiles/16 compressor (notice the similarity of claims!) if you're unconvinced. You can find it in Section 8 of Part 1 [faqs.org] of the FAQ.

    --Joe
  • team members (Score:3, Interesting)

    by loudici ( 27971 ) on Tuesday January 08, 2002 @10:07AM (#2803451) Homepage
    navigating through the flash rubbish you can reach a list of team members that includes steve smale from berkeley and richard stanley from MIT who both are existing senior academics.

    so either someone has lent their names to weirdoes without paying attention or there is something of substance hidden behind the PR ugliness. after all the PR is aimed toward investors, not toward sentient human beings, and is most probably not under the control of the scientific team.
  • by Thagg ( 9904 ) <thadbeier@gmail.com> on Tuesday January 08, 2002 @10:10AM (#2803467) Journal
    I was wondering as I read the headline and summary on slashdot "how can these sleazeballs possibly promote this scam, because it would be easy to show counterexamples?" This shows, once again, that I lack the imagination and chutzpah of a real con artist.

    The beauty of this scam is that zeospace claims that they can't even do it themselves, yet. They've only managed to compress very short strings. So, they can't be called to compress large random files because, well gosh, they just haven't gotten the big file compressor work yet. So, you can't prove that they are full of shit.

    Beautiful flash animation, though. I particularly like the fact that clicking the 'skip intro' button does absolutely nothing -- you get the flash garbage anyway.

    thad
  • by mblase ( 200735 ) on Tuesday January 08, 2002 @10:22AM (#2803538)
    Existing compression technologies are currently dependent upon the mapping and encoding of redundantly occurring mathematical structures, which are limited in application to single or several pass reduction. ZeoSync's approach to the encoding of practically random sequences is expected to evolve into the reduction of already reduced information across many reduction iterations, producing a previously unattainable reduction capability. ZeoSync intentionally randomizes naturally occurring patterns to form entropy-like random sequences through its patent pending technology known as Zero Space Tuner?. Once randomized, ZeoSync's BinaryAccelerator? encodes these singular-bit-variance strings within complex combinatorial series to result in massively reduced BitPerfect? equivalents. The combined TunerAccelerator? is expected to be commercially available during 2003.
    Now, I'm not as geeky as some, but this looks suspiciously like technobabble designed to impress a bunch of investors and provide long-term promises which can easily be evaded by the end of the next fiscal year. I mean, if they really did have such a technology available today, why is it going to take them an entire twelve months to integrate it into a piece of commercial software?
  • by Anarchofascist ( 4820 ) on Tuesday January 08, 2002 @10:30AM (#2803574) Homepage Journal
    "When you send me a one-byte copy of, say, The Matrix, you also have to tell me how many times it was compressed so I know how many times to run the decompressor!"

    Not true! You don't need an extra byte for the number of times the compression has been run, as long as you compress files that are no larger than a certain size.

    If each pass reduces the size by two orders of magnitude, then 256 compressions will compress down by a factor of (on average) 10^512 = one hundred million billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion times. That's enough to compress a 1024 x 768 movie (at 50 fps and 24 bit colour) into a single byte, as long as the movie runs for less than fifty five billion eight hundred million billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion billion times the current age of the UNIVERSE.

    Therefore, I should easily be able to compress The Matrix into a single byte with 256 passes.

    I don't need to encode the number of compressions, every decompression consists of decompressing 256 times.

  • by softsign ( 120322 ) on Tuesday January 08, 2002 @11:31AM (#2803896)
    I'm not sure if I understand your point, but from what I do understand, it seems to me you are missing it.

    If you look at this sequence as a one-dimensional series: 00101101, it's pretty hard (at least for a processor) to distinguish a pattern there... it's a pseudo-random sequence. But if I paint it this way, in 2d: (0,0) (1,0) (1,1) (0,1), I can step back and see a square with sides of length one.

    AFAIK, what these people are claiming is that they've developed a way to step WAY back, to n-dimensions, and have patterns emerge from seemingly random data.

    It's not the random-number generation that's significant here... it's the purported ability to compress a seemingly random sequence. RLE typically doesn't fare very well with pure random data because it only looks for certain types of redundancy.

    If I haven't missed the boat here, it's really a very interesting achievment.

  • by wberry ( 549228 ) on Tuesday January 08, 2002 @01:02PM (#2804321) Homepage

    Back in 1991 or 1992, in the days of 2400 bps modems, MS-DOS 5.0, and BBS'es, a "radical new compression tool" called OWS made the rounds. It claimed to have been written by some guy in Japan and use breakthroughs in fractal compression, often achieving 99% compression! "Better than ARJ! Better than PKzip!" Of course all my friends and I downloaded it immediately. Now we can send gam^H^H^Hfiles to each other in 10 minutes instead of 10 hours!

    Now I was in the ninth grade, and compression technology was a complete mystery to me then, so I suspected nothing at first. I installed it and read the docs. The commands and such were pretty much like PKzip. I promptly took one of my favorite ga^H^Hdirectories, *copied it to a different place*, compressed it, deleted it, and uncompressed it without problems. The compressed file was exactly 1024 bytes. Hmm, what a coincidence!

    The output looked kind of funny though:
    Compressing file abc.wad by 99%.
    Compressing file cde.wad by 99%.
    Compressing file start.bat by 99%.
    etc. Wait, start.bat is only 10 characters, that's like one bit! And why is *every* file compressed by 99%? Oh well, must be a display bug.

    So I called my friend and arranged to send him this g^Hfile via Zmodem, and it took only a few seconds. But he couldn't uncompress it on the other side. "Sector Not Found", he said. Oh well, try it again. Same result. Another bug.

    So I decided that this wasn't working out and stopped using OWS. Their user interface needed some work anyway, plus I was a little suspicious of compression bugs. The evidence was right there for me to make the now-obvious conclusion, but it didn't hit me until a few *weeks* later when all the BBS sysops were posting bulletins warning that OWS was a hoax.

    As it turns out, OWS was storing the FAT information in the compressed files, so that when people do reality checks it will appear to re-create the deleted files, as it did for me. But when they try to uncompress a file that actually isn't there or has had its FAT entries moved around, you get the "Sector Not Found" error and you're screwed. If I hadn't tried to send a compressed file to a friend I might have been duped into "compressing" and deleting half my software or more.

    All in all, a pretty cruel but effective joke. If it happened today somebody would be in federal pound-me-in-the-ass prison. Maybe it happened then too...

    (Yes, this is slightly off-topic, but where else am I going to post this?)

  • by Lictor ( 535015 ) on Tuesday January 08, 2002 @01:17PM (#2804375)
    I think the following statement in the press release pretty much says it all:

    >We perceive this advancement as a significant
    >breakthrough to the historical limitations of
    >digital communications as it was originally
    >detailed by
    >Dr. Claude Shannon in his treatise on Information
    >Theory."

    How about algorithmic information theory? Kolmogorov, Solomonov, Chaitin? The statement above indicates that the most recent word on compression is an old Bell Labs tech report by Claude Shannon... not to put Shannon down, that work *is* a landmark, but there has certainly been more work done since.

    Try compressing the number Pi using Shannons theory... you can't do it. On the other hand, using Kolmogorov complexity, you can compress it quite nicely.

    The fact that this statement appears in the press release seems to indicate a great deal of ignorance on the part of this corporations researchers. Part of any good research program is to familiarize yourself with previous work done in the field... and AIT is *not* some obscure backwater idea... there are several conferences on this topic every year and just about every CS graduate student has seen at least Kolmogorov complexity.

    This is a pretty serious credibility robber. (Not to mention that from a mathematical standpoint, compressing totally random data is impossible under our current axioms... so if we *can* compress completely random data... its time for a new theory of the foundations of mathematics. At the risk of sounding dogmatic: do you *really* think some dot-com startup is capable of this?

    Perhaps they are, but I'm going to need to see the proofs written up nice and formally before I run out and buy snake-oi... I mean *stock*.)
  • by MobiusKlein ( 58188 ) on Tuesday January 08, 2002 @03:38PM (#2805293)
    If you read the Reuters article carefully, it does not say a digital -> digital compression of 1:100, but implies a better way of encoding / compressing digital -> analog -> digital, with the analog bandwidth being much greater than today.

    Thats all the stuff where they talk about Dr. Claude Shannon and information theory. (They could have been clearer about it, but that's PR flacks for you.)

    examine the quote
    '"What we've developed is a new plateau in communications theory," St. George said. "We are expecting to produce the enormous capacity of analog signaling, with the benefit of the noise-free integrity of digital communications."'

    Sounds like they are trying to shove more data into an analog stream, using wacky math, than would normally be allowed.

    rbb
  • by Rubbersoul ( 199583 ) on Tuesday January 08, 2002 @06:58PM (#2806711)
    This may have already been posted, and if it has sorry, but I thought this may be of interest to some of you.

    Jean-loup Gailly (one of the creators of gzip) has written an article on a patent that was granted for compression of truly random data, and how it is not mathematically possible. You can read it here [gailly.net] for those that are interested.

Stellar rays prove fibbing never pays. Embezzlement is another matter.

Working...