Den onsdagen den 30:e oktober 2013 kl. 20:35:59 UTC+1 skrev Tim Delaney:
> On 31 October 2013 05:21,  <jonas.t...@gmail.com> wrote:
> 
> I am searching for the program or algorithm that makes the best possible of 
> completly (diffused data/random noise) and wonder what the state of art 
> compression is.
> 
> 
> 
> 
> I understand this is not the correct forum but since i think i have an 
> algorithm that can do this very good, and do not know where to turn for such 
> question i was thinking to start here.
> 
> 
> 
> It is of course lossless compression i am speaking of.
> 
> 
> 
> This is not an appropriate forum for this question. If you know it's an 
> inappropriate forum (as you stated) then do not post the question here. Do a 
> search with your preferred search engine and look up compression on lossless 
> Wikipedia. And read and understand the following link:
> 
> 
> 
> http://www.catb.org/esr/faqs/smart-questions.html
> 
> 
> 
> paying special attention to the following parts:
> 
> 
> 
> http://www.catb.org/esr/faqs/smart-questions.html#forum
> http://www.catb.org/esr/faqs/smart-questions.html#prune
> 
> 
> http://www.catb.org/esr/faqs/smart-questions.html#courtesy
> 
> http://www.catb.org/esr/faqs/smart-questions.html#keepcool
> 
> 
> http://www.catb.org/esr/faqs/smart-questions.html#classic
> 
> 
> 
> If you have *python* code implementing this algorithm and want help, post the 
> parts you want help with (and preferably post the entire algorithm in a 
> repository).
> 
> 
> 
> 
> However, having just seen the following from you in a reply to Mark ("I do 
> not follow instructions, i make them accesible to anyone"), I am not not 
> going to give a second chance - fail to learn from the above advice and 
> you'll meet my spam filter.
> 
> 
> 
> If the data is truly completely random noise, then there is very little that 
> lossless compression can do. On any individual truly random data set you 
> might get a lot of compression, a small amount of compression, or even 
> expansion, depending on what patterns have randomly occurred in the data set. 
> But there is no current lossless compression algorithm that can take truly 
> random data and systematically compress it to be smaller than the original.
> 
> 
> 
> If you think you have an algorithm that can do this on truly random data, 
> you're probably wrong - either your data is has patterns the algorithm can 
> exploit, or you've simply been lucky with the randomness of your data so far.
> 
> 
> 
> Tim Delaney

No i am not wrong.
End of story
-- 
https://mail.python.org/mailman/listinfo/python-list

Reply via email to