About 10,000, 100,000, 1,000,000, 10,000,000 bytes: 0.44888181407357763 0.433259276532548 0.35976039165384843 0.43851910590941723
must be an issue somewhere Well, I guess it was not sure at first, hence the large score, it doesn't change rapidly the more samples it uses see the upper 3.....and it slowly gets better ya....hmm......So if each correct letter prob it picks is 100%, then average would be high; 1.00000000....high=better predictor code is so small now: for m in range(256): if m == ord(window[-1]): list.append(predict[m]) print(round((sum(list) / len(list)), 200)) Now, we have to include program size....so...mine is basically 5,000 bytes not compressed......Hmm, if I'm compressing 100MBs, or rather trying to get score 1.0, then, we could say if my code was 100MB big then maybe I am storing every BPC answer (not that this is BPC I'm doing), so 5KBs is like, take 100MB and minus 5KBs, and that gives you the discount to remove from my 0.44888 score above I got.....so 100,000,000 - 5,000 = 0.00005....therefore 0.44888 - 0.00005 = 0.44883 RIGHT? Isn't this cool? ------------------------------------------ Artificial General Intelligence List: AGI Permalink: https://agi.topicbox.com/groups/agi/Td13a829978c4c9f3-Mf59bb4ee48af19e335f636d6 Delivery options: https://agi.topicbox.com/groups/agi/subscription