Are you saying BPC = Lossless Compression? So openAI had GPT do lossless 
compression then to get a BPC score? But I don't think they did lossless 
compression...

Doesn't BPC need an arithmetic function? I mean you have a set of predicted 
probabilities for each letter (or word/token, which need to be divided down to 
letters), then....oh...wait I see something here....this is to make the 
file/dataset smaller to prove it has a good predictor, while:

BPC seems to beĀ  way to show how good it predicts, without any compression... 
For example say we predict a=60% likely next, b=30%, c=10%. Now, instead of an 
arithmitic coder to give us a long d5D83nF#jsiv that encodes each letter, we 
instead say OK, next letter is 'c' (it sucks at predicting this one), so we got 
10% here then.....then for each next letter: 30%, 20%, 40%, 10%, 10%....then we 
can add em all up or average to get let's see: we have 500% to eat at down to 
be perfect predictor, and ate down 110% in total, so: 0.22 score. Right??? This 
seems cool, then can't we do this instead of compression? It'd be much 
faster/simpler I think.
------------------------------------------
Artificial General Intelligence List: AGI
Permalink: 
https://agi.topicbox.com/groups/agi/Td13a829978c4c9f3-M9a6ec6d0a7a512a93976da73
Delivery options: https://agi.topicbox.com/groups/agi/subscription

Reply via email to