I should note that as it digests new data in 800GB it will end up throwing away 
more more often as it already stores something similar to it already or doesn't 
favorite what it read or didn't read it enough in the data to learn it. As 
said, it can't store all of Earth data losslessly, only some; lossy. As it 
trains on the given small 2GB dataset, it starts off saving on space by using 
general patterns, it basically 'knows it all', hence performance is high after 
training a man for 20 years but performance shrinks if you store every data of 
Earth. Of course their paper says error lowers, rises, lowers, so it is clearly 
'storing things' that aren't that re-used. It could be that as it learns a 
higher level layer, it piles up cost to learn all the alphabet or all the 
words, but then it just slashes all the rest of the training by pointing to its 
stored nodes now. So rule of thumb is find as much diverse data as you can, 
store all different_enough patterns to 'know it all' without storing the full 
universe and don't store rare/ unfavorited,/ unrelated features and don't store 
features that have few parents utilizing them except for higher layer nodes. 
You won't know all data of Earth losslessly, only that your and Tom's cat drink 
milk, only that you have blonde hair, only that there is 7 billion humans on 
Earth. The low memory is all that's needed to understand the universe.
------------------------------------------
Artificial General Intelligence List: AGI
Permalink: 
https://agi.topicbox.com/groups/agi/Tc1f2c133ae3e4762-Mdf9620b27c84a83edbaa6f9b
Delivery options: https://agi.topicbox.com/groups/agi/subscription

Reply via email to