On Sunday, July 18, 2021, at 8:59 PM, YKY (Yan King Yin, 甄景贤) wrote: > Final version of my paper (corrected a lot of inaccuracies): > https://drive.google.com/file/d/1P0D9814ivR0MScowcmWh9ISpBETlUnq-/view?usp=sharing > You seem to mix together 1) BERT's prediction of next word in text with 2) prediction of next item in IQ tests but these two are not exactly the same...? You may argue that humans can do both, and indeed an AGI should be able to do that too. Typically, it would require multiple steps of reasoning, you were just looking at 1 layer of Transformer or Attention Mechanism, that corresponds to a single step of inference. Because prediction is prediction, the use of patterns is all that AGI does. There is different pattern finders for different types of problems. Saying to a robot "if you (see 1 cat) or (see 1 dog and a plane) walk forward else walk backwards" is a sort of match and then prediction as output. Nonetheless your paper is too un-unified, I'd like to see something way more simpler and evil and easy that explains everything about AGI. You can do a lot of damage with just 100 words and commonly used words instead of "topos theory". If we want to unite ideas to become openAI2, explain your work using common words/ sentences.
Question: I've asked a few people a question around forums and either they are too busy Transforming, or can't answer the question properly because really they didn't invent the Transformer architecture and are just users, lots of users albeit. I saw DALL-E/ IGPT predicts properly, whether given half a boat or half a word ex. 't h a n k s g i v ?' it is able to recognize it even if only saw a small version of a boat or word thanksgiving. This can't be just delay recognition with discounts for uneven timely activations, it needs a pattern of error in time errors, Hinton calls it "equivalence". And the thing is, BACKPROP and whatever cannot "learn" this on their own because it is too specific a function....it must be in the code, the code is like 6000 lines of code for heavens sake, which is VerY annoying and I refuse to understand it for now cuz mine is ~120 lines and half way there. ------------------------------------------ Artificial General Intelligence List: AGI Permalink: https://agi.topicbox.com/groups/agi/Tb5526c8a9151713b-M73ede6bd825fd107b122a902 Delivery options: https://agi.topicbox.com/groups/agi/subscription
