On Saturday, May 18, 2024, at 6:53 PM, Matt Mahoney wrote: > Surely you are aware of the 100% failure rate of symbolic AI over the last 70 > years? It should work in theory, but we have a long history of > underestimating the cost, lured by the early false success of covering half > of the cases with just a few hundred rules. >
I view LLM’s as systems within symbolic systems. Why? Simply that we exist in a spacetime environment and ALL COMMUNICATION is symbolic. And sub-symbolic representation is required for computation. All bits are symbols based on probabilities. Then as LLM’s become more intelligent the physical power consumption required to produce similar results will decrease as their symbolic networks grow and optimize. Could be wrong but It makes sense to me… saying everything is symbolic eliminates the argument. I know it's lazy but that's often how developers look at things in order to code them up :) Laziness is a form of optimization... ------------------------------------------ Artificial General Intelligence List: AGI Permalink: https://agi.topicbox.com/groups/agi/T682a307a763c1ced-M2252941b1c7cca5b59b32c1f Delivery options: https://agi.topicbox.com/groups/agi/subscription
