Why are you envious on me ? Because your life is a failure ? Don't you think a more productive way would be to actually do something about your life instead of hating random people for your own failures ?
On Tuesday, 17 December 2024 at 22:48:37 UTC+2 Brent Meeker wrote: > You're one who lies about it. Which no doubt makes you happy since it > supports your conviction that you're smarter than everyone else. > > Brent > > > > > On 12/17/2024 1:34 AM, 'Cosmin Visan' via Everything List wrote: > > What makes you lie about the existence of telepathy ? Does it make you > happy ? Does it replace the lack of vvahmen in your life ? > > On Monday, 16 December 2024 at 21:45:15 UTC+2 Brent Meeker wrote: > >> So you use non-existent telepathy. Well I guess that's easiest kind to >> obtain. >> >> I'm quite clear on the meaning of epistemology and ontology. Having >> knowledge of Cosmin Visan does mean he exists. >> >> Brent >> >> >> >> On 12/15/2024 2:37 AM, 'Cosmin Visan' via Everything List wrote: >> >> You make the classical confusion between epistemology and ontology. Only >> because you can use something, it doesn't mean that that something exists. >> Only because you watch a movie with Spider-Man, it doesn't mean Spider-Man >> exists. >> >> Also, I highly recommend you to perform for yourself such telepathy >> experiment. Thank you for reading my papers! >> >> On Saturday, 14 December 2024 at 22:57:55 UTC+2 Brent Meeker wrote: >> >>> I can successfully test my belief in matter. The fact that you did not >>> already know this casts strong doubt on your telepathic powers. >>> >>> "The second reason was that I already knew that I have telepathies when >>> I’m in relationships, thus I wanted to see what kind of telepathies appear >>> if I involve more than one girl." >>> From https://philpapers.org/archive/VISMAC-3.pdf >>> >>> I guess if I were to write such stupid drivel I wouldn't use my real >>> name either. >>> >>> Brent >>> >>> >>> >>> On 12/14/2024 1:21 AM, 'Cosmin Visan' via Everything List wrote: >>> >>> @Brent The only woo-woo is your belief in "matter". >>> >>> On Saturday, 14 December 2024 at 01:46:07 UTC+2 Brent Meeker wrote: >>> >>>> >>>> >>>> >>>> On 12/13/2024 1:18 AM, 'Cosmin Visan' via Everything List wrote: >>>> >>>> When you base an invention on the world of finite forms, of course that >>>> invention will be limited. You will never replicate the powers of >>>> consciousness, because consciousness draws its powers from the infinite >>>> world of the formless. And drawing from an infinite source, it is able to >>>> produce infinite forms and it doesn't need quazillions of forms to learn. >>>> >>>> Let's see you produce and infinite form or two. >>>> >>>> >>>> A baby learns to speak from just a few examples, because what the >>>> parents to is not to provide raw data to the baby, >>>> >>>> Twins often invent their own language which the speak to each other. >>>> Evolution has provided the raw data to create language. >>>> >>>> >>>> but to stimulate the baby's consciousness to access the formless source >>>> and to draw from there whatever forms it needs in order to be able to >>>> speak >>>> and generally learn anything. >>>> >>>> Woo-Woo magic. >>>> >>>> Brent >>>> >>>> >>>> On Friday, 13 December 2024 at 09:29:37 UTC+2 Alan Grayson wrote: >>>> >>>>> On Thursday, December 12, 2024 at 7:38:11 PM UTC-7 Brent Meeker wrote: >>>>> >>>>> Magic is always the explanation of those who can't understand. >>>>> >>>>> Brent >>>>> >>>>> >>>>> *There's plenty of magic, under a different name, in physics. Another >>>>> pitfall is religating hidden knowledge, aka occult knowledge, such as the >>>>> Chakras in Yoga, to de facto magic or someone's overactive imagination. >>>>> AG * >>>>> >>>>> On 12/12/2024 1:39 PM, 'Cosmin Visan' via Everything List wrote: >>>>> >>>>> Magic! >>>>> >>>>> On Thursday, 12 December 2024 at 20:00:58 UTC+2 John Clark wrote: >>>>> >>>>> *The number of "tokens" (words or parts of words) used to train LLMs >>>>> is 100 times larger than it was in 2020, the largest are now using tens >>>>> of >>>>> trillions. if you only consider text then the entire Internet only >>>>> contains about 3,100 trillion tokens. The amount of text LLMs train on is >>>>> doubling every year but the amount of human generated text on the >>>>> Internet >>>>> is only growing at about 10% a year, if that trend continues AIs will run >>>>> out of text somewhere around 2028. Does that mean AI progress is about >>>>> to >>>>> hit a wall? I don't think so for the following reasons:* >>>>> >>>>> *For one thing, because of improvements in algorithms, the computing >>>>> power needed for a Large Language Model to achieve the same performance >>>>> has halved about every 8 months. * >>>>> >>>>> *ALGORITHMIC PROGRESS IN LANGUAGE MODELS* >>>>> <https://arxiv.org/pdf/2403.05812> >>>>> >>>>> >>>>> *And computer chips specialized for AI rather than general computing, >>>>> like those made by Nvidia and other companies, are getting faster even >>>>> more >>>>> rapidly than Moore's Law. Also, the rate of growth of specialized data >>>>> sets, such as astronomical and biological data, are growing much much >>>>> more >>>>> quickly than text is; that's how AIs got so good at predicting how >>>>> proteins >>>>> fold up. * >>>>> >>>>> *And there is vastly more information if AI's are trained on other >>>>> types of data besides text, and some AI's are already being trained on >>>>> unlabeled images and videos. Yann LeCun, chief AI scientist at Meta, >>>>> said >>>>> that "although the 10^13 tokens used to train a LLM sounds like a lot >>>>> (it would take a human 170,000 years to read that much) , a 4-year-old >>>>> child has absorbed a volume of data 50 times greater than that just by >>>>> looking at objects during his waking hours. We’re never going to get to >>>>> human-level AI by just training on language, that’s just not happening".* >>>>> >>>>> *And then there's synthetic data. AlphaGeometry was trained to solve >>>>> geometry problems using 100 million computer generated synthetic examples >>>>> with no human demonstrations, and it ended up being as good at solving >>>>> difficult geometry problems as the very best high school students in the >>>>> entire nation. * >>>>> >>>>> *Solving olympiad geometry without human demonstrations* >>>>> <https://www.nature.com/articles/s41586-023-06747-5> >>>>> >>>>> *AI researchers are starting to change their strategy and have their >>>>> AI's reread their training set many times because AI's operate in a >>>>> statistical way so rereading improves performance * >>>>> >>>>> >>>>> *Scaling Data-Constrained Language Models* >>>>> <https://arxiv.org/pdf/2305.16264> >>>>> >>>>> >>>>> *Andy Zou at Carnegie Mellon University says "once an AI has got a >>>>> foundational knowledge base that’s probably greater than any single >>>>> person >>>>> could have, it no longer needs more data to get smarter. It just needs to >>>>> sit and think. I think we’re probably pretty close to that point.”* >>>>> >>>>> *John K Clark See what's on my new list at Extropolis >>>>> <https://groups.google.com/g/extropolis>* >>>>> >>>>> >>>>> -- >>>> You received this message because you are subscribed to the Google >>>> Groups "Everything List" group. >>>> To unsubscribe from this group and stop receiving emails from it, send >>>> an email to [email protected]. >>>> >>>> To view this discussion visit >>>> https://groups.google.com/d/msgid/everything-list/d190939b-b49f-4bd8-a77f-2cec16f8816dn%40googlegroups.com >>>> >>>> <https://groups.google.com/d/msgid/everything-list/d190939b-b49f-4bd8-a77f-2cec16f8816dn%40googlegroups.com?utm_medium=email&utm_source=footer> >>>> . >>>> >>>> >>>> -- >>> You received this message because you are subscribed to the Google >>> Groups "Everything List" group. >>> To unsubscribe from this group and stop receiving emails from it, send >>> an email to [email protected]. >>> >>> To view this discussion visit >>> https://groups.google.com/d/msgid/everything-list/e90a4ee5-c8b8-4178-afb9-8c7a9ac09690n%40googlegroups.com >>> >>> <https://groups.google.com/d/msgid/everything-list/e90a4ee5-c8b8-4178-afb9-8c7a9ac09690n%40googlegroups.com?utm_medium=email&utm_source=footer> >>> . >>> >>> >>> -- >> You received this message because you are subscribed to the Google Groups >> "Everything List" group. >> To unsubscribe from this group and stop receiving emails from it, send an >> email to [email protected]. >> >> To view this discussion visit >> https://groups.google.com/d/msgid/everything-list/046a5797-d14e-4321-959a-eb0362178e27n%40googlegroups.com >> >> <https://groups.google.com/d/msgid/everything-list/046a5797-d14e-4321-959a-eb0362178e27n%40googlegroups.com?utm_medium=email&utm_source=footer> >> . >> >> >> -- > You received this message because you are subscribed to the Google Groups > "Everything List" group. > To unsubscribe from this group and stop receiving emails from it, send an > email to [email protected]. > > To view this discussion visit > https://groups.google.com/d/msgid/everything-list/c638e101-1a26-47fe-8723-314433eea49en%40googlegroups.com > > <https://groups.google.com/d/msgid/everything-list/c638e101-1a26-47fe-8723-314433eea49en%40googlegroups.com?utm_medium=email&utm_source=footer> > . > > > -- You received this message because you are subscribed to the Google Groups "Everything List" group. To unsubscribe from this group and stop receiving emails from it, send an email to [email protected]. To view this discussion visit https://groups.google.com/d/msgid/everything-list/b2d8614a-8856-40ed-8cc1-31d2f31beb00n%40googlegroups.com.

