I can successfully test my belief in matter. The fact that you did not
already know this casts strong doubt on your telepathic powers.
"The second reason was that I already knew that I have telepathies when
I’m in relationships, thus I wanted to see what kind of telepathies
appear if I involve more than one girl."
From https://philpapers.org/archive/VISMAC-3.pdf
I guess if I were to write such stupid drivel I wouldn't use my real
name either.
Brent
On 12/14/2024 1:21 AM, 'Cosmin Visan' via Everything List wrote:
@Brent The only woo-woo is your belief in "matter".
On Saturday, 14 December 2024 at 01:46:07 UTC+2 Brent Meeker wrote:
On 12/13/2024 1:18 AM, 'Cosmin Visan' via Everything List wrote:
When you base an invention on the world of finite forms, of
course that invention will be limited. You will never replicate
the powers of consciousness, because consciousness draws its
powers from the infinite world of the formless. And drawing from
an infinite source, it is able to produce infinite forms and it
doesn't need quazillions of forms to learn.
Let's see you produce and infinite form or two.
A baby learns to speak from just a few examples, because what the
parents to is not to provide raw data to the baby,
Twins often invent their own language which the speak to each
other. Evolution has provided the raw data to create language.
but to stimulate the baby's consciousness to access the formless
source and to draw from there whatever forms it needs in order to
be able to speak and generally learn anything.
Woo-Woo magic.
Brent
On Friday, 13 December 2024 at 09:29:37 UTC+2 Alan Grayson wrote:
On Thursday, December 12, 2024 at 7:38:11 PM UTC-7 Brent
Meeker wrote:
Magic is always the explanation of those who can't
understand.
Brent
*There's plenty of magic, under a different name, in physics.
Another pitfall is religating hidden knowledge, aka occult
knowledge, such as the Chakras in Yoga, to de facto magic or
someone's overactive imagination. AG *
*
*
On 12/12/2024 1:39 PM, 'Cosmin Visan' via Everything List
wrote:
Magic!
On Thursday, 12 December 2024 at 20:00:58 UTC+2 John
Clark wrote:
*The number of "tokens" (words or parts of words)
used to train LLMs is 100 times larger than it was
in 2020, the largest are now using tens of
trillions. if you only consider text then the
entire Internet only contains about 3,100 trillion
tokens. The amount of text LLMs train on is doubling
every year but the amount of human generated text on
the Internet is only growing at about 10% a year, if
that trend continues AIs will run out of text
somewhere around 2028. Does that mean AI progress
is about to hit a wall? I don't think so for the
following reasons:*
*For one thing, because of improvements in
algorithms, the computing power needed for a Large
Language Model to achieve the same performance has
halved about every 8 months. *
*
*
*ALGORITHMIC PROGRESS IN LANGUAGE MODELS*
<https://arxiv.org/pdf/2403.05812>
*And computer chips specialized for AI rather than
general computing, like those made by Nvidia and
other companies, are getting faster even more
rapidly than Moore's Law. Also, the rate of growth
of specialized data sets, such as astronomical and
biological data, are growing much much more quickly
than text is; that's how AIs got so good at
predicting how proteins fold up. *
*And there is vastly more information if AI's are
trained on other types of data besides text, and
some AI's are already being trained on unlabeled
images and videos. Yann LeCun, chief AI scientist
at Meta, said that "/although the 10^13 tokens used
to train a LLM sounds like a lot /(it would take a
human 170,000 years to read that much)/, a
4-year-old child has absorbed a volume of data 50
times greater than that just by looking at objects
during his waking hours. We’re never going to get to
human-level AI by just training on language, that’s
just not happening/".*
*And then there's synthetic data. AlphaGeometry was
trained to solve geometry problems using 100 million
computer generated synthetic examples with no human
demonstrations, and it ended up being as good at
solving difficult geometry problems as the very best
high school students in the entire nation. *
*Solving olympiad geometry without human
demonstrations*
<https://www.nature.com/articles/s41586-023-06747-5>
*AI researchers are starting to change their
strategy and have their AI's reread their training
set many times because AI's operate in a statistical
way so rereading improves performance *
*Scaling Data-Constrained Language Models*
<https://arxiv.org/pdf/2305.16264>
*Andy Zouat Carnegie Mellon Universitysays "/once
//an AI has got a foundational knowledge base that’s
probably greater than any single person could
have,it no longer needs more data to get smarter. It
just needs to sit and think. I think we’re probably
pretty close to that point/.”*
*
*
*John K Clark See what's on my new list at
Extropolis <https://groups.google.com/g/extropolis>*
--
You received this message because you are subscribed to the
Google Groups "Everything List" group.
To unsubscribe from this group and stop receiving emails from it,
send an email to [email protected].
To view this discussion visit
https://groups.google.com/d/msgid/everything-list/d190939b-b49f-4bd8-a77f-2cec16f8816dn%40googlegroups.com
<https://groups.google.com/d/msgid/everything-list/d190939b-b49f-4bd8-a77f-2cec16f8816dn%40googlegroups.com?utm_medium=email&utm_source=footer>.
--
You received this message because you are subscribed to the Google
Groups "Everything List" group.
To unsubscribe from this group and stop receiving emails from it, send
an email to [email protected].
To view this discussion visit
https://groups.google.com/d/msgid/everything-list/e90a4ee5-c8b8-4178-afb9-8c7a9ac09690n%40googlegroups.com
<https://groups.google.com/d/msgid/everything-list/e90a4ee5-c8b8-4178-afb9-8c7a9ac09690n%40googlegroups.com?utm_medium=email&utm_source=footer>.
--
You received this message because you are subscribed to the Google Groups
"Everything List" group.
To unsubscribe from this group and stop receiving emails from it, send an email
to [email protected].
To view this discussion visit
https://groups.google.com/d/msgid/everything-list/bfb4373e-91c2-4c3c-a1e3-4111c40b8384%40gmail.com.