The book sounds intriguing, but it's not in my lending library.

Happily, there are lots of Tanner Lecture videos available online:
    https://www.youtube.com/watch?v=yI-MCqeCILs Wendy Brown from 2019
    https://www.youtube.com/watch?v=NJprCvmrpoY Kim Stanley Robinson from
March 2023

-- rec --


-- rec --

On Sun, Oct 8, 2023 at 11:56 AM David Eric Smith <desm...@santafe.edu>
wrote:

> It’s a good point/counterpinet, Pieter.
>
> I don’t want to stand against using new tools to relieve a scarcity
> constraint, and surely education has been a severe one.
>
> I have what appears to be an enormous suspicion of a bait and switch, in
> which suddenly we find ourselves in a world where simultaneously:
>
> 1. Sal Khan and Daphne Koller own everything, and nobody else owns
> anything; and
>
> 2. People have no skills (say I, as I use the scary-good google translate
> to take enormous pressure, though less shame, off me for not learning a
> language I should have learned by now).
>
> Or, as a popular bumper sticker in Santa Fe used to say post-2016: Where
> are we going?  And what am I doing in this handbasket?
>
> I should mention, on this point, that part of my cast of mind was shaped
> by an article I read some time ago in some education chronicle, about how
> the command-and-control interface to the computer was taking up more and
> more space in classrooms, and kids were starting to show severe systematic
> deficits in particular areas.  They could command anything they wanted into
> existence on the computer, and failed to learn that the rest of the world
> of people doesn’t work that way.  Negotiating skills were mentioned in the
> article.  My description of a dystopian non-education was meant to suggest
> that for several students to have to hear exactly the same thing would
> require them to come to terms with the fact that it will fit them
> differently, and they should become used to handling that too.
>
> It is interesting that I could never use a chatbot as a first stage of
> writing something, as the optimistic chatbots suggest everybody will do,
> and which I can believe for some people will be a good fit.  For me, I
> can’t even take looking at a page of my own writing when I am in a tangle,
> as a start to getting out of it.  All that pre-formed text is in the way of
> whatever part of my brain tries to bring into focus what I should do next.
> I know that chatbots and writing is not your point about education, and
> don’t mean to derail those points; the chatbots-and-journalism theme was
> one of the things from the original thread that put me onto the education
> sidetrack.  What (at least in my imagination) they have in common is that
> there are certain mental operations that I only do when I have “space”, and
> many of these tools seem designed to make sure nobody ever has to leave any
> “space” unfilled by some external stimulus, ever again.  I worry about it
> as a development-weakening prosthetic, though I understand that the
> argument for that would need to be made case by case, versus the likely
> (and I believe you, in many cases demonstrated) economies and advantages of
> personalization.
>
> Fortunately, I am neither designing anything nor voting on anything just
> now, so there is no danger I will hold anybody else back who is trying to
> make something work.
>
> Eric
>
>
> On Oct 8, 2023, at 7:45 AM, Pieter Steenekamp <piet...@randcontrols.co.za>
> wrote:
>
> I'd like to reflect how LLM's and AI could disrupt education.
>
> My paradigm is not to think in terms of the current educational model, but
> to throw out all preconceived ideas and to design it from scratch. There
> are obviously zillions of ways to do it and a good model will need to
> evolve, but the following are merely two examples to trigger new and novel
> approaches, one from Sal Khan and one from my own experience.
>
> 1 Sal Khan's approach to use AI to personalize education is based on the
> idea that every student can benefit from a customized learning path that
> adapts to their needs and preferences. He argues that AI can help scale the
> benefits of one-to-one tutoring, which has been shown to improve students'
> performance significantly. He also believes that AI can enhance the role of
> teachers by providing them with useful feedback, insights, and assistance.
> He envisions a future where every student has access to an AI-powered
> personal tutor and every teacher has an AI teaching assistant.
>
> Sal Khan is the founder and CEO of Khan Academy, a nonprofit education
> platform that offers free online courses, videos, and exercises for
> learners of all ages. He has been a pioneer in using technology to
> democratize education and make it accessible to anyone, anywhere. He has
> also been exploring the potential of AI to transform education for the
> better. In his 2023 TED talk¹, he demonstrated Khanmigo, an AI-powered
> guide that can help students with various subjects, skills, and tasks.
> Khanmigo can detect students' mistakes and misconceptions, provide
> effective feedback and explanations, encourage Socratic dialogue and
> debate, and assist with writing and storytelling. Khanmigo also acts as a
> teaching assistant for teachers, helping them with lesson planning,
> grading, and progress reports.
>
> Sal Khan's approach to use AI to personalize education is based on his
> vision of creating a more humanistic and learner-centered education system.
> He believes that AI can empower students to learn at their own pace,
> explore their interests, and develop their creativity. He also believes
> that AI can support teachers to focus more on the social and emotional
> aspects of teaching, such as building relationships, fostering curiosity,
> and inspiring passion. He hopes that by using AI to personalize education,
> we can create a more equitable and inclusive learning environment for
> everyone..
>
> Source: Conversation with Bing, 2023/10/08
> (1) . https://bing.com/search?q=Sal+Khan+AI+education.
> (2) Sal Khan’s 2023 TED Talk: AI in the classroom can transform education.
> https://blog.khanacademy.org/sal-khans-2023-ted-talk-ai-in-the-classroom-can-transform-education/
> .
> (3) Sal Khan: How AI could save (not destroy) education | TED Talk.
> https://www.ted.com/talks/sal_khan_how_ai_could_save_not_destroy_education/c
> .
> (4) The Future Of AI And Education: Insights From Sal Khan.
> https://www.openbusinesscouncil.org/the-future-of-ai-and-education-insights-from-sal-khan-in-conversation-with-bill-gates
> <https://linkprotect.cudasvc.com/url?a=https%3a%2f%2fwww.openbusinesscouncil.org%2fthe-future-of-ai-and-education-insights-from-sal-khan-in-conversation-with-bill-gates&c=E,1,2zh-o7p-ZRG5J2pDpv8lLYBDKVFUgyvlCSvOCnrGzy0OEqHE_M-1ND6g52xVjiUgyBncevhmKdnOnnuuTkeZ66oyJXl1nY6Pan2Zr01pkGQWF2U,&typo=1>
> .
> (5) undefined.
> https://www.ted.com/podcasts/rethinking-with-adam-grant/khan-academy-founder-sal-khan-on-ai-and-the-future-of-education-transcript
> .
> (6) undefined.
> https://podcasts.apple.com/us/podcast/how-ai-could-save-not-destroy-education-sal-khan/id160904630?i=1000611325934
> .
> (7) undefined. https://www.ted.com/contact%29.
> (8) en.wikipedia.org. https://en.wikipedia.org/wiki/Sal_Khan.
>
> 2 In South Africa, we are currently facing a significant education crisis.
> While there are some excellent private schools, they come with a high price
> tag. On the other hand, the state of public schools is a complete disaster.
>
> During the COVID-19 lockdowns, I had the opportunity to work with children
> ranging from 6 to 16 years old, and I decided to adopt a radically
> different educational approach. This approach was inspired by Sugata
> Mittra's belief that teachers may not be necessary if the learning
> environment is well-managed (as elaborated below).
>
> As part of our informal schooling setup, we incorporated a daily "speech"
> session at 11 am. This session was not limited to just speaking; the
> children were also required to submit handwritten copies of their speeches
> and present them to the group. Following each child's presentation, we
> engaged in a Q&A and discussion session.
>
> Even though we didn't have access to ChatGPT at that time, if I were to
> recreate this experience today, I would encourage the children to utilize
> resources like ChatGPT and other freely available AI tools on the internet.
> This approach aligns with preparing them for the demands of today's
> workplace.
>
> Children possess an innate desire to learn, seek challenges, and take
> pride in their accomplishments. Providing them with maximum freedom within
> defined boundaries is an excellent approach to help them flourish.
>
> Reference, info on Sugata Mitra from the internet
> The School in the Cloud is a concept developed by Sugata Mitra, an
> educational researcher and TED Prize winner. He proposes that children can
> learn anything by themselves, with the help of the Internet and their
> peers, without the need for formal teaching or supervision. He calls this
> method Self-Organized Learning Environments (SOLE)¹.
>
> Mitra conducted several experiments to test his idea, such as the "Hole in
> the Wall" project, where he installed computers in slums and villages in
> India and observed how children learned to use them without any guidance¹.
> He also created learning labs in different locations, where children could
> access online resources and mentors, called the "Granny Cloud", who
> supported their curiosity and exploration².
>
> Mitra's vision is to transform the way children learn, by giving them more
> autonomy, creativity and collaboration. He believes that this approach can
> prepare them for the future, where memorization of facts is less important
> than critical thinking and problem-solving. He also invites anyone,
> anywhere, to join his global experiment in self-organized learning, by
> creating their own SOLEs and sharing their discoveries².
>
> The School in the Cloud is also the subject of a documentary by Jerry
> Rothwell, which follows Mitra's three-year experiment and the impact of his
> work on children and educators around the world³. You can also read more
> about Mitra's ideas and experiences in his book "The School in the Cloud:
> The Emerging Future of Learning"⁴..
>
> Source: Conversation with Bing, 2023/10/08
> (1) Sugata Mitra: Build a School in the Cloud | TED Talk.
> https://www.ted.com/talks/sugata_mitra_build_a_school_in_the_cloud.
> (2) Sugata Mitra creates a School in the Cloud | Prize-winning wishes |
> TED ....
> https://www.ted.com/participate/ted-prize/prize-winning-wishes/school-in-the-cloud
> .
> (3) The School in the Cloud - United Nations Western Europe.
> https://unric.org/en/the-school-in-the-cloud/
> <https://linkprotect.cudasvc.com/url?a=https%3a%2f%2funric.org%2fen%2fthe-school-in-the-cloud%2f&c=E,1,P04gf7AkTCBw5Pu_pIZ15Sa4DD1-gfP_8k-B2baFWOI2jWFHYyPbbaBsn0OsnINICl6TI9QThLvR2A5qOglBMS7zz_Tmtt90LZBHRFp-&typo=1>
> .
> (4) The School in the Cloud by Sugata Mitra (ebook) - eBooks.com
> <https://linkprotect.cudasvc.com/url?a=https%3a%2f%2feBooks.com&c=E,1,wsixh_9R4QKLkD8DtVqVcLh8kGp_7BfjsaQ5St9_onGfLCa4Ou6JZjv0T2UFtSbIYPqSp0o5p5t2ZJssBHfXMnLUgHb4gPP2_oEKNle93g,,&typo=1&ancr_add=1>
> https://www.ebooks.com/en-us/book/209765777/the-school-in-the-cloud/sugata-mitra/
> <https://linkprotect.cudasvc.com/url?a=https%3a%2f%2fwww.ebooks.com%2fen-us%2fbook%2f209765777%2fthe-school-in-the-cloud%2fsugata-mitra%2f&c=E,1,ALPLtj1Pd8Kx3SVvle5l5lE2km7dGj6PWIWblhALEROHD-5OllWa3vIAV_dpaEw1Tt9GJyLnLfyIb5iufGiV8vo3uAWeoE-zwANhJFQ5BQ,,&typo=1>
> .
> (5) undefined. https://www.ted.com/contact%29.
> (6) en.wikipedia.org. https://en.wikipedia.org/wiki/Sugata_Mitra.
>
> On Sun, 8 Oct 2023 at 11:20, David Eric Smith <desm...@santafe.edu> wrote:
>
>> In reading about the role of LLMs in personalization, I just realized how
>> to further destroy education.  As in many of these things, one starts with
>> an insight that is basically good, and then hands it to a SV company to be
>> turned into something profoundly nihilistic.
>>
>> The lecturer sits at “the front” of wherever this happens.  The body of
>> students sit in something that makes them manifestly a group.  But what the
>> lecturer says isn’t what they hear.  The lecturer speaks into a computer,
>> and the students all wear headsets (probably full VR immersions).  An
>> intervening LLM, which has been trained on exchanges with each student and
>> with the lecturer, replaces whatever the lecturer says with a
>> “personalized” version of the “content” which is then fed to each student.
>>
>> No two people will have heard quite the same version of anything.  I
>> guess, if they ever wanted to discuss the material with each other, another
>> LLM could be recruited to stand between them to convert what either of them
>> said into a different stream for the other to hear.  That can be an
>> additional service for being “in the class”, which one can purchase for a
>> fee, like the fee to get on an airplane earlier in the line than some other
>> passenger who paid a lower fee.
>>
>> The only thing I find weird is that I haven’t already head this suggested
>> 1000 times even in the popular press.  It seems like the first idea that
>> would occur to any SV engineer, or (though probably requiring a cognitive
>> exoskeleton) any university administrator.
>>
>> Eric
>>
>> Oh yeah, p.s.:  The insight that wasn’t evil is that this is the
>> difference between lectures and books or other broadcast media.  In a real
>> lecture, in the old world where people spent time with each other, the
>> teacher gets to know the students enough to learn what they can use, and
>> renders the material with empathy for that.  I’ll leave the reader to put
>> together which parts of the resulting exchange get left out of my LLM
>> implementation.
>>
>> p.p.s.  I should mention, has any of you read this book?
>> [image: 9780674279384-lg.jpg]
>>
>> Nihilistic Times — Wendy Brown
>> <https://www.hup.harvard.edu/catalog.php?isbn=9780674279384>
>> hup.harvard.edu
>> <https://www.hup.harvard.edu/catalog.php?isbn=9780674279384>
>> <https://www.hup.harvard.edu/catalog.php?isbn=9780674279384>
>>
>> I got to spend a little time in the orbit of this woman last year (didn’t
>> really “meet”, since I was just a quiet sitter-in in a group’s weekly
>> discussion), and she seems to me like just a fine mind.  The book is a
>> chautauqua of sorts, not aspiring to a destination per se.  But it is a
>> good combination of empathy, analytic clarity, and erudition, nicely
>> combined.  Bears on how many things make up the machinery of nihilism,
>> extending well beyond relations I recognized as part of an integration,
>> though mentioning many things I rail against in daily life.
>>
>>
>>
>>
>> On Oct 7, 2023, at 8:36 PM, Steve Smith <sasm...@swcp.com> wrote:
>>
>> Tom -
>>
>> This is probably a little less relevant to the business of journalism
>> than you deserve but this is a pretty good (IMO) fleshing out of some of
>> the technical details of (L)LMs which might fill in some blanks:
>>
>> https://developers.google.com/machine-learning/resources/intro-llms
>>
>> There must be (by now) some (yet) more relevant resources for Journalists
>> (for example), not only to de-Jargonize a little but also to make it (yet
>> more) relevant to your domain.
>>
>> Here, straight from the *Horses' Mouth* (GPT 3.5) itself ( I leave it to
>> you to tell us how well these answers work for you?).  It is probably worth
>> engaging with chatGPT yourself and asking more and more focused
>> questions...   a good self-study in "prompt engineering"!
>>
>> http://chat.openai.com
>> <https://linkprotect.cudasvc.com/url?a=http%3a%2f%2fchat.openai.com&c=E,1,5AQR_YXXM2e5NRiZ2jwiMxV8_p-R8Bm41a-VJIvezvBxlpk5Dl9j1kvUr04X8zxM96dWITUcudcORIY8fwPup8paM-YALfjNmU7xfsuVh_3ViFZl3ak,&typo=1>
>>
>> ***
>>
>> *Introduction to Modern Language Models for Veteran Journalists*
>>
>> *In today's digital age, the field of journalism is undergoing a
>> significant transformation, largely driven by advancements in technology.
>> One key innovation that has had a profound impact on the industry is the
>> development of modern language models, often abbreviated as LLMs. While you
>> may not be a tech expert, understanding the basics of LLMs can be
>> invaluable for journalists. So, let's break it down in simple terms.*
>>
>> ***What Are Language Models?***
>>
>> *At its core, a language model is like a super-smart digital writer. It's
>> a computer program trained to understand and generate human language. Think
>> of it as a tool that helps computers understand and use words, sentences,
>> and paragraphs just like humans do.*
>>
>> ***The Power of Large Language Models (LLMs)***
>>
>> *Now, let's talk about Large Language Models (LLMs). These are the big
>> players in the world of language models. They're like language
>> supercomputers because they're trained on massive amounts of text from the
>> internet. This means they've learned a lot about how people write and talk.*
>>
>> ***How LLMs Help Journalists***
>>
>> *1. **Content Generation**: LLMs can generate written content, from news
>> articles to summaries and even creative pieces. Journalists can use them to
>> draft stories faster and get ideas flowing.*
>>
>> *2. **Fact-Checking**: They're handy for fact-checking, helping you
>> quickly verify information by scanning a vast database of knowledge.*
>>
>> *3. **Summarization**: LLMs can take long, complex documents and create
>> concise summaries, saving you time on research.*
>>
>> *4. **Translation**: They can also translate text between languages,
>> which can be incredibly useful for international reporting.*
>>
>> *5. **Personalization**: Some news websites use LLMs to tailor content to
>> individual readers, providing a more personalized news experience.*
>>
>> *6. **Social Media Analysis**: LLMs can sift through social media data,
>> helping journalists understand public sentiment and trends.*
>>
>> ***Potential Concerns***
>>
>> *While LLMs are incredibly powerful, there are some concerns to keep in
>> mind. They might unintentionally spread misinformation if not used
>> carefully, and they can't replace human judgment or critical thinking.*
>>
>> ***Conclusion***
>>
>> *In a nutshell, modern language models (LLMs) are like digital writing
>> assistants that can help journalists write better, research faster, and
>> adapt to the changing landscape of digital journalism. While they're
>> powerful tools, it's important to use them responsibly, keeping in mind
>> their strengths and limitations, as well as the ethical considerations of
>> automated content creation in journalism.*
>>
>> Can you elaborate on some of the technical jargon?
>>
>> *Certainly! Here are some common technical jargon related to Large
>> Language Models (LLMs) explained in journalist-friendly terms:*
>>
>> *1. **Pretraining**: This is the initial phase where the LLM learns from
>> a massive amount of text data on the internet. Think of it as the model
>> absorbing a vast library of information to understand language better.*
>>
>> *2. **Fine-tuning**: After pretraining, LLMs undergo fine-tuning to
>> specialize them for specific tasks. It's like giving the model extra
>> training in a particular subject, like finance or healthcare, to make it
>> more useful for specialized reporting.*
>>
>> *3. **Token**: In LLMs, text is broken down into smaller units called
>> tokens. A token can be as short as one character or as long as one word.
>> Knowing the number of tokens is important because it affects how much text
>> the model can handle at once.*
>>
>> *4. **Inference**: This term refers to the process of using a trained LLM
>> to generate text or make predictions. Journalists use inference to create
>> content, answer questions, or analyze data.*
>>
>> *5. **Prompt**: A prompt is the input you provide to an LLM to get a
>> response. It's like giving the model a question or a starting point for it
>> to generate text. Journalists often craft prompts to get specific
>> information or insights.*
>>
>> *6. **Bias**: LLMs can sometimes produce biased or unfair content because
>> they've learned from biased internet data. Journalists should be aware of
>> this and critically evaluate the model's outputs for any biases.*
>>
>> *7. **Algorithm**: This is a fancy word for a set of rules or
>> instructions that an LLM follows to perform tasks. Algorithms guide how the
>> model generates text and makes decisions.*
>>
>> *8. **NLP (Natural Language Processing)**: NLP is the field of study that
>> focuses on teaching computers to understand and generate human language.
>> LLMs are a product of NLP research.*
>>
>> *9. **ML (Machine Learning)**: LLMs are a type of machine learning model.
>> Machine learning is a way to make computers learn from data and improve
>> their performance over time.*
>>
>> *10. **Training Data**: This is the text data used to teach an LLM. It's
>> like the model's textbooks, and the quality and diversity of this data
>> influence how well the LLM performs.*
>>
>> *11. **Bias Mitigation**: This is the process of reducing bias in LLMs.
>> It's important in journalism to ensure that the information provided by the
>> model is fair and unbiased.*
>>
>> *12. **Ethical AI**: This refers to the practice of using LLMs
>> responsibly and ethically. Journalists should consider the ethical
>> implications of using these tools, especially when it comes to issues like
>> privacy and misinformation.*
>>
>> *13. **API (Application Programming Interface)**: Some LLMs offer APIs
>> that allow journalists to interact with the model through code, enabling
>> custom applications and integrations.*
>>
>> *14. **Fine-Tuned Models**: These are LLMs that have been trained for
>> specific tasks, like summarization, translation, or question-answering.
>> They are more specialized and can be particularly useful for journalism.*
>>
>> *Understanding these terms can help journalists effectively leverage LLMs
>> in their work while also being mindful of their limitations and ethical
>> considerations.*
>>
>>
>> - Steve
>>
>> On 10/7/23 5:13 PM, Tom Johnson wrote:
>>
>> Thanks. Those are impressive numbers.
>> Tom
>>
>> On Sat, Oct 7, 2023 at 4:11 PM Marcus Daniels <mar...@snoutfarm.com>
>> wrote:
>>
>>> The “large” refers to the number of parameters used.  A smaller large
>>> language model – a deep neural net -- start about 3 billion parameters, but
>>> larger ones like Claude 2 (the latest large language model of the company
>>> that wrote the paper Steve mentioned) have more than 130 billion
>>> parameters.   Amazingly, it is possible using (rooms of) GPUs and other
>>> accelerators to optimize in this a space of this size.   The billions of
>>> parameters come from the vocabulary size – the number of tokens that need
>>> to be discriminated, the many layers of transformers that are needed to
>>> capture the complexity of human and non-human languages (like DNA), and the
>>> context window size – how many paragraphs or pages the model is trained on
>>> at a time.   A small language model might be suitable for understanding the
>>> geometries of chemicals, say.
>>>
>>>
>>> *From:* Friam <friam-boun...@redfish.com> *On Behalf Of *Tom Johnson
>>> *Sent:* Saturday, October 7, 2023 2:38 PM
>>> *To:* The Friday Morning Applied Complexity Coffee Group <
>>> friam@redfish.com>
>>> *Subject:* Re: [FRIAM] Language Model Understanding
>>>
>>>
>>> Thanks for passing this along, Steve. I wish, however, the authors of
>>> this short piece would have included a definition of, in their usage,
>>> "Large Language Models" and "Small Language Models."  Perhaps I can find
>>> those in the larger paper.
>>>
>>> Tom
>>>
>>>
>>> On Sat, Oct 7, 2023 at 12:34 PM Steve Smith <sasm...@swcp.com> wrote:
>>>
>>> This popular-press article came through my Google News feed recently
>>> which I thought might be useful to the Journalists/English-Majors on the
>>> list to help understand how LLMs work, etc.   When I read it in detail
>>> (forwarded from my TS (TinyScreenPhone) on my LS (Large Screen Laptop)) I
>>> found it a bit more detailed and technical than I'd expected, but
>>> nevertheless rewarding and possibly offering some traction to
>>> Journalism/English majors as well as those with a larger investment in the
>>> CS/Math implied.
>>>
>>> Decomposing Language Models into Understandable Components
>>>
>>> <https://linkprotect.cudasvc.com/url?a=https%3a%2f%2fwww.anthropic.com%2findex%2fdecomposing-language-models-into-understandable-components&c=E,1,m6E2NyQiQqZTmLRighr8jjGwfjDVwi00TYuM_4iskHnmAw78zWd55VlGTXwTdzWWkVGzInmkQvZFTQYuycBPX1Ii3Lxo1Bb5DZUFcSW0rUdv4Ma-WSM,&typo=1>
>>>
>>> and the (more) technical paper behind the article
>>>
>>> https://transformer-circuits.pub/2023/monosemantic-features/index.html
>>>
>>> <https://linkprotect.cudasvc.com/url?a=https%3a%2f%2ftransformer-circuits.pub%2f2023%2fmonosemantic-features%2findex.html&c=E,1,PIXQOjp_g2tr-D-_MpOt_B6uVdfil6N3FViwvJ78HkMFtSu7LpKOXlQSpqz_GQuJyD31X5eqiYtu2d2_VgXFqA4pekVbchFYvvf9GwiQ1zIgLfEt&typo=1>
>>>
>>> Despite having sent a few dogs into vaguely similar scuffles in my
>>> careen(r):
>>>
>>> Faceted Ontologies for Pre Incident Indicator Analysis
>>> <https://linkprotect.cudasvc.com/url?a=https%3a%2f%2fapps.dtic.mil%2fsti%2ftr%2fpdf%2fADA588086.pdf&c=E,1,nPyDgDuU0nMwlV3W24mUTteOTIfbrs9SipIKre1b1_XIxOIWJ1ok_IKz181AKiB6geR5Mb1z6-DrUslqOaBkYUxiG5O8Zhc2mG3ALmDJQjCULUXzR9v8IUhidw,,&typo=1>
>>> SpindleViz
>>> <https://linkprotect.cudasvc.com/url?a=https%3a%2f%2fwww.ehu.eus%2fccwintco%2fuploads%2fc%2fc6%2fHAIS2010_925.pdf&c=E,1,YwucoE_ffMQO_zO0A6WWAg8Y0J7wyHAGiJfcntyFBphchxgl4Tx6INQAtF7qA9XB07R52Jwnqfs4YO9s2MHP3Lnh-nUkpbULJfatlSIig_XB5ixy7yxF&typo=1>
>>> ...
>>>
>>> ... I admit to finding this both intriguing and well over my head on
>>> casual inspection...  the (metaphorical?) keywords that drew me in  most
>>> strongly included *Superposition* and *Thought Vectors*, though they
>>> are (nod to Glen) probably riddled (heaped, overflowing, bursting, bloated
>>> ... )  with excess meaning.
>>>
>>> https://gabgoh.github.io/ThoughtVectors/
>>> <https://linkprotect.cudasvc.com/url?a=https%3a%2f%2fgabgoh.github.io%2fThoughtVectors%2f&c=E,1,WnnKNla0iVte02BV2Imnk1J5revdzixjz7Z3w4XFowr2Ti3fkKSk8kKj8BQdnL11w3MtGPIKl2FP5Y-6zSsJ5O9A0luCjDPRrFpM_kDhF0jacJbhFA,,&typo=1>
>>>
>>> This leads me (surprise!) to an open ended discursive series of thoughts
>>> probably better left for a separate posting (probably rendered in a
>>> semasiographic language like Heptapod B
>>> <https://en.wikipedia.org/wiki/Heptapod_languages#Orthography>).
>>>
>>> <must... stop... now... >
>>>
>>> - Steve
>>>
>>> -. --- - / ...- .- .-.. .. -.. / -- --- .-. ... . / -.-. --- -.. .
>>> FRIAM Applied Complexity Group listserv
>>> Fridays 9a-12p Friday St. Johns Cafe   /   Thursdays 9a-12p Zoom
>>> https://bit.ly/virtualfriam
>>> <https://linkprotect.cudasvc.com/url?a=https%3a%2f%2fbit.ly%2fvirtualfriam&c=E,1,7Se1kU5K0qRIvWJRD4XXT6OVdefOkjmd8iBwB4_ScBm8nmwT9MUlDa2OlXu7gsTvtdy7EwO5VnLshsDYL-lYhTu-S1P_ak0_11QPeoudf_Pg&typo=1>
>>> to (un)subscribe http://redfish.com/mailman/listinfo/friam_redfish.com
>>> <https://linkprotect.cudasvc.com/url?a=http%3a%2f%2fredfish.com%2fmailman%2flistinfo%2ffriam_redfish.com&c=E,1,XgC24CHCqH-MiGAYACD7z6PtLS8L3PA6Ntnnm8-GFS-8K9DlJz1wbcz4YhFucq-Ucbr9oeSbxnsUcDSZ0XfT5aalKerL34Pq0R57j_mDeX4RrFvxbg,,&typo=1>
>>> FRIAM-COMIC http://friam-comic.blogspot.com/
>>> <https://linkprotect.cudasvc.com/url?a=http%3a%2f%2ffriam-comic.blogspot.com%2f&c=E,1,1GkLbYQODhOcaZjv5c91-iHu4986bkpXTFuaNhmX9fFK-EzJOVs6x1AKJrFBa7SMS9GHwxtjJZQ_CrzTmsDpTJ-KHGBzYStX4zSsywCdFT2GG8q4MO_gHwI,&typo=1>
>>> archives:  5/2017 thru present
>>> https://redfish.com/pipermail/friam_redfish.com/
>>> <https://linkprotect.cudasvc.com/url?a=https%3a%2f%2fredfish.com%2fpipermail%2ffriam_redfish.com%2f&c=E,1,pPB6Gx8uMJadpxScEftegu18poh3jvMC3dByqfb7-kH5q9a1uN-Hx18Hpf2mWKYR1aATb5b9IUgqwfoulMiH2oh8K03lnbespQNIBfFm-D_etq1aSira&typo=1>
>>>   1/2003 thru 6/2021  http://friam.383.s1.nabble.com/
>>>
>>> -. --- - / ...- .- .-.. .. -.. / -- --- .-. ... . / -.-. --- -.. .
>>> FRIAM Applied Complexity Group listserv
>>> Fridays 9a-12p Friday St. Johns Cafe   /   Thursdays 9a-12p Zoom
>>> https://bit.ly/virtualfriam
>>> <https://linkprotect.cudasvc.com/url?a=https%3a%2f%2fbit.ly%2fvirtualfriam&c=E,1,9R_t1NDn9DvauNKttOqQgxrGaUrdMJxeYm5N63GEWkEvm1uQHRIPj5EryTCG4F8-FBeu18NA6Iqbyzr0qZmJUE2HF57NAov7tcuXG1CFAJ7ISv0,&typo=1>
>>> to (un)subscribe http://redfish.com/mailman/listinfo/friam_redfish.com
>>> <https://linkprotect.cudasvc.com/url?a=http%3a%2f%2fredfish.com%2fmailman%2flistinfo%2ffriam_redfish.com&c=E,1,lFTqZFH-R9D5jiRQ0m9XGZNNd5Rsr-TYbo8WKOmosICjS8S7fpAggB06ojoO4DIXx07lOw0UyZuhpBT_IsxNgIEKtnTtzzaLELhd606vpB5mrMmIKA,,&typo=1>
>>> FRIAM-COMIC http://friam-comic.blogspot.com/
>>> <https://linkprotect.cudasvc.com/url?a=http%3a%2f%2ffriam-comic.blogspot.com%2f&c=E,1,xRMd22AvNM1vduoxz-Cd6-edLdgmkb_sifY1_C_kq0J3nQo_3JxTOJwh26ybgsQDy5PV7BrokU2_MLcMFRgluXk3WJSSCAoZVLgtMCpIcY07gVC_vTKmq4s,&typo=1>
>>> archives:  5/2017 thru present
>>> https://redfish.com/pipermail/friam_redfish.com/
>>> <https://linkprotect.cudasvc.com/url?a=https%3a%2f%2fredfish.com%2fpipermail%2ffriam_redfish.com%2f&c=E,1,KQ2mCnilVXt2leucw3caOQ8GNGViF5aUsirFIbdwFgZJeCh48I4mlovIE97rJIszL6HttaJg58sbOf85zTBi0aq7vxRF9a4Q94NIPMK5kpk3Gbb-tTFwcOZ0pNE,&typo=1>
>>>   1/2003 thru 6/2021  http://friam.383.s1.nabble.com/
>>>
>>
>> -. --- - / ...- .- .-.. .. -.. / -- --- .-. ... . / -.-. --- -.. .
>> FRIAM Applied Complexity Group listserv
>> Fridays 9a-12p Friday St. Johns Cafe   /   Thursdays 9a-12p Zoom 
>> https://bit.ly/virtualfriam 
>> <https://linkprotect.cudasvc.com/url?a=https%3a%2f%2fbit.ly%2fvirtualfriam&c=E,1,JZG5Ry4qLJlJfRcyvniC-8nt5cXg6jFMhE3BkxTASN407yQ-YlvbXpA2Zd_l4ZYoxSROdQ-iJzUyJID9bt1wqtH1fEcb3cdiL6OmzuOItbP4tmY,&typo=1>
>> to (un)subscribe http://redfish.com/mailman/listinfo/friam_redfish.com 
>> <https://linkprotect.cudasvc.com/url?a=http%3a%2f%2fredfish.com%2fmailman%2flistinfo%2ffriam_redfish.com&c=E,1,iThoLpYpqt8aVw23fOZyibIh1jKmvxMDkX9wTBRonsR8X4ezwG2-HhOu2w5yW0lyX3zmbFgbfy7jDe1gj23sTQK_OWFEcLCsJGKJhFl5dSqaSII6Om4d0nzZQBo,&typo=1>
>> FRIAM-COMIC http://friam-comic.blogspot.com/ 
>> <https://linkprotect.cudasvc.com/url?a=http%3a%2f%2ffriam-comic.blogspot.com%2f&c=E,1,HYzkgvDtQjTwOa5_t7MZnsd_eXg7gQdb3c3PK_qnRyw9jdfOnrNAu3PtDOXQrjRLIJbkJDvSCzh2me61wHHIThc6Y2xNFIS7zHhMQAzK&typo=1>
>> archives:  5/2017 thru present 
>> https://redfish.com/pipermail/friam_redfish.com/ 
>> <https://linkprotect.cudasvc.com/url?a=https%3a%2f%2fredfish.com%2fpipermail%2ffriam_redfish.com%2f&c=E,1,tQoVq56CqO6ncvHKTN9hbcVTg93ckGFVj0aCH7mntnycdBP8xmtUmWQk6vfH6tXlmj4efSkAfKSAQslYZ9xESLjg64AhcTLO8gILEjPUH05xuw,,&typo=1>
>>   1/2003 thru 6/2021  http://friam.383.s1.nabble.com/
>>
>> -. --- - / ...- .- .-.. .. -.. / -- --- .-. ... . / -.-. --- -.. .
>> FRIAM Applied Complexity Group listserv
>> Fridays 9a-12p Friday St. Johns Cafe   /   Thursdays 9a-12p Zoom
>> https://linkprotect.cudasvc.com/url?a=https%3a%2f%2fbit.ly%2fvirtualfriam&c=E,1,Wmm9ThSA6Mz3l6f44HjYltfKE5iYDLD_dMQDomanWFe33hPZtOZx6UMs9HfvlE9qqDrpJHTiJq-6I7rnKON9ssVZvrW3vYhYayTxfEkBPA,,&typo=1
>> to (un)subscribe
>> https://linkprotect.cudasvc.com/url?a=http%3a%2f%2fredfish.com%2fmailman%2flistinfo%2ffriam_redfish.com&c=E,1,SCq73BTJlGKctIkbzRtoaDbciyn7yKRCVtCpmNnPBqLEajOEfO3ae1cuIeemIU9GwwAYpzKzxvR6walpGzl7rZXKHBzAtoeNy6j6mqtwhLpEMFRuv0XoITA,&typo=1
>> FRIAM-COMIC
>> https://linkprotect.cudasvc.com/url?a=http%3a%2f%2ffriam-comic.blogspot.com%2f&c=E,1,RNz26jMqvi3Xb0sjg5eWouWJOXinbcqzg63a-tE4P-K1Rmv4GlXJtlAaD8-Ij1iv6C6L7JF9LZzTygimbNX8G_fg7e6fKklz4FYSTigA5Q,,&typo=1
>> archives:  5/2017 thru present
>> https://linkprotect.cudasvc.com/url?a=https%3a%2f%2fredfish.com%2fpipermail%2ffriam_redfish.com%2f&c=E,1,S1eeYxStqxoVg4yeuKl0cDdLaMoPcak1Ri2w3UANtBaag5riZ9K-IGLJIia0qNnuf-ChNm1tzsBtHZOtE9Wqe-yu2oKWkUjMwJoTFYVpi6adLzNpgY4iVPs,&typo=1
>>  1/2003 thru 6/2021  http://friam.383.s1.nabble.com/
>>
>>
>> -. --- - / ...- .- .-.. .. -.. / -- --- .-. ... . / -.-. --- -.. .
>> FRIAM Applied Complexity Group listserv
>> Fridays 9a-12p Friday St. Johns Cafe   /   Thursdays 9a-12p Zoom
>> https://bit.ly/virtualfriam
>> <https://linkprotect.cudasvc.com/url?a=https%3a%2f%2fbit.ly%2fvirtualfriam&c=E,1,tYXe7u30UizIRX7sCP5z134Vp_rCgZrx-HG1QllHUyYO2zfuWxe9oJB0rJBMD_fbOkbLIYPpr2aium2bSNugjk7zYsHr5bk-DHPLj56i&typo=1>
>> to (un)subscribe http://redfish.com/mailman/listinfo/friam_redfish.com
>> <https://linkprotect.cudasvc.com/url?a=http%3a%2f%2fredfish.com%2fmailman%2flistinfo%2ffriam_redfish.com&c=E,1,owxu2tT3Pnf6XPyExtN00oNk919A19XBuIfw_2ugWaBuigwyq7DFJlupgseFZ7dOIgMUnH_x2z2nLX7lUkJsFKOeHX0DTNUNT-mUAW5rsP8,&typo=1>
>> FRIAM-COMIC http://friam-comic.blogspot.com/
>> <https://linkprotect.cudasvc.com/url?a=http%3a%2f%2ffriam-comic.blogspot.com%2f&c=E,1,0Dk21FivClJzcKxULsvq2ahHAaXAtwwsVxvRPMjYjKT8CN4LNIQgsTKYNRQbVsufJP8QWIPIZCVEAlkV78bQiWk7RfliP9Sp2Abk7ij92YlC6TmsJg,,&typo=1>
>> archives:  5/2017 thru present
>> https://redfish.com/pipermail/friam_redfish.com/
>> <https://linkprotect.cudasvc.com/url?a=https%3a%2f%2fredfish.com%2fpipermail%2ffriam_redfish.com%2f&c=E,1,xoH9CNhv1c36upLgpvFH9gpB7s-7TuKIBNfryFN2vs8QLRc8b7-KRpdiWxQKXL0o5ljIufdg0z0pvgn77zbqUXF-rGapyf9MPmtRBoxD6vgoFz4,&typo=1>
>>   1/2003 thru 6/2021  http://friam.383.s1.nabble.com/
>>
> <9780674279384-lg.jpg><9780674279384-lg.jpg>-. --- - / ...- .- .-.. ..
> -.. / -- --- .-. ... . / -.-. --- -.. .
> FRIAM Applied Complexity Group listserv
> Fridays 9a-12p Friday St. Johns Cafe   /   Thursdays 9a-12p Zoom
> https://linkprotect.cudasvc.com/url?a=https%3a%2f%2fbit.ly%2fvirtualfriam&c=E,1,2ZlC62KWqH1N6zMvqal4Ed0Wx5XQGMEddd41-jtZtFjgO1SirsDGzGIktZ5uuoAX15NFz0taAZNEgGPxd34ZeId3Oi-wtbhVUHXBugIhOw36&typo=1
> to (un)subscribe
> https://linkprotect.cudasvc.com/url?a=http%3a%2f%2fredfish.com%2fmailman%2flistinfo%2ffriam_redfish.com&c=E,1,QZaSg2mbVJz2oN4YNuVpvuZjAkolp8mlWkppzQ0-aoqWNBnx1M7v90qS6byWzuUXDqyzMJNRAQn26_WVFsAlr_i3r1TLR4E0nLYs6alQCEdr-CStPCw,&typo=1
> FRIAM-COMIC
> https://linkprotect.cudasvc.com/url?a=http%3a%2f%2ffriam-comic.blogspot.com%2f&c=E,1,6pvfI2nyTTYQmON_tKGUbLCRCzgFGvS9QLKctMYsBFDU0Jezuy2pxUKnT-jJ7-TUHFK5mRy3GkmvuVKm8PhtSGxRaWgt-3n8l-1NT599Ul54D_JdL5xw&typo=1
> archives:  5/2017 thru present
> https://linkprotect.cudasvc.com/url?a=https%3a%2f%2fredfish.com%2fpipermail%2ffriam_redfish.com%2f&c=E,1,Od6NS37sWb2LiDsHMSlWPpueHD3xdUe4deNxdljKnlJcfHERLOFE6iw9WOQqHlD3uRhr7Si6KXivdjch6wdVhh-QOKc1CsO9-2NvVI5dfgGO3M8R2Wt2C_VpvA,,&typo=1
>  1/2003 thru 6/2021  http://friam.383.s1.nabble.com/
>
>
> -. --- - / ...- .- .-.. .. -.. / -- --- .-. ... . / -.-. --- -.. .
> FRIAM Applied Complexity Group listserv
> Fridays 9a-12p Friday St. Johns Cafe   /   Thursdays 9a-12p Zoom
> https://bit.ly/virtualfriam
> to (un)subscribe http://redfish.com/mailman/listinfo/friam_redfish.com
> FRIAM-COMIC http://friam-comic.blogspot.com/
> archives:  5/2017 thru present
> https://redfish.com/pipermail/friam_redfish.com/
>   1/2003 thru 6/2021  http://friam.383.s1.nabble.com/
>
-. --- - / ...- .- .-.. .. -.. / -- --- .-. ... . / -.-. --- -.. .
FRIAM Applied Complexity Group listserv
Fridays 9a-12p Friday St. Johns Cafe   /   Thursdays 9a-12p Zoom 
https://bit.ly/virtualfriam
to (un)subscribe http://redfish.com/mailman/listinfo/friam_redfish.com
FRIAM-COMIC http://friam-comic.blogspot.com/
archives:  5/2017 thru present https://redfish.com/pipermail/friam_redfish.com/
  1/2003 thru 6/2021  http://friam.383.s1.nabble.com/

Reply via email to