The book sounds intriguing, but it's not in my lending library. Happily, there are lots of Tanner Lecture videos available online: https://www.youtube.com/watch?v=yI-MCqeCILs Wendy Brown from 2019 https://www.youtube.com/watch?v=NJprCvmrpoY Kim Stanley Robinson from March 2023
-- rec -- -- rec -- On Sun, Oct 8, 2023 at 11:56 AM David Eric Smith <desm...@santafe.edu> wrote: > It’s a good point/counterpinet, Pieter. > > I don’t want to stand against using new tools to relieve a scarcity > constraint, and surely education has been a severe one. > > I have what appears to be an enormous suspicion of a bait and switch, in > which suddenly we find ourselves in a world where simultaneously: > > 1. Sal Khan and Daphne Koller own everything, and nobody else owns > anything; and > > 2. People have no skills (say I, as I use the scary-good google translate > to take enormous pressure, though less shame, off me for not learning a > language I should have learned by now). > > Or, as a popular bumper sticker in Santa Fe used to say post-2016: Where > are we going? And what am I doing in this handbasket? > > I should mention, on this point, that part of my cast of mind was shaped > by an article I read some time ago in some education chronicle, about how > the command-and-control interface to the computer was taking up more and > more space in classrooms, and kids were starting to show severe systematic > deficits in particular areas. They could command anything they wanted into > existence on the computer, and failed to learn that the rest of the world > of people doesn’t work that way. Negotiating skills were mentioned in the > article. My description of a dystopian non-education was meant to suggest > that for several students to have to hear exactly the same thing would > require them to come to terms with the fact that it will fit them > differently, and they should become used to handling that too. > > It is interesting that I could never use a chatbot as a first stage of > writing something, as the optimistic chatbots suggest everybody will do, > and which I can believe for some people will be a good fit. For me, I > can’t even take looking at a page of my own writing when I am in a tangle, > as a start to getting out of it. All that pre-formed text is in the way of > whatever part of my brain tries to bring into focus what I should do next. > I know that chatbots and writing is not your point about education, and > don’t mean to derail those points; the chatbots-and-journalism theme was > one of the things from the original thread that put me onto the education > sidetrack. What (at least in my imagination) they have in common is that > there are certain mental operations that I only do when I have “space”, and > many of these tools seem designed to make sure nobody ever has to leave any > “space” unfilled by some external stimulus, ever again. I worry about it > as a development-weakening prosthetic, though I understand that the > argument for that would need to be made case by case, versus the likely > (and I believe you, in many cases demonstrated) economies and advantages of > personalization. > > Fortunately, I am neither designing anything nor voting on anything just > now, so there is no danger I will hold anybody else back who is trying to > make something work. > > Eric > > > On Oct 8, 2023, at 7:45 AM, Pieter Steenekamp <piet...@randcontrols.co.za> > wrote: > > I'd like to reflect how LLM's and AI could disrupt education. > > My paradigm is not to think in terms of the current educational model, but > to throw out all preconceived ideas and to design it from scratch. There > are obviously zillions of ways to do it and a good model will need to > evolve, but the following are merely two examples to trigger new and novel > approaches, one from Sal Khan and one from my own experience. > > 1 Sal Khan's approach to use AI to personalize education is based on the > idea that every student can benefit from a customized learning path that > adapts to their needs and preferences. He argues that AI can help scale the > benefits of one-to-one tutoring, which has been shown to improve students' > performance significantly. He also believes that AI can enhance the role of > teachers by providing them with useful feedback, insights, and assistance. > He envisions a future where every student has access to an AI-powered > personal tutor and every teacher has an AI teaching assistant. > > Sal Khan is the founder and CEO of Khan Academy, a nonprofit education > platform that offers free online courses, videos, and exercises for > learners of all ages. He has been a pioneer in using technology to > democratize education and make it accessible to anyone, anywhere. He has > also been exploring the potential of AI to transform education for the > better. In his 2023 TED talk¹, he demonstrated Khanmigo, an AI-powered > guide that can help students with various subjects, skills, and tasks. > Khanmigo can detect students' mistakes and misconceptions, provide > effective feedback and explanations, encourage Socratic dialogue and > debate, and assist with writing and storytelling. Khanmigo also acts as a > teaching assistant for teachers, helping them with lesson planning, > grading, and progress reports. > > Sal Khan's approach to use AI to personalize education is based on his > vision of creating a more humanistic and learner-centered education system. > He believes that AI can empower students to learn at their own pace, > explore their interests, and develop their creativity. He also believes > that AI can support teachers to focus more on the social and emotional > aspects of teaching, such as building relationships, fostering curiosity, > and inspiring passion. He hopes that by using AI to personalize education, > we can create a more equitable and inclusive learning environment for > everyone.. > > Source: Conversation with Bing, 2023/10/08 > (1) . https://bing.com/search?q=Sal+Khan+AI+education. > (2) Sal Khan’s 2023 TED Talk: AI in the classroom can transform education. > https://blog.khanacademy.org/sal-khans-2023-ted-talk-ai-in-the-classroom-can-transform-education/ > . > (3) Sal Khan: How AI could save (not destroy) education | TED Talk. > https://www.ted.com/talks/sal_khan_how_ai_could_save_not_destroy_education/c > . > (4) The Future Of AI And Education: Insights From Sal Khan. > https://www.openbusinesscouncil.org/the-future-of-ai-and-education-insights-from-sal-khan-in-conversation-with-bill-gates > <https://linkprotect.cudasvc.com/url?a=https%3a%2f%2fwww.openbusinesscouncil.org%2fthe-future-of-ai-and-education-insights-from-sal-khan-in-conversation-with-bill-gates&c=E,1,2zh-o7p-ZRG5J2pDpv8lLYBDKVFUgyvlCSvOCnrGzy0OEqHE_M-1ND6g52xVjiUgyBncevhmKdnOnnuuTkeZ66oyJXl1nY6Pan2Zr01pkGQWF2U,&typo=1> > . > (5) undefined. > https://www.ted.com/podcasts/rethinking-with-adam-grant/khan-academy-founder-sal-khan-on-ai-and-the-future-of-education-transcript > . > (6) undefined. > https://podcasts.apple.com/us/podcast/how-ai-could-save-not-destroy-education-sal-khan/id160904630?i=1000611325934 > . > (7) undefined. https://www.ted.com/contact%29. > (8) en.wikipedia.org. https://en.wikipedia.org/wiki/Sal_Khan. > > 2 In South Africa, we are currently facing a significant education crisis. > While there are some excellent private schools, they come with a high price > tag. On the other hand, the state of public schools is a complete disaster. > > During the COVID-19 lockdowns, I had the opportunity to work with children > ranging from 6 to 16 years old, and I decided to adopt a radically > different educational approach. This approach was inspired by Sugata > Mittra's belief that teachers may not be necessary if the learning > environment is well-managed (as elaborated below). > > As part of our informal schooling setup, we incorporated a daily "speech" > session at 11 am. This session was not limited to just speaking; the > children were also required to submit handwritten copies of their speeches > and present them to the group. Following each child's presentation, we > engaged in a Q&A and discussion session. > > Even though we didn't have access to ChatGPT at that time, if I were to > recreate this experience today, I would encourage the children to utilize > resources like ChatGPT and other freely available AI tools on the internet. > This approach aligns with preparing them for the demands of today's > workplace. > > Children possess an innate desire to learn, seek challenges, and take > pride in their accomplishments. Providing them with maximum freedom within > defined boundaries is an excellent approach to help them flourish. > > Reference, info on Sugata Mitra from the internet > The School in the Cloud is a concept developed by Sugata Mitra, an > educational researcher and TED Prize winner. He proposes that children can > learn anything by themselves, with the help of the Internet and their > peers, without the need for formal teaching or supervision. He calls this > method Self-Organized Learning Environments (SOLE)¹. > > Mitra conducted several experiments to test his idea, such as the "Hole in > the Wall" project, where he installed computers in slums and villages in > India and observed how children learned to use them without any guidance¹. > He also created learning labs in different locations, where children could > access online resources and mentors, called the "Granny Cloud", who > supported their curiosity and exploration². > > Mitra's vision is to transform the way children learn, by giving them more > autonomy, creativity and collaboration. He believes that this approach can > prepare them for the future, where memorization of facts is less important > than critical thinking and problem-solving. He also invites anyone, > anywhere, to join his global experiment in self-organized learning, by > creating their own SOLEs and sharing their discoveries². > > The School in the Cloud is also the subject of a documentary by Jerry > Rothwell, which follows Mitra's three-year experiment and the impact of his > work on children and educators around the world³. You can also read more > about Mitra's ideas and experiences in his book "The School in the Cloud: > The Emerging Future of Learning"⁴.. > > Source: Conversation with Bing, 2023/10/08 > (1) Sugata Mitra: Build a School in the Cloud | TED Talk. > https://www.ted.com/talks/sugata_mitra_build_a_school_in_the_cloud. > (2) Sugata Mitra creates a School in the Cloud | Prize-winning wishes | > TED .... > https://www.ted.com/participate/ted-prize/prize-winning-wishes/school-in-the-cloud > . > (3) The School in the Cloud - United Nations Western Europe. > https://unric.org/en/the-school-in-the-cloud/ > <https://linkprotect.cudasvc.com/url?a=https%3a%2f%2funric.org%2fen%2fthe-school-in-the-cloud%2f&c=E,1,P04gf7AkTCBw5Pu_pIZ15Sa4DD1-gfP_8k-B2baFWOI2jWFHYyPbbaBsn0OsnINICl6TI9QThLvR2A5qOglBMS7zz_Tmtt90LZBHRFp-&typo=1> > . > (4) The School in the Cloud by Sugata Mitra (ebook) - eBooks.com > <https://linkprotect.cudasvc.com/url?a=https%3a%2f%2feBooks.com&c=E,1,wsixh_9R4QKLkD8DtVqVcLh8kGp_7BfjsaQ5St9_onGfLCa4Ou6JZjv0T2UFtSbIYPqSp0o5p5t2ZJssBHfXMnLUgHb4gPP2_oEKNle93g,,&typo=1&ancr_add=1> > https://www.ebooks.com/en-us/book/209765777/the-school-in-the-cloud/sugata-mitra/ > <https://linkprotect.cudasvc.com/url?a=https%3a%2f%2fwww.ebooks.com%2fen-us%2fbook%2f209765777%2fthe-school-in-the-cloud%2fsugata-mitra%2f&c=E,1,ALPLtj1Pd8Kx3SVvle5l5lE2km7dGj6PWIWblhALEROHD-5OllWa3vIAV_dpaEw1Tt9GJyLnLfyIb5iufGiV8vo3uAWeoE-zwANhJFQ5BQ,,&typo=1> > . > (5) undefined. https://www.ted.com/contact%29. > (6) en.wikipedia.org. https://en.wikipedia.org/wiki/Sugata_Mitra. > > On Sun, 8 Oct 2023 at 11:20, David Eric Smith <desm...@santafe.edu> wrote: > >> In reading about the role of LLMs in personalization, I just realized how >> to further destroy education. As in many of these things, one starts with >> an insight that is basically good, and then hands it to a SV company to be >> turned into something profoundly nihilistic. >> >> The lecturer sits at “the front” of wherever this happens. The body of >> students sit in something that makes them manifestly a group. But what the >> lecturer says isn’t what they hear. The lecturer speaks into a computer, >> and the students all wear headsets (probably full VR immersions). An >> intervening LLM, which has been trained on exchanges with each student and >> with the lecturer, replaces whatever the lecturer says with a >> “personalized” version of the “content” which is then fed to each student. >> >> No two people will have heard quite the same version of anything. I >> guess, if they ever wanted to discuss the material with each other, another >> LLM could be recruited to stand between them to convert what either of them >> said into a different stream for the other to hear. That can be an >> additional service for being “in the class”, which one can purchase for a >> fee, like the fee to get on an airplane earlier in the line than some other >> passenger who paid a lower fee. >> >> The only thing I find weird is that I haven’t already head this suggested >> 1000 times even in the popular press. It seems like the first idea that >> would occur to any SV engineer, or (though probably requiring a cognitive >> exoskeleton) any university administrator. >> >> Eric >> >> Oh yeah, p.s.: The insight that wasn’t evil is that this is the >> difference between lectures and books or other broadcast media. In a real >> lecture, in the old world where people spent time with each other, the >> teacher gets to know the students enough to learn what they can use, and >> renders the material with empathy for that. I’ll leave the reader to put >> together which parts of the resulting exchange get left out of my LLM >> implementation. >> >> p.p.s. I should mention, has any of you read this book? >> [image: 9780674279384-lg.jpg] >> >> Nihilistic Times — Wendy Brown >> <https://www.hup.harvard.edu/catalog.php?isbn=9780674279384> >> hup.harvard.edu >> <https://www.hup.harvard.edu/catalog.php?isbn=9780674279384> >> <https://www.hup.harvard.edu/catalog.php?isbn=9780674279384> >> >> I got to spend a little time in the orbit of this woman last year (didn’t >> really “meet”, since I was just a quiet sitter-in in a group’s weekly >> discussion), and she seems to me like just a fine mind. The book is a >> chautauqua of sorts, not aspiring to a destination per se. But it is a >> good combination of empathy, analytic clarity, and erudition, nicely >> combined. Bears on how many things make up the machinery of nihilism, >> extending well beyond relations I recognized as part of an integration, >> though mentioning many things I rail against in daily life. >> >> >> >> >> On Oct 7, 2023, at 8:36 PM, Steve Smith <sasm...@swcp.com> wrote: >> >> Tom - >> >> This is probably a little less relevant to the business of journalism >> than you deserve but this is a pretty good (IMO) fleshing out of some of >> the technical details of (L)LMs which might fill in some blanks: >> >> https://developers.google.com/machine-learning/resources/intro-llms >> >> There must be (by now) some (yet) more relevant resources for Journalists >> (for example), not only to de-Jargonize a little but also to make it (yet >> more) relevant to your domain. >> >> Here, straight from the *Horses' Mouth* (GPT 3.5) itself ( I leave it to >> you to tell us how well these answers work for you?). It is probably worth >> engaging with chatGPT yourself and asking more and more focused >> questions... a good self-study in "prompt engineering"! >> >> http://chat.openai.com >> <https://linkprotect.cudasvc.com/url?a=http%3a%2f%2fchat.openai.com&c=E,1,5AQR_YXXM2e5NRiZ2jwiMxV8_p-R8Bm41a-VJIvezvBxlpk5Dl9j1kvUr04X8zxM96dWITUcudcORIY8fwPup8paM-YALfjNmU7xfsuVh_3ViFZl3ak,&typo=1> >> >> *** >> >> *Introduction to Modern Language Models for Veteran Journalists* >> >> *In today's digital age, the field of journalism is undergoing a >> significant transformation, largely driven by advancements in technology. >> One key innovation that has had a profound impact on the industry is the >> development of modern language models, often abbreviated as LLMs. While you >> may not be a tech expert, understanding the basics of LLMs can be >> invaluable for journalists. So, let's break it down in simple terms.* >> >> ***What Are Language Models?*** >> >> *At its core, a language model is like a super-smart digital writer. It's >> a computer program trained to understand and generate human language. Think >> of it as a tool that helps computers understand and use words, sentences, >> and paragraphs just like humans do.* >> >> ***The Power of Large Language Models (LLMs)*** >> >> *Now, let's talk about Large Language Models (LLMs). These are the big >> players in the world of language models. They're like language >> supercomputers because they're trained on massive amounts of text from the >> internet. This means they've learned a lot about how people write and talk.* >> >> ***How LLMs Help Journalists*** >> >> *1. **Content Generation**: LLMs can generate written content, from news >> articles to summaries and even creative pieces. Journalists can use them to >> draft stories faster and get ideas flowing.* >> >> *2. **Fact-Checking**: They're handy for fact-checking, helping you >> quickly verify information by scanning a vast database of knowledge.* >> >> *3. **Summarization**: LLMs can take long, complex documents and create >> concise summaries, saving you time on research.* >> >> *4. **Translation**: They can also translate text between languages, >> which can be incredibly useful for international reporting.* >> >> *5. **Personalization**: Some news websites use LLMs to tailor content to >> individual readers, providing a more personalized news experience.* >> >> *6. **Social Media Analysis**: LLMs can sift through social media data, >> helping journalists understand public sentiment and trends.* >> >> ***Potential Concerns*** >> >> *While LLMs are incredibly powerful, there are some concerns to keep in >> mind. They might unintentionally spread misinformation if not used >> carefully, and they can't replace human judgment or critical thinking.* >> >> ***Conclusion*** >> >> *In a nutshell, modern language models (LLMs) are like digital writing >> assistants that can help journalists write better, research faster, and >> adapt to the changing landscape of digital journalism. While they're >> powerful tools, it's important to use them responsibly, keeping in mind >> their strengths and limitations, as well as the ethical considerations of >> automated content creation in journalism.* >> >> Can you elaborate on some of the technical jargon? >> >> *Certainly! Here are some common technical jargon related to Large >> Language Models (LLMs) explained in journalist-friendly terms:* >> >> *1. **Pretraining**: This is the initial phase where the LLM learns from >> a massive amount of text data on the internet. Think of it as the model >> absorbing a vast library of information to understand language better.* >> >> *2. **Fine-tuning**: After pretraining, LLMs undergo fine-tuning to >> specialize them for specific tasks. It's like giving the model extra >> training in a particular subject, like finance or healthcare, to make it >> more useful for specialized reporting.* >> >> *3. **Token**: In LLMs, text is broken down into smaller units called >> tokens. A token can be as short as one character or as long as one word. >> Knowing the number of tokens is important because it affects how much text >> the model can handle at once.* >> >> *4. **Inference**: This term refers to the process of using a trained LLM >> to generate text or make predictions. Journalists use inference to create >> content, answer questions, or analyze data.* >> >> *5. **Prompt**: A prompt is the input you provide to an LLM to get a >> response. It's like giving the model a question or a starting point for it >> to generate text. Journalists often craft prompts to get specific >> information or insights.* >> >> *6. **Bias**: LLMs can sometimes produce biased or unfair content because >> they've learned from biased internet data. Journalists should be aware of >> this and critically evaluate the model's outputs for any biases.* >> >> *7. **Algorithm**: This is a fancy word for a set of rules or >> instructions that an LLM follows to perform tasks. Algorithms guide how the >> model generates text and makes decisions.* >> >> *8. **NLP (Natural Language Processing)**: NLP is the field of study that >> focuses on teaching computers to understand and generate human language. >> LLMs are a product of NLP research.* >> >> *9. **ML (Machine Learning)**: LLMs are a type of machine learning model. >> Machine learning is a way to make computers learn from data and improve >> their performance over time.* >> >> *10. **Training Data**: This is the text data used to teach an LLM. It's >> like the model's textbooks, and the quality and diversity of this data >> influence how well the LLM performs.* >> >> *11. **Bias Mitigation**: This is the process of reducing bias in LLMs. >> It's important in journalism to ensure that the information provided by the >> model is fair and unbiased.* >> >> *12. **Ethical AI**: This refers to the practice of using LLMs >> responsibly and ethically. Journalists should consider the ethical >> implications of using these tools, especially when it comes to issues like >> privacy and misinformation.* >> >> *13. **API (Application Programming Interface)**: Some LLMs offer APIs >> that allow journalists to interact with the model through code, enabling >> custom applications and integrations.* >> >> *14. **Fine-Tuned Models**: These are LLMs that have been trained for >> specific tasks, like summarization, translation, or question-answering. >> They are more specialized and can be particularly useful for journalism.* >> >> *Understanding these terms can help journalists effectively leverage LLMs >> in their work while also being mindful of their limitations and ethical >> considerations.* >> >> >> - Steve >> >> On 10/7/23 5:13 PM, Tom Johnson wrote: >> >> Thanks. Those are impressive numbers. >> Tom >> >> On Sat, Oct 7, 2023 at 4:11 PM Marcus Daniels <mar...@snoutfarm.com> >> wrote: >> >>> The “large” refers to the number of parameters used. A smaller large >>> language model – a deep neural net -- start about 3 billion parameters, but >>> larger ones like Claude 2 (the latest large language model of the company >>> that wrote the paper Steve mentioned) have more than 130 billion >>> parameters. Amazingly, it is possible using (rooms of) GPUs and other >>> accelerators to optimize in this a space of this size. The billions of >>> parameters come from the vocabulary size – the number of tokens that need >>> to be discriminated, the many layers of transformers that are needed to >>> capture the complexity of human and non-human languages (like DNA), and the >>> context window size – how many paragraphs or pages the model is trained on >>> at a time. A small language model might be suitable for understanding the >>> geometries of chemicals, say. >>> >>> >>> *From:* Friam <friam-boun...@redfish.com> *On Behalf Of *Tom Johnson >>> *Sent:* Saturday, October 7, 2023 2:38 PM >>> *To:* The Friday Morning Applied Complexity Coffee Group < >>> friam@redfish.com> >>> *Subject:* Re: [FRIAM] Language Model Understanding >>> >>> >>> Thanks for passing this along, Steve. I wish, however, the authors of >>> this short piece would have included a definition of, in their usage, >>> "Large Language Models" and "Small Language Models." Perhaps I can find >>> those in the larger paper. >>> >>> Tom >>> >>> >>> On Sat, Oct 7, 2023 at 12:34 PM Steve Smith <sasm...@swcp.com> wrote: >>> >>> This popular-press article came through my Google News feed recently >>> which I thought might be useful to the Journalists/English-Majors on the >>> list to help understand how LLMs work, etc. When I read it in detail >>> (forwarded from my TS (TinyScreenPhone) on my LS (Large Screen Laptop)) I >>> found it a bit more detailed and technical than I'd expected, but >>> nevertheless rewarding and possibly offering some traction to >>> Journalism/English majors as well as those with a larger investment in the >>> CS/Math implied. >>> >>> Decomposing Language Models into Understandable Components >>> >>> <https://linkprotect.cudasvc.com/url?a=https%3a%2f%2fwww.anthropic.com%2findex%2fdecomposing-language-models-into-understandable-components&c=E,1,m6E2NyQiQqZTmLRighr8jjGwfjDVwi00TYuM_4iskHnmAw78zWd55VlGTXwTdzWWkVGzInmkQvZFTQYuycBPX1Ii3Lxo1Bb5DZUFcSW0rUdv4Ma-WSM,&typo=1> >>> >>> and the (more) technical paper behind the article >>> >>> https://transformer-circuits.pub/2023/monosemantic-features/index.html >>> >>> <https://linkprotect.cudasvc.com/url?a=https%3a%2f%2ftransformer-circuits.pub%2f2023%2fmonosemantic-features%2findex.html&c=E,1,PIXQOjp_g2tr-D-_MpOt_B6uVdfil6N3FViwvJ78HkMFtSu7LpKOXlQSpqz_GQuJyD31X5eqiYtu2d2_VgXFqA4pekVbchFYvvf9GwiQ1zIgLfEt&typo=1> >>> >>> Despite having sent a few dogs into vaguely similar scuffles in my >>> careen(r): >>> >>> Faceted Ontologies for Pre Incident Indicator Analysis >>> <https://linkprotect.cudasvc.com/url?a=https%3a%2f%2fapps.dtic.mil%2fsti%2ftr%2fpdf%2fADA588086.pdf&c=E,1,nPyDgDuU0nMwlV3W24mUTteOTIfbrs9SipIKre1b1_XIxOIWJ1ok_IKz181AKiB6geR5Mb1z6-DrUslqOaBkYUxiG5O8Zhc2mG3ALmDJQjCULUXzR9v8IUhidw,,&typo=1> >>> SpindleViz >>> <https://linkprotect.cudasvc.com/url?a=https%3a%2f%2fwww.ehu.eus%2fccwintco%2fuploads%2fc%2fc6%2fHAIS2010_925.pdf&c=E,1,YwucoE_ffMQO_zO0A6WWAg8Y0J7wyHAGiJfcntyFBphchxgl4Tx6INQAtF7qA9XB07R52Jwnqfs4YO9s2MHP3Lnh-nUkpbULJfatlSIig_XB5ixy7yxF&typo=1> >>> ... >>> >>> ... I admit to finding this both intriguing and well over my head on >>> casual inspection... the (metaphorical?) keywords that drew me in most >>> strongly included *Superposition* and *Thought Vectors*, though they >>> are (nod to Glen) probably riddled (heaped, overflowing, bursting, bloated >>> ... ) with excess meaning. >>> >>> https://gabgoh.github.io/ThoughtVectors/ >>> <https://linkprotect.cudasvc.com/url?a=https%3a%2f%2fgabgoh.github.io%2fThoughtVectors%2f&c=E,1,WnnKNla0iVte02BV2Imnk1J5revdzixjz7Z3w4XFowr2Ti3fkKSk8kKj8BQdnL11w3MtGPIKl2FP5Y-6zSsJ5O9A0luCjDPRrFpM_kDhF0jacJbhFA,,&typo=1> >>> >>> This leads me (surprise!) to an open ended discursive series of thoughts >>> probably better left for a separate posting (probably rendered in a >>> semasiographic language like Heptapod B >>> <https://en.wikipedia.org/wiki/Heptapod_languages#Orthography>). >>> >>> <must... stop... now... > >>> >>> - Steve >>> >>> -. --- - / ...- .- .-.. .. -.. / -- --- .-. ... . / -.-. --- -.. . >>> FRIAM Applied Complexity Group listserv >>> Fridays 9a-12p Friday St. Johns Cafe / Thursdays 9a-12p Zoom >>> https://bit.ly/virtualfriam >>> <https://linkprotect.cudasvc.com/url?a=https%3a%2f%2fbit.ly%2fvirtualfriam&c=E,1,7Se1kU5K0qRIvWJRD4XXT6OVdefOkjmd8iBwB4_ScBm8nmwT9MUlDa2OlXu7gsTvtdy7EwO5VnLshsDYL-lYhTu-S1P_ak0_11QPeoudf_Pg&typo=1> >>> to (un)subscribe http://redfish.com/mailman/listinfo/friam_redfish.com >>> <https://linkprotect.cudasvc.com/url?a=http%3a%2f%2fredfish.com%2fmailman%2flistinfo%2ffriam_redfish.com&c=E,1,XgC24CHCqH-MiGAYACD7z6PtLS8L3PA6Ntnnm8-GFS-8K9DlJz1wbcz4YhFucq-Ucbr9oeSbxnsUcDSZ0XfT5aalKerL34Pq0R57j_mDeX4RrFvxbg,,&typo=1> >>> FRIAM-COMIC http://friam-comic.blogspot.com/ >>> <https://linkprotect.cudasvc.com/url?a=http%3a%2f%2ffriam-comic.blogspot.com%2f&c=E,1,1GkLbYQODhOcaZjv5c91-iHu4986bkpXTFuaNhmX9fFK-EzJOVs6x1AKJrFBa7SMS9GHwxtjJZQ_CrzTmsDpTJ-KHGBzYStX4zSsywCdFT2GG8q4MO_gHwI,&typo=1> >>> archives: 5/2017 thru present >>> https://redfish.com/pipermail/friam_redfish.com/ >>> <https://linkprotect.cudasvc.com/url?a=https%3a%2f%2fredfish.com%2fpipermail%2ffriam_redfish.com%2f&c=E,1,pPB6Gx8uMJadpxScEftegu18poh3jvMC3dByqfb7-kH5q9a1uN-Hx18Hpf2mWKYR1aATb5b9IUgqwfoulMiH2oh8K03lnbespQNIBfFm-D_etq1aSira&typo=1> >>> 1/2003 thru 6/2021 http://friam.383.s1.nabble.com/ >>> >>> -. --- - / ...- .- .-.. .. -.. / -- --- .-. ... . / -.-. --- -.. . >>> FRIAM Applied Complexity Group listserv >>> Fridays 9a-12p Friday St. Johns Cafe / Thursdays 9a-12p Zoom >>> https://bit.ly/virtualfriam >>> <https://linkprotect.cudasvc.com/url?a=https%3a%2f%2fbit.ly%2fvirtualfriam&c=E,1,9R_t1NDn9DvauNKttOqQgxrGaUrdMJxeYm5N63GEWkEvm1uQHRIPj5EryTCG4F8-FBeu18NA6Iqbyzr0qZmJUE2HF57NAov7tcuXG1CFAJ7ISv0,&typo=1> >>> to (un)subscribe http://redfish.com/mailman/listinfo/friam_redfish.com >>> <https://linkprotect.cudasvc.com/url?a=http%3a%2f%2fredfish.com%2fmailman%2flistinfo%2ffriam_redfish.com&c=E,1,lFTqZFH-R9D5jiRQ0m9XGZNNd5Rsr-TYbo8WKOmosICjS8S7fpAggB06ojoO4DIXx07lOw0UyZuhpBT_IsxNgIEKtnTtzzaLELhd606vpB5mrMmIKA,,&typo=1> >>> FRIAM-COMIC http://friam-comic.blogspot.com/ >>> <https://linkprotect.cudasvc.com/url?a=http%3a%2f%2ffriam-comic.blogspot.com%2f&c=E,1,xRMd22AvNM1vduoxz-Cd6-edLdgmkb_sifY1_C_kq0J3nQo_3JxTOJwh26ybgsQDy5PV7BrokU2_MLcMFRgluXk3WJSSCAoZVLgtMCpIcY07gVC_vTKmq4s,&typo=1> >>> archives: 5/2017 thru present >>> https://redfish.com/pipermail/friam_redfish.com/ >>> <https://linkprotect.cudasvc.com/url?a=https%3a%2f%2fredfish.com%2fpipermail%2ffriam_redfish.com%2f&c=E,1,KQ2mCnilVXt2leucw3caOQ8GNGViF5aUsirFIbdwFgZJeCh48I4mlovIE97rJIszL6HttaJg58sbOf85zTBi0aq7vxRF9a4Q94NIPMK5kpk3Gbb-tTFwcOZ0pNE,&typo=1> >>> 1/2003 thru 6/2021 http://friam.383.s1.nabble.com/ >>> >> >> -. --- - / ...- .- .-.. .. -.. / -- --- .-. ... . / -.-. --- -.. . >> FRIAM Applied Complexity Group listserv >> Fridays 9a-12p Friday St. Johns Cafe / Thursdays 9a-12p Zoom >> https://bit.ly/virtualfriam >> <https://linkprotect.cudasvc.com/url?a=https%3a%2f%2fbit.ly%2fvirtualfriam&c=E,1,JZG5Ry4qLJlJfRcyvniC-8nt5cXg6jFMhE3BkxTASN407yQ-YlvbXpA2Zd_l4ZYoxSROdQ-iJzUyJID9bt1wqtH1fEcb3cdiL6OmzuOItbP4tmY,&typo=1> >> to (un)subscribe http://redfish.com/mailman/listinfo/friam_redfish.com >> <https://linkprotect.cudasvc.com/url?a=http%3a%2f%2fredfish.com%2fmailman%2flistinfo%2ffriam_redfish.com&c=E,1,iThoLpYpqt8aVw23fOZyibIh1jKmvxMDkX9wTBRonsR8X4ezwG2-HhOu2w5yW0lyX3zmbFgbfy7jDe1gj23sTQK_OWFEcLCsJGKJhFl5dSqaSII6Om4d0nzZQBo,&typo=1> >> FRIAM-COMIC http://friam-comic.blogspot.com/ >> <https://linkprotect.cudasvc.com/url?a=http%3a%2f%2ffriam-comic.blogspot.com%2f&c=E,1,HYzkgvDtQjTwOa5_t7MZnsd_eXg7gQdb3c3PK_qnRyw9jdfOnrNAu3PtDOXQrjRLIJbkJDvSCzh2me61wHHIThc6Y2xNFIS7zHhMQAzK&typo=1> >> archives: 5/2017 thru present >> https://redfish.com/pipermail/friam_redfish.com/ >> <https://linkprotect.cudasvc.com/url?a=https%3a%2f%2fredfish.com%2fpipermail%2ffriam_redfish.com%2f&c=E,1,tQoVq56CqO6ncvHKTN9hbcVTg93ckGFVj0aCH7mntnycdBP8xmtUmWQk6vfH6tXlmj4efSkAfKSAQslYZ9xESLjg64AhcTLO8gILEjPUH05xuw,,&typo=1> >> 1/2003 thru 6/2021 http://friam.383.s1.nabble.com/ >> >> -. --- - / ...- .- .-.. .. -.. / -- --- .-. ... . / -.-. --- -.. . >> FRIAM Applied Complexity Group listserv >> Fridays 9a-12p Friday St. Johns Cafe / Thursdays 9a-12p Zoom >> https://linkprotect.cudasvc.com/url?a=https%3a%2f%2fbit.ly%2fvirtualfriam&c=E,1,Wmm9ThSA6Mz3l6f44HjYltfKE5iYDLD_dMQDomanWFe33hPZtOZx6UMs9HfvlE9qqDrpJHTiJq-6I7rnKON9ssVZvrW3vYhYayTxfEkBPA,,&typo=1 >> to (un)subscribe >> https://linkprotect.cudasvc.com/url?a=http%3a%2f%2fredfish.com%2fmailman%2flistinfo%2ffriam_redfish.com&c=E,1,SCq73BTJlGKctIkbzRtoaDbciyn7yKRCVtCpmNnPBqLEajOEfO3ae1cuIeemIU9GwwAYpzKzxvR6walpGzl7rZXKHBzAtoeNy6j6mqtwhLpEMFRuv0XoITA,&typo=1 >> FRIAM-COMIC >> https://linkprotect.cudasvc.com/url?a=http%3a%2f%2ffriam-comic.blogspot.com%2f&c=E,1,RNz26jMqvi3Xb0sjg5eWouWJOXinbcqzg63a-tE4P-K1Rmv4GlXJtlAaD8-Ij1iv6C6L7JF9LZzTygimbNX8G_fg7e6fKklz4FYSTigA5Q,,&typo=1 >> archives: 5/2017 thru present >> https://linkprotect.cudasvc.com/url?a=https%3a%2f%2fredfish.com%2fpipermail%2ffriam_redfish.com%2f&c=E,1,S1eeYxStqxoVg4yeuKl0cDdLaMoPcak1Ri2w3UANtBaag5riZ9K-IGLJIia0qNnuf-ChNm1tzsBtHZOtE9Wqe-yu2oKWkUjMwJoTFYVpi6adLzNpgY4iVPs,&typo=1 >> 1/2003 thru 6/2021 http://friam.383.s1.nabble.com/ >> >> >> -. --- - / ...- .- .-.. .. -.. / -- --- .-. ... . / -.-. --- -.. . >> FRIAM Applied Complexity Group listserv >> Fridays 9a-12p Friday St. Johns Cafe / Thursdays 9a-12p Zoom >> https://bit.ly/virtualfriam >> <https://linkprotect.cudasvc.com/url?a=https%3a%2f%2fbit.ly%2fvirtualfriam&c=E,1,tYXe7u30UizIRX7sCP5z134Vp_rCgZrx-HG1QllHUyYO2zfuWxe9oJB0rJBMD_fbOkbLIYPpr2aium2bSNugjk7zYsHr5bk-DHPLj56i&typo=1> >> to (un)subscribe http://redfish.com/mailman/listinfo/friam_redfish.com >> <https://linkprotect.cudasvc.com/url?a=http%3a%2f%2fredfish.com%2fmailman%2flistinfo%2ffriam_redfish.com&c=E,1,owxu2tT3Pnf6XPyExtN00oNk919A19XBuIfw_2ugWaBuigwyq7DFJlupgseFZ7dOIgMUnH_x2z2nLX7lUkJsFKOeHX0DTNUNT-mUAW5rsP8,&typo=1> >> FRIAM-COMIC http://friam-comic.blogspot.com/ >> <https://linkprotect.cudasvc.com/url?a=http%3a%2f%2ffriam-comic.blogspot.com%2f&c=E,1,0Dk21FivClJzcKxULsvq2ahHAaXAtwwsVxvRPMjYjKT8CN4LNIQgsTKYNRQbVsufJP8QWIPIZCVEAlkV78bQiWk7RfliP9Sp2Abk7ij92YlC6TmsJg,,&typo=1> >> archives: 5/2017 thru present >> https://redfish.com/pipermail/friam_redfish.com/ >> <https://linkprotect.cudasvc.com/url?a=https%3a%2f%2fredfish.com%2fpipermail%2ffriam_redfish.com%2f&c=E,1,xoH9CNhv1c36upLgpvFH9gpB7s-7TuKIBNfryFN2vs8QLRc8b7-KRpdiWxQKXL0o5ljIufdg0z0pvgn77zbqUXF-rGapyf9MPmtRBoxD6vgoFz4,&typo=1> >> 1/2003 thru 6/2021 http://friam.383.s1.nabble.com/ >> > <9780674279384-lg.jpg><9780674279384-lg.jpg>-. --- - / ...- .- .-.. .. > -.. / -- --- .-. ... . / -.-. --- -.. . > FRIAM Applied Complexity Group listserv > Fridays 9a-12p Friday St. Johns Cafe / Thursdays 9a-12p Zoom > https://linkprotect.cudasvc.com/url?a=https%3a%2f%2fbit.ly%2fvirtualfriam&c=E,1,2ZlC62KWqH1N6zMvqal4Ed0Wx5XQGMEddd41-jtZtFjgO1SirsDGzGIktZ5uuoAX15NFz0taAZNEgGPxd34ZeId3Oi-wtbhVUHXBugIhOw36&typo=1 > to (un)subscribe > https://linkprotect.cudasvc.com/url?a=http%3a%2f%2fredfish.com%2fmailman%2flistinfo%2ffriam_redfish.com&c=E,1,QZaSg2mbVJz2oN4YNuVpvuZjAkolp8mlWkppzQ0-aoqWNBnx1M7v90qS6byWzuUXDqyzMJNRAQn26_WVFsAlr_i3r1TLR4E0nLYs6alQCEdr-CStPCw,&typo=1 > FRIAM-COMIC > https://linkprotect.cudasvc.com/url?a=http%3a%2f%2ffriam-comic.blogspot.com%2f&c=E,1,6pvfI2nyTTYQmON_tKGUbLCRCzgFGvS9QLKctMYsBFDU0Jezuy2pxUKnT-jJ7-TUHFK5mRy3GkmvuVKm8PhtSGxRaWgt-3n8l-1NT599Ul54D_JdL5xw&typo=1 > archives: 5/2017 thru present > https://linkprotect.cudasvc.com/url?a=https%3a%2f%2fredfish.com%2fpipermail%2ffriam_redfish.com%2f&c=E,1,Od6NS37sWb2LiDsHMSlWPpueHD3xdUe4deNxdljKnlJcfHERLOFE6iw9WOQqHlD3uRhr7Si6KXivdjch6wdVhh-QOKc1CsO9-2NvVI5dfgGO3M8R2Wt2C_VpvA,,&typo=1 > 1/2003 thru 6/2021 http://friam.383.s1.nabble.com/ > > > -. --- - / ...- .- .-.. .. -.. / -- --- .-. ... . / -.-. --- -.. . > FRIAM Applied Complexity Group listserv > Fridays 9a-12p Friday St. Johns Cafe / Thursdays 9a-12p Zoom > https://bit.ly/virtualfriam > to (un)subscribe http://redfish.com/mailman/listinfo/friam_redfish.com > FRIAM-COMIC http://friam-comic.blogspot.com/ > archives: 5/2017 thru present > https://redfish.com/pipermail/friam_redfish.com/ > 1/2003 thru 6/2021 http://friam.383.s1.nabble.com/ >
-. --- - / ...- .- .-.. .. -.. / -- --- .-. ... . / -.-. --- -.. . FRIAM Applied Complexity Group listserv Fridays 9a-12p Friday St. Johns Cafe / Thursdays 9a-12p Zoom https://bit.ly/virtualfriam to (un)subscribe http://redfish.com/mailman/listinfo/friam_redfish.com FRIAM-COMIC http://friam-comic.blogspot.com/ archives: 5/2017 thru present https://redfish.com/pipermail/friam_redfish.com/ 1/2003 thru 6/2021 http://friam.383.s1.nabble.com/