Re: [FRIAM] Language Model Understanding

2023-10-07 Thread Steve Smith
Tom - This is probably a little less relevant to the business of journalism than you deserve but this is a pretty good (IMO) fleshing out of some of the technical details of (L)LMs which might fill in some blanks: https://developers.google.com/machine-learning/resources/intro-llms There

Re: [FRIAM] Language Model Understanding

2023-10-07 Thread Tom Johnson
Thanks. Those are impressive numbers. Tom On Sat, Oct 7, 2023 at 4:11 PM Marcus Daniels wrote: > The “large” refers to the number of parameters used. A smaller large > language model – a deep neural net -- start about 3 billion parameters, but > larger ones like Claude 2 (the latest large

Re: [FRIAM] Language Model Understanding

2023-10-07 Thread Marcus Daniels
The “large” refers to the number of parameters used. A smaller large language model – a deep neural net -- start about 3 billion parameters, but larger ones like Claude 2 (the latest large language model of the company that wrote the paper Steve mentioned) have more than 130 billion

Re: [FRIAM] Language Model Understanding

2023-10-07 Thread Tom Johnson
Thanks for passing this along, Steve. I wish, however, the authors of this short piece would have included a definition of, in their usage, "Large Language Models" and "Small Language Models." Perhaps I can find those in the larger paper. Tom On Sat, Oct 7, 2023 at 12:34 PM Steve Smith wrote:

[FRIAM] Language Model Understanding

2023-10-07 Thread Steve Smith
This popular-press article came through my Google News feed recently which I thought might be useful to the Journalists/English-Majors on the list to help understand how LLMs work, etc.   When I read it in detail (forwarded from my TS (TinyScreenPhone) on my LS (Large Screen Laptop)) I found