Yes, the money/expertise bar is still pretty high. But TANSTAAFL still applies. And the overwhelming evidence is 
coming in that specific models do better than those trained up on diverse data sets, "better" meaning 
less prone to subtle bullsh¡t. What I find fascinating is tools like OpenAI *facilitate* trespassing. We have a 
wonderful bloom of non-experts claiming they understand things like "deep learning". But do they? An 
old internet meme is brought to mind: "Do you even Linear Algebra, bro?" >8^D

On 1/8/23 01:06, Jochen Fromm wrote:
I have finished a number of Coursera courses recently, including "Deep Learning & 
Neural Networks with Keras" which was ok but not great. The problems with deep learning 
are

* to achieve impressive results like chatGPT from OpenAi or LaMDA from Goggle 
you need to spend millions on hardware
* only big organisations can afford to create such expensive models
* the resulting network is s black box and it is unclear why it works the way 
it does

In the end it is just the same old back propagation that has been known for decades, just 
on more computers and trained on more data. Peter Norvig calls it "The unreasonable 
effectiveness of data"
https://research.google.com/pubs/archive/35179.pdf

-J.


-------- Original message --------
From: Russ Abbott <[email protected]>
Date: 1/8/23 12:20 AM (GMT+01:00)
To: The Friday Morning Applied Complexity Coffee Group <[email protected]>
Subject: Re: [FRIAM] Deep learning training material

Hi Pieter,

A few comments.

  * Much of the actual deep learning material looks like it came from the Kaggle "Deep 
Learning <https://www.kaggle.com/learn/intro-to-deep-learning>" sequence.
  * In my opinion, R is an ugly and /ad hoc/ language. I'd stick to Python.
  * More importantly, I would put the How-to-use-Python stuff into a 
preliminary class. Assume your audience knows how to use Python and focus on 
Deep Learning. Given that, there is only a minimal amount of information about 
Deep Learning in the write-up. If I were to attend the workshop and thought I 
would be learning about Deep Learning, I would be disappointed--at least with 
what's covered in the write-up.

I say this because I've been looking for a good intro to Deep Learning. Even though I taught Computer Science for many years, and am now retired, I avoided Deep Learning because it was so non-symbolic. My focus has always been on symbolic computing. But Deep Learning has produced so many extraordinarily impressive results, I decided I should learn more about it. I haven't found any really good material. If you are interested, I'd be more than happy to work with you on developing some introductory Deep Learning material.
-- Russ Abbott
Professor Emeritus, Computer Science
California State University, Los Angeles


On Thu, Jan 5, 2023 at 11:31 AM Pieter Steenekamp <[email protected] 
<mailto:[email protected]>> wrote:

    Thanks to the kind support of OpenAI's chatGPT, I am in the process of 
gathering materials for a comprehensive and hands-on deep learning workshop. 
Although it is still a work in progress, I welcome any interested parties to 
take a look and provide their valuable input. Thank you!

    You can get it from:
    
https://www.dropbox.com/s/eyx4iumb0439wlx/deep%20learning%20training%20rev%2005012023.zip?dl=0
 
<https://www.dropbox.com/s/eyx4iumb0439wlx/deep%20learning%20training%20rev%2005012023.zip?dl=0>


--
ꙮ Mɥǝu ǝlǝdɥɐuʇs ɟᴉƃɥʇ' ʇɥǝ ƃɹɐss snɟɟǝɹs˙ ꙮ

-. --- - / ...- .- .-.. .. -.. / -- --- .-. ... . / -.-. --- -.. .
FRIAM Applied Complexity Group listserv
Fridays 9a-12p Friday St. Johns Cafe   /   Thursdays 9a-12p Zoom 
https://bit.ly/virtualfriam
to (un)subscribe http://redfish.com/mailman/listinfo/friam_redfish.com
FRIAM-COMIC http://friam-comic.blogspot.com/
archives:  5/2017 thru present https://redfish.com/pipermail/friam_redfish.com/
 1/2003 thru 6/2021  http://friam.383.s1.nabble.com/

Reply via email to