On Mon, May 20, 2024 at 07:07 Undescribed Horrific Abuse, One Victim & Survivor of Many <[email protected]> wrote:
> # we improved the stability of low-data machine learning model training > > ## why is this useful? > > low-data model training would let individual users make AI models without > access to big data, such as those models used at scale to addict users to > products, services, or behaviors. > > ability to recreate these models with the user in control would give them > many more options for personal freedom. > > additionally low-data training makes more powerful AI. > <snip> ## in short > > to make a big model on a small dataset, train a training model that > improves generalization to more data. > > this is more powerful but more difficult if it also improves > generalization of itself. > additionally: this may have been tried in research papers somewhere already, if attempting it is unreasonable. i might start with keywords around hyper-parameter metamodel for generalization or such
