Matt Harrison <matthewharri...@gmail.com> added the comment:

The ML world has collapsed on the terms X and y. (With that capitalization). 
Moreover, most (Python libraries) follow the interface of scikit-learn [0].

Training a model looks like this:

    model = LinearRegression()
    model.fit(X, y)

After that, the model instance has attribute that end in "_" that were learned 
from fitting. For linear regression[1] you get:

    model.coef_        # slope
    model.intercept_   # intercept

To make predictions you call .predict:

    y_hat = model.predict(X)

One bonus of leveraging the .fit/.predict interface (which other libraries such 
as XGBoost have also adopted) is that if your model is in the correct layout, 
you can trivially try different models.


0 - 
https://scikit-learn.org/stable/tutorial/basic/tutorial.html#learning-and-predicting

1 - 
https://scikit-learn.org/stable/modules/generated/sklearn.linear_model.LinearRegression.html#sklearn.linear_model.LinearRegression

----------
nosy: +matthewharrison

_______________________________________
Python tracker <rep...@bugs.python.org>
<https://bugs.python.org/issue44151>
_______________________________________
_______________________________________________
Python-bugs-list mailing list
Unsubscribe: 
https://mail.python.org/mailman/options/python-bugs-list/archive%40mail-archive.com

Reply via email to