Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

cross validation on all ML models #42

Open
fusr opened this issue Sep 12, 2023 · 1 comment
Open

cross validation on all ML models #42

fusr opened this issue Sep 12, 2023 · 1 comment

Comments

@fusr
Copy link
Owner

fusr commented Sep 12, 2023

No description provided.

@fusr
Copy link
Owner Author

fusr commented Sep 12, 2023

splitting data set:

Data Splitting: The first step is to split your dataset into two or more subsets: a training set and a testing (or validation) set. The training set is used to train the model, while the testing set is used to evaluate its performance. However, in cross-validation, you typically split the data into multiple subsets, often referred to as "folds."

K-Fold Cross-Validation: The most common form of cross-validation is k-fold cross-validation, where the data is divided into 'k' equal-sized subsets or folds. The model is trained and evaluated 'k' times, each time using a different fold as the testing set and the remaining folds as the training set. The performance scores (e.g., accuracy, mean squared error) from each fold are then averaged to provide an overall performance estimate.

more info in the hands on machine learning notebook. Ana will send the link on slack.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant