this post was submitted on 24 Jun 2025
315 points (98.2% liked)
Science Memes
15468 readers
2102 users here now
Welcome to c/science_memes @ Mander.xyz!
A place for majestic STEMLORD peacocking, as well as memes about the realities of working in a lab.
Rules
- Don't throw mud. Behave like an intellectual and remember the human.
- Keep it rooted (on topic).
- No spam.
- Infographics welcome, get schooled.
This is a science community. We use the Dawkins definition of meme.
Research Committee
Other Mander Communities
Science and Research
Biology and Life Sciences
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- !reptiles and [email protected]
Physical Sciences
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
Humanities and Social Sciences
Practical and Applied Sciences
- !exercise-and [email protected]
- [email protected]
- !self [email protected]
- [email protected]
- [email protected]
- [email protected]
Memes
Miscellaneous
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
Cross validation: "What am I a joke to you?"
What's cross validation?
Cross validation is a way of calculating the likely uncertainty of any model (it doesn't have to be a machine learning model).
A common cross validation approach is LOOCV (leave one out cross validation), for small datasets. Another is K-folds cross validation. In any case, the basics is to leave out "some amount" of your training data, totally removed from the training process, then you train your model, then you validate it on the trained model. You then repeat this process over the k-folds or each unit of your training data to create a valid uncertainty.
So a few things. First, this a standard approach in machine learning, because once you get stop making the assumptions of frequentism (and you probably should), you no longer get things like uncertainty for free, because the assumptions aren't met.
In some approaches in machine learning, this is necessary because there really isn't a tractable way to get uncertainty from the model (although in others, like random forest, you get cross validation for free).
Cross validation is great because you really don't need to understand anything about the model itself; you just implement the validation strategy and you get a valid answer for the model uncertainty.
A joke to you.