Overfitting: Difference between revisions

From Computer Science Wiki
No edit summary
 
(4 intermediate revisions by 2 users not shown)
Line 1: Line 1:
<center>
<blockquote style="padding: 5px; background-color: #FFF8DC; border: solid thin gray;">
  [[File:Exclamation.png]] This is student work which has not yet been approved as correct by the instructor
</blockquote>
</center>
[[file:Studying.png|right|frame|Case study notes<ref>http://www.flaticon.com/</ref>]]
[[file:Studying.png|right|frame|Case study notes<ref>http://www.flaticon.com/</ref>]]


== Introduction ==
== Introduction ==


Overfitting is when a statistical model takes too many parameters. (continue general meaning)
Overfitting is a phenomenon that occurs when a machine learning model is trained too well on the training data. This can cause the model to perform poorly on unseen data, because it has learned patterns in the training data that do not generalize to new data.
 
Please write a clear, concise description of your topic here.You will likely reference your introduction from somewhere else. Please use the following syntax at the end of each of your ideas. '''IT IS CRITICAL YOU ATTRIBUTE''' others work. Your introduction should be factual. No more than 3 or 4 sentences, please. Because you are not an expert in your topic, I expect you to triangulate your information. LOTS OF LINK TO OTHER RESOURCES PLEASE!
<nowiki>
<ref> the url I cited by material from </ref>
</nowiki>
 
== How does it work or a deeper look ==
 
* If you are discussing a THING YOU CAN TOUCH, you must explain how it works, and the parts it is made of. Google around for an "exploded technical diagram" of your thing, [http://cdiok.com/wp-content/uploads/2012/01/MRI-Technology.jpg maybe like this example of an MRI]  It is likely you will reference outside links. Please attribute your work.
* If you are discussing a PROCESS OR ABSTRACT CONCEPT (like [[fuzzy logic]]) you must deeply explain how it works.
 
== Examples ==
 
Please include some example of how your concept is actually used. Your example must include WHERE it is used, and WHAT IS BENEFIT of it being used.
 
== Pictures, diagrams ==
 
Pictures and diagrams go a LONG way to helping someone understand a topic. Especially if your topic is a little abstract or complex. Using a picture or diagram is a two part process:


# [https://www.mediawiki.org/wiki/Help:Managing_files upload a file]
In overfitting, the model becomes too complex and fits the noise in the training data rather than the underlying relationship. This can lead to poor generalization performance, as the model will be sensitive to the noise in the training data and may not be able to generalize to unseen data.
# [https://www.mediawiki.org/wiki/Help:Images use the file on a wiki page]


== External links ==
One way to mitigate overfitting is to use techniques such as regularization, which helps to constrain the model and reduce its complexity. Other techniques include using a larger training dataset, using cross-validation to tune the model, and early stopping to prevent the model from becoming too complex.
* https://en.wikipedia.org/wiki/Overfitting
* It would be helpful
* to include many links
* to other internet resources
* to help fellow students
* Please make sure the content is good
* and don't link to a google search results, please


== References ==
== References ==

Latest revision as of 09:12, 7 January 2023

Case study notes[1]

Introduction[edit]

Overfitting is a phenomenon that occurs when a machine learning model is trained too well on the training data. This can cause the model to perform poorly on unseen data, because it has learned patterns in the training data that do not generalize to new data.

In overfitting, the model becomes too complex and fits the noise in the training data rather than the underlying relationship. This can lead to poor generalization performance, as the model will be sensitive to the noise in the training data and may not be able to generalize to unseen data.

One way to mitigate overfitting is to use techniques such as regularization, which helps to constrain the model and reduce its complexity. Other techniques include using a larger training dataset, using cross-validation to tune the model, and early stopping to prevent the model from becoming too complex.

References[edit]