Backpropagation
This is student work which has not yet been approved as correct by the instructor
Introduction[edit]
Backpropagation is a method to calculate the gradient of the loss function with respect to the weights in an artificial neural network. It is commonly used as a part of algorithms that optimize the performance of the network by adjusting the weights, for example in the gradient descent algorithm. It is also called backward propagation of errors.[2]
<ref> the url I cited by material from </ref>
How does it work or a deeper look[edit]
- If you are discussing a THING YOU CAN TOUCH, you must explain how it works, and the parts it is made of. Google around for an "exploded technical diagram" of your thing, maybe like this example of an MRI It is likely you will reference outside links. Please attribute your work.
- If you are discussing a PROCESS OR ABSTRACT CONCEPT (like fuzzy logic) you must deeply explain how it works.
Examples[edit]
Please include some example of how your concept is actually used. Your example must include WHERE it is used, and WHAT IS BENEFIT of it being used.
Pictures, diagrams[edit]
Pictures and diagrams go a LONG way to helping someone understand a topic. Especially if your topic is a little abstract or complex. Using a picture or diagram is a two part process:
External links[edit]
- It would be helpful
- to include many links
- to other internet resources
- to help fellow students
- Please make sure the content is good
- and don't link to a google search results, please