Hey there!

There are a few easy ways you can reach me if you ever need anything or just want to say hi.

**Email**

matthew.h.mazur [at] gmail.com

Note that due to the quantity of emails I receive about neural networks thanks to this tutorial, **I am not able to help with backpropagation questions** – sorry about that! I’d recommend anyone wanting to learn more to check out PyImageSearch.com.

**Twitter**

@mhmazur

**Skype**

m.h.mazur

**LinkedIn**

http://www.linkedin.com/in/mhmazur

### Like this:

Like Loading...

Thank you so much, I can’t understand back propagation algorithm before read your article.

I seriously can’t thank you enough for the step by step backprop explanation! Easy to follow and you didn’t skip steps. This has helped so much with my AI class. Thank you

Thanks ,It was a very clear explanation of back propagation .

Thank you so much for the detailed explanation on back propagation

I spent entire day doing what you did on paper myself. It was an awesome day since you inspired me. Appreciate your brave and kind efforts.

Very good explanation. Thanks.

hi, Sir I have read your article and I got an understanding on artificial neural network and it is really valuable article but what I have a question is how we can train if the input is not in binary or if it is an ASCII character?

Doubt: do we need to update biasing

Thanks! This explains it very well! A couple things that would be neat to see as improvements would be to alter your java program you use to show the neural net in action and have it be a bit more interactive. By this I mean be able to step through say the first handful of calculations or just be able to stop it at will and then have it display the equations with real numbers as you click or mouse over the different parts of the model.

You are a great teacher!

how to determin the initial weight and bais dependend on the actual input and output

i just have one question about biases.

bias’s weight doesn’t change?

The tutorials I read when writing this didn’t update it so I didn’t here. It is something you can add on though.

Best articles to learn deep learning | Learn for Master

Thanks Mazur, “A Step by Step Backpropagation Example” is very well written compared to many other similar ones found in the net. I have a question, lets say the target output values are ’12’ and ’78’. Should I normalized these values to [0 ,1] ? If not, how do I calculate the output error where the training result is in range [0,1] ?

It would be even better if you could talk about the iterations

I mean you just took 1 training example

Please explain how to iterate all examples in dataset and epochs.

I am going to join the others on here and say thanks for great explanation of back propagation!

I don’t know how will I thanks to you for this excellent step by step explanation for the Backpropagation algorithm . Its a excellent work .and helpful for future work.

Concerning the backpropagation example, it was great for me to understand it. However, it doesn’t seem to mention the change of the biases in the process. Moreover, the image shows that the same bias is used on all neurons in the same layer which confused me a bit since each neuron usually has its own bias. It would be nice if you manage to modify the article based on these two ideas.

Very nice! I’ve used it to check my own implementation in Max/MSP and Python. However, after the first iteration my MSE equals 0.291027773694 instead of your 0.291027924. If I use the exact weight representations as in your example my MSE becomes 0.291027773717. Can’t exactly figure out what causes the difference(s).

thanks you are great i confused before know is better so please tell me where can i dowload for this exciting example

Example Max/MSP patcher using the included Mac 64-bit externals as attached.

Mac external sources on GitHub: https://github.com/mdehoogh/max_nn.git

Best,

Marc de Hoogh

Delft University of Technology, The Netherlands

Thank you for the explanation of MLP.

This is great :) Thank you!

This is great :) Thank you Matt!

Thank you ! about “step by step backprop explanation”!

Thanks for great example, but I have a question. When you calculated a partial derivative dEtotal/dOut1 you’ve got in the result a (-1) multiplier. Why?

(x^n)’ = n x^(n-1)

Hi, a small typo in net-h1, it should be 0.25*0.1, instead of 0.2 * 0.1