Here you will find daily news and tutorials about R, contributed by over 750 bloggers.
There are many ways to follow us - By e-mail:On Facebook: If you are an R blogger yourself you are invited to add your own R content feed to this site (Non-English R bloggers should add themselves- here)

Neural networks have been a very important area of scientific study that has evolved by different disciplines such as mathematics, biology, psychology, computer science, etc. The study of neural networks leapt from theory to practice with the emergence of computers. Training a neural network by adjusting the weights of the connections is computationally very expensive so its application to practical problems took until the mid-80s when a more efficient algorithm was discovered. That algorithm is now known as back-propagation errors or simply backpropagation. One of the most cited articles on this algorithm is:

Learning representations by back-propagating errors David E. Rumelhart^{*}, Geoffrey E. Hinton^{†} & Ronald J. Williams^{*} Nature323, 533 – 536 (09 October 1986)

Although it is a very technical article, anyone who wants to study and understand neural networks is obliged to pass through this material. I share the entire article in: https://github.com/pakinja/Data-R-Value

Related

To leave a comment for the author, please follow the link and comment on their blog: Data R Value.