Current Slide

Small screen detected. You are viewing the mobile version of SlideWiki. If you wish to edit slides you will need to use a larger device.

Summary

  • Most brains have lots of neurons, each neuron approximates a linear-threshold unit.
  • Perceptrons (one-layer networks) approximate neurons, but are as such insufficiently expressive.
  • Multi-layer networks are sufficiently expressive; can be trained to deal with generalized data sets, i.e. via error back-propagation.
  • Multi-layer networks allow for the modeling of arbitrary separation boundaries, while single-layer perceptrons only provide linear boundaries.
  • Endless number of applications: Handwriting Recognition, Time Series Prediction, Bioinformatics, Kernel Machines (Support Vectore Machines), Data Compression, Financial Predication, Speech Recognition, Computer Vision, Protein Structures...

Speaker notes:

Content Tools

Sources

There are currently no sources for this slide.