fbpx
Wikipedia

Neural network

A neural network is a neural circuit of biological neurons, sometimes also called a biological neural network, or a network of artificial neurons or nodes in the case of an artificial neural network.[1]

Simplified view of a feedforward artificial neural network

Artificial neural networks are used for solving artificial intelligence (AI) problems; they model connections of biological neurons as weights between nodes. A positive weight reflects an excitatory connection, while negative values mean inhibitory connections. All inputs are modified by a weight and summed. This activity is referred to as a linear combination. Finally, an activation function controls the amplitude of the output. For example, an acceptable range of output is usually between 0 and 1, or it could be −1 and 1.

These artificial networks may be used for predictive modeling, adaptive control and applications where they can be trained via a dataset. Self-learning resulting from experience can occur within networks, which can derive conclusions from a complex and seemingly unrelated set of information.[2]

Overview Edit

A biological neural network is composed of a group of chemically connected or functionally associated neurons. A single neuron may be connected to many other neurons and the total number of neurons and connections in a network may be extensive. Connections, called synapses, are usually formed from axons to dendrites, though dendrodendritic synapses[3] and other connections are possible. Apart from electrical signalling, there are other forms of signalling that arise from neurotransmitter diffusion.

Artificial intelligence, cognitive modelling, and neural networks are information processing paradigms inspired by how biological neural systems process data. Artificial intelligence and cognitive modelling try to simulate some properties of biological neural networks. In the artificial intelligence field, artificial neural networks have been applied successfully to speech recognition, image analysis and adaptive control, in order to construct software agents (in computer and video games) or autonomous robots.

Historically, digital computers evolved from the von Neumann model, and operate via the execution of explicit instructions via access to memory by a number of processors. On the other hand, the origins of neural networks are based on efforts to model information processing in biological systems. Unlike the von Neumann model, neural network computing does not separate memory and processing.

Neural network theory has served to identify better how the neurons in the brain function and provide the basis for efforts to create artificial intelligence.

History Edit

The preliminary theoretical base for contemporary neural networks was independently proposed by Alexander Bain[4] (1873) and William James[5] (1890). In their work, both thoughts and body activity resulted from interactions among neurons within the brain.

 
Computer simulation of the branching architecture of the dendrites of pyramidal neurons.[6]

For Bain,[4] every activity led to the firing of a certain set of neurons. When activities were repeated, the connections between those neurons strengthened. According to his theory, this repetition was what led to the formation of memory. The general scientific community at the time was skeptical of Bain's[4] theory because it required what appeared to be an inordinate number of neural connections within the brain. It is now apparent that the brain is exceedingly complex and that the same brain “wiring” can handle multiple problems and inputs.

James'[5] theory was similar to Bain's;[4] however, he suggested that memories and actions resulted from electrical currents flowing among the neurons in the brain. His model, by focusing on the flow of electrical currents, did not require individual neural connections for each memory or action.

C. S. Sherrington[7] (1898) conducted experiments to test James' theory. He ran electrical currents down the spinal cords of rats. However, instead of demonstrating an increase in electrical current as projected by James, Sherrington found that the electrical current strength decreased as the testing continued over time. Importantly, this work led to the discovery of the concept of habituation.

Wilhelm Lenz (1920) and Ernst Ising (1925) created and analyzed the Ising model[8] which is essentially a non-learning artificial recurrent neural network (RNN) consisting of neuron-like threshold elements.[9] In 1972, Shun'ichi Amari made this architecture adaptive.[10][9] His learning RNN was popularised by John Hopfield in 1982.[11]McCulloch and Pitts[12] (1943) also created a computational model for neural networks based on mathematics and algorithms. They called this model threshold logic. These early models paved the way for neural network research to split into two distinct approaches. One approach focused on biological processes in the brain and the other focused on the application of neural networks to artificial intelligence.

In the late 1940s psychologist Donald Hebb[13] created a hypothesis of learning based on the mechanism of neural plasticity that is now known as Hebbian learning. Hebbian learning is considered to be a 'typical' unsupervised learning rule and its later variants were early models for long term potentiation. These ideas started being applied to computational models in 1948 with Turing's B-type machines.

Farley and Clark[14] (1954) first used computational machines, then called calculators, to simulate a Hebbian network at MIT. Other neural network computational machines were created by Rochester, Holland, Habit, and Duda[15] (1956).

Frank Rosenblatt[16] (1958) created the perceptron, an algorithm for pattern recognition based on a two-layer learning computer network using simple addition and subtraction. With mathematical notation, Rosenblatt also described circuitry not in the basic perceptron, such as the exclusive-or circuit.

Some say that neural network research stagnated after the publication of machine learning research by Marvin Minsky and Seymour Papert[17] (1969). They discovered two key issues with the computational machines that processed neural networks. The first issue was that single-layer neural networks were incapable of processing the exclusive-or circuit. The second significant issue was that computers were not sophisticated enough to effectively handle the long run time required by large neural networks. However, by the time this book came out, methods for training multilayer perceptrons (MLPs) were already known. The first deep learning MLP was published by Alexey Grigorevich Ivakhnenko and Valentin Lapa in 1965.[18][19][9] The first deep learning MLP trained by stochastic gradient descent[20] was published in 1967 by Shun'ichi Amari.[21][9] In computer experiments conducted by Amari's student Saito, a five layer MLP with two modifiable layers learned useful internal representations to classify non-linearily separable pattern classes.[9]

Neural network research was boosted when computers achieved greater processing power. Also key in later advances was the backpropagation algorithm. It is an efficient application of the Leibniz chain rule (1673)[22] to networks of differentiable nodes.[9] It is also known as the reverse mode of automatic differentiation or reverse accumulation, due to Seppo Linnainmaa (1970).[23][24][25][26][9] The term "back-propagating errors" was introduced in 1962 by Frank Rosenblatt,[27][9] but he did not have an implementation of this procedure, although Henry J. Kelley had a continuous precursor of backpropagation[28] already in 1960 in the context of control theory.[9] In 1982, Paul Werbos applied backpropagation to MLPs in the way that has become standard.[29]

In the late 1970s to early 1980s, interest briefly emerged in theoretically investigating the Ising model by Wilhelm Lenz (1920) and Ernst Ising (1925)[8] in relation to Cayley tree topologies and large neural networks. In 1981, the Ising model was solved exactly by Peter Barth for the general case of closed Cayley trees (with loops) with an arbitrary branching ratio[30] and found to exhibit unusual phase transition behavior in its local-apex and long-range site-site correlations.[31][32]

The parallel distributed processing of the mid-1980s became popular under the name connectionism. The text by Rumelhart and McClelland[33] (1986) provided a full exposition on the use of connectionism in computers to simulate neural processes.

Neural networks, as used in artificial intelligence, have traditionally been viewed as simplified models of neural processing in the brain, even though the relation between this model and brain biological architecture is debated, as it is not clear to what degree artificial neural networks mirror brain function.[34]

Artificial intelligence Edit

A neural network (NN), in the case of artificial neurons called artificial neural network (ANN) or simulated neural network (SNN), is an interconnected group of natural or artificial neurons that uses a mathematical or computational model for information processing based on a connectionistic approach to computation. In most cases, an ANN is an adaptive system that changes its structure based on external or internal information that flows through the network.

In more practical terms, neural networks are non-linear statistical data modeling or decision making tools. They can be used to model complex relationships between inputs and outputs or to find patterns in data.

An artificial neural network involves a network of processing elements (artificial neurons) which can exhibit complex global behavior, determined by the connections between the processing elements and element parameters. Artificial neurons were first proposed in 1943 by Warren McCulloch, a neurophysiologist, and Walter Pitts, a logician, who first collaborated at the University of Chicago.[35]

One classical type of artificial neural network is the recurrent Hopfield network.

The concept of a neural network appears to have first been proposed by Alan Turing in his 1948 paper Intelligent Machinery in which he called them "B-type unorganised machines".[36]

The utility of artificial neural network models lies in the fact that they can be used to infer a function from observations and also to use it. Unsupervised neural networks can also be used to learn representations of the input that capture the salient characteristics of the input distribution, e.g., see the Boltzmann machine (1983), and more recently, deep learning algorithms, which can implicitly learn the distribution function of the observed data. Learning in neural networks is particularly useful in applications where the complexity of the data or task makes the design of such functions by hand impractical.

Applications Edit

Neural networks can be used in different fields. The tasks to which artificial neural networks are applied tend to fall within the following broad categories:

Application areas of ANNs include nonlinear system identification[37] and control (vehicle control, process control), game-playing and decision making (backgammon, chess, racing), pattern recognition (radar systems, face identification, object recognition), sequence recognition (gesture, speech, handwritten text recognition), medical diagnosis, financial applications, data mining (or knowledge discovery in databases, "KDD"), visualization and e-mail spam filtering. For example, it is possible to create a semantic profile of user's interests emerging from pictures trained for object recognition.[38]

Neuroscience Edit

Theoretical and computational neuroscience is the field concerned with the analysis and computational modeling of biological neural systems. Since neural systems are intimately related to cognitive processes and behaviour, the field is closely related to cognitive and behavioural modeling.

The aim of the field is to create models of biological neural systems in order to understand how biological systems work. To gain this understanding, neuroscientists strive to make a link between observed biological processes (data), biologically plausible mechanisms for neural processing and learning (biological neural network models) and theory (statistical learning theory and information theory).

Types of models Edit

Many models are used; defined at different levels of abstraction, and modeling different aspects of neural systems. They range from models of the short-term behaviour of individual neurons, through models of the dynamics of neural circuitry arising from interactions between individual neurons, to models of behaviour arising from abstract neural modules that represent complete subsystems. These include models of the long-term and short-term plasticity of neural systems and their relation to learning and memory, from the individual neuron to the system level.

Connectivity Edit

In August 2020 scientists reported that bi-directional connections, or added appropriate feedback connections, can accelerate and improve communication between and in modular neural networks of the brain's cerebral cortex and lower the threshold for their successful communication. They showed that adding feedback connections between a resonance pair can support successful propagation of a single pulse packet throughout the entire network.[39][40]

Criticism Edit

Historically, a common criticism of neural networks, particularly in robotics, was that they require a large diversity of training samples for real-world operation (although if single dimension, only 8 samples required to get practically to low of 12% error). This is not surprising, since any learning machine needs sufficient representative examples in order to capture the underlying structure that allows it to generalize to new cases. Dean Pomerleau, in his research presented in the paper "Knowledge-based Training of Artificial Neural Networks for Autonomous Robot Driving," uses a neural network to train a robotic vehicle to drive on multiple types of roads (single lane, multi-lane, dirt, etc.). A large amount of his research is devoted to (1) extrapolating multiple training scenarios from a single training experience, and (2) preserving past training diversity so that the system does not become overtrained (if, for example, it is presented with a series of right turns—it should not learn to always turn right). These issues are common in neural networks that must decide from amongst a wide variety of responses, but can be dealt with in several ways, for example by randomly shuffling the training examples, by using a numerical optimization algorithm that does not take too large steps when changing the network connections following an example, or by grouping examples in so-called mini-batches.

A. K. Dewdney, a former Scientific American columnist, wrote in 1997, "Although neural nets do solve a few toy problems, their powers of computation are so limited that I am surprised anyone takes them seriously as a general problem-solving tool."[41]

Arguments for Dewdney's position are that to implement large and effective software neural networks, much processing and storage resources need to be committed. While the brain has hardware tailored to the task of processing signals through a graph of neurons, simulating even a most simplified form on Von Neumann technology may compel a neural network designer to fill many millions of database rows for its connections—which can consume vast amounts of computer memory and data storage capacity. Furthermore, the designer of neural network systems will often need to simulate the transmission of signals through many of these connections and their associated neurons—which must often be matched with incredible amounts of CPU processing power and time. While neural networks often yield effective programs, they too often do so at the cost of efficiency (they tend to consume considerable amounts of time and money).

Arguments against Dewdney's position are that neural nets have been successfully used to solve many complex and diverse tasks, such as autonomously flying aircraft.[42]

Technology writer Roger Bridgman commented on Dewdney's statements about neural nets:

Neural networks, for instance, are in the dock not only because they have been hyped to high heaven, (what hasn't?) but also because you could create a successful net without understanding how it worked: the bunch of numbers that captures its behaviour would in all probability be "an opaque, unreadable table...valueless as a scientific resource".

In spite of his emphatic declaration that science is not technology, Dewdney seems here to pillory neural nets as bad science when most of those devising them are just trying to be good engineers. An unreadable table that a useful machine could read would still be well worth having.[43]

Although it is true that analyzing what has been learned by an artificial neural network is difficult, it is much easier to do so than to analyze what has been learned by a biological neural network. Moreover, recent emphasis on the explainability of AI has contributed towards the development of methods, notably those based on attention mechanisms, for visualizing and explaining learned neural networks. Furthermore, researchers involved in exploring learning algorithms for neural networks are gradually uncovering generic principles that allow a learning machine to be successful. For example, Bengio and LeCun (2007) wrote an article regarding local vs non-local learning, as well as shallow vs deep architecture.[44]

Some other criticisms came from believers of hybrid models (combining neural networks and symbolic approaches). They advocate the intermix of these two approaches and believe that hybrid models can better capture the mechanisms of the human mind (Sun and Bookman, 1990).[full citation needed]

Recent improvements Edit

While initially research had been concerned mostly with the electrical characteristics of neurons, a particularly important part of the investigation in recent years has been the exploration of the role of neuromodulators such as dopamine, acetylcholine, and serotonin on behaviour and learning.[citation needed]

Biophysical models, such as BCM theory, have been important in understanding mechanisms for synaptic plasticity, and have had applications in both computer science and neuroscience. Research is ongoing in understanding the computational algorithms used in the brain, with some recent biological evidence for radial basis networks and neural backpropagation as mechanisms for processing data.[citation needed]

Computational devices have been created in CMOS for both biophysical simulation and neuromorphic computing. More recent efforts show promise for creating nanodevices for very large scale principal components analyses and convolution.[45] If successful, these efforts could usher in a new era of neural computing that is a step beyond digital computing,[46] because it depends on learning rather than programming and because it is fundamentally analog rather than digital even though the first instantiations may in fact be with CMOS digital devices.

Between 2009 and 2012, the recurrent neural networks and deep feedforward neural networks developed in the research group of Jürgen Schmidhuber at the Swiss AI Lab IDSIA have won eight international competitions in pattern recognition and machine learning.[47] For example, multi-dimensional long short term memory (LSTM)[48][49] won three competitions in connected handwriting recognition at the 2009 International Conference on Document Analysis and Recognition (ICDAR), without any prior knowledge about the three different languages to be learned.

Variants of the back-propagation algorithm, as well as unsupervised methods by Geoff Hinton and colleagues at the University of Toronto, can be used to train deep, highly nonlinear neural architectures,[50] similar to the 1980 Neocognitron by Kunihiko Fukushima,[51] and the "standard architecture of vision",[52] inspired by the simple and complex cells identified by David H. Hubel and Torsten Wiesel in the primary visual cortex.

Radial basis function and wavelet networks have also been introduced. These can be shown to offer best approximation properties and have been applied in nonlinear system identification and classification applications.[37]

Deep learning feedforward networks alternate convolutional layers and max-pooling layers, topped by several pure classification layers. Fast GPU-based implementations of this approach have won several pattern recognition contests, including the IJCNN 2011 Traffic Sign Recognition Competition[53] and the ISBI 2012 Segmentation of Neuronal Structures in Electron Microscopy Stacks challenge.[54] Such neural networks also were the first artificial pattern recognizers to achieve human-competitive or even superhuman performance[55] on benchmarks such as traffic sign recognition (IJCNN 2012), or the MNIST handwritten digits problem of Yann LeCun and colleagues at NYU.

Analytical and computational techniques derived from statistical physics of disordered systems, can be extended to large-scale problems, including machine learning, e.g., to analyze the weight space of deep neural networks.[56]

See also Edit

References Edit

  1. ^ Hopfield, J. J. (1982). "Neural networks and physical systems with emergent collective computational abilities". Proc. Natl. Acad. Sci. U.S.A. 79 (8): 2554–2558. Bibcode:1982PNAS...79.2554H. doi:10.1073/pnas.79.8.2554. PMC 346238. PMID 6953413.
  2. ^ "Neural Net or Neural Network – Gartner IT Glossary". gartner.com.
  3. ^ Arbib, p.666
  4. ^ a b c d Bain (1873). Mind and Body: The Theories of Their Relation. New York: D. Appleton and Company.
  5. ^ a b James (1890). The Principles of Psychology. New York: H. Holt and Company.
  6. ^ Cuntz, Hermann (2010). "PLoS Computational Biology Issue Image | Vol. 6(8) August 2010". PLOS Computational Biology. 6 (8): ev06.i08. doi:10.1371/image.pcbi.v06.i08.
  7. ^ Sherrington, C.S. (1898). "Experiments in Examination of the Peripheral Distribution of the Fibers of the Posterior Roots of Some Spinal Nerves". Proceedings of the Royal Society of London. 190: 45–186. doi:10.1098/rstb.1898.0002.
  8. ^ a b Brush, Stephen G. (1967). "History of the Lenz-Ising Model". Reviews of Modern Physics. 39 (4): 883–893. Bibcode:1967RvMP...39..883B. doi:10.1103/RevModPhys.39.883.
  9. ^ a b c d e f g h i Schmidhuber, Juergen (2022). "Annotated History of Modern AI and Deep Learning". arXiv:2212.11279 [cs.NE].
  10. ^ Amari, Shun-Ichi (1972). "Learning patterns and pattern sequences by self-organizing nets of threshold elements". IEEE Transactions. C (21): 1197–1206.
  11. ^ Hopfield, J. J. (1982). "Neural networks and physical systems with emergent collective computational abilities". Proceedings of the National Academy of Sciences. 79 (8): 2554–2558. Bibcode:1982PNAS...79.2554H. doi:10.1073/pnas.79.8.2554. PMC 346238. PMID 6953413.
  12. ^ McCulloch, Warren; Walter Pitts (1943). "A Logical Calculus of Ideas Immanent in Nervous Activity". Bulletin of Mathematical Biophysics. 5 (4): 115–133. doi:10.1007/BF02478259.
  13. ^ Hebb, Donald (1949). The Organization of Behavior. New York: Wiley.
  14. ^ Farley, B.; W.A. Clark (1954). "Simulation of Self-Organizing Systems by Digital Computer". IRE Transactions on Information Theory. 4 (4): 76–84. doi:10.1109/TIT.1954.1057468.
  15. ^ Rochester, N.; J.H. Holland, L.H. Habit and W.L. Duda (1956). "Tests on a cell assembly theory of the action of the brain, using a large digital computer". IRE Transactions on Information Theory. 2 (3): 80–93. doi:10.1109/TIT.1956.1056810.
  16. ^ Rosenblatt, F. (1958). "The Perceptron: A Probabilistic Model For Information Storage And Organization In The Brain". Psychological Review. 65 (6): 386–408. CiteSeerX 10.1.1.588.3775. doi:10.1037/h0042519. PMID 13602029. S2CID 12781225.
  17. ^ Minsky, M.; S. Papert (1969). An Introduction to Computational Geometry. MIT Press. ISBN 978-0-262-63022-1.
  18. ^ Ivakhnenko, A. G. (1973). Cybernetic Predicting Devices. CCM Information Corporation.
  19. ^ Ivakhnenko, A. G.; Grigorʹevich Lapa, Valentin (1967). Cybernetics and forecasting techniques. American Elsevier Pub. Co.
  20. ^ Robbins, H.; Monro, S. (1951). "A Stochastic Approximation Method". The Annals of Mathematical Statistics. 22 (3): 400. doi:10.1214/aoms/1177729586.
  21. ^ Amari, Shun'ichi (1967). "A theory of adaptive pattern classifier". IEEE Transactions. EC (16): 279–307.
  22. ^ Leibniz, Gottfried Wilhelm Freiherr von (1920). The Early Mathematical Manuscripts of Leibniz: Translated from the Latin Texts Published by Carl Immanuel Gerhardt with Critical and Historical Notes (Leibniz published the chain rule in a 1676 memoir). Open court publishing Company. ISBN 9780598818461.
  23. ^ Linnainmaa, Seppo (1970). The representation of the cumulative rounding error of an algorithm as a Taylor expansion of the local rounding errors (Masters) (in Finnish). University of Helsinki. pp. 6–7.
  24. ^ Linnainmaa, Seppo (1976). "Taylor expansion of the accumulated rounding error". BIT Numerical Mathematics. 16 (2): 146–160. doi:10.1007/bf01931367. S2CID 122357351.
  25. ^ Griewank, Andreas (2012). "Who Invented the Reverse Mode of Differentiation?". Optimization Stories. Documenta Matematica, Extra Volume ISMP. pp. 389–400. S2CID 15568746.
  26. ^ Griewank, Andreas; Walther, Andrea (2008). Evaluating Derivatives: Principles and Techniques of Algorithmic Differentiation, Second Edition. SIAM. ISBN 978-0-89871-776-1.
  27. ^ Rosenblatt, Frank (1962). Principles of Neurodynamics. Spartan, New York.
  28. ^ Kelley, Henry J. (1960). "Gradient theory of optimal flight paths". ARS Journal. 30 (10): 947–954. doi:10.2514/8.5282.
  29. ^ Werbos, Paul (1982). "Applications of advances in nonlinear sensitivity analysis" (PDF). System modeling and optimization. Springer. pp. 762–770. (PDF) from the original on April 14, 2016. Retrieved July 2, 2017.
  30. ^ Barth, Peter F. (1981). Cooperativity and the Transition Behavior of Large Neural Nets (MS thesis). Burlington: University of Vermont. OCLC 8231704.
  31. ^ Krizan, J.E.; Barth, P.F.; Glasser, M.L. (1983). "Exact Phase Transitions for the Ising Model on the Closed Cayley Tree". Physica. North-Holland Publishing Co. 119A: 230–242. doi:10.1016/0378-4371(83)90157-7.
  32. ^ Glasser, M.L.; Goldberg, M. (1983), "The Ising model on a closed Cayley tree", Physica, 117A (2–3): 670–672, Bibcode:1983PhyA..117..670G, doi:10.1016/0378-4371(83)90138-3
  33. ^ Rumelhart, D.E.; James McClelland (1986). Parallel Distributed Processing: Explorations in the Microstructure of Cognition. Cambridge: MIT Press.
  34. ^ Russell, Ingrid. . Archived from the original on May 29, 2014.
  35. ^ McCulloch, Warren; Pitts, Walter (1943). "A Logical Calculus of Ideas Immanent in Nervous Activity". Bulletin of Mathematical Biophysics. 5 (4): 115–133. doi:10.1007/BF02478259.
  36. ^ Copeland, B. Jack, ed. (2004). The Essential Turing. Oxford University Press. p. 403. ISBN 978-0-19-825080-7.
  37. ^ a b Billings, S. A. (2013). Nonlinear System Identification: NARMAX Methods in the Time, Frequency, and Spatio-Temporal Domains. Wiley. ISBN 978-1-119-94359-4.
  38. ^ Wieczorek, Szymon; Filipiak, Dominik; Filipowska, Agata (2018). "Semantic Image-Based Profiling of Users' Interests with Neural Networks". Studies on the Semantic Web. 36 (Emerging Topics in Semantic Technologies). doi:10.3233/978-1-61499-894-5-179.
  39. ^ "Neuroscientists demonstrate how to improve communication between different regions of the brain". medicalxpress.com. Retrieved September 6, 2020.
  40. ^ Rezaei, Hedyeh; Aertsen, Ad; Kumar, Arvind; Valizadeh, Alireza (August 10, 2020). "Facilitating the propagation of spiking activity in feedforward networks by including feedback". PLOS Computational Biology. 16 (8): e1008033. Bibcode:2020PLSCB..16E8033R. doi:10.1371/journal.pcbi.1008033. ISSN 1553-7358. PMC 7444537. PMID 32776924. S2CID 221100528.   Text and images are available under a Creative Commons Attribution 4.0 International License.
  41. ^ Dewdney, A. K. (1997). Yes, we have no neutrons : an eye-opening tour through the twists and turns of bad science. New York: Wiley. ISBN 0-471-10806-5. OCLC 35558945.
  42. ^ Administrator, NASA (June 5, 2013). "Dryden Flight Research Center - News Room: News Releases: NASA NEURAL NETWORK PROJECT PASSES MILESTONE". NASA.
  43. ^ . Archived from the original on March 19, 2012. Retrieved August 1, 2006.
  44. ^ "Scaling Learning Algorithms towards {AI} - LISA - Publications - Aigaion 2.0". www.iro.umontreal.ca.
  45. ^ Yang, J. J.; et al. (2008). "Memristive switching mechanism for metal/oxide/metal nanodevices". Nat. Nanotechnol. 3 (7): 429–433. doi:10.1038/nnano.2008.160. PMID 18654568.
  46. ^ Strukov, D. B.; et al. (2008). "The missing memristor found". Nature. 453 (7191): 80–83. Bibcode:2008Natur.453...80S. doi:10.1038/nature06932. PMID 18451858. S2CID 4367148.
  47. ^ . Archived from the original on August 31, 2018. Retrieved December 10, 2012.
  48. ^ Graves, Alex; Schmidhuber, Jürgen (2008). "Offline Handwriting Recognition with Multidimensional Recurrent Neural Networks". In Bengio, Yoshua; Schuurmans, Dale; Lafferty, John; Williams, Chris K. I.; Culotta, Aron (eds.). Advances in Neural Information Processing Systems 21 (NIPS'21). Vol. 21. Neural Information Processing Systems (NIPS) Foundation. pp. 545–552.
  49. ^ Graves, A.; Liwicki, M.; Fernandez, S.; Bertolami, R.; Bunke, H.; Schmidhuber, J. (2009). "A Novel Connectionist System for Improved Unconstrained Handwriting Recognition". IEEE Transactions on Pattern Analysis and Machine Intelligence. 31 (5): 855–868. CiteSeerX 10.1.1.139.4502. doi:10.1109/TPAMI.2008.137. PMID 19299860. S2CID 14635907.
  50. ^ Hinton, G. E.; Osindero, S.; Teh, Y. (2006). "A fast learning algorithm for deep belief nets" (PDF). Neural Computation. 18 (7): 1527–1554. CiteSeerX 10.1.1.76.1541. doi:10.1162/neco.2006.18.7.1527. PMID 16764513. S2CID 2309950.
  51. ^ Fukushima, K. (1980). "Neocognitron: A self-organizing neural network model for a mechanism of pattern recognition unaffected by shift in position". Biological Cybernetics. 36 (4): 93–202. doi:10.1007/BF00344251. PMID 7370364. S2CID 206775608.
  52. ^ Riesenhuber, M.; Poggio, T. (1999). "Hierarchical models of object recognition in cortex". Nature Neuroscience. 2 (11): 1019–1025. doi:10.1038/14819. PMID 10526343. S2CID 8920227.
  53. ^ D. C. Ciresan, U. Meier, J. Masci, J. Schmidhuber. Multi-Column Deep Neural Network for Traffic Sign Classification. Neural Networks, 2012.
  54. ^ D. Ciresan, A. Giusti, L. Gambardella, J. Schmidhuber. Deep Neural Networks Segment Neuronal Membranes in Electron Microscopy Images. In Advances in Neural Information Processing Systems (NIPS 2012), Lake Tahoe, 2012.
  55. ^ D. C. Ciresan, U. Meier, J. Schmidhuber. Multi-column Deep Neural Networks for Image Classification. IEEE Conf. on Computer Vision and Pattern Recognition CVPR 2012.
  56. ^ [Ramezanpour, A.; Beam, A.L.; Chen, J.H.; Mashaghi, A. Statistical Physics for Medical Diagnostics: Learning, Inference, and Optimization Algorithms. Diagnostics 2020, 10, 972. ]

External links Edit

Listen to this article (31 minutes)
 
This audio file was created from a revision of this article dated 27 November 2011 (2011-11-27), and does not reflect subsequent edits.
  • A Brief Introduction to Neural Networks (D. Kriesel) - Illustrated, bilingual manuscript about artificial neural networks; Topics so far: Perceptrons, Backpropagation, Radial Basis Functions, Recurrent Neural Networks, Self Organizing Maps, Hopfield Networks.
  • Review of Neural Networks in Materials Science
  • Next Generation of Neural Networks - Google Tech Talks
  • Performance of Neural Networks
  • Neural Networks and Information
  • Sanderson, Grant (October 5, 2017). "But what is a Neural Network?". 3Blue1Brown. Archived from the original on November 7, 2021 – via YouTube.

neural, network, other, uses, disambiguation, neural, network, neural, circuit, biological, neurons, sometimes, also, called, biological, neural, network, network, artificial, neurons, nodes, case, artificial, neural, network, simplified, view, feedforward, ar. For other uses see Neural network disambiguation A neural network is a neural circuit of biological neurons sometimes also called a biological neural network or a network of artificial neurons or nodes in the case of an artificial neural network 1 Simplified view of a feedforward artificial neural networkArtificial neural networks are used for solving artificial intelligence AI problems they model connections of biological neurons as weights between nodes A positive weight reflects an excitatory connection while negative values mean inhibitory connections All inputs are modified by a weight and summed This activity is referred to as a linear combination Finally an activation function controls the amplitude of the output For example an acceptable range of output is usually between 0 and 1 or it could be 1 and 1 These artificial networks may be used for predictive modeling adaptive control and applications where they can be trained via a dataset Self learning resulting from experience can occur within networks which can derive conclusions from a complex and seemingly unrelated set of information 2 Contents 1 Overview 2 History 3 Artificial intelligence 4 Applications 5 Neuroscience 5 1 Types of models 5 2 Connectivity 6 Criticism 7 Recent improvements 8 See also 9 References 10 External linksOverview EditA biological neural network is composed of a group of chemically connected or functionally associated neurons A single neuron may be connected to many other neurons and the total number of neurons and connections in a network may be extensive Connections called synapses are usually formed from axons to dendrites though dendrodendritic synapses 3 and other connections are possible Apart from electrical signalling there are other forms of signalling that arise from neurotransmitter diffusion Artificial intelligence cognitive modelling and neural networks are information processing paradigms inspired by how biological neural systems process data Artificial intelligence and cognitive modelling try to simulate some properties of biological neural networks In the artificial intelligence field artificial neural networks have been applied successfully to speech recognition image analysis and adaptive control in order to construct software agents in computer and video games or autonomous robots Historically digital computers evolved from the von Neumann model and operate via the execution of explicit instructions via access to memory by a number of processors On the other hand the origins of neural networks are based on efforts to model information processing in biological systems Unlike the von Neumann model neural network computing does not separate memory and processing Neural network theory has served to identify better how the neurons in the brain function and provide the basis for efforts to create artificial intelligence History EditThe preliminary theoretical base for contemporary neural networks was independently proposed by Alexander Bain 4 1873 and William James 5 1890 In their work both thoughts and body activity resulted from interactions among neurons within the brain nbsp Computer simulation of the branching architecture of the dendrites of pyramidal neurons 6 For Bain 4 every activity led to the firing of a certain set of neurons When activities were repeated the connections between those neurons strengthened According to his theory this repetition was what led to the formation of memory The general scientific community at the time was skeptical of Bain s 4 theory because it required what appeared to be an inordinate number of neural connections within the brain It is now apparent that the brain is exceedingly complex and that the same brain wiring can handle multiple problems and inputs James 5 theory was similar to Bain s 4 however he suggested that memories and actions resulted from electrical currents flowing among the neurons in the brain His model by focusing on the flow of electrical currents did not require individual neural connections for each memory or action C S Sherrington 7 1898 conducted experiments to test James theory He ran electrical currents down the spinal cords of rats However instead of demonstrating an increase in electrical current as projected by James Sherrington found that the electrical current strength decreased as the testing continued over time Importantly this work led to the discovery of the concept of habituation Wilhelm Lenz 1920 and Ernst Ising 1925 created and analyzed the Ising model 8 which is essentially a non learning artificial recurrent neural network RNN consisting of neuron like threshold elements 9 In 1972 Shun ichi Amari made this architecture adaptive 10 9 His learning RNN was popularised by John Hopfield in 1982 11 McCulloch and Pitts 12 1943 also created a computational model for neural networks based on mathematics and algorithms They called this model threshold logic These early models paved the way for neural network research to split into two distinct approaches One approach focused on biological processes in the brain and the other focused on the application of neural networks to artificial intelligence In the late 1940s psychologist Donald Hebb 13 created a hypothesis of learning based on the mechanism of neural plasticity that is now known as Hebbian learning Hebbian learning is considered to be a typical unsupervised learning rule and its later variants were early models for long term potentiation These ideas started being applied to computational models in 1948 with Turing s B type machines Farley and Clark 14 1954 first used computational machines then called calculators to simulate a Hebbian network at MIT Other neural network computational machines were created by Rochester Holland Habit and Duda 15 1956 Frank Rosenblatt 16 1958 created the perceptron an algorithm for pattern recognition based on a two layer learning computer network using simple addition and subtraction With mathematical notation Rosenblatt also described circuitry not in the basic perceptron such as the exclusive or circuit Some say that neural network research stagnated after the publication of machine learning research by Marvin Minsky and Seymour Papert 17 1969 They discovered two key issues with the computational machines that processed neural networks The first issue was that single layer neural networks were incapable of processing the exclusive or circuit The second significant issue was that computers were not sophisticated enough to effectively handle the long run time required by large neural networks However by the time this book came out methods for training multilayer perceptrons MLPs were already known The first deep learning MLP was published by Alexey Grigorevich Ivakhnenko and Valentin Lapa in 1965 18 19 9 The first deep learning MLP trained by stochastic gradient descent 20 was published in 1967 by Shun ichi Amari 21 9 In computer experiments conducted by Amari s student Saito a five layer MLP with two modifiable layers learned useful internal representations to classify non linearily separable pattern classes 9 Neural network research was boosted when computers achieved greater processing power Also key in later advances was the backpropagation algorithm It is an efficient application of the Leibniz chain rule 1673 22 to networks of differentiable nodes 9 It is also known as the reverse mode of automatic differentiation or reverse accumulation due to Seppo Linnainmaa 1970 23 24 25 26 9 The term back propagating errors was introduced in 1962 by Frank Rosenblatt 27 9 but he did not have an implementation of this procedure although Henry J Kelley had a continuous precursor of backpropagation 28 already in 1960 in the context of control theory 9 In 1982 Paul Werbos applied backpropagation to MLPs in the way that has become standard 29 In the late 1970s to early 1980s interest briefly emerged in theoretically investigating the Ising model by Wilhelm Lenz 1920 and Ernst Ising 1925 8 in relation to Cayley tree topologies and large neural networks In 1981 the Ising model was solved exactly by Peter Barth for the general case of closed Cayley trees with loops with an arbitrary branching ratio 30 and found to exhibit unusual phase transition behavior in its local apex and long range site site correlations 31 32 The parallel distributed processing of the mid 1980s became popular under the name connectionism The text by Rumelhart and McClelland 33 1986 provided a full exposition on the use of connectionism in computers to simulate neural processes Neural networks as used in artificial intelligence have traditionally been viewed as simplified models of neural processing in the brain even though the relation between this model and brain biological architecture is debated as it is not clear to what degree artificial neural networks mirror brain function 34 Artificial intelligence EditMain article Artificial neural network A neural network NN in the case of artificial neurons called artificial neural network ANN or simulated neural network SNN is an interconnected group of natural or artificial neurons that uses a mathematical or computational model for information processing based on a connectionistic approach to computation In most cases an ANN is an adaptive system that changes its structure based on external or internal information that flows through the network In more practical terms neural networks are non linear statistical data modeling or decision making tools They can be used to model complex relationships between inputs and outputs or to find patterns in data An artificial neural network involves a network of processing elements artificial neurons which can exhibit complex global behavior determined by the connections between the processing elements and element parameters Artificial neurons were first proposed in 1943 by Warren McCulloch a neurophysiologist and Walter Pitts a logician who first collaborated at the University of Chicago 35 One classical type of artificial neural network is the recurrent Hopfield network The concept of a neural network appears to have first been proposed by Alan Turing in his 1948 paper Intelligent Machinery in which he called them B type unorganised machines 36 The utility of artificial neural network models lies in the fact that they can be used to infer a function from observations and also to use it Unsupervised neural networks can also be used to learn representations of the input that capture the salient characteristics of the input distribution e g see the Boltzmann machine 1983 and more recently deep learning algorithms which can implicitly learn the distribution function of the observed data Learning in neural networks is particularly useful in applications where the complexity of the data or task makes the design of such functions by hand impractical Applications EditNeural networks can be used in different fields The tasks to which artificial neural networks are applied tend to fall within the following broad categories Function approximation or regression analysis including time series prediction and modeling Classification including pattern and sequence recognition novelty detection and sequential decision making Data processing including filtering clustering blind signal separation and compression Application areas of ANNs include nonlinear system identification 37 and control vehicle control process control game playing and decision making backgammon chess racing pattern recognition radar systems face identification object recognition sequence recognition gesture speech handwritten text recognition medical diagnosis financial applications data mining or knowledge discovery in databases KDD visualization and e mail spam filtering For example it is possible to create a semantic profile of user s interests emerging from pictures trained for object recognition 38 Neuroscience EditTheoretical and computational neuroscience is the field concerned with the analysis and computational modeling of biological neural systems Since neural systems are intimately related to cognitive processes and behaviour the field is closely related to cognitive and behavioural modeling The aim of the field is to create models of biological neural systems in order to understand how biological systems work To gain this understanding neuroscientists strive to make a link between observed biological processes data biologically plausible mechanisms for neural processing and learning biological neural network models and theory statistical learning theory and information theory Types of models Edit Many models are used defined at different levels of abstraction and modeling different aspects of neural systems They range from models of the short term behaviour of individual neurons through models of the dynamics of neural circuitry arising from interactions between individual neurons to models of behaviour arising from abstract neural modules that represent complete subsystems These include models of the long term and short term plasticity of neural systems and their relation to learning and memory from the individual neuron to the system level Connectivity Edit See also Brain connectivity estimators In August 2020 scientists reported that bi directional connections or added appropriate feedback connections can accelerate and improve communication between and in modular neural networks of the brain s cerebral cortex and lower the threshold for their successful communication They showed that adding feedback connections between a resonance pair can support successful propagation of a single pulse packet throughout the entire network 39 40 Criticism EditHistorically a common criticism of neural networks particularly in robotics was that they require a large diversity of training samples for real world operation although if single dimension only 8 samples required to get practically to low of 12 error This is not surprising since any learning machine needs sufficient representative examples in order to capture the underlying structure that allows it to generalize to new cases Dean Pomerleau in his research presented in the paper Knowledge based Training of Artificial Neural Networks for Autonomous Robot Driving uses a neural network to train a robotic vehicle to drive on multiple types of roads single lane multi lane dirt etc A large amount of his research is devoted to 1 extrapolating multiple training scenarios from a single training experience and 2 preserving past training diversity so that the system does not become overtrained if for example it is presented with a series of right turns it should not learn to always turn right These issues are common in neural networks that must decide from amongst a wide variety of responses but can be dealt with in several ways for example by randomly shuffling the training examples by using a numerical optimization algorithm that does not take too large steps when changing the network connections following an example or by grouping examples in so called mini batches A K Dewdney a former Scientific American columnist wrote in 1997 Although neural nets do solve a few toy problems their powers of computation are so limited that I am surprised anyone takes them seriously as a general problem solving tool 41 Arguments for Dewdney s position are that to implement large and effective software neural networks much processing and storage resources need to be committed While the brain has hardware tailored to the task of processing signals through a graph of neurons simulating even a most simplified form on Von Neumann technology may compel a neural network designer to fill many millions of database rows for its connections which can consume vast amounts of computer memory and data storage capacity Furthermore the designer of neural network systems will often need to simulate the transmission of signals through many of these connections and their associated neurons which must often be matched with incredible amounts of CPU processing power and time While neural networks often yield effective programs they too often do so at the cost of efficiency they tend to consume considerable amounts of time and money Arguments against Dewdney s position are that neural nets have been successfully used to solve many complex and diverse tasks such as autonomously flying aircraft 42 Technology writer Roger Bridgman commented on Dewdney s statements about neural nets Neural networks for instance are in the dock not only because they have been hyped to high heaven what hasn t but also because you could create a successful net without understanding how it worked the bunch of numbers that captures its behaviour would in all probability be an opaque unreadable table valueless as a scientific resource In spite of his emphatic declaration that science is not technology Dewdney seems here to pillory neural nets as bad science when most of those devising them are just trying to be good engineers An unreadable table that a useful machine could read would still be well worth having 43 Although it is true that analyzing what has been learned by an artificial neural network is difficult it is much easier to do so than to analyze what has been learned by a biological neural network Moreover recent emphasis on the explainability of AI has contributed towards the development of methods notably those based on attention mechanisms for visualizing and explaining learned neural networks Furthermore researchers involved in exploring learning algorithms for neural networks are gradually uncovering generic principles that allow a learning machine to be successful For example Bengio and LeCun 2007 wrote an article regarding local vs non local learning as well as shallow vs deep architecture 44 Some other criticisms came from believers of hybrid models combining neural networks and symbolic approaches They advocate the intermix of these two approaches and believe that hybrid models can better capture the mechanisms of the human mind Sun and Bookman 1990 full citation needed Recent improvements EditWhile initially research had been concerned mostly with the electrical characteristics of neurons a particularly important part of the investigation in recent years has been the exploration of the role of neuromodulators such as dopamine acetylcholine and serotonin on behaviour and learning citation needed Biophysical models such as BCM theory have been important in understanding mechanisms for synaptic plasticity and have had applications in both computer science and neuroscience Research is ongoing in understanding the computational algorithms used in the brain with some recent biological evidence for radial basis networks and neural backpropagation as mechanisms for processing data citation needed Computational devices have been created in CMOS for both biophysical simulation and neuromorphic computing More recent efforts show promise for creating nanodevices for very large scale principal components analyses and convolution 45 If successful these efforts could usher in a new era of neural computing that is a step beyond digital computing 46 because it depends on learning rather than programming and because it is fundamentally analog rather than digital even though the first instantiations may in fact be with CMOS digital devices Between 2009 and 2012 the recurrent neural networks and deep feedforward neural networks developed in the research group of Jurgen Schmidhuber at the Swiss AI Lab IDSIA have won eight international competitions in pattern recognition and machine learning 47 For example multi dimensional long short term memory LSTM 48 49 won three competitions in connected handwriting recognition at the 2009 International Conference on Document Analysis and Recognition ICDAR without any prior knowledge about the three different languages to be learned Variants of the back propagation algorithm as well as unsupervised methods by Geoff Hinton and colleagues at the University of Toronto can be used to train deep highly nonlinear neural architectures 50 similar to the 1980 Neocognitron by Kunihiko Fukushima 51 and the standard architecture of vision 52 inspired by the simple and complex cells identified by David H Hubel and Torsten Wiesel in the primary visual cortex Radial basis function and wavelet networks have also been introduced These can be shown to offer best approximation properties and have been applied in nonlinear system identification and classification applications 37 Deep learning feedforward networks alternate convolutional layers and max pooling layers topped by several pure classification layers Fast GPU based implementations of this approach have won several pattern recognition contests including the IJCNN 2011 Traffic Sign Recognition Competition 53 and the ISBI 2012 Segmentation of Neuronal Structures in Electron Microscopy Stacks challenge 54 Such neural networks also were the first artificial pattern recognizers to achieve human competitive or even superhuman performance 55 on benchmarks such as traffic sign recognition IJCNN 2012 or the MNIST handwritten digits problem of Yann LeCun and colleagues at NYU Analytical and computational techniques derived from statistical physics of disordered systems can be extended to large scale problems including machine learning e g to analyze the weight space of deep neural networks 56 See also EditADALINE Adaptive resonance theory Biological cybernetics Biologically inspired computing Cerebellar model articulation controller Cognitive architecture Cognitive science Connectionism Connectomics Cultured neuronal networks Deep learning Deep Image Prior Digital morphogenesis Efficiently updatable neural network Exclusive or Evolutionary algorithm Genetic algorithm Gene expression programming Generative adversarial network Group method of data handling Habituation In situ adaptive tabulation Memristor Neural network software Nonlinear system identification Parallel constraint satisfaction processes Parallel distributed processing Predictive analytics Radial basis function network Self organizing map Simulated reality Support vector machine Time delay neural network BackpropagationReferences Edit Hopfield J J 1982 Neural networks and physical systems with emergent collective computational abilities Proc Natl Acad Sci U S A 79 8 2554 2558 Bibcode 1982PNAS 79 2554H doi 10 1073 pnas 79 8 2554 PMC 346238 PMID 6953413 Neural Net or Neural Network Gartner IT Glossary gartner com Arbib p 666 a b c d Bain 1873 Mind and Body The Theories of Their Relation New York D Appleton and Company a b James 1890 The Principles of Psychology New York H Holt and Company Cuntz Hermann 2010 PLoS Computational Biology Issue Image Vol 6 8 August 2010 PLOS Computational Biology 6 8 ev06 i08 doi 10 1371 image pcbi v06 i08 Sherrington C S 1898 Experiments in Examination of the Peripheral Distribution of the Fibers of the Posterior Roots of Some Spinal Nerves Proceedings of the Royal Society of London 190 45 186 doi 10 1098 rstb 1898 0002 a b Brush Stephen G 1967 History of the Lenz Ising Model Reviews of Modern Physics 39 4 883 893 Bibcode 1967RvMP 39 883B doi 10 1103 RevModPhys 39 883 a b c d e f g h i Schmidhuber Juergen 2022 Annotated History of Modern AI and Deep Learning arXiv 2212 11279 cs NE Amari Shun Ichi 1972 Learning patterns and pattern sequences by self organizing nets of threshold elements IEEE Transactions C 21 1197 1206 Hopfield J J 1982 Neural networks and physical systems with emergent collective computational abilities Proceedings of the National Academy of Sciences 79 8 2554 2558 Bibcode 1982PNAS 79 2554H doi 10 1073 pnas 79 8 2554 PMC 346238 PMID 6953413 McCulloch Warren Walter Pitts 1943 A Logical Calculus of Ideas Immanent in Nervous Activity Bulletin of Mathematical Biophysics 5 4 115 133 doi 10 1007 BF02478259 Hebb Donald 1949 The Organization of Behavior New York Wiley Farley B W A Clark 1954 Simulation of Self Organizing Systems by Digital Computer IRE Transactions on Information Theory 4 4 76 84 doi 10 1109 TIT 1954 1057468 Rochester N J H Holland L H Habit and W L Duda 1956 Tests on a cell assembly theory of the action of the brain using a large digital computer IRE Transactions on Information Theory 2 3 80 93 doi 10 1109 TIT 1956 1056810 Rosenblatt F 1958 The Perceptron A Probabilistic Model For Information Storage And Organization In The Brain Psychological Review 65 6 386 408 CiteSeerX 10 1 1 588 3775 doi 10 1037 h0042519 PMID 13602029 S2CID 12781225 Minsky M S Papert 1969 An Introduction to Computational Geometry MIT Press ISBN 978 0 262 63022 1 Ivakhnenko A G 1973 Cybernetic Predicting Devices CCM Information Corporation Ivakhnenko A G Grigorʹevich Lapa Valentin 1967 Cybernetics and forecasting techniques American Elsevier Pub Co Robbins H Monro S 1951 A Stochastic Approximation Method The Annals of Mathematical Statistics 22 3 400 doi 10 1214 aoms 1177729586 Amari Shun ichi 1967 A theory of adaptive pattern classifier IEEE Transactions EC 16 279 307 Leibniz Gottfried Wilhelm Freiherr von 1920 The Early Mathematical Manuscripts of Leibniz Translated from the Latin Texts Published by Carl Immanuel Gerhardt with Critical and Historical Notes Leibniz published the chain rule in a 1676 memoir Open court publishing Company ISBN 9780598818461 Linnainmaa Seppo 1970 The representation of the cumulative rounding error of an algorithm as a Taylor expansion of the local rounding errors Masters in Finnish University of Helsinki pp 6 7 Linnainmaa Seppo 1976 Taylor expansion of the accumulated rounding error BIT Numerical Mathematics 16 2 146 160 doi 10 1007 bf01931367 S2CID 122357351 Griewank Andreas 2012 Who Invented the Reverse Mode of Differentiation Optimization Stories Documenta Matematica Extra Volume ISMP pp 389 400 S2CID 15568746 Griewank Andreas Walther Andrea 2008 Evaluating Derivatives Principles and Techniques of Algorithmic Differentiation Second Edition SIAM ISBN 978 0 89871 776 1 Rosenblatt Frank 1962 Principles of Neurodynamics Spartan New York Kelley Henry J 1960 Gradient theory of optimal flight paths ARS Journal 30 10 947 954 doi 10 2514 8 5282 Werbos Paul 1982 Applications of advances in nonlinear sensitivity analysis PDF System modeling and optimization Springer pp 762 770 Archived PDF from the original on April 14 2016 Retrieved July 2 2017 Barth Peter F 1981 Cooperativity and the Transition Behavior of Large Neural Nets MS thesis Burlington University of Vermont OCLC 8231704 Krizan J E Barth P F Glasser M L 1983 Exact Phase Transitions for the Ising Model on the Closed Cayley Tree Physica North Holland Publishing Co 119A 230 242 doi 10 1016 0378 4371 83 90157 7 Glasser M L Goldberg M 1983 The Ising model on a closed Cayley tree Physica 117A 2 3 670 672 Bibcode 1983PhyA 117 670G doi 10 1016 0378 4371 83 90138 3 Rumelhart D E James McClelland 1986 Parallel Distributed Processing Explorations in the Microstructure of Cognition Cambridge MIT Press Russell Ingrid Neural Networks Module Archived from the original on May 29 2014 McCulloch Warren Pitts Walter 1943 A Logical Calculus of Ideas Immanent in Nervous Activity Bulletin of Mathematical Biophysics 5 4 115 133 doi 10 1007 BF02478259 Copeland B Jack ed 2004 The Essential Turing Oxford University Press p 403 ISBN 978 0 19 825080 7 a b Billings S A 2013 Nonlinear System Identification NARMAX Methods in the Time Frequency and Spatio Temporal Domains Wiley ISBN 978 1 119 94359 4 Wieczorek Szymon Filipiak Dominik Filipowska Agata 2018 Semantic Image Based Profiling of Users Interests with Neural Networks Studies on the Semantic Web 36 Emerging Topics in Semantic Technologies doi 10 3233 978 1 61499 894 5 179 Neuroscientists demonstrate how to improve communication between different regions of the brain medicalxpress com Retrieved September 6 2020 Rezaei Hedyeh Aertsen Ad Kumar Arvind Valizadeh Alireza August 10 2020 Facilitating the propagation of spiking activity in feedforward networks by including feedback PLOS Computational Biology 16 8 e1008033 Bibcode 2020PLSCB 16E8033R doi 10 1371 journal pcbi 1008033 ISSN 1553 7358 PMC 7444537 PMID 32776924 S2CID 221100528 nbsp Text and images are available under a Creative Commons Attribution 4 0 International License Dewdney A K 1997 Yes we have no neutrons an eye opening tour through the twists and turns of bad science New York Wiley ISBN 0 471 10806 5 OCLC 35558945 Administrator NASA June 5 2013 Dryden Flight Research Center News Room News Releases NASA NEURAL NETWORK PROJECT PASSES MILESTONE NASA Roger Bridgman s defence of neural networks Archived from the original on March 19 2012 Retrieved August 1 2006 Scaling Learning Algorithms towards AI LISA Publications Aigaion 2 0 www iro umontreal ca Yang J J et al 2008 Memristive switching mechanism for metal oxide metal nanodevices Nat Nanotechnol 3 7 429 433 doi 10 1038 nnano 2008 160 PMID 18654568 Strukov D B et al 2008 The missing memristor found Nature 453 7191 80 83 Bibcode 2008Natur 453 80S doi 10 1038 nature06932 PMID 18451858 S2CID 4367148 2012 Kurzweil AI Interview with Jurgen Schmidhuber on the eight competitions won by his Deep Learning team 2009 2012 Archived from the original on August 31 2018 Retrieved December 10 2012 Graves Alex Schmidhuber Jurgen 2008 Offline Handwriting Recognition with Multidimensional Recurrent Neural Networks In Bengio Yoshua Schuurmans Dale Lafferty John Williams Chris K I Culotta Aron eds Advances in Neural Information Processing Systems 21 NIPS 21 Vol 21 Neural Information Processing Systems NIPS Foundation pp 545 552 Graves A Liwicki M Fernandez S Bertolami R Bunke H Schmidhuber J 2009 A Novel Connectionist System for Improved Unconstrained Handwriting Recognition IEEE Transactions on Pattern Analysis and Machine Intelligence 31 5 855 868 CiteSeerX 10 1 1 139 4502 doi 10 1109 TPAMI 2008 137 PMID 19299860 S2CID 14635907 Hinton G E Osindero S Teh Y 2006 A fast learning algorithm for deep belief nets PDF Neural Computation 18 7 1527 1554 CiteSeerX 10 1 1 76 1541 doi 10 1162 neco 2006 18 7 1527 PMID 16764513 S2CID 2309950 Fukushima K 1980 Neocognitron A self organizing neural network model for a mechanism of pattern recognition unaffected by shift in position Biological Cybernetics 36 4 93 202 doi 10 1007 BF00344251 PMID 7370364 S2CID 206775608 Riesenhuber M Poggio T 1999 Hierarchical models of object recognition in cortex Nature Neuroscience 2 11 1019 1025 doi 10 1038 14819 PMID 10526343 S2CID 8920227 D C Ciresan U Meier J Masci J Schmidhuber Multi Column Deep Neural Network for Traffic Sign Classification Neural Networks 2012 D Ciresan A Giusti L Gambardella J Schmidhuber Deep Neural Networks Segment Neuronal Membranes in Electron Microscopy Images In Advances in Neural Information Processing Systems NIPS 2012 Lake Tahoe 2012 D C Ciresan U Meier J Schmidhuber Multi column Deep Neural Networks for Image Classification IEEE Conf on Computer Vision and Pattern Recognition CVPR 2012 Ramezanpour A Beam A L Chen J H Mashaghi A Statistical Physics for Medical Diagnostics Learning Inference and Optimization Algorithms Diagnostics 2020 10 972 External links EditListen to this article 31 minutes source source nbsp This audio file was created from a revision of this article dated 27 November 2011 2011 11 27 and does not reflect subsequent edits Audio help More spoken articles A Brief Introduction to Neural Networks D Kriesel Illustrated bilingual manuscript about artificial neural networks Topics so far Perceptrons Backpropagation Radial Basis Functions Recurrent Neural Networks Self Organizing Maps Hopfield Networks Review of Neural Networks in Materials Science Artificial Neural Networks Tutorial in three languages Univ Politecnica de Madrid Another introduction to ANN Next Generation of Neural Networks Google Tech Talks Performance of Neural Networks Neural Networks and Information Sanderson Grant October 5 2017 But what is a Neural Network 3Blue1Brown Archived from the original on November 7 2021 via YouTube Retrieved from https en wikipedia org w index php title Neural network amp oldid 1180911913, wikipedia, wiki, book, books, library,

article

, read, download, free, free download, mp3, video, mp4, 3gp, jpg, jpeg, gif, png, picture, music, song, movie, book, game, games.