fbpx
Wikipedia

Recursive neural network

A recursive neural network is a kind of deep neural network created by applying the same set of weights recursively over a structured input, to produce a structured prediction over variable-size input structures, or a scalar prediction on it, by traversing a given structure in topological order. Recursive neural networks, sometimes abbreviated as RvNNs, have been successful, for instance, in learning sequence and tree structures in natural language processing, mainly phrase and sentence continuous representations based on word embedding. RvNNs have first been introduced to learn distributed representations of structure, such as logical terms.[1] Models and general frameworks have been developed in further works since the 1990s.[2][3]

Architectures edit

Basic edit

 
A simple recursive neural network architecture

In the most simple architecture, nodes are combined into parents using a weight matrix that is shared across the whole network, and a non-linearity such as tanh. If c1 and c2 are n-dimensional vector representation of nodes, their parent will also be an n-dimensional vector, calculated as

 

Where W is a learned   weight matrix.

This architecture, with a few improvements, has been used for successfully parsing natural scenes, syntactic parsing of natural language sentences,[4] and recursive autoencoding and generative modeling of 3D shape structures in the form of cuboid abstractions.[5]

Recursive cascade correlation (RecCC) edit

RecCC is a constructive neural network approach to deal with tree domains[2] with pioneering applications to chemistry[6] and extension to directed acyclic graphs.[7]

Unsupervised RNN edit

A framework for unsupervised RNN has been introduced in 2004.[8][9]

Tensor edit

Recursive neural tensor networks use one, tensor-based composition function for all nodes in the tree.[10]

Training edit

Stochastic gradient descent edit

Typically, stochastic gradient descent (SGD) is used to train the network. The gradient is computed using backpropagation through structure (BPTS), a variant of backpropagation through time used for recurrent neural networks.

Properties edit

Universal approximation capability of RNN over trees has been proved in literature.[11][12]

Related models edit

Recurrent neural networks edit

Recurrent neural networks are recursive artificial neural networks with a certain structure: that of a linear chain. Whereas recursive neural networks operate on any hierarchical structure, combining child representations into parent representations, recurrent neural networks operate on the linear progression of time, combining the previous time step and a hidden representation into the representation for the current time step.

Tree Echo State Networks edit

An efficient approach to implement recursive neural networks is given by the Tree Echo State Network[13] within the reservoir computing paradigm.

Extension to graphs edit

Extensions to graphs include graph neural network (GNN),[14] Neural Network for Graphs (NN4G),[15] and more recently convolutional neural networks for graphs.

References edit

  1. ^ Goller, C.; Küchler, A. (1996). "Learning task-dependent distributed representations by backpropagation through structure". Proceedings of International Conference on Neural Networks (ICNN'96). Vol. 1. pp. 347–352. CiteSeerX 10.1.1.52.4759. doi:10.1109/ICNN.1996.548916. ISBN 978-0-7803-3210-2. S2CID 6536466.
  2. ^ a b Sperduti, A.; Starita, A. (1997-05-01). "Supervised neural networks for the classification of structures". IEEE Transactions on Neural Networks. 8 (3): 714–735. doi:10.1109/72.572108. ISSN 1045-9227. PMID 18255672.
  3. ^ Frasconi, P.; Gori, M.; Sperduti, A. (1998-09-01). "A general framework for adaptive processing of data structures". IEEE Transactions on Neural Networks. 9 (5): 768–786. CiteSeerX 10.1.1.64.2580. doi:10.1109/72.712151. ISSN 1045-9227. PMID 18255765.
  4. ^ Socher, Richard; Lin, Cliff; Ng, Andrew Y.; Manning, Christopher D. "Parsing Natural Scenes and Natural Language with Recursive Neural Networks" (PDF). The 28th International Conference on Machine Learning (ICML 2011).
  5. ^ Li, Jun; Xu, Kai; Chaudhuri, Siddhartha; Yumer, Ersin; Zhang, Hao; Guibas, Leonadis (2017). "GRASS: Generative Recursive Autoencoders for Shape Structures" (PDF). ACM Transactions on Graphics. 36 (4): 52. arXiv:1705.02090. doi:10.1145/3072959.3073613. S2CID 20432407.
  6. ^ Bianucci, Anna Maria; Micheli, Alessio; Sperduti, Alessandro; Starita, Antonina (2000). "Application of Cascade Correlation Networks for Structures to Chemistry". Applied Intelligence. 12 (1–2): 117–147. doi:10.1023/A:1008368105614. ISSN 0924-669X. S2CID 10031212.
  7. ^ Micheli, A.; Sona, D.; Sperduti, A. (2004-11-01). "Contextual processing of structured data by recursive cascade correlation". IEEE Transactions on Neural Networks. 15 (6): 1396–1410. CiteSeerX 10.1.1.135.8772. doi:10.1109/TNN.2004.837783. ISSN 1045-9227. PMID 15565768. S2CID 12370239.
  8. ^ Hammer, Barbara; Micheli, Alessio; Sperduti, Alessandro; Strickert, Marc (2004). "Recursive self-organizing network models". Neural Networks. 17 (8–9): 1061–1085. CiteSeerX 10.1.1.129.6155. doi:10.1016/j.neunet.2004.06.009. PMID 15555852.
  9. ^ Hammer, Barbara; Micheli, Alessio; Sperduti, Alessandro; Strickert, Marc (2004-03-01). "A general framework for unsupervised processing of structured data". Neurocomputing. 57: 3–35. CiteSeerX 10.1.1.3.984. doi:10.1016/j.neucom.2004.01.008.
  10. ^ Socher, Richard; Perelygin, Alex; Y. Wu, Jean; Chuang, Jason; D. Manning, Christopher; Y. Ng, Andrew; Potts, Christopher. "Recursive Deep Models for Semantic Compositionality Over a Sentiment Treebank" (PDF). EMNLP 2013.
  11. ^ Hammer, Barbara (2007-10-03). Learning with Recurrent Neural Networks. Springer. ISBN 9781846285677.
  12. ^ Hammer, Barbara; Micheli, Alessio; Sperduti, Alessandro (2005-05-01). "Universal Approximation Capability of Cascade Correlation for Structures". Neural Computation. 17 (5): 1109–1159. CiteSeerX 10.1.1.138.2224. doi:10.1162/0899766053491878. S2CID 10845957.
  13. ^ Gallicchio, Claudio; Micheli, Alessio (2013-02-04). "Tree Echo State Networks". Neurocomputing. 101: 319–337. doi:10.1016/j.neucom.2012.08.017. hdl:11568/158480.
  14. ^ Scarselli, F.; Gori, M.; Tsoi, A. C.; Hagenbuchner, M.; Monfardini, G. (2009-01-01). "The Graph Neural Network Model". IEEE Transactions on Neural Networks. 20 (1): 61–80. doi:10.1109/TNN.2008.2005605. ISSN 1045-9227. PMID 19068426. S2CID 206756462.
  15. ^ Micheli, A. (2009-03-01). "Neural Network for Graphs: A Contextual Constructive Approach". IEEE Transactions on Neural Networks. 20 (3): 498–511. doi:10.1109/TNN.2008.2010350. ISSN 1045-9227. PMID 19193509. S2CID 17486263.


recursive, neural, network, confused, with, recurrent, neural, network, recursive, neural, network, kind, deep, neural, network, created, applying, same, weights, recursively, over, structured, input, produce, structured, prediction, over, variable, size, inpu. Not to be confused with recurrent neural network A recursive neural network is a kind of deep neural network created by applying the same set of weights recursively over a structured input to produce a structured prediction over variable size input structures or a scalar prediction on it by traversing a given structure in topological order Recursive neural networks sometimes abbreviated as RvNNs have been successful for instance in learning sequence and tree structures in natural language processing mainly phrase and sentence continuous representations based on word embedding RvNNs have first been introduced to learn distributed representations of structure such as logical terms 1 Models and general frameworks have been developed in further works since the 1990s 2 3 Contents 1 Architectures 1 1 Basic 1 2 Recursive cascade correlation RecCC 1 3 Unsupervised RNN 1 4 Tensor 2 Training 2 1 Stochastic gradient descent 3 Properties 4 Related models 4 1 Recurrent neural networks 4 2 Tree Echo State Networks 4 3 Extension to graphs 5 ReferencesArchitectures editBasic edit nbsp A simple recursive neural network architecture In the most simple architecture nodes are combined into parents using a weight matrix that is shared across the whole network and a non linearity such as tanh If c1 and c2 are n dimensional vector representation of nodes their parent will also be an n dimensional vector calculated asp 1 2 tanh W c 1 c 2 displaystyle p 1 2 tanh left W c 1 c 2 right nbsp Where W is a learned n 2 n displaystyle n times 2n nbsp weight matrix This architecture with a few improvements has been used for successfully parsing natural scenes syntactic parsing of natural language sentences 4 and recursive autoencoding and generative modeling of 3D shape structures in the form of cuboid abstractions 5 Recursive cascade correlation RecCC edit RecCC is a constructive neural network approach to deal with tree domains 2 with pioneering applications to chemistry 6 and extension to directed acyclic graphs 7 Unsupervised RNN edit A framework for unsupervised RNN has been introduced in 2004 8 9 Tensor edit Recursive neural tensor networks use one tensor based composition function for all nodes in the tree 10 Training editStochastic gradient descent edit Typically stochastic gradient descent SGD is used to train the network The gradient is computed using backpropagation through structure BPTS a variant of backpropagation through time used for recurrent neural networks Properties editUniversal approximation capability of RNN over trees has been proved in literature 11 12 Related models editRecurrent neural networks edit Main article Recurrent neural network Recurrent neural networks are recursive artificial neural networks with a certain structure that of a linear chain Whereas recursive neural networks operate on any hierarchical structure combining child representations into parent representations recurrent neural networks operate on the linear progression of time combining the previous time step and a hidden representation into the representation for the current time step Tree Echo State Networks edit An efficient approach to implement recursive neural networks is given by the Tree Echo State Network 13 within the reservoir computing paradigm Extension to graphs edit Extensions to graphs include graph neural network GNN 14 Neural Network for Graphs NN4G 15 and more recently convolutional neural networks for graphs References edit Goller C Kuchler A 1996 Learning task dependent distributed representations by backpropagation through structure Proceedings of International Conference on Neural Networks ICNN 96 Vol 1 pp 347 352 CiteSeerX 10 1 1 52 4759 doi 10 1109 ICNN 1996 548916 ISBN 978 0 7803 3210 2 S2CID 6536466 a b Sperduti A Starita A 1997 05 01 Supervised neural networks for the classification of structures IEEE Transactions on Neural Networks 8 3 714 735 doi 10 1109 72 572108 ISSN 1045 9227 PMID 18255672 Frasconi P Gori M Sperduti A 1998 09 01 A general framework for adaptive processing of data structures IEEE Transactions on Neural Networks 9 5 768 786 CiteSeerX 10 1 1 64 2580 doi 10 1109 72 712151 ISSN 1045 9227 PMID 18255765 Socher Richard Lin Cliff Ng Andrew Y Manning Christopher D Parsing Natural Scenes and Natural Language with Recursive Neural Networks PDF The 28th International Conference on Machine Learning ICML 2011 Li Jun Xu Kai Chaudhuri Siddhartha Yumer Ersin Zhang Hao Guibas Leonadis 2017 GRASS Generative Recursive Autoencoders for Shape Structures PDF ACM Transactions on Graphics 36 4 52 arXiv 1705 02090 doi 10 1145 3072959 3073613 S2CID 20432407 Bianucci Anna Maria Micheli Alessio Sperduti Alessandro Starita Antonina 2000 Application of Cascade Correlation Networks for Structures to Chemistry Applied Intelligence 12 1 2 117 147 doi 10 1023 A 1008368105614 ISSN 0924 669X S2CID 10031212 Micheli A Sona D Sperduti A 2004 11 01 Contextual processing of structured data by recursive cascade correlation IEEE Transactions on Neural Networks 15 6 1396 1410 CiteSeerX 10 1 1 135 8772 doi 10 1109 TNN 2004 837783 ISSN 1045 9227 PMID 15565768 S2CID 12370239 Hammer Barbara Micheli Alessio Sperduti Alessandro Strickert Marc 2004 Recursive self organizing network models Neural Networks 17 8 9 1061 1085 CiteSeerX 10 1 1 129 6155 doi 10 1016 j neunet 2004 06 009 PMID 15555852 Hammer Barbara Micheli Alessio Sperduti Alessandro Strickert Marc 2004 03 01 A general framework for unsupervised processing of structured data Neurocomputing 57 3 35 CiteSeerX 10 1 1 3 984 doi 10 1016 j neucom 2004 01 008 Socher Richard Perelygin Alex Y Wu Jean Chuang Jason D Manning Christopher Y Ng Andrew Potts Christopher Recursive Deep Models for Semantic Compositionality Over a Sentiment Treebank PDF EMNLP 2013 Hammer Barbara 2007 10 03 Learning with Recurrent Neural Networks Springer ISBN 9781846285677 Hammer Barbara Micheli Alessio Sperduti Alessandro 2005 05 01 Universal Approximation Capability of Cascade Correlation for Structures Neural Computation 17 5 1109 1159 CiteSeerX 10 1 1 138 2224 doi 10 1162 0899766053491878 S2CID 10845957 Gallicchio Claudio Micheli Alessio 2013 02 04 Tree Echo State Networks Neurocomputing 101 319 337 doi 10 1016 j neucom 2012 08 017 hdl 11568 158480 Scarselli F Gori M Tsoi A C Hagenbuchner M Monfardini G 2009 01 01 The Graph Neural Network Model IEEE Transactions on Neural Networks 20 1 61 80 doi 10 1109 TNN 2008 2005605 ISSN 1045 9227 PMID 19068426 S2CID 206756462 Micheli A 2009 03 01 Neural Network for Graphs A Contextual Constructive Approach IEEE Transactions on Neural Networks 20 3 498 511 doi 10 1109 TNN 2008 2010350 ISSN 1045 9227 PMID 19193509 S2CID 17486263 nbsp This artificial intelligence related article is a stub You can help Wikipedia by expanding it vte Retrieved from https en wikipedia org w index php title Recursive neural network amp oldid 1129503968, wikipedia, wiki, book, books, library,

article

, read, download, free, free download, mp3, video, mp4, 3gp, jpg, jpeg, gif, png, picture, music, song, movie, book, game, games.