DocumentCode
1950557
Title
Optimising the Hystereses of a Two Context Layer RNN for Text Classification
Author
Arevian, Garen ; Panchev, Christo
Author_Institution
Sunderland Univ., Sunderland
fYear
2007
fDate
12-17 Aug. 2007
Firstpage
2936
Lastpage
2941
Abstract
Established techniques from information retrieval (IR) and machine learning (ML) have shown varying degrees of success in the automatic classification of real-world text. The capabilities of an extended version of the Simple recurrent network (SRN) for classifying news titles from the Reuters-21578 Corpus are explored. The architecture is composed of two hidden layers where each layer has an associated context layer that takes copies of previous activation states and integrates them with current activations. This results in improved performance, stability and generalisation by the adjustment of the percentage of previous activation strengths kept "in memory" by what is defined as the hysteresis parameter. The study demonstrates that this partial feedback of activations must be carefully fine-tuned to maintain optimal performance. Correctly adjusting the hysteresis values for very long and noisy text sequences is critical as classification performance degrades catastrophic ally when values are not optimally set.
Keywords
classification; information retrieval; learning (artificial intelligence); recurrent neural nets; text analysis; information retrieval; machine learning; recurrent neural network; text classification; two context layer RNN; Computer architecture; Degradation; Hysteresis; Information retrieval; Machine learning; Neural networks; Recurrent neural networks; Stability; Text categorization; USA Councils;
fLanguage
English
Publisher
ieee
Conference_Titel
Neural Networks, 2007. IJCNN 2007. International Joint Conference on
Conference_Location
Orlando, FL
ISSN
1098-7576
Print_ISBN
978-1-4244-1379-9
Electronic_ISBN
1098-7576
Type
conf
DOI
10.1109/IJCNN.2007.4371427
Filename
4371427
Link To Document