DocumentCode :
1950557
Title :
Optimising the Hystereses of a Two Context Layer RNN for Text Classification
Author :
Arevian, Garen ; Panchev, Christo
Author_Institution :
Sunderland Univ., Sunderland
fYear :
2007
fDate :
12-17 Aug. 2007
Firstpage :
2936
Lastpage :
2941
Abstract :
Established techniques from information retrieval (IR) and machine learning (ML) have shown varying degrees of success in the automatic classification of real-world text. The capabilities of an extended version of the Simple recurrent network (SRN) for classifying news titles from the Reuters-21578 Corpus are explored. The architecture is composed of two hidden layers where each layer has an associated context layer that takes copies of previous activation states and integrates them with current activations. This results in improved performance, stability and generalisation by the adjustment of the percentage of previous activation strengths kept "in memory" by what is defined as the hysteresis parameter. The study demonstrates that this partial feedback of activations must be carefully fine-tuned to maintain optimal performance. Correctly adjusting the hysteresis values for very long and noisy text sequences is critical as classification performance degrades catastrophic ally when values are not optimally set.
Keywords :
classification; information retrieval; learning (artificial intelligence); recurrent neural nets; text analysis; information retrieval; machine learning; recurrent neural network; text classification; two context layer RNN; Computer architecture; Degradation; Hysteresis; Information retrieval; Machine learning; Neural networks; Recurrent neural networks; Stability; Text categorization; USA Councils;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks, 2007. IJCNN 2007. International Joint Conference on
Conference_Location :
Orlando, FL
ISSN :
1098-7576
Print_ISBN :
978-1-4244-1379-9
Electronic_ISBN :
1098-7576
Type :
conf
DOI :
10.1109/IJCNN.2007.4371427
Filename :
4371427
Link To Document :
بازگشت