The Behavior of the Complex Integral Neural Network (original) (raw)
- 5117 Accesses
Abstract
In this paper, we introduce an Integral Neural Network based on complex domain. We describe the model of the neuron, analyze the behavior of the neuron, and indicate that in certain conditions it performs the calculation of Fourier Integration. Have studied on the neural network with hidden layers, we obtain the following facts: 1. This kind of structure can memorize a time variant function; 2. It calculates the convolution of input series and the function the neural network memorized; 3. This neural network structure also can calculate the correlation function; 4. In the case of many hidden layers, it can perform the Fourier Transform with many variants.
Preview
Unable to display preview. Download preview PDF.
Similar content being viewed by others
Reference
- Iku Nemoto, Makoto Kubono, “Complex Associative Memory”, Neural Networks, Vol. 9, No. 2 (1996) 253–261.
Article Google Scholar - T. Nitta, “An Extension of the Back-Propagation Algorithm to Complex Numbers”, Neural Networks, Vol. 9, No. 2 (1996) 1392–1414
Google Scholar - Stanislaw Jankowski, andrzej Lozowski, and Jacek M. Zurada, “Complex-Valued Multistate Neural Associative Memory”, IEEE Trans, on Neural Networks, Vol.7, No.6, (Nov. 1996) 1491–1496
Article Google Scholar - Srinivasa V. Chakravarthy and Joydeep Ghosh, “A Neural Network-based Associative Memory for Storing Complex-valued Patterns”, Systems, Man, and Cybernetics, 1994, Intel. Conf., 2213–2218
Google Scholar - Widrow, B., McCool, J., & Ball, M., “The complex LMS algorithm”, Proceedings of the IEEE, 63(4) (1975) 719–720
Google Scholar - A. Hirose, “Proposal of fully complex-valued Neural Networks”, In Proceedings of the International Joint Conference on Neural Networks, Vol. 4 (1992) 152–157
Google Scholar - Pan Yong, Li lei and Shi Hongbao, “An Integral Neural Network Based on Complex Domain”, Advances in Computer Science and Technology, Proceeding of the Fifth International Conference for Young Computer Scientists, Vol.1 (1999) 537–539
Google Scholar - Pan Yong, Li lei and Shi Hongbao, “An Integral Neural Network Based on Complex Domain and its applications”, Proceeding of International Symposium on Domain Theory (1999)
Google Scholar - Price, D et al, Pairwise Neural Network Classifiers with Probabilistic Outputs, NIPS (1996) 1009–1116
Google Scholar - Ulbricht C, Dorffner G., Canu S., Guillemyn D., Marijuan G., Olarte J., Rodriguez C, Martin L, Mechanisms for Handling Sequences with Neural Networks, Intelligent Engineering Systems through Artificial Neural Networks, Volume 2, ASME Press, New York, 1992.
Google Scholar - Ulbricht C, Handling Sequences with a Competitive Recurrent Network, Proceedings of the International Joint Conference on Neural Networks, Baltimore, 1992.
Google Scholar
Author information
Authors and Affiliations
- Institute of Computing Technology, Tongji University(West Campus), 200331, Shanghai
Pan Yong - Suzhou Railway Teachers College, Suzhou, 215009, Jiangsu
Shi Hongbao & Li Lei
Authors
- Pan Yong
- Shi Hongbao
- Li Lei
Editor information
Editors and Affiliations
- Department of Computer Science, University of Regina Regina, S4S 0A2, Saskatchewan, Canada
Wojciech Ziarko & Yiyu Yao &
Rights and permissions
Copyright information
© 2001 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Yong, P., Hongbao, S., Lei, L. (2001). The Behavior of the Complex Integral Neural Network. In: Ziarko, W., Yao, Y. (eds) Rough Sets and Current Trends in Computing. RSCTC 2000. Lecture Notes in Computer Science(), vol 2005. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-45554-X\_79
Download citation
- .RIS
- .ENW
- .BIB
- DOI: https://doi.org/10.1007/3-540-45554-X\_79
- Published: 18 December 2001
- Publisher Name: Springer, Berlin, Heidelberg
- Print ISBN: 978-3-540-43074-2
- Online ISBN: 978-3-540-45554-7
- eBook Packages: Springer Book Archive
Keywords
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.