We analytically determine that the backward-error-propagation learning algorithm has a well-defined region of convergence in neural learning-parameter space for two classes of photorefractive-based optical neural-network architectures. The first class uses electric-field amplitude encoding of signals and weights in a fully coherent system, whereas the second class uses intensity encoding of signals and weights in an incoherent/coherent system. Under typical assumptions on the grating formation in photorefractive materials used in adaptive optical interconnections, we compute weight updates for both classes of architectures. Using these weight updates, we derive a set of conditions that are sufficient for such a network to operate within the region of convergence. The results are verified empirically by simulations of the XOR sample problem. The computed weight updates for both classes of architectures contain two neural learning parameters: a learning-rate coefficient and a weight-decay coefficient. We show that these learning parameters are directly related to two important design parameters: system gain and exposure energy. The system gain determines the ratio of the learning-rate parameter to decay-rate parameter, and the exposure energy determines the size of the decay-rate parameter. We conclude that convergence is guaranteed (assuming no spurious local minima in the error function) by using a sufficiently high gain and a sufficiently low exposure energy per weight update.
© 1996 Optical Society of AmericaFull Article | PDF Article
Chau-Jern Cheng, Pochi Yeh, and Ken Yuh Hsu
J. Opt. Soc. Am. B 11(9) 1619-1624 (1994)
Kelvin Wagner and Demetri Psaltis
Appl. Opt. 26(23) 5061-5076 (1987)
Demetri Psaltis, David Brady, and Kelvin Wagner
Appl. Opt. 27(9) 1752-1759 (1988)