We address the problem of soliton phase jitter caused by fiber amplifier
noise in a communication channel. An analytical expression for phase jitter
is derived by use of perturbation theory. In the absence of filters the variance
of the phase is found to be proportional to the cube of the propagation distance,
but it increases only linearly if in-line filters are used. We verify with
simulations that in-line filters maintain the phase jitter at a level that
is tolerable for transoceanic links.
© 1999 Optical Society of America
Equations on this page are rendered with MathJax. Learn more.