Abstract
A low-complexity model for signal quality prediction in a nonlinear fiber-optic network is developed. The model, which builds on the Gaussian noise model, takes into account the signal degradation caused by a combination of chromatic dispersion, nonlinear signal distortion, and amplifier noise. The center frequencies, bandwidths, and transmit powers can be chosen independently for each channel, which makes the model suitable for analysis and optimization of resource allocation and routing in large-scale optical networks applying flexible-grid wavelength-division multiplexing.
© 2014 IEEE
PDF Article
More Like This
Cited By
You do not have subscription access to this journal. Cited by links are available to subscribers only. You may subscribe either as an Optica member, or as an authorized user of your institution.
Contact your librarian or system administrator
or
Login to access Optica Member Subscription