A denoising autoencoder for pretraining sequence-to-sequence models. BART is structurally similar to a traditional transformer encoder-decoder model, but is pre-trained like a denoising autoencoder. 27.07.2023 17:54 aior