Texture and color distortion are common in existing learning-based dehazing algorithms, and it is argued that one of the major reasons is that the shallow features of fog images are underutilized, and the deep features of fog images are insufficient for single image dehazing. In order to provide more texture and color information for image restoration, more shallow features need to be added in the process of image decoding. Therefore, a multi-scale feature fusion pyramid attention network (PAN) for single image dehazing is proposed. In PAN, combined with the attention mechanism, a shallow and deep feature fusion (SDF) strategy is designed. SDF considers multi-scale as well as channel-level fusion to provide feature information under different receptive fields while also highlighting important channels, such as texture and color information. DC is designed as a latent space mapping module to learn a mapping relationship between the latent space representation of the hazy image at low resolution and the corresponding latent space representation of the haze-free image. Additionally, network deconvolution (ND) and deformed convolution network (DCN) are introduced into PAN. The ND module can remove pixel-wise and channel-wise correlation of features, reduce data redundancy to obtain sparse representation of features, and speed up network convergence. The DCN module can use its adaptive receptive field to focus on the area of interest for calculation and play a role in texture feature enhancement. Finally, the perceptual loss is chosen as the regularization item of the loss function, which makes style features of the restored image closer to the real fog-free image. Extensive experiments reveal that the proposed PAN outperforms other existing dehazing methods on real-world and synthetic datasets.