In order to address the issues of insufficient acoustic feature extraction and severe decoding feature loss in single-channel speech enhancement networks based on convolutional encoder-decoder architecture, a single-channel speech enhancement network called Multi-Channel Information Aggregation and Collaborative Decoding (MIACD) was proposed. A dual-channel encoder was utilized to extract the speech magnitude spectrum and complex spectrum features, which were enriched with Self-Supervised Learning (SSL) representations. A four-layer Conformer block was employed to model the extracted features in time and frequency domains. By incorporating residual connections, the speech magnitude and complex features extracted by the dual-channel encoder were introduced into a three-channel information aggregation decoder. Additionally, a Channel-Time-Frequency Attention (CTF-Attention) mechanism was proposed to adjust the aggregated information in the decoder based on the distribution of speech energy, effectively alleviating the problem of severe acoustic information loss during decoding. Experimental results on the publicly available dataset Voice Bank DEMAND demonstrate that, compared to Glance and Gaze: a collaborative learning framework for Single-channel speech enhancement (GaGNet), the proposed method achieves a 5.1% improvement on the objective metric WB-PESQ (Wide Band Perceptual Evaluation of Speech Quality) and 96.7% on STOI (Short-Time Objective Intelligibility), validating that the proposed method effectively utilizes speech information for signal reconstruction, noise suppression, and speech intelligibility enhancement.