Spatially-coupled low-density parity-check(SC-LDPC)codes are prominent candidates for fu-ture communication standards due to their‘threshold saturation’properties.However,when facing burst erasures,the decoding proc...Spatially-coupled low-density parity-check(SC-LDPC)codes are prominent candidates for fu-ture communication standards due to their‘threshold saturation’properties.However,when facing burst erasures,the decoding process will stop and the decoding performances will dramatically de-grade.To improve the ability of burst erasure corrections,this paper proposes a two-dimensional SC-LDPC(2D-SC-LDPC)codes constructed by parallelly connecting two asymmetric SC-LDPC coupled chains for resistance to burst erasures.Density evolution algorithm is presented to evaluate the as-ymptotic performances against burst erasures,by which the maximum correctable burst erasure length can be computed.The analysis results show that the maximum correctable burst erasure lengths of the proposed 2D-SC-LDPC codes are much larger than the SC-LDPC codes and the asym-metric SC-LDPC codes.Finite-length performance simulation results of the 2D-SC-LDPC codes over the burst erasure channel confirm the excellent asymptotic performances.展开更多
Network coding (NC), introduced at the turn of the century, enables nodes in a network to combine data algebraically before either sending or forwarding them. Random network coding has gained popularity over the years...Network coding (NC), introduced at the turn of the century, enables nodes in a network to combine data algebraically before either sending or forwarding them. Random network coding has gained popularity over the years by combining the received packet randomly before forwarding them, resulting in a complex Jordan Gaussian Elimination (JGE) decoding process. The effectiveness of random NC is through cooperation among nodes. In this paper, we propose a simple, low-complexity cooperative protocol that exploits NC in a deterministic manner resulting in improved diversity, data rate, and less complex JGE decoding process. The proposed system is applied over a lossy wireless network. The scenario under investigation is as follows: M users must send their information to a common destination D and to exchange the information between each others, over erasure channels;typically the channels between the users and the destination are worse than the channels between users. It is possible to significantly reduce the traffic among users and destination, achieving significant bandwidth savings, by combining packets from different users in simple, deterministic ways without resorting to extensive header information before being forwarded to the destination and the M users. The key problem we try to address is how to efficiently combine the packets at each user while exploiting user cooperation and the probability of successfully recovering information from all users at D with k < 2M unique linear equations, accounting for the fact that the remaining packets will be lost in the network and there are two transmission stages. Simulation results show the behaviour for two and three transmission stages. Our results show that applying NC protocols in two or three stages decreases the traffic significantly, beside the fact that the proposed protocols enable the system to retrieve the lost packets rather than asking for ARQ, resulting in improved data flow, and less power consumption. In fact, in some protocols the ARQ dropped from the rate 10ˉ<sup>1</sup> to 10ˉ<sup>4</sup>, because of the proposed combining algorithm that enables the nodes to generate additional unique linear equations to broadcast rather than repeating the same ones via ARQ. Moreover, the number of the transmitted packets in each cooperative stage dropped from M (M - 1) to just M packets, resulting to 2 M packets instead 2 (M<sup>2</sup> - 1) when three stages of transmission system are used instead of one stage (two cooperative stages).展开更多
As a new class of forward error correcting encoding algorithm,Luby Transform codes are suitable for the erasure channel environment based on the packet communication.The encoding,decoding algorithms and the implementa...As a new class of forward error correcting encoding algorithm,Luby Transform codes are suitable for the erasure channel environment based on the packet communication.The encoding,decoding algorithms and the implementation of LT codes are summarized in the paper.Meanwhile simulations of the ideal soliton distribution and robust soliton distribution are conducted to evaluate the performance of LT codes in terms of successful decoding probability,mean degree and decoding time over the erasure channel.The parameter optimization rules of LT codes are deeply discussed and proposed in the paper.The research results are of great practical importance for improving the real time performance in the erasure correction applications.展开更多
基金Supported by the National Natural Science Foundation of China(No.U19B2015,62271386,61801371).
文摘Spatially-coupled low-density parity-check(SC-LDPC)codes are prominent candidates for fu-ture communication standards due to their‘threshold saturation’properties.However,when facing burst erasures,the decoding process will stop and the decoding performances will dramatically de-grade.To improve the ability of burst erasure corrections,this paper proposes a two-dimensional SC-LDPC(2D-SC-LDPC)codes constructed by parallelly connecting two asymmetric SC-LDPC coupled chains for resistance to burst erasures.Density evolution algorithm is presented to evaluate the as-ymptotic performances against burst erasures,by which the maximum correctable burst erasure length can be computed.The analysis results show that the maximum correctable burst erasure lengths of the proposed 2D-SC-LDPC codes are much larger than the SC-LDPC codes and the asym-metric SC-LDPC codes.Finite-length performance simulation results of the 2D-SC-LDPC codes over the burst erasure channel confirm the excellent asymptotic performances.
文摘Network coding (NC), introduced at the turn of the century, enables nodes in a network to combine data algebraically before either sending or forwarding them. Random network coding has gained popularity over the years by combining the received packet randomly before forwarding them, resulting in a complex Jordan Gaussian Elimination (JGE) decoding process. The effectiveness of random NC is through cooperation among nodes. In this paper, we propose a simple, low-complexity cooperative protocol that exploits NC in a deterministic manner resulting in improved diversity, data rate, and less complex JGE decoding process. The proposed system is applied over a lossy wireless network. The scenario under investigation is as follows: M users must send their information to a common destination D and to exchange the information between each others, over erasure channels;typically the channels between the users and the destination are worse than the channels between users. It is possible to significantly reduce the traffic among users and destination, achieving significant bandwidth savings, by combining packets from different users in simple, deterministic ways without resorting to extensive header information before being forwarded to the destination and the M users. The key problem we try to address is how to efficiently combine the packets at each user while exploiting user cooperation and the probability of successfully recovering information from all users at D with k < 2M unique linear equations, accounting for the fact that the remaining packets will be lost in the network and there are two transmission stages. Simulation results show the behaviour for two and three transmission stages. Our results show that applying NC protocols in two or three stages decreases the traffic significantly, beside the fact that the proposed protocols enable the system to retrieve the lost packets rather than asking for ARQ, resulting in improved data flow, and less power consumption. In fact, in some protocols the ARQ dropped from the rate 10ˉ<sup>1</sup> to 10ˉ<sup>4</sup>, because of the proposed combining algorithm that enables the nodes to generate additional unique linear equations to broadcast rather than repeating the same ones via ARQ. Moreover, the number of the transmitted packets in each cooperative stage dropped from M (M - 1) to just M packets, resulting to 2 M packets instead 2 (M<sup>2</sup> - 1) when three stages of transmission system are used instead of one stage (two cooperative stages).
基金supported by Zhongguancun Haidian Science Park Postdoctoral Special Fund
文摘As a new class of forward error correcting encoding algorithm,Luby Transform codes are suitable for the erasure channel environment based on the packet communication.The encoding,decoding algorithms and the implementation of LT codes are summarized in the paper.Meanwhile simulations of the ideal soliton distribution and robust soliton distribution are conducted to evaluate the performance of LT codes in terms of successful decoding probability,mean degree and decoding time over the erasure channel.The parameter optimization rules of LT codes are deeply discussed and proposed in the paper.The research results are of great practical importance for improving the real time performance in the erasure correction applications.