Spatially-coupled low-density parity-check(SC-LDPC)codes are prominent candidates for fu-ture communication standards due to their‘threshold saturation’properties.However,when facing burst erasures,the decoding proc...Spatially-coupled low-density parity-check(SC-LDPC)codes are prominent candidates for fu-ture communication standards due to their‘threshold saturation’properties.However,when facing burst erasures,the decoding process will stop and the decoding performances will dramatically de-grade.To improve the ability of burst erasure corrections,this paper proposes a two-dimensional SC-LDPC(2D-SC-LDPC)codes constructed by parallelly connecting two asymmetric SC-LDPC coupled chains for resistance to burst erasures.Density evolution algorithm is presented to evaluate the as-ymptotic performances against burst erasures,by which the maximum correctable burst erasure length can be computed.The analysis results show that the maximum correctable burst erasure lengths of the proposed 2D-SC-LDPC codes are much larger than the SC-LDPC codes and the asym-metric SC-LDPC codes.Finite-length performance simulation results of the 2D-SC-LDPC codes over the burst erasure channel confirm the excellent asymptotic performances.展开更多
A modified type of Hybrid ARQ system with erasures correction and parity bits retransmission is considered. Performance of the system is analyzed under assumption that the forward channel suffers from Nakagami common ...A modified type of Hybrid ARQ system with erasures correction and parity bits retransmission is considered. Performance of the system is analyzed under assumption that the forward channel suffers from Nakagami common fading and additive white Gaussian noise. A good agreement between theoretical results and simulation is achieved. The proposed ARQ protocol is compared with other known Hybrid ARQ algorithms. It demonstrates significantly higher throughput efficiency in a range of SNR.展开更多
Multi-channel can be used to provide higher transmission ability to the bandwidth-intensive and delay-sensitive real-time streams. However, traditional channel capacity theories and coding schemes are seldom designed ...Multi-channel can be used to provide higher transmission ability to the bandwidth-intensive and delay-sensitive real-time streams. However, traditional channel capacity theories and coding schemes are seldom designed for the real-time streams with strict delay constraint, especially in multi-channel context. This paper considers a real-time stream system, where real-time messages with different importance should be transmitted through several packet erasure channels, and be decoded by the receiver within a fixed delay. Based on window erasure channels and i.i.d.(identically and independently distributed) erasure channels, we derive the Multi-channel Real-time Stream Transmission(MRST) capacity models for Symmetric Real-time(SR) streams and Asymmetric Real-time(AR) streams respectively. Moreover, for window erasures, a Maximum Equilibrium Intra-session Code(MEIC) is presented for SR and AR streams, and is shown able to asymptotically achieve the theoretical MRST capacity. For i.i.d. erasures, we propose an Adaptive Maximum Equilibrium Intra-session Code(AMEIC), and then prove AMEIC can closely approach the MRST transmission capacity. Finally, the performances of the proposed codes are verified by simulations.展开更多
In an earlier work, we proposed a frame-based kernel analysis approach to the problem of recovering erasures from unknown locations. The new approach led to the stability question on recovering a signal from noisy par...In an earlier work, we proposed a frame-based kernel analysis approach to the problem of recovering erasures from unknown locations. The new approach led to the stability question on recovering a signal from noisy partial frame coefficients with erasures occurring at unknown locations. In this continuing work, we settle this problem by obtaining a complete characterization of frames that provide stable reconstructions. We show that an encoding frame provides a stable signal recovery from noisy partial frame coefficients at unknown locations if and only if it is totally robust with respect to erasures. We present several characterizations for either totally robust frames or almost robust frames. Based on these characterizations several explicit construction algorithms for totally robust and almost robust frames are proposed. As a consequence of the construction methods, we obtain that the probability for a randomly generated frame to be totally robust with respect to a fixed number of erasures is one.展开更多
The sensitive data stored in the public cloud by privileged users,such as corporate companies and government agencies are highly vulnerable in the hands of cloud providers and hackers.The proposed Virtual Cloud Storag...The sensitive data stored in the public cloud by privileged users,such as corporate companies and government agencies are highly vulnerable in the hands of cloud providers and hackers.The proposed Virtual Cloud Storage Archi-tecture is primarily concerned with data integrity and confidentiality,as well as availability.To provide confidentiality and availability,thefile to be stored in cloud storage should be encrypted using an auto-generated key and then encoded into distinct chunks.Hashing the encoded chunks ensured thefile integrity,and a newly proposed Circular Shift Chunk Allocation technique was used to determine the order of chunk storage.Thefile could be retrieved by performing the opera-tions in reverse.Using the regenerating code,the model could regenerate the missing and corrupted chunks from the cloud.The proposed architecture adds an extra layer of security while maintaining a reasonable response time and sto-rage capacity.Experimental results analysis show that the proposed model has been tested with storage space and response time for storage and retrieval.The VCSA model consumes 1.5x(150%)storage space.It was found that total storage required for the VCSA model is very low when compared with 2x Replication and completely satisfies the CIA model.The response time VCSA model was tested with different sizedfiles starting from 2 to 16 MB.The response time for storing and retrieving a 2 MBfile is 4.96 and 3.77 s respectively,and for a 16 MBfile,the response times are 11.06 s for storage and 5.6 s for retrieval.展开更多
This paper tries to analyze Philip Larkin's poem"High Window"from a deconstructive perspective. It is to show that in the poem the key words/signifiers are always under erasure, and thus the chain of sig...This paper tries to analyze Philip Larkin's poem"High Window"from a deconstructive perspective. It is to show that in the poem the key words/signifiers are always under erasure, and thus the chain of signification is endless since the poem is self-deconstructing. Then, the paper argues that the linguistic features of the poem paradoxically meaningful in the sense that it reflects the poet's skepticism and anxiety.展开更多
Frame erasure concealment is studied to solve the problem of rapid speech quality reduction due to the loss of speech parameters during speech transmission. A large hidden Markov model is applied to model the immittan...Frame erasure concealment is studied to solve the problem of rapid speech quality reduction due to the loss of speech parameters during speech transmission. A large hidden Markov model is applied to model the immittance spectral frequency (ISF) parameters in AMR-WB codec to optimally estimate the lost ISFs based on the minimum mean square error (MMSE) rule. The estimated ISFs are weighted with the ones of their previous neighbors to smooth the speech, resulting in the actual concealed ISF vectors. They are used instead of the lost ISFs in the speech synthesis on the receiver. Comparison is made between the speech concealed by this algorithm and by Annex I of G. 722. 2 specification, and simulation shows that the proposed concealment algorithm can lead to better performance in terms of frequency-weighted spectral distortion and signal-to-noise ratio compared to the baseline method, with an increase of 2.41 dB in signal-to-noise ratio (SNR) and a reduction of 0. 885 dB in frequency-weighted spectral distortion.展开更多
To reduce the time required to complete the regeneration process of erasure codes, we propose a Tree-structured Parallel Regeneration (TPR) scheme for multiple data losses in distributed storage systems. Under the sch...To reduce the time required to complete the regeneration process of erasure codes, we propose a Tree-structured Parallel Regeneration (TPR) scheme for multiple data losses in distributed storage systems. Under the scheme, two algorithms are proposed for the construction of multiple regeneration trees, namely the edge-disjoint algorithm and edge-sharing algorithm. The edge-disjoint algorithm constructs multiple independent trees, and is simple and appropriate for environments where newcomers and their providers are distributed over a large area and have few intersections. The edge-sharing algorithm constructs multiple trees that compete to utilize the bandwidth, and make a better utilization of the bandwidth, although it needs to measure the available band-width and deal with the bandwidth changes; it is therefore difficult to implement in practical systems. The parallel regeneration for multiple data losses of TPR primarily includes two optimizations: firstly, transferring the data through the bandwidth optimized-paths in a pipe-line manner; secondly, executing data regeneration over multiple trees in parallel. To evaluate the proposal, we implement an event-based simulator and make a detailed comparison with some popular regeneration methods. The quantitative comparison results show that the use of TPR employing either the edge-disjoint algorithm or edge-sharing algorithm reduces the regeneration time significantly.展开更多
Network coding (NC), introduced at the turn of the century, enables nodes in a network to combine data algebraically before either sending or forwarding them. Random network coding has gained popularity over the years...Network coding (NC), introduced at the turn of the century, enables nodes in a network to combine data algebraically before either sending or forwarding them. Random network coding has gained popularity over the years by combining the received packet randomly before forwarding them, resulting in a complex Jordan Gaussian Elimination (JGE) decoding process. The effectiveness of random NC is through cooperation among nodes. In this paper, we propose a simple, low-complexity cooperative protocol that exploits NC in a deterministic manner resulting in improved diversity, data rate, and less complex JGE decoding process. The proposed system is applied over a lossy wireless network. The scenario under investigation is as follows: M users must send their information to a common destination D and to exchange the information between each others, over erasure channels;typically the channels between the users and the destination are worse than the channels between users. It is possible to significantly reduce the traffic among users and destination, achieving significant bandwidth savings, by combining packets from different users in simple, deterministic ways without resorting to extensive header information before being forwarded to the destination and the M users. The key problem we try to address is how to efficiently combine the packets at each user while exploiting user cooperation and the probability of successfully recovering information from all users at D with k < 2M unique linear equations, accounting for the fact that the remaining packets will be lost in the network and there are two transmission stages. Simulation results show the behaviour for two and three transmission stages. Our results show that applying NC protocols in two or three stages decreases the traffic significantly, beside the fact that the proposed protocols enable the system to retrieve the lost packets rather than asking for ARQ, resulting in improved data flow, and less power consumption. In fact, in some protocols the ARQ dropped from the rate 10ˉ<sup>1</sup> to 10ˉ<sup>4</sup>, because of the proposed combining algorithm that enables the nodes to generate additional unique linear equations to broadcast rather than repeating the same ones via ARQ. Moreover, the number of the transmitted packets in each cooperative stage dropped from M (M - 1) to just M packets, resulting to 2 M packets instead 2 (M<sup>2</sup> - 1) when three stages of transmission system are used instead of one stage (two cooperative stages).展开更多
Fault-tolerance is increasingly significant for large-scale storage systems in which Byzantine failure of storage nodes may happen. Traditional Byzantine Quorum systems that tolerate Byzantine failures by using replic...Fault-tolerance is increasingly significant for large-scale storage systems in which Byzantine failure of storage nodes may happen. Traditional Byzantine Quorum systems that tolerate Byzantine failures by using replication have two main limitations: low space-efficiency and static quorum variables. We propose an Erasure-code Byzantine Fault-tolerance Quorum that can provide high reliability with far lower storage overhead than replication by adopting erasure code as redundancy scheme. Through read/write operations of clients and diagnose operation of supervisor, our Quorum system can detect Byzantine nodes, and dynamically adjust system size and fault threshold. Simulation results show that our method improves performance for the Quorum with relatively small quorums.展开更多
As a new class of forward error correcting encoding algorithm,Luby Transform codes are suitable for the erasure channel environment based on the packet communication.The encoding,decoding algorithms and the implementa...As a new class of forward error correcting encoding algorithm,Luby Transform codes are suitable for the erasure channel environment based on the packet communication.The encoding,decoding algorithms and the implementation of LT codes are summarized in the paper.Meanwhile simulations of the ideal soliton distribution and robust soliton distribution are conducted to evaluate the performance of LT codes in terms of successful decoding probability,mean degree and decoding time over the erasure channel.The parameter optimization rules of LT codes are deeply discussed and proposed in the paper.The research results are of great practical importance for improving the real time performance in the erasure correction applications.展开更多
基金Supported by the National Natural Science Foundation of China(No.U19B2015,62271386,61801371).
文摘Spatially-coupled low-density parity-check(SC-LDPC)codes are prominent candidates for fu-ture communication standards due to their‘threshold saturation’properties.However,when facing burst erasures,the decoding process will stop and the decoding performances will dramatically de-grade.To improve the ability of burst erasure corrections,this paper proposes a two-dimensional SC-LDPC(2D-SC-LDPC)codes constructed by parallelly connecting two asymmetric SC-LDPC coupled chains for resistance to burst erasures.Density evolution algorithm is presented to evaluate the as-ymptotic performances against burst erasures,by which the maximum correctable burst erasure length can be computed.The analysis results show that the maximum correctable burst erasure lengths of the proposed 2D-SC-LDPC codes are much larger than the SC-LDPC codes and the asym-metric SC-LDPC codes.Finite-length performance simulation results of the 2D-SC-LDPC codes over the burst erasure channel confirm the excellent asymptotic performances.
文摘A modified type of Hybrid ARQ system with erasures correction and parity bits retransmission is considered. Performance of the system is analyzed under assumption that the forward channel suffers from Nakagami common fading and additive white Gaussian noise. A good agreement between theoretical results and simulation is achieved. The proposed ARQ protocol is compared with other known Hybrid ARQ algorithms. It demonstrates significantly higher throughput efficiency in a range of SNR.
基金supported by National Key Technology Research and Development Program of China under Grant No.2015BAH08F01the joint fund of the Ministry of Education of People's Republic of China and China Mobile Communications Corporation under Grant No.MCM20160304
文摘Multi-channel can be used to provide higher transmission ability to the bandwidth-intensive and delay-sensitive real-time streams. However, traditional channel capacity theories and coding schemes are seldom designed for the real-time streams with strict delay constraint, especially in multi-channel context. This paper considers a real-time stream system, where real-time messages with different importance should be transmitted through several packet erasure channels, and be decoded by the receiver within a fixed delay. Based on window erasure channels and i.i.d.(identically and independently distributed) erasure channels, we derive the Multi-channel Real-time Stream Transmission(MRST) capacity models for Symmetric Real-time(SR) streams and Asymmetric Real-time(AR) streams respectively. Moreover, for window erasures, a Maximum Equilibrium Intra-session Code(MEIC) is presented for SR and AR streams, and is shown able to asymptotically achieve the theoretical MRST capacity. For i.i.d. erasures, we propose an Adaptive Maximum Equilibrium Intra-session Code(AMEIC), and then prove AMEIC can closely approach the MRST transmission capacity. Finally, the performances of the proposed codes are verified by simulations.
基金supported by National Science Foundation of USA(Grant Nos.DMS-1403400 and DMS-1712602)National Natural Science Foundation of China(Grant Nos.11171151,11371200,11525104 and 11531013)
文摘In an earlier work, we proposed a frame-based kernel analysis approach to the problem of recovering erasures from unknown locations. The new approach led to the stability question on recovering a signal from noisy partial frame coefficients with erasures occurring at unknown locations. In this continuing work, we settle this problem by obtaining a complete characterization of frames that provide stable reconstructions. We show that an encoding frame provides a stable signal recovery from noisy partial frame coefficients at unknown locations if and only if it is totally robust with respect to erasures. We present several characterizations for either totally robust frames or almost robust frames. Based on these characterizations several explicit construction algorithms for totally robust and almost robust frames are proposed. As a consequence of the construction methods, we obtain that the probability for a randomly generated frame to be totally robust with respect to a fixed number of erasures is one.
文摘The sensitive data stored in the public cloud by privileged users,such as corporate companies and government agencies are highly vulnerable in the hands of cloud providers and hackers.The proposed Virtual Cloud Storage Archi-tecture is primarily concerned with data integrity and confidentiality,as well as availability.To provide confidentiality and availability,thefile to be stored in cloud storage should be encrypted using an auto-generated key and then encoded into distinct chunks.Hashing the encoded chunks ensured thefile integrity,and a newly proposed Circular Shift Chunk Allocation technique was used to determine the order of chunk storage.Thefile could be retrieved by performing the opera-tions in reverse.Using the regenerating code,the model could regenerate the missing and corrupted chunks from the cloud.The proposed architecture adds an extra layer of security while maintaining a reasonable response time and sto-rage capacity.Experimental results analysis show that the proposed model has been tested with storage space and response time for storage and retrieval.The VCSA model consumes 1.5x(150%)storage space.It was found that total storage required for the VCSA model is very low when compared with 2x Replication and completely satisfies the CIA model.The response time VCSA model was tested with different sizedfiles starting from 2 to 16 MB.The response time for storing and retrieving a 2 MBfile is 4.96 and 3.77 s respectively,and for a 16 MBfile,the response times are 11.06 s for storage and 5.6 s for retrieval.
文摘This paper tries to analyze Philip Larkin's poem"High Window"from a deconstructive perspective. It is to show that in the poem the key words/signifiers are always under erasure, and thus the chain of signification is endless since the poem is self-deconstructing. Then, the paper argues that the linguistic features of the poem paradoxically meaningful in the sense that it reflects the poet's skepticism and anxiety.
基金The Science Foundation of Southeast University(No.XJ0704268)the Natural Science Foundation of the Education Department of Anhui Province(No.KJ2007B088)
文摘Frame erasure concealment is studied to solve the problem of rapid speech quality reduction due to the loss of speech parameters during speech transmission. A large hidden Markov model is applied to model the immittance spectral frequency (ISF) parameters in AMR-WB codec to optimally estimate the lost ISFs based on the minimum mean square error (MMSE) rule. The estimated ISFs are weighted with the ones of their previous neighbors to smooth the speech, resulting in the actual concealed ISF vectors. They are used instead of the lost ISFs in the speech synthesis on the receiver. Comparison is made between the speech concealed by this algorithm and by Annex I of G. 722. 2 specification, and simulation shows that the proposed concealment algorithm can lead to better performance in terms of frequency-weighted spectral distortion and signal-to-noise ratio compared to the baseline method, with an increase of 2.41 dB in signal-to-noise ratio (SNR) and a reduction of 0. 885 dB in frequency-weighted spectral distortion.
基金supported by the National Grand Fundamental Research of China (973 Program) under Grant No. 2011CB302601the National High Technology Research and Development of China (863 Program) under GrantNo. 2013AA01A213+2 种基金the National Natural Science Foundation of China under Grant No. 60873215the Natural Science Foundation for Distinguished Young Scholars of Hunan Province under Grant No. S2010J5050Specialized Research Fund for the Doctoral Program of Higher Education under Grant No. 20124307110015
文摘To reduce the time required to complete the regeneration process of erasure codes, we propose a Tree-structured Parallel Regeneration (TPR) scheme for multiple data losses in distributed storage systems. Under the scheme, two algorithms are proposed for the construction of multiple regeneration trees, namely the edge-disjoint algorithm and edge-sharing algorithm. The edge-disjoint algorithm constructs multiple independent trees, and is simple and appropriate for environments where newcomers and their providers are distributed over a large area and have few intersections. The edge-sharing algorithm constructs multiple trees that compete to utilize the bandwidth, and make a better utilization of the bandwidth, although it needs to measure the available band-width and deal with the bandwidth changes; it is therefore difficult to implement in practical systems. The parallel regeneration for multiple data losses of TPR primarily includes two optimizations: firstly, transferring the data through the bandwidth optimized-paths in a pipe-line manner; secondly, executing data regeneration over multiple trees in parallel. To evaluate the proposal, we implement an event-based simulator and make a detailed comparison with some popular regeneration methods. The quantitative comparison results show that the use of TPR employing either the edge-disjoint algorithm or edge-sharing algorithm reduces the regeneration time significantly.
文摘Network coding (NC), introduced at the turn of the century, enables nodes in a network to combine data algebraically before either sending or forwarding them. Random network coding has gained popularity over the years by combining the received packet randomly before forwarding them, resulting in a complex Jordan Gaussian Elimination (JGE) decoding process. The effectiveness of random NC is through cooperation among nodes. In this paper, we propose a simple, low-complexity cooperative protocol that exploits NC in a deterministic manner resulting in improved diversity, data rate, and less complex JGE decoding process. The proposed system is applied over a lossy wireless network. The scenario under investigation is as follows: M users must send their information to a common destination D and to exchange the information between each others, over erasure channels;typically the channels between the users and the destination are worse than the channels between users. It is possible to significantly reduce the traffic among users and destination, achieving significant bandwidth savings, by combining packets from different users in simple, deterministic ways without resorting to extensive header information before being forwarded to the destination and the M users. The key problem we try to address is how to efficiently combine the packets at each user while exploiting user cooperation and the probability of successfully recovering information from all users at D with k < 2M unique linear equations, accounting for the fact that the remaining packets will be lost in the network and there are two transmission stages. Simulation results show the behaviour for two and three transmission stages. Our results show that applying NC protocols in two or three stages decreases the traffic significantly, beside the fact that the proposed protocols enable the system to retrieve the lost packets rather than asking for ARQ, resulting in improved data flow, and less power consumption. In fact, in some protocols the ARQ dropped from the rate 10ˉ<sup>1</sup> to 10ˉ<sup>4</sup>, because of the proposed combining algorithm that enables the nodes to generate additional unique linear equations to broadcast rather than repeating the same ones via ARQ. Moreover, the number of the transmitted packets in each cooperative stage dropped from M (M - 1) to just M packets, resulting to 2 M packets instead 2 (M<sup>2</sup> - 1) when three stages of transmission system are used instead of one stage (two cooperative stages).
基金Supported by the National Natural Science Foun-dation of China (60373088)
文摘Fault-tolerance is increasingly significant for large-scale storage systems in which Byzantine failure of storage nodes may happen. Traditional Byzantine Quorum systems that tolerate Byzantine failures by using replication have two main limitations: low space-efficiency and static quorum variables. We propose an Erasure-code Byzantine Fault-tolerance Quorum that can provide high reliability with far lower storage overhead than replication by adopting erasure code as redundancy scheme. Through read/write operations of clients and diagnose operation of supervisor, our Quorum system can detect Byzantine nodes, and dynamically adjust system size and fault threshold. Simulation results show that our method improves performance for the Quorum with relatively small quorums.
基金supported by Zhongguancun Haidian Science Park Postdoctoral Special Fund
文摘As a new class of forward error correcting encoding algorithm,Luby Transform codes are suitable for the erasure channel environment based on the packet communication.The encoding,decoding algorithms and the implementation of LT codes are summarized in the paper.Meanwhile simulations of the ideal soliton distribution and robust soliton distribution are conducted to evaluate the performance of LT codes in terms of successful decoding probability,mean degree and decoding time over the erasure channel.The parameter optimization rules of LT codes are deeply discussed and proposed in the paper.The research results are of great practical importance for improving the real time performance in the erasure correction applications.