International Journal of Communications, Network and System Sciences
Vol.08 No.05(2015), Article ID:55695,21 pages
10.4236/ijcns.2015.85013
Compression of ECG Signal Based on Compressive Sensing and the Extraction of Significant Features
Mohammed M. Abo-Zahhad1, Aziza I. Hussein2, Abdelfatah M. Mohamed1
1Department of Electrical and Electronics Engineering, Faculty of Engineering, Assiut University, Assiut, Egypt
2Department of Computer and Systems Engineering, Faculty of Engineering, Minia University, Minia, Egypt
Email: eng_mmz_egy@yahoo.com, afmm52@yahoo.com, aziza_hu@yahoo.com
Copyright © 2015 by authors and Scientific Research Publishing Inc.
This work is licensed under the Creative Commons Attribution International License (CC BY).
http://creativecommons.org/licenses/by/4.0/



Received 19 February 2015; accepted 14 April 2015; published 15 April 2015
ABSTRACT
Diagnoses of heart diseases can be done effectively on long term recordings of ECG signals that preserve the signals’ morphologies. In these cases, the volume of the ECG data produced by the monitoring systems grows significantly. To make the mobile healthcare possible, the need for efficient ECG signal compression algorithms to store and/or transmit the signal efficiently has been rising exponentially. Currently, ECG signal is acquired at Nyquist rate or higher, thus introducing redundancies between adjacent heartbeats due to its quasi-periodic structure. Existing compression methods remove these redundancies by achieving compression and facilitate transmission of the patient’s imperative information. Based on the fact that these signals can be approximated by a linear combination of a few coefficients taken from different basis, an alternative new compression scheme based on Compressive Sensing (CS) has been proposed. CS provides a new approach concerned with signal compression and recovery by exploiting the fact that ECG signal can be reconstructed by acquiring a relatively small number of samples in the “sparse” domains through well-developed optimization procedures. In this paper, a single-lead ECG compression method has been proposed based on improving the signal sparisty through the extraction of the signal significant features. The proposed method starts with a preprocessing stage that detects the peaks and periods of the Q, R and S waves of each beat. Then, the QRS-complex for each signal beat is estimated. The estimated QRS -complexes are subtracted from the original ECG signal and the resulting error signal is compressed using the CS technique. Throughout this process, DWT sparsifying dictionaries have been adopted. The performance of the proposed algorithm, in terms of the reconstructed signal quality and compression ratio, is evaluated by adopting DWT spatial domain basis applied to ECG records extracted from the MIT-BIH Arrhythmia Database. The results indicate that average compression ratio of 11:1 with PRD1 = 1.2% are obtained. Moreover, the quality of the retrieved signal is guaranteed and the compression ratio achieved is an improvement over those obtained by previously reported algorithms. Simulation results suggest that CS should be considered as an acceptable methodology for ECG compression.
Keywords:
Compressed Sensing, ECG Signal Compression, Sparsity, Coherence, Spatial Domain

1. Introduction
Heart disease is the leading cause of mortality in the world. The ageing population makes heart diseases and other cardiovascular diseases (CVD) an increasing heavy burden on the healthcare systems of developing countries. The electrocardiogram is widely used for the diagnoses of these diseases because it is a noninvasive way to establish clinical diagnosis of heart diseases. It reveals a lot of important clinical information about the heart, and it is considered as the gold standard for the diagnosis of cardiac arrhythmias. Long-term records have become commonly used to detect information from the heart signals; thus the volume of the ECG data produced by monitoring systems can be quite large over a long period of time. In these cases, the quantity of data grows significantly and compression is required for reducing the storage space and transmission times. Thus, ECG data compression is often needed for efficient storage and transmission for telemedicine applications. Recently, to make the mobile healthcare possible, the need for an efficient ECG signal compression algorithms has been raising exponentially [1] [2] .
In technical literature, many compression algorithms have shown some success in ECG compression; however, algorithms that produce better compression ratios and less loss of data in the reconstructed data are needed. These algorithms can be classified into two major groups: the lossless and the lossy algorithms. The traditional approach of compressing and reconstructing signals or images from measured data follows the well-known Shan- non sampling theorem, which states that the sampling rate must be twice the highest frequency. Similarly, the fundamental theorem of linear algebra suggests that the number of collected samples (measurements) of a discrete finite-dimensional signal should be at least as large as its length (its dimension) in order to ensure reconstruction. In recent years, CS theory [3] has generated significant interest in the signal processing community because of its potential to enable signal reconstruction from significantly fewer data samples than suggested by conventional sampling theory. The novel theory of compressive sensing-also known compressed sensing, compressive sampling or sparse recovery-provides a fundamentally new approach to data acquisition and compression simultaneously [4] . Compared to conventional ECG compression algorithms, CS has some important advantages: 1) It transfers the computational burden from the encoder to the decoder, and thus offers simpler hardware implementations for the encoder; 2) the location of the largest coefficients in the wavelet domain does not need to be encoded.
Compressed sensing is a pioneering paradigm that enables to reconstruct sparse or compressible signals from a small number of linear projections. CS research currently advances in three major fronts: 1) the design of CS measurement matrices, 2) the development of new and efficient reconstruction techniques, and finally 3) the application of CS theory to novel problems and hardware implementations. The first two topics have already achieved a certain level of maturity, and many advanced methods have been developed. Currently, very high efficiency CS measurement systems have been developed with different characteristics (deterministic/non-de- terministic, adaptive/non-adaptive) that can be adopted in a variety of signal acquisition applications. On the other hand, reconstruction methods span a wide range of techniques that include Matching Pursuit/Greedy, Basis Pursuit/Linear Programming, Bayesian, Iterative Thresholding, among others [5] . Which method is selected de- pends on the application of interest performance and speed needs.
The application of CS in ECG compression is still at its early stages, it has already led to important results [6] . For example, in [7] the ability of CS to continually and blindly compress ECG signals at compression factors of 10× has been demonstrated. In [8] several design considerations for CS-based ECG telemonitoring, including the encoder architecture and the design of the measurement matrix has been studied by Dixon et al. Their results show high compression ratios using a 1-bit Bernoulli measurement matrix. In [9] [10] new contributions to the area include CS-based algorithms for ECG compression with focus on algorithms enabling joint reconstruction of ECG cycles by exploiting correlation between adjacent heartbeats. In addition, a CS-based method to reconstruct ECG signals in the presence of EMG noise using symmetric α-stable distributions to model the EMG interference [11] was proposed as an extension to the work presented in [9] [10] . A hidden problem when trying to reconstruct ECG signals using CS-based methods is the inability to accurately recover the low-magnitude coefficients of the wavelet representation [12] . To alleviate this problem, the prior information about the magnitude decay of the wavelet coefficients across subbands in the reconstruction algorithm was incorporated [13] . More precisely, a weighted l1-minimization algorithm with a weighting scheme based on the standard deviation of the wavelet coefficients at different scales was derived. In addition, the weighting scheme also takes into consideration the fact that the approximation subband coefficients accumulate most of the signal energy.
In this paper, a single-lead compression method has been proposed. It is based on improving the signal sparisty through the extraction of the significant ECG signal features. The proposed method starts with a preprocessing stage that detects the peaks and periods of the Q, R and S waves of each beat. Then, the QRS-complex for each signal beat is estimated. The estimated QRS -complexes are subtracted from the original ECG signal and the resulting error signal is compressed using CS technique. Throughout this process DWT sparsifying dictionaries have been adopted. The performance of the proposed algorithm in terms of the amount of compression and the reconstructed signal quality is evaluated using records extracted from the MIT-BIH Arrhythmia Database. Simulation results validate the superior performance of the proposed algorithm compared to other published algorithms. The rest of the paper is organized as follows. Section 2 introduces the compressed sensing Framework. Section 3 details the compressed sensing of ECG signal. Controlling the ECG signal sparisty using DWT basis is explained in Section 4. The solutions of CS problem including greedy algorithms, l1-minimization, and TV minimization are presented in Section 5. Section 6 introduces the methodology used for improving the ECG signal sparisty using QRS-complex estimation. Section 7 details the metrics adopted for measuring the performance of the proposed CS ECG signal compression algorithm. Sections 8 and 9 present the simulation results and the main conclusions respectively.
2. Compressed Sensing Problem
In a traditional ECG acquisition system, all samples of the original signal are acquired. Thus the number of signal samples can be in the order of millions. The acquisition process is followed by compression, which takes advantage of the redundancy (or the structure) in the signal to represent it in a domain where most of the signal coefficients can be discarded with little or no loss in quality. Hence, traditional acquisition systems first acquire a huge amount of data, a significant portion of which is immediately discarded. This creates important inefficiency in many practical applications. Compressive sensing addresses this inefficiency by effectively combining the acquisition and compression processes. Traditional decoding is replaced by recovery algorithms that exploit the underlying structure of the data [3] [4] [14] .
CS has become a very active research area in recent years due to its interesting theoretical nature and its practical utility in a wide range of applications; especially in wireless telemonitoring of ECG signals. Compared to traditional data compression technologies, it consumes much less energy thereby extending sensor lifetime, making it attractive to wireless body-area networks. In the following we provide a brief overview of the basic principles of CS, since they will form the basis of the proposed ECG compression algorithms. The basic CS framework is an underdetermined inverse problem, which can be expressed as
(1)
where, in the context of data compression,
is the
vector representing the ECG signal, 
is a user designed sensing matrix or measurement matrix,
is sensor noise, and
is the compressed signal. The matrix
here plays as a sensor that acquire information from the input signal
. The compressed signal,
, is sent to the receiver side where the original signal
is recovered by a CS algorithm using yand
. To successfully recover the ECG signal,
is required to be sparse. When
is not sparse, one generally seeks a sparsifying matrix or sparsifying dictionary
such that
can be sparsely represented with the dictionary matrix, i.e., 








In fact, the measurement process is not adaptive, meaning that 






a) Instead of point evaluations of the signal, the system takes M inner products of the signal with the basis vector
b) The number of measurements 

If















a highly sparse representation [3] [4] .
After the acquisition process, an estimate of the signal is obtained by a reconstruction algorithm. A common and practical approach used to determine the sparse solution is to solve this problem as a convex optimization problem. The original work on CS employed regularization based on 

where 





The 


3. Measurement Matrices
The measurement matrix 




That is, the matrix 




A related condition, referred to as incoherence, requires that the rows of 








with a lower bound given by

We say that a dictionary is incoherent if 
Direct construction of a measurement matrix 

each of the 


RIP and incoherence can be achieved with high probability simply by selecting 




The matrix 







The matrix 


4. Signal Recovery from Incomplete Measurements
CS theory also proposes that rather than acquire the entire signal and then compress, it should be possible to capture only the useful information to begin with. The challenge then is how to recover the signal from what

Figure 1. (a) Compressive sensing measurement process with a random Gaussian measurement matrix 



would traditionally seem to be an incomplete set of measurements. The ECG signal 









The question is now how we can actually recover 








It has been proven that computing the sparsest solution directly generally requires prohibitive computations of exponential complexity [22] , so several heuristic methods have been developed in literature, such as Matching Pursuit (MP) [23] , Basis Pursuit (BP) [24] , log barrier method [20] , iterative thresholding method [25] , and so forth. Most of these methods or algorithms fall into three distinct categories: greedy algorithms, 
4.1. Greedy Algorithms
Generally speaking, a greedy algorithm refers to any algorithm following the metaheuristic of choosing the best immediate or local optimum at each stage and expecting to find the global optimum at the end. It can find the global optimum for some optimization problems, but not for all [20] . This algorithm decomposes any signal into a linear combination of waveforms in a redundant dictionary of functions so that selected waveforms optimally match the structure of the signal. MP is easy to implement and has an exponential rate of convergence and good approximation properties. However, there is no theoretical guarantee that MP can achieve sparse representations. In [26] , the authors proposed a variant of MP, Orthogonal Matching Pursuit (OMP), which guarantees the nearly sparse solution under some conditions. A primary drawback of MP and its variants is the incapability of attaining truly sparse representations. The failure is usually caused by an inappropriate initial guess. This shortcoming also motivated the development of algorithms based on 
4.2. l1-Minimization Algorithms
In [3] some early results related to 






The solution to the above problem can be found with relative ease. There are methods that will find the solution to the BP problem but does it lead to a sparse solution? The answer in general is no but under the right conditions it can be guaranteed that BP will find a sparse solution or even the sparsest solution. This is because 















4.3. TV Minimization Algorithms
In the broad area of compressive sensing, 




5. Sparse Representation of ECG Signal
The time-domain representation of the ECG signal has low signal sparisty. Thus, ECG signal is not the true signal itself but its representation under a certain basis is sparse or compressible. Various researchers have reported ECG signals to be sparse in other bases [6] , [15] . A variety of compression algorithms represent ECG signals in suitable or thogonal basis and exploit signal redundancy in the transformed domain. Indeed, success of a compression algorithm depends on how compactly the signal is represented upon transformation. In this context, many transforming methods for representing signals in sparsity bases are proposed recently, for instance, FFT, DWT and DCT, etc. Generally, the biophysical signals are continuous and regular in nature. Hence, it can be represented by aforementioned transforms. To testify this, we exploit the DWT to guarantee the signal sparsity. Every transformation basis is able to provide a way of recovering the signal and should provide a way for re- trieving diagnostic information from the signal to form patient’s report of the medical case study. All the data we used are from MIT-BIH online distribution, a standard ECG database for arrhythmia diagnosis and research. This section introduces the wavelet transformation technique used for controlling the ECG signal sparisty.
The wavelet transform describes a multi-resolution decomposition process in terms of expansion of a signal onto a set of wavelet basis functions. Wavelet transforms have become an attractive and efficient tool in many applications especially in coding and compression of signals because of multi-resolution and high-energy compaction properties. Wavelets allow both time and frequency analysis of signals simultaneously because of the fact that energy of wavelet is concentrated in time and still possesses the wave like characteristics. As a result wavelet representation provides a versatile mathematical tool to analyze ECG signals.
Discrete Wavelet Transformation (DWT) has its own excellent space frequency localization property. The key issues in DWT and inverse DWT are signal decomposition and reconstruction, respectively. The basic idea behind decomposition and reconstruction is low-pass and high-pass filtering with the use of down sampling and up sampling respectively. The result of wavelet decomposition is hierarchically organized decompositions. One can choose the level of decomposition 

















so that the output of the inverse DWT is identical to the input of the forward DWT. In this environment, ECG signal representation using a wide variety of wavelets, drawn from various families including symlets and Daubechies’ bases has been adopted. In this context, the transform coefficients are arranged in decreasing order of magnitude, and count the number of coefficients accounting for 99% of the signal energy (as parser representation requires less number). “Symlet” and “Daubechies” families generally offer more compact representation com- pared to Meyer wavelet as well as biorthogonal and reverse biorthogonal families. In particular, the sparsest re-


Figure 2. A three-level two-channel iterative filter bank (a) forward DWT (b) inverse DWT.
presentation is provided by the “sym4” (closely followed by the “db4”) wavelet basis for abroad class of ECG signals [30] - [32] .
6. Improving ECG Signal Sparisty Using QRS-Complex Estimation
The correlation between the consecutive ECG beats can be exploited to improve the ECG signal sparisty. For this purpose, in this paper, the QRS-complex has been estimated based on the peaks and locations of Q, R and S waves. Then the estimated QRS-complex is subtracted from the original ECG signal and the resulting differential signal is manipulated using CS technique. The proposed compression scheme is presented in Figure 3. The details of the purposed compression algorithm are illustrated in subsequent steps as follows.
1) The signal is decomposed into windows; each of length 1024 samples. This short window length is considered in order to generate an approximate real time transmission. At the same time, many heartbeats in the window are incorporated to recover the signal with fewer samples.
2) The signal is preprocessed to determine the amplitudes and locations of the Q, R and S peaks. These parameters are used to estimate the QRS-complexes.
3) From the estimated QRS-complexes and the locations of the R-peaks locations, the error signal with more sparisity compared to the original ECG signal is determined as the difference between the original ECG signal and the estimated QRS-complexes.
4) Fewer measurements are determined from the resulting error signal and the sensing matrix.
5) The amplitudes and locations of the Q, R and S peaks and the measurement matrix are quantized.
6) The resulting quantized values are packetized for possible storage and/or transmission.
6.1. Preprocessing
In this section, the signal sparisty is controlled through the extraction of the significant ECG signal features. These features are extracted by estimating the QRS-complex for each signal beat. Then, the estimated QRS -complex is subtracted from the original ECG signal. After that, the resulting error signal is transformed into DWT domain and the resulting transformed coefficients are compressed using CS technique. A typical scalar ECG heartbeat is shown in Figure 4. The significant features of the ECG waveform are the P, Q, R, S and T waves and the duration of each wave. A typical ECG tracing of electrocardiogram baseline voltage is known as the isoelectric line. It is measured as the portion of the tracing following the T wave and preceding the next P wave. In addition to the QRS -complex, the ECG waveform contains P and T waves, 50-Hz noise from power line interference, EMG signal from muscles, motion artifact from the electrode and skin interface, and possibly other interference from
Figure 3. Block diagram of the proposed method.
Figure 4. Typical ECG signal.
electro surgery equipment. The power spectrum of the ECG signal can provide useful information about the QRS -complex estimation. Figure 5 summarizes the relative power spectra, based on the FFT, of the ECG , QRS - complex, P and T waves, motion artifact, and muscle noise taken for a set of 512 sample points that contain approximately two heartbeats [33] . It is visible that QRS -complex power spectrum involves the major part of the ECG heartbeat. Normal QRS -complex is 0.06 to 0.1 sec in duration and not every QRS -complex contains a Q wave, R wave, and S wave. By convention, any combination of these waves can be referred to as a QRS-com- plex. This portion can be represented by Q, R and S values, the Q-R and R-S durations and the event time of R as shown in Figure 4. These values can be extracted from the original signal.
6.2. The QRS-Complex Detection and Estimation
The aim of the QRS -complex estimation is to produce typical QRS -complex waveform using parameters extracted from the original ECG signal [34] . The estimation algorithm is a Matlab based estimator and is able to produce normal QRS waveform. A single heartbeat of ECG signal is a mixture of triangular and sinusoidal wave- forms. The QRS -complex waveform can be represented by shifted and scaled versions of these waveforms. Figure 6 illustrates a sinc function that looks like the QRS-complex. However, in QRS-complex the QR-dura- tion is not equal to the RS-duration. So, the QRS estimation is divided into two parts: QR part and RS part. The QR part can be generated using a sinc function described by

where, 


Figure 5. Relative power spectra of QRS-complex, P and T waves, and muscle noise and motion artifacts [33] .
Figure 6. The similarity between the QRS-complex and sinc wave.
where, 

To illustrate the QRS-complex estimation process, consider 1200 samples ECG signal extracted from record 103 of MIT-BIH arrhythmia database [35] . Investigation of this signal shows that its mean is 0.6973 and its maximum value is 1394. Figure 7(a) illustrates the signal after normalization and mean removal. This signal has 4 periods and the Q, R, and S values together with QR, and RS periods as given in Table 1. The first R-peak is located at 266 and durations between the four successive peaks are 311, 301 and 304 respectively. From these data and Equations (10) and (11) the QRS-complexes can be estimated. Figure 7(b) and Figure 7(c) illustrate the estimated ECG signal and the difference between the original signal and the estimated one respectively. Comparison between Figure 7(a) and Figure 7(c) show that the difference between the original and estimated ECG signals is much sparser compared to the original signal.



Figure 7. The first 1000 sample of record 100. (a) The original signal; (b) The estimated QRS- complex; (c) Difference between the original and the estimated QRS-complex signal.
Table 1. The values of peaks of the Q, R and S waves as well as the RR-periods’ durations.
7. Performance Metrics & Quality Measurement
A practical compression algorithm should not focus totally on compression itself. Many applications have requirements for the quality of the de-compressed signal. A robust compression algorithm should have the ability to maintain the quality of the de-compressed signal while achieving reasonable compression ratio. This is because only good quality signal reconstruction makes sense in reality. The evaluation of performance for testing ECG compression algorithms includes three components: compression efficiency, reconstruction error and com- putational complexity. All data compression algorithms minimizes data storage by reducing the redundancy wherever possible, thereby increasing the Compression Ratio (CR). Thus, the compression efficiency is measured by the CR. The compression ratio and the reconstruction error are usually dependent on each other. The computational complexity component is part of the practical implementation consideration. The following evaluation metrics were employed to determine the performance measures of the proposed method [1] [2] . The CR is defined as the ratio of the number of bits representing the original signal to the number of bits required to store the compressed signal.

where 

Quality of lossy compression schemes is usually determined by comparing the de-compressed data and the original data. If there are no differences at all, then the compression is lossless. Conventional measurements are based on mathematical distortions, such as percentage root mean square difference (PRD) and signal-to-noise ratio (SNR), etc. These measurements are not specific for ECG signals; they reflect the distortion of signal by statistics criteria. They are of general purpose, so the criteria may not be very accurate to describe the characteristics of a specific signal type.
For example, the ECG signal is for medical use, so what concerns medical specification most is the diagnostic feature, which is not covered in the general mathematical descriptions. Thus, in [37] the Weighted Diagnostic Distortion (WDD) is proposed. It uses diagnostic features as criteria and has better feedback from the perspective of medical specialists than other measurements. However, WDD may not bring the benefit and it is far more complex. Table 2 shows a comparison between different quality measures. Among those, PRD is the most widely used measure in ECG data compression. PRD, SNR and STD are mathematically related. PRD is derived from RMSE, which measures the power of errors between the original data and the reconstructed data and is used frequently for prediction. The advantage of the PRD over RMS is its scale-independence. That makes PRD more accurate across different data sets. The PRD indicates the error between the original ECG samples and the reconstructed data. This metric is commonly used for measuring the distortions in reconstructed biomedical sig- nals such as ECG signals and EEG signals. There are subtle differences in calculation; PRD has 3 types for ECG data compression, which are numbered 0, 1, and 2. The definitions are described by Equations (13), (14) and (15) for signal x of length N samples.



Table 2. Comparison between different quality measures.
where, 









The CR and PRD have the close relationship in the lossy compression algorithms. In general, the CR goes higher with the higher lossy level, while the error rate goes up. The final goal of the proposed compression algorithm is to keep the PRD value smaller than that of the conventional methods while maintaining the similar CR. Thus, quality score defined as the ratio between CR and PRD (QS = CR/PRD) is sometimes used to quantify the overall performance of the compression algorithm, considering both the CR and the error rate. A high score represents a good compression performance. Another distortion metric is the root mean square error (RMSE). In data compression, we are interested in finding an optimal approximation for minimizing this metric as defined by the following formula:

Since the similarity between the reconstructed and original signal is crucial from the clinical point of view, the cross correlation (CC) is used to evaluate the similarity between the original signal and its reconstruction.

where 

der to understand the local distortions between the original and the reconstructed signals, two metrics, the maximum error (MAXERR) and the peak amplitude related error (PARE), should be computed. The maximum error metric is defined as

and it shows how large the error is between every sample of the original and reconstructed signals. This metric should ideally be small if both signals are similar. The PARE is defined as

By plotting PARE, one will be able to understand the locations and magnitudes of the errors between the original and reconstructed signals.
8. Experimental Results
Compressive sensing directly acquires a compressed signal representation without going through the intermediate stage of acquiring N samples, where N is the signal length. Since CS-based techniques are still in early
Table 3. 
stages of research and development, specially the development of Analog to Information Conversion (AIC) hardware, signals that are used for experimentation are acquired in the traditional way. The CS measurements of these data are computed from the original ECG signal. Tests were conducted using 10-min long single-lead ECG signal extracted from records 100, 107, 115 and 117 in the MIT-BIH Arrhythmia database. Record 115 is included in the data set to evaluate the performance of the algorithm in the case of irregular heartbeats. The data are sampled with 360 Hz and each sample is represented with 11-bit resolution. The records are split into non- overlapping 1024 samples windows that are processed successively. Then the characteristic points of the ECG waveforms are detected using the procedure introduced in Section 6.2. It relies on the extraction of Q, R and S peaks and locations and the estimation of the QRS-complex. We begin by first detecting the R peaks, since they yield modulus maxima with highest amplitudes. This enables the segmentation of the original ECG record into individual beats. Then, multi scale wavelet decomposition is performed on the difference between the given ECG window, and the estimated QRS-complexes.
To evaluate the performance of the proposed method concerning the amount of data compression, the sparisity of the ECG signal in both the time-domain and the wavelet-domain are measured. As it has been mentioned before, the sparisity of an array x is defined as the number of non-zero entries in x. For this purpose two ECG signals, each of length 6 seconds, extracted from records 100 and 106 of the MIT-BIH database are considered. Figure 8 illustrates the time-domain representation of the two ECG signals, their estimated QRS-complexes and the differences between them. Figure 9 illustrates the threshold wavelet coefficients of the original ECG signal, and that of the differences between the original signal and the estimated QRS-complex for the two records. The bi-orthogonal wavelet filter “bior4.4” has been adopted in the wavelet transformation process where the decomposition has been carried out up to the 8th level. Thresholding is performed such that 98% of the total coefficients’ energy is kept and small coefficients are thrown away.
Figure 10 illustrates the effect of varying the signal length on the sparisity of the time-domain representation of ECG signals and the signals differences of the records after thresholding each of them such that 98% of the total energy is kept and small samples are thrown away. From this Figure it can be concluded that increasing the signal length increases its sparisity. Moreover, the signals differences are sparser compared to the original signals. Figure 11 shows the effect of varying the signal length on the sparisity of the wavelet representation of the ECG signals and the signals differences of the two records after thresholding both of them. Comparing the results presented in Figure 10 and Figure 11 shows that the wavelet transformed signals are sparser compared to the signal in time-domain.

Figure 8. Time-domain representation of the original ECG signal, the estimated QRS-complex and the differences between the original signal, the estimated QRS-complex for two MIT-BIH records. (a) For MIT-BIH record 100; (b) For MIT-BIH record 106.

Figure 9. Wavelet representation of both the original ECG signal, and the differences between the original signal, the estimated QRS-complex for two MIT-BIH records. (a) For MIT-BIH record 100; (c) For MIT-BIH record 106.

Figure 10. Effect of varying the signal length on the sparisity of the time-domain representation of ECG signals extracted from two MIT-BIH records. (a) For MIT-BIH record 100; (c) For MIT-BIH record 106.
Figure 12 illustrates the effect of varying the number of decomposition levels on the sparisity of the wavelet- domain representation of the original ECG signals and the signal differences of two MIT-BIH records. It shows that for record 100, the original signal is sparser than the signal differences for decomposition levels below 7. Otherwise, the signal differences are sparser. However, for record 106, the original signal is sparser than the signal differences for decomposition levels below 5. Otherwise, the signal differences are sparser. This indicates that the sparisity of the signal differences depends on the number of decomposition levels.
Figure 13 illustrates the effect of selecting the wavelet filters on the sparisity of the wavelet-domain representation of the original ECG signals and the signal differences extracted from two MIT-BIH records. It shows that for both records, the signal differences are sparser compared to the original signal for all Daubechies’ filters (from db2 up to db10). Comparing the results presented in Figure 13(a) and Figure 13(b) shows that the sparisity of the wavelet transformed signals depends on the ECG record to be analyzed. This has been checked by considering other MIT-BIH records.

Figure 11. Effect of varying the signal length on the sparisity of the wavelet-domain representation of ECG signals extracted from two MIT-BIH records. (a) For MIT-BIH record 100; (c) For MIT-BIH record 106.

Figure 12. Effect of varying the number of decomposition levels on the sparisity of the wavelet-domain representation of the original ECG signals and the signal differences extracted from two MIT-BIH records. (a) For MIT-BIH record 100; (b) For MIT-BIH record 106.
To explore the effect of controlling the signal sparisity using other wavelet families, sparse representation is checked using bi-orthogonal (bior4.4) DWT filters and compared with that using Daubechies (db6) filters. In both cases the transformation is performed with four detailed levels and one approximation level. Another important process is also considered to improve the signal sparasity; that is by thresholding the wavelet coefficients in different decomposition levels according to the energy content in each subband. In this case, the wavelet coefficients have been thresholded to preserve 98%, 96%, 94% and 90% of the coefficients energy in the 1st, 2nd, 3rd, and 4th details subbands respectively. Moreover, the coefficients in the approximation subband are kept without threshold. After that, the resulting wavelet coefficients are mapped into significant and insignificant coefficients by ones and zeros respectively.
Figure 14 illustrates wavelet coefficients before and after thresholding of the transformed difference between

Figure 13. Effect of varying the Daubechies’ filters on the wavelet sparisty of the wavelet-domain representation of the original ECG signals and the signal differences extracted from two MIT-BIH records. (a) For MIT-BIH record 100; (b) For MIT-BIH record 106.

Figure 14. The wavelet coefficients before and after thresholding of the transformed difference between the original and the estimated QRS-complex together with the mapping of the significant coefficients. (a) For Daubechies db6 wavelet filter; (b) For Biorthogonal bior4.4 wavelet filter.
the original ECG signal and the estimated QRS-complex together with the mapping of the significant and insignificant coefficients for the two wavelet filters. The ECG signal considered here is of length 1200 samples extracted from record 103.
Next, the performance of the proposed compression method has been compared with traditional wavelet based ECG compression techniques. Table 4 illustrates the comparison between the performances of the proposed method and two other methods for the compression of four MIT-BIH ECG records (100, 107, 115 and 117) [15] [36] . The proposed method is tested for the compression of the original signal and the signal differences using the CS technique and bior4.4 wavelet filters and eight decomposition levels. For all cases, the same signal length of 1024 samples is used. As illustrated in the table, the proposed method achieves better results in terms of the CR and PRD. Moreover, the results of the proposed indicate that the signal differences is compressed more than the original signal and smaller
Figure 15. The performance of the proposed CS based technique.
Table 4. Comparison of the proposed method using Compressed Sensing with two other compression methods for four MIT-BIH ECG records (100, 107, 115 and 117).
in compressing 1024 samples of the signal differences. This figure indicates that the proposed method achieved low reconstruction error. This corresponds to high quality recovered ECG signal and even outperforms the two other techniques [15] [36] .
9. Conclusion
This paper investigates CS approach as a revolutionary acquisition and processing theory that enables reconstruction of ECG signals from a set of non-adaptive measurements sampled at a much lower rate than required by the Nyquist-Shannon theorem. This results in both shorter acquisition times and reduced amounts of ECG data. At the core of the CS theory is the notion of signals sparseness. The information contained in ECG signals is represented more concisely in DWT transform domain and its performances in compressing ECG signals are evaluated. By acquiring a relatively small number of samples in the “sparse” domain, the ECG signal can be reconstructed with high accuracy through well-developed optimization procedures. Simulation results validate the superior performance of the proposed algorithm compared to other published algorithms. The performance of the proposed algorithm, in terms of the reconstructed signal quality and compression ratio, is evaluated by adopting DWT spatial domain basis applied to ECG records extracted from the MIT-BIH Arrhythmia Database. The results indicate that average compression ratio of 11:1 with 
Cite this paper
Mohammed M. Abo-Zahhad,Aziza I. Hussein,Abdelfatah M. Mohamed, (2015) Compression of ECG Signal Based on Compressive Sensing and the Extraction of Significant Features. International Journal of Communications, Network and System Sciences,08,97-117. doi: 10.4236/ijcns.2015.85013
References
- 1. Addison, P.S. (2005) Wavelet Transforms and the ECG: A Review. Physiological Measurement, 26, R155-R199.
http://dx.doi.org/10.1088/0967-3334/26/5/R01 - 2. Abo-Zahhad, M.M., Abdel-Hamid, T.K. and Mohamed, A.M. (2014) Compression of ECG Signals Based on DWT and Exploiting the Correlation between ECG Signal Samples. International Journal of Communications, Network and System Sciences, 7, 53-70.
http://dx.doi.org/10.4236/ijcns.2014.71007 - 3. Donoho, D.L. (2006) Compressed Sensing. IEEE Transactions on Information Theory, 52, 1289-1306.
http://dx.doi.org/10.1109/TIT.2006.871582 - 4. Candes, E., Romberg, J. and Tao, T. (2006) Robust Uncertainty Principles: Exact Signal Reconstruction from Highly Incomplete Frequency Information. IEEE Transactions on Information Theory, 52, 489-509.
http://dx.doi.org/10.1109/TIT.2005.862083 - 5. Candes, E.J. and Wakin, M.B. (2008) An Introduction to Compressive Sampling. IEEE Signal Processing Magazine, 25, 21-30.
http://dx.doi.org/10.1109/MSP.2007.914731 - 6. Mamaghanian, H., Khaled, N., Atienza, D. and Vandergheynst, P. (2011) Compressed Sensing for Real-Time Energy-Efficient ECG Compression on Wireless Body Sensor Nodes. IEEE Transactions on Biomedical Engineering, 58, 2456-2466.
http://dx.doi.org/10.1109/TBME.2011.2156795 - 7. Chen, F., Chandrakasan, A.P. and Stojanovic, V.M. (2012) Design and Analysis of a Hardware-Efficient Compressed Sensing Architecture for Data Compression in Wireless Sensors. IEEE Journal of Solid-State Circuits, 47, 744-756.
http://dx.doi.org/10.1109/JSSC.2011.2179451 - 8. Dixon, A.M.R., Allstot, E.G., Gangopadhyay, D. and Allstot, D.J. (2012) Compressed Sensing System Considerations for ECG and EMG Wireless Biosensors. IEEE Transactions on Biomedical Circuits and Systems, 6, 156-166.
http://dx.doi.org/10.1109/TBCAS.2012.2193668 - 9. Polania, L.F., Carrillo, R.E., Blanco-Velasco, M. and Barner, K.E. (2012) Compressive Sensing Exploiting Wavelet Domain Dependencies for ECG Compression. Proceedings of the SPIE Defense, Security, and Sensing, Baltimore, 23-27 April 2012, 83650E.
- 10. Polania, L.F., Carrillo, R.E., Blanco-Velasco, M. and Barner, K.E. (2012) On Exploiting Interbeat Correlation in Compressive Sensing-Based ECG Compression. Proceedings of the SPIE Defense, Security, and Sensing, Baltimore, 23-27 April 2012, 83650D
- 11. Polania, L.F., Carrillo, R.E., Blanco-Velasco, M. and Barner, K.E. (2012) Compressive Sensing for ECG Signals in the Presence of Electromyography Noise. Proceedings of the 38th Annual Northeast Bioengineering Conference (NEBEC), Philadelphia, 16-18 March 2012, 295-296.
http://dx.doi.org/10.1109/NEBC.2012.6207081 - 12. Zhang, Z., Jung, T., Makeig, S. and Rao, B.D. (2013) Compressed Sensing for Energy-Efficient Wireless Telemonitoring of Non-Invasive Fetal ECG via Block Sparse Bayesian Learning. IEEE Transactions on Biomedical Engineering, 60, 300-309.
http://dx.doi.org/10.1109/TBME.2012.2226175 - 13. Polanía, L.F., Carrillo, R.E., Blanco-Velasco, M. and Barner, K.E. (2011) ECG Compression via Matrix Completion. Proceedings of the European Signal Processing Conference, Barcelona, 29 August-3 September 2011, 1-5.
- 14. Candes, E.J. and Tao, T. (2005) Decoding by Linear Programming. IEEE Transactions on Information Theory, 51, 4203-4215.
http://dx.doi.org/10.1109/TIT.2005.858979 - 15. Polania, L.F., Carrillo, R.E., Blanco-Velasco, M. and Barner, K.E. (2011) Compressed Sensing Based Method for ECG Compression. Proceedings of the 2011 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Prague, 22-27 May 2011, 761-764.
- 16. Candes, E. and Tao, T. (2006) Near-Optimal Signal Recovery from Random Projections: Universal Encoding Strategies? IEEE Transactions on Information Theory, 52, 5406-5425.
http://dx.doi.org/10.1109/TIT.2006.885507 - 17. Calderbank, R., Howard, S. and Jafarpour, S. (2010) Construction of a Large Class of Deterministic Sensing Matrices That Satisfy a Statistical Isometry Property. IEEE Journal on Selected Topics in Signal Processing, 4, 358-374.
http://dx.doi.org/10.1109/JSTSP.2010.2043161 - 18. Amini, A., Montazerhodjat, V. and Marvasti, F. (2012) Matrices with Small Coherence Using Parry Block Codes. IEEE Transactions on Signal Processing, 60, 172-181.
http://dx.doi.org/10.1109/TSP.2011.2169249 - 19. Nguyen, T.L. and Shin, Y. (2013) Deterministic Sensing Matrices in Compressive Sensing: A Survey. The Scientific World Journal, 2013, 1-6.
- 20. Candes, E.J. and Romberg, J. (2015) l1-MAGIC: Recovery of Sparse Signals via Convex Programming.
http://www.acm.caltech.edu/l1magic - 21. Boyd, S. and Vandenberghe, L. (2009) Convex Optimization. 7th Edition, Cambridge University Press, Cambridge.
http://www.stanford.edu/~boyd/cvxbook/
http://dx.doi.org/10.1017/CBO9780511804441 - 22. Natarajan, B.K. (1995) Sparse Approximate Solutions to Linear Systems. SIAM Journal on Computing, 24, 227-234.
http://dx.doi.org/10.1137/S0097539792240406 - 23. Mallat, S.G. and Zhang, Z. (1993) Matching Pursuits with Time-Frequency Dictionaries. IEEE Transactions on Signal Processing, 41, 3397-3415.
http://dx.doi.org/10.1109/78.258082 - 24. Chen, S.S., Donoho, D.L. and Saunders, M.A. (2001) Atomic Decomposition by Basis Pursuit. SIAM Review, 43, 129-159.
http://dx.doi.org/10.1137/S003614450037906X - 25. Bioucas-Dias, J. and Figueiredo, M. (2007) A New TwIST: Two-Step Iterative Shrinkage/Thresholding Algorithms for Image Restoration. IEEE Transactions on Image Processing, 16, 2992-3004.
http://dx.doi.org/10.1109/TIP.2007.909319 - 26. Pati, Y.C., Rezaiifar, R. and Krishnaprasad, P.S. (1993) Orthogonal Matching Pursuit: Recursive Function Approximation with Applications to Wavelet Decomposition. Proceedings of the 27th Asilomar Conference on Signals, Systems and Computers, Pacific Grove, 1-3 November 1993, 40-44.
http://dx.doi.org/10.1109/ACSSC.1993.342465 - 27. Grant, M. and Boyd, S. (2008) CVX: Matlab Software for Disciplined Convex Programming.
http://stanford.edu/boyd/cvx - 28. Chambolle, A. and Lions, P.L. (1997) Image Recovery via Total Variation Minimization and Related Problems. Numerische Mathematik, 76, 167-188. http://dx.doi.org/10.1007/s002110050258
- 29. Chambolle, A. (2004) An Algorithm for Total Variation Minimization and Applications. Journal of Mathematical Imaging and Vision, 20, 89-97.
http://dx.doi.org/10.1023/B:JMIV.0000011321.19549.88 - 30. Singh, B.N. and Tiwari, A.K. (2006) Optimal Selection of Wavelet Basis Function Applied to ECG Signal Denoising. Digital Signal Processing, 16, 275-287.
http://dx.doi.org/10.1016/j.dsp.2005.12.003 - 31. Besar, R., Eswaran, C., Sahib, S. and Simpson, R.J. (2000) On the Choice of the Wavelets for ECG Data Compression. Proceedings of the 2000 IEEE International Conference on Acoustics, Speech, and Signal Processing, Istanbul, 4-6 June 2000, 3614-3617.
http://dx.doi.org/10.1109/ICASSP.2000.860184 - 32. Roy, A.B., Dey, D., Mohanty, B. and Banerjee, D. (2012) Comparison of FFT, DCT, DWT, WHT Compression Techniques on Electrocardiogram and Photoplethysmography Signals. IJCA Special Issue on International Conference on Computing, Communication and Sensor Network CCSN, 2012, 6-11.
- 33. Thakor, N.V., Webster, J.G. and Tompkins, W.J. (1983) Optimal QRS Detector. Medical Biological Engineering Computing, 21, 343-250.
- 34. Okada, M. (1979) A Digital Filter for the QRS Complex Detection. Transactions on Biomedical Engineering, 26, 700-703.
- 35. MIT-BIH Arrhythmia Database, 2014.
www. physionet.org/physiobank/database/mitdb - 36. Lu, Z., Kim, D.Y. and Pearlman, W.A. (2000) Wavelet Compression of ECG Signals by the Set Partitioning in Hierarchical Trees Algorithm. IEEE Transactions on Biomedical Engineering, 47, 849-856.
http://dx.doi.org/10.1109/10.846678 - 37. Zigel, Y., Cohen, A. and Katz, A. (2000) The Weighted Diagnostic Distortion (WDD) Measure for ECG Signal Compression. IEEE Transactions on Biomedical Engineering, 47, 1422-1430.
http://dx.doi.org/10.1109/TBME.2000.880093















