Introduction
Remote photoplethysmography (rPPG) is a non-contact optical method that measures the intensity of the light reflected from the skin by the means of a digital camera. Formation of the rPPG signal is thought to be attributed to the pulsatile variations of the light absorption due to the pulsatility of the blood volume inside capillaries, which are due to the pumping of the blood by the heart. However, the results of some research (Jonathan & Leahy, 2011; Kamshilin et al., 2011) are not consistent with the aforementioned model and therefore a new model of rPPG signal formation was proposed (Kamshilin et al., 2015). Its concept is as follows (Kamshilin et al., 2015): during the systole the capillary bed is filled with blood. This filling increases the transmural pressure in the capillaries, which results in the compression of the connective tissue and subsequently, the density of the capillary bed is increased. As a result, the amount of the absorbed light increases. In the end-diastole phase the transmural pressure is the lowest and the amount of back-reflected light is maximal. So, in this model the periodic changes of light scattering and absorption due to the deformation of the connective tissue of the skin are presumed to form the rPPG signal (Kamshilin et al., 2015). rPPG signal, just like the conventional contact photoplethysmogram (PPG), consists of the stationary (DC) and varying (AC) part. The former is related to the skin structure and the average blood volume in arterial and venous system (Tamura et al., 2014). The latter reflects the subtle pulsatile changes between systolic and diastolic phase of the cardiac cycle (Tamura et al., 2014). Detection of these minute changes offers assessment of the pulse rate (PR) as well as other physiological parameters: blood pressure (Jeong & Finkelstein, 2016), oxygen blood saturation (SpO2) (Verkruysse et al., 2017; Wieringa, Mastik & Van Der Steen, 2005), respiration rate (Zhao et al., 2013), etc.
The important advantage of rPPG in comparison to the conventional contact-based methods, such as electrocardiography (ECG) and PPG is the absence of the physical contact between the measurement device and a subject. This is a highly desirable characteristic, since contact methods impair mobility and can cause bias, whereas patches/cuffs may induce irritation, discomfort, stress and/or pain. In the clinical environment additional disadvantages occur: attached sensors may damage the skin (epidermal stripping), bonding between a parent and a child can be impaired, and sensors/wires can appear as artifacts on medical imaging (Aarts et al., 2013). The aforementioned bias in the contact methods is due to the vasodilation that occurs when external pressure is applied to the skin (Fromy, Abraham & Saumet, 1998). This physiological phenomenon serves as a protective mechanism which prevents pressure-induced ischemia (Fromy, Abraham & Saumet, 1998), but it also changes the physiological state of the human body and consequently influences the measurements. rPPG has therefore huge potential in various medical applications. These include telemedicine, triage, neonatal care, intensive care, operative settings, epilepsy monitoring, sleep studies, mental health monitoring, etc. Besides its use in medicine, rPPG could also be implemented in the driver monitoring systems, affective computing, studies of human-robot and human-computer-robot interaction, sports, surveillance systems, polygraph testing, authentication systems, animal health monitoring, etc.
Remote photoplethysmography can be implemented in two modes, reflectance and transmittance. In the latter, a digital camera and a light source are positioned opposite to each other, with observed tissue in-between, while in the former they are positioned on the same side of the observed tissue. The majority of the proposed rPPG systems operate in reflectance mode and comprise of a digital camera, dedicated lighting and skin. All the challenges arise from these three components and include camera properties, intensity and spectra of light source, skin-tone, skin composition and motion artifacts (Wang, 2017). General framework for the PR extraction from the video recordings consists of five steps: definition of the region-of-interest (ROI) that includes skin pixels, pre-processing of the raw RGB signal extracted from the ROI, extraction of the rPPG signal, post-processing of the extracted rPPG signal and PR estimation.
Related works
The feasibility of measuring PR by the means of digital camera was first shown on the time-lapse grayscale images (Takano & Ohta, 2007). In search of the most robust algorithm for the extraction of the pulse signal from video recordings, numerous methods have been proposed afterward: color-space-based, blind-source-separation-based (BSS-based), model-based and data-based. Color-space-based extracts the pulse signal from different channels of standard color spaces (Tsouri & Li, 2015; Verkruysse, Svaasand & Nelson, 2008). BSS-based methods can be divided into Independent-Component-Analysis-driven (ICA) (Poh, McDuff & Picard, 2010) and Principal-Component-Analysis-driven (PCA) (Lewandowska et al., 2011) approaches. Model-based methods include chrominance-based (CHROM) (De Haan & Jeanne, 2013), Blood-Volume-Pulse-vector-based (PBV) (De Haan & Van Leest, 2014) and Plane-Orthogonal-to-Skin (POS) (Wang et al., 2017a) algorithms. The last group, data-based methods, includes the Spatial-Subspace-Rotation (2SR) approach (Wang, Stuijk & De Haan, 2015a). Interested readers may refer to thorough reviews of rPPG methods (McDuff et al., 2015; Rouast et al., 2017; Sikdar, Behera & Dogra, 2016), to their algorithmic principles (Wang et al., 2017a) or to the comparison of different steps of PR extraction from the video recordings (Unakafov, 2017).
An important issue regarding the feasibility of the implementation of rPPG in real-world applications is its high sensitivity to different sources of noise. Most of the existing algorithms derive pulse signal from three-dimensional RGB signals and therefore only two distortions can be eliminated by linear combination of all three color channels which is usually not sufficient. This can either be solved by using a camera with multiple bands (McDuff, Gontarek & Picard, 2014) or by algorithmically increasing the dimensionality of the given RGB signals (Wang et al., 2017c). The latter is done by decomposing the raw RGB signals into multiple components by the means of Fourier Transform. The distortion signals are then suppressed in each of the components and the resulting signals are combined into a single one-dimensional rPPG signal. This algorithm is known as Sub-band rPPG (SB) (Wang et al., 2017c).
Outline of the paper
In this paper, a novel algorithm called Continuous-Wavelet-Transform-based Sub-Band rPPG (SB-CWT) is proposed. SB-CWT uses a different filter bank for the decomposition of the RGB signals than the SB algorithm, namely Continuous Wavelet Transform (CWT) using the analytic generalized Morse wavelet. In CWT, the target signal is correlated with the shifted (translated) and compressed/stretched (also termed dilated or scaled) versions of the analyzing wavelet. By varying the values of the scale and translation parameters the CWT coefficients are obtained. Multiplying these coefficients with their corresponding scaled and translated wavelets provides the constituent wavelets of the analyzed signal. Wavelet transform has the potential to increase the number of the degrees of freedom of the distortion elimination in comparison to the Short Time Fourier Transform (STFT) that was applied in SB. Additionally, wavelet transform provides more accurate information about the time and the frequency of an analyzed signal than STFT. This is achieved by varying the aspect ratio (i.e., ratio between the length of the time window and width of the frequency band) (Ganesan, Das & Venkataraman, 2004). The proposed algorithm is evaluated on the publicly available dataset MMSE-HR (Tulyakov et al., 2016) and compared with the state-of-the-art SB algorithm (Wang et al., 2017c).
The structure of the article is as follows. In the next section, Materials and Methods, the used MSSE-HR dataset is presented, the state-of-the-art SB algorithm used for evaluation of the proposed algorithm is described, the proposed SB-CWT is presented step-by-step, preprocessing steps of the reference signals are shown and the evaluation metrics are defined. In the Results and Discussion sections, the experimental results are presented and discussed. The conclusions are derived in the last section.
Materials and Methods Dataset description
In our research the subset data from the Multimodal Spontaneous Emotion database (Zhang et al., 2016), which includes 2D frontal face video recordings and blood volume pulse (BVP) measurements—MMSE-HR (Tulyakov et al., 2016)—was used. This dataset consists of 102 RGB image sets and corresponding physiological signals of 40 subjects (23 females and 17 males of various ethnic/racial ancestries) performing six different tasks. Table 1 lists all the tasks together with the emotions arousing from them. Image sets (image size: 1,040 × 1,392 pixels) were recorded with RGB 2D color camera (frame rate of 25 fps) of Di3D dynamic imaging system (Dimensional Imaging Ltd, Glasgow, Scotland). BVP signals were measured using Biopac NIBP100D (Biopac System, Inc., Goleta, CA, USA) with the sampling rate of one kHz and collected by Biopac MP150 data acquisition system (Biopac System, Inc., Goleta, CA, USA). All the recorded data were synchronized.
Task no. | Task identification | Task description | Target emotion |
---|---|---|---|
1 | T1 | Listen to a funny joke (interview) | Happiness, amusement |
2 | T8 | Improvise a silly song | Embarrassment |
3 | T9 | Follow-up task similar to T8 | Embarrassment |
4 | T10 | Experience physical threat in dart game | Fear, nervousness |
5 | T11 | Submerge hand into ice water | Physical pain |
6 | T14 | Experience smelly odour | Disgust |
DOI: 10.7717/peerj.5859/table-1
Note:
Adopted from Zhang et al. (2016).
Sub-Band rPPG (SB) method
The proposed SB-CWT method is benchmarked against the SB algorithm (Wang et al., 2017c). This method uses Fourier Transform for transforming the raw RGB signals into the frequency domain. The signals are then projected onto the plane termed as POS (Wang et al., 2017a), the axes of which most likely encapsulate the pulsatile region. The exact projection direction is then localized using the alpha-tuning (De Haan & Jeanne, 2013). Projection to POS together with alpha-tuning serves for the extraction of the rPPG (pulse) signal from the decomposed raw RGB signals. Next, the individual sub-band pulse signals are weighted using the ratio between the pulsatile amplitude and intensity variation amplitude as the weighting function. Finally, Inverse Fourier Transform is used for the transformation of the signal back into the time domain.
Choosing the analyzing wavelet
The general framework of the proposed SB-CWT algorithm is shown in Fig. 1 and the schematic representation of the algorithm is shown in Fig. 2. The framework consists of 13 steps all of which are explained throughout this section. The first question that emerges in any wavelet-based research is which wavelet is the most appropriate for a given application. In the case of using CWT there are several candidate analytic wavelets, such as Morlet, derivative of Gaussian, Shannon, log Gabor, etc. It was shown that there is a wavelet superfamily called generalized Morse wavelets in which all of the aforementioned wavelets can be subsumed (Lilly & Olhede, 2012). Furthermore, it is even possible to objectively define which wavelet within this family is the most appropriate one to be used (Lilly & Olhede, 2012). General Fourier-domain form of the generalized Morse wavelets is: [Formula omitted. See PDF.] where U(ω) is the unit step function, aβγ is the normalizing constant, ω denotes frequency, β is the compactness or the order parameter and γ is the symmetry or the family parameter. Increasing β results in the increased number of oscillations in the time domain and consequent narrowing of the wavelet in the frequency domain. Symmetry parameter γ defines the shape of the wavelet in the frequency domain. The Eq. (1) holds for ω > 0 with [Formula omitted. See PDF.] elsewhere (due to the unit step function). Generalized Morse wavelets are controlled by two parameters, β and γ. The next step was therefore to adjust their values. Symmetry parameter γ was set to 3, because the family of wavelets with this γ value possesses a few desirable properties, which make it recommended for general use (Lilly & Olhede, 2012). The γ = 3 family is known as Airy family. Interested readers may refer to Lilly & Olhede (2012) for more details. The only parameter left that needed to be defined was β. This parameter can be defined as P2/γ, where P2 is the time-bandwidth product. In our example, P2 value was defined for each of the four window lengths l (32, 64, 128 and 256 frames) in such a way that the value of the minimum wavelet bandpass frequency corresponds to the minimum frequency within the human heart rate band (40, 240) BPM. The β value was determined accordingly. The relation between a scale and a pseudo-frequency ωa (this notation is used because the mapping between a wavelet scale to a frequency can be only done in a general sense) is defined as: [Formula omitted. See PDF.] where [Formula omitted. See PDF.] is a camera frame rate, a denotes a scale and FF is the Fourier factor. The maximum scale (corresponds to the minimum wavelet bandpass frequency) can be calculated as N/2σt, where N is signal length and σt is time-domain standard deviation of generalized Morse wavelet. The latter can be approximated by [Formula omitted. See PDF.] . For the definition of σt refer to Eq. (21) in Lilly & Olhede (2009). Fourier factor is defined as [Formula omitted. See PDF.] . For example, given ωa = 40 min−1 and [Formula omitted. See PDF.] , the maximum scale that corresponds to the ωa equals 6.8224 is reached at P2 = 11.
Figure 2: Schematic representation of the proposed SB-CWT method. The illustration shows the entire procedure of extracting the pulse rate signal from facial recordings using the proposed SB-CWT. Notations above the arrows refer to the steps of the proposed algorithm. Notation s2 refers to the face detection, s3 to the ROI size adjustment, s4 to the spatial averaging of the pixel values inside the ROI in each frame, s5–7 to the signal inversion, windowing (l = 256) and temporal normalization, s8 to the continuous wavelet transform, s9–10 to the projection of wavelet coefficients on POS plane and alpha tuning, s11 to the application of weighting function to the wavelet coefficients, s12–13 to the inverse continuous wavelet transform with scale dependent thresholding and overlap-adding. The numbering of all the steps closely follows the numbering in the framework of the proposed algorithm depicted in Fig. 1. Notes: a.u. denotes the arbitrary units, R, G and B denote red, green and blue color channels, respectively. Colors of the signals in amplitude vs. time plots correspond to the RGB color channels. The output plot of s4 shows zero-mean raw RGB signals. DOI: 10.7717/peerj.5859/fig-2
The next required step was to define how finely the scales were to be discretized. This step was important in our application, because discretization indirectly defines the number of sub-bands. In CWT the base scale is usually defined as 21/ν, where ν is the number of voices per octave parameter. The name of this parameter refers to the fact that ν intermediate steps (i.e., scales) are needed if the scale is increased by an octave (i.e., if it is doubled). By increasing the value of ν, the scales are discretized more finely, which results in more sub-bands and consequently in more degrees of freedom for noise suppression. In the present research ν was set to 10, which means that the base scale was set to 21/10. The reason for choosing ν = 10 (common values of ν are 10, 12, 14, 16 and 32) was that we did not want to increase the computational time. Increasing ν namely requires more computations, which results in longer processing time of the algorithm. The described procedure of the selection of the wavelet parameters is depicted by step 1 in Fig. 1.
Estimation of the pulse rate from video recordings
Viola-Jones face detector (Viola & Jones, 2001) was used to detect faces from the first image in each image set (step 2 in Fig. 1). The width of the box containing the detected face was resized to 60%, while the height was kept unchanged (Poh, McDuff & Picard, 2011) (step 3 in Fig. 1). This was done to decrease the number of non-facial pixels in the ROI. Position of the ROI was kept constant throughout the entire image set. Raw RGB signals were obtained by spatial averaging of the RGB pixel intensity values inside the ROI in each frame (De Haan & Jeanne, 2013; Verkruysse, Svaasand & Nelson, 2008) (step 4 in Fig. 1): [Formula omitted. See PDF.] where Ci denotes i-th spatial RGB mean color channel signal (also termed raw RGB signal), [Formula omitted. See PDF.] denotes pixel intensity value I of a pixel located in j-th row and k-th column in i-th color channel, i = 1, 2, 3 and denotes R, G and B channels, respectively. Spatial averaging reduces the camera quantization noise. In the next step, the raw RGB signals were inverted (step 5 in Fig. 1) in order to enable a better comparison with the reference pulse signals derived from the BVP signals. Pulse signals were namely obtained in reflection rPPG mode and therefore the signal reaches its peaks in each end-diastole phase, when the intensity of the diffuse- or body-reflection component is maximal. On the other side, the BVP signals from the MMSE-HR database were acquired using the measurement device based on the Penaz’s principle (Penaz, 1973) or Vascular Unloading Technique. This is an approach where the digital blood flow is held constant by automatic inflation or deflation of the cuff attached to the finger. The applied pressure is directly related to the arterial pressure and is the highest during the ejection phase of the systole. Further processing was not carried out on the full-length raw RGB signals, but on the windowed signals using four different window lengths l = 32, l = 64, l = 128 and l = 256 (step 6 in Fig. 1). The windowed signals were temporally normalized (De Haan & Jeanne, 2013) (step 7 in Fig. 1) in order to remove the dependency of the raw RGB signals on the average skin reflection color. This normalization is defined as: [Formula omitted. See PDF.] where [Formula omitted. See PDF.] is the i-th temporally normalized windowed RGB color channel signal, Cwi is the i-th windowed spatial RGB mean color channel signal, μ(·) is the averaging operator, and i = 1, 2, 3 and denotes R, G and B channels, respectively.
Temporally normalized windowed signals were then decomposed using the CWT (step 8 in Fig. 1). Since all [Formula omitted. See PDF.] signals are discrete, and the generalized Morse wavelet is defined in Fourier domain, the discretized version of the CWT expressed as an inverse Fourier transform was used: [Formula omitted. See PDF.] where a and b (b = 0, 1, 2 … N−1) are dilation and translation parameters, respectively, [Formula omitted. See PDF.] is the discrete Fourier transform of the i-th temporally normalized windowed RGB signal, [Formula omitted. See PDF.] is the complex conjugate of the Fourier-domain form of the wavelet ψa, N denotes signal length and k denotes index of frequency (k = 0, 1, 2 … N−1). To ensure unit energy at each scale, the following normalization was used: [Formula omitted. See PDF.] where [Formula omitted. See PDF.] and [Formula omitted. See PDF.] Combining Eq. (5) with Eq. (6) the CWT can be expressed as: [Formula omitted. See PDF.]
The Eq. (7) was repeated for all scales a and all values b and the final result Wi, where i = 1, 2, 3, was a matrix of wavelet coefficients with the number of rows equal to the total number of scales and the number of columns equal to the signal length N.
In order to obtain the pulse signal all the wavelet coefficients were then projected on POS plane (Wang et al., 2017a) (step 9 in Fig. 1). POS is based on the physiological reasoning and uses a projection plane that is orthogonal to the temporally normalized skin-tone direction. The POS algorithm was chosen because it offers the best general performance in extracting PR from RGB signals (Wang et al., 2017a). The axes of the POS projection plane (Pp) are defined as: [Formula omitted. See PDF.] Projection of the wavelet coefficients on Pp results in two projected components, S1 and S2: [Formula omitted. See PDF.] The components S1 and S2 have in-phase pulsatile and anti-phase specular components which can be algorithmically separated (Wang et al., 2017a). This can be done by the means of the so called alpha-tuning (De Haan & Jeanne, 2013) (step 10 in Fig. 1), which can, for the case of the applied POS algorithm, be expressed as (Wang et al., 2017a): [Formula omitted. See PDF.] where α = |S1|/|S2|, where |·| denotes complex magnitude operator. In the case when the pulsatile variation is stronger than the specular (S1 and S2 are in in-phase), the strength of the resulting signal h is enhanced (Wang et al., 2017a). In the case when specular variations prevail (S1 and S2 are in anti-phase), the α moves the specular variation strength of one signal closer to the other, which results in the specular variation being cancelled out (due to the addition of two signals that are in anti-phase) (Wang et al., 2017a). In the next step, the obtained sub-band pulse signals were weighted using the global (scale-dependent) energy distribution function (step 11 in Fig. 1) defined as the sum of the original wavelet coefficients (the ones obtained by the CWT) at each scale (Bousefsaf, Maaoui & Pruski, 2013). The i-th weighted sub-band (i.e., at i-th scale) is defined as: [Formula omitted. See PDF.] where [Formula omitted. See PDF.] denotes i-th weighted sub-band pulse signal, hij denotes the CWT coefficient at i-th scale and j-th translation parameter, hi* denotes all CWT coefficients at i-th scale and n is the total number of translation parameters. After the weighting, the discretized Inverse Continuous Wavelet Transform (ICWT; step 12 in Fig. 1) together with scale-dependent thresholding was performed in order to transform the weighted sub-band pulse signals back into the time domain. Scale-dependent thresholding allows transformation over selected scales only. The inverses of the selected scales (i.e., frequencies) lie within the range of human heart rate band (i.e., 40–240 BPM). The transformation was performed by summing of the scaled weighted sub-band signals [Formula omitted. See PDF.] over the selected scales only: [Formula omitted. See PDF.] where i runs from imin that is equal to the index of the scale that corresponds to the pseudo-frequency 240 BPM to the index of the last scale, which corresponds to the pseudo-frequency 40 BPM. The scaled weighted sub-band signals are defined as: [Formula omitted. See PDF.] where real(·) denotes the operator that takes only the real part of the complex input and Cψ is the admissibility constant defined as [Formula omitted. See PDF.] (Γ denotes gamma function). This constant ensures that the wavelet of interest has zero mean. Final pulse signal was obtained by overlap adding zero-mean segments of the pulse signal [Formula omitted. See PDF.] using a sliding window with a sliding step of a one time sample (De Haan & Jeanne, 2013; Wang et al., 2017a) (step 13 in Fig. 1).
Average PR estimation from the obtained rPPG pulse signals was based on the average length of interbeat intervals (IBIs) of the entire pulse signal. The IBIs are defined as the intervals between the consecutive systolic peaks. These were detected using the derivative-based method for finding local maxima. The minimum peak-to-peak distance was set to 0.25 s, which corresponds to the upper limit of the human heart rate band (i.e., 240 BPM). This was the only constraint applied in the rPPG signal peak detection algorithm. Extraction of the pulse signal using the state-of-the-art SB method closely followed its original algorithm (Wang et al., 2017a). The only difference is that before the average PR estimation, the pulse signal was inverted due to the same reasoning as explained above.
To summarize, in the proposed SB-CWT the raw RGB signals extracted from the facial images were decomposed applying the CWT. Wavelet transform coefficients were then projected to the POS plane, the axes of which most likely encapsulate the pulsatile region (Wang et al., 2017a). Next, alpha-tuning (De Haan & Jeanne, 2013) was used to localize the exact projection direction. After that the projected coefficients were weighted using the global (scale-dependent) energy distribution function and finally inverted using the ICWT. The comparison of the algorithmic steps of SB-CWT and SB is shown in Table 2. From this table it can be seen that the two methods differ in signal decomposition step, weighting of the sub-band signals and in the transformation of the sub-band signals back into the time domain.
SB (Wang et al., 2017c) | SB-CWT | |
---|---|---|
Signal decomposition | FFT | CWT using generalized Morse wavelet |
rPPG signal extraction | POS (Wang et al., 2017a) + alpha tuning (De Haan & Jeanne, 2013) | POS (Wang et al., 2017a) + alpha tuning (De Haan & Jeanne, 2013) |
Weighting function | Ratio between the pulsatile amplitude and intensity variation amplitude | Scale-dependent energy distribution |
Signal transformation back into the time domain | IFFT | ICWT |
DOI: 10.7717/peerj.5859/table-2
Note:
CWT, continuous wavelet transform; FFT, fast Fourier transform; ICWT, inverse continuous wavelet transform; IFFT, inverse fast Fourier transform; POS, Plane-Orthogonal-to-Skin; rPPG, remote photoplethysmography.
Processing of the reference signals
Pre-processing of the reference signals included the following steps: bandpass filtering, clipping and squaring. Applied bandpass filter was zero phase first-order Butterworth filter with a lower cut-off frequency of 0.67 Hz and upper cut-off frequency of two Hz (these frequencies correspond to 40 and 120 BPM). Filtered signals were then clipped in order to keep only the parts of the signals that are above zero. By subsequent squaring of the signals, noise and diastolic waves were suppressed, while the systolic peaks were emphasized. All the pre-processing steps served for a robust detection of systolic peaks, needed for the estimation of the reference average PR. This was done in the same manner as described above for the rPPG signals, except for the minimum peak-to-peak distance between two consecutive peaks being empirically set to 0.3 s.
Evaluation metrics
In order to evaluate the performance of the proposed algorithm several metrics were used: Signal-to-Noise ratio (SNR), Mean Absolute Error (MAE), Mean Percentage Error (MPE), Root Mean Square Error (RMSE) and coefficient of determination r2. SNR was defined as the ratio of the energy spectrum within five bins around the PR frequency (defined as the peak frequency in the spectrum of the reference BVP signal) and the remaining energy contained in the spectrum. Due to the fact that pulse signal is pseudo-periodic, SNR was calculated within a time interval defined by the 256-frames-long sliding window and sliding step of one frame. The chosen SNR metric is a modified version of the metric proposed by De Haan & Jeanne (2013). Mathematical expression of the chosen SNR metric is: [Formula omitted. See PDF.] where wt (f) is a binary window, S (f) is the spectrum of the rPPG signal obtained by the Fourier transform and f is the frequency in BPM.
Mean absolute error measures the absolute average errors’ magnitude. Its formula is: [Formula omitted. See PDF.] where [Formula omitted. See PDF.] denotes the i-th average PR value obtained from rPPG signals, [Formula omitted. See PDF.] is the i-th reference average PR value and n denotes the total number of facial recordings.
Root mean square error is defined as the square root of the average of squared differences between the estimated [Formula omitted. See PDF.] and actual values [Formula omitted. See PDF.] : [Formula omitted. See PDF.] Mean percentage error is the average of the percentage error by which the estimated values [Formula omitted. See PDF.] differ from the reference values [Formula omitted. See PDF.] : [Formula omitted. See PDF.]
Coefficient of determination r2 defines the proportion of the variance in the dependent variable x predictable from the independent variable y. In the case of a linear least square regression, r2 is equal to the square of the Pearson correlation coefficient r, which is a measure of the linear dependence of two variables: [Formula omitted. See PDF.] where [Formula omitted. See PDF.] and [Formula omitted. See PDF.] are the mean values of the estimated and reference PR values, respectively.
The agreement between the reference pulse signal and rPPG signals was assessed using the Bland–Altman plots. In the original plot the differences between the results of the two methods are plotted against their averages (Bland & Altman, 1986). We however used the plots in which the differences between the results are plotted against the reference method (Krouwer, 2008).
Results
We used 101 RGB image sets from the MMSE-HR database in our study. One image set was eliminated (subject F020 performing task T1), because of the corrupt BVP reference signal preventing the successful extraction of the reference pulse signal. The average size of the ROI was 712 ± 44 × 427 ± 27 pixel (expressed as mean ± standard deviation).
Figure 3 compares the performance of the state-of-the-art SB method and the proposed SB-CWT method for four different window lengths. Figure 3A shows the results of the full original implementation of SB and SB-CWT, whereas Fig. 3B shows the results when weighting (step 11 in Fig. 1) was not applied in neither of the algorithms. The removal of the weighting step was used to study only the effect of the number of sub-bands on the PR estimation. Results in Fig. 3A show that the proposed SB-CWT algorithm outperforms the SB algorithm from the perspective of the median SNR values for all the chosen window lengths. The lowest SNR for both algorithms is achieved at the shortest window length (6.63 dB for SB-CWT and 4.23 dB for SB). The SNR increases with the increasing window length and at l = 256 equals 10.39 dB for SB-CWT (6.24 dB for SB). The highest achieved SNR ratio of the SB (at l = 256) is exceeded by the SB-CWT at the shortest window length l = 32. Results of the paired t-test for all window lengths reject the null hypothesis that pairwise difference between SNR data from SB and SB-CWT has a mean value equal to zero (p < 0.001). These results indicate that there is a statistically significant difference between the SNR values of SB and SB-CWT. In the case when weighting function was not applied in neither of the two studied algorithms (Fig. 3B), SB-CWT still outperforms SB for all window lengths. The highest SNR for SB-CWT is achieved at l = 32 (5.94 dB) and for SB at l =256 (5.53 dB). The trend of increasing SNR with increasing window length from Fig. 3A is not present in Fig. 3B for SB-CWT. There is a statistically significant difference between the SNR values of SB and SB-CWT without applied weighting functions at l = 32, l = 64 and l = 128.
Figure 3: SNR comparison of the SB and the proposed SB-CWT method for four different window lengths (l). (A) Results of the original proposed SB-CWT and state-of-the-art SB methods and (B) Results of the modified SB-CWT and SB methods (without the weighting of the sub-band signals). The red lines in each box indicate the median SNR values, bottom and box borders indicate the 25th and 75th percentiles, the whiskers extend to the largest/smallest values not considered as outliers and the red plus signs mark the outliers. DOI: 10.7717/peerj.5859/fig-3
Figure 4 shows the scatter plots examining the relation between the estimated average PRs obtained by both rPPG methods (denoted by PRrPPG) and the reference average PRs (PRref) for different window lengths. Linear regression lines together with their equations and goodness of fit (coefficients of determination) are also shown. Figures 4A–4D show the results of the full implementation of both rPPG algorithms, whereas Figs. 4E–4H show the results of both studied algorithms without applied weighting function. Results in Figs. 4A–4D show that with the increasing window length the agreement between the PRrPPG estimated using the SB-CWT and the PRref is getting higher. In the case of SB the described trend is disrupted at l = 256, since the agreement is lower than at l = 128. The proposed SB-CWT outperforms SB for all window lengths. Just as it was the case with the SNR, the SB-CWT at the shortest window length outperforms SB at the longest window length. All the correlation coefficients from which coefficients of the determination were calculated are statistically significant (p < 0.001). Results in Figs. 4E–4H show that the agreement between the PRrPPG estimated using the SB-CWT and the PRref is decreasing with the increasing window length. In the case of the SB method, there is no evident trend with increasing window length when it comes to the agreements between PRrPPG and PRref. The range of r2 values for SB from Figs. 4E–4H are comparable with the values from Figs. 4A–4D.
Figure 4: Scatter plots and regression line plots comparing the estimated pulse rates obtained by the proposed SB-CWT and state-of-the-art SB algorithms (PRrPPG) with the pulse rates obtained from the reference signal (PRref). Pulse rates estimated from the rPPG signals (PRrPPG) are plotted against the pulse rates estimated from the reference pulse signals (PRref) for different window lengths and implementations of the algorithms: (A) Results of the original SB-CWT and SB methods for l = 32, (B) Results of the original SB-CWT and SB methods for l = 64, (C) Results of the original SB-CWT and SB methods for l = 128, (D) Results of the original SB-CWT and SB methods for l = 256, (E) Results of the modified SB-CWT and SB methods (without the weighting of the sub-band signals) for l = 32, (F) Results of the modified SB-CWT and SB methods (without the weighting of the sub-band signals) for l = 64, (G) Results of the modified SB-CWT and SB methods (without the weighting of the sub-band signals) for l = 128, (H) Results of the modified SB-CWT and SB methods (without the weighting of the sub-band signals) for l = 256. Each subplot shows the equations of the regression lines and coefficients of determination (r2). DOI: 10.7717/peerj.5859/fig-4
Figure 5 shows the Bland–Altman plots depicting the agreements between the PRrPPG (obtained by SB and SB-CWT) and the PRref. The subfigures in Fig. 5 are labelled similarly as in Fig. 4. The Bland–Altman plots in Figs. 5A–5D are in agreement with the results in Figs. 4A–4D. The mean differences decrease with the increasing window length. The differences range from 5.37 to 1.82 BPM for SB-CWT and from 22.18 to 18.80 BPM for SB. For the window length l = 32, 95% of the measurements fall within −6.11 and 16.85 BPM for SB-CWT (−1.65 and 46.01 BPM for SB). In the case of l = 256, 95% of the measurements fall within −6.04 to 9.68 BPM (−10.58 and 48.19 BPM for SB). With the increasing window length, the agreement limits in SB-CWT are becoming narrower. On average, both methods show positive bias on the estimated PRs, but the bias on the proposed SB-CWT is lower than the bias on the SB. The defined intervals of agreement however do not say anything about the acceptability of the SB-CWT. Results in Figs. 5E–5H show increasing mean differences for SB-CWT (ranging from 10.13 to 14.23 BPM) and decreasing mean differences for SB (ranging from 17.52 to 15.47 BPM) with the increasing window lengths. The positive bias is shown on the PR estimation for both methods and for all window lengths. The agreement limits on Figs. 5E–5H are wider than the limits on Figs. 5A–5D.
Figure 5: The Bland–Altman plots showing agreements between the estimated pulse rates obtained by the proposed SB-CWT and state-of-the-art SB algorithms (PRrPPG) and the pulse rates obtained from the reference signals (PRref). Figure shows the relation between the pulse rates estimated from the rPPG signals (PRrPPG) and the reference signals (PRref) for different window lengths and implementations of the algorithms: (A) Results of the original SB-CWT and SB methods for l = 32, (B) Results of the original SB-CWT and SB methods for l = 64, (C) Results of the original SB-CWT and SB methods for l = 128, (D) Results of the original SB-CWT and SB methods for l = 256, (E) Results of the modified SB-CWT and SB methods (without the weighting of the sub-band signals) for l = 32, (F) Results of the modified SB-CWT and SB methods (without the weighting of the sub-band signals) for l = 64, (G) Results of the modified SB-CWT and SB methods (without the weighting of the sub-band signals) for l = 128, (H) Results of the modified SB-CWT and SB methods (without the weighting of the sub-band signals) for l = 256. The solid lines denote the mean differences. Between the upper and lower limits of the agreement (shown as dashed lines) lie 95% of all the differences. DOI: 10.7717/peerj.5859/fig-5
Statistical comparison of the SB and proposed SB-CWT methods (full implementation of the algorithms) in terms of MAE, MPE and RMSE is shown in Table 3. MAE, MPE and RMSE values decrease with the increasing window length. At l = 256, MAE, MPE and RMSE of the estimated PRs using SB-CWT are all more than 7 times lower than MAE, MPE and RMSE of the PRs estimated using SB (at l = 256). These results all indicate better performance of SB-CWT when compared to SB.
MAE (BPM) | MPE (%) | RMSE (BPM) | ||||
---|---|---|---|---|---|---|
SB (Wang et al., 2017c) | SB-CWT | SB (Wang et al., 2017c) | SB-CWT | SB (Wang et al., 2017c) | SB-CWT | |
l = 32 | 22.2 | 5.7 | −28.7 | −7.1 | 222.9 | 54.0 |
l = 64 | 20.7 | 4.1 | −26.3 | −4.6 | 208.0 | 35.1 |
l = 128 | 19.6 | 3.0 | −24.6 | −3.2 | 196.2 | 23.9 |
l = 256 | 18.9 | 2.4 | −23.6 | −2.5 | 189.0 | 18.3 |
DOI: 10.7717/peerj.5859/table-3
Note:
MAE, mean absolute error; MPE, mean percentage error; RMSE, root mean square error.
The computational times of SB-CWT and SB algorithms for the chosen facial recording of the subject F010 performing task 11 for all four window lengths are shown in Table 4. The chosen recording consists of 999 frames, which translates to the total duration of 39.96 s. This video was selected because its duration is equal to the median duration of all the used recordings from the MMSE-HR database. The computational times are expressed as mean ± standard deviation and calculated based on the times of five consecutive executions of the entire code that extracts final pulse signal from the facial recordings. The results show that SB-CWT is computationally more demanding than SB algorithm. All the computations and calculations were performed using the PC with an Intel Core i7 processor and 16 GB of RAM memory. The used software was written in MATLAB R2018a (MathWorks, Inc., Natick, MA, USA).
tl = 32 (s) | tl = 64 (s) | tl = 128 (s) | tl = 256 (s) | |
---|---|---|---|---|
SB (Wang et al., 2017c) | 11.25 ± 0.16 | 11.43 ± 0.30 | 11.44 ± 0.17 | 11.32 ± 0.09 |
SB-CWT | 16.51 ± 0.12 | 17.00 ± 0.19 | 17.48 ± 0.19 | 17.41 ± 0.28 |
DOI: 10.7717/peerj.5859/table-4
Note:
Values are expressed as mean ± standard deviation and are calculated based on the computational times of five consecutive executions of the code.
Discussion
The main reason for comparing the proposed SB-CWT with SB was that SB outperforms POS algorithm (Wang et al., 2017c), whereas POS was reported to have the best overall performance in a large benchmark (Wang et al., 2017a). POS was benchmarked against the single wavelength method (G) (Verkruysse, Svaasand & Nelson, 2008), the method combining two channels (G-R) (Hülsbusch, 2008), PCA (Lewandowska et al., 2011), ICA (Poh, McDuff & Picard, 2011), CHROM (De Haan & Jeanne, 2013), PBV (De Haan & Van Leest, 2014) and 2SR (Wang, Stuijk & De Haan, 2015a) algorithms. All these algorithms were the most commonly used state-of-the-art algorithms at that time.
The proposed SB-CWT algorithm was evaluated on the publicly available MMSE-HR dataset. There are numerous datasets that are in general suitable for the evaluation of the rPPG algorithms (see Table 5), but to the best of our knowledge MMSE-HR is the only one that offers uncompressed facial recordings. This type of recordings is needed for the evaluation of the methods that exploit model-based algorithms for rPPG signal extraction (as is the case with POS algorithm applied in both, SB-CWT and SB). The compression may namely eliminate the pulsating component of the rPPG signal that is related to the blood pulsations and can also pollute the signal with additional artifacts (Wang et al., 2017a). The main reason for publishing the list of datasets that can be used in rPPG studies is to encourage the rPPG community to start using them more commonly. These databases are unfortunately only rarely used by the rPPG researchers, even though most of the proposed algorithms are not based on model-based methods for rPPG signal extraction.
DEAP (Koelstra et al., 2012) |
Participants: 32 subjects (16 F, 16 M), age: 26.9 (mean) |
Recorded Physiological Data: BVP, EOG, EMG, EEG, GSR, skin temp., resp. rate |
Video Data: RGB videos, 720 × 576 @ 50 fps, DV PAL (AVI; h264 codec), duration: 1 min, distance: 1 m, no. of videos: 120 |
Illumination Conditions: controlled illumination (two different indoor environments) |
Citations in rPPG researches: Qi et al. (2017), Sarkar (2017) and Unakafov (2017) |
MAHNOB-HCI (Soleymani et al., 2012) |
Participants: 30 subjects (17 F, 13 M), age: 26.06 ± 4.39 (mean ± st. dev.) |
Recorded Physiological Data: ECG, EEG, GSR, resp. rate, skin temp., eye gaze data, audio signals |
Video Data: RGB videos, 750 × 580 @ 61 fps, distance: 0.4 m, duration: various |
Illumination Conditions: controlled illumination + uncontrolled illumination from the LCD screen |
Citations in rPPG researches: Hassan et al. (2017a, 2017b), Heusch, Anjos & Marcel (2017), Hsu, Ambikapathi & Chen (2017), Li et al. (2014), Niu et al. (2017) and Tulyakov et al. (2016) |
PURE (Stricker, Müller & Gross, 2014) |
Participants: 10 (2 F, 8 M) |
Recorded Physiological Data: PPG (pulse rate and SpO2) |
Video Data: RGB videos, 640 × 480 @ 30 fps, distance: 1.1 m, duration: 1 min, no. of videos: 60 |
Illumination Conditions: daylight only |
Citations in rPPG researches: Hsu, Ambikapathi & Chen (2017) |
DECAF (Abadi et al., 2015) |
Participants: 30 subjects (14 F, 16 M) |
Recorded Physiological Data: MEG, EOG, ECG, EMG |
Video Data: NIR videos, 20 fps |
Illumination Conditions: controlled illumination |
Citations in rPPG researches: none |
MMSE-HR (Tulyakov et al., 2016) |
Participants: 40 (23 F, 17 M) |
Recorded Physiological Data: BVP |
Video Data: RGB image sequences, 1,040 × 1,392 @ 25 fps, duration: various, no. of videos: 102 |
Illumination Conditions: controlled illumination |
Citations in rPPG researches: Macwan, Benezeth & Mansouri (2018) |
AMIGOS (Miranda-Correa et al., 2017) |
Participants: 40 subjects (13 F, 27 M), age: 28.3 (mean) |
Recorded Physiological Data: audio signals, EEG, GSR, ECG |
Video Data: RGB videos, 1,280 × 720 @ 25 fps, HD (MOV; h264 codec), recordings of individuals and 4 subjects group |
Illumination Conditions: controlled illumination |
Citations in rPPG researches: none |
COHFACE (Heusch, Anjos & Marcel, 2017) |
Participants: 40 subjects* (12 F, 28 M), age: 35.60 ± 11.77 (mean ± st. dev.) |
Recorded Physiological Data: BVP, resp. rate |
Video Data: RGB videos, 640 × 480 @ 20 fps, MP4 format inside AVI movie container, duration: 1 min, no. of videos: 140 |
Illumination Conditions: (1) controlled illumination only and (2) daylight only |
Citations in rPPG researches: none |
UBFC-RPPG (Bobbia et al., in press) |
Participants: no info |
Recorded Physiological Data: PPG (pulse rate and SpO2) |
Video Data: RGB videos, 640 × 480 @ 30 fps, distance: 1 m, duration: 2 min, no. of videos: 43 |
Illumination Conditions: no info |
Citations in rPPG researches: Macwan, Benezeth & Mansouri (2018) |
A dual-mode sleep video database (Hu et al., 2018) |
Participants: 12 (2 F, 10 M), age: 21–38 |
Recorded Physiological Data: ECG |
Video Data: thermal and IR videos, 640 × 480, distance: 1–3 m, no. of videos: 56 |
Illumination Conditions: controlled IR illumination |
Citations in rPPG researches: none |
DOI: 10.7717/peerj.5859/table-5
Notes:
BP, blood pressure; BVP, blood volume pulse; ECG, electrocardiography; EEG, electroencephalography; EMG, electromyography; EOG, electrooculography; F, female; GSR, galvanic skin response; IR, infrared; M, male; NIR, near-infrared; PPG, photoplethysmography; resp. rate, respiratory rate; RGB, red, green, blue color space; skin temp., skin temperature; SpO2, blood oxygen saturation; st. dev., standard deviation.
*Videos are available for only 22 subjects.
Results from Figs. 3A, 4A–4D, 5A–5D and Table 3 indicate that the proposed SB-CWT outperforms the state-of-the-art SB algorithm. The reasoning behind the better performance is multifaceted. Firstly, at l = 32, l = 64 and l = 128, SB-CWT segments the input rPPG signal on more sub-bands, which in turn increases the effectiveness of the separation of the pulse signal from noise components. SB-CWT (SB) offers 25 (5) sub-bands at l = 32, 26 (10) at l = 64, 26 (21) at l = 128 and 26 (43) at window length l = 256. It can be seen that with increasing window length the number of sub-bands in SB is increasing, while it does not change in SB-CWT. The only exception is at l = 64, when SB-CWT has one sub-band more than l = 32. This is however attributable to the discretization of scale values and their correspondence with the limits of human heart rate frequency band. Secondly, in SB algorithm the weighting function is based on the assumption that noise related to motion causes large intensity variations (Wang et al., 2017c). In the proposed SB-CWT, the weighting function however assumes that the amplitudes of the pulse signals are bigger than those of the noise signals (Bousefsaf, Maaoui & Pruski, 2013). Since the MMSE-HR database does not include very challenging recordings from the perspective of motion artifacts, the selected weighting function in the proposed SB-CWT is more likely more appropriate than the weighting function applied in the SB. This is partially confirmed by the fact that at l = 256 the number of sub-bands in SB-CWT is lower than the number in SB, but SB-CWT still performs significantly better.
In order to assess only the effect of the number of sub-bands on the performance of both algorithms, SB-CWT and SB methods were modified in the way that the weighting functions were not applied (step 11 in Fig. 1 was omitted). The results of modified SB and SB-CWT algorithms are shown in Figs. 3B, 4E–4H and 5E–5H. It is interesting that the median SNR value of SB-CWT at l = 256 is comparable to that of SB (see Fig. 3B), even though SB has 17 more sub-bands at that window length. It seems that the reason behind this phenomenon is that there are more degrees of freedom than independent distortions. When this is the case, increasing the number of sub-bands does not improve the performance of the algorithm. If weighting functions are applied (Fig. 3A), SB-CWT shows however significantly better performance than SB at l = 256. The reasoning behind this fact is that the weighting function applied in SB-CWT assumes that the pulse signal exhibits the strongest amplitude compared to noise signals, which is most likely the case in the studied facial recordings. The results from Figs. 4E–4H and 5E–5H also show that the agreement between the PRrPPG and PRref does not increase with the increasing window length. This is due to the fact that the applied weighting functions in SB and SB-CWT serve for combining the individual sub-band pulse signals into a single output pulse signal. If there is no weighting applied, the noise-dominated sub-band signals are not suppressed from the individual sub-band pulse signals, which results in a noisier output pulse signal. This is more pronounced at longer window lengths, when the number of frequency bins and consequently the number of sub-band signals is larger. Furthermore, the applied method of estimating average PR from IBIs is very sensitive to noise. This consequently further aggravates the agreement between the PRrPPG and PRref.
Despite the improved performance of SB-CWT in comparison to SB there are some examples, when SB-CWT did not perform better than SB. These examples are shown as outliers in Fig. 3A. These arise from the following subject/task pairs: F013/T8, F018/T1, M007/T10 and M016/T10. There are two main reasons for the poor performance of both: the proposed SB-CWT and the state-of-the-art SB at these examples. Firstly, three out of four subjects listed above are of Hispanic/Latino or African American ethnicities. Their darker skin tone increases the amount of the absorbed light, because of the higher melanin content in the skin. Therefore, the amount of the light that reaches pulsatile blood vessels is reduced which weakens the pulsatile part of the rPPG signal (Wang, Stuijk & De Haan, 2015b). Poorer performance of SB-CWT in the case of darker skin tone is somehow expected, since the algorithm is not designed for addressing the low pulsatility issues, just as it is the case with SB. Secondly, tasks T1 and T8 include more facial and head motion than the other tasks. The motion reduces the amount of the skin pixels inside the ROI with the fixed position and at the same time increases the amount of the non-skin pixels. The effect of poorer facial mask could be however improved by implementing face tracker and/or skin pixel segmentation.
The major advantage of the proposed SB-CWT method is, in general, its ability to offer more sub-bands (by increasing the number of voices per octave) than the SB method, which means that it offers suppression of more noise signals. Next, SB-CWT enables the usage of a weighting function based on the global (scale-dependent) energy distribution, which serves as an additional filter of undesired signal components. Both the advantages result in a more accurate assessment of PR from facial recordings compared to SB. Additionally, an important advantage of SB-CWT can be also seen when real-time applicability of SB-CWT and SB is discussed. SB-CWT namely offers more robust pulse-rate assessment at lower latency when compared to SB. Considering 25 fps camera recordings, the latency at l = 32 is 1.28 s, at l = 64 2.56 s, at l = 128 5.12 s and at l = 256 the latency equals 10.24 s. In order to achieve a comparable performance of SB-CWT at l = 32, the window length has to be increased to l = 256 for SB. In this case, there is an almost 9 s difference between the latencies. Finally, the method was evaluated on publicly available datasets. Therefore, it is easy to reproduce the results and compare them with the results of other algorithms.
On the other hand, SB-CWT has some disadvantages. Firstly, its inherent disadvantage, which is the same as with the SB, is that the algorithm cannot suppress motion that is of the same frequency as the pulse. However, this problem may be solved by applying Amplitude-Selective Filtering (Wang et al., 2017b), which exploits the fact that relative pulsatile amplitude of the rPPG signal varies within a specific range. Secondly, the weighting function based on the global (scale-dependent) energy distribution function assumes that the pulse wave signal exhibits the strongest amplitude compared to the noise signals. This is however not always the case, especially when longer portions of noisy signal are examined. Thirdly, SB-CWT is computationally more demanding than SB. However, the computational times (see Table 4) are still more than two times shorter than the actual duration of the sample facial recording, therefore the processing latency due to the windowing is of greater importance when it comes to assessing the feasibility of SB-CWT in real-world applications. Lastly, the proposed SB-CWT was not tested on the most challenging use case scenarios that involve a lot of motion. This is mostly due to the fact that to the best of our knowledge there is no appropriate publicly available datasets for the motion robustness testing of the proposed SB-CWT. However, it can be assumed that, since SB-CWT offers more sub-bands, it has the potential to offer significant motion robustness. This is however to be studied in the future work.
Conclusions
The extent of the elimination of the noise signals from the pulse signal in rPPG depends on the dimensionality of the acquired video signal. In order to increase the dimensionality of the RGB signals, SB-CWT algorithm based on the wavelet signal decomposition using generalized Morse wavelet was applied in the present work. The performance of the proposed algorithm was tested on the 101 facial videos from the publicly available dataset MMSE-HR and compared to the state-of-the-art SB algorithm using different window lengths. The results indicate that the proposed SB-CWT algorithm outperforms the state-of-the-art SB algorithm in terms of SNR and agreement between the estimated and the reference PRs. In order to further assess the feasibility of the proposed SB-CWT method in real-world applications, the method needs evaluation on recordings with significant motion.
Additional Information and Declarations
Competing Interests
The authors declare that they have no competing interests.
Author Contributions
Miha Finžgar conceived and designed the experiments, performed the experiments, analyzed the data, contributed reagents/materials/analysis tools, prepared figures and/or tables, authored or reviewed drafts of the paper, approved the final draft.
Primož Podržaj contributed reagents/materials/analysis tools, prepared figures and/or tables, authored or reviewed drafts of the paper, approved the final draft.
Data Availability
The following information was supplied regarding data availability:
Raw data is provided in the Supplemental Files.
Funding
This work of the Young Researcher Miha Finžgar was supported by the Slovenian Research Agency (ARRS). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
Aarts LA, Jeanne V, Cleary JP, Lieber C, Nelson JS, Bambang Oetomo S, Verkruysse W. 2013. Non-contact heart rate monitoring utilizing camera photoplethysmography in the neonatal intensive care unit—a pilot study. Early Human Development 89(12):943-948
Abadi MK, Subramanian R, Kia SM, Avesani P, Patras I, Sebe N. 2015. DECAF: MEG-based multimodal database for decoding affective physiological responses. IEEE Transactions on Affective Computing 6(3):209-222
Bland JM, Altman DG. 1986. Statistical methods for assessing agreement between two methods of clinical measurement. Lancet 327(8476):307-310
Bobbia S, Macwan R, Benezeth Y, Mansouri A, Dubois J. in press. Unsupervised skin tissue segmentation for remote photoplethysmography. Pattern Recognition Letters
Bousefsaf F, Maaoui C, Pruski A. 2013. Continuous wavelet filtering on webcam photoplethysmographic signals to remotely assess the instantaneous heart rate. Biomedical Signal Processing and Control 8(6):568-574
De Haan G, Jeanne V. 2013. Robust pulse rate from chrominance-based rPPG. IEEE Transactions on Biomedical Engineering 60(10):2878-2886
De Haan G, Van Leest A. 2014. Improved motion robustness of remote-PPG by using the blood volume pulse signature. Physiological Measurement 35(9):1913-1926
Fromy B, Abraham P, Saumet JL. 1998. Non-nociceptive capsaicin-sensitive nerve terminal stimulation allows for an original vasodilatory reflex in the human skin. Brain Research 811(1–2):166-168
Ganesan R, Das TK, Venkataraman V. 2004. Wavelet-based multiscale statistical process monitoring: a literature review. IIE Transactions 36(9):787-806
Hassan MA, Malik AS, Fofi D, Saad NM, Ali YS, Meriaudeau F. 2017b. Video-based heartbeat rate measuring method using Ballistocardiography. IEEE Sensors Journal 17(14):4544-4557
Hassan MA, Malik AS, Fofi D, Saad N, Meriaudeau F. 2017a. Novel health monitoring method using an RGB camera. Biomedical Optics Express 8(11):4838-4854
Heusch G, Anjos A, Marcel S. 2017. A reproducible study on remote heart rate measurement. preprint
Hsu GS, Ambikapathi A, Chen MS. 2017. Deep learning with time-frequency representation for pulse estimation from facial videos.
Hu M, Zhai G, Li D, Fan Y, Duan H, Zhu W, Yang X. 2018. Combination of near-infrared and thermal imaging techniques for the remote and simultaneous measurements of breathing and heart rates under sleep situation. PLOS ONE 13(1):e0190466
Hülsbusch M. 2008. An image-based functional method for opto-electronic detection of skin perfusion. RWTH Aachen University. D. Phil. Thesis (in German)
Jeong IC, Finkelstein J. 2016. Introducing contactless blood pressure assessment using a high speed video camera. Journal of Medical Systems 40(4):77
Jonathan E, Leahy MJ. 2011. Cellular phone-based photoplethysmographic imaging. Journal of Biophotonics 4(5):293-296
Kamshilin AA, Miridonov S, Teplov V, Saarenheimo R, Nippolainen E. 2011. Photoplethysmographic imaging of high spatial resolution. Biomedical Optics Express 2(4):996-1006
Kamshilin AA, Nippolainen E, Sidorov IS, Vasilev PV, Erofeev NP, Podolian NP, Romashko RV. 2015. A new look at the essence of the imaging photoplethysmography. Scientific Reports 5(1):10494
Koelstra S, Muhl C, Soleymani M, Lee JS, Yazdani A, Ebrahimi T, Pun T, Nijholt A, Patras I. 2012. DEAP: a database for emotion analysis; using physiological signals. IEEE Transactions on Affective Computing 3(1):18-31
Krouwer JS. 2008. Why Bland–Altman plots should use X, not (Y+X)/2 when X is a reference method. Statistics in Medicine 27(5):778-780
Lewandowska M, Rumiński J, Kocejko T, Nowak J. 2011. Measuring pulse rate with a webcam—a non-contact method for evaluating cardiac activity.
Li X, Chen J, Zhao G, Pietikäinen M. 2014. Remote heart rate measurement from face videos under realistic situations.
Lilly JM, Olhede SC. 2009. Higher-order properties of analytic wavelets. IEEE Transactions on Signal Processing 57(1):146-160
Lilly JM, Olhede SC. 2012. Generalized morse wavelets as a superfamily of analytic wavelets. IEEE Transactions on Signal Processing 60(11):6036-6041
Macwan R, Benezeth Y, Mansouri A. 2018. Remote photoplethysmography with constrained ICA using periodicity and chrominance constraints. BioMedical Engineering OnLine 17(1):22
McDuff DJ, Estepp JR, Piasecki AM, Blackford EB. 2015. A survey of remote optical photoplethysmographic imaging methods.
McDuff D, Gontarek S, Picard RW. 2014. Improvements in remote cardiopulmonary measurement using a five band digital camera. IEEE Transactions on Biomedical Engineering 61(10):2593-2601
Miranda-Correa JA, Abadi MK, Sebe N, Patras I. 2017. AMIGOS: a dataset for mood, personality and affect research on individuals and groups. preprint
Niu X, Han H, Shan S, Chen X. 2017. Continuous heart rate measurement from face: a robust rPPG approach with distribution learning.
Penaz J. 1973. Photoelectric measurement of blood pressure, volume and flow in the finger.
Poh MZ, McDuff DJ, Picard RW. 2010. Non-contact, automated cardiac pulse measurements using video imaging and blind source separation. Optics Express 18(10):10762-10774
Poh MZ, McDuff DJ, Picard RW. 2011. Advancements in noncontact, multiparameter physiological measurements using a webcam. IEEE Transactions on Biomedical Engineering 58(1):7-11
Qi H, Guo Z, Chen X, Shen Z, Jane Wang Z. 2017. Video-based human heart rate measurement using joint blind source separation. Biomedical Signal Processing and Control 31:309-320
Rouast PV, Adam MTP, Chiong R, Cornforth D, Lux E. 2017. Remote heart rate measurement using low-cost RGB face video: a technical literature review. Frontiers of Computer Science 12(5):858-872
Sarkar A. 2017. Cardiac signals: remote measurement and applications. Virginia Tech. D. Phil. Thesis
Sikdar A, Behera SK, Dogra DP. 2016. Computer-vision-guided human pulse rate estimation: a review. IEEE Reviews in Biomedical Engineering 9:91-105
Soleymani M, Lichtenauer J, Pun T, Pantic M. 2012. A multimodal database for affect recognition and implicit tagging. IEEE Transactions on Affective Computing 3(1):42-55
Stricker R, Müller S, Gross H-M. 2014. Non-contact video-based pulse rate measurement on a mobile service robot.
Takano C, Ohta Y. 2007. Heart rate measurement based on a time-lapse image. Medical Engineering & Physics 29(8):853-857
Tamura T, Maeda Y, Sekine M, Yoshida M. 2014. Wearable photoplethysmographic sensors—past and present. Electronics 3(2):282-302
Tsouri GR, Li Z. 2015. On the benefits of alternative color spaces for noncontact heart rate measurements using standard red-green-blue cameras. Journal of Biomedical Optics 20(4):048002
Tulyakov S, Alameda-Pineda X, Ricci E, Yin L, Cohn JF, Sebe N. 2016. Self-adaptive matrix completion for heart rate estimation from face videos under realistic conditions.
Unakafov AM. 2017. Pulse rate estimation using imaging photoplethysmography: generic framework and comparison of methods on a publicly available dataset. preprint
Verkruysse W, Bartula M, Bresch E, Rocque M, Meftah M, Kirenko I. 2017. Calibration of contactless pulse oximetry. Anesthesia & Analgesia 124(1):136-145
Verkruysse W, Svaasand LO, Nelson JS. 2008. Remote plethysmographic imaging using ambient light. Optics Express 16(26):21434-21445
Viola P, Jones M. 2001. Rapid object detection using a boosted cascade of simple features.
Wang W. 2017. Robust and automatic remote photoplethysmography. Technische Universiteit Eindhoven. D. Phil. Thesis
Wang W, Den Brinker AC, Stuijk S, De Haan G. 2017a. Algorithmic principles of remote PPG. IEEE Transactions on Biomedical Engineering 64(7):1479-1491
Wang W, Den Brinker AC, Stuijk S, De Haan G. 2017b. Amplitude-selective filtering for remote-PPG. Biomedical Optics Express 8(3):1965-1980
Wang W, Den Brinker AC, Stuijk S, De Haan G. 2017c. Robust heart rate from fitness videos. Physiological Measurement 38(6):1023-1044
Wang W, Stuijk S, De Haan G. 2015a. A novel algorithm for remote photoplethysmography: spatial subspace rotation. IEEE Transactions on Biomedical Engineering 63(9):1974-1984
Wang W, Stuijk S, De Haan G. 2015b. Unsupervised subject detection via remote PPG. IEEE Transactions on Biomedical Engineering 62(11):2629-2637
Wieringa FP, Mastik F, Van Der Steen AF. 2005. Contactless multiple wavelength photoplethysmographic imaging: a first step toward “SpO2 camera” technology. Annals of Biomedical Engineering 33(8):1034-1041
Zhang Z, Girard JM, Wu Y, Zhang X, Liu P, Ciftci U, Canavan S, Reale M, Horowitz A, Yang H, Cohn JF, Ji Q, Yin L+3 more. 2016. Multimodal spontaneous emotion corpus for human behavior analysis.
Zhao F, Li M, Qian Y, Tsien JZ. 2013. Remote measurements of heart and respiration rates for telemedicine. PLOS ONE 8(10):e71384
Faculty of Mechanical Engineering, University of Ljubljana, Ljubljana, Slovenia
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
© 2018 Finžgar and Podržaj. This is an open access article distributed under the terms of the Creative Commons Attribution License: http://creativecommons.org/licenses/by/4.0/ (the “License”), which permits unrestricted use, distribution, reproduction and adaptation in any medium and for any purpose provided that it is properly attributed. For attribution, the original author(s), title, publication source (PeerJ) and either DOI or URL of the article must be cited. Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Abstract
Background
Remote photoplethysmography (rPPG) is a promising optical method for non-contact assessment of pulse rate (PR) from video recordings. In order to implement the method in real-time applications, it is necessary for the rPPG algorithms to be capable of eliminating as many distortions from the pulse signal as possible.
Methods
In order to increase the degrees-of-freedom of the distortion elimination, the dimensionality of the RGB video signals is increased by the wavelet transform decomposition using the generalized Morse wavelet. The proposed Continuous-Wavelet-Transform-based Sub-Band rPPG method (SB-CWT) is evaluated on the 101 publicly available RGB facial video recordings and corresponding reference blood volume pulse (BVP) signals taken from the MMSE-HR database. The performance of the SB-CWT is compared with the performance of the state-of-the-art Sub-band rPPG (SB).
Results
Median signal-to-noise ratio (SNR) for the proposed SB-CWT ranges from 6.63 to 10.39 dB and for the SB from 4.23 to 6.24 dB. The agreement between the estimated PRs from rPPG pulse signals and the reference signals in terms of the coefficients of determination ranges from 0.81 to 0.91 for SB-CWT and from 0.41 to 0.47 for SB. All the correlation coefficients are statistically significant (p < 0.001). The Bland–Altman plots show that mean difference range from 5.37 to 1.82 BPM for SB-CWT and from 22.18 to 18.80 BPM for SB.
Discussion
The results show that the proposed SB-CWT outperforms SB in terms of SNR and the agreement between the estimated PRs from RGB video signals and PRs from the reference BVP signals.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer