1. Introduction
Wi-Fi received signal strength indicator (RSSI) is one of the basic sensory observations widely used for indoor localization. Due to its nonlinear and random properties, many machine learning approaches have been applied to Wi-Fi RSSI localization [1,2,3,4,5]. In particular, semi-supervised learning algorithms have been suggested to improve the efficiency, which reduces the human effort necessary for calibrating training data. For example, a large amount of unlabeled data can be easily collected by recording only Wi-Fi RSSI measurements, without assigning position labels, which can save resources for collection and calibration. By contrast, labeled training data must be created manually. Adding a large amount of unlabeled data in the semi-supervised learning framework can prevent the decrement in localization accuracy when using a small amount of labeled data.
The majority of the existing semi-supervised learning algorithms use the unlabeled data for the manifold regularization that captures the intrinsic geometric structure of the whole training data [6,7,8,9,10]. More progressed usage of unlabeled data is the pseudolabeling where unlabeled data are artificially labeled, and it is used for learning the estimation model. In [11,12,13], pseudolabels are iteratively updated, and the localization model is learned by the final pseudolabels. In [14], usage of unlabeled data avoids the biased parameter estimation against a small number of labeled data points, to obtain the probabilistic location-RSSI model. In [15,16], unlabeled data are used to find an embedding function from RSSI signal space to 2D location space. Semi-supervised deep learning approaches [17,18,19,20] improve positioning accuracy by using the unlabeled data for extracting features from the RSSI measurements in deep neutral network framework. Compared to discriminative model such as support vector machine (SVM); however, the deep learning methods normally take much time to finish the learning.
This paper proposes a new semi-supervised learning algorithm by interpreting the unlabeled data as spatio-temporal data. In the previous paper [21], the older version of the semi-supervised learning algorithm was applied to a mobile robot in a room size toy experiment. In this paper, the scalable algorithm has been developed for public indoor localization. The algorithm has two separate pseudolabeling and learning processes. First, in the pseudolabeling process, the time-series graph Laplacian SVM optimization is constructed, which estimates labels of the unlabeled data. The existing Laplacian SVM algorithms [6,7,8,9,10] consider only spatial relationship for the training data. To add the temporal meaning to the unlabeled data, this paper attaches the time-series regularization into the Laplacian SVM framework. Time-series learning is reasonable for the indoor localization targeting a smoothly moving human, and the corresponding data comes in a chronological order. As a result, the accurate pseudolabels are made by the proposed time-series semi-supervised learning.
Second, in the learning process, another learning framework to estimate position is constructed. Because the pseudolabels are artificial, the pseudolabeled data cannot be reliable as much as the labeled data. Therefore, it is desirable to limit the reliance on the pseudolabeled data. This can be dealt with a balancing optimization by weighting different balancing parameters between the labeled and pseudolabeled data. The existing semi-supervised methods [22], Ref. [11] might address this balance issue. In [22], the optimization framework consists of a singular value decomposition, a manifold regularization, and a loss function. In [11], the intermediate variable is introduced as a substitute of the original labeled data based on the graph Laplacian optimization framework. However, because many balancing parameters in [22], Ref. [11] are coupled to both the labeled and the unlabeled (or pseudolabeled) terms, it is difficult to adjust the balance. Also, as fewer labeled data are used, the pseudolabels become inaccurate. The proposed method solves this imbalance problem by adding Laplacian Least Square (LapLS) that is produced in this paper by combining the manifold regularization into the transductive SVM (TSVM [23]) structure. Because the pseudolabels are used as the learning input of LapLS, the proposed optimization becomes a linear problem that can be solved fast. In this learning process, two decoupled balancing parameters are individually weighed to the labeled term and the pseudolabeled term, separately, which makes it simple to balance the labeled and the pseudolabeled data. This balancing optimization adjusts the reliance on the pseudolabeled data relative to the labeled data. Outstanding performance of the proposed method is found even when a small amount of labeled training data are used.
The proposed algorithm is evaluated to estimate location of a smartphone carried by a user, and it is compared with some SVM-oriented semi-supervised algorithms. The accurate performance is validated by the analysis of the impact of the learning parameters. Also, according to the variation of the number of the labeled training data and the Wi-Fi access points, the proposed algorithm gives the best localization performance without sacrificing the computation time. In addition, the combination of the proposed learning-based estimation and the particle filter is implemented.
This paper is organized as follows. Section 2 overviews the learning-based indoor localization problem with description of experimental setup for Wi-Fi RSSI use. Section 3 presents the semi-supervised learning algorithms based on the graph Laplacian manifold learning. Section 4 devotes to introduce the proposed algorithm. Section 5 reports empirical results with the parameter setting, the localization with the variance of the number of the training data and access points, the computational time, and the filter-combined localization. Finally, concluding remarks are given in Section 6.
2. Learning-Based Indoor Localization
Figure 1a shows the localization setup where 8 Wi-Fi access points (APs) are deployed in the 37 × 18m2floor. The APs are pre-installed by Korean telecommunication companies such as SKT and LGT. By distinguishing APs’ unique MAC (media access control) addresses, it can define a concentrated set by assigning each RSSI measurement sent from the different APs to the specific location in the set. The experimental floor consists of 3 rooms, open space, and aisle. Each room is filled with desks and chairs, and there are some obstacles such as a vase and a rubbish bin in places on the floor. 5 different people join to collect the training data. In case of collecting the labeled data, the trainers are guided to record the labeled data points on every 1.5 × 1.5m2grid. During the calibration, the repeated measurement sets whose labels are the same location are averaged. As a result, 283 labeled training data are obtained. Similarly, another user is employed to collect 283 labeled test data, which are not used for any training algorithm. The smartphone device is Samsung Galaxy S7 operated by Android OS. The smartphone application to measure the Wi-Fi RSSIs and the locations is developed by Java Eclipse. The Wi-Fi scan function in the Android platform provides the information of MAC address and name of AP, and the RSSI value in dBm. Also, in this experiment, only 2.4 GHz RSSI signal can be collected.
Labeled training data are obtained by placing the receiver at different locations. Let us define the Wi-Fi observation set asxi={zi1,…,zin}∈Rnreceived from n different APs (n=8in this paper), wherezij(1≤j≤n) is a scalar RSSI measurement sent from the j-th access point corresponding to the user’s location(yXi,yYi). Total of the l labeled training data are given by{xi}i=1lwithxi∈X⊆Rn, and{yXi}i=1l,{yYi}i=1l. The unlabeled data set{xi}i=l+1l+uconsists of only the RSSI measurements, without position labels. The training phase builds separate mappingsfX:X→RandfY:X→Rwhich refer to relationships between Wi-Fi signal strength and location, using the labeled training data{(xi,yXi)}i=1land{(xi,yYi)}i=1l, respectively, and the unlabeled data{xi}i=l+1l+u. Because the modelsfXandfYare learned independently, we omit the subscripts offX,fY, andyX,yY, for simplification.
Figure 1c illustrates overview of the proposed indoor localization architecture using Wi-Fi RSSI. The main purpose of the semi-supervised learning-based localization is to learn the accurate positioning model even when a small amount of labeled data is used. For example, Figure 1b shows the Wi-Fi RSSI distribution of the access point #3 that is in a room. Contribution of the semi-supervised learning is to prevent the distribution from being distorted when using much less labeled training data. This will be restated in Example 1 of Section 4.1.
3. Semi-Supervised Learning
This section describes basic semi-supervised learning framework in Section 3.1 and reviews Laplacian least square SVR (LapLS) in Section 3.2 and Laplacian embedded regression least square (LapERLS) in Section 3.3. Key ideas from these algorithms will be applied for the proposed algorithm in the next Section 4.
3.1. Basic Semi-Supervised SVM Framework
Given a set of l labeled samples{(xi,yi)}i=1land a set of u unlabeled samples{xi}i=l+1l+u , Laplacian semi-supervised learning aims to establish a mapping f by the following regularized minimization functional [24]:
f*=argminf∈HkC∑iV(xi,yi,f)+γA ∥f∥A2+γI ∥f∥I2,
where V is a loss function,∥f∥A2is the norm of the function in the Reproducing Kernel Hilbert Space (RKHS)Hk,∥f∥I2is the norm of the function in the low-dimensional manifold, and C,γA,γIare the regularization weight parameters.
The solution of (1) is defined as an expansion of kernel function over the labeled and the unlabeled data, given by
f(x)=∑i=1l+uαiK(xi,x)+b,
with the bias term b and the kernel functionK(xi,xj)=〈ϕ(xi),ϕ(xj)〉, whereϕ(·)is a nonlinear mapping to RKHS.
The regularization term∥f∥A2associated with RKHS is defined as
∥f∥A2=(Φα)T(Φα)=αTKα,
whereΦ=[ϕ(x1),…,ϕ(xl+u)],α=[α1,…,αl+u]T, and K is the(l+u)×(l+u)kernel matrix whose element isKij. We adopt Gaussian kernel given by
Kij=K(xi,xj)=exp−∥xi−xj ∥2/σk2,
whereσk2is the kernel width parameter.
According to the manifold regularization, data points are samples obtained from a low-dimensional manifold embedded in a high-dimensional space. This is represented by the graph Laplacian:
∥f∥I2=1(l+u)2∑i=1l+u ∑j=1l+u Wij f(xi)−f(xj)2,=1(l+u)2fTLf,
where L is the normalized graph Laplacian given byL=D−1/2(D−W)D−1/2,f=[f(x1),…,f(xl+u)]T, W is the adjacency matrix of the data graph, and D is the diagonal matrix given byDii=∑j=1l+u Wij. In general, the edge weightsWijare defined as Gaussian function of Euclidean distance, given by
Wij=exp−∥xi−xj ∥2/σw2,
whereσw2is the kernel width parameter.
Minimizing∥f∥I2is equivalent to penalizing the rapid changes of the regression function evaluated between two data points. Therefore,γI ∥f∥I2 in (1) controls the smoothness of the data geometric structure.
3.2. Laplacian Least Square (LapLS)
This section describes developing of the LapLS algorithm algorithm by combining manifold regularization defined in (5) and transductive SVM (TSVM) [23]. In LapLS, the loss function V in (1) is defined by
V(xi,yi,f)=ei=yi−∑i=1l+uαiK(xi,x)+b.
LapLS finds optimal parametersα, b, and the labelsy1*,…,yu*of the unlabeled data when regularization parameters C andC*are given:
minα,e,e*,b,y1*,…,yu*C2∑i=1l ei2+C*2∑j=1u (ej*)2+γA αTKα+γI αTKLKα,subjectto:yi−∑k=1l+u αk Kik−b−ei=0,i=1,…,l.yj*−∑k=1l+u αk Kjk−b−ej*=0,j=1,…,u.
Optimization in (8) with respect to ally1*,…,yu* is a combinatorial problem [23]. To find the solution, we must search over all possible2ulabels of the unlabeled data. Therefore, this method is not useful when a large amount of the unlabeled data is applied.
3.3. Laplacian Embedded Regularized Least Square (LapERLS)
LapERLS introduces an intermediate decision variableg∈R(l+u)and additional regularization parameterγC into the Laplacian semi-supervised framework (1), as follows [11]:
minf∈Hk,g∈R(l+u)C∑i=1l+uV(xi,gi,f)+γC∑i=1l(gi−yi)2+γA ∥f∥A2+γI ∥g∥I2.
The optimization problem of (9) enforces the intermediate decision variable g to be close to the labeled data and to be smooth with respect to the graph manifold.
Loss function is given by:
V(xi,gi,f)=ξi=gi−∑j=1l+uαjK(xi,xj).
After reorganizing the terms in (9) with respect to manifold regularization and decision function and corresponding parameter, the primal optimization problem is as follows:
minα,g,ξ∈R(l+u)C2∑i=1l+u ξi2+αTKα+12(g−y)TΛ(g−y)+12μgTLg,subjectto:ξi=gi−∑k=1l+u αk Kik,i=1,…,l+u,
whereΛis a diagonal matrix of trade-off parameters withΛii=λifxiis a labeled data point, andΛii=0ifxiis unlabeled,y=[y1,…,yl,0,…,0]T∈R(l+u), and C,λ,μare tuning parameters.
Also, dual formulation of (10) is given by:
minβ∈R(l+u)12βTQ˜β+βTy˜,
where
Q˜=K+(Λ+μL)−1,y˜=(Λ+μL)−1Λy,β=−α.
The main characteristics of this method lies in usingy˜as the input to learning, unlike standard semi-supervised learning that usesy=[y1,…,yl,0,…,0]T. In other words, zero values in y are modified to some values denoting pseudolabels of unlabeled data.
It is noted that when the original labeled set y is replaced with the intermediate decision variable g, the accuracy of the learning may decrease. Moreover, when amount of available labeled data is small, the accuracy of the pseudolabels, which is difference between true labels of unlabeled data points and pseudolabels of the unlabeled data points, decreases significantly.
4. Time-Series Semi-Supervised Learning
This section describes a new algorithm by extracting key ideas from LapLS and LapERLS introduced in the previous section. In Section 4.1, a time-series representation is added to the unlabeled data to obtain pseudolabels. In Section 4.2, the pseudolabels are used in LapLS structure to derive an optimal solution by balancing the pseudolabeled and the labeled data. Notations are equivalent to the previous section.
4.1. Time-Series LapERLS
In [25], a time-series learning optimization problem is suggested by applying Hodric-Prescott (H-P) filter [26], which can capture a smoothed-curve representation of a time-series from the training data, given by
minf∑i=1tf(xi)−yi2+γT∑i=3tf(xi)+f(xi−2)−2f(xi−1)2,
where{(xi,yi)}i=1tis the time-series labeled training data. The second term is to make the sequential pointsf(xi),f(xi−1),f(xi−2) on a line. The solution of (13) in the matrix form is,
f=I+γTDDT−1y,
where
D=00⋯⋯⋯⋯0000⋯⋯⋯01−210⋯⋯001−210⋯0⋮⋮⋮⋮⋮⋮⋮0⋯⋯01−21t×t.
The main idea for a new semi-supervised learning is to assign additional temporal meaning to the unlabeled data. In the proposed algorithm, the H-P filter is added into the LapERLS formulation (9) in the following optimization:
minf∈Hk,g∈R(l+u)C∑i=1l+uV(xi,gi,f)+γC∑i=1l(gi−yi)2+γA ∥f∥A2+γI ∥g∥I2+γT∑i=3l+ug(xi)+g(xi−2)−2g(xi−1)2.
After rearranging (15) using the process similar to Equations (9)–(11), the optimization form of the proposed time-series LapERLS is given by:
minβ∈R(l+u)12βTQ˜β+βTy˜,
where
Q˜=K+(Λ+μ1L+μ2DDT)−1,
y˜=(Λ+μ1L+μ2DDT)−1Λy,
β=-α.
In comparison with (12) of the standard LapERLS,μ2DDT is added to (18). In the following, the experimental examples for describing the difference of the standard LapERLS and the time-series LapERLS are introduced.
Example 1.
Figure 2 examines the accuracy comparison between the standard LapERLS and the time-series LapERLS. In this example, a time-series training data set is collected as a user moves along the path illustrated in Figure 2a. This simulation uses 20% of the labeled training data and 80% of the unlabeled data among total 283 training data. Figure 2 illustrates estimations of the pseudolabels using each time-series LapERLS and standard LapERLS. As shown in Figure 2b,c, the pseudolabels produced by the time-series LapERLS are accurate while the standard LapERLS does not show meaningful pseudolabels. Therefore, the trajectory of pseudolabels from the time-series LapERLS can recover the true trajectory while the standard LapERLS cannot. Obviously, many incorrect pseudolabels such as Figure 2c will derive inaccurate localization performance.
The other physical interpretation about the pseudolabels can be seen from Figure 3. Figure 3a shows the RSSI distribution as the ground truth by using the entire labeled training data, Figure 3b shows the RSSI distribution recovered by the pseudolabels obtained by the time-series LapERLS, and Figure 3c is the RSSI distribution estimated by the pseudolabels of the standard LapERLS. In case of the standard LapERLS as shown in Figure 3c, the distribution is severely distorted due to the incorrectly estimated pseudolabels. On the other hand, the time-series LapERLS gives the very similar distribution to the original distribution.
Example 2.
Accuracy of the pseudolabels is examined by the three cases with comparison to a linear interpolation.
1. Sinusoidal trajectory:
A situation that a user moves a sinusoidal trajectory as described in Figure 4a is considered. The linear interpolation produces the pseudolabels laying on the straight line between two consecutive labeled points. On the other hand, because the proposed algorithm considers both spatial and temporal relation by using manifold and time-series learning, the result of the suggested algorithm can generate accurate pseudolabels as shown in Figure 4b.
2. Wandered trajectory:
The other case in which the linear interpolation is not useful is when the user does not walk straight forward to a destination. For example, in Figure 5, the user wanders around middle of the path. Because there are only two labeled data (one is at bottom, and the other is at top), the linear interpolation could not represent the wandered trajectory. In Figure 5b, it is shown that the developed algorithm generates accurate pseudolabels with respect to the wandered motion of the user.
3. Revisiting the learned trajectory:
The user used to revisit the same site during collecting training data. Suppose that the locations corresponding to the Wi-Fi RSSI measurements are not recorded during walking a path, except the start and end points as shown in Figure 6. It is assumed that we have already learned those area. The result under this situation is shown in Figure 6, where the developed algorithm generates accurate pseudolabels, while the linear interpolation cannot reflect the reality.
4.2. Balancing Labeled and Pseudolabeled Data
Regardless of how accurate the pseudolabels are, it is impossible to regard the pseudolabels as the labeled data, because true labels of the unlabeled data are unknown. One desirable approach is to properly balance the labeled and the pseudolabeled data in the learning process. This is feasible by applying LapLS structure in Section 3.2, which can control the balance of training data by the decoupled parameters C andC* introduced in (8).
Example 3.
Figure 7 illustrates an estimation of the sine function using LapLS in Section 3.2, where we divide the labeled training set in half and use the different values of the parameters, i.e.,C=0.5andC*=0.1. In the latter part, the estimation withC*=0.1 is not accurate. This result can be validated from (8). As the parameterC(*)becomes smaller, the related termC(*) ∑i (ei(*))2becomes also smaller. In other words, the optimization focuses less on the training data points with the smaller parameter value ofC(*).
For the final step in the suggested algorithm, the idea is to treat pseudolabelsy˜ in (18) as the labels of the unlabeled data in the LapLS (8) framework, which forms the following optimization:
minα∈R(l+u),e∈Rl,e*∈Ru,b∈RC2∑i=1l ei2+C*2∑j=1u (e˜j*)2+γA αTKα+γI αTKLKα,subjectto:yi−∑k=1l+u αk Kik−b−ei=0,i=1,…,l.y˜j*−∑k=1l+u αk Kjk−b−e˜j*=0,j=1,…,u,
wherey˜j*are pseudolabels of the unlabeled data fromy˜ (18). Therefore, the non-convex problem of LapLS (8) is modified to a convex problem due to insertion of the pseudolabels. After KKT conditions, we obtain the following linear system:
AX=Y,
with
A=K+Γ1(l+u)×111×(l+u)0,X=αb,Y=Y˜0,
where K is the kernel matrix in (4),α=[α1,…,αl+u]∈R(l+u),b∈R in (2) is the bias value,1(l+u)×1=[1,…,1]T∈R(l+u)is the one vector, andΓis the diagonal matrix withΓii=1/Cfori=1,…,landΓii=1/C*fori=l+1,…,l+u. The pseudolabel vectorY˜∈R(l+u)is a time-series set of the labeled and pseudolabeled data. The optimal solutionα*andb* obtained after solving (20) becomes the final parameters of the localization model, which is defined in (2). In the test phase, when a user queries location by sending a RSSI measurement set, the location is estimated based on the learned localization model.
The proposed algorithm in (20) improves the performance of LapERLS by combining the structure of LapLS. First, the pseudolabels are accurately estimated by assigning temporal-spatio representation into the unlabeled data. Second, it is easy to adjust the balance between the pseudolabels and the labeled data. Moreover, by incorporating the pseudolabels into the LapLS structure, the non-convex problem is transformed to a convex problem that can be computed far faster. The proposed algorithm is summarized in Algorithm 1.
Algorithm 1 Proposed semi-supervised learning for localization |
Step 1: Collect labeled training data set{(xi,yi)}i=1land unlabeled data set{xj}j=1uin a time-series. Step 2: Obtain kernel matrix K in (4), normalized Laplacian matrix L in (5) and matrixΛ in (10). Step 3: Choose values ofμ1andμ2 in (18), and then calculate the pseudolabels in (18). Step 4: Choose values of C andC* , and then solve the linear equation in (20). Step 5: Based on the optimal solutionα*andb* from Step 4, builds the localization model in (2). |
To evaluate the proposed algorithm in comparison with other semi-supervised learning algorithms, two kinds of error are defined. First is the pseudolabel error defined as the average of the distance errors between unlabeled data and pseudolabeled data. Second is the test error is defined as the average of the distance errors between true locations and estimates.
This section consists of the parameter setting in Section 5.1 and the compared localization results according to the variation of the training data in Section 5.2. In Section 5.3, the result according to the variation of the number of Wi-Fi access points and the computation analysis are given. Finally, Section 5.4 describes a combination of particle filter and the suggested learning algorithm.
5.1. Parameter Setting
The hyper-parameters in machine learning algorithms are generally tuned by cross validation that selects parameters values to minimize training error of some split training data sets, e.g., 10-fold cross validation [27]. However, most semi-supervised learning applications use a small number of the labeled data, so it is not suitable to employ the cross validation. This section provides a guideline for selecting all the parameters used in the developed algorithm, by describing the physical meaning of each parameter.
First,λ, i.e., the diagonal elements ofΛ in Step 2 of Algorithm 1, can be interpreted as the importance of the labeled data relative to the unlabeled data by (12) and (18). Ifλis relatively smaller than the particular value defined asλcritical, resultant pseudolabels of the labeled data are different from the true labels. Therefore, it is decided to select value ofλ larger than such critical value. Figure 8a shows the pseudolabel error according to the variation ofλ, whereλcritical=10becomes the proper parameter selection.
Second, selection ofμ1andμ2in Step 3 of Algorithm 1 represents a trade-off relationship between spatial and temporal correlation. If it is desirable to weigh the temporal meaning more than the spatial meaning,μ2is selected larger thanμ1 . Figure 8b shows the error of the pseudolabels according to the variation ofμ2at fixedμ1, which suggests that1<μ2<10 is proper for this data set. Also, Figure 8c gives the impact of the ratioμ2/μ1, whereμ2/μ1<1is proper to this data set.
Third, selection of C andC*in Step 4 of Algorithm 1 represents a trade-off relationship about the relative importance between the labeled data and the pseudolabeled data, as described in V-B. It is highlighted again thatC*should be smaller than C if we want to reduce the reliance of the pseudolabeled data relative to the labeled data. We used 10%, 25%, and 50% labeled data points among total of 283 labeled data points to test an effect of the parameters C andC* whose results are shown in Figure 9a,b. From Figure 9a, it is observed thatC*smaller than C can reduce the test error. In particular, the same test error is found atC*=1 , which implies that the algorithm with only 10% labeled data can have the same performance when using large amount of 50% used labeled data. This result highlights the importance of balancing the labeled and the pseudolabeled data and proves why the proposed algorithm can give good results when using a small number of the labeled data points. Also, Figure 9b shows the impact of values of C andC*when the ratioC/C*is fixed. It is shown that10<C/C*<50 is proper for this data set. Lastly, Figure 9c shows the test error with respect to the variance ofσk used for the kernel matrix. Table 1. summarizes the parameter values used for the rest of the experiments.
5.2. Variation of Number of Training Data
This experiments compares the semi-supervise algorithms, i.e., SSL [28], SSC [22], LapERLS [11] to the proposed algorithm with respect to the variation of the number of labeled training data at the fixed total of the training data, and the results are shown in Figure 10. For example, in case of 75% labeled data, 25% unlabeled data is used. The total of 283 and 93 training data points are used in Figure 10a,b, respectively. From the both results, we can observe that our algorithm outperforms the compared methods. In cases of 100% and 75% labeled data in Figure 10b, SSC gives slightly smaller error than the others. However, considering the advantage given to only SSC, i.e., the information of locations of the Wi-Fi access points, this error reduction is not noticeable.
The major contribution of our algorithm is found from the results when a small amount of labeled data is used. From Figure 10a,b, our algorithm shows the slightly increasing errors as less labeled data are used, while the others give the highly increasing error from 25% percentage. To analyze this result, we check the error of the pseudolabeled data points made in our algorithm, in Figure 10c. When using many labeled training data points such as 50%∼75%, accurate pseudolabels are made, so good learning result is obvious. Even though the pseudolabeled data is inaccurate such that only 10% labeled data is used as in Figure 10c, the test error is still small as in Figure 10a,b. This is because we balanced the pseudolabeled data by reducing the reliance on the pseudolabeled data relative to the labeled data.
5.3. Variation of Number of Wi-Fi Access Points and Computational Time
Figure 11 compares the results when the number of Wi-Fi access points (AP) is changed from 2 to 7, where the APs depicted in Figure 1 are removed by this order #1 → #6 → #3 → #7 → #8 → #4.
Figure 12 summarizes the computational training times of the compared algorithms. Computational complexity of the optimization of SSL, LapERLS, and the proposed algorithm isO(dn2+n3) [29] by the matrix computation and its inversion in (18), where d is the dimension of training data and n is the number of training data. In case of SSC, it additionally needs the computation of harmonic function using k-NN search algorithm which isO(nk(n+d)log(n+d)) [30]. The computation time of SSC increases significantly according to the increasing amount of the labeled data while the others remain lower-bounded in 0.5 s. The proposed algorithm needs little more time than LapERLS and SSL (at most 0.2 s).
5.4. Combination with Particle Filter
While the previous sections have shown the results of localization using the Wi-Fi RSSIs based on the learning algorithms, this section provides a combination with a filtering algorithm. An application of the particle filter can use other kinds of positioning information such as an IMU measurement and map information, to improve indoor localization. Because a detailed description about the particle filter can be found in many works [31,32,33], this paper simply describes how that information with the learning-based localization can be combined in the particle filter framework.
First, a dynamic model given an IMU sensor data can be defined by:
Xt+1=Xt+V^tΔtcos(θ^t)sin(θ^t),
whereXtis a 2-D location at time step t, andV^tandθ^t are the velocity and direction of an user, respectively. However, in (21), accurate estimation for the velocity and the direction is difficult to be obtained for the indoor positioning. For example, for obtaining accurate velocity, the attitude of the smartphone needs to be kept fixed without rotating. More specifically, the estimates of a step detection and velocity of the user have the significant error when the user swings the smartphone by walking. Also, the estimation of the direction is much biased in the indoor environment due to the ferrous material and the integral gyro error.
For such reasons, an alternative dynamic model without employing IMU measurement is the random walk defined by the Gaussian distribution as follows:
Xt+1∼NXt,Σd.
Next, the particle filter uses the estimated location obtained from the learning algorithm as the observation model in the following:
Zt=HXt+Γ,
whereZt=[fX,fY]Tis the estimated location from the learning algorithmfX,fY, and H is the identity matrix, andΓis a Gaussian noise whose mean is zero and variance isΣo.
The key point of the particle filter is to update the weightswtiof each particleXtifori=1,…,m, where m is the number of the particles. It decides the estimationX^tas the weighted summationX^t=∑i=1m wti Xti. The weights are updated in the following:
wti=wt−1i·P(Zt|Xti)·P(Xti|Xt−1i),
whereP(Zt|Xti)∼N(Zt−HXti,Σo). The map information can be used in the probabilityP(Xti|Xt−1i)in the following:
P(Xti|Xt−1i)=Pmifaparticlecrossedawall1−Pmifaparticledidnotcrossawall,
wherePmsimply set to zero because human cannot pass the wall.
Evaluation of the particle filter based on the learning algorithm is shown in Figure 13. It is assumed that the user walks constant speed along the designated path. By the records of timestamp using the smartphone, we calculate the average 5.2 m localization error when using only learning method, and 2.9 m localization error when combining the particle filtering.
6. Conclusions
This paper proposes a new semi-supervised learning algorithm by combining core concepts of LapERLS’s pseudolabeling, time-series learning, and LapLS-based balancing optimization. From the indoor localization experiment, the suggested algorithm achieves high accuracy by using only a small amount of the labeled training data in comparison with the other semi-supervised algorithms. Also, this paper provides a guidance and an analysis for selecting all the parameters used in the proposed algorithm. In addition to the Wi-Fi RSSI-based localization, there are various types of indoor localization methods using IMU, Bluetooth, UWB (Ultrawide band) and camera. This paper has shown the combination with the particle filter from which expandability to involve other types of information is validated. Thus, it is expected that a fusion algorithm with the other positioning approaches might improve localization accuracy for future work.
[Image omitted. See PDF.]
[Image omitted. See PDF.]
[Image omitted. See PDF.]
[Image omitted. See PDF.]
[Image omitted. See PDF.]
[Image omitted. See PDF.]
[Image omitted. See PDF.]
[Image omitted. See PDF.]
[Image omitted. See PDF.]
[Image omitted. See PDF.]
[Image omitted. See PDF.]
[Image omitted. See PDF.]
[Image omitted. See PDF.]
[Image omitted. See PDF.]
Value ofλ | Value ofμ1 | Value ofμ2 | Value ofC | Value ofC* | Value ofσk |
---|---|---|---|---|---|
10 | 5 | 2 | 10 | 1 | 30 |
Funding
This work was supported by Electronics and Telecommunications Research Institute(ETRI) grant funded by ICT R&D program of MSIP/IITP, 2017-0-00543, Development of Precise Positioning Technology for the Enhancement of Pedestrian’s Position/Spatial Cognition and Sports Competition Analysis.
Conflicts of Interest
The author declares no conflict of interest.
1. Hernández, N.; Ocaña, M.; Alonso, J.; Kim, E. Continuous space estimation: Increasing WiFi-based indoor localization resolution without increasing the site-survey effort. Sensors 2017, 17, 147.
2. Zheng, L.; Hu, B.; Chen, H. A high accuracy time-reversal based WiFi indoor localization approach with a single antenna. Sensors 2018, 18, 3437.
3. Wang, Y.; Xiu, C.; Zhang, X.; Yang, D. WiFi indoor localization with CSI fingerprinting-based random forest. Sensors 2018, 18, 2869.
4. Nuño-Maganda, M.; Herrera-Rivas, H.; Torres-Huitzil, C.; Marisol Marin-Castro, H.; Coronado-Pérez, Y. On-Device learning of indoor location for WiFi fingerprint approach. Sensors 2018, 18, 2202.
5. Botta, M.; Simek, M. Adaptive distance estimation based on RSSI in 802.15. 4 network. Radioengineering 2013, 22, 1162–1168.
6. Zhou, M.; Tang, Y.; Nie, W.; Xie, L.; Yang, X. GrassMA: Graph-based semi-supervised manifold alignment for indoor WLAN localization. IEEE Sens. J. 2017, 17, 7086–7095.
7. Zhang, L.; Valaee, S.; Xu, Y.; Ma, L.; Vedadi, F. Graph-based semi-supervised learning for indoor localization using crowdsourced data. Appl. Sci. 2017, 7, 467.
8. Du, B.; Xinyao, T.; Wang, Z.; Zhang, L.; Tao, D. Robust graph-based semisupervised learning for noisy labeled data via maximum correntropy criterion. IEEE Trans. Cybern. 2018, 49, 1440–1453.
9. Wang, M.; Fu, W.; Hao, S.; Tao, D.; Wu, X. Scalable semi-supervised learning by efficient anchor graph regularization. IEEE Trans. Knowl. Data Eng. 2016, 28, 1864–1877.
10. Yoo, J.; Kim, H. Target localization in wireless sensor networks using online semi-supervised support vector regression. Sensors 2015, 15, 12539–12559.
11. Chen, L.; Tsang, I.W.; Xu, D. Laplacian embedded regression for scalable manifold regularization. IEEE Trans. Neural Netw. Learn. Syst. 2012, 23, 902–915.
12. Nie, F.; Xu, D.; Li, X.; Xiang, S. Semisupervised dimensionality reduction and classification through virtual label regression. IEEE Trans. Syst. Man Cybern. Part B Cybern. 2011, 41, 675–685.
13. Kumar Mallapragada, P.; Jin, R.; Jain, A.K.; Liu, Y. Semiboost: Boosting for semi-supervised learning. IEEE Trans. Pattern Anal. Mach. Intell. 2009, 31, 2000–2014.
14. Ouyang, R.W.; Wong, A.K.S.; Lea, C.T.; Chiang, M. Indoor location estimation with reduced calibration exploiting unlabeled data via hybrid generative/discriminative learning. IEEE Trans. Mobile Comput. 2012, 11, 1613–1626.
15. Jain, V.K.; Tapaswi, S.; Shukla, A. RSS Fingerprints Based Distributed Semi-Supervised Locally Linear Embedding (DSSLLE) Location Estimation System for Indoor WLAN. Wirel. Pers. Commun. 2013, 71, 1175–1192.
16. Xia, Y.; Ma, L.; Zhang, Z.; Wang, Y. Semi-Supervised Positioning Algorithm in Indoor WLAN Environment. In Proceedings of the IEEE Vehicular Technology Conference, Glasgow, UK, 11–14 May 2015; pp. 1–5.
17. Mohammadi, M.; Al-Fuqaha, A.; Guizani, M.; Oh, J.S. Semisupervised deep reinforcement learning in support of IoT and smart city services. IEEE Internet Things J. 2017, 5, 624–635.
18. Gu, Y.; Chen, Y.; Liu, J.; Jiang, X. Semi-supervised deep extreme learning machine for Wi-Fi based localization. Neurocomputing 2015, 166, 282–293.
19. Khatab, Z.E.; Hajihoseini, A.; Ghorashi, S.A. A fingerprint method for indoor localization using autoencoder based deep extreme learning machine. IEEE Sens. Lett. 2017, 2, 1–4.
20. Jiang, X.; Chen, Y.; Liu, J.; Gu, Y.; Hu, L. FSELM: Fusion semi-supervised extreme learning machine for indoor localization with Wi-Fi and Bluetooth fingerprints. Soft Comput. 2018, 22, 3621–3635.
21. Yoo, J.; Johansson, K.H. Semi-supervised learning for mobile robot localization using wireless signal strengths. In Proceedings of the 2017 International Conference on Indoor Positioning and Indoor Navigation, Sapporo, Japan, 18–21 September 2017; pp. 1–8.
22. Pan, J.J.; Pan, S.J.; Yin, J.; Ni, L.M.; Yang, Q. Tracking mobile users in wireless networks via semi-supervised colocalization. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 34, 587–600.
23. Chapelle, O.; Vapnik, V.; Weston, J. Transductive Inference for Estimating Values of Functions. In Proceedings of the Neural Information Processing Systems, Denver, CO, USA, 29 November–4 December 1999; Volume 12, pp. 421–427.
24. Belkin, M.; Niyogi, P. Semi-supervised learning on Riemannian manifolds. Mach. Learn. 2004, 56, 209–239.
25. Tran, D.A.; Zhang, T. Fingerprint-based location tracking with Hodrick-Prescott filtering. In Proceedings of the IFIP Wireless and Mobile Networking Conference, Vilamoura, Portugal, 20–22 May 2014; pp. 1–8.
26. Ravn, M.O.; Uhlig, H. On adjusting the Hodrick-Prescott filter for the frequency of observations. Rev. Econ. Stat. 2002, 84, 371–376.
27. Kohavi, R. A study of cross-validation and bootstrap for accuracy estimation and model selection. In Proceedings of the IJCAI, Montreal, QC, Canada, 20–25 August 1995; Volume 14, pp. 1137–1145.
28. Yoo, J.; Kim, H.J. Online estimation using semi-supervised least square svr. In Proceedings of the IEEE International Conference on Systems, Man, and Cybernetics, San Diego, CA, USA, 5–8 October 2014; pp. 1624–1629.
29. Chapelle, O. Training a support vector machine in the primal. Neural Comput. 2007, 19, 1155–1178.
30. Chapelle, O.; Zien, A. Semi-Supervised Classification by Low Density Separation. In Proceedings of the AISTATS, Bridgetown, Barbados, 6–8 January 2005; pp. 57–64.
31. Gao, Z.; Mu, D.; Zhong, Y.; Gu, C. Constrained Unscented Particle Filter for SINS/GNSS/ADS Integrated Airship Navigation in the Presence of Wind Field Disturbance. Sensors 2019, 19, 471.
32. Dampf, J.; Frankl, K.; Pany, T. Optimal particle filter weight for bayesian direct position estimation in a gnss receiver. Sensors 2018, 18, 2736.
33. Gao, W.; Wang, W.; Zhu, H.; Huang, G.; Wu, D.; Du, Z. Robust Radiation Sources Localization Based on the Peak Suppressed Particle Filter for Mixed Multi-Modal Environments. Sensors 2018, 18, 3784.
Department of Electrical, Electronic and Control Engineering, Hankyong National University, Anseoung 17579, Korea
† This paper is an extended version of our paper published in Yoo, J.; Johansson, K.H. Semi-supervised learning for mobile robot localization using wireless signal strengths. In Proceedings of the 2017 International Conference on Indoor Positioning and Indoor Navigation (IPIN), Sapporo, Japan, 18–21 September 2017.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
© 2019. This work is licensed under https://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Abstract
Machine learning-based indoor localization used to suffer from the collection, construction, and maintenance of labeled training databases for practical implementation. Semi-supervised learning methods have been developed as efficient indoor localization methods to reduce use of labeled training data. To boost the efficiency and the accuracy of indoor localization, this paper proposes a new time-series semi-supervised learning algorithm. The key aspect of the developed method, which distinguishes it from conventional semi-supervised algorithms, is the use of unlabeled data. The learning algorithm finds spatio-temporal relationships in the unlabeled data, and pseudolabels are generated to compensate for the lack of labeled training data. In the next step, another balancing-optimization learning algorithm learns a positioning model. The proposed method is evaluated for estimating the location of a smartphone user by using a Wi-Fi received signal strength indicator (RSSI) measurement. The experimental results show that the developed learning algorithm outperforms some existing semi-supervised algorithms according to the variation of the number of training data and access points. Also, the proposed method is discussed in terms of why it gives better performance, by the analysis of the impact of the learning parameters. Moreover, the extended localization scheme in conjunction with a particle filter is executed to include additional information, such as a floor plan.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer