DeepPos: Deep Supervised AutoencoderNetwork for CSI Based Indoor Localization

DeepPos: Deep Supervised Autoencoder
Network for CSI Based Indoor Localization

Peyman Yazdanian and Vahid Pourahmadi
Electrical Engineering Department, Amirkabir University of Technology, Tehran, Iran
Abstract

The widespread of mobile devices facilitated emergence of many new applications and services. Among them are location based services (LBS) that provide services based on user’s location. Several techniques have been presented to enable LBS even in indoor environments where Global Positioning System (GPS) has low localization accuracy. These methods use some environment measurements (like Channel State Information (CSI) or Received Signal Strength (RSS)) for user localization. In this paper, we will use CSI and a novel deep learning algorithm to design a robust and efficient system for indoor localization. More precisely, we use supervised autoencoder (SAE) to model the environment using the data collected during the training phase. Then, during the testing phase, we use the trained model and estimate the coordinates of the unknown point by checking different possible labels. Unlike the previous fingerprinting approaches, in this work we do not store the CSI/RSS of fingerprints and instead we model the environment only with a single SAE. The performance of the proposed scheme is then evaluated in two indoor environments and compared with that of similar approaches.

Indoor Localization, Deep Supervised Auto Encoder, Received Signal Strength, Channel State Information
thanks: Corresponding Author: v.pourahmadi@aut.ac.ir.

I Introduction

Wireless technology today has a much wider meaning than the past. Apart from its main application, sending and receiving information, it can be used to facilitate emergence of new applications in various fields [1]. In closed environments where GPS satellites are not available, several wireless approaches have been proposed for tasks such as user indoor positioning and navigation [1].

In wireless communications, channel state information (CSI) describes how a signal propagates from a transmitter to a receiver and represents the combined effect of environment such as scattering, fading, and power decay with distance.

Most previous WiFi based localization approaches, mainly used Received Signal Strength (RSS), a parameter that represents the signal strength received from the Access Point (AP). Although collecting RSS and implementing algorithms based on it is very easy due to its availability and low hardware requirements, unfortunately due to multi-path effects and shadowing, it is not a stable measurement for localization, and its values may change over time irregularly. In addition, RSS values only provide coarse channel information and does not exploit the information existing in many subcarriers of the orthogonal frequency-devision multiplexing (OFDM) system that have been used as transmission technology of recent wireless standards [2].

As an alternative we can use CSI, which is more robust and due to the description of wireless channel in all of its subcarriers, contains more fine-grained characteristics [2]. In addition, the CSI values are different for each transmitting-receiving antenna pair, so in cases that the AP or user has multiple antennas, we will have more that one set of CSI measurements at each location which can be useful for tasks such as indoor positioning [3].

Indoor localization approaches are mainly categorized in four major methods which will be explained in next section in detail, but among them fingerprinting based localization schemes is of great popularity. Fingerprinting based approaches usually have two phases [4]: a) Off-line phase or training stage, which try to construct a database/model with survey data collected at the environment that the localization wanted to be performed afterwards. In this stage, machine learning methods can be applied to determine a compact model for the system instead of storing all the measurements. Several algorithms including recent deep learning method can be used in this stage. b) On-line phase or testing stage, where real time data collected by a user and transmitted to a server. The server then somehow look into all training data and find the most closely matched one(s) and using the selected point(s) suggests an estimation for the coordinates of the user.

Another CSI-based fingerprinting approach is DeepFi [2], which proposes to use CSI measurements for each of the training points and model their statistics with a Deep Belief Network (DBN). In other words, in this method, instead of saving the raw CSI measurements at each training point, it stores the weights of its corresponding DBN. In on-line stage, CSI data of the test point collected and its statistics is compared with all sample points using the fingerprints of the trained models, and the closest matches (with best input CSI reconstruction) are reported as the candidates for target location, and final coordinates is obtained with a probabilistic method. Despite high accuracy, due to the necessity of saving and running all the models/networks, the complexity of the method in both off-line and on-line phases is high and it may be not efficient (specially in environments with large number of training points).

In this paper, we present DeepPos, an indoor localization scheme which is mainly categorized as a fingerprinting scheme, but unlike common other approaches in this group, we do not store fingerprint information for each training points in our database. Instead DeepPos constructs and keeps only one model that represents the whole environment. It is probably the main distinctive feature that separates the proposed scheme from previously known localization methods. To build such model we have used supervised autoencoder (SAE) architecture that has been recently suggested in machine learning community. DeepPos works based on CSI information and it mainly has two stages:

More specifically, in the training phase, we train a supervised autoencoder (SAE) using the raw CSI data that has been collected at a few sample points (SPs) from the environment . Supervised training happens by providing the SP’s ID as the input of the middle layer of the SAE, (more on that in the following sections). After training, the resulted parameters of the network (e.g. its weights and biases) are stored as the environment model.

Afterwards, in the on-line (testing) phase, raw CSI information from the test point is collected and fed as the input of the SAE network generated in the previous stage. In order to estimate the coordinates of the unknown point, the SAE tries reconstructing the input CSI by changing its assumption on the point’s possible label and calculate the square error between the reconstructed CSI and the input raw CSI for each label. The resulted values are then used in final coordinate estimation of the test point.

The rest of this paper is organized as follows. In Section II, we review the related works in the field of indoor localization. This section is followed by the background information and the preliminaries of our work. Section IV presents the network structure of the DeepPos system. Implementation of DeepPos and experimental evaluations are presented in Section V. Finally, section VI concludes the paper.

Ii Related Work

There is a long history of studies in the field of indoor localization and generally they can grouped into four categories based on Ranging, angle of arrival (AoA), time of arrival (ToA) and fingerprints schemes.

Ii-a Ranging-based localization

Power Ranging-based approaches use at least three access points and the log-distance path loss (LDPL) model and try to estimate location of a node based on received signal strength (RSS) [5]. Different approaches have been considered to improve the accuracy of RSS-based methods, mainly by training the LPDL model’s parameters [5]. EZ [6] presents a no pre-deployment localization schema and uses genetic algorithm to solve RSS-distance identities [2]. In another group, range estimation has been carried out using channel state information (CSI) instead of RSS. For instance, FILA [7] first uses CSI in different time slots to train the parameters of LPDL model and make it more dynamic. This work also presents some ideas to find the relation between the effective CSI (defined in [7]) and distances between node and AP. The trained model is then used for user location estimation.

Ii-B AoA-based localization

By emergence of WiFi APs with multiple antennas (to support MIMO communications), antenna array based localization techniques, gained lots of interest as well [8]. Indoor localization based on angle-of-arrival (AOA), utilizes multiple antennas to calculate the AoAs of the multipath signals received at each AP. These methods first find the AoA of the direct path to the target, and then apply triangulation to localize the object [8, 9, 10, 11, 12, 13, 14]. The main challenge of such techniques is how to improve the resolution of the antenna array [2]. SpotFi technique [8], incorporates super-resolution algorithms that can accurately compute the AoAs of multipath components even when the AP has only three antennas. In this technique there are several APs in the environment, and for each AP, it first estimates the AoAs of all multi-paths in the environment and then finds the AoA of direct path to the target. SpotFi then localizes unknown point using the data it gathered from all APs.

Ii-C Fingerprinting-based Localization

Fingerprinting-based approaches usually require a training phase to survey the floor plan and a testing phase to search for the best matched fingerprints for location estimation [2]. These fingerprints can be built by WiFi [15], RFID [16] and Bluetooth technologies. Among different methods and due to availability of WiFi APs in most environments, WiFi based schemes are more common than others. Most previous WiFi based works was mostly done with RSSI. RADAR [4] is first deterministic work based on WiFi which tries to build fingerprints of RSS using one or more APs and applies K-Nearest Neighborhood (KNN) for position estimation [2]. Horus [15] is another RSS based scheme, where the RSS from an AP is modeled as a random variable over time and space [2]. It identifies different causes for the wireless channel variations and uses probabilistic techniques to achieve its high accuracy.

RSS based fingerprinting has two main shortcomings [2]: First, Due to the multipath effects and shadowing, RSSI is not even fixed for a single point, and they change over time. Second, its values only provide coarse channel information. Instead of RSS, channel impulse response of WiFi is considered as a location-related parameter and contains rich information of wireless channels [3]. Similar to ranging based methods, there have been recent studies on application of CSI information (instead of RSS) in fingerprinting application too. For instance, FIFS [1], presents an idea to find the relation between the effective (defined in [7]) and the distances of the node and the AP. It also proposes a probabilistic algorithm based on coherence bandwidth which enhances the accuracy of mapping the test point to the fingerprints. The shortcoming of this model is that it needs a large amount of calibration points to build the fingerprint database [2].

DeepFi [2] is another powerful CSI-based fingerprinting system. This method first collects CSI at some locations as training spots and for each of the them, constructs a deep neural network to model the statistics of that point and store all the models of all training spots as its database. For localization, CSI data is collected at a test point and its statistics is compared with the models which have been constructed in the training phase for each point and the closest matches are then selected as the candidates for the test point. Although it has high accuracy, the issue of this scheme is its very high complexity in both off-line and on-line phases to construct and check all models. It is also necessary to save parameters of all trained networks for each sample point, which may be not that efficient, specially when the number of training points increases (to have higher localization accuracy).

Iii Background Theory And Preliminaries

Here we first bring an overview of modulation technique in WLAN and its channel status information (CSI). Then we bring a section that briefly reviews basics of deep neural networks specially AEs.

Iii-a Channel State Information in OFDM

Orthogonal Frequency Division Multiplexing (OFDM) technique has been widely used in different technologies like IEEE802.11a/g/n, WiMAX, LTE standards to access the high data rate. In this technique, first, in the transmitter, the input data stream is broken into several streams each of which transmitted over non-overlapped subcarriers after they modulated with a certain modulation scheme. The data then converted back to the time domain by an inverse Fast Fourier Transform (IFFT). The data is then serialized and converted to analog signal and sent through the channel. At the receiver, after returning to the digital domain by Fast Fourier Transform (FFT), the signal is equalized (the effect of channel is removed) in the frequency domain. Demodulation and decoding are next steps in the receiver chain [1].

In WiFi Network Interface cards (NIC) usually all the above steps are happening inside of the chipset and are not accessible from outside. Thanks to an open source Linux kernel [17], and with few NICs, such as Intel IWL 5300, it is now possible to access the CSI from outside of the NIC and record the channel variation experienced during transmission of one OFDM sample over all subcarriers. Without CSI. As for notation, let and denote the transmitted and received signal at each subcarrier. We have

(1)

where denote the number of subcarrier, is the additive white Gaussian noise and represents the channel gain at the subcarrier.

The WiFi channel at the 2.4 GHz band can be considered as collection of 56 narrowband sub-channels. From these 56 subcarriers, the Intel WiFi Link 5300 NIC driver implementation is able to report back the CSI of 30 subcarriers. Its reported channel frequency response is a complex value fro each subcarrier, which is the real and imaginary part of the :

(2)

In this paper, the proposed DeepPos framework is based on the CSI of 30 subcarriers, which can reveal much richer channel properties than RSSI.

Iii-B Deep Learning Approaches

Application of machine Learning techniques in different problems is growing very fast due to their vast capability for solving real problems. Availability of massive amounts of computational power and emergence of some methods for solving the vanishing gradient problem of neural networks, make it practical to use large neural networks, and in particular deep networks [18]. Autoencoder (AE) is an important class of deep learning networks which usually are categorized as unsupervised learning algorithm and its main objective is to train network such that the target values (at the output nodes) be equal to the inputs.

A deep AE is composed of two symmetrical deep-belief networks that typically have four or five fully connected layers representing the encoding half of the net, and second set of four or five layers that make up the decoding half. Figure 1 shows a simplified schema of a deep AE’s structure. As an unsupervised learning algorithm, AE determines the weights of the network based on back propagation such that it minimizes the loss function of , where and are the output and input of the AE respectively.

An AE in essence is trying to learn an approximation to the identity function , so as to output is equal to input [19]. The hidden layers are Restricted Boltzmann Machines (RBM), which is the building blocks of deep-belief networks [20]. The middle layer of the AE (which usually has the fewest number of neurons) generally can be considered as the low order representation of the input vector.

Fig. 1: Structure of a Deep AE

Iv DeepPos System

DeepPos has two major phases, off-line training stage, and on-line location estimation that we will discuss them in the following subsections. First, in training phase, we model the environment via a deep neural network in the form of SAE. Then, in on-line stage, using this trained model, we select a few candidate points which will be then used to estimate the actual coordinates of the test point.

Iv-a DeepPos Architecture

Supervised autoencoder (SAE) which is a new architecture in deep learning community, is the core of our scheme. As mentioned in previous section, the main distinctive feature of DeepPos is training only one model for the whole environment and so there is no need to store many models or fingerprints information for each SP.

Additionally, in previous deep learning models like DeepFi [2], the individual Deep Belief Networks (constructed for each SP) is designed in an unsupervised manner (it does not have any information about the location of the collected CSI). Figure 2 shows the overall architecture of the DeepFi system. In contrast, DeepPos transforming raw input CSI to a lower dimension in a supervised manner, i.e. it tries to extract features that not only they are good for input reconstruction (like autoencoder) but also these features should keep information related to the location of the collected CSIs. Due to such structure (that has both lower dimension and the location information), it is possible to have only one single model for the whole environment.

Figure 3 depicts the overall architecture of DeepPos. As can be seen, DeepPos mainly have 3 parts: first, the encoding layers acts as a feature extractor. Then, the extracted features are concatenated with an one-hot vector representing the label for the training data. Finally, the combined vector is used as the input of the decoding layers which aims to reconstruct the original CSI vector. Detail explanation of each part is presented in section IV-B

Fig. 2: DeepFi Main Architecture
Fig. 3: DeepPos Main Architecture

Iv-B Training Stage With Deep Learning

In this phase, raw CSI data are collected at some SPs of the environment, each contains CSI measurements of the 30 OFDM subcarriers of 3 antennas. The diversity of data makes it possible for the system to learn more about the environment and to produce a more comprehensive model. In this work we only consider the amplitude responses of collected CSIs and the values are normalized between (0,1).

The collected data is used to train a a deep supervised autoencoder which aims to reconstruct the input CSI values. Figure 4 shows overall architecture of DeepPos training stage. A deep SAE network with four hidden layers is used to train and fit a model. In order to reduce the dimension of raw CSI data, we assume that the number of neurons in hidden layers are gradually decreasing. We denote the number of neurons in the first to the fourth layer by where .

Fig. 4: DeepPos Training Stage Schema

The pseudocode for the training stage is presented as in algorithm 1. To model the scenario with only one deep neural network, in each training iteration, after encoding of input data from each of the CSI measurements (lines 6-9), we append its corresponding label as a one-hot vector to the output of the middle layer of the network (lines 10-11). The resulted vector is then fed to the decoding layers to reconstruct the input CSI data (lines 12-17). The aim is to find network weights minimizing the reconstruction error.

We define as hidden variable in encoding layer , for = 1,2,3,4, and let denote the measurement vector of the SP. vector is of size 90, corresponding to 30 subcarriers measurement at 3 antennas, collected at a training spot. At each step, gets in one of the measured vectors, , as the network input. Furthermore, as Sigmoid function is used as the activation function of each layer and so we have:

(3)

and are the weights and biases matrix for encoding layer respectively.

After the encoding operation, in layer 4, we append with appropriate label as an one-hot vector. For example, if the input CSI was the CSI measurement of the SP (from total of SPs), we append an one-hot vector corresponding to (A vector of size with all zeros except at location which has one). So can be defined as:

(4)

Where is the encoded representation of and is the label one-hot vector for that packet. Note that the main difference between the common autoencoder and the supervised autoencoder is in addition of the labels at the middle layer of the network.

Finally, for the decoding stage we have:

(5)

Where and are the weights and bias matrices for decoding layer , respectively. Like common autoencoders, the goal is that the output of the network () be the best possible reconstruction of the input CSI vector. So, at it each epoc, the network weights are updated such that they minimize the cost function defined as:

(6)

Where is the output of the network for the input of . In this work we have used RMSprop optimizer as the deep network optimization technique. The resulted network guaranties that for each input CSI measurement of a SP, it gets the lowest reconstruction error for a label that matches the point’s ID.

Input :  measurements [] each with 90 CSI values collected and stored for each of the training spots
Output : Model parameters (Trained Network’s weights and biases)
1 Initialize weights of all layers , , , = randn // randn is standard Gaussian distribution function
2 for  : max-epoch do
3       for  :  do
4             for  :  do
5                   = the measurment vector of the SP ;
6                   // encoding
7                   for  :  do
8                         = sigmoid (. + )
9                        
10                   end for
11                  // append appropriate label for each input vector
12                   = (,);
13                   // decoding
14                   = ;
15                   for  :  do
16                         = sigmoid (. + )
17                        
18                   end for
19                  Obtain the error between input data and reconstructed data ;
20                  
21             end for
22            
23       end for
24      //fine-tuning;
25       Update network weight and bias based on the resulted errors with back-propagation with Adam optimaizer;
26 end for
Algorithm 1 Off-line Training Stage

Iv-C On-line Location Estimation With Consider KNN Points

The procedure of the DeepPos on-line stage is presented in algorithm 2. First, DeepPos takes a few CSI measurements at the test point and feed them to the trained SAE model. Note that for SAE to reconstruct the CSI, it needs to have a label (in the form of one-hot vector) input at its middle layer, but for the test location we do not have that label (since we do not know the location of the test point). The way that DeepPos proceed is to loop over all possibilities of one-hot vectors (corresponding to different SPs) and for each case, it calculates the mean square error (MSE) of the reconstructed CSI for the collected measurements. The location of the test point is then calculated using the weighted sum of the SP’s coordinates corresponding to the lowest MSEs.

In algorithm 2, the encoding layers construct the coded features of the unknown point’s CSI raw data (lines 1-7). Then, for each of the SPs, its corresponding one-hot vector is appended to this coded features. The decoding layers generate the reconstructed vector of the input (lines 7-16), and the mean square of the reconstruction error is evaluated. To have a more robust estimation, the MSE calculation is performed for CSI measurements (at the test point), and averaged over all of them. Labels that lead to a lower reconstruction MSEs are the candidate points from the network’s perspective for the test spot (lines 17-21). Finally, the test point’s coordinates are evaluated as the weighted mean of these candidate points (line 23). Figure 5 shows overall schema of DeepPos on-line stage.

Fig. 5: DeepPos On-line Location Estimation Stage Schema
Input :  measurements each with 90 CSI values collected at the test point,
Trained SAE model
Output : Estimated test point’s coordinates;
1 // encoding
2 for  :  do
3       = measurment vector of the test point ;
4       for  :  do
5             = sigmoid (. + )
6            
7       end for
8      
9 end for
10// append different labels to low order representation vector (), and find the reconstruction of the input data
11 for  :  do
12       for  :  do
13             = (,);
14             // decoding
15             = ;
16             for  :  do
17                   = sigmoid (. + )
18                  
19             end for
20            
21            
22       end for
23      
24      
25 end for
26// Find R of the least s
27 // Estimation coordinates of unknown point
28 = ;

Algorithm 2 On-line Location Estimation Stage

V Experimental Evaluation

We implemented DeepPos with 2.4 GHz commodity WiFi Intel 5300 NIC and carry out different experiments to validate its performance. We have used a Dell laptop equipped with an Intel 5300 NIC as the access point, and a Huawei handset as the mobile device. We have used the modified Linux driver developed by [17] to measure and store CSI information into a file. These files are sent to the server (which is a linux machine with 8 GB of memory and Intel® Core™ i5-3210M processor), and the CSI data for each packet from each training points are stored in a PostgreSQL database. Note that if the training SPs are too sparse, it may result in low localization accuracy in testing phase as there are not enough data to model. On the other hand, if we choose dense training spots, it will cost a lot of efforts on pre-training data collection. Based on our experiments, the distance between two spots is set to about 1m, which can maintain the balance between localization accuracy and pre-processing cost.

V-a Experiment Configuration

We verify the performance of DeepPos in two indoor scenarios. In each case, we collect data from a few SPs and at each SP we measure 30 packets continuously each of them contains 90 raw CSI data (30 subcarriers in 3 antennas). All the measurements are collected at a same height, which construct a 2D platform. First one is an empty classroom where there is a LoS path between the AP and mobile device. Second location is a room and corridor around, which is more challenging scenario as there are some points without LoS path. We evaluate overall system performance in both settings by using one leave-out cross-validation test. The experimental results for both scenarios are presented in the following.

V-A1 Class Room in the University

Figure 6 shows the layout of our first test scenario. In this 67 classroom, the AP is located on the table at one end of the room, and all the training and test samples are collected at almost same height. As shown by red dots in figure 6, SP positions are chosen uniformly scattered in the room. We collect 30 packets for each of the 19 SP. We choose deep network’s parameters as and .

Fig. 6: Classroom Layout
Fig. 7: Hall and Corridor Layout

V-A2 Empty Hall and corridor around

The hall and its entrance corridor is the second scenario in which there is some points are not in the LOS of the AP. The layout of this location is presented in figure 7 with dimensions of almost 57 . As shown in figure 7, The layout of this location .The SAE structure is similar to the first scenario and and .

V-B System Performance With Cross-Validation Test

To test the overall functionality of the system, leave one out cross-validation (LOOCV) test is applied. Specially when we have limited number of samples, LOOCV often gives us more accurate results compared with simple partitioning of data into train and test sets. More precisely, in LOOCV, at each step, the data of one SP is set aside (for testing) and the SAE is trained with the data of remaining SPs. Then in the on-line stage, we randomly select measurement data from the test SP and feed that into the the trained model. The model then reports back 2 (or 3) best candidate SPs that has less reconstruction MSE when we use their corresponding one-hot vector in the SAE. The final estimated location of the test point is then reported back as the weighted mean of these candidate spots.

The Euclidean distance between the estimated coordinates and the actual one of the test point is defined as the localization error for that SP. Denoting the estimated coordinates for the SP by and the actual position by we will have . The mean of all localization errors for all SPs is what we reported as the performance of the system.

To see the performance of the DeepPos, we compared its localization error with that of DeepFi [2]. As a quick reminder, DeepFi was one of the most recent indoor positioning scheme that uses deep learning method but it constructs a separate deep model for each of the SPs; thus, each model represents one of them and we need to have N deep models. Unlike DeepPos, where we only have a single deep network which represents the whole environment (not a single SP).

Figure 8 shows the CDF of localization errors when we use 2NN DeepPos and DeepFi in the first environment (classroom scenario). DeepPos has mean and median localization error in this case is about 1.87m and 1.57m respectively; while they are 1.91m and 1.38m for DeepFi respectively. In this scenario, DeepPos has 80% of localization errors under about 2m.

Fig. 8: CDF of localization error in Classroom

Similar experiment has been conducted for evaluating DeepPos performance in the second environment. Figure 9 shows CDF of distance errors in this place for 3NN DeepPos and DeepFi. In this case, DeepPos and DeepFi have the mean error of 1.82 and 1.81, respectively. For DeepPos, 80% of the localization error is less than 2m.

Fig. 9: CDF of localization error in Hall and Corridor

Table I summarizes the overall performance of DeepPos and DeepFi in both scenarios. This results shows that two systems have close localization performance, however, different from DeepFi, the proposed DeepPos system does not require to train a deep neural network for every training spots, which makes the model very complex. Instead, by using a single deep supervised autoencoder, we model the whole environment. Table I also presents the overall average times spent for LOOCV test of two systems both exceeded on the same environment and with the same dataset, which strongly shows simplicity of the proposed method compared to DeepFi.

                           Classroom

Method Mean error (m) Std. error (m) Average Execution Time (s)
DeepPos 1.872 1.331 1.1449

DeepFi
1.915 1.295 3.7072
                           Hall and Corridor
Method Mean error (m) Std. error (m) Average Execution Time (s)
DeepPos 1.824 1.240 1.053

DeepFi
1.815 1.287 3.457
TABLE I:
Compare of Mean, Std. of error and Exe. time (s)

V-C Number Of Sample Points And Overhead Analysis

In this section, we study the complexity of the localization framework in both off-line and on-line phases.

V-C1 Evaluating Training Overhead

In this part, we evaluate the training stage complexity and overhead in terms of total time required to train a model and memory usage as the number of SPs increases. To get an average time/RAM usage/accuracy, in each case, we have selected all possible combinations for training SPs and averaged the results over all cases. For example, if we want to report the average training time when we have SPs, we select all possible combinations of SPs from all SPs (in that scenario) and report the average training time of all cases.

Note that, as DeepFi constructs an individual model for each of the SPs, we can assume that these models are trained sequentially which keep the memory usage of that scheme in a normal range but the training time increases linearly as the number of SPs grows. With this assumption, figures 10 and 11 shows the results for both DeepPos and DeepFi. As can be seen, DeepPos needs much less training time (specially when number of SPs increased), while they have almost similar memory usage.

Another possible assumption for DeepFi is to train SPs’ models in parallel. The average training time and memory usage for this case is depicted in figures 12 and 13. As expected, in this case the training time of both schemes are getting closer together; however, the memory usage of DeepPos is much better than DeepFi (due to the need for training of one versus several models). Note that the extra training time of the DeepFi is probably because it uses more memory, and due to memory limitation of our computer (8GB), it gets some penalty due to virtual memory speed and heavier CPU utilization; Otherwise, DeepFi should have similar training time as DeepPos and the superiority of DeepPos in this mode is mainly on lower memory usage during training stage.

a) Classroom
b) Hall and Corridor
Fig. 10: Training stage time spent for sequential test
a) Classroom
b) Hall and Corridor
Fig. 11: Training stage memory usage for sequential test
a) Classroom
b) Hall and Corridor
Fig. 12: Training stage time spent for parallel test
a) Classroom
b) Hall and Corridor
Fig. 13: Training stage memory usage for parallel test

V-C2 Evaluating Testing Overhead

Increasing the number of SPs can boost the accuracy of the system. The point to consider though is by the extra training SPs, the model should not face much timing overhead, specially in testing stage.

To study this behavior for both DeepFi and DeepPos systems, we conducted an experiment that we increased the number of SPs and for each case calculated the average (as mentioned in previous section) localization accuracy and also measure the execution time of the testing phase.

The results are depicted in figures 14 and 15. Comparing the results, it can be verified that by addition of new SPs, the overall precision of both schemes increases almost similarly; however, DeepPos outperforms DeepFi in terms of prediction time.

a) Classroom
b) Hall and Corridor
Fig. 14: Mean of localization error
a) Classroom
b) Hall and Corridor
Fig. 15: Average time of localization stage

Vi Conclusion and future works

In this paper, we proposed DeepPos, a deep supervised autoencoder based system for indoor localization. DeepPos uses 2.4GHz WiFi fingerprinting measurements for training the model, but unlike similar approaches in this method, fingerprints are not stored for each of the training points. Alternatively, we fit only one model to the environment that uses deep SAE structure (the encoding layers act as a feature extractor, user locations are labels which are appended to this extracted features, and decoding layers play the role of a supervised neural network). The coordinates of a test location is then determined by finding the labels of the SAE that results in lowest CSI regeneration error. The performance of the proposed scheme is evaluated in two indoor environments and we have shown it keeps the high accuracy while it has much less complexity compared to other deep learning based indoor positioning methods.

References

  • [1] J. Xiao, K. Wu, Y. Yi, and L. M. Ni, “Fifs: Fine-grained indoor fingerprinting system,” in 2012 21st International Conference on Computer Communications and Networks (ICCCN), July 2012, pp. 1–7.
  • [2] X. Wang, L. Gao, S. Mao, and S. Pandey, “Csi-based fingerprinting for indoor localization: A deep learning approach,” IEEE Transactions on Vehicular Technology, vol. 66, no. 1, pp. 763–776, Jan 2017.
  • [3] X. Wang, L. Gao, and S. Mao, “Phasefi: Phase fingerprinting for indoor localization with a deep learning approach,” in 2015 IEEE Global Communications Conference (GLOBECOM), Dec 2015, pp. 1–6.
  • [4] P. Bahl and V. N. Padmanabhan, “Radar: an in-building rf-based user location and tracking system,” in Proceedings IEEE INFOCOM 2000. Conference on Computer Communications. Nineteenth Annual Joint Conference of the IEEE Computer and Communications Societies (Cat. No.00CH37064), vol. 2, 2000, pp. 775–784 vol.2.
  • [5] H. Lim, L.-C. Kung, J. C. Hou, and H. Luo, “Zero-configuration indoor localization over ieee 802.11 wireless infrastructure,” Wirel. Netw., vol. 16, no. 2, pp. 405–420, Feb. 2010.
  • [6] K. Chintalapudi, A. Padmanabha Iyer, and V. N. Padmanabhan, “Indoor localization without the pain,” in Proceedings of the Sixteenth Annual International Conference on Mobile Computing and Networking, ser. MobiCom ’10.   New York, NY, USA: ACM, 2010, pp. 173–184.
  • [7] K. Wu, J. Xiao, Y. Yi, M. Gao, and L. M. Ni, “Fila: Fine-grained indoor localization,” in 2012 Proceedings IEEE INFOCOM, March 2012, pp. 2210–2218.
  • [8] M. Kotaru, K. Joshi, D. Bharadia, and S. Katti, “Spotfi: Decimeter level localization using wifi,” in ACM SIGCOMM Computer Communication Review, vol. 45, no. 4.   ACM, 2015, pp. 269–282.
  • [9] S. Kumar, E. Omar Hussein Hamed, D. Katabi, and L. Erran Li, “Lte radio analytics made easy and accessible,” vol. 44, 08 2014.
  • [10] J. Gjengset, J. Xiong, G. McPhillips, and K. Jamieson, “Phaser: Enabling phased array signal processing on commodity wi-fi access points,” vol. 19, pp. 6–9, 08 2015.
  • [11] S. Kumar, S. Gil, D. Katabi, and D. Rus, “Accurate indoor localization with zero start-up cost,” 09 2014.
  • [12] J. Xiong and K. Jamieson, “Arraytrack: A fine-grained indoor location system,” pp. 71–84, 04 2013.
  • [13] S. Sen, J. Lee, K.-H. Kim, and P. Congdon, “Avoiding multipath to revive inbuilding wifi localization,” in MobiSys, 2013.
  • [14] D. Niculescu and B. Nath, “Vor base stations for indoor 802.11 positioning,” pp. 58–69, 01 2004.
  • [15] M. A. Youssef and A. Agrawala, “The horus wlan location determination system,” in In Communication Networks and Distributed Systems Modeling and Simulation Conference, 2005, pp. 205–218.
  • [16] L. M. Ni, Y. Liu, Y. C. Lau, and A. P. Patil, “Landmarc: indoor location sensing using active rfid,” in Proceedings of the First IEEE International Conference on Pervasive Computing and Communications, 2003. (PerCom 2003)., March 2003, pp. 407–415.
  • [17] D. Halperin, W. Hu, A. Sheth, and D. Wetherall, “Tool release: Gathering 802.11n traces with channel state information,” ACM SIGCOMM CCR, vol. 41, no. 1, p. 53, Jan. 2011.
  • [18] R. D. Hof, “Is Artificial Intelligence Finally Coming into Its Own? MIT Technology Review,” Available: https://www.technologyreview.com/s/513696/deep-learning/, 29-Mar-2016, [Online; Accessed: 25-Mar-2018].
  • [19] “Autoencoders, Unsupervised Feature Learning and Deep Learning Tutorial,” Available: http://ufldl.stanford.edu/tutorial/unsupervised/Autoencoders/, [Online; Accessed 25-Mar-2018].
  • [20] C. V. Nicholson and A. Gibson, “A Beginner’s Tutorial for Restricted Boltzmann Machines,” Available: https://deeplearning4j.org/restrictedboltzmannmachine, [Online; Accessed 25-Mar-2018].
Comments 0
Request Comment
You are adding the first comment!
How to quickly get a good reply:
  • Give credit where it’s due by listing out the positive aspects of a paper before getting into which changes should be made.
  • Be specific in your critique, and provide supporting evidence with appropriate references to substantiate general statements.
  • Your comment should inspire ideas to flow and help the author improves the paper.

The better we are at sharing our knowledge with each other, the faster we move forward.
""
The feedback must be of minimum 40 characters and the title a minimum of 5 characters
   
Add comment
Cancel
Loading ...
321563
This is a comment super asjknd jkasnjk adsnkj
Upvote
Downvote
""
The feedback must be of minumum 40 characters
The feedback must be of minumum 40 characters
Submit
Cancel

You are asking your first question!
How to quickly get a good answer:
  • Keep your question short and to the point
  • Check for grammar or spelling errors.
  • Phrase it like a question
Test
Test description