Fast Gaussian Process Occupancy Maps

Fast Gaussian Process Occupancy Maps


In this paper, we demonstrate our work on Gaussian Process Occupancy Mapping (GPOM). We concentrate on the inefficiency of the frame computation of the classical GPOM approaches. In robotics, most of the algorithms are required to run in real time. However, the high cost of computation makes the classical GPOM less useful. In this paper we don’t try to optimize the Gaussian Process itself, instead, we focus on the application. By analyzing the time cost of each step of the algorithm, we find a way that to reduce the cost while maintaining a good performance compared to the general GPOM framework. From our experiments, we can find that our model enables GPOM to run online and achieve a relatively better quality than the classical GPOM.

I Introduction

Mapping is an important research area in the field of robotics. Occupancy grid mapping has been introduced already about 3 decades ago and it has been widely used in the field of robotics for its simplicity and computation efficiency. It is quite often at the heart of a Simultaneous Localization and Mapping (SLAM) system, where it integrates the sensor data (most often (laster) range data) into a map. Occupancy grid mapping is simple in assuming that, upon updating the map from a new scan, the probability of occupation of a cell is solely determined by the ray passing through. This is loosing the spatial context information, because in reality the occupancy of a cell can often be related to its neighboring cells [1].

Gaussian Process Occupancy Mapping (GPOM) is a mapping algorithm which is updating all relevant cells when integrating a scan. GPOM has been developed for about a decade. It has its advantage over general occupancy mapping and some other probability-based maps to better model the sparsely sampled points and its fit for structural information. In [1] the authors build the whole map at a time, which takes quite a long time, because holding all of the range sensor data in one formula takes too much computation. What’s more, the algorithm cannot incrementally integrate new data, which is limiting the utility of GPOM.

Inspired by Bayesian Committee Machine (BCM), [5] clusters laser scans into groups. Instead of building the whole map in one go, as [1], they separate the task into many subproblems. By building the map on each group and fusing them with BCM, the computation cost of each group is tremendously reduced and, more importantly, this approach makes the algorithm incremental. As an improvement, [7] and [8] incrementally update the map in each frame, which allows for exploring in dynamically expanding environments.

After that, a more scalable method, Hilbert Maps [6], has been proposed. Comparing with GPOM, the Hilbert Maps method is a parametric method that can be trained with Stochastic gradient descent (SGD). However, it is less accurate than GPOM and, more importantly, it has to pass all the data multiple times and does not run incrementally.

Inspired by the indoor mapping survey [2], many indoor mapping algorithms are incremental such that it is possible to run them in real time. The authors of this paper were wondering why GPOM has not been applied in real time, even though it can run incrementally already. The answer is, that, even for the current incremental GPOM algorithms, the computing requirements are still intractable.

The Gaussian Process regression (GP) is the source of restrictions. Generally, the cost to build the model is , while the prediction only takes . While GPOM also requires variance, the prediction of GPOM will take .

Most of the work to optimize GP is done on its training process. But the speed limit may not be the same in GPOM application. In order to find the bottleneck that makes GPOM not meet the needs of real-time processing, we analyzed the GPOM pipeline as follows:

  1. Extract positive and negative samples

  2. Set the testing sample

  3. GP model Building

  4. Prediction on the test set to achieve and (see Section II)

  5. Fuse and of the global frame (size ) with the local frame (size )

  6. Build the image with logistic regression function

  7. Publish with ROS

In our experiments we found, that most of the computation cost is not in the third step (model building). This is a surprising result which is contradicting most of the works on GP. Instead, we found that the cost in prediction, step 4, is the highest.

In this paper, we are thus proposing a GPOM algorithm that optimizes the prediction step, which then makes our Fast-GPOM can run at least 10 times faster than GPOM in that step.

Additionally, we observed that GPOM has a design defect. In Fig. (b)b a wide range of space outside GPOM has high a probability of beginning occupied. Out Fast-GPOM can mitigate such problems.

(a) Fast-GPOM
(b) GPOM
(c) Ground Truth
(d) ROC
Fig. 1: Our Fast-GPOM and GPOM on a synthetic map. The color bar shows the probability to be occupied.

This paper is organized as follows: In the Section II, we review the GP and GPOM algorithm to make the further analysis clear. In Section III, by concentrating on the bottleneck, we propose our improved algorithm. After that, Section IV presents experiments to show the properties of our method. We conclude our work in Section V.

Ii Background

Ii-a Gaussian Process

Gaussian Processes (GP) generally have been applied to many tasks, but mainly for regression and classification. By comparing the similarity between features of observation and the inference sample, GP can provide both estimation and uncertainty in one step.

With pairs , the regression function to solve here in GP is


The latent function in the above formula is defined as


We can then obtain


with the assumption that the target data is jointly Gaussian.

With the kernel function , the predicted mean and standard derivation in the above formula should be:

1:, ,
7:return ,
Algorithm 1 GP Regression

To make it more clear how the computation is separated in training and inferencing, we follow [4] to write the algorithm in Algorithm 1. In Algorithm 1, are the observed pairs, is the inference point, and is the noise level.

Lines 1 and 2 can be computed before we have the inference data. Lines 3 to 5 are for deriving the mean and variance. For the Cholesky decomposition in line 1 and solving the triangular system in lines 2 and 4, the time complexity is and , respectively.

Ii-B Gaussian Process Occupancy Maps (GPOM)

There are many GPOM frameworks, but most of them only optimize on the Gaussian Process part, and do not make changes in the pipeline, so the framework is still the same.

The GPOM we will follow is as Algorithm 2. For its input, is the robot pose, is the measurement, is the mean, and is the variance. And the output is the occupancy probability.

1:, , ,
2: Extract Occupied and free sample with and
3:if  then
4:     MAP on GP model
5: Extract inference data with
9:return , ,
Algorithm 2 GPOM for each iteration

We extract the observed points from the range sensor. From odometry (or an integrated SLAM algorithm) we obtain the robot pose of each frame . Using the pose we can update the map using the information from the range sensor. In Fig. (a)a orange stars represent the scanned obstacle points. By sampling points on the laser line from the robot with an interval we generate the free samples (blue stars in Fig. (a)a). It should be noted that we don’t add free samples that are closer than to the occupied point. The second step is to estimate the parameters of the kernel. The first frame is handled specially. Subsequently, we extract the inference data from a window that has the robot pose as the center. This step is just taking many indexes, so it does not take much time. Next, we build the GP model, from 4, 5. For that we have to take the inverse of the kernel, which takes . The prediction should take , because it is required to calculate the covariance with training data.

(a) Positive and Negative samples
(b) Rings
Fig. 2: Extracting samples in GPOM. In the top figure, positive (orange star) and negative (blue star) samples are extracted from a laser range scan. In the bottom image, the space is separated by two rings: the ring around the obstacles and the ring of the last free sampled points. Thus the space is split into three regions: A, B and C.

In the end, BCM is used to fuse the prediction and logistic regression to squash the cell values into .

Iii Our Work

In our work, we first analyze the computation time for each step of the pipeline and then make optimizations on the identified bottleneck.

Iii-a Pipeline Analysis on GPOM

The pipeline of GPOM on each iteration is listed as in Section I.

The parameter estimation is only for the first frame and will not be included in this analysis. What’s more, we generate the probability map on each frame just for demonstration, it is not really necessary for the GPOM algorithm. Publishing the map in ROS is also extra work that is ignored here.

The time cost for all the steps is listed in Table I, as measured for three different maps. Those maps are shown in Fig. 4. We can see that there is a conflict with the usual intuition about which step should be optimized. It is not on the GP model building step, which has theoretically a complexity of . Instead, the prediction part with a complexity of takes the most of the time. This is not a contradiction, because the two n’s represent different variables.

So let’s be more specific about the computation of GPOM and use an example. We assume a scan with laser beams (range measurements) and furthermore assume that we extract 5 points on each beam. In experiments, we choose one out of ten lasers to use. So the number of training samples in each iteration should thus be . The inference data could be from a window with pixels and then the inference number is .

The computation of the sample extraction is linear with the number of beams , with a small constant weight. The inference data extraction is also linear to the window size. The GP model building and prediction will take and while tends to be much greater. After that, BCM and logistic squash are also linear to .

The inference computation is comparatively much bigger in applications like GPOM, because it predicts a local map at each frame. This is not the common application scene with GP regression.

GPOM cost in milliseconds Fast-GPOM cost in milliseconds
sparse_obstacles simple_rooms robocup sparse_obstacles simple_rooms robocup
Extract 4.63 4.98 4.65 10.57 5.76 65.89
Extract 0.56 0.21 0.59 21.72 2.23 11.46
Build GP model 9.45 7.55 7.94 8.02 3.05 5.36
Predict 1607.59 253.34 1821.46 157.97 8.41 52.03
BCM 2.12 0.37 2.14 4.68 0.38 2.12
Logistic 113.00 17.58 113.871 13.90 1.27 7.48
Build Map 1951.92 284.31 1737.20 219.10 21.40 86.02
Publish Map 86.87 14.52 88.72 149.65 14.44 92.68
TABLE I: Time cost on each step in GPOM and Fast-GPOM on three test maps. The time cost of the six steps of the pipeline, the total cost of map building and the cost to publish the map in ROS are demonstrated in this table. The value is an average of the frames after running out of scan data.

Iii-B Model with Heuristic of spatial information

Inspired by the above analysis and Table I, we believe the key to reducing the computation cost is to optimize on the data selection with context-awareness. In this work, we group the space around the robot into three regions.

Spatially, the uncertainty of the laser range sensor has a bound . It has a high probability to indicate that space away from the obstacle at a distance is free. To formulate differently, we can assume that space is free with a small variance. In this paper, we assume that space is free between the robot to away from the measured obstacle.

We don’t observe the space behind the obstacles, so we set the space behind the observed point as unknown. The defect of the classical GPOM is, when it does the model training, in the direction of the laser scan, there should be no observable sample after the occupied sample. Thus we cannot control the value outside the occupied space.

We also assume that if a scan does not provide any useful information on a certain position, the old value should be kept. In Fig. (b)b we can find that for GPOM the space behind the wall has a high probability. But that space has not been observed and thus should not be updated.

First, we first build two rings, as shown in Fig. 2. The outer ring is depicted green, it is the polygon border formed with the occupied samples. The inner ring in red is the polygon border formed by the free samples that are closest to the occupied sample on each ray.

The inner and outer rings separate the space into three regions: A (free space with low variance), B (Uncertain region in between the two rings), C (Unknown space).

The strategy we apply for the corresponding regions is:

  1. Training sample

    1. Star points in inner ring as negative samples

    2. Star points in outer ring as positive samples

  2. Prediction

    1. Directly set A region as free space in local frame

    2. Keep the C region value in local frame

    3. Only regression region B

Then we updated Algorithm 2 as Algorithm 3.

1:, , ,
2: Extract Occupied and free sample with and
3:, Extract Inner ring and outer ring
4: Selected samples on ring
5:if  then
6:     MAP on GP model with
7:, , Set inference data in region A, B, and C
Algorithm 3 Fast-GPOM

Iv Experiments

To highlight the performance of our method, we compare it with the classical GPOM method in synthetic environments.

Iv-a Experiment Environment

Fig. 3: Histogram of the average time cost to build the map for a single scan.

In the experiments, we use a PC with the following configuration: Intel Core i7-6700 3.4GHz and 16GiB memory with Ubuntu 16.04. Our algorithm is implemented based on the Robot Operating System (ROS) and using the Simple Two Dimensional Robot Simulator (STDR) [9] as our simulation environment. The STDR Simulator is a simple, flexible and scalable robot simulator for use within ROS. It provides a variety of different types of synthetic maps and the ground truth of these maps. It is reasonable to evaluate our algorithm and the performance of the maps in a simulation because then the ground truth is available and we only need to consider the noise of the simulated laser scan data.

In STDR, we utilize a simulated Hokuyo-laser scanner, the simulator robot will provide accurate odometry data. The properties of the simulated Hokuyo laser scanner are the same as real laser scanner and Gaussian noise is applied with a mean, 0.5 and standard deviation, 0.05.

In the experiment, we choose three maps that are provided by the STDR Simulator. The names of the maps are sparse_obstacles, simple_rooms and robocup. The sparse_obstacles map is an indoor scene with several obstacles, the map size is 775 746 and the resolution of occupancy map is 0.02m. The simple_rooms map is a basic indoor environment with several empty rooms. This is a relatively simple scenario, and the map size is 600 600 and the resolution of occupancy map is 0.05m. And the robocup map is a hand-draw narrow corridor scenario seem like the environment of robocup rescue league, the map size is 769 729 and the resolution of occupancy map is 0.02. On each map we record laser scan data and odometry data as a dataset. Then we compare with our algorithm with the general GPOM algorithm using these datasets.

The Gaussian process model is from pyGPs [10], which is a widely used library for Gaussian process in python.

For the implementation of the algorithms, GPOM and Fast-GPOM share the same setting, with the same , , , , and Matern (in pyGPs, ) kernel. In the squashing step we use , . The resolution of the map is determined to be the same as the arena dataset.

The code, dataset, environment setting and the ROS package of the Fast-GPOM are available on Github1.

Iv-B Comparison

Fig. 4: Map building with GPOM and Fast-GPOM. We show three rows, corresponding to the three maps: simple_room, sparse_obstacles and robocup. The first column is for GPOM, the second for our Fast-GPOM. The last two columns are for the ground truth maps and the ROC diagrams, respectively.

In the experiments we concentrate on three aspects:

  1. Time cost analysis on GPOM and Fast-GPOM

  2. Map quality comparison on GPOM and Fast-GPOM

  3. Map quality comparison on Fast-GPOM with different

In the map quality comparison, we utilize AUC (area under the curve) and ROC (receiver operating characteristic) as metrics to illustrate the performance.

There are different approaches to map quality measurement[11]. [3] uses SLAM to estimate the pose and remove some noisy points as the ground truth. In contrast, [1] only compares the quality to a synthetic map. We follow the latter approach and are using a ground truth map and synthetic sensor data to measure the mapping quality. Our Fast-GPOM is fast enough to work in real time ROS with listening and publishing. But for the experiments, instead of running live, we use recorded data from a bag file, which we read frame by frame.

Comparison on Time Cost

The comparison of mapping time cost here is on each simulation arenas. The time cost of each part of the classical GPOM and our Fast-GPOM algorithm are shown in Table I. For the map building part, the time cost comparison is illustrated in Fig 3. We can see that the time cost of our algorithm is reduced tremendously compared to the classical GPOM algorithm.

From the table, we can see the prediction time of GPOM is in a different order of magnitude compared to the other steps. Thus map building takes much longer than publishing. In contrast, the cost of time in Fast-GPOM is relatively close in among the steps. What’s more, in each arena, the build map cost of Fast-GPOM is in the same magnitude than publishing, in two of those maps it even takes less time. Which means that, if we run map building and publishing in parallel as a ROS package, the map can be displayed smoothly.

The frequency of the simulated laser scan is . In those three maps with the resolution of 0.02m, 0.05m and 0.02m, the frequency to process data can be about , and . So the Fast-GPOM is adequate for many cases.

Fig. 5: The generated robocup maps with .

Comparison of Map quality

sparse_obstacles simple_rooms robocup
GPOM 0.92 0.95 0.85
Fast-GPOM 0.93 0.93 0.87

In this part we compare the performance of the mapping between our algorithm and general GPOM by using ROC and AUC. The results of three scenes are showed as Fig 4 and each pixel of the maps is the probability of occupancy. To compute ROC and AUC of each map, we aligned and cropped the resulting map to ensure a pixel to pixel matching with the ground truth. During the evaluation, those pixels that are unknown in ground truth will be neglected. In the Fig. (d)d, the AUC of GPOM is higher than our algorithm, because the general GPOM algorithm estimates some parts of the area outside the wall to be occupancy area. That makes the map more like the ground truth, but that is a disadvantage in most scenarios. For example, in Fig. (i)i, there is a large part of unknown area estimated to be occupied area, but our algorithm could avoid the situations. In the most maps, the AUC of our algorithm is higher than the general GPOM algorithm. That means that our algorithm could guarantee to get a preferable map while it sped up the map building time. The AUC is shown in the ROC figures. To emphasize the comparison, we pick the AUC and put them in Table II.

Map Quality with different

In this experiment, we test our Fast-GPOM algorithm with different values for d of the robocup map: . We visualize the results and ROC curve in Fig. 5.

We can observe that the lower d values tend to provide thinner walls while maybe causing some blur in some edges. The ROC curves indicate that lower d values achieved better AUC value.

V Conclusions

In this work, we proposed a Fast-GPOM to make the incremental algorithm GPOM much faster. For that, we analyzed the classical Gaussian Process Occupancy Mapping (GPOM) in detail and evaluated the time cost of each step in order to find the most expensive step. Taking the spatial context into consideration, our improvement makes it possible to run GPOM in real time while at the same time reducing the wrong predictions in unknown space. From our experiments, we learned that our method provides a similar or better map quality compared to GPOM while speeding up the computation by an order of magnitude.




  1. O’Callaghan, S. T., & Ramos, F. T. (2012). Gaussian process occupancy maps. The International Journal of Robotics Research, 31(1), 42-62.
  2. Thrun, S. (2002). Robotic mapping: A survey. Exploring artificial intelligence in the new millennium, 1(1-35), 1.
  3. Jadidi, M. G., Miro, J. V., & Dissanayake, G. (2018). Gaussian processes autonomous mapping and exploration for range-sensing mobile robots. Autonomous Robots, 42(2), 273-290.
  4. Williams, C. K., & Rasmussen, C. E. (2006). Gaussian processes for machine learning. the MIT Press, 2(3), 4.Williams, C. K., & Rasmussen, C. E. (2006). Gaussian processes for machine learning. the MIT Press, 2(3), 4.
  5. Kim, S., & Kim, J. (2012, May). Building occupancy maps with a mixture of Gaussian processes. In Robotics and Automation (ICRA), 2012 IEEE International Conference on (pp. 4756-4761). IEEE.
  6. Ramos, F., & Ott, L. (2016). Hilbert maps: scalable continuous occupancy mapping with stochastic gradient descent. The International Journal of Robotics Research, 35(14), 1717-1730.
  7. Jadidi, M. G., Valls Miró, J., Valencia Carreño, R., Andrade-Cetto, J., & Dissanayake, G. (2013). Exploration in information distribution maps. In Proceedings of the 2013 RSS Workshop on Robotic Exploration, Monitoring, and Information Collection (pp. 1-8).
  8. Jadidi, M. G., Valls Miró, J., Valencia, R., Andrade-Cetto, J., & Dissanayake, G. (2013). Exploration using an information-based reaction-diffusion process.
  10. Neumann, M., Huang, S., Marthaler, D. E., & Kersting, K. (2015). pygps: A python library for gaussian process regression and classification. The Journal of Machine Learning Research, 16(1), 2611-2616.
  11. Schwertfeger, S., & Birk, A. (2015). Map evaluation using matched topology graphs. Autonomous Robots, Springer
Comments 0
Request Comment
You are adding the first comment!
How to quickly get a good reply:
  • Give credit where it’s due by listing out the positive aspects of a paper before getting into which changes should be made.
  • Be specific in your critique, and provide supporting evidence with appropriate references to substantiate general statements.
  • Your comment should inspire ideas to flow and help the author improves the paper.

The better we are at sharing our knowledge with each other, the faster we move forward.
The feedback must be of minimum 40 characters and the title a minimum of 5 characters
Add comment
Loading ...
This is a comment super asjknd jkasnjk adsnkj
The feedback must be of minumum 40 characters
The feedback must be of minumum 40 characters

You are asking your first question!
How to quickly get a good answer:
  • Keep your question short and to the point
  • Check for grammar or spelling errors.
  • Phrase it like a question
Test description