Skip to content

Commit

Permalink
Update index.md
Browse files Browse the repository at this point in the history
Adding publication and performance section
  • Loading branch information
CathIAS authored Jun 14, 2020
1 parent 60376db commit 385d05c
Showing 1 changed file with 37 additions and 7 deletions.
44 changes: 37 additions & 7 deletions index.md
Original file line number Diff line number Diff line change
@@ -1,17 +1,47 @@
### *Welcome to TLIO home page!*
#### TLIO is an IMU-only pedestrian navigation approach combining deep learning and Bayesian filter. We recognize human motion patterns from 60 hours of training data consisting of various pedestrian activities, and fuse the learned relative state estimates and uncertainties into an EKF framework solving for pose, velocity and sensor biases.

## Publication

![Introdunction Trajectory](images/IntroductionTrajAfterReview.png)
**IMU-Based Pedestrian Dead Reckoning with Learned Motion Model** \
Wenxin Liu, David Caruso, Eddy Ilg, Jing Dong, Anastasios I. Mourikis, Kostas Daniilidis, Vijay Kumar, Jakob Engel

<p align="center">
<img width="700" src=images/IntroductionTrajAfterReview.png>
</p>

**Abstract**: In this work we propose a tightly-coupled Extended Kalman Filter framework for IMU-only state estimation. Strap-down IMU measurements provide relative state estimates based on IMU kinematic motion model. However the integration of measurements is sensitive to sensor bias and noise, causing significant drift within seconds. Recent research by Yan et al. (RoNIN) and Chen et al. (IONet) showed the capability of using trained neural networks to obtain accurate 2D displacement estimates from segments of IMU data and obtained good position estimates from concatenating them. This paper demonstrates a network that regresses 3D displacement estimates and its uncertainty, giving us the ability to tightly fuse the relative state measurement into a stochastic cloning EKF to solve for pose, velocity and sensor biases. We show that our network, trained with pedestrian data from a headset, can produce statistically consistent measurement and uncertainty to be used as update step in the filter, and the tightly-coupled system outperforms velocity integration approaches in position estimates, and AHRS attitude filter in orientation estimates.

[Paper]
[Video]
[Code]


## Performance

We compare TLIO to our [RONIN](https://ronin.cs.sfu.ca/) implementation in 3D. The ground truth for the error metrics is from a state-of-the-art Visual Inertial Odometry implementation based on [MSCKF](https://www-users.cs.umn.edu/~stergios/papers/ICRA07-MSCKF.pdf).

<details>
<summary> Show figures </summary>

Accurate bias estimates from the filter result in better orientation estimates than AHRS attitude filter. RONIN uses AHRS rotation while TLIO estimates the rotation as part of the state. TLIO also obtains better position estimates.
![system-perf](images/system-perf.png)

This image shows sample trajectories in 2D. 3.a and 3.b are failure cases of side-stepping and rolling on a chair.
![exam-traj](images/examplesTraj2.png)

</details>

Checkout these supplementary tracking videos (Blue: TLIO; Green: VIO)

[Staircase](https://drive.google.com/open?id=1NIZilMaIGx05EUPfztoMxiR2g8P3C0TM)
[Symposium]
[Pool]
[Apartment]


This work aims at solving pedestrian state estimation problem with only IMU data, leveraging the combination of deep learning and Bayesian filters.

We propose a tightly-coupled Extended Kalman Filter framework for IMU-only state estimation. We train a network that regresses 3D displacement estimates and its uncertainty, and tightly fuse the relative state measurement into a stochastic cloning EKF solving for pose, velocity and sensor biases. We call this method **TLIO: Tight Learned Inertial Odometry**.

Here is a supplementary video showing tracking on staircases. (Blue: TLIO; Green: VIO)

[Staircase video](https://drive.google.com/open?id=1NIZilMaIGx05EUPfztoMxiR2g8P3C0TM)

More videos will be disclosed on this website soon!


0 comments on commit 385d05c

Please sign in to comment.