Flow chart of Tracking thread:
Main functions corresponding to each process (from wubo @ bubble robot):
Overall flow chart of Tracking
The picture above shows tracking CC speaks very clearly.
After the tracking thread obtains the image data, it will pass it to the function GrabImageStereo, GrabImageRGBD or ...
Posted by Burns on Sat, 26 Feb 2022 06:54:12 +0100
There are multiple radar function packages under the lidar folder.
It also includes a udev rule conversion script. The radar uses cp2102usb to serial port chip.
Multiple devices are connected to the host through usb, so it is difficult to distinguish the devices corresponding to each port name.
Set rules for the chip of USB to serial por ...
Posted by New Coder on Mon, 21 Feb 2022 19:58:49 +0100
Edgeslam environment installation configuration
Since the tutorials and records of environment installation and configuration cannot be found in edgeslam's environment at present, we can only grope a little bit here.
In order to ensure that the software version does not conflict directly, this paper completely follows the environment version ...
Posted by Garath531 on Fri, 11 Feb 2022 23:53:32 +0100
Cartogragper source code reading
In the future, we will study and record the source code of Cartogragper
1, Knowledge reserve
Gflags concise tutorial GFlags usage document GFlags is an open source library for command line parameter processing, including C + + version and python version. Unlike libraries such as getopt(), the ...
Posted by AlanG on Wed, 26 Jan 2022 22:29:22 +0100
stay SLAM GMapping (4) SLAM processor Each particle updated by particle filter independently records a possible robot trajectory and environment map
In order to improve efficiency, GMapping designs a data structure of trajectory tree The trajectory of the robot can be obtained by tracing from the leaf node to the root node ...
Posted by phpscriptcoder on Wed, 26 Jan 2022 20:17:05 +0100
Through the introduction in the previous section, we can calculate the camera motion, but we don't know the specific position of feature points in space. In monocular SLAM, we cannot obtain the depth information of pixels only by relying on a single image. We need to estimate the depth of map points by triangulation.
Posted by ChrisMayhew on Mon, 24 Jan 2022 09:35:09 +0100
The first tone is to write so much. The second tone is to sort out some contents of global optimization, but it is estimated that it will take some time. This record is purely personal understanding. If there is any difference, welcome to discuss ~ as for the part of using optical flow for dynamic object tracking in the paper, I am not sure whe ...
Posted by stevel on Fri, 14 Jan 2022 04:46:59 +0100
R3live is the relevant achievement of the MARS experiment of the University of Hong Kong on Livox radar. It was opened to the public on December 31, 2021. Its main goal is to make a multi-sensor fusion scheme with Livox radar as the core. The following is the personal environment configuration and simple operation record under Ubuntu system.
Posted by itshim on Mon, 10 Jan 2022 05:16:50 +0100
After understanding the principle and process of SLAM, individuals often wonder how to design and write a set of SLAM framework that can meet our needs from scratch. The author believes that Ceres, Eigen, Sophus and G2O function libraries cannot be avoided, especially Ceres function libraries are widely used in the optimization of l ...
Posted by thientanchuong on Wed, 05 Jan 2022 17:19:03 +0100