IP : 192.168.1.5
Subnet Mask : 255.255.255.0
Subsequently, to visualize the point clouds, we can make use of the LIVOX-Viewer tool, which is provided by Livox Tech. This software allows us to view and analyze the captured point cloud data from the LIVOX-AVIA LiDAR system. Download the Livox-Viewer version 0.10 or 0.11 from the following link.
https://www.livoxtech.com/downloads
Once you have downloaded the LIVOX-Viewer software, extract the files from the downloaded package. After extraction, navigate to the folder where the LIVOX-Viewer is located. In this folder, you will find the executable or application file that you can run to launch the LIVOX-Viewer interface. This interface will enable you to interact with and visualize the point cloud data recorded by the LIVOX-AVIA LiDAR.
Run the livox viewer by running :
./livox_viewer.sh
Upon opening the LIVOX-Viewer, you will be able to see the list of available LIVOX LiDAR devices. Once your LIVOX-AVIA LiDAR is connected and recognized by the viewer, it will appear in the list. To observe the point clouds in real-time, simply click on the "play" button provided within the LIVOX-Viewer interface. This action will initiate the visualization of live point cloud data captured by the LIVOX-AVIA LiDAR in real-time.
Livox Tech offers a Software Development Kit (SDK) for their LiDAR devices, which enables developers to perform various operations and utilize the LiDAR's functionalities in their applications. For comprehensive details and information about the SDK, you can access the following link (provide the specific link to the Livox Tech website or SDK documentation page). There, you will find documentation, code samples, and resources to help you integrate and use the Livox LiDAR SDK effectively in your projects. Make sure to follow each step carefully.
https://github.com/Livox-SDK/Livox-SDK
Please make sure that the fixed_frame in RVIZ is changed to livox_frame
For seamless operation, it is essential to have both the CUDA version and the PyTorch Cuda version identical. This ensures compatibility and enables efficient utilization of GPU resources for accelerated processing.
Please make sure that the fixed_frame in RVIZ is changed to livox_frame
There are two methods to obtain data from the LiDAR: using the Livox Viewer
or the Livox SDK
. In both approaches, the data is saved in the .lvx
file format. To further process and utilize the data for training with deep learning models, you can convert the .lvx
files to .bag
files. For comprehensive information and guidelines, please refer to the Livox SDK repository, where you can find detailed instructions and resources to aid you in this process.
Please run the following command to convert rosbag to pointcloud.
rosrun pcl_ros bag_to_pcd <input_file.bag> <topic> <output_directory>
lidarLabeler
Once you have launched MATLAB Lidar Labeler, start a new session by selecting the "New Session" option. In the new session, you can proceed to upload the point cloud files that you want to label. This will allow you to begin the labeling process and annotate objects of interest within the point cloud data. Select the label as Cuboid and label the objects from the point cloud.
To automate the labeling process in MATLAB Lidar Labeler, follow these steps:
-
Open MATLAB Lidar Labeler and load the point cloud data for labeling.
-
Manually label one object of interest in each frame for around 10-20 frames. Ensure that you label the same object consistently in each frame.
-
After manual labeling, use one of the provided automation algorithms in MATLAB Lidar Labeler. This algorithm will attempt to label the selected object automatically in the subsequent frames.
-
Review the automated labels and correct any inaccuracies or mistakes. This step is crucial to ensure the accuracy of the automated labeling.
-
Repeat the process for each object you want to label automatically. For each object, select one instance of it in each frame and apply the automation algorithm followed by manual corrections.
By following this approach, you can accelerate the labeling process while still maintaining the accuracy of the annotations by using the provided automation algorithms and making manual corrections as needed.
M-by-9 numeric matrix with rows of the form [xctr, yctr, zctr, xlen, ylen, zlen, xrot, yrot, zrot], where:
-
M is the number of labels in the frame.
-
xctr, yctr, and zctr specify the center of the cuboid.
-
xlen, ylen, and zlen specify the length of the cuboid along the x-axis, y-axis, and z-axis, respectively, before rotation has been applied.
-
xrot, yrot, and zrot specify the rotation angles for the cuboid along the x-axis, y-axis, and z-axis, respectively. These angles are clockwise-positive when looking in the forward direction of their corresponding axes.
The figure shows how these values determine the position of a cuboid.
For converting the labels to .txt
file export the labels in workspace and use the provided label_save.mlx
script in matlab inside the labels_python folder. It will convert the labels to .txt
file. After that use the [data_process.py](labels_python/data_process.py)
to visualize the pointclouds and labels in python. You can see similar thing as the picture below.
https://fr.mathworks.com/help/lidar/ug/object-detection-using-pointpillars-network.html
For our data please follow the given [train_pcd.mlx](train_matlab/train_pcd.mlx)
script inside [train_matlab](train_matlab)
folder.
For using point pillers network in matlab please follow the instructions given in the link below.
https://fr.mathworks.com/help/deeplearning/ug/lidar-object-detection-using-complex-yolov4.html
For our data please follow the given [train_pcd.mlx](train_matlab/yolov4.mlx)
script inside [train_matlab](train_matlab)
folder.
- Object Detection
- Distance Measurement
- Distance Measurement in real time
- Data Collection for custom dataset
- Train deep learning models on small dataset in MATLAB and measure distance of the objects
- Labeling all the collected data
- Train for multiclass objects
- Change the labeling format of the custom dataset to train in python
- Test on realtime scenarios
- Improve yolov4 accuracy