This shows you the differences between two versions of the page.
Both sides previous revisionPrevious revisionNext revision | Previous revision | ||
en:ros:simulations:iseauto [2020/03/23 08:46] – tomykalm | en:ros:simulations:iseauto [Unknown date] (current) – external edit (Unknown date) 127.0.0.1 | ||
---|---|---|---|
Line 1: | Line 1: | ||
- | ====== | + | ====== |
+ | Self-driving vehicles can fall into many different categories and levels of automation as described in previous chapters. One of the higher-level self-driving vehicles available nowadays is a Level 4 self-driving robot bus or Automated Vehicle (AV shuttle), also called a last-mile autonomous vehicle. The following describes an AV shuttle designed and developed in Tallinn University of Technology in cooperation with industrial partners in Estonia. | ||
===== Vehicle Description ===== | ===== Vehicle Description ===== | ||
+ | The iseAuto project was a cooperation project between industry and university which has a range of objectives from both sides as well as a very practical outcome. The project started in June 2017, when TalTech and Silberauto Estonia agreed to jointly develop a self-driving bus that had its public demonstration in September 2018. The purpose from the company’s side was to get involved with self-driving technology to be aware of the future of the automotive industry and also get experience in manufacturing a special purpose car body, as that is one of the main activities of one of the branches of the company. | ||
+ | |||
+ | {{ : | ||
+ | |||
+ | **Vehicle specifications** | ||
- | Vehicle specifications | ||
* Carriage of luggage or persons for 4 + 2 persons | * Carriage of luggage or persons for 4 + 2 persons | ||
* Average driving speed is 10-20 km/h | * Average driving speed is 10-20 km/h | ||
Line 11: | Line 15: | ||
* Main engine power 47 kW | * Main engine power 47 kW | ||
* Battery pack 16 kWh | * Battery pack 16 kWh | ||
- | {{: | ||
- | Gauges: | + | **Sensors:** |
- | * Height 2.4 m | + | |
- | * Length 3.5 m | + | |
- | * 1.5 m wide | + | |
- | Sensors: | + | * Front-top LiDAR Velodyne VLP-32 |
- | * LiDAR Velodyne VLP-16 | + | |
- | | + | |
- | * LiDAR in front | + | |
- | * Short distance radar in front | + | |
* Cameras | * Cameras | ||
* RTK-GNSS | * RTK-GNSS | ||
* IMU | * IMU | ||
- | Software: | + | {{ : |
+ | |||
+ | The iseAuto 3D model and its Lidar sensors are illustrated in the above figure. Two Velodyne VLP-16 and VLP-32 are installed at the top back and front of the vehicle respectively. Furthermore, | ||
+ | |||
+ | **Software:** | ||
* Robot Operating System (ROS) | * Robot Operating System (ROS) | ||
* Autoware | * Autoware | ||
- | * Yolo | ||
- | ===== Simulating the car ===== | + | ===== Simulating the Self-driving Vehicle ===== |
+ | TalTech iseAuto has a simulation model which can be simulated in the Gazebo environment by utilizing its physics and sensors. Nevertheless, | ||
+ | |||
+ | {{ : | ||
+ | |||
+ | On the other hand, CARLA and SVL are modern open-source simulators based on the game engines, Unreal and Unity respectively, | ||
+ | |||
+ | {{ : | ||
+ | |||
+ | The above figure shows a full map of the simulation workflow and the relation between Autoware and the simulator. Vehicle 3D models and the virtual environment, | ||
+ | |||
+ | The 3D mesh model of the vehicle should be imported to Unity to define physics components like collider and wheel actuation, in addition, to assign other features like lights and materials for appearance. Finally, the built vehicle exported from Unity is used in the simulator. Later, all the sensor configurations are defined via a JSON file inside the simulator. | ||
+ | |||
+ | ==== Virtual Environment Creation ==== | ||
+ | In order to create a virtual environment as a testbed for simulating the autonomous vehicle, the aerial mapping approach to map the desired environment is needed. In this way, a drone is used to capture images from the real environment, | ||
+ | |||
+ | {{ : | ||
+ | |||
+ | Taking aerial photos is one of the most important steps in the mapping process as it will significantly affect the outcome of the process and the amount of work to be done to process those images. The images taken are georeferenced by the drone. The onboard IMU provides the pictures with the orientation so that later they can be stitched together and used for photogrammetric processing. Third-party software aligns and creates the dense point cloud from the pictures that were captured. Once the dense point cloud is created, the segmentation and classification of the points are needed in order to separate unwanted objects and vegetation from the point cloud data. The below figure shows the three main steps to generate the Unity train from geospatial data. | ||
+ | |||
+ | {{ : | ||
+ | |||
+ | Now it is ready to load the final map that is built by Unity for the SVL simulator. | ||
+ | |||
+ | ==== Run the Simulation ==== | ||
+ | Now, by using a high-fidelity simulator we can simulate different scenarios close to the real to evaluate the control algorithm performance and safety. In terms of defining these scenarios, the SVL simulator provides a python API for spawning different objects like NPC (Non-player character) cars and pedestrians inside the virtual environment with different motion plans. | ||
+ | |||
+ | {{ : | ||
+ | |||
+ | The above-left figure shows the vehicle inside the simulator that is stopped behind an NPC car, while, at the same time the right figure shows the point cloud of all Lidars that is derived from the current scene in the environment. | ||
+ | |||
+ | {{ : | ||
+ | |||
+ | ==== Summary | ||
+ | Simulation is a more and more important aspect of developing robots and intelligent vehicles like self-driving cars. The importance is rising in parallel with the complexity of control algorithms and the integration of mechatronic systems in the vehicle. Simulation helps to reduce drastically the development cost and increase the general safety of the end result. It is clear that early prototypes may easily cause accidents if the testing is done on real traffic but if thousands of miles are conducted on the simulations and most of the software bugs and electromechanical incompatibilities are eliminated it is safer to start open road tests. Therefore simulations during the self-driving vehicle development are almost must be today' | ||
- | {{: | ||
- | To download the Iseauto Simulator, contact Iseauto Project Manager or your lecturer. The self-driving simulator is located in the project git repository, managed by the Iseauto project coordinators and the software development team leader. |