Repository: https://huggingface.co/datasets/Leandro4002/LEANDRONE_V2
Download zip: https://public.saraivam.ch/static/LEANDRONE_V2.zip
Description
The LEANDRONE_V2 dataset is a collection of 240 labelled images triying to mimic the photo taken by the front camera of a Bitcraze AI deck 1.1 mounted on a Crazyflie 2.1 nanodrone. The camera model is a Himax HM01B0.
This dataset is aimed to create a machine learning model for an autonomous line-following drone. A secondary objective was to count the number of bottles along the track, but it has been abandonned and the number of bottles is not indicated in the labels. The track is approximately 3.5m x 1.8m. The images are monochrome, 324x244 (The specs of the camera says it is 320x320, but in reality it is 324x244). The images are generated ~15-30 cm from the ground.
Each iteration has random variations of the line, number of bottles and bottles positions. For each iteration, there is images from the "drone" going around the track. There are 6 iterations with each 40 images which gives us a total of 240 images. Every even iteration, the track is done clockwise, and every odd iteration it is done anti-clockwise. In the image name, the first 2 digits is the iteration number and the 2 next digits is for the image number. They are separated by a '_'.
File structure
|
|
|---render
| "Black and white images, named from '00_00.png' to '05_39.png'"
|
|
|
|---labels.csv
"csv list containing 2 values per row: The name of the image and the angle where the drone should turn next.
A negative angle indicate a left turn and a positive angle indicates a right turn."
Example of sample
Extract from labels.csv:
image;angle
00_00.png;-0.04881754187845043
00_01.png;0.04906715233908746
00_02.png;0.08485470738964143
00_03.png;0.08043066958386594
Light sensitivity
The camera of the AI-deck sets it's light sensitivity at the start, so the images captured by the camera will be different depending on the camera's exposure to light when starting the drone.
Here is a comparison when the drone is started with the camera covered and white the camera looking at a light source:
The images in this dataset are generated for a camera that has been covered at the start.
Comparison generated vs actual
Here is a comparison between generated in the dataset and actual photos taken by the drone:
Small light artifacts may appear on the dataset's images.
Counting bottles
There has been an attempt to track the number of bottles around the track using the yolov8n model. This model assigns an id to an object when it is confident enough to have identify it correctly. In order to test the model, we try applying this model to the images of an iteration of the track and converting them into a video. In a video containing 7 bottles, the model could detect only 2:
This is probably due to the fact that this model is trained on bottles with standard shapes, and there are sometimes energy drinks and thermos flasks in the dataset, which disrupted the model. Also. the image resolution isn't very high, it's only in monochrome and the 3D model's rendering engine can't handle transparency and simply displays gray. As a result, a lot of detail is lost, and the model is no longer able to detect bottles accurately.
How it has been made
A 3d model of the room has been made with photogrammetry using Meshroom 2023.3.0:
Then, we import this model in Blender 4.1 and we move the camera along the track and compute the angle between the camera and the nex points in order to obtain the label and the generated image from the point of view of the camera at this point:
Author
Made in Blender 4.1.1 by Leandro SARAIVA MAIA
30 May 2024
This work is marked with CC0 1.0 Universal
- Downloads last month
- 136