Filtern
Dokumenttyp
- Video (57)
- Konferenzveröffentlichung (8)
Schlagworte
- Robotik (30)
- Flugkörper (21)
- UAV (21)
- Rettungsrobotik (8)
- 3D Modell (7)
- Rasenmäher (4)
- DRZ (3)
- Deutsches Rettungsrobotik-Zentrum (3)
- SLAM (3)
- Kartierung (2)
- Mapping (2)
- Point Clouds (2)
- Ubuntu (2)
- Vegetationsbrand (2)
- 360° Panorama (1)
- 3D-Printer (1)
- Bildverarbeitung (1)
- Inferenz <Künstliche Intelligenz> (1)
- NeRF (1)
- Ortsbestimmung (1)
- Rescue Robotics (1)
- Small UAVs (1)
- Travelling-salesman-Problem (1)
- Vegetationsbrandübung (1)
- Vegetatonsbrandübung (1)
- Visual Monocular SLAM (1)
Institut
- Informatik und Kommunikation (65) (entfernen)
Wie können mit Luftbildaufnahmen 3D Modelle generiert werden?
- Planen von kreisförmigen und einen rasterförmigen Flug Trajektorien.
- Autonomes Abfliegen und Aufnahme der Bilder
- Verortung der Bilder mittels GPS und Structure from Motion Algorithmen.
- Generierung von 3D Modellen mithilfe von Multi-View Stereo Algorithmen.
Venice 2018: Tradr Review
(2018)
The video shows an orthopoto and a textured 3D model of the location. 300 images were recorded in two short flights with a Mavic Pro in 50 meter height. The first one was a single grid while the camera facing down and the second one was a double grid facing the camera at an 60 degree angle. The 3D model is computed with OpenDroneMap.
In this paper, we present a method for detecting objects of interest, including cars, humans, and fire, in aerial images captured by unmanned aerial vehicles (UAVs) usually during vegetation fires. To achieve this, we use artificial neural networks and create a dataset for supervised learning. We accomplish the assisted labeling of the dataset through the implementation of an object detection pipeline that combines classic image processing techniques with pretrained neural networks. In addition, we develop a data augmentation pipeline to augment the dataset with utomatically labeled images. Finally, we evaluate the performance of different neural networks.
Sperical UAV: Crash Test with 1/2 liter bottle from 2 meters
The video shows the first test of a small spherical UAV (35 cm) with 4 rotors for missions in complex environments such as buildings, caves or tunnels. The spherical design protects the vehicle's internal components and allows the UAV to roll over the ground when the environment allows. The drone can land and take off in any position and come into contact with objects without endangering the propellers and can restart even after crashes.
This technical report is about the architecture and integration of very small commercial UAVs (< 40 cm diagonal) in indoor Search and Rescue missions. One UAV is manually controlled by only one single human operator delivering live video streams and image series for later 3D scene modelling and inspection. In order to assist the operator who has to simultaneously observe the environment and navigate through it we use multiple deep neural networks to provide guided autonomy, automatic object detection and classification and local 3D scene modelling. Our methods help to reduce the cognitive load of the operator. We describe a framework for quick integration of new methods from the field of Deep Learning, enabling for rapid evaluation in real scenarios, including the interaction of methods.
In the realm of digital situational awareness during disaster situations, accurate digital representations,
like 3D models, play an indispensable role. To ensure the
safety of rescue teams, robotic platforms are often deployed
to generate these models. In this paper, we introduce an
innovative approach that synergizes the capabilities of compact Unmaned Arial Vehicles (UAVs), smaller than 30 cm, equipped with 360° cameras and the advances of Neural Radiance Fields (NeRFs). A NeRF, a specialized neural network, can deduce a 3D representation of any scene using 2D images and then synthesize it from various angles upon request. This method is especially tailored for urban environments which have experienced significant destruction, where the structural integrity of buildings is compromised to the point of barring entry—commonly observed post-earthquakes and after severe fires. We have tested our approach through recent post-fire scenario, underlining the efficacy of NeRFs even in challenging outdoor environments characterized by water, snow, varying light conditions, and reflective surfaces.
9 Panoramen, das erste ist aus größerer Höhe aufgenommen und enthält im Himmel eine Karte mit den Positionen der aufgenommenen Punkte (gelb). Das aktuelle Bild ist im Fadenkreuz (rot). Zusätzlich noch ein paar Details zu dem aktuellen Punkt. Jedes Panorama ist 10 Sekunden lang.
Zum Betrachten die höchste Auflösungsstufe wählen und die Pausetaste verwenden. Mit dem gedrückten linken Button kann man sich im Bild bewegen.
The video showcases a 3D model of a chemical company following a tank explosion that occurred on August 17, 2023, in Kempen computed with the AI algorithm Neural Radiance Field (NeRF). Captured by a compact mini drone measuring 18cm x 18cm and equipped with a 360° camera, these images offer an intricate perspective of the aftermath. After a comprehensive aerial survey and inspection of the 360° images taken within the facility, authorities confirmed that it was safe for the evacuated residents to return to their homes. See also:
https://www1.wdr.de/fernsehen/aktuelle-stunde/alle-videos/video-grosser-chemieunfall-in-kempen-100.html
Nerf(acto) for the 3D modeling of the Computer Science building of Westfälische Hochschule GE
(2023)
The video shows a very high resolution 3D point cloud !!! of the computer science building of the University of Applied Science Gelsenkirchen. For the recording a 3 minute flight with a M30T was performed. The 105 images taken by the wide-angle camera during this flight were localized within 3 minutes using colmap and processed using Neural Radiance Fields (NeRF). The nerfacto model of Nerfstudio was trained on an Nvidia RTX 4090 for 8 minutes. Thus, a top 3D model is available after about 15 minutes.
The video shown here shows a free camera path rendered at 60 hz (Full HD).