Filtern
Erscheinungsjahr
Dokumenttyp
- Video (55)
- Konferenzveröffentlichung (14)
- Wissenschaftlicher Artikel (12)
- Teil eines Buches (Kapitel) (8)
- Vorlesung (7)
- Masterarbeit (2)
- Bericht (2)
- Sonstiges (1)
- Arbeitspapier (1)
Volltext vorhanden
- ja (102) (entfernen)
Schlagworte
- Robotik (27)
- Flugkörper (18)
- UAV (18)
- 3D Modell (7)
- Akkreditierung (6)
- E-Learning (6)
- Radio-Feature (6)
- Rettungsrobotik (5)
- Virtuelle Hochschule (5)
- Rasenmäher (4)
Institut
- Informatik und Kommunikation (102) (entfernen)
This technical report is about the architecture and integration of very small commercial UAVs (< 40 cm diagonal) in indoor Search and Rescue missions. One UAV is manually controlled by only one single human operator delivering live video streams and image series for later 3D scene modelling and inspection. In order to assist the operator who has to simultaneously observe the environment and navigate through it we use multiple deep neural networks to provide guided autonomy, automatic object detection and classification and local 3D scene modelling. Our methods help to reduce the cognitive load of the operator. We describe a framework for quick integration of new methods from the field of Deep Learning, enabling for rapid evaluation in real scenarios, including the interaction of methods.
In the realm of digital situational awareness during disaster situations, accurate digital representations,
like 3D models, play an indispensable role. To ensure the
safety of rescue teams, robotic platforms are often deployed
to generate these models. In this paper, we introduce an
innovative approach that synergizes the capabilities of compact Unmaned Arial Vehicles (UAVs), smaller than 30 cm, equipped with 360° cameras and the advances of Neural Radiance Fields (NeRFs). A NeRF, a specialized neural network, can deduce a 3D representation of any scene using 2D images and then synthesize it from various angles upon request. This method is especially tailored for urban environments which have experienced significant destruction, where the structural integrity of buildings is compromised to the point of barring entry—commonly observed post-earthquakes and after severe fires. We have tested our approach through recent post-fire scenario, underlining the efficacy of NeRFs even in challenging outdoor environments characterized by water, snow, varying light conditions, and reflective surfaces.
Durch Panoramen in Kombination mit dem ORB-SLAM ist ein schnelles Tracking möglich, liefert jedoch ausschließlich spärliche Daten. Durch die Kombination mit einem neuronalen Netz soll der SLAM Algorithmus zu einem RGBD-SLAM erweitert werden, um ein besseres Tracking und eine dichtere Punktwolke zu gewährleisten.
The two churches, San Francesco and Sant'Agostino in Amatrice, Italy was hit by an earthquake on August 24 2016. Both churches are in a state of partial collapse, in need of shoring to prevent potential further destruction and to preserve the national heritage. The video show the mission at 1.Sept.2016 in clips of 10 seconds.
The TRADR project was asked by the Italian firebrigade Vigili del Fuoco to provide 3D textured models of two churches.
The team entered San Francesco with two UGVs (ground robots) and one UAV (drone, flown by Prof. Surmann), teleoperating them entirely out of line of sight and partially in collaboration. We entered Sant'Agostino with one UAV (also flown by Prof. Surmann) while two other UAVs were providing a view from different angles to facilitate maneuvering them entirely out of line of sight.
Venice 2018: Tradr Review
(2018)
The video shows an orthopoto and a textured 3D model of the location. 300 images were recorded in two short flights with a Mavic Pro in 50 meter height. The first one was a single grid while the camera facing down and the second one was a double grid facing the camera at an 60 degree angle. The 3D model is computed with OpenDroneMap.
This video shows a model computed from 124 images taken at the Tjex 2015 of the trade project (www.tradr-project.eu). The images were acquired by walking around the object and reconstruct the structure with VisualSfm software.
The video shows a snapshot of a 16 minute flight of a DJI Phantom 3 professional over the Schloss Birlinghoven at Sankt Augustin, Germany. The castle is located at the Fraunhofer Campus at Sankt Augustin. The 3D model is generated out of 400 key frames of the 4k video which are cut out with ffmpeg. The work is part of an evaluation in the Tradr Project (www.tradr-project.eu)
Challenging visual localization of an UAV while flying out of a room into a snowy environment (~ 4:50). The UAV is equipped with a 360° camera. The localization is done with OpenVSLAM.
The video was recorded in Jan. 2019 at the Fire Brigade training center in Dortmund
To achieve nearly real time conditions the original resolution of 5k (30 fps) was reduced to 2k (ffmpeg -i video.mp4 -vf scale=1920:-1 -crf 25 vido-small.mp4) with high compression (-crf 25). This reduce the original size from 3.2 GB to 93MB (~ 4 MBit/s which could be transmitted online via a radio link). The localization shown did not use frameskip. With a frameskip above 1 the localization fails while the UAV is flying through the window. Indoor localization can be done with a frameskip of 3 in real time.
9 Panoramen, das erste ist aus größerer Höhe aufgenommen und enthält im Himmel eine Karte mit den Positionen der aufgenommenen Punkte (gelb). Das aktuelle Bild ist im Fadenkreuz (rot). Zusätzlich noch ein paar Details zu dem aktuellen Punkt. Jedes Panorama ist 10 Sekunden lang.
Zum Betrachten die höchste Auflösungsstufe wählen und die Pausetaste verwenden. Mit dem gedrückten linken Button kann man sich im Bild bewegen.