Informatik und Kommunikation
Filtern
Erscheinungsjahr
Dokumenttyp
- Video (60)
- Konferenzveröffentlichung (45)
- Teil eines Buches (Kapitel) (39)
- Wissenschaftlicher Artikel (32)
- Lehrmaterial (13)
- Buch (Monographie) (11)
- Vorlesung (7)
- Beitrag zu einer (nichtwissenschaftlichen) Zeitung oder Zeitschrift (4)
- Sonstiges (3)
- Masterarbeit (2)
Schlagworte
- Robotik (30)
- Flugkörper (21)
- UAV (21)
- Journalismus (15)
- Rettungsrobotik (8)
- 3D Modell (7)
- Akkreditierung (7)
- E-Learning (7)
- Marketing (6)
- Radio-Feature (6)
Institut
ARGUS is a tool for the systematic acquisition, documentation and evaluation of drone flights in rescue operations. In addition to the very fast generation of RGB and IR orthophotos, a trained AI can automatically detect fire, people and cars in the images captured by the drones. The video gives a short introduction to the Aerial Rescue and Geospatial Utility System -- ARGUS
Check out our Github repository under
https://github.com/RoblabWh/argus/
You can find the dataset on kaggle under
https://www.kaggle.com/datasets/julienmeine/rescue-object-detection
Selbstständig und selbstbestimmt essen und trinken zu können gehört zu den Grundbedürfnissen des Menschen und wird den Aktivitäten des täglichen Lebens (ATLs) zugeordnet. Körperliche Beeinträchtigungen, die mit Funktionsverlusten in Armen, Händen und ggf. der Beweglichkeit des Oberkörpers einhergehen, schränken die selbstständige Nahrungszufuhr erheblich ein. Die Betroffenen sind darauf angewiesen, dass ihnen Getränke und Mahlzeiten zubereitet, bereitgestellt und angereicht werden. Zu dieser Personengruppe gehören Menschen mit querschnittbedingter Tetraplegie, Multiple Sklerose, Muskeldystrophie und Erkrankungen mit ähnlichen Auswirkungen. Derzeit gibt es verschiedene assistive Technologien, die das selbstständige Essen und Trinken wieder ermöglichen sollen.Wie aber muss ein Interaktionsdesign für einen Roboterarm gestaltet sein, damit er von den Betroffenen zur Nahrungsaufnahme genutzt werden kann? Welche Anforderungen gibt es und welche Aspekte sind in Bezug auf die Akzeptanz eines Roboterarms zu berücksichtigen?
The two churches, San Francesco and Sant'Agostino in Amatrice, Italy was hit by an earthquake on August 24 2016. Both churches are in a state of partial collapse, in need of shoring to prevent potential further destruction and to preserve the national heritage. The video show the mission at 1.Sept.2016 in clips of 10 seconds.
The TRADR project was asked by the Italian firebrigade Vigili del Fuoco to provide 3D textured models of two churches.
The team entered San Francesco with two UGVs (ground robots) and one UAV (drone, flown by Prof. Surmann), teleoperating them entirely out of line of sight and partially in collaboration. We entered Sant'Agostino with one UAV (also flown by Prof. Surmann) while two other UAVs were providing a view from different angles to facilitate maneuvering them entirely out of line of sight.
Robot arms are one of many assistive technologies used by people with motor impairments. Assistive robot arms can allow people to perform activities of daily living (ADL) involving grasping and manipulating objects in their environment without the assistance of caregivers. Suitable input devices (e.g., joysticks) mostly have two Degrees of Freedom (DoF), while most assistive robot arms have six or more. This results in time-consuming and cognitively demanding mode switches to change the mapping of DoFs to control the robot. One option to decrease the difficulty of controlling a high-DoF assistive robot arm using a low-DoF input device is to assign different combinations of movement-DoFs to the device’s input DoFs depending on the current situation (adaptive control). To explore this method of control, we designed two adaptive control methods for a realistic virtual 3D environment. We evaluated our methods against a commonly used non-adaptive control method that requires the user to switch controls manually. This was conducted in a simulated remote study that used Virtual Reality and involved 39 non-disabled participants. Our results show that the number of mode switches necessary to complete a simple pick-and-place task decreases significantl when using an adaptive control type. In contrast, the task completion time and workload stay the same. A thematic analysis of qualitative feedback of our participants suggests that a longer period of training could further improve the performance of adaptive control methods.
A Robust Interface for Head Motion based Control of a Robot Arm using MARG and Visual Sensors
(2018)
Head-controlled human machine interfaces have gained popularity over the past years, especially in the restoration of the autonomy of severely disabled people, like tetraplegics. These interfaces need to be reliable and robust regarding the environmental conditions to guarantee safety of the user and enable a direct interaction between a human and a machine. This paper presents a hybrid MARG and visual sensor system for head orientation estimation which is in this case used to teleoperate a robotic arm. The system contains a Magnetic Angular Rate Gravity (MARG)-sensor and a Tobii eye tracker 4C. A MARG sensor consists of tri-axis accelerometer, gyroscope as well as a magnetometer which enable a complete measurement of orientation relative to the direction of gravity and magnetic field of the earth. The tri-axis magnetometer is sensitive to external magnetic fields which result in incorrect orientation estimation from the sensor fusion process. In this work the Tobii eye tracker 4C is used to increase head orientation estimation because it also features head tracking even though it is commonly used for eye tracking. This type of visual sensor does not suffer magnetic drift. However, it computes orientation data only, if a user is detectable. Within this work a state machine is presented which enables data fusion of the MARG and visual sensor to improve orientation estimation. The fusion of the orientation data of MARG and visual sensors enables a robust interface, which is immune against external magnetic fields. Therefore, it increases the safety of the human machine interaction.