Refine
Year of publication
Institute
- MASKOR Institut für Mobile Autonome Systeme und Kognitive Robotik (74) (remove)
Document Type
- Conference Proceeding (49)
- Article (16)
- Part of a Book (8)
- Contribution to a Periodical (1)
Keywords
- Gamification (3)
- Industry 4.0 (3)
- Automation (2)
- Autonomous mobile robots (2)
- Digital Twin (2)
- IO-Link (2)
- Multi-robot systems (2)
- Smart factory (2)
- 10BASE-T1L (1)
- Adaptive Systems (1)
The development of protype applications with sensors and actuators in the automation industry requires tools that are independent of manufacturer, and are flexible enough to be modified or extended for any specific requirements. Currently, developing prototypes with industrial sensors and actuators is not straightforward. First of all, the exchange of information depends on the industrial protocol that these devices have. Second, a specific configuration and installation is done based on the hardware that is used, such as automation controllers or industrial gateways. This means that the development for a specific industrial protocol, highly depends on the hardware and the software that vendors provide. In this work we propose a rapid-prototyping framework based on Arduino to solve this problem. For this project we have focused to work with the IO-Link protocol. The framework consists of an Arduino shield that acts as the physical layer, and a software that implements the IO-Link Master protocol. The main advantage of such framework is that an application with industrial devices can be rapid-prototyped with ease as its vendor independent, open-source and can be ported easily to other Arduino compatible boards. In comparison, a typical approach requires proprietary hardware, is not easy to port to another system and is closed-source.
The production and assembly of customized products increases the demand for flexible automation systems. One approach is to remove the safety fences that separate human and industrial robot to combine their skills. This collaboration possesses a certain risk for the human co-worker, leading to numerous safety concepts to protect him. The human needs to be monitored and tracked by a safety system using different sensors. The proposed system consists of a RGBD camera for surveillance of the common working area, an array of optical distance sensors to compensate shadowing effects of the RGBD camera and a laser range finder to detect the co-worker when approaching the work cell. The software for collision detection, path planning, robot control and predicting the behaviour of the co-worker is based on the Robot Operating System (ROS). A first prototype of the work cell shows that with advanced algorithms from the field of mobile robotics a very flexible safety concept can be realized: the robot not simply stops its movement when detecting a collision, but plans and executes an alternative path around the obstacle.
With autonomous mobile robots receiving increased
attention in industrial contexts, the need for benchmarks
becomes more and more an urgent matter. The RoboCup
Logistics League (RCLL) is one specific industry-inspired scenario
focusing on production logistics within a Smart Factory.
In this paper, we describe how the RCLL allows to assess the
performance of a group of robots within the scenario as a
whole, focusing specifically on the coordination and cooperation
strategies and the methods and components to achieve them.
We report on recent efforts to analyze performance of teams in
2014 to understand the implications of the current grading
scheme, and derived criteria and metrics for performance
assessment based on Key Performance Indicators (KPI) adapted
from classic factory evaluation. We reflect on differences and
compatibility towards RoCKIn, a recent major benchmarking
European project.
The Scarab Project
(2015)
Urban Search and Rescue (USAR) is an active research
field in the robotics community. Despite recent advances
for many open research questions, these kind of systems are
not widely used in real rescue missions. One reason is that such
systems are complex and not (yet) very reliable; another is that
one has to be an robotic expert to run such a system. Moreover,
available rescue robots are very expensive and the benefits of
using them are still limited.
In this paper, we present the Scarab robot, an alternative
design for a USAR robot. The robot is light weight, humanpackable
and its primary purpose is that of extending the
rescuer’s capability to sense the disaster site. The idea is that a
responder throws the robot to a certain spot. The robot survives
the impact with the ground and relays sensor data such as
camera images or thermal images to the responder’s hand-held
control unit from which the robot can be remotely controlled.
We present a robotic tool that autonomously follows a conversation to enable remote presence in video conferencing. When humans participate in a meeting with the help of video conferencing tools, it is crucial that they are able to follow the conversation both with acoustic and visual input. To this end, we design and implement a video conferencing tool robot that uses binaural sound source localization as its main source to autonomously orient towards the currently talking speaker. To increase robustness of the acoustic cue against noise we supplement the sound localization with a source detection stage. Also, we include a simple onset detector to retain fast response times. Since we only use two microphones, we are confronted with ambiguities on whether a source is in front or behind the device. We resolve these ambiguities with the help of face detection and additional moves. We tailor the system to our target scenarios in experiments with a four minute scripted conversation. In these experiments we evaluate the influence of different system settings on the responsiveness and accuracy of the device.
Autonomous agents require rich environment models for fulfilling their missions. High-definition maps are a well-established map format which allows for representing semantic information besides the usual geometric information of the environment. These are, for instance, road shapes, road markings, traffic signs or barriers. The geometric resolution of HD maps can be as precise as of centimetre level. In this paper, we report on our approach of using HD maps as a map representation for autonomous load-haul-dump vehicles in open-pit mining operations. As the mine undergoes constant change, we also need to constantly update the map. Therefore, we follow a lifelong mapping approach for updating the HD maps based on camera-based object detection and GPS data. We show our mapping algorithm based on the Lanelet 2 map format and show our integration with the navigation stack of the Robot Operating System. We present experimental results on our lifelong mapping approach from a real open-pit mine.