Refine
Year of publication
Institute
- MASKOR Institut für Mobile Autonome Systeme und Kognitive Robotik (71) (remove)
Document Type
- Conference Proceeding (48)
- Article (16)
- Part of a Book (6)
- Contribution to a Periodical (1)
Keywords
- Gamification (3)
- Industry 4.0 (3)
- Autonomous mobile robots (2)
- Digital Twin (2)
- IO-Link (2)
- Multi-robot systems (2)
- Smart factory (2)
- 10BASE-T1L (1)
- Adaptive Systems (1)
- Arduino (1)
With autonomous mobile robots receiving increased
attention in industrial contexts, the need for benchmarks
becomes more and more an urgent matter. The RoboCup
Logistics League (RCLL) is one specific industry-inspired scenario
focusing on production logistics within a Smart Factory.
In this paper, we describe how the RCLL allows to assess the
performance of a group of robots within the scenario as a
whole, focusing specifically on the coordination and cooperation
strategies and the methods and components to achieve them.
We report on recent efforts to analyze performance of teams in
2014 to understand the implications of the current grading
scheme, and derived criteria and metrics for performance
assessment based on Key Performance Indicators (KPI) adapted
from classic factory evaluation. We reflect on differences and
compatibility towards RoCKIn, a recent major benchmarking
European project.
In the future, we expect manufacturing companies to follow a new paradigm that mandates more automation and autonomy in production processes. Such smart factories will offer a variety of production technologies as services that can be combined ad hoc to produce a large number of different product types and variants cost-effectively even in small lot sizes. This is enabled by cyber-physical systems that feature flexible automated planning methods for production scheduling, execution control, and in-factory logistics.
During development, testbeds are required to determine the applicability of integrated systems in such scenarios. Furthermore, benchmarks are needed to quantify and compare system performance in these industry-inspired scenarios at a comprehensible and manageable size which is, at the same time, complex enough to yield meaningful results.
In this chapter, based on our experience in the RoboCup Logistics League (RCLL) as a specific example, we derive a generic blueprint for how a holistic benchmark can be developed, which combines a specific scenario with a set of key performance indicators as metrics to evaluate the overall integrated system and its components.
Ground or aerial robots equipped with advanced sensing technologies, such as three-dimensional laser scanners and advanced mapping algorithms, are deemed useful as a supporting technology for first responders. A great deal of excellent research in the field exists, but practical applications at real disaster sites are scarce. Many projects concentrate on equipping robots with advanced capabilities, such as autonomous exploration or object manipulation. In spite of this, realistic application areas for such robots are limited to teleoperated reconnaissance or search. In this paper, we investigate how well state-of-the-art and off-the-shelf components and algorithms are suited for reconnaissance in current disaster-relief scenarios. The basic idea is to make use of some of the most common sensors and deploy some widely used algorithms in a disaster situation, and to evaluate how well the components work for these scenarios. We acquired the sensor data from two field experiments, one from a disaster-relief operation in a motorway tunnel, and one from a mapping experiment in a partly closed down motorway tunnel. Based on these data, which we make publicly available, we evaluate state-of-the-art and off-the-shelf mapping approaches. In our analysis, we integrate opinions and replies from first responders as well as from some algorithm developers on the usefulness of the data and the limitations of the deployed approaches, respectively. We discuss the lessons we learned during the two missions. These lessons are interesting for the community working in similar areas of urban search and rescue, particularly reconnaissance and search.
The field of Cognitive Robotics aims at intelligent decision making of autonomous robots. It has matured over the last 25 or so years quite a bit. That is, a number of high-level control languages and architectures have emerged from the field. One concern in this regard is the action language GOLOG. GOLOG has been used in a rather large number of applications as a high-level control language ranging from intelligent service robots to soccer robots. For the lower level robot software, the Robot Operating System (ROS) has been around for more than a decade now and it has developed into the standard middleware for robot applications. ROS provides a large number of packages for standard tasks in robotics like localisation, navigation, and object recognition. Interestingly enough, only little work within ROS has gone into the high-level control of robots. In this paper, we describe our approach to marry the GOLOG action language with ROS. In particular, we present our architecture on inte grating golog++, which is based on the GOLOG dialect Readylog, with the Robot Operating System. With an example application on the Pepper service robot, we show how primitive actions can be easily mapped to the ROS ActionLib framework and present our control architecture in detail.
The production and assembly of customized products increases the demand for flexible automation systems. One approach is to remove the safety fences that separate human and industrial robot to combine their skills. This collaboration possesses a certain risk for the human co-worker, leading to numerous safety concepts to protect him. The human needs to be monitored and tracked by a safety system using different sensors. The proposed system consists of a RGBD camera for surveillance of the common working area, an array of optical distance sensors to compensate shadowing effects of the RGBD camera and a laser range finder to detect the co-worker when approaching the work cell. The software for collision detection, path planning, robot control and predicting the behaviour of the co-worker is based on the Robot Operating System (ROS). A first prototype of the work cell shows that with advanced algorithms from the field of mobile robotics a very flexible safety concept can be realized: the robot not simply stops its movement when detecting a collision, but plans and executes an alternative path around the obstacle.
Project work and inter disciplinarity are integral parts of today's engineering work. It is therefore important to incorporate these aspects into the curriculum of academic studies of engineering. At the faculty of Electrical Engineering and Information Technology an interdisciplinary project is part of the bachelor program to address these topics. Since the summer term 2020 most courses changed to online mode during the Covid-19 crisis including the interdisciplinary projects. This online mode introduces additional challenges to the execution of the projects, both for the students as well as for the lecture. The challenges, but also the risks and chances of this kind of project courses are subject of this paper, based on five different interdisciplinary projects
We present a robotic tool that autonomously follows a conversation to enable remote presence in video conferencing. When humans participate in a meeting with the help of video conferencing tools, it is crucial that they are able to follow the conversation both with acoustic and visual input. To this end, we design and implement a video conferencing tool robot that uses binaural sound source localization as its main source to autonomously orient towards the currently talking speaker. To increase robustness of the acoustic cue against noise we supplement the sound localization with a source detection stage. Also, we include a simple onset detector to retain fast response times. Since we only use two microphones, we are confronted with ambiguities on whether a source is in front or behind the device. We resolve these ambiguities with the help of face detection and additional moves. We tailor the system to our target scenarios in experiments with a four minute scripted conversation. In these experiments we evaluate the influence of different system settings on the responsiveness and accuracy of the device.
The maintenance of wind turbines is of growing importance considering the transition to renewable energy. This paper presents a multi-robot-approach for automated wind turbine maintenance including a novel climbing robot. Currently, wind turbine maintenance remains a manual task, which is monotonous, dangerous, and also physically demanding due to the large scale of wind turbines. Technical climbers are required to work at significant heights, even in bad weather conditions. Furthermore, a skilled labor force with sufficient knowledge in repairing fiber composite material is rare. Autonomous mobile systems enable the digitization of the maintenance process. They can be designed for weather-independent operations. This work contributes to the development and experimental validation of a maintenance system consisting of multiple robotic platforms for a variety of tasks, such as wind turbine tower and rotor blade service. In this work, multicopters with vision and LiDAR sensors for global inspection are used to guide slower climbing robots. Light-weight magnetic climbers with surface contact were used to analyze structure parts with non-destructive inspection methods and to locally repair smaller defects. Localization was enabled by adapting odometry for conical-shaped surfaces considering additional navigation sensors. Magnets were suitable for steel towers to clamp onto the surface. A friction-based climbing ring robot (SMART— Scanning, Monitoring, Analyzing, Repair and Transportation) completed the set-up for higher payload. The maintenance period could be extended by using weather-proofed maintenance robots. The multi-robot-system was running the Robot Operating System (ROS). Additionally, first steps towards machine learning would enable maintenance staff to use pattern classification for fault diagnosis in order to operate safely from the ground in the future.
This summer, RoboCup competitions were held for the 20th time in Leipzig, Germany. It was the second time that RoboCup took place in Germany, 10 years after the 2006 RoboCup in Bremen. In this article, we give an overview on the latest developments of RoboCup and what happened in the different leagues over the last decade. With its 20th edition, RoboCup clearly is a success story and a role model for robotics competitions. From our personal view point, we acknowledge this by giving a retrospection about what makes RoboCup such a success.
The Robot Operating System (ROS) is the current de-facto standard in robot middlewares. The steadily increasing size of the user base results in a greater demand for training as well. User groups range from students in academia to industry professionals with a broad spectrum of developers in between. To deliver high quality training and education to any of these audiences, educators need to tailor individual curricula for any such training. In this paper, we present an approach to ease compiling curricula for ROS trainings based on a taxonomy of the teaching contents. The instructor can select a set of dedicated learning units and the system will automatically compile the teaching material based on the dependencies of the units selected and a set of parameters for a particular training. We walk through an example training to illustrate our work.