Refine
Year of publication
- 2024 (3)
- 2023 (17)
- 2022 (12)
- 2021 (24)
- 2020 (20)
- 2019 (30)
- 2018 (35)
- 2017 (25)
- 2016 (26)
- 2015 (22)
- 2014 (24)
- 2013 (27)
- 2012 (17)
- 2011 (29)
- 2010 (24)
- 2009 (25)
- 2008 (26)
- 2007 (28)
- 2006 (17)
- 2005 (15)
- 2004 (19)
- 2003 (12)
- 2002 (20)
- 2001 (20)
- 2000 (19)
- 1999 (23)
- 1998 (16)
- 1997 (13)
- 1996 (10)
- 1995 (11)
- 1994 (15)
- 1993 (9)
- 1992 (8)
- 1991 (4)
- 1990 (6)
- 1989 (5)
- 1988 (5)
- 1987 (2)
- 1986 (1)
- 1985 (6)
- 1984 (1)
- 1982 (2)
- 1981 (1)
- 1980 (4)
- 1979 (2)
- 1978 (3)
- 1976 (1)
- 1975 (1)
- 1974 (1)
Document Type
- Article (400)
- Conference Proceeding (224)
- Part of a Book (35)
- Book (23)
- Patent (2)
- Doctoral Thesis (1)
- Poster (1)
Language
- English (686) (remove)
Has Fulltext
- no (686) (remove)
Keywords
- Enterprise Architecture (5)
- MINLP (5)
- Engineering optimization (4)
- Optimization (3)
- Powertrain (3)
- Technical Operations Research (3)
- Telecommunication (3)
- Competence Developing Games (2)
- Energy efficiency (2)
- Experimental validation (2)
- Gamification (2)
- Hot S-parameter (2)
- Machine Learning (2)
- Optimal Topology (2)
- Process engineering (2)
- Pump System (2)
- Robotic Process Automation (2)
- Serious Game (2)
- Ventilation System (2)
- Water (2)
Institute
- Fachbereich Elektrotechnik und Informationstechnik (686) (remove)
Today, the assembly of laser systems requires a large share of manual operations due to its complexity regarding the optimal alignment of optics. Although the feasibility of automated alignment of laser optics has been shown in research labs, the development effort for the automation of assembly does not meet economic requirements – especially for low-volume laser production. This paper presents a model-based and sensor-integrated assembly execution approach for flexible assembly cells consisting of a macro-positioner covering a large workspace and a compact micromanipulator with camera attached to the positioner. In order to make full use of available models from computer-aided design (CAD) and optical simulation, sensor systems at different levels of accuracy are used for matching perceived information with model data. This approach is named "chain of refined perception", and it allows for automated planning of complex assembly tasks along all major phases of assembly such as collision-free path planning, part feeding, and active and passive alignment. The focus of the paper is put on the in-process image-based metrology and information extraction used for identifying and calibrating local coordinate systems as well as the exploitation of that information for a part feeding process for micro-optics. Results will be presented regarding the processes of automated calibration of the robot camera as well as the local coordinate systems of part feeding area and robot base.
Ground or aerial robots equipped with advanced sensing technologies, such as three-dimensional laser scanners and advanced mapping algorithms, are deemed useful as a supporting technology for first responders. A great deal of excellent research in the field exists, but practical applications at real disaster sites are scarce. Many projects concentrate on equipping robots with advanced capabilities, such as autonomous exploration or object manipulation. In spite of this, realistic application areas for such robots are limited to teleoperated reconnaissance or search. In this paper, we investigate how well state-of-the-art and off-the-shelf components and algorithms are suited for reconnaissance in current disaster-relief scenarios. The basic idea is to make use of some of the most common sensors and deploy some widely used algorithms in a disaster situation, and to evaluate how well the components work for these scenarios. We acquired the sensor data from two field experiments, one from a disaster-relief operation in a motorway tunnel, and one from a mapping experiment in a partly closed down motorway tunnel. Based on these data, which we make publicly available, we evaluate state-of-the-art and off-the-shelf mapping approaches. In our analysis, we integrate opinions and replies from first responders as well as from some algorithm developers on the usefulness of the data and the limitations of the deployed approaches, respectively. We discuss the lessons we learned during the two missions. These lessons are interesting for the community working in similar areas of urban search and rescue, particularly reconnaissance and search.