Refine
Year of publication
Institute
- Fachbereich Elektrotechnik und Informationstechnik (1184) (remove)
Language
- English (719)
- German (464)
- Multiple languages (1)
Document Type
- Article (637)
- Conference Proceeding (302)
- Book (114)
- Part of a Book (64)
- Patent (17)
- Report (9)
- Other (8)
- Conference: Meeting Abstract (6)
- Contribution to a Periodical (6)
- Course Material (6)
Keywords
- Multimediamarkt (7)
- Enterprise Architecture (5)
- MINLP (5)
- Engineering optimization (4)
- Gamification (4)
- Serious Game (4)
- Auslenkung (3)
- Digitale Transformation (3)
- Digitalisierung (3)
- Education (3)
7T MR Safety
(2021)
Daten und Informationen sind die wichtigsten Ressourcen vieler Unternehmen und müssen daher entsprechend geschützt werden. Getrieben durch die erhöhte Vernetzung von Informationstechnologie, die höhere Offenheit infolge datengetriebener Dienstleistungen und eine starke Zunahme an Datenquellen, rücken die Gefahren von Informationsdiebstahl, -manipulation und -verlust in den Fokus von produzierenden Unternehmen. Auf dem Weg zum lern- und wandlungsfähigen Unternehmen kann dies zu einem großen Hindernis werden, da einerseits zu hohe Sicherheitsanforderungen neue Entwicklungen beschränken, andererseits wegen des Mangels an ausreichenden Informationssicherheitskonzepten Unternehmen weniger Innovationen wagen. Deshalb bedarf es individuell angepasster Konzepte für die Bereiche IT-Security, IT-Safety und Datenschutz für vernetzte Produkte, Produktion und Arbeitsplätze. Bei der Entwicklung und Durchsetzung dieser Konzepte steht der Faktor Mensch im Zentrum aller Überlegungen.
In diesem Kapitel wird dargestellt, wie der Faktor Mensch bei der Erstellung von Informationssicherheitskonzepten in verschiedenen Phasen zu beachten ist. Beginnend mit der Integration von Informationssystemen und damit verbundenen Sicherheitsmaßnahmen, über die Administration, bis hin zur Anwendung durch den Endnutzer, werden Methoden beschrieben, die den Menschen, verbunden mit seinem Mehrwert wie auch den Risiken, einschließen. Dabei werden sowohl Grundlagen aufgezeigt als auch Konzepte vorgestellt, mit denen Entscheider in der Unternehmens-IT Leitlinien für die Informationssicherheit festlegen können.
Universelle hydraulische Try-Out Presse / Teilprojekt A6 / Wiendahl, H.-P. ; Doege, E. ; Engels, E.
(1999)
This article describes the functionality of a MATLAB® library that can be used to develop motion-logic applications in MATLAB programming language for industrial drive and control systems using the well known sercos automation bus. Therewith MATLAB's functionality is extended to designing automation applications from single axis machines up to multi-kinematic robots.
Mit freundlicher Genehmigung der Autoren und des Oldenbourg Industrieverlags https://www.oldenbourg-industrieverlag.de/de/9783835633223-33223 erschienen als Beitrag im Tagungsband zur AALE-Tagung 2012. 9. Fachkonferenz 4.-5. Mai 2012, Aachen, Fachhochschule. ISBN 9783835633223 S 6-2. S. 165-174 Zusammenfassung der Autoren: "Abschließend kann festgehalten werden, dass die gesamten durchgeführten Überlegungen von den Messergebnissen der Inbetriebnahme positiv bestätigt werden. Es wird durch diese Topologie eine maximale Systemverfügbarkeit garantiert und die Zellen werden optimal betrieben. Ein Blick in die Zukunft für dieses Projekt zeigt viele Erweiterungs- und Optimierungsaufgaben. So muss der DC/DC-Wandler für einen bidirektionalen Betrieb erweitert werden, um einen Rekuperationsbetrieb zu gewährleisten bzw. die Zellen laden zu können. Dadurch können gleichzeitig die Gleichrichterdioden im Sekundärleistungsteil durch IGBTs oder MOSFETs ersetzt werden. Die voll funktionsfähige Version des DC/DC-Wandlers könnte höchstwahrscheinlich direkt auf die Anschlusspole der Zelle montiert werden, wodurch das Gewicht gesenkt wird. Durch die verringerte Verlustleistung in den Zuleitungen könnte der Wirkungsgrad noch einmal gesteigert werden.
Mit freundlicher Genehmigung der Autoren und des Oldenbourg Industrieverlags https://www.oldenbourg-industrieverlag.de/de/9783835633223-33223 erschienen als Beitrag im Tagungsband zur AALE-Tagung 2012. 9. Fachkonferenz 4.-5. Mai 2012, Aachen, Fachhochschule. ISBN 9783835633223 S 2-1 S. 37-46 Original-Abstract der Autoren: "Dieser Artikel beschreibt eine universelle Programmierschnittstelle zur Entwicklung von Motion-Logic Applikationen für industrielle Motion-Controller. Sie ermöglicht die Entwicklung von Software mit modernen Hochsprachen, wie C/C++. Visual Basic, C#, Java, Objective-Cund visuellen Programmiersprachen, wie LabVIEW in Koexistenz mit klassischen Programmiersprachen für speicherprogrammierbare Steuerungen nach IEC 61131-3. Daraus ergeben sich neue Möglichkeiten industrielle Antriebs- und Steuerungssysteme mit ihrem IT-Umfeld zu verbinden."
Pressen - Vergleich zwischen Raumlenker- und Exzenterantrieb / Doege, Eckhart ; Engels, Elmar
(1999)
Pressen - Hubzahlregelung an Schnelläuferpressen mit Fuzzy-Reglern / Doege, Eckhart ; Engels, Elmar
(1998)
This paper introduces a hardware setup to measure efficiency maps of low-power electric motors and their associated inverters. Here, the power of the device under test (DUT) ranges from some Watts to a few hundred Watts. The torque and speed of the DUT are measured independent of voltage and current in multiple load points. A Matlab-based software approach in combination with an open Texas-Instruments (TI) hardware setup ensures flexibility. Exemplarily, the efficiency field of a Permanent Magnet Synchronous Machine (PMSM) is measured to proof the concept. Brushless-DC (BLDC) motors can be tested as well. The nomenclature in this paper is based on the new European standard DIN EN 50598. Special attention is paid to the calculation of the measurement error.
In this paper we present SMART-FACTORY, a setup for a research and teaching facility in industrial robotics that is based on the RoboCup Logistics League. It is driven by the need for developing and applying solutions for digital production. Digitization receives constantly increasing attention in many areas, especially in industry. The common theme is to make things smart by using intelligent computer technology. Especially in the last decade there have been many attempts to improve existing processes in factories, for example, in production logistics, also with deploying cyber-physical systems. An initiative that explores challenges and opportunities for robots in such a setting is the RoboCup Logistics League. Since its foundation in 2012 it is an international effort for research and education in an intra-warehouse logistics scenario. During seven years of competition a lot of knowledge and experience regarding autonomous robots was gained. This knowledge and experience shall provide the basis for further research in challenges of future production. The focus of our SMART-FACTORY is to create a stimulating environment for research on logistics robotics, for teaching activities in computer science and electrical engineering programmes as well as for industrial users to study and explore the feasibility of future technologies. Building on a very successful history in the RoboCup Logistics League we aim to provide stakeholders with a dedicated facility oriented at their individual needs.
Der Erfolg eines Softwarenentwicklungsprojektes insbesondere eines Systemintegrationsprojektes wird mit der Erfüllung des „Teufelsdreiecks“, „In-Time“, „In-Budget“, „In-Quality“ gemessen. Hierzu ist die Kenntnis der Software- und Prozessqualität essenziell, um die Einhaltung der Qualitätskriterien festzustellen, aber auch, um eine Vorhersage hinsichtlich Termin- und Budgettreue zu treffen. Zu diesem Zweck wurde in der T-Systems Systems Integration ein System aus verschiedenen Key Performance Indikatoren entworfen und in der Organisation implementiert, das genau das leistet und die Kriterien für CMMI Level 3 erfüllt.
This paper addresses the pixel based recognition of 3D objects with bidirectional associative memories. Computational power and memory requirements for this approach are identified and compared to the performance of current computer architectures by benchmarking different processors. It is shown, that the performance of special purpose hardware, like neurocomputers, is between one and two orders of magnitude higher than the performance of mainstream hardware. On the other hand, the calculation of small neural networks is performed more efficiently on mainstream processors. Based on these results a novel concept is developed, which is tailored for the efficient calculation of bidirectional associative memories. The computational efficiency is further enhanced by the application of algorithms and storage techniques which are matched to characteristics of the application at hand.
This paper describes the realization of a novel neurocomputer which is based on the concepts of a coprocessor. In contrast to existing neurocomputers the main interest was the realization of a scalable, flexible system, which is capable of computing neural networks of arbitrary topology and scale, with full independence of special hardware from the software's point of view. On the other hand, computational power should be added, whenever needed and flexibly adapted to the requirements of the application. Hardware independence is achieved by a run time system which is capable of using all available computing power, including multiple host CPUs and an arbitrary number of neural coprocessors autonomously. The realization of arbitrary neural topologies is provided through the implementation of the elementary operations which can be found in most neural topologies.
In der Vergangenheit basierten große Systemintegrationsprojekte in der Regel auf Individualentwicklungen für einzelne Kunden. Getrieben durch Kostendruck steigt aber der Bedarf nach standardisierten Lösungen, die gleichzeitig die individuellen Anforderungen des jeweiligen Umfelds berücksichtigen. T-Systems GEI GmbH wird beiden Anforderungen mit Produktkerneln gerecht. Neben den technischen Aspekten der Kernelentwicklung spielen besonders organisatorische Aspekte eine Rolle, um Kernel effizient und qualitativ hochwertig zu entwickeln, ohne deren Funktionalitäten ins Uferlose wachsen zu lassen. Umgesetzt hat T-Systems dieses Konzept für Flughafeninformationssysteme. Damit kann dem wachsenden Bedarf der Flughafenbetreiber nach einer effizienten und kostengünstigen Softwarelösung zur Unterstützung Ihrer Geschäftsprozesse entsprochen werden.
This paper addresses the pixel based classification of three dimensional objects from arbitrary views. To perform this task a coding strategy, inspired by the biological model of human vision, for pixel data is described. The coding strategy ensures that the input data is invariant against shift, scale and rotation of the object in the input domain. The image data is used as input to a class of self organizing neural networks, the Kohonen-maps or self-organizing feature maps (SOFM). To verify this approach two test sets have been generated: the first set, consisting of artificially generated images, is used to examine the classification properties of the SOFMs; the second test set examines the clustering capabilities of the SOFM when real world image data is applied to the network after it has been preprocessed to be invariant against shift, scale and rotation. It is shown that the clustering capability of the SOFM is strongly dependant on the invariance coding of the images.
Tool supported requirements analysis for the user centered development of mobile enterprise software
(2008)
A user centered development method has proved satisfactory for the development of mobile enterprise software. To make use of this method, detailed information about the user and the place where the user interacts with his mobile device is required. This article describes how both can be modeled by a stereotypical and conceptual extended UML extension. Finally, a software tool is presented that supports the developed UML extension.
Design and Development of a Hot S-Parameter Measurement System for Plasma and Magnetron Applications
(2020)
This paper presents the design, development and calibration procedures of a novel hot S-parameter measurement system for plasma and magnetron applications with power level up to 6 kW. Based on a vector network analyzer, a power amplifier and two directional couplers, the input matching hotS 11 and transmission hotS 21 of the device under test are measured at 2.45 GHz center frequency and 300MHz bandwidth, while the device is driven by the magnetron. This measurement system opens a new horizon to develop many new industrial applications such as microwave plasma jets, dryer systems, dryers and so forth. Furthermore, the developing, controlling and monitoring a 2kW 2.45GHz plasma jet and a dryer system using the measurement system are presented and explained.
The development and analysis of three waveguides for the exposure of small biological in vitro samples to mobile communication signals at 900 MHz (GSM, Global System for Mobile Communications), 1.8 GHz (GSM), and 2 GHz (UMTS, Universal Mobile Telecommunications System) is presented. The waveguides were based on a fin-line concept and the chamber containing the samples bathed in extracellular solution was placed onto two fins with a slot in between, where the exposure field concentrates. Measures were taken to allow for patch clamp recordings during radiofrequency (RF) exposure. The necessary power for the achievement of the maximum desired specific absorption rate (SAR) of 20 W/kg (average over the mass of the solution) was approximately Pin = 50 mW, Pin = 19 mW, and Pin = 18 mW for the 900 MHz, 1800 MHz, and 2 GHz devices, respectively. At 20 W/kg, a slight RF-induced temperature elevation in the solution of no more than 0.3 °C was detected, while no thermal offsets due to the electromagnetic exposure could be detected at the lower SAR settings (2, 0.2, and 0.02 W/kg). A deviation of 10% from the intended solution volume yielded a calculated SAR deviation of 8% from the desired value. A maximum ±10% variation in the local SAR could occur when the position of the patch clamp electrode was altered within the area where the cells to be investigated were located.
Autonomous agents require rich environment models for fulfilling their missions. High-definition maps are a well-established map format which allows for representing semantic information besides the usual geometric information of the environment. These are, for instance, road shapes, road markings, traffic signs or barriers. The geometric resolution of HD maps can be as precise as of centimetre level. In this paper, we report on our approach of using HD maps as a map representation for autonomous load-haul-dump vehicles in open-pit mining operations. As the mine undergoes constant change, we also need to constantly update the map. Therefore, we follow a lifelong mapping approach for updating the HD maps based on camera-based object detection and GPS data. We show our mapping algorithm based on the Lanelet 2 map format and show our integration with the navigation stack of the Robot Operating System. We present experimental results on our lifelong mapping approach from a real open-pit mine.
Software development projects often fail because of insufficient code quality. It is now well documented that the task of testing software, for example, is perceived as uninteresting and rather boring, leading to poor software quality and major challenges to software development companies. One promising approach to increase the motivation for considering software quality is the use of gamification. Initial research works already investigated the effects of gamification on software developers and come to promising. Nevertheless, a lack of results from field experiments exists, which motivates the chapter at hand. By conducting a gamification experiment with five student software projects and by interviewing the project members, the chapter provides insights into the changing programming behavior of information systems students when confronted with a leaderboard. The results reveal a motivational effect as well as a reduction of code smells.
Meitner-Auger-electron emitters have a promising potential for targeted radionuclide therapy of cancer because of their short range and the high linear energy transfer of Meitner-Auger-electrons (MAE). One promising MAE candidate is 197m/gHg with its half-life of 23.8 h and 64.1 h, respectively, and high MAE yield. Gold nanoparticles (AuNPs) that are labelled with 197m/gHg could be a helpful tool for radiation treatment of glioblastoma multiforme when infused into the surgical cavity after resection to prevent recurrence. To produce such AuNPs, 197m/gHg was embedded into pristine AuNPs. Two different syntheses were tested starting from irradiated gold containing trace amounts of 197m/gHg. When sodium citrate was used as reducing agent, no 197m/gHg labelled AuNPs were formed, but with tannic acid, 197m/gHg labeled AuNPs were produced. The method was optimized by neutralizing the pH (pH = 7) of the Au/197m/gHg solution, which led to labelled AuNPs with a size of 12.3 ± 2.0 nm as measured by transmission electron microscopy. The labelled AuNPs had a concentration of 50 μg (gold)/mL with an activity of 151 ± 93 kBq/mL (197gHg, time corrected to the end of bombardment).
Competence Developing Games (CDGs) are a new concept of how to think about games with serious intentions. In order to emphasize on this topic, a new framework has been developed. It basically relies on learning and motivation theories. This ‘motivational Competence Developing Game Framework’ demonstrates how it is possible to use these theories in a CDG development process. The theoretical derivation and use of the framework is explained in this paper.
With the many achievements of Machine Learning in the past years, it is likely that the sub-area of Deep Learning will continue to deliver major technological breakthroughs [1]. In order to achieve best results, it is important to know the various different Deep Learning frameworks and their respective properties. This paper provides a comparative overview of some of the most popular frameworks. First, the comparison methods and criteria are introduced and described with a focus on computer vision applications: Features and Uses are examined by evaluating papers and articles, Adoption and Popularity is determined by analyzing a data science study. Then, the frameworks TensorFlow, Keras, PyTorch and Caffe are compared based on the previously described criteria to highlight properties and differences. Advantages and disadvantages are compared, enabling researchers and developers to choose a framework according to their specific needs.
Purpose
To design and evaluate a four-channel cardiac transceiver coil array for functional cardiac imaging at 7T.
Materials and Methods
A four-element cardiac transceiver surface coil array was developed with two rectangular loops mounted on an anterior former and two rectangular loops on a posterior former. specific absorption rate (SAR) simulations were performed and a Burn:x-wiley:10531807:media:JMRI22451:tex2gif-stack-1 calibration method was applied prior to obtain 2D FLASH CINE (mSENSE, R = 2) images from nine healthy volunteers with a spatial resolution of up to 1 × 1 × 2.5 mm3.
Results
Tuning and matching was found to be better than 10 dB for all subjects. The decoupling (S21) was measured to be >18 dB between neighboring loops, >20 dB for opposite loops, and >30 dB for other loop combinations. SAR values were well within the limits provided by the IEC. Imaging provided clinically acceptable signal homogeneity with an excellent blood-myocardium contrast applying the Burn:x-wiley:10531807:media:JMRI22451:tex2gif-stack-2 calibration approach.
Conclusion
A four-channel cardiac transceiver coil array for 7T was built, allowing for cardiac imaging with clinically acceptable signal homogeneity and an excellent blood-myocardium contrast. Minor anatomic structures, such as pericardium, mitral, and tricuspid valves and their apparatus, as well as trabeculae, were accurately delineated.
In this paper we report on CO2 Meter, a do-it-yourself carbon dioxide measuring device for the classroom. Part of the current measures for dealing with the SARS-CoV-2 pandemic is proper ventilation in indoor settings. This is especially important in schools with students coming back to the classroom even with high incidents rates. Static ventilation patterns do not consider the individual situation for a particular class. Influencing factors like the type of activity, the physical structure or the room occupancy are not incorporated. Also, existing devices are rather expensive and often provide only limited information and only locally without any networking. This leaves the potential of analysing the situation across different settings untapped. Carbon dioxide level can be used as an indicator of air quality, in general, and of aerosol load in particular. Since, according to the latest findings, SARS-CoV-2 can be transmitted primarily in the form of aerosols, carbon dioxide may be used as a proxy for the risk of a virus infection. Hence, schools could improve the indoor air quality and potentially reduce the infection risk if they actually had measuring devices available in the classroom. Our device supports schools in ventilation and it allows for collecting data over the Internet to enable a detailed data analysis and model generation. First deployments in schools at different levels were received very positively. A pilot installation with a larger data collection and analysis is underway.
Digitale Mobilfunksysteme
(1996)
Digitale Mobilfunksysteme
(2013)
Architecture for platform- and hardware-independent mesh networks : how to unify the channels
(2013)
This paper will prove that mesh networks among different platforms and hardware channels can help to channel valuable information even if public telecommunication infrastructure is not available due to arbitrary reasons. Therefore, results of a simulation for mesh networks on mass events will be provided, followed by the developed architecture and an outlook on future research. The developed architecture is currently being implemented and field tested on mass events.
Die Informationsbroschüre "Konzept für einen spielerischen Ansatz zur multimodalen Mobilitätsplanung“ richtet sich an Spiele- und System-Designer sowie Entwickler. In dieser Broschüre werden mögliche Potenziale im Bereich des allgemeinen Mobilitätsmanagements aufgezeigt, Automobilhersteller vernetzten sich zunehmend mit Technologie-Unternehmen.
Der Telekommunikationsmarkt erfährt substanzielle Veränderungen. Neue Geschäftsmodelle, innovative Dienstleistungen und Technologien erfordern Reengineering, Transformation und Prozessstandardisierung. Mit der Enhanced Telecom Operation Map (eTOM) bietet das TM Forum ein international anerkanntes de facto Referenz-Prozess-Framework basierend auf spezifischen Anforderungen und Ausprägungen der Telekommunikationsindustrie an. Allerdings enthält dieses Referenz-Framework nur eine hierarchische Sammlung von Prozessen auf unterschiedlichen Abstraktionsebenen. Eine Kontrollsicht verstanden als sequenzielle Anordnung von Aktivitäten und daraus resultierend ein realer Prozessablauf fehlt ebenso wie eine Ende-zu-Ende-Sicht auf den Kunden. In diesem Artikel erweitern wir das eTOM-Referenzmodell durch Referenzprozessabläufe, in welchen wir das Wissen über Prozesse in Telekommunikationsunternehmen abstrahieren und generalisieren. Durch die Referenzprozessabläufe werden Unternehmen bei dem strukturierten und transparenten (Re-)Design ihrer Prozesse unterstützt. Wir demonstrieren die Anwendbarkeit und Nützlichkeit unserer Referenzprozessabläufe in zwei Fallstudien und evaluieren diese anhand von Kriterien für die Bewertung von Referenzmodellen. Die Referenzprozessabläufe wurden vom TM Forum in den Standard aufgenommen und als Teil von eTOM Version 9 veröffentlicht. Darüber hinaus diskutieren wir die Komponenten unseres Ansatzes, die auch außerhalb der Telekommunikationsindustrie angewandt werden können.
The potential of electronic markets in enabling innovative product bundles through flexible and sustainable partnerships is not yet fully exploited in the telecommunication industry. One reason is that bundling requires seamless de-assembling and re-assembling of business processes, whilst processes in telecommunication companies are often product-dependent and hard to virtualize. We propose a framework for the planning of the virtualization of processes, intended to assist the decision maker in prioritizing the processes to be virtualized: (a) we transfer the virtualization pre-requisites stated by the Process Virtualization Theory in the context of customer-oriented processes in the telecommunication industry and assess their importance in this context, (b) we derive IT-oriented requirements for the removal of virtualization barriers and highlight their demand on changes at different levels of the organization. We present a first evaluation of our approach in a case study and report on lessons learned and further steps to be performed.
Die Veränderungen des Telekommunikationsmarktes haben in der Praxis zu einer Vielzahl von Transformationsprojekten geführt. Was gehört aber zu einem “Transformationsprojekt”, welche Prozesse und Systeme werden verändert? Zur Beantwortung dieser Frage haben wir 184 Berichte zu Projekten analysiert, die als "Transformationsprojekte" bezeichnet waren. Für die Analyse haben wir einen Kodierungsrahmen konzipiert und anhand dessen die Berichte mit einem hierarchischen Clustering-Verfahren in Themen gruppiert. Die Ergebnisse liefern Hinweise über die in der Praxis gesetzten Schwerpunkte und Prioritäten. Sie können
somit als Unterstützung für Unternehmen dienen, die ein Transformationsprojekt planen. Sie weisen zudem darauf hin, in welchen Bereichen eines Unternehmens Unterstützung durch wissenschaftlich erprobte Werkzeuge und Modelle nötig ist.
Market changes have forced telecommunication companies to transform their business. Increased competition, short innovation cycles, changed usage patterns, increased customer expectations and cost reduction are the main drivers. Our objective is to analyze to what extend transformation projects have improved the orientation towards the end-customers. Therefore, we selected 38 real-life case studies that are dealing with customer orientation. Our analysis is based on a telecommunication-specific framework that aligns strategy, business processes and information systems. The result of our analysis shows the following: transformation projects that aim to improve the customer orientation are combined with clear goals on costs and revenue of the enterprise. These projects are usually directly linked to the customer touch points, but also to the development and provisioning of products. Furthermore, the analysis shows that customer orientation is not the sole trigger for transformation. There is no one-fits-all solution; rather, improved customer orientation needs aligned changes of business processes as well as information systems related to different parts of the company.
As the potential of a next generation network (NGN) is recognised, telecommunication companies consider switching to it. Although the implementation of an NGN seems to be merely a modification of the network infrastructure, it may trigger or require changes in the whole company, because it builds upon the separation between service and transport, a flexible bundling of services to products and the streamlining of the IT infrastructure. We propose a holistic framework, structured into the layers ‘strategy’, ‘processes’ and ‘information systems’ and incorporate into each layer all concepts necessary for the implementation of an NGN, as well as the alignment of these concepts. As a first proof-of-concept for our framework we have performed a case study on the introduction of NGN in a large telecommunication company; we show that our framework captures all topics that are affected by an NGN implementation.
Subject of this case is Deutsche Telekom Services Europe (DTSE), a service center for administrative processes. Due to the high volume of repetitive tasks (e.g., 100k manual uploads of offer documents into SAP per year), automation was identified as an important strategic target with a high management attention and commitment. DTSE has to work with various backend application systems without any possibility to change those systems. Furthermore, the complexity of administrative processes differed. When it comes to the transfer of unstructured data (e.g., offer documents) to structured data (e.g., MS Excel files), further cognitive technologies were needed.
How does the implementation of a next generation network influence a telecommunication company?
(2009)
As the potential of a Next Generation Network (NGN) is recognized, telecommunication companies consider switching to it. Although the implementation of an NGN seems to be merely a modification of the network infrastructure, it may trigger or require changes in the whole company and even influence the company strategy. To capture the effects of NGN we propose a framework based on concepts of business engineering and technical recommendations for the introduction of NGN technology. The specific design of solutions for the layers "Strategy", "Processes" and "Information Systems" as well as their interdependencies are an essential characteristic of the developed framework. We have per-formed a case study on NGN implementation and observed that all layers captured by our framework are influenced by the introduction of an NGN.
Unternehmen sind in der Regel überzeugt, dass sie die Bedürfnisse ihrer Kunden in den Mittelpunkt stellen. Aber in der direkten Interaktion mit dem Kunden zeigen sie häufig Schwächen. Der folgende Beitrag illustriert, wie durch eine konsequente Ausrichtung der Wertschöpfungsprozesse auf die zentralen Kundenbedürfnisse ein Dreifacheffekt erzielt werden kann: Nachhaltig erhöhte Kundenzufriedenheit, gesteigerte Effizienz und eine Differenzierung im Wettbewerb.
Kundenanforderungen an Netzwerke haben sich in den vergangenen Jahren stark verändert. Mit NFV und SDN sind Unternehmen technisch in der Lage, diesen gerecht zu werden. Die Provider stehen jedoch vor großen Herausforderungen: Insbesondere Produkte und Prozesse müssen angepasst und agiler werden, um die Stärken von NFV und SDN zum Kundenvorteil auszuspielen.
Robotic process automation (RPA) has attracted increasing attention in research and practice. This chapter positions, structures, and frames the topic as an introduction to this book. RPA is understood as a broad concept that comprises a variety of concrete solutions. From a management perspective RPA offers an innovative approach for realizing automation potentials, whereas from a technical perspective the implementation based on software products and the impact of artificial intelligence (AI) and machine learning (ML) are relevant. RPA is industry-independent and can be used, for example, in finance, telecommunications, and the public sector. With respect to RPA this chapter discusses definitions, related approaches, a structuring framework, a research framework, and an inside as well as outside architectural view. Furthermore, it provides an overview of the book combined with short summaries of each chapter.
The telecommunications industry is currently going through a major transformation. In this context, the enhanced Telecom Operations Map (eTOM) is a domain-specific process reference model that is offered by the industry organization TM Forum. In practice, eTOM is well accepted and confirmed as de facto standard. It provides process definitions and process flows on different levels of detail. This article discusses the reference modeling of eTOM, i.e., the design, the resulting artifact, and its evaluation based on three project cases. The application of eTOM in three projects illustrates the design approach and concrete models on strategic and operational levels. The article follows the Design Science Research (DSR) paradigm. It contributes with concrete design artifacts to the transformational needs of the telecommunications industry and offers lessons-learned from a general DSR perspective.
Am Beispiel der Telekommunikationsindustrie zeigt der Beitrag eine konkrete Ausgestaltung anwendungsorientierter Forschung, die sowohl für die Praxis als auch für die Wissenschaft nutzen- und erkenntnisbringend ist. Forschungsgegenstand sind die Referenzmodelle des Industriegremiums TM Forum, die von vielen Telekommunikationsunternehmen zur Transformation ihrer Strukturen und Systeme genutzt werden. Es wird die langjährige Forschungstätigkeit bei der Weiterentwicklung und Anwendung dieser Referenzmodelle beschrieben. Dabei wird ein konsequent gestaltungsorientierter Forschungsansatz verfolgt. Das Zusammenspiel aus kontinuierlicher Weiterentwicklung in Zusammenarbeit mit einem Industriegremium und der Anwendung in vielfältigen Praxisprojekten führt zu einer erfolgreichen Symbiose aus praktischer Nutzengenerierung sowie wissenschaftlichem Erkenntnisgewinn. Der Beitrag stellt den gewählten Forschungsansatz anhand konkreter Beispiele vor. Darauf basierend werden Empfehlungen und Herausforderungen für eine gestaltungs- und praxisorientierte Forschung diskutiert.
This book reflects the tremendous changes in the telecommunications industry in the course of the past few decades – shorter innovation cycles, stiffer competition and new communication products. It analyzes the transformation of processes, applications and network technologies that are now expected to take place under enormous time pressure. The International Telecommunication Union (ITU) and the TM Forum have provided reference solutions that are broadly recognized and used throughout the value chain of the telecommunications industry, and which can be considered the de facto standard. The book describes how these reference solutions can be used in a practical context: it presents the latest insights into their development, highlights lessons learned from numerous international projects and combines them with well-founded research results in enterprise architecture management and reference modeling. The complete architectural transformation is explained, from the planning and set-up stage to the implementation. Featuring a wealth of examples and illustrations, the book offers a valuable resource for telecommunication professionals, enterprise architects and project managers alike.
Im Rahmen der Digitalisierung ist die zunehmende Automatisierung von bisher manuellen Prozessschritten ein Aspekt, der massive Auswirkungen auf die zukünftige Arbeitswelt haben wird. In diesem Kontext werden an den Einsatz von Softwarerobotern zur Prozessautomatisierung hohe Erwartungen geknüpft. Bei den Implementierungsansätzen wird die Diskussion aktuell insbesondere durch Robotic Process Automation (RPA) und Chatbots geprägt. Beide Ansätze verfolgen das gemeinsame Ziel einer 1:1-Automatisierung von menschlichen Handlungen und dadurch ein direktes Ersetzen von Mitarbeitern durch Maschinen. Bei RPA werden Prozesse durch Softwareroboter erlernt und automatisiert ausgeführt. Dabei emulieren RPA-Roboter die Eingaben auf der bestehenden Präsentationsschicht, so dass keine Änderungen an vorhandenen Anwendungssystemen notwendig sind. Am Markt werden bereits unterschiedliche RPA-Lösungen als Softwareprodukte angeboten. Durch Chatbots werden Ein- und Ausgaben von Anwendungssystemen über natürliche Sprache realisiert. Dadurch ist die Automatisierung von unternehmensexterner Kommunikation (z. B. mit Kunden) aber auch von unternehmensinternen Assistenztätigkeiten möglich. Der Beitrag diskutiert die Auswirkungen von Softwarerobotern auf die Arbeitswelt anhand von Anwendungsbeispielen und erläutert die unternehmensindividuelle Entscheidung über den Einsatz von Softwarerobotern anhand von Effektivitäts- und Effizienzzielen.
Im Rahmen der digitalen Transformation werden innovative Technologiekonzepte, wie z. B. das Internet der Dinge und Cloud Computing als Treiber für weitreichende Veränderungen von Organisationen und Geschäftsmodellen angesehen. In diesem Kontext ist Robotic Process Automation (RPA) ein neuartiger Ansatz zur Prozessautomatisierung, bei dem manuelle Tätigkeiten durch sogenannte Softwareroboter erlernt und automatisiert ausgeführt werden. Dabei emulieren Softwareroboter die Eingaben auf der bestehenden Präsentationsschicht, so dass keine Änderungen an vorhandenen Anwendungssystemen notwendig sind. Die innovative Idee ist die Transformation der bestehenden Prozessausführung von manuell zu digital, was RPA von traditionellen Ansätzen des Business Process Managements (BPM) unterscheidet, bei denen z. B. prozessgetriebene
Anpassungen auf Ebene der Geschäftslogik notwendig sind. Am Markt werden bereits unterschiedliche RPA-Lösungen als Softwareprodukte angeboten. Gerade bei operativen Prozessen mit sich wiederholenden Verarbeitungsschritten in unterschiedlichen Anwendungssystemen sind gute Ergebnisse durch RPA dokumentiert, wie z. B. die Automatisierung von 35 % der Backoffice-Prozesse bei Telefonica. Durch den vergleichsweise niedrigen Implementierungsaufwand verbunden mit einem hohen Automatisierungspotenzial ist in der Praxis (z. B. Banken, Telekommunikation, Energieversorgung) ein hohes Interesse an RPA vorhanden. Der Beitrag diskutiert RPA als innovativen Ansatz zur
Prozessdigitalisierung und gibt konkrete Handlungsempfehlungen für die Praxis. Dazu wird zwischen modellgetriebenen und selbstlernenden Ansätzen unterschieden. Anhand von generellen Architekturen von RPA-Systemen werden Anwendungsszenarien sowie deren Automatisierungspotenziale, aber auch Einschränkungen, diskutiert. Es folgt ein strukturierter Marktüberblick ausgewählter RPA-Produkte. Anhand von drei konkreten Anwendungsbeispielen wird die Nutzung von RPA in der Praxis verdeutlicht.
Zur Unterstützung des Transformationsbedarfs von Telekommunikationsunternehmen sind die Referenzmodelle des TM Forums in der Praxis weltweit anerkannt. Dabei findet jedoch meist eine losgelöste Nutzung für spezifische Einzelthemen statt. Daher führt dieser Artikel die bestehenden Inhalte in einer industriespezifischen, übergreifenden Referenzarchitektur zusammen. Der Fokus liegt auf den Ebenen Aufbauorganisation, Prozesse, Applikationen und Daten. Darüber hinaus werden inhaltliche Architekturdomänen zur Strukturierung angeboten. Die Referenzarchitektur ist hierarchisch aufgebaut und wird hier beispielhaft für ausgewählte, aggregierte Inhalte beschrieben. Als erste Evaluation wird die Anwendung der Referenzarchitektur in drei Praxisprojekten erläutert.
Durch die Fragmentierung von Wertschöpfungsketten ergeben sich neue Herausforderungen für das Management von Kundenbeziehungen. Die Dissertation untersucht die daraus resultierenden Anforderungen an eine übergreifende Integration von Customer Relationship Management in der
Telekommunikationsindustrie. Ziel ist es, durch Anwendung von Methoden eines Enterprise Architecture Framework eine übergreifend Lösung zu gestalten. Grundlegende Prämisse dabei ist, dass die übergreifende Gestaltung eines Customer Relationship Management für alle an der
Wertschöpfung beteiligten Unternehmen vorteilhaft ist.
Die Telekommunikationsindustrie hat in den letzten Jahrzehnten einen enormen Wandel vollzogen. Für Telekommunikationsunternehmen erfordert dies fundamentale Umstrukturierungen von Strategie, Prozessen, Anwendungssystemen und Netzwerktechnologien. Dabei spielen Unternehmensarchitekturen und Referenzmodelle eine wichtige Rolle. Zwar existieren in der Praxis anerkannte Referenzmodelle, aber wie sind diese für eine systematische Transformation zu gestalten? Wie sieht eine konkrete Lösung für die Telekommunikationsindustrie aus?
Als Antwort stellt Christian Czarnecki in seinem Buch eine referenzmodellbasierte Unternehmensarchitektur vor. Basierend auf einer umfangreichen Untersuchung von Transformationsprojekten werden Probleme und Anforderungen der Praxis identifiziert, für die mit Methoden der Unternehmenstransformation, Referenzmodellierung und Unternehmensarchitektur ein Lösungsvorschlag entwickelt und evaluiert wird. Dieser besteht u. a. aus detaillierten Anwendungsfällen, Referenzprozessabläufen, einer Zuordnung von Prozessen zu Anwendungssystemen sowie Handlungsempfehlungen zur Virtualisierung.
Für Wissenschaftler und Studierende der Wirtschaftsinformatik zeigt das Buch neue Erkenntnisse einer anwendungsorientierten Referenzmodellierung. Für Praktiker liefert es eine methodisch fundierte Lösung für die aktuellen Transformationsbedarfe der Telekommunikationsindustrie. Christian Czarnecki arbeitet seit 2004 als Unternehmensberater und hat viele Telekommunikationsunternehmen bei deren Transformation begleitet. In 2013 erfolgte die Promotion zum Doktoringenieur an der Otto-von-Guericke-Universität Magdeburg.
Because of customer churn, strong competition, and operational inefficiencies, the telecommunications operator ME Telco (fictitious name due to confidentiality) launched a strategic transformation program that included a Business Process Management (BPM) project. Major problems were silo-oriented process management and missing cross-functional transparency. Process improvements were not consistently planned and aligned with corporate targets. Measurable inefficiencies were observed on an operational level, e.g., high lead times and reassignment rates of the incident management process.
Intelligent autonomous software robots replacing human activities and performing administrative processes are reality in today’s corporate world. This includes, for example, decisions about invoice payments, identification of customers for a marketing campaign, and answering customer complaints. What happens if such a software robot causes a damage? Due to the complete absence of human activities, the question is not trivial. It could even happen that no one is liable for a damage towards a third party, which could create an uncalculatable legal risk for business partners. Furthermore, the implementation and operation of those software robots involves various stakeholders, which result in the unsolvable endeavor of identifying the originator of a damage. Overall it is advisable to all involved parties to carefully consider the legal situation. This chapter discusses the liability of software robots from an interdisciplinary perspective. Based on different technical scenarios the legal aspects of liability are discussed.
Automated driving is now possible in diverse road and traffic conditions. However, there are still situations that automated vehicles cannot handle safely and efficiently. In this case, a Transition of Control (ToC) is necessary so that the driver takes control of the driving. Executing a ToC requires the driver to get full situation awareness of the driving environment. If the driver fails to get back the control in a limited time, a Minimum Risk Maneuver (MRM) is executed to bring the vehicle into a safe state (e.g., decelerating to full stop). The execution of ToCs requires some time and can cause traffic disruption and safety risks that increase if several vehicles execute ToCs/MRMs at similar times and in the same area. This study proposes to use novel C-ITS traffic management measures where the infrastructure exploits V2X communications to assist Connected and Automated Vehicles (CAVs) in the execution of ToCs. The infrastructure can suggest a spatial distribution of ToCs, and inform vehicles of the locations where they could execute a safe stop in case of MRM. This paper reports the first field operational tests that validate the feasibility and quantify the benefits of the proposed infrastructure-assisted ToC and MRM management. The paper also presents the CAV and roadside infrastructure prototypes implemented and used in the trials. The conducted field trials demonstrate that infrastructure-assisted traffic management solutions can reduce safety risks and traffic disruptions.
This paper describes the potential for developing a digital twin of society- a dynamic model that can be used to observe, analyze, and predict the evolution of various societal aspects. Such a digital twin can help governmental agencies and policy makers in interpreting trends, understanding challenges, and making decisions regarding investments or policies necessary to support societal development and ensure future prosperity. The paper reviews related work regarding the digital twin paradigm and its applications. The paper presents a motivating case study- an analysis of opportunities and challenges faced by the German federal employment agency, Bundesagentur f¨ur Arbeit (BA), proposes solutions using digital twins, and describes initial proofs of concept for such solutions.
Objective
This study assesses and quantifies impairment of postoperative magnetic resonance imaging (MRI) at 7 Tesla (T) after implantation of titanium cranial fixation plates (CFPs) for neurosurgical bone flap fixation.
Materials and methods
The study group comprised five patients who were intra-individually examined with 3 and 7 T MRI preoperatively and postoperatively (within 72 h/3 months) after implantation of CFPs. Acquired sequences included T₁-weighted magnetization-prepared rapid-acquisition gradient-echo (MPRAGE), T₂-weighted turbo-spin-echo (TSE) imaging, and susceptibility-weighted imaging (SWI). Two experienced neurosurgeons and a neuroradiologist rated image quality and the presence of artifacts in consensus reading.
Results
Minor artifacts occurred around the CFPs in MPRAGE and T2 TSE at both field strengths, with no significant differences between 3 and 7 T. In SWI, artifacts were accentuated in the early postoperative scans at both field strengths due to intracranial air and hemorrhagic remnants. After resorption, the brain tissue directly adjacent to skull bone could still be assessed. Image quality after 3 months was equal to the preoperative examinations at 3 and 7 T.
Conclusion
Image quality after CFP implantation was not significantly impaired in 7 T MRI, and artifacts were comparable to those in 3 T MRI.
One central challenge for self-driving cars is a proper path-planning. Once a trajectory has been found, the next challenge is to accurately and safely follow the precalculated path. The model-predictive controller (MPC) is a common approach for the lateral control of autonomous vehicles. The MPC uses a vehicle dynamics model to predict the future states of the vehicle for a given prediction horizon. However, in order to achieve real-time path control, the computational load is usually large, which leads to short prediction horizons. To deal with the computational load, the control algorithm can be parallelized on the graphics processing unit (GPU). In contrast to the widely used stochastic methods, in this paper we propose a deterministic approach based on grid search. Our approach focuses on systematically discovering the search area with different levels of granularity. To achieve this, we split the optimization algorithm into multiple iterations. The best sequence of each iteration is then used as an initial solution to the next iteration. The granularity increases, resulting in smooth and predictable steering angle sequences. We present a novel GPU-based algorithm and show its accuracy and realtime abilities with a number of real-world experiments.