Conference Proceeding
Refine
Year of publication
Document Type
- Conference Proceeding (130) (remove)
Language
- English (130) (remove)
Has Fulltext
- yes (130) (remove)
Keywords
- Biosensor (25)
- CAD (7)
- Finite-Elemente-Methode (7)
- civil engineering (7)
- Bauingenieurwesen (6)
- Blitzschutz (6)
- Clusterion (4)
- Sonde (4)
- Telekommunikationsmarkt (4)
- Air purification (3)
- Einspielen <Werkstoff> (3)
- Eisschicht (3)
- Elektromagnetischer Schutzschild (3)
- Führung (3)
- Hämoglobin (3)
- Leadership (3)
- Lightning (3)
- Limit analysis (3)
- Luftreiniger (3)
- Plasmacluster ion technology (3)
Institute
- Fachbereich Medizintechnik und Technomathematik (77)
- IfB - Institut für Bioengineering (38)
- Fachbereich Energietechnik (19)
- Fachbereich Maschinenbau und Mechatronik (9)
- Fachbereich Wirtschaftswissenschaften (9)
- Fachbereich Luft- und Raumfahrttechnik (7)
- Fachbereich Elektrotechnik und Informationstechnik (6)
- INB - Institut für Nano- und Biotechnologien (6)
- MASKOR Institut für Mobile Autonome Systeme und Kognitive Robotik (4)
- Fachbereich Bauingenieurwesen (3)
- Solar-Institut Jülich (2)
- Fachbereich Gestaltung (1)
Novel organic membrane-based thin-film microsensors for the determination of heavy metal cations
(2006)
A first step towards the fabrication and electrochemical evaluation of thin-film microsensors based on organic PVC membranes for the determination of Hg(II), Cd(II), Pb(II) and Cu(II) ions in solutions has been realised. The membrane-coating mixture used in the preparation of this new type of microsensors is incorporating PVC as supporting matrix, o-nitrophenyloctylether (o-NPOE) as solvent mediator and a recently synthesized Hg[dimethylglyoxime(phene)]2+ and Bis-(4-hydroxyacetophenone)-ethylenediamine as electroactive materials for Hg(II) and Cd(II), respectively. A set of three commercialised ionophores for Cd(II), Pb(II) and Cu(II) has been also used for comparison. Thin-film microsensors based on these membranes showed a Nernstian response of slope (26-30 mV/dec.) for the respective tested cations. The potentiometric response characteristics (linear range, pH range, detection limit and response time) are comparable with those obtained by conventional membranes as well as coated wire electrodes prepared from the same membrane. The realisation of the new organic membrane-based thin-film microsensors overcomes the problem of an insufficient selectivity of solid-state-based thinfilm sensors.
One of interesting but not well known water properties is related to appearance of highly ordered structures in response to strong electrical field. In 1893 Sir William Armstrong placed a cotton thread between two wine glasses filled with chemically pure water. When high DC voltage was applied between the glasses, a connection consisting of water formed, producing a "water bridge"
Tests with palm tree leaves have just started yet and scan data are in the process to be analyzed. The final goal of future project for palm tree gender and species recognition will be to develop optical scanning technology to be applied to date palm tree leaves for in–situ screening purposes. Depending on the software used and the particular requirements of the users the technology potentially shall be able to identify palm tree diseases, palm tree gender, and species of young date palm trees by scanning leaves.
Recent analysis of scientific data from Cassini and earth-based observations gave evidence for a global ocean under a surrounding solid ice shell on Saturn's moon Enceladus. Images of Enceladus' South Pole showed several fissures in the ice shell with plumes constantly exhausting frozen water particles, building up the E-Ring, one of the outer rings of Saturn. In this southern region of Enceladus, the ice shell is considered to be as thin as 2 km, about an order of magnitude thinner than on the rest of the moon. Under the ice shell, there is a global ocean consisting of liquid water. Scientists are discussing different approaches the possibilities of taking samples of water, i.e. by melting through the ice using a melting probe. FH Aachen UAS developed a prototype of maneuverable melting probe which can navigate through the ice that has already been tested successfully in a terrestrial environment. This means no atmosphere and or ambient pressure, low ice temperatures of around 100 to 150K (near the South Pole) and a very low gravity of 0,114 m/s^2 or 1100 μg. Two of these influencing measures are about to be investigated at FH Aachen UAS in 2017, low ice temperature and low ambient pressure below the triple point of water. Low gravity cannot be easily simulated inside a large experiment chamber, though. Numerical simulations of the melting process at RWTH Aachen however are showing a gravity dependence of melting behavior. Considering this aspect, VIPER provides a link between large-scale experimental simulations at FH Aachen UAS and numerical simulations at RWTH Aachen. To analyze the melting process, about 90 seconds of experiment time in reduced gravity and low ambient pressure is provided by the REXUS rocket. In this time frame, the melting speed and contact force between ice and probes are measured, as well as heating power and a two-dimensional array of ice temperatures. Additionally, visual and infrared cameras are used to observe the melting process.
A multi-functional device applying for the safe maintenance at high-altitude on wind turbines
(2015)
The propagation of mechanical waves in plates of isotropic elastic material is investigated. After a short introduction to the understanding of focussing of stress waves in a plate with a curved boundary the method of characteristics is applied to a plate of hyperelastic material. Using this method the propagation of acceleration waves is discussed. Based on this a numerical difference scheme is developed for solving initial-boundary-value problems and applied to two examples: propagation of a point disturbance in a homogeneously finitely strained non-linear elastic plate and geometrical focussing in al linear elastic plate.
The applicability of differential pulse voltammetry (DPV) and adsorptive stripping voltammetry (AdSV) at a non-toxic meniscus-modified silver solid amalgam electrode (m-AgSAE) for the determination of trace amounts of genotoxic substances was demonstrated on the determination of micromolar and submicromolar concentrations of 3-nitrofluoranthene using methanol - 0.01 mol L-1 NaOH (9:1) mixture as a base electrolyte and of Ostazine Orange using 0.01 mol L-1 NaOH as a base electrolyte.
This paper presents a two-dimensional-in-space mathematical model of biosensors based on an array of enzyme microreactors immobilised on a single electrode. The modeling system acts under amperometric conditions. The microreactors were modeled by particles and by strips. The model is based on the diffusion equations containing a nonlinear term related to the Michaelis-Menten kinetics of the enzymatic reaction. The model involves three regions: an array of enzyme microreactors where enzyme reaction as well as mass transport by diffusion takes place, a diffusion limiting region where only the diffusion takes place, and a convective region, where the analyte concentration is maintained constant. Using computer simulation, the influence of the geometry of the microreactors and of the diffusion region on the biosensor response was investigated. The digital simulation was carried out using the finite difference technique.
As a deduction from these results, we can conclude that proteins mainly in vitro, denaturate totally at a temperature between 57°C -62°C, and they also affected by NO and different ions types. In which mainly, NO cause earlier protein denaturation, which means that, NO has a destabilizing effect on proteins, and also different ions will alter the protein denaturation in which, some ions will cause earlier protein denaturation while others not.
This paper compares several blade element theory (BET) method-based propeller simulation tools, including an evaluation against static propeller ground tests and high-fidelity Reynolds-Average Navier Stokes (RANS) simulations. Two proprietary propeller geometries for paraglider applications are analysed in static and flight conditions. The RANS simulations are validated with the static test data and used as a reference for comparing the BET in flight conditions. The comparison includes the analysis of varying 2D aerodynamic airfoil parameters and different induced velocity calculation methods. The evaluation of the BET propeller simulation tools shows the strength of the BET tools compared to RANS simulations. The RANS simulations underpredict static experimental data within 10% relative error, while appropriate BET tools overpredict the RANS results by 15–20% relative error. A variation in 2D aerodynamic data depicts the need for highly accurate 2D data for accurate BET results. The nonlinear BET coupled with XFOIL for the 2D aerodynamic data matches best with RANS in static operation and flight conditions. The novel BET tool PropCODE combines both approaches and offers further correction models for highly accurate static and flight condition results.
A 3D finite element model of the female pelvic floor for the reconstruction of urinary incontinence
(2014)
The Scarab Project
(2015)
Urban Search and Rescue (USAR) is an active research
field in the robotics community. Despite recent advances
for many open research questions, these kind of systems are
not widely used in real rescue missions. One reason is that such
systems are complex and not (yet) very reliable; another is that
one has to be an robotic expert to run such a system. Moreover,
available rescue robots are very expensive and the benefits of
using them are still limited.
In this paper, we present the Scarab robot, an alternative
design for a USAR robot. The robot is light weight, humanpackable
and its primary purpose is that of extending the
rescuer’s capability to sense the disaster site. The idea is that a
responder throws the robot to a certain spot. The robot survives
the impact with the ground and relays sensor data such as
camera images or thermal images to the responder’s hand-held
control unit from which the robot can be remotely controlled.
The Ministry of Science and Research in North Rhine-Westphalia created eight platforms of excellence, one in the research area „Energy and Environment“ in 2002 at ACUAS. This platform concentrates the research and development of 13 professors in Jülich and Aachen and of two scientific institutes with different topics: – NOWUM-Energy with emphasis on efficient and economic energy conversion – The Solar Institute Jülich – SIJ – being the largest research institute in the field of renewables at a University of Applied Sciences in Germany With this platform each possible energy conversion – nuclear, fossil, renewable- can be dealt with to help solving the two most important problems of mankind, energy and potable water. At the CSE are presented the historical development, some research results and the combined master studies in „Energy Systems“ and „Nuclear Applications“
One of the most important parameters in a burning chamber - in power stations, in waste to energy plants - is the temperature. This temperature is in the range of 700-1500 °C - one of the most advanced measuring methods being the acoustic pyrometry with the possibility of producing temperature mapping in one level of the burning chamber - comparable to computer tomography. The results of these measurements discussed in the presentation can be used - to fulfil the legal requirements in the FRG or in the EU - to equalise the temperature in one level of the burning chamber to optimise the steam production (better efficiency of the plant) and to minimise the production of temperature controlled flue gas components (NO, CO a. o.) - to control the SNCR-process if used.
Multi-interface level sensors and new development in monitoring and control of oil separators
(2006)
In the oil industry, huge saving may be made if suitable multi-interface level measurement systems are employed for effectively monitoring crude oil separators and efficient control of their operation. A number of techniques, e.g. externally mounted displacers, differential pressure transmitters and capacitance rod devices, have been developed to measure the separation process with gas, oil, water and other components. Because of the unavailability of suitable multi-interface level measurement systems, oil separators are currently operated by the trial-and-error approach. In this paper some conventional techniques, which have been used for level measurement in industry, and new development are discussed.
A multi-sensor system is a chemical sensor system which quantitatively and qualitatively records gases with a combination of cross-sensitive gas sensor arrays and pattern recognition software. This paper addresses the issue of data analysis for identification of gases in a gas sensor array. We introduce a software tool for gas sensor array configuration and simulation. It concerns thereby about a modular software package for the acquisition of data of different sensors. A signal evaluation algorithm referred to as matrix method was used specifically for the software tool. This matrix method computes the gas concentrations from the signals of a sensor array. The software tool was used for the simulation of an array of five sensors to determine gas concentration of CH4, NH3, H2, CO and C2H5OH. The results of the present simulated sensor array indicate that the software tool is capable of the following: (a) identify a gas independently of its concentration; (b) estimate the concentration of the gas, even if the system was not previously exposed to this concentration; (c) tell when a gas concentration exceeds a certain value. A gas sensor data base was build for the configuration of the software. With the data base one can create, generate and manage scenarios and source files for the simulation. With the gas sensor data base and the simulation software an on-line Web-based version was developed, with which the user can configure and simulate sensor arrays on-line.
The ”IceMole“ is a novel maneuverable subsurface ice probe for clean in-situ analysis and sampling of subsurface ice and subglacial water/brine. It is developed and build at FH Aachen University of Applied Sciences’ Astronautical Laboratory. A first prototype was successfully tested on the Swiss Morteratsch glacier in 2010. Clean sampling is achieved with a hollow ice screw (as it is used in mountaineering) at the tip of the probe. Maneuverability is achieved with a differentially heated melting head. Funded by the German Space Agency (DLR), a consortium led by FH Aachen currently develops a much more advanced IceMole probe, which includes a sophisticated system for obstacle avoidance, target detection, and navigation in the ice. We intend to use this probe for taking clean samples of subglacial brine at the Blood Falls (McMurdo Dry Valleys, East Antarctica) for chemical and microbiological analysis. In our conference contribution, we 1) describe the IceMole design, 2) report the results of the field tests of the first prototype on the Morteratsch glacier, 3) discuss the probe’s potential for the clean in-situ analysis and sampling of subsurface ice and subglacial liquids, and 4) outline the way ahead in the development of this technology.
We present the novel concept of a combined drilling and melting probe for subsurface ice research. This probe, named “IceMole”, is currently developed, built, and tested at the FH Aachen University of Applied Sciences’ Astronautical Laboratory. Here, we describe its first prototype design and report the results of its field tests on the Swiss Morteratsch glacier. Although the IceMole design is currently adapted to terrestrial glaciers and ice shields, it may later be modified for the subsurface in-situ investigation of extraterrestrial ice, e.g., on Mars, Europa, and Enceladus. If life exists on those bodies, it may be present in the ice (as life can also be found in the deep ice of Earth).
Evaluation of fragility curves for a three-storey-reinforced-concrete mock-up of SMART 2013 project
(2016)
The paper deals with the development of the probabilistic approach to the assessment of risk due to lightning. Sources of damage, types of damage and types of loss are defined and, accordingly, the procedure for risk analysis and the way of assessment of different risk components is proposed. The way to evaluate the influence of different protection measures (lightning protection system; shielding of structure, cables and equipment; routing of internal wiring; surge protective device) in reducing such probabilities is considered. The paper has been prepared within the framework of the activity of IEC TC81-WG9/CLC TC81-WG4 directed to prepare the draft IEC 62305-2 Risk Management, in cooperation with the Secretary of IEC/CLC TC81.
A melting probe equipped with autofluorescence-based detection system combined with a light scattering unit, and, optionally, with a microarray chip would be ideally suited to probe icy environments like Europa’s ice layer as well as the polar ice layers of Earth and Mars for recent and extinct live.
Mechanical stimulation of the cells resulted in evident changes in the cell morphology, protein composition and gene expression. Microscopically, additional formation of stress fibers accompanied by cell re-arrangements in a monolayer was observed. Also, significant activation of p53 gene was revealed as compared to control. Interestingly, the use of CellTech membrane coating induced cell death after mechanical stress had been applied. Such an effect was not detected when fibronectin had been used as an adhesion substrate.
Recently, SHARP corporation has developed the world’s first “Plasma Cluster Ions (PCI)” air purification technology, which uses plasma discharge to generate cluster ions. The new plasma cluster device releases into the air positive and negative ions, which are harmless to humans and are able to decompose and deactivate airborne substances by chemical reactions. A lot of phenomenological tests of the PCI air purification technology on microbial cells have been conducted. And, in most cases, it has been shown that PCI demonstrate strongly pronounced killing effect. Although, the particular mechanisms of PCI action are still not evident. We studied variations in resistance to PCI among gram-positive airborne microorganisms, as well as some dose-dependent, spatial, cultural and biochemical properties of PCI action in respect of Staphylococcus spp, Enterococcus spp, Micrococcus spp.
Summary and Conclusions PCIs were clearly effective in terms of their antibacterial effects with the strains tested. This efficacy increased with the time the bacteries were exposed to PCIs. The bactericidal action has proved to be irreversible. PCIs were significantly less effective in shadowed areas. PCI exposure caused multiple protein damages as observed in SDS PAGE studies. There was no single but multiple molecular mechanism causing the bacterial death.
Recently, the SHARP Corporation, Japan, has developed the world’s first "Plasma Cluster Ions (PCI)" air purification technology using plasma discharge to generate cluster ions. The new plasma cluster device releases positive and negative ions into the air, which are able to decompose and deactivate harmful airborne substances by chemical reactions. Because cluster ions consist of positive and negative ions that normally exist in the natural world, they are completely harmless and safe to humans. The amount of ozone generated by cluster ions is less than 0.01 ppm, which is significantly less than the 0.05-ppm standard for industrial operations and consumer electronics. This amount, thus, has no harming effects whatsoever on the human body. But particular properties and chemical processes in PCI treatment are still under study. It has been shown that PCI in most cases show strongly pronounced irreversible killing effects in respect of airborne microflora due to free-radical induced reactions and can be considered as a potent technology to disinfect both home, medical and industrial appliances.
Recently, SHARP corporation has developed the world’s first "Plasma Cluster Ions® (PCI)" air purification technology, which uses plasma discharge to generate cluster ions. The new Plasma Cluster Device releases positive and negative ions into the air, which are harmless to humans and are able to decompose and deactivate airborne substances by chemical reactions. In the past, phenomenological tests on the efficacy of the PCI air purification technology on microbial cells have been conducted. In most cases, it has been shown that PCI demonstrated strongly pronounced killing effects on microorganisms. However, the particular mechanisms of PCI action still have to be uncovered.
Close interrelations between sound and image are not a mere phenomenon of today’s multimedia technology. The idea of the synthesis of different media lies at the core of the concept of the Gesamtkunstwerk in the second half of the 19th century and it can also be traced back to the synaesthesia debate at the beginning of the 20th century [...].
The main objective of our ROS Summer School series is to introduce MA level students to program mobile robots with the Robot Operating System (ROS). ROS is a robot middleware that is used my many research institutions world-wide. Therefore, many state-of-the-art algorithms of mobile robotics are available in ROS and can be deployed very easily. As a basic robot platform we deploy a 1/10 RC cart that is wquipped with an Arduino micro-controller to control the servo motors, and an embedded PC that runs ROS. In two weeks, participants get to learn the basics of mobile robotics hands-on. We describe our teaching concepts and our curriculum and report on the learning success of our students.
On 1st January 1998, the German telecom market was fully liberalised. Since then genuine competition between market participants has developed, based on a comprehensive legal and regulatory framework that provides for safeguards against unfair competition and market power by Deutsche Telekom. Today, about 10 years after the liberalisation of the telecommunications sector a revision of this regulatory approach has become necessary because at least on three dimensions the situation is quite different from the one 10 years ago: First, with numerous established alternative operators in the market monopolies have been successfully challenged and competition introduced. Second, not only is Cable TV becoming in large parts of Germany a viable alternative for the provision of broadband services but also mobile services are becoming increasingly a substitute for fixed services. Last but not least there are important technological changes under way, requiring huge investments in infrastructure upgrades for next generation networks. In the light of these new developments the question is to which extent the current regulatory approach of severe ex-ante regulatory intervention is still appropriate. Is any part of the network of the former incumbent still a bottleneck? A more light handed regulatory approach might be the right response to this new situation. The paper is organised as follows: The first section will briefly examine the economic rationale for regulating network access. Based on the assumption that regulation is always necessary when bottlenecks exist regulatory principles for an efficient network access regime will be derived. The second section compares the situation of the German market in early 1998 with the one of today. Thereby three dimensions will be considered: the degree of competition, the potential for substitution and technological developments. The third section will define some requirements for the future regulation of telecom markets. Proposals will be elaborated how to ensure competitive telecom markets in the light of new economic and technological challenges.
Working paper distributed at 2nd Annual Next Generation Telecommunications Conference 2009, 13th – 14th October 2009, Brussels 14 pages Abstract Governments all over Europe are in the process of adopting new broadband strategies. The objective is to create modern telecommunications networks based on powerful broadband infrastructures". In doing so, they aim for innovative and investment-friendly concepts. For instance, in a recently published consultation paper on the subject the German regulator BNetzA declared that it will take “greater account of … reducing risks, securing the investment and innovation power, providing planning certainty and transparency – in order to support and advance broadband rollout in Germany”. It further states that when regulating wholesale rates it has to be ensured that “… adequate incentives for network rollout are provided on the one hand, while sustainable and fair competition is ensured on the other”. Also an EC draft recommendation on regulated network access is about to set new standards for the regulation of next generation access networks. According to the recommendation the prices of new assets shall be based on costs plus a projectspecific risk premium to be included in the costs of capital for the investment risk incurred by the operator. This approach has been criticised from various sides. In particular it has been questioned whether such an approach is adequate to meet the objectives of encouraging both competition and investment into next generation access networks. Against this background, the concept of “long term risk sharing contracts” has been proposed recently as an approach which does not only incorporate the various additional risks involved in the deployment of NGA infrastructure, but has several other advantages. This paper will demonstrate that the concept allows for competition to evolve at both the retail and wholesale level on fair, objective, non-discriminatory and transparent terms and conditions. Moreover, it ensures the highest possible investment incentive in line with socially desirable outcome. The paper is organised as follows: The next section will briefly outline the importance of encouraging competition and investment in an NGA-environment. The third section will specify the design of long term risk sharing contracts in view of achieving these objectives. The fourth section will examine potential problems associated with the concept. In doing so a way of how to deal with them will be elaborated. The last section will look at arguments against long term risk sharing contracts. It will be shown that these arguments are not strong enough to build a case against introducing such contracts.
A key feature of future broadband markets will be diversity of access technologies, meaning that numerous technologies will be exploited for broadband communication. Various factors will affect the success of these future broadband markets, the regulatory policy being one amongst others. So far, a coherent regulatory approach does not exist as to broadband markets. First results of policies so far suggest that less sector-specific regulation is likely to occur. Instead, regulators must ensure that access to networks and services of potentially dominant providers in a relevant broadband market will satisfy requirements for openness and non-discrimination. In this environment the future challenge of regulationg broadband markets will be to set the right incentives for investment into new infrastructures. This paper examines whether there is a need for the regulation of future broadband access markets an if yes, what is the appropriate regulatory tool to do so. Thereby the focus is on the analysis of European broadband markets and the regulatory approaches applied. The first section provides a description of the characteristics of future broadband markets. The second section discusses possible bottlenecks on broadband markets an their regulatory implications. The third section will examine regulatory issues concerning access to broadband networks in more detail. This will be done by comparing the regulatory approaches of European countries and the results in terms of bradband penetration. The final section will give key recommendations for a regulatory strategy on brandband access markets.
Market data for the German telecom market shows that Deutsche Telekom as the former incumbent is constantly loosing shares on all arkets for voice telephony: the market for local calls, the market for long-distance calls and the market for international calls. At the same time prices decline steadily with the latest trend being that operators offer voice services free of charge, the costs of which are covered by a monthly subscription charge. Against this background the paper examines the state of policy and regulatory reform in the telecommunications sector in Germany almost 10 years after the liberalisation of the fixed telecommunications market. Thereby the focus is on the analysis of the competitive conditions that have been established on the German market for voice telephony services. If these retail markets are competitive, there might be a need to remove remaining regulatory provisions. In the new environment of converging markets the future challenge of regulating fixed telecom markets might be to ensure that access to the network and/or services of a potentially dominant provider in a relevant market will satisfy requirements for openness and non-discrimination.
To give the exchange of goods and services between the European Union (EU) and the United States (U.S.) new momentum the two parties are currently negotiating the transatlantic free trade agreement Transatlantic Trade and Investment Partnership (TTIP). The aim is to create the largest free trade area in the world. The agreement, once entered into force, will oblige EU countries and the U.S. to further liberalize their markets.
The negotiations on TTIP include a chapter on Electronic Communications/ Telecommunications. The challenge therein will be securing commitments for market access to Electronic Communications services. At the same time, these commitments must reflect the legitimate need for consumer protection issues. The need to reduce Electronic Communications-related non-tariff barriers to trade between the Parties is due to the fact that these markets are heavily regulated. Without transnational rules as to regulations national governments can abuse these regulations to deter the market entry by new (foreign) suppliers. Thus the free trade agreement TTIP affects in many respects regulatory provisions on and access to Electronic Communications markets. The objective of this paper is therefore to examine to what extend the regulatory principles for Electronic Communications markets envisaged under TTIP will result in trade facilitation and regulatory convergence between the EU and the U.S.
As to this question the result of the analysis is that the chapter on Electronic Communications will be an important step towards facilitating trade in Electronic Communications services. At the same time some regulatory convergence will take place, but this convergence will not lead to a (full) harmonization of regulations. Rather the norm, also after TTIP negotiations will have been concluded successfully, will be mutual recognition of different regulatory regimes. Different regulations being the optimal policy response in different market settings will continue to exist. Moreover, it is very unlikely that such regulatory principles for the Electronic Communications sector are a vehicle for a race to the bottom in levels of consumer protection.