Refine
Year of publication
Institute
- Fachbereich Medizintechnik und Technomathematik (77)
- IfB - Institut für Bioengineering (38)
- Fachbereich Energietechnik (19)
- Fachbereich Wirtschaftswissenschaften (12)
- Fachbereich Maschinenbau und Mechatronik (9)
- Fachbereich Elektrotechnik und Informationstechnik (6)
- Fachbereich Luft- und Raumfahrttechnik (6)
- INB - Institut für Nano- und Biotechnologien (6)
- MASKOR Institut für Mobile Autonome Systeme und Kognitive Robotik (4)
- Fachbereich Bauingenieurwesen (3)
Has Fulltext
- yes (132) (remove)
Language
- English (132) (remove)
Document Type
- Conference Proceeding (132) (remove)
Keywords
- Biosensor (25)
- CAD (7)
- Finite-Elemente-Methode (7)
- civil engineering (7)
- Bauingenieurwesen (6)
- Blitzschutz (6)
- Clusterion (4)
- Sonde (4)
- Telekommunikationsmarkt (4)
- Air purification (3)
ITCE-2003 - 4th Joint Symposium on Information Technology in Civil Engineering ed Flood, I., Seite 1-12, ASCE (CD-ROM), Nashville, USA In this paper we discussed graph based tools to support architects during the conceptual design phase. Conceptual Design is defined before constructive design; the used concepts are more abstract. We develop two graph based approaches, a topdown using the graph rewriting system PROGRES and a more industrially oriented approach, where we extend the CAD system ArchiCAD. In both approaches, knowledge can be defined by a knowledge engineer, in the top-down approach in the domain model graph, in the bottom-up approach in the in an XML file. The defined knowledge is used to incrementally check the sketch and to inform the architect about violations of the defined knowledge. Our goal is to discover design error as soon as possible and to support the architect to design buildings with consideration of conceptual knowledge.
In: Advances in intelligent computing in engineering : proceedings of the 9.International EG-ICE Workshop ; Darmstadt, (01 - 03 August) 2002 / Martina Schnellenbach-Held ... (eds.) . - Düsseldorf: VDI-Verl., 2002 .- Fortschritt-Berichte VDI, Reihe 4, Bauingenieurwesen ; 180 ; S. 1-35 The paper describes a novel way to support conceptual design in civil engineering. The designer uses semantical tools guaranteeing certain internal structures of the design result but also the fulfillment of various constraints. Two different approaches and corresponding tools are discussed: (a) Visually specified tools with automatic code generation to determine a design structure as well as fixing various constraints a design has to obey. These tools are also valuable for design knowledge specialist. (b) Extensions of existing CAD tools to provide semantical knowledge to be used by an architect. It is sketched how these different tools can be combined in the future. The main part of the paper discusses the concepts and realization of two prototypes following the two above approaches. The paper especially discusses that specific graphs and the specification of their structure are useful for both tool realization projects.
Applications of Graph Transformations with Industrial Relevance Lecture Notes in Computer Science, 2004, Volume 3062/2004, 434-439, DOI: http://dx.doi.org/10.1007/978-3-540-25959-6_33 This paper gives a brief overview of the tools we have developed to support conceptual design in civil engineering. Based on the UPGRADE framework, two applications, one for the knowledge engineer and another for architects allow to store domain specific knowledge and to use this knowledge during conceptual design. Consistency analyses check the design against the defined knowledge and inform the architect if rules are violated.
The workflow of a high throughput screening setup for the rapid identification of new and improved sensor materials is presented. The polyol method was applied to prepare nanoparticular metal oxides as base materials, which were functionalised by surface doping. Using multi-electrode substrates and high throughput impedance spectroscopy (HT-IS) a wide range of materials could be screened in a short time. Applying HT-IS in search of new selective gas sensing materials a NO2-tolerant NO sensing material with reduced sensitivities towards other test gases was identified based on iridium doped zinc oxide. Analogous behaviour was observed for iridium doped indium oxide.
IASSE-2004 - 13th International Conference on Intelligent and Adaptive Systems and Software Engineering eds. W. Dosch, N. Debnath, pp. 245-250, ISCA, Cary, NC, 1-3 July 2004, Nice, France We introduce a UML-based model for conceptual design support in civil engineering. Therefore, we identify required extensions to standard UML. Class diagrams are used for elaborating building typespecific knowledge: Object diagrams, implicitly contained in the architect’s sketch, are validated against the defined knowledge. To enable the use of industrial, domain-specific tools, we provide an integrated conceptual design extension. The developed tool support is based on graph rewriting. With our approach architects are enabled to deal with semantic objects during early design phase, assisted by incremental consistency checks.
Lightning protection design of a renewable energy hybrid-system without power mains connection
(2001)
In the year 2000 a direct lightning strike to the hybridsystem without power mains connection VATALI on the Greek island Crete results in the destruction and damage of some mechanical and electrical components. The hybrid-system VATALI was not lightning protected at that time. The hardware damage costs are approx. 60,000 €. The exposed site of the hybrid-system on top of a mountain was and still is the reason for a high risk of lightning strikes. Also in the future further lightning strikes have to be taken into consideration. In the paper a fundamental lightning protection design concept for renewable energy hybrid-systems without power mains connection and protection measures against direct strikes and overvoltages are shown in detail. The design concept was realized exemplarily for the hybrid-system VATALI. The hardware costs for the protection measures were about 15,000 €. About 50% of the costs are due to protection measures against direct strikes, 50% are due to overvoltage protection. Future extensions, new installations, or modifications have to be included into the lightning protection design concept of the hybrid-system.
In the paper a lightning protection design concept for renewable energy hybrid-systems without power mains connection is described. Based on a risk analysis protection measures against direct strikes and overvoltages are shown in an overview. The design concept is realized exemplarily for the hybrid-system VATALI on the Greek island Crete. VATALI, not lightning protected at that time, was a victim of a lightning strike in the year 2000 causing destructions and damages of some mechanical and electrical components with costs of approx. 60.000 €. The hardware costs for the protection measures were about 15.000 €: about 50% of the costs are due to protection measures against direct strikes, 50% are due to overvoltage protection.
In the paper the results obtained from experiments at a modelled reinforced building in case of a direct lightning strike are compared with calculations. The comparison includes peak values of the magnetic field Hmax, its derivative (dH/dt)max and of induced voltages umax in typical cable routings. The experiments are performed at a 1:6 scaled building and the results are extrapolated using the similarity relations theory. The calculations are based on the approximate formulae given in IEC 62305-4 and have to be supplemented by a rough estimation of the additional shielding effect of a second reinforcement layer. The comparison shows, that the measured peak values of the magnetic field and its derivative are mostly lower than the calculated. The induced voltages are in good agreement. Hence, calculations of the induced voltages based on IEC 62305-4 are a good method for lightning protection studies of buildings, where the reinforcement is used as a grid-like electromagnetic shield.
In the presented paper data collected from the field related to damage statistics of electrical and electronic apparatus in household are reported and investigated. These damages (total number approx. 74000 cases), registered by five German insurance companies in 2005 and 2006, were adviced by customers as caused by lightning overvoltages. With the use of stochastical methods it is possible, to reasses the collected data and to distinguish between cases, which are with high probability caused by lightning overvoltages, and those, which are not. If there was an indication for a direct strike, this case was excluded, so the focus was only on indirect lightning flashes, i.e. only flashes to ground near the structure and flashes to or nearby an incoming service line were investigated. The data from the field contain the location of damaged apparatus (residence of the policy holder) and the distances of the nearest cloud-to-ground stroke to the location of the damage registered by the German lightning location network BLIDS at the date of damage. The statistical data along with some complementary numerical simulations allow to verify the correspondence of the Standards rules used for IEC 62305-2 with the field data and to define some correction needs. The results could lead to a better understanding whether a damage reported to an insurance company is really caused by indirect lightning, or not.
Hydrophobic magnetic nanoparticles (NPs) consisting of undecanoate-capped magnetite (Fe3O4, average diameter ca. 5 nm) are used to control quantized electron transfer to surface-confined redox units and metal NPs. A two-phase system consisting of an aqueous electrolyte solution and a toluene phase that includes the suspended undecanoatecapped magnetic NPs is used to control the interfacial properties of the electrode surface. The attracted magnetic NPs form a hydrophobic layer on the electrode surface resulting in the change of the mechanisms of the surface-confined electrochemical processes. A quinone-monolayer modified Au electrode demonstrates an aqueous-type of the electrochemical process (2e-+2H+ redox mechanism) for the quinone units in the absence of the hydrophobic magnetic NPs, while the attraction of the magnetic NPs to the surface results in the stepwise single-electron transfer mechanism characteristic of a dry nonaqueous medium. Also, the attraction of the hydrophobic magnetic NPs to the Au electrode surface modified with Au NPs (ca. 1.4 nm) yields a microenvironment with a low dielectric constant that results in the single-electron quantum charging of the Au NPs.
Electromechanical model of hiPSC-derived ventricular cardiomyocytes cocultured with fibroblasts
(2018)
The CellDrum provides an experimental setup to study the mechanical effects of fibroblasts co-cultured with hiPSC-derived ventricular cardiomyocytes. Multi-scale computational models based on the Finite Element Method are developed. Coupled electrical cardiomyocyte-fibroblast models (cell level) are embedded into reaction-diffusion equations (tissue level) which compute the propagation of the action potential in the cardiac tissue. Electromechanical coupling is realised by an excitation-contraction model (cell level) and the active stress arising during contraction is added to the passive stress in the force balance, which determines the tissue displacement (tissue level). Tissue parameters in the model can be identified experimentally to the specific sample.
Abstracts of the ACHEMA 2000 - International Meeting on Chemical Engineering, Environmental Protection and Biotechnology, May 22 - 27, 2000. Frankfurt am Main. Achema 2000 : special edition / Linde. [Ed.: Linde AG. Red.: Volker R. Leski]. - Wiesbaden : Linde AG, 2000. - 56 p. : Ill., . - pp: 79 - 81
The Passivhaus building standard is a concept developed for the realization of energy-efficient and economical buildings with a simultaneous high utilization comfort under European climate conditions. Major elements of the Passivhaus concept are a high thermal insulation of the external walls, the use of heat and/or solar shading glazing as well as an airtight building envelope in combination with energy-efficient technical building installations and heating or cooling generators, such as an efficient energy-recovery in the building air-conditioning. The objective of this research project is the inquiry to determine the parameters or constraints under which the Passivhaus concept can be implemented under the arid climate conditions in the Arabian Peninsula to achieve an energy-efficient and economical building with high utilization comfort. In cooperation between the Qatar Green Building Council (QGBC), Barwa Real Estate (BRE) and Kahramaa the first Passivhaus was constructed in Qatar and on the Arabian Peninsula in 2013. The Solar-Institut Jülich of Aachen University of Applied Science supports the Qatar Green Building Council with a dynamic building and equipment simulation of the Passivhaus and the neighbouring reference building. This includes simulation studies with different component configurations for the building envelope and different control strategies for heating or cooling systems as well as the air conditioning of buildings to find an energetic-economical optimum. Part of these analyses is the evaluation of the energy efficiency of the used energy recovery system in the Passivhaus air-conditioning and identification of possible energy-saving effects by the use of a bypass function integrated in the heat exchanger. In this way it is expected that on an annual basis the complete electricity demand of the building can be covered by the roof-integrated PV generator.
Table of contents 1. Introduction 2. Multi-level Technology Transfer Infrastructure 2.1 Level 1: University Education – Encourage the Idea of becoming an Entrepreneur 2.2 Level 2: Post Graduate Education – Improve your skills and focus it on a product family. 2.3 Level 3: Birth of a Company – Focus your skills on a product and a market segment. 2.4 Level 4: Ready to stand alone – Set up your own business 2.5 Level 5: Grow to be Strong – Develop your business 2.6 Level 6: Competitive and independent – Stay innovative. 3. Samples 3.1 Sample 1: Laser Processing and Consulting Centre, LBBZ 3.2 Sample 2: Prototyping Centre, CP 4. Funding - Waste money or even lost Money? 5. Conclusion
GaAs-based Gunn diodes with graded AlGaAs hot electron injector heterostructures have been developed under the special needs in automotive applications. The fabrication of the Gunn diode chips was based on total substrate removal and processing of integrated Au heat sinks. Especially, the thermal and RF behavior of the diodes have been analyzed by DC, impedance and S-parameter measurements. The electrical investigations have revealed the functionality of the hot electron injector. An optimized layer structure could fulfill the requirements in adaptive cruise control (ACC) systems at 77 GHz with typical output power between 50 and 90 mW.
On 1st January 1998, the German telecom market was fully liberalised. Since then genuine competition between market participants has developed, based on a comprehensive legal and regulatory framework that provides for safeguards against unfair competition and market power by Deutsche Telekom. Today, about 10 years after the liberalisation of the telecommunications sector a revision of this regulatory approach has become necessary because at least on three dimensions the situation is quite different from the one 10 years ago: First, with numerous established alternative operators in the market monopolies have been successfully challenged and competition introduced. Second, not only is Cable TV becoming in large parts of Germany a viable alternative for the provision of broadband services but also mobile services are becoming increasingly a substitute for fixed services. Last but not least there are important technological changes under way, requiring huge investments in infrastructure upgrades for next generation networks. In the light of these new developments the question is to which extent the current regulatory approach of severe ex-ante regulatory intervention is still appropriate. Is any part of the network of the former incumbent still a bottleneck? A more light handed regulatory approach might be the right response to this new situation. The paper is organised as follows: The first section will briefly examine the economic rationale for regulating network access. Based on the assumption that regulation is always necessary when bottlenecks exist regulatory principles for an efficient network access regime will be derived. The second section compares the situation of the German market in early 1998 with the one of today. Thereby three dimensions will be considered: the degree of competition, the potential for substitution and technological developments. The third section will define some requirements for the future regulation of telecom markets. Proposals will be elaborated how to ensure competitive telecom markets in the light of new economic and technological challenges.
A key feature of future broadband markets will be diversity of access technologies, meaning that numerous technologies will be exploited for broadband communication. Various factors will affect the success of these future broadband markets, the regulatory policy being one amongst others. So far, a coherent regulatory approach does not exist as to broadband markets. First results of policies so far suggest that less sector-specific regulation is likely to occur. Instead, regulators must ensure that access to networks and services of potentially dominant providers in a relevant broadband market will satisfy requirements for openness and non-discrimination. In this environment the future challenge of regulationg broadband markets will be to set the right incentives for investment into new infrastructures. This paper examines whether there is a need for the regulation of future broadband access markets an if yes, what is the appropriate regulatory tool to do so. Thereby the focus is on the analysis of European broadband markets and the regulatory approaches applied. The first section provides a description of the characteristics of future broadband markets. The second section discusses possible bottlenecks on broadband markets an their regulatory implications. The third section will examine regulatory issues concerning access to broadband networks in more detail. This will be done by comparing the regulatory approaches of European countries and the results in terms of bradband penetration. The final section will give key recommendations for a regulatory strategy on brandband access markets.
Market data for the German telecom market shows that Deutsche Telekom as the former incumbent is constantly loosing shares on all arkets for voice telephony: the market for local calls, the market for long-distance calls and the market for international calls. At the same time prices decline steadily with the latest trend being that operators offer voice services free of charge, the costs of which are covered by a monthly subscription charge. Against this background the paper examines the state of policy and regulatory reform in the telecommunications sector in Germany almost 10 years after the liberalisation of the fixed telecommunications market. Thereby the focus is on the analysis of the competitive conditions that have been established on the German market for voice telephony services. If these retail markets are competitive, there might be a need to remove remaining regulatory provisions. In the new environment of converging markets the future challenge of regulating fixed telecom markets might be to ensure that access to the network and/or services of a potentially dominant provider in a relevant market will satisfy requirements for openness and non-discrimination.
Working paper distributed at 2nd Annual Next Generation Telecommunications Conference 2009, 13th – 14th October 2009, Brussels 14 pages Abstract Governments all over Europe are in the process of adopting new broadband strategies. The objective is to create modern telecommunications networks based on powerful broadband infrastructures". In doing so, they aim for innovative and investment-friendly concepts. For instance, in a recently published consultation paper on the subject the German regulator BNetzA declared that it will take “greater account of … reducing risks, securing the investment and innovation power, providing planning certainty and transparency – in order to support and advance broadband rollout in Germany”. It further states that when regulating wholesale rates it has to be ensured that “… adequate incentives for network rollout are provided on the one hand, while sustainable and fair competition is ensured on the other”. Also an EC draft recommendation on regulated network access is about to set new standards for the regulation of next generation access networks. According to the recommendation the prices of new assets shall be based on costs plus a projectspecific risk premium to be included in the costs of capital for the investment risk incurred by the operator. This approach has been criticised from various sides. In particular it has been questioned whether such an approach is adequate to meet the objectives of encouraging both competition and investment into next generation access networks. Against this background, the concept of “long term risk sharing contracts” has been proposed recently as an approach which does not only incorporate the various additional risks involved in the deployment of NGA infrastructure, but has several other advantages. This paper will demonstrate that the concept allows for competition to evolve at both the retail and wholesale level on fair, objective, non-discriminatory and transparent terms and conditions. Moreover, it ensures the highest possible investment incentive in line with socially desirable outcome. The paper is organised as follows: The next section will briefly outline the importance of encouraging competition and investment in an NGA-environment. The third section will specify the design of long term risk sharing contracts in view of achieving these objectives. The fourth section will examine potential problems associated with the concept. In doing so a way of how to deal with them will be elaborated. The last section will look at arguments against long term risk sharing contracts. It will be shown that these arguments are not strong enough to build a case against introducing such contracts.
To give the exchange of goods and services between the European Union (EU) and the United States (U.S.) new momentum the two parties are currently negotiating the transatlantic free trade agreement Transatlantic Trade and Investment Partnership (TTIP). The aim is to create the largest free trade area in the world. The agreement, once entered into force, will oblige EU countries and the U.S. to further liberalize their markets.
The negotiations on TTIP include a chapter on Electronic Communications/ Telecommunications. The challenge therein will be securing commitments for market access to Electronic Communications services. At the same time, these commitments must reflect the legitimate need for consumer protection issues. The need to reduce Electronic Communications-related non-tariff barriers to trade between the Parties is due to the fact that these markets are heavily regulated. Without transnational rules as to regulations national governments can abuse these regulations to deter the market entry by new (foreign) suppliers. Thus the free trade agreement TTIP affects in many respects regulatory provisions on and access to Electronic Communications markets. The objective of this paper is therefore to examine to what extend the regulatory principles for Electronic Communications markets envisaged under TTIP will result in trade facilitation and regulatory convergence between the EU and the U.S.
As to this question the result of the analysis is that the chapter on Electronic Communications will be an important step towards facilitating trade in Electronic Communications services. At the same time some regulatory convergence will take place, but this convergence will not lead to a (full) harmonization of regulations. Rather the norm, also after TTIP negotiations will have been concluded successfully, will be mutual recognition of different regulatory regimes. Different regulations being the optimal policy response in different market settings will continue to exist. Moreover, it is very unlikely that such regulatory principles for the Electronic Communications sector are a vehicle for a race to the bottom in levels of consumer protection.
The main objective of our ROS Summer School series is to introduce MA level students to program mobile robots with the Robot Operating System (ROS). ROS is a robot middleware that is used my many research institutions world-wide. Therefore, many state-of-the-art algorithms of mobile robotics are available in ROS and can be deployed very easily. As a basic robot platform we deploy a 1/10 RC cart that is wquipped with an Arduino micro-controller to control the servo motors, and an embedded PC that runs ROS. In two weeks, participants get to learn the basics of mobile robotics hands-on. We describe our teaching concepts and our curriculum and report on the learning success of our students.
Close interrelations between sound and image are not a mere phenomenon of today’s multimedia technology. The idea of the synthesis of different media lies at the core of the concept of the Gesamtkunstwerk in the second half of the 19th century and it can also be traced back to the synaesthesia debate at the beginning of the 20th century [...].
Recently, the SHARP Corporation, Japan, has developed the world’s first "Plasma Cluster Ions (PCI)" air purification technology using plasma discharge to generate cluster ions. The new plasma cluster device releases positive and negative ions into the air, which are able to decompose and deactivate harmful airborne substances by chemical reactions. Because cluster ions consist of positive and negative ions that normally exist in the natural world, they are completely harmless and safe to humans. The amount of ozone generated by cluster ions is less than 0.01 ppm, which is significantly less than the 0.05-ppm standard for industrial operations and consumer electronics. This amount, thus, has no harming effects whatsoever on the human body. But particular properties and chemical processes in PCI treatment are still under study. It has been shown that PCI in most cases show strongly pronounced irreversible killing effects in respect of airborne microflora due to free-radical induced reactions and can be considered as a potent technology to disinfect both home, medical and industrial appliances.
Summary and Conclusions PCIs were clearly effective in terms of their antibacterial effects with the strains tested. This efficacy increased with the time the bacteries were exposed to PCIs. The bactericidal action has proved to be irreversible. PCIs were significantly less effective in shadowed areas. PCI exposure caused multiple protein damages as observed in SDS PAGE studies. There was no single but multiple molecular mechanism causing the bacterial death.
Recently, SHARP corporation has developed the world’s first "Plasma Cluster Ions® (PCI)" air purification technology, which uses plasma discharge to generate cluster ions. The new Plasma Cluster Device releases positive and negative ions into the air, which are harmless to humans and are able to decompose and deactivate airborne substances by chemical reactions. In the past, phenomenological tests on the efficacy of the PCI air purification technology on microbial cells have been conducted. In most cases, it has been shown that PCI demonstrated strongly pronounced killing effects on microorganisms. However, the particular mechanisms of PCI action still have to be uncovered.
Recently, SHARP corporation has developed the world’s first “Plasma Cluster Ions (PCI)” air purification technology, which uses plasma discharge to generate cluster ions. The new plasma cluster device releases into the air positive and negative ions, which are harmless to humans and are able to decompose and deactivate airborne substances by chemical reactions. A lot of phenomenological tests of the PCI air purification technology on microbial cells have been conducted. And, in most cases, it has been shown that PCI demonstrate strongly pronounced killing effect. Although, the particular mechanisms of PCI action are still not evident. We studied variations in resistance to PCI among gram-positive airborne microorganisms, as well as some dose-dependent, spatial, cultural and biochemical properties of PCI action in respect of Staphylococcus spp, Enterococcus spp, Micrococcus spp.
Mechanical stimulation of the cells resulted in evident changes in the cell morphology, protein composition and gene expression. Microscopically, additional formation of stress fibers accompanied by cell re-arrangements in a monolayer was observed. Also, significant activation of p53 gene was revealed as compared to control. Interestingly, the use of CellTech membrane coating induced cell death after mechanical stress had been applied. Such an effect was not detected when fibronectin had been used as an adhesion substrate.
A melting probe equipped with autofluorescence-based detection system combined with a light scattering unit, and, optionally, with a microarray chip would be ideally suited to probe icy environments like Europa’s ice layer as well as the polar ice layers of Earth and Mars for recent and extinct live.
The paper deals with the development of the probabilistic approach to the assessment of risk due to lightning. Sources of damage, types of damage and types of loss are defined and, accordingly, the procedure for risk analysis and the way of assessment of different risk components is proposed. The way to evaluate the influence of different protection measures (lightning protection system; shielding of structure, cables and equipment; routing of internal wiring; surge protective device) in reducing such probabilities is considered. The paper has been prepared within the framework of the activity of IEC TC81-WG9/CLC TC81-WG4 directed to prepare the draft IEC 62305-2 Risk Management, in cooperation with the Secretary of IEC/CLC TC81.
Evaluation of fragility curves for a three-storey-reinforced-concrete mock-up of SMART 2013 project
(2016)
We present the novel concept of a combined drilling and melting probe for subsurface ice research. This probe, named “IceMole”, is currently developed, built, and tested at the FH Aachen University of Applied Sciences’ Astronautical Laboratory. Here, we describe its first prototype design and report the results of its field tests on the Swiss Morteratsch glacier. Although the IceMole design is currently adapted to terrestrial glaciers and ice shields, it may later be modified for the subsurface in-situ investigation of extraterrestrial ice, e.g., on Mars, Europa, and Enceladus. If life exists on those bodies, it may be present in the ice (as life can also be found in the deep ice of Earth).
The ”IceMole“ is a novel maneuverable subsurface ice probe for clean in-situ analysis and sampling of subsurface ice and subglacial water/brine. It is developed and build at FH Aachen University of Applied Sciences’ Astronautical Laboratory. A first prototype was successfully tested on the Swiss Morteratsch glacier in 2010. Clean sampling is achieved with a hollow ice screw (as it is used in mountaineering) at the tip of the probe. Maneuverability is achieved with a differentially heated melting head. Funded by the German Space Agency (DLR), a consortium led by FH Aachen currently develops a much more advanced IceMole probe, which includes a sophisticated system for obstacle avoidance, target detection, and navigation in the ice. We intend to use this probe for taking clean samples of subglacial brine at the Blood Falls (McMurdo Dry Valleys, East Antarctica) for chemical and microbiological analysis. In our conference contribution, we 1) describe the IceMole design, 2) report the results of the field tests of the first prototype on the Morteratsch glacier, 3) discuss the probe’s potential for the clean in-situ analysis and sampling of subsurface ice and subglacial liquids, and 4) outline the way ahead in the development of this technology.
Cure or blessing? The effect of (non-financial) signals on sustainable venture's funding success
(2022)
A multi-sensor system is a chemical sensor system which quantitatively and qualitatively records gases with a combination of cross-sensitive gas sensor arrays and pattern recognition software. This paper addresses the issue of data analysis for identification of gases in a gas sensor array. We introduce a software tool for gas sensor array configuration and simulation. It concerns thereby about a modular software package for the acquisition of data of different sensors. A signal evaluation algorithm referred to as matrix method was used specifically for the software tool. This matrix method computes the gas concentrations from the signals of a sensor array. The software tool was used for the simulation of an array of five sensors to determine gas concentration of CH4, NH3, H2, CO and C2H5OH. The results of the present simulated sensor array indicate that the software tool is capable of the following: (a) identify a gas independently of its concentration; (b) estimate the concentration of the gas, even if the system was not previously exposed to this concentration; (c) tell when a gas concentration exceeds a certain value. A gas sensor data base was build for the configuration of the software. With the data base one can create, generate and manage scenarios and source files for the simulation. With the gas sensor data base and the simulation software an on-line Web-based version was developed, with which the user can configure and simulate sensor arrays on-line.