Conference Proceeding
Refine
Year of publication
Institute
- Fachbereich Medizintechnik und Technomathematik (243) (remove)
Document Type
- Conference Proceeding (243) (remove)
Keywords
- Biosensor (25)
- CAD (11)
- Finite-Elemente-Methode (11)
- civil engineering (11)
- Bauingenieurwesen (10)
- Einspielen <Werkstoff> (6)
- shakedown analysis (6)
- Clusterion (4)
- Limit analysis (4)
- Natural language processing (4)
A procedure for the evaluation of the failure probability of elastic-plastic thin shell structures is presented. The procedure involves a deterministic limit and shakedown analysis for each probabilistic iteration which is based on the kinematical approach and the use the exact Ilyushin yield surface. Based on a direct definition of the limit state function, the non-linear problems may be efficiently solved by using the First and Second Order Reliabiblity Methods (Form/SORM). This direct approach reduces considerably the necessary knowledge of uncertain technological input data, computing costs and the numerical error. In: Computational plasticity / ed. by Eugenio Onate. Dordrecht: Springer 2007. VII, 265 S. (Computational Methods in Applied Sciences ; 7) (COMPLAS IX. Part 1 . International Center for Numerical Methods in Engineering (CIMNE)). ISBN 978-1-402-06576-7 S. 186-189
Useful market simulations are key to the evaluation of diferent market designs existing of multiple market mechanisms or rules. Yet a simulation framework which has a comparison of diferent market mechanisms in mind was not found. The need to create an objective view on different sets of market rules while investigating meaningful agent strategies concludes that such a simulation framework is needed to advance the research on this subject. An overview of diferent existing market simulation models is given which also shows the research gap and the missing capabilities of those systems. Finally, a methodology is outlined how a novel market simulation which can answer the research questions can be developed.
Messenger apps like WhatsApp and Telegram are frequently used for everyday communication, but they can also be utilized as a platform for illegal activity. Telegram allows public groups with up to 200.000 participants. Criminals use these public groups for trading illegal commodities and services, which becomes a concern for law enforcement agencies, who manually monitor suspicious activity in these chat rooms. This research demonstrates how natural language processing (NLP) can assist in analyzing these chat rooms, providing an explorative overview of the domain and facilitating purposeful analyses of user behavior. We provide a publicly available corpus of annotated text messages with entities and relations from four self-proclaimed black market chat rooms. Our pipeline approach aggregates the extracted product attributes from user messages to profiles and uses these with their sold products as features for clustering. The extracted structured information is the foundation for further data exploration, such as identifying the top vendors or fine-granular price analyses. Our evaluation shows that pretrained word vectors perform better for unsupervised clustering than state-of-the-art transformer models, while the latter is still superior for sequence labeling.
Functional testing and characterisation of ISFETs on wafer level by means of a micro-droplet cell
(2006)
A wafer-level functionality testing and characterisation system for ISFETs (ionsensitive field-effect transistor) is realised by means of integration of a specifically designed capillary electrochemical micro-droplet cell into a commercial wafer prober-station. The developed system allows the identification and selection of “good” ISFETs at the earliest stage and to avoid expensive bonding, encapsulation and packaging processes for nonfunctioning ISFETs and thus, to decrease costs, which are wasted for bad dies. The developed system is also feasible for wafer-level characterisation of ISFETs in terms of sensitivity, hysteresis and response time. Additionally, the system might be also utilised for wafer-level testing of further electrochemical sensors.
The nonlinear scalar constitutive equations of gases lead to a change in sound speed from point to point as would be found in linear inhomogeneous (and time dependent) media. The nonlinear tensor constitutive equations of solids introduce the additional local effect of solution dependent anisotropy. The speed of a wave passing through a point changes with propagation direction and its rays are inclined to the front. It is an open question whether the widely used operator splitting techniques achieve a dimensional splitting with physically reasonable results for these multi-dimensional problems. May be this is the main reason why the theoretical and numerical investigations of multi-dimensional wave propagation in nonlinear solids are so far behind gas dynamics. We hope to promote the subject a little by a discussion of some fundamental aspects of the solution of the equations of nonlinear elastodynamics. We use methods of characteristics because they only integrate mathematically exact equations which have a direct physical interpretation.
We report on the synthesis and CO gas-sensing properties of mesoporous tin(IV) oxides (SnO2). For the synthesis cetyltrimethylammonium bromide (CTABr) was used as a structure-directing agent; the resulting SnO2 powders were applied as films to commercially available sensor substrates by drop coating. Nitrogen physisorption shows specific surface areas up to 160 m2·g-1 and mean pore diameters of about 4 nm, as verified by TEM. The film conductance was measured in dependence on the CO concentration in humid synthetic air at a constant temperature of 300 °C. The sensors show a high sensitivity at low CO concentrations and turn out to be largely insensitive towards changes in the relative humidity. We compare the materials with commercially available SnO2-based sensors.
In: Advances in intelligent computing in engineering : proceedings of the 9.International EG-ICE Workshop ; Darmstadt, (01 - 03 August) 2002 / Martina Schnellenbach-Held ... (eds.) . - Düsseldorf: VDI-Verl., 2002 .- Fortschritt-Berichte VDI, Reihe 4, Bauingenieurwesen ; 180 ; S. 1-35 The paper describes a novel way to support conceptual design in civil engineering. The designer uses semantical tools guaranteeing certain internal structures of the design result but also the fulfillment of various constraints. Two different approaches and corresponding tools are discussed: (a) Visually specified tools with automatic code generation to determine a design structure as well as fixing various constraints a design has to obey. These tools are also valuable for design knowledge specialist. (b) Extensions of existing CAD tools to provide semantical knowledge to be used by an architect. It is sketched how these different tools can be combined in the future. The main part of the paper discusses the concepts and realization of two prototypes following the two above approaches. The paper especially discusses that specific graphs and the specification of their structure are useful for both tool realization projects.
In: Proceedings of the 39th Annual Hawaii International Conference on System Sciences, 2006. HICSS '06 http://dx.doi.org/10.1109/HICSS.2006.200 The conceptual design phase at the beginning of the building construction process is not adequately supported by any CAD-tool. Conceptual design support needs regarding two aspects: first, the architect must be able to develop conceptual sketches that provide abstraction from constructive details. Second, conceptually relevant knowledge should be available to check these conceptual sketches. The paper deals with knowledge to formalize for conceptual design. To enable domain experts formalizing knowledge, a graph-based specification is presented that allows the development of a domain ontology and design rules specific for one class of buildings at runtime. The provided tool support illustrates the introduced concepts and demonstrates the consistency analysis between knowledge and conceptual design.
Heavy metal detection with semiconductor devices based on PLD-prepared chalcogenide glass thin films
(2007)
An array of 50 MHz quartz microbalances (QMBs) coated with a dendronized polymer was used to detect small amounts of volatile organic compounds (VOCs) in the gas phase. The results were compared to those obtained with the commonly used 10 MHz QMBs. The 50 MHz QMBs proved to be a powerful tool for the detection of VOCs in the gas phase; therefore, they represent a promising alternative to the much more delicate surface acoustic wave devices (SAWs).
The propagation of mechanical waves in plates of isotropic elastic material is investigated. After a short introduction to the understanding of focussing of stress waves in a plate with a curved boundary the method of characteristics is applied to a plate of hyperelastic material. Using this method the propagation of acceleration waves is discussed. Based on this a numerical difference scheme is developed for solving initial-boundary-value problems and applied to two examples: propagation of a point disturbance in a homogeneously finitely strained non-linear elastic plate and geometrical focussing in al linear elastic plate.
Micromachined thermal heater platforms offer low electrical power consumption and high modulation speed, i.e. properties which are advantageous for realizing nondispersive infrared (NDIR) gas- and liquid monitoring systems. In this paper, we report on investigations on silicon-on-insulator (SOI) based infrared (IR) emitter devices heated by employing different kinds of metallic and semiconductor heater materials. Our results clearly reveal the superior high-temperature performance of semiconductor over metallic heater materials. Long-term stable emitter operation in the vicinity of 1300 K could be attained using heavily antimony-doped tin dioxide (SnO2:Sb) heater elements.
The human arm consists of the humerus (upper arm), the medial ulna and the lateral radius (forearm). The joint between the humerus and the ulna is called humeroulnar joint and the joint between the humerus and the radius is called humeroradial joint. Lateral and medial collateral ligaments stabilize the elbow. Statistically, 2.5 out of 10,000 people suffer from radial head fractures [1]. In these fractures the cartilage is often affected. Caused by the injured cartilage, degenerative diseases like posttraumatic arthrosis may occur. The resulting pain and reduced range of motion have an impact on the patient’s quality of life. Until now, there has not been a treatment which allows typical loads in daily life activities and offers good long-term results. A new surgical approach was developed with the motivation to reduce the progress of the posttraumatic arthrosis. Here, the radius is shortened by 3 mm in the proximal part [2]. By this means, the load of the radius is intended to be reduced due to a load shift to the ulna. Since the radius is the most important stabilizer of the elbow it has to be confirmed that the stability is not affected. In the first test (Fig. 1 left), pressure distributions within the humeroulnar and humeroradial joints a native and a shortened radius were measured using resistive pressure sensors (I5076 and I5027, Tekscan, USA). The humerus was loaded axially in a tension testing machine (Z010, Zwick Roell, Germany) in 50 N steps up to 400 N. From the humerus the load is transmitted through both the radius and the ulna into the hand which is fixed on the ground. In the second test (Fig. 1 right), the joint stability was investigated using a digital image correlation system to measure the displacement of the ulna. Here, the humerus is fixed with a desired flexion angle and the unconstrained forearm lies on the ground. A rope connects the load actuator with a hook fixed in the ulna. A guide roller is used so that the rope pulls the ulna horizontally when a tensile load is applied. This creates a moment about the elbow joint with a maximum value of 7.5 Nm. Measurements were performed with varying flexion angles (0°, 30°, 60°, 90°, 120°). For both tests and each measurement, seven specimens were used. Student ́s t-test was employed to determine whether the mean values of the measurements in native specimen and operated specimens differ significantly.
In: Net-distributed Co-operation : Xth International Conference on Computing in Civil and Building Engineering, Weimar, June 02 - 04, 2004 ; proceedings / [ed. by Karl Beuke ...] . - Weimar: Bauhaus-Univ. Weimar 2004. - 1. Aufl. . Seite 1-14 ISBN 3-86068-213-X International Conference on Computing in Civil and Building Engineering <10, 2004, Weimar> Summary In our project, we develop new tools for the conceptual design phase. During conceptual design, the coarse functionality and organization of a building is more important than a detailed worked out construction. We identify two roles, first the knowledge engineer who is responsible for knowledge definition and maintenance; second the architect who elaborates the conceptual de-sign. The tool for the knowledge engineer is based on graph technology, it is specified using PROGRES and the UPGRADE framework. The tools for the architect are integrated to the in-dustrial CAD tool ArchiCAD. Consistency between knowledge and conceptual design is en-sured by the constraint checker, another extension to ArchiCAD.
Effective training requires high muscle forces potentially leading to training-induced injuries. Thus, continuous monitoring and controlling of the loadings applied to the musculoskeletal system along the motion trajectory is required. In this paper, a norm-optimal iterative learning control algorithm for the robot-assisted training is developed. The algorithm aims at minimizing the external knee joint moment, which is commonly used to quantify the loading of the medial compartment. To estimate the external knee joint moment, a musculoskeletal lower extremity model is implemented in OpenSim and coupled with a model of an industrial robot and a force plate mounted at its end-effector. The algorithm is tested in simulation for patients with varus, normal and valgus alignment of the knee. The results show that the algorithm is able to minimize the external knee joint moment in all three cases and converges after less than seven iterations.
Due to the transition to renewable energies, electricity markets need to be made fit for purpose. To enable the comparison of different energy market designs, modeling tools covering market actors and their heterogeneous behavior are needed. Agent-based models are ideally suited for this task. Such models can be used to simulate and analyze changes to market design or market mechanisms and their impact on market dynamics. In this paper, we conduct an evaluation and comparison of two actively developed open-source energy market simulation models. The two models, namely AMIRIS and ASSUME, are both designed to simulate future energy markets using an agent-based approach. The assessment encompasses modelling features and techniques, model performance, as well as a comparison of model results, which can serve as a blueprint for future comparative studies of simulation models. The main comparison dataset includes data of Germany in 2019 and simulates the Day-Ahead market and participating actors as individual agents. Both models are comparable close to the benchmark dataset with a MAE between 5.6 and 6.4 €/MWh while also modeling the actual dispatch realistically.
Label-free Electrostatic Detection of DNA Amplification by PCR Using Capacitive Field-effect Devices
(2016)
A capacitive field-effect EIS (electrolyte-insulator-semiconductor) sensor modified with a positively charged weak polyelectrolyte of poly(allylamine hydrochloride) (PAH)/single-stranded probe DNA (ssDNA) bilayer has been used for a label-free electrostatic detection of pathogen-specific DNA amplification via polymerase chain reaction (PCR). The sensor is able to distinguish between positive and negative PCR solutions, to detect the existence of target DNA amplicons in PCR samples and thus, can be used as tool for a quick verification of DNA amplification and the successful PCR process.
Label-free sensing of biomolecules by their intrinsic molecular charge using field-effect devices
(2015)
Light-stimulated hydrogel actuators with incorporated graphene oxide for microfluidic applications
(2015)
Limit and shakedown theorems are exact theories of classical plasticity for the direct computation of safety factors or of the load carrying capacity under constant and varying loads. Simple versions of limit and shakedown analysis are the basis of all design codes for pressure vessels and pipings. Using Finite Element Methods more realistic modeling can be used for a more rational design. The methods can be extended to yield optimum plastic design. In this paper we present a first implementation in FE of limit and shakedown analyses for perfectly plastic material. Limit and shakedown analyses are done of a pipe–junction and a interaction diagram is calculated. The results are in good correspondence with the analytic solution we give in the appendix.
Direct methods comprising limit and shakedown analysis is a branch of computational mechanics. It plays a significant role in mechanical and civil engineering design. The concept of direct method aims to determinate the ultimate load bearing capacity of structures beyond the elastic range. For practical problems, the direct methods lead to nonlinear convex optimization problems with a large number of variables and onstraints. If strength and loading are random quantities, the problem of shakedown analysis is considered as stochastic programming. This paper presents a method so called chance constrained programming, an effective method of stochastic programming, to solve shakedown analysis problem under random condition of strength. In this our investigation, the loading is deterministic, the strength is distributed as normal or lognormal variables.
7th International Conference on Reliability of Materials and Structures (RELMAS 2008). June 17 - 20, 2008 ; Saint Petersburg, Russia. pp 354-358. Reprint with corrections in red Introduction Analysis of advanced structures working under extreme heavy loading such as nuclear power plants and piping system should take into account the randomness of loading, geometrical and material parameters. The existing reliability are restricted mostly to the elastic working regime, e.g. allowable local stresses. Development of the limit and shakedown reliability-based analysis and design methods, exploiting potential of the shakedown working regime, is highly needed. In this paper the application of a new algorithm of probabilistic limit and shakedown analysis for shell structures is presented, in which the loading and strength of the material as well as the thickness of the shell are considered as random variables. The reliability analysis problems may be efficiently solved by using a system combining the available FE codes, a deterministic limit and shakedown analysis, and the First and Second Order Reliability Methods (FORM/SORM). Non-linear sensitivity analyses are obtained directly from the solution of the deterministic problem without extra computational costs.
Traglast- und Einspielanalysen sind vereinfachte doch exakte Verfahren der Plastizität, die neben ausreichender Verformbarkeit keine einschränkenden Voraussetzungen beinhalten. Die Vereinfachungen betreffen die Beschaffung der Daten und Modelle für Details der Lastgeschichte und des Stoffverhaltens. Anders als die klassische Behandlung nichtlinearer Probleme der Strukturmechanik führt die Methode auf Optimierungsprobleme. Diese sind bei realistischen FEM-Modellen sehr groß. Das hat die industrielle Anwendung der Traglast- und Einspielanalysen stark verzögert. Diese Situation wird durch das Brite-EuRam Projekt LISA grundlegend geändert. In LISA entsteht auf der Basis des industriellen FEM-Programms PERMAS ein Verfahren zur direkten Berechnung der Tragfähigkeit duktiler Strukturen. Damit kann der Betriebsbereich von Komponenten und Bauwerken auf den plastischen Bereich erweitert werden, ohne den Aufwand gegenüber elastischen Analysen wesentlich zu erhöhen. Die beachtlichen Rechenzeitgewinne erlauben Parameterstudien und die Berechnung von Interaktionsdiagrammen, die einen schnellen Überblick über mögliche Betriebsbereiche vermitteln. Es zeigt sich, daß abhängig von der Komponente und ihren Belastungen teilweise entscheidende Sicherheitsgewinne zur Erweiterung der Betriebsbereiche erzielt werden können. Das Vorgehen erfordert vom Anwender oft ein gewisses Umdenken. Es werden keine Spannungen berechnet, um damit Sicherheit und Lebensdauer zu interpretieren. Statt dessen berechnet man direkt die gesuchte Sicherheit. Der Post-Prozessor wird nur noch zur Modell- und Rechenkontrolle benötigt. Das Vorgehen ist ähnlich der Stabilitätsanalyse (Knicken, Beulen). Durch namhafte industrielle Projektpartner werden Validierung und die Anwendbarkeit auf eine breite Palette technischer Probleme garantiert. Die ebenfalls in LISA entwickelten Zuverlässigkeitsanalysen sind nichlinear erst auf der Basis direkter Verfahren effektiv möglich. Ohne Traglast- und Einspielanalyse ist plastische Strukturoptimierung auch heute kaum durchführbar. Auf die vorgesehenen Erweiterungen der Werkstoffmodellierung für nichtlineare Verfestigung und für Schädigung konnte hier nicht eingegangen werden. Es herrscht ein deutlicher Mangel an Experimenten zum Nachweis der Grenzen zwischen elastischem Einspielen und dem Versagen durch LCF oder durch Ratchetting.
Traglast- und Einspielanalysen sind vereinfachte doch exakte Verfahren der Plastizität, die neben ausreichender Verformbarkeit keine einschränkenden Voraussetzungen beinhalten. Die Vereinfachungen betreffen die Beschaffung der Daten und Modelle für Details der Lastgeschichte und des Stoffverhaltens. Anders als die klassische Behandlung nichtlinearer Probleme der Strukturmechanik führt die Methode auf Optimierungsprobleme. Diese sind bei realistischen FEM-Modellen sehr groß. Das hat die industrielle Anwendung der Traglast- und Einspielanalysen stark verzögert. Diese Situation wird durch das Brite-EuRam Projekt LISA grundlegend geändert. Die Autoren möchten der Europäischen Kommission an dieser Stelle für die Förderung ausdrücklich danken. In LISA entsteht auf der Basis des industriellen FEM-Programms PERMAS ein Verfahren zur direkten Berechnung der Tragfähigkeit duktiler Strukturen. Damit kann der Betriebsbereich von Komponenten und Bauwerken auf den plastischen Bereich erweitert werden, ohne den Aufwand gegenüber elastischen Analysen wesentlich zu erhöhen. Die beachtlichen Rechenzeitgewinne erlauben Parameterstudien und die Berechnung von Interaktionsdiagrammen, die einen schnellen Überblick über mögliche Betriebsbereiche vermitteln. Es zeigt sich, daß abhängig von der Komponente und ihren Belastungen teilweise entscheidende Sicherheitsgewinne zur Erweiterung der Betriebsbereiche erzielt werden können. Das Vorgehen erfordert vom Anwender oft ein gewisses Umdenken. Es werden keine Spannungen berechnet, um damit Sicherheit und Lebensdauer zu interpretieren. Statt dessen berechnet man direkt die gesuchte Sicherheit. Der Post-Prozessor wird nur noch zur Modell- und Rechenkontrolle benötigt. Das Vorgehen ist änhlich der Stabilitätsanalyse (Knicken, Beulen). Durch namhafte industrielle Projektpartner werden Validierung und die Anwendbarkeit auf eine breite Palette technischer Probleme garantiert. Die ebenfalls in LISA geplante Zuverlässigkeitsanalyse ist erst auf der Basis direkter Verfahren effektiv möglich. Ohne Traglast- und Einspielanalyse ist plastische Strukturoptimierung auch heute kaum durchführbar.