Refine
Year of publication
- 2016 (284) (remove)
Institute
- Fachbereich Medizintechnik und Technomathematik (54)
- Fachbereich Chemie und Biotechnologie (44)
- Fachbereich Bauingenieurwesen (35)
- Fachbereich Elektrotechnik und Informationstechnik (35)
- IfB - Institut für Bioengineering (35)
- Fachbereich Wirtschaftswissenschaften (31)
- Fachbereich Luft- und Raumfahrttechnik (28)
- Fachbereich Maschinenbau und Mechatronik (26)
- Fachbereich Energietechnik (17)
- INB - Institut für Nano- und Biotechnologien (15)
Document Type
- Article (118)
- Conference Proceeding (81)
- Part of a Book (29)
- Book (25)
- Other (10)
- Conference: Meeting Abstract (8)
- Report (6)
- Doctoral Thesis (3)
- Part of a Periodical (2)
- Patent (1)
Keywords
- Technical Operations Research (2)
- Additive Manufacturing (1)
- Annulus Fibrosus (1)
- Archeology (1)
- Assessment (1)
- Asymptotic efficiency (1)
- Bacillus atrophaeus (1)
- Balance (1)
- Balanced hypergraph (1)
- Brandfall (1)
This paper describes the development of a capacitively coupled high-pressure lamp with input power between 20 and 43 W at 2.45 GHz, using a coaxial line network. Compared with other electrodeless lamp systems, no cavity has to be used and a reduction in the input power is achieved. Therefore, this lamp is an alternative to the halogen incandescent lamp for domestic lighting. To serve the demands of domestic lighting, the filling of the lamp is optimized over all other resulting requirements, such as high efficacy at low induced powers and fast startups. A workflow to develop RF-driven plasma applications is presented, which makes use of the hot S-parameter technique. Descriptions of the fitting process inside a circuit and FEM simulator are given. Results of the combined ignition and operation network from simulations and measurements are compared. An initial prototype is built and measurements of the lamp's lighting properties are presented along with an investigation of the efficacy optimizations using large signal amplitude modulation. With this lamp, an efficacy of 135 lmW -1 is achieved.
The composition of plant biomass varies depending on the feedstock and pre-treatment conditions and influences its processing in biorefineries. In order to ensure optimal process conditions, the quantitative proportion of the main polymeric components of the pre-treated biomass has to be determined. Current standard procedures for biomass compositional analysis are complex, the measurements are afflicted with errors and therefore often not comparable. Hence, new powerful analytical methods are urgently required to characterize biomass. In this contribution, Differential Scanning Calorimetry (DSC) was applied in combination with multivariate data analysis (MVA) to detect the cellulose content of the plant biomass pretreated by Liquid Hot Water (LHW) and Organosolv processes under various conditions. Unlike conventional techniques, the developed analytic method enables the accurate quantification of monosaccharide content of the plant biomass without any previous sample preparation. It is easy to handle and avoids errors in sample preparation.
For the successful implementation of microfluidic reaction systems, such as PCR and electrophoresis, the movement of small liquid volumes is essential. In conventional lab-on-a-chip-platforms, solvents and samples are passed through defined microfluidic channels with complex flow control installations. The droplet actuation platform presented here is a promising alternative. With it, it is possible to move a liquid drop (microreactor) on a planar surface of a reaction platform (lab-in-a-drop). The actuation of microreactors on the hydrophobic surface of the platform is based on the use of magnetic forces acting on the outer shell of the liquid drops which is made of a thin layer of superhydrophobic magnetite particles. The hydrophobic surface of the platform is needed to avoid any contact between the liquid core and the surface to allow a smooth movement of the microreactor. On the platform, one or more microreactors with volumes of 10 µL can be positioned and moved simultaneously. The platform itself consists of a 3 x 3 matrix of electrical double coils which accommodate either neodymium or iron cores. The magnetic field gradients are automatically controlled. By variation of the magnetic field gradients, the microreactors' magnetic hydrophobic shell can be manipulated automatically to move the microreactor or open the shell reversibly. Reactions of substrates and corresponding enzymes can be initiated by merging the microreactors or bringing them into contact with surface immobilized catalysts.
In the field of biotechnology and molecular biology, the use of small liquid volumes has significant advantages. In particular, screening and optimization runs with acceptable amounts of expensive and hardly available catalysts, reagents, or biomolecules are feasible with microfluidic technologies. The presented new microfluidic system is based on the inclusion of small liquid volumes by a protective shell of magnetizable microparticles. Hereby, discrete aqueous microreactor drops with volumes of 1–30 μL can be formed on a simple planar surface. A digital movement and manipulation of the microreactor is performed by overlapping magnetic forces. The magnetic forces are generated by an electrical coil matrix positioned below a glass plate. With the new platform technology, several discrete reaction compartments can be moved simultaneously on one surface. Due to the magnetic fields, the reactors can even be merged to initiate reactions by mixing or positioned above surface-immobilized catalysts and then opened by magnetic force. Comparative synthesis routes of the magnetizable shell particles and superhydrophobic glass slides including their performance and stability with the reaction platform are described. The influence of diffusive mass transport during the catalyzed reaction is discussed by evaluation finite element model of the microreactor. Furthermore, a first model dye reaction of the enzyme laccase has been established.
Evaluation of lignocellulosic material for butanol production using enzymatic hydrolysate medium
(2016)
Butanol is a promising gasoline additive and platform chemical that can be readily produced via acetone-butanolethanol (ABE) fermentation from pretreated lignocellulosic materials. This article examines lignocellulosic material from beech wood for ABE fermentation, using Clostridium acetobutylicum. First, the utilization of both C₅₋ (xylose) and C₆₋ (glucose) sugars as sole carbon source was investigated in static cultivation, using serum bottles and synthetic medium. The utilization of pentose sugar resulted in a solvent yield of 0.231 g·g_sugar⁻¹, compared to 0.262 g·g_sugar⁻¹ using hexose. Then, the Organosolv pretreated crude cellulose fibers (CF) were enzymatically decomposed, and the resulting hydrolysate medium was analyzed for inhibiting compounds (furans, organic acids, phenolics) and treated with ionexchangers for detoxification. Batch fermentation in a bioreactor using CF hydrolysate medium resulted in a total solvent yield of 0.20 gABE·g_sugar⁻¹.
The interplay of albumin (BSA) and lysozyme (LYZ) adsorbed simultaneously on titanium was analyzed by gel electrophoresis and BCA assay. It was found that BSA and lysozyme adsorb cooperatively. Additionally, the isoelectric point of the respective protein influences the adsorption. Also, the enzymatic activity of lysozyme and amylase (AMY) in mixtures with BSA was considered with respect to a possible influence of protein-protein interaction on enzyme activity. Indeed, an increase of lysozyme activity in the presence of BSA could be observed. In contrast, BSA does not influence the activity of amylase.
Schienenverkehrssysteme stehen in zunehmendem Wettbewerb, sowohl untereinander als auch mit anderen Verkehrsträgern. Als wichtiger Aspekt zur Steigerung der Kosteneffizienz wird die Digitalisierung des Betriebs und der Fahrzeuge betrachtet. Über eine Prognose der Ausfallwahrscheinlichkeit bzw. Restlebensdauer von Subsystemen können mittels Digitalisierung die Instandhaltungskosten gesenkt werden. Die geringen Fehlerraten im System Bahn machen die Nutzung besonderer Simulationstechniken notwendig. In diesem Beitrag wird gezeigt, wie sich die Subsystemverfügbarkeit aus den beobachteten Fehlerraten der Teilfunktionen vorhersagen lässt.
This work presents a methodology for automated
damage-sensitive feature extraction and anomaly
detection under multivariate operational variability
for in-flight assessment of wings. The
method uses a passive excitation approach, i. e.
without the need for artificial actuation. The
modal system properties (natural frequencies and
damping ratios) are used as damage-sensitive
features. Special emphasis is placed on the use
of Fiber Bragg Grating (FBG) sensing technology
and the consideration of Operational and
Environmental Variability (OEV). Measurements
from a wind tunnel investigation with a composite
cantilever equipped with FBG and piezoelectric
sensors are used to successfully detect an impact
damage. In addition, the feasibility of damage
localisation and severity estimation is evaluated
based on the coupling found between damageand
OEV-induced feature changes.
The main objective of the BATIMASS project was to address how the energy balance in relatively lightweight steel buildings can be improved by building in ‘active thermal mass’ (ATM) into the building fabric. This was achieved through concept design, dynamic thermal modelling and testing of a number of potentially viable systems and concepts. A significant programme of thermal simulation modelling was undertaken utilising the thermally equivalent slab (TES) concept to model the passive thermal capacity effect of profiled, composite metal floor decks. It is apparent from the modelling results that thermal mass is a highly complex phenomenon which is highly dependent upon building type, occupancy patterns, climate and many other aspects of the building design and servicing strategy. The ATM systems developed, both conceptually and for prototype testing, focussed on water-cooled composite slabs, the Cofradal floor system and the phase change material (PCM) Energain. In addition to laboratory testing of prototypes, whole building monitoring was undertaken at the Kubik building in Spain and the RWTH test building in Germany. Advanced thermal modelling was also undertaken to estimate the likely benefits of the ATM concept designs developed and for comparison with the test results. In addition to thermal testing, structural tests were conducted on composite floor specimens incorporating embedded water pipes. This Final Report presents the results of the activities carried out under this RFCS contract RFSR CT 2012 00033. The work carried out is reported in six major sections corresponding to the technical Work Packages of the project. Only summaries of the work carried out are provided in this report; all work undertaken is fully reported in the formal project deliverables.
Purpose
To assess the feasibility of prostate ¹H MR spectroscopic imaging (MRSI) using low-power spectral-spatial (SPSP) pulses at 7T, exploiting accurate spectral selection and spatial selectivity simultaneously.
Methods
A double spin-echo sequence was equipped with SPSP refocusing pulses with a spectral selectivity of 1 ppm. Three-dimensional prostate ¹H-MRSI at 7T was performed with the SPSP-MRSI sequence using an 8-channel transmit array coil and an endorectal receive coil in three patients with prostate cancer and in one healthy subject. No additional water or lipid suppression pulses were used.
Results
Prostate ¹H-MRSI could be obtained well within specific absorption rate (SAR) limits in a clinically feasible time (10 min). Next to the common citrate signals, the prostate spectra exhibited high spermine signals concealing creatine and sometimes also choline. Residual lipid signals were observed at the edges of the prostate because of limitations in spectral and spatial selectivity.
Conclusion
It is possible to perform prostate ¹H-MRSI at 7T with a SPSP-MRSI sequence while using separate transmit and receive coils. This low-SAR MRSI concept provides the opportunity to increase spatial resolution of MRSI within reasonable scan times.
Compared to peripheral pain, trigeminal pain elicits higher levels of fear, which is assumed to enhance the interruptive effects of pain on concomitant cognitive processes. In this fMRI study we examined the behavioral and neural effects of trigeminal (forehead) and peripheral (hand) pain on visual processing and memory encoding. Cerebral activity was measured in 23 healthy subjects performing a visual categorization task that was immediately followed by a surprise recognition task. During the categorization task subjects received concomitant noxious electrical stimulation on the forehead or hand. Our data show that fear ratings were significantly higher for trigeminal pain. Categorization and recognition performance did not differ between pictures that were presented with trigeminal and peripheral pain. However, object categorization in the presence of trigeminal pain was associated with stronger activity in task-relevant visual areas (lateral occipital complex, LOC), memory encoding areas (hippocampus and parahippocampus) and areas implicated in emotional processing (amygdala) compared to peripheral pain. Further, individual differences in neural activation between the trigeminal and the peripheral condition were positively related to differences in fear ratings between both conditions. Functional connectivity between amygdala and LOC was increased during trigeminal compared to peripheral painful stimulation. Fear-driven compensatory resource activation seems to be enhanced for trigeminal stimuli, presumably due to their exceptional biological relevance.
Objective
This study assesses and quantifies impairment of postoperative magnetic resonance imaging (MRI) at 7 Tesla (T) after implantation of titanium cranial fixation plates (CFPs) for neurosurgical bone flap fixation.
Materials and methods
The study group comprised five patients who were intra-individually examined with 3 and 7 T MRI preoperatively and postoperatively (within 72 h/3 months) after implantation of CFPs. Acquired sequences included T₁-weighted magnetization-prepared rapid-acquisition gradient-echo (MPRAGE), T₂-weighted turbo-spin-echo (TSE) imaging, and susceptibility-weighted imaging (SWI). Two experienced neurosurgeons and a neuroradiologist rated image quality and the presence of artifacts in consensus reading.
Results
Minor artifacts occurred around the CFPs in MPRAGE and T2 TSE at both field strengths, with no significant differences between 3 and 7 T. In SWI, artifacts were accentuated in the early postoperative scans at both field strengths due to intracranial air and hemorrhagic remnants. After resorption, the brain tissue directly adjacent to skull bone could still be assessed. Image quality after 3 months was equal to the preoperative examinations at 3 and 7 T.
Conclusion
Image quality after CFP implantation was not significantly impaired in 7 T MRI, and artifacts were comparable to those in 3 T MRI.
Purpose
To calculate local specific absorption rate (SAR) correctly, both the amplitude and phase of the signal in each transmit channel have to be known. In this work, we propose a method to derive a conservative upper bound for the local SAR, with a reasonable safety margin without knowledge of the transmit phases of the channels.
Methods
The proposed method uses virtual observation points (VOPs). Correction factors are calculated for each set of VOPs that prevent underestimation of local SAR when the VOPs are applied with the correct amplitudes but fixed phases.
Results
The proposed method proved to be superior to the worst-case calculation based on the maximum eigenvalue of the VOPs. The mean overestimation for six coil setups could be reduced, whereas no underestimation of the maximum local SAR occurred. In the best investigated case, the overestimation could be reduced from a factor of 3.3 to a factor of 1.7.
Conclusion
The upper bound for the local SAR calculated with the proposed method allows a fast estimation of the local SAR based on power measurements in the transmit channels and facilitates SAR monitoring in systems that do not have the capability to monitor transmit phases
Zum Zweck der Kapitalanlage beteiligte sich der Bekl. an einer Publikumsgesellschaft. Diese war als (mehrgliedrige) atypische stille Gesellschaft organisiert (vgl. a. BGHZ 199, 104 = DNotZ 2014, 374 = NZG 2013, 1422 = DStR 2014, 45 Rn. 18). Der Gesellschaftsvertrag („GV“) sah für diejenigen Gesellschafter, die wie der Bekl. ihre Einlage in Form einer Einmaleinlage erbracht hatten, eine jährliche gewinnunabhängige Ausschüttung vor. Es sollte sich dabei ausdrücklich nicht um eine Garantieverzinsung handeln. Ende 2009 wurde die stille Gesellschaft durch Mehrheitsbeschluss der Stillen aufgelöst. Nach dem GV waren die Stillen im Falle ihres Ausscheidens sowie bei „Liquidation des Unternehmens“ des Geschäftsinhabers verhältnismäßig an dem jeweils seit ihrem Beitritt gebildeten Vermögen einschließlich der stillen Reserven sowie eines evtl. Geschäftswerts zu beteiligen (Auseinandersetzungswert). Den sich hiernach für den Bekl. auf seinem Kapitalkonto ergebenden Negativsaldo sollte dieser durch Erstattung der von ihm erhaltenen gewinnunabhängigen Auszahlungen ausgleichen. Anders als das LG als Berufungsinstanz bejahte der BGH auf Grundlage des GV einen Rückerstattungsanspruch der klagenden Geschäftsinhaberin.
Robotergestütztes System für ein verbessertes neuromuskuläres Aufbautraining der Beinstrecker
(2016)
Neuromuskuläres Aufbautraining der Beinstrecker ist ein wichtiger Bestandteil in der Rehabilitation und Prävention von Muskel-Skelett-Erkrankungen. Effektives Training erfordert hohe Muskelkräfte, die gleichzeitig hohe Belastungen von bereits geschädigten Strukturen bedeuten. Um trainingsinduzierte Schädigungen zu vermeiden, müssen diese Kräfte kontrolliert werden. Mit heutigen Trainingsgeräten können diese Ziele allerdings nicht erreicht werden. Für ein sicheres und effektives Training sollen durch den Einsatz der Robotik, Sensorik, eines Regelkreises sowie Muskel-Skelett-Modellen Belastungen am Zielgewebe direkt berechnet und kontrolliert werden. Auf Basis zweier Vorstudien zu möglichen Stellgrößen wird der Aufbau eines robotischen Systems vorgestellt, das sowohl für Forschungszwecke als auch zur Entwicklung neuartiger Trainingsgeräte verwendet werden kann.
Für die Verarbeitung von natürlicher Sprache ist ein wichtiger Zwischenschritt das Parsing, bei dem für Sätze der natürlichen Sprache Ableitungsbäume bestimmt werden. Dieses Verfahren ist vergleichbar zum Parsen formaler Sprachen, wie z. B. das Parsen eines Quelltextes. Die Parsing-Methoden der formalen Sprachen, z. B. Bottom-up-Parser, können nicht auf das Parsen der natürlichen Sprache übertragen werden, da keine Formalisierung der natürlichen Sprachen existiert [3, 12, 23, 30].
In den ersten Programmen, die natürliche Sprache verarbeiten [32, 41], wurde versucht die natürliche Sprache mit festen Regelmengen zu verarbeiten. Dieser Ansatz stieß jedoch schnell an seine Grenzen, da die Regelmenge nicht vollständig sowie nicht minimal ist und wegen der benötigten Menge an Regeln schwer zu verwalten ist. Die Korpuslinguistik [22] bot die Möglichkeit, die Regelmenge durch Supervised-Machine-Learning-Verfahren [2] abzulösen.
Teil der Korpuslinguistik ist es, große Textkorpora zu erstellen und diese mit sprachlichen Strukturen zu annotieren. Zu diesen Strukturen gehören sowohl die Wortarten als auch die Ableitungsbäume der Sätze. Vorteil dieser Methodik ist es, dass repräsentative Daten zur Verfügung stehen. Diese Daten werden genutzt, um mit Supervised-Machine-Learning-Verfahren die Gesetzmäßigkeiten der natürliche Sprachen zu erlernen.
Das Maximum-Entropie-Verfahren ist ein Supervised-Machine-Learning-Verfahren, das genutzt wird, um natürliche Sprache zu erlernen. Ratnaparkhi [25] nutzt Maximum-Entropie, um Ableitungsbäume für Sätze der natürlichen Sprache zu erlernen. Dieses Verfahren macht es möglich, die natürliche Sprache (abgebildet als Σ∗) trotz einer fehlenden formalen Grammatik zu parsen.
This summer, RoboCup competitions were held for the 20th time in Leipzig, Germany. It was the second time that RoboCup took place in Germany, 10 years after the 2006 RoboCup in Bremen. In this article, we give an overview on the latest developments of RoboCup and what happened in the different leagues over the last decade. With its 20th edition, RoboCup clearly is a success story and a role model for robotics competitions. From our personal view point, we acknowledge this by giving a retrospection about what makes RoboCup such a success.
Mit steigenden Dämmstandards und höheren Komfortanforderungen der Nutzer gerät die Problematik der sommerlichen Überhitzung zunehmend in den Fokus. Um die Überhitzung möglichst gering zu halten, sind Maßnahmen und Lösungen zu entwickeln, die den potenziellen Kühlbedarf eines Gebäudes vermeiden sowie reduzieren. Im Rahmen des europäischen Forschungsprojektes BATIMASS wurden Techniken untersucht, die die sommerliche Raumtemperatur ohne zusätzliche Kühlung (passiv) oder aber mit energieeffizienter wasserbasierter Flächenkühlung (aktiv) reduzieren und die besonders für Gebäude in Stahl(leicht)bauweise geeignet sind. Dafür wurde die Methodik der thermisch äquivalenten Decke weiterentwickelt, um das thermische Verhalten von Profilblechdecken in Gebäuden für beide Lösungsansätze analysieren zu können. Darüber hinaus wurde der Einsatz von Phasenwechselmaterial (PCM) zur Steigerung der Speicherfähigkeit von leichten Decken mit besonders geringer thermischer Masse in Simulationen sowie im Labor untersucht und bewertet.
This article discusses the contrast between the information transportation companies provide to travellers and that of their brand messaging. Companies’ brand messaging often portrays the service they provide as pleasant, stress free and perfect. Customers and users of the service, on the other hand, often describe their experience of the service as a negative one. This article suggests that the brand value would be greater if transportation companies paid more attention to the users’ experience when designing their information systems, particularly in worst case scenarios.
The enormous diversity of seed traits is an intriguing feature and critical for the overwhelming success of higher plants. In particular, seed mass is generally regarded to be key for seedling development but is mostly approximated by using scanning methods delivering only two-dimensional data, often termed seed size. However, three-dimensional traits, such as the volume or mass of single seeds, are very rarely determined in routine measurements. Here, we introduce a device named phenoSeeder, which enables the handling and phenotyping of individual seeds of very different sizes. The system consists of a pick-and-place robot and a modular setup of sensors that can be versatilely extended. Basic biometric traits detected for individual seeds are two-dimensional data from projections, three-dimensional data from volumetric measures, and mass, from which seed density is also calculated. Each seed is tracked by an identifier and, after phenotyping, can be planted, sorted, or individually stored for further evaluation or processing (e.g. in routine seed-to-plant tracking pipelines). By investigating seeds of Arabidopsis (Arabidopsis thaliana), rapeseed (Brassica napus), and barley (Hordeum vulgare), we observed that, even for apparently round-shaped seeds of rapeseed, correlations between the projected area and the mass of seeds were much weaker than between volume and mass. This indicates that simple projections may not deliver good proxies for seed mass. Although throughput is limited, we expect that automated seed phenotyping on a single-seed basis can contribute valuable information for applications in a wide range of wild or crop species, including seed classification, seed sorting, and assessment of seed quality.
Wind-induced operational variability is one of the major challenges for structural health monitoring of slender engineering structures like aircraft wings or wind turbine blades. Damage sensitive features often show an even bigger sensitivity to operational variability. In this study a composite cantilever was subjected to multiple mass configurations, velocities and angles of attack in a controlled wind tunnel environment. A small-scale impact damage was introduced to the specimen and the structural response measurements were repeated. The proposed damage detection methodology is based on automated operational modal analysis. A novel baseline preparation procedure is described that reduces the amount of user interaction to the provision of a single consistency threshold. The procedure starts with an indeterminate number of operational modal analysis identifications from a large number of datasets and returns a complete baseline matrix of natural frequencies and damping ratios that is suitable for subsequent anomaly detection. Mahalanobis distance-based anomaly detection is then applied to successfully detect the damage under varying severities of operational variability and with various degrees of knowledge about the present operational conditions. The damage detection capabilities of the proposed methodology were found to be excellent under varying velocities and angles of attack. Damage detection was less successful under joint mass and wind variability but could be significantly improved through the provision of the currently encountered operational conditions.
Label-free Electrostatic Detection of DNA Amplification by PCR Using Capacitive Field-effect Devices
(2016)
A capacitive field-effect EIS (electrolyte-insulator-semiconductor) sensor modified with a positively charged weak polyelectrolyte of poly(allylamine hydrochloride) (PAH)/single-stranded probe DNA (ssDNA) bilayer has been used for a label-free electrostatic detection of pathogen-specific DNA amplification via polymerase chain reaction (PCR). The sensor is able to distinguish between positive and negative PCR solutions, to detect the existence of target DNA amplicons in PCR samples and thus, can be used as tool for a quick verification of DNA amplification and the successful PCR process.
A New Class of Biosensors Based on Tobacco Mosaic Virus and Coat Proteins as Enzyme Nanocarrier
(2016)
Air-water flows can be found in different engineering applications: from nuclear engineering to huge hydraulic structures. In this paper, a single tip fibre optical probe has been used to record high frequency (over 1 MHz) phase functions at different locations of a stepped spillway. These phase functions have been related to the interfacial velocities by means of Artificial Neural Networks (ANN) and the measurements of a classical double tip conductivity probe. Special attention has been put to the input selection and the ANN dimensions. Finally, ANN have shown to be able to link the signal rising times and plateau shapes to the air-water interfacial velocity.
Leitung und Konzeption: Thomas Tünnemann, Professor für Gestalten, FB Architektur, FH Aachen:
Mit Beiträgen von
Miriam Azzab, Larissa Rohr, Janna Steinhart, Felix Reymann, Kevin Osterkamp, Mark Kieckhefer, Pia Bienert, Sarah Schuhmann, Hermann Stuzmann, Silvana Hecklinger, Thorsten, Köllen, Jesse Dilworth, Silke Wanders, Tomas Tünnemann und Jan Waschinzki
Genussrechte als Instrument der Unternehmensfinanzierung von kommunalen Wohnungs- baugesellschaften
(2016)
Die kommunalen Wohnungsbaugesellschaften und ihr Geschäftszweck bzw. Versorgungsauftrag erfahren aus gesamtgesellschaftlichen Gründen aktuell einen erheblichen Bedeutungszuwachs. Der kommunale Wohnungsbau soll und muss intensiviert werden, da z.B. zunehmend Wohnungsbindungen auslaufen. Bezahlbare Neubauten fehlen speziell Familien, Senioren, Studenten und Flüchtlingen. Um das Investitionsvolumen zu bewältigen, bedarf es nicht nur der Förderkredite, sondern auch der Zuführung „frischen Kapitals“. Der Beitrag stellt dar, dass Genussrechte ein interessantes und oft unbekanntes Finanzierungsinstrument für kommunale Wohnungsbaugesellschaften mit passgenauen Ausgestaltungsmöglichkeiten sind. Mit Hilfe von Genussrechten können die notwendigen Neubau-Investitionen solide mit Quasi-Eigenkapital unterlegt werden und den sozialen Wohnungsbau wiederbeleben. Genussrechte als Hybridkapital lassen sich mangels gesetzlicher Eingrenzung exakt und individuell ebenso an die Bedürfnisse der kommunalen Wohnungsbaugesellschaft anpassen wie an die der Gesellschafter und des jeweiligen Neubauvorhabens.
The Dry-Low-NOₓ (DLN) Micromix combustion technology has been developed as low emission combustion principle for industrial gas turbines fueled with hydrogen or syngas. The combustion process is based on the phenomenon of jet-in-crossflow-mixing. Fuel is injected perpendicular into the air-cross-flow and burned in a multitude of miniaturized, diffusion-like flames. The miniaturization of the flames leads to a significant reduction of NOₓ emissions due to the very short residence time of reactants in the flame.
In the Micromix research approach, CFD analyses are validated towards experimental results. The combination of numerical and experimental methods allows an efficient design and optimization of DLN Micromix combustors concerning combustion stability and low NOₓ emissions.
The paper presents a comparison of several numerical combustion models for hydrogen and hydrogen-rich syngas. They differ in the complexity of the underlying reaction mechanism and the associated computational effort.
For pure hydrogen combustion a one-step global reaction is applied using a hybrid Eddy-Break-up model that incorporates finite rate kinetics. The model is evaluated and compared to a detailed hydrogen combustion mechanism derived by Li et al. including 9 species and 19 reversible elementary reactions. Based on this mechanism, reduction of the computational effort is achieved by applying the Flamelet Generated Manifolds (FGM) method while the accuracy of the detailed reaction scheme is maintained.
For hydrogen-rich syngas combustion (H₂-CO) numerical analyses based on a skeletal H₂/CO reaction mechanism derived by Hawkes et al. and a detailed reaction mechanism provided by Ranzi et al. are performed.
The comparison between combustion models and the validation of numerical results is based on exhaust gas compositions available from experimental investigation on DLN Micromix combustors.
The conducted evaluation confirms that the applied detailed combustion mechanisms are able to predict the general physics of the DLN-Micromix combustion process accurately. The Flamelet Generated Manifolds method proved to be generally suitable to reduce the computational effort while maintaining the accuracy of detailed chemistry.
Especially for reaction mechanisms with a high number of species accuracy and computational effort can be balanced using the FGM model.
Combined with the use of renewable energy sources for its production, Hydrogen represents a possible alternative gas turbine fuel within future low emission power generation. Due to the large difference in the physical properties of Hydrogen compared to other fuels such as natural gas, well established gas turbine combustion systems cannot be directly applied for Dry Low NOx (DLN) Hydrogen combustion. Thus, the development of DLN combustion technologies is an essential and challenging task for the future of Hydrogen fuelled gas turbines. The DLN Micromix combustion principle for hydrogen fuel has been developed to significantly reduce NOx-emissions. This combustion principle is based on cross-flow mixing of air and gaseous hydrogen which reacts in multiple miniaturized diffusion-type flames. The major advantages of this combustion principle are the inherent safety against flash-back and the low NOx-emissions due to a very short residence time of reactants in the flame region of the micro-flames. The Micromix Combustion technology has been already proven experimentally and numerically for pure Hydrogen fuel operation at different energy density levels. The aim of the present study is to analyze the influence of different geometry parameter variations on the flame structure and the NOx emission and to identify the most relevant design parameters, aiming to provide a physical understanding of the Micromix flame sensitivity to the burner design and identify further optimization potential of this innovative combustion technology while increasing its energy density and making it mature enough for real gas turbine application. The study reveals great optimization potential of the Micromix Combustion technology with respect to the DLN characteristics and gives insight into the impact of geometry modifications on flame structure and NOx emission. This allows to further increase the energy density of the Micromix burners and to integrate this technology in industrial gas turbines.
Combined with the use of renewable energy sources for its production, hydrogen represents a possible alternative gas turbine fuel for future low-emission power generation. Due to the difference in the physical properties of hydrogen compared to other fuels such as natural gas, well-established gas turbine combustion systems cannot be directly applied to dry low NOₓ (DLN) hydrogen combustion. The DLN micromix combustion of hydrogen has been under development for many years, since it has the promise to significantly reduce NOₓ emissions. This combustion principle for air-breathing engines is based on crossflow mixing of air and gaseous hydrogen. Air and hydrogen react in multiple miniaturized diffusion-type flames with an inherent safety against flashback and with low NOₓ emissions due to a very short residence time of the reactants in the flame region. The paper presents an advanced DLN micromix hydrogen application. The experimental and numerical study shows a combustor configuration with a significantly reduced number of enlarged fuel injectors with high-thermal power output at constant energy density. Larger fuel injectors reduce manufacturing costs, are more robust and less sensitive to fuel contamination and blockage in industrial environments. The experimental and numerical results confirm the successful application of high-energy injectors, while the DLN micromix characteristics of the design point, under part-load conditions, and under off-design operation are maintained. Atmospheric test rig data on NOₓ emissions, optical flame-structure, and combustor material temperatures are compared to numerical simulations and show good agreement. The impact of the applied scaling and design laws on the miniaturized micromix flamelets is particularly investigated numerically for the resulting flow field, the flame-structure, and NOₓ formation.
Turbulent dispersion in bounded horizontal jets : RANS capabilities and physical modeling comparison
(2016)
Low-end-Embedded-Plattformen stellen eine hohe Anforderung an die Entscheidungsfähigkeit des Entwicklers: Zum nächstgrößeren Prozessor greifen und ein Betriebssystem benutzen oder doch besser auf das Betriebssystem verzichten? Die Frage lässt sich einfach beantworten: Einen Nanokernel verwenden und das Embedded-System mit einem minimalen Footprint realisieren. Adam Dunkels Protothreads sind eine ausgesprochen effiziente Art, Mikrocontroller gut strukturiert zu programmieren und gleichzeitig auf Overhead zu verzichten. So können auch mit kleinen 8-bit-Prozessoren anspruchsvolle Aufgaben in einem Thread-Modell bearbeitet werden. Man muss also nicht immer das Rad neu erfinden oder gleich auf Linux-basierte Systeme zurückgreifen.
Optical flow estimation is known from Computer Vision where it is used to determine obstacle movements through a sequence of images following an assumption of brightness conservation. This paper presents the first study on application of the optical flow method to aerated stepped spillway flows. For this purpose, the flow is captured with a high-speed camera and illuminated with a synchronized LED light source. The flow velocities, obtained using a basic Horn–Schunck method for estimation of the optical flow coupled with an image pyramid multi-resolution approach for image filtering, compare well with data from intrusive conductivity probe measurements. Application of the Horn–Schunck method yields densely populated flow field data sets with velocity information for every pixel. It is found that the image pyramid approach has the most significant effect on the accuracy compared to other image processing techniques. However, the final results show some dependency on the pixel intensity distribution, with better accuracy found for grey values between 100 and 150.
IoT von der Stange
(2016)
To better understand what kinds of sports and exercise could be beneficial for the intervertebral disc (IVD), we performed a review to synthesise the literature on IVD adaptation with loading and exercise. The state of the literature did not permit a systematic review; therefore, we performed a narrative review. The majority of the available data come from cell or whole-disc loading models and animal exercise models. However, some studies have examined the impact of specific sports on IVD degeneration in humans and acute exercise on disc size. Based on the data available in the literature, loading types that are likely beneficial to the IVD are dynamic, axial, at slow to moderate movement speeds, and of a magnitude experienced in walking and jogging. Static loading, torsional loading, flexion with compression, rapid loading, high-impact loading and explosive tasks are likely detrimental for the IVD. Reduced physical activity and disuse appear to be detrimental for the IVD. We also consider the impact of genetics and the likelihood of a ‘critical period’ for the effect of exercise in IVD development. The current review summarises the literature to increase awareness amongst exercise, rehabilitation and ergonomic professionals regarding IVD health and provides recommendations on future directions in research.
Background and Objective
Effective leg extension training at a leg press requires high forces, which need to be controlled to avoid training-induced damage. In order to avoid high external knee adduction moments, which are one reason for unphysiological loadings on knee joint structures, both training movements and the whole reaction force vector need to be observed. In this study, the applicability of lateral and medial changes in foot orientation and position as possible manipulated variables to control external knee adduction moments is investigated. As secondary parameters both the medio-lateral position of the center of pressure and the frontal-plane orientation of the reaction force vector are analyzed.
Methods
Knee adduction moments are estimated using a dynamic model of the musculoskeletal system together with the measured reaction force vector and the motion of the subject by solving the inverse kinematic and dynamic problem. Six different foot conditions with varying positions and orientations of the foot in a static leg press are evaluated and compared to a neutral foot position.
Results
Both lateral and medial wedges under the foot and medial and lateral shifts of the foot can influence external knee adduction moments in the presented study with six healthy subjects. Different effects are observed with the varying conditions: the pose of the leg is changed and the direction and center of pressure of the reaction force vector is influenced. Each effect results in a different direction or center of pressure of the reaction force vector.
Conclusions
The results allow the conclusion that foot position and orientation can be used as manipulated variables in a control loop to actively control knee adduction moments in leg extension training.
In Deutschland liegt der Anteil der Windkraft an der Gesamtstromerzeugung bei 13,3% mit mehr als 25.000 installierten Windenergieanlagen (WEA). Weltweit erfährt die Windbranche ein rasantes Wachstum. Indien und China berichten eine jährliche Wachstumsrate an Neuinstallationen von 45%. Die Technologie zur Erzeugung elektrischer Energie aus Windkraft ist noch vergleichsweise jung. Durch die weltweit steigende Anzahl an Windenergieanlagen wächst zunehmend der Bedarf an innovativen Wartungslösungen. Komponenten wie Generator oder Getriebe sind inzwischen weitestgehend ausgereift. Der Fokus richtet sich zunehmend auf die wesentliche Kernkomponente - die Rotorblätter.
Industriekletterer inspizieren die Rotorblätter oder Türme i.d.R.
in einem zwei Jahres Rhythmus. Sie werden zunehmend durch Seilarbeitsbühnen unterstützt. Für größere Reparaturen kommen Kräne zum Einsatz, mit denen das Rotorblatt für die Instandhaltung demontiert wird. Die Standardinspektion besteht aus Sicht- und Klopfprüfung der Rotorblattoberfläche und ist nur bei sehr ruhiger Wetterlage durchführbar. Seit September 2014 wird das Forschungsprojekt SMART (Scanning, Monitoring, Analysis, Repair and Transportation), Entwicklung einer Wartungsplattform für WEA, vom BMWi gefördert. Das Konsortium besteht aus zwei Firmen und der
Fachhochschule Aachen. Die SMART-Anlage klettert reibschlüssig am Turm der WEA mittels speziellen Kettenfahrwerken (Abbildung) auf- und abwärts. Ein ringförmiges Spannsystems, basierend auf dem Konzept der „Nürnberger“-Schere, erzeugt die erforderliche Anpresskraft für den Kletterprozess. Wettergeschützte Arbeitskabinen ermöglichen die ganzjährige Instandhaltung von Rotorblättern und ebenso Türmen. Dadurch können Wartungsarbeiten auf 24 Stunden am Tag ausgeweitet werden. Der kombinierte Einsatz (Sensorfusion) bildgebender Messtechnik wie Thermografie, Ultraschall, und Terahertz in der Arbeitskabine kann die Dokumentation, Effizienz und Qualität der Instandhaltungsarbeiten erheblich verbessern. Langfristiges Ziel von SMART ist ein Condition Monitoring für Rotorblätter und Türme auf Basis digitalisierter dreidimensionaler Volumenscans. Der kooperative Einsatz mit UAVs erweitert die Instandhaltungsstrategie. UAVs ermöglichen die schnelle, kostengünstige globale optische Inspektion von Rotorblattoberflächen zur Detektion potentieller Fehlstellen. Der „Proof-of-Concept“ Meilenstein wurde mit der Demonstration eines funktionsfähigen Modells im Dezember 2015 erfolgreich abgeschlossen.
Robots are widely used as a vehicle to spark interest in science and technology in learners. A number of initiatives focus on this issue, for instance, the Roberta Initiative, the FIRST Lego League, the World Robot Olympiad and RoboCup Junior. Robotic competitions are valuable not only for school learners but also for university students, as the RoboCup initiative shows. Besides technical skills, the students get some project exposure and experience what it means to finish their tasks on time. But qualifying students for future high-tech areas should not only be for students from developed countries. In this article, we present our experiences with research and education in robotics within the RoboCup initiative, in Germany and South Africa; we report on our experiences with trying to get the RoboCup initiative in South Africa going. RoboCup has a huge support base of academic institutions in Germany; this is not the case in South Africa. We present our ‘north–south’ collaboration initiatives in RoboCup between Germany and South Africa and discuss some of the reasons why we think it is harder to run RoboCup in South Africa.
Replacement tissues, designed to fill in articular cartilage defects, should exhibit the same properties as the native material. The aim of this study is to foster the understanding of, firstly, the mechanical behavior of the material itself and, secondly, the influence of cultivation parameters on cell seeded implants as well as on cell migration into acellular implants. In this study, acellular cartilage replacement material is theoretically, numerically and experimentally investigated regarding its viscoelastic properties, where a phenomenological model for practical applications is developed. Furthermore, remodeling and cell migration are investigated.
To give the exchange of goods and services between the European Union (EU) and the United States (U.S.) new momentum the two parties are currently negotiating the transatlantic free trade agreement Transatlantic Trade and Investment Partnership (TTIP). The aim is to create the largest free trade area in the world. The agreement, once entered into force, will oblige EU countries and the U.S. to further liberalize their markets.
The negotiations on TTIP include a chapter on Electronic Communications/ Telecommunications. The challenge therein will be securing commitments for market access to Electronic Communications services. At the same time, these commitments must reflect the legitimate need for consumer protection issues. The need to reduce Electronic Communications-related non-tariff barriers to trade between the Parties is due to the fact that these markets are heavily regulated. Without transnational rules as to regulations national governments can abuse these regulations to deter the market entry by new (foreign) suppliers. Thus the free trade agreement TTIP affects in many respects regulatory provisions on and access to Electronic Communications markets. The objective of this paper is therefore to examine to what extend the regulatory principles for Electronic Communications markets envisaged under TTIP will result in trade facilitation and regulatory convergence between the EU and the U.S.
As to this question the result of the analysis is that the chapter on Electronic Communications will be an important step towards facilitating trade in Electronic Communications services. At the same time some regulatory convergence will take place, but this convergence will not lead to a (full) harmonization of regulations. Rather the norm, also after TTIP negotiations will have been concluded successfully, will be mutual recognition of different regulatory regimes. Different regulations being the optimal policy response in different market settings will continue to exist. Moreover, it is very unlikely that such regulatory principles for the Electronic Communications sector are a vehicle for a race to the bottom in levels of consumer protection.
We present a new Min-Max theorem for an optimization problem closely connected to matchings and vertex covers in balanced hypergraphs. The result generalizes Kőnig’s Theorem (Berge and Las Vergnas in Ann N Y Acad Sci 175:32–40, 1970; Fulkerson et al. in Math Progr Study 1:120–132, 1974) and Hall’s Theorem (Conforti et al. in Combinatorica 16:325–329, 1996) for balanced hypergraphs.
We prove characterizations of the existence of perfect ƒ-matchings in uniform mengerian and perfect hypergraphs. Moreover, we investigate the ƒ-factor problem in balanced hypergraphs. For uniform balanced hypergraphs we prove two existence theorems with purely combinatorial arguments, whereas for non-uniform balanced hypergraphs we show that the ƒ-factor problem is NP-hard.
An equitable graph coloring is a proper vertex coloring of a graph G where the sizes of the color classes differ by at most one. The equitable chromatic number is the smallest number k such that G admits such equitable k-coloring. We focus on enumerative algorithms for the computation of the equitable coloring number and propose a general scheme to derive pruning rules for them: We show how the extendability of a partial coloring into an equitable coloring can be modeled via network flows. Thus, we obtain pruning rules which can be checked via flow algorithms. Computational experiments show that the search tree of enumerative algorithms can be significantly reduced in size by these rules and, in most instances, such naive approach even yields a faster algorithm. Moreover, the stability, i.e., the number of solved instances within a given time limit, is greatly improved.
Since the execution of flow algorithms at each node of a search tree is time consuming, we derive arithmetic pruning rules (generalized Hall-conditions) from the network model. Adding these rules to an enumerative algorithm yields an even larger runtime improvement.
Analysis of the long-term effect of the MBST® nuclear magnetic resonance therapy on gonarthrosis
(2016)
Inhalt
06 Von Aachen nach Hollywood | Am 5. März 2016 lädt Shawn Bu den Star-Wars-Fanfilm, den er für seine Bachelorarbeit produziert hat, bei YouTube hoch. Was dann passiert, ist einfach unglaublich
16 Es geht aufwärts | FH entwickelt Kletterroboter zur Wartung und Reparatur von Windenergieanlagen
20 Auf den Millimeter genau | FH-Bauingenieure vermessen die Urftstaumauer
24 Auf legendären Rennstrecken | FH-Youngster treten in GT-Motorsportserie an
26 "Unten ist da, wo die Erde ist..." | Astronaut Dr. Thomas Reiter trägt an der FH Aachen vor
30 Von der Kita entspannt in den Hörsaal | Wie man den Spagat zwischen Kindererziehung und Studium meistert
32 "Solidarität ist keine Rosinenpickerei" | Martin Schulz und der Spiegel zu Gast an der FH Aachen
34 Ankommen | Drei Beispiele, wie Menschen an der FH Aachen und in der Region sich mit dem Thema Flucht beschäftigen
36 "Wir sind nicht die großen Retter" | Fotodokumentation von Stephan Klumpp
38 Mehr als nur ein Dach über dem Kopf | Dina Bharucha entwickelt ein ausklappbares Flüchlingshaus
40 "Es wird zu wenig über Fluchtursachen gesprochen" | Knut Paul, Pressesprecher der Bundespolizeiinspektion Aachen, über die Flüchtlingslage in der Aachener Region
42 Pinipas Abenteuer | Eine himmlische Pfannkuchensuche durch Europa
44 Warum es sich lohnt, in Nachwuchs zu investieren | Drei Beispiele aus dem FH-Stipendienprogramm
46 Rückenschmerzen ade | Das Start-up-Unternehmen physiosense entwickelt einen "intelligenten" Bürostuhl
48 Fahren und sparen | FH-Wissenschaftler Jörg Kreyer erforscht, wie der Kraftstoffverbrauch von Nutzfahrzeugen gesenkt werden kann
52 Informationsdesign für die Industrie 4.0 | "ProSense": Fachbereich Gestaltung entwickelt neues Verfahren zur visuellen Darstellung großer Datenmengen
58 Maßgeschneidert für unsere Region | Der Studiengang "Open Borders MBA" richtet sich an Führungskräfte, die grenzüberschreitend arbeiten wollen
60 Rennsport im Miniaturformat| FH unterstützt das Einhard-Gymnasium bei der Junior-Ingenieur-Akademie
61 Mathematik der Herzen
63 Impressum
Inhalt
06 Fast jeder Zweite ist betroffen | Psychosoziale Beratung hilft Studierenden
10 Hacken statt Knacken | IT-Sicherheit im Automobil: Kooperation der FH Aachen mit Interpol
13 Dritter Platz für die FH Aachen beim PR-Bild Award
14 Ich bin ein Berlinger | Tourbus Wolfgang chauffiert keine Rockstars, sondern 3D-Drucker
18 Licht aus, Spot an und ... studieren | Fachbereich Wirtschaftswissenschaften dreht Imagefilm
22 Ein Tropfen Blut genügt | FH-Absolventen forschen in Grundlagenprojekt zu mobilen Geräten, die Malaria diagnostizieren können
24 Die dritte Hand | Babor setzt "Pick-by-Vision"-Datenbrille als weltweit erstes Unternehmen ein - FH-Forscher Prof. Voß war an der Entwicklung beteiligt
30 3 Fragen an den neuen AStA-Vorsitzenden Paul Baumann
32 Maronen-Creme-Suppe
34 Schmerzfrei!? | FH-Mitarbeiter Herbert Meurers war schon tot - und kämpfte sich zurück ins Leben
40 Mit "Sally" über die Weichen | Schienenfahrzeugtechnik an der FH Aachen
44 Sicherheit, Freiheit - und Freude über die Bürokratie | Der Syrer Hasan al Jalloud studiert an der FH Aachen und blickt mit Sorge auf das Geschehen in seinem Heimatland
46 Diese Frau erfüllt keine Klischees | Ana Pérez ist die einzige Frau im Aixtreme-Racing-Team
50 Des Rätsels Lösung: das Bewahren | Gespräch mit Prof. Dr. Heinrich Hemme zum Thema Unterhaltungsmathematik
54 "Da weiß man einfach, was Sache ist" | Bauunternehmer Hubert Schlun über sein Leben, seine Werte, seine FH Aachen
58 Prof. Baumann ist neuer LRK-Vorsitzender | FH-Rektor macht sich für die Fachhochschulen in NRW stark
60 Lehrer mit einem Herz für Holz | Alexander Forbrig nimmt am BeLEK-Programm teil und verbindet sein Studium mit einer Ausbildung zum Berufsschullehrer
63 Impressum
Smoothed Finite Element Methods for Nonlinear Solid Mechanics Problems: 2D and 3D Case Studies
(2016)
The Smoothed Finite Element Method (SFEM) is presented as an edge-based and a facebased techniques for 2D and 3D boundary value problems, respectively. SFEMs avoid shortcomings of the standard Finite Element Method (FEM) with lower order elements such as overly stiff behavior, poor stress solution, and locking effects. Based on the idea of averaging spatially the standard strain field of the FEM over so-called smoothing domains SFEM calculates the stiffness matrix for the same number of degrees of freedom (DOFs) as those of the FEM. However, the SFEMs significantly improve accuracy and convergence even for distorted meshes and/or nearly incompressible materials.
Numerical results of the SFEMs for a cardiac tissue membrane (thin plate inflation) and an artery (tension of 3D tube) show clearly their advantageous properties in improving accuracy particularly for the distorted meshes and avoiding shear locking effects.
Unser Zuhause wird zunehmend intelligenter. Smart Homes bieten uns die Steuerung von Haus- oder Unterhaltungstechnik bequem vom Smartphone aus. Junge Familien nutzen die Technologie, um mittels vernetzten Babymonitorsystemen ihren Nachwuchs von überall aus im Blick zu haben. Davon auszugehen, dass solche Systeme mit einem Fokus auf Sicherheit entwickelt wurden, um die sehr persönlichen Daten zu schützen, ist jedoch ein Trugschluss. Die Untersuchung eines handelsüblichen und keineswegs billigen Systems zeigt, dass die Geräte sehr einfach kompromittiert und missbraucht werden können.
Malware auf Smartphones ist ein Problem, dem auch Strafverfolgungsbehörden immer häufiger gegenüberstehen.
Insbesondere Telefone, bei denen potentiell schadhafte Apps zu einem finanziellen Schaden
geführt haben, finden sich auf den Schreibtischen der Polizei wieder. Dabei müssen die Ermittler möglichst
schnell und gezielt erkennen können, ob eine App tatsächlich schadhaft manipuliert wurde, was
manipuliert wurde und mit wem die App kommuniziert. Klassische Malware-Erkennungsverfahren helfen
zwar bei der generellen Erkennung schadhafter Software, sind aber für die polizeiliche Praxis nicht
geeignet. Dieses Paper stellt ein Programm vor, welches gerade die forensischen Fragestellungen berücksichtigt
und so für den Einsatz in der Strafverfolgung in Frage kommt.
Regardless of size or destination, synthetic biology starts with com-parably small information units, which need to be combined and properly arranged in order to achieve a certain goal. This may be the de novo synthesis of individual genes from oligonucleotides, a shuffling of protein domains in order to create novel biocatalysts, the assembly of multiple enzyme encoding genes in metabolic pathway design, or strain development at the production stage. The CoLibry concept has been designed in order to close the gap between recombinant production of individual genes and genome editing.
Application of the optical flow method to velocity determination in hydraulic structure models
(2016)
The aim of this work was to perform a detailed investigation of the use of Selective Laser Melting (SLM) technology to process eutectic silver-copper alloy Ag 28 wt. % Cu (also called AgCu28). The processing occurred with a Realizer SLM 50 desktop machine. The powder analysis (SEM-topography, EDX, particle distribution) was reported as well as the absorption rates for the near-infrared (NIR) spectrum. Microscope imaging showed the surface topography of the manufactured parts. Furthermore, microsections were conducted for the analysis of porosity. The Design of Experiments approach used the response surface method in order to model the statistical relationship between laser power, spot distance and pulse time.
Für das Auftreten extremer Wetterereignisse werden für Kernkraftwerke Eintrittshäufigkeiten für nicht mehr beherrschbare Zustände von unter 10⁻⁴/a gefordert. Dies gilt auch für die Einwirkung von Blitzeinschlägen. Die bisherige Nachweisführung zu Blitz- und Überspannungsschutz eines KKW in Deutschland ist deterministisch. In diesem Bericht werden das Vorgehen für einen entsprechenden Nachweis für leittechnische Einrichtungen der Sicherheitstechnik von KKW, der zur geforderten Zielgröße der Eintrittshäufigkeit führt. Die Ergebnisse werden zusammenfassend bewertet.
In Anbetracht weltweit zunehmend strengerer klimapolitischer
Ziele steigt auch der Druck für Nutzfahrzeughersteller, effizientere und umweltfreundlichere
Technologien zu entwickeln. Den Blick bei der Bewertung dieser
ausschließlich auf die Fahrzeugnutzung zu richten, ist längst nicht mehr zufriedenstellend.
Im Rahmen dieser Analyse wird ein gegenwärtig auf dem Markt erwerblicher
und in deutschen Städten bereits seit Jahren betriebener Hybridbus
energetisch und ökologisch mit einem konventionell angetriebenen, nahezu baugleichen
Modell entlang des Lebensweges bewertet.
Nach Definition von Ziel und Untersuchungsrahmen wird ein Überblick auf bereits
durchgeführte Lebenszyklusanalysen zu Hybridbussen im Stadtverkehr gegeben
und Schlussfolgerungen für die anschließende Analyse abgeleitet. Diese
wird im Rahmen einer energetischen und ökologischen Bewertung beider Produktsysteme
anhand der Parameter "Primärenergieeinsatz" und "CO2äq Emissionen"
praktiziert. Der Fahrzeugrumpf beider Fahrzeuge des gleichen Modells
wird dabei als einheitlich angenommen, sodass bei dem Vergleich der Herstellung
vereinfacht nur die sich unterscheidenden Komponenten des Antriebstranges
berücksichtigt werden. Die Resultate der Wirkungsabschätzung werden als
Differenz des Hybridbusses gegenüber dem Referenzfahrzeug über die einzelnen
Lebenszyklusphasen dargestellt. Schließlich werden Prognosen getroffen, ab
welcher Strecke die bei der Herstellung erzeugten höheren CO2äq Emissionen
des Hybridantriebstranges gegenüber dem Referenzmodell ausgeglichen werden.
Mit der vorliegenden, parallel entsprechend in 10 weiteren Verfahren ergangenen Entscheidung behandelte der BGH zum wiederholten Male das Geschäftsmodell der Accessio Wertpapierhandelshaus AG („A AG“, früher: Wertpapierhandelshaus Driver & Bengsch AG). Die klagenden Anleger, zunächst nur akquiriert durch ein Tagesgeldkonto mit besonders attraktiven Zinsen, schlossen im Weiteren mit dieser einen Vermögensverwaltungsvertrag ab. Zur Abwicklung der Wertpapiergeschäfte eröffneten sie über die A AG zugleich ein Depotkonto bei der beklagten Discount-Brokerin. Für dieses erhielt die A AG eine Transaktionsvollmacht. Die Discount-Brokerin schuldete nach den Vertragsdokumenten über die gesetzlichen Aufklärungs- und Erkundigungspflichten bei Auftragsausführung hinaus keine Anlageberatung („execution-only-business“). Durch nach ihrer Behauptung fehlerhafte Anlageberatung der A AG erlitten die Anleger einen Schaden. In dem Rechtsstreit verlangten sie dessen Ersatz von der Discount-Brokerin, da die A AG zwischenzeitlich insolvent wurde.
Retinal Vessel Analysis (RVA) in the context of subarachnoid hemorrhage: A proof of concept study
(2016)
Background
Timely detection of impending delayed cerebral ischemia after subarachnoid hemorrhage (SAH) is essential to improve outcome, but poses a diagnostic challenge. Retinal vessels as an embryological part of the intracranial vasculature are easily accessible for analysis and may hold the key to a new and non-invasive monitoring technique. This investigation aims to determine the feasibility of standardized retinal vessel analysis (RVA) in the context of SAH.
Methods
In a prospective pilot study, we performed RVA in six patients awake and cooperative with SAH in the acute phase (day 2–14) and eight patients at the time of follow-up (mean 4.6±1.7months after SAH), and included 33 age-matched healthy controls. Data was acquired using a manoeuvrable Dynamic Vessel Analyzer (Imedos Systems UG, Jena) for examination of retinal vessel dimension and neurovascular coupling.
Results
Image quality was satisfactory in the majority of cases (93.3%). In the acute phase after SAH, retinal arteries were significantly dilated when compared to the control group (124.2±4.3MU vs 110.9±11.4MU, p<0.01), a difference that persisted to a lesser extent in the later stage of the disease (122.7±17.2MU, p<0.05). Testing for neurovascular coupling showed a trend towards impaired primary vasodilation and secondary vasoconstriction (p = 0.08, p = 0.09 resp.) initially and partial recovery at the time of follow-up, indicating a relative improvement in a time-dependent fashion.
Conclusion
RVA is technically feasible in patients with SAH and can detect fluctuations in vessel diameter and autoregulation even in less severely affected patients. Preliminary data suggests potential for RVA as a new and non-invasive tool for advanced SAH monitoring, but clinical relevance and prognostic value will have to be determined in a larger cohort.
Hintergrund
Die Anwendung und das Verständnis von Statistik sind sehr wichtig für die biomedizinische Forschung und für die klinische Praxis. Dies gilt insbesondere auch zur Abschätzung der Möglichkeiten unterschiedlichster Diagnostik- und Therapieoptionen beim Glaukom. Die scheinbare Komplexität der Statistik, die zum Teil dem „gesunden Menschenverstand“ zu widersprechen scheint, zusammen mit der nur vorsichtigen Akzeptanz der Statistik bei vielen Medizinern können zu bewussten und unbewussten Manipulationen bei der Datendarstellung und -interpretation führen.
Ziel der Arbeit
Ziel ist die verständliche Darstellung einiger typischer Fehler in der medizinisch-statistischen Datenbehandlung.
Material und Methoden
Anhand hypothetischer Beispiele aus der Glaukomdiagnostik erfolgen die Darstellung der Wirkung eines hypotensiven Medikamentes sowie die Beurteilung der Ergebnisse eines diagnostischen Tests. Es werden die typischsten statistischen Einsatzbereiche und Irrtumsquellen ausführlich und verständlich analysiert
Ergebnisse
Mechanismen von Datenmanipulation und falscher Dateninterpretation werden aufgeklärt. Typische Irrtumsquellen bei der statistischen Auswertung und Datendarstellung werden dabei erläutert.
Schlussfolgerungen
Die erläuterten praktischen Beispiele zeigen die Notwendigkeit, die Grundlagen der Statistik zu verstehen und korrekt anwenden zu können. Fehlendes Grundlagenwissen und Halbwissen der medizinischen Statistik können zu folgenschweren Missverständnissen und falschen Entscheidungen in der medizinischen Forschung, aber auch in der klinischen Praxis führen.
Four members of a homologous series of chlorinated poly(vinyl ester) oligomers CCl₃–(CH₂CH (OCO(CH₂)ₘCH₃))ₙ–Cl with degrees of polymerization of 10 and 20 were prepared by telomerisation using carbon tetrachloride. The number of side chain carbon atoms ranges from 2 (poly(vinyl acetate) to 18 (poly(vinyl stearate)). The effect of the n-alkyl side chain length and of the degree of polymerization on the thermal stability and crystallization behaviour of the synthesized compounds was investigated.
All oligomers degrade in two major steps by first losing HCl and side chains with subsequent breakdown of the backbone. The members with short side chains, up to poly(vinyl octanoate), are amorphous and show internal plasticization, whereas those with high number of side chain carbon atoms are semi-crystalline due to side-chain crystallization. A better packing for poly(vinyl stearate) is also noticeable. The glass transition and melting temperatures as well as the onset temperature of decomposition are influenced to a larger extent by the side chain length than by the degree of polymerization. Thermal stability is improved if both the size and number of side chains increase, but only a long side chain causes a significant increase of the resistance to degradation. This results in a stabilization of PVAc so that oligomers from poly(vinyl octanoate) on are stable under atmospheric conditions. Thus, the way to design stable, chlorinated PVEs oligomers is to use a long n-alkyl side chain.
Sensitivity of turbulent Schmidt number and turbulence model to simulations of jets in crossflow
(2016)
Environmental discharges have been traditionally designed by means of cost-intensive and time-consuming experimental studies. Some extensively validated models based on an integral approach have been often employed for water quality problems, as recommended by USEPA (i.e.: CORMIX). In this study, FLOW-3D is employed for a full 3D RANS modelling of two turbulent jet-to-crossflow cases, including free surface jet impingement. Results are compared to both physical modelling and CORMIX to better assess model performance. Turbulence measurements have been collected for a better understanding of turbulent diffusion's parameter sensitivity. Although both studied models are generally able to reproduce jet trajectory, jet separation downstream of the impingement has been reproduced only by RANS modelling. Additionally, concentrations are better reproduced by FLOW-3D when the proper turbulent Schmidt number is used. This study provides a recommendation on the selection of the turbulence model and the turbulent Schmidt number for future outfall structures design studies.
The performance and biomass yield of the perennial energy plant Sida hermaphrodita (hereafter referred to as Sida) as a feedstock for biogas and solid fuel was evaluated throughout one entire growing period at agricultural field conditions. A Sida plant development code was established to allow comparison of the plant growth stages and biomass composition. Four scenarios were evaluated to determine the use of Sida biomass with regard to plant development and harvest time: (i) one harvest for solid fuel only; (ii) one harvest for biogas production only; (iii) one harvest for biogas production, followed by a harvest of the regrown biomass for solid fuel; and (iv) two consecutive harvests for biogas production. To determine Sida's value as a feedstock for combustion, we assessed the caloric value, the ash quality, and melting point with regard to DIN EN ISO norms. The results showed highest total dry biomass yields of max. 25 t ha⁻¹, whereas the highest dry matter of 70% to 80% was obtained at the end of the growing period. Scenario (i) clearly indicated the highest energy recovery, accounting for 439 288 MJ ha⁻¹; the energy recovery of the four scenarios from highest to lowest followed this order: (i) ≫ (iii) ≫ (iv) > (ii). Analysis of the Sida ashes showed a high melting point of >1500 °C, associated with a net calorific value of 16.5–17.2 MJ kg⁻¹. All prerequisites for DIN EN ISO norms were achieved, indicating Sida's advantage as a solid energy carrier without any post-treatment after harvesting. Cell wall analysis of the stems showed a constant lignin content after sampling week 16 (July), whereas cellulose had already reached a plateau in sampling week 4 (April). The results highlight Sida as a promising woody, perennial plant, providing biomass for flexible and multipurpose energy applications.
The Android operating system powers the majority of the world’s mobile devices and has been becoming increasingly important in day-to-day digital forensics. Therefore, technicians and analysts are in need of reliable methods for extracting and analyzing memory images from live Android systems. This paper takes different existing, extraction methods and derives a universal, reproducible, reliably documented method for both extraction and analysis. In addition the VOLIX II front-end for the Volatility Framework is extended with additional functionality to make the analysis of Android memory images easier for technically non-adept users.