@article{MonakhovaDiehl2021, author = {Monakhova, Yulia and Diehl, Bernd W. K.}, title = {Simplification of NMR Workflows by Standardization Using 2H Integral of Deuterated Solvent as Applied to Aloe vera Preparations}, series = {Applied Magnetic Resonance}, volume = {52}, journal = {Applied Magnetic Resonance}, number = {11}, publisher = {Springer}, address = {Cham}, issn = {1613-7507}, doi = {10.1007/s00723-021-01393-4}, pages = {1591 -- 1600}, year = {2021}, abstract = {In this study, a recently proposed NMR standardization approach by 2H integral of deuterated solvent for quantitative multicomponent analysis of complex mixtures is presented. As a proof of principle, the existing NMR routine for the analysis of Aloe vera products was modified. Instead of using absolute integrals of targeted compounds and internal standard (nicotinamide) from 1H-NMR spectra, quantification was performed based on the ratio of a particular 1H-NMR compound integral and 2H-NMR signal of deuterated solvent D2O. Validation characteristics (linearity, repeatability, accuracy) were evaluated and the results showed that the method has the same precision as internal standardization in case of multicomponent screening. Moreover, a dehydration process by freeze drying is not necessary for the new routine. Now, our NMR profiling of A. vera products needs only limited sample preparation and data processing. The new standardization methodology provides an appealing alternative for multicomponent NMR screening. In general, this novel approach, using standardization by 2H integral, benefits from reduced sample preparation steps and uncertainties, and is recommended in different application areas (purity determination, forensics, pharmaceutical analysis, etc.).}, language = {en} } @article{BurmistrovaSobolevaMonakhova2021, author = {Burmistrova, Natalia A. and Soboleva, Polina M. and Monakhova, Yulia}, title = {Is infrared spectroscopy combined with multivariate analysis a promising tool for heparin authentication?}, series = {Journal of Pharmaceutical and Biomedical Analysis}, volume = {194}, journal = {Journal of Pharmaceutical and Biomedical Analysis}, number = {Article number: 113811}, publisher = {Elsevier}, address = {Amsterdam}, isbn = {0731-7085}, doi = {10.1016/j.jpba.2020.113811}, year = {2021}, abstract = {The investigation of the possibility to determine various characteristics of powder heparin (n = 115) was carried out with infrared spectroscopy. The evaluation of heparin samples included several parameters such as purity grade, distributing company, animal source as well as heparin species (i.e. Na-heparin, Ca-heparin, and heparinoids). Multivariate analysis using principal component analysis (PCA), soft independent modelling of class analogy (SIMCA), and partial least squares - discriminant analysis (PLS-DA) were applied for the modelling of spectral data. Different pre-processing methods were applied to IR spectral data; multiplicative scatter correction (MSC) was chosen as the most relevant. Obtained results were confirmed by nuclear magnetic resonance (NMR) spectroscopy. Good predictive ability of this approach demonstrates the potential of IR spectroscopy and chemometrics for screening of heparin quality. This approach, however, is designed as a screening tool and is not considered as a replacement for either of the methods required by USP and FDA.}, language = {en} } @article{MonakhovaDiehl2021, author = {Monakhova, Yulia and Diehl, Bernd W.K.}, title = {Novel approach of qNMR workflow by standardization using 2H integral: Application to any intrinsic calibration standard}, series = {Talanta}, volume = {222}, journal = {Talanta}, number = {Article number: 121504}, publisher = {Elsevier}, isbn = {0039-9140}, doi = {10.1016/j.talanta.2020.121504}, year = {2021}, abstract = {Quantitative nuclear magnetic resonance (qNMR) is routinely performed by the internal or external standardization. The manuscript describes a simple alternative to these common workflows by using NMR signal of another active nuclei of calibration compound. For example, for any arbitrary compound quantification by NMR can be based on the use of an indirect concentration referencing that relies on a solvent having both 1H and 2H signals. To perform high-quality quantification, the deuteration level of the utilized deuterated solvent has to be estimated. In this contribution the new method was applied to the determination of deuteration levels in different deuterated solvents (MeOD, ACN, CDCl3, acetone, benzene, DMSO-d6). Isopropanol-d6, which contains a defined number of deuterons and protons, was used for standardization. Validation characteristics (precision, accuracy, robustness) were calculated and the results showed that the method can be used in routine practice. Uncertainty budget was also evaluated. In general, this novel approach, using standardization by 2H integral, benefits from reduced sample preparation steps and uncertainties, and can be applied in different application areas (purity determination, forensics, pharmaceutical analysis, etc.).}, language = {en} } @article{MonakhovaSobolevaFedotovaetal.2022, author = {Monakhova, Yulia and Soboleva, Polina M. and Fedotova, Elena S. and Musina, Kristina T. and Burmistrova, Natalia A.}, title = {Quantum chemical calculations of IR spectra of heparin disaccharide subunits}, series = {Computational and Theoretical Chemistry}, volume = {1217}, journal = {Computational and Theoretical Chemistry}, number = {Article number: 113891}, publisher = {Elsevier}, address = {New York, NY}, isbn = {2210-271X}, doi = {10.1016/j.comptc.2022.113891}, year = {2022}, abstract = {Heparin is a natural polysaccharide, which plays essential role in many biological processes. Alterations in building blocks can modify biological roles of commercial heparin products, due to significant changes in the conformation of the polymer chain. The variability structure of heparin leads to difficulty in quality control using different analytical methods, including infrared (IR) spectroscopy. In this paper molecular modelling of heparin disaccharide subunits was performed using quantum chemistry. The structural and spectral parameters of these disaccharides have been calculated using RHF/6-311G. In addition, over-sulphated chondroitin sulphate disaccharide was studied as one of the most widespread contaminants of heparin. Calculated IR spectra were analyzed with respect to specific structure parameters. IR spectroscopic fingerprint was found to be sensitive to substitution pattern of disaccharide subunits. Vibrational assignments of calculated spectra were correlated with experimental IR spectral bands of native heparin. Chemometrics was used to perform multivariate analysis of simulated spectral data.}, language = {en} } @article{BurgerLindnerRumpfetal.2022, author = {Burger, Ren{\´e} and Lindner, Simon and Rumpf, Jessica and Do, Xuan Tung and Diehl, Bernd W.K. and Rehahn, Matthias and Monakhova, Yulia and Schulze, Margit}, title = {Benchtop versus high field NMR: Comparable performance found for the molecular weight determination of lignin}, series = {Journal of Pharmaceutical and Biomedical Analysis}, volume = {212}, journal = {Journal of Pharmaceutical and Biomedical Analysis}, number = {Article number: 114649}, publisher = {Elsevier}, address = {New York, NY}, isbn = {0731-7085}, doi = {10.1016/j.jpba.2022.114649}, year = {2022}, abstract = {Lignin is a promising renewable biopolymer being investigated worldwide as an environmentally benign substitute of fossil-based aromatic compounds, e.g. for the use as an excipient with antioxidant and antimicrobial properties in drug delivery or even as active compound. For its successful implementation into process streams, a quick, easy, and reliable method is needed for its molecular weight determination. Here we present a method using 1H spectra of benchtop as well as conventional NMR systems in combination with multivariate data analysis, to determine lignin's molecular weight (Mw and Mn) and polydispersity index (PDI). A set of 36 organosolv lignin samples (from Miscanthus x giganteus, Paulownia tomentosa and Silphium perfoliatum) was used for the calibration and cross validation, and 17 samples were used as external validation set. Validation errors between 5.6\% and 12.9\% were achieved for all parameters on all NMR devices (43, 60, 500 and 600 MHz). Surprisingly, no significant difference in the performance of the benchtop and high-field devices was found. This facilitates the application of this method for determining lignin's molecular weight in an industrial environment because of the low maintenance expenditure, small footprint, ruggedness, and low cost of permanent magnet benchtop NMR systems.}, language = {en} } @article{MonakhovaDiehl2022, author = {Monakhova, Yulia and Diehl, Bernd W.K.}, title = {Multinuclear NMR screening of pharmaceuticals using standardization by 2H integral of a deuterated solvent}, series = {Journal of Pharmaceutical and Biomedical Analysis}, volume = {209}, journal = {Journal of Pharmaceutical and Biomedical Analysis}, number = {Article number: 114530}, publisher = {Elsevier}, isbn = {0731-7085}, doi = {10.1016/j.jpba.2021.114530}, year = {2022}, abstract = {NMR standardization approach that uses the 2H integral of deuterated solvent for quantitative multinuclear analysis of pharmaceuticals is described. As a proof of principle, the existing NMR procedure for the analysis of heparin products according to US Pharmacopeia monograph is extended to the determination of Na+ and Cl- content in this matrix. Quantification is performed based on the ratio of a 23Na (35Cl) NMR integral and 2H NMR signal of deuterated solvent, D2O, acquired using the specific spectrometer hardware. As an alternative, the possibility of 133Cs standardization using the addition of Cs2CO3 stock solution is shown. Validation characteristics (linearity, repeatability, sensitivity) are evaluated. A holistic NMR profiling of heparin products can now also be used for the quantitative determination of inorganic compounds in a single analytical run using a single sample. In general, the new standardization methodology provides an appealing alternative for the NMR screening of inorganic and organic components in pharmaceutical products.}, language = {en} } @article{LindnerBurgerRutledgeetal.2022, author = {Lindner, Simon and Burger, Ren{\´e} and Rutledge, Douglas N. and Do, Xuan Tung and Rumpf, Jessica and Diehl, Bernd W. K. and Schulze, Margit and Monakhova, Yulia}, title = {Is the calibration transfer of multivariate calibration models between high- and low-field NMR instruments possible? A case study of lignin molecular weight}, series = {Analytical chemistry}, volume = {94}, journal = {Analytical chemistry}, number = {9}, publisher = {ACS Publications}, address = {Washington, DC}, isbn = {1520-6882}, doi = {10.1021/acs.analchem.1c05125}, pages = {3997 -- 4004}, year = {2022}, abstract = {Although several successful applications of benchtop nuclear magnetic resonance (NMR) spectroscopy in quantitative mixture analysis exist, the possibility of calibration transfer remains mostly unexplored, especially between high- and low-field NMR. This study investigates for the first time the calibration transfer of partial least squares regressions [weight average molecular weight (Mw) of lignin] between high-field (600 MHz) NMR and benchtop NMR devices (43 and 60 MHz). For the transfer, piecewise direct standardization, calibration transfer based on canonical correlation analysis, and transfer via the extreme learning machine auto-encoder method are employed. Despite the immense resolution difference between high-field and low-field NMR instruments, the results demonstrate that the calibration transfer from high- to low-field is feasible in the case of a physical property, namely, the molecular weight, achieving validation errors close to the original calibration (down to only 1.2 times higher root mean square errors). These results introduce new perspectives for applications of benchtop NMR, in which existing calibrations from expensive high-field instruments can be transferred to cheaper benchtop instruments to economize.}, language = {en} } @inproceedings{BrockmannSultanowCzarnecki2018, author = {Brockmann, Carsten and Sultanow, Eldar and Czarnecki, Christian}, title = {Enterprise architectures between agility and traditional methodologies}, series = {GI Edition Proceedings Band 285 Workshops der INFORMATIK 2018}, booktitle = {GI Edition Proceedings Band 285 Workshops der INFORMATIK 2018}, editor = {Czarnecki, Christian and Brockmann, Carsten and Sultanow, Eldar and Koschmider, Agnes and Selzer, Annika and Gesellschaft f{\"u}r Informatik e. V.,}, publisher = {K{\"o}llen}, address = {Bonn}, isbn = {9783885796794}, issn = {1617-5468}, pages = {1 Seite}, year = {2018}, abstract = {For this year's workshop on Enterpirse Architecture in Research and Practice we have received eight submissions from which four have passed the rigorous peer-review. The acceptance quote of 50\% assures that only advancements in the field are included in our workshop.}, language = {en} } @inproceedings{BensbergAuthCzarneckietal.2018, author = {Bensberg, Frank and Auth, Gunnar and Czarnecki, Christian and W{\"o}rndle, Christopher}, title = {Transforming literature-intensive research processes through text analytics - design, implementation and lessons learned}, editor = {Kemal İlter, H.}, doi = {10.6084/m9.figshare.7582073.v1}, pages = {9 Seiten}, year = {2018}, abstract = {The continuing growth of scientific publications raises the question how research processes can be digitalized and thus realized more productively. Especially in information technology fields, research practice is characterized by a rapidly growing volume of publications. For the search process various information systems exist. However, the analysis of the published content is still a highly manual task. Therefore, we propose a text analytics system that allows a fully digitalized analysis of literature sources. We have realized a prototype by using EBSCO Discovery Service in combination with IBM Watson Explorer and demonstrated the results in real-life research projects. Potential addressees are research institutions, consulting firms, and decision-makers in politics and business practice.}, language = {en} } @inproceedings{BrockmannSultanowCzarnecki2019, author = {Brockmann, Carsten and Sultanow, Eldar and Czarnecki, Christian}, title = {Is enterprise architecture still relevant in the digital age?}, series = {GI Edition Proceedings Band 295 INFORMATIK 2019, Workshop-Beitr{\"a}ge}, booktitle = {GI Edition Proceedings Band 295 INFORMATIK 2019, Workshop-Beitr{\"a}ge}, editor = {Draude, Claude and Lange, Martin and Sick, Bernhard}, publisher = {K{\"o}llen}, address = {Bonn}, isbn = {9783885796893}, issn = {1617-5468}, doi = {10.18420/inf2019_ws01}, pages = {21 -- 21}, year = {2019}, language = {en} } @inproceedings{AuthCzarneckiBensberg2019, author = {Auth, Gunnar and Czarnecki, Christian and Bensberg, Frank}, title = {Impact of robotic process automation on enterprise architectures}, series = {GI Edition Proceedings Band 295 INFORMATIK 2019, Workshop-Beitr{\"a}ge}, booktitle = {GI Edition Proceedings Band 295 INFORMATIK 2019, Workshop-Beitr{\"a}ge}, editor = {Draude, Claude and Lange, Martin and Sick, Bernhard and Gesellschaft f{\"u}r Informatik e.V. (GI),}, publisher = {K{\"o}llen}, address = {Bonn}, isbn = {9783885796893}, issn = {1617-5468}, doi = {10.18420/inf2019_ws05}, pages = {59 -- 65}, year = {2019}, abstract = {The initial idea of Robotic Process Automation (RPA) is the automation of business processes through the presentation layer of existing application systems. For this simple emulation of user input and output by software robots, no changes of the systems and architecture is required. However, considering strategic aspects of aligning business and technology on an enterprise level as well as the growing capabilities of RPA driven by artificial intelligence, interrelations between RPA and Enterprise Architecture (EA) become visible and pose new questions. In this paper we discuss the relationship between RPA and EA in terms of perspectives and implications. As workin- progress we focus on identifying new questions and research opportunities related to RPA and EA.}, language = {en} } @inproceedings{RitschelStenzelCzarneckietal.2021, author = {Ritschel, Konstantin and Stenzel, Adina and Czarnecki, Christian and Hong, Chin-Gi}, title = {Realizing robotic process automation potentials: an architectural perspective on a real-life implementation case}, series = {GI Edition Proceedings Band 314 "INFORMATIK 2021" Computer Science \& Sustainability}, booktitle = {GI Edition Proceedings Band 314 "INFORMATIK 2021" Computer Science \& Sustainability}, editor = {Gesellschaft f{\"u}r Informatik e.V. (GI),}, publisher = {K{\"o}llen}, address = {Bonn}, isbn = {9783885797081}, issn = {1617-5468}, doi = {10.18420/informatik2021-108}, pages = {1303 -- 1311}, year = {2021}, abstract = {The initial idea of Robotic Process Automation (RPA) is the automation of business processes through a simple emulation of user input and output by software robots. Hence, it can be assumed that no changes of the used software systems and existing Enterprise Architecture (EA) is required. In this short, practical paper we discuss this assumption based on a real-life implementation project. We show that a successful RPA implementation might require architectural work during analysis, implementation, and migration. As practical paper we focus on exemplary lessons-learned and new questions related to RPA and EA.}, language = {en} } @inproceedings{AmirBauckhageChircuetal.2022, author = {Amir, Malik and Bauckhage, Christian and Chircu, Alina and Czarnecki, Christian and Knopf, Christian and Piatkowski, Nico and Sultanow, Eldar}, title = {What can we expect from quantum (digital) twins?}, series = {Wirtschaftsinformatik 2022 Proceedings}, booktitle = {Wirtschaftsinformatik 2022 Proceedings}, publisher = {AIS Electronic Library (AISeL)}, pages = {1 -- 14}, year = {2022}, abstract = {Digital twins enable the modeling and simulation of real-world entities (objects, processes or systems), resulting in improvements in the associated value chains. The emerging field of quantum computing holds tremendous promise forevolving this virtualization towards Quantum (Digital) Twins (QDT) and ultimately Quantum Twins (QT). The quantum (digital) twin concept is not a contradiction in terms - but instead describes a hybrid approach that can be implemented using the technologies available today by combining classicalcomputing and digital twin concepts with quantum processing. This paperpresents the status quo of research and practice on quantum (digital) twins. It alsodiscuses their potential to create competitive advantage through real-timesimulation of highly complex, interconnected entities that helps companies better address changes in their environment and differentiate their products andservices.}, language = {en} } @inproceedings{BlaneckBornheimGriegeretal.2022, author = {Blaneck, Patrick Gustav and Bornheim, Tobias and Grieger, Niklas and Bialonski, Stephan}, title = {Automatic readability assessment of german sentences with transformer ensembles}, series = {Proceedings of the GermEval 2022 Workshop on Text Complexity Assessment of German Text}, booktitle = {Proceedings of the GermEval 2022 Workshop on Text Complexity Assessment of German Text}, publisher = {Association for Computational Linguistics}, address = {Potsdam}, doi = {10.48550/arXiv.2209.04299}, pages = {57 -- 62}, year = {2022}, abstract = {Reliable methods for automatic readability assessment have the potential to impact a variety of fields, ranging from machine translation to self-informed learning. Recently, large language models for the German language (such as GBERT and GPT-2-Wechsel) have become available, allowing to develop Deep Learning based approaches that promise to further improve automatic readability assessment. In this contribution, we studied the ability of ensembles of fine-tuned GBERT and GPT-2-Wechsel models to reliably predict the readability of German sentences. We combined these models with linguistic features and investigated the dependence of prediction performance on ensemble size and composition. Mixed ensembles of GBERT and GPT-2-Wechsel performed better than ensembles of the same size consisting of only GBERT or GPT-2-Wechsel models. Our models were evaluated in the GermEval 2022 Shared Task on Text Complexity Assessment on data of German sentences. On out-of-sample data, our best ensemble achieved a root mean squared error of 0:435.}, language = {en} } @inproceedings{SteuerDankertSharmaBlecketal.2017, author = {Steuer-Dankert, Linda and Sharma, Mamta Rameshwarlal and Bleck, Wolfgang and Leicht-Scholten, Carmen}, title = {Innovation through Diversity - Development of a Diversity and Innovation management concept}, series = {International Conference on Innovation and Management : IAM23017S : Date: July 4-7, 2017, Osaka, Japan}, booktitle = {International Conference on Innovation and Management : IAM23017S : Date: July 4-7, 2017, Osaka, Japan}, editor = {Farn, C. K.}, publisher = {Kuang Hui Chiu}, address = {Osaka}, issn = {2218-6387}, pages = {Panel C}, year = {2017}, abstract = {Acknowledging that a diverse workforce could be a potential source of innovation, the current research deals with the fine details of why diversity management is central to achieving innovation in heterogeneous research groups and how this could be effectively realized in an organization. The types of heterogeneities addressed mainly include gender, qualification, academic discipline and intercultural perspectives. The type of organization being dealt with in this work is a complex association of research institutes at a technical university in Germany (RWTH Aachen University), namely a 'Cluster of Excellence', whereby several institutes of the university work collaboratively in different sub-projects. The 'Cluster of Excellence' is a part of the 'Excellence Initiative' of the German federal and state governments German Research Foundation (DFG) and German Council of Science and Humanities, with the ultimate aim of promoting cutting-edge research. To support interdisciplinary collaboration and thus the performance of the cluster, the development of a diversity and innovation management concept is presently in the conceptual phase and will be described in the frame of this paper. The 3-S-Diversity Model, composed of the three elements: skills, structure and strategy, serves as a basis for the development of the concept. The proposed concept consists of six phases; the first two phases lay the ground work by developing an understanding of the status quo on the forms of diversity in the Cluster of Excellence, the type of organizational structure of the member institutes and the varieties of specialist work cultures of the same. The third and the fourth phases build up on this foundation by means of qualitative and quantitative studies. While the third phase deals with the sensitization of the management level to the close connection between diversity and innovation; the need to manage them thereafter and find tailor-made methods of doing so, the fourth phase shall mainly focus on the mindset of the employees in this regard. The fifth phase shall consolidate the learnings and the ideas developed in the course of the first four phases into an implementable strategy. The ultimate phase shall be the implementation of this concept in the Cluster. The first three phases have been accomplished successfully and the preliminary results are already available.}, language = {en} } @article{Maurischat2022, author = {Maurischat, Andreas}, title = {Algebraic independence of the Carlitz period and its hyperderivatives}, series = {Journal of Number Theory}, volume = {240}, journal = {Journal of Number Theory}, publisher = {Elsevier}, address = {Orlando, Fla.}, issn = {0022-314X}, doi = {10.1016/j.jnt.2022.01.006}, pages = {145 -- 162}, year = {2022}, language = {en} } @article{KotliarOrtnerConradietal.2022, author = {Kotliar, Konstantin and Ortner, Marion and Conradi, Anna and Hacker, Patricia and Hauser, Christine and G{\"u}nthner, Roman and Moser, Michaela and Muggenthaler, Claudia and Diehl-Schmid, Janine and Priller, Josef and Schmaderer, Christoph and Grimmer, Timo}, title = {Altered retinal cerebral vessel oscillation frequencies in Alzheimer's disease compatible with impaired amyloid clearance}, series = {Neurobiology of Aging}, volume = {120}, journal = {Neurobiology of Aging}, publisher = {Elsevier}, address = {Amsterdam}, issn = {0197-4580}, doi = {10.1016/j.neurobiolaging.2022.08.012}, pages = {117 -- 127}, year = {2022}, abstract = {Retinal vessels are similar to cerebral vessels in their structure and function. Moderately low oscillation frequencies of around 0.1 Hz have been reported as the driving force for paravascular drainage in gray matter in mice and are known as the frequencies of lymphatic vessels in humans. We aimed to elucidate whether retinal vessel oscillations are altered in Alzheimer's disease (AD) at the stage of dementia or mild cognitive impairment (MCI). Seventeen patients with mild-to-moderate dementia due to AD (ADD); 23 patients with MCI due to AD, and 18 cognitively healthy controls (HC) were examined using Dynamic Retinal Vessel Analyzer. Oscillatory temporal changes of retinal vessel diameters were evaluated using mathematical signal analysis. Especially at moderately low frequencies around 0.1 Hz, arterial oscillations in ADD and MCI significantly prevailed over HC oscillations and correlated with disease severity. The pronounced retinal arterial vasomotion at moderately low frequencies in the ADD and MCI groups would be compatible with the view of a compensatory upregulation of paravascular drainage in AD and strengthen the amyloid clearance hypothesis.}, language = {en} } @article{ColomboDriraFrotscheretal.2022, author = {Colombo, Daniele and Drira, Slah and Frotscher, Ralf and Staat, Manfred}, title = {An element-based formulation for ES-FEM and FS-FEM models for implementation in standard solid mechanics finite element codes for 2D and 3D static analysis}, series = {International Journal for Numerical Methods in Engineering}, volume = {124}, journal = {International Journal for Numerical Methods in Engineering}, number = {2}, publisher = {Wiley}, address = {Chichester}, issn = {1097-0207}, doi = {10.1002/nme.7126}, pages = {402 -- 433}, year = {2022}, abstract = {Edge-based and face-based smoothed finite element methods (ES-FEM and FS-FEM, respectively) are modified versions of the finite element method allowing to achieve more accurate results and to reduce sensitivity to mesh distortion, at least for linear elements. These properties make the two methods very attractive. However, their implementation in a standard finite element code is nontrivial because it requires heavy and extensive modifications to the code architecture. In this article, we present an element-based formulation of ES-FEM and FS-FEM methods allowing to implement the two methods in a standard finite element code with no modifications to its architecture. Moreover, the element-based formulation permits to easily manage any type of element, especially in 3D models where, to the best of the authors' knowledge, only tetrahedral elements are used in FS-FEM applications found in the literature. Shape functions for non-simplex 3D elements are proposed in order to apply FS-FEM to any standard finite element.}, language = {en} } @article{BhattaraiMayStaatetal.2022, author = {Bhattarai, Aroj and May, Charlotte Anabell and Staat, Manfred and Kowalczyk, Wojciech and Tran, Thanh Ngoc}, title = {Layer-specific damage modeling of porcine large intestine under biaxial tension}, series = {Bioengineering}, volume = {9}, journal = {Bioengineering}, number = {10, Early Access}, publisher = {MDPI}, address = {Basel}, issn = {2306-5354}, doi = {10.3390/bioengineering9100528}, pages = {1 -- 17}, year = {2022}, abstract = {The mechanical behavior of the large intestine beyond the ultimate stress has never been investigated. Stretching beyond the ultimate stress may drastically impair the tissue microstructure, which consequently weakens its healthy state functions of absorption, temporary storage, and transportation for defecation. Due to closely similar microstructure and function with humans, biaxial tensile experiments on the porcine large intestine have been performed in this study. In this paper, we report hyperelastic characterization of the large intestine based on experiments in 102 specimens. We also report the theoretical analysis of the experimental results, including an exponential damage evolution function. The fracture energies and the threshold stresses are set as damage material parameters for the longitudinal muscular, the circumferential muscular and the submucosal collagenous layers. A biaxial tensile simulation of a linear brick element has been performed to validate the applicability of the estimated material parameters. The model successfully simulates the biomechanical response of the large intestine under physiological and non-physiological loads.}, language = {en} } @incollection{StriebingMuellerSchraudneretal.2022, author = {Striebing, Clemens and M{\"u}ller, J{\"o}rg and Schraudner, Martina and Gewinner, Irina Valerie and Guerrero Morales, Patricia and Hochfeld, Katharina and Hoffman, Shekinah and Kmec, Julie A. and Nguyen, Huu Minh and Schneider, Jannick and Sheridan, Jennifer and Steuer-Dankert, Linda and Trimble O'Connor, Lindsey and Vandevelde-Rougale, Agn{\`e}s}, title = {Promoting diversity and combatting discrimination in research organizations: a practitioner's guide}, series = {Diversity and discrimination in research organizations}, booktitle = {Diversity and discrimination in research organizations}, publisher = {Emerald Publishing Limited}, address = {Bingley}, isbn = {978-1-80117-959-1 (Print)}, doi = {10.1108/978-1-80117-956-020221012}, pages = {421 -- 442}, year = {2022}, abstract = {The essay is addressed to practitioners in research management and from academic leadership. It describes which measures can contribute to creating an inclusive climate for research teams and preventing and effectively dealing with discrimination. The practical recommendations consider the policy and organizational levels, as well as the individual perspective of research managers. Following a series of basic recommendations, six lessons learned are formulated, derived from the contributions to the edited collection on "Diversity and Discrimination in Research Organizations."}, language = {en} }