@article{SchreiberBarkschatKraftetal.2015, author = {Schreiber, Marc and Barkschat, Kai and Kraft, Bodo and Z{\"u}ndorf, Albert}, title = {Quick Pad Tagger : An Efficient Graphical User Interface for Building Annotated Corpora with Multiple Annotation Layers}, series = {Computer Science \& Information Technology (CS \& IT)}, volume = {5}, journal = {Computer Science \& Information Technology (CS \& IT)}, number = {4}, publisher = {Academy \& Industry Research Collaboration Center (AIRCC)}, isbn = {978-1-921987-32-8}, issn = {2231 - 5403}, doi = {10.5121/csit.2015.50413}, pages = {131 -- 143}, year = {2015}, language = {en} } @inproceedings{SchreiberKraftZuendorf2016, author = {Schreiber, Marc and Kraft, Bodo and Z{\"u}ndorf, Albert}, title = {Cost-efficient quality assurance of natural language processing tools through continuous monitoring with continuous integration}, series = {3rd International Workshop on Software Engineering Research and Industrial Practice}, booktitle = {3rd International Workshop on Software Engineering Research and Industrial Practice}, doi = {10.1145/2897022.2897029}, pages = {46 -- 52}, year = {2016}, language = {en} } @inproceedings{SchreiberKraftZuendorf2017, author = {Schreiber, Marc and Kraft, Bodo and Z{\"u}ndorf, Albert}, title = {Metrics Driven Research Collaboration: Focusing on Common Project Goals Continuously}, series = {39th International Conference on Software Engineering, May 20-28, 2017 - Buenos Aires, Argentina}, booktitle = {39th International Conference on Software Engineering, May 20-28, 2017 - Buenos Aires, Argentina}, pages = {8 Seiten}, year = {2017}, abstract = {Research collaborations provide opportunities for both practitioners and researchers: practitioners need solutions for difficult business challenges and researchers are looking for hard problems to solve and publish. Nevertheless, research collaborations carry the risk that practitioners focus on quick solutions too much and that researchers tackle theoretical problems, resulting in products which do not fulfill the project requirements. In this paper we introduce an approach extending the ideas of agile and lean software development. It helps practitioners and researchers keep track of their common research collaboration goal: a scientifically enriched software product which fulfills the needs of the practitioner's business model. This approach gives first-class status to application-oriented metrics that measure progress and success of a research collaboration continuously. Those metrics are derived from the collaboration requirements and help to focus on a commonly defined goal. An appropriate tool set evaluates and visualizes those metrics with minimal effort, and all participants will be pushed to focus on their tasks with appropriate effort. Thus project status, challenges and progress are transparent to all research collaboration members at any time.}, language = {en} } @inproceedings{SchreiberKraftZuendorf2017, author = {Schreiber, Marc and Kraft, Bodo and Z{\"u}ndorf, Albert}, title = {Metrics driven research collaboration: focusing on common project goals continuously}, series = {Proceedings : 2017 IEEE/ACM 4th International Workshop on Software Engineering Research and Industrial Practice : SER\&IP 2017 : 21 May 2017 Buenos Aires, Argentina}, booktitle = {Proceedings : 2017 IEEE/ACM 4th International Workshop on Software Engineering Research and Industrial Practice : SER\&IP 2017 : 21 May 2017 Buenos Aires, Argentina}, editor = {Bilof, Randall}, publisher = {IEEE Press}, address = {Piscataway, NJ}, isbn = {978-1-5386-2797-6}, doi = {10.1109/SER-IP.2017..6}, pages = {41 -- 47}, year = {2017}, language = {en} } @misc{NobisrathZuendorfGeorgeetal.2017, author = {Nobisrath, Ulrich and Z{\"u}ndorf, Albert and George, Tobias and Ruben, Jubeh and Kraft, Bodo}, title = {Software Stories Guide}, pages = {21}, year = {2017}, abstract = {Software Stories are a simple graphical notation for requirements analysis and design in agile software projects. Software Stories are based on example scenarios. Example scenarios facilitate the communication between lay people or domain experts and software experts.}, language = {en} } @inproceedings{SchreiberKraftZuendorf2018, author = {Schreiber, Marc and Kraft, Bodo and Z{\"u}ndorf, Albert}, title = {NLP Lean Programming Framework: Developing NLP Applications More Effectively}, series = {Proceedings of NAACL-HLT 2018: Demonstrations, New Orleans, Louisiana, June 2 - 4, 2018}, booktitle = {Proceedings of NAACL-HLT 2018: Demonstrations, New Orleans, Louisiana, June 2 - 4, 2018}, doi = {10.18653/v1/N18-5001 }, pages = {5 Seiten}, year = {2018}, abstract = {This paper presents NLP Lean Programming framework (NLPf), a new framework for creating custom natural language processing (NLP) models and pipelines by utilizing common software development build systems. This approach allows developers to train and integrate domain-specific NLP pipelines into their applications seamlessly. Additionally, NLPf provides an annotation tool which improves the annotation process significantly by providing a well-designed GUI and sophisticated way of using input devices. Due to NLPf's properties developers and domain experts are able to build domain-specific NLP applications more efficiently. NLPf is Opensource software and available at https:// gitlab.com/schrieveslaach/NLPf.}, language = {en} } @inproceedings{SchmidtsKraftSchreiberetal.2018, author = {Schmidts, Oliver and Kraft, Bodo and Schreiber, Marc and Z{\"u}ndorf, Albert}, title = {Continuously evaluated research projects in collaborative decoupled environments}, series = {2018 ACM/IEEE 5th International Workshop on Software Engineering Research and Industrial PracticePractice, May 29, 2018, Gothenburg, Sweden : SER\&IP' 18}, booktitle = {2018 ACM/IEEE 5th International Workshop on Software Engineering Research and Industrial PracticePractice, May 29, 2018, Gothenburg, Sweden : SER\&IP' 18}, publisher = {ACM}, address = {New York, NY}, pages = {1 -- 9}, year = {2018}, abstract = {Often, research results from collaboration projects are not transferred into productive environments even though approaches are proven to work in demonstration prototypes. These demonstration prototypes are usually too fragile and error-prone to be transferred easily into productive environments. A lot of additional work is required. Inspired by the idea of an incremental delivery process, we introduce an architecture pattern, which combines the approach of Metrics Driven Research Collaboration with microservices for the ease of integration. It enables keeping track of project goals over the course of the collaboration while every party may focus on their expert skills: researchers may focus on complex algorithms, practitioners may focus on their business goals. Through the simplified integration (intermediate) research results can be introduced into a productive environment which enables getting an early user feedback and allows for the early evaluation of different approaches. The practitioners' business model benefits throughout the full project duration.}, language = {en} } @inproceedings{SchmidtsKraftSiebigterothetal.2019, author = {Schmidts, Oliver and Kraft, Bodo and Siebigteroth, Ines and Z{\"u}ndorf, Albert}, title = {Schema Matching with Frequent Changes on Semi-Structured Input Files: A Machine Learning Approach on Biological Product Data}, series = {Proceedings of the 21st International Conference on Enterprise Information Systems - Volume 1: ICEIS}, booktitle = {Proceedings of the 21st International Conference on Enterprise Information Systems - Volume 1: ICEIS}, isbn = {978-989-758-372-8}, doi = {10.5220/0007723602080215}, pages = {208 -- 215}, year = {2019}, language = {en} } @inproceedings{SiebigterothKraftSchmidtsetal.2019, author = {Siebigteroth, Ines and Kraft, Bodo and Schmidts, Oliver and Z{\"u}ndorf, Albert}, title = {A Study on Improving Corpus Creation by Pair Annotation}, series = {Proceedings of the Poster Session of the 2nd Conference on Language, Data and Knowledge (LDK-PS 2019)}, booktitle = {Proceedings of the Poster Session of the 2nd Conference on Language, Data and Knowledge (LDK-PS 2019)}, issn = {1613-0073}, pages = {40 -- 44}, year = {2019}, language = {en} } @inproceedings{SchmidtsKraftWinkensetal.2020, author = {Schmidts, Oliver and Kraft, Bodo and Winkens, Marvin and Z{\"u}ndorf, Albert}, title = {Catalog integration of low-quality product data by attribute label ranking}, series = {Proceedings of the 9th International Conference on Data Science, Technology and Applications - Volume 1: DATA}, booktitle = {Proceedings of the 9th International Conference on Data Science, Technology and Applications - Volume 1: DATA}, isbn = {978-989-758-440-4}, doi = {10.5220/0009831000900101}, pages = {90 -- 101}, year = {2020}, language = {en} } @inproceedings{KloeserKohlKraftetal.2021, author = {Kl{\"o}ser, Lars and Kohl, Philipp and Kraft, Bodo and Z{\"u}ndorf, Albert}, title = {Multi-attribute relation extraction (MARE): simplifying the application of relation extraction}, series = {Proceedings of the 2nd International Conference on Deep Learning Theory and Applications - DeLTA}, booktitle = {Proceedings of the 2nd International Conference on Deep Learning Theory and Applications - DeLTA}, isbn = {978-989-758-526-5}, doi = {10.5220/0010559201480156}, pages = {148 -- 156}, year = {2021}, abstract = {Natural language understanding's relation extraction makes innovative and encouraging novel business concepts possible and facilitates new digitilized decision-making processes. Current approaches allow the extraction of relations with a fixed number of entities as attributes. Extracting relations with an arbitrary amount of attributes requires complex systems and costly relation-trigger annotations to assist these systems. We introduce multi-attribute relation extraction (MARE) as an assumption-less problem formulation with two approaches, facilitating an explicit mapping from business use cases to the data annotations. Avoiding elaborated annotation constraints simplifies the application of relation extraction approaches. The evaluation compares our models to current state-of-the-art event extraction and binary relation extraction methods. Our approaches show improvement compared to these on the extraction of general multi-attribute relations.}, language = {en} } @inproceedings{KohlSchmidtsKloeseretal.2021, author = {Kohl, Philipp and Schmidts, Oliver and Kl{\"o}ser, Lars and Werth, Henri and Kraft, Bodo and Z{\"u}ndorf, Albert}, title = {STAMP 4 NLP - an agile framework for rapid quality-driven NLP applications development}, series = {Quality of Information and Communications Technology. QUATIC 2021}, booktitle = {Quality of Information and Communications Technology. QUATIC 2021}, publisher = {Springer}, address = {Cham}, isbn = {978-3-030-85346-4}, doi = {10.1007/978-3-030-85347-1_12}, pages = {156 -- 166}, year = {2021}, abstract = {The progress in natural language processing (NLP) research over the last years, offers novel business opportunities for companies, as automated user interaction or improved data analysis. Building sophisticated NLP applications requires dealing with modern machine learning (ML) technologies, which impedes enterprises from establishing successful NLP projects. Our experience in applied NLP research projects shows that the continuous integration of research prototypes in production-like environments with quality assurance builds trust in the software and shows convenience and usefulness regarding the business goal. We introduce STAMP 4 NLP as an iterative and incremental process model for developing NLP applications. With STAMP 4 NLP, we merge software engineering principles with best practices from data science. Instantiating our process model allows efficiently creating prototypes by utilizing templates, conventions, and implementations, enabling developers and data scientists to focus on the business goals. Due to our iterative-incremental approach, businesses can deploy an enhanced version of the prototype to their software environment after every iteration, maximizing potential business value and trust early and avoiding the cost of successful yet never deployed experiments.}, language = {en} } @inproceedings{SchmidtsKraftWinkensetal.2021, author = {Schmidts, Oliver and Kraft, Bodo and Winkens, Marvin and Z{\"u}ndorf, Albert}, title = {Catalog integration of heterogeneous and volatile product data}, series = {DATA 2020: Data Management Technologies and Applications}, booktitle = {DATA 2020: Data Management Technologies and Applications}, publisher = {Springer}, address = {Cham}, isbn = {978-3-030-83013-7}, doi = {10.1007/978-3-030-83014-4_7}, pages = {134 -- 153}, year = {2021}, abstract = {The integration of frequently changing, volatile product data from different manufacturers into a single catalog is a significant challenge for small and medium-sized e-commerce companies. They rely on timely integrating product data to present them aggregated in an online shop without knowing format specifications, concept understanding of manufacturers, and data quality. Furthermore, format, concepts, and data quality may change at any time. Consequently, integrating product catalogs into a single standardized catalog is often a laborious manual task. Current strategies to streamline or automate catalog integration use techniques based on machine learning, word vectorization, or semantic similarity. However, most approaches struggle with low-quality or real-world data. We propose Attribute Label Ranking (ALR) as a recommendation engine to simplify the integration process of previously unknown, proprietary tabular format into a standardized catalog for practitioners. We evaluate ALR by focusing on the impact of different neural network architectures, language features, and semantic similarity. Additionally, we consider metrics for industrial application and present the impact of ALR in production and its limitations.}, language = {en} } @inproceedings{BuesgenKloeserKohletal.2022, author = {B{\"u}sgen, Andr{\´e} and Kl{\"o}ser, Lars and Kohl, Philipp and Schmidts, Oliver and Kraft, Bodo and Z{\"u}ndorf, Albert}, title = {Exploratory analysis of chat-based black market profiles with natural language processing}, series = {Proceedings of the 11th International Conference on Data Science, Technology and Applications}, booktitle = {Proceedings of the 11th International Conference on Data Science, Technology and Applications}, isbn = {978-989-758-583-8}, issn = {2184-285X}, doi = {10.5220/0011271400003269}, pages = {83 -- 94}, year = {2022}, abstract = {Messenger apps like WhatsApp or Telegram are an integral part of daily communication. Besides the various positive effects, those services extend the operating range of criminals. Open trading groups with many thousand participants emerged on Telegram. Law enforcement agencies monitor suspicious users in such chat rooms. This research shows that text analysis, based on natural language processing, facilitates this through a meaningful domain overview and detailed investigations. We crawled a corpus from such self-proclaimed black markets and annotated five attribute types products, money, payment methods, user names, and locations. Based on each message a user sends, we extract and group these attributes to build profiles. Then, we build features to cluster the profiles. Pretrained word vectors yield better unsupervised clustering results than current state-of-the-art transformer models. The result is a semantically meaningful high-level overview of the user landscape of black market chatrooms. Additionally, the extracted structured information serves as a foundation for further data exploration, for example, the most active users or preferred payment methods.}, language = {en} } @inproceedings{SildatkeKarwanniKraftetal.2020, author = {Sildatke, Michael and Karwanni, Hendrik and Kraft, Bodo and Schmidts, Oliver and Z{\"u}ndorf, Albert}, title = {Automated Software Quality Monitoring in Research Collaboration Projects}, series = {ICSEW'20: Proceedings of the IEEE/ACM 42nd International Conference on Software Engineering Workshops}, booktitle = {ICSEW'20: Proceedings of the IEEE/ACM 42nd International Conference on Software Engineering Workshops}, doi = {10.1145/3387940.3391478}, pages = {603 -- 610}, year = {2020}, language = {en} } @inproceedings{BuesgenKloeserKohletal.2023, author = {B{\"u}sgen, Andr{\´e} and Kl{\"o}ser, Lars and Kohl, Philipp and Schmidts, Oliver and Kraft, Bodo and Z{\"u}ndorf, Albert}, title = {From cracked accounts to fake IDs: user profiling on German telegram black market channels}, series = {Data Management Technologies and Applications}, booktitle = {Data Management Technologies and Applications}, editor = {Cuzzocrea, Alfredo and Gusikhin, Oleg and Hammoudi, Slimane and Quix, Christoph}, publisher = {Springer}, address = {Cham}, isbn = {978-3-031-37889-8 (Print)}, doi = {10.1007/978-3-031-37890-4_9}, pages = {176 -- 202}, year = {2023}, abstract = {Messenger apps like WhatsApp and Telegram are frequently used for everyday communication, but they can also be utilized as a platform for illegal activity. Telegram allows public groups with up to 200.000 participants. Criminals use these public groups for trading illegal commodities and services, which becomes a concern for law enforcement agencies, who manually monitor suspicious activity in these chat rooms. This research demonstrates how natural language processing (NLP) can assist in analyzing these chat rooms, providing an explorative overview of the domain and facilitating purposeful analyses of user behavior. We provide a publicly available corpus of annotated text messages with entities and relations from four self-proclaimed black market chat rooms. Our pipeline approach aggregates the extracted product attributes from user messages to profiles and uses these with their sold products as features for clustering. The extracted structured information is the foundation for further data exploration, such as identifying the top vendors or fine-granular price analyses. Our evaluation shows that pretrained word vectors perform better for unsupervised clustering than state-of-the-art transformer models, while the latter is still superior for sequence labeling.}, language = {en} } @inproceedings{KohlFreyerKraemeretal.2023, author = {Kohl, Philipp and Freyer, Nils and Kr{\"a}mer, Yoka and Werth, Henri and Wolf, Steffen and Kraft, Bodo and Meinecke, Matthias and Z{\"u}ndorf, Albert}, title = {ALE: a simulation-based active learning evaluation framework for the parameter-driven comparison of query strategies for NLP}, series = {Deep Learning Theory and Applications. DeLTA 2023. Communications in Computer and Information Science}, booktitle = {Deep Learning Theory and Applications. DeLTA 2023. Communications in Computer and Information Science}, editor = {Conte, Donatello and Fred, Ana and Gusikhin, Oleg and Sansone, Carlo}, publisher = {Springer}, address = {Cham}, isbn = {978-3-031-39058-6 (Print)}, doi = {978-3-031-39059-3}, pages = {235 -- 253}, year = {2023}, abstract = {Supervised machine learning and deep learning require a large amount of labeled data, which data scientists obtain in a manual, and time-consuming annotation process. To mitigate this challenge, Active Learning (AL) proposes promising data points to annotators they annotate next instead of a subsequent or random sample. This method is supposed to save annotation effort while maintaining model performance. However, practitioners face many AL strategies for different tasks and need an empirical basis to choose between them. Surveys categorize AL strategies into taxonomies without performance indications. Presentations of novel AL strategies compare the performance to a small subset of strategies. Our contribution addresses the empirical basis by introducing a reproducible active learning evaluation (ALE) framework for the comparative evaluation of AL strategies in NLP. The framework allows the implementation of AL strategies with low effort and a fair data-driven comparison through defining and tracking experiment parameters (e.g., initial dataset size, number of data points per query step, and the budget). ALE helps practitioners to make more informed decisions, and researchers can focus on developing new, effective AL strategies and deriving best practices for specific use cases. With best practices, practitioners can lower their annotation costs. We present a case study to illustrate how to use the framework.}, language = {en} } @inproceedings{KloeserBuesgenKohletal.2023, author = {Kl{\"o}ser, Lars and B{\"u}sgen, Andr{\´e} and Kohl, Philipp and Kraft, Bodo and Z{\"u}ndorf, Albert}, title = {Explaining relation classification models with semantic extents}, series = {DeLTA 2023: Deep Learning Theory and Applications}, booktitle = {DeLTA 2023: Deep Learning Theory and Applications}, editor = {Conte, Donatello and Fred, Ana and Gusikhin, Oleg and Sansone, Carlo}, publisher = {Springer}, address = {Cham}, isbn = {978-3-031-39058-6 (Print)}, doi = {10.1007/978-3-031-39059-3_13}, pages = {189 -- 208}, year = {2023}, abstract = {In recent years, the development of large pretrained language models, such as BERT and GPT, significantly improved information extraction systems on various tasks, including relation classification. State-of-the-art systems are highly accurate on scientific benchmarks. A lack of explainability is currently a complicating factor in many real-world applications. Comprehensible systems are necessary to prevent biased, counterintuitive, or harmful decisions. We introduce semantic extents, a concept to analyze decision patterns for the relation classification task. Semantic extents are the most influential parts of texts concerning classification decisions. Our definition allows similar procedures to determine semantic extents for humans and models. We provide an annotation tool and a software framework to determine semantic extents for humans and models conveniently and reproducibly. Comparing both reveals that models tend to learn shortcut patterns from data. These patterns are hard to detect with current interpretability methods, such as input reductions. Our approach can help detect and eliminate spurious decision patterns during model development. Semantic extents can increase the reliability and security of natural language processing systems. Semantic extents are an essential step in enabling applications in critical areas like healthcare or finance. Moreover, our work opens new research directions for developing methods to explain deep learning models.}, language = {en} } @article{SildatkeKarwanniKraftetal.2023, author = {Sildatke, Michael and Karwanni, Hendrik and Kraft, Bodo and Z{\"u}ndorf, Albert}, title = {A distributed microservice architecture pattern for the automated generation of information extraction pipelines}, series = {SN Computer Science}, journal = {SN Computer Science}, number = {4, Article number: 833}, publisher = {Springer Singapore}, address = {Singapore}, issn = {2661-8907}, doi = {10.1007/s42979-023-02256-4}, pages = {19 Seiten}, year = {2023}, abstract = {Companies often build their businesses based on product information and therefore try to automate the process of information extraction (IE). Since the information source is usually heterogeneous and non-standardized, classic extract, transform, load techniques reach their limits. Hence, companies must implement the newest findings from research to tackle the challenges of process automation. They require a flexible and robust system that is extendable and ensures the optimal processing of the different document types. This paper provides a distributed microservice architecture pattern that enables the automated generation of IE pipelines. Since their optimal design is individual for each input document, the system ensures the ad-hoc generation of pipelines depending on specific document characteristics at runtime. Furthermore, it introduces the automated quality determination of each available pipeline and controls the integration of new microservices based on their impact on the business value. The introduced system enables fast prototyping of the newest approaches from research and supports companies in automating their IE processes. Based on the automated quality determination, it ensures that the generated pipelines always meet defined business requirements when they come into productive use.}, language = {en} }