Refine
Year of publication
Document Type
- Conference Proceeding (65)
- Working Paper (65)
- Article (50)
- Bachelor Thesis (45)
- Report (34)
- Master's Thesis (21)
- Book (19)
- Study Thesis (10)
- Course Material (9)
- Part of a Book (6)
Has Fulltext
- yes (337)
Is part of the Bibliography
- no (337)
Keywords
- Digitalisierung (16)
- E-Learning (15)
- Agilität <Management> (11)
- Führung (11)
- Grader (11)
- Herbarium (11)
- Korruption (11)
- Programmierung (11)
- Autobewerter (10)
- Computerunterstütztes Lernen (10)
Institute
- Fakultät IV - Wirtschaft und Informatik (337) (remove)
The digital transformation with its new technologies and customer expectation has a significant effect on the customer channels in the insurance industry. The objective of this study is the identification of enabling and hindering factors for the adoption of online claim notification services that are an important part of the customer experience in insurance. For this purpose, we conducted a quantitative cross-sectional survey based on the exemplary scenario of car insurance in Germany and analyzed the data via structural equation modeling (SEM). The findings show that, besides classical technology acceptance factors such as perceived usefulness and ease of use, digital mindset and status quo behavior play a role: acceptance of digital innovations, lacking endurance as well as lacking frustration tolerance with the status quo lead to a higher intention for use. Moreover, the results are strongly moderated by the severity of the damage event—an insurance-specific factor that is sparsely considered so far. The latter discovery implies that customers prefer a communication channel choice based on the individual circumstances of the claim.
During the Corona-Pandemic, information (e.g. from the analysis of balance sheets and payment behavior) traditionally used for corporate credit risk analysis became less valuable because it represents only past circumstances. Therefore, the use of currently published data from social media platforms, which have shown to contain valuable information regarding the financial stability of companies, should be evaluated. In this data e. g. additional information from disappointed employees or customers can be present. In order to analyze in how far this data can improve the information base for corporate credit risk assessment, Twitter data regarding the ten greatest insolvencies of German companies in 2020 and solvent counterparts is analyzed in this paper. The results from t-tests show, that sentiment before the insolvencies is significantly worse than in the comparison group which is in alignment with previously conducted research endeavors. Furthermore, companies can be classified as prospectively solvent or insolvent with up to 70% accuracy by applying the k-nearest-neighbor algorithm to monthly aggregated sentiment scores. No significant differences in the number of Tweets for both groups can be proven, which is in contrast to findings from studies which were conducted before the Corona-Pandemic. The results can be utilized by practitioners and scientists in order to improve decision support systems in the domain of corporate credit risk analysis. From a scientific point of view, the results show, that the information asymmetry between lenders and borrowers in credit relationships, which are principals and agents according to the principal-agent-theory, can be reduced based on user generated content from social media platforms. In future studies, it should be evaluated in how far the data can be integrated in established processes for credit decision making. Furthermore, additional social media platforms as well as samples of companies should be analyzed. Lastly, the authenticity of user generated contend should be taken into account in order to ensure, that credit decisions rely on truthful information only.
Unternehmen, die sich ernsthaft mit Nachhaltigkeit beschäftigen, müssen den Nachweis erbringen, dass sie positive Effekte für die Gesellschaft erzielen. Damit ist eine ganzheitliche Wirkungsmessung unabdingbar. Sozialunternehmen sollten als Vorbild für eine solche Wirkungsmessung dienen. Eine wissenschaftliche Studie auf Basis der sog. „Ergebnispyramide“ kommt jedoch zu dem Schluss, dass selbst diese ihre Wirkung bisher kaum ganzheitlich messen.
In this paper we describe the selection of a modern build automation tool for an industry research partner of ours, namely an insurance company. Build automation has become increasingly important over the years. Today, build automation became one of the central concepts in topics such as cloud native development based on microservices and DevOps. Since more and more products for build automation have entered the market and existing tools have changed their functional scope, there is nowadays a large number of tools on the market that differ greatly in their functional scope. Based on requirements from our partner company, a build server analysis was conducted. This paper presents our analysis requirements, a detailed look at one of the examined tools and a summarizes our comparison of all three tools from our final comparison round.
Music streaming platforms offer music listeners an overwhelming choice of music. Therefore, users of streaming platforms need the support of music recommendation systems to find music that suits their personal taste. Currently, a new class of recommender systems based on knowledge graph embeddings promises to improve the quality of recommendations, in particular to provide diverse and novel recommendations. This paper investigates how knowledge graph embeddings can improve music recommendations. First, it is shown how a collaborative knowledge graph can be derived from open music data sources. Based on this knowledge graph, the music recommender system EARS (knowledge graph Embedding-based Artist Recommender System) is presented in detail, with particular emphasis on recommendation diversity and explainability. Finally, a comprehensive evaluation with real-world data is conducted, comparing of different embeddings and investigating the influence of different types of knowledge.
Die Angriffserkennung ist ein wesentlicher Bestandteil, Cyberangriffe zu verhindern und abzumildern. Dazu werden Daten aus verschiedenen Quellen gesammelt und auf Einbruchsspuren durchsucht. Die heutzutage produzierten Datenmengen sind ein wesentliches Problem für die Angriffserkennung. Besonders bei komplexen Cyberangriffen, die über einen längeren Zeitraum stattfinden, wächst die zu durchsuchende Datenmenge stark an und erschwert das Finden und Kombinieren der einzelnen Angriffsschritte.
Eine mögliche Lösung, um dem Problem entgegenzuwirken, ist die Reduktion der Datenmenge. Die Datenreduktion versucht, Daten herauszufiltern, die aus Sicht der Angriffserkennung irrelevant sind. Diese Ansätze werden unter dem Begriff Reduktionstechniken zusammengefasst. In dieser Arbeit werden Reduktionstechniken aus der Wissenschaft untersucht und auf Benchmark Datensätzen angewendet, um ihre Nutzbarkeit zu evaluieren. Dabei wird der Frage nachgegangen, ob die Reduktionstechniken in der Lage sind, irrelevante Daten ausfindig zu machen und zu reduzieren, ohne dass eine Beeinträchtigung der Angriffserkennung stattfindet. Die Evaluation der Angriffserkennung erfolgt durch ThreaTrace, welches eine Graph Neural Network basierte Methode ist.
Die Evaluierung zeigt, dass mehrere Reduktionstechniken die Datenmenge wesentlich reduzieren können, ohne die Angriffserkennung zu beeinträchtigen. Bei drei Techniken führt der Einsatz zu keinen nennenswerten Veränderungen der Erkennungsraten. Dabei wurden Reduktionsraten von bis zu 30 % erreicht. Bei der Anwendung einer Reduktionstechnik stieg die Erkennungsleistung sogar um 8 %. Lediglich bei zwei Techniken führt der Einsatz zum drastischen Absinken der Erkennungsrate.
Insgesamt zeigt die Arbeit, dass eine Datenreduktion angewandt werden kann, ohne die Angriffserkennung zu beeinträchtigen. In besonderen Fällen kann eine Datenreduktion, die Erkennungsleistung sogar verbessern. Allerdings ist der erfolgreiche Einsatz der Reduktionstechniken abhängig vom verwendeten Datensatz und der verwendeten Methode der Angriffserkennung.
The transfer of historically grown monolithic software architectures into modern service-oriented architectures creates a lot of loose coupling points. This can lead to an unforeseen system behavior and can significantly impede those continuous modernization processes, since it is not clear where bottlenecks in a system arise. It is therefore necessary to monitor such modernization processes with an adaptive monitoring concept to be able to correctly record and interpret unpredictable system dynamics. This contribution presents a generic QoS measurement framework for service-based systems. The framework consists of an XML-based specification for the measurement to be performed – the Information Model (IM) – and the QoS System, which provides an execution platform for the IM. The framework will be applied to a standard business process of the German insurance industry, and the concepts of the IM and their mapping to artifacts of the QoS System will be presented. Furtherm ore, design and implementation of the QoS System’s parser and generator module and the generated artifacts are explained in detail, e.g., event model, agents, measurement module and analyzer module.
Der Bachelor-Studiengang Mediendesigninformatik der Hochschule Hannover ist ein Informatikstudiengang mit dem speziellen Anwendungsgebiet Mediendesign. In Abgrenzung von Studiengängen der Medieninformatik liegt der Anwendungsfokus auf der kreativen Gestaltung etwa von 3D-Modellierungen, Animationen und Computerspielen. Absolvent*innen des Studiengangs sollen an der Schnittstelle zwischen Informatik und Mediendesign agieren können, zum Beispiel bei der Erstellung von Benutzungsschnittstellen und VR/AR-Anwendungen. Der Artikel stellt das Curriculum des interdisziplinären Studiengangs vor und reflektiert nach dem Abschluss der ersten beiden Studierendenkohorten die Erfahrungen, indem die ursprünglichen Ziele den Zahlen der Hochschulstatistik und den Ergebnissen zweier Studierendenbefragungen gegenübergestellt werden.
Even for the more traditional insurance industry, the Microservices Architecture (MSA) style plays an increasingly important role in provisioning insurance services. However, insurance businesses must operate legacy applications, enterprise software, and service-based applications in parallel for a more extended transition period. The ultimate goal of our ongoing research is to design a microservice reference architecture in cooperation with our industry partners from the insurance domain that provides an approach for the integration of applications from different architecture paradigms. In Germany, individual insurance services are classified as part of the critical infrastructure. Therefore, German insurance companies must comply with the Federal Office for Information Security requirements, which the Federal Supervisory Authority enforces. Additionally, insurance companies must comply with relevant laws, regulations, and standards as part of the business’s compliance requirements. Note: Since Germany is seen as relatively ’tough’ with respect to privacy and security demands, fullfilling those demands might well be suitable (if not even ’over-achieving’) for insurances in other countries as well. The question raises thus, of how insurance services can be secured in an application landscape shaped by the MSA style to comply with the architectural and security requirements depicted above. This article highlights the specific regulations, laws, and standards the insurance industry must comply with. We present initial architectural patterns to address authentication and authorization in an MSA tailored to the requirements of our insurance industry partners.
Cloud computing has become well established in private and public sector projects over the past few years, opening ever new opportunities for research and development, but also for education. One of these opportunities presents itself in the form of dynamically deployable, virtual lab environments, granting educational institutions increased flexibility with the allocation of their computing resources. These fully sandboxed labs provide students with their own, internal network and full access to all machines within, granting them the flexibility necessary to gather hands-on experience with building heterogeneous microservice architectures. The eduDScloud provides a private cloud infrastructure to which labs like the microservice lab outlined in this paper can be flexibly deployed at a moment’s notice.
In this paper the workflow of the project 'Untersuchungs-, Simulations- und Evaluationstool für Urbane Logistik` (USEfUL) is presented. Aiming to create a web-based decision support tool for urban logistics, the project needed to integrate multiple steps into a single workflow, which in turn needed to be executed multiple times. While a service-oriented system could not be created, the principles of service orientation was utilized to increase workflow efficiency and flexibility, allowing the workflow to be easily adapted to new concepts or research areas.
In the context of modern mobility, topics such as smart-cities, Car2Car-Communication, extensive vehicle sensor-data, e-mobility and charging point management systems have to be considered. These topics of modern mobility often have in common that they are characterized by complex and extensive data situations. Vehicle position data, sensor data or vehicle communication data must be preprocessed, aggregated and analyzed. In many cases, the data is interdependent. For example, the vehicle position data of electric vehicles and surrounding charging points have a dependence on one another and characterize a competition situation between the vehicles. In the case of Car2Car-Communication, the positions of the vehicles must also be viewed in relation to each other. The data are dependent on each other and will influence the ability to establish a communication. This dependency can provoke very complex and large data situations, which can no longer be treated efficiently. With this work, a model is presented in order to be able to map such typical data situations with a strong dependency of the data among each other. Microservices can help reduce complexity.
Microservices build a deeply distributed system. Although this offers significant flexibility for development teams and helps to find solutions for scalability or security questions, it also intensifies the drawbacks of a distributed system. This article offers a decision framework, which helps to increase the resiliency of microservices. A metamodel is used to represent services, resiliency patterns, and quality attributes. Furthermore, the general idea for a suggestion procedure is outlined.
Portable-micro-Combined-Heat-and-Power-units are a gateway technology bridging conventional vehicles and Battery Electric Vehicles (BEV). Being a new technology, new software has to be created that can be easily adapted to changing requirements. We propose and evaluate three different architectures based on three architectural paradigms. Using a scenario-based evaluation, we conclude that a Service-Oriented Architecture (SOA) using microservices provides a higher quality solution than a layered or Event-Driven Complex-Event-Processing (ED-CEP) approach. Future work will include implementation and simulation-driven evaluation.
The usage of microservices promises a lot of benefits concerning scalability and maintainability, rewriting large monoliths is however not always possible. Especially in scientific projects, pure microservice architectures are therefore not feasible in every project. We propose the utilization of microservice principles for the construction of microsimulations for urban transport. We present a prototypical architecture for the connection of MATSim and AnyLogic, two widely used simulation tools in the context of urban transport simulation. The proposed system combines the two tools into a singular tool supporting civil engineers in decision making on innovative urban transport concepts.
To avoid the shortcomings of traditional monolithic applications, the Microservices Architecture (MSA) style plays an increasingly important role in providing business services. This is true even for the more conventional insurance industry with its highly heterogeneous application landscape and sophisticated cross-domain business processes. Therefore, the question arises of how workflows can be implemented to grant the required flexibility and agility and, on the other hand, to exploit the potential of the MSA style. In this article, we present two different approaches – orchestration and choreography. Using an application scenario from the insurance domain, both concepts are discussed. We introduce a pattern that outlines the mapping of a workflow to a choreography.
Microservices are meanwhile an established software engineering vehicle, which more and more companies are examining and adopting for their development work. Naturally, reference architectures based on microservices come into mind as a valuable thing to utilize. Initial results for such architectures are published in generic and in domain-specific form. Missing to the best of our knowledge however, is a domain-specific reference architecture based on microservices, which takes into account specifics of the insurance industry domain. Jointly with partners from the German insurance industry, we take initial steps to fill this gap in the present article. Thus, we aim towards a microservices-based reference software architecture for (at least German) insurance companies. As the main results of this article we thus provide an initial such reference architecture together with a deeper look into two important parts of it.
The transfer of historically grown monolithic software architectures into modern service-oriented architectures creates a lot of loose coupling points. This can lead to an unforeseen system behavior and can significantly impede those continuous modernization processes, since it is not clear where bottlenecks in a system arise. It is therefore necessary to monitor such modernization processes with an adaptive monitoring concept in order to be able to correctly record and interpret unpredictable system dynamics. For this purpose, a general measurement methodology and a specific implementation concept are presented in this work.
In microservice architectures, data is often hold redundantly to create an overall resilient system. Although the synchronization of this data proposes a significant challenge, not much research has been done on this topic yet. This paper shows four general approaches for assuring consistency among services and demonstrates how to identify the best solution for a given architecture. For this, a microservice architecture, which implements the functionality of a mainframe-based legacy system from the insurance industry, serves as an example.
Microservices is an architectural style for complex application systems, promising some crucial benefits, e.g. better maintainability, flexible scalability, and fault tolerance. For this reason microservices has attracted attention in the software development departments of different industry sectors, such as ecommerce and streaming services. On the other hand, businesses have to face great challenges, which hamper the adoption of the architectural style. For instance, data are often persisted redundantly to provide fault tolerance. But the synchronization of those data for the sake of consistency is a major challenge. Our paper presents a case study from the insurance industry which focusses consistency issues when migrating a monolithic core application towards microservices. Based on the Domain Driven Design (DDD) methodology, we derive bounded contexts and a set of microservices assigned to these contexts. We discuss four different approaches to ensure consistency and propose a best practice to identify the most appropriate approach for a given scenario. Design and implementation details and compliance issues are presented as well.
In this paper we describe the selection of a modern build automation tool for an industry research partner of ours, namely an insurance company. Build automation has become increasingly important over the years. Today, build automation became one of the central concepts in topics such as cloud native development based on microservices and DevOps. Since more and more products for build automation have entered the market and existing tools have changed their functional scope, there is nowadays a large number of tools on the market that differ greatly in their functional scope. Based on requirements from our partner company, a build server analysis was conducted. This paper presents our analysis requirements, a detailed look at one of the examined tools and a summarized comparison of two tools.
A Look at Service Meshes
(2021)
Service meshes can be seen as an infrastructure layer for microservice-based applications that are specifically suited for distributed application architectures. It is the goal to introduce the concept of service meshes and its use for microservices with the example of an open source service mesh called Istio. This paper gives an introduction into the service mesh concept and its relation to microservices. It also gives an overview of selected features provided by Istio as relevant to the above concept and provides a small sample setup that demonstrates the core features.
Cloud Computing: Serverless
(2021)
A serverless architecture is a new approach to offering services over the Internet. It combines BaaS (Backend-as-a-service) and FaaS (Function-as-a-service). With the serverless architecture no own or rented infrastructures are needed anymore. In addition, the company does not have to worry about scaling any longer, as this happens automatically and immediately. Furthermore, there is no need any longer for maintenance work on the servers, as this is completely taken over by the provider. Administrators are also no longer needed for the same reason. Finally, many ready-made functions are offered, with which the development effort can be reduced. As a result, the serverless architecture is very well suited to many application scenarios, and it can save considerable costs (server costs, maintenance costs, personnel costs, electricity costs, etc.). The company only must subdivide the source code of the application and upload it to the provider’s server. The rest is done by the provider.
In this paper, we present a novel approach for real-time rendering of soft eclipse shadows cast by spherical, atmosphereless bodies. While this problem may seem simple at first, it is complicated by several factors. First, the extreme scale differences and huge mutual distances of the involved celestial bodies cause rendering artifacts in practice. Second, the surface of the Sun does not emit light evenly in all directions (an effect which is known as limb darkening). This makes it impossible to model the Sun as a uniform spherical light source. Finally, our intended applications include real-time rendering of solar eclipses in virtual reality, which require very high frame rates. As a solution to these problems, we precompute the amount of shadowing into an eclipse shadow map, which is parametrized so that it is independent of the position and size of the occluder. Hence, a single shadow map can be used for all spherical occluders in the Solar System. We assess the errors introduced by various simplifications and compare multiple approaches in terms of performance and precision. Last but not least, we compare our approaches to the state-of-the-art and to reference images. The implementation has been published under the MIT license.
On November 30th, 2022, OpenAI released the large language model ChatGPT, an extension of GPT-3. The AI chatbot provides real-time communication in response to users’ requests. The quality of ChatGPT’s natural speaking answers marks a major shift in how we will use AI-generated information in our day-to-day lives. For a software engineering student, the use cases for ChatGPT are manifold: assessment preparation, translation, and creation of specified source code, to name a few. It can even handle more complex aspects of scientific writing, such as summarizing literature and paraphrasing text. Hence, this position paper addresses the need for discussion of potential approaches for integrating ChatGPT into higher education. Therefore, we focus on articles that address the effects of ChatGPT on higher education in the areas of software engineering and scientific writing. As ChatGPT was only recently released, there have been no peer-reviewed articles on the subject. Thus, we performed a structured grey literature review using Google Scholar to identify preprints of primary studies. In total, five out of 55 preprints are used for our analysis. Furthermore, we held informal discussions and talks with other lecturers and researchers and took into account the authors’ test results from using ChatGPT. We present five challenges and three opportunities for the higher education context that emerge from the release of ChatGPT. The main contribution of this paper is a proposal for how to integrate ChatGPT into higher education in four main areas.
Since textual user generated content from social media platforms contains valuable information for decision support and especially corporate credit risk analysis, automated approaches for text classification such as the application of sentiment dictionaries and machine learning algorithms have received great attention in recent user generated content based research endeavors. While machine learning algorithms require individual training data sets for varying sources, sentiment dictionaries can be applied to texts immediately, whereby domain specific dictionaries attain better results than domain independent word lists. We evaluate by means of a literature review how sentiment dictionaries can be constructed for specific domains and languages. Then, we construct nine versions of German sentiment dictionaries relying on a process model which we developed based on the literature review. We apply the dictionaries to a manually classified German language data set from Twitter in which hints for financial (in)stability of companies have been proven. Based on their classification accuracy, we rank the dictionaries and verify their ranking by utilizing Mc Nemar’s test for significance. Our results indicate, that the significantly best dictionary is based on the German language dictionary SentiWortschatz and an extension approach by use of the lexical-semantic database GermaNet. It achieves a classification accuracy of 59,19 % in the underlying three-case-scenario, in which the Tweets are labelled as negative, neutral or positive. A random classification would attain an accuracy of 33,3 % in the same scenario and hence, automated coding by use of the sentiment dictionaries can lead to a reduction of manual efforts. Our process model can be adopted by other researchers when constructing sentiment dictionaries for various domains and languages. Furthermore, our established dictionaries can be used by practitioners especially in the domain of corporate credit risk analysis for automated text classification which has been conducted manually to a great extent up to today.
Nachhaltiges Anforderungsmanagement mit externen IT-Dienstleistern in der öffentlichen Verwaltung
(2023)
Kontext: Der Einfluss der Organisationskultur und -struktur auf das Outsourcing von IT-Prozessen in der öffentlichen Verwaltung wird untersucht.
Zielsetzung: Es wird die Wirkung der kulturellen und strukturellen Besonderheiten der öffentlichen Verwaltung auf die Zusammenarbeit mit externen IT-Dienstleistern erforscht. Ziel ist die Entwicklung eines nachhaltigen, zukunftswirksamen Konzepts des Anforderungsmanagements.
Methode: Expert*inneninterviews mit einer anschließenden qualitativen Inhaltsanalyse nach der Methodik von Gläser und Laudel werden durchgeführt.
Ergebnisse: Die Struktur der öffentlichen Verwaltung ist stark fragmentiert; die Kultur geprägt von Risikoaversion und bürokratischem Handeln.
Konklusion: Eine Kombination aus Kommunikations-, Wissens- und Beziehungsmanagement ermöglicht ein nachhaltiges Anforderungsmanagement.
The paper provides a comprehensive overview of modeling and pricing cyber insurance and includes clear and easily understandable explanations of the underlying mathematical concepts. We distinguish three main types of cyber risks: idiosyncratic, systematic, and systemic cyber risks. While for idiosyncratic and systematic cyber risks, classical actuarial and financial mathematics appear to be well-suited, systemic cyber risks require more sophisticated approaches that capture both network and strategic interactions. In the context of pricing cyber insurance policies, issues of interdependence arise for both systematic and systemic cyber risks; classical actuarial valuation needs to be extended to include more complex methods, such as concepts of risk-neutral valuation and (set-valued) monetary risk measures.
Mobile crowdsourcing refers to systems where the completion of tasks necessarily requires physical movement of crowdworkers in an on-demand workforce. Evidence suggests that in such systems, tasks often get assigned to crowdworkers who struggle to complete those tasks successfully, resulting in high failure rates and low service quality. A promising solution to ensure higher quality of service is to continuously adapt the assignment and respond to failure-causing events by transferring tasks to better-suited workers who use different routes or vehicles. However, implementing task transfers in mobile crowdsourcing is difficult because workers are autonomous and may reject transfer requests. Moreover, task outcomes are uncertain and need to be predicted. In this paper, we propose different mechanisms to achieve outcome prediction and task coordination in mobile crowdsourcing. First, we analyze different data stream learning approaches for the prediction of task outcomes. Second, based on the suggested prediction model, we propose and evaluate two different approaches for task coordination with different degrees of autonomy: an opportunistic approach for crowdshipping with collaborative, but non-autonomous workers, and a market-based model with autonomous workers for crowdsensing.
Begleitheft zum Seminar "Psychology at Work - Selbstmanagement im Unternehmen und Authentic Leadership". Checklisten, Übungen, Arbeitsblätter, Links und Foliensätze. Zusätzlich zu den Videos und den Live online Einheiten enthält das Begleitheft die Video-Übungen zur besseren Übersicht und Bearbeitung, sowie zusätzliche online-Links und vertiefende Arbeitsblätter.
Begleitheft zum Seminar "Psychology at work. Positive Psychologie im Unternehmen und Positive Leadership". Checklisten, Übungen, Arbeitsblätter, Links und Foliensätze.
Zusätzlich zu den Videos und den Live online Einheiten enthält das Begleitheft die Video-Übungen zur besseren Übersicht und Bearbeitung, sowie zusätzliche online-Links und vertiefende Arbeitsblätter.
High-performance firms typically have two features in common: (i) they produce in more than one country and (ii) they produce more than one product. In this paper, we analyze the internationalization strategies of multi-product firms. Guided by several new stylized facts, we develop a theoretical model to determine optimal modes of market access at the firm–product level. We find that the most productive firmssell core varieties via foreign direct investment and export products with intermediate productivity. Shocks to trade costs and technology affect the endogenous decision to export or produce abroad at the product-level and, in turn, the relative productivity between parents and affiliates.
Die gesetzlich vorgesehene Bereitstellung von Digitalisierungsangeboten stellt öffentliche Verwaltungen vor steigende Herausforderungen. Aufgrund der Heterogenität der Nutzerinnen und Nutzer ist es für öffentliche Verwaltungen häufig problematisch, klare Anforderungen zu erheben und zu erfüllen. Hinzukommen strukturelle und organisatorische Gegebenheiten wie beispielsweise ausgeprägte Entscheidungshierarchien, die eine nutzerzentrierte Vorgehensweise erschweren können. Darüber hinaus sieht sich die öffentliche Verwaltung zunehmend mit komplexer werdenden Problemen konfrontiert. Es stellt sich daher die Frage, wie in der öffentlichen Verwaltung ein moderner Ansatz zur Nutzerzentrierung und Problemlösung eingesetzt werden kann. Dieser Artikel präsentiert die Ergebnisse einer Einzelfallstudie bei der Niedersächsischen Landesbehörde für Straßenbau und Verkehr (NLStBV). Wir haben mit einer Fokusgruppe einen Design-Thinking-Workshop durchgeführt, um Potenziale und Anwendungsmöglichkeiten des Ansatzes in der öffentlichen Verwaltung zu identifizieren. Auf Basis einer SWOT-Analyse haben wir die Ergebnisse untersucht und geben vier konkrete Handlungsempfehlungen für die Einführung sowie Nutzung von Design Thinking.
Context: Agile software development (ASD) sets social aspects like communication and collaboration in focus. Thus, one may assume that the specific work organization of companies impacts the work of ASD teams. A major change in work organization is the switch to a 4-day work week, which some companies investigated in experiments. Also, recent studies show that ASD teams are affected by the switch to remote work since the Covid 19 pandemic outbreak in 2020.
Objective: Our study presents empirical findings on the effects on ASD teams operating remote in a 4-day work week organization. Method: We performed a qualitative single case study and conducted seven semi-structured interviews, observed 14 agile practices and screened eight project documents and protocols of agile practices.
Results: We found, that the teams adapted the agile method in use due to the change to a 4-day work week environment and the switch to remote work. The productivity of the two ASD teams did not decrease. Although the stress level of the ASD team member increased due to the 4-day work week, we found that the job satisfaction of the individual ASD team members is affected positively. Finally, we point to affects on social facets of the ASD teams.
Conclusion: The research community benefits from our results as the current state of research dealing with the effects of a 4-day work week on ASD teams is limited. Also, our findings provide several practical implications for ASD teams working remote in a 4-day work week.
Context: Companies adapt agile methods, practices or artifacts for their use in practice since more than two decades. This adaptions result in a wide variety of described agile practices. For instance, the Agile Alliance lists 75 different practices in its Agile Glossary. This situation may lead to misunderstandings, as agile practices with similar names can be interpreted and used differently.
Objective: This paper synthesize an integrated list of agile practices, both from primary and secondary sources.
Method: We performed a tertiary study to identify existing overviews and lists of agile practices in the literature. We identified 876 studies, of which 37 were included.
Results: The results of our paper show that certain agile practices are listed and used more often in existing studies. Our integrated list of agile practices comprises 38 entries structured in five categories. Conclusion: The high number of agile practices and thus, the wide variety increased steadily over the past decades due to the adaption of agile methods. Based on our findings, we present a comprehensive overview of agile practices. The research community benefits from our integrated list of agile practices as a potential basis for future research. Also, practitioners benefit from our findings, as the structured overview of agile practices provides the opportunity to select or adapt practices for their specific needs.
Gutes Arbeiten für Mitarbeiter ist je nach Arbeitskontext unterschiedlich zu bewerten, hängt jedoch von der Gestaltung bestimmter Kontextfaktoren ab. Die Kontextfaktoren guter Arbeit sind der zentrale Forschungsgegenstand dieser Arbeit. Dabei steht ein E‑Commerce-Team (EC-Team) von Otto im Fokus der Untersuchungen.
Das Ziel unseres Artikels ist es, die Kontextfaktoren zu analysieren, die dazu führen, dass gute Arbeit ermöglicht wird. Dabei ist eine auf Dauer funktionierende Arbeitsweise gesucht, welche eine hohe Arbeitsqualität und -quantität ermöglicht. Dazu sind die beiden primären Ziele zu definieren, was gutes Arbeiten ausmacht und zum anderen die Kontextfaktoren für gutes Arbeiten innerhalb des EC-Teams bei Otto zu identifizieren.
Unsere Forschungsfrage lautet: Welche Kontextfaktoren sind für gutes Arbeiten bei Otto im EC-Team in der derzeitigen Remote-Arbeit besonders relevant und entsprechend gestaltbar?
Um die Forschungsfrage beantworten zu können, wird zunächst eine Literaturrecherche zur Definition von guter Arbeit vorgenommen. Anschließend wird untersucht, welche Faktoren laut Literatur zu einer guten Arbeit beitragen, um aus den resultierenden Faktoren Cluster zu bilden.
Die Cluster werden dem Otto EC-Team zur Abstimmung mit der Mehrpunktabfrage über das virtuelle Kollaborations-Tool MiroFootnote 2 zur Verfügung gestellt. Aufbauend auf dem Ergebnis der Abstimmung, werden ein Gamification Board, Erinnerungsmails und ein Stimmungsbarometer erstellt, um die Auswirkungen des Clusters im Rahmen eines Experiments zu analysieren.
Diese Maßnahmen werden innerhalb von zwei Wochen durchgeführt. Um die Erfahrungen der Probanden zu sammeln, werden anschließend Interviews durchgeführt und ausgewertet. Die Ergebnisse der Interviews fließen in die anschließende Handlungsempfehlung ein.
Durch die Covid-19-Pandemie und die damit einhergehenden Effekte auf die Arbeitswelt ist die Belastung der Mitarbeitenden in einen stärkeren Fokus gerückt worden. Dieser Umstand trifft unter anderem durch den umfassenden Wechsel in die Remote Work auch auf agile Software-Entwicklungsteams in vielen Unternehmen zu. Eine zu hohe Arbeitsbelastung kann zu diversen negativen Effekten, wie einem erhöhten Krankenstand, mangelndem Wohlbefinden der Mitarbeitenden oder reduzierter Produktivität führen. Es ist zudem bekannt, dass sich die Arbeitsbelastung in der Wissensarbeit auf die Qualität der Arbeitsergebnisse auswirkt. Dieser Forschungsbeitrag identifiziert potenzielle Faktoren der Arbeitsbelastung der Mitglieder eines agilen Software-Entwicklungsteams bei der Otto GmbH & Co KG. Auf der Grundlage der Faktoren präsentieren wir Maßnahmen zur Reduzierung von Arbeitsbelastung und erläutern unsere Erkenntnisse, die wir im Rahmen eines Experiments validiert haben. Unsere Ergebnisse zeigen, dass bereits kleinteilige Maßnahmen, wie das Einführen von Ruhearbeitsphasen während des Arbeitstages, zu positiven Effekten bspw. hinsichtlich einer gesteigerten Konzentrationsfähigkeit führen und wie sich diese auf die Qualität der Arbeitsergebnisse auswirken.
Daten sind für jedes Unternehmen die treibende Kraft und die konsistenteste Quelle für qualifizierte Entscheidungsprozesse. Für die optimale Nutzung der vorliegenden Daten über alle Geschäftsbereiche hinweg wird das Datenmanagement benötigt, jedoch bringt dessen Einführung große Herausforderungen mit sich. Wird es nicht eingeführt bzw. umgesetzt hat dies Folgen für das Unternehmen wie z. B. Wettbewerbsnachteile und hohe Kosten. Ziel dieser Bachelorarbeit ist es von Datenmanagement, den Nutzen aufzuzeigen, die Herausforderungen zu identifizieren und Lösungsansätze zu untersuchen bzw. eigene zu entwickeln. Das Ergebnis ist eine Untersuchung und ein Vergleich des Nutzens, der Herausforderungen sowie der Lösungsansätze im Datenmanagement, zwischen Literatur und einem Anwendungsfall.
The purpose of this research is to explore results that are measured by social enterprises (= SEs) according to their mission and vision. Four SEs are examined for this reason. The status quo of aligned measurements was captured by conducting seven semi-structured interviews with persons from the middle and top management of the considered SEs. A conceptual framework, which categorizes output, outcome and impact measurements, is used as the basis for a structured content analysis. The findings imply that SEs’ measurements are not sufficiently aligned with their mission and vision. Outputs are measured by all considered SEs. However, they fail to measure outcomes with all its sublevels. Especially, measuring mindset change and behavior change outcomes are neglected by the examined SEs. That can lead to adjustments, where SEs only create more outputs but fail to create more outcomes and impact. Furthermore, neglecting outcome measurements makes existing but mostly unsystematic impact measurements invalid, since outputs, outcomes and impact build on each other. The research presented here provides one of the first investigations into the alignment of measurements with mission and vision in the context of SEs. Ultimately, the findings question SEs current measurements and aim to open further perspectives on improving the performance of SEs.
Die Covid-19 Pandemie hat zu einem signifikanten Anstieg der Remote Work geführt. Die Veränderung in der Interaktion und Kollaboration ist für viele agile Teams eine Herausforderung gewesen. Diverse Studien zeigen unterschiedliche Effekte und Auswirkungen auf die Zusammenarbeit agiler Teams während der Pandemie. So ist die Kommunikation sachlicher und zielgerichteter geworden. Ebenso wird eine Verminderung des sozialen Austauschs in den Teams berichtet. Unser Artikel thematisiert die Veränderung der Interaktion in agilen Teams durch die Remote Work. Wir haben eine qualitative Fallstudie bei einem agilen Software-Entwicklungsteam bei Otto durchgeführt. Unsere Ergebnisse zeigen einen Zusammenhang zwischen den Auswirkungen auf die Interaktion und der persönlichen Autonomie der Team-Mitglieder. Darüber hinaus haben wir keine signifikanten negativen Effekte durch die veränderte Interaktion auf die agile Arbeitsweise festgestellt.
Social skills are essential for a successful understanding of agile methods in software development. Several studies highlight the opportunities and advantages of integrating real-world projects and problems while collaborating with companies into higher education using agile methods. This integration comes with several opportunities and advantages for both the students and the company. The students are able to interact with real-world software development teams, analyze and understand their challenges and identify possible measures to tackle them. However, the integration of real-world problems and companies is complex and may come with a high effort in terms of coordination and preparation of the course. The challenges related to the interaction and communication with students are increased by virtual distance teaching during the Covid-19 pandemic as direct contact with students is missing. Also, we do not know how problem-based learning in virtual distance teaching is valued by the students. This paper presents our adapted eduScrum approach and learning outcome of integrating experiments with real-world software development teams from two companies into a Master of Science course organized in virtual distance teaching. The evaluation shows that students value analyzing real-world problems using agile methods. They highlight the interaction with real-world software development teams. Also, the students appreciate the organization of the course using an iterative approach with eduScrum. Based on our findings, we present four recommendations for the integration of agile methods and real world problems into higher education in virtual distance teaching settings. The results of our paper contribute to the practitioner and researcher/lecturer community, as we provide valuable insights how to fill the gap between practice and higher education in virtual distance settings.
An der Hochschule Hannover wird das Lern-Management-System Moodle eingesetzt. Für die dort verfügbare Quizfunktion wurde der neue Fragetyp MooPT (Moodle Programming Task) von der Abteilung Informatik der Fakultät IV entwickelt. Einem solchen Quiz muss immer ein Frageverhalten zugewiesen werden. Die Frageverhalten, welche Moodle standardmäßig anbietet, sind nicht mit dem MooPT-Fragetypen kompatibel. Aus diesem Grund müssen diese angepasst werden. Dies wurde bereits für zwei der acht Standard-Frageverhalten von der Abteilung Informatik durchgeführt.
Diese Arbeit beschäftigt sich mit der Entwicklung der restlichen Verhalten für den MooPT-Fragetypen. Dabei wird die Moodle-Quiz-Komponente und der MooPT-Fragetyp mit seinen Frageverhalten analysiert. Anschließen wird auf Basis dieser Analyse die Entwicklung der restlichen Frageverhalten erläutert.
Der Quellcode der Moodle-Plattform ist in PHP geschrieben und somit auch der des MooPT-Fragetypen und der der Frageverhalten. Auf GitHub stehen die Implementierungen der entwickelten Frageverhalten unter der GPL-Lizenz zur Verfügung:
https://github.com/LennartRolfes/moodle-qbehaviour_adaptivemoopt
https://github.com/LennartRolfes/moodle-qbehaviour_adaptivemooptnopenalty
https://github.com/LennartRolfes/moodle-qbehaviour_deferredmooptcbm
https://github.com/LennartRolfes/moodle-qbehaviour_immediatemooptcbm
https://github.com/LennartRolfes/moodle-qbehaviour_interactivemoopt
Im ländlichen Raum können Mobilitätsbedarfe schwer über den öffentlichen Personennahverkehr gedeckt werden. Wie diese Bedarfslücke über den Einsatz kombinierter Transportkonzepte von Personen und Gütern reduziert werden kann, wird prototypisch über eine agentenbasierte Simulationsanwendung in der Simulationssoftware AnyLogic untersucht. Reale Mobilitätsdaten werden dabei jedoch nicht berücksichtigt.
Das Ziel der vorliegenden Arbeit ist die Verbesserung der Datengrundlage des Prototypen mit Hilfe von Machine Learning. Unter Verwendung des Forschungsansatzes Design Science Research wurden ML-Modelle entlang des CRISP-DM Frameworks entwickelt. Diese verarbeiten die zur Verfügung stehenden Mobilitätsdaten und können nach deren Integration in den Prototypen zur Parametrierung genutzt werden. Im Zuge der Arbeit werden dazu geeignete Parameter identifiziert, die Mobilitätsdaten beschafft und umfangreich für das Modelltraining in H2O Driverless AI transformiert. Das beste ML-Modell wird in den Prototypen integriert und es werden notwendige Anpassungen vorgenommen, um die Parametrierung zu ermöglichen. Die anschließende Evaluation der Simulationsanwendung zeigt eine datenbasierte und realitätsgetreuere Simulation des simultanen und kombinierten Transports von Personen und Gütern.
Hybride Arbeit: Wo und wie wollen Mitglieder agiler Softwareentwicklungsteams zukünftig arbeiten?
(2022)
Vor dem Ausbruch der COVID-19 Pandemie haben Mitglieder agiler Softwareentwicklungsteams üblicherweise gemeinsam im Büro gearbeitet. Durch die rasche Verbreitung des Coronavirus wurden diese Teams weltweit ins Home Office geschickt, um die Ausbreitung des Virus einzudämmen. Verschiedene Studien weisen darauf hin, dass viele Mitglieder agiler Softwareentwicklungsteams nach dem Ende der Pandemie nicht wieder in Vollzeit in das Büro zurückkehren wollen, sondern einen hybriden Ansatz zwischen dem Büro und dem Home Office präferieren. Das Ziel dieser Abschlussarbeit ist herauszufinden, wie Mitglieder agiler Softwareentwicklungsteams sich die Zukunft ihrer Arbeit vorstellen. Dazu wurde eine quantitative Datenerhebung in Form einer webbasierten Befragung mittels eines standardisierten Fragebogens unter Mitgliedern agiler Softwareentwicklungsteams durchgeführt. Das Ergebnis ist, dass eine flexible Wahl des Arbeitsortes erwartet wird, wobei das Home Office zukünftig der präferierte Arbeitsort sein wird. Einen Einfluss durch agile Rahmenwerke und den dazugehörigen Praktiken auf die Wahl des Arbeitsortes konnte nicht festgestellt werden. Für die Zukunft wird keine signifikante Produktivitätssteigerung durch ein hybrides Arbeitsmodell erwartet.
Dramatic increases in the number of cyber security attacks and breaches toward businesses and organizations have been experienced in recent years. The negative impacts of these breaches not only cause the stealing and compromising of sensitive information, malfunctioning of network devices, disruption of everyday operations, financial damage to the attacked business or organization itself, but also may navigate to peer businesses/organizations in the same industry. Therefore, prevention and early detection of these attacks play a significant role in the continuity of operations in IT-dependent organizations. At the same time detection of various types of attacks has become extremely difficult as attacks get more sophisticated, distributed and enabled by Artificial Intelligence (AI). Detection and handling of these attacks require sophisticated intrusion detection systems which run on powerful hardware and are administered by highly experienced security staff. Yet, these resources are costly to employ, especially for small and medium-sized enterprises (SMEs). To address these issues, we developed an architecture -within the GLACIER project- that can be realized as an in-house operated Security Information Event Management (SIEM) system for SMEs. It is affordable for SMEs as it is solely based on free and open-source components and thus does not require any licensing fees. Moreover, it is a Self-Contained System (SCS) and does not require too much management effort. It requires short configuration and learning phases after which it can be self-contained as long as the monitored infrastructure is stable (apart from a reaction to the generated alerts which may be outsourced to a service provider in SMEs, if necessary). Another main benefit of this system is to supply data to advanced detection algorithms, such as multidimensional analysis algorithms, in addition to traditional SIEMspecific tasks like data collection, normalization, enrichment, and storage. It supports the application of novel methods to detect security-related anomalies. The most distinct feature of this system that differentiates it from similar solutions in the market is its user feedback capability. Detected anomalies are displayed in a Graphical User Interface (GUI) to the security staff who are allowed to give feedback for anomalies. Subsequently, this feedback is utilized to fine-tune the anomaly detection algorithm. In addition, this GUI also provides access to network actors for quick incident responses. The system in general is suitable for both Information Technology (IT) and Operational Technology (OT) environments, while the detection algorithm must be specifically trained for each of these environments individually.
Bis heute ist völlig unbekannt, ob wir allein im Universum sind. Um auf dieses Thema eine Antwort zu finden, überprüft diese Bachelorarbeit, ob Convolutional (CNN) und Recurrent Neural Networks (RNN) für die Erkennung außerirdischer Signale geeignet sind.
Das Ziel war dabei, in einem Datensatz bestehend aus Spektrogrammen mehr als 50% aller außerirdischer Signale zu erkennen, da nur so ein Neuronales Netzwerk ein besseres Resultat als eine zufällige Klassifikation liefert, bei der im Mittel 50% aller Signale erkannt werden.
Dabei zeigte sich, dass sich mit beiden Varianten der Neuronalen Netzwerke bis zu 90% aller Signale erkennen lassen, die Vorhersagen von CNNs allerdings verlässlicher sind. RNNs bieten hingegen aufgrund ihrer geringeren Größe einen deutlich leichtgewichtigeren Ansatz und führen zu einer signifikanten Speicherersparnis.
Daraus folgt, dass Neuronale Netzwerke bei der Suche nach außerirdischem Leben im Universum helfen können, um die Frage „Sind wir allein im Universum?“ endgültig zu beantworten.
Recent developments in the field of deep learning have shown promising advances for a wide range of historically difficult computer vision problems. Using advanced deep learning techniques, researchers manage to perform high-quality single-image super-resolution, i.e., increasing the resolution of a given image without major losses in image quality, usually encountered when using traditional approaches such as standard interpolation. This thesis examines the process of deep learning super-resolution using convolutional neural networks and investigates whether the same deep learning models can be used to increase OCR results for low-quality text images.
In zahlreichen Fachbüchern wird „Digital Leadership“ zur Unterstützung der digitalen Transformation propagiert. Im Rahmen eines Forschungsprojektes wurden die hierin vertretenen Führungsansätze mit aktuellen wissenschaftlichen Erkenntnissen abgeglichen. Fazit: Die Bücher stellen durchgängig altbewährte Führungskonzepte vor - neuere wissenschaftliche Erkenntnisse enthalten sie kaum.