Refine
Year of publication
Document Type
- Conference Proceeding (21)
- Article (9)
Has Fulltext
- yes (30)
Is part of the Bibliography
- no (30)
Keywords
- Mikroservice (8)
- Serviceorientierte Architektur (8)
- SOA (7)
- Insurance Industry (5)
- Versicherungswirtschaft (5)
- Cloud Computing (3)
- Microservices (3)
- OSGi (3)
- microservices (3)
- Big Data (2)
- CEP (2)
- CI/CD (2)
- Consistency (2)
- DevOps (2)
- Dienstgüte (2)
- ECA (2)
- Indicator Measurement (2)
- Microservice (2)
- Microservices Architecture (2)
- Rechnernetz (2)
- Service-orientation (2)
- Urban Logistics (2)
- Versicherungsbetrieb (2)
- XML-Model (2)
- XML-Schema (2)
- build automation (2)
- build server (2)
- API (1)
- Adaptive IT Infrastructure (1)
- Angewandte Informatik (1)
- Application Programming Interface (1)
- Arbeitsablauf (1)
- Auswahl (1)
- Authentication (1)
- Authentifikation (1)
- Authorization (1)
- Autorisierung (1)
- BaaS (Backend-as-a-service) (1)
- Batteriefahrzeug (1)
- Battery Electric Vehicles (1)
- Big Data Analytics (1)
- Choreography (1)
- City-Logistik (1)
- Complex Event Processing (CEP) (1)
- Computersicherheit (1)
- Computersimulation (1)
- Continuous Delivery (1)
- Curriculumentwicklung (1)
- Decision Support Tool (1)
- Distributed file systems (1)
- Docker (1)
- Domain Driven Design (DDD) (1)
- EPN (1)
- Education (1)
- Elektromobilität (1)
- Entscheidungsunterstützungssystem (1)
- Event Admin (EA) (1)
- Event Processing Network (1)
- Event Processing Network Model (1)
- Event monitoring (1)
- FaaS (Function-as-a-service) (1)
- Fault tolerance (1)
- Function as a Service (1)
- Hadoop (1)
- Intelligentes Stromnetz (1)
- Interdiziplinäre Studiengänge (1)
- Istio (1)
- Kontinuierliche Integration (1)
- Kubernetes (1)
- LightSabre (1)
- MapReduce (1)
- MapReduce algorithm (1)
- Mediendesign (1)
- Mediendesignausbildung (1)
- Mediendesigninformatik (1)
- Mikro-Kraft-Wärme-Kopplung (1)
- Nachhaltigkeit (1)
- NoSQL databases. (1)
- Ontologies (1)
- OpenStack (1)
- Orchestration (1)
- PageRank (1)
- Personennahverkehr (1)
- Portable Micro-CHP Unit (1)
- Praxisprojekte (1)
- Pregel (1)
- Programmierausbildung (1)
- Prüfstand (1)
- Quality of Service (1)
- Quality of Service (QoS) (1)
- REST <Informatik> (1)
- RESTful (1)
- Reference Architecture (1)
- Referenzmodell (1)
- Representational State Transfer (1)
- Resiliency (1)
- Richardson Maturity Model (1)
- RuleCore (1)
- SOA co-existence (1)
- Security (1)
- Serverless Computing (1)
- Service Mesh (1)
- Service Orientation (1)
- Shortest Path (1)
- Simulation (1)
- Smart Grid (1)
- Software Architecture (1)
- Softwarearchitektur (1)
- Softwarewerkzeug (1)
- Test Bench (1)
- Verteiltes System (1)
- Virtuelles Laboratorium (1)
- Web services (1)
- Word Counting (1)
- Workflow (1)
- XML (1)
- aerospace engineering (1)
- asynchronous messaging (1)
- cashing (1)
- cloud computing (1)
- complex event processing (CEP) (1)
- data mapping (1)
- distributed environments (1)
- distributed systems (1)
- e-mobility (1)
- eduDScloud (1)
- evaluation (1)
- events (1)
- generic interface (1)
- large scale systems (1)
- load balancing (1)
- matrix calulations (1)
- pmCHP (1)
- private cloud (1)
- reliable message delivery (1)
- scaling (1)
- serverless architecture (1)
- serverless functions (1)
- service models (1)
- service-orientation (1)
- shopping cart system (1)
- system integration (1)
- tool evaluation (1)
- virtual lab (1)
- web services (1)
- workload decomposition (1)
Heterogeneity has to be taken into account when integrating a set of existing information sources into a distributed information system that are nowadays often based on Service- Oriented Architectures (SOA). This is also particularly applicable to distributed services such as event monitoring, which are useful in the context of Event Driven Architectures (EDA) and Complex Event Processing (CEP). Web services deal with this heterogeneity at a technical level, also providing little support for event processing. Our central thesis is that such a fully generic solution cannot provide complete support for event monitoring; instead, source specific semantics such as certain event types or support for certain event monitoring techniques have to be taken into account. Our core result is the design of a configurable event monitoring (Web) service that allows us to trade genericity for the exploitation of source specific characteristics. It thus delivers results for the areas of SOA, Web services, CEP and EDA.
The automated transfer of flight logbook information from aircrafts into aircraft maintenance systems leads to reduced ground and maintenance time and is thus desirable from an economical point of view. Until recently, flight logbooks have not been managed electronically in aircrafts or at least the data transfer from aircraft to ground maintenance system has been executed manually. Latest aircraft types such as the Airbus A380 or the Boeing 787 do support an electronic logbook and thus make an automated transfer possible. A generic flight logbook transfer system must deal with different data formats on the input side – due to different aircraft makes and models – as well as different, distributed aircraft maintenance systems for different airlines as aircraft operators. This article contributes the concept and top level distributed system architecture of such a generic system for automated flight log data transfer. It has been developed within a joint industry and applied research project. The architecture has already been successfully evaluated in a prototypical implementation.
OSGi is a popular Java-based platform, which has its roots in the area of embedded systems. However, nowadays it is used more and more in enterprise systems. To fit this new application area, OSGi has recently been extended with the Remote Services specification. This specification enables distribution, which OSGi was previously lacking. However, the specification provides means for synchronous communication only and leaves out asynchronous communication. As an attempt to fill a gap in this field, we propose, implement and evaluate an approach for the integration of asynchronous messaging into OSGi.
In this paper various techniques in relation to large-scale systems are presented. At first, explanation of large-scale systems and differences from traditional systems are given. Next, possible specifications and requirements on hardware and software are listed. Finally, examples of large-scale systems are presented.
OSGi in Cloud Environments
(2013)
In this paper, five ontologies are described, which include the event concepts. The paper provides an overview and comparison of existing event models. The main criteria for comparison are that there should be possibilities to model events with stretch in the time and location and participation of objects; however, there are other factors that should be taken into account as well. The paper also shows an example of using ontologies in complex event processing.
This article discusses event monitoring options for heterogeneous event sources as they are given in nowadays heterogeneous distributed information systems. It follows the central assumption, that a fully generic event monitoring solution cannot provide complete support for event monitoring; instead, event source specific semantics such as certain event types or support for certain event monitoring techniques have to be taken into account. Following from this, the core result of the work presented here is the extension of a configurable event monitoring (Web) service for a variety of event sources. A service approach allows us to trade genericity for the exploitation of source specific characteristics. It thus delivers results for the areas of SOA, Web services, CEP and EDA.
During the transition from conventional towards purely electrical, sustainable mobility, transitional technologies play a major part in the task of increasing adaption rates and decreasing range anxiety. Developing new concepts to meet this challenge requires adaptive test benches, which can easily be modified e.g. when progressing from one stage of development to the next, but also meet certain sustainability demands themselves.
The system architecture presented in this paper is built around a service-oriented software layer, connecting a modular hardware layer for direct access to sensors and actuators to an extensible set of client tools. Providing flexibility, serviceability and ease of use, while maintaining a high level of reusability for its constituent components and providing features to reduce the required overall run time of the test benches, it can effectively decrease the CO2 emissions of the test bench while increasing its sustainability and efficiency.
The transfer of historically grown monolithic software architectures into modern service-oriented architectures creates a lot of loose coupling points. This can lead to an unforeseen system behavior and can significantly impede those continuous modernization processes, since it is not clear where bottlenecks in a system arise. It is therefore necessary to monitor such modernization processes with an adaptive monitoring concept to be able to correctly record and interpret unpredictable system dynamics. This contribution presents a generic QoS measurement framework for service-based systems. The framework consists of an XML-based specification for the measurement to be performed – the Information Model (IM) – and the QoS System, which provides an execution platform for the IM. The framework will be applied to a standard business process of the German insurance industry, and the concepts of the IM and their mapping to artifacts of the QoS System will be presented. Furtherm ore, design and implementation of the QoS System’s parser and generator module and the generated artifacts are explained in detail, e.g., event model, agents, measurement module and analyzer module.
In the context of modern mobility, topics such as smart-cities, Car2Car-Communication, extensive vehicle sensor-data, e-mobility and charging point management systems have to be considered. These topics of modern mobility often have in common that they are characterized by complex and extensive data situations. Vehicle position data, sensor data or vehicle communication data must be preprocessed, aggregated and analyzed. In many cases, the data is interdependent. For example, the vehicle position data of electric vehicles and surrounding charging points have a dependence on one another and characterize a competition situation between the vehicles. In the case of Car2Car-Communication, the positions of the vehicles must also be viewed in relation to each other. The data are dependent on each other and will influence the ability to establish a communication. This dependency can provoke very complex and large data situations, which can no longer be treated efficiently. With this work, a model is presented in order to be able to map such typical data situations with a strong dependency of the data among each other. Microservices can help reduce complexity.
The transfer of historically grown monolithic software architectures into modern service-oriented architectures creates a lot of loose coupling points. This can lead to an unforeseen system behavior and can significantly impede those continuous modernization processes, since it is not clear where bottlenecks in a system arise. It is therefore necessary to monitor such modernization processes with an adaptive monitoring concept in order to be able to correctly record and interpret unpredictable system dynamics. For this purpose, a general measurement methodology and a specific implementation concept are presented in this work.
Portable-micro-Combined-Heat-and-Power-units are a gateway technology bridging conventional vehicles and Battery Electric Vehicles (BEV). Being a new technology, new software has to be created that can be easily adapted to changing requirements. We propose and evaluate three different architectures based on three architectural paradigms. Using a scenario-based evaluation, we conclude that a Service-Oriented Architecture (SOA) using microservices provides a higher quality solution than a layered or Event-Driven Complex-Event-Processing (ED-CEP) approach. Future work will include implementation and simulation-driven evaluation.
Cloud computing has become well established in private and public sector projects over the past few years, opening ever new opportunities for research and development, but also for education. One of these opportunities presents itself in the form of dynamically deployable, virtual lab environments, granting educational institutions increased flexibility with the allocation of their computing resources. These fully sandboxed labs provide students with their own, internal network and full access to all machines within, granting them the flexibility necessary to gather hands-on experience with building heterogeneous microservice architectures. The eduDScloud provides a private cloud infrastructure to which labs like the microservice lab outlined in this paper can be flexibly deployed at a moment’s notice.
Hadoop is a Java-based open source programming framework, which supports the processing and storage of large volumes of data sets in a distributed computing environment. On the other hand, an overwhelming majority of organizations are moving their big data processing and storing to the cloud to take advantage of cost reduction – the cloud eliminates the need for investing heavily in infrastructures, which may or may not be used by organizations. This paper shows how organizations can alleviate some of the obstacles faced when trying to make Hadoop run in the cloud.
Nowadays, REST is the most dominant architectural style of choice at least for newly created web services. So called RESTfulness is thus really a catchword for web application, which aim to expose parts of their functionality as RESTful web services. But are those web services RESTful indeed? This paper examines the RESTfulness of ten popular RESTful APIs (including Twitter and PayPal). For this examination, the paper defines REST, its characteristics as well as its pros and cons. Furthermore, Richardson's Maturity Model is shown and utilized to analyse those selected APIs regarding their RESTfulness. As an example, a simple, RESTful web service is provided as well.
Our work is motivated primarily by the lack of standardization in the area of Event Processing Network (EPN) models. We identify general requirements for such models. These requirements encompass the possibility to describe events in the real world, to establish temporal and causal relationships among the events, to aggregate the events, to organize the events into a hierarchy, to categorize the events into simple or complex, to create an EPN model in an easy and simple way and to use that model ad hoc. As the major contribution, this paper applies the identified requirements to the RuleCore model.
The usage of microservices promises a lot of benefits concerning scalability and maintainability, rewriting large monoliths is however not always possible. Especially in scientific projects, pure microservice architectures are therefore not feasible in every project. We propose the utilization of microservice principles for the construction of microsimulations for urban transport. We present a prototypical architecture for the connection of MATSim and AnyLogic, two widely used simulation tools in the context of urban transport simulation. The proposed system combines the two tools into a singular tool supporting civil engineers in decision making on innovative urban transport concepts.
In microservice architectures, data is often hold redundantly to create an overall resilient system. Although the synchronization of this data proposes a significant challenge, not much research has been done on this topic yet. This paper shows four general approaches for assuring consistency among services and demonstrates how to identify the best solution for a given architecture. For this, a microservice architecture, which implements the functionality of a mainframe-based legacy system from the insurance industry, serves as an example.
Microservices is an architectural style for complex application systems, promising some crucial benefits, e.g. better maintainability, flexible scalability, and fault tolerance. For this reason microservices has attracted attention in the software development departments of different industry sectors, such as ecommerce and streaming services. On the other hand, businesses have to face great challenges, which hamper the adoption of the architectural style. For instance, data are often persisted redundantly to provide fault tolerance. But the synchronization of those data for the sake of consistency is a major challenge. Our paper presents a case study from the insurance industry which focusses consistency issues when migrating a monolithic core application towards microservices. Based on the Domain Driven Design (DDD) methodology, we derive bounded contexts and a set of microservices assigned to these contexts. We discuss four different approaches to ensure consistency and propose a best practice to identify the most appropriate approach for a given scenario. Design and implementation details and compliance issues are presented as well.