Informatik
Refine
Year of publication
- 2018 (83) (remove)
Document Type
- Conference proceeding (59)
- Journal article (16)
- Book chapter (6)
- Doctoral Thesis (2)
Is part of the Bibliography
- yes (83)
Institute
- Informatik (83)
Publisher
- Springer (12)
- Hochschule Reutlingen (11)
- IEEE (11)
- RWTH Aachen (8)
- Gesellschaft für Informatik (3)
- Springer International Publishing (3)
- ACM (2)
- Deutsche Gesellschaft für Computer- und Roboterassistierte Chirurgie e.V. (2)
- Elsevier (2)
- HTWG Konstanz (2)
- Haufe (2)
- SCITEPRESS (2)
- Springer Science + Business Media B.V (2)
- ARVO (1)
- American Marketing Association (1)
- Association for Information Systems (AIS) (1)
- BioMed Central (1)
- De Gruyter (1)
- Emerald (1)
- Faculty of Organization and Informatics, University of Zagreb (1)
- IGI Global (1)
- Lund University (1)
- Riga Technical University Press (1)
- SPIE (1)
- Sage (1)
- Smart Home & Living Baden-Württemberg e.V. (1)
- Springer Gabler (1)
- The Association for Computing Machinery (1)
- Universidad Carlos III de Madrid (1)
- University of Konstanz, University Library (1)
- University of Portsmouth (1)
The basic idea behind a wearable robotic grasp assistancesystem is to support people that suffer from severe motor impairments in daily activities. Such a system needs to act mostly autonomously and according to the user’s intent. Vision-based hand pose estimation could be an integral part of a larger control and assistance framework. In this paper we evaluate the performance of egocentric monocular hand pose estimation for a robot-controlled hand exoskeleton in a simulation. For hand pose estimation we adopt a Convolutional Neural Network (CNN). We train and evaluate this network with computer graphics, created by our own data generator. In order to guide further design decisions we focus in our experiments on two egocentric camera viewpoints tested on synthetic data with the help of a 3D-scanned hand model, with and without an exoskeleton attached to it.We observe that hand pose estimation with a wrist-mounted camera performs more accurate than with a head-mounted camera in the context of our simulation. Further, a grasp assistance system attached to the hand alters visual appearance and can improve hand pose estimation. Our experiment provides useful insights for the integration of sensors into a context sensitive analysis framework for intelligent assistance.
The use of additive manufacturing technologies for industrial production is constantly growing. This technology differs from the known production proecdures. The areas for scheduling, detailed and sequence planning are particularly important for additive production due to the long print times and flexible use of the production area. Therefore, production-relevant variables are considered and used for the production planning and control (PPC) of additive manufacturing machines. For this purpose, an optimization model is presented which shows a time-oriented build space utilization. In the implementation, a nesting algorithm is used to check the combinability of different models for each individual print job.
The blockchain technology enables a common data basis between the participants. Entries are logged and the authenticity of the participants is guaranteed. In the case of a relationship between customers and producers, this would lead to verifiable cooperation, which would be a major step as companies enter into service contracts based on the flow of many small transactions through communication. This paper proposes an architecture that enables the creation and processing of orders between the customer and producers via a blockchain based production network. The handling of larger files which are traceable via the blockchain is also shown and the use of a public or permissioned blockchain for an application case is also considered.
This work is a report on practical experiences with the issue of interoperability in German practice management systems (PMSs) from an ongoing clinical trial on teledermatology, the TeleDerm project. A proprietary and established web-platform for store-and-forward telemedicine is integrated with the IT in the GPs’ offices for automatic exchange of basic patient data. Most of the 19 different PMSs included in the study sample lack support of modern health data exchange standards, therefore the relatively old but widely available German health data exchange interface “Gerätedatentransfer” (GDT) is used. Due to the lack of enforcement and regulation of the GDT standard, several obstacles to interoperability are encountered. As a partial, but reusable working solution to cope with these issues, we present a custom middleware which is used in conjunction with GDT. We describe the design, technical implementation and observed hindrances with the existing infrastructure. A discussion on health care interfacing standards and the current state of interoperability in German PMS software is given.
While there are several theoretical comparisons of Object Orientation (OO) and Service Orientation (SO), little empirical research on the maintainability of the two paradigms exists. To provide support for a generalizable comparison, we conducted a study with four related parts. Two functionally equivalent systems (one OO and one SO version) were analyzed with coupling and cohesion metrics as well as via a controlled experiment, where participants had to extend the systems. We also conducted a survey with 32 software professionals and interviewed 8 industry experts on the topic. Results indicate that the SO version of our system possesses a higher degree of cohesion, a lower degree of coupling, and could be extended faster. Survey and interview results suggest that industry sees systems built with SO as more loosely coupled, modifiable, and reusable. OO systems, however, were described as less complex and easier to test.
Maintainability assurance techniques are used to control this quality attribute and limit the accumulation of potentially unknown technical debt. Since the industry state of practice and especially the handling of service- and microservice-based systems in this regard are not well covered in scientific literature, we created a survey to gather evidence for a) used processes, tools, and metrics in the industry, b) maintainability-related treatment of systems based on service orientation, and c) influences on developer satisfaction w.r.t. maintainability. 60 software professionals responded to our online questionnaire. The results indicate that using explicit and systematic techniques has benefits for maintainability. The more sophisticated the applied methods the more satisfied participants were with the maintainability of their software while no link to a hindrance in productivity could be established. Other important findings were the absence of architecture-level evolvability control mechanisms as well as a significant neglect of service-oriented particularities for quality assurance. The results suggest that industry has to improve its quality control in these regards to avoid problems with long living service-based software systems.
To bring a pattern-based perspective to the SOA vs. microservices discussion, we qualitatively analyzed a total of 118 SOA patterns from 2 popular catalogs for their (partial) applicability to microservices. Patterns had to hold up to 5 derived microservices principles to be applicable. 74 patterns (63%) were categorized as fully applicable, 30 (25%) as partially applicable, and 14 (12%) as not applicable. Most frequently violated microservices characteristics werde Decentralization and Single System. The findings suggest that microservices and SOA share a large set of architectural principles and solutions in the general space of service-based systems while only having a small set of differences in specific areas.
Lots of movies are produced every year, too many to watch all of them and in particular, to get an overview about the evolution of typical movie genres and actors playing in them. Moreover, it is a challenging problem to detect correlations among the movies and the actors in those movies, in particular, if we are interested in time-varying data patterns like trends, countertrends, or anomalies and outliers. Those correlations are specifically interesting if they can be inspected on different levels of granularity, e.g., temporal, but also hierarchical in form of country- or continent-based correlations. In this paper we describe the IMDb Explorer, a webbased visualization tool that consists of two major views denoted by the movie cosmos and the career lines. Both views are linked and interactively manipulable while a list of user-defined metrics are explorable. We illustrate the usefulness of the visualization tool by applying it to the entire movie database provided by IMDb.
Perceptual integration of kinematic components in the recognition of emotional facial expressions
(2018)
According to a long-standing hypothesis in motor control, complex body motion is organized in terms of movement primitives, reducing massively the dimensionality of the underlying control problems. For body movements, this low dimensional organization has been convincingly demonstrated by the learning of low-dimensional representations from kinematic and EMG data. In contrast, the effective dimensionality of dynamic facial expressions is unknown, and dominant analysis approaches have been based on heuristically defined facial ‘‘action units,’’ which reflect contributions of individual face muscles. We determined the effective dimensionality of dynamic facial expressions by learning of a low dimensional model from 11 facial expressions. We found an amazingly low dimensionality with only two movement primitives being sufficient to simulate these dynamic expressions with high accuracy. This low dimensionality is confirmed statistically, by Bayesian model comparison of models with different numbers of primitives, and by a psychophysical experiment that demonstrates that expressions, simulated with only two primitives, are indistinguishable from natural ones.
In addition, we find statistically optimal integration of the emotion information specified by these primitives in visual perception. Taken together, our results indicate that facial expressions might be controlled by a very small number of independent control units, permitting very low dimensional parametrization of the associated facial expression.
Die Bedeutung und Stellung der Informationstechnologie erlebte in den letzten 60 Jahren einen fortlaufenden Wandel. Der anfänglich rein unterstützende Charakter entwickelte sich immer mehr zu einem wichtigen Bestandteil der Aufbau- und Ablauforganisation im Unternehmen. Ein definiertes IT-Servicemanagement im Unternehmen sieht sich mittlerweile gleichgeordnet mit den restlichen Fachabteilungen, tritt mit seinen Leistungen als Dienstleister auf und betrachtet die Fachabteilungen als „Kunde“. Neue Technologien und Innovationen und die daraus resultierenden Neudefinitionen bestehender Anforderungen sollen im Rahmen der Digitalisierung in Unternehmen positive Effekte zeigen. IT Infrastructure Library (ITIL) wird als Framework für IT-Servicemanagement in der Industrie und im öffentlichen Dienst genutzt. Der Ansatz von ITIL unterstützte den Kulturwandel und sensibilisierte das Management und die Mitarbeiter darin, serviceorientiert zu denken. Da dieser Ansatz einen vordefinierten, zyklischen Ablauf hat, könnten schnell eintreffende Kundenanforderungen nicht fristgerecht umgesetzt werden, weshalb agile Methoden wie der DevOps-Ansatz in den Vordergrund rücken. Die Herausforderung besteht darin, den Kulturwandel bei der Einführung von DevOps in bestehenden ITIL-Strukturen in Unternehmen zu fördern.
Motor-based theories of facial expression recognition propose that the visual perception of facial expression is aided by sensorimotor processes that are also used for the production of the same expression. Accordingly, sensorimotor and visual processes should provide congruent emotional information about a facial expression. Here, we report evidence that challenges this view. Specifically, the repeated execution of facial expressions has the opposite effect on the recognition of a subsequent facial expression than the repeated viewing of facial expressions. Moreover, the findings of the motor condition, but not of the visual condition, were correlated with a nonsensory condition in which participants imagined an emotional situation. These results can be well accounted for by the idea that facial expression recognition is not always mediated by motor processes but can also be recognized on visual information alone.
Free-floating e-scooter sharing is an upcoming trend in mobility, which has been spreading since 2015 in various German cities. Unlike the more scientifically explorend car sharing, the usage patterns and behaviors of e-scooter sharing customers are yet to be analyzed. This presumably discovers better ways to attract customers as well as adaptions of the business model in order to increase scooter utilization and therefore the profit of the e-scooter providers. As most of the customer's journey, from registration to scooter reservation and the ride itself, is digitally traceable, large datasets are available allowing for understanding of customers' needs and motivations. Based on these datasets of an e-scooter provider operating in a big German city we propose a customer clustering that identifies four different customer segments, which enables multiple conclusions to be drawn for business development and improving the problem-solution fit of the e-scooter sharing model.
Software engineering education is supposed to provide students with industry-relevant knowledge and skills. Educators must address issues beyond exercises and theories that can be directly rehearsed in small settings. A way to experience such effects and to increase the relevance of software engineering education is to apply empirical studies in teaching. In our article, we show how different types of empirical studies can be used for educational purposes in software engineering. We give examples illustrating how to utilize empirical studies, discuss challenges, and derive an initial guideline that supports teachers to include empirical studies in software engineering courses.
Empirical software engineering experts on the use of students and professionals in experiments
(2018)
Using students as participants remains a valid simplification of reality needed in laboratory contexts. It is an effective way to advance software engineering theories and technologies but, like any other aspect of study settings, should be carefully considered during the design, execution, interpretation, and reporting of an experiment. The key is to understand which developer population portion is being represented by the participants in an experiment. Thus, a proposal for describing experimental participants is put forward.
The relative pros and cons of using students or practitioners in experiments in empirical software engineering have been discussed for a long time and continue to be an important topic. Following the recent publication of “Empirical software engineering experts on the use of students and professionals in experiments” by Falessi, Juristo, Wohlin, Turhan, Münch, Jedlitschka, and Oivo (EMSE, February 2018) we received a commentary by Sjøberg and Bergersen. Given that the topic is of great methodological interest to the community and requires nuanced treatment, we invited two editorial board members, Martin Shepperd and Per Runeson, respectively, to provide additional views.
Sleep study can be used for detection of sleep quality and in general bed behaviors. These results can helpful for regulating sleep and recognizing different sleeping disorders of human. In comparison to the leading standard measuring system, which is Polysomnography (PSG), the system proposed in this work is a non-invasive sleep monitoring device. For continuous analysis or home use, the PSG or wearable Actigraphy devices tends to be uncomfortable. Besides, these methods not only decrease practicality due to the process of having to put them on, but they are also very expensive. The system proposed in this paper classifies respiration and body movement with only one type of sensor and also in a noninvasive way. The sensor used is a pressure sensor. This sensor is low cost and can be used for commercial proposes. The system was tested by carrying out an experiment that recorded the sleep process of a subject. These recordings showed excellent results in the classification of breathing rate and body movements.
Eines der gängigsten bildgebenden Verfahren in der Medizin ist die Sonografie. Jedoch ist die Reproduzierbarkeit der Ultraschalldiagnostik bis heute noch immer ein Problem, wodurch Fehldiagnosen gestellt werden. Durch das in diesem Papier vorgestellte prototypische System zur Unterstützung für Medizinstudenten in Ultraschallseminaren sollen Anforderungen zur Reproduzierbarkeit einer Ultraschalluntersuchung definiert werden. Durch Experteninterviews wurden Einblicke in die klinischen Abläufe und den Krankenhaus-Alltag gewonnen, welche Inhalte relevant sind, um die Reproduzierbarkeit von Ultraschalluntersuchungen zu ermöglichen.
We present an approach for segmenting individual cells and lamellipodia in epithelial cell clusters using fully convolutional neural networks. The method will set the basis for measuring cell cluster dynamics and expansion to improve the investigation of collective cell migration phenomena. The fully learning-based front-end avoids classical feature engineering, yet the network architecture needs to be designed carefully. Our network predicts how likely each pixel belongs to one of the classes and, thus, is able to segment the image. Besides characterizing segmentation performance, we discuss how the network will be further employed.
As production workspaces become more mobile and dynamic it becomes increasingly important to reliably monitor the overall state of the environment. Therein manipulators or other robotic systems likely have to be able to act autonomously together with humans and other systems within a joint workspace. Such interactions require that all components in non-stationary environments are able to perceive the state relative to each other. As vision-sensors provide a rich source of information to accomplish this, we present RoPose, a convolutional neural network (CNN) based approach, to estimate the two dimensional joint configuration of a simulated industrial manipulator from a camera image. This pose information can further be used by a novel targetless calibration setup to estimate the pose of the camera relative to the manipulator’s space. We present a pipeline to automatically generate synthetic training data and conclude with a discussion of the potential usage of the same pipeline to acquire real image datasets of physically existent robots.
Creating new business models, products or services is challenging in fast changing unpredictable environments. Often, product teams need to make many assumptions (e.g., assumptions about future demands) that might not be true. These assumptions impose risks to the success and these risks need to be mitigated early. One of the principles of the Lean Startup approach is to identify and prioritize the riskiest assumptions in order to validate them as early as possible. This helps to avoid wasting effort and time. In the literature there are several different methods for identifying and prioritizing the riskiest assumptions reported. However, only little research exists about the practical application of these methods in practice and how to teach them. In this paper, we present and empirically analyze a workshop format that we have developed for teaching the prioritization of Lean Startup assumptions. We aim at raising the awareness for assumption thinking among the participants and teach them through group work how to prioritize assumptions. The results of the analysis of a multitude of conducted workshops show that the applied method did lead to reasonable results and accompanying learning effects. In addition, the participants got aware of assumption thinking and liked learning in a practical way.
Early reduction of risks in a startup or an innovation project is highly important. Appropriate means for risk reduction, such as testing business models with different kinds of experiments exist. However, deciding what to test and how to select the right test, is challenging for many startups and innovation projects. This article presents the so-called Business Experiments Navigator (BEN), a toolkit to assist startup and innovation processes. It compliments other tools such as the Business Model Canvas or the Lean Startup process. The main contribution of BEN is to bridge the gap between the riskiest assumptions of a business model and the multitude of available testing techniques by providing assumption templates. The Business Experiments Navigator has been validated in several workshops. Results show that it creates awareness among the workshop participants that a business model is based on assumptions which impose risks and need to be validated. Further, users of BEN were able to identify relevant assumptions and map different kinds of assumptions to appropriate testing techniques. The process applied in the workshops, as well as the assumption templates, helped the participants understand the main concepts and transfer their learnings, to their own business ideas.
Die Segmentierung und das Tracking von minimal-invasiven robotergeführten Instrumenten ist ein wesentlicher Bestandteil für verschiedene computer assistierte Eingriffe. Allerdings treten in der minimal-invasiven Chirurgie, die das Anwendungsfeld für den hier beschriebenen Ansatz darstellt, häufig Schwierigkeiten durch Reflexionen, Schatten oder visuelle Verdeckungen durch Rauch und Organe auf und erschweren die Segmentierung und das Tracking der Instrumente.
Dieser Beitrag stellt einen Deep Learning Ansatz für ein markerloses Tracking von minimal-invasiven Instrumenten vor und wird sowohl auf simulierten als auch realen Daten getestet. Es wird ein simulierter als auch realer Datensatz mit Ground Truth Kennzeichnung für die binäre Segmentierung von Instrument und Hintergrund erstellt. Für den simulierten Datensatz werden Bilder aus einem simulierten Instrument und realem Hintergrund zusammengesetzt. Im Falle des realen Datensatzes spricht man von der Zusammensetzung der Bilder aus einem realen Instrument und Hintergrund. Insgesamt wird auf den simulierten Daten eine Pixelgenauigkeit von 94.70 Prozent und auf den realen Daten eine Pixelgenauigkeit von 87.30 Prozent erreicht.
The need for creating digitally enhanced products, services, and experiences as well as the emergence of new or modified business models has a significant impact on the automotive domain. Innovative solutions and new topics such as Smart Mobility or Connectivity require current automotive development processes to undergo major changes. They need to be redesigned in a way that it is possible to learn and adapt continuously at a fast pace. Agile methods are promising approaches to address these new challenges. However, agile methods are not tailored to the specific characteristics of the automotive domain such as software product line (SPLs) development. Although, there have been efforts to apply agile methods in the automotive domain, widespread adoptions have not yet taken place.
Back to the future: origins and directions of the “Agile Manifesto” – views of the originators
(2018)
In 2001, seventeen professionals set up the manifesto for agile software development. They wanted to define values and basic principles for better software development. On top of brought into focus, the manifesto has been widely adopted by developers, in software-developing organizations and outside the world of IT. Agile principles and their implementation in practice have paved the way for radical new and innovative ways of software and product development. In parallel, the understanding of the manifesto’s underlying principles evolved over time. This, in turn, may affect current and future applications of agile principles. This article presents results from a survey and an interview study in collaboration with the original contributors of the manifesto for agile software development. Furthermore, it comprises the results from a workshop with one of the original authors. This publication focuses on the origins of the manifesto, the contributors’ views from today’s perspective, and their outlook on future directions. We evaluated 11 responses from the survey and 14 interviews to understand the viewpoint of the contributors. They emphasize that agile methods need to be carefully selected and agile should not be seen as a silver bullet. They underline the importance of considering the variety of different practices and methods that had an influence on the manifesto. Furthermore, they mention that people should question their current understanding of "agile" and recommend reconsidering the core ideas of the manifesto.
An assessment model to foster the adoption of agile software product lines in the automotive domain
(2018)
A software product line is commonly used for the software development in large automotive organizations. A strategic reuse of software is needed to handle the increasing complexity of the development and to maintain the quality of numerous software variants. However, the development process needs to be continuously adapted at a fast pace to satisfy the changing market demands. Introducing agile software development methods promise the flexibility to react on customers’ change requests and market demands to deliver high quality software. Despite this need, it is still challenging to combine agile software development and product lines. The maturity of an agile adoption is often hard to determine. Assessing the current situation regarding the combination is a first step towards a successful inclusion of agile methods into automotive software product lines. Based on an interview study with 16 participants and a literature review, we build the so-called ASPLA Model allowing self-assessments within the team to determine the current state of agile software development in combination with software product lines. The model comprises seven areas of improvement and recommends a possibility to improve the current status.
Combining agile development and software product lines in automotive: challenges and recommendations
(2018)
Software product lines (SPLs) are used throughout the automotive industry. SPLs help to manage the large number of variants and to improve quality by reuse. In order to develop high quality software faster, agile software development (ASD) practices are introduced. From both the research and the management point of view it is still not clear how these two approaches can be combined. We derive recommendations to combine ASD and SPLs based on challenges identified for an automotive specific model. This study combines the outcome of a literature review and a qualitative interview study with 16 practitioners from the automotive domain. We evaluate the results and analyze the relationship between ASD and SPLs in the automotive domain. Furthermore, we derive recommendations to combine ASD and SPLs based on challenges identified in the automotive domain. This study identifies 86 individual challenges. Important challenges address supplier collaboration and faster software release cycles without loss of quality. The identified challenges and the derived recommendations show that the combination of ASD and SPL in the automotive industry is promising but not trivial. There is a need for an automotive-specific approach that combines ASD and SPL.
Context: Software product lines are widely used in automotive embedded software development. This software paradigm improves the quality of software variants by reuse. The combination of agile software development practices with software product lines promises a faster delivery of high quality software. However, the set up of an agile software product line is still challenging, especially in the automotive domain. Goal: This publication aims to evaluate to what extend agility fits to automotive product line engineering. Method: Based on previous work and two workshops, agility is mapped to software product line concerns. Results: This publication presents important principles of software product lines, and examines how agile approaches fit to those principles. Additionally, the principles are related to one of the four major concerns of software product line engineering: Business, Architecture, Process, and Organization. Conclusion: Agile software product line engineering is promising and can add value to existing development approaches. The identified commonalities and hindering factors need to be considered when defining a combined agile product line engineering approach.
Modeling interactive Enterprise Architecture visualizations: an extended architecture description
(2018)
Enterprise architectures consist of a multitude of architecture elements, which relate in manifold ways to each other. Due to the high number of relationships between these elements, architectural analysis mechanisms are essential for all stakeholders to keep track and to work out relevant model characteristics. In practice EAs are often analyzed using visualizations by hand. However, the visualizations are often static and there are only few interaction possibilities. As a result, new visualizations have to be created or configured by experts if information demands change. In addition, hardly any tools are used for analysis of complex model characteristics. In this article we introduce an extended conceptualization of the architecture description that defines the structure of interactive visualizations and the integration of further tools to flexibly respond to the information demands of stakeholders. In addition, we develop a so-called Architecture Cockpit that realizes the extended conceptualization in a prototype. At the end we demonstrate and evaluate our approach through a practical test in a company in the finance and insurance industry.
Unternehmen stehen aktuell aufgrund der Digitalisierung, des stetigen technologischen Fortschritts und immer kürzer werdenden Produktlebenszyklen vor großen Herausforderungen. Um am Markt bestehen zu können, müssen Geschäftsmodelle öfter und schneller an sich verändernde Marktverhältnisse angepasst werden als dies früher der Fall war. Eine schnelle Anpassungsfähigkeit, auch Agilität genannt, ist in der heutigen Zeit ein entscheidender Wettbewerbsfaktor. Aufgrund des stetig wachsenden IT Anteils in Produkten sowie der Tatsache, dass diese IT-gestützt hergestellt werden, haben Änderungen des Geschäftsmodells große Auswirkungen auf die Unternehmensarchitektur eines Unternehmens.
Eine Unternehmensarchitektur umspannt das Unternehmen, indem diese die fachlichen und technischen Strukturen, insbesondere die gesamte IT, des Unternehmens beinhaltet und integriert. Das Management der Unternehmensarchitektur ist die Disziplin zur Beherrschung und Abstimmung dieser Strukturen. An der Gestaltung der Unternehmensarchitektur wirken viele Stakeholder mit individuellen und teils gegensätzlichen Interessen aus den unterschiedlichsten Bereichen des Unternehmens mit. Dies macht die Entscheidungsfindung zu einer komplexen Aufgabe.
Die in dieser Arbeit entworfene integrative Methode für die Entscheidungsfindung hat das Ziel, die Betroffenen und Beteiligten, im Folgenden Stakeholder, bei ihren Entscheidungen zu unterstützen. Die Grundidee hierbei ist die systematische Einbeziehung der Interessen der Stakeholder und davon abgeleiteter Visualisierungen. Dies verleiht der Methode ihren integrativen Charakter und hilft Abhängigkeiten zwischen Stakeholdern zu erkennen. Dadurch wird die Zusammenarbeit zwischen den an Entscheidungen beteiligten Stakeholdern gefördert. Neben der systematischen Einbeziehung von Visualisierungen wird im Rahmen dieser Arbeit das Konzept der Technik eingeführt. Techniken werden ebenfalls von den Interessen der Stakeholder abgeleitet und dienen der Unterstützung bei der Durchführung von Aktivitäten der Entscheidungsfindung, indem Vorgehensweisen bei bestimmten Aufgaben vorgegeben oder Teilprozesse der Entscheidungsfindung sogar automatisiert durchgeführt werden. Das Konzept der Technik, die systematische Ableitung von den Interessen der Stakeholder sowie das Zusammenspiel mit Visualisierungen wird in dieser Arbeit in Form einer erweiterten Konzeptualisierung der Architekturbeschreibung definiert.
Da die Werkzeugunterstützung in der Praxis häufig eine Herausforderung darstellt, rundet diese Arbeit ein eigens konzipiertes und prototypisch validiertes Architekturcockpit ab. Das Cockpit ist eine auf einem elektronischen Sitzungsraum basierende Werkzeugunterstützung der eingeführten integrativen Methode.
Enterprise Governance, Risk and Compliance (GRC) systems are key to managing risks threatening modern enterprises from many different angles. Key constituent to GRC systems is the definition of controls that are implemented on the different layers of an Enterprise Architecture (EA). As part of the compliance aspect of GRC, the effectiveness of these controls is assessed and reported to relevant management bodies within the enterprise. In this paper we present a metamodel which links controls to the affected elements of an EA and supplies a way of expressing associated assessment techniques and results. We complement the metamodel with an expository instantiation in a cockpit for control compliance applied in an international enterprise in the insurance industry.
The focus of the developed maturity model was set on processes. The concept of the widespread CMM and its practices has been transferred to the perioperative domain and the concept of the new maturity model. Additional optimization goals and technological as well as networking-specific aspects enable a process- and object-focused view of the maturity model in order to ensure broad coverage of different subareas. The evaluation showed that the model is applicable to the perioperative field. Adjustments and extensions of the maturity model are future steps to improve the rating and classification of the new maturity model.
Container virtualization evolved into a key technology for deployment automation in line with the DevOps paradigm. Whereas container management systems facilitate the deployment of cloud applications by employing container based artifacts, parts of the deployment logic have been applied before to build these artifacts. Current approaches do not integrate these two deployment phases in a comprehensive manner. Limited knowledge on application software and middleware encapsulated in container-based artifacts leads to maintainability and configuration issues. Besides, the deployment of cloud applications is based on custom orchestration solutions leading to lock in problems. In this paper, we propose a two-phase deployment method based on the TOSCA standard. We present integration concepts for TOSCA-based orchestration and deployment automation using container-based artifacts. Our two-phase deployment method enables capturing and aligning all the deployment logic related to a software release leading to better maintainability. Furthermore, we build a container management system, which is composed of a TOSCA-based orchestrator on Apache Mesos, to deploy container-based cloud applications automatically.
The state of the art proposes the microservices architectural style to build applications. Additionally, container virtualization and container management systems evolved into the perfect fit for developing, deploying, and operating microservices in line with the DevOps paradigm. Container virtualization facilitates deployment by ensuring independence from the runtime environment. However, microservices store their configuration in the environment. Therefore, software developers have to wire their microservice implementation with technologies provided by the target runtime environment such as configuration stores and service registries. These technological dependencies counteract the portability benefit of using container virtualization. In this paper, we present AUTOGENIC - a model-based approach to assist software developers in building microservices as self configuring containers without being bound to operational technologies. We provide developers with a simple configuration model to specify configuration operations of containers and automatically generate a self-configuring microservice tailored for the targeted runtime environment. Our approach is supported by a method, which describes the steps to automate the generation of self-configuring microservices. Additionally, we present and evaluate a prototype, which leverages the emerging TOSCA standard.
With the digital transformation being one of the most discussed topics in the business world today, many enterprises – especially small and medium sized ones – find themselves struggling with the understanding of new digital technologies and thus the potential benefits and risks for their companies. New technologies like the Internet of Things, Blockchain or Machine Learning have great potential for businesses. However, carefully evaluating and selecting purposeful technologies – aligned to the digital strategy – is the key to success. Technologies appear, change and also vanish so rapidly in the digital age, that a proper understanding is crucial for a sustainable technological foundation. Focusing on the characteristic features of technologies, the presented approach promises to create a better technological understanding for decision makers in small and medium-sized enterprises (SMEs) in a playful manner: With a serious game that fosters insight and allays fears of digitalization.
Background: Internationally, teledermatology has proven to be a viable alternative to conventional physical referrals. Travel cost and referral times are reduced while patient safety is preserved. Especially patients from rural areas benefit from this healthcare innovation. Despite these established facts and positive experiences from EU neighboring countries like the Netherlands or the United Kingdom, Germany has not yet implemented store-and-forward teledermatology in routine care.
Methods: The TeleDerm study will implement and evaluate store-and-forward teledermatology in 50 general practitioner (GP) practices as an alternative to conventional referrals. TeleDerm aims to confirm that the possibility of store-and-forward teledermatology in GP practices is going to lead to a 15% (n = 260) reduction in referrals in the intervention arm. The study uses a cluster-randomized controlled trial design. Randomization is planned for the cluster “county”. The main observational unit is the GP practice. Poisson distribution of referrals is assumed. The evaluation of secondary outcomes like acceptance, enablers and barriers uses a mixed methods design with questionnaires and interviews.
Discussion: Due to the heterogeneity of GP practice organization, patient management software, information technology service providers, GP personal technical affinity and training, we expect several challenges in implementing teledermatology in German GP routine care. Therefore, we plan to recruit 30% more GPs than required by the power calculation. The implementation design and accompanying evaluation is expected to deliver vital insights into the specifics of implementing telemedicine in German routine care.
Decentralized energy systems are characterized by an ad hoc planing. The missing integration of energy objectives into business strategy creates difficulties resulting in inefficient energy architectures and decisions. Practice-proven methods such as balanced scorecard, enterprise architecture management and value network approach supports the transformation path towards an effective decentralized system. The methods are evaluated based on a case study. Managing multi-dimensionality, high complexity and multiple actors are the main drivers for an effective and efficient energy management system. The underlying basis to gain the positive impacts of these methods on decentralized corporate energy systems is digitization of energy data and processes.
Software and system development faces numerous challenges of rapidly changing markets. To address such challenges, companies and projects design and adopt specific development approaches by combining well-structured methods and flexible agile practices. Yet, the number of methods and practices is large and the actual process composition is often carried out in an ad-hoc manner. This paper reports on a survey on hybrid software development approaches. We study which approaches are used in practice, how different approaches are combined, and what contextual factors influence the use and combination of hybrid software development approaches.
Software engineering courses have to deliver theoretical and technical knowledge and skills while establishing links to practice. However, due to course goals or resource limitations, it is not always possible or even meaningful to set up complete projects and let students work on a real piece of software. For instance, if students shall understand the impact of group dynamics on productivity, a particular software to be developed is of less interest than an environment in which students can learn about team-related phenomena. To address this issue, we use experimentation as a teaching tool in software engineering courses. Experiments help to precisely characterize and study a problem in a systematic way, to observe phenomena, and to develop and evaluate solutions. Furthermore, experiments help establishing short feedback and learning cycles, and they also allow for experiencing risk and failure scenarios in a controlled environment. In this paper, we report on three courses in which we implemented different experiments and we share our experiences and lessons learned. Using these courses, we demonstrate how to use classroom experiments, and we provide a discussion on the feasibility based on formal and informal course evaluations. This experience report thus aims to help teachers integrating small- and medium sized experiments in their courses.
Rapid prototyping platforms reduce development time by allowing quick prototyping of a prototype idea and achieve more time for actual application development with user interfaces. This approach has long been followed in technical platforms, such as the Arduino. To transfer this form of prototyping to wearables, WearIT is presented in this paper.WearIT consists of four components as a wearable prototyping platform: (1) a vest, (2) sensor and actuator shields, (3) its own library and (4) a motherboard consisting of Arduino, Raspberry Pi, a board and a GPS module. As a result, a wearable prototype can be quickly developed by attaching sensor and actuator shields to the WearIT vest. These sensor and actuator shields can then be programmed through the WearIT library. Via Virtual Network Computing (VNC) with a remote computer, the screen contents of the Raspberry Pi can be accessed and the Arduino be programmed.
Engineers of the research project “Digital Product Life-Cycle” are using a graph-based design language to model all aspects of the product they are working on. This abstract model is the base for all further investigations, developments and implementations. In particular at early stages of development, collaborative decision making is very important. We propose a semantic augmented knowledge space by means of mixed reality technology, to support engineering teams. Therefore we present an interaction prototype consisting of a pico projector and a camera. In our usage scenario engineers are augmenting different artefacts in a virtual working environment. The concept of our prototype contains both an interaction and a technical concept. To realise implicit and natural interactions, we conducted two prototype tests: (1) A test with a low-fidelity prototype and (2) a test by using the method Wizard of Oz. As a result, we present a prototype with interaction selection using augmentation spotlighting and an interaction zoom as a semantic zoom.
Im Rahmen dieser Arbeit wurde eine Software-Architektur entwickelt, mit der sich Interaktionen zwischen autonomen Fahrzeugen und Passanten im Straßenverkehr in einer simulierten Umgebung untersuchen lassen. Hierbei wird das autonome Fahrzeug durch einen externen Fahrsimulator gesteuert. Der Einsatz eines Motion-Capture-Systems ermöglicht dabei die Aufzeichnung und Übertragungen der Bewegungsdaten von Passant und Fahrer in die virtuelle Umgebung. Durch den Einsatz von head-mounted Displays sollen Akteure die virtuelle Umgebung möglichst als real empfinden. Auf Basis der entwickelten Software-Architektur wurde eine Simulationsumgebung realisiert, in der Interaktionen zwischen einem Passant und einem autonomem Fahrzeug untersucht werden können. Das Projekt soll das Potential von Motion-Capture gestützten Simulationsumgebungen für die Konzeption und Entwicklung von autonomen Fahrsystemen aufzeigen.
Recognizing actions of humans, reliably inferring their meaning and being able to potentially exchange mutual social information are core challenges for autonomous systems when they directly share the same space with humans. Today’s technical perception solutions have been developed and tested mostly on standard vision benchmark datasets where manual labeling of sensory ground truth is a tedious but necessary task. Furthermore, rarely occurring human activities are underrepresented in such data leading to algorithms not recognizing such activities. For this purpose, we introduce a modular simulation framework which offers to train and validate algorithms on various environmental conditions. For this paper we created a dataset, containing rare human activities in urban areas, on which a current state of the art algorithm for pose estimation fails and demonstrate how to train such rare poses with simulated data only.
Context: Organizations increasingly develop software in a distributed manner. The cloud provides an environment to create and maintain software-based products and services. Currently, it is unknown which software processes are suited for cloud-based development and what their effects in specific contexts are.
Objective: We aim at better understanding the software process applied to distributed software development using the cloud as development environment. We further aim at providing an instrument which helps project managers comparing different solution approaches and to adapt team processes to improve future project activities and outcomes.
Method: We provide a simulation model which helps analyzing different project parameters and their impact on projects performed in the cloud. To evaluate the simulation model, we conduct different analyses using a Scrumban process and data from a project executed in Finland and Spain. An extra adaptation of the simulation model for Scrum and Kanban was used to evaluate the suitability of the simulation model to cover further process models.
Results: A comparison of the real project data with the results obtaind from the different simulation runs shows the simulation producing results close to the real data, and we could successfully replicate a distributed software project. Furthermore, we could show that the simulation model is suitable to address further process models.
Conclusion: The simulator helps reproducing activities, developers, and events in the project, and it helps analyzing potential tradeoffs, e.g., regarding throughput, total time, project size, team size and work-in-progress limits. Furthermore, the simulation model supports project managers selecting the most suitable planning alternative thus supporting decision-making processes.
Am Körper getragenen Geräte, sog. Wearables, kommunizieren in der Regel über Bluetooth-Low-Energy (BLE) mit dem Smartphone. Viele Anwendungen, insbesondere im Bereich Gesundheit und AAL, basieren auf der Zusammenarbeit von Wearables mit SmartHome-Geräten. Diese Arbeit präsentiert die Definition und Implementierung von einem neuen BLE Profil für EKG, das Streaming der Signal zum SmartPhone und die Möglichkeit, mehrere solcher Biosignale parallel zu streamen, besitzt. Die Datenarchitektur der App erlaubt eine konfigurierbare Synchronisation der Signal mit dem SmartHome.
Objective: This paper aims at getting an understanding of current problems and challenges with roadmapping processes in companies that are facing volatile markets with innovative products. It also aims at gathering ideas and attempts on how to react to those challenges.
Method: As an initial step towards the objectice a semi-structured expert interview study with a case company in the Smart Home domain was conducted. Four employees from the case company with different roles around product roadmaps have been interviewed and a content analysis of the data has been performed.
Results: The study shows a significant consensus among the interviewees about several major challenges and the necessity to change the traditional roadmapping process and format. The interviewees stated that based on their experience traditional feature-based product roadmaps are increasingly losing their benefits (such as good planning certainty) in volatile environments. Furthermore, the ability to understand customer needs and behaviors has become highly important for creating and adjusting product roadmaps. The interviewees see the need for both, sufficiently stable goals on the roadmap and flexibility with respect to products or features to be developed. To reach this target the interviewees proposed to create roadmaps based on outcome goals instead of product features. In addition, it was proposed to decrease the level of detail of the roadmaps and to emphasize the long-term view. Decisions about which feature to develop should be open as long as possible. Expected benefits of such a new way of product roadmapping are higher user centricity, a stable overall direction, more flexibility with respect to development decisions, and less breaking of commitments.
Workshops and tutorials
(2018)
The 19th International Conference on Product-Focused Software Process Improvement (PROFES 2018) hosted two workshops and three tutorials. The workshops and tutorials complemented and enhanced the main conference program, offering a wider knowledge perspective around the conference topics. The topics of the two workshops were Hybrid Development Approaches in Software Systems Development (HELENA) and Managing Quality in Agile & Rapid Software Development Processes (QUaSD). The topics of the tutorials were The human factor in agile transitions – using the personas concept in agile oaching, Process Management 4.0 – Best Practices, and Domain-specific languages for specification, development, and testing of autonomous systems.
A transaction is a demarcated sequence of application operations, for which the following properties are guaranteed by the underlying transaction processing system (TPS): atomicity, consistency, isolation, and durability (ACID). Transactions are therefore a general abstraction, provided by TPS that simplifies application development by relieving transactional applications from the burden of concurrency and failure handling. Apart from the ACID properties, a TPS must guarantee high and robust performance (high transactional throughput and low response times), high reliability (no data loss, ability to recover last consistent state, fault tolerance), and high availability (infrequent outages, short recovery times).
The architectures and workhorse algorithms of a high-performance TPS are built around the properties of the underlying hardware. The introduction of nonvolatile memories (NVM) as novel storage technology opens an entire new problem space, with the need to revise aspects such as the virtual memory hierarchy, storage management and data placement, access paths, and indexing. NVM are also referred to as storage-class memory (SCM).
Active storage
(2018)
In brief, Active Storage refers to an architectural hardware and software paradigm, based on collocation storage and compute units. Ideally, it will allow to execute application-defined data ... within the physical data storage. Thus Active Storage seeks to minimize expensive data movement, improving performance, scalability, and resource efficiency. The effective use of Active Storage mandates new architectures, algorithms, interfaces, and development toolchains.