Informatik
Refine
Document Type
- Conference proceeding (498)
- Journal article (186)
- Book chapter (48)
- Doctoral Thesis (10)
- Book (3)
- Report (1)
- Working Paper (1)
Has full text
- yes (747) (remove)
Is part of the Bibliography
- yes (747)
Institute
- Informatik (747)
- Technik (2)
Publisher
- Springer (160)
- Hochschule Reutlingen (95)
- IEEE (89)
- Gesellschaft für Informatik e.V (59)
- Elsevier (42)
- Association for Computing Machinery (37)
- IARIA (26)
- RWTH Aachen (14)
- De Gruyter (13)
- SciTePress (11)
- MDPI (10)
- Università Politecnica delle Marche (10)
- Deutsche Gesellschaft für Computer- und Roboterassistierte Chirurgie e.V. (9)
- Association for Information Systems (8)
- University of Hawai'i at Manoa (8)
- Emerald (7)
- IOP Publishing (7)
- American Marketing Association (4)
- OpenProceedings (4)
- Riga Technical University Press (4)
- SPIE. The International Society for Optical Engineering (4)
- University of Hawaii at Manoa (4)
- University of Zagreb (4)
- Wiley (4)
- EuroMed Press (3)
- International Academy of Business Disciplines (3)
- PeerJ Inc. (3)
- Sage Publishing (3)
- Taylor & Francis (3)
- Universität Konstanz (3)
- Universität Tübingen (3)
- Academic Conferences International (2)
- CSW-Verlag (2)
- Deutsche Aktuarvereinigung (DAV) e.V. (2)
- HTWG Konstanz (2)
- IADIS Press (2)
- IBM Research Division (2)
- IGI Global (2)
- IGI Publishing (2)
- International Association for Development of the Information Society (2)
- International Society for Photogrammetry and Remote Sensing (2)
- Smart Home & Living Baden-Württemberg e.V. (2)
- The Association for Computing Machinery, Inc. (2)
- Thieme (2)
- Universität Stuttgart (2)
- 3m5.Media GmbH (1)
- ARVO (1)
- Academic Conferences International Limited (1)
- Association for Computing Machinery ACM (1)
- Association of Computing Machinery (1)
- CIDR (1)
- Cambridge University Press (1)
- Circle International (1)
- Cornell Universiy (1)
- DGMP (1)
- DIMECC Oy (1)
- Deutsche Gesellschaft für Computer- und Roboterassistierte Chirurgie e. V. (1)
- EDP Sciences (1)
- Ed2.0Work (1)
- Elektronikpraxis, Vogel Business Media GmbH & Co. KG (1)
- Eurographics Association (1)
- Fachausschuß Management der Anwendungsentwicklung und -wartung (1)
- Frontiers Media (1)
- Frontiers Research Foundation (1)
- GITO Verlag (1)
- Gabler (1)
- Hochschule Heilbronn (1)
- Hochschule der Medien (1)
- IADIS (1)
- JMIR Publications (1)
- Johannes Kepler University Linz (1)
- Karlsruher Institut für Technologie (1)
- MFG Stiftung Baden-Württemberg (1)
- Morressier (1)
- NextMed (1)
- PLOS (1)
- Pallas Press (1)
- SISSA (1)
- SciKA (1)
- Society for Science and Education (1)
- Springer Nature (1)
- Technical University (1)
- Technische Universität Darmstadt (1)
- The Association for Computing Machinery (1)
- Tomas Bata University in Zlín (1)
- University of Jaén (1)
- University of Minho (1)
- University of Portsmouth (1)
- University of Zagreb Faculty of Organization and Informatics (1)
- University of the West of Scotland (1)
- Universität Trier (1)
- Universität des Saarlandes (1)
- imc (1)
- vwh Verlag Werner Hülsbusch (1)
Transaction processing is of growing importance for mobile computing. Booking tickets, flight reservation, banking, ePayment, and booking holiday arrangements are just a few examples for mobile transactions. Due to temporarily disconnected situations the synchronisation and consistent transaction processing are key issues. Serializability is a too strong criteria for correctness when the semantics of a transaction is known. We introduce a transaction model that allows higher concurrency for a certain class of transactions defined by its semantic. The transaction results are ”escrow serializable” and the synchronisation mechanism is non-blocking. Experimental implementation showed higher concurrency, transaction throughput, and less resources used than common locking or optimistic protocols.
Modern web-based applications are often built as multi-tier architecture using persistence middleware. Middleware technology providers recommend the use of Optimistic Concurrency Control (OCC) mechanism to avoid the risk of blocked resources. However, most vendors of relational database management systems implement only locking schemes for concurrency control. As consequence a kind of OCC has to be implemented at client or middleware side.
A simple Row Version Verification (RVV) mechanism has been proposed to implement an OCC at client side. For performance reasons the middleware uses buffers (cache) of its own to avoid network traffic and possible disk I/O. This caching however complicates the use of RVV because the data in the middleware cache may be stale (outdated). We investigate various data access technologies, including the new Java Persistence API (JPA) and Microsoft’s LINQ technologies for their ability to use the RVV programming discipline.
The use of persistence middleware that tries to relieve the programmer from the low level transaction programming turns out to even complicate the situation in some cases.Programmed examples show how to use SQL data access patterns to solve the problem.
Die Informatics Inside-Konferenz findet in diesem Jahr zum dritten Mal statt. Mit dem Thema "Grenzen überwinden – Virtualität erweitert Realität" stellt sich die Veranstaltung einem aktuellen Schwerpunkt, der viele Interessierte aus Wirtschaft, Wissenschaft und Forschung anzieht. Die Konferenz hat sich von einer Veranstaltung für die Masterstudenten des Studiengangs Medien- und Kommunikationsinformatik zu einer offenen Studentenkonferenz entwickelt. Um die Qualität weiter zu steigern wurde parallel dazu ein zweistufiges Review-Verfahren für Beiträge dieses Tagungsbandes eingeführt.
The Third International Conference on Advances in Databases, Knowledge, and Data Applications (DBKDA 2011) held on January 23-27, 2011 in St. Maarten, The Netherlands Antilles, continued a series of international events covering a large spectrum of topics related to advances in fundamentals on databases, evolution of relation between databases and other domains, data base technologies and content processing, as well as specifics in applications domains databases. Advances in different technologies and domains related to databases triggered substantial improvements for content processing, information indexing, and data, process and knowledge mining. The push came from Web services, artificial intelligence, and agent technologies, as well as from the generalization of the XML adoption. High-speed communications and computations, large storage capacities, and load-balancing for distributed databases access allow new approaches for content processing with incomplete patterns, advanced ranking algorithms and advanced indexing methods. Evolution on e-business, ehealth and telemedicine, bioinformatics, finance and marketing, geographical positioning systems put pressure on database communities to push the ‘de facto’ methods to support new requirements in terms of scalability, privacy, performance, indexing, and heterogeneity of both content and technology. We take this opportunity to thank all the members of the DBKDA 2011 Technical Program Committee as well as the numerous reviewers. The creation of such a broad and high-quality conference program would not have been possible without their involvement. We also kindly thank all the authors who dedicated much of their time and efforts to contribute to the DBKDA 2011. We truly believe that, thanks to all these efforts, the final conference program consists of top quality contributions. This event could also not have been a reality without the support of many individuals, organizations, and sponsors. We are grateful to the members of the DBKDA 2011 organizing committee for their help in handling the logistics and for their work to make this professional meeting a success. We hope that DBKDA 2011 was a successful international forum for the exchange of ideas and results between academia and industry and for the promotion of progress in database research. We are convinced that the participants found the event useful and communications very open. The beautiful places of St. Maarten surely provided a pleasant environment during the conference and we hope you had a chance to visit the surroundings.
This work presents a disconnected transaction model able to cope with the increased complexity of longliving, hierarchically structured, and disconnected transactions. Wecombine an Open and Closed Nested Transaction Model with Optimistic Concurrency Control and interrelate flat transactions with the aforementioned complex nature. Despite temporary inconsistencies during a transaction’s execution our model ensures consistency.
Social Media
(2011)
In jüngerer Zeit gewinnt die Nutzung des Internet für das Inbound Marketing zunehmend an Bedeutung. Dabei liegt der Fokus auf den so genannten Social Media Plattformen wie Facebook, YouTube, MySpace, XING, LinkedIn, Twitter, SlideShare und Posterous. Die Entwicklung dieser Medien ist auf eine Veränderung bei der Nutzung des Internet zurückzuführen, die häufig unter dem Schlagwort Web 2.0 zusammengefasst wird. Das gewandelte Mediennutzungsverhalten der Kunden induziert Chancen und Risiken für das Marketing.
Das Internet gewinnt für das Marketing zunehmend an Bedeutung. Dabei liegt der Fokus auf sogenannten Social-Media-Anwendungen wie Facebook, Twitter oder XING. Für Unternehmen stellt sich die Frage, ob das veränderte Mediennutzungsverhalten der Kunden eine neue Marketinglogik induziert. Eine aktuelle Untersuchung gibt Einblicke in die Chancen und Risiken, Anwendungsbedingungen und Kontextfaktoren für die Nutzung von Social Media im Marketing.
The Fourth International Conference on Advances in Databases, Knowledge, and Data Applications [DBKDA 2012], held between February 29th and March 5th, 2012 in Saint Gilles, Reunion Island, continued a series of international events covering a large spectrum of topics related to advances in fundamentals on databases, evolution of relation between databases and other domains, data base technologies and content processing, as well as specifics in applications domains databases. Advances in different technologies and domains related to databases triggered substantial improvements for content processing, information indexing, and data, process and knowledge mining. The push came from Web services, artificial intelligence, and agent technologies, as well as from the generalization of the XML adoption. High-speed communications and computations, large storage capacities, and loadbalancing for distributed databases access allow new approaches for content processing with incomplete patterns, advanced ranking algorithms and advanced indexing methods. Evolution on e-business, e-health and telemedicine, bioinformatics, finance and marketing, geographical positioning systems put pressure on database communities to push the ‘de facto’ methods to support new requirements in terms of scalability, privacy, performance, indexing, and heterogeneity of both content and technology. We take here the opportunity to warmly thank all the members of the DBKDA 2012 Technical Program Committee, as well as the numerous reviewers. The creation of such a broad and high quality conference program would not have been possible without their involvement. We also kindly thank all the authors who dedicated much of their time and efforts to contribute to DBKDA 2012. We truly believe that, thanks to all these efforts, the final conference program consisted of top quality contributions. Also, this event could not have been a reality without the support of many individuals, organizations, and sponsors. We are grateful to the members of the DBKDA 2012 organizing committee for their help in handling the logistics and for their work to make this professional meeting a success. We hope that DBKDA 2012 was a successful international forum for the exchange of ideas and results between academia and industry and for the promotion of progress in the fields of databases, knowledge, and data applications. We are convinced that the participants found the event useful and communications very open. We also hope the attendees enjoyed the charm of Saint Gilles, Reunion Island.
Redirected walking techniques allow people to walk in a larger virtual space than the physical extents of the laboratory. We describe two experiments conducted to investigate human sensitivity to walking on a curved path and to validate a new redirected walking technique. In a psychophysical experiment, we found that sensitivity to walking on a curved path was significantly lower for slower walking speeds (radius of 10 meters versus 22 meters). In an applied study, we investigated the influence of a velocity-dependent dynamic gain controller and an avatar controller on the average distance that participants were able to freely walk before needing to be reoriented. The mean walked distance was significantly greater in the dynamic gain controller condition, as compared to the static controller (22 meters versus 15 meters). Our results demonstrate that perceptually motivated dynamic redirected walking techniques, in combination with reorientation techniques, allow for unaided exploration of a large virtual city model.
Das Motto der diesjährigen Konferenz lautet "Reality++: Tomorrow comes today!". Unter diesem fast schon visionären Thema werden die Ergebnisse der verschiedenen Vertiefungsarbeiten aus den vergangenen Monaten präsentiert. Das Programm wird vervollständigt durch Beiträge von Experten aus der Forschung und Industrie.
Multi-dimensional patient data, such as time varying volume data, data of different imaging modalities, surface segmentations etc. are of growing importance in the clinical routine. For many use cases, it is of major importance to replicate a certain visualization of a data set created on one machine on a different computer using different software tools. Up until now, there exists no standardized methodology for this consistent presentation. We propose an extension of the Digital Imaging und Communications in Medicine (DICOM) called “Multi dimensional Presentation State” and outline scope and first results of the standardization process.
Energy-efficiency and safety became an important factor for car manufacturers. Thus, the cars have been optimised regarding the energy consumption and safety by optimising for example the power train or the engine. Besides the optimisation of the car itself, energy-efficiency and safety can also be increased by adapting the individual driving behaviour to the current driving situation. This paper introduces a driving system, which is in development. Its goal is to optimise the driving behaviour in terms of energy-efficiency and safety by giving recommendations to the driver. For the creation of a recommendation the driving system monitors the driver and the current driving situation as well as the car using in-vehicle sensors and serial-bus systems. On the basis of the acquired data, the driving system will give individual energy-efficiency and safety recommendations in real-time. This will allow eliminating bad driving habits, while considering the driver needs.
Telemedicine is becoming an increasingly important approach to diagnostic, treat or prevent diseases. However, the usage of Information Communication Technologies in healthcare results in a considerable amount of data that must be efficiently and securely transmitted. Many manufacturers provide telemedicine platforms without regarding interoperability, mobility and collaboration. This paper describes a collaborative mobile telemonitoring platform that can use the IEEE 11073 and HL7 communication standards or adapt proprietary protocols. The proposed platform also covers the security and modularity aspects. Furthermore this work introduces an Android-based prototype implementation
The workshop aims to discuss leading edge contributions to the interdisciplinary research area of ambient intelligence (AmI) applied to the domains of telemedicine and driving assistance. AmI refers to human centered environments attributed with sensors. The development of AmI in the two application domains of the workshop shares several commonalities: the extensive usage of networked devices and sensors, the design of artificial intelligence algorithms for diagnosis, including recommendation systems and qualitative reasoning or the application of mobile and wireless communication to their distributed systems. Together with the presentation of common aspects of Ambient Intelligence, a further goal of the workshop is to stimulate synergies among both application domains and present examples. The telemedicine domain can benefit from methodologies in designing complex devices, real-time conform system design, audiovisual or computer vision system design used in automotive driving assistance. Furthermore, the automotive domain can benefit from the usercentric view, biometric sensor data design, multi-user data bases for aggregation and diagnosis using big data like used in telemedicine. The German Government supports these research lines in its Hightec-Strategie under the domains “Health and Nutrition” and “Climate and Energy”. In Spain the term “Spanish Program for R&D Challenged Oriented Society – Challenge in energy safe, efficient and clean & Challenge in sustainable transport, smart and integrated” is used. Scientific contributions to the event are peer-reviewed by a suited program committee having members from Germany and Spain. The same committee is serving the JARCA workshop (Jornadas sobre Sistemas cualitativos y sus Aplicaciones en Diagnosis, Robótica e Inteligencia Ambiental - Conference on Qualitative Systems and their Applications in Diagnoses, Robotics and Ambient Intelligence) since 15 years. This workshop is sponsored by the German Academic Exchange Service (DAAD) under contract number 57070010.
When forecasting sales figures, not only the sales history but also the future price of a product will influence the sales quantity. At first sight, multivariate time series seem to be the appropriate model for this task. Nonetheless, in real life history is not always repeatable, i.e., in the case of sales history there is only one price for a product at a given time. This complicates the design of a multivariate time series. However, for some seasonal or perishable products the price is rather a function of the expiration date than of the sales history. This additional information can help to design a more accurate and causal time series model. The proposed solution uses an univariate time series model but takes the price of a product as a parameter that influences systematically the prediction based on a calculated periodicity. The price influence is computed based on historical sales data using correlation analysis and adjustable price ranges to identify products with comparable history. The periodicity is calculated based on a novel approach that is based on data folding and Pearson Correlation. Compared to other techniques this approach is easy to compute and allows to preset the price parameter for predictions and simulations. Tests with data from the Data Mining Cup 2012 as well as artificial data demonstrate better results than established sophisticated time series methods.
"Learning by doing" in Higher Education in technical disciplines is mostly realized by hands-on labs. It challenges the exploratory aptitude and curiosity of a person. But, exploratory learning is hindered by technical situations that are not easy to establish and to verify. Technical skills are, however, mandatory for employees in this area. On the other side, theoretical concepts are often compromised by commercial products. The challenge is to contrast and reconcile theory with practice. Another challenge is to implement a self-assessment and grading scheme that keeps up with the scalability of e-learning courses. In addition, it should allow the use of different commercial products in the labs and still grade the assignment results automatically in a uniform way. In two European Union funded projects we designed, implemented, and evaluated a unique e-learning reference model, which realizes a modularized teaching concept that provides easily reproducible virtual hands-on labs. The novelty of the approach is to use software products of industrial relevance to compare with theory and to contrast different implementations. In a sample case study, we demonstrate the automated assessment for the creative database modeling and design task. Pilot applications in several European countries demonstrated that the participants gained highly sustainable competences that improved their attractiveness for employment.
The Fifth International Conference on Advances in Databases, Knowledge, and Data Applications [DBKDA 2013], held between January 27th- February 1st, 2013 in Seville, Spain, continued a series of international events covering a large spectrum of topics related to advances in fundamentals on databases, evolution of relation between databases and other domains, data base technologies and content processing, as well as specifics in applications domains databases. Advances in different technologies and domains related to databases triggered substantial improvements for content processing, information indexing, and data, process and knowledge mining. The push came from Web services, artificial intelligence, and agent technologies, as well as from the generalization of the XML adoption. High-speed communications and computations, large storage capacities, and loadbalancing for distributed databases access allow new approaches for content processing with incomplete patterns, advanced ranking algorithms and advanced indexing methods. Evolution on e-business, ehealth and telemedicine, bioinformatics, finance and marketing, geographical positioning systems put pressure on database communities to push the ‘de facto’ methods to support new requirements in terms of scalability, privacy, performance, indexing, and heterogeneity of both content and technology. We take here the opportunity to warmly thank all the members of the DBKDA 2013 Technical Program Committee, as well as the numerous reviewers. The creation of such a high quality conference program would not have been possible without their involvement. We also kindly thank all the authors who dedicated much of their time and efforts to contribute to DBKDA 2013. We truly believe that, thanks to all these efforts, the final conference program consisted of top quality contributions. Also, this event could not have been a reality without the support of many individuals, organizations, and sponsors. We are grateful to the members of the DBKDA 2013 organizing committee for their help in handling the logistics and for their work to make this professional meeting a success. We hope that DBKDA 2013 was a successful international forum for the exchange of ideas and results between academia and industry and for the promotion of progress in the fields of databases, knowledge and data applications. We are convinced that the participants found the event useful and communications very open. We also hope the attendees enjoyed the charm of Seville, Spain.
New storage technologies, such as Flash and Non- Volatile Memories, with fundamentally different properties are appearing. Leveraging their performance and endurance requires a redesign of existing architecture and algorithms in modern high performance databases. Multi-Version Concurrency Control (MVCC) approaches in database systems, maintain multiple timestamped versions of a tuple. Once a transaction reads a tuple the database system tracks and returns the respective version eliminating lock-requests. Hence under MVCC reads are never blocked, which leverages well the excellent read performance (high throughput, low latency) of new storage technologies. Upon tuple updates, however, established implementations of MVCC approaches (such as Snapshot Isolation) lead to multiple random writes – caused by (i) creation of the new and (ii) in-place invalidation of the old version – thus generating suboptimal access patterns for the new storage media. The combination of an append based storage manager operating with tuple granularity and snapshot isolation addresses asymmetry and in-place updates. In this paper, we highlight novel aspects of log-based storage, in multi-version database systems on new storage media. We claim that multi-versioning and append-based storage can be used to effectively address asymmetry and endurance. We identify multi-versioning as the approach to address dataplacement in complex memory hierarchies. We focus on: version handling, (physical) version placement, compression and collocation of tuple versions on Flash storage and in complex memory hierarchies. We identify possible read- and cacherelated optimizations.
When forecasting sales figures, not only the sales history but also the future price of a product will influence the sales quantity. At first sight, multivariate time series seem to be the appropriate model for this task. Nontheless, in real life history is not always repeatable, i.e. in the case of sales history there is only one price for a product at a given time. This complicates the design of a multivariate time series. However, for some seasonal or perishable products the price is rather a function of the expiration date than of the sales history. This additional information can help to design a more accurate and causal time series model. The proposed solution uses an univariate time series model but takes the price of a product as a parameter that influences systematically the prediction. The price influence is computed based on historical sales data using correlation analysis and adjustable price ranges to identify products with comparable history. Compared to other techniques this novel approach is easy to compute and allows to preset the price parameter for predictions and simulations. Tests with data from the Data Mining Cup 2012 demonstrate better results than established sophisticated time series methods.
Der Beitrag gibt einen Überblick zum Stand der Vertrauensforschung in Marketing und Vertrieb. Dabei ist Vertrauen als Gegenstand der Forschung innerhalb des Relationship Marketing Ansatzes sehr gut etabliert. Bei der Definition des Vertrauensbegriffs stützt sich das Marketing auf die Erkenntnisse der sozialwirtschaftlichen Nachbardisziplinen. Soweit Kunden ihren Anbietern vertrauen, gehen sie grundsätzlich ein Risiko ein und machen sich hierdurch angreifbar. Man vertraut in einen Anbieter, ohne vorab genau zu wissen, ob das gewünschte Resultat einer Kooperation mit Sicherheit eintritt. Dies gilt umgekehrt auch für den Anbieter, der zum Teil erhebliche Vorinvestitionen tätigen muss, ohne vorab zu wissen, ob tatsächlich eine Geschäftsbeziehung mit einem Kunden entsteht. Vertrauen ist daher v.a. in komplexen und langfristigen Beziehungen zwischen Anbietern und Kunden eine wesentliche Ressource. Entsprechend thematisiert der Beitrag die Bedingungen und Auswirkungen von Vertrauen auf unterschiedlichen Ebenen. Dabei dominiert in Marketing und Vertrieb noch immer eine interpersonale Perspektive. Die Potentiale organisationaler Beziehungsstrategien sind zum gegenwärtigen Zeitpunkt eher schwach beleuchtet, jedoch greift der Beitrag einige Trends für die weitere Ausrichtung der Vertrauensforschung auf, die zukünftig stärker an Bedeutung gewinnen werden. Dabei ist grundsätzlich davon auszugehen, dass bei zunehmend volatilen Rahmenbedingungen das Interesse an Vertrauensfragen auch in Marketing und Vertrieb weiter zunimmt.