Refine
Year of publication
Document Type
- Part of Periodical (132)
- Bachelor Thesis (68)
- Diploma Thesis (47)
- Master's Thesis (34)
- Study Thesis (9)
- Conference Proceedings (5)
- Doctoral Thesis (2)
- Article (1)
Keywords
- Simulation (5)
- ontology (5)
- Bluetooth (4)
- computer clusters (4)
- Android <Systemplattform> (3)
- Augmented Reality (3)
- Customer Relationship Management (3)
- Enterprise 2.0 (3)
- Informatik (3)
- Knowledge Compilation (3)
Institute
- Fachbereich 4 (298) (remove)
In order to enhance the company’s appeal for potential employees and improve the satisfaction of already salaried workers, it is necessary to offer a variety of work-life balance measures. But as their implementation causes time and financial costs, a prioritization of measures is needed. To express a recommendation for companies, this study is led by the questions if there are work-life balance measures which have more impact on employee satisfaction than others, how big the relative impact of work-life balance measures on job satisfaction in comparison to other work and private life variables is, if there is a relation between the effectiveness of measures and their use and if there is a difference between the measures which are most important from the employees’ perspective and the companies’ offers.
These questions are formulated in eight research hypotheses which are examined in a quantitative research design with online survey data from 289 employees of fifteen different German companies. The formation of a hierarchy of the effectiveness of measures towards job satisfaction as well as the investigation of the relative impact in comparison to other variables is performed using a multiple regression analysis, whilst the differences between employees’ expectations and the availability of offers are examined with t-tests.
Support in childcare, support in voluntary activities and teambuilding events have a significantly higher impact on job satisfaction than other work-life balance measures, and their potential use is higher than the actual use which leads to the conclusion that there is yet potential for companies to improve their employees’ satisfaction by implementing these measures. In addition, flexible work hours, flexible work locations and free time and overtime accounts are the most important measures from the employees’ point of view and already widely offered by the surveyed companies. In general, the overall use of the available measures and the quantity of offered measures are more important with regard to job satisfaction than the specific kind of measure. In addition, work-life balance measures are more important towards job satisfaction for younger people.
Despite widespread plans of big companies like Amazon and Google to develop unmanned delivery drones, scholarly research in this field is scarce, especially in the information systems field. From technical and legal perspectives, drone delivery in last-mile scenarios is in a quite mature state. However, estimates of user acceptance are varying between high skepticism and exaggerated optimism. This research follows a mixed method approach consisting both qualitative and quantitative research, to identify and test determinants of consumer delivery drone service adoption. The qualitative part rests on ten interviews among average consumers, who use delivery services on a regular basis. Insights gained from the qualitative part were used to develop an online survey and to assess the influence of associated risks on adoption intentions. The quantitative results show that especially financial and physical risks impede drone delivery service adoption. Delivery companies who are currently thinking about providing a delivery drone service may find these results useful when evaluating usage behaviors in the future market for delivery drones.
Diese Arbeit beschreibt einen Ansatz zur webbasierten und GPU-unterstützten medizinischen Visualisierung. Der Schwerpunkt liegt auf der client-seitigen Ausführung von direktem Volumen-Rendering mittels WebGL und der Übertragung von medizinischen Datensätzen von Server zu Client. Die Motivation dieser Arbeit liegt vor allem in den neuesten Entwicklungen von Webtechnologien begründet, da es bisher nicht möglich war hardwarebeschleunigte 3D-Grafik direkt im Webbrowser darzustellen. Erst seit der Entwicklung der 3D-Grafik-Programmierschnittstelle WebGL besteht die Möglichkeit GPU-unterstütztes Volumenrendering im Browser-Kontext ohne den Einsatz zusätzlicher Software zu realisieren. Es wird ein webbasiertes Volumenrendering-System vorgestellt, das die Umsetzung von Volumen-Raycasting mit WebGL zur direkten Darstellung von Volumendaten in Echtzeit behandelt. Für die technische Umsetzung wurden das Google Web Toolkit und die Google App Engine als Infrastruktur verwendet.
Die weltweite Zugänglichkeit und umfangreiche Nutzung des Internets machen dieses Medium zu einem effizienten und beliebten Informations-, Kommunikations-, und Verkaufsinstrument. Immer mehr Menschen und Organisationen versuchen, diese Vorzüge durch eine eigene Website für ihre Zwecke zu verwenden. Als hilfreiches Mittel zur Optimierung von Webpräsenzen bewährte sich in den letzten Jahren der Einsatz von Web-Analytics-Software. Durch diese Software sind Websitebetreiber in der Lage, Informationen über die Besucher ihrer Website und deren Nutzungsverhalten zu sammeln und zu messen. Das angestrebte Resultat sind Optimierungsentscheidungen auf Basis von Daten an Stelle von Annahmen und wirkungsvolle Testmöglichkeiten.
Für den Bereich des E-Commerce existieren bis dato zahlreiche wissenschaftliche und praxiserprobte Hilfestellungen für Web-Analytics-Projekte. Informationswebsites hingegen werden trotz ihrer Wichtigkeit nur vereinzelt thematisiert. Um diesem Defizit entgegenzuwirken, hat Hausmann 2012 das Framework for Web Analytics entwickelt, welches dem Anwender ein hilfreiches Referenzmodell für sein Web Analytics-Vorhaben bietet. Diesen Ansatz weiter voranzutreiben ist das Ziel der Abschlussarbeit. Dazu wird mithilfe einer Literaturanalyse und einer Fallstudie das Framework validiert und ergänzt, sowie weitere Handlungsempfehlungen identifiziert. Als Ergebnis werden die wichtigsten Erkenntnisse dieser Forschung zusammengefasst und für den zukünftigen Gebrauch festgehalten.
This paper introduces Vocville, a causal online game for learning vocabularies. I am creating this application for my master thesis of my career as a "Computervisualist" (computer visions) for the University of Koblenz - Landau. The application is an online browser game based on the idea of the really successful Facebook game FarmVille. The application is seperated in two parts; a Grails application manages a database which holds the game objects like vocabulary, a Flex/Flash application generates the actual game by using these data. The user can create his own home with everything in it. For creating things, the user has to give the correct translation of the object he wants to create several times. After every query he has to wait a certain amount of time to be queried again. When the correct answer is given sufficient times, the object is builded. After building one object the user is allowed to build others. After building enough objects in one area (i.e. a room, a street etc.) the user can activate other areas by translating all the vocabularies of the previous area. Users can also interact with other users by adding them as neighbors and then visiting their homes or sending them gifts, for which they have to fill in the correct word in a given sentence.
In dieser Arbeit wurde die Realisierung einer mobilen Sicherheitslösung für Überwachungszwecke vorgestellt, welche unter Zuhilfenahme des staatlich geförderten Forschungsprojekts CamInSens entwickelt wurde. CamInSens soll erreichen, in überwachten Gefahrenbereichen die erhaltenen Video- und Sensorendaten so zu analysieren, dass Bedrohungen möglichst frühzeitig erkannt und behandelt werden können. Das Ziel dieser Arbeit war, auf Basis von verarbeiteten Daten aus CamInSens eine Interaktion und Visualisierung zu entwickeln, die bei einem späteren Praxiseinsatz mobilem Sicherheitspersonal dabei hilft, in Bedrohungssituationen fundierte Entscheidungen treffen zu können. Zu diesem Zweck wurde nicht nur eine Software implementiert, sondern auch eine Marktsichtung hinsichtlich geeigneter Geräte und einsetzbarer Softwarebibliotheken durchgeführt.
Virtual Goods + ODRL 2012
(2012)
This is the 10th international workshop for technical, economic, and legal aspects of business models for virtual goods incorporating the 8th ODRL community group meeting. This year we did not call for completed research results, but we invited PhD students to present and discuss their ongoing research work. In the traditional international group of virtual goods and ODRL researchers we discussed PhD research from Belgium, Brazil, and Germany. The topics focused on research questions about rights management in the Internet and e-business stimulation. In the center of rights management stands the conception of a formal policy expression that can be used for human readable policy transparency, as well as for machine readable support of policy conformant systems behavior up to automatic policy enforcement. ODRL has proven to be an ideal basis for policy expressions, not only for digital copy rights, but also for the more general "Policy Awareness in the World of Virtual Goods". In this sense, policies support the communication of virtual goods, and they are a virtualization of rules-governed behavior themselves.
This thesis deals with the distribution of large virtual networks to multiple physical hosts using the virtualization tools VNUML and EDIV. While VNUML is responsible for simulating the network, EDIV has its focus primarily on the distributiion. After a short explanation of basic terms and concepts from the field of network simulation, the two previously mentioned virtualization tools are described in detail. This starts in both cases with a detailed description regarding the proper installation and configuration, followed by a demonstration of the most important functions, such as starting or stopping a simulation. The creation of a suitable network scenario and the information scripts provided by EDIV are two further points described in this context. To use the presented possibilities of those tools in practice, an own network scenario is designed and distributed to several hosts, so that the different functions of EDIV can be shown and described accurately.
Das Zeitfenster ab 1964 bis heute verzeichnet einen Geburtenrückgang. Diese Tatsache hat insofern negative Auswirkung, als dass es für die sozialen Sicherungen Gefahren birgt, nicht mehr finanziert werden zu können. Des Weiteren ist über einen langen Zeitraum mit einem Fachkräftemangel zu rechnen, sollte sich der Geburtenrückgang weiterhin fortsetzen.
In dieser Ausarbeitung wird der Zusammenhang zwischen dem Bildungsgrad der Frau und dem Anteil kindeloser Frauen und der Anzahl an Kindern untersucht. Sollte dabei ein Zusammenhang bestehen, so müssten vor allem Akademikerinnen den höchsten Anteil an Kinderlosigkeit und die niedrigste Kinderzahl je Frau aufweisen. Im Umkehrschluss müssten Frauen mit Hauptschulabschluss oder Frauen ohne Schulabschluss die niedrigsten Anteile an Frauen ohne Kinder und die höchsten Zahlen an Kindern je Frau vermelden.
Das Ziel dieser Bachelorarbeit ist es ein Schätzkonzept zu validieren, das im Rahmen eines Projektes auf der Universität Koblenz-Landau entwickelt wurde, um den Anteil kinderloser Frauen und die Anzahl an Kindern je Frau zu schätzen
Hybrid systems are the result of merging the two most commonly used models for dynamical systems, namely continuous dynamical systems defined by differential equations and discrete-event systems defined by automata. One can view hybrid systems as constrained systems, where the constraints describe the possible process flows, invariants within states, and transitions on the one hand, and to characterize certain parts of the state space (e.g. the set of initial states, or the set of unsafe states) on the other hand. Therefore, it is advantageous to use constraint logic programming (CLP) as an approach to model hybrid systems. In this paper, we provide CLP implementations, that model hybrid systems comprising several concurrent hybrid automata, whose size is only straight proportional to the size of the given system description. Furthermore, we allow different levels of abstraction by making use of hierarchies as in UML statecharts. In consequence, the CLP model can be used for analyzing and testing the absence or existence of (un)wanted behaviors in hybrid systems. Thus in summary, we get a procedure for the formal verification of hybrid systems by model checking, employing logic programming with constraints.
In the "Institut MTI Mittelrhein" the impact of training and experience on motion and motion planning in sports is examined. As a part of these investigations, the motion behaviour of a sports climber is analyzed on a sports climbing wall. The aim of this thesis is the segmentation and marker-based detection of the movement of a climber in video recordings. The segmented markers from each image are tracked by a tracking procedure over time. The motion tracks will be exported to the motion analysis software Simi Motion in a format suitable for further analysis.
Remote rendering services offer the possibility to stream high quality images to lower powered devices. Due to the transmission of data the interactivity of applications is afflicted with a delay. A method to reduce delay of the camera manipulation on the client is called 3d-warping. This method causes artifacts. In this thesis different approaches of remote rendering setups will be shown. The artifacts and improvements of the warping method will be described. Methods to reduce the artifacts will be implemented and analyzed.
Das allgemeine Erreichbarkeitsproblem in Stellen/Transitions-Netzen behandelt die Frage, ob eine bestimmte Zielmarkierung aus einer anderen Markierung erreichbar ist. Die lineare Algebra als Teilgebiet der Mathematik kann zur Beschreibung und Analyse von Petri-Netzen herangezogen werden. Jedoch stellen Kreise in Netzen bei der Erreichbarkeitsanalyse ein Problem für sie dar: ob für bestimmte Transitionen erforderliche Marken sofort, erst nach dem Durchlaufen bestimmter Kreise oder gar nicht zur Verfügung stehen, ist nicht erkennbar, hat jedoch unmittelbare Auswirkungen auf die Erreichbarkeit. Es gibt verschiedene Möglichkeiten, dieses Problem linear-algebraisch anzugehen. In dieser Arbeit werden vorhandene Verfahren sowie neue Lösungsansätze auf Basis der Reproduzierbarkeit der leeren Markierung diskutiert.
This thesis deals with problems, which occure when rendering stereoscopic contents. These problems are elaborated, simulated with the help of a program developed in this thesis and evaluated by a group of volunteers. Thereby it shall be determined, whether the errors are noticeable and how much they influence the 3D effect of the stereoscopic images. Each error is simulated using different camera assemblies and evaluated depending on the choosen assembly.
This work deals with the migration of software systems towards the use of the character set defined in the Unicode standard. The work is performed as a case study on the document-management-system PROXESS. A conversion process will be designed that defines the working-steps of the migration for the entire system as well as an arbitrary decomposition of the system into individual modules. The working-steps for each module can be performed chronologically independent of each other to a great extend. For the conversion of the implementation, an approach of automatic recognition of usage patterns is applied. The approach aims at searching the abstract syntax tree for sequences of program instructions that can be assigned to a certain usage pattern. The usage pattern defines another sequence of instructions that acts as an sample solution for that usage pattern. The sample solution demonstrates the Unicode-based management of strings for that usage pattern. By applying a transformation rule, the original sequence of instructions is transferred to the sequence of instructions exposed by the sample solution of the related usage pattern. This mechanism is a starting point for the development of tools that perform this transformation automatically.
Die Computergrafik befasst sich mit der Erzeugung von virtuellen Bildern. Im Bereich der 3D-Computergrafik werden die dargestellten Objekte im dreidimensionalen Raum beschrieben. Dazu bedient man sich diverser Generierungsverfahren. Einer dieser so genannten Renderer ist das Raytracing-Verfahren. Es erfreut sich in der Computergrafik wegen der erreichten Bildqualität bei ueberschaubarer Komplexität großer Beliebtheit. Dabei wird versucht, immer realistischere Ergebnisse zu erreichen. In der Vergangenheit wurde Raytracing deswegen beispielsweise um globale Beleuchtungsmodelle oder um reflektierende beziehungsweise um transparente Objekte erweitert. Dabei wurde aber ein wichtiger Punkt häufig vernachlässigt, welcher ebenfalls den Grad an Realismus deutlich erhöhen kann: die Kamera. Meistens geht man auch heutzutage von einem vereinfachten Lochkameramodell aus. Aus diesem Grund genügen solche Modelle nicht den Ansprüchen physikalisch-korrekter Renderingverfahren. Eine wirklich umfassend korrekte Abbildung von Szenen darf also nicht vernachlässigen, dass ein generiertes Bild durch ein Linsensystem noch einmal entscheidend beeinflusst wird. In dieser Arbeit wird deswegen ein physikalisch korrektes Kameramodell vorgestellt, welches die geometrischen Eigenschaften des Linsensystems berücksichtigt und die Belichtung auf der Bildebene korrekt berechnet.
UML models and OWL ontologies constitute modeling approaches with different strength and weaknesses that make them appropriate for use of specifying different aspects of software systems. In particular, OWL ontologies are well suited to specify classes using an expressive logical language with highly flexible, dynamic and polymorphic class membership, while UML diagrams are much more suitable for specifying not only static models including classes and associations, but also dynamic behavior. Though MOF based metamodels and UML profiles for OWL have been proposed in the past, an integrated use of both modeling approaches in a coherent framework has been lacking so far. We present such a framework, TwoUse, for developing integrated models, comprising the benefits of UML models and OWL ontologies
The STOR project aims at the development of a scientific component system employing models and knowledge for object recognition in images. This interim report elaborates on the requirements for such a component system, structures the application area by identifying a large set of basic operations, and shows how a set of appropriate data structures and components can be derived. A small case studies exemplifies the approach.
Dualizing marked Petri nets results in tokens for transitions (t-tokens). A marked transition can strictly not be enabled, even if there are sufficient "enabling" tokens (p-tokens) on its input places. On the other hand, t-tokens can be moved by the firing of places. This permits flows of t-tokens which describe sequences of non-events. Their benefiit to simulation is the possibility to model (and observe) causes and effects of non-events, e.g. if something is broken down.
With the ongoing process of building business networks in today- economy, business to-business integration (B2B Integration) has become a strategic tool for utilizing and optimizing information exchange between business partners. Industry and academia have made remarkable progress in implementing and conceptualizing different kinds of electronic inter-company relationships in the last years. Nevertheless, academic findings generally focus exclusively on certain aspects of the research object, e.g. document standards, process integration or other descriptive criteria. Without arncommon framework these results stay unrelated and their mutual impact on each other remains largely unexplained. In this paper we explore motivational factors of B2B integration in practice. In a research project using a uniform taxonomy (eXperience methodology) we classified real-world B2B integration projects from a pool of over 400 case studies using a pre-developed framework for integration scenarios. The result of our partly exploratory research shows the influence of the role of a company in the supply chain and its motive to invest in a B2B solution.
This volume contains those research papers presented at the Second International Conference on Tests and Proofs (TAP 2008) that were not included in the main conference proceedings. TAP was the second conference devoted to the convergence of proofs and tests. It combines ideas from both areas for the advancement of software quality. To prove the correctness of a program is to demonstrate, through impeccable mathematical techniques, that it has no bugs; to test a program is to run it with the expectation of discovering bugs. On the surface, the two techniques seem contradictory: if you have proved your program, it is fruitless to comb it for bugs; and if you are testing it, that is surely a sign that you have given up on any hope of proving its correctness. Accordingly, proofs and tests have, since the onset of software engineering research, been pursued by distinct communities using rather different techniques and tools. And yet the development of both approaches leads to the discovery of common issues and to the realization that each may need the other. The emergence of model checking has been one of the first signs that contradiction may yield to complementarity, but in the past few years an increasing number of research efforts have encountered the need for combining proofs and tests, dropping earlier dogmatic views of their incompatibility and taking instead the best of what each of these software engineering domains has to offer. The first TAP conference (held at ETH Zurich in February 2007) was an attempt to provide a forum for the cross-fertilization of ideas and approaches from the testing and proving communities. For the 2008 edition we found the Monash University Prato Centre near Florence to be an ideal place providing a stimulating environment. We wish to sincerely thank all the authors who submitted their work for consideration. And we would like to thank the Program Committee members as well as additional referees for their great effort and professional work in the review and selection process. Their names are listed on the following pages. In addition to the contributed papers, the program included three excellent keynote talks. We are grateful to Michael Hennell (LDRA Ltd., Cheshire, UK), Orna Kupferman (Hebrew University, Israel), and Elaine Weyuker (AT&T Labs Inc., USA) for accepting the invitation to address the conference. Two very interesting tutorials were part of TAP 2008: "Parameterized Unit Testing with Pex" (J. de Halleux, N. Tillmann) and "Integrating Verification and Testing of Object-Oriented Software" (C. Engel, C. Gladisch, V. Klebanov, and P. Rümmer). We would like to express our thanks to the tutorial presenters for their contribution. It was a team effort that made the conference so successful. We are grateful to the Conference Chair and the Steering Committee members for their support. And we particularly thank Christoph Gladisch, Beate Körner, and Philipp Rümmer for their hard work and help in making the conference a success. In addition, we gratefully acknowledge the generous support of Microsoft Research Redmond, who financed an invited speaker.
Terrainklassifikation mit Markov-Zufallsfeldern auf Basis von fusionierten Kamera- und Laserdaten
(2011)
A mobile system, that has to navigate automatically in an outdoor environment,rnneeds to have knowledge about its surrounding terrain. Laser-range-finders, sometimes combined with cameras, are often used to analyse terrain. Several problems, like missing or noisy data, lead to erroneous identification of environment.
The target of this work is to add a context-sensitive classification component and data of other sensors to a procedure, based on 3d-data, obtained by a laser-range-finder. The first upgrade consists of a Markov Random Field, which is used to model the relationships between neighbouring terrainsegments and allows a segmentation of the whole terrain.
The second upgrade fuses laserdata with pictures from cameras to obtain additional terrainfeatures.
Web-programming is a huge field of different technologies and concepts. Each technology implements a web-application requirement like content generation or client-server communication. Different technologies within one application are organized by concepts, for example architectural patterns. The thesis describes an approach for creating a taxonomy about these web-programming components using the free encyclopaedia Wikipedia. Our 101companies project uses implementations to identify and classify the different technology sets and concepts behind a web-application framework. These classifications can be used to create taxonomies and ontologies within the project. The thesis also describes, how we priorize useful web-application frameworks with the help of Wikipedia. Finally, the created implementations concerning web-programming are documented.
Taktstraße
(2008)
Eine Taktstraße ermöglicht eine automatisierte Verarbeitung eines Werkstückes mit Hilfe von Förderbändern, Lichtschranken, Schiebern und Bearbeitungsstationen. Für eine vorgegebene Taktstraße wird eine Ansteuerung entwickelt. Dazu wird der Mikrocontroller ATMega16 von Atmel eingesetzt. Ein externer Controller sendet über den TWI-Bus Steuerbefehle an den mit der Taktstraße verbundenen Controller. Um die Taktstraße bedienbar zu machen, wird eine geeignete Platine entworfen sowie eine LCD-Bibliothek als Ausgabe- und Informationsmedium. Die Arbeit umfasst alle für ein Projekt im Rahmen eines Informatikstudiums benötigten Entwicklungsstadien von der Projektplanung über die Aneignung von spezifischem Grundlagenwissen, die Hard- und Softwareentwicklung bis hin zu ausführlichen Entwicklungs- und Testphasen.
Generalized methods for automated theorem proving can be used to compute formula transformations such as projection elimination and knowledge compilation. We present a framework based on clausal tableaux suited for such tasks. These tableaux are characterized independently of particular construction methods, but important features of empirically successful methods are taken into account, especially dependency directed backjumping and branch local operation. As an instance of that framework an adaption of DPLL is described. We show that knowledge compilation methods can be essentially improved by weaving projection elimination partially into the compilation phase.
The annotation of digital media is no new area of research, instead it is widely investigated. There are many innovative ideas for creating the process of annotation. The most extensive segment of related work is about semi automatic annotation. One characteristic is common in the related work: None of them put the user in focus. If you want to build an interface, which is supporting and satsfying the user, you will have to do a user evaluation first. Whithin this thesis we want to analyze, which features an interface should or should not have to meet these requirements of support, user satisfaction and beeing intuitive. After collecting many ideas and arguing with a team of experts, we determined only a few of them. Different combination of these determined variables form the interfaces, we have to investigate in our usability study. The results of the usability leads to the assumption, that autocompletion and suggestion features supports the user. Furthermore coloring tags for grouping them into categories is not disturbing to the user, but has a tendency of being supportive. Same tendencies emerge for an interface consisting of two user interface elements. There is also an example given for the definition differences of being intuitive. This thesis leads to the concolusion that for reasons of user satisfaction and support it is allowed to differ from classical annotation interface features and to implement further usability studies in the section of annotation interfaces.
Large and unknown data sets can be easily and systematically discovered by using faceted search. If implementing applications for smartphones, it needs to be considered that unlike desktop applications you can only use smaller screen sizes and there are limited possibilities for interaction between user and smartphone. These limitations can negatively influence the usability of an application. With FaThumb and MobileFacets, two mobile applications exist, which implement and use faceted search, although only MobileFacets is designed for current smartphones with touchscreen. However, FaThumb provides a novel facet navigation, which is newly realized in MFacets for present smartphones within this work.
Moreover, this work deals with the performance of a summative evaluation between both applications, MFacets and MobileFacets, with regards to usability and presents the evaluated results.
Die folgende Arbeit soll einen Überblick über bestehende Lösungen zur Interaktion in Erweiterten Realitäten (Augmented Reality) schaffen. Hierzu werden anhand dreier grundlegender Betrachtungsweisen unterschiedliche Interaktionskonzepte und -umsetzungen sowohl von der technischen, als auch von der konzeptuellen Seite her, vorgestellt. Neben Fragen der Visualisierung werden unterschiedliche Typen von Benutzungsschnittstellen vorstellt. Den größten Teil nehmen die drei typischen Interaktionsaufgaben Selektion- und Manipulation, Navigation und Systemkontrolle und die damit verbundenen Interaktionstechniken ein. Die Inhalte des Arbeitsberichts beschränken sich auf den Einsatz von Interaktionelementen in Augmented Reality Umgebungen. Dies geschieht in Abgrenzung zu Forschungsarbeiten auf dem Gebiet zu Interaktionstechniken in Virtual Reality Umgebungen (vollimmersiv oder auch desktoporientiert). Zwar standen und stehen viele Interaktionstechniken aus dem Bereich VR in der AR Pate, doch haben sich gerade im Bereich der AR neue Techniken und Konzepte entwickelt. Folglich sollen VR Techniken nur dann betrachtet werden, wenn Sie in AR Anwendungen angewendet wurden bzw. wenn ihre Anwendung sinnvoll erscheint.
Especially e-government projects have a complex stakeholder structure: governments, businesses, non-profits and private stakeholders in different forms and roles are involved or affected by e-government projects. Consequently, the success of ICT projects critically depends on the integration of different stakeholder groups in the design processes of e-government solutions. In this context, stakeholder participation is sought therefore useful initiatives like open government and good governance drive this process forward. Although it is important to analyze the objectives, expectations and power characteristics of project participants and project stakeholders, the determinations of the stakeholder groups through a so called stakeholder analysis are insufficiently methodically developed.
The aim of this bachelor thesis is to pick up the stakeholder theories, stakeholder participation theories, stakeholder analysis and transfer it to an administrative level as well as to check their applicability. This will be done by explaining and systematizing appropriate procedures for the classification and mapping of stakeholder groups, through well-known stakeholder theories of Freeman, Mitchel and Rowley. In order to generate recommendations for future e-government projects, the application of stakeholder theories in two e-government projects with stakeholder involvement will be reviewed. It is also important to work out how stakeholder groups can be actively involved in the developmental processes in concrete e-government projects by using Web 2.0 possibilities. The role of Web 2.0 applications as an additional channel of communication and participation tool will be presented and evaluated.
The result is a guide, which supports successful participation of stakeholders in projects, by analyzing the process for stakeholder identification, stakeholder grouping and prioritization and showing instructions and benefits for using Web 2.0.
Das Einsatzgebiet des im Rahmen von insgesamt drei Studienarbeiten eingeführten "Spontaneous Guest Access (SpoGA) & Extended Invitation Management System (E-IMS)" ist im Firmenumfeld angesiedelt. Gästen einer Firma kann durch Nutzung des Systems ohne großen Aufwand ein Zugang zu einem gesicherten, drahtlosen Firmennetzwerk ermöglicht werden, über welches auch der Zugriff aufs Internet erfolgen kann. Das System soll auch die Planung und Durchführung von Veranstaltungen wie insbesondere Konferenzen erleichtern. SpoGA soll es einem Gastgeber ermöglichen, die Netzwerkgeräte befugter Gäste für einen vorgegebenen Zeitraum automatisiert freizuschalten. Somit kann für bestimmte Geräte komfortabel und ohne großen Aufwand ein temporärer Zugang geschaffen werden, ohne die Sicherheit des Netzes zu beeinflussen. E-IMS soll den Verwaltungsaufwand bei der Organisation und Durchführung von Veranstaltungen so weit wie möglich reduzieren. Durch die Anbindung an SpoGA wird außerdem eine automatisierte Freischaltung der Netzwerkzugänge der Konferenzteilnehmer möglich. Auch Aspekte der Teilnehmer- und Ressourcenverwaltung werden berücksichtigt: automatisierter Versand von Einladungen, selbstständige Buchung der von den Teilnehmern jeweils benötigten Geräte wie Beamer und Projektoren und Übersicht über die angefallenen bzw. laufenden Kosten der Veranstaltung. Gegenstand dieser Arbeit ist die Umsetzung des zweiten Teilbereichs des E-IMS, welcher die Realisierung der Veranstaltungsverwaltung sowie der Funktion zum Abrechnen in Anspruch genommener Ressourcen umfasst. Um Nutzern einen mobilen Zugriff auf die Veranstaltungsdaten zu ermöglichen, wurde die entsprechende Anwendung prototypisch für Mobiltelefone realisiert.
In Silico simulation of biological systems is an important sub area of computational biology (system biology), and becomes more and more an inherent part for research. Therefore, different kinds of software tools are required. At present, a multitude of tools for several areas exists, but the problem is that most of the tools are essentially application specific and cannot be combined. For instance, a software tool for the simulation of biochemical processes is not able to interact with tools for the morphology simulation and vice versa. In order to obtain realistic results with computer-aided simulations it is important to regard the biological system in its entirety. The objective is to develop a software framework, which provides an interface structure to combine existing simulation tools, and to offer an interaction between all affiliated systems. Consequently, it is possible to re-use existing models and simulation programs. Additionally, dependencies between those can be defined. The system is designed to interoperate as an extendable architecture for various tools. The thesis shows the usability and applicability of the software and discusses potential improvements.
Social networking platforms as creativity fostering systems: research model and exploratory study
(2008)
Social networking platforms are enabling users to create their own content, share this content with anyone they invite and organize connections with existing or new online contacts. Within these electronic environments users voluntarily add comments on virtual boards, distribute their search results or add information about their expertise areas to their social networking profiles and thereby share it with acquaintances, friends and increasingly even with colleagues in the corporate world. As a result, it is most likely that the underlying knowledge sharing processes result in many new and creative ideas. The objective of our research therefore is to understand if and how social social networking platforms can enforce creativity. In addition, we look at how these processes could be embedded within the organizational structures that influence innovative knowledge sharing behavior. The basis for our research is a framework which focuses on the relations between intrinsic motivation, creativity and social networking platforms. First results of our empirical investigation of a social software platform called "StudiVZ.net" proved that our two propositions are valid.
SOA-Security
(2007)
This paper is a part of the ASG project (Adaptive Services Grid) and addresses some IT security issues of service oriented architectures. It defines a service-oriented security concept, it explores the SOA security challenge, it describes the existing WS-Security standard, and it undertakes a first step into a survey on best practice examples. In particular, the ASG middleware platform technology (JBossWS) is analyzed with respect to its ability to handle security functions.
The advantage of virtual networks is to test routing-algorithms or distributed software in an easy-to-use, cheap and well-arranged way. Optimized Link State Routing (OLSR) and Babel are two modern routing-algorithms for mobile ad-hoc networks that distinguish faulty links from good link connections in their routing process. Typically virtual networks provide nearly perfect link connections without any network traffic disruptions. The Traffic Control (tc) tool is able to manipulate connections in terms of delay and packet loss.
In this bachelor thesis virtual network scenarios are created with VNUML and Qemu and manipulated with tc in order to analyze the behavior of ad-hoc routing algorithms like Babel and OLSR. The behavior of well-known TCP congestion control algorithms, e.g., TCP Vegas and TCP Reno are analyzed, too.
In the period between 1985 and 1994 ethnic conflicts had a part for more than 50% of the total number of conflicts. Oftentimes, conflicts may arise between different ethnic groups or between one ethnic group and the state. The reasons for this are varied in ranging from social deprivation to the threat of collective identity. An essential key role is played by the political elite, which takes advantage of political and economic crises in order to mobilize certain ethnic groups.
This thesis attempts to investigate such ethnic conflicts on the case of former Yugoslavia, focusing in particular on the role of the political elite. For this purpose, the approach of an agent-based simulation is been followed. The agent-based model, which is carried by ArcGIS, EmIL and Repast allows a thorough investigation of such conflicts. The main feature lies in the ability of agents to be able to develop a normative behavior, so that compex social behaviors can be simulated.
This paper describes results of the simulation of social objects, the dependence of schoolchildren's professional abilities on their personal characteristics. The simulation tool is the artificial neural network (ANN) technology. Results of a comparison of the time expense for training the ANN and for calculating the weight coefficients with serial and parallel algorithms, respectively, are presented.
This paper originates from the FP6 project "Emergence in the Loop (EMIL)" which explores the emergence of norms in artificial societies. Part of work package 3 of this project is a simulator that allows for simulation experiments in different scenarios, one of which is collaborative writing. The agents in this still prototypical implementation are able to perform certain actions, such as writing short texts, submitting them to a central collection of texts (the "encyclopaedia") or adding their texts to texts formerly prepared by other agents. At the same time they are able to comment upon others' texts, for instance checking for correct spelling, for double entries in the encyclopaedia or for plagiarisms. Findings of this kind lead to reproaching the original authors of blamable texts. Under certain conditions blamable activities are no longer performed after some time.
In dieser Arbeit wird ein Datenmodell für Shared Annotations vorgestellt. Basierend auf einem bereits existierenden Datenmodell für Annotationen, wird dieses erweitert um die Möglichkeit der Modellierung von Shared Annotations. Daraufhin werden Funktionen von Shared Annotations erläutert, die über das neue Annotationsmodell möglich sind.
This paper presents a method for the evolution of SHI ABoxes which is based on a compilation technique of the knowledge base. For this the ABox is regarded as an interpretation of the TBox which is close to a model. It is shown, that the ABox can be used for a semantically guided transformation resulting in an equisatisfiable knowledge base. We use the result of this transformation to effciently delete assertions from the ABox. Furthermore, insertion of assertions as well as repair of inconsistent ABoxes is addressed. For the computation of the necessary actions for deletion, insertion and repair, the E-KRHyper theorem prover is used.
Diese Masterarbeit beschreibt und evaluiert die im Rahmen der Arbeit entwickelten Erweiterungen zur Schellings Segregationsimulation. Der Nobelpreisträger Thomas C. Schelling simulierte als erster ab 1969 die Segregation. Die dabei entstandenen Ergebnisse erlaubten eine genauere Analyse der auftretenden Phänomene. Uri Wilensky, Mitarbeiter der "Northwestern University", entwickelte seit 1999 eine Segregationssimulation in NetLogo.
Im Rahmen meiner Masterarbeit habe ich das NetLogo-Model von Uri Wilensky in sechs unterschiedlichen Szenarien weiterentwickelt.
This paper describes the development of security requirements for non-political Internet voting. The practical background is our experience with the Internet voting within the Gesellschaft für Informatik (GI - Informatics Society) 2004 and 2005. The theoretical background is the international state-of-the-art of requirements about electronic voting, especially in the US and in Europe. A focus of this paper is on the user community driven standardization of security requirements by means of a Protection Profile of the international Common Criteria standard.
Distance vector routing protocols are interior gateway protocols in which every router sets up a routing table with the help of the information it receives from its neighboring routers. The routing table contains the next hops and associated distances on the shortest paths to every other router in the network. Security mechanisms implemented in distance vector routing protocols are insufficient. It is rather assumed that the environment is trustworthy. However, routers can be malicious for several reasons and manipulate routing by injecting false routing updates. Authenticity and integrity of transmitted routing updates have to be guaranteed and at the same time performance and benefits should be well-balanced.
In this paper several approaches that aim at meeting the above mentioned conditions are examined and their advantages and disadvantages are compared.
This master thesis deals basically with the design and implementation of a path planning system based on rapidly exploring search trees for general-n-trailers. This is a probabilistic method that is characterized by a fast and uniform exploration. The method is well established, however, has been applied only to vehicles with simple kinematics to date. General-n-trailers represent a particular challenge as their controllability is limited. For this reason the focus of this thesis rests on the application of the mentioned procedure to general-n-trailers. In this context systematic correlations between the characteristics of general-n-trailers and the possibilities for the realization and application of the method are analyzed.
Der neue Personalausweis bietet mit der eID-Funktion die Möglichkeit sich online gegenüber anderen auszuweisen. Dafür ist eine Software notwendig, die auf dem lokalen Computer installiert ist. Diese Arbeit zeigt eine mögliche Vorgehensweise beim Entwurf und der Implementierung der notwendigen Funktionen solch einer Software.
Im Gesundheitswesen geht es nicht nur um die Vertraulichkeit von Patientendaten, sondern auch um ihre Integrität, von der die richtige Behandlung und Pflege des Patienten abhängen. Eine Maßnahme zur Absicherung eines Krankenhausinformationssystems (KIS) gegen unautorisierte Angreifer ist eine flexibel organisierte Zugriffskontrolle zum Schutz der patientenbezogenen Daten. Dieser Artikel beschreibt das Konzept einer rollenbasierten Zugriffskontrolle (RBAC - Role Based Access Control) und seine Einsatzmöglichkeiten im Gesundheitswesen.
This paper describes the robots TIAGo and Lisa used by
team homer@UniKoblenz of the University of Koblenz-Landau, Germany,
for the participation at the RoboCup@Home 2019 in Sydney,
Australia. We ended up first at RoboCup@Home 2019 in the Open Platform
League and won the competition in our league now three times
in a row (four times in total) which makes our team the most successful
in RoboCup@Home. We demonstrated approaches for learning from
demonstration, touch enforcing manipulation and autonomous semantic
exploration in the finals. A special focus is put on novel system components
and the open source contributions of our team. We have released
packages for object recognition, a robot face including speech synthesis,
mapping and navigation, speech recognition interface, gesture recognition
and imitation learning. The packages are available (and new packages
will be released) on http://homer.uni-koblenz.de.
This paper describes the robots TIAGo and Lisa used by team homer@UniKoblenz of the University of Koblenz-Landau, Germany, for the participation at the RoboCup@Home 2018 in Montreal, Canada. Further this paper serves as qualification material for the RoboCup-@Home participation in 2018. A special focus is put on novel system components and the open source contributions of our team. This year the team from Koblenz won the biggest annual scientianc robot competition in Montreal in the RoboCup@Home Open Platform track for the third time and also won the RoboCup@Home German Open for the second time. As a research highlight a novel symbolic imitation learning approach was demonstrated during the annals. The TIAGo robotic research platform was used for the first time by the team. We have released packages for object recognition, a robot face including speech synthesis, mapping and navigation, speech recognition interface via android and a GUI. The packages are available (and new packages will be released) on http://wiki.ros.org/agas-ros-pkg. Further information can be found on our project page http://homer.uni-koblenz.de.
This paper describes the robot Lisa used by team homer@UniKoblenz of the University of Koblenz Landau, Germany, for the participation at the RoboCup@Home 2017 in Nagoya, Japan. A special focus is put on novel system components and the open source contributions of our team. We have released packages for object recognition, a robot face including speech synthesis, mapping and navigation, speech recognition interface via android and a GUI. The packages are available (and new packages will be released) on
http://wiki.ros.org/agas-ros-pkg.
This paper describes the robot Lisa used by team
homer@UniKoblenz of the University of Koblenz Landau, Germany, for the participation at the RoboCup@Home 2016 in Leipzig, Germany. A special focus is put on novel system components and the open source contributions of our team. We have released packages for object recognition, a robot face including speech synthesis, mapping and navigation, speech recognition interface via android and a GUI. The packages are available (and new packages will be released) on http://wiki.ros.org/agas-ros-pkg.
The University of Koblenz-Landau would like to apply for participation in the RoboCup Mixed Reality League in Suzhou, China 2008. Our team is composed of ten team members and two supervisors. All members are graduate students of Computational Visualistics. Our supervisors are Ph.D. candidates currently researching in the working groups of artificial intelligence and computer graphics.
Rescueanalyser
(2007)
Im Forschungsgebiet der Künstlichen Intelligenz und Multiagenten-Systeme stellt die Erkennung von Plänen eine besondere Herausforderung dar. Es gilt aus einer Menge von Beobachtungen Agentenpläne effektiv, zuverlässig und eindeutig zu identifizieren. Im Bereich des RoboCup ist es von entscheidender Bedeutung, dass Agenten über die Pläne und Intensionen anderer Agenten schlussfolgern können, um gemeinsame Ziele zu erreichen. Diese Diplomarbeit umfasst die Konzeption und Entwicklung eines Werkzeugs für die RoboCupRescue-Simulation, welches die Abläufe einer Rescue-Simulation analysiert und die beobachteten Aktionsfolgen Plänen einer vorspezifizierten Planbibliothek zuordnet. Das entwickelte Analysewerkzeug zeichnet sich einmal durch seine Unabhängigkeit von den verschiedenen Rescue-Agententeams aus, sowie durch die Verwendung neuester Methoden zur effektiven Planerkennung. Damit ermöglicht der entwickelte Rescueanalyser sowohl die Fehlersuche und Optimierung des eigenen Agententeams als auch die Erforschung und Auswertung unterschiedlicher Strategien der Rescue-Agenten konkurrierender Teams. Neben der Visualisierung der Ergebnisse mittels eines optimierten Monitors des Rescue-Simulators, können die Resultate und Auswertungen der Planerkennung auch durch textbasierte Ausgaben detailliert eingesehen werden.
The paper is devoted to solving a problem of the development of the website of Russian municipal policlinics and provides a selection of a set of elements which should be posted on a website. Such elements are necessary to provide citizens with correct and ergonomic e-services. The insufficient development of an infrastructure of institutions of public and municipal administration (particularly, healthcare institutions) in Russia made it necessary to analyze webresources used in different countries at different levels of providing medical services. The information resources of medical treatment facilities of the United Kingdom, of the United Statesrnof America and of the Federal Republic of Germany were researched separately for three existing economic models of healthcare. A set of criteria for the assessment of medical webresources was developed.
API migration refers to the change of a used API to a different API in a program. A special case is called wrapper-based API migration. The API change is done without touching the program but the old API is reimplemented by means of the from now on used one. This so called wrapper has the interface of the reimplemented API but uses the implementation of the new one.
This is an interesting approach because the wrapper can be used in each program, which uses the old API.
To make this approach reproducible we study a ranking-based method for implementing a wrapper, where we always implement the method with the highest priority depending on failing test cases. Thus, we can reconstruct each implementation step.
We first develop an infrastructure to run and log test suites of Java projects that use an API, which we want to change.
We then build a wrapper for a given API using the ranking-based approach.
Replikation einer Multi-Agenten-Simulationsumgebung zur Überprüfung auf Integrität und Konsistenz
(2012)
In dieser Master -Arbeit möchte ich zunächst eine Simulation vorstellen, mit der das Verhalten von Agenten untersucht wird, die in einer generierten Welt versuchen zu über leben und dazu einige Handlungsmöglichkeiten zur Auswahl haben. Anschließend werde ich kurz die theoretischen Aspekte beleuchten, welche hier zu Grunde liegen. Der Hauptteil meiner Arbeit ist meine Replikation einer Simulation, die von Andreas König im Jahr 2000 in Java angefertigt worden ist [Kö2000] . Ich werde hier seine Arbeit in stark verkürzter Form darstellen und anschließend auf meine eigene Entwicklung eingehen.
Im Schlussteil der Arbeit werde ich die Ergebnisse meiner Simulation mit denen von Andreas König vergleichen und die verwendeten Werkzeuge (Java und NetLogo) besprechen. Zum Abschluss werde ich in einem Fazit mein Vorhaben kurz zusammenfassen und berichten was sich umsetzen ließ, was nicht funktioniert hat und warum.
Gartner predicts that 33 per cent of total accesses to Business Intelligence (BI) functionalities will be done until 2013 using mobile devices. While mobile devices have already spread to a high extent in private life they have recently become more and more part of the professional life as well and businesses are seeking to develop smart scenarios on how to integrate the new technologies in their processes. The development of mobile software applications is also booming in the field of BI. The research topic of Mobile Business Intelligence (mBI) is focussing on how to expand traditional BI concepts to the new mobility dimension thus enabling to apply BI functionalities on mobile devices, such as smartphones and tablet computers. The increasing significance of mBI for research in combination with the tremendous economic growth potential forecasted for the next years is the driver for writing this thesis. The main aspect of the present thesis is a concept on how to develop a consistent mBI Strategy aligning all structures, processes and stakeholders to one standard set of goals and guidelines. Crucial points when implementing an mBI strategy are the definition of goals, the identification of risks and opportunities, the adaptation of mBI to the user context, the integration of mBI into the existing IT landscape, as well as a proper set-up of system architectures and security concepts. Besides analyzing the aspect of mBI strategy this thesis also considers the market situation of mBI. Clearly defined goals and the systematic development of a consistent mBI strategy which is in line with these goals can be considered as a crucial prerequisite for companies to successfully implement mBI.
Parallelmanipulatoren, welche den Stewartmechanismus nutzen, ermöglichen die präzise Ausführung von Aufgaben in einem begrenzten Arbeitsraum. Durch die Nutzung von sechs Freiheitsgraden wird eine hohe Flexibilität der Positionierung erreicht. Die robuste Konstruktion sorgt zudem für ein sehr gutes Verhältnis von Gewicht zu Nutzlast.
Diese Bachelorarbeit befasst sich mit der Entwicklung einer flexiblen Softwarelösung zur Ansteuerung einer Stewartplattform. Dies umfasst ein Modell der Plattform, welches zu Testzwecken dient. Es werden zunächst die mathematischen Grundlagen der Inversen Kinematik erarbeitet aufbauend auf einem zuvor definierten Bewegungsmodell. Es folgt die Entwicklung einer generischen Architektur zur Übermittlung und Auswertung von Steuerkommandos vom PC. Die Implementierung geschieht in C und wird in verschiedene Module aufgeteilt, welche jeweils einen Aufgabenbereich der Positionskontrolle oder der Hardwarekommunikation abdecken. Es wird zudem eine graphische Nutzeroberfläche vorgestellt, über die man die Position der Plattform manuell verändern kann. Eine automatische Ansteuerung wird im folgenden Anwendungsbeispiel beschrieben, wo die Plattform mit frequentiellen Beschleunigungswerten einer Achterbahnsimulation beliefert wird.
This bachelor thesis deals with the topic "user-friendly design of applications (apps)" on mobile devices, a subdomain of software-ergonomics. In the process, two applications are being analyzed with the aim of developing a solution on how support on a mobile device should be conducted. This study focuses primarily on appropriate gestures to coordinate the 'help function' on a mobile device. The study results show that the test persons request a customized help function, but reject an extensive help description, as this seems to be overwhelming for the user.
In this thesis the possibilities for real-time visualization of OpenVDB
files are investigated. The basics of OpenVDB, its possibilities, as well
as NanoVDB and its GPU port, were studied. A system was developed
using PNanoVDB, the graphics API port of OpenVDB. Techniques were
explored to improve and accelerate a single ray approach of ray tracing.
To prove real-time capability, two single scattering approaches were
also implemented. One of these was selected, further investigated and
optimized to achieve interactive real-time rendering.
It is important to give artists immediate feedback on their adjustments, as
well as the possibility to change all parameters to ensure a user friendly
creation process.
In addition to the optical rendering, corresponding benchmarks were
collected to compare different improvement approaches and to prove
their relevance. Attention was paid to the rendering times and memory
consumption on the GPU to ensure optimal use. A special focus, when
rendering OpenVDB files, was put on the integrability and extensibility of
the program to allow easy integration into an existing real-time renderer
like U-Render.
Querying for meta knowledge
(2008)
The Semantic Web is based on accessing and reusing RDF data from many different sources, which one may assign different levels of authority and credibility. Existing Semantic Web query languages, like SPARQL, have targeted the retrieval, combination and reuse of facts, but have so far ignored all aspects of meta knowledge, such as origins, authorship, recency or certainty of data, to name but a few. In this paper, we present an original, generic, formalized and implemented approach for managing many dimensions of meta knowledge, like source, authorship, certainty and others. The approach re-uses existing RDF modeling possibilities in order to represent meta knowledge. Then, it extends SPARQL query processing in such a way that given a SPARQL query for data, one may request meta knowledge without modifying the query proper. Thus, our approach achieves highly flexible and automatically coordinated querying for data and meta knowledge, while completely separating the two areas of concern.
Within the scope of this bachelor thesis, a survey targeting the alumni of the Department 4: Computer Science of the University of Koblenz-Landau was planned, realized and evaluated. The goal was to support the Task Force Bachelor Master that was in charge of the re-accreditation process of the study courses. At first, the theoretical fundament of the survey design was acquired via desk research. Moreover, the analysis of past surveys of similar character lead to getting an impression of the requirements. Under consideration of recent changes, a survey prototype was created and improved following the insights from a pre-test. Using the open source tool LimeSurvey, the final version was implemented. The platform was used as the technical basis of the survey. The recipients included members of alumni clubs as well as recent alumni from the last years.
The survey lead to insights about the satisfaction of the alumni with their study course and the study situation in general. Furthermore, the opinion regarding two new master courses, E-Government and Web Science, was requested. The feedback from four of the study courses was enough to give significant results, for the other five courses it was only possible to interpret the general statements. All in all, there was a high rate of satisfaction with the studies.
Additionally, it was possible to collect suggestions for improvements and criticism. The main topics were internationality, emphasis on study topics, freedom of choice/specialization and relevancy to practice. As a result of the survey, a recommendation was verbalized, that should lead to an improvement in quality and need of the teaching in the Department 4: Computer Science in combination with the detailed results.
Im Rahmen dieser Abschlußarbeit wurde ein Plugin zur Visualisierung/Simulation von Public-Key-Infrastrukturen für die Kryptographie-Lernsoftware JCryp-Tool entwickelt und implementiert. Public-Key-Infrastrukturen stellen einen Bereich in der Kryptographie dar, mit dem viele Anwender von IT-Systemen in Berührung kommen.
In dieser Studienarbeit werden neben den Grundlagen der Web Services, Komponenten und APIs zur Realisierung des Sticky Loggings aufgezeigt. Es wird ein Szenario zum Testen des Sticky Loggings beschrieben und als Web Services implementiert. Der Sticky-Logging-Formalismus wird erklärt und es wird eine API zur Erstellung der StickyLogs implementiert. Die StickyLogs werden innerhalb des SOAP-Attachments der SOAP-Nachrichten zwischen den Web Services ausgetauscht. Dazu wird eine Realisierung mit einem Messagehandler unter JAX-WS programmiert und erläutert.
Echtzeitsysteme spielen in der heutigen Zeit in vielen Anwendungsbereichen eine bedeutende Rolle. Beispiele hierfür finden sich in der Robotik, der Fabrikautomation, der Medizintechnik, aber auch in Bereichen aus dem täglichen Leben, wie im Automobilbereich oder in der Mobilkommunikation. Trotz der ständig wiederkehrenden Begegnung mit ihnen im Alltag, sind Echtzeitsysteme häufig innerhalb von größeren Systemen eingebettet und können somit nur selten vom jeweiligen Nutzer bewusst als eigenständig wahrgenommen werden.
Um spezifische Eigenschaften des Gebiets Echtzeitsysteme im Schulunterricht oder in der Ausbildung an Hochschulen den Lernenden sichtbar zu machen, existiert im Labor der Arbeitsgruppe Echtzeitsysteme der Universität Koblenz-Landau ein Versuchsaufbau mit dem Namen "Wippe", bei dem eine Kamera die Bewegung einer Kugel auf einer ebenen Fläche aufnehmen und vermessen kann. Durch die Neigung der Fläche in zwei Achsen ist die Wippe prinzipiell in der Lage, die Kugel zu bewegen und zum Halten zu bringen. Insbesondere soll verhindert werden, dass die Kugel von der Fläche fällt. Die vorliegende Arbeit beschäftigt sich mit der Analyse und Überarbeitung des Wippe-Systems. Außerdem wird ein Regelalgorithmus für eine automatische Steuerung des Systems realisiert.
Das Wippe-Experiment soll für didaktische Zwecke eingesetzt werden und ein offenes System darstellen, welches die Merkmale und Eigenschaften eines Echtzeitsystems exemplarisch verdeutlicht.
In dieser Ausarbeitung beschreibe ich die Ergebnisse meiner Untersuchungen zur Erweiterung des LogAnswer-Systemsmit nutzerspezifischen Profilinformationen. LogAnswer ist ein natürlichsprachliches open-domain Frage-Antwort-System. Das heißt: es beantwortet Fragen zu beliebigen Themen und liefert dabei konkrete (möglichst knappe und korrekte) Antworten zurück. Das System wird im Rahmen eines Gemeinschaftsprojekts der Arbeitsgruppe für künstliche Intelligenz von Professor Ulrich Furbach an der Universität Koblenz-Landau und der Arbeitsgruppe Intelligent Information and Communication Systems (IICS) von Professor Hermann Helbig an der Fernuniversität Hagen entwickelt. Die Motivation meiner Arbeit war die Idee, dass der Prozess der Antwortfindung optimiert werden kann, wenn das Themengebiet, auf das die Frage abzielt, im Vorhinein bestimmt werden kann. Dazu versuchte ich im Rahmen meiner Arbeit die Interessensgebiete von Nutzern basierend auf Profilinformationen zu bestimmen. Das Semantic Desktop System NEPOMUK wurde verwendet um diese Profilinformationen zu erhalten. NEPOMUK wird verwendet um alle Daten, Dokumente und Informationen, die ein Nutzer auf seinem Rechner hat zu strukturieren. Dazu nutzt das System ein sogenanntes Personal Information Model (PIMO) in Form einer Ontologie. Diese Ontologie enthält unter anderem eine Klasse "Topic", welche die wichtigste Grundlage für das Erstellen der in meiner Arbeit verwendeten Nutzerprofile bildete. Konkret wurde die RDF-Anfragesprache SPARQL verwendet, um eine Liste aller für den Nutzer relevanten Themen aus der Ontologie zu filtern. Die zentrale Idee meiner Arbeit war es nun diese Profilinformationen zur Optimierung des Ranking von Antwortkandidaten einzusetzen. In LogAnswer werden zu jeder gestellten Frage bis zu 200 potentiell relevante Textstellen aus der deutschen Wikipedia extrahiert. Diese Textstellen werden auf Basis von Eigenschaften (wie z.B. lexikalische Übereinstimmungen zwischen Frage und Textstelle) geordnet, da innerhalb des zur Verfügung stehenden Zeitlimits nicht alle Kandidaten bearbeitet werden können.
Mein Ansatz verfolgte das Ziel, diesen Algorithmus durch Nutzerprofile so zu erweitern, dass Antwortkandidaten, welche für den Benutzer relevante Informationen enthalten, höher in der Rangfolge eingeordnet werden. Zur Umsetzung dieser Idee musste eine Methode gefunden werden, um zu bestimmen ob ein Antwortkandidat mit dem Profil übereinstimmt. Da sich die in einer Textstelle enthaltenen Informationen in den meisten Fällen auf das übergeordnete Thema des Artikels beziehen, ohne den Namen des Artikels explizit zu erwähnen, wurde in meiner Implementierung der Artikelname betrachtet, um zu ermitteln, zu welchem Themengebiet die Textstelle Informationen liefert. Als zusätzliches Hilfsmittel wurde außerdem die DBpedia-Ontologie eingesetzt, welche die Informationen der Wikipedia strukturiert im RDF Format enthält. Mit Hilfe dieser Ontologie war es möglich, jeden Artikel in Kategorien einzuordnen, die dann mit den im Profil enthaltenen Stichworten verglichen wurden. Zur Untersuchung der Auswirkungen des Ansatzes auf das Ranking-Verfahren wurden mehrere Testläufe mit je 200 Testfragen durchgeführt. Die erste Testmenge bestand aus zufällig ausgewählten Fragen, die mit meinem eigenen Nutzerprofil getestet wurden. Dieser Testlauf lieferte kaum nutzbare Ergebnisse, da nur bei 29 der getesteten Fragen überhaupt ein Antwortkandidat mit dem Profil in Verbindung gebracht werden konnte. Außerdem konnte eine potentielle Verbesserung der Ergebnisse nur bei einer dieser 29 Fragen festgestellt werden, was zu der Schlussfolgerung führte, dass der Einsatz von Profildaten nicht für Anwendungsfälle geeignet ist, in denen die Fragen keine Korrelation mit dem genutzten Profil aufweisen.
Da die Grundannahme meiner Arbeit war, dass Nutzer in erster Linie Fragen zu den Interessensgebieten stellen, welche sich aus ihrem Profil ableiten lassen, sollten die weiteren Testläufe genau diesen Fall beleuchten. Dazu wurden 200 Testfragen aus dem Bereich Sport ausgewählt und mit einem Profil getestet, welches Stichworte zu unterschiedlichen Sportarten enthielt. Die Tests mit den Sportfragen waren wesentlich aussagekräftiger. Auch hier deuteten die Ergebnisse darauf hin, dass der Ansatz kein großes Potential zur Verbesserung des Rankings hat. Eine genauere Betrachtung einiger ausgewählter Beispiele zeigte allerdings, dass die Integration von Profildaten für bestimmte Anwendungsfälle, wie z.B. offene Fragen für die es mehr als eine korrekte Antwort gibt, durchaus zu einer Verbesserung der Ergebnisse führen kann. Außerdem wurde festgestellt, dass viele der schlechten Ergebnisse auf Inkosistenzen in der DBpedia-Ontologie und grundsätzliche Probleme im Umgang mit Wissensbasen in natürlicher Sprache beruhen.
Die Schlussfolgerung meiner Arbeit ist, dass der in dieser Arbeit vorgestellte Ansatz zur Integration von Profilinformationen für den aktuellen Anwendungsfall von LogAnswer nicht geeignet ist, da vor allem Faktenwissen aus sehr unterschiedlichen Domänen abgefragt wird und offene Fragen nur einen geringen Anteil ausmachen.
In dieser Diplomarbeit wurde ein System entwickelt, dass eine Navigation für Fußgänger ermöglicht. Wie beabsichtigt wurde das System für die Benutzung auf einem iPhone realisiert.
Obwohl die Karten der neuen Generation vonNAVTEQ sich noch in Entwicklung befinden, konnten erste Eindrücke gesammelt werden, wie die Navigation der Fußgänger in Zukunft aussehen wird. Das System arbeitet aber auch mit Kartendaten von OpenStreetMap, die die klassische Repräsentation der Kartendaten haben. Die Positionierung kann später, wenn der Galileo-Empfänger zur Verfügung steht, umgestellt werden und Positionsdaten mit höherer Präzision für die Navigation bereitstellen. Die Routenberechnung konnte mit einem CH-ähnlichen Verfahren durch Vorberechnung beschleunigt werden und erlaubt trotzdem eine Änderung des Profils, ohne dass eine neue Vorberechnung nötig ist. Anders als beim einfachen CH, wo nach einer Änderung des Profils für die Berechnung der Routenkosten eine neue Vorberechnung nötig ist, die im Vergleich zu der Berechnung einer Route aufwendig ist. In Abbildung 8.1 ist die demonstrative Navigation mit dem fertigen System abgebildet. Diese zeigt eine Berechnung der Routen für 3 unterschiedliche Profile, die unterschiedliche Steigungsgrade bevorzugen. Daraus ergeben sich unterschiedliche Routen.
Process Commodities : Entwicklung eines Reifegradmodells alsBasis für Outsourcingentscheidungen
(2009)
Der vorliegende Arbeitsbericht behandelt die Anforderungen an Prozessmanagement als Basis für Outsourcingentscheidungen. Zu dieser Fragestellung erschien im Jahr 2005 ein Artikel von Thomas Davenport mit dem Titel "The Coming Commoditization of Processes". In diesem wird auf das Fehlen eines einheitlichen Verständnisses über Inhalte und Beschreibungen von Prozessen und Prozessschritten, Leistungsindikatoren sowie Prozessmanagementpraktiken hingewiesen. Der vorliegende Arbeitsbericht versucht diese Lücke in Teilen zu schließen, indem er ein Reifegradmodell für Prozessmanagement vorstellt. Mit Hilfe des Modells können Unternehmen einschätzen, in wieweit sie vorhandene Methoden des Prozessmanagements umsetzen und damit die Grundlage für Prozessoutsourcing geschaffen haben.
The natural and the artificial environment of mankind is of enormous complexity, and our means of understanding this complex environment are restricted unless we make use of simplified (but not oversimplified) dynamical models with the help of which we can explicate and communicate what we have understood in order to discuss among ourselves how to re-shape reality according to what our simulation models make us believe to be possible. Being both a science and an art, modelling and simulation isrnstill one of the core tools of extended thought experiments, and its use is still spreading into new application areas, particularly as the increasing availability of massive computational resources allows for simulating more and more complex target systems.
In the early summer of 2012, the 26th European Conference on Modelling andrnSimulation (ECMS) once again brings together the best experts and scientists in the field to present their ideas and research, and to discuss new challenges and directions for the field.
The 2012 edition of ECMS includes three new tracks, namely Simulation-BasedrnBusiness Research, Policy Modelling and Social Dynamics and Collective Behaviour, and extended the classical Finance and Economics track with Social Science. It attracted more than 110 papers, 125 participants from 21 countries and backgrounds ranging from electrical engineering to sociology.
This book was inspired by the event, and it was prepared to compile the most recent concepts, advances, challenges and ideas associated with modelling and computer simulation. It contains all papers carefully selected from the large number of submissions by the programme committee for presentation during the conference and is organised according to the still growing number tracks which shaped the event. The book is complemented by two invited pieces from other experts that discussed an emerging approach to modelling and a specialised application. rnrnWe hope these proceedings will serve as a reference to researchers and practitioners in the ever growing field as well as an inspiration to newcomers to the area of modelling and computer simulation. The editors are honoured and proud to present you with this carefully compiled selection of topics and publications in the field.
The paper deals with a specific introduction into probability propagation nets. Starting from dependency nets (which in a way can be considered the maximum information which follows from the directed graph structure of Bayesian networks), the probability propagation nets are constructed by joining a dependency net and (a slightly adapted version of) its dual net. Probability propagation nets are the Petri net version of Bayesian networks. In contrast to Bayesian networks, Petri nets are transparent and easy to operate. The high degree of transparency is due to the fact that every state in a process is visible as a marking of the Petri net. The convenient operability consists in the fact that there is no algorithm apart from the firing rule of Petri net transitions. Besides the structural importance of the Petri net duality there is a semantic matter; common sense in the form of probabilities and evidencebased likelihoods are dual to each other.
Probability propagation nets
(2007)
A class of high level Petri nets, called "probability propagation nets", is introduced which is particularly useful for modeling probability and evidence propagation. These nets themselves are well suited to represent the probabilistic Horn abduction, whereas specific foldings of them will be used for representing the flows of probabilities and likelihoods in Bayesian networks.
Potential vermögender Kunden
(2011)
This bachelor thesis deals with the potential of wealthy clients in the private banking sector, as well as with financial services offered in this domain. For this purpose a comparison of the international potential in the German-speaking world was conducted which was ascertained with the help of statistical series.
The reference period for this comparison extends over thernyears from 2000 to 2009. Especially in this time period one was able to notice capital restructuring due to the finance crisis and its impacts. The results out of the analysis of the changing potential are then used to find possibilities of reorganizing offered financial services in the private banking sector. While doing so, the achievement of a high service level has the highest priority as this is the main reason for customers to remain with a certain bank. In addition to that this thesis also mentions new possibilities concerning the banking sector as well as consequences of different models of consultant payments.
Pose-Tracking
(2009)
Die bildbasierte automatische Bestimmung der Pose, d. h. der Position und Blickrichtung, einer Kamera in derWelt, ist eine relevante, aber immer noch unzureichend gelöste Aufgabe im Rechnersehen. In dem diesem Bericht zugrunde liegenden Projekt werden höhere markante Merkmale in den Bildern sicherer lokalisiert, sowie deren semantische Signifikanz vom Rechner bestimmt. Eine Posebestimmung wird durch eine Registrierung der elementaren Bestandteile dieser markanten Merkmale im Bild mit Merkmalen im 3-D-Modell erreicht. Dazu werden neue Algorithmen zur Detektion, Lokalisation und Registrierung der markanten Merkmale entwickelt bzw. vorhandene Algorithmen weiter verbessert. Modelle, wie sie aus der Rekonstruktion von Mehrfachansichten entstehen, werden durch weitere Semantik angereichert. Als Anwendungsszenario wird die Posebestimmung auf dem Campusgelände unter Verwendung von Bildern und einem semantischen CG-Modell des Campus gewählt. Die allgemeinen Verfahren werden an diesem Beispiel getestet und ihre Tragfähigkeit wird in Experimenten belegt. Im modularen System entstehen problemunabhängige Einzelbausteine zur Detektion markanter Merkmale und zur 3-D-Rekonstruktion und Posebestimmung aus Merkmalen, Punkten und Linien. Damit wird belegt, dass eine robuste Detektion markanter Merkmale möglich ist und zu einer effektiven Rekonstruktion und Posebestimmung auch in teilweise wenig strukturierten Außengebieten genutzt werden kann.
Polsearchine: Implementation of a policy-based search engine for regulating information flows
(2013)
Many search engines regulate Internet communication in some way. It is often difficult for end users to notice such regulation, as well as obtaining background information for it. Additionally, the regulation can usually be circumvented easily. This bachelor thesis presents the prototypical metasearch engine "Polsearchine" for addressing these weaknesses. Its regulation is established through InFO, a model for regulating information flows developed by Kasten and Scherp. More precisely, the extension for regulating search engines SEFCO is being used. For retrieving search results, Polsearchine uses an external search engine API. The API can be interchanged easily to make this metasearch engine independent from one specific API.
Today you can find smartphones everywhere. This situation created a hype for Augmented Reality and AR Apps. The big question is: Do these applications provide a real added value? To make AR pratically it is important to add the computational power of a computer to the advantages of AR. An easy and fast way of interaction is essential.
A Poker-Assistance-Software is an ideal test area for an AR Application with real added value. The estimation of the winning probability and a fast automated tracking of the playing cards is the perfect field of investigation.
In this discussion it is interesting to evaluate the added value of AR Applications in common.
Human detection is a key element for human-robot interaction. More and more robots are used in human environments, and are expected to react to the behavior of people. Before a robot can interact with a person, it must be able to detect it at first. This thesis presents a system for the detection of humans and their hands using a RGB-D camera. First, a model based hypotheses for possible positions of humans are created to recognize a person. By using the upper parts of the body are used to extract, new features based on relief and width of a person- head and shoulders are extracted. The hypotheses are checked by classifying the features with a support vector machine (SVM). The system is able to detect people in different poses. Both sitting and standing humans are found, by using the visible upper parts of the person. Moreover, the system is able to recognize if a human is facing or averting the sensor. If the human is facing the sensor, the color information and the distance between hand and body are used to detect the positions of the person- hands. This information is useful for gestures recognition and thus can further enhances human-robot interaction.
Diese Diplomarbeit beschreibt die Entwicklung einer Web Service orientierten Umgebung zur Verarbeitung von Pegeldaten mehrerer Gewässer in Zusammenarbeit mit der Bundesanstalt für Gewässerkunde (BfG). Sie dokumentiert den gesamten Entwicklungsablauf und gibt eine Einführung in die verwendeten Technologien. Für XHydro, ein Austauschformat zum wirtschaftlichen und organisationsübergreifenden Austausch von Zeitreihen, wird zusätzlich eine Bewertung des Systems inklusive Verbesserungsvorschlägen gegeben.
This bachelor thesis deals with the comparison related to the similarity of recorded WiFi patterns during the tracing of a path through the streets of a large city. Both MAC address only comparison has been investigated as well as the incorporation of RSSI values, whereby the localization accuracy has been evaluated. Methods for the detection of different types and combinations of loops in the path are demonstrated likewise the attempt to estimate the degree of urban development in the environment of the user by assessing the received signal strength and signal-to-noise ratio of GPS satellites and GSM cell towers.
In order to observe a user- proximity to a certain spot on a large public square the absorption of WiFi signals by the human body has been taken into account. Finally, the results of a comparison of the computing performance of a modern smartphone versus the alternative of remote calculation on a server including data transmission via cellular data network are presented.
Die Arbeit stellt Path Tracing zum Rendern von Bildern mitrnglobaler Beleuchtung vor. Durch die Berechnung der Rendergleichung, mithilfe von Zufallsexperimenten, ist das Verfahren physikalisch plausibel. Entscheidend für die Qualität der Ergebnisse ist Sampling. Der Schwerpunktrnder Arbeit ist die Untersuchung verschiedener Samplingstrategien. Dazu werden die Ergebnisse unterschiedlicher Dichtefunktionen verglichenrnund die Methoden bewertet. Außerdem werden Effekte, wie beispielsweise Depth of Field, mittels Sampling visualisiert.
In this diploma thesis a skeleton-based matching technique for 2D shapes is introduced. First, current approaches for the matching of shapes will be presented. The basics of skeleton-based matchings will be introduced. In the context of this thesis, a skeleton-based matching approach was implemented as presented in the original paper. This implementation is evaluated by performing a similarity search in three shape databases. Strengths and limitations of the approach are pointed out. In addition, the introduced algorithm will be examined with respect to extending it towards matching of 3D objects. In particular, the approach is applied to medical data sets: Pre- and postoperative CT images of the abdominal aorta of one patient will be compared. Problems and approaches for matching of 3D objects in general and blood vessels in particular will be presented.
This thesis addresses the implementation of a particle simulation of an explosion. The simulation will be displayed via ray tracing in near real time. The implementation makes use of the openCL standard. The focus of research in this thesis is to analyse the performance of this combination of components.
The processing of data is often restricted by contractual and legal requirements for protecting privacy and IPRs. Policies provide means to control how and by whom data is processed. Conditions of policies may depend on the previous processing of the data. However, existing policy languages do not provide means to express such conditions. In this work we present a formal model and language allowing for specifying conditions based on the history of data processing. We base the model and language on XACML.
Opinion Mining : Using Twitter as a source of opinion for the prediction of stock market prices
(2012)
Neben den theoretischen Grundkonzepten der automatisierten Fließtextanalyse, die das Fundament dieser Arbeit bilden, soll ein Überblick in den derzeitigen Forschungsstand bei der Analyse von Twitter-Nachrichten gegeben werden. Hierzu werden verschiedene Forschungsergebnisse der, derzeit verfügbaren wissenschaftlichen Literatur erläutert, miteinander verglichen und kritisch hinterfragt. Deren Ergebnisse und Vorgehensweisen sollen in unsere eigene Forschung mit eingehen, soweit sie sinnvoll erscheinen. Ziel ist es hierbei, den derzeitigen Forschungsstand möglichst gut zu nutzen.
Ein weiteres Ziel ist es, dem Leser einen Überblick über verschiedene maschinelle Datenanalysemethoden zur Erkennung von Meinungen zu geben. Dies ist notwendig, um die Bedeutung der im späteren Verlauf der Arbeit eingesetzten Analysemethoden in ihrem wissenschaftlichen Kontext besser verstehen zu können. Da diese Methoden auf verschiedene Arten durchgeführt werden können, werden verschiedene Analysemethoden vorgestellt und miteinander verglichen. Hierdurch soll die Machbarkeit der folgenden Meinungsauswertung bewiesen werden. Um eine hinreichende Genauigkeit bei der folgenden Untersuchung zu gewährleisten, wird auf ein bereits bestehendes und evaluiertes Framework zurückgegriffen. Dieses ist als API 1 verfügbar und wird daher zusätzlich behandelt. Der Kern Inhalt dieser Arbeit wird sich der Analyse von Twitternachrichten mit den Methoden des Opinion Mining widmen.
Es soll untersucht werden, ob sich Korrelationen zwischen der Meinungsausprägung von Twitternachrichten und dem Börsenkurs eines Unternehmens finden lassen. Es soll dabei die Stimmungslage der Firma Google Inc. über einen Zeitraum von einem Monat untersucht und die dadurch gefunden Erkenntnisse mit dem Börsenkurs des Unternehmens verglichen werden. Ziel ist es, die Erkenntnisse von (Sprenger & Welpe, 2010) und (Taytal & Komaragiri, 2009) auf diesem Gebiet zu überprüfen und weitere Fragestellungen zu beantworten.
Ziel dieser Forschungsarbeit ist die Auswahl und Evaluierung von Open-Source ERPSystemen auf effiziente Anwendbarkeit in Unternehmen zum Zwecke des Aufbaus eines "ERP-Future-Labs", in welchem mittelständische Handelsunternehmen das/die installierte(n) System(e) testen können. Den Projektabschluss bildet hierbei die Installation eines lauffähigen Systems, auf welchem die vom Auftraggeber vorgegebenen Geschäftsprozesse abgewickelt werden können. Ferner sollen die Auftraggeber auf dem System geschult, eine Dokumentation der Software (Installation/Bedienung) und des Projekts erstellt werden.
This work addresses the challenge of calibrating multiple solid-state LIDAR systems. The study focuses on three different solid-state LIDAR sensors that implement different hardware designs, leading to distinct scanning patterns for each system. Consequently, detecting corresponding points between the point clouds generated by these LIDAR systems—as required for calibration—is a complex task. To overcome this challenge, this paper proposes a method that involves several steps. First, the measurement data are preprocessed to enhance its quality. Next, features are extracted from the acquired point clouds using the Fast Point Feature Histogram method, which categorizes important characteristics of the data. Finally, the extrinsic parameters are computed using the Fast Global Registration technique. The best set of parameters for the pipeline and the calibration success are evaluated using the normalized root mean square error. In a static real-world indoor scenario, a minimum root mean square error of 7 cm was achieved. Importantly, the paper demonstrates that the presented approach is suitable for online use, indicating its potential for real-time applications. By effectively calibrating the solid-state LIDAR systems and establishing point correspondences, this research contributes to the advancement of multi-LIDAR fusion and facilitates accurate perception and mapping in various fields such as autonomous driving, robotics, and environmental monitoring.
The measurement of the benefits of collaborative technologies is a challenge caused by sociological influences on these systems. In this work, the introduction of a collaborative system into an existing IT environment is examined, so that a value can be identified for the company. The analysis of frameworks has shown that the benefits can be reaped through better management and organization of in-company resources and optimization of the underlying process.
Recipients" youtube comments to the five most successful songs of 2011 and 2012 are tested for nostalgic content. These nostalgic relevant comments are analyzed by content and finally interpreted. It shall be found out, whether nostalgic music content is a factor for success. By using the uses and gratifications theory the recipients" purpose of consuming nostalgic-evoking music will be identified. Music is a clearly stronger trigger for evoking nostalgia than the music video whereas nostalgia triggers positive and/or negative affect. Furthermore personal nostalgia is much more evident than historical nostalgia. Moreover the lyrics have a considerably higher potential to elicit nostalgia than any other song units. Persons and momentous events are the most frequent objects in personal nostalgic reverie. The purpose of consuming nostalgic music is the intended evocation of positive and/or negative affect. Hence nostalgia in music seems to satisfy certain needs and it can be assumed that nostalgia is a factor of success in music industry.
Due to the increasing pervasiveness of the mobile web, it is possible to send and receive mails with mobile devices. Content of digital communication should be encrypted to prevent eavesdropping and manipulation. Corresponding procedures use cryptographic keys, which have to be exchange previously. It has to be ensured, that a cryptographic key really belongs to the person, who it is supposedly assigned to. Within the scope of this thesis a concept for a smartphone application to exchange cryptographic keys was designed. The concept consists of a specification of a component-based framework, which can be used to securely exchange data in general. This framework was extended and used as the basis for a smartphone application. The application allows creating, managing and exchanging cryptographic keys. The Near Field Communication is used for the exchange. Implemented security measures prevent eavesdropping and specific manipulation. In the future the concept and the application can be extended and adjusted to be used in other contexts.
QEMU ist eine Open-Source Virtualisierungssoftware, durch die Computersysteme simuliert werden können. Durch VDE (Virtual Distributed Ethernet) ist es möglich, mehrere durch QEMU virtualisierte Computer miteinander zu verbinden und so ein virtuelles Netzwerk zu erzeugen. Mithilfe von Virtualisierung können Netze zum Testen von Netzwerkanwendungen und -protokollen erzeugt werden, deren Aufbau oder Nutzung ohne Virtualisierung impraktikabel oder schlicht unerschwinglich wäre. Ziel dieser Arbeit ist es, ein Programm zu entwickeln, welches eine in einer Szenario-Datei hinterlegte Netzwerktopologie mit QEMU und VDE-Switches aufbauen kann. Es soll ein Vergleich angestellt werden zwischen Netzen, die mit QEMU aufgebaut wurden und solchen, die über VNUML aufgebaut wurden.
Networked RDF graphs
(2007)
Networked graphs are defined in this paper as a small syntactic extension of named graphs in RDF. They allow for the definition of a graph by explicitly listing triples as well as by SPARQL queries on one or multiple other graphs. By this extension it becomes possible to define a graph including a view onto other graphs and to define the meaning of a set of graphs by the way they reference each other. The semantics of networked graphs is defined by their mapping into logic programs. The expressiveness and computational complexity of networked graphs, varying by the set of constraints imposed on the underlying SPARQL queries, is investigated. We demonstrate the capabilities of networked graphs by a simple use case.
Hybrid automata are used as standard means for the specification and analysis of dynamical systems. Several researches have approached them to formally specify reactive Multi-agent systems situated in a physical environment, where the agents react continuously to their environment. The specified systems, in turn, are formally checked with the help of existing hybrid automata verification tools. However, when dealing with multi-agent systems, two problems may be raised. The first problem is a state space problem raised due to the composition process, where the agents have to be parallel composed into an agent capturing all possible behaviors of the multi-agent system prior to the verification phase. The second problem concerns the expressiveness of verification tools when modeling and verifying certain behaviors. Therefore, this paper tackles these problems by showing how multi-agent systems, specified as hybrid automata, can be modeled and verified using constraint logic programming(CLP). In particular, a CLP framework is presented to show how the composition of multi-agent behaviors can be captured dynamically during the verification phase. This can relieve the state space complexity that may occur as a result of the composition process. Additionally, the expressiveness of the CLP model flexibly allows not only to model multi-agent systems, but also to check various properties by means of the reachability analysis. Experiments are promising to show the feasibility of our approach.
Das Ziel dieser Arbeit ist die Entwicklung eines MP3-Players mit SD-Karten-Slot. Zudem soll er ein Farbdisplay besitzen und über eine komfortable Bedienmöglichkeit verfügen. In dieser Dokumentation wird die verwendete Hardware und ihr Zusammenspiel erklärt. Die für die Realisierung notwendige Software wird vorgestellt und deren Benutzung mithilfe von Codebeispielen verdeutlicht. Insbesondere wird auf die extra für dieses Projekt entworfenen Codebausteine eingegangen. Dies soll die Wiederverwendung der Bausteine in anderen Projekten vereinfachen.
In this thesis, a first prototype of a mobile instruction device with mixed reality (MR) funktionality is developed. This system shall be capable to support training on the job through interaction with the work item. The concept corresponds to a didactic approach presented by Martens-Parree that combines constructivism with situated learning. As an application example, the training of glider pilots checked out on a new type was chosen. Whether the MR device could increase the competence, or facilitiate the completion of certain tasks, was examined in a survey with fifteen testers. The results of the study show that in general, the didactic approach of Martens-Parree is valid. While an increase in fact knowledge has been observed, it was not (yet) possible to demonstrate an increase in skills with respect to the work tasks.
The impact of innovation in order to secure a company- existence is increasing. Especially in the field of information and communication systems, modified conditions, reduced product life cycles (in some markets e.g. in the software market down to a few months), and increased competition force many companies to adapt their innovation speed. Long-term profit will just be obtained by those companies that will manage to consolidate or further to enlarge their competitive position with technological innovation. The increasing relevance of new developments requires a separate treatment of technological innovations and demands for a technology and innovation management. Not every innovation is thereby a result of a single, spontaneous idea. Many innovations evolve from the integration or combination of already existing technologies. Especially in markets with insignificant growing rates, this kind of innovation is quite important. The purpose of this article is a case-oriented identification of existing deficiencies within established recommendations for a strategic technology and innovation management and the discussion about the role of modularity and technology integration as an alternative way to innovate.