Refine
Year of publication
- 2010 (27) (remove)
Document Type
- Part of Periodical (11)
- Doctoral Thesis (4)
- Study Thesis (4)
- Diploma Thesis (3)
- Master's Thesis (3)
- Bachelor Thesis (2)
Language
- English (27) (remove)
Keywords
- API (2)
- hybrid automata (2)
- multi-agent systems (2)
- ontology (2)
- Abbildung <Mathematik> (1)
- Absolutismus (1)
- Agentenorientiertes Software Engineering (1)
- Algorithm Engineering (1)
- Artificial Neural Networks (1)
- Autoritarismus (1)
- Beschaffung (1)
- Bildanalyse (1)
- Biometric Authentication (1)
- Biotransformation (1)
- Container Entity Modell (1)
- Context-aware processes (1)
- Core Ontology on Multimedia (1)
- Datenaustausch (1)
- Decision-support (1)
- Description Logic (1)
- Description Logics (1)
- Digitale Bilder (1)
- Dijkstras Algorithmus (1)
- Distributed process execution (1)
- EU (1)
- Eclipse <Programmierumgebung> (1)
- Entscheidungsunterstützung (1)
- Europäischer Schadensbericht (1)
- Fingerprint Recognition (1)
- Formale Ontologie (1)
- ICM (1)
- JML (1)
- Java Modeling Language (1)
- Java. Programmiersprache (1)
- Knowledge Compilation (1)
- Kollektivismus (1)
- Künstliche Neuronale Netze (1)
- Mapping <Mathematics> (1)
- Maßtheorie (1)
- Measure-theory (1)
- Metamodel (1)
- Model-Driven Engineering (1)
- Multi-Agenten-Systeme (1)
- Multiagentensysteme (1)
- Multimedia Metadata Ontology (1)
- Nationalismus (1)
- Networks (1)
- Neuronales Netz (1)
- OCL <Programmiersprache> (1)
- OWL (1)
- OWL <Informatik> (1)
- Ontologie. Wissensverarbeitung (1)
- Ontology (1)
- Ontology API model (1)
- Oracle Generation (1)
- Oraklegenerierung (1)
- PEPPOL (1)
- POIs (1)
- Pan European Public Procurement OnLine (1)
- Petri net (1)
- Pfadplanung (1)
- Plug in (1)
- Probability propagation nets (1)
- Process tracing (1)
- Religiosität (1)
- RoboCup (1)
- Robotik (1)
- STOF Model (1)
- Service-orientierte Architektur (1)
- Software Engineering (1)
- Softwaretest (1)
- Softwaretesting (1)
- Sozialwissenschaftliche Simulation (1)
- Speaker Recognition (1)
- Specification (1)
- Spezifikation (1)
- Test Generation (1)
- Testgenerierung (1)
- Text (1)
- Texterkennung (1)
- UML (1)
- Unterwasserwelt (1)
- VCD (1)
- Virtual Company Dossier (1)
- absolutism (1)
- application programming interfaces (1)
- aquatic environment (1)
- authoritarianism (1)
- backpropagation (1)
- bias (1)
- biotransformation (1)
- collectivism (1)
- constraint logic programming (1)
- cultural dimensions (1)
- distributed information systems (1)
- events (1)
- hybride Automaten (1)
- iCity project (1)
- mPayments (1)
- micro-agent (1)
- mobile interaction (1)
- mobile phone (1)
- model-driven engineering (1)
- modulares System (1)
- multiagent systems (1)
- multimedia metadata (1)
- nationalism (1)
- networks (1)
- neural (1)
- path planning (1)
- points of interest (1)
- privacy protection (1)
- public key infrastructure (1)
- recurrent (1)
- religiousness (1)
- rich multimedia presentations (1)
- robotics (1)
- semantic annotation (1)
- semantics (1)
- social simulation (1)
- visualization (1)
Institute
This dissertation provides an interdisciplinary contribution to the project ReGLaN-Health & Logistics. ReGLaN-Health & Logistics, is an international cooperation deriving benefits from the capabilities of scientists working on different fields. The aim of the project is the development of a socalled SDSS that supports decision makers working within health systems with a special focus on rural areas. In this dissertation, one important component for the development of the DSS named EWARS is proposed and described in detail. This component called SPATTB is developed with the intention of dealing with spatial data, i.e. data with additional geocoded information with regard to the special requirements of the EWARS.rnrnAn important component in the process of developing the EWARS is the concept of GIS. Classically, geocoded information with a vectorial character numerically describing spatial phenomena is managed and processed in a GIS. For the development of the EWARS, the manageability of the type of data exemplarily given by (x,y,o) with coordinates x,y ) and Ozon-concentration o is not sufficient. It is described, that the manageable data has to be extended to data of type (x,y,f ), where (x,y) are the geocoded information, but where f is not only a numerical value but a functional description of a certain phenomenom. An example for the existence and appearance of that type of data is the geocoded information about the variation of the Ozon-concentration in time or depending on temperature. A knowledge-base as important subsystem of DSS containing expert knowledge is mentioned. This expert-knowledge can be made manageable when using methods from the field of fuzzy logic. Thereby mappings, socalled fuzzy-sets, are generated. Within the EWARS, these mappings will be used with respect to additional geocoded data. The knowledge about the geocoded mapping information only at a finite set of locations (x,y) associated with mapping information f is not sufficient in applications that need continuous statements in a certain geographical area. To provide a contribution towards solving this problem, methods from the field of computer geometry and CAD, so-called Bezier-methods, are used for interpolating this geocoded mapping information. Classically, these methods operates on vectors a the multidimensional vector-space whose elements contain real-valued components but in terms of dealing with mapping information, there has to be an extension on topological vector spaces since mapping spaces can be defined as such spaces. This builds a new perspective and possibility in the application of these methods. Therefore, the according algorithms have to be extended; this work is presented. The field of Artificial Neural Networks plays an important role for the processing and management of the data within the EWARS, where features of biological processes and structures are modeled and implemented as algorithms. Generally, the developed methods can be divided as usable in terms of interpolation or approximation functional coherences and in such being applicable to classification problems. In this dissertation one method from each type is regarded in more detailed. Thereby, the classical algorithms of the so-called Backpropagation-Networks for approximation and the Kohonen-Networks for classification are described. Within the thesis, an extension of these algorithms is then proposed using coherences from mathematical measure-theory and approximation theory. The mentioned extension of these algorithms is based on a preprocessing of the mapping data using integration methods from measure theory.
This thesis presents an analysis of API usage in a large corpus of Java software retrieved from the open source repositories hosted at SourceForge. Most larger software projects use software libraries, which offer a public "application programming interface" or API as an interface for the programmer. In order to facilitate the transition between different APIs, there are emerging research projects in the field of automated API migration. However, there is a lack of basic statistical background information about in-the-wild usage of APIs as such measurements have, until now, only been done on rather small corpora. We thus present an analysis method suitable for measurements with large corpora. First, we create a corpus of open source projects hosted on SourceForge, as well as a corpus of software libraries. Then, all projects in the corpus are compiled with an instrumented compiler. We use a compiler plugin for javac that gives detailed information about every method created by the compiler. This information is stored in a database and analyzed.
Mobile payment has been a payment option in the market for a long time now and was predicted to become a widely used payment method. However, over the years, the market penetration rate of mPayments has been relatively low, despite it having all characteristics required of a convenient payment method. The primaryrnreason for this has been cited as a lack of customer acceptance mainly caused due to the lack of perceived security by the end-user. Although biometric authentication is not a new technology, it is experiencing a revival in the light of the present day terror threats and increased security requirements in various industries. The application of biometric authentication in mPayments is analysed here and a suitable biometric authentication method for use with mPayments is recommended. The issue of enrolment, human and technical factors to be considered are discussed and the STOF business model is applied to a BiMoP (biometric mPayment) application.
Knowledge compilation is a common technique for propositional logic knowledge bases. A given knowledge base is transformed into a normal form, for which queries can be answered efficiently. This precompilation step is expensive, but it only has to be performed once. We apply this technique to concepts defined in the Description Logic ALC. We introduce a normal form called linkless normal form for ALC concepts and discuss an efficient satisability test for concepts given in this normal form. Furthermore, we will show how to efficiently calculate uniform interpolants of precompiled concepts w.r.t. a given signature.
We propose a new approach for mobile visualization and interaction of temporal information by integrating support for time with today's most prevalent visualization of spatial information, the map. Our approach allows for an easy and precise selection of the time that is of interest and provides immediate feedback to the users when interacting with it. It has been developed in an evolutionary process gaining formative feedback from end users.
Existing tools for generating application programming interfaces (APIs) for ontologies lack sophisticated support for mapping the logics-based concepts of the ontology to an appropriate object-oriented implementation of the API. Such a mapping has to overcome the fundamental differences between the semantics described in the ontology and the pragmatics, i.e., structure, functionalities, and behavior implemented in the API. Typically, concepts from the ontology are mapped one-to-one to classes in the targeted programming language. Such a mapping only produces concept representations but not an API at the desired level of granularity expected by an application developer. We present a Model-Driven Engineering (MDE) process to generate customized APIs for ontologies. This API generation is based on the semantics defined in the ontology but also leverages additional information the ontology provides. This can be the inheritance structure of the ontology concepts, the scope of relevance of an ontology concept, or design patterns defined in the ontology.
In recent years ontologies have become common on the WWW to provide high-level descriptions of specific domains. These descriptions could be effectively used to build applications with the ability to find implicit consequences of their represented knowledge. The W3C developed the Resource Description Framework RDF, a language to describe the semantics of the data on the web, and the Ontology Web Language OWL, a family of knowledge representation languages for authoring ontologies. In this thesis we propose an ontology API engineering framework that makes use of the state-of-the-art ontology modeling technologies as well as of software engineering technologies. This system simplifies the design and implementation process of developing dedicated APIs for ontologies. Developers of semantic web applications usually face the problem of mapping entities or complex relations described in the ontology to object-oriented representations. Mapping complex relationship structures that come with complex ontologies to a useful API requires more complicated API representations than does the mere mapping of concepts to classes. The implementation of correct object persistence functions in such class representations also becomes quite complex.
In recent development, attempts have been made to integrate UML and OWL into one hybrid modeling language, namely TwoUse. This aims at making use of the benefits of both modeling languages and overcoming the restrictions of each. In order to create a modeling language that will actually be used in software development an integration with OCL is needed. This integration has already been described at the contextual level in, however an implementation is lacking so far. The scope of this paper is the programatical implementation of the integration of TwoUse with OCL. In order to achieve this, two different OCL implementations that already provide parsing and interpretation functionalities for expressions over regular UML. This paper presents two attempts to extend existing OCL implementations, as well as a comparison of the existing approaches.
Unlocking the semantics of multimedia presentations in the web with the multimedia metadata ontology
(2010)
The semantics of rich multimedia presentations in the web such as SMIL, SVG and Flash cannot or only to a very limited extend be understood by search engines today. This hampers the retrieval of such presentations and makes their archival and management a difficult task. Existing metadata models and metadata standards are either conceptually too narrow, focus on a specific media type only, cannot be used and combined together, or are not practically applicable for the semantic description of rich multimedia presentations. In this paper, we propose the Multimedia Metadata Ontology (M3O) for annotating rich, structured multimedia presentations. The M3O provides a generic modeling framework for representing sophisticated multimedia metadata. It allows for integrating the features provided by the existing metadata models and metadata standards. Our approach bases on Semantic Web technologies and can be easily integrated with multimedia formats such as the W3C standards SMIL and SVG. With the M3O, we unlock the semantics of rich multimedia presentations in the web by making the semantics machine-readable and machine-understandable. The M3O is used with our SemanticMM4U framework for the multi-channel generation of semantically-rich multimedia presentations.
Im Rahmen dieser Bachelorarbeit wurde ein Back-Office für die elektronische Version des Europäischen Schadensberichtes erstellt. Es wurde bereits in anderen Arbeiten ein mobiler Client, welcher auf einem Windows Mobile Handy läuft, sowie ein Polizei Client erstellt. Diese greifen auf das Back-Office zu, um Daten, wie z.B. die Autodaten (Automarke, der Typ, das Baujahr und Bilder eines 3D-Modells des Autos) zu einem bestimmten Kennzeichen oder die Personendaten des jeweiligen Autobesitzers zu erhalten. Der mobile Client sendet zudem die Unfallakte an das Back-Office, damit die Daten über einen Unfall in diesem abgespeichert und weiter bearbeitet werden können. Ziel der Arbeit war es ein erweiterbares, modulares System zu entwickeln, welches später um weitere Module ergänzt werden kann, um neue Funktionen bereitstellen zu können. Diese Module können jeweils beliebige Daten in einer Datenbank abspeichern und diese von der Datenbank auch wieder abfragen, sowie verändern, ohne dass das relationale Schema der Datenbank verändert werden muss.