Filtern
Erscheinungsjahr
Dokumenttyp
- Ausgabe (Heft) zu einer Zeitschrift (84) (entfernen)
Sprache
- Englisch (84) (entfernen)
Schlagworte
- Bluetooth (4)
- ontology (4)
- Knowledge Compilation (3)
- computer clusters (3)
- Campus Information System (2)
- E-KRHyper (2)
- Equality (2)
- Linked Open Data (2)
- OWL (2)
- Ontology (2)
- Petri-Netze (2)
- Semantic Web (2)
- Theorem Proving (2)
- University (2)
- artificial neural networks (2)
- classification (2)
- constraint logic programming (2)
- mobile phone (2)
- multimedia metadata (2)
- parallel algorithms (2)
- probability propagation nets (2)
- risk (2)
- social media (2)
- ABox (1)
- Adaptive Services Grid (ASG) (1)
- Amazon Mechanical Turks (1)
- Augmented Reality (1)
- Automated Theorem Proving (1)
- Automated Theorem Proving Systems (1)
- Bayes Procedures (1)
- COVID-19 (1)
- Calculus (1)
- Cloud Computing (1)
- Computer Supported Cooperative Work (1)
- Conference (1)
- Context-aware processes (1)
- Core Ontology on Multimedia (1)
- Core Ontology on Multimedia (COMM) (1)
- Creativity (1)
- Crowdsourcing (1)
- DPLL procedure (1)
- Description Logic (1)
- Description Logics (1)
- Discussion Forums (1)
- Distributed process execution (1)
- E-KRHyper theorem prover (1)
- E-government (1)
- E-services (1)
- Enhanced Reality (1)
- Enterprise Systems (1)
- Generative Model (1)
- Graph Technology (1)
- Healthcare institution (1)
- Heimarbeit (1)
- Horn Clauses (1)
- Hyper Tableau Calculus (1)
- IASON (1)
- IT Outsourcing (1)
- IT Security (1)
- IT Services (1)
- Image (1)
- Intelligent Information Network (1)
- Internet (1)
- Internet Voting (1)
- KRHyper (1)
- Knowledge Sharing (1)
- Linked Data Modeling (1)
- MIA (1)
- MPEG-7 (1)
- Mixed method (1)
- Mobile Information Systems (1)
- Model-Driven Engineering (1)
- Multi-robot System (1)
- Multiagent System (1)
- Multimedia Metadata Ontology (1)
- Neuronales Netz (1)
- ODRL (1)
- Object Recognition (1)
- Online Community (1)
- Ontology alignment (1)
- POIs (1)
- Personalised Information Systems (1)
- Petri Nets (1)
- Petri net (1)
- Petrinetz (1)
- Probability (1)
- Probability propagation nets (1)
- Process tracing (1)
- Propagation (1)
- Quality assessment system (1)
- RDF (1)
- RDF Graphs (1)
- RDF modeling (1)
- Railway Research (1)
- Railway Research Topics (1)
- Railway Safety (1)
- Railway Safety Research (1)
- ReDSeeDS-Project (1)
- Resource Description Framework (RDF) (1)
- Robocup 2008 (1)
- Routing Information Protocol (RIP) (1)
- Routing Loops (1)
- Routing with Metric based Topology Investigation (RMTI) (1)
- SPARQL (1)
- Schema Information (1)
- Search engine (1)
- Security (1)
- Semantics (1)
- Service-oriented Architectures (SOA) (1)
- Social Networking Platforms (1)
- Software Development (1)
- Software techniques for object recognition (STOR) (1)
- Stochastic Logic (1)
- Support System (1)
- Survey Research (1)
- TAP (1)
- TBox (1)
- Tableau Calculus (1)
- Telearbeit (1)
- Theorem prover (1)
- Tokens (1)
- Traceability (1)
- UML (1)
- Umfrage (1)
- Unified Modeling Language (UML ) (1)
- Vocabulary Mapping (1)
- Vocabulary Reuse (1)
- Web Ontology Language (OWL) (1)
- Website (1)
- Wechselkursänderung (1)
- Werbung (1)
- Word-of-Mouth (1)
- XML (1)
- adaptive resonance theory (1)
- application programming interfaces (1)
- artifcial neural networks (1)
- artiffficial neural networks (1)
- artififfcial neural networks (1)
- assessment model (1)
- blood analysis (1)
- business process management (1)
- categorisation (1)
- core ontologies (1)
- currency exchange rates (1)
- delivery drone (1)
- design thinking (1)
- digital workplace (1)
- directed acyclic graphs (1)
- drone (1)
- e-Commerce (1)
- e-learning (1)
- e-service (1)
- e-service quality (1)
- entrepreneurial design thinking (1)
- entrepreneurial thinking (1)
- entrepreneurship education (1)
- estimation of algorithm efficiency (1)
- event model (1)
- event-based systems (1)
- events (1)
- faceted search (1)
- finite state automata (1)
- first-order logic (1)
- gaze information (1)
- governance (1)
- gradient method of training weight coefficients (1)
- hybrid automata (1)
- hybrid systems (1)
- hybrid work (1)
- iCity project (1)
- image processing (1)
- image semantics (1)
- information system (1)
- knowledge management system (1)
- knowledge work (1)
- living book (1)
- mathematical model (1)
- media competence model (1)
- metadata formats (1)
- metadata standards (1)
- methodology (1)
- minimum self-contained graphs (1)
- mobile application (1)
- mobile devices (1)
- mobile facets (1)
- mobile interaction (1)
- model generation (1)
- multi-agent systems (1)
- parallel calculations (1)
- personal information management (1)
- persönliches Informationsmanagement (1)
- points of interest (1)
- privacy and personal data (1)
- privacy competence model (1)
- privacy protection (1)
- public key infrastructure (1)
- regression analysis (1)
- regular dag languages (1)
- remote work (1)
- rich multimedia presentations (1)
- risks (1)
- scene analysis (1)
- security awareness (1)
- semantic annotation (1)
- semantic desktop (1)
- semantics (1)
- semantischer Desktop (1)
- sequent calculi (1)
- social media data (1)
- social object (1)
- social simulation (1)
- summative evaluation (1)
- tagging (1)
- teams (1)
- technology acceptance model (1)
- time series (1)
- tracking (1)
- virtual goods (1)
- visualization (1)
- web-portal medical e-services (1)
- work from anywhere (1)
- work from home (1)
Iterative Signing of RDF(S) Graphs, Named Graphs, and OWL Graphs: Formalization and Application
(2013)
When publishing graph data on the web such as vocabulariesrnusing RDF(S) or OWL, one has only limited means to verify the authenticity and integrity of the graph data. Today's approaches require a high signature overhead and do not allow for an iterative signing of graph data. This paper presents a formally defined framework for signing arbitrary graph data provided in RDF(S), Named Graphs, or OWL. Our framework supports signing graph data at different levels of granularity: minimum self-contained graphs (MSG), sets of MSGs, and entire graphs. It supports for an iterative signing of graph data, e. g., when different parties provide different parts of a common graph, and allows for signing multiple graphs. Both can be done with a constant, low overhead for the signature graph, even when iteratively signing graph data.
Six and Gimmler have identified concrete capabilities that enable users to use the Internet in a competent way. Their media competence model can be used for the didactical design of media usage in secondary schools. However, the special challenge of security awareness is not addressed by the model. In this paper, the important dimension of risk and risk assessment will be introduced into the model. This is especially relevant for the risk of the protection of personal data and privacy. This paper will apply the method of IT risk analysis in order to select those dimensions of the Six/Gimmler media competence model that are appropriate to describe privacy aware Internet usage. Privacy risk aware decisions for or against the Internet usage is made visible by the trust model of Mayer et al.. The privacy extension of the competence model will lead to a measurement of the existing privacy awareness in secondary schools, which, in turn, can serve as a didactically well-reasoned design of Informatics modules in secondary schools. This paper will provide the privacy-extended competence model, while empirical measurement and module design is planned for further research activities.
The aim of this paper is to identify and understand the risks and issues companies are experiencing from the business use of social media and to develop a framework for describing and categorising those social media risks. The goal is to contribute to the evolving theorisation of social media risk and to provide a foundation for the further development of social media risk management strategies and processes. The study findings identify thirty risk types organised into five categories (technical, human, content, compliance and reputational). A risk-chain is used to illustrate the complex interrelated, multi-stakeholder nature of these risks and directions for future work are identified.
Ontologies play an important role in knowledge representation for sharing information and collaboratively developing knowledge bases. They are changed, adapted and reused in different applications and domains resulting in multiple versions of an ontology. The comparison of different versions and the analysis of changes at a higher level of abstraction may be insightful to understand the changes that were applied to an ontology. While there is existing work on detecting (syntactical) differences and changes in ontologies, there is still a need in analyzing ontology changes at a higher level of abstraction like ontology evolution or refactoring pattern. In our approach we start from a classification of model refactoring patterns found in software engineering for identifying such refactoring patterns in OWL ontologies using DL reasoning to recognize these patterns.
SOA-Security
(2007)
This paper is a part of the ASG project (Adaptive Services Grid) and addresses some IT security issues of service oriented architectures. It defines a service-oriented security concept, it explores the SOA security challenge, it describes the existing WS-Security standard, and it undertakes a first step into a survey on best practice examples. In particular, the ASG middleware platform technology (JBossWS) is analyzed with respect to its ability to handle security functions.
This paper describes the development of security requirements for non-political Internet voting. The practical background is our experience with the Internet voting within the Gesellschaft für Informatik (GI - Informatics Society) 2004 and 2005. The theoretical background is the international state-of-the-art of requirements about electronic voting, especially in the US and in Europe. A focus of this paper is on the user community driven standardization of security requirements by means of a Protection Profile of the international Common Criteria standard.
Virtual Goods + ODRL 2012
(2012)
This is the 10th international workshop for technical, economic, and legal aspects of business models for virtual goods incorporating the 8th ODRL community group meeting. This year we did not call for completed research results, but we invited PhD students to present and discuss their ongoing research work. In the traditional international group of virtual goods and ODRL researchers we discussed PhD research from Belgium, Brazil, and Germany. The topics focused on research questions about rights management in the Internet and e-business stimulation. In the center of rights management stands the conception of a formal policy expression that can be used for human readable policy transparency, as well as for machine readable support of policy conformant systems behavior up to automatic policy enforcement. ODRL has proven to be an ideal basis for policy expressions, not only for digital copy rights, but also for the more general "Policy Awareness in the World of Virtual Goods". In this sense, policies support the communication of virtual goods, and they are a virtualization of rules-governed behavior themselves.
Various best practices and principles guide an ontology engineer when modeling Linked Data. The choice of appropriate vocabularies is one essential aspect in the guidelines, as it leads to better interpretation, querying, and consumption of the data by Linked Data applications and users.
In this paper, we present the various types of support features for an ontology engineer to model a Linked Data dataset, discuss existing tools and services with respect to these support features, and propose LOVER: a novel approach to support the ontology engineer in modeling a Linked Data dataset. We demonstrate that none of the existing tools and services incorporate all types of supporting features and illustrate the concept of LOVER, which supports the engineer by recommending appropriate classes and properties from existing and actively used vocabularies. Hereby, the recommendations are made on the basis of an iterative multimodal search. LOVER uses different, orthogonal information sources for finding terms, e.g. based on a best string match or schema information on other datasets published in the Linked Open Data cloud. We describe LOVER's recommendation mechanism in general and illustrate it alongrna real-life example from the social sciences domain.
This paper presents a method for the evolution of SHI ABoxes which is based on a compilation technique of the knowledge base. For this the ABox is regarded as an interpretation of the TBox which is close to a model. It is shown, that the ABox can be used for a semantically guided transformation resulting in an equisatisfiable knowledge base. We use the result of this transformation to effciently delete assertions from the ABox. Furthermore, insertion of assertions as well as repair of inconsistent ABoxes is addressed. For the computation of the necessary actions for deletion, insertion and repair, the E-KRHyper theorem prover is used.
Knowledge compilation is a common technique for propositional logic knowledge bases. The idea is to transform a given knowledge base into a special normal form ([MR03],[DH05]), for which queries can be answered efficiently. This precompilation step is very expensive but it only has to be performed once. We propose to apply this technique to knowledge bases defined in Description Logics. For this, we introduce a normal form, called linkless concept descriptions, for ALC concepts. Further we present an algorithm, based on path dissolution, which can be used to transform a given concept description into an equivalent linkless concept description. Finally we discuss a linear satisfiability test as well as a subsumption test for linkless concept descriptions.
This paper offers an informal overview and discussion on first order predicate logic reasoning systems together with a description of applications which are carried out in the Artificial Intelligence Research Group of the University in Koblenz. Furthermore the technique of knowledge compilation is shortly introduced.
This paper shows how multiagent systems can be modeled by a combination of UML statecharts and hybrid automata. This allows formal system specification on different levels of abstraction on the one hand, and expressing real-time system behavior with continuous variables on the other hand. It is not only shown how multi-robot systems can be modeled by a combination of hybrid automata and hierarchical state machines, but also how model checking techniques for hybrid automata can be applied. An enhanced synchronization concept is introduced that allows synchronization taking time and avoids state explosion to a certain extent.
In this paper we describe a series of projects on location based and personalised information systems. We start wit a basic research project and we show how we came with the help of two other more application oriented project to a product. This is developed by a consortium of enterprises and it already is in use in the city of Koblenz.
In this paper we describe a network for distributing personalized information within a pervasive university. We discuss the system architecture of our Bluetooth-based CampusNews-system, both, from the administrator and the user viewpoint. We furthermore present first statistical data about the usage of the partial installation at the Koblenz campus together with an outlook to future work.
In this paper we describe a network for distributing personalized Information in a metropolitan area. We discuss the system architecture of our Bluetooth-based information system as well as the reasoning process that fits users" needs with potential messages. We furthermore present our findings on parallelizing Bluetooth connection setup and performance.
Semantic desktop environments aim at improving the effectiveness and efficiency of users carrying out daily tasks within their personal information management infrastructure (PIM). They support the user by transferring and exploiting the explicit semantics of data items across different PIM applications. Whether such an approach does indeed reach its aim of facilitating users" life and—if so—to which extent, however, remains an open question that we address in this paper with the first summative evaluation of a semantic desktop approach. We approach the research question exploiting our own semantic desktop infrastructure, X-COSIM. As data corpus, we have used over 100 emails and 50 documents extracted from the organizers of a conference-like event at our university. The evaluation has been carried out with 18 subjects. We have developed a test environment to evaluate COSIMail and COSIFile, two semantic PIM applications based on X-COSIM. As result, we have found a significant improvement for typical PIM tasks compared to a standard desktop environment.
The STOR project aims at the development of a scientific component system employing models and knowledge for object recognition in images. This interim report elaborates on the requirements for such a component system, structures the application area by identifying a large set of basic operations, and shows how a set of appropriate data structures and components can be derived. A small case studies exemplifies the approach.
Das Interesse am Konzept Crowdfunding ist in den letzten Jahren sowohl aus der Wirtschaft als auch aus der Wissenschaft gestiegen. Neben Künstlern und Entrepreneuren finanzieren nun auch Wissenschaftler ihre Projekte durch zahlreiche kleine Beiträge aus der Crowd. Allerdings spiegelt die wahrgenommene Nutzung in Deutschland nicht die Vorteile einer Crowdfunding-Kampagne wider, insbesondere nicht im internationalen Vergleich. Die vorliegende Studie untersucht diesen Umstand, indem sie die Motive und Barrieren für eine Finanzierung durch Crowdfunding identifiziert, um Empfehlungen für Forschungseinrichtungen zur Förderung von Crowdfunding zu formulieren. Durch eine Literaturanalyse werden erste Erkenntnisse ermittelt, welche dann für die Durchführung von qualitativen Interviews mit elf Wissenschaftlern, die bereits eine Crowdfunding-Kampagne erfolgreich abgeschlossen haben, verwendet werden. Die Ergebnisse zeigen, dass ForscherInnen in Deutschland Crowdfunding in erster Linie dazu verwenden, Aufmerksamkeit für das Thema und die wissenschaftliche Gemeinschaft im Allgemeinen zu schaffen. Die größten Barrieren sind der enorme Aufwand, der mit einer Kampagne verbunden ist, und die mangelnde Reputation des Konzepts des Crowdfunding seitens der deutschen Wissenschaft. Zudem eignen sich nur Themen und Projekte mit einer hohen Öffentlichkeitswirksamkeit und einem Fördervolumen von maximal fünfstelligen Beträgen. Darüber hinaus konnte die öffentliche Wahrnehmung der Experten während der Kampagne als zusätzliche Barriere identifiziert werden.
Diese Ergebnisse führen zu drei Empfehlungen für Forschungseinrichtungen zur Förderung von Crowdfunding: Es wird empfohlen das Bewusstsein für das Thema Crowdfunding als zusätzliche Form der Forschungsfinanzierung zu sensibilisieren und die Vorteile einer Crowdfunding-Kampagne hervorzuheben. Universitäten sollten mit Crowdfunding Akteuren (bspw. Plattformen und Experten) zusammenarbeiten und die Netzwerkeffekte einer Universität sollten genutzt werden. Die Universitäten sollten Unterstützung leisten, um die Arbeitsbelastung auf interdisziplinäre Teams zu verteilen, um so das Verhältnis zwischen Aufwand und Ertrag zu optimieren. Die gewählte Methodik und der Geltungsbereich der Thesis eröffnen weitere Forschungsmöglichkeiten, die die Perspektive der Universitäten und die Bedingungen in anderen Ländern untersuchen könnten. Darüber hinaus ist eine groß angelegte quantitative Studie erforderlich, um die identifizierten Konzepte statistisch zu validieren.
Expert-driven business process management is an established means for improving efficiency of organizational knowledge work. Implicit procedural knowledge in the organization is made explicit by defining processes. This approach is not applicable to individual knowledge work due to its high complexity and variability. However, without explicitly described processes there is no analysis and efficient communication of best practices of individual knowledge work within the organization. In addition, the activities of the individual knowledge work cannot be synchronized with the activities in the organizational knowledge work.rnrnSolution to this problem is the semantic integration of individual knowledgernwork and organizational knowledge work by means of the patternbased core ontology strukt. The ontology allows for defining and managing the dynamic tasks of individual knowledge work in a formal way and to synchronize them with organizational business processes. Using the strukt ontology, we have implemented a prototype application for knowledge workers and have evaluated it at the use case of an architectural fifirm conducting construction projects.
With the Multimedia Metadata Ontology (M3O), we have developed a sophisticated model for representing among others the annotation, decomposition, and provenance of multimedia metadata. The goal of the M3O is to integrate the existing metadata standards and metadata formats rather than replacing them. To this end, the M3O provides a scaffold needed to represent multimedia metadata. Being an abstract model for multimedia metadata, it is not straightforward how to use and specialize the M3O for concrete application requirements and existing metadata formats and metadata standards. In this paper, we present a step-by-step alignment method describing how to integrate and leverage existing multimedia metadata standards and metadata formats in the M3O in order to use them in a concrete application. We demonstrate our approach by integrating three existing metadata models: the Core Ontology on Multimedia (COMM), which is a formalization of the multimedia metadata standard MPEG-7, the Ontology for Media Resource of the W3C, and the widely known industry standard EXIF for image metadata