Filtern
Erscheinungsjahr
Dokumenttyp
- Ausgabe (Heft) zu einer Zeitschrift (355) (entfernen)
Sprache
- Deutsch (270)
- Englisch (84)
- Mehrsprachig (1)
Volltext vorhanden
- ja (355) (entfernen)
Schlagworte
- Vorlesungsverzeichnis (55)
- Forschen in Koblenz (7)
- Junges Forschen (7)
- Wissenschaft Koblenz (7)
- Bluetooth (4)
- computer clusters (4)
- ontology (4)
- Knowledge Compilation (3)
- Petri-Netze (3)
- parallel algorithms (3)
- Augmented Reality (2)
- Campus Information System (2)
- Customer Relationship Management (2)
- DOCHOUSE (2)
- Datensicherheit (2)
- E-KRHyper (2)
- Enhanced Reality (2)
- Equality (2)
- Fallstudie (2)
- IBM Lotus Quickr (2)
- Informatik (2)
- Kryptographie (2)
- Linked Open Data (2)
- OWL (2)
- Ontology (2)
- Semantic Web (2)
- Telearbeit (2)
- Theorem Proving (2)
- University (2)
- artifcial neural networks (2)
- artificial neural networks (2)
- classification (2)
- constraint logic programming (2)
- mobile phone (2)
- multimedia metadata (2)
- privacy protection (2)
- probability propagation nets (2)
- risk (2)
- social media (2)
- 2D/3D-Segmentierung (1)
- 3-D-Modell (1)
- 3D-Color-Structure-Code (1)
- 3D-RETISEG (1)
- ABox (1)
- Absorptive capacity (1)
- Adaptive Services Grid (ASG) (1)
- Allgemeine Systemtheorie (1)
- Amazon Mechanical Turks (1)
- Annotationsmodell (1)
- Ausstellung (1)
- Authentifizierung (1)
- Authentisierung (1)
- Automated Theorem Proving (1)
- Automated Theorem Proving Systems (1)
- B2B (1)
- Bachelorstudiengang (1)
- Bayes Procedures (1)
- Bilderkennung (1)
- CCRDMT (1)
- COVID-19 (1)
- Calculus (1)
- Campus Koblenz (1)
- Case Study (1)
- Cloud Computing (1)
- Colonoscopy (1)
- Common Annotation Framework (1)
- Computer Supported Cooperative Work (1)
- Conference (1)
- Context-aware processes (1)
- Core Ontology on Multimedia (1)
- Core Ontology on Multimedia (COMM) (1)
- Creativity (1)
- Crowdsourcing (1)
- DPLL procedure (1)
- Darmspiegelung (1)
- Datenbrille (1)
- Datenschutz (1)
- Delphi-Studie (1)
- Description Logic (1)
- Description Logics (1)
- Dienstleistungsproduktivität (1)
- Discussion Forums (1)
- Distributed process execution (1)
- Dokumentation (1)
- E-KRHyper theorem prover (1)
- E-Mail (1)
- E-government (1)
- E-services (1)
- ERP-Systeme (1)
- Eigenschaften von Teammitgliedern (1)
- Eisenbahn-Unfalluntersuchungsstellen (1)
- Eisenbahnunfalldatenbank (1)
- Eisenbahnunfallkosten (1)
- Enterprise Systems (1)
- Enterprise-Resource-Planning (1)
- Entscheidungsprozess (1)
- Evaluation (1)
- Fernarbeitsregelungen (1)
- Forensik (1)
- Forschungsprojekt KMU 2.0 (1)
- Generation Z (1)
- Generative Model (1)
- Gesundheitswesen (1)
- Graph Technology (1)
- GraphUML (1)
- Healthcare institution (1)
- Heimarbeit (1)
- Herausforderungen (1)
- Homeoffice (1)
- Horn Clauses (1)
- Hyper Tableau Calculus (1)
- IASON (1)
- IEC 61131 (1)
- IP-Adressen (1)
- IPTV (1)
- IT Outsourcing (1)
- IT Security (1)
- IT Services (1)
- IT security analysis (1)
- IT-Risk-Management (1)
- IT-Sicherheitsmodelle (1)
- Image (1)
- Innovation (1)
- Innovationsmanagement (1)
- Intelligent Information Network (1)
- Interaktionselemente (1)
- Internet (1)
- Internet Voting (1)
- Java (1)
- Jungunternehmen (1)
- KRHyper (1)
- KSAO (1)
- Kalibrierung (1)
- Klimawandel (1)
- Knowledge Sharing (1)
- Koloskopie (1)
- Kontextbezogener Dienst (1)
- Krankenhausinformationssystem (1)
- Leistungsmessung (1)
- Linked Data Modeling (1)
- Logistik (1)
- MIA (1)
- MPEG-7 (1)
- Masterstudiengang (1)
- Mehrbenutzer-Annotationssystem (1)
- Mixed method (1)
- Mobile Arbeit (1)
- Mobile Information Systems (1)
- Mobiles Umfeld (1)
- Model-Driven Engineering (1)
- Multi-robot System (1)
- Multiagent System (1)
- Multimedia Metadata Ontology (1)
- National Investigation Bodies (1)
- Netzwerk kleiner und mittlerer Unternehmen (1)
- Neuronales Netz (1)
- Nutzen Condition Monitoring (1)
- ODRL (1)
- Object Recognition (1)
- Online Community (1)
- Online-Überwachung (1)
- Onlinewahl (1)
- Ontology alignment (1)
- Open Innovation (1)
- Open Source (1)
- Outsourcing (1)
- POIs (1)
- Personal Key Infrastruktur (1)
- Personalausweis (1)
- Personalised Information Systems (1)
- Petri Nets (1)
- Petri net (1)
- Petrinetz (1)
- Polypendetektion (1)
- Pose (1)
- Predictive Maintenance (1)
- Probability (1)
- Probability propagation nets (1)
- Process tracing (1)
- Programmiersprachen (1)
- Projekt iCity (1)
- Propagation (1)
- Prozessmanagement (1)
- Quality assessment system (1)
- RDF (1)
- RDF Graphs (1)
- RDF modeling (1)
- RDMT (1)
- RMIS (1)
- Railway Diagnostics & Monitoring Technologies (1)
- Railway Research (1)
- Railway Research Topics (1)
- Railway Safety (1)
- Railway Safety Research (1)
- ReDSeeDS-Project (1)
- Regionalbanken (1)
- Resource Description Framework (RDF) (1)
- Risikomanagement-Informationssysteme (1)
- Robocup 2008 (1)
- Router (1)
- Routing Information Protocol (RIP) (1)
- Routing Loops (1)
- Routing with Metric based Topology Investigation (RMTI) (1)
- SOAMIG (1)
- SOMA (1)
- SPARQL (1)
- Schema Information (1)
- Schutzprofil (1)
- Search engine (1)
- Security (1)
- Semantics (1)
- Semantik (1)
- Service-Oriented Architectures (SOA) (1)
- Service-oriented Architectures (SOA) (1)
- Shared Annotation Model (1)
- Shared Annotations (1)
- Social Networking Platforms (1)
- Software (1)
- Software Development (1)
- Software techniques for object recognition (STOR) (1)
- Software-Customizing (1)
- Startups (1)
- Stochastic Logic (1)
- Support System (1)
- Survey Research (1)
- Syntax (1)
- Systemtheorie (1)
- TAP (1)
- TBox (1)
- Tableau Calculus (1)
- Technikfolgenabschätzung (1)
- Technologiefrüherkennung (1)
- Technologiemanagement (1)
- Theorem prover (1)
- Tokens (1)
- Traceability (1)
- Transformation (1)
- UML (1)
- Umfrage (1)
- Unfallkostenberechnung (1)
- Unified Modeling Language (UML ) (1)
- United Internet AG (1)
- Universität Koblenz-Landau (1)
- Virtuelle Teams (1)
- Virtuelle Zusammenarbeit (1)
- Vocabulary Mapping (1)
- Vocabulary Reuse (1)
- Vorausschauende Instandhaltung (1)
- Wachstum (1)
- Wayside Train Monitoring Systems (1)
- Web 2.0 (1)
- Web Ontology Language (OWL) (1)
- Website (1)
- Wechselkursänderung (1)
- Werbung (1)
- Word-of-Mouth (1)
- XML (1)
- XSD (1)
- Zertifikat (1)
- activation functions of neurons (1)
- adaptive resonance theory (1)
- application programming interfaces (1)
- artiffficial neural networks (1)
- artififfcial neural networks (1)
- assessment model (1)
- behavior of employers (1)
- blood analysis (1)
- business process management (1)
- categorisation (1)
- core ontologies (1)
- currency exchange rates (1)
- delivery drone (1)
- design thinking (1)
- deutschsprachiger Markt (1)
- digital workplace (1)
- directed acyclic graphs (1)
- disability-adjusted life year (1)
- drone (1)
- e-Commerce (1)
- e-learning (1)
- e-service (1)
- e-service quality (1)
- elektronische Anwendung (1)
- elektronische Anwendungen (1)
- entrepreneurial design thinking (1)
- entrepreneurial thinking (1)
- entrepreneurship education (1)
- estimation of algorithm efficiency (1)
- event model (1)
- event-based systems (1)
- events (1)
- expectations (1)
- faceted search (1)
- finite state automata (1)
- first-order logic (1)
- gaze information (1)
- governance (1)
- gradient method of training weight coefficients (1)
- human resources management (1)
- hybrid automata (1)
- hybrid systems (1)
- hybrid work (1)
- iCity project (1)
- identification of organization (1)
- image processing (1)
- image semantics (1)
- information system (1)
- knowledge management system (1)
- knowledge work (1)
- kollaboratives Lernen (1)
- leader (1)
- leadership (1)
- legacy software (1)
- living book (1)
- mathematical model (1)
- media competence model (1)
- metadata formats (1)
- metadata standards (1)
- methodology (1)
- migration (1)
- minimum self-contained graphs (1)
- mobile application (1)
- mobile devices (1)
- mobile facets (1)
- mobile interaction (1)
- mobile phones (1)
- model generation (1)
- multi-agent systems (1)
- myAnnotations (1)
- nachhaltigkeitsorientierter Führungsstil (1)
- parallel calculations (1)
- personal information management (1)
- personality characteristics (1)
- persönliches Informationsmanagement (1)
- points of interest (1)
- polyp detection (1)
- privacy and personal data (1)
- privacy competence model (1)
- public key infrastructure (1)
- quality adjusted life year (1)
- regression analysis (1)
- regular dag languages (1)
- remote work (1)
- reputation of company (1)
- rich multimedia presentations (1)
- risks (1)
- scene analysis (1)
- security awareness (1)
- semantic annotation (1)
- semantic desktop (1)
- semantics (1)
- semantischer Desktop (1)
- sensor data (1)
- sequent calculi (1)
- social media data (1)
- social networks (1)
- social object (1)
- social simulation (1)
- summative evaluation (1)
- tagging (1)
- teams (1)
- technology acceptance model (1)
- time series (1)
- tracking (1)
- traffic survey (1)
- value of a statistical life (1)
- virtual goods (1)
- visualization (1)
- web-portal medical e-services (1)
- work from anywhere (1)
- work from home (1)
Institut
- Zentrale Einrichtungen (171)
- Fachbereich 4 (132)
- Institut für Informatik (38)
- Institut für Wirtschafts- und Verwaltungsinformatik (35)
- Institut für Management (32)
- Institute for Web Science and Technologies (15)
- Institut für Computervisualistik (12)
- Institut für Anglistik und Amerikanistik (7)
- Institut für Evangelische Theologie (7)
- Institut für Kulturwissenschaft (7)
UML models and OWL ontologies constitute modeling approaches with different strength and weaknesses that make them appropriate for use of specifying different aspects of software systems. In particular, OWL ontologies are well suited to specify classes using an expressive logical language with highly flexible, dynamic and polymorphic class membership, while UML diagrams are much more suitable for specifying not only static models including classes and associations, but also dynamic behavior. Though MOF based metamodels and UML profiles for OWL have been proposed in the past, an integrated use of both modeling approaches in a coherent framework has been lacking so far. We present such a framework, TwoUse, for developing integrated models, comprising the benefits of UML models and OWL ontologies
In this paper we describe a network for distributing personalized Information in a metropolitan area. We discuss the system architecture of our Bluetooth-based information system as well as the reasoning process that fits users" needs with potential messages. We furthermore present our findings on parallelizing Bluetooth connection setup and performance.
Semantic descriptions of non-textual media available on the web can be used to facilitate retrieval and presentation of media assets and documents containing them. While technologies for multimedia semantic descriptions already exist, there is as yet no formal description of a high quality multimedia ontology that is compatible with existing (semantic) web technologies. We explain the complexity of the problem using an annotation scenario. We then derive a number of requirements for specifying a formal multimedia ontology, including: compatibility with MPEG-7, embedding in foundational ontologies, and modularisation including separation of document structure from domain knowledge. We then present the developed ontology and discuss it with respect to our requirements.
Generalized methods for automated theorem proving can be used to compute formula transformations such as projection elimination and knowledge compilation. We present a framework based on clausal tableaux suited for such tasks. These tableaux are characterized independently of particular construction methods, but important features of empirically successful methods are taken into account, especially dependency directed backjumping and branch local operation. As an instance of that framework an adaption of DPLL is described. We show that knowledge compilation methods can be essentially improved by weaving projection elimination partially into the compilation phase.
In Enhanced Reality Umgebungen nimmt der Betrachter die Umgebung und zusätzliche, in eine halbdurchsichtige Datenbrille eingeblendete Informationen wahr. Das Kalibrierungsproblem der Datenbrille ist die Aufgabe, die eingeblendete Information mit dem korrekten realen Hintergrund zur Deckung zu bringen. Heutige Datenbrillen sind vergleichsweise klobig und schwer, deshalb kommt es häufig zu leichtem Verrutschen der Brille. Wird dieses Verrutschen nicht in die Position der Einblendung in der Brille einbezogen, so passt die Einblendung nicht mehr zum realen Hintergrund. Dies wird in Abbildung 1.1 exemplarisch dargestellt. Nach initialer Kalibrierung der teildurchsichtigen Datenbrille auf das Auge des Betrachters soll deshalb bei relativer Lageveränderung der Datenbrille zum Auge (durch leichtes Verrutschen) eine Re-Kalibrierung in Echtzeit automatisch erfolgen. Eine automatische Re-Kalibrierung bei Verrutschen wird von uns erstmalig erforscht.
This paper offers an informal overview and discussion on first order predicate logic reasoning systems together with a description of applications which are carried out in the Artificial Intelligence Research Group of the University in Koblenz. Furthermore the technique of knowledge compilation is shortly introduced.
Im Gesundheitswesen geht es nicht nur um die Vertraulichkeit von Patientendaten, sondern auch um ihre Integrität, von der die richtige Behandlung und Pflege des Patienten abhängen. Eine Maßnahme zur Absicherung eines Krankenhausinformationssystems (KIS) gegen unautorisierte Angreifer ist eine flexibel organisierte Zugriffskontrolle zum Schutz der patientenbezogenen Daten. Dieser Artikel beschreibt das Konzept einer rollenbasierten Zugriffskontrolle (RBAC - Role Based Access Control) und seine Einsatzmöglichkeiten im Gesundheitswesen.
Die Workshop-Reihe 'Algorithmen und Werkzeuge für Petrinetze' wurde 1994 mit dem Ziel initiiert, in der deutschsprachigen Petrinetz-Community den fachlichen Austausch und die inhaltliche Zusammenarbeit zwischen den mit der Entwicklung und Analyse von Algorithmen beschäftigten Arbeitsgruppen und den im Bereich der Implementierung von Werkzeugen tätigen Arbeitsgruppen zu fördern. Der vorliegende Sammelband enthält die Vorträge, die auf dem Workshop präsentiert worden sind. Um auch die Vorstellung von noch unfertigen Ideen oder von in Entwicklung befindlichen Werkzeugen zu ermöglichen, fand wie in den vergangenen Jahren kein formaler Begutachtungsprozess statt. Die eingereichten Beiträge wurden lediglich auf ihre Relevanz für das Workshop-Thema hin geprüft.
The E-KRHyper system is a model generator and theorem prover for first-order logic with equality. It implements the new E-hyper tableau calculus, which integrates a superposition-based handling of equality into the hyper tableau calculus. E-KRHyper extends our previous KRHyper system, which has been used in a number of applications in the field of knowledge representation. In contrast to most first order theorem provers, it supports features important for such applications, for example queries with predicate extensions as answers, handling of large sets of uniformly structured input facts, arithmetic evaluation and stratified negation as failure. It is our goal to extend the range of application possibilities of KRHyper by adding equality reasoning.
We introduce linear expressions for unrestricted dags (directed acyclic graphs) and finite deterministic and nondeterministic automata operating on them. Those dag automata are a conservative extension of the Tu,u-automata of Courcelle on unranked, unordered trees and forests. Several examples of dag languages acceptable and not acceptable by dag automata and some closure properties are given.
Die Bedeutung von Innovation für die Sicherung der Unternehmensexistenz nimmt zu. Gerade im Bereich der Informations- und Kommunikationssysteme zwingen veränderte Rahmenbedingungen, verkürzte Produktlebenszyklen und verstärkter Wettbewerb viele Unternehmen zur Anpassung ihrer Innovationsgeschwindigkeit. Langfristiger Erfolg wird sich nur für die Unternehmen einstellen, denen es gelingt, durch technologische Innovationen ihre Wettbewerbsposition zu festigen oder auszubauen. Dies erfordert einen gesonderten Umgang mit technologischen Innovationen und verlangt nach einem Technologie- und Innovationsmanagement. Dabei ist nicht jede Innovation das Produkt einer einmaligen, möglicherweise gar spontanen Idee. Viele Innovationen entstehen aus der Integration oder Kombination bereits bewährter Technologien. Dies ist besonders interessant, wenn die originären Technologien Branchen entstammen, die nur noch marginale Wachstumsraten aufweisen. Aufbauend auf den Besonderheiten des Fallbeispiels Decision on Demand werden in diesem Artikel existierende Mängel etablierter Vorschläge für ein strategisches Technologie- und Innovationsmanagement identifiziert und die Rolle von modularer Technologieintegration als Alternative diskutiert.
The model evolution calculus
(2004)
The DPLL procedure is the basis of some of the most successful propositional satisfiability solvers to date. Although originally devised as a proof procedure for first-order logic, it has been used almost exclusively for propositional logic so far because of its highly inefficient treatment of quantifiers, based on instantiation into ground formulas. The recent FDPLL calculus by Baumgartner was the first successful attempt to lift the procedure to the first-order level without resorting to ground instantiations. FDPLL lifts to the first-order case the core of the DPLL procedure, the splitting rule, but ignores other aspects of the procedure that, although not necessary for completeness, are crucial for its effectiveness in practice. In this paper, we present a new calculus loosely based on FDPLL that lifts these aspects as well. In addition to being a more faithful litfing of the DPLL procedure, the new calculus contains a more systematic treatment of universal literals, one of FDPLL's optimizations, and so has the potential of leading to much faster implementations.
The Living Book is a system for the management of personalized and scenario specific teaching material. The main goal of the system is to support the active, explorative and selfdetermined learning in lectures, tutorials and self study. The Living Book includes a course on 'logic for computer scientists' with a uniform access to various tools like theorem provers and an interactive tableau editor. It is routinely used within teaching undergraduate courses at our university. This paper describes the Living Book and the use of theorem proving technology as a core component in the knowledge management system (KMS) of the Living Book. The KMS provides a scenario management component where teachers may describe those parts of given documents that are relevant in order to achieve a certain learning goal. The task of the KMS is to assemble new documents from a database of elementary units called 'slices' (definitions, theorems, and so on) in a scenario-based way (like 'I want to prepare for an exam and need to learn about resolution'). The computation of such assemblies is carried out by a model-generating theorem prover for first-order logic with a default negation principle. Its input consists of meta data that describe the dependencies between different slices, and logic-programming style rules that describe the scenario-specific composition of slices. Additionally, a user model is taken into account that contains information about topics and slices that are known or unknown to a student. A model computed by the system for such input then directly specifies the document to be assembled. This paper introduces the elearning context we are faced with, motivates our choice of logic and presents the newly developed calculus used in the KMS.
We aim to demonstrate that automated deduction techniques, in particular those following the model computation paradigm, are very well suited for database schema/query reasoning. Specifically, we present an approach to compute completed paths for database or XPath queries. The database schema and a query are transformed to disjunctive logic programs with default negation, using a description logic as an intermediate language. Our underlying deduction system, KRHyper, then detects if a query is satisfiable or not. In case of a satisfiable query, all completed paths -- those that fulfill all given constraints -- are returned as part of the computed models. The purpose of our approach is to dramatically reduce the workload on the query processor. Without the path completion, a usual XML query processor would search the database for solutions to the query. In the paper we describe the transformation in detail and explain how to extract the solution to the original task from the computed models. We understand this paper as a first step, that covers a basic schema/query reaÂsoning task by model-based deduction. Due to the underlying expressive logic formalism we expect our approach to easily adapt to more sophisticated problem settings, like type hierarchies as they evolve within the XML world.