Filtern
Erscheinungsjahr
Dokumenttyp
- Ausgabe (Heft) zu einer Zeitschrift (72)
- Diplomarbeit (14)
- Bachelorarbeit (11)
- Masterarbeit (9)
- Konferenzveröffentlichung (5)
- Dissertation (2)
- Wissenschaftlicher Artikel (1)
- Studienarbeit (1)
Sprache
- Englisch (115) (entfernen)
Volltext vorhanden
- ja (115) (entfernen)
Schlagworte
- Bluetooth (4)
- ontology (4)
- Knowledge Compilation (3)
- University (3)
- computer clusters (3)
- Campus Information System (2)
- Cloud Computing (2)
- E-KRHyper (2)
- Enterprise 2.0 (2)
- Linked Open Data (2)
- OWL (2)
- Ontology (2)
- Petri-Netze (2)
- RDF (2)
- Semantic Web (2)
- Simulation (2)
- artificial neural networks (2)
- classification (2)
- constraint logic programming (2)
- governance (2)
- mobile phone (2)
- parallel algorithms (2)
- probability propagation nets (2)
- risk (2)
- social media (2)
- 101companies (1)
- 8C model (1)
- ABox (1)
- API (1)
- API Analysis (1)
- API Migratiom (1)
- Adaptive Services Grid (ASG) (1)
- Adobe Flex (1)
- Agenten (1)
- Agents (1)
- Amazon Mechanical Turks (1)
- Android <Systemplattform> (1)
- Annotation (1)
- Archivierung (1)
- Artificial Intelligence (1)
- Augmented Reality (1)
- Automated Theorem Proving (1)
- Automated Theorem Proving Systems (1)
- Automatisches Beweisverfahren (1)
- B2B Integration (1)
- Bayes Procedures (1)
- Bebauungsdichte (1)
- Beschreibungslogik (1)
- Business Collaboration (1)
- Calculus (1)
- Casual Games (1)
- Chaos (1)
- Computational biology (1)
- Computerspiel (1)
- Conference (1)
- Content Analysis (1)
- Content Management (1)
- Context-aware processes (1)
- Core Ontology on Multimedia (1)
- Core Ontology on Multimedia (COMM) (1)
- Creativity (1)
- Criteria Matrix (1)
- Crowdsourcing (1)
- Data Mining (1)
- Defi-Now! (1)
- Defibrillator (1)
- Delta (1)
- Description Logic (1)
- Description Logics (1)
- Design Pattern (1)
- Diabetes (1)
- Diabetische Retinopathie (1)
- Discussion Forums (1)
- Distance Vector Routing (1)
- Distanz Vektor Routing (1)
- Distributed process execution (1)
- Drahtloses lokales Netz (1)
- Dreidimensionale Bildverarbeitung (1)
- Dynamische Analyse (1)
- E-Business (1)
- E-Hyper Tableau (1)
- E-KRHyper theorem prover (1)
- E-government (1)
- E-services (1)
- ECMS 2012 (1)
- Emergenz (1)
- Enhanced Reality (1)
- Enterprise Information Management (1)
- Enterprise Systems (1)
- Entwurfsmuster (1)
- Erste Hilfe (1)
- European Conference on Modelling and Simulation (1)
- Facebook Application (1)
- Farbkalibrierung (1)
- First aid (1)
- FlexViz (1)
- GPGPU (1)
- GPS (1)
- GSM-Standard (1)
- Generative Model (1)
- Geocaching (1)
- Germany (1)
- Grails (1)
- Grails 1.2 (1)
- Graph Technology (1)
- Haskell (1)
- Healthcare institution (1)
- Horn Clauses (1)
- Hypertableaux (1)
- IASON (1)
- IT Outsourcing (1)
- IT Security (1)
- IT Services (1)
- Image (1)
- Incremental Algorithms (1)
- Intelligent Information Network (1)
- Internet (1)
- Internet Voting (1)
- Java <Programmiersprache> (1)
- Kantenverfolgung (1)
- Katastrophentheorie (1)
- Knowledge Sharing (1)
- Konzept (1)
- Linked Data Modeling (1)
- Logik (1)
- Lokalisation (1)
- MIA (1)
- MPEG-7 (1)
- MapReduce (1)
- Maschinelles Sehen (1)
- Matching (1)
- Mediator framework (1)
- Medizinische Bildverarbeitung (1)
- Migration (1)
- Mister X (1)
- Mixed method (1)
- Mobile Information Systems (1)
- Model-Driven Engineering (1)
- Morphologische Operatoren (1)
- Motivation (1)
- Multi-robot System (1)
- Multiagent System (1)
- Multimedia Metadata Ontology (1)
- Natural Feature Tracking (1)
- Netzwerk Routing (1)
- Neuronales Netz (1)
- ODRL (1)
- OWL <Informatik> (1)
- Object Recognition (1)
- Online Community (1)
- Ontologie <Wissensverarbeitung> (1)
- Ontology alignment (1)
- OpenVDB (1)
- POIs (1)
- Path Tracing (1)
- Personalised Information Systems (1)
- Petri Nets (1)
- Petri net (1)
- Petrinetz (1)
- Pfadnachverfolgung (1)
- Politik (1)
- Probability (1)
- Probability propagation nets (1)
- Proceedings (1)
- Process tracing (1)
- Propagation (1)
- Prozedurale Synthese (1)
- Quality assessment system (1)
- RDF Graphs (1)
- RDF modeling (1)
- ReDSeeDS-Project (1)
- Real-Life Game (1)
- Real-Life Spiel (1)
- Real-Time (1)
- Reengineering (1)
- Registratur (1)
- Resource Description Framework (RDF) (1)
- Retina Befundbilder (1)
- Robocup 2008 (1)
- Routing Information Protocol (1)
- Routing Information Protocol (RIP) (1)
- Routing Information Protokoll (1)
- Routing Loops (1)
- Routing with Metric based Topology Investigation (RMTI) (1)
- SPARQL (1)
- Schema Information (1)
- Schlussfolgern (1)
- Search engine (1)
- Security (1)
- Security Routing (1)
- Selbstorganisation (1)
- Semantics (1)
- Service identification (1)
- Service-Identifizierung (1)
- Service-oriented Architectures (SOA) (1)
- Serviceorientierte Architektur (1)
- Sicherheit Routing (1)
- Simulationswerkzeug (1)
- Smartphone (1)
- Smartphone Applikation (1)
- Social Games (1)
- Social Networking Platforms (1)
- Software Development (1)
- Software migration (1)
- Software techniques for object recognition (STOR) (1)
- Software-Migration (1)
- Softwareentwicklung (1)
- Softwareergonomie (1)
- Softwarewartung (1)
- Soziales System (1)
- Steuerung (1)
- Stochastic Logic (1)
- Strassenkreuzung (1)
- Support System (1)
- Survey Research (1)
- Systembiologie (1)
- TAP (1)
- TBox (1)
- Taxonomy (1)
- Text Analysis (1)
- Text Mining (1)
- Theorem prover (1)
- Tokens (1)
- Tool Evaluation (1)
- Traceability (1)
- Tracing (1)
- UML (1)
- Unified Modeling Language (UML ) (1)
- Usability (1)
- Visibility Skeleton (1)
- Vocabulary Mapping (1)
- Vocabulary Reuse (1)
- Vocabulary Trainer (1)
- WLAN Fingerprinting (1)
- Web Analytics (1)
- Web Analytics Framework (1)
- Web Mining (1)
- Web Ontology Language (OWL) (1)
- Web-application framework (1)
- Web-programming technologies (1)
- Website (1)
- Wechselkursänderung (1)
- Werbung (1)
- WiFi Fingerprinting (1)
- Wikipedia (1)
- Wissensbasis (1)
- Word-of-Mouth (1)
- Wrapping (1)
- adaptive resonance theory (1)
- agent-based simulation (1)
- application programming interfaces (1)
- archiving (1)
- artifcial neural networks (1)
- artiffficial neural networks (1)
- artififfcial neural networks (1)
- benefits (1)
- blood analysis (1)
- business process management (1)
- by-stander effect (1)
- catastrophy theory (1)
- categorisation (1)
- chaos (1)
- collaborative technologies (1)
- colour calibration (1)
- concept (1)
- contexts of use (1)
- core ontologies (1)
- currency exchange rates (1)
- delivery drone (1)
- description logic (1)
- design thinking (1)
- deutsche Hochschulen (1)
- diabetic retinopathy (1)
- directed acyclic graphs (1)
- distinct object identifiers (1)
- drone (1)
- e-Commerce (1)
- eXperience methodology (1)
- edge linking (1)
- emergence (1)
- entrepreneurial design thinking (1)
- entrepreneurial thinking (1)
- entrepreneurship education (1)
- estimation of algorithm efficiency (1)
- event model (1)
- event-based systems (1)
- events (1)
- faceted search (1)
- finite state automata (1)
- gradient method of training weight coefficients (1)
- hybrid automata (1)
- hybrid systems (1)
- hypertableaux (1)
- iCity project (1)
- image processing (1)
- information system (1)
- jOWL (1)
- knowledge base (1)
- knowledge work (1)
- mathematical model (1)
- media competence model (1)
- medical image processing (1)
- metadata formats (1)
- metadata standards (1)
- minimum self-contained graphs (1)
- mobile application (1)
- mobile devices (1)
- mobile facets (1)
- mobile interaction (1)
- morphological operators (1)
- multi-agent systems (1)
- multimedia metadata (1)
- norm (1)
- parallel calculations (1)
- personal information management (1)
- persönliches Informationsmanagement (1)
- points of interest (1)
- policy modelling (1)
- privacy and personal data (1)
- privacy competence model (1)
- privacy protection (1)
- procedural content generation (1)
- public key infrastructure (1)
- regular dag languages (1)
- retina fundus images (1)
- risks (1)
- scene analysis (1)
- security awareness (1)
- self-organisation (1)
- semantic desktop (1)
- semantics (1)
- semantischer Desktop (1)
- shiq (1)
- simulation (1)
- smartphone app (1)
- social media data (1)
- social object (1)
- social simulation (1)
- social system (1)
- summative evaluation (1)
- teams (1)
- technology acceptance model (1)
- time series (1)
- tool-integration (1)
- tracking (1)
- unique name assumption (1)
- usability study (1)
- video games (1)
- virtual goods (1)
- visualization (1)
Institut
- Fachbereich 4 (115) (entfernen)
Zur Entwicklung von Webanwendungen und Webseiten existieren viele verschiedene Technologien und Konzepte. Jede dieser Technologien implementiert bestimmte Anforderungen, wie z.B. das Erzeugen von Inhalten oder die Kommunikation zwischen Client und Server. Verschiedene Konzepte helfen, diese Technologien innerhalb einer Webanwendung zusammenzufügen. Nicht zuletzt Architekturstile und -muster gehören zu diesen Konzepten. Die Diplomarbeit beschreibt einen Ansatz zur Erstellung einer Taxonomie dieser Technologien und Konzepte unter Zuhilfenahme der freien Enzyclopädie Wikipedia, im speziellen der Kategorie "Web-Application Framework". Unser 101companies Projekt benutzt Implementationen, um die einem Web-Application-Framework zugrunde liegenden Technologien zu identifizieren und zu klassifizieren. Innerhalb des Projekts werden Taxonomien und Ontologien mit Hilfe dieser Klassifikationen erstellt. Zusätzlich beschreibt die Ausarbeitung, wie nützliche Web-Application-Frameworks mit der Hilfe von Wikipedia priorisiert werden. Abschließend enthält die Diplomarbeit auch die Dokumentation der betreffenden Implementationen.
The estimation of various social objects is necessary in different fields of social life, science, education, etc. This estimation is usually used for forecasting, for evaluating of different properties and for other goals in complex man-machine systems. At present this estimation is possible by means of computer and mathematical simulation methods which is connected with significant difficulties, such as: - time-distributed process of receiving information about the object; - determination of a corresponding mathematical device and structure identification of the mathematical model; - approximation of the mathematical model to real data, generalization and parametric identification of the mathematical model; - identification of the structure of the links of the real social object. The solution of these problems is impossible without a special intellectual information system which combines different processes and allows predicting the behaviour of such an object. However, most existing information systems lead to the solution of only one special problem. From this point of view the development of a more general technology of designing such systems is very important. The technology of intellectual information system development for estimation and forecasting the professional ability of respondents in the sphere of education can be a concrete example of such a technology. Job orientation is necessary and topical in present economic conditions. It helps tornsolve the problem of expediency of investments to a certain sphere of education. Scientifically validated combined diagnostic methods of job orientation are necessary to carry out professional selection in higher education establishments. The requirements of a modern society are growing, with the earlier developed techniques being unable to correspond to them sufficiently. All these techniques lack an opportunity to account all necessary professional and personal characteristics. Therefore, it is necessary to use a system of various tests. Thus, the development of new methods of job orientation for entrants is necessary. The information model of the process of job orientation is necessary for this purpose. Therefore, it would be desirable to have an information system capable of giving recommendations concerning the choice of a trade on the basis of complex personal characteristics of entrants.
As Enterprise 2.0 (E2.0) initiatives are gradually moving out of the early experimentation phase it is time to focus greater attention on examining the structures, processes and operations surrounding E2.0 projects. In this paper we present the findings of an empirical study to investigate and understand the reasons for initiating E2.0 projects and the benefits being derived from them. Our study comprises seven in-depth case studies of E2.0 implementations. We develop a classification and means of visualising the scope of E2.0 initiatives and use these methods to analyse and compare projects.
Our findings indicate a wide range of motivations and combinations of technology in use and show a strong emphasis towards the content management functionality of E2.0 technologies.
Generalized methods for automated theorem proving can be used to compute formula transformations such as projection elimination and knowledge compilation. We present a framework based on clausal tableaux suited for such tasks. These tableaux are characterized independently of particular construction methods, but important features of empirically successful methods are taken into account, especially dependency directed backjumping and branch local operation. As an instance of that framework an adaption of DPLL is described. We show that knowledge compilation methods can be essentially improved by weaving projection elimination partially into the compilation phase.
Autonome Systeme, wie Roboter, sind bereits Teil unseres täglichen Lebens. Eine Sache, in der Menschen diesen Maschinen überlegen sind, ist die Fähigkeit, auf sein Gegenüber angemessen zu reagieren. Dies besteht nicht nur aus der Fähigkeit zu hören, was eine Person sagt, sondern auch daraus, ihre Mimik zu erkennen und zu interpretieren.
In dieser Bachelorarbeit wird ein System entwickelt, welches automatisch Gesichtsausdrücke erkennt und einer Emotion zuordnet. Das System arbeitet mit statischen Bildern und benutzt merkmalsbasierte Methoden zur Beschreibung von Gesichtsdaten. In dieser Arbeit werden gebräuchliche Schritte analysiert und aktuelle Methoden vorgestellt. Das beschriebene System basiert auf 2D-Merkmalen. Diese Merkmale werden im Gesicht detektiert. Ein neutraler Gesichtsausdruck wird nicht als Referenzbild benötigt. Das System extrahiert zwei Arten von Gesichtsparametern. Zum einen sind es Distanzen, die zwischen den Merkmalspunkten liegen. Zum anderen sind es Winkel, die zwischen den Linien liegen, die die Merkmalspunkte verbinden. Beide Arten von Parametern werden implementiert und getestet. Der Parametertyp, der die besten Ergebnisse liefert, wird schließlich in dem System benutzt. Eine Support Vector Machine (SVM) mit mehreren Klassen klassifiziert die Parameter. Das Ergebnis sind Kennzeichen von Action Units des Facial Action Coding Systems (FACS). Diese Kennzeichen werden einer Gesichtsemotion zugeordnet.
Diese Arbeit befasst sich mit den sechs Basisgesichtsausdrücken (glücklich, überrascht, traurig, ängstlich, wütend und angeekelt) plus dem neutralen Gesichtsausdruck. Das vorgestellte System wird in C++ implementiert und an das Robot Operating System (ROS) angebunden.
The natural and the artificial environment of mankind is of enormous complexity, and our means of understanding this complex environment are restricted unless we make use of simplified (but not oversimplified) dynamical models with the help of which we can explicate and communicate what we have understood in order to discuss among ourselves how to re-shape reality according to what our simulation models make us believe to be possible. Being both a science and an art, modelling and simulation isrnstill one of the core tools of extended thought experiments, and its use is still spreading into new application areas, particularly as the increasing availability of massive computational resources allows for simulating more and more complex target systems.
In the early summer of 2012, the 26th European Conference on Modelling andrnSimulation (ECMS) once again brings together the best experts and scientists in the field to present their ideas and research, and to discuss new challenges and directions for the field.
The 2012 edition of ECMS includes three new tracks, namely Simulation-BasedrnBusiness Research, Policy Modelling and Social Dynamics and Collective Behaviour, and extended the classical Finance and Economics track with Social Science. It attracted more than 110 papers, 125 participants from 21 countries and backgrounds ranging from electrical engineering to sociology.
This book was inspired by the event, and it was prepared to compile the most recent concepts, advances, challenges and ideas associated with modelling and computer simulation. It contains all papers carefully selected from the large number of submissions by the programme committee for presentation during the conference and is organised according to the still growing number tracks which shaped the event. The book is complemented by two invited pieces from other experts that discussed an emerging approach to modelling and a specialised application. rnrnWe hope these proceedings will serve as a reference to researchers and practitioners in the ever growing field as well as an inspiration to newcomers to the area of modelling and computer simulation. The editors are honoured and proud to present you with this carefully compiled selection of topics and publications in the field.
From September 4 to 11, 1992, a fiirst meeting between Ukrainian and German scientists interested in mathematical and computer modeling of social processes was held at Vorzel' near Kiev. The meeting had been planned for nearly three years by Igor V. Chernenko and Mikhail V. Kuz'min, then members of the research group on mathematical modeling in sociology at the Institute of Sociology of the Academy of Science of the Ukrainian Republic, and had to be postponed twice due to the political development in the former Soviet Union, but thanks to the organizers' perseverance (and in spite of a strike of the airport personell at Kiev Borispol Airport on the eve of the conference) the conference could at last be realized.rnThe main purpose of the conference was to discuss a synergetic interpretation of large-scale destructive social processes as catastrophic phenomena in self-organized systems.
This paper originates from the FP6 project "Emergence in the Loop (EMIL)" which explores the emergence of norms in artificial societies. Part of work package 3 of this project is a simulator that allows for simulation experiments in different scenarios, one of which is collaborative writing. The agents in this still prototypical implementation are able to perform certain actions, such as writing short texts, submitting them to a central collection of texts (the "encyclopaedia") or adding their texts to texts formerly prepared by other agents. At the same time they are able to comment upon others' texts, for instance checking for correct spelling, for double entries in the encyclopaedia or for plagiarisms. Findings of this kind lead to reproaching the original authors of blamable texts. Under certain conditions blamable activities are no longer performed after some time.
Particle Swarm Optimization ist ein Optimierungsverfahren, das auf der Simulation von Schwärmen basiert.
In dieser Arbeit wird ein modifizierter Algorithmus, der durch Khan et al. 2010 eingeführt wurde, zur Schätzung der lokalen Kamerapose in 6DOF verwendet. Die Poseschätzung basiert auf kontinuierlichen Farb-und Tiefendaten, die durch einen RGB-D Sensor zur Verfügung gestellt werden. Daten werden von unterschiedlichen Posen aufgenommen und als gemeinsames Model registriert. Die Genauigkeit und Berechnungsdauer der Implementierung wird mit aktuellen Algorithmen verglichen und in unterschiedlichen Konfigurationen evaluiert.
In Silico simulation of biological systems is an important sub area of computational biology (system biology), and becomes more and more an inherent part for research. Therefore, different kinds of software tools are required. At present, a multitude of tools for several areas exists, but the problem is that most of the tools are essentially application specific and cannot be combined. For instance, a software tool for the simulation of biochemical processes is not able to interact with tools for the morphology simulation and vice versa. In order to obtain realistic results with computer-aided simulations it is important to regard the biological system in its entirety. The objective is to develop a software framework, which provides an interface structure to combine existing simulation tools, and to offer an interaction between all affiliated systems. Consequently, it is possible to re-use existing models and simulation programs. Additionally, dependencies between those can be defined. The system is designed to interoperate as an extendable architecture for various tools. The thesis shows the usability and applicability of the software and discusses potential improvements.
In dieser Arbeit werden vier der derzeit verfügbaren webbasierten Visualisierungstools evaluiert (FlexViz, Jambalaya applet, Experimental jOWL TouchGraph, Plone ontology).
Im Rahmen dieser Arbeit werden weiterhin folgende Fragen erforscht: 1. Welche Anforderungen an die Visualisierung von Ontologien können definiert werden? 2. Wie kann eine Bewertung und Analyse durchgeführt werden? 3. Wie kann eine ausgewählte Methode getestet werden? Basierend auf den Ergebnissen dieser Fragestellungen stellte sich die Visualisierungsmethode FlexViz im gegebenen Szenario als beste heraus. FlexViz wurde eingesetzt, indem es in das Content Management System Plone integriert wurde. Es wurde mit VCD Ontologien getestet und schließlich aus der Sicht verschiedener Interessengruppen analysiert. Es zeigte sich, dass FlexViz ein gutes Tool zum Visualisieren, Verstehen und Analysieren von bereits entwickelten Ontologien ist, da die Benutzeroberfläche benutzerfreundlich und interaktiv ist. Jedoch zeigt FlexViz deutliche Schwächen bzgl. dem Bearbeiten, Erweitern und Entwickeln von Ontologien.
Prozedurale Synthese, das Erzeugen von Computerspielinhalten durch die Verwendung von pseudo-zufälligen Algorithmen, ist ein Themenbereich mit wachsendem Interesse in wirtschaftlichen und akademischen Kreisen, verdankt sowohl durch ihre Eignung zur Senkung von Entwicklungszeit und -kosten als auch durch die Möglichkeit, prozedurale Synthese zur Erzeugung von interessanten und einmaligen Spielwelten einzusetzen. Obwohl viele aktuelle Computerspiele prozedural generierte Inhalte verwenden, gebrauchen nur wenige diese Methoden zur Erzeugung realistischer Weltraum-Spielumgebungen, und die Umsetzbarkeit der Anwendung prozeduraler Synthese zur Entwicklung eines solchen Spiels wurde untersucht. Aktuelle Modelle aus der Forschung wurden verwendet, um eine in Echtzeit laufende astronomische Simulation in Python zu entwickeln, welche Stern- und Planetenobjekte in einer fiktiven Galaxie prozedural erzeugt, die als Spielwelt eines einfachen 2DWeltraumspiels dient, in welchem der Spieler nach intelligentem Leben suchen muss.
Querying for meta knowledge
(2008)
The Semantic Web is based on accessing and reusing RDF data from many different sources, which one may assign different levels of authority and credibility. Existing Semantic Web query languages, like SPARQL, have targeted the retrieval, combination and reuse of facts, but have so far ignored all aspects of meta knowledge, such as origins, authorship, recency or certainty of data, to name but a few. In this paper, we present an original, generic, formalized and implemented approach for managing many dimensions of meta knowledge, like source, authorship, certainty and others. The approach re-uses existing RDF modeling possibilities in order to represent meta knowledge. Then, it extends SPARQL query processing in such a way that given a SPARQL query for data, one may request meta knowledge without modifying the query proper. Thus, our approach achieves highly flexible and automatically coordinated querying for data and meta knowledge, while completely separating the two areas of concern.
In dieser Arbeit wird ein skelettbasiertes Matching-Verfahren für 2D-Objekte vorgestellt. Zunächst werden aktuelle Ansätze zum Matchen von Objekten vorgestellt, anschließend werden die Grundlagen von skelettbasiertem Matching erklärt. Ein skelettbasiertes Verfahren wurde im Rahmen dieser Arbeit gemäß dem vorliegenden Original-Paper neu implementiert. Diese Implementierung wird anhand einer Ähnlichkeitssuche in drei Bild-Datenbanken evaluiert. Stärken und Schwächen des Verfahrens werden herausgearbeitet. Des weiteren wird der vorgestellte Algorithmus auf Erweiterungen untersucht, die das Matchen von 3D-Objekten ermöglichen sollen. Im speziellen wird das Verfahren auf medizinische Daten angewendet: CT-Aufnahmen der abdominalen Aorta eines Patienten vor und nach einer Operation werden miteinander verglichen. Problemfälle und Erweiterungsansätze für das Matchen von 3D-Objekten im Allgemeinen und von Blutgefäßen im Speziellen werden vorgestellt.
In recent years, traceability has been more and more universally accepted as being a key factor for the success of software development projects. However, the multitude of different, not well-integrated taxonomies, approaches and technologies impedes the application of traceability techniques in practice. This paper presents a comprehensive view on traceability, pertaining to the whole software development process. Based on graph technology, it derives a seamless approach which combines all activities related to traceability information, namely definition, recording, identification, maintenance, retrieval, and utilization in one single conceptual framework. The presented approach is validated in the context of the ReDSeeDS-project aiming at requirements-based software reuse.
The way information is presented to users in online community platforms has an influence on the way the users create new information. This is the case, for instance, in question-answering fora, crowdsourcing platforms or other social computation settings. To better understand the effects of presentation policies on user activity, we introduce a generative model of user behaviour in this paper. Running simulations based on this user behaviour we demonstrate the ability of the model to evoke macro phenomena comparable to the ones observed on real world data.
Cloud Computing is a topic that has gained momentum in the last years. Current studies show that an increasing number of companies is evaluating the promised advantages and considering making use of cloud services. In this paper we investigate the phenomenon of cloud computing and its importance for the operation of ERP systems. We argue that the phenomenon of cloud computing could lead to a decisive change in the way business software is deployed in companies. Our reference framework contains three levels (IaaS, PaaS, SaaS) and clarifies the meaning of public, private and hybrid clouds. The three levels of cloud computing and their impact on ERP systems operation are discussed. From the literature we identify areas for future research and propose a research agenda.
Knowledge compilation is a common technique for propositional logic knowledge bases. A given knowledge base is transformed into a normal form, for which queries can be answered efficiently. This precompilation step is expensive, but it only has to be performed once. We apply this technique to concepts defined in the Description Logic ALC. We introduce a normal form called linkless normal form for ALC concepts and discuss an efficient satisability test for concepts given in this normal form. Furthermore, we will show how to efficiently calculate uniform interpolants of precompiled concepts w.r.t. a given signature.
In this paper, we compare two approaches for exploring large,rnhierarchical data spaces of social media data on mobile devicesrnusing facets. While the first approach arranges thernfacets in a 3x3 grid, the second approach makes use of arnscrollable list of facets for exploring the data. We have conductedrna between-group experiment of the two approachesrnwith 24 subjects (20 male, 4 female) executing the same set ofrntasks of typical mobile users" information needs. The resultsrnshow that the grid-based approach requires significantly morernclicks, but subjects need less time for completing the tasks.rnFurthermore, it shows that the additional clicks do not hamperrnthe subjects" satisfaction. Thus, the results suggest thatrnthe grid-based approach is a better choice for faceted searchrnon touchscreen mobile devices. To the best of our knowledge,rnsuch a summative evaluation of different approaches for facetedrnsearch on mobile devices has not been done so far.
In der vorliegenden Arbeit untersuche ich eine breite Spanne von Libraries, Frameworks und anderer Technologien für die Programmiersprache Haskell. Ich demonstriere verschiedene praktische Anwendungen von Haskell und trage durch Implementationen und Taxonomieeinheiten zum 101companies System bei. Dies tue ich, indem ich einen großen Teil des 101companies feature models abdecke und damit verknüpfte Definitionen von Begriffen und Technologien in dieses System einpflege. Die Implementationen zeigen, wie verschiedene Sprachkonzepte von Haskell wie ein sehr starkes Typensystem, Polymorphismus, Funktionen höherer Ordnungen und Monaden effektiv bei der Entwicklung von Informationssystemen benutzt werden können. In diesem Zusammenhang zeige ich sowohl Vorteile als auch Einschränkungen der verschiedenen Technologien auf.