Master's Thesis
Refine
Document Type
- Master's Thesis (34) (remove)
Keywords
- Computersimulation (2)
- Alex Bernstein (1)
- Artificial Intelligence (1)
- Classification (1)
- Clustering (1)
- Criteria Matrix (1)
- Data Mining (1)
- Data analysis (1)
- Datenanalyse (1)
- Dynamische Analyse (1)
- Echtzeitphysik (1)
- Interpersonale Kommunikation (1)
- Klassifikation (1)
- Massenkommunikation (1)
- Mehragentensystem (1)
- Menschenmengen-Evakuierung (1)
- Mikrocontroller (1)
- Mindestlohn (1)
- Minimum wage (1)
- Mobile BI (1)
- Mobile Business Intelligence (1)
- Modell (1)
- Netlogo (1)
- OpenVDB (1)
- Politische Kampagne (1)
- Politische Meinungsbildung (1)
- Real-Time (1)
- Reengineering (1)
- Regression (1)
- Robert P. Abelson (1)
- Service identification (1)
- Service-Identifizierung (1)
- Serviceorientierte Architektur (1)
- Simulation (1)
- Software migration (1)
- Software-Migration (1)
- Sport (1)
- Sports (1)
- Text Analysis (1)
- Text Mining (1)
- Tool Evaluation (1)
- USB (1)
- Videospiele (1)
- Web Analytics (1)
- Web Analytics Framework (1)
- Web Mining (1)
- crowd evacuation (1)
- physical interaction (1)
- physische Interaktion (1)
- psychisches Verhalten (1)
- psychological behavior (1)
- realtime physics (1)
- revenue management (1)
- simulation (1)
- videogames (1)
Institute
- Fachbereich 4 (34) (remove)
Particle swarm optimization is an optimization technique based on simulation of the social behavior of swarms.
The goal of this thesis is to solve 6DOF local pose estimation using a modified particle swarm technique introduced by Khan et al. in 2010. Local pose estimation is achieved by using continuous depth and color data from a RGB-D sensor. Datasets are aquired from different camera poses and registered into a common model. Accuracy and computation time of the implementation is compared to state of the art algorithms and evaluated in different configurations.
Diese Masterarbeit beschreibt und evaluiert die im Rahmen der Arbeit entwickelten Erweiterungen zur Schellings Segregationsimulation. Der Nobelpreisträger Thomas C. Schelling simulierte als erster ab 1969 die Segregation. Die dabei entstandenen Ergebnisse erlaubten eine genauere Analyse der auftretenden Phänomene. Uri Wilensky, Mitarbeiter der "Northwestern University", entwickelte seit 1999 eine Segregationssimulation in NetLogo.
Im Rahmen meiner Masterarbeit habe ich das NetLogo-Model von Uri Wilensky in sechs unterschiedlichen Szenarien weiterentwickelt.
The purpose of this master thesis is to enable the Robot Lisa to process complex commands and extract the necessary information in order to perform a complex task as a sequence of smaller tasks. This is intended to be achieved by the improvement of the understanding that Lisa has of her environment by adding semantics to the maps that she builds. The complex command itself will be expected to be already parsed. Therefore the way the input is processed to become a parsed command is out of the scope of this work. Maps that Lisa builds will be improved by the addition of semantic annotations that can include any kind of information that might be useful for the performance of generic tasks. This can include (but not necessarily limited to) hierarchical classifications of locations, objects and surfaces. The processing of the command in addition to some information of the environment shall trigger the performance of a sequence of actions. These actions are expected to be included in Lisa- currently implemented tasks and will rely on the currently existing modules that perform them.
Nevertheless the aim of this work is not only to be able to use currently implemented tasks in a more complex sequence of actions but also make it easier to add new tasks to the complex commands that Lisa can perform.
This thesis aims to clarify the question of how stakeholders can participate in a project. We discuss methods on how to select the relevant stakeholders from the whole set of the stakeholders and how to involve these stakeholders in the participatory process. The participation process does not only classical methods of participation such as workshops, especially before modern Web 2.0 techniques.
The World Wide Web (WWW) has become a very important communication channel. Its usage has steadily grown within the past. Interest by website owners in identifying user behaviour has been around since Tim Berners-Lee developed the first web browser in 1990. But as the influence of the online channel today eclipses all other media the interest in monitoring website usage and user activities has intensified as well. Gathering and analysing data about the usage of websites can help to understand customer behaviour, improve services and potentially increase profit.
It is further essential for ensuring effective website design and management, efficient mass customization and effective marketing. Web Analytics (WA) is the area addressing these considerations. However, changing technologies and evolving Web Analytic methods and processes present a challenge to organisations starting with Web Analytic programmes. Because of lacking resources in different areas and other types of websites especially small and medium-sized enterprises (SME) as well as non-profit organisations struggle to operate WA in an effective manner.
This research project aims to identify the existing gap between theory, tool possibilities and business needs for undertaking Web Analytic programmes. Therefore the topic was looked at from three different ways: the academic literature, Web Analytic tools and an interpretative case study. The researcher utilized an action research approach to investigate Web Analytics presenting an holistic overview and to identify the gaps that exists. The outcome of this research project is an overall framework, which provides guidance for SMEs who operate information websites on how to proceed in a Web Analytic programme.
Large amounts of qualitative data make the utilization of computer-assisted methods for their analysis inevitable. In this thesis Text Mining as an interdisciplinary approach, as well as the methods established in the empirical social sciences for analyzing written utterances are introduced. On this basis a process of extracting concept networks from texts is outlined and the possibilities of utilitzing natural language processing methods within are highlighted. The core of this process is text processing, to whose execution software solutions supporting manual as well as automated work are necessary. The requirements to be met by these solutions, against the background of the initiating project GLODERS, which is devoted to investigating extortion racket systems as part of the global fiσnancial system, are presented, and their fulσlment by the two most preeminent candidates reviewed. The gap between theory and pratical application is closed by a prototypical application of the method to a data set of the research project utilizing the two given software solutions.
ERP market analysis
(2013)
Der aktuelle ERP Markt wird dominiert von den fünf größten Anbietern SAP, Oracle, Microsoft, Infor und Sage. Da der Markt und die angebotenen Lösungen vielfältig sind, bedarf es einer fundierten Analyse der Systeme. Die Arbeit beleuchtet dabei anhand ausgesuchter Literatur und Kennzahlen der verschiedenen Unternehmen die theoretische Seite der angebotenen Lösungen der fünf großen ERP Anbieter. Daneben wird die Nutzung der Systeme in der Praxis anhand der Befragung von sechs Anwendern analysiert und die Systeme miteinander verglichen.
Ziel der Arbeit ist es, dass die Forschungsfragen beantwortet werden und dass es bezogen auf die Systeme dem Leser der Arbeit ersichtlich wird, welches ERP System für welche Unternehmensbranche und Unternehmensgröße am besten geeignet ist.
Des Weiteren gibt die Arbeit Aufschluss darüber, welche Trends für ERP Systeme für die Zukunft zu erwarten sind und welche Herausforderungen sich dadurch für die Unternehmen stellen.
Im Rahmen des "Design Thinking"-Prozesses kommen unterschiedliche Varianten kreativitätsfördernder Techniken zum Einsatz. Aufgrund der zunehmenden Globalisierung ergeben sich immer häufiger Kollaborationen, bei denen sich die jeweiligen Projektteilnehmer an verteilten Standorten befinden. Somit erweist sich die Digitalisierung des Design-Prozesses als durchaus erstrebenswert. Ziel der hier vorliegenden Untersuchung ist daher die Erstellung eines Bewertungsschemas, welches die Eignung digitaler Kreativitätstechniken in Bezug auf das "Entrepreneurial Design Thinking" misst. Des Weiteren soll geprüft werden, inwiefern sich der Einsatz von e-Learning-Systemen in Kombination mit der Verwendung digitaler Kreativitätstechniken eignet. Diese Prüfung soll am Beispiel der e-Learning Software "WebCT" konkretisiert werden. Hieraus ergibt sich die folgende Fragestellung: Welche digitalen Kreativitätstechniken eignen sich für die Anwendung im Bereich des "Entrepreneurial Design Thinkings" unter Einsatz der e-Learning Plattform "WebCT"? Zunächst wird eine Literaturanalyse bezüglich des "Entrepreneurial Design Thinkings", der klassische und digitale Kreativitätstechniken sowie des Arbeitens in Gruppen, was auch das Content Management, e-Learning-Systeme und die Plattform "WebCT" beinhaltet, durchgeführt. Im Anschluss daran wird eine qualitative Untersuchung durchgeführt. Auf Basis bereits bestehender Literatur, soll ein Bewertungsschema erstellt werden, welches misst, welche der behandelten digitalen Kreativitätstechniken für den Einsatz im "Entrepreneurial Design Thinking" am besten geeignet ist. Darauf aufbauend erfolgt die Verknüpfung des digitalisierten "Design Thinking"-Prozesses mit der e-Learning Plattform "WebCT". Abschließend wird diskutiert, in wie fern diese Zusammenführung als sinnvoll erachtet werden kann.
In a software reengineering task legacy systems are adapted computer-aided to new requirements. For this an efficient representation of all data and information is needed. TGraphs are a suitable representation because all vertices and edges are typed and may have attributes. Further more there exists a global sequence of all graph elements and for each vertex exists a sequence of all incidences. In this thesis the "Extractor Description Language" (EDL) was developed. It can be used to generate an extractor out of a syntax description, which is extended by semantic actions. The generated extractor can be used to create a TGraph representation of the input data. In contrast to classical parser generators EDL support ambiguous grammars, modularization, symbol table stacks and island grammars. These features simplify the creation of the syntax description. The collected requirements for EDL are used to determine an existing parser generator which is suitable to realize the requirements.
After that the syntax and semantics of EDL are described and implemented using the suitable parser generator. Following two extractors one for XML and one for Java are created with help of EDL. Finally the time they need to process some input data is measured.
This master thesis deals basically with the design and implementation of a path planning system based on rapidly exploring search trees for general-n-trailers. This is a probabilistic method that is characterized by a fast and uniform exploration. The method is well established, however, has been applied only to vehicles with simple kinematics to date. General-n-trailers represent a particular challenge as their controllability is limited. For this reason the focus of this thesis rests on the application of the mentioned procedure to general-n-trailers. In this context systematic correlations between the characteristics of general-n-trailers and the possibilities for the realization and application of the method are analyzed.