Refine
Year of publication
- 2014 (24) (remove)
Document Type
- Bachelor Thesis (18)
- Master's Thesis (5)
- Doctoral Thesis (1)
Keywords
- OpenGL (3)
- Android (2)
- Graphik (2)
- Smartphone (2)
- VOXEL (2)
- 3D-Visualisierung (1)
- App (1)
- Applikation (1)
- Beleuchtung (1)
- Beschleunigungsensor (1)
Institute
- Institut für Computervisualistik (24) (remove)
Online Handschrifterkennung chinesischer Schriftzeichen auf androidfähigen mobilen Endgeräten
(2014)
Usage of mobile dictionaries or translators requires an input. This input has to be processed and recognized beforehand. Chinese characters are more suited for a handwritten input than a keyboard based one. Reason for that are the characters consisting mostly of pictograms or ideograms.
This thesis deals with an implementation of a prototypical recognition system on a mobile device. The recognition process should be online and therefore running while writing. It can save time for the user, because suggestions are made during runtime.
Basics and an overview over the current state of the art in online handwriting recognition will be given. An approach will be chosen and implemented, such that the recognition process is fast and needs little memory. The implementation will be tested and it will show, that a fast recognition can be possible on small devices. Suggestions for expansions and improvements will be given, including a future work part.
In this bachelor thesis, the question of whether or not a jump'n run game with sensor control for android devices is useful, is handled. To this end, a game was developed, which is once controlled with and without sensors at different levels. In a second version, the game is completely controlled by means of sensors, so that the controls can later be compared. It is explained how the game was planned, designed and investigated. In addition, it is checked whether games with sensor control already exist. The engine, which was used to developed the game, is also introduced. Finally, the evaluation is carried out for an elaborated user test on the playability of the game in terms of control.
In the man-machine interaction tracking and identification of individuals plays an important role. In this work, a framework for the service-robot Lisa, of the Active Vision Group, has been created to combine different methods for the detection, tracking and identification of individuals. First leg detection is performed to establish hypotheses for people using a 2D-laserscan. This assumption needs to be confirmed by an analysis of the Kinect point cloud. After successful confirmation online-boosting on RGB-data is performed for identification. The leg data will also be used with a linear Kalman filter to estimate the movement of people. Through the combination of of Kalman filter with leg detection and online-boosting people tracking should be enabled. Further receiving an interchange of persons should - by brief occlusion or faulty associate of legs - can be prevented.
Die Medizinische Visualisierung komplexer Gefäßbäume hat das Potential den klinischen Alltag in der Gefäßchirurgie zu erleichtern.
Dazu sind exakte, hochaufgelöste Darstellungen und echtzeitfähige Berechnungsmethoden notwendig. Bekannte Ansätze aus den Bereichen der direkten (z.B. Raycasting) und indirekten
(z.B. Marching Cubes) Volumenvisualisierung sind nicht in der Lage alle Anforderungen zufriedenstellend zu erfüllen. Verbesserte
Ergebnisse können mit hybriden Methoden erzielt werden, die unterschiedliche Visualisierungsverfahren kombinieren.
Im Rahmen dieser Arbeit wurde ein hybrides Renderingsystem zur Darstellung von Blutgefäßen entwickelt, das die Bildqualität durch Integration einer Marching Cubes Oberfläche in ein Raycasting–System optimiert, dabei Detailstrukturen erhält und ausreichende Performanz zur Interaktion bietet. Die Ergebnissezeigen die verbesserte Plastizität und Genauigkeit der Darstellung.Anhand von Experten– und Laienbefragungen konnte der Nutzen des Systems vor allem für die Patientenaufklärung nachgewiesen werden. Die Erschließung zusätzlicher Anwendungsgebiete ist durch die Weiterentwicklung des Renderers möglich.
The goal of this work is evaluation and optimization of several eye-tracking algorithms for estimation of relevant features regarding accuracy. The extracted features are pupil- and glintcenters. The algorithms are applicable to off the shelf cameras. A synthetic model of the eye was modified and utilized. The model was used to supply ground truth for the evaluation of the methods.
Der Markt der mobilen Endgeräte entwickelt sich schnell weiter und es kommen schon Kinder im frühsten Alter mit solchen Technologien in Berührung. Daher ist es wichtig, Kinder richtig an die Geräte heranzuführen. Von Vorteil wäre eine Einbindung von Smartphones und Tablets, im Bezug auf den Lernprozess, in den Unterricht. Die vorliegende Arbeit behandelt deshalb das Konzept einer Lernspielapp, die durch Pädagogen konfiguriert werden kann. Die Evaluation soll Aufschluss über die Motivation der Kinder geben und die Aufgeschlossenheit der Pädagogen gegenüber neuen Medien ermitteln.
This thesis covers the mathematical background of ray-casting as well as an exemplary implementation on graphics processing units, using a modern programming interface. The implementation is embedded within an editor, which enables the user to activate optimizations of the algorithm. Techniques like transfer functions and local illumination are available for a more realistic visualization of materials. Moreover, the user interface gives access to features like importing volumes, let one define a custom transfer function, holds controls to adjust parameters of rendering and allows to activate further techniques, which are also subject of discussion in this thesis. Benefit of all shown techniques is measured, whether it is expected to be visual or on the part of performance.
In den letzten Jahren ist eine steigende Verbreitung von Touchscreen-Geräten zu verzeichnen. Ihre Bedienung unterscheidet sich grundlegend von der mit Maus und Tastatur. Durch die Eingabe mit Gesten oder mehreren Fingern kann es schwierig sein den Aktionen eines Anderen zu folgen. Probleme entstehen durch die Verdeckung des Bildschirms mit der Eingabehand. Sieht man nur den Bildschirminhalt, zum Beispiel bei einer Videoübertragung, gehen Informationen über die Eingabe verloren.
In dieser Arbeit wird ein System entwickelt, das die kollaborative Arbeit an voneinander entfernten Touchscreen-Geräten verbessern soll. Dazu wird aus den Tiefendaten eines Kinect Sensors eine grafische Repräsentation der Eingabehand erstellt. Durch Einblendung dieser Visualisierung soll es einem Anwender erleichtert werden den Eingaben eines Anwenders zu folgen. Bedienkonzepte, wie zum Beispiel Gesten, sollen dadurch besser vermittelt werden. Außerdem soll so die Möglichkeit geschaffen werden, Informationen über eine gemeinsame Problematik effizienter auszutauschen. Deshalb wurde ein Testsystem mit zwei Arbeitsplätzen entwickelt. Darin übernimmt ein Anwender die Rolle des Erklärenden und führt einen zweiten Anwender, den Ausführenden, durch verschiedene Testszenarien. Hierbei stehen ihm bei einem Teil der Aufgaben die Visualisierung der Hand zur Verfügung, während er in anderen Aufgaben nur verbal mit seinem Gegenüber kommunizieren kann.
Im Rahmen einer Evaluation wird das System auf seine Effizienz zur Bedienung von Touchscreen-Systemen überprüft. Des Weiteren wird untersucht, inwieweit die grafische Qualität den gestellten Anforderungen genügt, um einen Mehrwert für die Anwendung zu bieten.
3D-models are getting more important in many areas such as multimedia applications, robotics or film industries. Of particular interest is the creation of 3D-models from a series of monocular images. This is because the cameras that are required for this purpose are becoming cheaper, smaller and more sophisticated at the same time. Increasingly often, suitable cameras are already integrated in devices like smartphones, tablet PCs or cars for example. Hence, there is a great potential for applications of this reconstruction technique.
This thesis is based on the use of a series of images that were taken with arncalibrated camera. The first step is to extract point correspondences from this image series making use of the well-known SURF- and A-KAZE-features. Starting from the point correspondences, it is possible to reconstruct a 3D-Modell with different algorithms that consists of a point cloud and camera poses. To reduce errors in the 3D-model, this thesis especially focuses on explaining the bundle adjustment algorithm, which is being used for a non-linear error minimization of a cost function.
The thesis also introduces the application for the 3D-reconstruction and the visualization of the results, that was developed in the course of this thesis.
The implemented system is evaluated based on statistics and the newly aquiredrnknowledge is presented. The thesis concludes with a summary of its results, and a number of ideas for potential future applications and developments.
Object recognition is a well-investigated area in image-based computer vision and several methods have been developed. Approaches based on Implicit Shape Models have recently become popular for recognizing objects in 2D images, which separate objects into fundamental visual object parts and spatial relationships between the individual parts. This knowledge is then used to identify unknown object instances. However, since the emergence of aσordable depth cameras like Microsoft Kinect, recognizing unknown objects in 3D point clouds has become an increasingly important task. In the context of indoor robot vision, an algorithm is developed that extends existing methods based on Implicit Shape Model approaches to the task of 3D object recognition.
The goal of this thesis is the development of methods for augmented image synthesis using 3D photo collections. 3D photo collections are representations of real scenes automatically generated from single photos and describe a scene as a set of images with known camera poses as well as a sparse point-based model of the scene geometry. The main goal is to perform a photo-realistic augmented image synthesis of real and virtual parts, where the real scene is provided as a 3D photo collection. Therefore, three main problems are addressed.
Since the photos may be represented in different device-specific RGB color spaces, a color characterization of the 3D photo collections is necessary to gain correct color information that is consistent with human perception. The proposed novel method automatically transforms all images into a common RGB color space and thereby simplifies color characterization of 3D photo collections.
As a main problem for augmented image synthesis, all environmental lighting has to be known in order to apply illumination to virtual parts that is consistent with the real portions shown in the photos. To solve this problem, two novel methods were developed to reconstruct the lighting from 3D photo collections.
In order to perform image synthesis for arbitrary views on the scene, an image-based approach was developed that generates new views in 3D photo collections making direct use of its point cloud. The novel method creates new views in real-time and allows free-navigation.
In conclusion, the proposed novel methods show that 3D photo collections are a useful representation for real scenes in Augmented Reality and they can be used to perform a realistic image synthesis of real and virtual portions.
The present thesis covers the implementation and optimization of global illumination in three-dimensional scenes. Global illumination does not only consider direct illumination dependent on one or more light sources, but also indirect illumination which is emitted by surrounding objects in the scene. The thesis focuses on the implementation of a global illumination method and its improvement using OpenGL 4.4. This is done by a voxelization of the scene. By traversing the resulting voxel structure, additional information is taken from the scene, which contributes to a plausible global illumination.
Abstract 3 This paper explains the convolution reverb, a method that enables users to add realistic sounding reverberation to audio material that was recorded in neutral sounding rooms. In particular, the possibility of computing the effect on the GPU using OpenCL is discussed, to make use of the high concurrency of the problem. This paper aims at the development of a VST plugin that utilizes the GPU accelerated convolution algorithm, so that it can be used for audio software solutions.
Zusätzlich zum Rendern wird die Rechenleistung moderner Grafikkarten immer häufiger auch für allgemeine Berechnungen (GPGPU) genutzt. Für die Umsetzung stehen verschiedene Möglichkeiten zur Verfügung, die von der Verwendung der Renderingpipeline bis zu eigenständigen Schnittstellen reichen. In dieser Arbeit werden mit Render-To-Texture, Transform Feedback, Compute Shader und OpenCL vier verschiedene GPGPU-Methoden untersucht. Anhand von Partikelsystemen werden sie hinsichtlich der benötigten Berechnungszeit, der GPU-Auslastung, Lines of Code und Portierbarkeit miteinander verglichen. Dazu wurden sowohl das N-Körper Problem, Smoothed Particle Hydrodynamics und ein Partikelschwarm als Partikelsysteme umgesetzt. Es konnte gezeigt werden, dass insbesondere OpenCL und Compute Shader sehr gute Ergebnisse liefern.
This thesis presents a prototypical application for speech therapy as a therapeutical tool, simulating the articulation of German phonemes and the swallowing reflex. A special attention is given to a three-dimensional visualization of anatomical models of the pharyngolaryngeal region, which can be used in an interactive way. For examining the benefits of such an application in relation to conventional therapy mediums, the thesis considers iteratively the requirements of speech therapists.
This work presents an application for simulation objects, which can change their aggregate states between solid and liquid using a temperature system. The focal points are the simulation of fluids with a particle system, the generation of a surface and the visualization of metal. The application should be interactive and match the real time conditions. Different types of Shader are used for the parallelized computations on the GPU. Also more options to use the application and possible improvements are presented.
Ziel der Bachelorarbeit war es, eine moderne Art des Geländespiels "Schnitzeljagd" zu entwickeln. Dieses sollte möglichst auf jedem aktuellen Smartphone spielbar sein, welches mit dem Betriebssystem Android arbeitet. Das Gelände ist begrenzt auf den Universitätscampus Koblenz, somit dient das Spiel auch dazu, den Campus besser kennenzulernen.
Den Benutzern der Campusjagd wird eine mobile Applikation geboten, welche sie anhand von Hinweisen und Rätseln über den kompletten Campus führt, um letztendlich an einen Zielort zu gelangen, an welchem sich ein "Schatz" befindet. Anstatt wie üblich auf dem Gelände Schnitzel bzw. Schnipsel mit Hinweisen zu verstecken, werden bei der Campusjagd QR-Codes aufgehängt, um so den Weg zu markieren. Zur Täuschung sind auch irrelevante Codes zu finden. Die Codes müssen in der richtigen Reihenfolge abgelaufen werden, d. h. von einem Code erhält der Spieler den Hinweis zum jeweiligen Nächsten. Außerdem ist es möglich, dass aus einem QR-Code mehrere Hinweise für nächste Stationen angezeigt werden.
Diese Arbeit beschreibt den Entwurf und die Umsetzung einer prototypischen Android-Applikation zum Dokumentieren von Reiserouten. Nach einer kurzen Einführung in die grundlegenden Aspekte des Reisens, der notwendigen Hilfsmittel sowie das Betriebssystem Android wird das Konzept und die Anforderungen an die Anwendung vorgestellt. Anschließend werden die Realisierung der Implementierung erläutert und die Ergebnisse der Evaluation offengelegt.
Erweiterung der Konzeption und Implementierung einer Screening Applikation für mobile Endgeräte
(2014)
In this bachelor thesis an existing generic concept and an existing prototype for a smartphone application to record, monitor and document physical symptoms or observations of the human body are being extended. The existing funktionalities are being complemented by analysis of the previous Prototype. The concept and its Function modules, which are implemented in the existing prototype for the mobile platform Android, are being extended based on their analysed weaknesses. The resulting prototype and generic concept are evaluated and optimizations and extensions are being collected for further projects.
Tiny waves driven by wind, shallow, long waves, head overlapping sea, all of these waves occur in every ocean and even in small lakes. The surface of water is one of the most versatile phenomenas of nature. Not only the movement of waves, but also the reflection of sky, sun and coastline makes the surface of water unique. Exactly this complexity is what brings its own challenges to the simulation of water surfaces. That is why simulation of water occupies mathematicians with a challenge for nearly 400 years now.
In the last fifty years this challenge has more and more shifted to computer science. Computer graphic designers have tried to visualise water in a realistic manner for centuries. Science in this field expends from simple noise filters to mathematically complex solutions like Fourier Transformation.
In the following work historical background of todays wave theories, as well as mathematical fundamentals are given. The focus of this work is set on the implementation of these methods in OpenGL 3.3.