Refine
Year of publication
Document Type
- Bachelor Thesis (100) (remove)
Keywords
- Computervisualistik (8)
- OpenGL (7)
- Computergrafik (5)
- Android (4)
- Graphik (4)
- Line Space (4)
- Augmented Reality (3)
- C++ (3)
- Compute Shader (3)
- Computergraphik (3)
- GPGPU (3)
- Linespace (3)
- OpenGL Shading Language (3)
- Programmierung (3)
- Virtual Reality (3)
- Applikation (2)
- Bildverarbeitung (2)
- Datenstruktur (2)
- Global Illumination (2)
- Grafikkarte (2)
- Immersion (2)
- Informatik (2)
- Path Tracing (2)
- Physiksimulation (2)
- Raytracing (2)
- Realistische Computergrafik (2)
- Rendering (2)
- Shader (2)
- Smartphone (2)
- VOXEL (2)
- 3D (1)
- Android development (1)
- Androidentwicklung (1)
- App (1)
- Assembly (1)
- Astrophysik (1)
- Automatische Klassifikation (1)
- Autorensystem (1)
- Beleuchtung (1)
- Beschleunigungsensor (1)
- Bildanalyse (1)
- Blickpunktabhängig (1)
- Bounding Volume Hierarchy (1)
- Brechung (1)
- Coloskopie (1)
- Compute-Shader (1)
- Computer Graphics (1)
- Constitutive Model (1)
- Cybersickness (1)
- Darmpolyp (1)
- Digitale Bilder (1)
- ECSA (1)
- Echtzeit Raytracing (1)
- Entity Component System Architecture (1)
- Entwicklung (1)
- Erweiterte Realität (1)
- Eulerian grid (1)
- Extended Reality (1)
- Fabric Simulation (1)
- Fast Convolution (1)
- Fell (1)
- Fell Rendering (1)
- Flow-Erlebnis (1)
- Fluidsimulation (1)
- Fourier (1)
- Fresnel (1)
- GPU (1)
- Gerstner (1)
- Geste (1)
- Gesteninterpretation (1)
- Glint Center Estimation Algorithmus (1)
- Globale Beleuchtung (1)
- Grafik (1)
- Grafikprogrammierung (1)
- Graphicsprogramming (1)
- Grundschule (1)
- Haare (1)
- Halbschatten (1)
- HoloLens (1)
- Innenarchitektur (1)
- Jump'n Run-Spiel (1)
- Kartesisches Euler-Gitter (1)
- Kartesisches Gitter (1)
- Kinect (1)
- Kollaboration (1)
- Kollisionserkennung (1)
- Konturfindung (1)
- Leap-Motion (1)
- Lehrer (1)
- Lernspiel (1)
- Light Injection (1)
- Material Point Method (1)
- Material point method (1)
- Materialmodell (1)
- Materie-Punkt-Methode (1)
- Merkmalsdetektion (1)
- Modellierungstool (1)
- Montage (1)
- Motion Sickness (1)
- Musik (1)
- N-Body Simulation (1)
- N-Körper Simulation (1)
- Nvidia CUDA (1)
- OpenCL (1)
- Par (1)
- Partikel (1)
- Partikelsystem (1)
- Pathtracing (1)
- Pre-Integration (1)
- Pupil Center Estimation Algorithmus (1)
- Ray Tracer (1)
- Ray tracing (1)
- Ray-Casting (1)
- Reflections (1)
- Reflektionen (1)
- Reflexion (1)
- Reverberation (1)
- Schnee (1)
- Schnelle Faltung (1)
- Sensor (1)
- Sensorensteuerung (1)
- Specular (1)
- Spiel (1)
- Stadt (1)
- Starrkörper (1)
- Sterne (1)
- Steuerung (1)
- Stoffsimulation (1)
- Text (1)
- Texterkennung (1)
- Touchscreen (1)
- Transfer function (1)
- Transferfunktion (1)
- VR (1)
- VST (1)
- Virtuelle Realität (1)
- Visualisierung (1)
- Volumen (1)
- Volumen-Rendering (1)
- Volumendaten (1)
- Voxelisierung (1)
- Wasseroberfläche (1)
- Wellen (1)
- Wiederhall (1)
- accelerometer (1)
- android (1)
- area light (1)
- fur rendering (1)
- general purpose (1)
- graphics card (1)
- indirect lighting (1)
- indirektes Licht (1)
- interactive application (1)
- interaktive Applikation (1)
- interaktiven (1)
- jump'n run game (1)
- line space (1)
- menschliches Auge (1)
- multidimensional transfer function (1)
- multidimensionale Transferfunktion (1)
- physics simulation (1)
- prozedural (1)
- ray casting (1)
- raycasting (1)
- raytracing (1)
- reflection (1)
- refraction (1)
- rigid body (1)
- sensor control (1)
- shader (1)
- snow (1)
- soft shadow (1)
- volume rendering (1)
- volume visualization (1)
- volumenrendering (1)
- voxelization (1)
- water surface (1)
- waves (1)
Institute
- Institut für Computervisualistik (100) (remove)
Colonoscopy is the gold standard for the detection of colorectal polyps that can progress into cancer. In such an examination, physicians search for polyps in endoscopic images. Thereby polyps can be removed. To support experts with a computer-aided diagnosis system, the University of Koblenz-Landau currently makes some efforts in research different methods for automatic detection. Comparable to traditional pattern recognition systems, features are initially extracted and a classifier is trained on such data. Afterwards, unknown endoscopic images can be classified with the previously trained classifier. This thesis concentrates on the extension of the feature extraction module in the existing system. New detection methods are compared to existing techniques. Several features are implemented, incorporating Graylevel Co-occurrence Matrices, Local Binary Patterns and Discrte Wavelet Transform. Different modifications on those features are applied and evaaluated.
This thesis deals with the development of an authoring system for modeling 3D environments with physical description. In contrast to creating scenes in other common modeling tools, one can now compute and describe physical entities of a scene additional to the usual geometry. It is very important for those authoring systems to be extendable and customizable for specific requirement of the user. The focus lies on developing simple program architecture, which is easy to extend and to modify.
There are a few systems high and low-cost ones for gaze tracking. Normally low-cost systems go in hand with low-resolution cameras. Here the image quality is poor, so the algorithms for detecting the gaze have to work more precisely. But how to test and analyse them correctly, when there is a bad image quality and no reference point known? The idea of this work is, to generate synthetic eye images, where the reference points are known, because they are mainly manually set and then to test and analyse the algorithms with these synthetic images. By switching on features like gaussian noise or a second glint-like reflection point, it is possible to stepwise approximate the synthetic images close to reality. In fact the experiments will lead to an improvement of the algorithms used in a low-resolution system environment.
The present work starts with an introduction of methods for three-dimensional curve skeletonization. Different kinds of historic and recent skeletonization approaches are analysed in detail. Later on, a state-of-the-art skeletonization algorithm is introduced. This algorithm deals as a basis for the own approach presented subsequently. After the description and definition of a new method improving the state-of-the-art algorithm, experiments are conducted to get appraisable results. Next, a ground truth is described which has been set up manually by humans. The human similarity evaluations are compared with the results of the automatic computer-based similarity measures provided by the own approach. For this comparison, standard evaluation criteria from the field of information retrieval have been used.
Texture-based text detection in digital images using wavelet features and support vector machines
(2010)
In this bachelor thesis a new texture-based approach for the detection of text in digital images is presented. The procedure can be essentially divided into two main tasks, in detection of text blocks and detection of individual words, whereby the individual words are extracted from the detected text blocks. Roughly, the developed method acts with multiple support vector machines, which classify possible text regions of an image into real text regions, using wavelet-based features. In the process the possible text regions are defifined by edge projections with diσerent orientations. The results of the approach are X/Y coordinates, width and height of rectangular regions of an image, which contains individual words. This knowledge can be further processed, for example by an optical character recognition software to get the important and useful text information.
The following thesis analyses the functionality and programming capabilitiesrnof compute shaders. For this purpose, chapter 2 gives an introductionrnto compute shaders by showing how they work and how they can be programmed. In addition, the interaction of compute shaders and OpenGL 4.3 is shown through two introductory examples. Chapter 3 describes an NBodyrnsimulation that has been implemented in order to show the computational power of compute shaders and the use of shared memory. Then it is shown in chapter 4 how compute shaders can be used for physical simulationsrnand where problems may arise. In chapter 5 a specially conceived and implemented algorithm for detecting lines in images is described and then compared with the Hough transform. Lastly, a final conclusion is drawn in chapter 6.
Das Kommunikationsverhalten hat sich in den letzten Jahren durch die Smartphonenutzung verändert. Die Nutzer kommunizieren oft nur noch über den elektronischen Weg. Die persönliche Kommunikation, außerhalb des Smartphones, nimmt ab. Das Umfeld gerät unterdessen in Vergessenheit. In der vorliegenden Arbeit werden verschiedene Spielkonzepte entwickelt, welche die Kommunikation steigern sollen. Realisiert wird der Ansatz in einer prototypischen Stadtführer-App, nach den Spielkonzepten von "Scotland Yard" und "Schnitzeljagd". Während der Nutzung müssen die Spieler verschiedene Aufgaben lösen. Welches Spielkonzept sich in Bezug auf die Kommunikationsförderung am besten eignet, wird in einer Evaluation analysiert.
Der Markt der mobilen Endgeräte entwickelt sich schnell weiter und es kommen schon Kinder im frühsten Alter mit solchen Technologien in Berührung. Daher ist es wichtig, Kinder richtig an die Geräte heranzuführen. Von Vorteil wäre eine Einbindung von Smartphones und Tablets, im Bezug auf den Lernprozess, in den Unterricht. Die vorliegende Arbeit behandelt deshalb das Konzept einer Lernspielapp, die durch Pädagogen konfiguriert werden kann. Die Evaluation soll Aufschluss über die Motivation der Kinder geben und die Aufgeschlossenheit der Pädagogen gegenüber neuen Medien ermitteln.
3D-models are getting more important in many areas such as multimedia applications, robotics or film industries. Of particular interest is the creation of 3D-models from a series of monocular images. This is because the cameras that are required for this purpose are becoming cheaper, smaller and more sophisticated at the same time. Increasingly often, suitable cameras are already integrated in devices like smartphones, tablet PCs or cars for example. Hence, there is a great potential for applications of this reconstruction technique.
This thesis is based on the use of a series of images that were taken with arncalibrated camera. The first step is to extract point correspondences from this image series making use of the well-known SURF- and A-KAZE-features. Starting from the point correspondences, it is possible to reconstruct a 3D-Modell with different algorithms that consists of a point cloud and camera poses. To reduce errors in the 3D-model, this thesis especially focuses on explaining the bundle adjustment algorithm, which is being used for a non-linear error minimization of a cost function.
The thesis also introduces the application for the 3D-reconstruction and the visualization of the results, that was developed in the course of this thesis.
The implemented system is evaluated based on statistics and the newly aquiredrnknowledge is presented. The thesis concludes with a summary of its results, and a number of ideas for potential future applications and developments.
This thesis imparts a general view of the mechanics and implementation of latest voxelization strategies using the GPU. In addition to established voxelization procedures using the rasterization pipeline, new possibilities arising from GPGPU programming are examined. On the basis of the programming language C++ and the graphics library OpenGL the implementation of several methods is explained. The methods are compared in terms of performance and quality of the resulting voxelization and are evaluated critically with regards to possible use cases. Furthermore, two exemplary applications are detailed that use a voxelized scene in such a way that the augmentation of established techniques of real time graphics are facilitated. To this end, the concepts and the implementations of Transmittance Shadow Mapping and of Reflective Shadow Mapping utilizing a voxel based ambient occlusion effect is explained. Finally, the prolonging relevance of voxelization is put into prospect, by addressing latest research and further enhancements and applications of the presented methods.