Refine
Year of publication
Document Type
- Doctoral Thesis (470)
- Part of Periodical (354)
- Bachelor Thesis (275)
- Diploma Thesis (196)
- Master's Thesis (190)
- Study Thesis (138)
- Article (14)
- Conference Proceedings (12)
- Other (8)
- Report (8)
Language
- German (1137)
- English (534)
- Multiple languages (4)
- Spanish (2)
- (1)
Keywords
- Vorlesungsverzeichnis (55)
- Bildverarbeitung (16)
- Augmented Reality (15)
- Simulation (12)
- Computergraphik (10)
- Computersimulation (10)
- Pestizid (10)
- Robotik (10)
- Computergrafik (9)
- Computervisualistik (9)
Institute
- Institut für Computervisualistik (336)
- Fachbereich 4 (298)
- Zentrale Einrichtungen (176)
- Institut für Wirtschafts- und Verwaltungsinformatik (147)
- Institut für Informatik (143)
- Institut für Management (116)
- Fachbereich 7 (99)
- Institute for Web Science and Technologies (57)
- Institut für Softwaretechnik (54)
- Fachbereich 8 (47)
Social media platforms such as Twitter or Reddit allow users almost unrestricted access to publish their opinions on recent events or discuss trending topics. While the majority of users approach these platforms innocently, some groups have set their mind on spreading misinformation and influencing or manipulating public opinion. These groups disguise as native users from various countries to spread frequently manufactured articles, strong polarizing opinions in the political spectrum and possibly become providers of hate-speech or extremely political positions. This thesis aims to implement an AutoML pipeline for identifying second language speakers from English social media texts. We investigate style differences of text in different topics and across the platforms Reddit and Twitter, and analyse linguistic features. We employ feature-based models with datasets from Reddit, which include mostly English conversation from European users, and Twitter, which was newly created by collecting English tweets from selected trending topics in different countries. The pipeline classifies language family, native language and origin (Native or non-Native English speakers) of a given textual input. We evaluate the resulting classifications by comparing prediction accuracy, precision and F1 scores of our classification pipeline to traditional machine learning processes. Lastly, we compare the results from each dataset and find differences in language use for topics and platforms. We obtained high prediction accuracy for all categories on the Twitter dataset and observed high variance in features such as average text length especially for Balto-Slavic countries.
Will Eisners Graphic Novels zeugen von einer tiefgehenden Identitifation mit dem Judentum als Volkszugehörigkeit, Religion und Kultur und spiegeln das Judentum in all seinen Facetten wider. Dabei ist besonders hervorzuheben, dass die Entwicklung des Gesamtwerks parallel verläuft zur Geschichte der Emanzipation der jüdischen Bevölkerung in New York City. Der Band clustert die jüdischen Aspekte in Eisners Werk in beispielsweise Faktoren kollektiver Erinnerung, Umgang mit und Kampf gegen Antisemitismus und religiöse Besonderheiten.
As Enterprise 2.0 (E2.0) initiatives are gradually moving out of the early experimentation phase it is time to focus greater attention on examining the structures, processes and operations surrounding E2.0 projects. In this paper we present the findings of an empirical study to investigate and understand the reasons for initiating E2.0 projects and the benefits being derived from them. Our study comprises seven in-depth case studies of E2.0 implementations. We develop a classification and means of visualising the scope of E2.0 initiatives and use these methods to analyse and compare projects.
Our findings indicate a wide range of motivations and combinations of technology in use and show a strong emphasis towards the content management functionality of E2.0 technologies.
Remote Working Study 2022
(2022)
The Remote Working Study 2022 is focused on the transition to work from home (WFH) triggered by the stay at home directives of 2020. These directives required employees to work in their private premises wherever possible to reduce the transmission of the coronavirus. The study, conducted by the Center for Enterprise Information Research (CEIR) at the University of Koblenz from December 2021 to January 2022, explores the transition to remote working.
The objective of the survey is to collect baseline information about organisations’ remote work experiences during and immediately following the COVID-19 lockdowns. The survey was completed by the key persons responsible for the implementation and/or management of the digital workplace in 19 German and Swiss organisations.
The data presented in this report was collected from member organisations of the IndustryConnect initiative. IndustryConnect is a university-industry research programme that is coordinated by researchers from the University of Koblenz. It focuses on research in the areas of the digital workplace and enterprise collaboration technologies, and facilitates the generation of new research insights and the exchange of experiences among user companies.
The Internet of Things (IoT) is a concept in which connected physical objects are integrated into the virtual world to become active partakers of businesses and everyday processes (Uckelmann, Harrison and Michahelles, 2011; Shrouf, Ordieres and Miragliotta, 2014). It is expected to have a major impact on businesses (Council, Nic and Intelligence, 2008), but small and medium enterprises’ business models are threatened if they do not adopt the new concept (Sommer, 2015). Thus, this thesis aims to showcase a sample implementation of connected devices in a small enterprise, demonstrating its added benefits for the business.
Design Science Research (DSR) is used to develop a prototype based on a use case provided by a carpentry. The prototype comprises a hardware sensor and a web application which can be used by the wood shop to improve their processes. The thesis documents the iterative process of developing a prototype from the grounds up to useable hard- and software.
This contribution provides an example of how IoT can be used and implemented at a small business.
Im Vergleich zu herkömmlicher Computergrafik (perspektivische Projektion) bietet Raytracing entscheidende Vorteile, die hauptsächlich in der vergleichsweise hohen physikalischen Korrektheit der Methode begründet sind. Die Schwächen liegen hingegen im immensen Rechenaufwand.
Ein Raytracer ist vergleichsweise so rechenintensiv, weil für jeden Pixel mindestens ein Strahl verschickt werden muss. Dieser muss gegen alle Objekte im Raum geschnitten werden. Hinzu kommen noch die Strahlen, die entstehen, wenn Strahlen an Objekten reflektiert werden (Rekursion). Um diesen Rechenaufwand zu verkleinern und zusätzlich ein besseres Bild zu erzeugen, soll der adaptive Sampler den Raytracer unterstützen. Der adaptive Sampler soll während des Rendervorgangs den progressiven Fortschritt in der Bildgenerierung beobachten und Pixel von der weiteren Berechnung ausschließen, für die sich ein zusätzliches Verschießen von Strahlen nicht mehr lohnt.
Anders als der rein progressive Raytracer hört der adaptive Sampler mit dem Konvergieren des Bildes auf zu rechnen. Der adaptive Sampler soll so dafür sorgen, dass schneller ein besseres Bild erzeugt wird und somit die Performanz gesteigert wird. Insgesamt erwartet man sich vom adaptiven Sampler Vorteile bei der Berechnung von bestimmten Szenen. Unter anderem eine Verbesserung bei Szenen mit rein diffus beleuchteten Bildbereichen, sowie eine Verbesserung bei Szenen mit unterschiedlich rechenintensiven Bildbereichen. Ein normaler Raytracer kann nicht beurteilen, wie sinnvoll seine Schüsse sind. Er kann nur mehr Strahlen verschießen, in der Hoffnung, das Bild damit effektiv zu verbessern.
Es gibt jedoch viele Szenarien, bei denen eine linear steigende Schussanzahl pro Pixel keine gleichmäßige Verbesserung im Bild erzeugt. Das bedeutet, dass Bereiche im Bild schon gut aussehen, während andere noch sehr verrauscht sind. Man möchte nun Bildbereiche, die bereits konvergiert sind, in denen sich ein weiterer Beschuss also nicht mehr bemerkbar macht, ausschließen und die Rechenleistung dort nutzen, wo man sie noch braucht.
Wichtig dabei ist, dass Pixel nicht ungewollt zu früh von der Berechnung ausgeschlossen werden, die nicht weit genug konvergiert sind. Der adaptive Sampler soll so lange arbeiten, bis jeder Pixel dauerhaft keine Änderungen mehr vorweist. Das bedeutet, dass die Wahrscheinlichkeit für eine signifikante Farbänderung eines Pixels durch Verschießen eines Strahls (bei mehreren Lichtquellen in RenderGin mehrere Strahlen pro Pixel) klein genug ist. Es wird zwar intern keine Wahrscheinlichkeit berechnet, jedoch bekommt der Raytracer eine Art Gedächtnis: Er speichert die Veränderungen im beleuchteten Bild und deren Verlauf in eigenen Gedächtnisbildern. Das "Gedächtnis" für das alte Bild (Zustand des Bildes in der letzten Iteration über die Pixel) repräsentiert dabei das Kurzzeitgedächtnis. Es ist absolut genau. Das Langzeitgedächtnis wird von drei verschiedenen Bildern repräsentiert. Das erste gibt die Anzahl der verschossenen Strahlen pro Pixel an. Das zweite ist ein Wahrheitswertebild, das für jeden Pixel angibt, ob dieser noch in die Berechnung einbezogen werden soll. Das dritte Bild gibt an, wie oft jeder Pixel eine Farbänderung vollzogen hat, die geringer ist als der geforderte Maximalabstand eines Pixels zu sich selbst (vor und nach dem Verschießen eines weiteren Strahls).
Mit diesen drei Bildern ist es möglich, zusätzliche quantitative Informationen zu den qualitativen Informationen des Vergleichs vom neuen und alten Bild zu berücksichtigen.
In dieser Arbeit kläre ich die Frage, ob die gewünschten Effekte eintreten und ob bei Integration in die bestehende Struktur von RenderGin ein Performanzgewinn möglich ist. Die Umsetzung eines adaptiven Samplers ist als Plug-In in der Software RenderGin von Numenus GmbH geschehen. RenderGin ist ein echtzeitfähiger, progressiver Raytracer, der sich durch seine Performanz auszeichnet. Die Bildgenerierung geschieht allein auf der CPU, die Grafikkarte wird lediglich zur Anzeige des erzeugten Bildes benötigt.
Die Umsetzung und Programmierung des Plug-Ins ist in Microsoft Visual Studio 2010 geschehen unter Verwendung des RenderGin SDK der Numenus GmbH.
In der vorliegenden Untersuchung stehen geometrische Aufgaben und die in den seit 2004 national verbindlichen Bildungsstandards im Fach Mathematik für den Primarbereich formulierten Anforderungsbereiche im Zentrum. Diese zeigen die kognitiven Anforderungen an Schülerinnen und Schüler bei der Bearbeitung von Aufgaben auf, wobei zwischen „Reproduzieren", „Zusammenhänge herstellen" und „Verallgemeinern und Reflektieren" unterschieden wird (KMK, 2005a, S. 13).
Durch die drei Anforderungsbereiche sollen Lehrkräfte unter anderem die Chance zur Entwicklung einer anforderungsbezogenen Aufgabenkultur erhalten. Des Weiteren soll die Integration von Aufgaben aus allen drei Anforderungsbereichen im Unterricht angeregt und einem einseitig ausgerichteten Unterricht entgegen gewirkt werden.
Da die Anforderungsbereiche bislang nicht empirisch validiert wurden und in den Veröffentlichungen der Kultusministerkonferenz nicht klar zur Schwierigkeit von Aufgaben abgegrenzt werden (KMK, 2005a, S. 13; KMK, 2005b, S. 17; KMK, 2004b, S. 13), wurde in der vorliegenden Untersuchung zum einen die Möglichkeit der eindeutigen Zuordnung geometrischer Aufgaben zu den drei Anforderungsbereichen geprüft.
Zum anderen wurde untersucht, inwiefern die in den geometrischen Aufgaben enthaltenen kognitiven Anforderungen in Zusammenhang mit der empirischen Schwierigkeit von Aufgaben, der mathematischen Leistungsfähigkeit von Schülerinnen und Schülern, dem Geschlecht und den Anforderungen der im Unterricht gestellten Aufgaben stehen.
Vor dem Hintergrund der dem deutschen Mathematikunterricht nachgesagten Kalkül- beziehungsweise Fertigkeitsorientierung (Baumert et al., 2001, S. 296; Granzer & Walther, 2008, S. 9) und den damit einhergehenden Stärken deutscher Schülerinnen und Schüler im Bereich von Routineaufgaben und Schwächen im Bereich von Aufgaben mit höheren kognitiven Anforderungen (Grassmann et al., 2014, S. 11; Reiss & Hammer, 2013, S. 82; Schütte, 2008, S. 41) wurde zudem die Verteilung der im Rahmen der Untersuchung gewonnenen, schriftlich fixierten geometrischen Schulbuch- und Unterrichtsaufgaben auf die drei Anforderungsbereiche analysiert.
Durch die Betrachtung geometrischer Aufgaben konnte stichprobenartig der quantitative Geometrieanteil in den Schulbüchern und im Unterricht der vierten Jahrgangsstufe ermittelt werden, um so den Forschungsstand zum Stellenwert des Geometrieunterrichts (Maier, 1999; Backe-Neuwald, 2000; Roick, Gölitz & Hasselhorn, 2004) zu aktualisieren beziehungsweise zu ergänzen.
In this work has been examined, how the existing model of the simulation of cables and hoses can be advanced. Therefore an investigation has been made on the main influences to the shape simulation and the factors of constraints and side conditions were analyzed. For the validation of the accuracy, the simulation has to be compared to real specimen behavior. To obtain a very precise digitalization of the shape, the choice was made to use a laser scanner that converts the pointcloud into a .vrml file which can be imported into the simulation environment. The assumption was that the simulation method itself has the highest impact to the simulated shape. This is why the capabilities of the most sophisticated methods have been analyzed. The main criterion for the success of a simulation approach proved not to be accuracy, as expected. Process integration and usability showed to be of higher interest for the efficient exertion. Other factors like the pricing, the functionality and the real-time capability were assayed as well. The analyzed methods are based on the solution of the equations of elasticity with different ways of discetization, finite-elements and a spring-impulse-system. Since the finite-element-system takes several minutes for the computation of the shape and the spring-impulse-system reacts retarded on user manipulation, the competitiveness of these approaches is low. The other methods distinguish more in real-time performance, data interfaces and functionality than in accuracy. For the accuracy of a system, the consideration of other factors proved to be very important. As one of these main factors, the accurate assignment of the material properties was indicated. Until the start of this work, only the finite-element-approach dealt with this factor, but no documentation or validation is provided. In the publications of the other methods, the material properties are estimated to obtain a plausible simulation shape. Therefore the specific material values of real specimen have been measured and assigned to the simulation. With the comparison to the real shape it has been proven that the accuracy is very high with the measured properties. Since these measurements are very costly and time consuming, an investigation on a faster and cheaper way to obtain these values has been made. It has been assumed that with the knowledge of the cross-section it should be possible to compute the specimen behavior. Since the braid distribution changes individually from specimen to specimen, a more general way to obtain the values needed to be found. The program composer has been developed, where only the number of the different braids and the taping is entered. It computes with very high precision the stiffness, the density and the final diameter of the bundle. With the measured values and the fitting to the real shape it has been proven that the simulation approach reflects the precise behavior of cables and hoses. Since the stiffness of the single braids is wasteful to measure, a measurement setup was created where the stiffness has a large impact to the shape. With known density, the stiffness of the specimen can be reconstructed precisely. Thus a fast and beneficial way of obtaining the stiffness of a cable has been invented. The poissons ratio of cables and bundles cannot be measured with a tensile test, since the inner structure is very complex. For hoses, the variation of the inner diameter has been measured during the tensile test as well. The resulting values were reasonable, but their accuracy could not be proven. For cables and hoses, it has been tried to obtain the poissons ratio via the computation of the cross section, but the influence of individual changes and the crosstalk of the braids is very high. Therefore a setup was constructed where the torsion stiffness can be measured. For cables and hoses, the individual cross-sections and taping lead to varying results. For hoses, expected and repeatable good values for the poissons ratio were obtained. The low influence of the poisons ratio in the range between 0 and 0.5 has been proven. Therefore we decided to follow the advice of [Old06] and our own experiences to set the poisons ratio for cables and bundles to 0.25. With the knowledge of the measurability and the capabilities of the developed program composer, a procedure to obtain material properties for bundles has been designed. 1. Measurement of the braid density with via pyknometer or mass, length and diameter. 2. Empirical reconstruction of the stiffness with the designed setup. 3. Composing the bundle with the program composer. 4. Adding a factor for the taping and transfer the values to the simulation. The model of the cable simulation has been improved as follows: The main influences in the simulation of cables and hoses are the simulation method, the material properties and the geometric constraints. To obtain higher accuracy, an investigation on the correct material properties is indispensable. The scientific determination of material properties for the simulation of cables, bundles and hoses has been performed for the first time. The influence of geometrical constraints has been analyzed and documented. The next steps are the analysis of pre-deformation and further investigations to the determination of the poisons ratio with a more precise torsion test. All analysis were made with the simulation approach fleXengine. A comparison to other simulation methods would be of high interest.
Agricultural land-use may lead to brief pulse exposures of pesticides in edge-of-field streams, potentially resulting in adverse effects on aquatic macrophytes, invertebrates and ecosystem functions. The higher tier risk assessment is mainly based on pond mesocosms which are not designed to mimic stream-typical conditions. Relatively little is known on exposure and effect assessment using stream mesocosms.
Thus the present thesis evaluates the appliacability of the stream mesocosms to mimic stream-typical pulse exposures, to assess resulting effects on flora and fauna and to evaluate aquatic-terrestrial food web coupling. The first objective was to mimic stream-typical pulse exposure scenarios with different durations (≤ 1 to ≥ 24 hours). These exposure scenarios established using a fluorescence tracer were the methodological basis for the effect assessment of an herbicide and an insecticide. In order to evaluate the applicability of stream mesocosms for regulatory purposes, the second objective was to assess effects on two aquatic macrophytes following a 24-h pulse exposure with the herbicide iofensulfuron-sodium (1, 3, 10 and 30 µg/L; n = 3). Growth inhibition of up to 66 and 45% was observed for the total shoot length of Myriophyllum spicatum and Elodea canadensis, respectively. Recovery of this endpoint could be demonstrated within 42 days for both macrophytes. The third objective was to assess effects on structural and functional endpoints following a 6-h pulse exposure of the pyrethroid ether etofenprox (0.05, 0.5 and 5 µg/L; n = 4). The most sensitive structural (abundance of Cloeon simile) and functional (feeding rates of Asellus aquaticus) endpoint revealed significant effects at 0.05 µg/L etofenprox. This concentration was below field-measured etofenprox concentrations and thus suggests that pulse exposures adversely affect invertebrate populations and ecosystem functions in streams. Such pollutions of streams may also result in decreased emergence of aquatic insects and potentially lead to an insect-mediated transfer of pollutants to adjacent food webs. Test systems capable to assess aquatic-terrestrial effects are not yet integrated in mesocosm approaches but might be of interest for substances with bioaccumulation potential. Here, the fourth part provides an aquatic-terrestrial model ecosystem capable to assess cross-ecosystem effects. Information on the riparian food web such as the contribution of aquatic (up to 71%) and terrestrial (up to 29%) insect prey to the diet of the riparian spider Tetragnatha extensa was assessed via stable isotope ratios (δ13C and δ15N). Thus, the present thesis provides the methodological basis to assess aquatic-terrestrial pollutant transfer and effects on the riparian food web.
Overall the results of this thesis indicate, that stream mesocosms can be used to mimic stream-typical pulse exposures of pesticides, to assess resulting effects on macrophytes and invertebrates within prospective environmental risk assessment (ERA) and to evaluate changes in riparian food webs.
Since software influences nearly every aspect of everyday life, the security of software systems is more important than ever before. The evaluation of the security of a software system still poses a significant challenge in practice, mostly due to the lack of metrics, which can map the security properties of source code onto numeric values. It is a common assumption, that the occurrence of security vulnerabilities and the quality of the software design stand in direct correlation, but there is currently no clear evidence to support this. A proof of an existing correlation could help to optimize the measurements of program security, making it possible to apply quality measurements to evaluate it. For this purpose, this work evaluates fifty open-source android applications, using three security and seven quality metrics. It also considers the correlations between the metrics. The quality metrics range from simple code metrics to high-level metrics such as object-oriented anti-patterns, which together provide a comprehensive picture of the quality. Two visibility metrics, along with a metric that computes the minimal permission request for mobile applications, were selected to illustrate the security. Using the evaluation projects, it was found that there is a clear correlation between most quality metrics. By contrast, no significant correlations were found using the security metrics. This work discusses the correlations and their causes as well as further recommendations based on the findings.
The work presented in this thesis investigated interactions of selected biophysical processes that affect zooplankton ecology at smaller scales. In this endeavour, the extent of changes in swimming behaviour and fluid disturbances produced by swimming Daphnia in response to changing physical environments were quantified. In the first research question addressed within this context, size and energetics of hydrodynamic trails produced by Daphnia swimming in non-stratified still waters were characterized and quantified as a function of organisms’ size and their swimming patterns.
The results revealed that neither size nor the swimming pattern of Daphnia affects the width of induced trails or dissipation rates. Nevertheless, as the size and swimming velocity of the organisms increased, trail volume increased in proportional to the cubic power of Reynolds number, and the biggest trail volume was about 500 times the body volume of the largest daphnids. Larger spatial extent of fluid perturbation and prolonged period to decay caused by bigger trail volumes would play a significant role in zooplankton ecology, e.g. increasing the risk of predation.
The study also found that increased trail volume brought about significantly enhanced total dissipated power at higher Reynolds number, and the magnitudes of total dissipated power observed varied in the range of (1.3-10)X10-9 W.
Furthermore, this study provided strong evidence that swimming speed of Daphnia and total dissipated power in Daphnia trails exceeded those of some other selected zooplankton species.
In recognizing turbulence as an intrinsic environmental perturbation in aquatic habitats, this thesis also examined the response of Daphnia to a range of turbulence flows, which correspond to turbu-lence levels that zooplankton generally encounter in their habitats. Results indicated that within the range of turbulent intensities to which the Daphnia are likely to be exposed in their natural habitats, increasing turbulence compelled the organisms to enhance their swimming activity and swim-ming speed. However, as the turbulence increased to extremely high values (10-4 m2s-3), Daphnia began to withdraw from their active swimming behaviour. Findings of this work also demonstrated that the threshold level of turbulence at which animals start to alleviate from largely active swimming is about 10-6 m2s-3. The study further illustrated that during the intermediate range of turbu-lence; 10-7 - 10-6 m2s-3, kinetic energy dissipation rates in the vicinity of the organisms is consistently one order of magnitude higher than that of the background turbulent flow.
Swarming, a common conspicuous behavioural trait observed in many zooplankton species, is considered to play a significant role in defining freshwater ecology of their habitats from food exploitation, mate encountering to avoiding predators through hydrodynamic flow structures produced by them, therefore, this thesis also investigated implications of Daphnia swarms at varied abundance & swarm densities on their swimming kinematics and induced flow field.
The results showed that Daphnia aggregated in swarms with swarm densities of (1.1-2.3)x103 L-1, which exceeded the abundance densities by two orders of magnitude (i.e. 1.7 - 6.7 L-1). The estimated swarm volume decreased from 52 cm3 to 6.5 cm3, and the mean neighbouring distance dropped from 9.9 to 6.4 body lengths. The findings of this work also showed that mean swimming trajectories were primarily horizontal concentric circles around the light source. Mean flow speeds found to be one order of magnitude lower than the corresponding swimming speeds of Daphnia. Furthermore, this study provided evidences that the flow fields produced by swarming Daphnia differed considerably between unidirectional vortex swarming and bidirectional swimming at low and high abundances respectively.
Increasingly, problematic smartphone use behavior (PSU) and excessive consumption are reported. In this study, an experiment was developed to investigate the influence of screen coloration using the grayscale setting on smartphone usage time in repeated measurements. We also investigated how individuals perceived suffering correlates with smartphone usage time and PSU, and whether differences exist by smartphone usage type (social, process, habitual). 240 subjects completed a questionnaire about smartphone usage time, PSU, perceived suffering, and smartphone usage types. Afterward, their smartphones were switched to grayscale setting for at least 24h, and thereafter 92 of these participants completed the second questionnaire. Analyses showed that grayscale setting decreases usage time and that there is a positive correlation between PSU, smartphone usage duration, and perceived suffering. The types of use (process and habitual) influence one’s perceived suffering. Thus, it shows that individuals are aware of their PSU and suffer from it. Using grayscale setting is effective in reducing smartphone use time.
Performanz von RIP-MTIfi
(2009)
Diese Diplomarbeit beschäftigt sich mit der Performanz von RIP-MTI, insbesondere mit der Performanz der Schleifenerkennung. Ziel der Arbeit ist es, die Zeitdauer der Schleifenerkennung von RIP-MTI zu untersuchen und Probleme, welche bei der Erkennung von Schleifen auftreten könen, aufzudecken und zu lösen.
The maintenance strategy “predictive maintenance”, which is characterized by predicting the failure behavior of technical units based on modern sensor technology, plays a key role in smart factories against the background of an industry 4.0. This paper contains an evaluation of the current state of research on this strategy and gives an overview of the areas of application to date. With the aid of a qualitative video analysis, the implementation in the industries and company divisions involved and the type of goods monitored are examined. The analyzed video clips were uploaded to YouTube for example for marketing purposes by various companies with different perspectives on predictive maintenance. The video analysis was realized by applying a previously defined coding plan to the video material. The results show a predominant application in the manufacturing industry, in which predictive maintenance is used to monitor plants and machines. In addition, the strategy is also mainly applied to means of transport used for freight and passenger transport in various infrastructures. As a result of the video analysis, the currently high need for explanation of predictive maintenance becomes visible. By looking at these explanations, one also learns something about the special features that distinguish it from other maintenance strategies.
More than 10,000 organic chemicals such as pharmaceuticals, ingredients of personal care products and biocides are ubiquitously used in every day life. After their application, many of these chemicals enter the domestic sewer. Research has shown that conventional biological wastewater treatment in municipal wastewater treatment plants (WWTPs) is an insufficient barrier for the release of most of these anthropogenic chemicals into the receiving waters.
This bears unforeseen risks for aquatic wildlife and drinking water resources. Especially for recently introduced and/or detected compounds (so called emerging micropollutants), there is a growing need to investigate the occurrence and fate in WWTPs. In order to get a comprehensive picture on the behavior in municipal wastewater treatment, the following groups of emerging organic micropollutants, spanning a broad range of applications and physico-chemical properties, were selected as target compounds: pharmaceuticals (beta blockers, psycho-active drugs), UV-filters, vulcanization accelerators (benzothiazoles), biocides (anti-dandruffs, preservatives, disinfectants) and pesticides (phenylurea and triazine herbicides).
Forschungsergebnisse zum Männerchorwesen Deutschlands im 19. Jahrhundert belegen dessen gesellschaftliche und politische Relevanz. Das so genannte Sängerwesen leistete demnach einen wesentlichen Beitrag zur Nationsbildung in Deutschland, da die Sänger durch ihren Gesang sowie durch ihre Aktivitäten im Verein und in der Öffentlichkeit zur inneren Einigung der Bevölkerung beitrugen und somit halfen, eine einheitliche Nation zu formen. Im Gegensatz dazu gab es bislang kaum Erkenntnisse über die gesellschaftlichen und politischen Hintergründe des Männerchorwesens der Pfalz im gleichen Zeitraum. Um diese Lücke zu schließen, wurde mit der vorliegenden Arbeit die Geschichte des Männerchorwesens der Pfalz erforscht, insbesondere hinsichtlich seiner Bedeutung für die Nationsbildung Deutschlands. Der Untersuchungszeitraum erstreckt sich von 1816, dem Jahr, in dem die Pfalz zum bayerischen Staatsgebiet wurde, bis zur Gründung des Deutschen Reiches im Jahre 1871.
Zunächst wird die Entwicklung des pfälzischen Sängerwesens bezüglich der Zahl der gegründeten Vereinen in den einzelnen Jahren und Orten sowie bezüglich des Feierns lokaler und regionaler Sängerfeste im Überblick und im Vergleich zum Musikvereinswesen der Pfalz im gleichen Zeitraum dargestellt. Dieser Betrachtung des pfälzischen Männerchorwesens als Ganzem folgt die Untersuchung seiner Einzelteile, der Personen und Ereignisse innerhalb einzelner Sängervereine und innerhalb bestimmter Zeitabschnitte, vor dem Hintergrund der jeweiligen politischen und gesellschaftlichen Situation. Bedeutend sind in diesem Zusammenhang vor allem die Auswirkungen der politisch-gesellschaftlichen Großereignisse Hambacher Fest von 1832 sowie Revolution von 1848/49 auf die laienmusikalischen Vereinskulturen.
Schließlich werden die einzelnen Phänomene und die Gesamtentwicklung aufeineander bezogen. Der Anhang der Forschungsarbeit beinhaltet, neben Auszügen aus den Protokollbüchern des "Cäcilienverein-Liedertafel Dürkheim" sowie Plakaten von pfälzischen Musik- und Sängerfesten der 1840er Jahre, Übersichtstabellen mit Informationen zu den Pfälzischen Musikfesten des 19. Jahrhunderts sowie zu den im Untersuchungszeitraum gegründeten Sängervereinen, außerdem Kartenmaterial zur räumlichen Verbreitung der Vereine sowie Notenmaterial zum freimaurerischen "Weihelied" des Kaiserslauterer Seminarlehrers Philipp Walter.
Das Organische Qualitätsmanagement (OQM) hat seine Wurzeln in der "Natürlichen Gemeindeentwicklung" (NGE). Der evangelische Theologe Christian A. Schwarz und der Diplompsychologe Christoph Schalk gingen im Rahmen eines internationalen Forschungsprojektes der Frage nach, ob es universell gültige Prinzipien für das qualitative und quantitative Wachstum von Kirchengemeinden gibt. Diese Studie wurde zu einem der größten Forschungsprojekte, das jemals über das Wachstum von Gemeinden durchgeführt wurde. Bis Februar 2011 wurden in über 70 Ländern 71.512 Profile für Kirchengemeinden unterschiedlichster Prägung und Konfession erstellt.
Das Ergebnis dieser fortschreitenden Studie bietet eine wissenschaftlich zu verifizierende Antwort auf die Frage: "Was sind die Wachstumsprinzipien, die unabhängig von Kultur, theologischer Prägung und Frömmigkeitsrichtung gelten und können sich diese Prinzipien auch positiv auf Non Profit Organisationen und säkulare Wirtschaftsunternehmen anwenden lassen? Kapitel 1 beschreibt zunächst die Entstehung, Herkunft und Entwicklung des OQM, d.h. den Übertragungsprozess aller Erkenntnisse aus der Entwicklung von Kirchengemeinden auf die Realität von kirchlichen Organisationen und später auf die Anforderungen eines Wirtschaftsunternehmens. Kapitel 2 geht ausführlich auf die Beschreibung der acht Qualitätsmerkmale ein. Hierbei stehen weniger die Substantive (Leitung, Mitarbeiter, Strukturen, Beziehungen etc.) im Mittelpunkt, als vielmehr die Adjektive (bevollmächtigend, gabenorientiert, zweckmäßig, vertrauensvoll etc.). Sie beschreiben, auf was es in Veränderungsprozessen wirklich ankommt.
Zu Beginn des Projektes war nicht abzusehen, ob und in welcher Weise sich die Qualitätsmerkmale und die christlichen Sozialprinzipien als Erfolgsfaktoren auch in anderen Unternehmensformen beweisen können. In Kapitel 3 werden Praxisbeispiele zeigen, dass ein dialogisch arbeitendes OQM nicht nur Kirchengemeinden sondern auch caritativen Unternehmen und sogar Wirtschaftsunternehmen helfen kann zu wachsen und somit erfolgreich zu sein.
Generalized methods for automated theorem proving can be used to compute formula transformations such as projection elimination and knowledge compilation. We present a framework based on clausal tableaux suited for such tasks. These tableaux are characterized independently of particular construction methods, but important features of empirically successful methods are taken into account, especially dependency directed backjumping and branch local operation. As an instance of that framework an adaption of DPLL is described. We show that knowledge compilation methods can be essentially improved by weaving projection elimination partially into the compilation phase.
Clubs, such as Scouts, rely on the work of their volunteer members, who have a variety of tasks to accomplish. Often there are sudden changes in their organization teams and offices, whereby planning steps are lost and inexperience in planning occurs. Since the special requirements are not covered by already existing tools, ScOuT, a planning tool for the organization administration, is designed and developed in this work to support clubs with regard to the mentioned problems. The focus was on identifying and using various suitable guidelines and heuristic methods to create a usable interface. The developed product was evaluated empirically by a user survey in terms of usability.
The result of this study shows that already a high degree of the desired goal could be reached by the inclusion of the guidelines and methods. From this it can be concluded that with the help of user-specific concept ideas and the application of suitable guidelines and methods, a suitable basis for a usable application to support clubs can be created.