Refine
Year of publication
- 2020 (69) (remove)
Document Type
- Master's Thesis (29)
- Article (21)
- Conference Proceeding (11)
- Part of a Book (4)
- Doctoral Thesis (2)
- Other (1)
- Report (1)
Institute
- Forschungszentrum Mikrotechnik (15)
- Forschungszentrum Business Informatics (10)
- Technik | Engineering & Technology (6)
- Department of Computer Science (Ende 2021 aufgelöst; Integration in die übergeordnete OE Technik) (5)
- Soziales & Gesundheit (5)
- Forschungszentrum Energie (4)
- Wirtschaft (2)
- Department of Engineering (Ende 2021 aufgelöst; Integration in die übergeordnete OE Technik) (1)
- Forschung (1)
- Forschungszentrum Digital Factory Vorarlberg (1)
Language
- English (69) (remove)
Keywords
- Optical coherence tomography (3)
- Arrayed waveguide gratings (2)
- Human needs (2)
- International business (2)
- OCT (2)
- ROS (2)
- SME (2)
- Y-branch splitter (2)
- A priori & a posteriori segmentation (1)
- AI on the Edge (1)
This thesis aims to support the product development process. Therefore, an approach is developed, implemented as a prototype and evaluated, for automated solution space exploration of formally predefined design automation tasks holding the product knowledge of engineers. For this reason, a classification of product development tasks related to the representation of the mathematical model is evaluated based on the parameters defined in this thesis. In a second step, the mathematical model should be solved. A Solver is identified able to handle the given problem class.
Due to the context of this work, System Modelling Language (SysML) is chosen for the product knowledge formalisation. In the next step the given SysML model has to be translated into an object-oriented model. This translation is implemented by extracting information of a ".xml"-file using the XML Metadata Interchanging (XMI) standard. The information contained in the file is structured using the Unified Modelling Language (UML) profile for SysML. Afterwards a mathematical model in MiniZinc language is generated. MiniZinc is a mathematical modelling language interpretable by many different Solvers. The generated mathematical model is classified related to the Variable Type and Linearity of the Constraints and Objective of the generated mathematical model. The output is stored in a ".txt"-file.
To evaluate the functionality of the prototype, time consumption of the different performed procedures is measured. This data shows that models containing Continuous Variables need a longer time to be classified and optimised. Another observation shows that the transformation into an object-oriented model and the translation of this model into a mathematical representation are dependent on the number of SysML model elements. Using MiniZinc resulted in the restriction that models which use non-linear functions and Boolean Expressions cannot be solved. This is because the implementation of non-linear Solvers at MiniZinc is still in the development phase. An investigation of the optimally of the results, provided by the Solvers, was left for further work.
The Digital Factory Vorarlberg is the youngest Research Center of Vorarlberg University of Applied Sciences. In the lab of the research center a research and learning factory has been established for educating students and employees of industrial partners. Showcases and best practice scenarios for various topics of digitalization in the manufacturing industry are demonstrated. In addition, novel methods and technologies for digital production, cloud-based manufacturing, data analytics, IT- and OT-security or digital twins are being developed. The factory comprises only a minimum core of logistics and fabrication processes to guarantee manageability within an academic setup. As a product, fidget spinners are being fabricated. A webshop allows customers to individually design their products and directly place orders in the factory. A centralized SCADA-System is the core data hub for the factory. Various data analytic tools and methods and a novel database for IoT-applications are connected to the SCADA-System. As an alternative to on premise manufacturing, orders can be pushed into a cloud-based manufacturing platform, which has been developed at the Digital Factory. A broker system allows fabrication in distributed facilities and offers various optimization services. Concepts, such as outsourcing product configuration to customers or new types of engineering services in cloud-based manufacturing can be explored and demonstrated. In this paper, we present the basic concept of the Digital Factory Vorarlberg, as well as some of the newly developed topics.
A modified matrix adaptation evolution strategy with restarts for constrained real-world problems
(2020)
In combination with successful constraint handling techniques, a Matrix Adaptation Evolution Strategy (MA-ES) variant (the εMAg-ES) turned out to be a competitive algorithm on the constrained optimization problems proposed for the CEC 2018 competition on constrained single objective real-parameter optimization. A subsequent analysis points to additional potential in terms of robustness and solution quality. The consideration of a restart scheme and adjustments in the constraint handling techniques put this into effect and simplify the configuration. The resulting BP-εMAg-ES algorithm is applied to the constrained problems proposed for the IEEE CEC 2020 competition on Real-World Single-Objective Constrained optimization. The novel MA-ES variant realizes improvements over the original εMAg-ES in terms of feasibility and effectiveness on many of the real-world benchmarks. The BP-εMAg-ES realizes a feasibility rate of 100% on 44 out of 57 real-world problems and improves the best-known solution in 5 cases.
A novel calorimetric technique for the analysis of gas-releasing endothermic dissociation reactions
(2020)
Issues with professional conduct and discrimination against Lesbian, Gay, Bisexual, Transgender (LGBT+) people in health and social care, continue to exist in most EU countries and worldwide.
The project IENE9 titled: “Developing a culturally competent and compassionate LGBT+ curriculum in health and social care education” aims to enable teacher/trainers of theory and practice to enhance their skills regarding LGBT+ issues and develop teaching tools to support the inclusion of LGBT+ issues within health and social care curricula. The newly culturally competent and compassionate LGBT+ curriculum will be delivered though a Massive Open Online Course (MOOC) which is aimed at health and social care workers, professionals and learners across Europe and worldwide.
We have identified educational policies and guidelines at institutions teaching in health and social care, taken into account for developing the learning/teaching resources. The MOOC will be an innovative training model based on the Papadopoulos (2014) model for “Culturally Competent Compassion”. The module provides a logical and easy to follow structure based on its four constructs 'Culturally Aware and Compassionate Learning', 'Culturally Knowledgeable and Compassionate Learning', 'Culturally Sensitive and Compassionate Learning', 'Culturally Competent and Compassionate Learning'.
Specific training may result in better knowledge and skills of the health and social care workforce, which helps to reduce inequalities and communication with LGBT+ people, as well as diminishing the feelings of stigma or discrimination experienced.
An electrochemical study with three redox substances on a carbon based nanogap electrode array
(2020)
An implementation approach of the gap navigation tree using the TurtleBot 3 Burger and ROS Kinetic
(2020)
The creation of a spatial model of the environment is an important task to allow the planning of routes through the environment. Depending on the number of sensor inputs different ways of creating a spatial environment model are possible. This thesis introduces an implementation approach of the Gap Navigation Tree which is aimed for usage with robots that have a limited amount of sensors. The Gap Navigation Tree is a tree structure based on depth discontinuities constructed from the data of a laser scanner. Using the simulated TurtleBot 3 Burger and ROS kinetic a framework is created that implements the theory of the Gap Navigation Tree. The framework is structured in a way that allows using different robots with different sensor types by separating the detection of depth discontinuities from the building and updating of the Gap Navigation Tree.
Zeros can cause many issues in data analysis and dealing with them requires specialized procedures. We differentiate between rounded zeros, structural zeros and missing values. Rounded zeros occur when the true value of a variable is hidden because of a detection limit in whatever mechanism was used to acquire the data. Structural zeros are values which are truly zero, often coming about due to a hidden mechanism separate from the one which generates values greater than 0. Missing values are values that are completely missing for unknown or known reasons. This thesis outlines various methods for dealing with different kinds of zeros in different contexts. Many of these methods are very specific in their ideal usecase. They are separated based on which kind of zero they are intended for and if they are better suited for compositional or for standard data.
For rounded zeros we impute the zeros with an estimated value below the detection limit. The author describes multiplicative replacement, a simple procedure that imputes values at a fixed fraction of the detection limit. As a more advanced technique, the author describes Kaplan Meier smoothing spline replacement, which interpolates a spline on a Kaplan Meier curve and uses the spline below the detection limit to impute values in a more natural distribution. Rounded zeros cannot be imputed with the same techniques that would be used for regular missing values, since there is more information available on the true value of a rounded zero than there would be for a regular missing value.
Structural zeros cannot be imputed since they are a true zero. Imputing them would falsify their values and produce a value where there should be none. Because of this, we apply modelling techniques that can work around structural zeros and incorporate them. For standard data, the zero inflated Poisson model is presented. This model utilizes a mixture of a logistic and a Poisson distribution to accurately model data with a large amount of structural zeros. While the Poisson distribution is only applicable to count data, the zero inflation concept can be applied to different kinds of distributions. For compositional data, the zero adjusted Dirichlet model is introduced. This model mixes Dirichlet distributions for every pattern of zeros found within the data. Non-algorithmic techniques to reduce the amount of structural zeros present are also shown. These techniques being amalgamation, which combines columns with structural zeros into more broad descriptors and classification, which changes columns into categorical values based on a structural zero being present or not.
Missing values are values that are completely missing for various known or unknown reasons. Different imputation techniques are introduced. For standard data, MissForest imputation is introduced, which utilizes a RandomForest regression to impute mixed type missing values. Another imputation technique shown utilizes both a genetic algorithm and a neural network to impute values based on the genetic algorithm minimizing the error of an autoencoder neural network. In the case of compositional data, knn imputation is presented, which utilizes the knn concept also found in knn clustering to impute the values based on the closest samples with a value available.
All of these methods are explained and demonstrated to give readers a guide to finding the suitable methods to use in different scenarios.
The thesis also provides a general guide on dealing with zeros in data, with decision flowcharts and more detailed descriptions for both compositional and standard data being presented. General tips on getting better results when zeros are involved are also given and explained. This general guide was then applied to a dataset to show it in action.