Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Principal component analysis (PCA) is a popular dimension-reduction method to reduce the complexity and obtain the informative aspects of high-dimensional datasets. When the data distribution is skewed, data transformation is commonly used prior to applying PCA. Such transformation is usually obtained from previous studies, prior knowledge, or trial-and-error. In this work, we develop a model-based method that integrates data transformation in PCA and finds an appropriate data transformation using the maximum profile likelihood. Extensions of the method to handle functional data and missing values are also developed. Several numerical algorithms are provided for efficient computation. The proposed method is illustrated using simulated and real-world data examples. Supplementary materials for this article are available online.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
To achieve true data interoperability is to eliminate format and data model barriers, allowing you to seamlessly access, convert, and model any data, independent of format. The ArcGIS Data Interoperability extension is based on the powerful data transformation capabilities of the Feature Manipulation Engine (FME), giving you the data you want, when and where you want it.In this course, you will learn how to leverage the ArcGIS Data Interoperability extension within ArcCatalog and ArcMap, enabling you to directly read, translate, and transform spatial data according to your independent needs. In addition to components that allow you to work openly with a multitude of formats, the extension also provides a complex data model solution with a level of control that would otherwise require custom software.After completing this course, you will be able to:Recognize when you need to use the Data Interoperability tool to view or edit your data.Choose and apply the correct method of reading data with the Data Interoperability tool in ArcCatalog and ArcMap.Choose the correct Data Interoperability tool and be able to use it to convert your data between formats.Edit a data model, or schema, using the Spatial ETL tool.Perform any desired transformations on your data's attributes and geometry using the Spatial ETL tool.Verify your data transformations before, after, and during a translation by inspecting your data.Apply best practices when creating a workflow using the Data Interoperability extension.
The INTEGRATE (Inverse Network Transformations for Efficient Generation of Robust Airfoil and Turbine Enhancements) project is developing a new inverse-design capability for the aerodynamic design of wind turbine rotors using invertible neural networks. This AI-based design technology can capture complex non-linear aerodynamic effects while being 100 times faster than design approaches based on computational fluid dynamics. This project enables innovation in wind turbine design by accelerating time to market through higher-accuracy early design iterations to reduce the levelized cost of energy. INVERTIBLE NEURAL NETWORKS Researchers are leveraging a specialized invertible neural network (INN) architecture along with the novel dimension-reduction methods and airfoil/blade shape representations developed by collaborators at the National Institute of Standards and Technology (NIST) learns complex relationships between airfoil or blade shapes and their associated aerodynamic and structural properties. This INN architecture will accelerate designs by providing a cost-effective alternative to current industrial aerodynamic design processes, including: Blade element momentum (BEM) theory models: limited effectiveness for design of offshore rotors with large, flexible blades where nonlinear aerodynamic effects dominate Direct design using computational fluid dynamics (CFD): cost-prohibitive Inverse-design models based on deep neural networks (DNNs): attractive alternative to CFD for 2D design problems, but quickly overwhelmed by the increased number of design variables in 3D problems AUTOMATED COMPUTATIONAL FLUID DYNAMICS FOR TRAINING DATA GENERATION - MERCURY FRAMEWORK The INN is trained on data obtained using the University of Marylands (UMD) Mercury Framework, which has with robust automated mesh generation capabilities and advanced turbulence and transition models validated for wind energy applications. Mercury is a multi-mesh paradigm, heterogeneous CPU-GPU framework. The framework incorporates three flow solvers at UMD, 1) OverTURNS, a structured solver on CPUs, 2) HAMSTR, a line based unstructured solver on CPUs, and 3) GARFIELD, a structured solver on GPUs. The framework is based on Python, that is often used to wrap C or Fortran codes for interoperability with other solvers. Communication between multiple solvers is accomplished with a Topology Independent Overset Grid Assembler (TIOGA). NOVEL AIRFOIL SHAPE REPRESENTATIONS USING GRASSMAN SPACES We developed a novel representation of shapes which decouples affine-style deformations from a rich set of data-driven deformations over a submanifold of the Grassmannian. The Grassmannian representation as an analytic generative model, informed by a database of physically relevant airfoils, offers (i) a rich set of novel 2D airfoil deformations not previously captured in the data , (ii) improved low-dimensional parameter domain for inferential statistics informing design/manufacturing, and (iii) consistent 3D blade representation and perturbation over a sequence of nominal shapes. TECHNOLOGY TRANSFER DEMONSTRATION - COUPLING WITH NREL WISDEM Researchers have integrated the inverse-design tool for 2D airfoils (INN-Airfoil) into WISDEM (Wind Plant Integrated Systems Design and Engineering Model), a multidisciplinary design and optimization framework for assessing the cost of energy, as part of tech-transfer demonstration. The integration of INN-Airfoil into WISDEM allows for the design of airfoils along with the blades that meet the dynamic design constraints on cost of energy, annual energy production, and the capital costs. Through preliminary studies, researchers have shown that the coupled INN-Airfoil + WISDEM approach reduces the cost of energy by around 1% compared to the conventional design approach. This page will serve as a place to easily access all the publications from this work and the repositories for the software developed and released through this project.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Cardiovascular diseases continue to be a significant global health threat. The electrocardiogram (ECG) signal is a physiological signal that plays a major role in preventing severe and even fatal heart diseases. The purpose of this research is to explore a simple mathematical feature transformation that could be applied to ECG signal segments in order to improve the detection accuracy of heartbeats, which could facilitate automated heart disease diagnosis. Six different mathematical transformation methods were examined and analyzed using 10s-length ECG segments, which showed that a reciprocal transformation results in consistently better classification performance for normal vs. atrial fibrillation beats and normal vs. atrial premature beats, when compared to untransformed features. The second best data transformation in terms of heartbeat detection accuracy was the cubic transformation. Results showed that applying the logarithmic transformation, which is considered the go-to data transformation, was not optimal among the six data transformations. Using the optimal data transformation, the reciprocal, can lead to a 35.6% accuracy improvement. According to the overall comparison tested by different feature engineering methods, classifiers, and different dataset sizes, performance improvement also reached 4.7%. Therefore, adding a simple data transformation step, such as the reciprocal or cubic, to the extracted features can improve current automated heartbeat classification in a timely manner.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
List of data transformation methods.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Dataset and roof-of-concept method used for proceeding titled "Data Transformation for Clustering Utilization
for Feature Detection in MS" of IWBBIO 22 conference.
MIT Licensehttps://opensource.org/licenses/MIT
License information was derived automatically
A recently discovered universal rank-based matrix method to extract trends from noisy time series is described in Ierley and Kostinski (2019) but the formula for the output matrix elements, implemented there as an open-access supplement MATLAB computer code, is O(N^4), with N the matrix dimension. This can become prohibitively large for time series with hundreds of sample points or more. Based on recurrence relations, here we derive a much faster O(N^2) algorithm and provide code implementations in MATLAB and in open-source JULIA. In some cases one has the output matrix and needs to solve an inverse problem to obtain the input matrix. A fast algorithm and code for this companion problem, also based on the recurrence relations, are given. Finally, in the narrower, but common, domains of (i) trend detection and (ii) parameter estimation of a linear trend, users require, not the individual matrix elements, but simply their accumulated mean value. For this latter case we provide a yet faster O(N) heuristic approximation that relies on a series of rank one matrices. These algorithms are illustrated on a time series of high energy cosmic rays with N > 4 x 10^4 .
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Data transformation methods, hyperparameter optimization and feature selection used in prior studies.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Description of the data transformation methods for compositional data and forecasting models.
https://www.elsevier.com/about/policies/open-access-licenses/elsevier-user-license/cpc-license/https://www.elsevier.com/about/policies/open-access-licenses/elsevier-user-license/cpc-license/
Title of program: TRANSFORM Catalogue Id: AADT_v1_0
Nature of problem The natural coupling scheme for multi-configurational Dirac-Fock calculations is the jj coupling scheme. For comparison of results of such calculations with experiment, other coupling schemes are often more useful. The program produces the transformation between a jj coupling scheme and an alternative scheme.
Versions of this program held in the CPC repository in Mendeley Data AADT_v1_0; TRANSFORM; 10.1016/0010-4655(86)90169-4
This program has been imported from the CPC Program Library held at Queen's University Belfast (1969-2019)
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
The dataset used for evaluating the hybrid feature location technique presented in the paper: "A Hybrid Feature Location Technique for Re-engineering Single Systems into Software Product Lines". This enables reproducibility, evaluation, and comparison of our study.
Folder "Dataset" contains for each subject system used:
(i) the artificial variants and their configurations;
(ii) the ECCO repository containing the traces;
(iii) the ground truth and composed variants;
(iv) the metrics results.
Folder "Scenarios" contains for each subject system used:
(i) the videos recorded from exercising features on GUI.
Apache License, v2.0https://www.apache.org/licenses/LICENSE-2.0
License information was derived automatically
This dataset is related to the problem of geometric dimension reduction, from 2D profile to 1D midcurve.
The format of original/raw dataset is very simplistic and points based. But to make it more appropriate, this dataset makes it line based.
Line based Shape Format is [((x1,y1,..),(x2,y2,..)), ..]
For example: L-profile raw data looks like: 5.0 5.0 10.0 5.0 10.0 30.0 35.0 30.0 35.0 35.0 5.0 35.0
Output Profile Format: [((5.0,5.0), (10.0,5.0)), ((10.0,5.0), (10.0,30.0)), ((10.0,30.0), (35.0,30.0)), ((35.0,30.0), (35.0, 35.0)), ((35.0, 35.0), (5.0,35.0)), ((5.0,35.0), (5.0,5.0))]
You can see that list line has been added to close the polygon.
For L-midcurve raw-data looks like: 7.5 5.0 7.5 32.5 35.0 32.5 7.5 32.5
Output Midcurve Format: [((7.5,5.0), (7.5, 32.5)), ((7.5, 32.5), (35.0, 32.5)), ((35.0, 32.5) (7.5, 32.5))]
No closure is done. Care has been taken for shapes like 'T' and 'X' where midcurve poly-lines are not sequential.
Once base geometries are generated then, various transformations are used to populate augmented dataset.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
The folder contains two subfolders: one with EyeLink 1000 Plus eye-tracking data, and the other with Tobii Nano Pro data. Each of these folders includes a file named "Gaze_position_raw", which is an Excel file containing all the raw data collected from participants. In this file, different trials are stored in separate sheets, and each sheet contains columns for the displayed target location and the corresponding gaze location.A separate folder called "Processed_data_EyeLink/Tobii" contains the results after performing a linear transformation. In this folder, each trial is saved as a separate Excel file. These files include the target location, the original gaze location, and the corrected gaze location after the corresponding linear transformation.
CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
License information was derived automatically
Abstract In this paper, the author presents fundamentally new methods for transforming the value of commodities into original and equilibrium prices of production using Karl Marx's five-sector tables from the third volume of Capital. Mathematical verification of methods is performed using sequential iterations and the program Wolfram Mathematics. For the first time, the method of inverse transformation of production prices of commodities into value prices is presented. It is proved that the pricing systems based on the principles of value and price of production are not mutually exclusive. They complement each other, representing a single whole. A comprehensive solution to the transformation problem shows that Karl Marx does not have the mistakes attributed to him by critics. JEL classification: B14, B16, B24, E11, E20, E21, E22, P16, P17 Keywords: transformation problem, original transformation, the individual sphere of production, the equilibrium price of production, inverse transformation
CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
License information was derived automatically
In this project we have reviewed existing methods used to homogenize data and developed several new methods for dealing with this diversity in survey questions on the same subject. The project is a spin-off from the World Database of Happiness, the main aim of which is to collate and make available research findings on the subjective enjoyment of life and to prepare these data for research synthesis. The first methods we discuss were proposed in the book ‘Happiness in Nations’ and which were used at the inception of the World Database of Happiness. Some 10 years later a new method was introduced: the International Happiness Scale Interval Study (HSIS). Taking the HSIS as a basis the Continuum Approach was developed. Then, building on this approach, we developed the Reference Distribution Method.
https://www.law.cornell.edu/uscode/text/17/106https://www.law.cornell.edu/uscode/text/17/106
Per- and polyfluoroalkyl substances (PFAS) represent a class of synthetic compounds widely utilized in consumer products and prevalent across various environmental matrices. They are typically analyzed individually or comprehensively as a chemical family, primarily identified by their abundant fluorine content. Over the last two decades, there has been a notable trend towards utilizing partially fluorinated PFAS, which emulate the characteristics of fully fluorinated terminal PFAS but are subject to less stringent regulatory oversight. To provide context for experimental work, Chapter 1 introduces the dissertation’s objectives and overarching research questions, while Chapter 2 presents a comprehensive review of PFAS and fluorine analysis techniques. This review highlights current methodological strengths, limitations, and key parameters, laying the groundwork for the two primary research projects. The first project, described in Chapter 3, investigates the ambient transformation of PFAS precursors. It begins by exploring an elimination pathway under controlled laboratory conditions using a base, establishing a baseline understanding of the reaction. A similar experiment is then conducted using a consumer-grade cleaning agent, demonstrating that comparable transformation is possible in indoor environments. Using both targeted and non-targeted analysis, the study reveals challenges in accurately assessing PFAS composition and species abundance due to the transformation of precursors to terminal PFAS. The second project, detailed in Chapter 4, focuses on developing analytical methods for total fluorine (TF) detection using Particle-Induced Gamma-ray Emission (PIGE) spectrometry, a relatively novel technique in PFAS analysis that lacks standardized protocols for this chemical class. Despite this, PIGE demonstrated promising results comparable to those of established analytical instruments. Successful TF analysis methods were applied to various matrices relevant to PFAS treatment, including granular activated carbon, Ottawa sand, and brass. Finally, Chapter 5 synthesizes the dissertation’s findings, discusses broader implications, and identifies opportunities for future PFAS research.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
This is the dataset used for the review article "Transforming towards what? A review of futures thinking applied in the quest for navigating sustainability transformations". The spreadsheet contains the bibliographic records and the data used and organized for its analysis.
This project explored and compared recent changes in primary care in Scotland and China, focusing on how these developments addressed the needs of ageing populations and sought to reduce health inequalities. The study identified key facilitators and barriers to progress in both countries and highlighted opportunities for mutual learning.
Data were collected by GP and patient surveys and individual qualitative interviews with GPs, patients and primary care multidisciplinary team members.
The data cannot be made available for future reuse as participants were not informed about secondary reuse.
Abstract. In this paper, the author presents fundamentally new methods for transforming the value of commodities into original and equilibrium prices of production using Karl Marx's five-sector tables from the third volume of Capital. Mathematical verification of methods is performed using sequential iterations and the program Wolfram Mathematics. For the first time, the method of inverse transformation of production prices of commodities into value prices is presented. It is proved that the pricing systems based on the principles of value and price of production are not mutually exclusive. They complement each other, representing a single whole. A comprehensive solution to the transformation problem shows that Karl Marx does not have the mistakes attributed to him by critics. JEL classification: B14, B16, B24, E11, E20, E21, E22, P16, P17 Keywords: transformation problem, original transformation, the individual sphere of production, the equilibrium price of production, inverse transformation
https://spdx.org/licenses/CC0-1.0.htmlhttps://spdx.org/licenses/CC0-1.0.html
Advances in technology allow the acquisition of data with high spatial and temporal resolution. These datasets are usually accompanied by estimates of the measurement uncertainty, which may be spatially or temporally varying and should be taken into consideration when making decisions based on the data. At the same time, various transformations are commonly implemented to reduce the dimensionality of the datasets for post-processing, or to extract significant features. However, the corresponding uncertainty is not usually represented in the low-dimensional or feature vector space. A method is proposed that maps the measurement uncertainty into the equivalent low-dimensional space with the aid of approximate Bayesian computation, resulting in a distribution that can be used to make statistical inferences. The method involves no assumptions about the probability distribution of the measurement error and is independent of the feature extraction process as demonstrated in three examples. In the first two examples Chebyshev polynomials were used to analyse structural displacements and soil moisture measurements; while in the third, principal component analysis was used to decompose global ocean temperature data. The uses of the method range from supporting decision making in model validation or confirmation, model updating or calibration and tracking changes in condition, such as the characterisation of the El Niño Southern Oscillation.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Principal component analysis (PCA) is a popular dimension-reduction method to reduce the complexity and obtain the informative aspects of high-dimensional datasets. When the data distribution is skewed, data transformation is commonly used prior to applying PCA. Such transformation is usually obtained from previous studies, prior knowledge, or trial-and-error. In this work, we develop a model-based method that integrates data transformation in PCA and finds an appropriate data transformation using the maximum profile likelihood. Extensions of the method to handle functional data and missing values are also developed. Several numerical algorithms are provided for efficient computation. The proposed method is illustrated using simulated and real-world data examples. Supplementary materials for this article are available online.