According to a 2022 survey, 84 percent of global respondents said that social media made it much easier to manipulate people. Overall, 91 percent of respondents in the Netherlands agreed with this statement, as did 90 percent of those in Australia. A little over half of respondents in Malaysia agreed that social media made it easier to manipulate people.
In 2020, it was found that the number of countries using social media to spread computational propaganda and disinformation about politics was at an all-time high. Governments and political parties of 81 countries were using social media manipulation to influence public attitudes and to spread disinformation.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
MethodsThe objective of this project was to determine the capability of a federated analysis approach using DataSHIELD to maintain the level of results of a classical centralized analysis in a real-world setting. This research was carried out on an anonymous synthetic longitudinal real-world oncology cohort randomly splitted in three local databases, mimicking three healthcare organizations, stored in a federated data platform integrating DataSHIELD. No individual data transfer, statistics were calculated simultaneously but in parallel within each healthcare organization and only summary statistics (aggregates) were provided back to the federated data analyst.Descriptive statistics, survival analysis, regression models and correlation were first performed on the centralized approach and then reproduced on the federated approach. The results were then compared between the two approaches.ResultsThe cohort was splitted in three samples (N1 = 157 patients, N2 = 94 and N3 = 64), 11 derived variables and four types of analyses were generated. All analyses were successfully reproduced using DataSHIELD, except for one descriptive variable due to data disclosure limitation in the federated environment, showing the good capability of DataSHIELD. For descriptive statistics, exactly equivalent results were found for the federated and centralized approaches, except some differences for position measures. Estimates of univariate regression models were similar, with a loss of accuracy observed for multivariate models due to source database variability.ConclusionOur project showed a practical implementation and use case of a real-world federated approach using DataSHIELD. The capability and accuracy of common data manipulation and analysis were satisfying, and the flexibility of the tool enabled the production of a variety of analyses while preserving the privacy of individual data. The DataSHIELD forum was also a practical source of information and support. In order to find the right balance between privacy and accuracy of the analysis, set-up of privacy requirements should be established prior to the start of the analysis, as well as a data quality review of the participating healthcare organization.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Data for benchmarking SPC against other process monitoring methods. The data consist of a one-dimensional timeseries of floats (x.csv). Addititionally information whether the data are within the specifications are provided as another time series (y.csv). The data are generated by solving an optimization problem for each time to generate a mixture distribution of different probability distributions. Then for each timestep one record is sampled. Inputs for the optimization problem are the given probability distributions, the lower and upper limit of the tolerance interval as well as the desired median of the data. Additionally weights of the different probability distributions can be given as boundary condions for the different time steps. Metadata generated from the solving are stored in k_matrix.csv (wheights at each time step) and distribs (probability distribution objects). The data consists of phases with data from a stable mixture distribution and phases with data from a mixture distribution that do not fulfill the stability criteria.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
MethodsThe objective of this project was to determine the capability of a federated analysis approach using DataSHIELD to maintain the level of results of a classical centralized analysis in a real-world setting. This research was carried out on an anonymous synthetic longitudinal real-world oncology cohort randomly splitted in three local databases, mimicking three healthcare organizations, stored in a federated data platform integrating DataSHIELD. No individual data transfer, statistics were calculated simultaneously but in parallel within each healthcare organization and only summary statistics (aggregates) were provided back to the federated data analyst.Descriptive statistics, survival analysis, regression models and correlation were first performed on the centralized approach and then reproduced on the federated approach. The results were then compared between the two approaches.ResultsThe cohort was splitted in three samples (N1 = 157 patients, N2 = 94 and N3 = 64), 11 derived variables and four types of analyses were generated. All analyses were successfully reproduced using DataSHIELD, except for one descriptive variable due to data disclosure limitation in the federated environment, showing the good capability of DataSHIELD. For descriptive statistics, exactly equivalent results were found for the federated and centralized approaches, except some differences for position measures. Estimates of univariate regression models were similar, with a loss of accuracy observed for multivariate models due to source database variability.ConclusionOur project showed a practical implementation and use case of a real-world federated approach using DataSHIELD. The capability and accuracy of common data manipulation and analysis were satisfying, and the flexibility of the tool enabled the production of a variety of analyses while preserving the privacy of individual data. The DataSHIELD forum was also a practical source of information and support. In order to find the right balance between privacy and accuracy of the analysis, set-up of privacy requirements should be established prior to the start of the analysis, as well as a data quality review of the participating healthcare organization.
IND is applicable to most data sets consisting of independent instances, each described by a fixed length vector of attribute values. An attribute value may be a number, one of a set of attribute specific symbols, or omitted. One of the attributes is delegated the 'target' and IND grows trees to predict the target. Prediction can then be done on new data or the decision tree printed out for inspection.
CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
License information was derived automatically
Data for publication:Dopaminergic action prediction errors serve as a value-free teaching signalAnimals’ choice behavior is characterized by two main tendencies: taking actions that led to rewards and repeating past actions. Theory suggests these strategies may be reinforced by different types of dopaminergic teaching signals: reward prediction error to reinforce value-based associations and movement-based action prediction errors to reinforce value-free repetitive associations. Here we use an auditory-discrimination task in mice to show that movement-related dopamine activity in the tail of the striatum encodes the hypothesized action prediction error signal. Causal manipulations reveal that this prediction error serves as a value-free teaching signal that supports learning by reinforcing repeated associations. Computational modelling and experiments demonstrate that action prediction errors alone cannot support reward-guided learning but when paired with the reward prediction error circuitry they serve to consolidate stable sound-action associations in a value-free manner. Together we show that there are two types of dopaminergic prediction errors that work in tandem to support learning, each reinforcing different types of association in different striatal areas.These datasets generate main Figures 1, 4, and supplementary panels
About this webinar We rarely receive the research data in an appropriate form. Often data is messy. Sometimes it is incomplete. And sometimes there’s too much of it. Frequently, it has errors. This webinar targets beginners and presents a quick demonstration of using the most widespread data wrangling tool, Microsoft Excel, to sort, filter, copy, protect, transform, aggregate, summarise, and visualise research data. Webinar Topics Introduction to Microsoft Excel user interface Interpret data using sorting, filtering, and conditional formatting Summarise data using functions Analyse data using pivot tables Manipulate and visualise data Handy tips to speed up your work Licence Copyright © 2021 Intersect Australia Ltd. All rights reserved.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
This file can be used to manipulate the storage technologies cost data for the Starter Data Kits.
Attribution 3.0 (CC BY 3.0)https://creativecommons.org/licenses/by/3.0/
License information was derived automatically
The shark control program (SCP) relies on nets or drumlines, or a combination of both, to minimise the threat of shark attack on humans in particular locations. Following is information on numbers and locations of sharks that have been caught by the SCP.
This dataset contains details of non-target numbers in the Shark Control program by species, date of capture, and location from 2001.
It is important to reduce the inadvertent impacts of the SCP on other marine animals (bycatch) without compromising human safety. Bycatch levels are carefully monitored and research is focused on minimising impacts on non-target species.
Open Government Licence - Canada 2.0https://open.canada.ca/en/open-government-licence-canada
License information was derived automatically
The detailed financial statistics by North American Industry Classification System (NAICS) 511130 book publishers which include all members under detailed financial statistics and by country of control, (dollars X 1,000,000), every 2 years, for five years of data.
Excel spreadsheet that contains the raw fecundity data used to conduct power simulations specific to the MEOGRT reproductive assessment. This dataset is associated with the following publication: Flynn, K., J. Swintek, and R. Johnson. The influence of control group reproduction on the statistical power of the Environmental Protection Agency’s medaka Extended One-Generation Reproduction Test (MEOGRT). ECOTOXICOLOGY AND ENVIRONMENTAL SAFETY. Elsevier Science Ltd, NY, USA, 136: 8-13, (2016).
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
corrMatrix.m: This Matlab function computes the correlation matrix of w-test statistics.
KMC.m: This Matlab function computes the critical values for max-w test statistic based on Monte Carlo method. It is needed to run corrMatrix.m before use it.
kNN.m: This Matlab function based on neural networks allows anyone to obtain the desired critical value with good control of type I error. In that case, you need to download file SBPNN.mat and save it in your folder. It is needed to run corrMatrix.m before use it.
SBPNN.mat: MATLAB's flexible network object type (called SBPNN.mat) that allows anyone to obtain the desired critical value with good control of type I error.
Examples.txt: File containing examples of both design and covariance matrices in adjustment problems of geodetic networks.
rawMC.txt: Monte-Carlo-based critical values for the following signifiance levels: α′= 0.001, α′= 0.01, α′= 0.05, α′= 0.1 and α′= 0.5. The number of the observations (n) were fixed for each α ′with n = 5 to n= 100 by a increment of 5. For each "n" the correlation between the w-tests (ρwi,wj) were also fixed from ρwi,wj = 0.00 to ρwi,wj = 1.00, by increment of 0.1, considering also taking into account the correlation ρwi,wj = 0.999. For each combination of α′,"n" and ρwi,wj, m= 5,000,000 Monte Carlo experiments were run.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
The boost of available digital media has led to a significant increase in derivative work. With tools for manipulating objects becoming more and more mature, it can be very difficult to determine whether one piece of media was derived from another one or tampered with. As derivations can be done with malicious intent, there is an urgent need for reliable and easily usable tampering detection methods. However, even media considered semantically untampered by humans might have already undergone compression steps or light post-processing, making automated detection of tampering susceptible to false positives. In this paper, we present the PS-Battles dataset which is gathered from a large community of image manipulation enthusiasts and provides a basis for media derivation and manipulation detection in the visual domain. The dataset consists of 102'028 images grouped into 11'142 subsets, each containing the original image as well as a varying number of manipulated derivatives.
Mirror of the git repository: https://github.com/dbisUnibas/PS-Battles
Paper on arxiv: https://arxiv.org/abs/1804.04866
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
In this paper, we show that concept of Statistical Process Control tools was thoroughly examined and the definitions of quality control concepts were presented. This is significant because of it is anticipated that this study will contribute to the literature as an exemplary application that demonstrates the role of statistical process control (SPC) tools in quality improvement in the evaluation and decision-making phase.
This is significant because of this study is to investigate applications of quality control, to clarify statistical control methods and problem-solving procedures, to generate proposals for problem-solving approaches, and to disseminate improvement studies in the ready-to-wear industry. The basic Statistical Process Control tools used in the study, the most repetitive faults were detected and these faults were divided into sub-headings for more detailed analysis. In this way, it was tried to prevent the repetition of faults by going down to the root causes of any detected fault. With this different perspective, it is expected that the study will contribute to other fields.
We give consent for the publication of identifiable details, which can include photograph(s) and case history and details within the text (“Material”) to be published in the Journal of Quality Technology. We confirm that have seen and been given the opportunity to read both the Material and the Article (as attached) to be published by Taylor & Francis.
This document contains the following information: Control of immigration: statistics United Kingdom 2005.
This Command Paper was laid before Parliament by a Government Minister by Command of Her Majesty. Command Papers are considered by the Government to be of interest to Parliament but are not required to be presented by legislation.
This statistic shows the number of diesel vehicles that were equipped with software designed to defeat emissions controls. As of September 30, 2015, Volkswagen confirmed that some 1.2 million Skoda vehicles were affected by the scandal. In September 2015, Volkswagen recalled some 482,000 cars in the United States.
The Geometry Manipulation Protocol (GMP) is a library which serializes datatypes between XML and ANSI C data structures to support CFD applications. This library currently provides a description of geometric configurations, general moving-body scenarios (prescribed and/or 6-DOF), and control surface settings.
CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
License information was derived automatically
The reseach objective is to present a microfluidic approach to achieve the dynamic control of particle pathlines within a flow through microfluidic device. Our approach combines three key aspects: the design of a flow-through microfluidic flow cell with the ability to manipulate the streamlines of the flow, an optimization procedure to find a priori optimal particle path-lines, and a Proportion-Integral-Derivative-based (PID) feedback controller to provide real time control over the particle manipulations. The experimental raw images were recorded with a sCMOS camera (PCO) with a pixel pitch of 6.5 μm. The camera was mounted on a microscope (Nikon Eclipse Ti) with a 1x objective. The
acquisition frequency was 5 Hz corresponding to an average in-plane displacement of 4-6 pixels between two consecutive recordings. The zip file contains the raw images and the MATLAB script used to do an experiment of two particles coming close to each other by only using the hydrodynamic forcing in a Hele-Shaw cell.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Source data covering changepoints in UAG and offtake nodes
According to a 2022 survey, 84 percent of global respondents said that social media made it much easier to manipulate people. Overall, 91 percent of respondents in the Netherlands agreed with this statement, as did 90 percent of those in Australia. A little over half of respondents in Malaysia agreed that social media made it easier to manipulate people.