88 datasets found
  1. Italy: perceived manipulation of people online 2019

    • statista.com
    Updated Jul 7, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Statista (2022). Italy: perceived manipulation of people online 2019 [Dataset]. https://www.statista.com/statistics/1015132/perceived-people-manipulation-online-in-italy/
    Explore at:
    Dataset updated
    Jul 7, 2022
    Dataset authored and provided by
    Statistahttp://statista.com/
    Time period covered
    2019
    Area covered
    Italy
    Description

    This statistic depicts the results of a survey about how easy it is to manipulate people online in Italy in 2019. According to data, the largest group of users (29.3 percent) agreed that online people could be influenced easily, whereas 29.4 percent completely believed in this potential. Only 3.1 percent of respondents were barely of the same opinion.

  2. Global adults on social media making it easier to manipulate people 2022

    • statista.com
    Updated Apr 23, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Statista (2024). Global adults on social media making it easier to manipulate people 2022 [Dataset]. https://www.statista.com/statistics/1462913/social-media-easier-to-manipulate-people/
    Explore at:
    Dataset updated
    Apr 23, 2024
    Dataset authored and provided by
    Statistahttp://statista.com/
    Area covered
    Worldwide
    Description

    According to a 2022 survey, 84 percent of global respondents said that social media made it much easier to manipulate people. Overall, 91 percent of respondents in the Netherlands agreed with this statement, as did 90 percent of those in Australia. A little over half of respondents in Malaysia agreed that social media made it easier to manipulate people.

  3. f

    Collection of example datasets used for the book - R Programming -...

    • figshare.com
    txt
    Updated Dec 4, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Kingsley Okoye; Samira Hosseini (2023). Collection of example datasets used for the book - R Programming - Statistical Data Analysis in Research [Dataset]. http://doi.org/10.6084/m9.figshare.24728073.v1
    Explore at:
    txtAvailable download formats
    Dataset updated
    Dec 4, 2023
    Dataset provided by
    figshare
    Authors
    Kingsley Okoye; Samira Hosseini
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    This book is written for statisticians, data analysts, programmers, researchers, teachers, students, professionals, and general consumers on how to perform different types of statistical data analysis for research purposes using the R programming language. R is an open-source software and object-oriented programming language with a development environment (IDE) called RStudio for computing statistics and graphical displays through data manipulation, modelling, and calculation. R packages and supported libraries provides a wide range of functions for programming and analyzing of data. Unlike many of the existing statistical softwares, R has the added benefit of allowing the users to write more efficient codes by using command-line scripting and vectors. It has several built-in functions and libraries that are extensible and allows the users to define their own (customized) functions on how they expect the program to behave while handling the data, which can also be stored in the simple object system.For all intents and purposes, this book serves as both textbook and manual for R statistics particularly in academic research, data analytics, and computer programming targeted to help inform and guide the work of the R users or statisticians. It provides information about different types of statistical data analysis and methods, and the best scenarios for use of each case in R. It gives a hands-on step-by-step practical guide on how to identify and conduct the different parametric and non-parametric procedures. This includes a description of the different conditions or assumptions that are necessary for performing the various statistical methods or tests, and how to understand the results of the methods. The book also covers the different data formats and sources, and how to test for reliability and validity of the available datasets. Different research experiments, case scenarios and examples are explained in this book. It is the first book to provide a comprehensive description and step-by-step practical hands-on guide to carrying out the different types of statistical analysis in R particularly for research purposes with examples. Ranging from how to import and store datasets in R as Objects, how to code and call the methods or functions for manipulating the datasets or objects, factorization, and vectorization, to better reasoning, interpretation, and storage of the results for future use, and graphical visualizations and representations. Thus, congruence of Statistics and Computer programming for Research.

  4. d

    Data from: Precipitation manipulation experiments may be confounded by water...

    • catalog.data.gov
    • datasets.ai
    • +2more
    Updated Apr 21, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Agricultural Research Service (2025). Data from: Precipitation manipulation experiments may be confounded by water source [Dataset]. https://catalog.data.gov/dataset/data-from-precipitation-manipulation-experiments-may-be-confounded-by-water-source-7d7bc
    Explore at:
    Dataset updated
    Apr 21, 2025
    Dataset provided by
    Agricultural Research Service
    Description

    This is digital research data corresponding to the manuscript, Reinhart, K.O., Vermeire, L.T. Precipitation Manipulation Experiments May Be Confounded by Water Source. J Soil Sci Plant Nutr (2023). https://doi.org/10.1007/s42729-023-01298-0 Files for a 3x2x2 factorial field experiment and water quality data used to create Table 1. Data for the experiment were used for the statistical analysis and generation of summary statistics for Figure 2. Purpose: This study aims to investigate the consequences of performing precipitation manipulation experiments with mineralized water in place of rainwater (i.e. demineralized water). Limited attention has been paid to the effects of water mineralization on plant and soil properties, even when the experiments are in a rainfed context. Methods: We conducted a 6-yr experiment with a gradient in spring rainfall (70, 100, and 130% of ambient). We tested effects of rainfall treatments on plant biomass and six soil properties and interpreted the confounding effects of dissolved solids in irrigation water. Results: Rainfall treatments affected all response variables. Sulfate was the most common dissolved solid in irrigation water and was 41 times more abundant in irrigated (i.e. 130% of ambient) than other plots. Soils of irrigated plots also had elevated iron (16.5 µg × 10 cm-2 × 60-d vs 8.9) and pH (7.0 vs 6.8). The rainfall gradient also had a nonlinear (hump-shaped) effect on plant available phosphorus (P). Plant and microbial biomasses are often limited by and positively associated with available P, suggesting the predicted positive linear relationship between plant biomass and P was confounded by additions of mineralized water. In other words, the unexpected nonlinear relationship was likely driven by components of mineralized irrigation water (i.e. calcium, iron) and/or shifts in soil pH that immobilized P. Conclusions: Our results suggest robust precipitation manipulation experiments should either capture rainwater when possible (or use demineralized water) or consider the confounding effects of mineralized water on plant and soil properties. Resources in this dataset: Resource Title: Readme file- Data dictionary File Name: README.txt Resource Description: File contains data dictionary to accompany data files for a research study. Resource Title: 3x2x2 factorial dataset.csv File Name: 3x2x2 factorial dataset.csv Resource Description: Dataset is for a 3x2x2 factorial field experiment (factors: rainfall variability, mowing seasons, mowing intensity) conducted in northern mixed-grass prairie vegetation in eastern Montana, USA. Data include activity of 5 plant available nutrients, soil pH, and plant biomass metrics. Data from 2018. Resource Title: water quality dataset.csv File Name: water quality dataset.csv Resource Description: Water properties (pH and common dissolved solids) of samples from Yellowstone River collected near Miles City, Montana. Data extracted from Rinella MJ, Muscha JM, Reinhart KO, Petersen MK (2021) Water quality for livestock in northern Great Plains rangelands. Rangeland Ecol. Manage. 75: 29-34.

  5. ARC Code TI: IND: Creation and Manipulation of Decision Trees from Data

    • catalog.data.gov
    • s.cnmilf.com
    • +1more
    Updated Apr 10, 2025
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Ames Research Center (2025). ARC Code TI: IND: Creation and Manipulation of Decision Trees from Data [Dataset]. https://catalog.data.gov/dataset/arc-code-ti-ind-creation-and-manipulation-of-decision-trees-from-data
    Explore at:
    Dataset updated
    Apr 10, 2025
    Dataset provided by
    Ames Research Centerhttps://nasa.gov/ames/
    Description

    IND is applicable to most data sets consisting of independent instances, each described by a fixed length vector of attribute values. An attribute value may be a number, one of a set of attribute specific symbols, or omitted. One of the attributes is delegated the 'target' and IND grows trees to predict the target. Prediction can then be done on new data or the decision tree printed out for inspection.

  6. Data manipulation and visualization exercise

    • kaggle.com
    Updated Oct 8, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Pawan Saini (2023). Data manipulation and visualization exercise [Dataset]. https://www.kaggle.com/datasets/pawansaini01/data-manipulation-and-visualization-exercise
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Oct 8, 2023
    Dataset provided by
    Kagglehttp://kaggle.com/
    Authors
    Pawan Saini
    License

    https://creativecommons.org/publicdomain/zero/1.0/https://creativecommons.org/publicdomain/zero/1.0/

    Description

    Dataset

    This dataset was created by Pawan Saini

    Released under CC0: Public Domain

    Contents

  7. S

    Assessment Data for VR manipulation from Six Factos

    • scidb.cn
    Updated Mar 26, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Liu Gangling (2024). Assessment Data for VR manipulation from Six Factos [Dataset]. http://doi.org/10.57760/sciencedb.17190
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Mar 26, 2024
    Dataset provided by
    Science Data Bank
    Authors
    Liu Gangling
    Description

    Assessment Data for VR manipulation from Six Factos

  8. f

    R code generate analysis centralized.

    • plos.figshare.com
    txt
    Updated Nov 14, 2024
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Romain Jégou; Camille Bachot; Charles Monteil; Eric Boernert; Jacek Chmiel; Mathieu Boucher; David Pau (2024). R code generate analysis centralized. [Dataset]. http://doi.org/10.1371/journal.pone.0312697.s012
    Explore at:
    txtAvailable download formats
    Dataset updated
    Nov 14, 2024
    Dataset provided by
    PLOS ONE
    Authors
    Romain Jégou; Camille Bachot; Charles Monteil; Eric Boernert; Jacek Chmiel; Mathieu Boucher; David Pau
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    MethodsThe objective of this project was to determine the capability of a federated analysis approach using DataSHIELD to maintain the level of results of a classical centralized analysis in a real-world setting. This research was carried out on an anonymous synthetic longitudinal real-world oncology cohort randomly splitted in three local databases, mimicking three healthcare organizations, stored in a federated data platform integrating DataSHIELD. No individual data transfer, statistics were calculated simultaneously but in parallel within each healthcare organization and only summary statistics (aggregates) were provided back to the federated data analyst.Descriptive statistics, survival analysis, regression models and correlation were first performed on the centralized approach and then reproduced on the federated approach. The results were then compared between the two approaches.ResultsThe cohort was splitted in three samples (N1 = 157 patients, N2 = 94 and N3 = 64), 11 derived variables and four types of analyses were generated. All analyses were successfully reproduced using DataSHIELD, except for one descriptive variable due to data disclosure limitation in the federated environment, showing the good capability of DataSHIELD. For descriptive statistics, exactly equivalent results were found for the federated and centralized approaches, except some differences for position measures. Estimates of univariate regression models were similar, with a loss of accuracy observed for multivariate models due to source database variability.ConclusionOur project showed a practical implementation and use case of a real-world federated approach using DataSHIELD. The capability and accuracy of common data manipulation and analysis were satisfying, and the flexibility of the tool enabled the production of a variety of analyses while preserving the privacy of individual data. The DataSHIELD forum was also a practical source of information and support. In order to find the right balance between privacy and accuracy of the analysis, set-up of privacy requirements should be established prior to the start of the analysis, as well as a data quality review of the participating healthcare organization.

  9. Countries using organized social media manipulation campaigns 2017-2020

    • statista.com
    Updated Apr 28, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Statista (2022). Countries using organized social media manipulation campaigns 2017-2020 [Dataset]. https://www.statista.com/statistics/1023881/organized-social-media-manipulation-campaigns-worldwide/
    Explore at:
    Dataset updated
    Apr 28, 2022
    Dataset authored and provided by
    Statistahttp://statista.com/
    Area covered
    Worldwide
    Description

    In 2020, it was found that the number of countries using social media to spread computational propaganda and disinformation about politics was at an all-time high. Governments and political parties of 81 countries were using social media manipulation to influence public attitudes and to spread disinformation.

  10. u

    Behavioral and manipulation data

    • rdr.ucl.ac.uk
    csv
    Updated Mar 31, 2025
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Hernando Martínez Vergara (2025). Behavioral and manipulation data [Dataset]. http://doi.org/10.5522/04/28654883.v1
    Explore at:
    csvAvailable download formats
    Dataset updated
    Mar 31, 2025
    Dataset provided by
    University College London
    Authors
    Hernando Martínez Vergara
    License

    CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
    License information was derived automatically

    Description

    Data for publication:Dopaminergic action prediction errors serve as a value-free teaching signalAnimals’ choice behavior is characterized by two main tendencies: taking actions that led to rewards and repeating past actions. Theory suggests these strategies may be reinforced by different types of dopaminergic teaching signals: reward prediction error to reinforce value-based associations and movement-based action prediction errors to reinforce value-free repetitive associations. Here we use an auditory-discrimination task in mice to show that movement-related dopamine activity in the tail of the striatum encodes the hypothesized action prediction error signal. Causal manipulations reveal that this prediction error serves as a value-free teaching signal that supports learning by reinforcing repeated associations. Computational modelling and experiments demonstrate that action prediction errors alone cannot support reward-guided learning but when paired with the reward prediction error circuitry they serve to consolidate stable sound-action associations in a value-free manner. Together we show that there are two types of dopaminergic prediction errors that work in tandem to support learning, each reinforcing different types of association in different striatal areas.These datasets generate main Figures 1, 4, and supplementary panels

  11. d

    Database on failure risk and financial data manipulation indicators for...

    • search.dataone.org
    Updated Nov 8, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Achim, Monica Violeta; Lucuț-Capras, Isabella (2023). Database on failure risk and financial data manipulation indicators for Romanian non-financial listed companies [Dataset]. http://doi.org/10.7910/DVN/28DTRY
    Explore at:
    Dataset updated
    Nov 8, 2023
    Dataset provided by
    Harvard Dataverse
    Authors
    Achim, Monica Violeta; Lucuț-Capras, Isabella
    Description

    These database collect data about failure risk and data manipulation indicators for a number of 63 non-financial Romanian companies listed on the Bucharest Stock Exchange, for the period 2015- 2020. The database contains all the required financial data of the 63 non financial companies, along with the calculations of Beneish scores and Altman scores for each of the company in the sample.

  12. o

    Replication data for: Big Data: New Tricks for Econometrics

    • openicpsr.org
    Updated May 1, 2014
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Hal R. Varian (2014). Replication data for: Big Data: New Tricks for Econometrics [Dataset]. http://doi.org/10.3886/E113925V1
    Explore at:
    Dataset updated
    May 1, 2014
    Dataset provided by
    American Economic Association
    Authors
    Hal R. Varian
    Time period covered
    May 1, 2014
    Description

    Computers are now involved in many economic transactions and can capture data associated with these transactions, which can then be manipulated and analyzed. Conventional statistical and econometric techniques such as regression often work well, but there are issues unique to big datasets that may require different tools. First, the sheer size of the data involved may require more powerful data manipulation tools. Second, we may have more potential predictors than appropriate for estimation, so we need to do some kind of variable selection. Third, large datasets may allow for more flexible relationships than simple linear models. Machine learning techniques such as decision trees, support vector machines, neural nets, deep learning, and so on may allow for more effective ways to model complex relationships. In this essay, I will describe a few of these tools for manipulating and analyzing big data. I believe that these methods have a lot to offer and should be more widely known and used by economists.

  13. d

    cmomy: A python package to calculate and manipulate Central (co)moments.

    • datasets.ai
    • res1catalogd-o-tdatad-o-tgov.vcapture.xyz
    • +1more
    0, 33
    Updated Aug 27, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    National Institute of Standards and Technology (2024). cmomy: A python package to calculate and manipulate Central (co)moments. [Dataset]. https://datasets.ai/datasets/cmomy-a-python-package-to-calculate-and-manipulate-central-comoments-dcd00
    Explore at:
    0, 33Available download formats
    Dataset updated
    Aug 27, 2024
    Dataset authored and provided by
    National Institute of Standards and Technology
    Description

    cmomy is a python package to calculate central moments and co-moments in a numerical stable and direct way. Behind the scenes, cmomy makes use of Numba to rapidly calculate moments. cmomy provides utilities to calculate central moments from individual samples, precomputed central moments, and precomputed raw moments. It also provides routines to perform bootstrap resampling based on raw data, or precomputed moments. cmomy has numpy array and xarray DataArray interfaces.

  14. o

    Mobile co-manipulation data

    • explore.openaire.eu
    • zenodo.org
    • +1more
    Updated Aug 5, 2021
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Aitor Ibarguren (2021). Mobile co-manipulation data [Dataset]. http://doi.org/10.5281/zenodo.5163197
    Explore at:
    Dataset updated
    Aug 5, 2021
    Authors
    Aitor Ibarguren
    Description

    Data acquired during large part co-manipulation processes. Specifically, trajectory percentage and trajectory deviation. Notation of files (i.e. "u2_AB_500.csv"): User: u2 would be the second subject of the experiment. Path: Two options, AB (station A to station B) and BA (station B to A). Maximum allowed distance: Value which defined the width of the lane.

  15. d

    Replication Data for: How dropping subjects who failed manipulation checks...

    • search.dataone.org
    • dataverse.harvard.edu
    Updated Nov 8, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Varaine, Simon (2023). Replication Data for: How dropping subjects who failed manipulation checks can bias your experimental results. An illustrative case [Dataset]. http://doi.org/10.7910/DVN/7DXBGG
    Explore at:
    Dataset updated
    Nov 8, 2023
    Dataset provided by
    Harvard Dataverse
    Authors
    Varaine, Simon
    Description

    Manipulations checks are post-experimental measures widely used to verify that subjects understood the treatment. Some researchers drop subjects who failed manipulation checks in order to limit the analyses to attentive subjects. This short report offers a novel illustration on how this practice may bias experimental results: in the present case, through confirming a hypothesis that is likely false. In a survey experiment, subjects were primed with fictional news stories depicting an economic decline versus prosperity. Subjects were then asked whether the news story depicted an economic decline or prosperity. Results indicate that responses to this manipulation check captured subjects’ pre-existing beliefs about the economic situation. As a consequence, dropping subjects who failed the manipulation check mixes the effects of pre-existing and induced beliefs, increasing the risk of false positive findings. Researchers should avoid dropping subjects based on post-treatment measures and rely on pre-treatment measures of attentiveness.

  16. Data from: Robotic manipulation datasets for offline compositional...

    • data.niaid.nih.gov
    • search.dataone.org
    • +1more
    zip
    Updated Jun 6, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Marcel Hussing; Jorge Mendez; Anisha Singrodia; Cassandra Kent; Eric Eaton (2024). Robotic manipulation datasets for offline compositional reinforcement learning [Dataset]. http://doi.org/10.5061/dryad.9cnp5hqps
    Explore at:
    zipAvailable download formats
    Dataset updated
    Jun 6, 2024
    Dataset provided by
    University of Pennsylvania
    Massachusetts Institute of Technology
    Authors
    Marcel Hussing; Jorge Mendez; Anisha Singrodia; Cassandra Kent; Eric Eaton
    License

    https://spdx.org/licenses/CC0-1.0.htmlhttps://spdx.org/licenses/CC0-1.0.html

    Description

    Offline reinforcement learning (RL) is a promising direction that allows RL agents to be pre-trained from large datasets avoiding recurrence of expensive data collection. To advance the field, it is crucial to generate large-scale datasets. Compositional RL is particularly appealing for generating such large datasets, since 1) it permits creating many tasks from few components, and 2) the task structure may enable trained agents to solve new tasks by combining relevant learned components. This submission provides four offline RL datasets for simulated robotic manipulation created using the 256 tasks from CompoSuite Mendez et al., 2022. In every task in CompoSuite, a robot arm is used to manipulate an object to achieve an objective all while trying to avoid an obstacle. There are for components for each of these four axes that can be combined arbitrarily leading to a total of 256 tasks. The component choices are * Robot: IIWA, Jaco, Kinova3, Panda* Object: Hollow box, box, dumbbell, plate* Objective: Push, pick and place, put in shelf, put in trashcan* Obstacle: None, wall between robot and object, wall between goal and object, door between goal and object The four included datasets are collected using separate agents each trained to a different degree of performance, and each dataset consists of 256 million transitions. The degrees of performance are expert data, medium data, warmstart data and replay data: * Expert dataset: Transitions from an expert agent that was trained to achieve 90% success on every task.* Medium dataset: Transitions from a medium agent that was trained to achieve 30% success on every task.* Warmstart dataset: Transitions from a Soft-actor critic agent trained for a fixed duration of one million steps.* Medium-replay-subsampled dataset: Transitions that were stored during the training of a medium agent up to 30% success. These datasets are intended for the combined study of compositional generalization and offline reinforcement learning. Methods The datasets were collected by using several deep reinforcement learning agents trained to the various degrees of performance described above on the CompoSuite benchmark (https://github.com/Lifelong-ML/CompoSuite) which builds on top of robosuite (https://github.com/ARISE-Initiative/robosuite) and uses the MuJoCo simulator (https://github.com/deepmind/mujoco). During reinforcement learning training, we stored the data that was collected by each agent in a separate buffer for post-processing. Then, after training, to collect the expert and medium dataset, we run the trained agents for 2000 trajectories of length 500 online in the CompoSuite benchmark and store the trajectories. These add up to a total of 1 million state-transitions tuples per dataset, totalling a full 256 million datapoints per dataset. The warmstart and medium-replay-subsampled dataset contain trajectories from the stored training buffer of the SAC agent trained for a fixed duration and the medium agent respectively. For medium-replay-subsampled data, we uniformly sample trajectories from the training buffer until we reach more than 1 million transitions. Since some of the tasks have termination conditions, some of these trajectories are trunctated and not of length 500. This sometimes results in a number of sampled transitions larger than 1 million. Therefore, after sub-sampling, we artificially truncate the last trajectory and place a timeout at the final position. This can in some rare cases lead to one incorrect trajectory if the datasets are used for finite horizon experimentation. However, this truncation is required to ensure consistent dataset sizes, easy data readability and compatibility with other standard code implementations. The four datasets are split into four tar.gz folders each yielding a total of 12 compressed folders. Every sub-folder contains all the tasks for one of the four robot arms for that dataset. In other words, every tar.gz folder contains a total of 64 tasks using the same robot arm and four tar.gz files form a full dataset. This is done to enable people to only download a part of the dataset in case they do not need all 256 tasks. For every task, the data is separately stored in an hdf5 file allowing for the usage of arbitrary task combinations and mixing of data qualities across the four datasets. Every task is contained in a folder that is named after the CompoSuite elements it uses. In other words, every task is represented as a folder named

  17. H

    Replication Data for: State-Linked Manipulated Media in the Time of...

    • dataverse.harvard.edu
    Updated Nov 22, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Benjamin Bagozzi; Karthik Balasubramanian; Rajni Goel; Christopher Parker (2024). Replication Data for: State-Linked Manipulated Media in the Time of Covid-19: A Look at Iran [Dataset]. http://doi.org/10.7910/DVN/MJHKNI
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Nov 22, 2024
    Dataset provided by
    Harvard Dataverse
    Authors
    Benjamin Bagozzi; Karthik Balasubramanian; Rajni Goel; Christopher Parker
    License

    CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
    License information was derived automatically

    Area covered
    Iran
    Description

    What drives changes in the thematic focus of state-linked manipulated media? We study this question in relation to a long-running Iranian state-linked manipulated media campaign that was uncovered by Twitter in 2021. Using a variety of machine learning methods, we uncover and analyze how this manipulation campaign's topical themes changed in relation to rising Covid-19 cases in Iran. By using the topics of the tweets in a novel way, we find that increases in domestic Covid-19 cases engendered a shift in Iran's manipulated media focus away from Covid-19 themes and towards international finance- and investment-focused themes. These findings underscore (i) the potential for state-linked manipulated media campaigns to be used for diversionary purposes and (ii) the promise of machine learning methods for detecting such behaviors.

  18. d

    Replication Data for: A Note on Dropping Experimental Subjects who Fail a...

    • search.dataone.org
    Updated Nov 22, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Aronow, Peter M.; Baron, Jonathon; Pinson, Lauren (2023). Replication Data for: A Note on Dropping Experimental Subjects who Fail a Manipulation Check [Dataset]. http://doi.org/10.7910/DVN/GXXYMH
    Explore at:
    Dataset updated
    Nov 22, 2023
    Dataset provided by
    Harvard Dataverse
    Authors
    Aronow, Peter M.; Baron, Jonathon; Pinson, Lauren
    Description

    This repository mirrors the Yale ISPS Data Archive containing the relevant replication files pertaining to Aronow, Baron, Pinson (2018), "A Note on Dropping Experimental Subjects who Fail a Manipulation Check." https://isps.yale.edu/research/data/d150. Please direct inquiries to isps@yale.edu.

  19. U

    Data set for: Molecular and atomic manipulation mediated by electronic...

    • researchdata.bath.ac.uk
    sxm, txt, xlsx
    Updated Nov 15, 2016
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Kristina Rusimova; Peter Sloan (2016). Data set for: Molecular and atomic manipulation mediated by electronic excitation of the underlying Si(111)-7x7 surface [Dataset]. http://doi.org/10.15125/BATH-00315
    Explore at:
    txt, sxm, xlsxAvailable download formats
    Dataset updated
    Nov 15, 2016
    Dataset provided by
    University of Bath
    Authors
    Kristina Rusimova; Peter Sloan
    Dataset funded by
    Engineering and Physical Sciences Research Council
    Description

    All the data associated with this publication, please see publication for methodology and details.

  20. Data from: The PS-Battles Dataset - an Image Collection for Image...

    • zenodo.org
    bin, sh, tsv
    Updated Dec 21, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Silvan Heller; Luca Rossetto; Heiko Schuldt; Silvan Heller; Luca Rossetto; Heiko Schuldt (2022). The PS-Battles Dataset - an Image Collection for Image Manipulation Detection [Dataset]. http://doi.org/10.5281/zenodo.7467933
    Explore at:
    sh, tsv, binAvailable download formats
    Dataset updated
    Dec 21, 2022
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Silvan Heller; Luca Rossetto; Heiko Schuldt; Silvan Heller; Luca Rossetto; Heiko Schuldt
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    The boost of available digital media has led to a significant increase in derivative work. With tools for manipulating objects becoming more and more mature, it can be very difficult to determine whether one piece of media was derived from another one or tampered with. As derivations can be done with malicious intent, there is an urgent need for reliable and easily usable tampering detection methods. However, even media considered semantically untampered by humans might have already undergone compression steps or light post-processing, making automated detection of tampering susceptible to false positives. In this paper, we present the PS-Battles dataset which is gathered from a large community of image manipulation enthusiasts and provides a basis for media derivation and manipulation detection in the visual domain. The dataset consists of 102'028 images grouped into 11'142 subsets, each containing the original image as well as a varying number of manipulated derivatives.

    Mirror of the git repository: https://github.com/dbisUnibas/PS-Battles

    Paper on arxiv: https://arxiv.org/abs/1804.04866

Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
Statista (2022). Italy: perceived manipulation of people online 2019 [Dataset]. https://www.statista.com/statistics/1015132/perceived-people-manipulation-online-in-italy/
Organization logo

Italy: perceived manipulation of people online 2019

Explore at:
Dataset updated
Jul 7, 2022
Dataset authored and provided by
Statistahttp://statista.com/
Time period covered
2019
Area covered
Italy
Description

This statistic depicts the results of a survey about how easy it is to manipulate people online in Italy in 2019. According to data, the largest group of users (29.3 percent) agreed that online people could be influenced easily, whereas 29.4 percent completely believed in this potential. Only 3.1 percent of respondents were barely of the same opinion.

Search
Clear search
Close search
Google apps
Main menu