94 datasets found
  1. f

    Reasons given for a report as a percentage of the overall number of reports....

    • plos.figshare.com
    xls
    Updated Jun 4, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Barbara Strobl; Simon Etter; Ilja van Meerveld; Jan Seibert (2023). Reasons given for a report as a percentage of the overall number of reports. [Dataset]. http://doi.org/10.1371/journal.pone.0222579.t001
    Explore at:
    xlsAvailable download formats
    Dataset updated
    Jun 4, 2023
    Dataset provided by
    PLOS ONE
    Authors
    Barbara Strobl; Simon Etter; Ilja van Meerveld; Jan Seibert
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Reasons given for a report as a percentage of the overall number of reports.

  2. A global reference database of crowdsourced cropland data collected using...

    • zenodo.org
    • doi.pangaea.de
    • +1more
    zip
    Updated Jul 16, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Linda See; Linda See (2024). A global reference database of crowdsourced cropland data collected using the Geo-Wiki platform [Dataset]. http://doi.org/10.1594/pangaea.873912
    Explore at:
    zipAvailable download formats
    Dataset updated
    Jul 16, 2024
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Linda See; Linda See
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    A global reference dataset on cropland was collected through a crowdsourcing campaign implemented using Geo-Wiki. This reference dataset is based on a systematic sample at latitude and longitude intersections, enhanced in locations where the cropland probability varies between 25-75% for a better representation of cropland globally. Over a three week period, around 36K samples of cropland were collected. For the purpose of quality assessment, additional datasets are provided. One is a control dataset of 1793 sample locations that have been validated by students trained in image interpretation. This dataset was used to assess the quality of the crowd validations as the campaign progressed. Another set of data contains 60 expert or gold standard validations for additional evaluation of the quality of the participants. These three datasets have two parts, one showing cropland only and one where it is compiled per location and user. This reference dataset will be used to validate and compare medium and high resolution cropland maps that have been generated using remote sensing. The dataset can also be used to train classification algorithms in developing new maps of land cover and cropland extent.

  3. Z

    Data from: A Dataset of Crowdsourced Smarthome Requirements with Creativity...

    • data.niaid.nih.gov
    Updated Jan 24, 2020
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Ajmeri, Nirav (2020). A Dataset of Crowdsourced Smarthome Requirements with Creativity Ratings [Dataset]. https://data.niaid.nih.gov/resources?id=zenodo_3550720
    Explore at:
    Dataset updated
    Jan 24, 2020
    Dataset provided by
    Ajmeri, Nirav
    Rahman, Muhammad Fazalul
    Singh, Munindar P.
    Murukannaiah, Pradeep K.
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    This dataset contains the data analyzed in the paper titled "Crowdsourcing Requirements: Does Teamwork Enhance Crowd Creativity?"

    The dataset contains the following csv files:

    1. presurvey-questions: List of presurvey questions to collect demographics

    2. disc-questions: List of DISC personality questions to cause a crowd worker’s personality. Each group has a set of 4 statements out of which the worker was expected to select one

    3. post-survey-questions: List of postsurvey questions

    4. users: List of crowd workers in the study; values 1 and 2 of the column ‘group_type’ correspond to workers in solo and interacting teams respectively

    5. presurvey-responses: Workers' responses to the presurvey

    6. personality_data: Workers’ IPIP (O, C, E, A, N metrics) and DISC (raw and normalized) scores

    7. post-survey-responses: Workers' responses to the postsurvey

    8. all_requirements: Requirements in a user story format, elicited by the crowd workers

    9. creativity-ratings.csv: Authors’ average ratings for each requirement for the metrics ‘detailedness’, ‘novelty’ and ‘usefulness’

  4. Crowdsourcing Guidelines for a Successful Data Event

    • osf.io
    Updated Jul 6, 2016
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Jamene Brooks-Kieffer; Brianna Marshall; Michelle Hudson; Mark Laufersweiler; Steve Van Tuyl; Andrew Johnson; Scott Martin (2016). Crowdsourcing Guidelines for a Successful Data Event [Dataset]. https://osf.io/uwja6
    Explore at:
    Dataset updated
    Jul 6, 2016
    Dataset provided by
    Center for Open Sciencehttps://cos.io/
    Authors
    Jamene Brooks-Kieffer; Brianna Marshall; Michelle Hudson; Mark Laufersweiler; Steve Van Tuyl; Andrew Johnson; Scott Martin
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    This project originated as an unconference-style panel at the 2016 Research Data Access and Preservation Summit. It collects case studies of research data-related events hosted or co-hosted by academic libraries. Case studies collected for the panel describe four flavors of event: the Center for Open Science Workshop on Reproducible Research, Data Carpentry, Software Carpentry, and Day of Data. Subsequent case study contributions describe more instances of these events as well as other events specific to their host institutions. Libraries and other potential sponsors are encouraged to use these case studies as resources for planning their own data-related events. The project organizers will periodically solicit more contributions to this collection of event case studies.

  5. FPCA-II. Food Price Crowdsourcing Africa-expansion

    • data.europa.eu
    csv, zip
    Updated Jan 21, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Joint Research Centre (2023). FPCA-II. Food Price Crowdsourcing Africa-expansion [Dataset]. https://data.europa.eu/data/datasets/f3bc86b0-be5f-4441-8370-c2ccb739029e?locale=en
    Explore at:
    zip, csvAvailable download formats
    Dataset updated
    Jan 21, 2023
    Dataset authored and provided by
    Joint Research Centrehttps://joint-research-centre.ec.europa.eu/index_en
    License

    http://data.europa.eu/eli/dec/2011/833/ojhttp://data.europa.eu/eli/dec/2011/833/oj

    Description

    Timely and reliable monitoring of commodity food prices is an essential requirement for assessing market and food security risks and establishing early warning systems, especially in developing economies. However, data from regional or national systems for tracking changes in food prices in sub-Saharan Africa lacks the temporal or spatial richness and is often insufficient to inform targeted interventions. In addition to limited opportunity for [near-]real-time assessment of food prices, various stages in the commodity supply chain are mostly unrepresented, thereby limiting insights on stage-related price evolution. Yet, governments and market stakeholders rely on commodity price data to make decisions on appropriate interventions or commodity-focused investments. Recent rapid technological development indicates that digital devices and connectivity services are becoming affordable for many, including in remote areas of developing economies. This offers a great opportunity for harvesting price data (via new data collection methodologies, such as crowdsourcing/crowdsensing — i.e. citizen-generated data — using mobile apps/devices) and disseminating it (via web dashboards or other means) in real-time. This real-time data can support decisions at various levels and related policy-making processes. However, market information that aims at improving the functioning of markets and supply chains requires a continuous data flow as well as quality, accessibility and trust. More data does not necessarily translate into better information. Citizen-based data-generation systems are often confronted by challenges related to data quality and citizen participation, which may be further complicated by the volume of data generated compared to traditional approaches. Following the food price hikes during the first noughties of the 21st century, the European Commission's Joint Research Centre (JRC) started working on innovative methodologies for real-time food price data collection and analysis in developing countries. The work carried out so far includes a pilot initiative to crowdsource data from selected markets across several African countries, two workshops (with relevant stakeholders and experts), and the development of a spatial statistical quality methodology to facilitate the best possible exploitation of geo-located data. Based on the latter, the JRC designed the Food Price Crowdsourcing Africa (FPCA) project and implemented it initially within two states in Northern Nigeria, then expanded to two further states. The FPCA is a credible methodology, based on the voluntary provision of data by a crowd (people living in urban, suburban, and rural areas) using a mobile app, leveraging monetary and non-monetary incentives to enhance contribution, which makes it possible to collect, analyse and validate, and disseminate staple food price data in real time across market segments. The granularity and high frequency of the crowdsourcing data open the door to real-time space-time analysis, which can be essential for policy and decision making and rapid response on specific geographic regions.

  6. d

    Open e-commerce 1.0: Five years of crowdsourced U.S. Amazon purchase...

    • search.dataone.org
    • dataverse.harvard.edu
    Updated Dec 16, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Alex Berke; Dan Calacci; Robert Mahari; Takahiro Yabe; Kent Larson; Sandy Pentland (2023). Open e-commerce 1.0: Five years of crowdsourced U.S. Amazon purchase histories with user demographics [Dataset]. http://doi.org/10.7910/DVN/YGLYDY
    Explore at:
    Dataset updated
    Dec 16, 2023
    Dataset provided by
    Harvard Dataverse
    Authors
    Alex Berke; Dan Calacci; Robert Mahari; Takahiro Yabe; Kent Larson; Sandy Pentland
    Description

    This dataset contains longitudinal purchases data from 5027 Amazon.com users in the US, spanning 2018 through 2022: amazon-purchases.csv It also includes demographic data and other consumer level variables for each user with data in the dataset. These consumer level variables were collected through an online survey and are included in survey.csv fields.csv describes the columns in the survey.csv file, where fields/survey columns correspond to survey questions. The dataset also contains the survey instrument used to collect the data. More details about the survey questions and possible responses, and the format in which they were presented can be found by viewing the survey instrument. A 'Survey ResponseID' column is present in both the amazon-purchases.csv and survey.csv files. It links a user's survey responses to their Amazon.com purchases. The 'Survey ResponseID' was randomly generated at the time of data collection. amazon-purchases.csv Each row in this file corresponds to an Amazon order. Each such row has the following columns: Survey ResponseID Order date Shipping address state Purchase price per unit Quantity ASIN/ISBN (Product Code) Title Category The data were exported by the Amazon users from Amazon.com and shared by users with their informed consent. PII and other information not listed above were stripped from the data. This processing occurred on users' machines before sharing with researchers.

  7. H

    Data from: The Crowdsourced Replication Initiative Participant Survey

    • dataverse.harvard.edu
    Updated Nov 11, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Nate Breznau; Eike Mark Rinke; Alexander Wuttke (2024). The Crowdsourced Replication Initiative Participant Survey [Dataset]. http://doi.org/10.7910/DVN/UUP8CX
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Nov 11, 2024
    Dataset provided by
    Harvard Dataverse
    Authors
    Nate Breznau; Eike Mark Rinke; Alexander Wuttke
    License

    CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
    License information was derived automatically

    Description

    The Crowdsourced Replication Initiative (CRI) involved 204 researchers who volunteered to engage in a replication of a well-known study on immigration and social policy preferences. In this project, the participants were surveyed four times between August 20th, 2018 and January 20th, 2019. Survey questions with identifying features have been removed to protect participant anonymity and the data are available in the file cri_survey_long_public with labels or *_nolabs, without. The survey included both objective criteria, such as experience with methods and the substantive topic of the replication, and subjective criteria, such as the participants own beliefs about the hypothesis and immigration in general. In addition, they were asked questions about their time commitment, constraints they faced and some other feedback about the process of crowdsourcing. As of 2024, we provide data on the participants’ reviews of the other teams’ models. These review scores were initially not directly useable due to some problems with the 4th wave of the participant survey. The participants were given model descriptions that did not always match with the models they should have reflected. However, we have now used these paragraphs to match descriptions. We were able to match roughly 95% of all models. The new data file peer_model_dyad allows users to analyze data that are in participant-model dyad format. These data are linkable to both the participant survey here, and the CRI model specification and results data on Github (https://github.com/nbreznau/CRI). Because of matching and uneven numbers of models per team, there are some participants whose rankings apply to dozens of models and others only a few. The variable descriptions for these data are in the peer_model_dyad_codebook file. We also now provide dyadic data that matches each participant with each model specification produced by their team in df_dyad. These data contain all model specifications and the AME (Average Marginal Effect) produced by that model.

  8. d

    Commodifying infrastructure spatial dynamics with crowdsourced smartphone...

    • datadryad.org
    • data.niaid.nih.gov
    zip
    Updated Sep 9, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Liam Cronin; Soheil Sadeghi; Thomas Matarazzo; Sebastiano Milardo; Iman Dabbaghchian; Paolo Santi; Umberto Fugiglando; Shamim Pakzad (2024). Commodifying infrastructure spatial dynamics with crowdsourced smartphone data [Dataset]. http://doi.org/10.5061/dryad.15dv41p49
    Explore at:
    zipAvailable download formats
    Dataset updated
    Sep 9, 2024
    Dataset provided by
    Dryad
    Authors
    Liam Cronin; Soheil Sadeghi; Thomas Matarazzo; Sebastiano Milardo; Iman Dabbaghchian; Paolo Santi; Umberto Fugiglando; Shamim Pakzad
    Description

    This data set was collected from various sources: the research team, ANAS employees, and Uber drivers. The method for data collection and data processing for each dataset can be found in the related works.

  9. Z

    Data from: Mapping Cropland in Ethiopia Using Crowdsourcing

    • data.niaid.nih.gov
    • zenodo.org
    • +1more
    Updated Jul 16, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Kraxner, Florian (2024). Mapping Cropland in Ethiopia Using Crowdsourcing [Dataset]. https://data.niaid.nih.gov/resources?id=zenodo_6597347
    Explore at:
    Dataset updated
    Jul 16, 2024
    Dataset provided by
    Fritz, Steffen
    Obersteiner, Michael
    Mill, Nitashree
    Baruah, Ujjal Deka
    See, Linda
    McCallum, Ian
    Kraxner, Florian
    Perger, Christoph
    Kalita, Nripen Ram
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Area covered
    Ethiopia
    Description

    The spatial distribution of cropland is an important input to many applications including food security monitoring and economic land use modeling. Global land cover maps derived from remote sensing are one source of cropland but they are currently not accurate enough in the cropland domain to meet the needs of the user community. Moreover, when compared with one another, these land cover products show large areas of spatial disagreement, which makes the choice very difficult regarding which land cover product to use. This paper takes an entirely different approach to mapping cropland, using crowdsourcing of Google Earth imagery via tools in Geo-Wiki. Using sample data generated by a crowdsourcing campaign for the collection of the degree of cultivation and settlement in Ethiopia, a cropland map was created using simple inverse distance weighted interpolation. The map was validated using data from the GOFC-GOLD validation portal and an independent crowdsourced dataset from Geo-Wiki. The results show that the crowdsourced cropland map for Ethiopia has a higher overall accuracy than the individual global land cover products for this country. Such an approach has great potential for mapping cropland in other countries where such data do not currently exist. Not only is the approach inexpensive but the data can be collected over a very short period of time using an existing network of volunteers.

  10. n

    2019 Crowdsourced Photos Public Feature Layer View

    • prep-response-portal.napsgfoundation.org
    • cest-cusec.hub.arcgis.com
    • +3more
    Updated Jul 11, 2019
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    NAPSG Foundation (2019). 2019 Crowdsourced Photos Public Feature Layer View [Dataset]. https://prep-response-portal.napsgfoundation.org/datasets/2019-crowdsourced-photos-public-feature-layer-view/api
    Explore at:
    Dataset updated
    Jul 11, 2019
    Dataset authored and provided by
    NAPSG Foundation
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Area covered
    Description

    Purpose: This is the 2019 Hurricanes Crowdsourced Photos Public Feature Layer View. This is a live publicly accessible layer for the Crowdsource Story Map accessible here: This layer cannot be edited, it is view only. ShareHidden Field: 0 = Needs Review, 1 = Already Reviewed, 2 = Hidden (not available in this public view).Audience: GIS Staff and Technologists who would like to add this layer to their own web maps and apps. If you need access to this layer in other formats, see the Open Data link. Please send us an email at triage@publicsafetygis.org to tell us if you are going to use this layer and if you have any questions or need assistance with this layer.Need to download the photos? See this technical support article.

  11. Data from: A global dataset of crowdsourced land cover and land use...

    • doi.pangaea.de
    • search.dataone.org
    html, tsv
    Updated Dec 21, 2016
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Steffen Fritz; Linda See; Christoph Perger; Ian McCallum; Christian Schill; Dmitry Schepaschenko; Martina Duerauer; Mathias Karner; Christopher Dresel; Juan-Carlos Laso-Bayas; Myroslava Lesiv; Inian Moorthy; Carl F Salk; Olha Danylo; Tobias Sturn; Franziska Albrecht; Liangzhi You; Florian Kraxner; Michael Obersteiner (2016). A global dataset of crowdsourced land cover and land use reference data (2011-2012) [Dataset]. http://doi.org/10.1594/PANGAEA.869680
    Explore at:
    html, tsvAvailable download formats
    Dataset updated
    Dec 21, 2016
    Dataset provided by
    PANGAEA
    Authors
    Steffen Fritz; Linda See; Christoph Perger; Ian McCallum; Christian Schill; Dmitry Schepaschenko; Martina Duerauer; Mathias Karner; Christopher Dresel; Juan-Carlos Laso-Bayas; Myroslava Lesiv; Inian Moorthy; Carl F Salk; Olha Danylo; Tobias Sturn; Franziska Albrecht; Liangzhi You; Florian Kraxner; Michael Obersteiner
    License

    Attribution 3.0 (CC BY 3.0)https://creativecommons.org/licenses/by/3.0/
    License information was derived automatically

    Time period covered
    Jan 1, 1911 - Aug 27, 2095
    Area covered
    Variables measured
    Code, Size, LATITUDE, DATE/TIME, LONGITUDE, Confidence, Percentage, Resolution, Human impact, Identification, and 1 more
    Description

    This dataset is about: A global dataset of crowdsourced land cover and land use reference data (2011-2012). Please consult parent dataset @ https://doi.org/10.1594/PANGAEA.869682 for more information.

  12. CrowdMag Visualization Web Map

    • noaa.hub.arcgis.com
    Updated May 15, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    NOAA GeoPlatform (2023). CrowdMag Visualization Web Map [Dataset]. https://noaa.hub.arcgis.com/maps/f8e24dd400c94d4e8275417f2e8a2070
    Explore at:
    Dataset updated
    May 15, 2023
    Dataset provided by
    National Oceanic and Atmospheric Administrationhttp://www.noaa.gov/
    Authors
    NOAA GeoPlatform
    Area covered
    Description

    This web map is a component of the CrowdMag Visualization App.NOAA's CrowdMag is a crowdsourced data collection project that uses a mobile app to collect geomagnetic data from the magnetometers that modern smartphones use as part of their navigation systems. NCEI collects these data from citizen scientists around the world and provides quality control services before making them available through a series of aggregated maps and charts. These data have the potential to provide a high resolution alternative to geomagnetic satellite data, as well as near real-time information about changes in the magnetic field.This map shows data collected from phones around the world! Displayed are the Crowdsourced magnetic data within a tolerance level of prediction by World Magnetic Model. We have added some uncertainty to each data point shown to ensure the privacy of our contributors. The data points are grouped together (or "aggregated") into small areas , and we display the median data value across all the readings for each point.

    This map is updated every day. Layers are available for Median Intensity, Median Horizontal Component (Y), and Median Vertical Component (Z).
    
    
    Use the time slider to select the date range. Select the different layers under the "Crowdmag Observations" menu. View a color scale using the legend tool. Zoom to your location using the "Find my Location" tool. Click or tap on a data point to view a popup containing more information.
    
  13. f

    DataSheet_1_Millimeter-Level Plant Disease Detection From Aerial Photographs...

    • frontiersin.figshare.com
    docx
    Updated May 30, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Tyr Wiesner-Hanks; Harvey Wu; Ethan Stewart; Chad DeChant; Nicholas Kaczmar; Hod Lipson; Michael A. Gore; Rebecca J. Nelson (2023). DataSheet_1_Millimeter-Level Plant Disease Detection From Aerial Photographs via Deep Learning and Crowdsourced Data.docx [Dataset]. http://doi.org/10.3389/fpls.2019.01550.s001
    Explore at:
    docxAvailable download formats
    Dataset updated
    May 30, 2023
    Dataset provided by
    Frontiers
    Authors
    Tyr Wiesner-Hanks; Harvey Wu; Ethan Stewart; Chad DeChant; Nicholas Kaczmar; Hod Lipson; Michael A. Gore; Rebecca J. Nelson
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Computer vision models that can recognize plant diseases in the field would be valuable tools for disease management and resistance breeding. Generating enough data to train these models is difficult, however, since only trained experts can accurately identify symptoms. In this study, we describe and implement a two-step method for generating a large amount of high-quality training data with minimal expert input. First, experts located symptoms of northern leaf blight (NLB) in field images taken by unmanned aerial vehicles (UAVs), annotating them quickly at low resolution. Second, non-experts were asked to draw polygons around the identified diseased areas, producing high-resolution ground truths that were automatically screened based on agreement between multiple workers. We then used these crowdsourced data to train a convolutional neural network (CNN), feeding the output into a conditional random field (CRF) to segment images into lesion and non-lesion regions with accuracy of 0.9979 and F1 score of 0.7153. The CNN trained on crowdsourced data showed greatly improved spatial resolution compared to one trained on expert-generated data, despite using only one fifth as many expert annotations. The final model was able to accurately delineate lesions down to the millimeter level from UAV-collected images, the finest scale of aerial plant disease detection achieved to date. The two-step approach to generating training data is a promising method to streamline deep learning approaches for plant disease detection, and for complex plant phenotyping tasks in general.

  14. n

    Data from: Crowdsourcing the identification of organisms: a case-study of...

    • data.niaid.nih.gov
    • zenodo.org
    • +1more
    zip
    Updated Jan 21, 2016
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Jonathan Silvertown; Martin Harvey; Richard Greenwood; Mike Dodd; Jon Rosewell; Tony Rebelo; Janice Ansine; Kevin McConway (2016). Crowdsourcing the identification of organisms: a case-study of iSpot [Dataset]. http://doi.org/10.5061/dryad.r0005
    Explore at:
    zipAvailable download formats
    Dataset updated
    Jan 21, 2016
    Dataset provided by
    The Open University
    University of Edinburgh
    South African National Biodiversity Institute
    Authors
    Jonathan Silvertown; Martin Harvey; Richard Greenwood; Mike Dodd; Jon Rosewell; Tony Rebelo; Janice Ansine; Kevin McConway
    License

    https://spdx.org/licenses/CC0-1.0.htmlhttps://spdx.org/licenses/CC0-1.0.html

    Area covered
    United Kingdom, Southern Africa
    Description

    Accurate species identification is fundamental to biodiversity science, but the natural history skills required for this are neglected in formal education at all levels. In this paper we describe how the web application ispotnature.org and its sister site ispot.org.za (collectively, "iSpot") are helping to solve this problem by combining learning technology with crowdsourcing to connect beginners with experts. Over 94% of observations submitted to iSpot receive a determination. To date (2014), iSpot has crowdsourced the identification of 30,000 taxa (>80% at species level) in > 390,000 observations with a global community numbering > 42,000 registered participants. More than half the observations on ispotnature.org were named within an hour of submission. iSpot uses a unique, 9-dimensional reputation system to motivate and reward participants and to verify determinations. Taxon-specific reputation points are earned when a parti! cipant proposes an identification that achieves agreement from other participants, weighted by the agreers' own reputation scores for the taxon. This system is able to discriminate effectively between competing determinations when two or more are proposed for the same observation. In 57% of such cases the reputation system improved the accuracy of the determination, while in the remainder it either improved precision (e.g. by adding a species name to a genus) or revealed false precision, for example where a determination to species level was not supported by the available evidence. We propose that the success of iSpot arises from the structure of its social network which efficiently connects beginners and experts, overcoming the social as well as geographic barriers that normally separate the two.

  15. D

    Data outputs of the SKILLNET letter metadata crowdsourcing project CEMROL:...

    • dataverse.nl
    csv, json, txt, xlsx
    Updated Jan 7, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    SKILLNET project; SKILLNET project (2023). Data outputs of the SKILLNET letter metadata crowdsourcing project CEMROL: raw data [Dataset]. http://doi.org/10.34894/CQSNXV
    Explore at:
    json(107447), csv(33309), xlsx(6424), xlsx(10579), json(54554), csv(177), json(1357812), json(777223), xlsx(6687), json(929024), json(1916467), csv(30508), json(993019), xlsx(39788), json(362254), json(77359), json(69069), json(1037783), json(237003), xlsx(5427), csv(36282), json(585293), json(901297), txt(55690), csv(10076), xlsx(5426), xlsx(6265), json(921924), xlsx(6418), csv(17723), csv(14124), json(562178), xlsx(6705), xlsx(6084), json(1312537), csv(189101), json(465994), xlsx(18297), json(1109720), xlsx(15404), json(16049), xlsx(6210), xlsx(7866), json(2097019), json(1392306), json(273170), json(840855), xlsx(31662), csv(119057), csv(69114), json(761764), csv(151818), xlsx(19905), json(1170746), csv(41163), json(483576), csv(14929), csv(32361), csv(238357), json(1367233), xlsx(6310), json(758222), csv(67555), json(723235), xlsx(6604), csv(1298), json(1389471), csv(8456), xlsx(6500), json(725685), csv(26254), xlsx(6922), json(1514054), csv(24726), xlsx(13137), json(421578), xlsx(42294), json(148099), csv(36537), xlsx(32710), json(228561), csv(11645), csv(219515), csv(96973), xlsx(25279), xlsx(7884), csv(62491), csv(166059), json(913047), xlsx(14896), csv(45859), json(2889444), csv(29115), json(1949825), csv(77721), xlsx(6682), csv(45560), json(261660), xlsx(7396), csv(2294), json(359585), csv(45452), json(802756), json(574896), csv(32009), json(4366069), xlsx(6976), json(598402), json(71331), json(541716), xlsx(6561), xlsx(6029), xlsx(14909), json(1394362), xlsx(52011), xlsx(38603), json(1726265), xlsx(6329), xlsx(21894), csv(47097), xlsx(6246), csv(65146), json(1293233), json(1605224), json(821698), csv(130937), xlsx(18665), xlsx(19363), json(1473803), json(109853), csv(129360), xlsx(23351), xlsx(6802), xlsx(13874), json(645338), csv(53204), json(641830), csv(120564), csv(25051), csv(9983), json(103176), json(651677), xlsx(27102), json(715154), json(761851), json(383546), xlsx(9275), csv(45889), json(819552), json(1772418), json(564070), csv(19334), json(227361), csv(48215), xlsx(6088), xlsx(23195), xlsx(6666), xlsx(6481), xlsx(6031), json(300784), csv(134477), json(782115), csv(7758), csv(91665), csv(80268), json(141662), json(20337), json(1528093), xlsx(9186), json(54082), json(1466262), xlsx(16801), csv(115988), csv(118156), xlsx(8945), json(941056), xlsx(10464), json(2733777), xlsx(6563), csv(18371), json(1928208), csv(90921), csv(6021), xlsx(8333), xlsx(6493), xlsx(6307), json(1233216), json(1760412), json(4466010), json(1593222), csv(37604), json(443334), xlsx(7739), json(753255), json(737976), xlsx(10482), csv(30969), json(161661), xlsx(6325), xlsx(10338), json(161855), xlsx(6235), json(1352197), json(205302), xlsx(7695), json(137652), xlsx(6703), xlsx(6292), xlsx(6664), json(129348), json(1559991), json(152220), xlsx(6348), xlsx(25312), xlsx(7105), json(656041), xlsx(7680), json(1189557), json(4023908), xlsx(6128), csv(4857), xlsx(6621), csv(24127), json(1105413), xlsx(6702), xlsx(17538), xlsx(18553), csv(92191), json(1507541), json(1003549), xlsx(6842), xlsx(6040), csv(700), csv(32819), csv(68169), csv(130605), csv(12698), json(359308), csv(55815), xlsx(6854), json(86683), xlsx(23026), csv(14553), json(1055536), xlsx(6337), json(110242), json(621250), json(909536), json(983467), json(3351530), csv(40379), json(76071), json(1244185), json(930928), json(484056), csv(67109), json(730270), csv(35011), xlsx(40834), csv(36024), xlsx(32467), json(1555874), json(860618), xlsx(6564), csv(76068), json(1405371), csv(34429), xlsx(21116), csv(28146), xlsx(30604), json(981338), csv(116428), json(1086517), json(488742), csv(127562), xlsx(18944), csv(112917), json(412661), csv(106589), xlsx(11127), xlsx(6412), json(1126639), csv(9066), csv(8270), xlsx(34613), json(67336), xlsx(10003), xlsx(7560), json(373939), csv(77604), json(4462655), xlsx(6614), csv(91426), json(2836743), json(2293219), csv(79155), json(211258), csv(49788), csv(104447), csv(54636), json(14531), json(1688234), xlsx(13016), csv(45016), csv(145332), csv(46425), csv(71064), xlsx(19131), json(488854), json(784562), json(291907), json(1649271), json(186260), xlsx(30324), csv(162465), json(958731), csv(14166), xlsx(6225), json(297812), csv(18006), xlsx(8785), csv(28151), json(621681), json(1376683), json(1376794), xlsx(6332), json(348119), json(432922), csv(184379), json(1404620), xlsx(6787), xlsx(51351), json(1251673), json(751442), json(387732), json(995329), csv(134580), json(1110949), csv(14363), xlsx(7059), json(706368), json(1223368), csv(109680), json(857104), xlsx(8559), csv(73114), json(834431), xlsx(6046), json(127960), xlsx(24825), xlsx(6624), xlsx(6253), xlsx(6276), csv(71577), csv(53708), json(38197), csv(137304), json(718777), json(7105), json(232715), csv(97906), json(48771), csv(84168), json(2925628), json(515176), json(38057), csv(129310), csv(53347), csv(39458), xlsx(6541), xlsx(13422), json(163586), csv(69384), csv(3844), json(60809), csv(17244), json(2010392), json(608095), xlsx(19541), csv(25708), csv(38853), json(742570), xlsx(6830), xlsx(6330), xlsx(25486), json(1014759), csv(26222), json(146453), json(1088845), csv(109313), csv(46806), json(1644498), xlsx(6362), json(2377654), json(881680), xlsx(6413), xlsx(6257), csv(136652), json(1625414), json(154237), json(62233), json(2132226), json(92363), json(789358), csv(34852), csv(94344), xlsx(25518), xlsx(6000), csv(50766), csv(34144), csv(39626), json(1690454), csv(117848), csv(61140), json(831435), json(1114236), xlsx(26589), json(308048), xlsx(8005), json(1423398), json(108072), json(1817285), csv(17987), xlsx(6354), xlsx(27970), csv(107012), json(490197)Available download formats
    Dataset updated
    Jan 7, 2023
    Dataset provided by
    DataverseNL
    Authors
    SKILLNET project; SKILLNET project
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    This dataset contains all the exported data contributed by volunteers during the SKILLNET project via the crowdsourcing platform CERMROL: https://cemrol.hum.uu.nl/#/ (more information also available here: https://skillnet.nl/cemrol/). Via this platform, volunteers helped in selecting (marking) and adding basic metadata to every letter in a series of selected letter editions. The raw, unprocessed files, are provided in this dataset. A selection of this raw data was manually cleaned by a student assistant, it is made available as a separate dataset: https://doi.org/10.34894/NJKUF0 Note: The latest version of the CEMROL's code developed by Sheean Spoel from the Digital Humanities Lab at Utrecht University is publicly available at https://github.com/UUDigitalHumanitieslab/scribeAPI and also deposited in Zenodo. The historical date entry is public at https://github.com/UUDigitalHumanitieslab/historical-dates and https://github.com/UUDigitalHumanitieslab/historical-dates-ui. Sheean indicates that the date calculator itself is not open source (the user interface is). The calculations are based on Axel Findling's Roman date converter and Nikolaus A. Bär's easter date calculator.

  16. f

    Time periods for testing and non-testing.

    • figshare.com
    xls
    Updated Jun 3, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Vinayak Dixit; Divya Jayakumar Nair; Sai Chand; Michael W. Levin (2023). Time periods for testing and non-testing. [Dataset]. http://doi.org/10.1371/journal.pone.0230598.t001
    Explore at:
    xlsAvailable download formats
    Dataset updated
    Jun 3, 2023
    Dataset provided by
    PLOS ONE
    Authors
    Vinayak Dixit; Divya Jayakumar Nair; Sai Chand; Michael W. Levin
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Time periods for testing and non-testing.

  17. C

    Crowdsourced Testing Industry Report

    • datainsightsmarket.com
    doc, pdf, ppt
    Updated Jan 15, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Data Insights Market (2025). Crowdsourced Testing Industry Report [Dataset]. https://www.datainsightsmarket.com/reports/crowdsourced-testing-industry-12750
    Explore at:
    ppt, pdf, docAvailable download formats
    Dataset updated
    Jan 15, 2025
    Dataset authored and provided by
    Data Insights Market
    License

    https://www.datainsightsmarket.com/privacy-policyhttps://www.datainsightsmarket.com/privacy-policy

    Time period covered
    2025 - 2033
    Area covered
    Global
    Variables measured
    Market Size
    Description

    The size of the Crowdsourced Testing Industry market was valued at USD XX Million in 2023 and is projected to reach USD XXX Million by 2032, with an expected CAGR of 10.50% during the forecast period.A crowdsourced testing technique is used through the testing activity with an independent network of test operators. Engaging a broad crowd to perform applications on several devices and systems on which they may operate helps assess the usability, functionality, and performance of those applications. Utilizing a crowd will give more exposure to organizations about testing, enable them to shorten their cycles, and minimize their testing expenses.Crowdsourced testing is most useful for organizations that want to enhance the user experience, detect bugs early in the development cycle, and ensure cross-platform and cross-device compatibility. Recent developments include: January 2022: Testlio, the pioneer of networked testing, introduced fused testing. This new methodology combined expert manual testing with the efficiency of test automation, assisting engineering and product leaders in meeting increased customer demands for exceptional digital experiences., February 2021: Applause App Quality Inc., a crowdsourced testing solutions company, announced the launch of its Excellent Product Platform, which would provide customers with enterprise-grade software-as-a-service infrastructure, digital testing solutions, and access to the world's largest community of digital experts.. Key drivers for this market are: Rise in the Number of Operating Systems, Devices, and Applications, Demand for Scaling Quality Assurance of Software to Magnify Customer Experience. Potential restraints include: Concerns Over Data Privacy Regulations Over the Globe. Notable trends are: Large Enterprises to Constitute a Significant Market Size.

  18. Reply & Supply: Experiment 2 (Synonym Proposal Task) Crowdsourcing Data

    • figshare.com
    txt
    Updated May 31, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    James Bagrow (2023). Reply & Supply: Experiment 2 (Synonym Proposal Task) Crowdsourcing Data [Dataset]. http://doi.org/10.6084/m9.figshare.5183710.v1
    Explore at:
    txtAvailable download formats
    Dataset updated
    May 31, 2023
    Dataset provided by
    figshare
    Figsharehttp://figshare.com/
    Authors
    James Bagrow
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    These data support the synonymy proposal task (SPT) experiment described in the paper "Reply & Supply: Efficient crowdsourcing when workers do more than answer questions", https://arxiv.org/abs/1611.00954.The data consist of two CSV files, one describing the questions built by the crowd as they work, the other recording the responses of workers when presented with questions. A "question id" (qid) field links these data. The IDs of Mechanical Turk workers were deidentified. The task interface is described in Fig. 2 of the paper.Three question sampling algorithms were tested in the experiment. These are recorded in the algorithm. Note that workers may participate in multiple algorithms, and that qid is only unique for a given algorithm (qid 1 under the random algorithm and qid 1 under the binomial algorithm are not the same question).

  19. d

    Data from: Using molecular and crowd-sourcing methods to assess breeding...

    • datadryad.org
    • data.niaid.nih.gov
    • +2more
    zip
    Updated Jul 27, 2020
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Lowell Mills; Jeremy D. Wilson; Anke Lange; Karen Moore; Barry Henwood; Hazel Knipe; Dominique Chaput; Charles Tyler (2020). Using molecular and crowd-sourcing methods to assess breeding ground diet of a migratory brood parasite of conservation concern [Dataset]. http://doi.org/10.5061/dryad.v6wwpzgsp
    Explore at:
    zipAvailable download formats
    Dataset updated
    Jul 27, 2020
    Dataset provided by
    Dryad
    Authors
    Lowell Mills; Jeremy D. Wilson; Anke Lange; Karen Moore; Barry Henwood; Hazel Knipe; Dominique Chaput; Charles Tyler
    Time period covered
    2020
    Description

    NCBI Sequence Read Archive under BioProject number PRJNA606798

  20. B

    Replication Data for: Crowd sourcing remote comparative lameness assessments...

    • borealisdata.ca
    • open.library.ubc.ca
    • +1more
    Updated May 10, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Kehan Sheng; Borbala Foris; Marina von Keyserlingk; John Gardenier; Cameron Clark; Daniel Weary (2024). Replication Data for: Crowd sourcing remote comparative lameness assessments for dairy cattle [Dataset]. http://doi.org/10.5683/SP3/3EEQEQ
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    May 10, 2024
    Dataset provided by
    Borealis
    Authors
    Kehan Sheng; Borbala Foris; Marina von Keyserlingk; John Gardenier; Cameron Clark; Daniel Weary
    License

    CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
    License information was derived automatically

    Dataset funded by
    NSERC Industrial Research Chair
    Description

    Lameness assessments are rarely conducted routinely on dairy farms and when completed typically underestimate lameness prevalence, hampering early diagnosis and treatment. A well-known feature of many perceptual tasks is that relative assessments are more accurate than absolute assessments, suggesting that creating methods that allow for the relative scoring of ‘which cow is more lame’ will allow for reliable lameness assessments. Here we developed and tested a remote comparative lameness assessment method: we recruited non-experienced crowd workers via an online platform and asked them to watch two videos side-by-side, each showing a cow walking, and to identify which cow was more lame and by how much (on a scale of -3 to 3). We created 11 tasks, each with 10 video pairs for comparison, and recruited 50 workers per task. All tasks were also completed by 5 experienced cattle lameness assessors. We evaluated data filtering and clustering methods based on worker responses and determined the agreement among workers, among experienced assessors, and between these groups. A moderate to high interobserver reliability was observed (intraclass correlation coefficient, ICC=0.46 to 0.77) for crowd workers and agreement was high among the experienced assessors (ICC=0.87). Average crowd worker responses showed excellent agreement with the average of experienced assessor responses (ICC= 0.89 to 0.91), regardless of data processing method. To investigate if we could use fewer workers per task while still retaining high agreement with experienced assessors, we randomly sub-sampled 2 to 43 (1 less than the minimum number of workers retained per task after data cleaning) workers from each task. The agreement with experienced assessors increased substantially as we increased the number of workers from 2 to 10, but little increase was observed after 10 or more workers were used (ICC>0.80). The proposed method provides a fast and cost-effective way to assess lameness in commercial herds. In addition, this method allows for large-scale data collection useful for training computer vision algorithms that could be used to automate lameness assessments on farm.

Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
Barbara Strobl; Simon Etter; Ilja van Meerveld; Jan Seibert (2023). Reasons given for a report as a percentage of the overall number of reports. [Dataset]. http://doi.org/10.1371/journal.pone.0222579.t001

Reasons given for a report as a percentage of the overall number of reports.

Related Article
Explore at:
xlsAvailable download formats
Dataset updated
Jun 4, 2023
Dataset provided by
PLOS ONE
Authors
Barbara Strobl; Simon Etter; Ilja van Meerveld; Jan Seibert
License

Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically

Description

Reasons given for a report as a percentage of the overall number of reports.

Search
Clear search
Close search
Google apps
Main menu