46 datasets found
  1. d

    MetOp-C ASCAT Level 2 25.0km Ocean Surface Wind Vectors in Full Orbit Swath

    • catalog.data.gov
    • datasets.ai
    • +6more
    Updated Apr 10, 2025
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    NASA/JPL/PODAAC (2025). MetOp-C ASCAT Level 2 25.0km Ocean Surface Wind Vectors in Full Orbit Swath [Dataset]. https://catalog.data.gov/dataset/metop-c-ascat-level-2-25-0km-ocean-surface-wind-vectors-in-full-orbit-swath-d26e1
    Explore at:
    Dataset updated
    Apr 10, 2025
    Dataset provided by
    NASA/JPL/PODAAC
    Description

    This dataset contains operational near-real-time Level 2 ocean surface wind vector retrievals from the Advanced Scatterometer (ASCAT) on MetOp-C at 25 km sampling resolution (note: the effective resolution is 50 km). It is a product of the European Organization for the Exploitation of Meteorological Satellites (EUMETSAT) Ocean and Sea Ice Satellite Application Facility (OSI SAF) provided through the Royal Netherlands Meteorological Institute (KNMI). The wind vector retrievals are currently processed using the CMOD7.n geophysical model function using a Hamming filter to spatially average the Sigma-0 data in the ASCAT L1B data. Each file is provided in netCDF version 3 format, and contains one full orbit derived from 3-minute orbit granules. Latency is approximately 2 hours from the latest measurement. The beginning of the orbit is defined by the first wind vector cell measurement within the first 3-minute orbit granule that starts north of the Equator in the ascending node. ASCAT is a C-band dual swath fan beam radar scatterometer providing two independent swaths of backscatter retrievals in sun-synchronous polar orbit aboard the MetOp-C platform. For more information about the MetOp-C platform and mission, please refer to: https://www.eumetsat.int/our-satellites/metop-series . For more timely announcements, users are encouraged to register with the KNMI scatterometer email list: scat@knmi.nl. Users are also highly advised to check the dataset user guide periodically for updates and new information on known problems and issues. All intellectual property rights of the OSI SAF products belong to EUMETSAT. The use of these products is granted to every interested user, free of charge. If you wish to use these products, EUMETSAT's copyright credit must be shown by displaying the words "copyright (year) EUMETSAT" on each of the products used.

  2. Z

    A large database of motor imagery EEG signals and users' demographic,...

    • data.niaid.nih.gov
    • zenodo.org
    Updated Sep 13, 2023
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Rimbert Sébastien (2023). A large database of motor imagery EEG signals and users' demographic, personality and cognitive profile information for Brain-Computer Interface research [Dataset]. https://data.niaid.nih.gov/resources?id=zenodo_7516450
    Explore at:
    Dataset updated
    Sep 13, 2023
    Dataset provided by
    Pillette Léa
    Lotte Fabien
    Rimbert Sébastien
    Roc Aline
    Dreyer Pauline
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Context : We share a large database containing electroencephalographic signals from 87 human participants, with more than 20,800 trials in total representing about 70 hours of recording. It was collected during brain-computer interface (BCI) experiments and organized into 3 datasets (A, B, and C) that were all recorded following the same protocol: right and left hand motor imagery (MI) tasks during one single day session. It includes the performance of the associated BCI users, detailed information about the demographics, personality and cognitive user’s profile, and the experimental instructions and codes (executed in the open-source platform OpenViBE). Such database could prove useful for various studies, including but not limited to: 1) studying the relationships between BCI users' profiles and their BCI performances, 2) studying how EEG signals properties varies for different users' profiles and MI tasks, 3) using the large number of participants to design cross-user BCI machine learning algorithms or 4) incorporating users' profile information into the design of EEG signal classification algorithms.

    Sixty participants (Dataset A) performed the first experiment, designed in order to investigated the impact of experimenters' and users' gender on MI-BCI user training outcomes, i.e., users performance and experience, (Pillette & al). Twenty one participants (Dataset B) performed the second one, designed to examined the relationship between users' online performance (i.e., classification accuracy) and the characteristics of the chosen user-specific Most Discriminant Frequency Band (MDFB) (Benaroch & al). The only difference between the two experiments lies in the algorithm used to select the MDFB. Dataset C contains 6 additional participants who completed one of the two experiments described above. Physiological signals were measured using a g.USBAmp (g.tec, Austria), sampled at 512 Hz, and processed online using OpenViBE 2.1.0 (Dataset A) & OpenVIBE 2.2.0 (Dataset B). For Dataset C, participants C83 and C85 were collected with OpenViBE 2.1.0 and the remaining 4 participants with OpenViBE 2.2.0. Experiments were recorded at Inria Bordeaux sud-ouest, France.

    Duration : Each participant's folder is composed of approximately 48 minutes EEG recording. Meaning six 7-minutes runs and a 6-minutes baseline.

    Documents Instructions: checklist read by experimenters during the experiments. Questionnaires: the Mental Rotation test used, the translation of 4 questionnaires, notably the Demographic and Social information, the Pre and Post-session questionnaires, and the Index of Learning style. English and french version Performance: The online OpenViBE BCI classification performances obtained by each participant are provided for each run, as well as answers to all questionnaires Scenarios/scripts : set of OpenViBE scenarios used to perform each of the steps of the MI-BCI protocol, e.g., acquire training data, calibrate the classifier or run the online MI-BCI

    Database : raw signals Dataset A : N=60 participants Dataset B : N=21 participants Dataset C : N=6 participants

  3. Data from: Use characteristics, visitor preferences, and conflict between...

    • catalog.data.gov
    • healthdata.gov
    • +6more
    Updated Apr 21, 2025
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    U.S. Forest Service (2025). Use characteristics, visitor preferences, and conflict between horse users and hikers in the Charles C. Deam Wilderness Area: 1990-1991 visitor survey data [Dataset]. https://catalog.data.gov/dataset/use-characteristics-visitor-preferences-and-conflict-between-horse-users-and-hikers-in-the-0e712
    Explore at:
    Dataset updated
    Apr 21, 2025
    Dataset provided by
    U.S. Department of Agriculture Forest Servicehttp://fs.fed.us/
    Description

    The Charles C. Deam Wilderness area is located on the Hoosier National Forest in Indiana. There has been historic use of this area by both hikers and horse riders, however there was substantial concern about the interaction and conflict between these two groups in this wilderness area. Between the summers of 1990 and 1991 a mailback questionnaire was sent to people visiting the Charles C. Deam Wilderness are for recreational purposes to investigate visitor use characterstics and preferences about their wilderness experience. Data include visitor activity, visitor characteristics, interaction with other groups, opinions regarding management policies, current wilderness conditions, preferred wilderness conditions, items influencing visitor quality, as well as user perception of similarities and differences between hikers and horse users.

  4. R

    Identification Dataset

    • universe.roboflow.com
    zip
    Updated Apr 24, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Roderick Cai (2022). Identification Dataset [Dataset]. https://universe.roboflow.com/roderick-cai/identification-5lyed/dataset/15
    Explore at:
    zipAvailable download formats
    Dataset updated
    Apr 24, 2022
    Dataset authored and provided by
    Roderick Cai
    License

    CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
    License information was derived automatically

    Variables measured
    Identification Bounding Boxes
    Description

    Here are a few use cases for this project:

    1. Smart Security Surveillance: This model can be utilized to monitor security cameras in real-time and identify people, cars, bicycles, and other objects, enabling it to track suspicious behaviors or misplaced items within the scene. It could also detect unauthorized personnel or unrecognized items entering restricted areas.

    2. Traffic Monitoring and Management: The "identification" model can be employed to analyze traffic in a city or on highways, identifying vehicle types (cars, bicycles) and estimating traffic density. This data can be relayed to traffic control centers to optimize traffic signals, enforce traffic laws, and reduce congestion.

    3. Automatic Access Control: Integrating the model into parking facilities and gated communities can help automate entry access, identifying and allowing authorized vehicles and personnel to enter while keeping unregistered ones outside. It can also be used to control bicycle parking areas, umbrellas in public spaces, and potted plants in shared gardens or office spaces.

    4. Object-Based Image Retrieval: The model can be utilized to create a labeled database of images, allowing users to query images based on visual content for various purposes. For example, identifying and classifying photographs based on the presence of people, cars, bicycles, umbrella, or potted plants for easier retrieval.

    5. Assistance for Visually Impaired Individuals: Integrating the "identification" model into an assistance device or smartphone app can help visually impaired users better navigate their surroundings. The model can provide audio cues or descriptions of the detected objects (people, cars, bicycles, umbrellas, potted plants) to help users avoid obstacles and better understand their environment.

  5. The Canada Trademarks Dataset

    • zenodo.org
    pdf, zip
    Updated Jul 19, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Jeremy Sheff; Jeremy Sheff (2024). The Canada Trademarks Dataset [Dataset]. http://doi.org/10.5281/zenodo.4999655
    Explore at:
    zip, pdfAvailable download formats
    Dataset updated
    Jul 19, 2024
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Jeremy Sheff; Jeremy Sheff
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Area covered
    Canada
    Description

    The Canada Trademarks Dataset

    18 Journal of Empirical Legal Studies 908 (2021), prepublication draft available at https://papers.ssrn.com/abstract=3782655, published version available at https://onlinelibrary.wiley.com/share/author/CHG3HC6GTFMMRU8UJFRR?target=10.1111/jels.12303

    Dataset Selection and Arrangement (c) 2021 Jeremy Sheff

    Python and Stata Scripts (c) 2021 Jeremy Sheff

    Contains data licensed by Her Majesty the Queen in right of Canada, as represented by the Minister of Industry, the minister responsible for the administration of the Canadian Intellectual Property Office.

    This individual-application-level dataset includes records of all applications for registered trademarks in Canada since approximately 1980, and of many preserved applications and registrations dating back to the beginning of Canada’s trademark registry in 1865, totaling over 1.6 million application records. It includes comprehensive bibliographic and lifecycle data; trademark characteristics; goods and services claims; identification of applicants, attorneys, and other interested parties (including address data); detailed prosecution history event data; and data on application, registration, and use claims in countries other than Canada. The dataset has been constructed from public records made available by the Canadian Intellectual Property Office. Both the dataset and the code used to build and analyze it are presented for public use on open-access terms.

    Scripts are licensed for reuse subject to the Creative Commons Attribution License 4.0 (CC-BY-4.0), https://creativecommons.org/licenses/by/4.0/. Data files are licensed for reuse subject to the Creative Commons Attribution License 4.0 (CC-BY-4.0), https://creativecommons.org/licenses/by/4.0/, and also subject to additional conditions imposed by the Canadian Intellectual Property Office (CIPO) as described below.

    Terms of Use:

    As per the terms of use of CIPO's government data, all users are required to include the above-quoted attribution to CIPO in any reproductions of this dataset. They are further required to cease using any record within the datasets that has been modified by CIPO and for which CIPO has issued a notice on its website in accordance with its Terms and Conditions, and to use the datasets in compliance with applicable laws. These requirements are in addition to the terms of the CC-BY-4.0 license, which require attribution to the author (among other terms). For further information on CIPO’s terms and conditions, see https://www.ic.gc.ca/eic/site/cipointernet-internetopic.nsf/eng/wr01935.html. For further information on the CC-BY-4.0 license, see https://creativecommons.org/licenses/by/4.0/.

    The following attribution statement, if included by users of this dataset, is satisfactory to the author, but the author makes no representations as to whether it may be satisfactory to CIPO:

    The Canada Trademarks Dataset is (c) 2021 by Jeremy Sheff and licensed under a CC-BY-4.0 license, subject to additional terms imposed by the Canadian Intellectual Property Office. It contains data licensed by Her Majesty the Queen in right of Canada, as represented by the Minister of Industry, the minister responsible for the administration of the Canadian Intellectual Property Office. For further information, see https://creativecommons.org/licenses/by/4.0/ and https://www.ic.gc.ca/eic/site/cipointernet-internetopic.nsf/eng/wr01935.html.

    Details of Repository Contents:

    This repository includes a number of .zip archives which expand into folders containing either scripts for construction and analysis of the dataset or data files comprising the dataset itself. These folders are as follows:

    • /csv: contains the .csv versions of the data files
    • /do: contains Stata do-files used to convert the .csv files to .dta format and perform the statistical analyses set forth in the paper reporting this dataset
    • /dta: contains the .dta versions of the data files
    • /py: contains the python scripts used to download CIPO’s historical trademarks data via SFTP and generate the .csv data files

    If users wish to construct rather than download the datafiles, the first script that they should run is /py/sftp_secure.py. This script will prompt the user to enter their IP Horizons SFTP credentials; these can be obtained by registering with CIPO at https://ised-isde.survey-sondage.ca/f/s.aspx?s=59f3b3a4-2fb5-49a4-b064-645a5e3a752d&lang=EN&ds=SFTP. The script will also prompt the user to identify a target directory for the data downloads. Because the data archives are quite large, users are advised to create a target directory in advance and ensure they have at least 70GB of available storage on the media in which the directory is located.

    The sftp_secure.py script will generate a new subfolder in the user’s target directory called /XML_raw. Users should note the full path of this directory, which they will be prompted to provide when running the remaining python scripts. Each of the remaining scripts, the filenames of which begin with “iterparse”, corresponds to one of the data files in the dataset, as indicated in the script’s filename. After running one of these scripts, the user’s target directory should include a /csv subdirectory containing the data file corresponding to the script; after running all the iterparse scripts the user’s /csv directory should be identical to the /csv directory in this repository. Users are invited to modify these scripts as they see fit, subject to the terms of the licenses set forth above.

    With respect to the Stata do-files, only one of them is relevant to construction of the dataset itself. This is /do/CA_TM_csv_cleanup.do, which converts the .csv versions of the data files to .dta format, and uses Stata’s labeling functionality to reduce the size of the resulting files while preserving information. The other do-files generate the analyses and graphics presented in the paper describing the dataset (Jeremy N. Sheff, The Canada Trademarks Dataset, 18 J. Empirical Leg. Studies (forthcoming 2021)), available at https://papers.ssrn.com/abstract=3782655). These do-files are also licensed for reuse subject to the terms of the CC-BY-4.0 license, and users are invited to adapt the scripts to their needs.

    The python and Stata scripts included in this repository are separately maintained and updated on Github at https://github.com/jnsheff/CanadaTM.

    This repository also includes a copy of the current version of CIPO's data dictionary for its historical XML trademarks archive as of the date of construction of this dataset.

  6. cai-conversation

    • huggingface.co
    Updated Dec 6, 2023
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Hugging Face H4 (2023). cai-conversation [Dataset]. https://huggingface.co/datasets/HuggingFaceH4/cai-conversation
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Dec 6, 2023
    Dataset provided by
    Hugging Facehttps://huggingface.co/
    Authors
    Hugging Face H4
    Description

    Dataset Card for "cai-conversation"

    More Information needed

  7. u

    Point-of-care Hepatitis C testing in Ghana : dataset from primary care...

    • researchdata.up.ac.za
    application/x-rar
    Updated Jun 5, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Evans Duah; Kuhlula Maluleke (2025). Point-of-care Hepatitis C testing in Ghana : dataset from primary care evaluation [Dataset]. http://doi.org/10.25403/UPresearchdata.29209832.v2
    Explore at:
    application/x-rarAvailable download formats
    Dataset updated
    Jun 5, 2025
    Dataset provided by
    University of Pretoria
    Authors
    Evans Duah; Kuhlula Maluleke
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Area covered
    Ghana
    Description

    This dataset supports a diagnostic trial that evaluated the Bioline™ Hepatitis C Virus (HCV) point-of-care (POC) test in Ghana’s primary healthcare settings. The study aimed to assess the diagnostic performance, usability, acceptability, and cost-effectiveness of the Bioline™ HCV test using the REASSURED criteria. Data were collected in three phases: diagnostic accuracy testing among 516 participants, a usability and acceptability assessment with 81 healthcare workers using mixed methods, and a cost analysis of different HCV testing models. The dataset includes quantitative performance metrics, usability scores, qualitative interview data, and cost comparisons. These findings contribute to understanding the feasibility of decentralized HCV testing in resource-limited settings and inform efforts toward achieving the WHO hepatitis elimination targets.

  8. E

    Data from: Global hydrological dataset of daily streamflow data from the...

    • catalogue.ceh.ac.uk
    • hosted-metadata.bgs.ac.uk
    • +2more
    zip
    Updated May 28, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    S. Turner; J. Hannaford; L.J. Barker; G. Suman; R. Armitage; A. Killeen; A. Griffin; H. Davies; A. Kumar; H. Dixon; M.T.D. Albuquerque; N. Almeida Ribeiro; C. Alvarez-Garreton; E. Amoussou; B. Arheimer; Y. Asano; T. Berezowski; A. Bodian; H. Boutaghane; R. Capell; H. Dakhaoui; J. Daňhelka; H.X. Do; C. Ekkawatpanit; E.M. El Khalki; A.K. Fleig; R. Fonseca; J.D. Giraldo-Osorio; A.B.T. Goula; M. Hanel; G Hodgkins; S. Horton; C. Kan; D.G. Kingston; G. Laaha; R. Laugesen; W. Lopes; S. Mager; Y. Markonis; L. Mediero; G. Midgley; C. Murphy; P. O'Connor; A.I. Pedersen; H.T. Pham; M. Piniewski; M. Rachdane; B. Renard; M.E. Saidi; P. Schmocker-Facker; K. Stahl; M. Thyler; M. Toucher; Y. Tramblay; J. Uusikivi; N. Venegas-Cordero; S. Vissesri; A. Watson; S. Westra; P.H. Whitfield (2024). Global hydrological dataset of daily streamflow data from the Reference Observatory of Basins for INternational hydrological climate change detection (ROBIN), 1863 - 2022 [Dataset]. http://doi.org/10.5285/3b077711-f183-42f1-bac6-c892922c81f4
    Explore at:
    zipAvailable download formats
    Dataset updated
    May 28, 2024
    Dataset provided by
    NERC EDS Environmental Information Data Centre
    Authors
    S. Turner; J. Hannaford; L.J. Barker; G. Suman; R. Armitage; A. Killeen; A. Griffin; H. Davies; A. Kumar; H. Dixon; M.T.D. Albuquerque; N. Almeida Ribeiro; C. Alvarez-Garreton; E. Amoussou; B. Arheimer; Y. Asano; T. Berezowski; A. Bodian; H. Boutaghane; R. Capell; H. Dakhaoui; J. Daňhelka; H.X. Do; C. Ekkawatpanit; E.M. El Khalki; A.K. Fleig; R. Fonseca; J.D. Giraldo-Osorio; A.B.T. Goula; M. Hanel; G Hodgkins; S. Horton; C. Kan; D.G. Kingston; G. Laaha; R. Laugesen; W. Lopes; S. Mager; Y. Markonis; L. Mediero; G. Midgley; C. Murphy; P. O'Connor; A.I. Pedersen; H.T. Pham; M. Piniewski; M. Rachdane; B. Renard; M.E. Saidi; P. Schmocker-Facker; K. Stahl; M. Thyler; M. Toucher; Y. Tramblay; J. Uusikivi; N. Venegas-Cordero; S. Vissesri; A. Watson; S. Westra; P.H. Whitfield
    Time period covered
    Jan 1, 1863 - Dec 31, 2022
    Area covered
    Earth
    Dataset funded by
    Natural Environment Research Councilhttps://www.ukri.org/councils/nerc
    Description

    The Reference Observatory of Basins for INternational hydrological climate change detection (ROBIN) dataset is a global hydrological dataset containing publicly available daily flow data for 2,386 gauging stations across the globe which have natural or near-natural catchments. Metadata is also provided alongside these stations for the Full ROBIN Dataset consisting of 3,060 gauging stations. Data were quality controlled by the central ROBIN team before being added to the dataset, and two levels of data quality are applied to guide users towards appropriate the data usage. Most records have data of at least 40 years with minimal missing data with data records starting in the late 19th Century for some sites through to 2022. ROBIN represents a significant advance in global-scale, accessible streamflow data. The project was funded the UK Natural Environment Research Council Global Partnership Seedcorn Fund - NE/W004038/1 and the NC-International programme [NE/X006247/1] delivering National Capability

  9. Road safety statistics: data tables

    • gov.uk
    Updated Dec 19, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Department for Transport (2024). Road safety statistics: data tables [Dataset]. https://www.gov.uk/government/statistical-data-sets/reported-road-accidents-vehicles-and-casualties-tables-for-great-britain
    Explore at:
    Dataset updated
    Dec 19, 2024
    Dataset provided by
    GOV.UKhttp://gov.uk/
    Authors
    Department for Transport
    Description

    These tables present high-level breakdowns and time series. A list of all tables, including those discontinued, is available in the table index. More detailed data is available in our data tools, or by downloading the open dataset.

    Latest data and table index

    The tables below are the latest final annual statistics for 2023. The latest data currently available are provisional figures for 2024. These are available from the latest provisional statistics.

    A list of all reported road collisions and casualties data tables and variables in our data download tool is available in the https://assets.publishing.service.gov.uk/media/683709928ade4d13a63236df/reported-road-casualties-gb-index-of-tables.ods">Tables index (ODS, 30.1 KB).

    All collision, casualty and vehicle tables

    https://assets.publishing.service.gov.uk/media/66f44e29c71e42688b65ec43/ras-all-tables-excel.zip">Reported road collisions and casualties data tables (zip file) (ZIP, 16.6 MB)

    Historic trends (RAS01)

    RAS0101: https://assets.publishing.service.gov.uk/media/66f44bd130536cb927482733/ras0101.ods">Collisions, casualties and vehicles involved by road user type since 1926 (ODS, 52.1 KB)

    RAS0102: https://assets.publishing.service.gov.uk/media/66f44bd1080bdf716392e8ec/ras0102.ods">Casualties and casualty rates, by road user type and age group, since 1979 (ODS, 142 KB)

    Road user type (RAS02)

    RAS0201: https://assets.publishing.service.gov.uk/media/66f44bd1a31f45a9c765ec1f/ras0201.ods">Numbers and rates (ODS, 60.7 KB)

    RAS0202: https://assets.publishing.service.gov.uk/media/66f44bd1e84ae1fd8592e8f0/ras0202.ods">Sex and age group (ODS, 167 KB)

    RAS0203: https://assets.publishing.service.gov.uk/media/67600227b745d5f7a053ef74/ras0203.ods">Rates by mode, including air, water and rail modes (ODS, 24.2 KB)

    Road type (RAS03)

    RAS0301: https://assets.publishing.service.gov.uk/media/66f44bd1c71e42688b65ec3e/ras0301.ods">Speed limit, built-up and non-built-up roads (ODS, 49.3 KB)

    RAS0302: https://assets.publishing.service.gov.uk/media/66f44bd1080bdf716392e8ee/ras0302.ods">Urban and rural roa

  10. Transaction Graph Dataset for the Bitcoin Blockchain - Part 2 of 4 - Dataset...

    • cryptodata.center
    Updated Dec 4, 2024
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    cryptodata.center (2024). Transaction Graph Dataset for the Bitcoin Blockchain - Part 2 of 4 - Dataset - CryptoData Hub [Dataset]. https://cryptodata.center/dataset/transaction-graph-dataset-for-the-bitcoin-blockchain-part-2-of-4
    Explore at:
    Dataset updated
    Dec 4, 2024
    Dataset provided by
    CryptoDATA
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    This dataset contains bitcoin transfer transactions extracted from the Bitcoin Mainnet blockchain. Details of the datasets are given below: FILENAME FORMAT: The filenames have the following format: btc-tx- where For example file btc-tx-100000-149999-aa.bz2 and the rest of the parts if any contain transactions from block 100000 to block 149999 inclusive. The files are compressed with bzip2. They can be uncompressed using command bunzip2. TRANSACTION FORMAT: Each line in a file corresponds to a transaction. The transaction has the following format: BLOCK TIME FORMAT: The block time file has the following format: IMPORTANT NOTE: Public Bitcoin Mainnet blockchain data is open and can be obtained by connecting as a node on the blockchain or by using the block explorer web sites such as https://btcscan.org . The downloaders and users of this dataset accept the full responsibility of using the data in GDPR compliant manner or any other regulations. We provide the data as is and we cannot be held responsible for anything. NOTE: If you use this dataset, please do not forget to add the DOI number to the citation. If you use our dataset in your research, please also cite our paper: https://link.springer.com/chapter/10.1007/978-3-030-94590-9_14 @incollection{kilicc2022analyzing, title={Analyzing Large-Scale Blockchain Transaction Graphs for Fraudulent Activities}, author={K{\i}l{\i}{\c{c}}, Baran and {"O}zturan, Can and {\c{S}}en, Alper}, booktitle={Big Data and Artificial Intelligence in Digital Finance}, pages={253--267}, year={2022}, publisher={Springer, Cham} }

  11. o

    Data from: LifeSnaps: a 4-month multi-modal dataset capturing unobtrusive...

    • explore.openaire.eu
    • zenodo.org
    Updated Jul 13, 2022
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Sofia Yfantidou; Christina Karagianni; Stefanos Efstathiou; Athena Vakali; Joao Palotti; Dimitrios Panteleimon Giakatos; Thomas Marchioro; Andrei Kazlouski; Elena Ferrari; Šarūnas Girdzijauskas (2022). LifeSnaps: a 4-month multi-modal dataset capturing unobtrusive snapshots of our lives in the wild [Dataset]. http://doi.org/10.5281/zenodo.6832242
    Explore at:
    Dataset updated
    Jul 13, 2022
    Authors
    Sofia Yfantidou; Christina Karagianni; Stefanos Efstathiou; Athena Vakali; Joao Palotti; Dimitrios Panteleimon Giakatos; Thomas Marchioro; Andrei Kazlouski; Elena Ferrari; Šarūnas Girdzijauskas
    Description

    LifeSnaps Dataset Documentation Ubiquitous self-tracking technologies have penetrated various aspects of our lives, from physical and mental health monitoring to fitness and entertainment. Yet, limited data exist on the association between in the wild large-scale physical activity patterns, sleep, stress, and overall health, and behavioral patterns and psychological measurements due to challenges in collecting and releasing such datasets, such as waning user engagement, privacy considerations, and diversity in data modalities. In this paper, we present the LifeSnaps dataset, a multi-modal, longitudinal, and geographically-distributed dataset, containing a plethora of anthropological data, collected unobtrusively for the total course of more than 4 months by n=71 participants, under the European H2020 RAIS project. LifeSnaps contains more than 35 different data types from second to daily granularity, totaling more than 71M rows of data. The participants contributed their data through numerous validated surveys, real-time ecological momentary assessments, and a Fitbit Sense smartwatch, and consented to make these data available openly to empower future research. We envision that releasing this large-scale dataset of multi-modal real-world data, will open novel research opportunities and potential applications in the fields of medical digital innovations, data privacy and valorization, mental and physical well-being, psychology and behavioral sciences, machine learning, and human-computer interaction. The following instructions will get you started with the LifeSnaps dataset and are complementary to the original publication. Data Import: Reading CSV For ease of use, we provide CSV files containing Fitbit, SEMA, and survey data at daily and/or hourly granularity. You can read the files via any programming language. For example, in Python, you can read the files into a Pandas DataFrame with the pandas.read_csv() command. Data Import: Setting up a MongoDB (Recommended) To take full advantage of the LifeSnaps dataset, we recommend that you use the raw, complete data via importing the LifeSnaps MongoDB database. To do so, open the terminal/command prompt and run the following command for each collection in the DB. Ensure you have MongoDB Database Tools installed from here. For the Fitbit data, run the following: mongorestore --host localhost:27017 -d rais_anonymized -c fitbit

  12. U

    Cell Maps for Artificial Intelligence - June 2025 Data Release (Beta)

    • dataverse.lib.virginia.edu
    Updated Jul 1, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Clark T; Clark T; Parker J; Parker J; Al Manir S; Al Manir S; Axelsson U; Navarro ero Navarro F; Navarro ero Navarro F; Chinn B; Churas CP; Dailamy A; Dailamy A; Doctor Y; Doctor Y; Fall J; Forget A; Forget A; Gao J; Gao J; Hansen JN; Hansen JN; Hu M; Johannesson A; Khaliq H; Lee YH; Lee YH; Lenkiewicz J; Levinson MA; Levinson MA; Marquez C; Marquez C; Metallo C; Metallo C; Muralidharan M; Nourreddine S; Niestroy J; Niestroy J; Obernier K; Obernier K; Polacco B; Pratt D; Pratt D; Qian G; Qian G; Schaffer L; Schaffer L; Sigaeva A; Sigaeva A; Thaker S; Thaker S; Zhang Y; Bélisle-Pipon JC; Bélisle-Pipon JC; Brandt C; Brandt C; Chen JY; Chen JY; Ding Y; Ding Y; Fodeh S; Fodeh S; Krogan N; Krogan N; Lundberg E; Lundberg E; Mali P; Payne-Foster P; Payne-Foster P; Ratcliffe S; Ratcliffe S; Ravitsky V; Ravitsky V; Sali A; Sali A; Schulz W; Schulz W; Ideker T; Ideker T; Axelsson U; Chinn B; Churas CP; Fall J; Hu M; Johannesson A; Khaliq H; Lenkiewicz J; Muralidharan M; Nourreddine S; Polacco B; Zhang Y; Mali P (2025). Cell Maps for Artificial Intelligence - June 2025 Data Release (Beta) [Dataset]. http://doi.org/10.18130/V3/F3TD5R
    Explore at:
    json(395980), html(25393), html(42795), json(351611), html(91694), html(40150), html(40093), zip(2777853502), json(38057), html(48532), html(42658), html(44677), zip(3419931819), html(291382), zip(3057986837), application/ld+json; profile="http://www.w3.org/ns/json-ld#flattened http://www.w3.org/ns/json-ld#compacted https://w3id.org/ro/crate"(35868), html(40141), json(1213999), text/comma-separated-values(25061), html(284877), html(356805)Available download formats
    Dataset updated
    Jul 1, 2025
    Dataset provided by
    University of Virginia Dataverse
    Authors
    Clark T; Clark T; Parker J; Parker J; Al Manir S; Al Manir S; Axelsson U; Navarro ero Navarro F; Navarro ero Navarro F; Chinn B; Churas CP; Dailamy A; Dailamy A; Doctor Y; Doctor Y; Fall J; Forget A; Forget A; Gao J; Gao J; Hansen JN; Hansen JN; Hu M; Johannesson A; Khaliq H; Lee YH; Lee YH; Lenkiewicz J; Levinson MA; Levinson MA; Marquez C; Marquez C; Metallo C; Metallo C; Muralidharan M; Nourreddine S; Niestroy J; Niestroy J; Obernier K; Obernier K; Polacco B; Pratt D; Pratt D; Qian G; Qian G; Schaffer L; Schaffer L; Sigaeva A; Sigaeva A; Thaker S; Thaker S; Zhang Y; Bélisle-Pipon JC; Bélisle-Pipon JC; Brandt C; Brandt C; Chen JY; Chen JY; Ding Y; Ding Y; Fodeh S; Fodeh S; Krogan N; Krogan N; Lundberg E; Lundberg E; Mali P; Payne-Foster P; Payne-Foster P; Ratcliffe S; Ratcliffe S; Ravitsky V; Ravitsky V; Sali A; Sali A; Schulz W; Schulz W; Ideker T; Ideker T; Axelsson U; Chinn B; Churas CP; Fall J; Hu M; Johannesson A; Khaliq H; Lenkiewicz J; Muralidharan M; Nourreddine S; Polacco B; Zhang Y; Mali P
    License

    Attribution-NonCommercial-ShareAlike 4.0 (CC BY-NC-SA 4.0)https://creativecommons.org/licenses/by-nc-sa/4.0/
    License information was derived automatically

    Dataset funded by
    National Institutes of Health
    Description

    Description This dataset is the June 2025 Data Release of Cell Maps for Artificial Intelligence (CM4AI; CM4AI.org), the Functional Genomics Grand Challenge in the NIH Bridge2AI program. This Beta release includes perturb-seq data in undifferentiated KOLF2.1J iPSCs; SEC-MS data in undifferentiated KOLF2.1J iPSCs, iPSC-derived NPCs, neurons, cardiomyocytes, and treated and untreated MDA-MB468 breast cancer cells; and IF images in MDA-MB-468 breast cancer cells in the presence and absence of chemotherapy (vorinostat and paclitaxel). External Data Links Access external data resources related to this dataset: Sequence Read Archive (SRA) Data: NCBI BioProject Mass Spectrometry Data (Human iPSCs): MassIVE Repository Mass Spectrometry Data (Human Cancer Cells): MassIVE Repository Data Governance & Ethics Human Subjects: No De-identified Samples: Yes FDA Regulated: No Data Governance Committee: Jillian Parker (jillianparker@health.ucsd.edu) Ethical Review: Vardit Ravitsky (ravitskyv@thehastingscenter.org) and Jean-Christophe Belisle-Pipon (jean-christophe_belisle-pipon@sfu.ca) Completeness These data are not yet in completed final form: Some datasets are under temporary pre-publication embargo Protein-protein interaction (SEC-MS), protein localization (IF imaging), and CRISPRi perturbSeq data interrogate sets of proteins which incompletely overlap Computed cell maps not included in this release Maintenance Plan Dataset will be regularly updated and augmented through the end of the project in November 2026 Updates on a quarterly basis Long term preservation in the University of Virginia Dataverse, supported by committed institutional funds Intended Use This dataset is intended for: AI-ready datasets to support research in functional genomics AI model training Cellular process analysis Cell architectural changes and interactions in presence of specific disease processes, treatment conditions, or genetic perturbations Limitations Researchers should be aware of inherent limitations: This is an interim release Does not contain predicted cell maps, which will be added in future releases The current release is most suitable for bioinformatics analysis of the individual datasets Requires domain expertise for meaningful analysis Prohibited Uses These laboratory data are not to be used in clinical decision-making or in any context involving patient care without appropriate regulatory oversight and approval Potential Sources of Bias Users should be aware of potential biases: Data in this release was derived from commercially available de-identified human cell lines Does not represent all biological variants which may be seen in the population at large

  13. u

    Data from: Satellite remote sensing dataset of Sentinel-2 for phenology...

    • observatorio-cientifico.ua.es
    • producciocientifica.uv.es
    • +1more
    Updated 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Ganeva, Dessislava; Graf Valentin, Lukas; Prikaziuk, Egor; Koren, Gerbrand; Tomelleri, Enrico; Verrelst, Jochem; Berger, Katja; Belda, Santiago; Cai, Zhanzhang; Silva Figueira, Cláudio; Ganeva, Dessislava; Graf Valentin, Lukas; Prikaziuk, Egor; Koren, Gerbrand; Tomelleri, Enrico; Verrelst, Jochem; Berger, Katja; Belda, Santiago; Cai, Zhanzhang; Silva Figueira, Cláudio (2023). Satellite remote sensing dataset of Sentinel-2 for phenology metrics extraction from sites in Bulgaria and France [Dataset]. https://observatorio-cientifico.ua.es/documentos/668fc44fb9e7c03b01bd98c9
    Explore at:
    Dataset updated
    2023
    Authors
    Ganeva, Dessislava; Graf Valentin, Lukas; Prikaziuk, Egor; Koren, Gerbrand; Tomelleri, Enrico; Verrelst, Jochem; Berger, Katja; Belda, Santiago; Cai, Zhanzhang; Silva Figueira, Cláudio; Ganeva, Dessislava; Graf Valentin, Lukas; Prikaziuk, Egor; Koren, Gerbrand; Tomelleri, Enrico; Verrelst, Jochem; Berger, Katja; Belda, Santiago; Cai, Zhanzhang; Silva Figueira, Cláudio
    Area covered
    Bulgaria, France
    Description

    Site Description: In this dataset, there are seventeen production crop fields in Bulgaria where winter rapeseed and wheat were grown and two research fields in France where winter wheat – rapeseed – barley – sunflower and winter wheat – irrigated maize crop rotation is used. The full description of those fields is in the database "In-situ crop phenology dataset from sites in Bulgaria and France" (doi.org/10.5281/zenodo.7875440). Methodology and Data Description: Remote sensing data is extracted from Sentinel-2 tiles 35TNJ for Bulgarian sites and 31TCJ for French sites on the day of the overpass since September 2015 for Sentinel-2 derived vegetation indices and since October 2016 for HR-VPP products. To suppress spectral mixing effects at the parcel boundaries, as highlighted by Meier et al., 2020, the values from all datasets were subgrouped per field and then aggregated to a single median value for further analysis. Sentinel-2 data was downloaded for all test sites from CREODIAS (https://creodias.eu/) in L2A processing level using a maximum scene-wide cloudy cover threshold of 75%. Scenes before 2017 were available in L1C processing level only. Scenes in L1C processing level were corrected for atmospheric effects after downloading using Sen2Cor (v2.9) with default settings. This was the same version used for the L2A scenes obtained intermediately from CREODIAS. Next, the data was extracted from the Sentinel-2 scenes for each field parcel where only SCL classes 4 (vegetation) and 5 (bare soil) pixels were kept. We resampled the 20m band B8A to match the spatial resolution of the green and red band (10m) using nearest neighbor interpolation. The entire image processing chain was carried out using the open-source Python Earth Observation Data Analysis Library (EOdal) (Graf et al., 2022). Apart from the widely used Normalized Difference Vegetation Index (NDVI) and Enhanced Vegetation Index (EVI), we included two recently proposed indices that were reported to have a higher correlation with photosynthesis and drought response of vegetation: These were the Near-Infrared Reflection of Vegetation (NIRv) (Badgley et al., 2017) and Kernel NDVI (kNDVI) (Camps-Valls et al., 2021). We calculated the vegetation indices in two different ways: First, we used B08 as near-infrared (NIR) band which comes in a native spatial resolution of 10 m. B08 (central wavelength 833 nm) has a relatively coarse spectral resolution with a bandwidth of 106 nm. Second, we used B8A which is available at 20 m spatial resolution. B8A differs from B08 in its central wavelength (864 nm) and has a narrower bandwidth (21 nm or 22 nm in the case of Sentinel-2A and 2B, respectively) compared to B08. The High Resolution Vegetation Phenology and Productivity (HR-VPP) dataset from Copernicus Land Monitoring Service (CLMS) has three 10-m set products of Sentinel-2: vegetation indices, vegetation phenology and productivity parameters and seasonal trajectories (Tian et al., 2021). Both vegetation indices, Normalized Vegetation Index (NDVI) and Plant Phenology (PPI) and plant parameters, Fraction of Absorbed Photosynthetic Active Radiation (FAPAR) and Leaf Area Index (LAI) were computed for the time of Sentinel-2 overpass by the data provider. NDVI is computed directly from B04 and B08 and PPI is computed using Difference Vegetation Index (DVI = B08 - B04) and its seasonal maximum value per pixel. FAPAR and LAI are retrieved from B03 and B04 and B08 with neural network training on PROSAIL model simulations. The dataset has a quality flag product (QFLAG2) which is a 16-bit that extends the scene classification band (SCL) of the Sentinel-2 Level-2 products. A “medium” filter was used to mask out QFLAG2 values from 2 to 1022, leaving land pixels (bit 1) within or outside cloud proximity (bits 11 and 13) or cloud shadow proximity (bits 12 and 14). The HR-VPP daily raw vegetation indices products are described in detail in the user manual (Smets et al., 2022) and the computations details of PPI are given by Jin and Eklundh (2014). Seasonal trajectories refer to the 10-daily smoothed time-series of PPI used for vegetation phenology and productivity parameters retrieval with TIMESAT (Jönsson and Eklundh 2002, 2004). HR-VPP data was downloaded through the WEkEO Copernicus Data and Information Access Services (DIAS) system with a Python 3.8.10 harmonized data access (HDA) API 0.2.1. Zonal statistics [’min’, ’max’, ’mean’, ’median’, ’count’, ’std’, ’majority’] were computed on non-masked pixel values within field boundaries with rasterstats Python package 0.17.00. The Start of season date (SOSD), end of season date (EOSD) and length of seasons (LENGTH) were extracted from the annual Vegetation Phenology and Productivity Parameters (VPP) dataset as an additional source for comparison. These data are a product of the Vegetation Phenology and Productivity Parameters, see (https://land.copernicus.eu/pan-european/biophysical-parameters/high-resolution-vegetation-phenology-and-productivity/vegetation-phenology-and-productivity) for detailed information. File Description: 4 datasets: 1_senseco_data_S2_B08_Bulgaria_France; 1_senseco_data_S2_B8A_Bulgaria_France; 1_senseco_data_HR_VPP_Bulgaria_France; 1_senseco_data_phenology_VPP_Bulgaria_France 3 metadata: 2_senseco_metadata_S2_B08_B8A_Bulgaria_France; 2_senseco_metadata_HR_VPP_Bulgaria_France; 2_senseco_metadata_phenology_VPP_Bulgaria_France The dataset files “1_senseco_data_S2_B8_Bulgaria_France” and “1_senseco_data_S2_B8A_Bulgaria_France” concerns all vegetation indices (EVI, NDVI, kNDVI, NIRv) data values and related information, and metadata file “2_senseco_metadata_S2_B08_B8A_Bulgaria_France” describes all the existing variables. Both “1_senseco_data_S2_B8_Bulgaria_France” and “1_senseco_data_S2_B8A_Bulgaria_France” have the same column variable names and for that reason, they share the same metadata file “2_senseco_metadata_S2_B08_B8A_Bulgaria_France”. The dataset file “1_senseco_data_HR_VPP_Bulgaria_France” concerns vegetation indices (NDVI, PPI) and plant parameters (LAI, FAPAR) data values and related information, and metadata file “2_senseco_metadata_HRVPP_Bulgaria_France” describes all the existing variables. The dataset file “1_senseco_data_phenology_VPP_Bulgaria_France” concerns the vegetation phenology and productivity parameters (LENGTH, SOSD, EOSD) values and related information, and metadata file “2_senseco_metadata_VPP_Bulgaria_France” describes all the existing variables. Bibliography G. Badgley, C.B. Field, J.A. Berry, Canopy near-infrared reflectance and terrestrial photosynthesis, Sci. Adv. 3 (2017) e1602244. https://doi.org/10.1126/sciadv.1602244. G. Camps-Valls, M. Campos-Taberner, Á. Moreno-Martínez, S. Walther, G. Duveiller, A. Cescatti, M.D. Mahecha, J. Muñoz-Marí, F.J. García-Haro, L. Guanter, M. Jung, J.A. Gamon, M. Reichstein, S.W. Running, A unified vegetation index for quantifying the terrestrial biosphere, Sci. Adv. 7 (2021) eabc7447. https://doi.org/10.1126/sciadv.abc7447. L.V. Graf, G. Perich, H. Aasen, EOdal: An open-source Python package for large-scale agroecological research using Earth Observation and gridded environmental data, Comput. Electron. Agric. 203 (2022) 107487. https://doi.org/10.1016/j.compag.2022.107487. H. Jin, L. Eklundh, A physically based vegetation index for improved monitoring of plant phenology, Remote Sens. Environ. 152 (2014) 512–525. https://doi.org/10.1016/j.rse.2014.07.010. P. Jonsson, L. Eklundh, Seasonality extraction by function fitting to time-series of satellite sensor data, IEEE Trans. Geosci. Remote Sens. 40 (2002) 1824–1832. https://doi.org/10.1109/TGRS.2002.802519. P. Jönsson, L. Eklundh, TIMESAT—a program for analyzing time-series of satellite sensor data, Comput. Geosci. 30 (2004) 833–845. https://doi.org/10.1016/j.cageo.2004.05.006. J. Meier, W. Mauser, T. Hank, H. Bach, Assessments on the impact of high-resolution-sensor pixel sizes for common agricultural policy and smart farming services in European regions, Comput. Electron. Agric. 169 (2020) 105205. https://doi.org/10.1016/j.compag.2019.105205. B. Smets, Z. Cai, L. Eklund, F. Tian, K. Bonte, R. Van Hoost, R. Van De Kerchove, S. Adriaensen, B. De Roo, T. Jacobs, F. Camacho, J. Sánchez-Zapero, S. Else, H. Scheifinger, K. Hufkens, P. Jönsson, HR-VPP Product User Manual Vegetation Indices, 2022. F. Tian, Z. Cai, H. Jin, K. Hufkens, H. Scheifinger, T. Tagesson, B. Smets, R. Van Hoolst, K. Bonte, E. Ivits, X. Tong, J. Ardö, L. Eklundh, Calibrating vegetation phenology from Sentinel-2 using eddy covariance, PhenoCam, and PEP725 networks across Europe, Remote Sens. Environ. 260 (2021) 112456. https://doi.org/10.1016/j.rse.2021.112456.

  14. Book-Crossing Dataset

    • kaggle.com
    zip
    Updated Sep 7, 2019
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    somnambWl (2019). Book-Crossing Dataset [Dataset]. https://www.kaggle.com/somnambwl/bookcrossing-dataset
    Explore at:
    zip(17632108 bytes)Available download formats
    Dataset updated
    Sep 7, 2019
    Authors
    somnambWl
    Description

    Book-Crossing dataset mined by Cai-Nicolas Ziegler

    Freely available for research use when acknowledged with the following reference (further details on the dataset are given in this publication):

    • PDF

    • Cai-Nicolas Ziegler, Sean M. McNee, Joseph A. Konstan, Georg Lausen; Proceedings of the 14th International World Wide Web Conference (WWW '05), May 10-14, 2005, Chiba, Japan. To appear.

    Further information and the original dataset can be found at the original webpage.

    Changes to the dataset:

    • Location removed as it comes in different formats not in default (city, state, country).
    • Transferred from ISO-8859-1 to UTF-8
    • Manually fixed a few rows with incorrect number of columns

    Note:

    • out of 278859 users:
      • only 99053 rated at least 1 book
      • only 43385 rated at least 2 books.
      • only 12306 rated at least 10 books.
    • out of 271379 books:
      • only 270171 are rated at least once.
      • only 124513 have at least 2 ratings.
      • only 17480 have at least 10 ratings.
  15. o

    Data from: A large EEG database with users' profile information for motor...

    • explore.openaire.eu
    • zenodo.org
    Updated Jan 9, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Dreyer Pauline; Roc Aline; Rimbert Sébastien; Pillette Léa; Lotte Fabien (2023). A large EEG database with users' profile information for motor imagery Brain-Computer Interface research [Dataset]. http://doi.org/10.5281/zenodo.7554429
    Explore at:
    Dataset updated
    Jan 9, 2023
    Authors
    Dreyer Pauline; Roc Aline; Rimbert Sébastien; Pillette Léa; Lotte Fabien
    Description

    Context : We share a large database containing electroencephalographic signals from 87 human participants, with more than 20,800 trials in total representing about 70 hours of recording. It was collected during brain-computer interface (BCI) experiments and organized into 3 datasets (A, B, and C) that were all recorded following the same protocol: right and left hand motor imagery (MI) tasks during one single day session. It includes the performance of the associated BCI users, detailed information about the demographics, personality and cognitive user’s profile, and the experimental instructions and codes (executed in the open-source platform OpenViBE). Such database could prove useful for various studies, including but not limited to: 1) studying the relationships between BCI users' profiles and their BCI performances, 2) studying how EEG signals properties varies for different users' profiles and MI tasks, 3) using the large number of participants to design cross-user BCI machine learning algorithms or 4) incorporating users' profile information into the design of EEG signal classification algorithms. Sixty participants (Dataset A) performed the first experiment, designed in order to investigated the impact of experimenters' and users' gender on MI-BCI user training outcomes, i.e., users performance and experience, (Pillette & al). Twenty one participants (Dataset B) performed the second one, designed to examined the relationship between users' online performance (i.e., classification accuracy) and the characteristics of the chosen user-specific Most Discriminant Frequency Band (MDFB) (Benaroch & al). The only difference between the two experiments lies in the algorithm used to select the MDFB. Dataset C contains 6 additional participants who completed one of the two experiments described above. Physiological signals were measured using a g.USBAmp (g.tec, Austria), sampled at 512 Hz, and processed online using OpenViBE 2.1.0 (Dataset A) & OpenVIBE 2.2.0 (Dataset B). For Dataset C, participants C83 and C85 were collected with OpenViBE 2.1.0 and the remaining 4 participants with OpenViBE 2.2.0. Experiments were recorded at Inria Bordeaux sud-ouest, France. Duration : Each participant's folder is composed of approximately 48 minutes EEG recording. Meaning six 7-minutes runs and a 6-minutes baseline. Documents Instructions: checklist read by experimenters during the experiments. Questionnaires: the Mental Rotation test used, the translation of 4 questionnaires, notably the Demographic and Social information, the Pre and Post-session questionnaires, and the Index of Learning style. English and french version Performance: The online OpenViBE BCI classification performances obtained by each participant are provided for each run, as well as answers to all questionnaires Scenarios/scripts : set of OpenViBE scenarios used to perform each of the steps of the MI-BCI protocol, e.g., acquire training data, calibrate the classifier or run the online MI-BCI Database : raw signals Dataset A : N=60 participants Dataset B : N=21 participants Dataset C : N=6 participants The article that expained the database is available here: Dreyer, P., Roc, A., Pillette, L. et al. A large EEG database with users’ profile information for motor imagery brain-computer interface research. Sci Data 10, 580 (2023). https://doi.org/10.1038/s41597-023-02445-z {"references": ["Pillette & al (2021). Experimenters Influence on Mental-Imagery based Brain-Computer Interface User Training. International Journal of Human-Computer Studies, pp.102603.", "Camille Benaroch & al (2022). When should MI-BCI feature optimization include prior knowledge, and which one?. Brain-Computer Interfaces, 9 (2), pp.115-128"]} https://doi.org/10.1038/s41597-023-02445-z

  16. P

    Bottles and Cups Dataset | Household Objects Dataset

    • paperswithcode.com
    • kaggle.com
    Updated Jul 26, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    (2022). Bottles and Cups Dataset | Household Objects Dataset [Dataset]. https://paperswithcode.com/dataset/bottles-and-cups-dataset-household-objects
    Explore at:
    Dataset updated
    Jul 26, 2022
    Description

    This dataset consists of images of bottles and cups.

    Introduction Dataset consists of images of bottle and cups captured using mobile phones in a real-world scenario. Images were captured under a wide variety of indoor lighting conditions. This dataset can be used for the detection of a wide variety of bottles and cups made up of a variety of materials from a lot of different of viewpoints, locations, orientations, etc.

    Dataset Features

    Captured by 3000+ unique users Captured using mobile phones Variety of different material of bottles and cups HD Resolution Highly diverse Various lighting conditions Indoor scenes

    Dataset Features

    Classification and detection annotations available Multiple category annotations possible COCO, PASCAL VOC and YOLO formats

    To download full datasets or to submit a request for your dataset needs, please ping us at sales@datacluster.ai Visit www.datacluster.ai to know more.

    Note: All the images are manually captured and verified by a large contributor base on DataCluster platform

  17. Z

    Data from: Hcropland30: A hybrid 30-m global cropland map by leveraging...

    • data.niaid.nih.gov
    • zenodo.org
    Updated Aug 3, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Li, Zexuan (2024). Hcropland30: A hybrid 30-m global cropland map by leveraging global land cover products and Landsat data based on a deep learning model [Dataset]. https://data.niaid.nih.gov/resources?id=zenodo_13169747
    Explore at:
    Dataset updated
    Aug 3, 2024
    Dataset provided by
    Fritz, Steffen
    Zhang, Xinyu
    You, Liangzhi
    Yin, He
    Cai, Zhiwen
    Xu, Baodong
    Li, Zexuan
    Wei, Haodong
    Yang, Jingya
    Wu, Hao
    Hu, Qiong
    Wu, Wenbin
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Hcropland30:A 30-m global cropland map by leveraging global land cover products and Landsat data based on a deep learning model

    Please note this dataset is undergoing peer review

    Version: 1.0

    Authors: Qiong Hu a, 1, Zhiwen Cai b, 1, Liangzhi You c, d, Steffen Fritz e, Xinyu Zhang c, He Yin f, Haodong Weic, Jingya Yang g, Zexuan Li a, Qiangyi Yu g, Hao Wu a, Baodong Xu b *, Wenbin Wu g, *

    a Key Laboratory for Geographical Process Analysis & Simulation of Hubei Province/College of Urban and Environmental Sciences, Central China Normal University, Wuhan 430079, China

    b College of Resources and Environment, Huazhong Agricultural University, Wuhan 430070, China

    c Macro Agriculture Research Institute, College of Plant Science and Technology, Huazhong Agricultural University, Wuhan 430070, China

    d International Food Policy Research Institute, 1201 I Street, NW, Washington, DC 20005, USA

    e Novel Data Ecosystems for sustainability Research Group, International Institute for Applied Systems Analysis (IIASA), Schlossplatz 1, Laxenburg A-2361, Austria

    f Department of Geography, Kent State University, 325 S. Lincoln Street, Kent, OH 44242, USA

    g State Key Laboratory of Efficient Utilization of Arid and Semi-arid Arable Land in Northern China, the Institute of Agricultural Resources and Regional Planning, Chinese Academy of Agricultural Sciences, Beijing 100081, China

    Introduction

    We are pleased to introduce a comprehensive global cropland mapping dataset (named Hcropland30) in 2020, meticulously curated to support a wide range of research and analysis applications related to agricultural land and environmental assessment. This dataset encompasses the entire globe, divided into 16,284 grids, each measuring an area of 1°×1°. Hcropland30 was produced by leveraging global land cover products and Landsat data based on a deep learning model. Initially, we established a hierarchal sampling strategy that used the simulated annealing method to identify the representative 1°×1° grids globally and the sparse point-level samples within these selected 1°×1°grids. Subsequently, we employed an ensemble learning technique to expand these sparse point-level samples into the densely pixel-wise labels, creating the area-level 1°×1° cropland labels. These area-level labels were then used to train a U-Net model for predicting global cropland distribution, followed by a comprehensive evaluation of the mapping accuracy.

    Dataset

    1. Hcropland30: A hybrid 30-m global cropland map in 2020

    ****Data format: GeoTiff

    ****Spatial resolution: 30 m

    ****Projection: EPSG: 4326 (WGS84)

    ****Values: 1 denotes cropland and 0 denotes non-cropland

    The dataset has been uploaded in 16,284 tiles. The extent of each tile can be found in the file of “Grids.shp”. Each file is named according to the grid’s Id number. For example, “000015.tif” corresponds to the cropland mapping result for the 15-th 1°×1° grid. This systematic naming convention ensures easy identification and retrieval of the specific grid data.

    1. 1°×1° Grids: This file contains all 16,284 1°×1° grids used in the dataset. The vector file includes 18 attribute fields, providing comprehensive metadata for each grid. These attributes are essential for users who need detailed information about each grid’s characteristics.

    ****Data format: ESRI shapefile

    ****Projection: EPSG: 4326 (WGS84)

    ****Attribute Fields:

    Id: The grid’s ID number.

    area: The area of the grid.

    mode: Indicates the representative sample grid.

    climate: The climate type the grid belongs to.

    dem: Average DEM value of the grid.

    ndvi_s1 to ndvi_s4: Average NDVI values for four seasons within the grid.

    esa, esri, fcs30, fromglc, glad, globeland30: Proportion of cropland pixels of different publicly available cropland products.

    inconsistent: Proportion of inconsistent pixels within the grid according to different public cropland products.

    hcropland30: Proportion of cropland pixels of our Hcropland30 dataset.

    1. Samples: The selected representative pixel-level samples, including 32,343 cropland and 67657 non-cropland samples. The category information of each sample was determined based on visual interpretation on Google Earth image and three-year NDVI time series curves from 2019-2021.

    ****Data format: ESRI shapefile

    ****Projection: EPSG: 4326 (WGS84)

    ****Attribute Fields:

    type: 1 denotes cropland sample and 0 denotes non-cropland sample.

    Citation

    If you use this dataset, please cite the following paper:

    Hu, Q., Cai, Z., You, L., Fritz, S., Zhang, X., Yin, H., Wei, H., Yang, J., Li, Z., Yu, Q., Wu, H., Xu, B., Wu, W. (2024). Hcropland30: A 30-m global cropland map by leveraging global land cover products and Landsat data based on a deep learning model, Remote Sensing of Environment, submitted.

    License

    The data is licensed under Creative Commons Attribution 4.0 International (CC BY 4.0).

    Disclaimer

    This dataset is provided as-is, without any warranty, express or implied. The dataset author is not

    responsible for any errors or omissions in the data, or for any consequences arising from the use

    of the data.

    Contact

    If you have any questions or feedback regarding the dataset, please contact the dataset author

    Qiong Hu (huqiong@ccnu.edu.cn)

  18. f

    Defects in C programs

    • figshare.com
    7z
    Updated Jun 20, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Yuan-An Xiao (2022). Defects in C programs [Dataset]. http://doi.org/10.6084/m9.figshare.20073119.v1
    Explore at:
    7zAvailable download formats
    Dataset updated
    Jun 20, 2022
    Dataset provided by
    figshare
    Authors
    Yuan-An Xiao
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    These are datasets containing software defects in C programs paired with corresponding patches and metadata, collected from public GitHub repositories.

    • GDD.7z contains 181722 general defects
    • MDD.7z contains 48076 memory-related defects

    [Note about compliance] These datasets are to help researchers evaluate the ability of deep learning in software engineering. They are not intended for commercial use, as repositories may have their own license. Users of these datasets should check the license of each defect on GitHub to see what is permitted. We have included the repository name of each defect in the corresponding metadata file for your convenience.

  19. a

    Annual Count of Extreme Summer Days - Projections (12km)

    • hub.arcgis.com
    • climatedataportal.metoffice.gov.uk
    Updated Feb 7, 2023
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Met Office (2023). Annual Count of Extreme Summer Days - Projections (12km) [Dataset]. https://hub.arcgis.com/datasets/2e0ede325c4540e59e02c351a51fa051
    Explore at:
    Dataset updated
    Feb 7, 2023
    Dataset authored and provided by
    Met Office
    Area covered
    Description

    [Updated 28/01/25 to fix an issue in the ‘Lower’ values, which were not fully representing the range of uncertainty. ‘Median’ and ‘Higher’ values remain unchanged. The size of the change varies by grid cell and fixed period/global warming levels but the average difference between the 'lower' values before and after this update is 0.0.]What does the data show? The Annual Count of Extreme Summer Days is the number of days per year where the maximum daily temperature is above 35°C. It measures how many times the threshold is exceeded (not by how much) in a year. Note, the term ‘extreme summer days’ is used to refer to the threshold and temperatures above 35°C outside the summer months also contribute to the annual count. The results should be interpreted as an approximation of the projected number of days when the threshold is exceeded as there will be many factors such as natural variability and local scale processes that the climate model is unable to represent.The Annual Count of Extreme Summer Days is calculated for two baseline (historical) periods 1981-2000 (corresponding to 0.51°C warming) and 2001-2020 (corresponding to 0.87°C warming) and for global warming levels of 1.5°C, 2.0°C, 2.5°C, 3.0°C, 4.0°C above the pre-industrial (1850-1900) period. This enables users to compare the future number of extreme summer days to previous values.What are the possible societal impacts?The Annual Count of Extreme Summer Days indicates increased health risks, transport disruption and damage to infrastructure from high temperatures. It is based on exceeding a maximum daily temperature of 35°C. Impacts include:Increased heat related illnesses, hospital admissions or death affecting not just the vulnerable. Transport disruption due to overheating of road and railway infrastructure.Other metrics such as the Annual Count of Summer Days (days above 25°C), Annual Count of Hot Summer Days (days above 30°C) and the Annual Count of Tropical Nights (where the minimum temperature does not fall below 20°C) also indicate impacts from high temperatures, however they use different temperature thresholds.What is a global warming level?The Annual Count of Extreme Summer Days is calculated from the UKCP18 regional climate projections using the high emissions scenario (RCP 8.5) where greenhouse gas emissions continue to grow. Instead of considering future climate change during specific time periods (e.g. decades) for this scenario, the dataset is calculated at various levels of global warming relative to the pre-industrial (1850-1900) period. The world has already warmed by around 1.1°C (between 1850–1900 and 2011–2020), whilst this dataset allows for the exploration of greater levels of warming. The global warming levels available in this dataset are 1.5°C, 2°C, 2.5°C, 3°C and 4°C. The data at each warming level was calculated using a 21 year period. These 21 year periods are calculated by taking 10 years either side of the first year at which the global warming level is reached. This time will be different for different model ensemble members. To calculate the value for the Annual Count of Extreme Summer Days, an average is taken across the 21 year period. Therefore, the Annual Count of Extreme Summer Days show the number of extreme summer days that could occur each year, for each given level of warming. We cannot provide a precise likelihood for particular emission scenarios being followed in the real world future. However, we do note that RCP8.5 corresponds to emissions considerably above those expected with current international policy agreements. The results are also expressed for several global warming levels because we do not yet know which level will be reached in the real climate as it will depend on future greenhouse emission choices and the sensitivity of the climate system, which is uncertain. Estimates based on the assumption of current international agreements on greenhouse gas emissions suggest a median warming level in the region of 2.4-2.8°C, but it could either be higher or lower than this level.What are the naming conventions and how do I explore the data?This data contains a field for each global warming level and two baselines. They are named ‘ESD’ (where ESD means Extreme Summer Days, the warming level or baseline, and ‘upper’ ‘median’ or ‘lower’ as per the description below. E.g. ‘Extreme Summer Days 2.5 median’ is the median value for the 2.5°C warming level. Decimal points are included in field aliases but not field names e.g. ‘Extreme Summer Days 2.5 median’ is ‘ExtremeSummerDays_25_median’. To understand how to explore the data, see this page: https://storymaps.arcgis.com/stories/457e7a2bc73e40b089fac0e47c63a578Please note, if viewing in ArcGIS Map Viewer, the map will default to ‘ESD 2.0°C median’ values.What do the ‘median’, ‘upper’, and ‘lower’ values mean?Climate models are numerical representations of the climate system. To capture uncertainty in projections for the future, an ensemble, or group, of climate models are run. Each ensemble member has slightly different starting conditions or model set-ups. Considering all of the model outcomes gives users a range of plausible conditions which could occur in the future. For this dataset, the model projections consist of 12 separate ensemble members. To select which ensemble members to use, the Annual Count of Extreme Summer Days was calculated for each ensemble member and they were then ranked in order from lowest to highest for each location. The ‘lower’ fields are the second lowest ranked ensemble member. The ‘upper’ fields are the second highest ranked ensemble member. The ‘median’ field is the central value of the ensemble.This gives a median value, and a spread of the ensemble members indicating the range of possible outcomes in the projections. This spread of outputs can be used to infer the uncertainty in the projections. The larger the difference between the lower and upper fields, the greater the uncertainty.‘Lower’, ‘median’ and ‘upper’ are also given for the baseline periods as these values also come from the model that was used to produce the projections. This allows a fair comparison between the model projections and recent past. Useful linksThis dataset was calculated following the methodology in the ‘Future Changes to high impact weather in the UK’ report and uses the same temperature thresholds as the 'State of the UK Climate' report.Further information on the UK Climate Projections (UKCP).Further information on understanding climate data within the Met Office Climate Data Portal.

  20. o

    Capture-24: Activity tracker dataset for human activity recognition

    • ora.ox.ac.uk
    • healthdatagateway.org
    Updated Jan 1, 2021
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Chan Chang, S; Walmsley, R; Gershuny , J; Harms, T; Thomas, E; Milton, K; Kelly, P; Foster, C; Wong, A; Gray, N; Haque, S; Hollowell, S; Doherty, A (2021). Capture-24: Activity tracker dataset for human activity recognition [Dataset]. http://doi.org/10.5287/bodleian:NGx0JOMP5
    Explore at:
    (6902652480)Available download formats
    Dataset updated
    Jan 1, 2021
    Dataset provided by
    University of Oxford
    Authors
    Chan Chang, S; Walmsley, R; Gershuny , J; Harms, T; Thomas, E; Milton, K; Kelly, P; Foster, C; Wong, A; Gray, N; Haque, S; Hollowell, S; Doherty, A
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Time period covered
    Jan 1, 2014 - Dec 31, 2016
    Area covered
    Oxford
    Description

    This dataset contains Axivity AX3 wrist-worn activity tracker data that were collected from 151 participants in 2014-2016 around the Oxfordshire area. Participants were asked to wear the device in daily living for a period of roughly 24 hours, amounting to a total of almost 4,000 hours. Vicon Autograph wearable cameras and Whitehall II sleep diaries were used to obtain the ground truth activities performed during the period (e.g. sitting watching TV, walking the dog, washing dishes, sleeping), resulting in more than 2,500 hours of labelled data. Accompanying code to analyse this data is available at https://github.com/activityMonitoring/capture24. The following papers describe the data collection protocol in full: i.) Gershuny J, Harms T, Doherty A, Thomas E, Milton K, Kelly P, Foster C (2020) Testing self-report time-use diaries against objective instruments in real time. Sociological Methodology doi: 10.1177/0081175019884591; ii.) Willetts M, Hollowell S, Aslett L, Holmes C, Doherty A. (2018) Statistical machine learning of sleep and physical activity phenotypes from sensor data in 96,220 UK Biobank participants. Scientific Reports. 8(1):7961. Regarding Data Protection, the Clinical Data Set will not include any direct subject identifiers. However, it is possible that the Data Set may contain certain information that could be used in combination with other information to identify a specific individual, such as a combination of activities specific to that individual ("Personal Data"). Accordingly, in the conduct of the Analysis, users will comply with all applicable laws and regulations relating to information privacy. Further, the user agrees to preserve the confidentiality of, and not attempt to identify, individuals in the Data Set.

Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
NASA/JPL/PODAAC (2025). MetOp-C ASCAT Level 2 25.0km Ocean Surface Wind Vectors in Full Orbit Swath [Dataset]. https://catalog.data.gov/dataset/metop-c-ascat-level-2-25-0km-ocean-surface-wind-vectors-in-full-orbit-swath-d26e1

MetOp-C ASCAT Level 2 25.0km Ocean Surface Wind Vectors in Full Orbit Swath

Explore at:
Dataset updated
Apr 10, 2025
Dataset provided by
NASA/JPL/PODAAC
Description

This dataset contains operational near-real-time Level 2 ocean surface wind vector retrievals from the Advanced Scatterometer (ASCAT) on MetOp-C at 25 km sampling resolution (note: the effective resolution is 50 km). It is a product of the European Organization for the Exploitation of Meteorological Satellites (EUMETSAT) Ocean and Sea Ice Satellite Application Facility (OSI SAF) provided through the Royal Netherlands Meteorological Institute (KNMI). The wind vector retrievals are currently processed using the CMOD7.n geophysical model function using a Hamming filter to spatially average the Sigma-0 data in the ASCAT L1B data. Each file is provided in netCDF version 3 format, and contains one full orbit derived from 3-minute orbit granules. Latency is approximately 2 hours from the latest measurement. The beginning of the orbit is defined by the first wind vector cell measurement within the first 3-minute orbit granule that starts north of the Equator in the ascending node. ASCAT is a C-band dual swath fan beam radar scatterometer providing two independent swaths of backscatter retrievals in sun-synchronous polar orbit aboard the MetOp-C platform. For more information about the MetOp-C platform and mission, please refer to: https://www.eumetsat.int/our-satellites/metop-series . For more timely announcements, users are encouraged to register with the KNMI scatterometer email list: scat@knmi.nl. Users are also highly advised to check the dataset user guide periodically for updates and new information on known problems and issues. All intellectual property rights of the OSI SAF products belong to EUMETSAT. The use of these products is granted to every interested user, free of charge. If you wish to use these products, EUMETSAT's copyright credit must be shown by displaying the words "copyright (year) EUMETSAT" on each of the products used.

Search
Clear search
Close search
Google apps
Main menu