100+ datasets found
  1. cmomy: A python package to calculate and manipulate Central (co)moments.

    • catalog.data.gov
    • datasets.ai
    Updated Sep 30, 2025
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    National Institute of Standards and Technology (2025). cmomy: A python package to calculate and manipulate Central (co)moments. [Dataset]. https://catalog.data.gov/dataset/cmomy-a-python-package-to-calculate-and-manipulate-central-comoments
    Explore at:
    Dataset updated
    Sep 30, 2025
    Dataset provided by
    National Institute of Standards and Technologyhttp://www.nist.gov/
    Description

    cmomy is a python package to calculate central moments and co-moments in a numerical stable and direct way. Behind the scenes, cmomy makes use of Numba to rapidly calculate moments. cmomy provides utilities to calculate central moments from individual samples, precomputed central moments, and precomputed raw moments. It also provides routines to perform bootstrap resampling based on raw data, or precomputed moments. cmomy has numpy array and xarray DataArray interfaces.

  2. Storage and Transit Time Data and Code

    • zenodo.org
    zip
    Updated Nov 15, 2024
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Andrew Felton; Andrew Felton (2024). Storage and Transit Time Data and Code [Dataset]. http://doi.org/10.5281/zenodo.14171251
    Explore at:
    zipAvailable download formats
    Dataset updated
    Nov 15, 2024
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Andrew Felton; Andrew Felton
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Author: Andrew J. Felton
    Date: 11/15/2024

    This R project contains the primary code and data (following pre-processing in python) used for data production, manipulation, visualization, and analysis, and figure production for the study entitled:

    "Global estimates of the storage and transit time of water through vegetation"

    Please note that 'turnover' and 'transit' are used interchangeably. Also please note that this R project has been updated multiple times as the analysis has updated throughout the peer review process.

    #Data information:

    The data folder contains key data sets used for analysis. In particular:

    "data/turnover_from_python/updated/august_2024_lc/" contains the core datasets used in this study including global arrays summarizing five year (2016-2020) averages of mean (annual) and minimum (monthly) transit time, storage, canopy transpiration, and number of months of data able as both an array (.nc) or data table (.csv). These data were produced in python using the python scripts found in the "supporting_code" folder. The remaining files in the "data" and "data/supporting_data" folder primarily contain ground-based estimates of storage and transit found in public databases or through a literature search, but have been extensively processed and filtered here. The "supporting_data"" folder also contains annual (2016-2020) MODIS land cover data used in the analysis and contains separate filters containing the original data (.hdf) and then the final process (filtered) data in .nc format. The resulting annual land cover distributions were used in the pre-processing of data in python.

    #Code information

    Python scripts can be found in the "supporting_code" folder.

    Each R script in this project has a role:

    "01_start.R": This script sets the working directory, loads in the tidyverse package (the remaining packages in this project are called using the `::` operator), and can run two other scripts: one that loads the customized functions (02_functions.R) and one for importing and processing the key dataset for this analysis (03_import_data.R).

    "02_functions.R": This script contains custom functions. Load this using the `source()` function in the 01_start.R script.

    "03_import_data.R": This script imports and processes the .csv transit data. It joins the mean (annual) transit time data with the minimum (monthly) transit data to generate one dataset for analysis: annual_turnover_2. Load this using the
    `source()` function in the 01_start.R script.

    "04_figures_tables.R": This is the main workhouse for figure/table production and supporting analyses. This script generates the key figures and summary statistics used in the study that then get saved in the "manuscript_figures" folder. Note that all maps were produced using Python code found in the "supporting_code"" folder. Also note that within the "manuscript_figures" folder there is an "extended_data" folder, which contains tables of the summary statistics (e.g., quartiles and sample sizes) behind figures containing box plots or depicting regression coefficients.

    "supporting_generate_data.R": This script processes supporting data used in the analysis, primarily the varying ground-based datasets of leaf water content.

    "supporting_process_land_cover.R": This takes annual MODIS land cover distributions and processes them through a multi-step filtering process so that they can be used in preprocessing of datasets in python.

  3. Z

    The dataset for the study of code change patterns in Python

    • data.niaid.nih.gov
    Updated Oct 19, 2021
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Anonymous (2021). The dataset for the study of code change patterns in Python [Dataset]. https://data.niaid.nih.gov/resources?id=zenodo_4004117
    Explore at:
    Dataset updated
    Oct 19, 2021
    Authors
    Anonymous
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    The dataset of Python projects used for the study of code change patterns and their automation. The dataset lists 120 projects, divided into four domains — Web, Media, Data, and ML+DL.

  4. Raw data and python scripts

    • figshare.com
    xlsx
    Updated May 24, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Shane Telfer (2023). Raw data and python scripts [Dataset]. http://doi.org/10.6084/m9.figshare.22682965.v2
    Explore at:
    xlsxAvailable download formats
    Dataset updated
    May 24, 2023
    Dataset provided by
    figshare
    Figsharehttp://figshare.com/
    Authors
    Shane Telfer
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Data for paper 'Hetero-interpenetrated metal-organic frameworks'

  5. q

    Data from: Get Started With GIS in Open Source Python Workshop

    • qubeshub.org
    Updated Oct 15, 2019
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Leah Wasser; Jenny Palomino; Joe McGlinchy (2019). Get Started With GIS in Open Source Python Workshop [Dataset]. https://qubeshub.org/publications/1441
    Explore at:
    Dataset updated
    Oct 15, 2019
    Dataset provided by
    QUBES
    Authors
    Leah Wasser; Jenny Palomino; Joe McGlinchy
    Description

    There are a suite of powerful open source python libraries that can be used to work with spatial data. Learn how to use geopandas, rasterio and matplotlib to plot and manipulate spatial data in Python.

  6. d

    Data from: Burmese python environmental DNA data, and environmental...

    • catalog.data.gov
    • data.usgs.gov
    • +1more
    Updated Nov 21, 2025
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    U.S. Geological Survey (2025). Burmese python environmental DNA data, and environmental covariates, collected from wading bird aggregations and control sites in the Greater Everglades Ecosystem, United States, in 2017 [Dataset]. https://catalog.data.gov/dataset/burmese-python-environmental-dna-data-and-environmental-covariates-collected-from-wading-b
    Explore at:
    Dataset updated
    Nov 21, 2025
    Dataset provided by
    United States Geological Surveyhttp://www.usgs.gov/
    Area covered
    Everglades, United States
    Description

    Environmental DNA (eDNA) water samples were collected at 15 tree islands containing wading bird breeding colonies (order Pelecaniformes) and 15 empty control islands in the central Everglades of Florida in spring of 2017 (April through June) and analyzed for the presence of eDNA from invasive Burmese pythons (Python bivittatus). The Burmese python is now established as a breeding population throughout south Florida, USA. Pythons can consume large quantities of prey and may be a particular threat to wading bird breeding colonies in the Everglades. To quantify python occupancy rates at tree islands where wading birds breed, we utilized environmental DNA (eDNA) analysis—a genetic tool which detects shed DNA in water samples and provides high detection probabilities compared to traditional survey methods. We fitted multi-scale Bayesian occupancy models to test the prediction that Burmese pythons occupy islands with wading bird colonies in the central Everglades at higher rates compared to representative control islands in the same region containing no breeding birds.

  7. Raw data and Python script

    • springernature.figshare.com
    zip
    Updated Jan 21, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Binbin Tu (2025). Raw data and Python script [Dataset]. http://doi.org/10.6084/m9.figshare.25243147.v1
    Explore at:
    zipAvailable download formats
    Dataset updated
    Jan 21, 2025
    Dataset provided by
    figshare
    Figsharehttp://figshare.com/
    Authors
    Binbin Tu
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Raw data and Python script

  8. Data from: Python Code

    • figshare.com
    text/x-python
    Updated Feb 7, 2019
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Paul Jones (2019). Python Code [Dataset]. http://doi.org/10.6084/m9.figshare.6470729.v1
    Explore at:
    text/x-pythonAvailable download formats
    Dataset updated
    Feb 7, 2019
    Dataset provided by
    figshare
    Figsharehttp://figshare.com/
    Authors
    Paul Jones
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Python scripts were used to identify interfaces and compute the vapor film thickness. To run the scripts, the unprocessed TIFF images for both the X-ray and visible light cameras are required. The python scripts are reported here for transparency.Order of code execution:(1) interface.py - Finds liquid-vapor and sphere-fluid interfaces(2) filmThickness.py - Calculates vapor film thickness(3) filmCompare.py - Compares vapor film thickness for each trial(4) make_video.py - Create multi-panel video of results(5) fft_interface.py - Compute discrete fast Fourier transformChange in naming convention: A different naming convention was used for storing and processing data from experimental trials than for reporting. For example, we used the name A1_roughSphere_xray_C1S0001 to denote data corresponding to the first experimental trial of the rough sphere. When reporting this information, we instead used the notation RO1 for simplicity. This was done for the smooth sphere, and thermoprobe as well.

  9. Creating_simple_Sintetic_dataset

    • kaggle.com
    zip
    Updated Jan 20, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Lala Ibadullayeva (2025). Creating_simple_Sintetic_dataset [Dataset]. https://www.kaggle.com/datasets/lalaibadullayeva/creating-simple-sintetic-dataset
    Explore at:
    zip(476698 bytes)Available download formats
    Dataset updated
    Jan 20, 2025
    Authors
    Lala Ibadullayeva
    License

    MIT Licensehttps://opensource.org/licenses/MIT
    License information was derived automatically

    Description

    Dataset Description

    Overview: This dataset contains three distinct fake datasets generated using the Faker and Mimesis libraries. These libraries are commonly used for generating realistic-looking synthetic data for testing, prototyping, and data science projects. The datasets were created to simulate real-world scenarios while ensuring no sensitive or private information is included.

    Data Generation Process: The data creation process is documented in the accompanying notebook, Creating_simple_Sintetic_data.ipynb. This notebook showcases the step-by-step procedure for generating synthetic datasets with customizable structures and fields using the Faker and Mimesis libraries.

    File Contents:

    Datasets: CSV files containing the three synthetic datasets. Notebook: Creating_simple_Sintetic_data.ipynb detailing the data generation process and the code used to create these datasets.

  10. o

    Demographic Analysis Workflow using Census API in Jupyter Notebook:...

    • openicpsr.org
    delimited
    Updated Jul 23, 2020
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Donghwan Gu; Nathanael Rosenheim (2020). Demographic Analysis Workflow using Census API in Jupyter Notebook: 1990-2000 Population Size and Change [Dataset]. http://doi.org/10.3886/E120381V1
    Explore at:
    delimitedAvailable download formats
    Dataset updated
    Jul 23, 2020
    Dataset provided by
    Texas A&M University
    Authors
    Donghwan Gu; Nathanael Rosenheim
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Area covered
    US Counties, Boone County, Kentucky
    Description

    This archive reproduces a table titled "Table 3.1 Boone county population size, 1990 and 2000" from Wang and vom Hofe (2007, p.58). The archive provides a Jupyter Notebook that uses Python and can be run in Google Colaboratory. The workflow uses Census API to retrieve data, reproduce the table, and ensure reproducibility for anyone accessing this archive.The Python code was developed in Google Colaboratory, or Google Colab for short, which is an Integrated Development Environment (IDE) of JupyterLab and streamlines package installation, code collaboration and management. The Census API is used to obtain population counts from the 1990 and 2000 Decennial Census (Summary File 1, 100% data). All downloaded data are maintained in the notebook's temporary working directory while in use. The data are also stored separately with this archive.The notebook features extensive explanations, comments, code snippets, and code output. The notebook can be viewed in a PDF format or downloaded and opened in Google Colab. References to external resources are also provided for the various functional components. The notebook features code to perform the following functions:install/import necessary Python packagesintroduce a Census API Querydownload Census data via CensusAPI manipulate Census tabular data calculate absolute change and percent changeformatting numbersexport the table to csvThe notebook can be modified to perform the same operations for any county in the United States by changing the State and County FIPS code parameters for the Census API downloads. The notebook could be adapted for use in other environments (i.e., Jupyter Notebook) as well as reading and writing files to a local or shared drive, or cloud drive (i.e., Google Drive).

  11. Overwrite Hosted Feature Services, v2.1.4

    • arc-gis-hub-home-arcgishub.hub.arcgis.com
    Updated Apr 16, 2019
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Esri (2019). Overwrite Hosted Feature Services, v2.1.4 [Dataset]. https://arc-gis-hub-home-arcgishub.hub.arcgis.com/content/d45f80eb53c748e7aa3d938a46b48836
    Explore at:
    Dataset updated
    Apr 16, 2019
    Dataset authored and provided by
    Esrihttp://esri.com/
    Description

    Want to keep the data in your Hosted Feature Service current? Not interested in writing a lot of code?Leverage this Python Script from the command line, Windows Scheduled Task, or from within your own code to automate the replacement of data in an existing Hosted Feature Service. It can also be leveraged by your Notebook environment and automatically managed by the MNCD Tool!See the Sampler Notebook that features the OverwriteFS tool run from Online to update a Feature Service. It leverages MNCD to cache the OverwriteFS script for import to the Notebook. A great way to jump start your Feature Service update workflow! RequirementsPython v3.xArcGIS Python APIStored Connection Profile, defined by Python API 'GIS' module. Also accepts 'pro', to specify using the active ArcGIS Pro connection. Will require ArcGIS Pro and Arcpy!Pre-Existing Hosted Feature ServiceCapabilitiesOverwrite a Feature Service, refreshing the Service Item and DataBackup and reapply Service, Layer, and Item properties - New at v2.0.0Manage Service to Service or Service to Data relationships - New at v2.0.0Repair Lost Service File Item to Service Relationships, re-enabling Service Overwrite - New at v2.0.0'Swap Layer' capability for Views, allowing two Services to support a View, acting as Active and Idle role during Updates - New at v2.0.0Data Conversion capability, able to invoke following a download and before Service update - New at v2.0.0Includes 'Rss2Json' Conversion routine, able to read a RSS or GeoRSS source and generate GeoJson for Service Update - New at v2.0.0Renamed 'Rss2Json' to 'Xml2GeoJSON' for its enhanced capabilities, 'Rss2Json' remains for compatability - Revised at v2.1.0Added 'Json2GeoJSON' Conversion routine, able to read and manipulate Json or GeoJSON data for Service Updates - New at v2.1.0Can update other File item types like PDF, Word, Excel, and so on - New at v2.1.0Supports ArcGIS Python API v2.0 - New at v2.1.2RevisionsSep 29, 2021: Long awaited update to v2.0.0!Sep 30, 2021: v2.0.1, Patch to correct Outcome Status when download or Coversion resulted in no change. Also updated documentation.Oct 7, 2021: v2.0.2, workflow Patch correcting Extent update of Views when Overwriting Service, discovered following recent ArcGIS Online update. Enhancements to 'datetimeUtil' Support script.Nov 30, 2021: v2.1.0, added new 'Json2GeoJSON' Converter, enhanced 'Xml2GeoJSON' Converter, retired 'Rss2Json' Converter, added new Option Switches 'IgnoreAge' and 'UpdateTarget' for source age control and QA/QC workflows, revised Optimization logic and CRC comparison on downloads.Dec 1, 2021: v2.1.1, Only a patch to Conversion routines: Corrected handling of null Z-values in Geometries (discovered immediately following release 2.1.0), improve error trapping while processing rows, and added deprecation message to retired 'Rss2Json' conversion routine.Feb 22, 2022: v2.1.2, Patch to detect and re-apply case-insensitive field indexes. Update to allow Swapping Layers to Service without an associated file item. Added cache refresh following updates. Patch to support Python API 2.0 service 'table' property. Patches to 'Json2GeoJSON' and 'Xml2GeoJSON' converter routines.Sep 5, 2024: v2.1.4, Patch service manager refresh failure issue. Added trace report to Convert execution on exception. Set 'ignore-DataItemCheck' property to True when 'GetTarget' action initiated. Hardened Async job status check. Update 'overwriteFeatureService' to support GeoPackage type and file item type when item.name includes a period, updated retry loop to try one final overwrite after del, fixed error stop issue on failed overwrite attempts. Removed restriction on uploading files larger than 2GB. Restores missing 'itemInfo' file on service File items. Corrected false swap success when view has no layers. Lifted restriction of Overwrite/Swap Layers for OGC. Added 'serviceDescription' to service detail backup. Added 'thumbnail' to item backup/restore logic. Added 'byLayerOrder' parameter to 'swapFeatureViewLayers'. Added 'SwapByOrder' action switch. Patch added to overwriteFeatureService 'status' check. Patch for June 2024 update made to 'managers.overwrite' API script that blocks uploads > 25MB, API v2.3.0.3. Patch 'overwriteFeatureService' to correctly identify overwrite file if service has multiple Service2Data relationships.Includes documentation updates!

  12. f

    Data from: Hard Potato: A Python Library to Control Commercial Potentiostats...

    • acs.figshare.com
    zip
    Updated Jun 1, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Oliver Rodríguez; Michael A. Pence; Joaquín Rodríguez-López (2023). Hard Potato: A Python Library to Control Commercial Potentiostats and to Automate Electrochemical Experiments [Dataset]. http://doi.org/10.1021/acs.analchem.2c04862.s003
    Explore at:
    zipAvailable download formats
    Dataset updated
    Jun 1, 2023
    Dataset provided by
    ACS Publications
    Authors
    Oliver Rodríguez; Michael A. Pence; Joaquín Rodríguez-López
    License

    Attribution-NonCommercial 4.0 (CC BY-NC 4.0)https://creativecommons.org/licenses/by-nc/4.0/
    License information was derived automatically

    Description

    Here, we develop and show the use of an open-source Python library to control commercial potentiostats. It standardizes the commands for different potentiostat models, opening the possibility to perform automated experiments independently of the instrument used. At the time of this writing, we have included potentiostats from CH Instruments (models 1205B, 1242B, 601E, and 760E) and PalmSens (model Emstat Pico), although the open-source nature of the library allows for more to be included in the future. To showcase the general workflow and implementation of a real experiment, we have automated the Randles–Ševčı́k methodology to determine the diffusion coefficient of a redox-active species in solution using cyclic voltammetry. This was accomplished by writing a Python script that includes data acquisition, data analysis, and simulation. The total run time was 1 min and 40 s, well below the time it would take even an experienced electrochemist to apply the methodology in a traditional manner. Our library has potential applications that expand beyond the automation of simple repetitive tasks; for example, it can interface with peripheral hardware and well-established third-party Python libraries as part of a more complex and intelligent setup that relies on laboratory automation, advanced optimization, and machine learning.

  13. H

    Using Python and Jupyter Notebook to Retrieve and Visualize the Water...

    • hydroshare.org
    zip
    Updated Apr 21, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Ali Farshid (2022). Using Python and Jupyter Notebook to Retrieve and Visualize the Water Temperature Data of the Logan River, Utah [Dataset]. https://www.hydroshare.org/resource/8c565dc2f9244182a575f91515e83d1d
    Explore at:
    zip(358.6 MB)Available download formats
    Dataset updated
    Apr 21, 2022
    Dataset provided by
    HydroShare
    Authors
    Ali Farshid
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Time period covered
    Jan 1, 2014 - Dec 18, 2021
    Area covered
    Description

    The mainstem Logan River is a suitable habitat for cold-water fishes such as native populations of cutthroat trout (Budy & Gaeta, 2018). On the other hand, high water temperatures can harm cold-water fish populations by creating physiological stresses, intensifying metabolic demands, and limiting suitable habitats (Williams & et al., 2015). In this regard, the State of Utah Department of Environmental Quality (UDEQ) has identified the Logan River as a suitable habitat for cold-water species, which can become unsuitable when the water temperature rises higher than 20 degrees Celsius (Rule R317-2, 2022). However, the UDEQ does not provide any details on how to evaluate the violations from the standard. One way to evaluate violations is to look at water temperature distributions (i.e., histograms) along the river from high elevations to low elevations at different locations. In this report, I used three different Python libraries to manipulate, extract, and explore the water temperature data of the Logan River from 2014 to 2021 obtained from the Logan River Observatory website. The results (i.e., the generated histograms by executing Jupyter Notebook in the HydroShare environment) show that the Logan River tends to experience higher water temperatures as its elevation drops regardless of the season. This can provide some insights for the UDEQ to simultaneously consider space and time in assessing violations from the standard.

  14. D

    Replication Data for: GPRat: Gaussian Process Regression with Asynchronous...

    • darus.uni-stuttgart.de
    Updated Apr 2, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Maksim Helmann; Alexander Strack; Dirk Pflüger (2025). Replication Data for: GPRat: Gaussian Process Regression with Asynchronous Tasks [Dataset]. http://doi.org/10.18419/DARUS-4743
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Apr 2, 2025
    Dataset provided by
    DaRUS
    Authors
    Maksim Helmann; Alexander Strack; Dirk Pflüger
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    This repository complements the identically titled paper submitted to WAMTA 2025 and allows to reproduce the published results. For a more description please consider the README.md file.

  15. Z

    MASCDB, a database of images, descriptors and microphysical properties of...

    • data.niaid.nih.gov
    • springerprofessional.de
    • +2more
    Updated Jul 5, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Grazioli, Jacopo; Ghiggi, Gionata; Berne, Alexis (2023). MASCDB, a database of images, descriptors and microphysical properties of individual snowflakes in free fall [Dataset]. https://data.niaid.nih.gov/resources?id=ZENODO_5578920
    Explore at:
    Dataset updated
    Jul 5, 2023
    Dataset provided by
    EPFL-ENAC-IIE-LTE
    Authors
    Grazioli, Jacopo; Ghiggi, Gionata; Berne, Alexis
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Dataset overview

    This dataset provides data and images of snowflakes in free fall collected with a Multi-Angle Snowflake Camera (MASC) The dataset includes, for each recorded snowflakes:

    A triplet of gray-scale images corresponding to the three cameras of the MASC

    A large quantity of geometrical, textural descriptors and the pre-compiled output of published retrieval algorithms as well as basic environmental information at the location and time of each measurement.

    The pre-computed descriptors and retrievals are available either individually for each camera view or, some of them, available as descriptors of the triplet as a whole. A non exhaustive list of precomputed quantities includes for example:

    Textural and geometrical descriptors as in Praz et al 2017

    Hydrometeor classification, riming degree estimation, melting identification, as in Praz et al 2017

    Blowing snow identification, as in Schaer et al 2020

    Mass, volume, gyration estimation, as in Leinonen et al 2021

    Data format and structure

    The dataset is divided into four .parquet file (for scalar descriptors) and a Zarr database (for the images). A detailed description of the data content and of the data records is available here.

    Supporting code

    A python-based API is available to manipulate, display and organize the data of our dataset. It can be found on GitHub. See also the code documentation on ReadTheDocs.

    Download notes

    All files available here for download should be stored in the same folder, if the python-based API is used

    MASCdb.zarr.zip must be unzipped after download

    Field campaigns

    A list of campaigns included in the dataset, with a minimal description is given in the following table

        Campaign_name
        Information
    

    Shielded / Not shielded

    DFIR = Double Fence Intercomparison Reference

    APRES3-2016 & APRES3-2017

        Instrument installed in Antarctica in the context of the APRES3 project. See for example Genthon et al, 2018 or Grazioli et al 2017
        Not shielded
    
    
        Davos-2015
        Instrument installed in the Swiss Alps within the context of SPICE (Solid Precipitation InterComparison Experiment)
        Shielded (DFIR)
    
    
        Davos-2019
        Instrument installed in the Swiss Alps within the context of RACLETS (Role of Aerosols and CLouds Enhanced by Topography on Snow)
        Not shielded
    
    
        ICEGENESIS-2021
        Instrument installed in the Swiss Jura in a MeteoSwiss ground measurement site, within the context of ICE-GENESIS. See for example Billault-Roux et al, 2023
        Not shielded
    
    
        ICEPOP-2018
        Instrument installed in Korea, in the context of ICEPOP. See for example Gehring et al 2021.
        Shielded (DFIR)
    
    
        Jura-2019 & Jura-2023
        Instrument installed in the Swiss Jura within a MeteoSwiss measurement site
        Not shielded
    
    
        Norway-2016
        Instrument installed in Norway during the High-Latitude Measurement of Snowfall (HiLaMS). See for example Cooper et al, 2022.
        Not shielded
    
    
        PLATO-2019
        Instrument installed in the "Davis" Antarctic base during the PLATO field campaign
        Not shielded
    
    
        POPE-2020
        Instrument installed in the "Princess Elizabeth Antarctica" base during the POPE campaign. See for example Ferrone et al, 2023.
        Not shielded
    
    
        Remoray-2022
        Instrument installed in the French Jura.
        Not shielded
    
    
        Valais-2016
        Instrument installed in the Swiss Alps in a ski resort.
        Not shielded
    

    Version

    1.0 - Two new campaigns ("Jura-2023", "Norway-2016") added. Added references and list of campaigns.

    0.3 - a new campaign is added to the dataset ("Remoray-2022")

    0.2 - rename of variables. Variable precision (digits) standardized

    0.1 - first upload

  16. H

    Techniques for Increased Automation of Aquatic Sensor Data Post Processing...

    • hydroshare.org
    • search.dataone.org
    zip
    Updated Sep 7, 2021
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Amber Spackman Jones; Jeffery S. Horsburgh; Tannner Jones (2021). Techniques for Increased Automation of Aquatic Sensor Data Post Processing in Python: Video Presentation [Dataset]. https://www.hydroshare.org/resource/bc5c616426214b60b068352ae028d963
    Explore at:
    zip(351.0 MB)Available download formats
    Dataset updated
    Sep 7, 2021
    Dataset provided by
    HydroShare
    Authors
    Amber Spackman Jones; Jeffery S. Horsburgh; Tannner Jones
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    This resource contains a video recording for a presentation given as part of the National Water Quality Monitoring Council conference in April 2021. The presentation covers the motivation for performing quality control for sensor data, the development of PyHydroQC, a Python package with functions for automating sensor quality control including anomaly detection and correction, and the performance of the algorithms applied to data from multiple sites in the Logan River Observatory.

    The initial abstract for the presentation: Water quality sensors deployed to aquatic environments make measurements at high frequency and commonly include artifacts that do not represent the environmental phenomena targeted by the sensor. Sensors are subject to fouling from environmental conditions, often exhibit drift and calibration shifts, and report anomalies and erroneous readings due to issues with datalogging, transmission, and other unknown causes. The suitability of data for analyses and decision making often depend on subjective and time-consuming quality control processes consisting of manual review and adjustment of data. Data driven and machine learning techniques have the potential to automate identification and correction of anomalous data, streamlining the quality control process. We explored documented approaches and selected several for implementation in a reusable, extensible Python package designed for anomaly detection for aquatic sensor data. Implemented techniques include regression approaches that estimate values in a time series, flag a point as anomalous if the difference between the sensor measurement exceeds a threshold, and offer replacement values for correcting anomalies. Additional algorithms that scaffold the central regression approaches include rules-based preprocessing, thresholds for determining anomalies that adjust with data variability, and the ability to detect and correct anomalies using forecasted and backcasted estimation. The techniques were developed and tested based on several years of data from aquatic sensors deployed at multiple sites in the Logan River Observatory in northern Utah, USA. Performance was assessed based on labels and corrections applied previously by trained technicians. In this presentation, we describe the techniques for detection and correction, report their performance, illustrate the workflow for applying to high frequency aquatic sensor data, and demonstrate the possibility for additional approaches to help increase automation of aquatic sensor data post processing.

  17. D

    HVAC Edge Controller With Python Runtime Market Research Report 2033

    • dataintelo.com
    csv, pdf, pptx
    Updated Oct 1, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Dataintelo (2025). HVAC Edge Controller With Python Runtime Market Research Report 2033 [Dataset]. https://dataintelo.com/report/hvac-edge-controller-with-python-runtime-market
    Explore at:
    pdf, csv, pptxAvailable download formats
    Dataset updated
    Oct 1, 2025
    Dataset authored and provided by
    Dataintelo
    License

    https://dataintelo.com/privacy-and-policyhttps://dataintelo.com/privacy-and-policy

    Time period covered
    2024 - 2032
    Area covered
    Global
    Description

    HVAC Edge Controller with Python Runtime Market Outlook



    According to our latest research, the global HVAC Edge Controller with Python Runtime market size reached USD 1.42 billion in 2024, reflecting the rapid digital transformation in building automation and intelligent climate control. The market is expected to grow at a robust CAGR of 17.8% from 2025 to 2033, with the total market projected to reach USD 6.14 billion by 2033. This surge is primarily driven by the increasing adoption of IoT-enabled HVAC solutions, the need for real-time data analytics, and the integration of programmable runtimes such as Python for enhanced system customization and interoperability.




    The growth of the HVAC Edge Controller with Python Runtime market is significantly influenced by the ongoing digitalization of commercial and industrial infrastructure worldwide. As organizations strive to optimize energy consumption, reduce operational costs, and meet stringent environmental regulations, the demand for advanced edge controllers capable of running Python scripts has escalated. These controllers enable seamless integration with IoT devices, advanced analytics, and AI-driven applications, allowing for smarter and more adaptive control of HVAC systems. The flexibility offered by Python runtime empowers facility managers and system integrators to develop custom algorithms for predictive maintenance, energy optimization, and fault detection, further accelerating market adoption.




    Another major growth factor is the rapid expansion of smart building initiatives, particularly in developed regions such as North America and Europe. Governments and private sector entities are investing heavily in smart infrastructure, which necessitates the deployment of intelligent HVAC solutions to ensure occupant comfort, energy efficiency, and regulatory compliance. The ability of HVAC edge controllers with Python runtime to interface with legacy systems and modern cloud platforms makes them an attractive choice for both retrofit and new construction projects. Moreover, the proliferation of edge computing paradigms in building automation is fueling demand for controllers that can process data locally, minimize latency, and enhance system reliability.




    The market is also benefiting from the rising adoption of cloud-based deployment models and wireless connectivity options. As remote monitoring and management become essential in the post-pandemic era, organizations are increasingly leveraging cloud-enabled HVAC edge controllers to gain real-time visibility and control over distributed assets. The Python runtime environment, in particular, allows for rapid application development and integration with third-party services, enabling a wide range of use cases from simple automation tasks to complex machine learning-driven optimizations. This trend is expected to continue as the ecosystem of Python-based libraries and frameworks for building automation expands.




    Regionally, Asia Pacific is emerging as the fastest-growing market for HVAC Edge Controller with Python Runtime solutions, driven by urbanization, industrialization, and the proliferation of smart city projects. Countries like China, Japan, and India are witnessing significant investments in infrastructure modernization, which is translating into increased demand for intelligent HVAC control systems. Meanwhile, North America maintains a dominant market share due to its mature building automation sector and early adoption of edge computing technologies. Europe is also a key market, characterized by stringent energy efficiency regulations and a strong focus on sustainability in the built environment.



    Component Analysis



    The component segment of the HVAC Edge Controller with Python Runtime market is divided into hardware, software, and services, each playing a pivotal role in shaping the overall ecosystem. Hardware forms the backbone of edge controllers, encompassing microprocessors, sensors, communication modules, and interface boards that enable real-time data acquisition and control. With the increasing complexity of building automation requirements, hardware providers are focusing on delivering robust, scalable, and energy-efficient platforms capable of running Python scripts natively. This hardware evolution is critical to ensuring compatibility with a wide range of HVAC equipment and facilitating seamless integration with both legacy and modern systems.



    <

  18. Lock-in amplifier data collected using Python instrument control suite of...

    • figshare.com
    zip
    Updated May 31, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Lucas Koerner (2023). Lock-in amplifier data collected using Python instrument control suite of Bluesky and instrbuilder [Dataset]. http://doi.org/10.6084/m9.figshare.7768352.v1
    Explore at:
    zipAvailable download formats
    Dataset updated
    May 31, 2023
    Dataset provided by
    figshare
    Figsharehttp://figshare.com/
    Authors
    Lucas Koerner
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    The archived data 'data_figshare.zip' produces figures 3-9 of the paper under review. Data includes .sqlite databases for each run, .json files that describe run starts, run stops, and events. An archive of numpy archives ('numpy_arrays.zip') that store captured time sequences (arrays) is also included, but is a much larger file and is only needed if modified post processing is to be applied on the ADA2200 data. See acquisition and analysis code for more details including information on how to configure databroker to run the analysis code. The attached local_file.yml should be placed into ~/.config/databroker/ and the placeholder 'your_directory' must be modified to point to the data_figshare directory.https://github.com/lucask07/instrbuilder/tree/master/instrbuilder/bluesky_demo/lockin_analysis(no legal or ethical requirements)

  19. d

    (HS 17) Automate Workflows using Jupyter notebook to create Large Spatial...

    • search.dataone.org
    • dataone.org
    Updated Dec 30, 2023
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Young-Don Choi (2023). (HS 17) Automate Workflows using Jupyter notebook to create Large Spatial Sample Datasets [Dataset]. https://search.dataone.org/view/sha256%3A031befe4052e42a42b569cdcb0e76542e5c5b163dbf4480db9d1a52481071759
    Explore at:
    Dataset updated
    Dec 30, 2023
    Dataset provided by
    Hydroshare
    Authors
    Young-Don Choi
    Description

    For the automated workflows, we create Jupyter notebooks for each state. In these workflows, GIS processing to merge, extract and project GeoTIFF data was the most important process. For this process, we used ArcPy which is a python package to perform geographic data analysis, data conversion, and data management in ArcGIS (Toms, 2015). After creating state-scale LSS datasets in GeoTIFF format, we convert GeoTIFF to NetCDF using xarray and rioxarray Python packages. Xarray is a Python package to work with multi-dimensional arrays and rioxarray is rasterio xarray extension. Rasterio is a Python library to read and write GeoTIFF and other raster formats. We used xarray to manipulate data type and add metadata in NetCDF file and rioxarray to save GeoTIFF to NetCDF format. Through these procedures, we created three composite HyddroShare resources to share state-scale LSS datasets. Due to the limitation of ArcGIS Pro license which is a commercial GIS software, we developed this Jupyter notebook on Windows OS.

  20. c

    Code supporting "Control data, Sankey diagrams, and exergy: Assessing the...

    • repository.cam.ac.uk
    bin, zip
    Updated Jan 23, 2019
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Gonzalez Hernandez, A; Lupton, Rick (2019). Code supporting "Control data, Sankey diagrams, and exergy: Assessing the resource efficiency of industrial plants" [Dataset]. http://doi.org/10.17863/CAM.20556
    Explore at:
    zip(15185 bytes), zip(13216 bytes), bin(1813 bytes), zip(37941 bytes), bin(1101 bytes)Available download formats
    Dataset updated
    Jan 23, 2019
    Dataset provided by
    University of Cambridge
    Apollo
    Authors
    Gonzalez Hernandez, A; Lupton, Rick
    License

    MIT Licensehttps://opensource.org/licenses/MIT
    License information was derived automatically

    Description

    This folder contains the Python code that was used to process the resource flow data of a basic oxygen steelmaking plant owned by Tata Steel.

Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
National Institute of Standards and Technology (2025). cmomy: A python package to calculate and manipulate Central (co)moments. [Dataset]. https://catalog.data.gov/dataset/cmomy-a-python-package-to-calculate-and-manipulate-central-comoments
Organization logo

cmomy: A python package to calculate and manipulate Central (co)moments.

Explore at:
Dataset updated
Sep 30, 2025
Dataset provided by
National Institute of Standards and Technologyhttp://www.nist.gov/
Description

cmomy is a python package to calculate central moments and co-moments in a numerical stable and direct way. Behind the scenes, cmomy makes use of Numba to rapidly calculate moments. cmomy provides utilities to calculate central moments from individual samples, precomputed central moments, and precomputed raw moments. It also provides routines to perform bootstrap resampling based on raw data, or precomputed moments. cmomy has numpy array and xarray DataArray interfaces.

Search
Clear search
Close search
Google apps
Main menu