Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Contains the DGM5 geological model and the VELMOD 3.1 velocity model as xarray datasets in UTM31 coordinates.
Original data:
Format:
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
This repository contains data for three different experiments presented in the paper:
(1) moose_feet (40 files): The moose leg experiments are labeled as ax_y.nc,
where 'a' indicates attached digits and 'f' indicates free digits. The
number 'x' is either 1 (front leg) or 2 (hind leg), and the number 'y'
is an increment from 0 to 9 representing the 10 samples of each set.
(2) synthetic_feet (120 files): The synthetic feet experiments are labeled
as lw_a_y.nc, where 'lw' (Low Water content) can be replaced by 'mw'
(Medium Water content) or 'vw' (Vast Water content). The 'a' can be 'o'
(Original Go1 foot), 'r' (Rigid extended foot), 'f' (Free digits anisotropic
foot), or 'a' (Attached digits). Similar to (1), the last number is an increment from 0 to 9.
(3) Go1 (15 files): The locomotion experiments of the quadruped robot on the
track are labeled as condition_y.nc, where 'condition' is either 'hard_ground'
for experiments on hard ground, 'bioinspired_feet' for the locomotion of the
quadruped on mud using bio-inspired anisotropic feet, or 'original_feet' for
experiments where the robot used the original Go1 feet. The 'y' is an increment from 0 to 4.
The files for moose_feet and synthetic_feet contain timestamp (s), position (m), and force (N) data.
The files for Go1 contain timestamp (s), position (rad), velocity (rad/s), torque (Nm) data for all 12 motors, and the distance traveled by the robot (m).
All files can be read using xarray datasets (https://docs.xarray.dev/en/stable/generated/xarray.Dataset.html).
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Fronts-toolbox is a collection of tools to detect oceanic fronts in Python.
Some front-detection algorithms are complex and thus may perform poorly when written directly in Python.
This library provides a framework of Numba accelerated functions that can be applied easily to Numpy arrays, Dask arrays, or Xarray data.
It could also support Cuda arrays if necessary.
This makes creating and modifying those functions easier (especially for non-specialists) than if they were written in Fortran or C extensions.
The data in this repository is to be used to test and showcase the various algorithms.
Renamed the "Unindexed dimensions" section in the Dataset
and DataArray
repr (added in v0.9.0) to "Dimensions without coordinates".
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
This repository contains the data cube (an xarray DataArray) used in Jiang et al. 2023 Revisiting ε Eridani with NEID: Identifying New Activity-Sensitive Lines in a Young K Dwarf Star (in press). The cube contains all line parameters (centroid, depth, FWHM, and integrated flux) for each line in the compiled line list over 32 NEID observations of ε Eridani spanning a six-month period from September 2021 to February 2022, as well as the measured RV and activity indices for each observation. For information on how the line parameters are measured, see the paper.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
'## This item has been replaced by the one which can be found at https://datashare.ed.ac.uk/handle/10283/4849 - https://doi.org/10.7488/ds/3843 ##' This archive contains the driving data and selected model outputs to accompany the manuscript: 'Resolving scale-variance in the carbon dynamics of fragmented, mixed-use landscapes estimated using Model-Data Fusion', submitted to Biogeosciences Discussions. The archive contains two zip files containing: (i) the observations and driving data assimilated into CARDAMOM; and (ii) a selection of model output, including the carbon (C) stocks for each DALEC pool, and a compilation of key C fluxes. Data and model output are stored as netcdf files. The xarray package (https://docs.xarray.dev/en/stable/index.html) provides a convenient starting point for using netcdf files within python environments. More details are provided in the document 'Milodowski_etal_dataset_description.pdf'
We implemented automated workflows using Jupyter notebooks for each state. The GIS processing, crucial for merging, extracting, and projecting GeoTIFF data, was performed using ArcPy—a Python package for geographic data analysis, conversion, and management within ArcGIS (Toms, 2015). After generating state-scale LES (large extent spatial) datasets in GeoTIFF format, we utilized the xarray and rioxarray Python packages to convert GeoTIFF to NetCDF. Xarray is a Python package to work with multi-dimensional arrays and rioxarray is rasterio xarray extension. Rasterio is a Python library to read and write GeoTIFF and other raster formats. Xarray facilitated data manipulation and metadata addition in the NetCDF file, while rioxarray was used to save GeoTIFF as NetCDF. These procedures resulted in the creation of three HydroShare resources (HS 3, HS 4 and HS 5) for sharing state-scale LES datasets. Notably, due to licensing constraints with ArcGIS Pro, a commercial GIS software, the Jupyter notebook development was undertaken on a Windows OS.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
96-well plate z-stack of zebrafish with lyz:EGFP expressing neutrophils acquired with a multi-camera array microscope (MCAM)(Ramona Optics Inc., Durham, NC, USA). Mesh well inserts are used and half of the zebrafish on the plate were injected with csf3r morpholino. The overall z-stack is broken into four files.
HDF5 files can be opened using open source Python software: https://docs.xarray.dev/
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Author: Andrew J. FeltonDate: 5/5/2024
This R project contains the primary code and data (following pre-processing in python) used for data production, manipulation, visualization, and analysis and figure production for the study entitled:
"Global estimates of the storage and transit time of water through vegetation"
Please note that 'turnover' and 'transit' are used interchangeably in this project.
Data information:
The data folder contains key data sets used for analysis. In particular:
"data/turnover_from_python/updated/annual/multi_year_average/average_annual_turnover.nc" contains a global array summarizing five year (2016-2020) averages of annual transit, storage, canopy transpiration, and number of months of data. This is the core dataset for the analysis; however, each folder has much more data, including a dataset for each year of the analysis. Data are also available is separate .csv files for each land cover type. Oterh data can be found for the minimum, monthly, and seasonal transit time found in their respective folders. These data were produced using the python code found in the "supporting_code" folder given the ease of working with .nc and EASE grid in the xarray python module. R was used primarily for data visualization purposes. The remaining files in the "data" and "data/supporting_data"" folder primarily contain ground-based estimates of storage and transit found in public databases or through a literature search, but have been extensively processed and filtered here.
Python scripts can be found in the "supporting_code" folder.
Each R script in this project has a particular function:
01_start.R: This script loads the R packages used in the analysis, sets thedirectory, and imports custom functions for the project. You can also load in the main transit time (turnover) datasets here using the source()
function.
02_functions.R: This script contains the custom function for this analysis, primarily to work with importing the seasonal transit data. Load this using the source()
function in the 01_start.R script.
03_generate_data.R: This script is not necessary to run and is primarilyfor documentation. The main role of this code was to import and wranglethe data needed to calculate ground-based estimates of aboveground water storage.
04_annual_turnover_storage_import.R: This script imports the annual turnover andstorage data for each landcover type. You load in these data from the 01_start.R scriptusing the source()
function.
05_minimum_turnover_storage_import.R: This script imports the minimum turnover andstorage data for each landcover type. Minimum is defined as the lowest monthlyestimate.You load in these data from the 01_start.R scriptusing the source()
function.
06_figures_tables.R: This is the main workhouse for figure/table production and supporting analyses. This script generates the key figures and summary statistics used in the study that then get saved in the manuscript_figures folder. Note that allmaps were produced using Python code found in the "supporting_code"" folder.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Test data for ASTE Release 1 integration with ECCOv4-py.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
96-well plate z-stack of zebrafish with lyz:EGFP expressing neutrophils acquired with a multi-camera array microscope (MCAM)(Ramona Optics Inc., Durham, NC, USA). Zebrafish larvae have been injected with csf3r morpholino.
HDF5 files can be opened using open source Python software: https://docs.xarray.dev/
This resource includes materials for two workshops: (1) FAIR Data Management and (2) Advanced Application of Python for Hydrology and Scientific Storytelling, both prepared for presentation at the NWC Summer Institute BootCamp 2024.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
This is a dataset of Sentinel-1 radiometric terrain corrected (RTC) imagery processed by the Alaska Satellite Facility covering a region within the Central Himalaya. It accompanies a tutorial demonstrating accessing and working with Sentinel-1 RTC imagery using xarray and other open source python packages.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Testing files for the xesmf remapping package.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
This resource includes materials for two workshops: (1) FAIR data management and collaborating on simulation data in the cloud (2) Advanced application of Python for working with high value environmental datasets (3) Configuring and running a NextGen simulation and analyzing model outputs
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
We present a globally consistent, satellite-derived dataset of CO_2 enhancement (ΔXCO_2), quantifying the spatially resolved excess in atmospheric CO_2 concentrations as a collective consequence of anthropogenic emissions and terrestrial carbon uptake. This dataset is generated from the deviations of NASA's OCO-3 satellite retrievals comprising 54 million observations across more than 200 countries from 2019 to 2023.
Dear reviewers, please download the datasets here and access using the password enclosed in the review documents. Many thanks!
Data Descriptions -----------------------------------------
# install pre-requests
! pip install netcdf4
! pip install h5netcdf# read co2 enhancement data
import xarray as xr
fn = './CO2_Enhancements_Global.nc'
data = xr.open_dataset(fn)
type(data)
Please cite at least one of the following for any use of the CO2E dataset.
Zhou, Y.*, Fan, P., Liu, J., Xu, Y., Huang, B., Webster, C. (2025). GloCE v1.0: Global CO2 Enhancement Dataset 2019-2023 [Data set]. Zenodo. https://doi.org/10.5281/zenodo.15209825
Fan, P., Liu, J., Xu, Y., Huang, B., Webster, C., & Zhou, Y*. (Under Review) A global dataset of CO2 enhancements during 2019-2023.
For any data inquiries, please email Yulun Zhou at yulunzhou@hku.hk.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
The model build using these datasets can be found at https://github.com/AstexUK/ESP_DNN/tree/master/esp_dnnThe dataset themselves can be opened using xarray Python library (http://xarray.pydata.org/en/stable/#)
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
This Zenodo repository contains all migration flow estimates associated with the paper "Deep learning four decades of human migration." Evaluation code, training data, trained neural networks, and smaller flow datasets are available in the main GitHub repository, which also provides detailed instructions on data sourcing. Due to file size limits, the larger datasets are archived here.
Data is available in both NetCDF (.nc
) and CSV (.csv
) formats. The NetCDF format is more compact and pre-indexed, making it suitable for large files. In Python, datasets can be opened as xarray.Dataset
objects, enabling coordinate-based data selection.
Each dataset uses the following coordinate conventions:
The following data files are provided:
T
summed over Birth ISO). Dimensions: Year, Origin ISO, Destination ISOAdditionally, two CSV files are provided for convenience:
imm
: Total immigration flowsemi
: Total emigration flowsnet
: Net migrationimm_pop
: Total immigrant population (non-native-born)emi_pop
: Total emigrant population (living abroad)mig_prev
: Total origin-destination flowsmig_brth
: Total birth-destination flows, where Origin ISO
reflects place of birthEach dataset includes a mean
variable (mean estimate) and a std
variable (standard deviation of the estimate).
An ISO3 conversion table is also provided.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Next Day Wildfire Spread Dataset
This dataset is an xarray version of the original Next Day Wildfire Spread dataset. It comes in three splits: train, eval and test. Note: Given the original dataset does not contain spatio-temporal information, the xarray coordinates has been set to arbitrary ranges (0-63 for spatial dimensions and 0-number_of_samples for the temporal dimension).
Example
To open a train split of the dataset and show an elevation plot at time=2137:… See the full description on the dataset page: https://huggingface.co/datasets/TheRootOf3/next-day-wildfire-spread.
Attribution-NonCommercial 4.0 (CC BY-NC 4.0)https://creativecommons.org/licenses/by-nc/4.0/
License information was derived automatically
This dataset provides simulated data on plastic and substance flows and stocks in buildings and infrastructure as described in the data article "Plastics in the German Building and Infrastructure Sector: A High-Resolution Dataset on Historical Flows, Stocks, and Legacy Substance Contamination". Besides simulated data, the repository contains input data and model files used to produce the simulated data.
Data & Data Visualization: The dataset contains input data and simulated data for the six main plastic applications in buildings and infrastructure in Germany in the period from 1950 to 2023, which are profiles, flooring, pipes, insulation material, cable insulations, and films. For each application the data are provided in a sub-directory (1_ ... 6_) following the structure described below.
Input Data:
The input data are stored in an xlsx-file with three sheets: flows, parameters, and data quality assessment. The data sources for all input data are detailed in the Supplementary Material of the linked Data in Brief article.
Simulated Data:
Simulated data are stored in a sub-folder, which contains:
Note: All files in the [product]/simulated_data folder are automatically replaced with updated model results upon execution of immec_dmfa_calculate_submodels.py.
To reduce storage requirements, data are stored in gzipped pickle files (.pkl.gz), while smaller files are provided as pickle files (.pkl). To open the files, users can use Python with the following code snippet:
import gzip
# Load a gzipped pickle file
with gzip.open("filename.pkl.gz", "rb") as f:
data = pickle.load(f)
# Load a regular pickle file
with open("filename.pkl", "rb") as f:
data = pickle.load(f)
Please note that opening pickle files requires compatible versions of numpy
and pandas
, as the files may have been created using version-specific data structures. If you encounter errors, ensure your package versions match those used during file creation (pandas: 2.2.3, numpy: 2.2.4).
Simulated data are provided as Xarray datasets, a data structure designed for efficient handling, analysis, and visualization of multi-dimensional labeled data. For more details on using Xarray, please refer to the official documentation: https://docs.xarray.dev/en/stable/
Core Model Files:
Computational Considerations:
During model execution, large arrays are generated, requiring significant memory. To enable computation on standard computers, Monte Carlo simulations are split into multiple chunks:
Dependencies
The model relies on the ODYM framework. To run the model, ODYM must be downloaded from https://github.com/IndEcol/ODYM (S. Pauliuk, N. Heeren, ODYM — An open software framework for studying dynamic material systems: Principles, implementation, and data structures, Journal of Industrial Ecology 24 (2020) 446–458. https://doi.org/10.1111/jiec.12952.)
7_Model_Structure:
8_Additional_Data: This folder contains supplementary data used in the model, including substance concentrations, data quality assessment scores, open-loop recycling distributions, and lifetime distributions.
The dataset was generated using a dynamic material flow analysis (dMFA) model. For a complete methodology description, refer to the Data in Brief article (add DOI).
If you use this dataset, please cite: Schmidt, S., Verni, X.-F., Gibon, T., Laner, D. (2025). Dataset for: Plastics in the German Building and Infrastructure Sector: A High-Resolution Dataset on Historical Flows, Stocks, and Legacy Substance Contamination, Zenodo. DOI: 10.5281/zenodo.15049210
This dataset is licensed under CC BY-NC 4.0, permitting use, modification, and distribution for non-commercial purposes, provided that proper attribution is given.
For questions or further details, please contact:
Sarah Schmidt
Center for Resource Management and Solid Waste Engineering
University of Kassel
Email: sarah.schmidt@uni-kassel.de
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Contains the DGM5 geological model and the VELMOD 3.1 velocity model as xarray datasets in UTM31 coordinates.
Original data:
Format: