29 datasets found
  1. H

    Time-Series Matrix (TSMx): A visualization tool for plotting multiscale...

    • dataverse.harvard.edu
    Updated Jul 8, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Georgios Boumis; Brad Peter (2024). Time-Series Matrix (TSMx): A visualization tool for plotting multiscale temporal trends [Dataset]. http://doi.org/10.7910/DVN/ZZDYM9
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Jul 8, 2024
    Dataset provided by
    Harvard Dataverse
    Authors
    Georgios Boumis; Brad Peter
    License

    CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
    License information was derived automatically

    Description

    Time-Series Matrix (TSMx): A visualization tool for plotting multiscale temporal trends TSMx is an R script that was developed to facilitate multi-temporal-scale visualizations of time-series data. The script requires only a two-column CSV of years and values to plot the slope of the linear regression line for all possible year combinations from the supplied temporal range. The outputs include a time-series matrix showing slope direction based on the linear regression, slope values plotted with colors indicating magnitude, and results of a Mann-Kendall test. The start year is indicated on the y-axis and the end year is indicated on the x-axis. In the example below, the cell in the top-right corner is the direction of the slope for the temporal range 2001–2019. The red line corresponds with the temporal range 2010–2019 and an arrow is drawn from the cell that represents that range. One cell is highlighted with a black border to demonstrate how to read the chart—that cell represents the slope for the temporal range 2004–2014. This publication entry also includes an excel template that produces the same visualizations without a need to interact with any code, though minor modifications will need to be made to accommodate year ranges other than what is provided. TSMx for R was developed by Georgios Boumis; TSMx was originally conceptualized and created by Brad G. Peter in Microsoft Excel. Please refer to the associated publication: Peter, B.G., Messina, J.P., Breeze, V., Fung, C.Y., Kapoor, A. and Fan, P., 2024. Perspectives on modifiable spatiotemporal unit problems in remote sensing of agriculture: evaluating rice production in Vietnam and tools for analysis. Frontiers in Remote Sensing, 5, p.1042624. https://www.frontiersin.org/journals/remote-sensing/articles/10.3389/frsen.2024.1042624 TSMx sample chart from the supplied Excel template. Data represent the productivity of rice agriculture in Vietnam as measured via EVI (enhanced vegetation index) from the NASA MODIS data product (MOD13Q1.V006). TSMx R script: # import packages library(dplyr) library(readr) library(ggplot2) library(tibble) library(tidyr) library(forcats) library(Kendall) options(warn = -1) # disable warnings # read data (.csv file with "Year" and "Value" columns) data <- read_csv("EVI.csv") # prepare row/column names for output matrices years <- data %>% pull("Year") r.names <- years[-length(years)] c.names <- years[-1] years <- years[-length(years)] # initialize output matrices sign.matrix <- matrix(data = NA, nrow = length(years), ncol = length(years)) pval.matrix <- matrix(data = NA, nrow = length(years), ncol = length(years)) slope.matrix <- matrix(data = NA, nrow = length(years), ncol = length(years)) # function to return remaining years given a start year getRemain <- function(start.year) { years <- data %>% pull("Year") start.ind <- which(data[["Year"]] == start.year) + 1 remain <- years[start.ind:length(years)] return (remain) } # function to subset data for a start/end year combination splitData <- function(end.year, start.year) { keep <- which(data[['Year']] >= start.year & data[['Year']] <= end.year) batch <- data[keep,] return(batch) } # function to fit linear regression and return slope direction fitReg <- function(batch) { trend <- lm(Value ~ Year, data = batch) slope <- coefficients(trend)[[2]] return(sign(slope)) } # function to fit linear regression and return slope magnitude fitRegv2 <- function(batch) { trend <- lm(Value ~ Year, data = batch) slope <- coefficients(trend)[[2]] return(slope) } # function to implement Mann-Kendall (MK) trend test and return significance # the test is implemented only for n>=8 getMann <- function(batch) { if (nrow(batch) >= 8) { mk <- MannKendall(batch[['Value']]) pval <- mk[['sl']] } else { pval <- NA } return(pval) } # function to return slope direction for all combinations given a start year getSign <- function(start.year) { remaining <- getRemain(start.year) combs <- lapply(remaining, splitData, start.year = start.year) signs <- lapply(combs, fitReg) return(signs) } # function to return MK significance for all combinations given a start year getPval <- function(start.year) { remaining <- getRemain(start.year) combs <- lapply(remaining, splitData, start.year = start.year) pvals <- lapply(combs, getMann) return(pvals) } # function to return slope magnitude for all combinations given a start year getMagn <- function(start.year) { remaining <- getRemain(start.year) combs <- lapply(remaining, splitData, start.year = start.year) magns <- lapply(combs, fitRegv2) return(magns) } # retrieve slope direction, MK significance, and slope magnitude signs <- lapply(years, getSign) pvals <- lapply(years, getPval) magns <- lapply(years, getMagn) # fill-in output matrices dimension <- nrow(sign.matrix) for (i in 1:dimension) { sign.matrix[i, i:dimension] <- unlist(signs[i]) pval.matrix[i, i:dimension] <- unlist(pvals[i]) slope.matrix[i, i:dimension] <- unlist(magns[i]) } sign.matrix <-...

  2. f

    Graph Input Data Example.xlsx

    • figshare.com
    xlsx
    Updated Dec 26, 2018
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Dr Corynen (2018). Graph Input Data Example.xlsx [Dataset]. http://doi.org/10.6084/m9.figshare.7506209.v1
    Explore at:
    xlsxAvailable download formats
    Dataset updated
    Dec 26, 2018
    Dataset provided by
    figshare
    Authors
    Dr Corynen
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    The various performance criteria applied in this analysis include the probability of reaching the ultimate target, the costs, elapsed times and system vulnerability resulting from any intrusion. This Excel file contains all the logical, probabilistic and statistical data entered by a user, and required for the evaluation of the criteria. It also reports the results of all the computations.

  3. T

    Nuclear Medicine National Headquarter System

    • datahub.va.gov
    • data.va.gov
    • +4more
    application/rdfxml +5
    Updated Sep 12, 2019
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    (2019). Nuclear Medicine National Headquarter System [Dataset]. https://www.datahub.va.gov/dataset/Nuclear-Medicine-National-Headquarter-System/x6z5-25xw
    Explore at:
    csv, xml, application/rssxml, json, tsv, application/rdfxmlAvailable download formats
    Dataset updated
    Sep 12, 2019
    Description

    The Nuclear Medicine National HQ System database is a series of MS Excel spreadsheets and Access Database Tables by fiscal year. They consist of information from all Veterans Affairs Medical Centers (VAMCs) performing or contracting nuclear medicine services in Veterans Affairs medical facilities. The medical centers are required to complete questionnaires annually (RCS 10-0010-Nuclear Medicine Service Annual Report). The information is then manually entered into the Access Tables, which includes: * Distribution and cost of in-house VA - Contract Physician Services, whether contracted services are made via sharing agreement (with another VA medical facility or other government medical providers) or with private providers. * Workload data for the performance and/or purchase of PET/CT studies. * Organizational structure of services. * Updated changes in key imaging service personnel (chiefs, chief technicians, radiation safety officers). * Workload data on the number and type of studies (scans) performed, including Medicare Relative Value Units (RVUs), also referred to as Weighted Work Units (WWUs). WWUs are a workload measure calculated as the product of a study's Current Procedural Terminology (CPT) code, which consists of total work costs (the cost of physician medical expertise and time), and total practice costs (the costs of running a practice, such as equipment, supplies, salaries, utilities etc). Medicare combines WWUs together with one other parameter to derive RVUs, a workload measure widely used in the health care industry. WWUs allow Nuclear Medicine to account for the complexity of each study in assessing workload, that some studies are more time consuming and require higher levels of expertise. This gives a more accurate picture of workload; productivity etc than using just 'total studies' would yield. * A detailed Full-Time Equivalent Employee (FTEE) grid, and staffing distributions of FTEEs across nuclear medicine services. * Information on Radiation Safety Committees and Radiation Safety Officers (RSOs). Beginning in 2011 this will include data collection on part-time and non VA (contract) RSOs; other affiliations they may have and if so to whom they report (supervision) at their VA medical center.Collection of data on nuclear medicine services' progress in meeting the special needs of our female veterans. Revolving documentation of all major VA-owned gamma cameras (by type) and computer systems, their specifications and ages. * Revolving data collection for PET/CT cameras owned or leased by VA; and the numbers and types of PET/CT studies performed on VA patients whether produced on-site, via mobile PET/CT contract or from non-VA providers in the community.* Types of educational training/certification programs available at VA sites * Ongoing funded research projects by Nuclear Medicine (NM) staff, identified by source of funding and research purpose. * Data on physician-specific quality indicators at each nuclear medicine service.* Academic achievements by NM staff, including published books/chapters, journals and abstracts. * Information from polling field sites re: relevant issues and programs Headquarters needs to address. * Results of a Congressionally mandated contracted quality assessment exercise, also known as a Proficiency study. Study results are analyzed for comparison within VA facilities (for example by mission or size), and against participating private sector health care groups. * Information collected on current issues in nuclear medicine as they arise. Radiation Safety Committee structures and membership, Radiation Safety Officer information and information on how nuclear medicine services provided for female Veterans are examples of current issues.The database is now stored completely within MS Access Database Tables with output still presented in the form of Excel graphs and tables.

  4. Z

    Dairy Supply Chain Sales Dataset

    • data.niaid.nih.gov
    • zenodo.org
    Updated Jul 12, 2024
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Dimitrios Pliatsios (2024). Dairy Supply Chain Sales Dataset [Dataset]. https://data.niaid.nih.gov/resources?id=zenodo_7853252
    Explore at:
    Dataset updated
    Jul 12, 2024
    Dataset provided by
    Christos Chaschatzis
    Dimitrios Pliatsios
    Panagiotis Sarigiannidis
    Thomas Lagkas
    Konstantinos Georgakidis
    Ilias Siniosoglou
    Vasileios Argyriou
    Dimitris Iatropoulos
    Athanasios Liatifis
    Anna Triantafyllou
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    1.Introduction

    Sales data collection is a crucial aspect of any manufacturing industry as it provides valuable insights about the performance of products, customer behaviour, and market trends. By gathering and analysing this data, manufacturers can make informed decisions about product development, pricing, and marketing strategies in Internet of Things (IoT) business environments like the dairy supply chain.

    One of the most important benefits of the sales data collection process is that it allows manufacturers to identify their most successful products and target their efforts towards those areas. For example, if a manufacturer could notice that a particular product is selling well in a certain region, this information could be utilised to develop new products, optimise the supply chain or improve existing ones to meet the changing needs of customers.

    This dataset includes information about 7 of MEVGAL’s products [1]. According to the above information the data published will help researchers to understand the dynamics of the dairy market and its consumption patterns, which is creating the fertile ground for synergies between academia and industry and eventually help the industry in making informed decisions regarding product development, pricing and market strategies in the IoT playground. The use of this dataset could also aim to understand the impact of various external factors on the dairy market such as the economic, environmental, and technological factors. It could help in understanding the current state of the dairy industry and identifying potential opportunities for growth and development.

    1. Citation

    Please cite the following papers when using this dataset:

    I. Siniosoglou, K. Xouveroudis, V. Argyriou, T. Lagkas, S. K. Goudos, K. E. Psannis and P. Sarigiannidis, "Evaluating the Effect of Volatile Federated Timeseries on Modern DNNs: Attention over Long/Short Memory," in the 12th International Conference on Circuits and Systems Technologies (MOCAST 2023), April 2023, Accepted

    1. Dataset Modalities

    The dataset includes data regarding the daily sales of a series of dairy product codes offered by MEVGAL. In particular, the dataset includes information gathered by the logistics division and agencies within the industrial infrastructures overseeing the production of each product code. The products included in this dataset represent the daily sales and logistics of a variety of yogurt-based stock. Each of the different files include the logistics for that product on a daily basis for three years, from 2020 to 2022.

    3.1 Data Collection

    The process of building this dataset involves several steps to ensure that the data is accurate, comprehensive and relevant.

    The first step is to determine the specific data that is needed to support the business objectives of the industry, i.e., in this publication’s case the daily sales data.

    Once the data requirements have been identified, the next step is to implement an effective sales data collection method. In MEVGAL’s case this is conducted through direct communication and reports generated each day by representatives & selling points.

    It is also important for MEVGAL to ensure that the data collection process conducted is in an ethical and compliant manner, adhering to data privacy laws and regulation. The industry also has a data management plan in place to ensure that the data is securely stored and protected from unauthorised access.

    The published dataset is consisted of 13 features providing information about the date and the number of products that have been sold. Finally, the dataset was anonymised in consideration to the privacy requirement of the data owner (MEVGAL).

    File

    Period

    Number of Samples (days)

    product 1 2020.xlsx

    01/01/2020–31/12/2020

    363

    product 1 2021.xlsx

    01/01/2021–31/12/2021

    364

    product 1 2022.xlsx

    01/01/2022–31/12/2022

    365

    product 2 2020.xlsx

    01/01/2020–31/12/2020

    363

    product 2 2021.xlsx

    01/01/2021–31/12/2021

    364

    product 2 2022.xlsx

    01/01/2022–31/12/2022

    365

    product 3 2020.xlsx

    01/01/2020–31/12/2020

    363

    product 3 2021.xlsx

    01/01/2021–31/12/2021

    364

    product 3 2022.xlsx

    01/01/2022–31/12/2022

    365

    product 4 2020.xlsx

    01/01/2020–31/12/2020

    363

    product 4 2021.xlsx

    01/01/2021–31/12/2021

    364

    product 4 2022.xlsx

    01/01/2022–31/12/2022

    364

    product 5 2020.xlsx

    01/01/2020–31/12/2020

    363

    product 5 2021.xlsx

    01/01/2021–31/12/2021

    364

    product 5 2022.xlsx

    01/01/2022–31/12/2022

    365

    product 6 2020.xlsx

    01/01/2020–31/12/2020

    362

    product 6 2021.xlsx

    01/01/2021–31/12/2021

    364

    product 6 2022.xlsx

    01/01/2022–31/12/2022

    365

    product 7 2020.xlsx

    01/01/2020–31/12/2020

    362

    product 7 2021.xlsx

    01/01/2021–31/12/2021

    364

    product 7 2022.xlsx

    01/01/2022–31/12/2022

    365

    3.2 Dataset Overview

    The following table enumerates and explains the features included across all of the included files.

    Feature

    Description

    Unit

    Day

    day of the month

    -

    Month

    Month

    -

    Year

    Year

    -

    daily_unit_sales

    Daily sales - the amount of products, measured in units, that during that specific day were sold

    units

    previous_year_daily_unit_sales

    Previous Year’s sales - the amount of products, measured in units, that during that specific day were sold the previous year

    units

    percentage_difference_daily_unit_sales

    The percentage difference between the two above values

    %

    daily_unit_sales_kg

    The amount of products, measured in kilograms, that during that specific day were sold

    kg

    previous_year_daily_unit_sales_kg

    Previous Year’s sales - the amount of products, measured in kilograms, that during that specific day were sold, the previous year

    kg

    percentage_difference_daily_unit_sales_kg

    The percentage difference between the two above values

    kg

    daily_unit_returns_kg

    The percentage of the products that were shipped to selling points and were returned

    %

    previous_year_daily_unit_returns_kg

    The percentage of the products that were shipped to selling points and were returned the previous year

    %

    points_of_distribution

    The amount of sales representatives through which the product was sold to the market for this year

    previous_year_points_of_distribution

    The amount of sales representatives through which the product was sold to the market for the same day for the previous year

    Table 1 – Dataset Feature Description

    1. Structure and Format

    4.1 Dataset Structure

    The provided dataset has the following structure:

    Where:

    Name

    Type

    Property

    Readme.docx

    Report

    A File that contains the documentation of the Dataset.

    product X

    Folder

    A folder containing the data of a product X.

    product X YYYY.xlsx

    Data file

    An excel file containing the sales data of product X for year YYYY.

    Table 2 - Dataset File Description

    1. Acknowledgement

    This project has received funding from the European Union’s Horizon 2020 research and innovation programme under grant agreement No. 957406 (TERMINET).

    References

    [1] MEVGAL is a Greek dairy production company

  5. d

    Supplemental data from: Hydraulic characterization of volcanic rocks in...

    • catalog.data.gov
    • data.usgs.gov
    • +1more
    Updated Jul 6, 2024
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    U.S. Geological Survey (2024). Supplemental data from: Hydraulic characterization of volcanic rocks in Pahute Mesa using an integrated analysis of 16 multiple-well aquifer tests, Nevada National Security Site, 2009–14 [Dataset]. https://catalog.data.gov/dataset/supplemental-data-from-hydraulic-characterization-of-volcanic-rocks-in-pahute-mesa-using-a
    Explore at:
    Dataset updated
    Jul 6, 2024
    Dataset provided by
    United States Geological Surveyhttp://www.usgs.gov/
    Description

    This USGS data release represents tabular data and water-level modeling files for the 16 Pahute Mesa multiple-well aquifer tests conducted from 2009–2014. The data release was produced in compliance with the new 'open data' requirements as a way to make the scientific products associated with USGS research efforts and publications available to the public. The dataset consists of 5 separate items: 1. Pumping time-series data (Tabular dataset) 2. Simplified pumping time-series data (Tabular dataset) 3. Drawdown time-series data (Tabular dataset) 4. Water-level models (Macro-enabled Excel spreadsheets) 5. SeriesSEE (Excel Add-In)

  6. Z

    Dataset: Evaluation of post-hoc interpretability methods in time-series...

    • data.niaid.nih.gov
    • zenodo.org
    Updated Jan 20, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Lovis, Christian (2023). Dataset: Evaluation of post-hoc interpretability methods in time-series classification [Dataset]. https://data.niaid.nih.gov/resources?id=ZENODO_7534769
    Explore at:
    Dataset updated
    Jan 20, 2023
    Dataset provided by
    Lovis, Christian
    Bjelogrlic, Mina
    Turbé, Hugues
    Mengaldo, Gianmarco
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    This repository contains the dataset, trained models as well as results for the article Evaluation of post-hoc interpretability methods in time-series classification.

    The code to reproduce the results presented in the article is available on GitHub. More details on the data and results can be found in the article.

    Files:

    datasets.zip: Include the three datasets used in the article:

    ECG: Processed version of the CPSC dataset from Classification of 12-lead ECGs: the PhysioNet - Computing in Cardiology Challenge 2020.

    fordA: Dataset from the UCR Time Series Classification Archive

    synthetic: Synthetic dataset developed specifically for the purpose of the article

    trained_models.zip: Include CNN, transformer and bi-lstm trained on the three datasets

    results_paper.zip: Computed relevance and evaluation metrics for the trained models

    model_interpretability: Include the relevance computed using the different interpretability methods as well as the computed metrics for each method

    summary_results: Summary of the evaluation metrics across all interpretability methods for each dataset as well as an excel file summarising the metrics across all datasets.

  7. FOI 29916 - Datasets - Open Data Portal

    • opendata.nhsbsa.net
    Updated Jan 3, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    nhsbsa.net (2023). FOI 29916 - Datasets - Open Data Portal [Dataset]. https://opendata.nhsbsa.net/dataset/foi-29916
    Explore at:
    Dataset updated
    Jan 3, 2023
    Dataset provided by
    NHS Business Services Authority
    Description

    CSVs with more than 1 million rows can be viewed using add-ons to existing software, such as the Microsoft PowerPivot add-on for Excel, to handle larger data sets. The Microsoft PowerPivot add-on for Excel is available using the link in the 'Related Links' section below. Once PowerPivot has been installed, to load the large files, please follow the instructions below. Note that it may take at least 20 to 30 minutes to load one monthly file. Start Excel as normal Click on the PowerPivot tab

  8. Example Student Data.xlsx

    • figshare.com
    xlsx
    Updated Jun 3, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Carrie Ellis (2022). Example Student Data.xlsx [Dataset]. http://doi.org/10.6084/m9.figshare.19985453.v1
    Explore at:
    xlsxAvailable download formats
    Dataset updated
    Jun 3, 2022
    Dataset provided by
    figshare
    Figsharehttp://figshare.com/
    Authors
    Carrie Ellis
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    In the attached Excel file, "Example Student Data", there are 6 sheets. There are three sheets with sample datasets, one for each of the three different exercise protocols described. Additionally, there are three sheets with sample graphs created using one of the three datasets. · Sheets 1 and 2: This is an example of a dataset and graph created from an exercise protocol designed to stress the creatine phosphate system. Here, the subject was a track and field athlete who threw the shot put for the DeSales University track team. The NIRS monitor was placed on the right triceps muscle, and the student threw the shot put six times with a minute rest in between throws. Data was collected telemetrically by the NIRS device and then downloaded after the student had completed the protocol. · Sheets 3 and 4: This is an example of a dataset and graph created from an exercise protocol designed to stress the glycolytic energy system. In this example, the subject performed continuous squat jumps for 30 seconds, followed by a 90 second rest period, for a total of three exercise bouts. The NIRS monitor was place on the left gastrocnemius muscle. Here again, data was collected telemetrically by the NIRS device and then downloaded after he had completed the protocol. · Sheets 5 and 6: In this example, the dataset and graph are from an exercise protocol designed to stress the oxidative system. Here, the student held a light-intensity, isometric biceps contraction (pushing against a table). The NIRS monitor was attached to the left biceps muscle belly. Here, data was collected by a student observing the SmO2 values displayed on a secondary device; specifically, a smartphone with the IPSensorMan APP displaying data. The recorder student observed and recorded the data on an Excel Spreadsheet, and marked the times that exercise began and ended on the Spreadsheet.

  9. o

    Quality Assurance and Quality Control (QA/QC) of Meteorological Time Series...

    • osti.gov
    • knb.ecoinformatics.org
    Updated Jan 1, 2021
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Environmental System Science Data Infrastructure for a Virtual Ecosystem (ESS-DIVE) (United States) (2021). Quality Assurance and Quality Control (QA/QC) of Meteorological Time Series Data for Billy Barr, East River, Colorado USA [Dataset]. http://doi.org/10.15485/1823516
    Explore at:
    Dataset updated
    Jan 1, 2021
    Dataset provided by
    Environmental System Science Data Infrastructure for a Virtual Ecosystem (ESS-DIVE) (United States)
    U.S. DOE > Office of Science > Biological and Environmental Research (BER)
    Area covered
    Colorado, East River, United States
    Description

    A comprehensive Quality Assurance (QA) and Quality Control (QC) statistical framework consists of three major phases: Phase 1—Preliminary raw data sets exploration, including time formatting and combining datasets of different lengths and different time intervals; Phase 2—QA of the datasets, including detecting and flagging of duplicates, outliers, and extreme values; and Phase 3—the development of time series of a desired frequency, imputation of missing values, visualization and a final statistical summary. The time series data collected at the Billy Barr meteorological station (East River Watershed, Colorado) were analyzed. The developed statistical framework is suitable for both real-time and post-data-collection QA/QC analysis of meteorological datasets.The files that are in this data package include one excel file, converted to CSV format (Billy_Barr_raw_qaqc.csv) that contains the raw meteorological data, i.e., input data used for the QA/QC analysis. The second CSV file (Billy_Barr_1hr.csv) is the QA/QC and flagged meteorological data, i.e., output data from the QA/QC analysis. The last file (QAQC_Billy_Barr_2021-03-22.R) is a script written in R that implements the QA/QC and flagging process. The purpose of the CSV data files included in this package is to provide input and output files implemented in the R script.

  10. d

    Historical volatility time series and Live prices on Equity Options

    • datarade.ai
    Updated Mar 9, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Canari (2023). Historical volatility time series and Live prices on Equity Options [Dataset]. https://datarade.ai/data-products/historical-volatility-time-series-and-live-prices-on-equity-o-canari
    Explore at:
    Dataset updated
    Mar 9, 2023
    Dataset authored and provided by
    Canari
    Area covered
    Switzerland, Norway, France, Belgium, Spain, Netherlands, Germany, Italy, Sweden, United Kingdom
    Description

    This dataset offers both live (delayed) prices and End Of Day time series on equity options

    1/ Live (delayed) prices for options on European stocks and indices including: Reference spot price, bid/ask screen price, fair value price (based on surface calibration), implicit volatility, forward Greeks : delta, vega Canari.dev computes AI-generated forecast signals indicating which option is over/underpriced, based on the holders strategy (buy and hold until maturity, 1 hour to 2 days holding horizon...). From these signals is derived a "Canari price" which is also available in this live tables.
    Visit our website (canari.dev ) for more details about our forecast signals.

    The delay ranges from 15 to 40 minutes depending on underlyings.

    2/ Historical time series: Implied vol Realized vol Smile Forward
    See a full API presentation here : https://youtu.be/qitPO-SFmY4 .

    These data are also readily accessible in Excel thanks the provided Add-in available on Github: https://github.com/canari-dev/Excel-macro-to-consume-Canari-API

    If you need help, contact us at: contact@canari.dev

    User Guide: You can get a preview of the API by typing "data.canari.dev" in your web browser. This will show you a free version of this API with limited data.

    Here are examples of possible syntaxes:

    For live options prices: data.canari.dev/OPT/DAI data.canari.dev/OPT/OESX/0923 The "csv" suffix to get a csv rather than html formating, for example: data.canari.dev/OPT/DB1/1223/csv For historical parameters: Implied vol : data.canari.dev/IV/BMW

    data.canari.dev/IV/ALV/1224

    data.canari.dev/IV/DTE/1224/csv

    Realized vol (intraday, maturity expressed as EWM, span in business days): data.canari.dev/RV/IFX ... Implied dividend flow: data.canari.dev/DIV/IBE ... Smile (vol spread between ATM strike and 90% strike, normalized to 1Y with factor 1/√T): data.canari.dev/SMI/DTE ... Forward: data.canari.dev/FWD/BNP ...

    List of available underlyings: Code Name OESX Eurostoxx50 ODAX DAX OSMI SMI (Swiss index) OESB Eurostoxx Banks OVS2 VSTOXX ITK AB Inbev ABBN ABB ASM ASML ADS Adidas AIR Air Liquide EAD Airbus ALV Allianz AXA Axa BAS BASF BBVD BBVA BMW BMW BNP BNP BAY Bayer DBK Deutsche Bank DB1 Deutsche Boerse DPW Deutsche Post DTE Deutsche Telekom EOA E.ON ENL5 Enel INN ING IBE Iberdrola IFX Infineon IES5 Intesa Sanpaolo PPX Kering LOR L Oreal MOH LVMH LIN Linde DAI Mercedes-Benz MUV2 Munich Re NESN Nestle NOVN Novartis PHI1 Philips REP Repsol ROG Roche SAP SAP SNW Sanofi BSD2 Santander SND Schneider SIE Siemens SGE Société Générale SREN Swiss Re TNE5 Telefonica TOTB TotalEnergies UBSN UBS CRI5 Unicredito SQU Vinci VO3 Volkswagen ANN Vonovia ZURN Zurich Insurance Group

  11. d

    Bioregional Assessment areas v05

    • data.gov.au
    • researchdata.edu.au
    • +1more
    zip
    Updated Nov 20, 2019
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Bioregional Assessment Program (2019). Bioregional Assessment areas v05 [Dataset]. https://data.gov.au/data/dataset/activity/25f89049-839d-4736-bd81-97d8e8a40f8e
    Explore at:
    zip(3880455)Available download formats
    Dataset updated
    Nov 20, 2019
    Dataset provided by
    Bioregional Assessment Program
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Abstract

    The dataset was derived by the Bioregional Assessment Programme from multiple source datasets. The source datasets are identified in the Lineage field in this metadata statement. The processes undertaken to produce this derived dataset are described in the History field in this metadata statement.

    Approved boundaries of the bioregions and subregions (version 5) for defining the reporting regions for bioregional assessments of impacts of coal seam gas and coal mining development on water resources.

    This is identical to Bioregional_Assessment_areas_v04 except that the attribute tables include areas (both sq. km and ha) based on the GDA 1994 Australia Albers projection. A spreadsheet version of the attribute table is also provided for the benefit of non ArcGIS users.

    Purpose

    Provides authoritative boundaries for defining bioregions and subregions to be reported on for the Bioregional Assessments and tabulation of Bioregion and subregion areas.

    Dataset History

    This dataset contains two spatial shapefiles: "ba_bioregion_alb_gda94_v05.shp" and "ba_subregion_alb_gda94_v05.shp".

    The shapefiles are copies of the previous versions (Bioregional Assessment Areas v04), with the following changes.

    Two fields have been added to each of the shapefiles' attribute tables.

    "albers_km2" and "albers_ha" which list the the bioregion/subregion areas in square kilometres and hectares respectively.

    The polygonal areas are calculated in ArcGIS and are based on the GDA_1994_Australia_Albers projection. Parameters as follows:

    Projected Coordinate System: GDA_1994_Australia_Albers

    Projection: Albers

    False_Easting: 0.00000000

    False_Northing: 0.00000000

    Central_Meridian: 132.00000000

    Standard_Parallel_1: -18.00000000

    Standard_Parallel_2: -36.00000000

    Latitude_Of_Origin: 0.00000000

    Linear Unit: Meter

    Geographic Coordinate System: GCS_GDA_1994

    Datum: D_GDA_1994

    Prime Meridian: Greenwich

    Angular Unit: Degree

    It should be noted that area calculations using a different projection (eg UTM or MGA) may yield slightly different areas to those published in this dataset.

    Version 5 of this dataset also includes Excel spreadsheet versions of each shapefiles' attrubute table, to enable non ArcGIS users to access the bioregion/subregion area information.

    Dataset Citation

    Bioregional Assessment Programme (2014) Bioregional Assessment areas v05. Bioregional Assessment Derived Dataset. Viewed 13 March 2019, http://data.bioregionalassessments.gov.au/dataset/25f89049-839d-4736-bd81-97d8e8a40f8e.

    Dataset Ancestors

  12. d

    Processed High Temperature Probe and Major Fluid Sampler Time Series Data...

    • datadiscoverystudio.org
    • marine-geo.org
    • +2more
    xls v.1
    Updated Jul 16, 2010
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    (2010). Processed High Temperature Probe and Major Fluid Sampler Time Series Data from the East Pacific Rise 9N site assembled as part of the EPR:9N_VonDamm Data Compilation (1991)Marine Geoscience Digital Library internal dataset identifiers [Dataset]. http://doi.org/10.1594/IEDA/316294
    Explore at:
    xls v.1Available download formats
    Dataset updated
    Jul 16, 2010
    Area covered
    Description

    This data set was acquired with a DSPL HOBO HighTemp Temperature Probe and Major Fluid Sampler assembled as part of the 1991 EPR:9N_VonDamm data compilation (Chief Scientist: Dr. Karen Von Damm; Investigators: Dr. Julie Bryce, Florencia Prado, and Dr. Karen Von Damm). The data files are in Microsoft Excel format and include Fluid Chemistry and Temperature time series data and were processed after data collection. Funding was provided by NSF grant OCE03-27126. This data was cited by Oosting and Von Damm, 1996, Von Damm et al., 1997, Ravizza et al., 2001, Von Damm, 2000, Von Damm, 2004, Von Damm and Lilley, 2004, and Haymon et al., 1993.

  13. g

    USGS Geochron: Data Compilation Templates | gimi9.com

    • gimi9.com
    Updated Mar 2, 2025
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    (2025). USGS Geochron: Data Compilation Templates | gimi9.com [Dataset]. https://gimi9.com/dataset/data-gov_usgs-geochron-data-compilation-templates
    Explore at:
    Dataset updated
    Mar 2, 2025
    Description

    USGS Geochron is a database of geochronological and thermochronological dates and data. The USGS Geochron: Data Compilation Templates data release hosts Microsoft Excel-based data compilation templates for the USGS Geochron database. Geochronological and thermochronological methods currently archived in the USGS Geochron database include radiocarbon, cosmogenic (10Be, 26Al, 3He), fission track, (U-Th)/He, U-series, U-Th-Pb, 40Ar/39Ar, K-Ar, Lu-Hf, Rb-Sr, Sm-Nd, and Re-Os dating methods. For questions or to submit data please contact geochron@usgs.gov

  14. Prescription Drugs Introduced to Market

    • data.chhs.ca.gov
    csv, xlsx, zip
    Updated Jun 6, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Department of Health Care Access and Information (2025). Prescription Drugs Introduced to Market [Dataset]. https://data.chhs.ca.gov/dataset/prescription-drugs-introduced-to-market
    Explore at:
    xlsx(78989), xlsx(97853), xlsx(88082), csv(4193), xlsx(56740), zip, csv(209944), xlsx(87563), xlsx(138801)Available download formats
    Dataset updated
    Jun 6, 2025
    Dataset authored and provided by
    Department of Health Care Access and Information
    Description

    This dataset provides data for new prescription drugs introduced to market in California with a Wholesale Acquisition Cost (WAC) that exceeds the Medicare Part D specialty drug cost threshold. Prescription drug manufacturers submit information to HCAI within a specified time period after a drug is introduced to market. Key data elements include the National Drug Code (NDC) administered by the FDA, a narrative description of marketing and pricing plans, and WAC, among other information. Manufacturers may withhold information that is not in the public domain. Note that prescription drug manufacturers are able to submit new drug reports for a prior quarter at any time. Therefore, the data set may include additional new drug report(s) from previous quarter(s).

    There are two types of New Drug data sets: Monthly and Annual. The Monthly data sets include the data in completed reports submitted by manufacturers for calendar year 2025, as of June 6, 2025. The Annual data sets include data in completed reports submitted by manufacturers for the specified year. The data sets may include reports that do not meet the specified minimum thresholds for reporting.

    The program regulations are available here: https://hcai.ca.gov/wp-content/uploads/2024/03/CTRx-Regulations-Text.pdf

    The data format and file specifications are available here: https://hcai.ca.gov/wp-content/uploads/2024/03/Format-and-File-Specifications-version-2.0-ada.pdf

    DATA NOTES: Due to recent changes in Excel capabilities, it is not recommended that you save these files to .csv format. If you do, when importing back into Excel the leading zeros in the NDC number column will be dropped. If you need to save it into a different format other than .xlsx it must be .txt

  15. N

    Excel, AL Median Income by Age Groups Dataset: A Comprehensive Breakdown of...

    • neilsberg.com
    csv, json
    Updated Feb 25, 2025
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Neilsberg Research (2025). Excel, AL Median Income by Age Groups Dataset: A Comprehensive Breakdown of Excel Annual Median Income Across 4 Key Age Groups // 2025 Edition [Dataset]. https://www.neilsberg.com/research/datasets/e931c85f-f353-11ef-8577-3860777c1fe6/
    Explore at:
    json, csvAvailable download formats
    Dataset updated
    Feb 25, 2025
    Dataset authored and provided by
    Neilsberg Research
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Area covered
    Excel
    Variables measured
    Income for householder under 25 years, Income for householder 65 years and over, Income for householder between 25 and 44 years, Income for householder between 45 and 64 years
    Measurement technique
    The data presented in this dataset is derived from the U.S. Census Bureau American Community Survey (ACS) 2019-2023 5-Year Estimates. It delineates income distributions across four age groups (Under 25 years, 25 to 44 years, 45 to 64 years, and 65 years and over) following an initial analysis and categorization. Subsequently, we adjusted these figures for inflation using the Consumer Price Index retroactive series via current methods (R-CPI-U-RS). For additional information about these estimations, please contact us via email at research@neilsberg.com
    Dataset funded by
    Neilsberg Research
    Description
    About this dataset

    Context

    The dataset presents the distribution of median household income among distinct age brackets of householders in Excel. Based on the latest 2019-2023 5-Year Estimates from the American Community Survey, it displays how income varies among householders of different ages in Excel. It showcases how household incomes typically rise as the head of the household gets older. The dataset can be utilized to gain insights into age-based household income trends and explore the variations in incomes across households.

    Key observations: Insights from 2023

    In terms of income distribution across age cohorts, in Excel, where there exist only two delineated age groups, the median household income is $83,750 for householders within the 25 to 44 years age group, compared to $58,958 for the 65 years and over age group.

    Content

    When available, the data consists of estimates from the U.S. Census Bureau American Community Survey (ACS) 2019-2023 5-Year Estimates. All incomes have been adjusting for inflation and are presented in 2023-inflation-adjusted dollars.

    Age groups classifications include:

    • Under 25 years
    • 25 to 44 years
    • 45 to 64 years
    • 65 years and over

    Variables / Data Columns

    • Age Of The Head Of Household: This column presents the age of the head of household
    • Median Household Income: Median household income, in 2023 inflation-adjusted dollars for the specific age group

    Good to know

    Margin of Error

    Data in the dataset are based on the estimates and are subject to sampling variability and thus a margin of error. Neilsberg Research recommends using caution when presening these estimates in your research.

    Custom data

    If you do need custom data for any of your research project, report or presentation, you can contact our research staff at research@neilsberg.com for a feasibility of a custom tabulation on a fee-for-service basis.

    Inspiration

    Neilsberg Research Team curates, analyze and publishes demographics and economic data from a variety of public and proprietary sources, each of which often includes multiple surveys and programs. The large majority of Neilsberg Research aggregated datasets and insights is made available for free download at https://www.neilsberg.com/research/.

    Recommended for further research

    This dataset is a part of the main dataset for Excel median household income by age. You can refer the same here

  16. N

    Excel Township, Minnesota Median Income by Age Groups Dataset: A...

    • neilsberg.com
    csv, json
    Updated Aug 7, 2024
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Neilsberg Research (2024). Excel Township, Minnesota Median Income by Age Groups Dataset: A Comprehensive Breakdown of Excel township Annual Median Income Across 4 Key Age Groups // 2024 Edition [Dataset]. https://www.neilsberg.com/research/datasets/a3d462e6-54ae-11ef-a42e-3860777c1fe6/
    Explore at:
    json, csvAvailable download formats
    Dataset updated
    Aug 7, 2024
    Dataset authored and provided by
    Neilsberg Research
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Area covered
    Excel Township
    Variables measured
    Income for householder under 25 years, Income for householder 65 years and over, Income for householder between 25 and 44 years, Income for householder between 45 and 64 years
    Measurement technique
    The data presented in this dataset is derived from the U.S. Census Bureau American Community Survey (ACS) 2018-2022 5-Year Estimates. It delineates income distributions across four age groups (Under 25 years, 25 to 44 years, 45 to 64 years, and 65 years and over) following an initial analysis and categorization. Subsequently, we adjusted these figures for inflation using the Consumer Price Index retroactive series via current methods (R-CPI-U-RS). For additional information about these estimations, please contact us via email at research@neilsberg.com
    Dataset funded by
    Neilsberg Research
    Description
    About this dataset

    Context

    The dataset presents the distribution of median household income among distinct age brackets of householders in Excel township. Based on the latest 2018-2022 5-Year Estimates from the American Community Survey, it displays how income varies among householders of different ages in Excel township. It showcases how household incomes typically rise as the head of the household gets older. The dataset can be utilized to gain insights into age-based household income trends and explore the variations in incomes across households.

    Key observations: Insights from 2022

    In terms of income distribution across age cohorts, in Excel township, the median household income stands at $130,807 for householders within the 45 to 64 years age group, followed by $96,316 for the 65 years and over age group. Notably, householders within the 25 to 44 years age group, had the lowest median household income at $95,014.

    Content

    When available, the data consists of estimates from the U.S. Census Bureau American Community Survey (ACS) 2018-2022 5-Year Estimates. All incomes have been adjusting for inflation and are presented in 2023-inflation-adjusted dollars.

    Age groups classifications include:

    • Under 25 years
    • 25 to 44 years
    • 45 to 64 years
    • 65 years and over

    Variables / Data Columns

    • Age Of The Head Of Household: This column presents the age of the head of household
    • Median Household Income: Median household income, in 2022 inflation-adjusted dollars for the specific age group

    Good to know

    Margin of Error

    Data in the dataset are based on the estimates and are subject to sampling variability and thus a margin of error. Neilsberg Research recommends using caution when presening these estimates in your research.

    Custom data

    If you do need custom data for any of your research project, report or presentation, you can contact our research staff at research@neilsberg.com for a feasibility of a custom tabulation on a fee-for-service basis.

    Inspiration

    Neilsberg Research Team curates, analyze and publishes demographics and economic data from a variety of public and proprietary sources, each of which often includes multiple surveys and programs. The large majority of Neilsberg Research aggregated datasets and insights is made available for free download at https://www.neilsberg.com/research/.

    Recommended for further research

    This dataset is a part of the main dataset for Excel township median household income by age. You can refer the same here

  17. d

    Data from: The Bronson Files, Dataset 6, Field 13, 2014

    • catalog.data.gov
    • agdatacommons.nal.usda.gov
    • +2more
    Updated Apr 21, 2025
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Agricultural Research Service (2025). The Bronson Files, Dataset 6, Field 13, 2014 [Dataset]. https://catalog.data.gov/dataset/the-bronson-files-dataset-6-field-13-2014-e1c41
    Explore at:
    Dataset updated
    Apr 21, 2025
    Dataset provided by
    Agricultural Research Service
    Description

    Dr. Kevin Bronson provides a unique nitrogen and water management in cotton agricultural research dataset for compute, including notation of field events and operations, an intermediate analysis mega-table of correlated and calculated parameters, and laboratory analysis results generated during the experimentation, plus high-resolution plot level intermediate data analysis tables of SAS process output, as well as the complete raw data sensor recorded logger outputs. This data was collected using a Hamby rig as a high-throughput proximal plant phenotyping platform. The Hamby 6000 rig Ellis W. Chenault, & Allen F. Wiese. (1989). Construction of a High-Clearance Plot Sprayer. Weed Technology, 3(4), 659–662. http://www.jstor.org/stable/3987560 Dr. Bronson modified an old high-clearance Hamby 6000 rig, adding a tank and pump with a rear boom, to perform precision liquid N applications. A Raven control unit with GPS supplied variable rate delivery options. The 12 volt Holland Scientific GeoScoutX data recorder and associated CropCircle ACS-470 sensors with GPS signal, was easy to mount and run on the vehicle as an attached rugged data acquisition module, and allowed the measuring of plants using custom proximal active optical reflectance sensing. The HS data logger was positioned near the operator, and sensors were positioned in front of the rig, on forward protruding armature attached to a hydraulic front boom assembly, facing downward in nadir view 1 m above the average canopy height. A 34-size class AGM battery sat under the operator and provided the data system electrical power supply. Data suffered reduced input from Conley. Although every effort was afforded to capture adequate quality across all metrics, experiment exterior considerations were such that canopy temperature data is absent, and canopy height is weak due to technical underperformance. Thankfully, reflectance data quality was maintained or improved through the implementation of new hardware by Bronson. See included README file for operational details and further description of the measured data signals. Summary: Active optical proximal cotton canopy sensing spatial data and including few additional related metrics and weak low-frequency ultrasonic derived height are presented. Agronomic nitrogen and irrigation management related field operations are listed. Unique research experimentation intermediate analysis table is made available, along with raw data. The raw data recordings, and annotated table outputs with calculated VIs are made available. Plot polygon coordinate designations allow a re-intersection spatial analysis. Data was collected in the 2014 season at Maricopa Agricultural Center, Arizona, USA. High throughput proximal plant phenotyping via electronic sampling and data processing method approach is exampled using a modified high-clearance Hamby spray-rig. Acquired data conforms to location standard methodologies of the plant phenotyping. SAS and GIS compute processing output tables, including Excel formatted examples are presented, where data tabulation and analysis is available. Additional ultrasonic data signal explanation is offered as annotated time-series charts. The weekly proximal sensing data collected include the primary canopy reflectance at six wavelengths. Lint and seed yields, first open boll biomass, and nitrogen uptake were also determined. Soil profile nitrate to 1.8 m depth was determined in 30-cm increments, before planting and after harvest. Nitrous oxide emissions were determined with 1-L vented chambers (samples taken at 0, 12, and 24 minutes). Nitrous oxide was determined by gas chromatography (electron detection detector).

  18. m

    Cross Regional Eucalyptus Growth and Environmental Data

    • data.mendeley.com
    Updated Oct 7, 2024
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Christopher Erasmus (2024). Cross Regional Eucalyptus Growth and Environmental Data [Dataset]. http://doi.org/10.17632/2m9rcy3dr9.3
    Explore at:
    Dataset updated
    Oct 7, 2024
    Authors
    Christopher Erasmus
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    The dataset is provided in a single .xlsx file named "eucalyptus_growth_environment_data_V2.xlsx" and consists of fifteen sheets:

    Codebook: This sheet details the index, values, and descriptions for each field within the dataset, providing a comprehensive guide to understanding the data structure.

    ALL NODES: Contains measurements from all devices, totalling 102,916 data points. This sheet aggregates the data across all nodes.

    GWD1 to GWD10: These subset sheets include measurements from individual nodes, labelled according to the abbreviation “Generic Wireless Dendrometer” followed by device IDs 1 through 10. Each sheet corresponds to a specific node, representing measurements from ten trees (or nodes).

    Metadata: Provides detailed metadata for each node, including species, initial diameter, location, measurement frequency, battery specifications, and irrigation status. This information is essential for identifying and differentiating the nodes and their specific attributes.

    Missing Data Intervals: Details gaps in the data stream, including start and end dates and times when data was not uploaded. It includes information on the total duration of each missing interval and the number of missing data points.

    Missing Intervals Distribution: Offers a summary of missing data intervals and their distribution, providing insight into data gaps and reasons for missing data.

    All nodes utilize LoRaWAN for data transmission. Please note that intermittent data gaps may occur due to connectivity issues between the gateway and the nodes, as well as maintenance activities or experimental procedures.

    Software considerations: The provided R code named “Simple_Dendro_Imputation_and_Analysis.R” is a comprehensive analysis workflow that processes and analyses Eucalyptus growth and environmental data from the "eucalyptus_growth_environment_data_V2.xlsx" dataset. The script begins by loading necessary libraries, setting the working directory, and reading the data from the specified Excel sheet. It then combines date and time information into a unified DateTime format and performs data type conversions for relevant columns. The analysis focuses on a specified device, allowing for the selection of neighbouring devices for imputation of missing data. A loop checks for gaps in the time series and fills in missing intervals based on a defined threshold, followed by a function that imputes missing values using the average from nearby devices. Outliers are identified and managed through linear interpolation. The code further calculates vapor pressure metrics and applies temperature corrections to the dendrometer data. Finally, it saves the cleaned and processed data into a new Excel file while conducting dendrometer analysis using the dendRoAnalyst package, which includes visualizations and calculations of daily growth metrics and correlations with environmental factors such as vapour pressure deficit (VPD).

  19. Customer Shopping Trends Dataset

    • kaggle.com
    Updated Oct 5, 2023
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Sourav Banerjee (2023). Customer Shopping Trends Dataset [Dataset]. https://www.kaggle.com/datasets/iamsouravbanerjee/customer-shopping-trends-dataset
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Oct 5, 2023
    Dataset provided by
    Kagglehttp://kaggle.com/
    Authors
    Sourav Banerjee
    Description

    Context

    The Customer Shopping Preferences Dataset offers valuable insights into consumer behavior and purchasing patterns. Understanding customer preferences and trends is critical for businesses to tailor their products, marketing strategies, and overall customer experience. This dataset captures a wide range of customer attributes including age, gender, purchase history, preferred payment methods, frequency of purchases, and more. Analyzing this data can help businesses make informed decisions, optimize product offerings, and enhance customer satisfaction. The dataset stands as a valuable resource for businesses aiming to align their strategies with customer needs and preferences. It's important to note that this dataset is a Synthetic Dataset Created for Beginners to learn more about Data Analysis and Machine Learning.

    Content

    This dataset encompasses various features related to customer shopping preferences, gathering essential information for businesses seeking to enhance their understanding of their customer base. The features include customer age, gender, purchase amount, preferred payment methods, frequency of purchases, and feedback ratings. Additionally, data on the type of items purchased, shopping frequency, preferred shopping seasons, and interactions with promotional offers is included. With a collection of 3900 records, this dataset serves as a foundation for businesses looking to apply data-driven insights for better decision-making and customer-centric strategies.

    Dataset Glossary (Column-wise)

    • Customer ID - Unique identifier for each customer
    • Age - Age of the customer
    • Gender - Gender of the customer (Male/Female)
    • Item Purchased - The item purchased by the customer
    • Category - Category of the item purchased
    • Purchase Amount (USD) - The amount of the purchase in USD
    • Location - Location where the purchase was made
    • Size - Size of the purchased item
    • Color - Color of the purchased item
    • Season - Season during which the purchase was made
    • Review Rating - Rating given by the customer for the purchased item
    • Subscription Status - Indicates if the customer has a subscription (Yes/No)
    • Shipping Type - Type of shipping chosen by the customer
    • Discount Applied - Indicates if a discount was applied to the purchase (Yes/No)
    • Promo Code Used - Indicates if a promo code was used for the purchase (Yes/No)
    • Previous Purchases - The total count of transactions concluded by the customer at the store, excluding the ongoing transaction
    • Payment Method - Customer's most preferred payment method
    • Frequency of Purchases - Frequency at which the customer makes purchases (e.g., Weekly, Fortnightly, Monthly)

    Structure of the Dataset

    https://i.imgur.com/6UEqejq.png" alt="">

    Acknowledgement

    This dataset is a synthetic creation generated using ChatGPT to simulate a realistic customer shopping experience. Its purpose is to provide a platform for beginners and data enthusiasts, allowing them to create, enjoy, practice, and learn from a dataset that mirrors real-world customer shopping behavior. The aim is to foster learning and experimentation in a simulated environment, encouraging a deeper understanding of data analysis and interpretation in the context of consumer preferences and retail scenarios.

    Cover Photo by: Freepik

    Thumbnail by: Clothing icons created by Flat Icons - Flaticon

  20. Immigration system statistics data tables

    • gov.uk
    • totalwrapture.com
    Updated May 22, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Home Office (2025). Immigration system statistics data tables [Dataset]. https://www.gov.uk/government/statistical-data-sets/immigration-system-statistics-data-tables
    Explore at:
    Dataset updated
    May 22, 2025
    Dataset provided by
    GOV.UKhttp://gov.uk/
    Authors
    Home Office
    Description

    List of the data tables as part of the Immigration System Statistics Home Office release. Summary and detailed data tables covering the immigration system, including out-of-country and in-country visas, asylum, detention, and returns.

    If you have any feedback, please email MigrationStatsEnquiries@homeoffice.gov.uk.

    Accessible file formats

    The Microsoft Excel .xlsx files may not be suitable for users of assistive technology.
    If you use assistive technology (such as a screen reader) and need a version of these documents in a more accessible format, please email MigrationStatsEnquiries@homeoffice.gov.uk
    Please tell us what format you need. It will help us if you say what assistive technology you use.

    Related content

    Immigration system statistics, year ending March 2025
    Immigration system statistics quarterly release
    Immigration system statistics user guide
    Publishing detailed data tables in migration statistics
    Policy and legislative changes affecting migration to the UK: timeline
    Immigration statistics data archives

    Passenger arrivals

    https://assets.publishing.service.gov.uk/media/68258d71aa3556876875ec80/passenger-arrivals-summary-mar-2025-tables.xlsx">Passenger arrivals summary tables, year ending March 2025 (MS Excel Spreadsheet, 66.5 KB)

    ‘Passengers refused entry at the border summary tables’ and ‘Passengers refused entry at the border detailed datasets’ have been discontinued. The latest published versions of these tables are from February 2025 and are available in the ‘Passenger refusals – release discontinued’ section. A similar data series, ‘Refused entry at port and subsequently departed’, is available within the Returns detailed and summary tables.

    Electronic travel authorisation

    https://assets.publishing.service.gov.uk/media/681e406753add7d476d8187f/electronic-travel-authorisation-datasets-mar-2025.xlsx">Electronic travel authorisation detailed datasets, year ending March 2025 (MS Excel Spreadsheet, 56.7 KB)
    ETA_D01: Applications for electronic travel authorisations, by nationality ETA_D02: Outcomes of applications for electronic travel authorisations, by nationality

    Entry clearance visas granted outside the UK

    https://assets.publishing.service.gov.uk/media/68247953b296b83ad5262ed7/visas-summary-mar-2025-tables.xlsx">Entry clearance visas summary tables, year ending March 2025 (MS Excel Spreadsheet, 113 KB)

    https://assets.publishing.service.gov.uk/media/682c4241010c5c28d1c7e820/entry-clearance-visa-outcomes-datasets-mar-2025.xlsx">Entry clearance visa applications and outcomes detailed datasets, year ending March 2025 (MS Excel Spreadsheet, 29.1 MB)
    Vis_D01: Entry clearance visa applications, by nationality and visa type
    Vis_D02: Outcomes of entry clearance visa applications, by nationality, visa type, and outcome

    Additional dat

Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
Georgios Boumis; Brad Peter (2024). Time-Series Matrix (TSMx): A visualization tool for plotting multiscale temporal trends [Dataset]. http://doi.org/10.7910/DVN/ZZDYM9

Time-Series Matrix (TSMx): A visualization tool for plotting multiscale temporal trends

Explore at:
2 scholarly articles cite this dataset (View in Google Scholar)
CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
Dataset updated
Jul 8, 2024
Dataset provided by
Harvard Dataverse
Authors
Georgios Boumis; Brad Peter
License

CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
License information was derived automatically

Description

Time-Series Matrix (TSMx): A visualization tool for plotting multiscale temporal trends TSMx is an R script that was developed to facilitate multi-temporal-scale visualizations of time-series data. The script requires only a two-column CSV of years and values to plot the slope of the linear regression line for all possible year combinations from the supplied temporal range. The outputs include a time-series matrix showing slope direction based on the linear regression, slope values plotted with colors indicating magnitude, and results of a Mann-Kendall test. The start year is indicated on the y-axis and the end year is indicated on the x-axis. In the example below, the cell in the top-right corner is the direction of the slope for the temporal range 2001–2019. The red line corresponds with the temporal range 2010–2019 and an arrow is drawn from the cell that represents that range. One cell is highlighted with a black border to demonstrate how to read the chart—that cell represents the slope for the temporal range 2004–2014. This publication entry also includes an excel template that produces the same visualizations without a need to interact with any code, though minor modifications will need to be made to accommodate year ranges other than what is provided. TSMx for R was developed by Georgios Boumis; TSMx was originally conceptualized and created by Brad G. Peter in Microsoft Excel. Please refer to the associated publication: Peter, B.G., Messina, J.P., Breeze, V., Fung, C.Y., Kapoor, A. and Fan, P., 2024. Perspectives on modifiable spatiotemporal unit problems in remote sensing of agriculture: evaluating rice production in Vietnam and tools for analysis. Frontiers in Remote Sensing, 5, p.1042624. https://www.frontiersin.org/journals/remote-sensing/articles/10.3389/frsen.2024.1042624 TSMx sample chart from the supplied Excel template. Data represent the productivity of rice agriculture in Vietnam as measured via EVI (enhanced vegetation index) from the NASA MODIS data product (MOD13Q1.V006). TSMx R script: # import packages library(dplyr) library(readr) library(ggplot2) library(tibble) library(tidyr) library(forcats) library(Kendall) options(warn = -1) # disable warnings # read data (.csv file with "Year" and "Value" columns) data <- read_csv("EVI.csv") # prepare row/column names for output matrices years <- data %>% pull("Year") r.names <- years[-length(years)] c.names <- years[-1] years <- years[-length(years)] # initialize output matrices sign.matrix <- matrix(data = NA, nrow = length(years), ncol = length(years)) pval.matrix <- matrix(data = NA, nrow = length(years), ncol = length(years)) slope.matrix <- matrix(data = NA, nrow = length(years), ncol = length(years)) # function to return remaining years given a start year getRemain <- function(start.year) { years <- data %>% pull("Year") start.ind <- which(data[["Year"]] == start.year) + 1 remain <- years[start.ind:length(years)] return (remain) } # function to subset data for a start/end year combination splitData <- function(end.year, start.year) { keep <- which(data[['Year']] >= start.year & data[['Year']] <= end.year) batch <- data[keep,] return(batch) } # function to fit linear regression and return slope direction fitReg <- function(batch) { trend <- lm(Value ~ Year, data = batch) slope <- coefficients(trend)[[2]] return(sign(slope)) } # function to fit linear regression and return slope magnitude fitRegv2 <- function(batch) { trend <- lm(Value ~ Year, data = batch) slope <- coefficients(trend)[[2]] return(slope) } # function to implement Mann-Kendall (MK) trend test and return significance # the test is implemented only for n>=8 getMann <- function(batch) { if (nrow(batch) >= 8) { mk <- MannKendall(batch[['Value']]) pval <- mk[['sl']] } else { pval <- NA } return(pval) } # function to return slope direction for all combinations given a start year getSign <- function(start.year) { remaining <- getRemain(start.year) combs <- lapply(remaining, splitData, start.year = start.year) signs <- lapply(combs, fitReg) return(signs) } # function to return MK significance for all combinations given a start year getPval <- function(start.year) { remaining <- getRemain(start.year) combs <- lapply(remaining, splitData, start.year = start.year) pvals <- lapply(combs, getMann) return(pvals) } # function to return slope magnitude for all combinations given a start year getMagn <- function(start.year) { remaining <- getRemain(start.year) combs <- lapply(remaining, splitData, start.year = start.year) magns <- lapply(combs, fitRegv2) return(magns) } # retrieve slope direction, MK significance, and slope magnitude signs <- lapply(years, getSign) pvals <- lapply(years, getPval) magns <- lapply(years, getMagn) # fill-in output matrices dimension <- nrow(sign.matrix) for (i in 1:dimension) { sign.matrix[i, i:dimension] <- unlist(signs[i]) pval.matrix[i, i:dimension] <- unlist(pvals[i]) slope.matrix[i, i:dimension] <- unlist(magns[i]) } sign.matrix <-...

Search
Clear search
Close search
Google apps
Main menu