18 datasets found
  1. Graph Input Data Example.xlsx

    • figshare.com
    xlsx
    Updated Dec 26, 2018
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Dr Corynen (2018). Graph Input Data Example.xlsx [Dataset]. http://doi.org/10.6084/m9.figshare.7506209.v1
    Explore at:
    xlsxAvailable download formats
    Dataset updated
    Dec 26, 2018
    Dataset provided by
    figshare
    Figsharehttp://figshare.com/
    Authors
    Dr Corynen
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    The various performance criteria applied in this analysis include the probability of reaching the ultimate target, the costs, elapsed times and system vulnerability resulting from any intrusion. This Excel file contains all the logical, probabilistic and statistical data entered by a user, and required for the evaluation of the criteria. It also reports the results of all the computations.

  2. Car-Sales-Analysis-Excel-Dashboard

    • kaggle.com
    zip
    Updated Feb 11, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Ibrahimryk (2025). Car-Sales-Analysis-Excel-Dashboard [Dataset]. https://www.kaggle.com/datasets/ibrahimryk/car-sales-analysis-excel-dashboard/code
    Explore at:
    zip(496747 bytes)Available download formats
    Dataset updated
    Feb 11, 2025
    Authors
    Ibrahimryk
    License

    https://creativecommons.org/publicdomain/zero/1.0/https://creativecommons.org/publicdomain/zero/1.0/

    Description

    his project involves the creation of an interactive Excel dashboard for SwiftAuto Traders to analyze and visualize car sales data. The dashboard includes several visualizations to provide insights into car sales, profits, and performance across different models and manufacturers. The project makes use of various charts and slicers in Excel for the analysis.

    Objective: The primary goal of this project is to showcase the ability to manipulate and visualize car sales data effectively using Excel. The dashboard aims to provide:

    Profit and Sales Analysis for each dealer. Sales Performance across various car models and manufacturers. Resale Value Analysis comparing prices and resale values. Insights into Retention Percentage by car models. Files in this Project: Car_Sales_Kaggle_DV0130EN_Lab3_Start.xlsx: The original dataset used to create the dashboard. dashboards.xlsx: The final Excel file that contains the complete dashboard with interactive charts and slicers. Key Visualizations: Average Price and Year Resale Value: A bar chart comparing the average price and resale value of various car models. Power Performance Factor: A column chart displaying the performance across different car models. Unit Sales by Model: A donut chart showcasing unit sales by car model. Retention Percentage: A pie chart illustrating customer retention by car model. Tools Used: Microsoft Excel for creating and organizing the visualizations and dashboard. Excel Slicers for interactive filtering. Charts: Bar charts, pie charts, column charts, and sunburst charts. How to Use: Download the Dataset: You can download the Car_Sales_Kaggle_DV0130EN_Lab3_Start.xlsx file from Kaggle and follow the steps to create a similar dashboard in Excel. Open the Dashboard: The dashboards.xlsx file contains the final version of the dashboard. Simply open it in Excel and start exploring the interactive charts and slicers.

  3. Petre_Slide_CategoricalScatterplotFigShare.pptx

    • figshare.com
    pptx
    Updated Sep 19, 2016
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Benj Petre; Aurore Coince; Sophien Kamoun (2016). Petre_Slide_CategoricalScatterplotFigShare.pptx [Dataset]. http://doi.org/10.6084/m9.figshare.3840102.v1
    Explore at:
    pptxAvailable download formats
    Dataset updated
    Sep 19, 2016
    Dataset provided by
    Figsharehttp://figshare.com/
    Authors
    Benj Petre; Aurore Coince; Sophien Kamoun
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Categorical scatterplots with R for biologists: a step-by-step guide

    Benjamin Petre1, Aurore Coince2, Sophien Kamoun1

    1 The Sainsbury Laboratory, Norwich, UK; 2 Earlham Institute, Norwich, UK

    Weissgerber and colleagues (2015) recently stated that ‘as scientists, we urgently need to change our practices for presenting continuous data in small sample size studies’. They called for more scatterplot and boxplot representations in scientific papers, which ‘allow readers to critically evaluate continuous data’ (Weissgerber et al., 2015). In the Kamoun Lab at The Sainsbury Laboratory, we recently implemented a protocol to generate categorical scatterplots (Petre et al., 2016; Dagdas et al., 2016). Here we describe the three steps of this protocol: 1) formatting of the data set in a .csv file, 2) execution of the R script to generate the graph, and 3) export of the graph as a .pdf file.

    Protocol

    • Step 1: format the data set as a .csv file. Store the data in a three-column excel file as shown in Powerpoint slide. The first column ‘Replicate’ indicates the biological replicates. In the example, the month and year during which the replicate was performed is indicated. The second column ‘Condition’ indicates the conditions of the experiment (in the example, a wild type and two mutants called A and B). The third column ‘Value’ contains continuous values. Save the Excel file as a .csv file (File -> Save as -> in ‘File Format’, select .csv). This .csv file is the input file to import in R.

    • Step 2: execute the R script (see Notes 1 and 2). Copy the script shown in Powerpoint slide and paste it in the R console. Execute the script. In the dialog box, select the input .csv file from step 1. The categorical scatterplot will appear in a separate window. Dots represent the values for each sample; colors indicate replicates. Boxplots are superimposed; black dots indicate outliers.

    • Step 3: save the graph as a .pdf file. Shape the window at your convenience and save the graph as a .pdf file (File -> Save as). See Powerpoint slide for an example.

    Notes

    • Note 1: install the ggplot2 package. The R script requires the package ‘ggplot2’ to be installed. To install it, Packages & Data -> Package Installer -> enter ‘ggplot2’ in the Package Search space and click on ‘Get List’. Select ‘ggplot2’ in the Package column and click on ‘Install Selected’. Install all dependencies as well.

    • Note 2: use a log scale for the y-axis. To use a log scale for the y-axis of the graph, use the command line below in place of command line #7 in the script.

    7 Display the graph in a separate window. Dot colors indicate

    replicates

    graph + geom_boxplot(outlier.colour='black', colour='black') + geom_jitter(aes(col=Replicate)) + scale_y_log10() + theme_bw()

    References

    Dagdas YF, Belhaj K, Maqbool A, Chaparro-Garcia A, Pandey P, Petre B, et al. (2016) An effector of the Irish potato famine pathogen antagonizes a host autophagy cargo receptor. eLife 5:e10856.

    Petre B, Saunders DGO, Sklenar J, Lorrain C, Krasileva KV, Win J, et al. (2016) Heterologous Expression Screens in Nicotiana benthamiana Identify a Candidate Effector of the Wheat Yellow Rust Pathogen that Associates with Processing Bodies. PLoS ONE 11(2):e0149035

    Weissgerber TL, Milic NM, Winham SJ, Garovic VD (2015) Beyond Bar and Line Graphs: Time for a New Data Presentation Paradigm. PLoS Biol 13(4):e1002128

    https://cran.r-project.org/

    http://ggplot2.org/

  4. m

    Dataset of development of business during the COVID-19 crisis

    • data.mendeley.com
    • narcis.nl
    Updated Nov 9, 2020
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Tatiana N. Litvinova (2020). Dataset of development of business during the COVID-19 crisis [Dataset]. http://doi.org/10.17632/9vvrd34f8t.1
    Explore at:
    Dataset updated
    Nov 9, 2020
    Authors
    Tatiana N. Litvinova
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    To create the dataset, the top 10 countries leading in the incidence of COVID-19 in the world were selected as of October 22, 2020 (on the eve of the second full of pandemics), which are presented in the Global 500 ranking for 2020: USA, India, Brazil, Russia, Spain, France and Mexico. For each of these countries, no more than 10 of the largest transnational corporations included in the Global 500 rating for 2020 and 2019 were selected separately. The arithmetic averages were calculated and the change (increase) in indicators such as profitability and profitability of enterprises, their ranking position (competitiveness), asset value and number of employees. The arithmetic mean values of these indicators for all countries of the sample were found, characterizing the situation in international entrepreneurship as a whole in the context of the COVID-19 crisis in 2020 on the eve of the second wave of the pandemic. The data is collected in a general Microsoft Excel table. Dataset is a unique database that combines COVID-19 statistics and entrepreneurship statistics. The dataset is flexible data that can be supplemented with data from other countries and newer statistics on the COVID-19 pandemic. Due to the fact that the data in the dataset are not ready-made numbers, but formulas, when adding and / or changing the values in the original table at the beginning of the dataset, most of the subsequent tables will be automatically recalculated and the graphs will be updated. This allows the dataset to be used not just as an array of data, but as an analytical tool for automating scientific research on the impact of the COVID-19 pandemic and crisis on international entrepreneurship. The dataset includes not only tabular data, but also charts that provide data visualization. The dataset contains not only actual, but also forecast data on morbidity and mortality from COVID-19 for the period of the second wave of the pandemic in 2020. The forecasts are presented in the form of a normal distribution of predicted values and the probability of their occurrence in practice. This allows for a broad scenario analysis of the impact of the COVID-19 pandemic and crisis on international entrepreneurship, substituting various predicted morbidity and mortality rates in risk assessment tables and obtaining automatically calculated consequences (changes) on the characteristics of international entrepreneurship. It is also possible to substitute the actual values identified in the process and following the results of the second wave of the pandemic to check the reliability of pre-made forecasts and conduct a plan-fact analysis. The dataset contains not only the numerical values of the initial and predicted values of the set of studied indicators, but also their qualitative interpretation, reflecting the presence and level of risks of a pandemic and COVID-19 crisis for international entrepreneurship.

  5. Sales Dashboard in Microsoft Excel

    • kaggle.com
    zip
    Updated Apr 14, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Bhavana Joshi (2023). Sales Dashboard in Microsoft Excel [Dataset]. https://www.kaggle.com/datasets/bhavanajoshij/sales-dashboard-in-microsoft-excel/discussion
    Explore at:
    zip(253363 bytes)Available download formats
    Dataset updated
    Apr 14, 2023
    Authors
    Bhavana Joshi
    Description

    This interactive sales dashboard is designed in Excel for B2C type of Businesses like Dmart, Walmart, Amazon, Shops & Supermarkets, etc. using Slicers, Pivot Tables & Pivot Chart.

    Dashboard Overview

    1. Sales dashboard ==> basically, it is designed for the B2C type of business. like Dmart, Walmart, Amazon, Shops & supermarkets, etc.
    2. Slices ==> slices are used to drill down the data, on the basis of yearly, monthly, by sales type, and by mode of payment.
    3. Total Sales/Total Profits ==> here is, the total sales, total profit, and profit percentage these all are combined into a monthly format and we can hide or unhide it to view it as individually or comparative.
    4. Product Visual ==> the visual indicates product-wise sales for the selected period. Only 10 products are visualized at a glance, and you can scroll up & down to view other products in the list.
    5. Daily Sales ==> It shows day-wise sales. (Area Chart)
    6. Sales Type/Payment Mode ==> It shows sales percentage contribution based on the type of selling and mode of payment.
    7. Top Product & Category ==> this is for the top-selling product and product category.
    8. Category ==> the final one is the category-wise sales contribution.

    Datasheets Overview

    1. The dataset has the master data sheet or you can call it a catalog. It is added in the table form.
    2. The first column is the product ID the list of items in this column is unique.
    3. Then we have the product column instead of these two columns, we can manage with only one also but I kept it separate because sometimes product names can be the same, but some parameters will be different, like price, supplier, etc.
    4. The next column is the category column, which is the product category. like cosmetics, foods, drinks, electronics, etc.
    5. Then we have 4th column which is the unit of measure (UOM) you can update it also, based on the products you have.
    6. And the last two columns are buying price and selling price, which means unit purchasing price and unit selling price.

    Input Sheet

    The first column is the date of Selling. The second column is the product ID. The third column is quantity. The fourth column is sales types, like direct selling, are purchased by a wholesaler or ordered online. The fifth column is a mode of payment, which is online or in cash. You can update these two as per requirements. The last one is a discount percentage. if you want to offer any discount, you can add it here.

    Analysis Sheet: where all backend calculations are performed.

    So, basically these are the four sheets mentioned above with different tasks.

    However, a sales dashboard enables organizations to visualize their real-time sales data and boost productivity.

    A dashboard is a very useful tool that brings together all the data in the forms of charts, graphs, statistics and many more visualizations which lead to data-driven and decision making.

    Questions & Answers

    1. What percentage of profit ratio of sales are displayed in the year 2021 and year 2022? ==> Total profit ratio of sales in the year 2021 is 19% with large sales of PRODUCT42, whereas profit ratio of sales for 2022 is 22% with large sales of PRODUCT30.
    2. Which is the top product that have large number of sales in year 2021-2022? ==> The top product in the year 2021 is PRODUCT42 with the total sales of $12,798 whereas in the year 2022 the top product is PRODUCT30 with the total sales of $13,888.
    3. In Area Chart which product is highly sold on 28th April 2022? ==> The large number of sales on 28th April 2022 is for PRODUCT14 with a 24% of profit ratio.
    4. What is the sales type and payment mode present? ==> The sale type and payment modes show the sales percentage contribution based on the type of selling and mode of payment. Here, the sale types are Direct Sales with 52%, Online Sales with 33% and Wholesaler with 15%. Also, the payment modes are Online mode and Cash equally distributed with 50%.
    5. In which month the direct sales are highest in the year 2022? ==> The highest direct sales can be easily identified which is designed by monthly format and it’s the November month where direct sales are highest with 28% as compared with other months.
    6. Which payment mode is highly received in the year 2021 and year 2022? ==> The payments received in the year 2021 are the cash payments with 52% as compared with online transactions which are 48%. Also, the cash payment highly received is in the month of March, July and October with direct sales of 42%, Online with 45% and wholesaler with 13% with large sales of PRODUCT24. ==> The payments received in the year 2022 are the Online payments with 52% as compared with cash payments which are 48%. Also, the online payment highly received is in the month of Jan, Sept and December with direct sales of 45%, Online with 37% and whole...
  6. Store Data Analysis using MS excel

    • kaggle.com
    zip
    Updated Mar 10, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    NisshaaChoudhary (2024). Store Data Analysis using MS excel [Dataset]. https://www.kaggle.com/datasets/nisshaachoudhary/store-data-analysis-using-ms-excel/discussion
    Explore at:
    zip(13048217 bytes)Available download formats
    Dataset updated
    Mar 10, 2024
    Authors
    NisshaaChoudhary
    License

    https://creativecommons.org/publicdomain/zero/1.0/https://creativecommons.org/publicdomain/zero/1.0/

    Description

    Vrinda Store: Interactive Ms Excel dashboardVrinda Store: Interactive Ms Excel dashboard Feb 2024 - Mar 2024Feb 2024 - Mar 2024 The owner of Vrinda store wants to create an annual sales report for 2022. So that their employees can understand their customers and grow more sales further. Questions asked by Owner of Vrinda store are as follows:- 1) Compare the sales and orders using single chart. 2) Which month got the highest sales and orders? 3) Who purchased more - women per men in 2022? 4) What are different order status in 2022?

    And some other questions related to business. The owner of Vrinda store wanted a visual story of their data. Which can depict all the real time progress and sales insight of the store. This project is a Ms Excel dashboard which presents an interactive visual story to help the Owner and employees in increasing their sales. Task performed : Data cleaning, Data processing, Data analysis, Data visualization, Report. Tool used : Ms Excel The owner of Vrinda store wants to create an annual sales report for 2022. So that their employees can understand their customers and grow more sales further. Questions asked by Owner of Vrinda store are as follows:- 1) Compare the sales and orders using single chart. 2) Which month got the highest sales and orders? 3) Who purchased more - women per men in 2022? 4) What are different order status in 2022? And some other questions related to business. The owner of Vrinda store wanted a visual story of their data. Which can depict all the real time progress and sales insight of the store. This project is a Ms Excel dashboard which presents an interactive visual story to help the Owner and employees in increasing their sales. Task performed : Data cleaning, Data processing, Data analysis, Data visualization, Report. Tool used : Ms Excel Skills: Data Analysis · Data Analytics · ms excel · Pivot Tables

  7. Project Priority Matrix (Dynamic Excel Template)

    • kaggle.com
    zip
    Updated Oct 24, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Asjad (2025). Project Priority Matrix (Dynamic Excel Template) [Dataset]. https://www.kaggle.com/datasets/asjadd/project-priority-matrix-dynamic-excel-template
    Explore at:
    zip(50515 bytes)Available download formats
    Dataset updated
    Oct 24, 2025
    Authors
    Asjad
    License

    https://creativecommons.org/publicdomain/zero/1.0/https://creativecommons.org/publicdomain/zero/1.0/

    Description

    Project Priority Matrix (Dynamic Excel Tool)

    Overview

    This dataset provides a dynamic Excel model for prioritizing projects based on Feasibility, Impact, and Size.
    It visualizes project data on a Bubble Chart that updates automatically when new projects are added.

    Use this tool to make data-driven prioritization decisions by identifying which projects are most feasible and high-impact.

    Goal

    Organizations often struggle to compare multiple initiatives objectively.
    This matrix helps teams quickly determine which projects to pursue first by visualizing:

    • Feasibility → How achievable a project is
    • Impact → The potential benefit or value it delivers
    • Size → The level of effort or resources required

    How It Works

    1. Each project is rated on a 1–10 scale for:
      • Feasibility
      • Impact
      • Size
    2. The Excel file uses a Bubble Chart:
      • X-axis: Feasibility
      • Y-axis: Impact
      • Bubble size: Project Size
    3. The chart automatically updates when new projects or scores are added.

    Example (partial data):

    CriteriaProject 1Project 2Project 3Project 4Project 5Project 6Project 7Project 8
    Feasibility79527268
    Impact84466777
    Size102374431

    Interpretation Guide

    QuadrantDescriptionAction
    High Feasibility / High ImpactQuick winsTop Priority
    High Impact / Low FeasibilityValuable but riskyPlan carefully
    Low Impact / High FeasibilityEasy but minor valueOptional
    Low Impact / Low FeasibilityLow returnDefer or drop

    Excel Features

    • Dynamic Bubble Chart (updates with new data)
    • Named Ranges for auto-expanding data
    • Optional Conditional Formatting
    • Data Validation for consistent scoring

    How to Use

    1. Download and open Project_Priority_Matrix.xlsx.
    2. Go to the Data sheet.
    3. Add your project names and scores (1–10).
    4. Watch the chart update instantly to reflect your data.

    You can use this for: - Portfolio management
    - Product or feature prioritization
    - Strategy planning workshops

    File Information

    • File: Project_Priority_Matrix.xlsx
    • Format: Excel (.xlsx)
    • Version: 1.0
    • Last Updated: October 2025

    License

    Free for personal and organizational use.
    Attribution is appreciated if you share or adapt this file.

    Author: [Asjad]
    Contact: [m.asjad2000@gmail.com]
    Compatible With: Microsoft Excel 2019+ / Office 365

  8. D

    Graph Database For Security Market Research Report 2033

    • dataintelo.com
    csv, pdf, pptx
    Updated Oct 1, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Dataintelo (2025). Graph Database For Security Market Research Report 2033 [Dataset]. https://dataintelo.com/report/graph-database-for-security-market
    Explore at:
    csv, pdf, pptxAvailable download formats
    Dataset updated
    Oct 1, 2025
    Dataset authored and provided by
    Dataintelo
    License

    https://dataintelo.com/privacy-and-policyhttps://dataintelo.com/privacy-and-policy

    Time period covered
    2024 - 2032
    Area covered
    Global
    Description

    Graph Database for Security Market Outlook



    According to our latest research, the global graph database for security market size reached USD 2.1 billion in 2024. This dynamic sector is expanding rapidly, supported by a robust compound annual growth rate (CAGR) of 22.7% from 2025 to 2033. By the end of the forecast period in 2033, the market is expected to attain a value of USD 16.3 billion. This impressive trajectory is primarily driven by escalating cyber threats, the proliferation of complex digital ecosystems, and the increasing demand for advanced analytics in security operations.




    One of the most significant growth factors for the graph database for security market is the exponential rise in cyberattacks and sophisticated threat vectors targeting organizations worldwide. As digital transformation accelerates across industries, enterprises are generating vast volumes of interconnected data, creating new vulnerabilities and attack surfaces. Traditional relational databases struggle to effectively manage and analyze such complex, highly connected datasets. In contrast, graph databases excel at mapping relationships and patterns, making them invaluable for identifying suspicious activities, tracking threat actors, and correlating diverse security events in real-time. The ability to visualize and traverse connections at scale empowers security teams to detect advanced persistent threats, insider attacks, and fraud schemes that would otherwise go unnoticed.




    Another pivotal driver is the increasing regulatory pressure and compliance requirements faced by organizations in sectors such as BFSI, healthcare, and government. Regulations including GDPR, HIPAA, and PCI DSS demand robust data protection, rigorous access controls, and comprehensive audit trails. Graph database technologies enable organizations to model complex access hierarchies, monitor user behaviors, and ensure compliance with evolving legal frameworks. By providing granular visibility into user roles, permissions, and interactions, these solutions facilitate proactive risk management and timely incident response. The integration of artificial intelligence and machine learning with graph databases further enhances predictive analytics and automation in security operations, reducing the burden on human analysts and improving overall resilience.




    The rapid adoption of cloud computing, IoT devices, and remote work models is reshaping the security landscape and fueling demand for graph database solutions. As organizations migrate workloads to multi-cloud and hybrid environments, the complexity of managing identities, access rights, and network flows increases exponentially. Graph databases provide a unified view of assets, users, and their interdependencies, enabling security teams to identify misconfigurations, detect lateral movement, and enforce zero-trust principles. The scalability and flexibility of cloud-based graph database offerings are particularly attractive to enterprises seeking to modernize their security infrastructure without incurring significant capital expenditures. Strategic investments in research and development, partnerships with cybersecurity vendors, and the emergence of managed graph database services are further propelling market growth.




    Regionally, North America dominates the graph database for security market, accounting for the largest revenue share in 2024. This leadership is attributed to the presence of major technology providers, high cybersecurity spending, and early adoption of advanced analytics solutions. Europe follows closely, driven by stringent data privacy regulations and a strong focus on digital sovereignty. The Asia Pacific region is witnessing the fastest growth, supported by rapid digitalization, government initiatives, and increased awareness of cybersecurity risks. Latin America and the Middle East & Africa are emerging as promising markets, although challenges such as limited infrastructure and skills gaps persist. Overall, regional dynamics are shaped by varying regulatory landscapes, industry maturity, and investment levels in digital security.



    Component Analysis



    The graph database for security market is segmented by component into software and services, each playing a critical role in the adoption and effectiveness of graph database solutions. The software segment comprises graph database management systems, visualization tools, analytics engines, and integration platforms. Thes

  9. c

    Niagara Open Data

    • catalog.civicdataecosystem.org
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Niagara Open Data [Dataset]. https://catalog.civicdataecosystem.org/dataset/niagara-open-data
    Explore at:
    Description

    The Ontario government, generates and maintains thousands of datasets. Since 2012, we have shared data with Ontarians via a data catalogue. Open data is data that is shared with the public. Click here to learn more about open data and why Ontario releases it. Ontario’s Open Data Directive states that all data must be open, unless there is good reason for it to remain confidential. Ontario’s Chief Digital and Data Officer also has the authority to make certain datasets available publicly. Datasets listed in the catalogue that are not open will have one of the following labels: If you want to use data you find in the catalogue, that data must have a licence – a set of rules that describes how you can use it. A licence: Most of the data available in the catalogue is released under Ontario’s Open Government Licence. However, each dataset may be shared with the public under other kinds of licences or no licence at all. If a dataset doesn’t have a licence, you don’t have the right to use the data. If you have questions about how you can use a specific dataset, please contact us. The Ontario Data Catalogue endeavors to publish open data in a machine readable format. For machine readable datasets, you can simply retrieve the file you need using the file URL. The Ontario Data Catalogue is built on CKAN, which means the catalogue has the following features you can use when building applications. APIs (Application programming interfaces) let software applications communicate directly with each other. If you are using the catalogue in a software application, you might want to extract data from the catalogue through the catalogue API. Note: All Datastore API requests to the Ontario Data Catalogue must be made server-side. The catalogue's collection of dataset metadata (and dataset files) is searchable through the CKAN API. The Ontario Data Catalogue has more than just CKAN's documented search fields. You can also search these custom fields. You can also use the CKAN API to retrieve metadata about a particular dataset and check for updated files. Read the complete documentation for CKAN's API. Some of the open data in the Ontario Data Catalogue is available through the Datastore API. You can also search and access the machine-readable open data that is available in the catalogue. How to use the API feature: Read the complete documentation for CKAN's Datastore API. The Ontario Data Catalogue contains a record for each dataset that the Government of Ontario possesses. Some of these datasets will be available to you as open data. Others will not be available to you. This is because the Government of Ontario is unable to share data that would break the law or put someone's safety at risk. You can search for a dataset with a word that might describe a dataset or topic. Use words like “taxes” or “hospital locations” to discover what datasets the catalogue contains. You can search for a dataset from 3 spots on the catalogue: the homepage, the dataset search page, or the menu bar available across the catalogue. On the dataset search page, you can also filter your search results. You can select filters on the left hand side of the page to limit your search for datasets with your favourite file format, datasets that are updated weekly, datasets released by a particular organization, or datasets that are released under a specific licence. Go to the dataset search page to see the filters that are available to make your search easier. You can also do a quick search by selecting one of the catalogue’s categories on the homepage. These categories can help you see the types of data we have on key topic areas. When you find the dataset you are looking for, click on it to go to the dataset record. Each dataset record will tell you whether the data is available, and, if so, tell you about the data available. An open dataset might contain several data files. These files might represent different periods of time, different sub-sets of the dataset, different regions, language translations, or other breakdowns. You can select a file and either download it or preview it. Make sure to read the licence agreement to make sure you have permission to use it the way you want. Read more about previewing data. A non-open dataset may be not available for many reasons. Read more about non-open data. Read more about restricted data. Data that is non-open may still be subject to freedom of information requests. The catalogue has tools that enable all users to visualize the data in the catalogue without leaving the catalogue – no additional software needed. Have a look at our walk-through of how to make a chart in the catalogue. Get automatic notifications when datasets are updated. You can choose to get notifications for individual datasets, an organization’s datasets or the full catalogue. You don’t have to provide and personal information – just subscribe to our feeds using any feed reader you like using the corresponding notification web addresses. Copy those addresses and paste them into your reader. Your feed reader will let you know when the catalogue has been updated. The catalogue provides open data in several file formats (e.g., spreadsheets, geospatial data, etc). Learn about each format and how you can access and use the data each file contains. A file that has a list of items and values separated by commas without formatting (e.g. colours, italics, etc.) or extra visual features. This format provides just the data that you would display in a table. XLSX (Excel) files may be converted to CSV so they can be opened in a text editor. How to access the data: Open with any spreadsheet software application (e.g., Open Office Calc, Microsoft Excel) or text editor. Note: This format is considered machine-readable, it can be easily processed and used by a computer. Files that have visual formatting (e.g. bolded headers and colour-coded rows) can be hard for machines to understand, these elements make a file more human-readable and less machine-readable. A file that provides information without formatted text or extra visual features that may not follow a pattern of separated values like a CSV. How to access the data: Open with any word processor or text editor available on your device (e.g., Microsoft Word, Notepad). A spreadsheet file that may also include charts, graphs, and formatting. How to access the data: Open with a spreadsheet software application that supports this format (e.g., Open Office Calc, Microsoft Excel). Data can be converted to a CSV for a non-proprietary format of the same data without formatted text or extra visual features. A shapefile provides geographic information that can be used to create a map or perform geospatial analysis based on location, points/lines and other data about the shape and features of the area. It includes required files (.shp, .shx, .dbt) and might include corresponding files (e.g., .prj). How to access the data: Open with a geographic information system (GIS) software program (e.g., QGIS). A package of files and folders. The package can contain any number of different file types. How to access the data: Open with an unzipping software application (e.g., WinZIP, 7Zip). Note: If a ZIP file contains .shp, .shx, and .dbt file types, it is an ArcGIS ZIP: a package of shapefiles which provide information to create maps or perform geospatial analysis that can be opened with ArcGIS (a geographic information system software program). A file that provides information related to a geographic area (e.g., phone number, address, average rainfall, number of owl sightings in 2011 etc.) and its geospatial location (i.e., points/lines). How to access the data: Open using a GIS software application to create a map or do geospatial analysis. It can also be opened with a text editor to view raw information. Note: This format is machine-readable, and it can be easily processed and used by a computer. Human-readable data (including visual formatting) is easy for users to read and understand. A text-based format for sharing data in a machine-readable way that can store data with more unconventional structures such as complex lists. How to access the data: Open with any text editor (e.g., Notepad) or access through a browser. Note: This format is machine-readable, and it can be easily processed and used by a computer. Human-readable data (including visual formatting) is easy for users to read and understand. A text-based format to store and organize data in a machine-readable way that can store data with more unconventional structures (not just data organized in tables). How to access the data: Open with any text editor (e.g., Notepad). Note: This format is machine-readable, and it can be easily processed and used by a computer. Human-readable data (including visual formatting) is easy for users to read and understand. A file that provides information related to an area (e.g., phone number, address, average rainfall, number of owl sightings in 2011 etc.) and its geospatial location (i.e., points/lines). How to access the data: Open with a geospatial software application that supports the KML format (e.g., Google Earth). Note: This format is machine-readable, and it can be easily processed and used by a computer. Human-readable data (including visual formatting) is easy for users to read and understand. This format contains files with data from tables used for statistical analysis and data visualization of Statistics Canada census data. How to access the data: Open with the Beyond 20/20 application. A database which links and combines data from different files or applications (including HTML, XML, Excel, etc.). The database file can be converted to a CSV/TXT to make the data machine-readable, but human-readable formatting will be lost. How to access the data: Open with Microsoft Office Access (a database management system used to develop application software). A file that keeps the original layout and

  10. g

    Data from: Stratigraphic Classification Table for the PetroPhysical Property...

    • dataservices.gfz-potsdam.de
    Updated 2019
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Kristian Bär; Philipp Mielke (2019). Stratigraphic Classification Table for the PetroPhysical Property Database P³ [Dataset]. http://doi.org/10.5880/gfz.4.8.2019.p3.s
    Explore at:
    Dataset updated
    2019
    Dataset provided by
    GFZ Data Services
    datacite
    Authors
    Kristian Bär; Philipp Mielke
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Area covered
    Earth
    Dataset funded by
    FP7 Energy
    Description

    This data publication is part of the 'P³-Petrophysical Property Database' project, which was developed within the EC funded project IMAGE (Integrated Methods for Advanced Geothermal Exploration, EU grant agreement No. 608553) and consists of a scientific paper, a full report on the database, the database as excel and .csv files and additional tables for a hierarchical classification of the petrography and stratigraphy of the investigated rock samples (see related references). This publication here provides a hierarchical interlinked stratigraphic classification according to the chronostratigraphical units of the international chronostratigraphic chart of the IUGS v2016/04 (Cohen et al. 2013, updated) according to international standardisation. As addition to this IUGS chart, which is also documented in GeoSciML, stratigraphic IDs and parent IDs were included to define the direct relationships between the stratigraphic terms. The P³ database aims at providing easily accessible, peer-reviewed information on physical rock properties relevant for geothermal exploration and reservoir characterization in one single compilation. Collected data include hydraulic, thermophysical and mechanical properties and, in addition, electrical resistivity and magnetic susceptibility. Each measured value is complemented by relevant meta-information such as the corresponding sample location, petrographic description, chronostratigraphic age and, most important, original citation. The original stratigraphic and petrographic descriptions are transferred to standardized catalogues following a hierarchical structure ensuring intercomparability for statistical analysis, of which the stratigraphic catalogue is presented here. These chronostratigraphic units are compiled to ensure that formations of a certain age are connected to the corresponding stratigraphic epoch, period or erathem. Thus, the chronostratigraphic units are directly correlated to each other by their stratigraphic ID and stratigraphic parent ID and can thus be used for interlinked data assessment of the petrophysical properties of samples of an according stratigraphic unit.

  11. ICSE 2025 - Artifact

    • figshare.com
    pdf
    Updated Jan 24, 2025
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    FARIDAH AKINOTCHO (2025). ICSE 2025 - Artifact [Dataset]. http://doi.org/10.6084/m9.figshare.28194605.v1
    Explore at:
    pdfAvailable download formats
    Dataset updated
    Jan 24, 2025
    Dataset provided by
    Figsharehttp://figshare.com/
    Authors
    FARIDAH AKINOTCHO
    License

    MIT Licensehttps://opensource.org/licenses/MIT
    License information was derived automatically

    Description

    Mobile Application Coverage: The 30% Curse and Ways Forward## Purpose In this artifact, we provide the information about our benchmarks used for manual and tool exploration. We include coverage results achieved by tools and human analysts as well as plots of the coverage progression over time for analysts. We further provide manual analysis results for our case study, more specifically extracted reasons for unreachability for the case study apps and extracted code-level properties, which constitute a ground truth for future work in coverage explainability. Finally, we identify a list of beyond-GUI exploration tools and categorize them for future work to take inspiration from. We are claiming available and reusable badges; the artifact is fully aligned with the results described in our paper and comprehensively documented.## ProvenanceThe paper preprint is available here: https://people.ece.ubc.ca/mjulia/publications/Mobile_Application_Coverage_ICSE2025.pdf## Data The artifact submission is organized into five parts:- 'BenchInfo' excel sheet describing our experiment dataset- 'Coverage' folder containing coverage results for tools and analysts (RQ1) - 'Reasons' excel sheet describing our manually extracted reasons for unreachability (RQ2)- 'ActivationProperties' excel sheet describing our manually extracted code properties of unreached activities (RQ3)- 'ActivationProperties-Graph' pdf which presents combinations of the extracted code properties in a graph format.- 'BeyondGUI' folder containing information about identified techniques which go beyond GUI exploration.The artifact requires about 15MB of storage.### Dataset: 'BenchInfo.xlsx'This file list the full application dataset used for experiments into three tabs: 'BenchNotGP' (apps from AndroTest dataset which are not on Google Play), 'BenchGP' (apps from AndroTest which are also on Google Play) and 'TopGP' (top ranked free apps from Google Play). Each tab contains the following information:- Application Name- Package Name- Version Used (Latest)- Original Version- # Activities- Minimum SDK- Target SDK- # Permissions (in Manifest)- List of Permissions (in Manifest)- # Features (in Manifest)- List of Features (in Manifest)The 'TopGP' sheet also includes Google-Play-specific information, namely:- Category (one of 32 app categories)- Downloads- Popularity RankThe 'BenchGP' and 'BenchNotGP' sheets also include the original version (included in the AndroTest benchmark) and the source (one of F-Droid, Github or Google Code Archives).### RQ1: 'Coverage'The 'Coverage' folder includes coverage results for tools and analysts, and is structured as follows:- 'CoverageResults.xlsx": An excel sheet containing the coverage results achieved by each human analysts and tool. - The first tab described the results over all apps for analysts combined, tools combined, and analysts + tools, which map to Table II in the paper. - Each of the following 42 tab, one per app in TopGP, marks the activities reached by Analyst 1, Analyst 2, Tool 1 (ape) and Tool 2 (fastbot), with an 'x' in the corresponding column to indicate that the activity was reached by the given agent.- 'Plots': A folder containing plots of the progressive coverage over time of analysts, split into one folder for 'Analyst1' and one for 'Analyst2'. - Each of the analysts' folder includes a subfolder per benchmark ('BenchNotGP', 'BenchGP' and 'TopGP'), containing as many png files as applications in the benchmark (respectively 47, 14 and 42 image files) named 'ANALYST_[X]_[APP_PACKAGE_NAME]'.png.### RQ2: 'Reasons.xslx'This file contains the extracted reasons for unreachability for the 11 apps manually analyzed. - The 'Summary' tab provides an overview of unreached activities per reasons over all apps and per app, which corresponds to Table III in the paper. - The following 11 tabs, each corresponding to and named after a single application, describe the reasons associated with each activity of that application. Each column corresponds to a single reason and 'x' indicates that the activity is unreached due to the reason in that column. The top row sums up the total number of activities unreached due to a given reason in each column.- The activities at the bottom which are greyed out correspond to activities that were reached during exploration, and are thus excluded from the reason extraction.### RQ3: 'ActivationProperties.xslx'This file contains the full list of activation properties extracted for each of the 185 activities analyzed for RQ2.The first half of the columns (columns C-M) correspond to the reasons (excluding Transitive, Inconclusive and No Caller) and the second half (columns N-AD) correspond to properties described in Figure 5 in the paper, namely:- Exported- Activation Location: - Code: GUI/lifecycle, Other Android or App-specific - Manifest- Activation Guards: - Enforcement: In Code or In Resources - Restriction: Mandatory or Discretionary- Data: - Type: Parameters, Execution Dependencies - Format: Primitive, Strings, ObjectsThe rows are grouped by applications, and each row correspond to an activity of that application. 'x' in a given column indicates the presence of the property in that column within the analyzed path to the activity. The third and fourth rows sums up the numbers and percentages for each property, as reported in Figure 5.### RQ3: 'ActivationProperties-Graph.pdf'This file shows combinations of the individual properties listed in 'ActivationProperties.xlsx' in a graph format, extending the combinations described in Table IV with data (types and format) and reasons for unreachability.### BeyondGUIThis folder includes:- 'ToolInfo.xlsx': an excel sheet listing the identified 22 beyond-GUI papers, the date of publication, availability, invasiveness (Source code, Bytecode, framework, OS) and their targeting strategy (None, Manual or Automated).- ToolClassification.pdf': a pdf file describing our paper selection methodology as well as a classication of the techniques in terms of Invocation Strategy, Navigation Strategy, Value Generation Strategy, and Value Generation Types. We fully introduced these categories in the pdf file.## Requirements & technology skills assumed by the reviewer evaluating the artifactThe artifact entirely consists of Excel sheets which can be opened with common Excel visualization software, i.e., Microsoft Excel, coverage plots as PNG files and PDF files. It requires about 15MB of storage in total.No other specific technology skills are required of the reviewer evaluating the artifact.

  12. Superstore Sales Analysis

    • kaggle.com
    zip
    Updated Oct 21, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Ali Reda Elblgihy (2023). Superstore Sales Analysis [Dataset]. https://www.kaggle.com/datasets/aliredaelblgihy/superstore-sales-analysis/versions/1
    Explore at:
    zip(3009057 bytes)Available download formats
    Dataset updated
    Oct 21, 2023
    Authors
    Ali Reda Elblgihy
    Description

    Analyzing sales data is essential for any business looking to make informed decisions and optimize its operations. In this project, we will utilize Microsoft Excel and Power Query to conduct a comprehensive analysis of Superstore sales data. Our primary objectives will be to establish meaningful connections between various data sheets, ensure data quality, and calculate critical metrics such as the Cost of Goods Sold (COGS) and discount values. Below are the key steps and elements of this analysis:

    1- Data Import and Transformation:

    • Gather and import relevant sales data from various sources into Excel.
    • Utilize Power Query to clean, transform, and structure the data for analysis.
    • Merge and link different data sheets to create a cohesive dataset, ensuring that all data fields are connected logically.

    2- Data Quality Assessment:

    • Perform data quality checks to identify and address issues like missing values, duplicates, outliers, and data inconsistencies.
    • Standardize data formats and ensure that all data is in a consistent, usable state.

    3- Calculating COGS:

    • Determine the Cost of Goods Sold (COGS) for each product sold by considering factors like purchase price, shipping costs, and any additional expenses.
    • Apply appropriate formulas and calculations to determine COGS accurately.

    4- Discount Analysis:

    • Analyze the discount values offered on products to understand their impact on sales and profitability.
    • Calculate the average discount percentage, identify trends, and visualize the data using charts or graphs.

    5- Sales Metrics:

    • Calculate and analyze various sales metrics, such as total revenue, profit margins, and sales growth.
    • Utilize Excel functions to compute these metrics and create visuals for better insights.

    6- Visualization:

    • Create visualizations, such as charts, graphs, and pivot tables, to present the data in an understandable and actionable format.
    • Visual representations can help identify trends, outliers, and patterns in the data.

    7- Report Generation:

    • Compile the findings and insights into a well-structured report or dashboard, making it easy for stakeholders to understand and make informed decisions.

    Throughout this analysis, the goal is to provide a clear and comprehensive understanding of the Superstore's sales performance. By using Excel and Power Query, we can efficiently manage and analyze the data, ensuring that the insights gained contribute to the store's growth and success.

  13. Bank Transaction Analytics Dashboard – SQL + Excel

    • kaggle.com
    zip
    Updated Aug 18, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Prachi Singh (2025). Bank Transaction Analytics Dashboard – SQL + Excel [Dataset]. https://www.kaggle.com/datasets/prachisingh29ds/bank-transaction-analytics-dashboard-sql-excel
    Explore at:
    zip(2856220 bytes)Available download formats
    Dataset updated
    Aug 18, 2025
    Authors
    Prachi Singh
    License

    https://creativecommons.org/publicdomain/zero/1.0/https://creativecommons.org/publicdomain/zero/1.0/

    Description

    📊 Bank Transaction Analytics Dashboard – SQL + Excel

    🔹 Overview

    This project focuses on Bank Transaction Analysis using a combination of SQL scripts and Excel dashboards. The goal is to provide insights into customer spending patterns, payment modes, suspicious transactions, and overall financial trends.

    The dataset and analysis files can help learners and professionals understand how SQL and Excel can be used together for business decision-making, customer behavior tracking, and data-driven insights.

    🔹 Contents

    The dataset includes the following resources:

    📂 SQL Scripts:

    Create & Insert tables

    15 Basic Queries

    15 Advanced Queries

    📂 CSV File:

    Bank Transaction Analytics.csv (main dataset)

    📂 Excel Charts:

    Pie, Bar, Column, Line, Doughnut charts

    Final Interactive Dashboard

    📂 Screenshots:

    Query outputs, Charts, and Final Dashboard visualization

    📂 PDF Reports:

    Project Report

    Dashboard Report

    📄 README.md:

    Complete documentation and step-by-step explanation

    🔹 Key Insights

    26–35 age group spent the most across categories.

    Amazon identified as the top merchant.

    NetBanking showed the highest share compared to POS/UPI.

    Travel & Shopping emerged as dominant categories.

    🔹 Applications

    Detecting suspicious transactions.

    Understanding customer behavior.

    Identifying top merchants and categories.

    Building business intelligence dashboards.

    🔹 How to Use

    Download the dataset and SQL scripts.

    Run Bank_Transaction_Analytics.SQL to create and insert data.

    Execute the queries (Basic + Advanced) for insights.

    Open Excel files to explore interactive charts and dashboards.

    Refer to Project Report PDF for documentation.

    🔹 Author

    👩‍💻 Created by: Prachi Singh

    GitHub: Bank Transaction Analytics Dashboard(https://github.com/prachi-singh-ds/Bank-Transaction-Analytics-Dashboard)

    ⚡This project is a complete SQL + Excel integration case study and is suitable for Data Science, Business Analytics, and Data Engineering portfolios.

  14. IPPS DRG Provider Summary

    • kaggle.com
    zip
    Updated Jan 23, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    The Devastator (2023). IPPS DRG Provider Summary [Dataset]. https://www.kaggle.com/datasets/thedevastator/ipps-drg-provider-summary
    Explore at:
    zip(8432015 bytes)Available download formats
    Dataset updated
    Jan 23, 2023
    Authors
    The Devastator
    License

    Open Database License (ODbL) v1.0https://www.opendatacommons.org/licenses/odbl/1.0/
    License information was derived automatically

    Description

    IPPS DRG Provider Summary

    Average Discharges, Charges, and Medicare Payments

    By Health [source]

    About this dataset

    This dataset is a valuable resource for gaining insight into Inpatient Prospective Payment System (IPPS) utilization, average charges and average Medicare payments across the top 100 Diagnosis-Related Groups (DRG). With column categories such as DRG Definition, Hospital Referral Region Description, Total Discharges, Average Covered Charges, Average Medicare Payments and Average Medicare Payments 2 this dataset enables researchers to discover and assess healthcare trends in areas such as provider payment comparsons by geographic location or compare service cost across hospital. Visualize the data using various methods to uncover unique information and drive further hospital research

    More Datasets

    For more datasets, click here.

    Featured Notebooks

    • 🚨 Your notebook can be here! 🚨!

    How to use the dataset

    This dataset provides a provider level summary of Inpatient Prospective Payment System (IPPS) discharges, average charges and average Medicare payments for the Top 100 Diagnosis-Related Groups (DRG). This data can be used to analyze cost and utilization trends across hospital DRGs.

    To make the most use of this dataset, here are some steps to consider:

    • Understand what each column means in the table: Each column provides different information from the DRG Definition to Hospital Referral Region Description and Average Medicare Payments.
    • Analyze the data by looking for patterns amongst the relevant columns: Compare different aspects such as total discharges or average Medicare payments by hospital referral region or DRG Definition. This can help identify any potential trends amongst different categories within your analysis.
    • Generate visualizations: Create charts, graphs, or maps that display your data in an easy-to-understand format using tools such as Microsoft Excel or Tableau. Such visuals may reveal more insights into patterns within your data than simply reading numerical values on a spreadsheet could provide alone.

    Research Ideas

    • Identifying potential areas of cost savings by drilling down to particular DRGs and hospital regions with the highest average covered charges compared to average Medicare payments.
    • Establishing benchmarks for typical charges and payments across different DRGs and hospital regions to help providers set market-appropriate prices.
    • Analyzing trends in total discharges, charges and Medicare payments over time, allowing healthcare organizations to measure their performance against regional peers

    Acknowledgements

    If you use this dataset in your research, please credit the original authors. Data Source

    License

    License: Open Database License (ODbL) v1.0 - You are free to: - Share - copy and redistribute the material in any medium or format. - Adapt - remix, transform, and build upon the material for any purpose, even commercially. - You must: - Give appropriate credit - Provide a link to the license, and indicate if changes were made. - ShareAlike - You must distribute your contributions under the same license as the original. - Keep intact - all notices that refer to this license, including copyright notices. - No Derivatives - If you remix, transform, or build upon the material, you may not distribute the modified material. - No additional restrictions - You may not apply legal terms or technological measures that legally restrict others from doing anything the license permits.

    Columns

    File: 97k6-zzx3.csv | Column name | Description | |:-----------------------------------------|:------------------------------------------------------| | drg_definition | Diagnosis-Related Group (DRG) definition. (String) | | average_medicare_payments | Average Medicare payments for each DRG. (Numeric) | | hospital_referral_region_description | Description of the hospital referral region. (String) | | total_discharges | Total number of discharges for each DRG. (Numeric) | | average_covered_charges | Average covered charges for each DRG. (Numeric) | | average_medicare_payments_2 | Average Medicare payments for each DRG. (Numeric) |

    **File: Inpatient_Prospective_Payment_System_IPPS_Provider_Summary_for_the_Top_100_Diagnosis-Related_Groups_DRG...

  15. 18 excel spreadsheets by species and year giving reproduction and growth...

    • catalog.data.gov
    • data.wu.ac.at
    Updated Aug 17, 2024
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    U.S. EPA Office of Research and Development (ORD) (2024). 18 excel spreadsheets by species and year giving reproduction and growth data. One excel spreadsheet of herbicide treatment chemistry. [Dataset]. https://catalog.data.gov/dataset/18-excel-spreadsheets-by-species-and-year-giving-reproduction-and-growth-data-one-excel-sp
    Explore at:
    Dataset updated
    Aug 17, 2024
    Dataset provided by
    United States Environmental Protection Agencyhttp://www.epa.gov/
    Description

    Excel spreadsheets by species (4 letter code is abbreviation for genus and species used in study, year 2010 or 2011 is year data collected, SH indicates data for Science Hub, date is date of file preparation). The data in a file are described in a read me file which is the first worksheet in each file. Each row in a species spreadsheet is for one plot (plant). The data themselves are in the data worksheet. One file includes a read me description of the column in the date set for chemical analysis. In this file one row is an herbicide treatment and sample for chemical analysis (if taken). This dataset is associated with the following publication: Olszyk , D., T. Pfleeger, T. Shiroyama, M. Blakely-Smith, E. Lee , and M. Plocher. Plant reproduction is altered by simulated herbicide drift toconstructed plant communities. ENVIRONMENTAL TOXICOLOGY AND CHEMISTRY. Society of Environmental Toxicology and Chemistry, Pensacola, FL, USA, 36(10): 2799-2813, (2017).

  16. NSE - Nifty 50 Index Minute data (2015 to 2025)

    • kaggle.com
    zip
    Updated Aug 6, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Deba (2025). NSE - Nifty 50 Index Minute data (2015 to 2025) [Dataset]. https://www.kaggle.com/datasets/debashis74017/nifty-50-minute-data
    Explore at:
    zip(184768242 bytes)Available download formats
    Dataset updated
    Aug 6, 2025
    Authors
    Deba
    License

    https://creativecommons.org/publicdomain/zero/1.0/https://creativecommons.org/publicdomain/zero/1.0/

    Description

    UPDATED EVERY WEEK Last Update - 26th July 2025

    Disclaimer!!! Data uploaded here are collected from the internet and some google drive. The sole purposes of uploading these data are to provide this Kaggle community with a good source of data for analysis and research. I don't own these datasets and am also not responsible for them legally by any means. I am not charging anything (either money or any favor) for this dataset. RESEARCH PURPOSE ONLY

    Context

    • The NIFTY 50 is a well-diversified 50 stock index and it represents 13 important sectors of the economy.
    • It is used for a variety of purposes such as benchmarking fund portfolios, index-based derivatives, and index funds.
    • NIFTY 50 is owned and managed by NSE Indices Limited.
    • The NIFTY 50 index has shaped up to be the largest single financial product in India.

    This data contains all the indices of NSE. NIFTY 50, NIFTY BANK, NIFTY 100, NIFTY COMMODITIES, NIFTY CONSUMPTION, NIFTY FIN SERVICE, NIFTY IT, NIFTY INFRA, NIFTY ENERGY, NIFTY FMCG, NIFTY AUTO, NIFTY 200, NIFTY ALPHA 50, NIFTY 500, NIFTY CPSE, NIFTY GS COMPSITE, NIFTY HEALTHCARE, NIFTY CONSR DURBL, NIFTY LARGEMID250, NIFTY INDIA MFG, NIFTY IND DIGITAL, INDIA VIX

    File Information and Column Descriptions.

    Nifty 50 index data with 1 minute data. The dataset contains OHLC (Open, High, Low, and Close) prices from Jan 2015 to Aug 2024. - This dataset can be used for time series analysis, regression problems, and time series forecasting both for one step and multi-step ahead in the future. - Options data can be integrated with this minute data, to get more insight about this data. - Different backtesting strategies can be built on this data.

    File Information

    • This dataset contains 6 files, each file contains nifty 50 data with different intervals.
    • Different intervals are - 1 min, 3 min, 5 min, 15 min, and 1 hour, Daily data from intervals of 2015 Jan to 2024 August.

    Column Descriptors

    • Each file contains OHLC (Open, High, Low, and Close) prices and Data time information. Since these are Nifty 50 index data, so volume is not present.

    Inspiration

    Time series forecasting - Predict stock price

    • Predict future stock price one step ahead and multi-step ahead in time.
    • Use different time series forecasting techniques for forecasting the future stock price. ### Machine learning and Deep learning techniques
    • Possible ML and DL models include Neural networks, RNNs, LSTMs, Transformers, Attention networks, etc.
    • Different error functions can be considered like RMSE, MAE, RMSEP etc. ### Feature engineering
    • Different augmented features can be created and that can be used for forecasting.
    • Correlation analysis, Feature importance to justify the important features.
  17. Amazon Financial Dataset

    • kaggle.com
    zip
    Updated Dec 18, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Krishna Yadu (2024). Amazon Financial Dataset [Dataset]. https://www.kaggle.com/datasets/krishnayadav456wrsty/amazon-financial-dataset
    Explore at:
    zip(7415 bytes)Available download formats
    Dataset updated
    Dec 18, 2024
    Authors
    Krishna Yadu
    License

    Apache License, v2.0https://www.apache.org/licenses/LICENSE-2.0
    License information was derived automatically

    Description

    Title:

    Amazon Financial Dataset: R&D, Marketing, Campaigns, and Profit

    Description:

    This dataset provides fictional yet insightful financial data of Amazon's business activities across all 50 states of the USA. It is specifically designed to help students, researchers, and practitioners perform various data analysis tasks such as log normalization, Gaussian distribution visualization, and financial performance comparisons.

    Each row represents a state and contains the following columns:
    - R&D Amount (in $): The investment made in research and development.
    - Marketing Amount (in $): The expenditure on marketing activities.
    - Campaign Amount (in $): The costs associated with promotional campaigns.
    - State: The state in which the data is recorded.
    - Profit (in $): The net profit generated from the state.

    Additional features include log-normalized and Z-score transformations for advanced analysis.

    Use Cases:

    This dataset is ideal for practicing:
    1. Log Transformation: Normalize skewed data for better modeling and analysis.
    2. Statistical Analysis: Explore relationships between financial investments and profit.
    3. Visualization: Create compelling graphs such as Gaussian distributions and standard normal distributions.
    4. Machine Learning Projects: Build regression models to predict profits based on R&D and marketing spend.

    File Information:

    • File Format: Excel (.xlsx)
    • Number of Records: 50 (one for each state of the USA)
    • Columns: 5 primary financial columns and additional preprocessed columns for normalization and Z-scores.

    Important Note:

    This dataset is synthetically generated and is not based on actual Amazon financial records. It is created solely for educational and practice purposes.

    Tags:

    • Financial Analysis
    • Data Visualization
    • Machine Learning
    • Statistical Analysis
    • Educational Dataset
  18. Bird Migration Dataset (Data Visualization / EDA)

    • kaggle.com
    zip
    Updated May 13, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Sahir Maharaj (2025). Bird Migration Dataset (Data Visualization / EDA) [Dataset]. https://www.kaggle.com/datasets/sahirmaharajj/bird-migration-dataset-data-visualization-eda
    Explore at:
    zip(3249826 bytes)Available download formats
    Dataset updated
    May 13, 2025
    Authors
    Sahir Maharaj
    License

    Apache License, v2.0https://www.apache.org/licenses/LICENSE-2.0
    License information was derived automatically

    Description

    This dataset contains 10,000 synthetic records simulating the migratory behavior of various bird species across global regions. Each entry represents a single bird tagged with a tracking device and includes detailed information such as flight distance, speed, altitude, weather conditions, tagging information, and migration outcomes.

    The data was entirely synthetically generated using randomized yet realistic values based on known ranges from ornithological studies. It is ideal for practicing data analysis and visualization techniques without privacy concerns or real-world data access restrictions. Because it’s artificial, the dataset can be freely used in education, portfolio projects, demo dashboards, machine learning pipelines, or business intelligence training.

    With over 40 columns, this dataset supports a wide array of analysis types. Analysts can explore questions like “Do certain species migrate in larger flocks?”, “How does weather impact nesting success?”, or “What conditions lead to migration interruptions?”. Users can also perform geospatial mapping of start and end locations, cluster birds by behavior, or build time series models based on migration months and environmental factors.

    For data visualization, tools like Power BI, Python (Matplotlib/Seaborn/Plotly), or Excel can be used to create insightful dashboards and interactive charts.

    Join the Fabric Community DataViz Contest | May 2025: https://community.fabric.microsoft.com/t5/Power-BI-Community-Blog/%EF%B8%8F-Fabric-Community-DataViz-Contest-May-2025/ba-p/4668560

  19. Not seeing a result you expected?
    Learn how you can add new datasets to our index.

Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
Dr Corynen (2018). Graph Input Data Example.xlsx [Dataset]. http://doi.org/10.6084/m9.figshare.7506209.v1
Organization logoOrganization logo

Graph Input Data Example.xlsx

Explore at:
xlsxAvailable download formats
Dataset updated
Dec 26, 2018
Dataset provided by
figshare
Figsharehttp://figshare.com/
Authors
Dr Corynen
License

Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically

Description

The various performance criteria applied in this analysis include the probability of reaching the ultimate target, the costs, elapsed times and system vulnerability resulting from any intrusion. This Excel file contains all the logical, probabilistic and statistical data entered by a user, and required for the evaluation of the criteria. It also reports the results of all the computations.

Search
Clear search
Close search
Google apps
Main menu