Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
The various performance criteria applied in this analysis include the probability of reaching the ultimate target, the costs, elapsed times and system vulnerability resulting from any intrusion. This Excel file contains all the logical, probabilistic and statistical data entered by a user, and required for the evaluation of the criteria. It also reports the results of all the computations.
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
This dataset contains sales transaction records used to create an interactive Excel Sales Performance Dashboard for business analytics practice.
It includes six columns capturing essential sales metrics such as date, region, product, quantity, sales revenue, and profit. The data is structured to help analysts and learners explore data visualization, PivotTable summarization, and dashboard design concepts in Excel.
The dataset was created for educational and demonstration purposes to help users:
Columns: Date – Transaction date (daily sales record) Region – Geographic area of the sale (East, West, North, South) Product – Product category or item sold Sales – Total revenue generated from the sale (USD) Profit – Net profit made per transaction Quantity – Number of units sold
Typical uses include: Excel or Power BI dashboard projects PivotTable practice for business reporting Data cleaning and chart-building exercises Portfolio development for business analytics students Built and tested in Microsoft Excel using PivotTables, Charts, and Conditional Formatting.
Facebook
Twitterhttps://creativecommons.org/publicdomain/zero/1.0/https://creativecommons.org/publicdomain/zero/1.0/
This dataset provides a dynamic Excel model for prioritizing projects based on Feasibility, Impact, and Size.
It visualizes project data on a Bubble Chart that updates automatically when new projects are added.
Use this tool to make data-driven prioritization decisions by identifying which projects are most feasible and high-impact.
Organizations often struggle to compare multiple initiatives objectively.
This matrix helps teams quickly determine which projects to pursue first by visualizing:
Example (partial data):
| Criteria | Project 1 | Project 2 | Project 3 | Project 4 | Project 5 | Project 6 | Project 7 | Project 8 |
|---|---|---|---|---|---|---|---|---|
| Feasibility | 7 | 9 | 5 | 2 | 7 | 2 | 6 | 8 |
| Impact | 8 | 4 | 4 | 6 | 6 | 7 | 7 | 7 |
| Size | 10 | 2 | 3 | 7 | 4 | 4 | 3 | 1 |
| Quadrant | Description | Action |
|---|---|---|
| High Feasibility / High Impact | Quick wins | Top Priority |
| High Impact / Low Feasibility | Valuable but risky | Plan carefully |
| Low Impact / High Feasibility | Easy but minor value | Optional |
| Low Impact / Low Feasibility | Low return | Defer or drop |
Project_Priority_Matrix.xlsx. You can use this for:
- Portfolio management
- Product or feature prioritization
- Strategy planning workshops
Project_Priority_Matrix.xlsxFree for personal and organizational use.
Attribution is appreciated if you share or adapt this file.
Author: [Asjad]
Contact: [m.asjad2000@gmail.com]
Compatible With: Microsoft Excel 2019+ / Office 365
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
This data publication is part of the 'P³-Petrophysical Property Database' project, which was developed within the EC funded project IMAGE (Integrated Methods for Advanced Geothermal Exploration, EU grant agreement No. 608553) and consists of a scientific paper, a full report on the database, the database as excel and .csv files and additional tables for a hierarchical classification of the petrography and stratigraphy of the investigated rock samples (see related references). This publication here provides a hierarchical interlinked stratigraphic classification according to the chronostratigraphical units of the international chronostratigraphic chart of the IUGS v2016/04 (Cohen et al. 2013, updated) according to international standardisation. As addition to this IUGS chart, which is also documented in GeoSciML, stratigraphic IDs and parent IDs were included to define the direct relationships between the stratigraphic terms. The P³ database aims at providing easily accessible, peer-reviewed information on physical rock properties relevant for geothermal exploration and reservoir characterization in one single compilation. Collected data include hydraulic, thermophysical and mechanical properties and, in addition, electrical resistivity and magnetic susceptibility. Each measured value is complemented by relevant meta-information such as the corresponding sample location, petrographic description, chronostratigraphic age and, most important, original citation. The original stratigraphic and petrographic descriptions are transferred to standardized catalogues following a hierarchical structure ensuring intercomparability for statistical analysis, of which the stratigraphic catalogue is presented here. These chronostratigraphic units are compiled to ensure that formations of a certain age are connected to the corresponding stratigraphic epoch, period or erathem. Thus, the chronostratigraphic units are directly correlated to each other by their stratigraphic ID and stratigraphic parent ID and can thus be used for interlinked data assessment of the petrophysical properties of samples of an according stratigraphic unit.
Facebook
TwitterThis interactive sales dashboard is designed in Excel for B2C type of Businesses like Dmart, Walmart, Amazon, Shops & Supermarkets, etc. using Slicers, Pivot Tables & Pivot Chart.
The first column is the date of Selling. The second column is the product ID. The third column is quantity. The fourth column is sales types, like direct selling, are purchased by a wholesaler or ordered online. The fifth column is a mode of payment, which is online or in cash. You can update these two as per requirements. The last one is a discount percentage. if you want to offer any discount, you can add it here.
So, basically these are the four sheets mentioned above with different tasks.
However, a sales dashboard enables organizations to visualize their real-time sales data and boost productivity.
A dashboard is a very useful tool that brings together all the data in the forms of charts, graphs, statistics and many more visualizations which lead to data-driven and decision making.
Questions & Answers
Facebook
TwitterContains EXCEL data for the charts in the article.
Facebook
TwitterThe Ontario government, generates and maintains thousands of datasets. Since 2012, we have shared data with Ontarians via a data catalogue. Open data is data that is shared with the public. Click here to learn more about open data and why Ontario releases it. Ontario’s Open Data Directive states that all data must be open, unless there is good reason for it to remain confidential. Ontario’s Chief Digital and Data Officer also has the authority to make certain datasets available publicly. Datasets listed in the catalogue that are not open will have one of the following labels: If you want to use data you find in the catalogue, that data must have a licence – a set of rules that describes how you can use it. A licence: Most of the data available in the catalogue is released under Ontario’s Open Government Licence. However, each dataset may be shared with the public under other kinds of licences or no licence at all. If a dataset doesn’t have a licence, you don’t have the right to use the data. If you have questions about how you can use a specific dataset, please contact us. The Ontario Data Catalogue endeavors to publish open data in a machine readable format. For machine readable datasets, you can simply retrieve the file you need using the file URL. The Ontario Data Catalogue is built on CKAN, which means the catalogue has the following features you can use when building applications. APIs (Application programming interfaces) let software applications communicate directly with each other. If you are using the catalogue in a software application, you might want to extract data from the catalogue through the catalogue API. Note: All Datastore API requests to the Ontario Data Catalogue must be made server-side. The catalogue's collection of dataset metadata (and dataset files) is searchable through the CKAN API. The Ontario Data Catalogue has more than just CKAN's documented search fields. You can also search these custom fields. You can also use the CKAN API to retrieve metadata about a particular dataset and check for updated files. Read the complete documentation for CKAN's API. Some of the open data in the Ontario Data Catalogue is available through the Datastore API. You can also search and access the machine-readable open data that is available in the catalogue. How to use the API feature: Read the complete documentation for CKAN's Datastore API. The Ontario Data Catalogue contains a record for each dataset that the Government of Ontario possesses. Some of these datasets will be available to you as open data. Others will not be available to you. This is because the Government of Ontario is unable to share data that would break the law or put someone's safety at risk. You can search for a dataset with a word that might describe a dataset or topic. Use words like “taxes” or “hospital locations” to discover what datasets the catalogue contains. You can search for a dataset from 3 spots on the catalogue: the homepage, the dataset search page, or the menu bar available across the catalogue. On the dataset search page, you can also filter your search results. You can select filters on the left hand side of the page to limit your search for datasets with your favourite file format, datasets that are updated weekly, datasets released by a particular organization, or datasets that are released under a specific licence. Go to the dataset search page to see the filters that are available to make your search easier. You can also do a quick search by selecting one of the catalogue’s categories on the homepage. These categories can help you see the types of data we have on key topic areas. When you find the dataset you are looking for, click on it to go to the dataset record. Each dataset record will tell you whether the data is available, and, if so, tell you about the data available. An open dataset might contain several data files. These files might represent different periods of time, different sub-sets of the dataset, different regions, language translations, or other breakdowns. You can select a file and either download it or preview it. Make sure to read the licence agreement to make sure you have permission to use it the way you want. Read more about previewing data. A non-open dataset may be not available for many reasons. Read more about non-open data. Read more about restricted data. Data that is non-open may still be subject to freedom of information requests. The catalogue has tools that enable all users to visualize the data in the catalogue without leaving the catalogue – no additional software needed. Have a look at our walk-through of how to make a chart in the catalogue. Get automatic notifications when datasets are updated. You can choose to get notifications for individual datasets, an organization’s datasets or the full catalogue. You don’t have to provide and personal information – just subscribe to our feeds using any feed reader you like using the corresponding notification web addresses. Copy those addresses and paste them into your reader. Your feed reader will let you know when the catalogue has been updated. The catalogue provides open data in several file formats (e.g., spreadsheets, geospatial data, etc). Learn about each format and how you can access and use the data each file contains. A file that has a list of items and values separated by commas without formatting (e.g. colours, italics, etc.) or extra visual features. This format provides just the data that you would display in a table. XLSX (Excel) files may be converted to CSV so they can be opened in a text editor. How to access the data: Open with any spreadsheet software application (e.g., Open Office Calc, Microsoft Excel) or text editor. Note: This format is considered machine-readable, it can be easily processed and used by a computer. Files that have visual formatting (e.g. bolded headers and colour-coded rows) can be hard for machines to understand, these elements make a file more human-readable and less machine-readable. A file that provides information without formatted text or extra visual features that may not follow a pattern of separated values like a CSV. How to access the data: Open with any word processor or text editor available on your device (e.g., Microsoft Word, Notepad). A spreadsheet file that may also include charts, graphs, and formatting. How to access the data: Open with a spreadsheet software application that supports this format (e.g., Open Office Calc, Microsoft Excel). Data can be converted to a CSV for a non-proprietary format of the same data without formatted text or extra visual features. A shapefile provides geographic information that can be used to create a map or perform geospatial analysis based on location, points/lines and other data about the shape and features of the area. It includes required files (.shp, .shx, .dbt) and might include corresponding files (e.g., .prj). How to access the data: Open with a geographic information system (GIS) software program (e.g., QGIS). A package of files and folders. The package can contain any number of different file types. How to access the data: Open with an unzipping software application (e.g., WinZIP, 7Zip). Note: If a ZIP file contains .shp, .shx, and .dbt file types, it is an ArcGIS ZIP: a package of shapefiles which provide information to create maps or perform geospatial analysis that can be opened with ArcGIS (a geographic information system software program). A file that provides information related to a geographic area (e.g., phone number, address, average rainfall, number of owl sightings in 2011 etc.) and its geospatial location (i.e., points/lines). How to access the data: Open using a GIS software application to create a map or do geospatial analysis. It can also be opened with a text editor to view raw information. Note: This format is machine-readable, and it can be easily processed and used by a computer. Human-readable data (including visual formatting) is easy for users to read and understand. A text-based format for sharing data in a machine-readable way that can store data with more unconventional structures such as complex lists. How to access the data: Open with any text editor (e.g., Notepad) or access through a browser. Note: This format is machine-readable, and it can be easily processed and used by a computer. Human-readable data (including visual formatting) is easy for users to read and understand. A text-based format to store and organize data in a machine-readable way that can store data with more unconventional structures (not just data organized in tables). How to access the data: Open with any text editor (e.g., Notepad). Note: This format is machine-readable, and it can be easily processed and used by a computer. Human-readable data (including visual formatting) is easy for users to read and understand. A file that provides information related to an area (e.g., phone number, address, average rainfall, number of owl sightings in 2011 etc.) and its geospatial location (i.e., points/lines). How to access the data: Open with a geospatial software application that supports the KML format (e.g., Google Earth). Note: This format is machine-readable, and it can be easily processed and used by a computer. Human-readable data (including visual formatting) is easy for users to read and understand. This format contains files with data from tables used for statistical analysis and data visualization of Statistics Canada census data. How to access the data: Open with the Beyond 20/20 application. A database which links and combines data from different files or applications (including HTML, XML, Excel, etc.). The database file can be converted to a CSV/TXT to make the data machine-readable, but human-readable formatting will be lost. How to access the data: Open with Microsoft Office Access (a database management system used to develop application software). A file that keeps the original layout and
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Mortality surveillance aids in identifying and addressing causes of death allowing health systems to adapt and respond effectively. An assessment of mortality surveillance in Uganda was conducted from November 2023 to June 2024 through data reviews and plenary discussions engaging various stakeholders in Uganda. Eight (8) workshops/meetings were conducted over a period of eight months to generate information on mortality data sources, processes of data generation and challenges affecting the system. Responses from the meetings and workshops were recorded and transcribed. Data were thematically analysed and presented as descriptive narratives. Quantitative data from district health information system version. 2 (DHIS2) was analyzed using excel and presented using charts and tables. The rapid assessment of mortality surveillance in Uganda highlighted opportunities for improved mortality surveillance through the existence of various sources of data. It was highlighted that 66.9% of the death occur in communities, however, there is a major data completeness gaps where suboptimal data from the community is feed into the national health statistics database (DHIS2) to enable stakeholder analysis and utilization. Furthermore, a number of data quality issues were identified in the health facility generated data where 33% of the deaths occur. These include: data completeness where the national referral specialized health institutes do not feed their data into the national data base, late reporting and the lack of coordination and standardisation of reporting among the various partners. The existence of structures to conduct mortality surveillance in Uganda presents an opportunity for improved mortality surveillance despite the highlighted gaps and challenges. Adoption of strategies aimed to enable the successful implementation of an efficient mortality surveillance program like: strengthening governance and operations of death reporting activities, establishing a clear definition of institutional roles and responsibilities, raising awareness and advocacy at all levels, building technical capacities, improving allocation of resources, and leveraging on shared interests by both implementing and development partners could improve mortality surveillance and the health of the population through utilisation of the generated data.
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
This online appendix provides the materials of an experiment regarding the performance of individual subjects (24 students) in the extraction of conceptual models from a specification expressed:
[IV1] as user stories or use cases
[IV2] for one of three types of fictitious systems: hospital management system, urban traffic simulator, and international football association portal
We measure the performance in terms of validity (DV1) and completeness (DV2) against models that were created by domain experts (the three authors of the paper and of this online appendix) from each of the specifications.
The materials include
The description of the three systems (folder System Descriptions)
An Excel spreadsheet that includes raw data, charts, and statistical results
The guidelines that the authors used in assessing the quality of the subjects' models against the expert models
For each student who participated in the experiment and gave consent,
The specification, either via user stories or use cases, created by the student
The student model created by the student
The expert model created by one of the paper authors Note that these files are highlighted to denote how we applied the tagging guidelines
Facebook
Twitterhttps://fred.stlouisfed.org/legal/#copyright-pre-approvalhttps://fred.stlouisfed.org/legal/#copyright-pre-approval
Graph and download economic data for Dow Jones Industrial Average (DJIA) from 2015-12-02 to 2025-12-01 about stock market, average, industry, and USA.
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
All Comparisons of Differentially Expressed Genes - excel sheet containing the annotations and fold change values of the all the differentially expressed genes between the different clone comparisonsFinal List of Common Genes - excel sheet containing the list of genes that were commonly differentially expressed between all the aphid clone comparisons. Also contains table and bar chart presenting the number of times each candidate gene selected from previous literature was found in each aphid clone comparison.Non-direct and Direct Competition - excel sheet containing number of nymphs produced by all 6 clones on the 3 host plants in the non-direct competition, and the number of nymphs produced by the two clones NS and Viola in the direct competition experiment.sterror - excel sheet containing the means and standard error values of the 6 grouped resistant and susceptible clones in the non-direct competition experiment, used to make the bar plot for the non-direct competition experiment.sterror2 - excel sheet containing the means and standard error values of the resistant clone Viola and susceptible clone NS in the direct competition experiment, used to make the bar plot for the direct competition experiment.cabbagettest - excel sheet containing the number of nymphs produce by the 6 grouped resistant and susceptible clones on the 3 host plants, used to conduct the unpaired t tests to compare the reproductive performance of resistant and susceptible clones on the 3 different host plants when in not in competitiondirectcompetition - excel sheet containing the number of nymphs produce by the resistant clone Viola and susceptible clone NS on the 3 host plants, used to conduct the unpaired t tests comparing the reproductive performance of resistant and susceptible clones on the 3 different host plants when in direct competitionAPHID HOST SHIFT DISS Rscript - R script containing all my statistical tests: unpaired t tests of resistant and susceptible clones on the 3 host plants when in direct and non direct competition, and kruskal Wallis tests and post hoc Dunns test to identify significant differences between individual and resistant and susceptible clones on the different host plants. Also contains all my code for my bar charts for the non-direct and direct competition experiments and the code for my box plots showing the significant differences between individual clones and resistant and susceptible clones on the different host plants.Up and Down-regulated Genes Graph - excel sheet containing the number of and and down regulated genes in each aphid clone comparison and the bar graph generated from this data.
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Microsatellites, also known as SSRs or STRs, are polymorphic DNA regions with tandem repetitions of a nucleotide motif of size 1–6 base pairs with a broad range of applications in many fields, such as comparative genomics, molecular biology, and forensics. However, the majority of researchers do not have computational training and struggle while running command-line tools or very limited web tools for their SSR research, spending a considerable amount of time learning how to execute the software and conducting the post-processing data tabulation in other tools or manually—time that could be used directly in data analysis. We present EasySSR, a user-friendly web tool with command-line full functionality, designed for practical use in batch identifying and comparing SSRs in sequences, draft, or complete genomes, not requiring previous bioinformatic skills to run. EasySSR requires only a FASTA and an optional GENBANK file of one or more genomes to identify and compare STRs. The tool can automatically analyze and compare SSRs in whole genomes, convert GenBank to PTT files, identify perfect and imperfect SSRs and coding and non-coding regions, compare their frequencies, abundancy, motifs, flanking sequences, and iterations, producing many outputs ready for download such as PTT files, interactive charts, and Excel tables, giving the user the data ready for further analysis in minutes. EasySSR was implemented as a web application, which can be executed from any browser and is available for free at https://computationalbiology.ufpa.br/easyssr/. Tutorials, usage notes, and download links to the source code can be found at https://github.com/engbiopct/EasySSR.
Facebook
Twitterhttps://creativecommons.org/publicdomain/zero/1.0/https://creativecommons.org/publicdomain/zero/1.0/
📊 Road Accident Data Analysis: Interactive Excel Dashboard 🚗
Excited to share my Kaggle project focusing on road accident data analysis. Leveraging Excel's power, I've developed an interactive dashboard offering comprehensive insights for safer roads.
Key Aspects:
Data Processing & Cleaning: Ensured data reliability through meticulous processing. KPIs: Primarily focused on Total Casualties, with detailed breakdowns for Fatal, Serious, Slight, and by Car type. Visualizations: Engaging charts - Doughnuts, Line, Bar, and Pie - offering a holistic view of accident trends. Interactivity: User-friendly features include Urban/Rural and Year filters for dynamic exploration. Unique Insights:
Monthly Trends: Line chart for a nuanced comparison of current vs. previous year casualties. Road Type Breakdown: Bar chart to showcase casualties distributed across different road types. Geospatial Analysis: Doughnut charts detailing casualties by location and area. Call for Collaboration: Seeking Kaggle community input for refinement and optimization. Let's collectively contribute to making our roads safer through data-driven insights!
Looking forward to your feedback and contributions! 🚀🌐
Facebook
Twitterhttps://creativecommons.org/publicdomain/zero/1.0/https://creativecommons.org/publicdomain/zero/1.0/
UPDATED EVERY WEEK Last Update - 26th July 2025
Disclaimer!!! Data uploaded here are collected from the internet and some google drive. The sole purposes of uploading these data are to provide this Kaggle community with a good source of data for analysis and research. I don't own these datasets and am also not responsible for them legally by any means. I am not charging anything (either money or any favor) for this dataset. RESEARCH PURPOSE ONLY
This data contains all the indices of NSE.
NIFTY 50,
NIFTY BANK,
NIFTY 100,
NIFTY COMMODITIES,
NIFTY CONSUMPTION,
NIFTY FIN SERVICE,
NIFTY IT,
NIFTY INFRA,
NIFTY ENERGY,
NIFTY FMCG,
NIFTY AUTO,
NIFTY 200,
NIFTY ALPHA 50,
NIFTY 500,
NIFTY CPSE,
NIFTY GS COMPSITE,
NIFTY HEALTHCARE,
NIFTY CONSR DURBL,
NIFTY LARGEMID250,
NIFTY INDIA MFG,
NIFTY IND DIGITAL,
INDIA VIX
Nifty 50 index data with 1 minute data. The dataset contains OHLC (Open, High, Low, and Close) prices from Jan 2015 to Aug 2024. - This dataset can be used for time series analysis, regression problems, and time series forecasting both for one step and multi-step ahead in the future. - Options data can be integrated with this minute data, to get more insight about this data. - Different backtesting strategies can be built on this data.
Facebook
Twitterhttps://fred.stlouisfed.org/legal/#copyright-public-domainhttps://fred.stlouisfed.org/legal/#copyright-public-domain
Graph and download economic data for Canadian Dollars to U.S. Dollar Spot Exchange Rate (EXCAUS) from Jan 1971 to Nov 2025 about Canada, exchange rate, currency, rate, and USA.
Not seeing a result you expected?
Learn how you can add new datasets to our index.
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
The various performance criteria applied in this analysis include the probability of reaching the ultimate target, the costs, elapsed times and system vulnerability resulting from any intrusion. This Excel file contains all the logical, probabilistic and statistical data entered by a user, and required for the evaluation of the criteria. It also reports the results of all the computations.