Facebook
TwitterThe USDA Agricultural Research Service (ARS) recently established SCINet , which consists of a shared high performance computing resource, Ceres, and the dedicated high-speed Internet2 network used to access Ceres. Current and potential SCINet users are using and generating very large datasets so SCINet needs to be provisioned with adequate data storage for their active computing. It is not designed to hold data beyond active research phases. At the same time, the National Agricultural Library has been developing the Ag Data Commons, a research data catalog and repository designed for public data release and professional data curation. Ag Data Commons needs to anticipate the size and nature of data it will be tasked with handling. The ARS Web-enabled Databases Working Group, organized under the SCINet initiative, conducted a study to establish baseline data storage needs and practices, and to make projections that could inform future infrastructure design, purchases, and policies. The SCINet Web-enabled Databases Working Group helped develop the survey which is the basis for an internal report. While the report was for internal use, the survey and resulting data may be generally useful and are being released publicly. From October 24 to November 8, 2016 we administered a 17-question survey (Appendix A) by emailing a Survey Monkey link to all ARS Research Leaders, intending to cover data storage needs of all 1,675 SY (Category 1 and Category 4) scientists. We designed the survey to accommodate either individual researcher responses or group responses. Research Leaders could decide, based on their unit's practices or their management preferences, whether to delegate response to a data management expert in their unit, to all members of their unit, or to themselves collate responses from their unit before reporting in the survey. Larger storage ranges cover vastly different amounts of data so the implications here could be significant depending on whether the true amount is at the lower or higher end of the range. Therefore, we requested more detail from "Big Data users," those 47 respondents who indicated they had more than 10 to 100 TB or over 100 TB total current data (Q5). All other respondents are called "Small Data users." Because not all of these follow-up requests were successful, we used actual follow-up responses to estimate likely responses for those who did not respond. We defined active data as data that would be used within the next six months. All other data would be considered inactive, or archival. To calculate per person storage needs we used the high end of the reported range divided by 1 for an individual response, or by G, the number of individuals in a group response. For Big Data users we used the actual reported values or estimated likely values. Resources in this dataset:Resource Title: Appendix A: ARS data storage survey questions. File Name: Appendix A.pdfResource Description: The full list of questions asked with the possible responses. The survey was not administered using this PDF but the PDF was generated directly from the administered survey using the Print option under Design Survey. Asterisked questions were required. A list of Research Units and their associated codes was provided in a drop down not shown here. Resource Software Recommended: Adobe Acrobat,url: https://get.adobe.com/reader/ Resource Title: CSV of Responses from ARS Researcher Data Storage Survey. File Name: Machine-readable survey response data.csvResource Description: CSV file includes raw responses from the administered survey, as downloaded unfiltered from Survey Monkey, including incomplete responses. Also includes additional classification and calculations to support analysis. Individual email addresses and IP addresses have been removed. This information is that same data as in the Excel spreadsheet (also provided).Resource Title: Responses from ARS Researcher Data Storage Survey. File Name: Data Storage Survey Data for public release.xlsxResource Description: MS Excel worksheet that Includes raw responses from the administered survey, as downloaded unfiltered from Survey Monkey, including incomplete responses. Also includes additional classification and calculations to support analysis. Individual email addresses and IP addresses have been removed.Resource Software Recommended: Microsoft Excel,url: https://products.office.com/en-us/excel
Facebook
Twitterhttps://creativecommons.org/publicdomain/zero/1.0/https://creativecommons.org/publicdomain/zero/1.0/
PROJECT OBJECTIVE
We are a part of XYZ Co Pvt Ltd company who is in the business of organizing the sports events at international level. Countries nominate sportsmen from different departments and our team has been given the responsibility to systematize the membership roster and generate different reports as per business requirements.
Questions (KPIs)
TASK 1: STANDARDIZING THE DATASET
TASK 2: DATA FORMATING
TASK 3: SUMMARIZE DATA - PIVOT TABLE (Use SPORTSMEN worksheet after attempting TASK 1) • Create a PIVOT table in the worksheet ANALYSIS, starting at cell B3,with the following details:
TASK 4: SUMMARIZE DATA - EXCEL FUNCTIONS (Use SPORTSMEN worksheet after attempting TASK 1)
• Create a SUMMARY table in the worksheet ANALYSIS,starting at cell G4, with the following details:
TASK 5: GENERATE REPORT - PIVOT TABLE (Use SPORTSMEN worksheet after attempting TASK 1)
• Create a PIVOT table report in the worksheet REPORT, starting at cell A3, with the following information:
Process
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
File name definitions:
'...v_50_175_250_300...' - dataset for velocity ranges [50, 175] + [250, 300] m/s
'...v_175_250...' - dataset for velocity range [175, 250] m/s
'ANNdevelop...' - used to perform 9 parametric sub-analyses where, in each one, many ANNs are developed (trained, validated and tested) and the one yielding the best results is selected
'ANNtest...' - used to test the best ANN from each aforementioned parametric sub-analysis, aiming to find the best ANN model; this dataset includes the 'ANNdevelop...' counterpart
Where to find the input (independent) and target (dependent) variable values for each dataset/excel ?
input values in 'IN' sheet
target values in 'TARGET' sheet
Where to find the results from the best ANN model (for each target/output variable and each velocity range)?
open the corresponding excel file and the expected (target) vs ANN (output) results are written in 'TARGET vs OUTPUT' sheet
Check reference below (to be added when the paper is published)
https://www.researchgate.net/publication/328849817_11_Neural_Networks_-_Max_Disp_-_Railway_Beams
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
About Dataset Safa S. Abdul-Jabbar, Alaa k. Farhan
Context This is the first Dataset for various ordinary patients in Iraq. The Dataset provides the patients’ Cell Blood Count test information that can be used to create a Hematology diagnosis/prediction system. Also, this Data was collected in 2022 from Al-Zahraa Al-Ahly Hospital. These data can be cleaned & analyzed using any programming language because it is provided in an excel file that can be accessed and manipulated easily. The user just needs to understand how rows and columns are arranged because the data was collected as images(CBC images) from the laboratories and then stored the extracted data in an excel file. Content This Dataset contains 500 rows. For each row (patient information), there are 21 columns containing CBC test features that can be described as follows:
ID: Patients Identifier
WBC: White Blood Cell, Normal Ranges: 4.0 to 10.0, Unit: 10^9/L.
LYMp: Lymphocytes percentage, which is a type of white blood cell, Normal Ranges: 20.0 to 40.0, Unit: %
MIDp: Indicates the percentage combined value of the other types of white blood cells not classified as lymphocytes or granulocytes, Normal Ranges: 1.0 to 15.0, Unit: %
NEUTp: Neutrophils are a type of white blood cell (leukocytes); neutrophils percentage, Normal Ranges: 50.0 to 70.0, Unit: %
LYMn: Lymphocytes number are a type of white blood cell, Normal Ranges: 0.6 to 4.1, Unit: 10^9/L.
MIDn: Indicates the combined number of other white blood cells not classified as lymphocytes or granulocytes, Normal Ranges: 0.1 to 1.8, Unit: 10^9/L.
NEUTn: Neutrophils Number, Normal Ranges: 2.0 to 7.8, Unit: 10^9/L.
RBC: Red Blood Cell, Normal Ranges: 3.50 to 5.50, Unit: 10^12/L
HGB: Hemoglobin, Normal Ranges: 11.0 to 16.0, Unit: g/dL
HCT: Hematocrit is the proportion, by volume, of the Blood that consists of red blood cells, Normal Ranges: 36.0 to 48.0, Unit: %
MCV: Mean Corpuscular Volume, Normal Ranges: 80.0 to 99.0, Unit: fL
MCH: Mean Corpuscular Hemoglobin is the average amount of haemoglobin in the average red cell, Normal Ranges: 26.0 to 32.0, Unit: pg
MCHC: Mean Corpuscular Hemoglobin Concentration, Normal Ranges: 32.0 to 36.0, Unit: g/dL
RDWSD: Red Blood Cell Distribution Width, Normal Ranges: 37.0 to 54.0, Unit: fL
RDWCV: Red blood cell distribution width, Normal Ranges: 11.5 to 14.5, Unit: %
PLT: Platelet Count, Normal Ranges: 100 to 400, Unit: 10^9/L
MPV: Mean Platelet Volume, Normal Ranges: 7.4 to 10.4, Unit: fL
PDW: Red Cell Distribution Width, Normal Ranges: 10.0 to 17.0, Unit: %
PCT: The level of Procalcitonin in the Blood, Normal Ranges: 0.10 to 0.28, Unit: %
PLCR: Platelet Large Cell Ratio, Normal Ranges: 13.0 to 43.0, Unit: %
Acknowledgements We thank the entire Al-Zahraa Al-Ahly Hospital Hospital team, especially the hospital manager, for cooperating with us in collecting this data while maintaining patients' confidentiality.
Facebook
TwitterRaw data from six species of raptors that evaluate the percentage of GPS fixes and the percentage of the home range area, or 95% kernel area, captured by the buffer circle. In each file, the first worksheet has details on the length of each year and season per each individual; the second worksheet counts the individual and average home range and the percentage of GPS fixes captured by the calculated buffer circle during each year and season; the third worksheet calculates the individual and average percentage of the area captured by the calculated buffer circle; while the fourth worksheet calculates the individual year and average annual percentage of the home range covered by the calculated buffer circle, and at what size does the buffer circle capture 95% of the species' home range.
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
The open repository consists of two folders; Dataset and Picture. The dataset folder consists file “AWS Dataset Pangandaraan.xlsx”. There are 10 columns with three first columns as time attributes and the other six as atmosphere datasets. Each parameter has 8085 data, and Each parameter has a parameter index at the bottom of the column we added, including mMinimum, mMaximum, and Average values.
For further use, the user can choose one or more parameters for calculating or analyzing. For example, wind data (speed and direction) can be utilized to calculate Waves using the Hindcast method. Furthermore, the user can filter data by using the feature in Excel to extract the exact time range for analyzing various phenomena considered correlated to atmosphere data around Pangandaran, Indonesia.
The second folder, named “Picture,” contains three figures, including the monthly distribution of datasets, temporal data, and wind rose. Furthermore, the user can filter data by using the feature in Excel sheet to extract the exact time range for analyzing various phenomena considered correlated to atmosphere data around Pangandaran, Indonesia
Facebook
TwitterExcel Age-Range creator for Office for National Statistics (ONS) Mid year population estimates (MYE) covering each year between 1999 and 2016 These files take into account the revised estimates for 2002-2010 released in April 2013 down to Local Authority level and the post 2011 estimates based on the Census results. Scotland and Northern Ireland data has not been revised, so Great Britain and United Kingdom totals comprise the original data for these plus revised England and Wales figures. This Excel based tool enables users to query the single year of age raw data so that any age range can easily be calculated without having to carry out often complex, and time consuming formulas that could also be open to human error. Simply select the lower and upper age range for both males and females and the spreadsheet will return the total population for the range. Please adhere to the terms and conditions of supply contained within the file. Tip: You can copy and paste the rows you are interested in to another worksheet by using the filters at the top of the columns and then select all by pressing Ctrl+A. Then simply copy and paste the cells to a new location. ONS Mid year population estimates Open Excel tool (London Boroughs, Regions and National, 1999-2016) Also available is a custom-age tool for all geographies in the UK. This full MYE dataset by single year of age (SYA) age and gender is available as a Datastore package here. Ward Level Population estimates Single year of age population tool for 2002 to 2015 for all wards in London. New 2014 Ward boundary estimates Ward boundary changes in May 2014 only affected three London boroughs - Hackney, Kensington and Chelsea, and Tower Hamlets. The estimates between 2001-2013 have been calculated by the GLA by taking the proportion of a the old ward that falls within the new ward based on the proportion of population living in each area at the 2011 Census. Therefore, these estimates are purely indicative and are not official statistics and not endorsed by ONS. From 2014 onwards, ONS began publishing official estimates for the new ward boundaries. Download here.
Facebook
TwitterAttribution-NonCommercial 4.0 (CC BY-NC 4.0)https://creativecommons.org/licenses/by-nc/4.0/
License information was derived automatically
The van der Waals volume is a widely used descriptor in modeling physicochemical properties. However, the calculation of the van der Waals volume (VvdW) is rather time-consuming, from Bondi group contributions, for a large data set. A new method for calculating van der Waals volume has been developed, based on Bondi radii. The method, termed Atomic and Bond Contributions of van der Waals volume (VABC), is very simple and fast. The only information needed for calculating VABC is atomic contributions and the number of atoms, bonds, and rings. Then, the van der Waals volume (Å3/molecule) can be calculated from the following formula: VvdW = ∑ all atom contributions − 5.92NB − 14.7RA − 3.8RNR (NB is the number of bonds, RA is the number of aromatic rings, and RNA is the number of nonaromatic rings). The number of bonds present (NB) can be simply calculated by NB = N − 1 + RA + RNA (where N is the total number of atoms). A simple Excel spread sheet has been made to calculate van der Waals volumes for a wide range of 677 organic compounds, including 237 drug compounds. The results show that the van der Waals volumes calculated from VABC are equivalent to the computer-calculated van der Waals volumes for organic compounds.
Facebook
TwitterUp to 10 individuals of a fin fish, shark, and crustacean species are collected and morphologically identified by Southeast Fisheries Science Center. Water-soluble sarcoplasmic proteins are extracted from the tissue; proteins are analyzed by microfluidic electrophoresis (Agilent Bioanalyzer 2100) to generate species-specific protein patterns. The protein patterns are entered into a database to determine which are the most abundant for each species. The proteins that are seen in all 10 individuals are transferred to the excel pattern matching library database that uses NSIL designed formulas to calculate the range of each protein (+/- 1.5 to 3%) and compare an unknown protein to pattern to those in the pattern matching library.
Facebook
Twitterhttp://reference.data.gov.uk/id/open-government-licencehttp://reference.data.gov.uk/id/open-government-licence
Excel Age-Range creator for Office for National Statistics (ONS) Mid year population estimates (MYE) covering each year between 1999 and 2014
https://londondatastore-upload.s3.amazonaws.com/mye-custom-tool.JPG" alt="" />
These files take into account the revised estimates for 2002-2010 released in April 2013 down to Local Authority level and the post 2011 estimates based on the Census results. Scotland and Northern Ireland data has not been revised, so Great Britain and United Kingdom totals comprise the original data for these plus revised England and Wales figures.
This Excel based tool enables users to query the single year of age raw data so that any age range can easily be calculated without having to carry out often complex, and time consuming formulas that could also be open to human error. Simply select the lower and upper age range for both males and females and the spreadsheet will return the total population for the range. Please adhere to the terms and conditions of supply contained within the file.
Tip: You can copy and paste the rows you are interested in to another worksheet by using the filters at the top of the columns and then select all by pressing Ctrl+A. Then simply copy and paste the cells to a new location.
ONS Mid year population estimates
Open Excel tool (London Boroughs, Regions and National, 1999-2014)
Also available is a custom-age tool for all geographies in the UK. Open the tool for all UK geographies (local authority and above) for: 2010, 2011, 2012, 2013, and 2014.
This full MYE dataset by single year of age (SYA) age and gender is available as a Datastore package here.
Ward Level Population estimates
Excel single year of age population tool for 2002 to 2013 for all wards in London.
New 2014 Ward boundary estimates
This data is only for wards in the three London boroughs that changed their ward boundaries in May 2014. The estimates in this spreadsheet have been calculated by the GLA by taking the proportion of a the old ward that falls within the new ward based on the proportion of population living in each area at the 2011 Census. Therefore, these estimates are purely indicative and are not official statistics and not endorsed by ONS.
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
A dataset of crystallographic texture results for both α (hexagonal close packed, hcp) and β (body-centred cubic, bcc) phases, measured from 31 different hot-rolled Ti-6Al-4V (Ti-64) materials and 3 differently orientated samples using synchrotron X-ray diffraction (SXRD). The aim of the work was to accurately quantify bulk macro-texture for both the α and β phases across a range of different processing conditions, and to compare results with electron backscatter diffraction (EBSD) measurements. The synchrotron intensities were extracted using a new Fourier-based peak fitting method from the Continuous-Peak-Fit Python package, and then directly used to calculate the pole figures, orientation distribution functions (ODFs) and numerical values for the texture indices in MTEX
Material
The Ti-64 materials had been hot-rolled at a range of different temperatures, and to different reductions, followed by air-cooling. Three samples of different orientation were cut from the centre of these rolled blocks, and from the starting material. The material and hot-rolling conditions are recorded in this analysis dataset as an excel spreadsheet and summarised in the table below.
A table recording the sample number and associated hot-rolling condition.
Sample Number
Rolling Condition
1
825ºC, 87.5% Reduction
2
865ºC, 87.5% Reduction
3
895ºC, 87.5% Reduction
4
915ºC, 87.5% Reduction
5
935ºC, 87.5% Reduction
6
950ºC, 87.5% Reduction
7
960ºC, 87.5% Reduction
8
975ºC, 87.5% Reduction
9
1020ºC, 87.5% Reduction
10
β-annealed, 825ºC, 87.5% Reduction
11
β-annealed, 915ºC, 87.5% Reduction
12
β-annealed, 975ºC, 87.5% Reduction
13
Reduced heating from 915ºC, 87.5% Reduction
14
Reduced heating from 975ºC, 87.5% Reduction
15
825ºC, 75% Reduction
16
865ºC, 75% Reduction
17
895ºC, 75% Reduction
18
915ºC, 75% Reduction
19
935ºC, 75% Reduction
20
950ºC, 75% Reduction
21
960ºC, 75% Reduction
22
975ºC, 75% Reduction
23
1020ºC, 75% Reduction
24
β-annealed, 825ºC, 75% Reduction
25
β-annealed, 915ºC, 75% Reduction
26
β-annealed, 975ºC, 75% Reduction
27
Reduced heating from 915ºC, 75% Reduction
28
Reduced heating from 975ºC, 75% Reduction
29
As-received
30
As-received, β-annealed
31
975ºC, 50% Reduction
MTEX Data Analysis
The lattice plane intensities for 22 α and 4 β phase peaks were extracted from the Continuous-Peak-Fit analysis, also included in this analysis dataset, and saved as text files in the form of pole figures. The lattice intensity text files were analysed in MTEX using scripts from the continuous-peak-fit-analysis package, to plot pole figures and ODF slices, and to calculate pole figure maxima, ODF maxima, texture indices and texture component phase fractions. A kernel half-width of 10° was found to produce optimal data fitting, for highly accurate texture strength intensity values.
Metadata
An accompanying YAML text file contains associated processing metadata for the SXRD analysis, recording information about the packages used to process the data, along with details about the different files contained within this results dataset.
Not seeing a result you expected?
Learn how you can add new datasets to our index.
Facebook
TwitterThe USDA Agricultural Research Service (ARS) recently established SCINet , which consists of a shared high performance computing resource, Ceres, and the dedicated high-speed Internet2 network used to access Ceres. Current and potential SCINet users are using and generating very large datasets so SCINet needs to be provisioned with adequate data storage for their active computing. It is not designed to hold data beyond active research phases. At the same time, the National Agricultural Library has been developing the Ag Data Commons, a research data catalog and repository designed for public data release and professional data curation. Ag Data Commons needs to anticipate the size and nature of data it will be tasked with handling. The ARS Web-enabled Databases Working Group, organized under the SCINet initiative, conducted a study to establish baseline data storage needs and practices, and to make projections that could inform future infrastructure design, purchases, and policies. The SCINet Web-enabled Databases Working Group helped develop the survey which is the basis for an internal report. While the report was for internal use, the survey and resulting data may be generally useful and are being released publicly. From October 24 to November 8, 2016 we administered a 17-question survey (Appendix A) by emailing a Survey Monkey link to all ARS Research Leaders, intending to cover data storage needs of all 1,675 SY (Category 1 and Category 4) scientists. We designed the survey to accommodate either individual researcher responses or group responses. Research Leaders could decide, based on their unit's practices or their management preferences, whether to delegate response to a data management expert in their unit, to all members of their unit, or to themselves collate responses from their unit before reporting in the survey. Larger storage ranges cover vastly different amounts of data so the implications here could be significant depending on whether the true amount is at the lower or higher end of the range. Therefore, we requested more detail from "Big Data users," those 47 respondents who indicated they had more than 10 to 100 TB or over 100 TB total current data (Q5). All other respondents are called "Small Data users." Because not all of these follow-up requests were successful, we used actual follow-up responses to estimate likely responses for those who did not respond. We defined active data as data that would be used within the next six months. All other data would be considered inactive, or archival. To calculate per person storage needs we used the high end of the reported range divided by 1 for an individual response, or by G, the number of individuals in a group response. For Big Data users we used the actual reported values or estimated likely values. Resources in this dataset:Resource Title: Appendix A: ARS data storage survey questions. File Name: Appendix A.pdfResource Description: The full list of questions asked with the possible responses. The survey was not administered using this PDF but the PDF was generated directly from the administered survey using the Print option under Design Survey. Asterisked questions were required. A list of Research Units and their associated codes was provided in a drop down not shown here. Resource Software Recommended: Adobe Acrobat,url: https://get.adobe.com/reader/ Resource Title: CSV of Responses from ARS Researcher Data Storage Survey. File Name: Machine-readable survey response data.csvResource Description: CSV file includes raw responses from the administered survey, as downloaded unfiltered from Survey Monkey, including incomplete responses. Also includes additional classification and calculations to support analysis. Individual email addresses and IP addresses have been removed. This information is that same data as in the Excel spreadsheet (also provided).Resource Title: Responses from ARS Researcher Data Storage Survey. File Name: Data Storage Survey Data for public release.xlsxResource Description: MS Excel worksheet that Includes raw responses from the administered survey, as downloaded unfiltered from Survey Monkey, including incomplete responses. Also includes additional classification and calculations to support analysis. Individual email addresses and IP addresses have been removed.Resource Software Recommended: Microsoft Excel,url: https://products.office.com/en-us/excel