Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
This section presents a discussion of the research data. The data was received as secondary data however, it was originally collected using the time study techniques. Data validation is a crucial step in the data analysis process to ensure that the data is accurate, complete, and reliable. Descriptive statistics was used to validate the data. The mean, mode, standard deviation, variance and range determined provides a summary of the data distribution and assists in identifying outliers or unusual patterns. The data presented in the dataset show the measures of central tendency which includes the mean, median and the mode. The mean signifies the average value of each of the factors presented in the tables. This is the balance point of the dataset, the typical value and behaviour of the dataset. The median is the middle value of the dataset for each of the factors presented. This is the point where the dataset is divided into two parts, half of the values lie below this value and the other half lie above this value. This is important for skewed distributions. The mode shows the most common value in the dataset. It was used to describe the most typical observation. These values are important as they describe the central value around which the data is distributed. The mean, mode and median give an indication of a skewed distribution as they are not similar nor are they close to one another. In the dataset, the results and discussion of the results is also presented. This section focuses on the customisation of the DMAIC (Define, Measure, Analyse, Improve, Control) framework to address the specific concerns outlined in the problem statement. To gain a comprehensive understanding of the current process, value stream mapping was employed, which is further enhanced by measuring the factors that contribute to inefficiencies. These factors are then analysed and ranked based on their impact, utilising factor analysis. To mitigate the impact of the most influential factor on project inefficiencies, a solution is proposed using the EOQ (Economic Order Quantity) model. The implementation of the 'CiteOps' software facilitates improved scheduling, monitoring, and task delegation in the construction project through digitalisation. Furthermore, project progress and efficiency are monitored remotely and in real time. In summary, the DMAIC framework was tailored to suit the requirements of the specific project, incorporating techniques from inventory management, project management, and statistics to effectively minimise inefficiencies within the construction project.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Context
The dataset presents median household incomes for various household sizes in South Range, MI, as reported by the U.S. Census Bureau. The dataset highlights the variation in median household income with the size of the family unit, offering valuable insights into economic trends and disparities within different household sizes, aiding in data analysis and decision-making.
Key observations
https://i.neilsberg.com/ch/south-range-mi-median-household-income-by-household-size.jpeg" alt="South Range, MI median household income, by household size (in 2022 inflation-adjusted dollars)">
When available, the data consists of estimates from the U.S. Census Bureau American Community Survey (ACS) 2017-2021 5-Year Estimates.
Household Sizes:
Variables / Data Columns
Good to know
Margin of Error
Data in the dataset are based on the estimates and are subject to sampling variability and thus a margin of error. Neilsberg Research recommends using caution when presening these estimates in your research.
Custom data
If you do need custom data for any of your research project, report or presentation, you can contact our research staff at research@neilsberg.com for a feasibility of a custom tabulation on a fee-for-service basis.
Neilsberg Research Team curates, analyze and publishes demographics and economic data from a variety of public and proprietary sources, each of which often includes multiple surveys and programs. The large majority of Neilsberg Research aggregated datasets and insights is made available for free download at https://www.neilsberg.com/research/.
This dataset is a part of the main dataset for South Range median household income. You can refer the same here
CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
License information was derived automatically
Mean-variance data collections for portfolio optimization problems based on time series of daily stock prices for New York stock exchange. These data collections can be used for investment portfolio optimization research.NYSE includes over 2000 stocks. These data includes randomly selected sets of size equal to 200, 300, 400, 500, 600, 700, 800 and 900. Each set include the succesive yearly data from 2014, 2015, 2016, 2017, 2018, and 2019 grouped in a single folder. While each folder includes the data saved in a text file following the format used by J. E. Beasley in OR Library (http://people.brunel.ac.uk/~mastjjb/jeb/orlib/portinfo.html).Example structure of files in the 200 set is presented below:- 2014 + JKMP2_200_2014_1 + JKMP2_200_2014_2 + JKMP2_200_2014_3 ... + JKMP2_200_2014_12- 2015 + JKMP2_200_2015_1 + JKMP2_200_2015_2 + JKMP2_200_2015_3 ... + JKMP2_200_2015_12- 2016 + JKMP2_200_2016_1 + JKMP2_200_2016_2 + JKMP2_200_2016_3 ... + JKMP2_200_2016_12- 2017 + JKMP2_200_2017_1 + JKMP2_200_2017_2 + JKMP2_200_2017_3 ... + JKMP2_200_2017_12- 2018 + JKMP2_200_2018_1 + JKMP2_200_2018_2 + JKMP2_200_2018_3 ... + JKMP2_200_2018_12- 2019 + JKMP2_200_2019_1 + JKMP2_200_2019_2 + JKMP2_200_2019_3 ... + JKMP2_200_2019_12A detailed description of the data collections can be found in the README file of the dataset. For each set of stocks, we estimated the correlation matrix and the vector of mean returns, based on the corresponding time series.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Social science commonly studies relationships among variables by employing survey questions. Answers to these questions will contain some degree of measurement error, distorting the relationships of interest. Such distortions can be removed by standard statistical methods, when these are provided knowledge of a question’s measurement error variance. However, acquiring this information routinely necessitates additional experimentation, which is infeasible in practice. We use three decades’ worth of survey experiments combined with machine learning methods to show that survey measurement error variance can be predicted from the way a question was asked. By predicting experimentally obtained estimates of survey measurement error variance from question characteristics, we enable researchers to obtain estimates of the extent of measurement error in a survey question without requiring additional data collection. Our results suggest only some commonly accepted best practices in survey design have a noticeable impact on study quality, and that predicting measurement error variance is a useful approach to removing this impact in future social surveys. This repository accompanies the full paper, and allows users to reproduce all results.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
This dataset is about book series. It has 1 row and is filtered where the books is Linkages between excess currency and stock market returns : Granger causality in mean and variance. It features 10 columns including number of authors, number of books, earliest publication date, and latest publication date.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
R Core Team. (2016). R: A language and environment for statistical computing. R Foundation for Statistical Computing.
Supplement to Occipital and left temporal instantaneous amplitude and frequency oscillations correlated with access and phenomenal consciousness (https://philpapers.org/rec/PEROAL-2).
Occipital and left temporal instantaneous amplitude and frequency oscillations correlated with access and phenomenal consciousness move from the features of the ERP characterized in Occipital and Left Temporal EEG Correlates of Phenomenal Consciousness (Pereira, 2015, https://doi.org/10.1016/b978-0-12-802508-6.00018-1, https://philpapers.org/rec/PEROAL) towards the instantaneous amplitude and frequency of event-related changes correlated with a contrast in access and in phenomenology.
Occipital and left temporal instantaneous amplitude and frequency oscillations correlated with access and phenomenal consciousness proceed as following.
In the first section, empirical mode decomposition (EMD) with post processing (Xie, G., Guo, Y., Tong, S., and Ma, L., 2014. Calculate excess mortality during heatwaves using Hilbert-Huang transform algorithm. BMC medical research methodology, 14, 35) Ensemble Empirical Mode Decomposition (postEEMD) and Hilbert-Huang Transform (HHT).
In the second section, calculated the variance inflation factor (VIF).
In the third section, partial least squares regression (PLSR): the minimal root mean squared error of prediction (RMSEP).
In the last section, partial least squares regression (PLSR): significance multivariate correlation (sMC) statistic.
This table (Variance Analysis Dataset - Yields N-level Environment) is part of a larger file dataset that contains processed data and information used in the meta-analysis “Yield development of German winter wheat between 1958 and 2015” of the Project “Data-Meta Analysis to assess the productivity development of cultivated plants” funded by the DFG. This table contains the final data used for the variance analysis in this project and derived from the entire dataset, which comprises the following data: - Winter wheat (Triticum aestivum) yields and nitrogen application amounts from nitrogen fertilization experiments of variable duration (1-6 years) carried out at 43 locations across Germany and between 1958 and 2015 found in 34 different sources in the literature. - The derived maximum yields (Ymax) and optimal nitrogen amounts (Nopt) from the nitrogen experiments, function coefficients, and statistics. - Geographical information (latitude, longitude, altitude) and other site specific information of the experimental sites (soil type, soil yield potential, mean annual temperature, mean annual precipitation, mean annual climatic water balance, soil climate region, cultivation region). - Processed phenological and climatic data for each experimental site.
This file contains the data set used to develop a random forest model predict background specific conductivity for stream segments in the contiguous United States. This Excel readable file contains 56 columns of parameters evaluated during development. The data dictionary provides the definition of the abbreviations and the measurement units. Each row is a unique sample described as R** which indicates the NHD Hydrologic Unit (underscore), up to a 7-digit COMID, (underscore) sequential sample month. To develop models that make stream-specific predictions across the contiguous United States, we used StreamCat data set and process (Hill et al. 2016; https://github.com/USEPA/StreamCat). The StreamCat data set is based on a network of stream segments from NHD+ (McKay et al. 2012). These stream segments drain an average area of 3.1 km2 and thus define the spatial grain size of this data set. The data set consists of minimally disturbed sites representing the natural variation in environmental conditions that occur in the contiguous 48 United States. More than 2.4 million SC observations were obtained from STORET (USEPA 2016b), state natural resource agencies, the U.S. Geological Survey (USGS) National Water Information System (NWIS) system (USGS 2016), and data used in Olson and Hawkins (2012) (Table S1). Data include observations made between 1 January 2001 and 31 December 2015 thus coincident with Moderate Resolution Imaging Spectroradiometer (MODIS) satellite data (https://modis.gsfc.nasa.gov/data/). Each observation was related to the nearest stream segment in the NHD+. Data were limited to one observation per stream segment per month. SC observations with ambiguous locations and repeat measurements along a stream segment in the same month were discarded. Using estimates of anthropogenic stress derived from the StreamCat database (Hill et al. 2016), segments were selected with minimal amounts of human activity (Stoddard et al. 2006) using criteria developed for each Level II Ecoregion (Omernik and Griffith 2014). Segments were considered as potentially minimally stressed where watersheds had 0 - 0.5% impervious surface, 0 – 5% urban, 0 – 10% agriculture, and population densities from 0.8 – 30 people/km2 (Table S3). Watersheds with observations with large residuals in initial models were identified and inspected for evidence of other human activities not represented in StreamCat (e.g., mining, logging, grazing, or oil/gas extraction). Observations were removed from disturbed watersheds, with a tidal influence or unusual geologic conditions such as hot springs. About 5% of SC observations in each National Rivers and Stream Assessment (NRSA) region were then randomly selected as independent validation data. The remaining observations became the large training data set for model calibration. This dataset is associated with the following publication: Olson, J., and S. Cormier. Modeling spatial and temporal variation in natural background specific conductivity. ENVIRONMENTAL SCIENCE & TECHNOLOGY. American Chemical Society, Washington, DC, USA, 53(8): 4316-4325, (2019).
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
In this work we present results of all the major global models and normalise the model results by looking at changes over time relative to a common base year value. We give an analysis of the variability across the models, both before and after normalisation in order to give insights into variance at national and regional level. A dataset of harmonised results (based on means) and measures of dispersion is presented, providing a baseline dataset for CBCA validation and analysis. The dataset is intended as a goto dataset for country and regional results of consumption and production based accounts. The normalised mean for each country/region is the principle result that can be used to assess the magnitude and trend in the emission accounts. However, an additional key element of the dataset are the measures of robustness and spread of the results across the source models. These metrics give insight into the amount of trust should be placed in the individual country/region results. Code at https://doi.org/10.5281/zenodo.3181930
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
A simulated SEND + EDX dataset along with the code used to produce it.
Added in newer version: the VAE processing of the SEND data has been included
The distribution of data records for the maximum horizontal stress orientation SHmax in the Earth’s crust is sparse and very unequally. In order to analyse the stress pattern and its wavelength or to predict the mean SHmax orientation on a regular grid, statistical interpolation as conducted e.g. by Coblentz and Richardson (1995), Müller et al. (2003), Heidbach and Höhne (2008), Heidbach et al. (2010) or Reiter et al. (2014) is necessary. Based on their work we wrote the Matlab® script Stress2Grid that provides several features to analyse the mean SHmax pattern. The script facilitates and speeds up this analysis and extends the functionality compared to aforementioned publications. The script is complemented by a number of example and input files as described in the WSM Technical Report (Ziegler and Heidbach, 2017, http://doi.org/10.2312/wsm.2017.002). The script provides two different concepts to calculate the mean SHmax orientation on a regular grid. The first is using a fixed search radius around the grid point and computes the mean SHmax orientation if sufficient data records are within the search radius. The larger the search radius the larger is the filtered wavelength of the stress pattern. The second approach is using variable search radii and determines the search radius for which the variance of the mean SHmax orientation is below a given threshold. This approach delivers mean SHmax orientations with a user-defined degree of reliability. It resolves local stress perturbations and is not available in areas with conflicting information that result in a large variance. Furthermore, the script can also estimate the deviation between plate motion direction and the mean SHmax orientation.
CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
License information was derived automatically
In scientific data analysis, clusters identified computationally often substantiate existing hypotheses or motivate new ones. Yet the combinatorial nature of the clustering result, which is a partition rather than a set of parameters or a function, blurs notions of mean and variance. This intrinsic difficulty hinders the development of methods to improve clustering by aggregation or to assess the uncertainty of clusters generated. We overcome that barrier by aligning clusters via optimal transport. Equipped with this technique, we propose a new algorithm to enhance clustering by any baseline method using bootstrap samples. Cluster alignment enables us to quantify variation in the clustering result at the levels of both overall partitions and individual clusters. Set relationships between clusters such as one-to-one match, split, and merge can be revealed. A covering point set for each cluster, a concept kin to the confidence interval, is proposed. The tools we have developed here will help address the crucial question of whether any cluster is an intrinsic or spurious pattern. Experimental results on both simulated and real datasets are provided.
CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
License information was derived automatically
The provided dataset contains results from Monte Carlo simulations related to variance swaps. The data is organized into multiple sheets, each focusing on different parameters and scenarios.Figure 1:Monte Carlo Simulations: This section presents the results of Monte Carlo simulations for both discretely-sampled and continuously-sampled variance swaps. The values are reported for different sample sizes (N=12 to N=322), showing how the estimated variance swap values converge as the number of samples increases.Sample 1 and Sample 2: These represent two different sets of simulation results, each showing the impact of varying sample sizes on the variance swap values.Figure 2:κθ (Kappa Theta): This section explores the impact of different values of κθ on the variance swap values. θ̃ (Theta Tilde): This part examines the effect of varying θ̃ on the variance swap values .σθ (Sigma Theta): This section analyzes the influence of σθ on the variance swap values .θ₀ (Theta Zero): This part investigates the impact of different initial volatility levels (θ₀) on the variance swap values .Sheet 3:λ (Lambda): This section studies the effect of varying λ on the variance swap values .η (Eta): This part examines the influence of η on the variance swap values .v (Nu): This section analyzes the impact of v on the variance swap values .δ (Delta): This part investigates the effect of varying δ on the variance swap values .Overall, the dataset provides a comprehensive analysis of how different parameters and sampling methods affect the valuation of variance swaps, offering insights into the sensitivity and convergence behavior of these financial instruments under various conditions.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Context
The dataset presents median household incomes for various household sizes in Will County, IL, as reported by the U.S. Census Bureau. The dataset highlights the variation in median household income with the size of the family unit, offering valuable insights into economic trends and disparities within different household sizes, aiding in data analysis and decision-making.
Key observations
https://i.neilsberg.com/ch/will-county-il-median-household-income-by-household-size.jpeg" alt="Will County, IL median household income, by household size (in 2022 inflation-adjusted dollars)">
When available, the data consists of estimates from the U.S. Census Bureau American Community Survey (ACS) 2017-2021 5-Year Estimates.
Household Sizes:
Variables / Data Columns
Good to know
Margin of Error
Data in the dataset are based on the estimates and are subject to sampling variability and thus a margin of error. Neilsberg Research recommends using caution when presening these estimates in your research.
Custom data
If you do need custom data for any of your research project, report or presentation, you can contact our research staff at research@neilsberg.com for a feasibility of a custom tabulation on a fee-for-service basis.
Neilsberg Research Team curates, analyze and publishes demographics and economic data from a variety of public and proprietary sources, each of which often includes multiple surveys and programs. The large majority of Neilsberg Research aggregated datasets and insights is made available for free download at https://www.neilsberg.com/research/.
This dataset is a part of the main dataset for Will County median household income. You can refer the same here
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
It has long been suspected that the rate of mutation varies across the human genome at a large scale based on the divergence between humans and other species. However, it is now possible to directly investigate this question using the large number of de novo mutations (DNMs) that have been discovered in humans through the sequencing of trios. We investigate a number of questions pertaining to the distribution of mutations using more than 130,000 DNMs from three large datasets. We demonstrate that the amount and pattern of variation differs between datasets at the 1MB and 100KB scales probably as a consequence of differences in sequencing technology and processing. In particular, datasets show different patterns of correlation to genomic variables such as replication time. Never-the-less there are many commonalities between datasets, which likely represent true patterns. We show that there is variation in the mutation rate at the 100KB, 1MB and 10MB scale that cannot be explained by variation at smaller scales, however the level of this variation is modest at large scales–at the 1MB scale we infer that ~90% of regions have a mutation rate within 50% of the mean. Different types of mutation show similar levels of variation and appear to vary in concert which suggests the pattern of mutation is relatively constant across the genome. We demonstrate that variation in the mutation rate does not generate large-scale variation in GC-content, and hence that mutation bias does not maintain the isochore structure of the human genome. We find that genomic features explain less than 40% of the explainable variance in the rate of DNM. As expected the rate of divergence between species is correlated to the rate of DNM. However, the correlations are weaker than expected if all the variation in divergence was due to variation in the mutation rate. We provide evidence that this is due the effect of biased gene conversion on the probability that a mutation will become fixed. In contrast to divergence, we find that most of the variation in diversity can be explained by variation in the mutation rate. Finally, we show that the correlation between divergence and DNM density declines as increasingly divergent species are considered.
We designed novel field experimental infrastructure to resolve the relative importance of changes in the climate mean and variance in regulating the structure and function of dryland populations, communities, and ecosystem processes. The Mean - Variance Climate Experiment (MVE) adds three novel elements to prior designs that have manipulated interannual variance in climate in the field (Gherardi & Sala, 2013) by (i) determining interactive effects of mean and variance with a factorial design that crosses reduced mean with increased variance, (ii) studying multiple dryland biomes to compare their susceptibility to transition under interactive climate drivers, and (iii) adding stochasticity to our treatments to permit the antecedent effects that occur under natural climate variability. This new infrastructure enables direct experimental tests of the hypothesis that interactions between the mean and variance of precipitation will have larger ecological impacts than either the mean or variance in precipitation alone.
This dataset includes plant species cover and height data measured in 1 m x 1 m quadrats at all Mean - Variance experiment sites. Quadrat locations span five important ecosystems of the American southwest: blue grama-dominated Plains grassland (est. fall 2019), black grama-dominated Chihuahuan Desert grassland (est. fall 2020), creosotebush dominated Chihuahuan Desert shrubland (est. fall 2021), juniper savanna (est. fall 2022) and pinon-juniper woodland (est. fall 2023). Data on plant cover and height for each plant species are collected per individual plant or patch (for clonal plants) within 1 m x 1 m quadrats. These data inform population dynamics of foundational and rare plant species. The cover and height of individual plants or patches are sampled twice yearly (spring and fall) in permanent 1m x 1m plots within each site or experiment. This data package includes plant cover and height only -- for species biomass estimates per quad see package knb-lter-sev.350.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Context
The dataset illustrates the median household income in Show Low, spanning the years from 2010 to 2021, with all figures adjusted to 2022 inflation-adjusted dollars. Based on the latest 2017-2021 5-Year Estimates from the American Community Survey, it displays how income varied over the last decade. The dataset can be utilized to gain insights into median household income trends and explore income variations.
Key observations:
From 2010 to 2021, the median household income for Show Low increased by $2,280 (4.10%), as per the American Community Survey estimates. In comparison, median household income for the United States increased by $4,559 (6.51%) between 2010 and 2021.
Analyzing the trend in median household income between the years 2010 and 2021, spanning 11 annual cycles, we observed that median household income, when adjusted for 2022 inflation using the Consumer Price Index retroactive series (R-CPI-U-RS), experienced growth year by year for 6 years and declined for 5 years.
https://i.neilsberg.com/ch/show-low-az-median-household-income-trend.jpeg" alt="Show Low, AZ median household income trend (2010-2021, in 2022 inflation-adjusted dollars)">
When available, the data consists of estimates from the U.S. Census Bureau American Community Survey (ACS) 2017-2021 5-Year Estimates. All incomes have been adjusting for inflation and are presented in 2022-inflation-adjusted dollars.
Years for which data is available:
Variables / Data Columns
Good to know
Margin of Error
Data in the dataset are based on the estimates and are subject to sampling variability and thus a margin of error. Neilsberg Research recommends using caution when presening these estimates in your research.
Custom data
If you do need custom data for any of your research project, report or presentation, you can contact our research staff at research@neilsberg.com for a feasibility of a custom tabulation on a fee-for-service basis.
Neilsberg Research Team curates, analyze and publishes demographics and economic data from a variety of public and proprietary sources, each of which often includes multiple surveys and programs. The large majority of Neilsberg Research aggregated datasets and insights is made available for free download at https://www.neilsberg.com/research/.
This dataset is a part of the main dataset for Show Low median household income. You can refer the same here
This dataset is comprised of mean and variance of the surface velocity field of the Gulf of Mexico, obtained from a large set of historical surface drifter data from the Gulf of Mexico—3761 trajectories spanning 27 years and more than a dozen data sources— which were, uniformly processed and quality controlled, and assimilated into a spatially and temporally gridded dataset. A gridded product, called GulfFlow, is created by averaging all available data from the GulfDrifters dataset within quarter-degree spatial bins, and within overlapping monthlong temporal bins having a semimonthly spacing. The data set runs from August 16, 1992 to August 1, 2019, for a total of 648 overlapping time slices. Odd- numbered slices correspond to calendar months, while even-numbered slices run from halfway through one month to halfway through the following month. A higher spatial resolution version, GulfFlow-1/12 degree is created in the identical way but using 1/12 degree bins instead of quarter-degree bins. In addition to the average velocities within each 3D bin, the count of sources contributing to each bin is also distributed, as is the subgridscale velocity variance discussed in the next section. The count variable is a four-dimensional array of integers, the fourth dimension of which has length 30. This variable gives the number of hourly observations from each source dataset contributing to each three-dimensional bin. Values 1–15 are the count of velocities from drifters from each of the 15 experiments that have not been flagged as having lost their drogues, while values 16–30 are for observation from drifters that have been flagged as having lost their drogue. Values above 15 are only populated for the GDP, HARGOS, LASER and some of the DWDE drifters, as a drogue presence flag is not always available. It is useful at this stage to introduce notation for different types of averages. For convenience we represent the velocity as a vector, u = [u v]T , where the superscript “T” denotes the transpose. Let an overbar, \(\overline {\bf u}\) , denote an average over a spatial bin and over all times, while angled brackets, <u>, denote an average over a spatial bin and a particular temporal bin. Thus, <u>, is a function of time while \(\overline {\bf u}\) is not. We refer to <u>, as the local average, \(\overline {\bf u}\) as the global average, and \(\overline {<\bf u>}\) as the double average. Given the inhomogeneity of the drifter data, turns out the global average is biased towards intensive but short duration programs, hence the double average results in a much better representation of the true mean velocity field. The dataset includes the global average \(\overline {<\bf u>}\), the local covariance defined as
\(\bf{ε}=<(u − )(𝐮−< 𝐮 >)^T>\)
and \(\epsilon^2\)which is the trace of \(\overline{\bf ε}\)
\(\epsilon^2\)=\(tr\{\overline{\bf ε}\}\)
The data is distributed in two separate netcdCDF files, one for each grid resolution.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Site data contained in the ScrIntrvls_AllSrcRefs_AllWellsRev.csv dataset define the top and bottom altitudes of well screens in 64,763 irrigation wells completed in the Mississippi River Valley alluvial aquifer (MRVA) that constitute a production zone in the Mississippi Alluvial Plain (MAP) extending across the midwestern and southern United States from Illinois to Louisiana. Each well entry contains an Enumerated Domain Value of the Attribute Label SrcRefNo to identify the state environmental agency that contributed to the database, and enumerated values are associated with specific state agencies by using the Enumerated Domain Value Definition. Screen-top and -bottom altitudes and land surface are referenced (corrected) to the National Elevation Dataset (NED) 10-meter digital elevation model (DEM; https://nationalmap.gov/elevation.html). The dataset identifies 50,103 screen-bottom altitudes and 50,457 screen-top altitudes that were used in subsequent geostatistical estimation after spatial analytics filtered out duplicate-coordinate wells, geographic and stratigraphic outliers, and incongruities of screen-top and -bottom altitudes compared with DEM land-surface altitude and the published digital surface of the bottom altitude of the MRVA (https://doi.org/10.3133/sim3426). Well entries are indexed in the dataset to identify use in four regional geostatistical models that collectively encompass the MAP extent and provide gridded estimates of screen-top and-bottom altitudes and estimation uncertainty (estimation variance) associated with each gridded altitude estimate. Digital surfaces of screen-top and -bottom estimates and estimation variances are represented as raster datasets that were converted to netCDF format and conform with the previously published National Hydrologic Grid (https://doi.org/10.5066/F7P84B24) at one-kilometer resolution. This dataset contains high-quality map images for gridded estimates of MRVA screen-top and -bottom altitude and for corresponding gridded estimation variances saved in the Tagged Image File (.tif) format.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
This section presents a discussion of the research data. The data was received as secondary data however, it was originally collected using the time study techniques. Data validation is a crucial step in the data analysis process to ensure that the data is accurate, complete, and reliable. Descriptive statistics was used to validate the data. The mean, mode, standard deviation, variance and range determined provides a summary of the data distribution and assists in identifying outliers or unusual patterns. The data presented in the dataset show the measures of central tendency which includes the mean, median and the mode. The mean signifies the average value of each of the factors presented in the tables. This is the balance point of the dataset, the typical value and behaviour of the dataset. The median is the middle value of the dataset for each of the factors presented. This is the point where the dataset is divided into two parts, half of the values lie below this value and the other half lie above this value. This is important for skewed distributions. The mode shows the most common value in the dataset. It was used to describe the most typical observation. These values are important as they describe the central value around which the data is distributed. The mean, mode and median give an indication of a skewed distribution as they are not similar nor are they close to one another. In the dataset, the results and discussion of the results is also presented. This section focuses on the customisation of the DMAIC (Define, Measure, Analyse, Improve, Control) framework to address the specific concerns outlined in the problem statement. To gain a comprehensive understanding of the current process, value stream mapping was employed, which is further enhanced by measuring the factors that contribute to inefficiencies. These factors are then analysed and ranked based on their impact, utilising factor analysis. To mitigate the impact of the most influential factor on project inefficiencies, a solution is proposed using the EOQ (Economic Order Quantity) model. The implementation of the 'CiteOps' software facilitates improved scheduling, monitoring, and task delegation in the construction project through digitalisation. Furthermore, project progress and efficiency are monitored remotely and in real time. In summary, the DMAIC framework was tailored to suit the requirements of the specific project, incorporating techniques from inventory management, project management, and statistics to effectively minimise inefficiencies within the construction project.