Facebook
TwitterDH represents 100% for the relative measure. Differences between medians and distributions were significant between all disciplines if indicated with * and were significantly different between GS and SG when marked with 1, significantly different between GS and DH if marked with 2 and significantly different between SG and DH if marked with 3. If no parameter was significantly different the column is empty. Columns marked with—indicate that the measure was not calculated.Median, interquartile range (IQR) and significance level of the difference between discipline medians and distributions for all parameters, and percentage of DH for GS and SG.
Facebook
TwitterS12 Fig. shows results are robust to the range of the latin hypercube sample.IQR, inter-quartile range.Mean values and inter-quartile ranges of plausible parameter sets.
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
The median (and interquartile range) of the individuals’ median and inter quartile range (range) of both the time on the stone and the number of steps on the stone in the standardized and nonstandardized configuration.
Facebook
TwitterScan 1 and scan 2 single-participant ROI median and interquartile range (IQR) ultrashort-T2* values and between-scan absolute and percent change calculated from the two scan sessions for one participant. The results calculated using all seven TE values and the subset of three TE values are presented for all voxels and for only voxels with acceptable ultrashort-T2* fit (R2 ≥ 0.5). The median and interquartile range values are calculated for the sample of all voxels within each ROI.
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Parameters are shown separately for the three different datasets (1, 2, pathological gamblers [PG]).
Facebook
TwitterThe exercise after this contains questions that are based on the housing dataset.
How many houses have a waterfront? a. 21000 b. 21450 c. 163 d. 173
How many houses have 2 floors? a. 2692 b. 8241 c. 10680 d. 161
How many houses built before 1960 have a waterfront? a. 80 b. 7309 c. 90 d. 92
What is the price of the most expensive house having more than 4 bathrooms? a. 7700000 b. 187000 c. 290000 d. 399000
For instance, if the ‘price’ column consists of outliers, how can you make the data clean and remove the redundancies? a. Calculate the IQR range and drop the values outside the range. b. Calculate the p-value and remove the values less than 0.05. c. Calculate the correlation coefficient of the price column and remove the values less than the correlation coefficient. d. Calculate the Z-score of the price column and remove the values less than the z-score.
What are the various parameters that can be used to determine the dependent variables in the housing data to determine the price of the house? a. Correlation coefficients b. Z-score c. IQR Range d. Range of the Features
If we get the r2 score as 0.38, what inferences can we make about the model and its efficiency? a. The model is 38% accurate, and shows poor efficiency. b. The model is showing 0.38% discrepancies in the outcomes. c. Low difference between observed and fitted values. d. High difference between observed and fitted values.
If the metrics show that the p-value for the grade column is 0.092, what all inferences can we make about the grade column? a. Significant in presence of other variables. b. Highly significant in presence of other variables c. insignificance in presence of other variables d. None of the above
If the Variance Inflation Factor value for a feature is considerably higher than the other features, what can we say about that column/feature? a. High multicollinearity b. Low multicollinearity c. Both A and B d. None of the above
Facebook
TwitterDistribution of Body Mass Index (BMI) (mean, standard deviation (SD), median and inter-quartile range (IQR)) and frequencies of major BMI categories across year of birth and contextual variables of Swiss conscripts.
Facebook
TwitterOur target was to predict gender, age and emotion from audio. We found audio labeled datasets on Mozilla and RAVDESS. So by using R programming language 20 statistical features were extracted and then after adding the labels these datasets were formed. Audio files were collected from "Mozilla Common Voice" and “Ryerson AudioVisual Database of Emotional Speech and Song (RAVDESS)”.
Datasets contains 20 feature columns and 1 column for denoting the label. The 20 statistical features were extracted through the Frequency Spectrum Analysis using R programming Language. They are: 1) meanfreq - The mean frequency (in kHz) is a pitch measure, that assesses the center of the distribution of power across frequencies. 2) sd - The standard deviation of frequency is a statistical measure that describes a dataset’s dispersion relative to its mean and is calculated as the variance’s square root. 3) median - The median frequency (in kHz) is the middle number in the sorted, ascending, or descending list of numbers. 4) Q25 - The first quartile (in kHz), referred to as Q1, is the median of the lower half of the data set. This means that about 25 percent of the data set numbers are below Q1, and about 75 percent are above Q1. 5) Q75 - The third quartile (in kHz), referred to as Q3, is the central point between the median and the highest distributions. 6) IQR - The interquartile range (in kHz) is a measure of statistical dispersion, equal to the difference between 75th and 25th percentiles or between upper and lower quartiles. 7) skew - The skewness is the degree of distortion from the normal distribution. It measures the lack of symmetry in the data distribution. 8) kurt - The kurtosis is a statistical measure that determines how much the tails of distribution vary from the tails of a normal distribution. It is actually the measure of outliers present in the data distribution. 9) sp.ent - The spectral entropy is a measure of signal irregularity that sums up the normalized signal’s spectral power. 10) sfm - The spectral flatness or tonality coefficient, also known as Wiener entropy, is a measure used for digital signal processing to characterize an audio spectrum. Spectral flatness is usually measured in decibels, which, instead of being noise-like, offers a way to calculate how tone-like a sound is. 11) mode - The mode frequency is the most frequently observed value in a data set. 12) centroid - The spectral centroid is a metric used to describe a spectrum in digital signal processing. It means where the spectrum’s center of mass is centered. 13) meanfun - The meanfun is the average of the fundamental frequency measured across the acoustic signal. 14) minfun - The minfun is the minimum fundamental frequency measured across the acoustic signal 15) maxfun - The maxfun is the maximum fundamental frequency measured across the acoustic signal. 16) meandom - The meandom is the average of dominant frequency measured across the acoustic signal. 17) mindom - The mindom is the minimum of dominant frequency measured across the acoustic signal. 18) maxdom - The maxdom is the maximum of dominant frequency measured across the acoustic signal 19) dfrange - The dfrange is the range of dominant frequency measured across the acoustic signal. 20) modindx - the modindx is the modulation index, which calculates the degree of frequency modulation expressed numerically as the ratio of the frequency deviation to the frequency of the modulating signal for a pure tone modulation.
Gender and Age Audio Data Souce: Link: https://commonvoice.mozilla.org/en Emotion Audio Data Souce: Link : https://smartlaboratory.org/ravdess/
Facebook
TwitterThis dataset includes elevation-based probability and depth statistics for estimating inundation under various sea-level rise and high tide flooding scenarios in and around the National Park Service’s De Soto National Memorial. These datasets were developed using 1-m digital elevation model (DEM) from the 3D Elevation program. This data release includes results from analyses of two local sea-level rise scenarios for two-time steps — the Intermediate-Low and Intermediate-High for 2050 and 2100 from Sweet and others (2022). Additionally, this data release includes maps of inundation probability under the minor, moderate, and major high tide flooding thresholds defined by the National oceanic and Atmospheric Administration (NOAA). We estimated the probability of an area being inundated under a given scenario using Monte Carlo simulations with 1,000 iterations. For an individual iteration, each pixel of the DEM was randomly propagated based on the lidar data uncertainty, while the sea-level rise and high tide flooding water level estimates were also propagated based on uncertainty in the estimate (Sweet and others, 2022) and tidal datum transformation, respectively. Moreover, the probability of a pixel being inundated was calculated by summing the binary simulation outputs and dividing by 1,000. Following, probability was binned into the following classes: 1) Unlikely, probability ≤0.33; 2) Likely as not, probability >0.33 and ≤0.66; and 3) Likely, probability >0.66. Finally, depth statistics were only recorded when depth was equal to or greater than 0. We calculated the median depth, 25th percentile, 75th percentile, and interquartile range using all the pixels that met this criterion. When utilizing the depth statistics, it is important to also consider the probability of this pixel being flooded. In other words, the depth layers may show some depth returns, but the pixel may have rarely been inundated for the 1,000 iterations.
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
The objective was to identify horn fly-susceptible and horn fly-resistant animals in a Sindhi herd by two different methods. The number of horn flies on 25 adult cows from a Sindhi herd was counted every 14 days. As it was an open herd, the trial period was divided into three stages based on cow composition, with the same cows maintained within each period: 2011-2012 (36 biweekly observations); 2012-2013 (26 biweekly observations); and 2013-2014 (22 biweekly observations). Only ten cows were present in the herd throughout the entire period from 2011-2014 (84 biweekly observations). The variables evaluated were the number of horn flies on the cows, the sampling date and a binary variable for rainy or dry season. Descriptive statistics were calculated, including the median, the interquartile range, and the minimum and maximum number of horn flies, for each observation day. For the present analysis, fly-susceptible cows were identified as those for which the infestation of flies appeared in the upper quartile for more than 50% of the weeks and in the lower quartile for less than 20% of the weeks. In contrast, fly-resistant cows were defined as those for which the fly counts appeared in the lower quartile for more than 50% of the weeks and in the upper quartile for less than 20% of the weeks. To identify resistant and susceptible cows for the best linear unbiased predictions analysis, three repeated measures linear mixed models (one for each period) were constructed with cow as a random effect intercept. The response variable was the log ten transformed counts of horn flies per cow, and the explanatory variable were the observation date and season. As the trail took place in a semiarid region with two seasons well stablished the season was evaluated monthly as a binary outcome, considering a rainy season if it rained more or equal than 50mm or dry season if the rain was less than 50mm. The Standardized residuals and the BLUPs of the random effects were obtained and assessed for normality, heteroscedasticity and outlying observations. Each cow’s BLUPs were plotted against the average quantile rank values that were determined as the difference between the number of weeks in the high-risk quartile group and the number of weeks in the low risk quartile group, averaged by the total number of weeks in each of the observation periods. A linear model fit for the values of BLUPS against the average rank values and the correlation between the two methods was tested using Spearman’s correlation coefficient. The animal effect values (BLUPs) were evaluated by percentiles, with 0 representing the lowest counts (or more resistant cows) and 10 representing the highest counts (or more susceptible cows). These BLUPs represented only the effect of cow and not the effect of day, season or other unmeasured counfounders.
Facebook
TwitterThis dataset includes elevation-based probability and depth statistics for estimating inundation under various sea-level rise and high tide flooding scenarios in and around the National Park Service’s Dry Tortugas National Park. These datasets were developed using digital elevation model (DEM) from National Oceanic and Atmospheric Administration (NOAA). This data release includes results from analyses of two local sea-level rise scenarios for two-time steps — the Intermediate-Low and Intermediate-High for 2050 and 2100 from Sweet and others (2022). Additionally, this data release includes maps of inundation probability under the minor, moderate, and major high tide flooding thresholds defined by NOAA. We estimated the probability of an area being inundated under a given scenario using Monte Carlo simulations with 1,000 iterations. For an individual iteration, each pixel of the DEM was randomly propagated based on the lidar data uncertainty, while the sea-level rise and high tide flooding water level estimates were also propagated based on uncertainty in the sea-level rise estimate (Sweet and others, 2022) and tidal datum transformation, respectively. Moreover, the probability of a pixel being inundated was calculated by summing the binary simulation outputs and dividing by 1,000. Following, probability was binned into the following classes: 1) Unlikely, probability ≤0.33; 2) Likely as not, probability >0.33 and ≤0.66; and 3) Likely, probability >0.66. Finally, depth statistics were only recorded when depth was equal to or greater than 0. We calculated the median depth, 25th percentile, 75th percentile, and interquartile range using all the pixels that met this criterion. When utilizing the depth statistics, it is important to also consider the probability of this pixel being flooded. In other words, the depth layers may show some depth returns, but the pixel may have rarely been inundated for the 1,000 iterations.
Facebook
TwitterThis dataset includes elevation-based probability and depth statistics for estimating inundation under various sea-level rise and high tide flooding scenarios in and around the National Park Service’s Big Cypress National Preserve. For information on the digital elevation model (DEM) source used to develop these datasets refer to the corresponding spatial metadata file (Danielson and others, 2023). This data release includes results from analyses of two local sea-level rise scenarios for two-time steps — the Intermediate-Low and Intermediate-High for 2050 and 2100 from Sweet and others (2022). Additionally, this data release includes maps of inundation probability under the minor, moderate, and major high tide flooding thresholds defined by the National Oceanic and Atmospheric Administration (NOAA). We estimated the probability of an area being inundated under a given scenario using Monte Carlo simulations with 1,000 iterations. For an individual iteration, each pixel of the DEM was randomly propagated based on the lidar data uncertainty, while the sea-level rise and high tide flooding water level estimates were also propagated based on uncertainty in the sea-level rise estimate (Sweet and others, 2022) and tidal datum transformation, respectively. Moreover, the probability of a pixel being inundated was calculated by summing the binary simulation outputs and dividing by 1,000. Following, probability was binned into the following classes: 1) Unlikely, probability ≤0.33; 2) Likely as not, probability >0.33 and ≤0.66; and 3) Likely, probability >0.66. Finally, depth statistics were only recorded when depth was equal to or greater than 0. We calculated the median depth, 25th percentile, 75th percentile, and interquartile range using all the pixels that met this criterion. When utilizing the depth statistics, it is important to also consider the probability of this pixel being flooded. In other words, the depth layers may show some depth returns, but the pixel may have rarely been inundated for the 1,000 iterations.
Facebook
TwitterLarval life history traits and geographic distribution for each thoracican barnacle species used in the study
The table "finalmergeddata.csv" contains life history and enironmental data as well as the calculated variance (IQR = interquartile range, se = standard error) summarized per species. The table "lifehistory.xls" contains the species-specific larval life history data we extracted from the literature. The first tab, "Taxonomy + larval mode" has one row per species. The taxonomy is taken from WoRMS (www.marinespecies.org). The following two tabs contain information on other larval traits and the known geographic distribution of the barnacle species. In these tabs, each species can occur several times, as we chose to give each reference a separate row. The references are detailed in the datatable_references file. The meaning of all columns is explained in the last tab "METADATA". Detailed references for the data sources are available in the last tab "Data sourc...
Facebook
TwitterThis dataset includes elevation-based probability and depth statistics for estimating inundation under various sea-level rise and high tide flooding scenarios in and around the National Park Service’s Biscayne National Park. For information on the digital elevation model (DEM) source used to develop these datasets refer to the corresponding spatial metadata file (Danielson and others, 2023). This data release includes results from analyses of two local sea-level rise scenarios for two-time steps — the Intermediate-Low and Intermediate-High for 2040 and 2080 from Sweet and others (2022). Additionally, this data release includes maps of inundation probability under the minor, moderate, and major hight tide flooding thresholds. We estimated the probability of an area being inundated under a given scenario using Monte Carlo simulations with 1,000 iterations. For an individual iteration, each pixel of the DEM was randomly propagated based on the lidar data uncertainty, while the sea-level rise and high tide flooding water level estimates were also propagated based on uncertainty in the estimate (Sweet and others, 2022) and tidal datum transformation. Moreover, the probability of a pixel being inundated was calculated by summing the binary simulation outputs and dividing by 1,000. Following, probability was binned into the following classes: 1) Unlikely, probability ≤0.33; 2) Likely as not, probability >0.33 and ≤0.66; and 3) Likely, probability >0.66. Finally, depth statistics were only recorded when depth was equal to or greater than 0. We calculated the median depth, 25th percentile, 75th percentile, and interquartile range using all the pixels that met this criterion. When utilizing the depth statistics, it is important to also consider the probability of this pixel being flooded. In other words, the depth layers may show some depth returns, but the pixel may have rarely been inundated for the 1,000 iterations.
Facebook
Twitterhttps://creativecommons.org/publicdomain/zero/1.0/https://creativecommons.org/publicdomain/zero/1.0/
Walmart Inc. is a multinational retail corporation that operates a chain of hypermarkets, discount department stores, and grocery stores. It is one of the world's largest companies by revenue and a key player in the retail sector. Walmart's stock is actively traded on major stock exchanges, making it an interesting subject for financial analysis.
This dataset contains historical stock price data for Walmart, sourced directly from Yahoo Finance using the yfinance Python API. The data covers daily stock prices and includes multiple key financial indicators.
This notebook performs an extensive EDA to uncover insights into Walmart's stock price trends, volatility, and overall behavior in the stock market. The following analysis steps are included:
This dataset and analysis can be useful for: - 📡 Stock Market Analysis – Evaluating Walmart’s stock price trends and volatility. - 🏦 Investment Research – Assisting traders and investors in making informed decisions. - 🎓 Educational Purposes – Teaching data science and financial analysis using real-world stock data. - 📊 Algorithmic Trading – Developing trading strategies based on historical stock price trends.
📥 Download the dataset and explore Walmart’s stock performance today! 🚀
Facebook
TwitterThis dataset includes elevation-based probability and depth statistics for estimating inundation under various sea-level rise and high tide flooding scenarios in and around the National Park Service’s San Juan National Historic Site. These datasets were developed using 1-m digital elevation model (DEM) from the 3D Elevation program. This data release includes results from analyses of two local sea-level rise scenarios for two-time steps — the Intermediate-Low and Intermediate-High for 2050 and 2100 from Sweet and others (2022). Additionally, this data release includes maps of inundation probability under the minor, moderate, and major high tide flooding thresholds defined by the National Oceanic and Atmospheric Administration (NOAA). We estimated the probability of an area being inundated under a given scenario using Monte Carlo simulations with 1,000 iterations. For an individual iteration, each pixel of the DEM was randomly propagated based on the lidar data uncertainty, while the sea-level rise and high tide flooding water level estimates were also propagated based on uncertainty in the sea-level rise estimate (Sweet and others, 2022) and tidal datum transformation, respectively. Moreover, the probability of a pixel being inundated was calculated by summing the binary simulation outputs and dividing by 1,000. Following, probability was binned into the following classes: 1) Unlikely, probability ≤0.33; 2) Likely as not, probability >0.33 and ≤0.66; and 3) Likely, probability >0.66. Finally, depth statistics were only recorded when depth was equal to or greater than 0. We calculated the median depth, 25th percentile, 75th percentile, and interquartile range using all the pixels that met this criterion. When utilizing the depth statistics, it is important to also consider the probability of this pixel being flooded. In other words, the depth layers may show some depth returns, but the pixel may have rarely been inundated for the 1,000 iterations.
Facebook
TwitterThe experiments have been carried out with a group of 30 volunteers within an age bracket of 19-48 years. Each person performed six activities (WALKING, WALKING_UPSTAIRS, WALKING_DOWNSTAIRS, SITTING, STANDING, LAYING) wearing a smartphone (Samsung Galaxy S II) on the waist. Using its embedded accelerometer and gyroscope, we captured 3-axial linear acceleration and 3-axial angular velocity at a constant rate of 50Hz. The experiments have been video-recorded to label the data manually. The obtained dataset has been randomly partitioned into two sets, where 70% of the volunteers was selected for generating the training data and 30% for the test data.
The sensor signals (accelerometer and gyroscope) were pre-processed by applying noise filters and then sampled in fixed-width sliding windows of 2.56 sec and 50% overlap (128 readings/window). The sensor acceleration signal, which has gravitational and body motion components, was separated using a Butterworth low-pass filter into body acceleration and gravity. The gravitational force is assumed to have only low-frequency components, therefore a filter with 0.3 Hz cutoff frequency was used. From each window, a vector of features was obtained by calculating variables from the time and frequency domain.
The features selected for this database come from the accelerometer and gyroscope 3-axial raw signals tAcc-XYZ and tGyro-XYZ. These time-domain signals (prefix 't' to denote time) were captured at a constant rate of 50 Hz. Then they were filtered using a median filter and a 3rd order low pass Butterworth filter with a corner frequency of 20 Hz to remove noise. Similarly, the acceleration signal was then separated into the body and gravity acceleration signals (tBodyAcc-XYZ and tGravityAcc-XYZ) using another low pass Butterworth filter with a corner frequency of 0.3 Hz.
Subsequently, the body l linear acceleration and angular velocity were derived in time to obtain Jerk signals (tBodyAccJerk-XYZ and tBodyGyroJerk-XYZ). Also the magnitude of these three-dimensional signals were calculated using the Euclidean norm (tBodyAccMag, tGravityAccMag, tBodyAccJerkMag, tBodyGyroMag, tBodyGyroJerkMag).
Finally a Fast Fourier Transform (FFT) was applied to some of these signals producing fBodyAcc-XYZ, fBodyAccJerk-XYZ, fBodyGyro-XYZ, fBodyAccJerkMag, fBodyGyroMag, fBodyGyroJerkMag. (Note the 'f' to indicate frequency domain signals).
These signals were used to estimate variables of the feature vector for each pattern: '-XYZ' is used to denote 3-axial signals in the X, Y, and Z directions.
tBodyAcc-XYZ tGravityAcc-XYZ tBodyAccJerk-XYZ tBodyGyro-XYZ tBodyGyroJerk-XYZ tBodyAccMag tGravityAccMag tBodyAccJerkMag tBodyGyroMag tBodyGyroJerkMag fBodyAcc-XYZ fBodyAccJerk-XYZ fBodyGyro-XYZ fBodyAccMag fBodyAccJerkMag fBodyGyroMag fBodyGyroJerkMag
The set of variables that were estimated from these signals are:
mean(): Mean value std(): Standard deviation mad(): Median absolute deviation max(): Largest value in array min(): Smallest value in array sma(): Signal magnitude area energy(): Energy measure. Sum of the squares divided by the number of values. iqr(): Interquartile range entropy(): Signal entropy arCoeff(): Autorregresion coefficients with Burg order equal to 4 correlation(): correlation coefficient between two signals maxInds(): index of the frequency component with the largest magnitude meanFreq(): Weighted average of the frequency components to obtain a mean frequency skewness(): skewness of the frequency domain signal kurtosis(): kurtosis of the frequency domain signal bandsEnergy(): Energy of a frequency interval within the 64 bins of the FFT of each window. angle(): Angle between two vectors.
Additional vectors are obtained by averaging the signals in a signal window sample. These are used on the angle() variable:
gravityMean tBodyAccMean tBodyAccJerkMean tBodyGyroMean tBodyGyroJerkMean
This data set consists of the following columns:
1 tBodyAcc-mean()-X 2 tBodyAcc-mean()-Y 3 tBodyAcc-mean()-Z 4 tBodyAcc-std()-X 5 tBodyAcc-std()-Y 6 tBodyAcc-std()-Z 7 tBodyAcc-mad()-X 8 tBodyAcc-mad()-Y 9 tBodyAcc-mad()-Z 10 tBodyAcc-max()-X 11 tBodyAcc-max()-Y 12 tBodyAcc-max()-Z 13 tBodyAcc-min()-X 14 tBodyAcc-min()-Y 15 tBodyAcc-min()-Z 16 tBodyAcc-sma() 17 tBodyAcc-energy()-X 18 tBodyAcc-energy()-Y 19 tBodyAcc-energy()-Z 20 tBodyAcc-iqr()-X 21 tBodyAcc-iqr()-Y 22 tBodyAcc-iqr()-Z 23 tBodyAcc-entropy()-X 24 tBodyAcc-entropy()-Y 25 tBodyAcc-entropy()-Z 26 tBodyAcc-arCoeff()-X,1 27 tBodyAcc-arCoeff()-X,2 28 tBodyAcc-arCoeff()-X,3 29 tBodyAcc-arCoeff()-X,4 30 tBodyAcc-arCoeff()-Y,1 31 tBodyAcc-arCoeff()-Y,2 32 tBodyAcc-arCoeff()-Y,3 33 tBodyAcc-arCoeff()-Y,4 34 tBodyAcc-arCoeff()-Z,1 35 tBodyAcc-arCoeff()-Z,2 36 tBodyAcc-arCoeff()-Z,3 37 tBodyAcc-arCoeff()-Z,4 38 tBodyAcc-correlation()-X,Y 39 tBodyAcc-correlation()-X,Z 40 tBodyAcc-correlation()-Y,Z 41 tGravityAcc-mean()-X 42 tGravit...
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
The risk for the occurrence of osteoporotic fractures according to quartiles of gamma-glutamyl transferase variability.
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
A live version of the data record, which will be kept up-to-date with new estimates, can be downloaded from the Humanitarian Data Exchange: https://data.humdata.org/dataset/covid-19-mobility-italy.
If you find the data helpful or you use the data for your research, please cite our work:
Pepe, E., Bajardi, P., Gauvin, L., Privitera, F., Lake, B., Cattuto, C., & Tizzoni, M. (2020). COVID-19 outbreak response, a dataset to assess mobility changes in Italy following national lockdown. Scientific Data 7, 230 (2020).
The data record is structured into 4 comma-separated value (CSV) files, as follows:
id_provinces_IT.csv. Table of the administrative codes of the 107 Italian provinces. The fields of the table are:
COD_PROV is an integer field that is used to identify a province in all other data records;
SIGLA is a two-letters code that identifies the province according to the ISO_3166-2 standard (https://en.wikipedia.org/wiki/ISO_3166-2:IT);
DEN_PCM is the full name of the province.
OD_Matrix_daily_flows_norm_full_2020_01_18_2020_04_17.csv. The file contains the daily fraction of users’ moving between Italian provinces. Each line corresponds to an entry of matrix (i, j). The fields of the table are:
p1: COD_PROV of origin,
p2: COD_PROV of destination,
day: in the format yyyy-mm-dd.
median_q1_q3_rog_2020_01_18_2020_04_17.csv. The file contains median and interquartile range (IQR) of users’ radius of gyration in a province by week. Each entry of the table fields of the table are:
COD_PROV of the province;
SIGLA of the province;
DEN_PCM of the province;
week: median value of the radius of gyration on week week, with week in the format dd/mm-DD/MM where dd/mm and DD/MM are the first and the last day of the week, respectively.
week Q1 first quartile (Q1) of the distribution of the radius of gyration on week week,
week Q3 third quartile (Q3) of the distribution of the radius of gyration on week week,
average_network_degree_2020_01_18_2020_04_17.csv. The file contains daily time-series of the average degree 〈k〉 of the proximity network. Each entry of the table is a value of 〈k〉 on a given day. The fields of the table are:
COD_PROV of the province;
SIGLA of the province;
DEN_PCM of the province;
day in the format yyyy-mm-dd.
ESRI shapefiles of the Italian provinces updated to the most recent definition are available from the website of the Italian National Office of Statistics (ISTAT): https://www.istat.it/it/archivio/222527.
Facebook
TwitterSupplementary Material Table S4: Results of a more detailed TTO analysis performed at both SOC and PT level. Min: minimum; Max: maximum; IQR: interquartile range; Q1: first quartile; Q3: third quartile; SD: standard deviation; SE: standard error.
Facebook
TwitterDH represents 100% for the relative measure. Differences between medians and distributions were significant between all disciplines if indicated with * and were significantly different between GS and SG when marked with 1, significantly different between GS and DH if marked with 2 and significantly different between SG and DH if marked with 3. If no parameter was significantly different the column is empty. Columns marked with—indicate that the measure was not calculated.Median, interquartile range (IQR) and significance level of the difference between discipline medians and distributions for all parameters, and percentage of DH for GS and SG.