This table shows overall ATCEMS response interval performance for entire fiscal years. Data in the table is broken out by incident response priority and service area (City of Austin or Travis County).
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
This RR interval dataset is derived from 10,000 cases of 24-hour Holter monitoring data sampled at 128 Hz. Among the cases, 9,500 are labeled as non-atrial fibrillation (NAF), and 500 as paroxysmal atrial fibrillation (PAF). These data have been used in the article "Clinician-AI Collaboration: A Win-Win solution for Efficiency and Reliability in Atrial Fibrillation Diagnosis".The dataset formated as CSV file consists of two columns:rr_interval: Represents the interval between consecutive R-peaks, measured in milliseconds.label: Categorical labels for the beats, where:1 indicates AF0 indicates NAF-1 indicates noise or artifactsEach case is named based on its category. NAF cases are labeled as NAF0001.csv through NAF9500.csv, while PAF cases are labeled as PAF0001.csv through PAF0500.csv.For any questions, please contact the email: hustzp@hust.edu.cn
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
This is the data set behind the Wind Generation Interactive Query Tool created by the CEC. The visualization tool interactively displays wind generation over different time intervals in three-dimensional space. The viewer can look across the state to understand generation patterns of regions with concentrations of wind power plants. The tool aids in understanding high and low periods of generation. Operation of the electric grid requires that generation and demand are balanced in each period.
Renewable energy resources like wind facilities vary in size and geographic distribution within each state. Resource planning, land use constraints, climate zones, and weather patterns limit availability of these resources and where they can be developed. National, state, and local policies also set limits on energy generation and use. An example of resource planning in California is the Desert Renewable Energy Conservation Plan.
By exploring the visualization, a viewer can gain a three-dimensional understanding of temporal variation in generation CFs, along with how the wind generation areas compare to one another. The viewer can observe that areas peak in generation in different periods. The large range in CFs is also visible.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
With the rapid development of data acquisition and storage space, massive datasets exhibited with large sample size emerge increasingly and make more advanced statistical tools urgently need. To accommodate such big volume in the analysis, a variety of methods have been proposed in the circumstances of complete or right censored survival data. However, existing development of big data methodology has not attended to interval-censored outcomes, which are ubiquitous in cross-sectional or periodical follow-up studies. In this work, we propose an easily implemented divide-and-combine approach for analyzing massive interval-censored survival data under the additive hazards model. We establish the asymptotic properties of the proposed estimator, including the consistency and asymptotic normality. In addition, the divide-and-combine estimator is shown to be asymptotically equivalent to the full-data-based estimator obtained from analyzing all data together. Simulation studies suggest that, relative to the full-data-based approach, the proposed divide-and-combine approach has desirable advantage in terms of computation time, making it more applicable to large-scale data analysis. An application to a set of interval-censored data also demonstrates the practical utility of the proposed method.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Symbolic data have become increasingly popular in the era of big data. In this paper, we consider density estimation and regression for interval-valued data, a special type of symbolic data, common in astronomy and official statistics. We propose kernel estimators with adaptive bandwidths to account for variability of each interval. Specifically, we derive cross-validation bandwidth selectors for density estimation and extend the Nadaraya–Watson estimator for regression with interval data. We assess the performance of the proposed methods in comparison with existing kernel methods by extensive simulation studies and real data analysis.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Motivated by a breast cancer study, we consider regression analysis of interval-censored failure time data in the presence of a random change point. Although a great deal of literature on interval-censored data has been established, there does not seem to exist an established method that can allow for the existence of random change points. Such data can occur in, for example, clinical trials where the risk of a disease may dramatically change when some biological indexes of the human body exceed certain thresholds. To fill the gap, we will first consider regression analysis of such data under a class of linear transformation models and provide a sieve maximum likelihood estimation procedure. Then a penalized method is proposed for simultaneous estimation and variable selection, and the asymptotic properties of the proposed method are established. An extensive simulation study is conducted and indicates that the proposed methods work well in practical situations. The approaches are applied to the real data from the breast cancer study mentioned above. Supplementary materials for this article are available online, including a standardized description of the materials available for reproducing the work.
These data document overstory tree height, diameter and growth for a prescribed burning study with unburned controls on the Malheur National Forest in the southern Blue Mountains of Oregon. The original prescribed fires were conducted in the fall of 1997 and spring of 1998 and were repeated at two intervals, five and fifteen years. Five year interval reburns have been repeated three times (four burns total) and the fifteen year interval a single time (two burns total). These data document tree conditions prior to and following the last reburns with reference data for tree growth from 1998. Tree diameter data are available on 10 meter radius plots and diameter, height and growth data are available on 0.2 hectare plots.
Linked data for every time interval and instant into the past and future, from years down to seconds. This is an infinite set of linked data. It includes government years and properly handles the transition to the Gregorian calendar within the UK.
Part of package:reference-data-gov-uk
U.S. Government Workshttps://www.usa.gov/government-works
License information was derived automatically
Includes accelerometer data using an ActiGraph to assess usual sedentary, moderate, vigorous, and very vigorous activity at baseline, 6 weeks, and 10 weeks. Includes relative reinforcing value (RRV) data showing how participants rated how much they would want to perform both physical and sedentary activities on a scale of 1-10 at baseline, week 6, and week 10. Includes data on the breakpoint, or Pmax of the RRV, which was the last schedule of reinforcement (i.e. 4, 8, 16, …) completed for the behavior (exercise or sedentary). For both Pmax and RRV score, greater scores indicated a greater reinforcing value, with scores exceeding 1.0 indicating increased exercise reinforcement. Includes questionnaire data regarding preference and tolerance for exercise intensity using the Preference for and Tolerance of Intensity of Exercise Questionnaire (PRETIEQ) and positive and negative outcome expectancy of exercise using the outcome expectancy scale (OES). Includes data on height, weight, and BMI. Includes demographic data such as gender and race/ethnicity. Resources in this dataset:Resource Title: Actigraph activity data. File Name: AGData.csvResource Description: Includes data from Actigraph accelerometer for each participant at baseline, 6 weeks, and 10 weeks.Resource Title: RRV Data. File Name: RRVData.csvResource Description: Includes data from RRV at baseline, 6 weeks, and 10 weeks, OES survey data, PRETIE-Q survey data, and demographic data (gender, weight, height, race, ethnicity, and age).
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
We present a Wilson interval for binomial proportions for use with multiple imputation for missing data. Using simulation studies, we show that it can have better repeated sampling properties than the usual confidence interval for binomial proportions based on Rubin’s combining rules. Further, in contrast to the usual multiple imputation confidence interval for proportions, the multiple imputation Wilson interval is always bounded by zero and one. Supplementary material is available online.
Subscribers can find out export and import data of 23 countries by HS code or product’s name. This demo is helpful for market analysis.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
The fire return interval departure (FRID) analysis quantifies the difference between current and pre-settlement fire frequencies, allowing managers to target areas at high risk of threshold-type responses owing to altered fire regimes and interactions with other factors. This is a measure of the extent to which contemporary fires (i.e. since 1908) are burning at frequencies similar to the frequencies that occurred prior to Euro-American settlement.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
feature_table: Features for learning flow-sensitivity heuristics. Features present syntactic or semantic properties of variables in C programs.
Figure 1: Figure 1 presents the learned flow-sensitivity heuristic (f0, f1). F0 presents the variables that will be analyzed flow-insensitively while F1 presents the variables that will be analyzed flow-sensitively.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Data File Including 48 Datasets with Values Used in Figs. 2 and 3 in PLOS ONE Article. Thirty of these datasets are the measured time series row data for stride interval in the PD subjects; the remaining 18 datasets are the measured time series row data for stride interval in the control subjects. Each dataset is composed of a single comma separated values (CSV) file and each CSV file stores a data array of two columns: the first column is a time stamp in seconds, the second column is time series data of stride interval.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Predicted water level heights at Portland Roads at regular time intervals.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
ABSTRACT
The issue of diagnosing psychotic diseases, including schizophrenia and bipolar disorder, in particular, the objectification of symptom severity assessment, is still a problem requiring the attention of researchers. Two measures that can be helpful in patient diagnosis are heart rate variability calculated based on electrocardiographic signal and accelerometer mobility data. The following dataset contains data from 30 psychiatric ward patients having schizophrenia or bipolar disorder and 30 healthy persons. The duration of the measurements for individuals was usually between 1.5 and 2 hours. R-R intervals necessary for heart rate variability calculation were collected simultaneously with accelerometer data using a wearable Polar H10 device. The Positive and Negative Syndrome Scale (PANSS) test was performed for each patient participating in the experiment, and its results were attached to the dataset. Furthermore, the code for loading and preprocessing data, as well as for statistical analysis, was included on the corresponding GitHub repository.
BACKGROUND
Heart rate variability (HRV), calculated based on electrocardiographic (ECG) recordings of R-R intervals stemming from the heart's electrical activity, may be used as a biomarker of mental illnesses, including schizophrenia and bipolar disorder (BD) [Benjamin et al]. The variations of R-R interval values correspond to the heart's autonomic regulation changes [Berntson et al, Stogios et al]. Moreover, the HRV measure reflects the activity of the sympathetic and parasympathetic parts of the autonomous nervous system (ANS) [Task Force of the European Society of Cardiology the North American Society of Pacing Electrophysiology, Matusik et al]. Patients with psychotic mental disorders show a tendency for a change in the centrally regulated ANS balance in the direction of less dynamic changes in the ANS activity in response to different environmental conditions [Stogios et al]. Larger sympathetic activity relative to the parasympathetic one leads to lower HRV, while, on the other hand, higher parasympathetic activity translates to higher HRV. This loss of dynamic response may be an indicator of mental health. Additional benefits may come from measuring the daily activity of patients using accelerometry. This may be used to register periods of physical activity and inactivity or withdrawal for further correlation with HRV values recorded at the same time.
EXPERIMENTS
In our experiment, the participants were 30 psychiatric ward patients with schizophrenia or BD and 30 healthy people. All measurements were performed using a Polar H10 wearable device. The sensor collects ECG recordings and accelerometer data and, additionally, prepares a detection of R wave peaks. Participants of the experiment had to wear the sensor for a given time. Basically, it was between 1.5 and 2 hours, but the shortest recording was 70 minutes. During this time, evaluated persons could perform any activity a few minutes after starting the measurement. Participants were encouraged to undertake physical activity and, more specifically, to take a walk. Due to patients being in the medical ward, they received instruction to take a walk in the corridors at the beginning of the experiment. They were to repeat the walk 30 minutes and 1 hour after the first walk. The subsequent walks were to be slightly longer (about 3, 5 and 7 minutes, respectively). We did not remind or supervise the command during the experiment, both in the treatment and the control group. Seven persons from the control group did not receive this order and their measurements correspond to freely selected activities with rest periods but at least three of them performed physical activities during this time. Nevertheless, at the start of the experiment, all participants were requested to rest in a sitting position for 5 minutes. Moreover, for each patient, the disease severity was assessed using the PANSS test and its scores are attached to the dataset.
The data from sensors were collected using Polar Sensor Logger application [Happonen]. Such extracted measurements were then preprocessed and analyzed using the code prepared by the authors of the experiment. It is publicly available on the GitHub repository [Książek et al].
Firstly, we performed a manual artifact detection to remove abnormal heartbeats due to non-sinus beats and technical issues of the device (e.g. temporary disconnections and inappropriate electrode readings). We also performed anomaly detection using Daubechies wavelet transform. Nevertheless, the dataset includes raw data, while a full code necessary to reproduce our anomaly detection approach is available in the repository. Optionally, it is also possible to perform cubic spline data interpolation. After that step, rolling windows of a particular size and time intervals between them are created. Then, a statistical analysis is prepared, e.g. mean HRV calculation using the RMSSD (Root Mean Square of Successive Differences) approach, measuring a relationship between mean HRV and PANSS scores, mobility coefficient calculation based on accelerometer data and verification of dependencies between HRV and mobility scores.
DATA DESCRIPTION
The structure of the dataset is as follows. One folder, called HRV_anonymized_data contains values of R-R intervals together with timestamps for each experiment participant. The data was properly anonymized, i.e. the day of the measurement was removed to prevent person identification. Files concerned with patients have the name treatment_X.csv, where X is the number of the person, while files related to the healthy controls are named control_Y.csv, where Y is the identification number of the person. Furthermore, for visualization purposes, an image of the raw RR intervals for each participant is presented. Its name is raw_RR_{control,treatment}_N.png, where N is the number of the person from the control/treatment group. The collected data are raw, i.e. before the anomaly removal. The code enabling reproducing the anomaly detection stage and removing suspicious heartbeats is publicly available in the repository [Książek et al]. The structure of consecutive files collecting R-R intervals is following:
Phone timestamp
RR-interval [ms]
12:43:26.538000
651
12:43:27.189000
632
12:43:27.821000
618
12:43:28.439000
621
12:43:29.060000
661
...
...
The first column contains the timestamp for which the distance between two consecutive R peaks was registered. The corresponding R-R interval is presented in the second column of the file and is expressed in milliseconds.
The second folder, called accelerometer_anonymized_data contains values of accelerometer data collected at the same time as R-R intervals. The naming convention is similar to that of the R-R interval data: treatment_X.csv and control_X.csv represent the data coming from the persons from the treatment and control group, respectively, while X is the identification number of the selected participant. The numbers are exactly the same as for R-R intervals. The structure of the files with accelerometer recordings is as follows:
Phone timestamp
X [mg]
Y [mg]
Z [mg]
13:00:17.196000
-961
-23
182
13:00:17.205000
-965
-21
181
13:00:17.215000
-966
-22
187
13:00:17.225000
-967
-26
193
13:00:17.235000
-965
-27
191
...
...
...
...
The first column contains a timestamp, while the next three columns correspond to the currently registered acceleration in three axes: X, Y and Z, in milli-g unit.
We also attached a file with the PANSS test scores (PANSS.csv) for all patients participating in the measurement. The structure of this file is as follows:
no_of_person
PANSS_P
PANSS_N
PANSS_G
PANSS_total
1
8
13
22
43
2
11
7
18
36
3
14
30
44
88
4
18
13
27
58
...
...
...
...
..
The first column contains the identification number of the patient, while the three following columns refer to the PANSS scores related to positive, negative and general symptoms, respectively.
USAGE NOTES
All the files necessary to run the HRV and/or accelerometer data analysis are available on the GitHub repository [Książek et al]. HRV data loading, preprocessing (i.e. anomaly detection and removal), as well as the calculation of mean HRV values in terms of the RMSSD, is performed in the main.py file. Also, Pearson's correlation coefficients between HRV values and PANSS scores and the statistical tests (Levene's and Mann-Whitney U tests) comparing the treatment and control groups are computed. By default, a sensitivity analysis is made, i.e. running the full pipeline for different settings of the window size for which the HRV is calculated and various time intervals between consecutive windows. Preparing the heatmaps of correlation coefficients and corresponding p-values can be done by running the utils_advanced_plots.py file after performing the sensitivity analysis. Furthermore, a detailed analysis for the one selected set of hyperparameters may be prepared (by setting sensitivity_analysis = False), i.e. for 15-minute window sizes, 1-minute time intervals between consecutive windows and without data interpolation method. Also, patients taking quetiapine may be excluded from further calculations by setting exclude_quetiapine = True because this medicine can have a strong impact on HRV [Hattori et al].
The accelerometer data processing may be performed using the utils_accelerometer.py file. In this case, accelerometer recordings are downsampled to ensure the same timestamps as for R-R intervals and, for each participant, the mobility coefficient is calculated. Then, a correlation
https://data.gov.tw/licensehttps://data.gov.tw/license
The accumulated rainfall chart with wide intervals will be updated for download starting from September 15, 2023. Please update before December 31, 2023 as the old links will expire after that. If you need to download a large amount of data, please apply for membership at the Meteorological Data Open Platform https://opendata.cwa.gov.tw/index.
This dataset provides information about the number of properties, residents, and average property values for Interval Street cross streets in Chantilly, VA.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Data Snooping (DS) is the best-established method to identify gross errors (outliers) in geodetic data analysis, with a given probability. The power of the test is the probability of DS correctly identifying a gross error, while the confidence interval is the probability of DS not rejecting an observation uncontaminated by gross error. In practice, the power of the test is always unknown. Thus, the objective of this paper is to present a theoretical review of how to determine the minimum power of the test, and bound values for the confidence interval of the DS procedure in an n-dimensional scenario, i.e., considering all observations involved. Along with the theoretical review, a numerical example involving a simulated leveling network is presented. The results obtained in the experiments agreed with the previously calculated theoretical values, i.e., the revised methodology showed satisfactory performance in practice. The example also shows the importance of the revised methodology in the planning stage (or pre-analysis) of geodetic networks
https://data.4tu.nl/info/fileadmin/user_upload/Documenten/4TU.ResearchData_Restricted_Data_2022.pdfhttps://data.4tu.nl/info/fileadmin/user_upload/Documenten/4TU.ResearchData_Restricted_Data_2022.pdf
This repository contains the input data associated with the research paper titled "The Effects of PLU Polling Interval on the Reconstructed OD Matrix: A Dutch Case Study Using a Data-Driven Method". The paper is published in the TRR. The input_data.rar includes all relevant raw data. This data has the overall purpose of understanding the effect of temporal resolution of data on the estimated origin destination matrix. The 1.experienced_plans.xml file contains the input data from MATSim with the travel diaries of users in Amsterdam. amsterdamMezuroZones.rar contains the geographical information of the data used from MATSiM . The files inside can be opened in python.
This table shows overall ATCEMS response interval performance for entire fiscal years. Data in the table is broken out by incident response priority and service area (City of Austin or Travis County).