Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Experimental data can broadly be divided in discrete or continuous data. Continuous data are obtained from measurements that are performed as a function of another quantitative variable, e.g., time, length, concentration, or wavelength. The results from these types of experiments are often used to generate plots that visualize the measured variable on a continuous, quantitative scale. To simplify state-of-the-art data visualization and annotation of data from such experiments, an open-source tool was created with R/shiny that does not require coding skills to operate it. The freely available web app accepts wide (spreadsheet) and tidy data and offers a range of options to normalize the data. The data from individual objects can be shown in 3 different ways: (1) lines with unique colors, (2) small multiples, and (3) heatmap-style display. Next to this, the mean can be displayed with a 95% confidence interval for the visual comparison of different conditions. Several color-blind-friendly palettes are available to label the data and/or statistics. The plots can be annotated with graphical features and/or text to indicate any perturbations that are relevant. All user-defined settings can be stored for reproducibility of the data visualization. The app is dubbed PlotTwist and runs locally or online: https://huygens.science.uva.nl/PlotTwist
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
We hypothesized that (1) correlation of (A) the output of instrumentation to generate quantitative continuous measurements of movements and (B) the quantitative measurements of trained examiners using structured ratings of movements would generate the tools to differentiate the movements of (A) Parkinson's disease (PD), (B) parkinsonian syndromes, and health, and (2) continuous quantitative measurements of movements would improve the ratings generated by visual observations of trained raters, and provide pathognomonic signatures to identify PD and parkinsonian syndromes.
A protocol for a low-cost quantitative continuous measurement of movements in the extremities of people with PD (McKay, et al., 2019) was administered to people with PD and multiple system atrophy-parkinsonian type (MSA-P) and age- and sex-matched healthy control participants. Data from instrumentation was saved as WinDaq files (Dataq Instruments, Inc., Akron, Ohio) and converted into Excel files (McKay, et al., 2019) using the WinDaq Waveform Data Browser (Dataq Instruments, Inc., Akron, Ohio).
Participants were asked to sit in a straight-back chair with arms approximately six inches from the wall to minimize the risk of hitting the wall. The examiner sat in a similar chair facing the participant. The examiner asked the technologist and the videographer to begin recording immediately before instructing the participant to perform each item.
Items were scored live by the examiner at the same time that the quantitative continuous measurements of movements were recorded by the instrumentation.
Healthy control participants were matched for age and sex with participants with PD. The key identifies the diagnosis (PD = Parkinson's disease, MSA-P = Multiple system atrophy - parkinsonian type, HC = healthy control, 1 = male, 0 = female). Participants with PD completed a single test session (0002, 0005, 0007-0009, 0012, 0017-0018, and 0021), a test and a retest session (0001, 0003, 0006, 0010-0011, 0013, 0015, 0019, 0022-0023), or a test and two retest sessions (0014). HC participants completed test and retest sessions (0020, 0024-0030). A participant with MSA-P (0004) completed a test session. Individual files for the WinDaq, Excel, and coding forms for each testing are entered in the dataset. The Excel files for the five repetitive items were converted to fast Fourier transforms (FFTs) and continuous wavelet transforms (CWTs) (MatLab).
None of the files underwent filtering.
Healthy participants exhibited some of the features of disease.
The data provide the basis to determine how a session may predict future performance.
Facebook
TwitterThe recent surge in enthusiasm for simultaneously inferring relationships from extinct and extant species has reinvigorated interest in statistical approaches for modelling morphological evolution. Current statistical methods use the Mk model to describe substitutions between discrete character states. Although representing a significant step forward, the Mk model presents challenges in biological interpretation, and its adequacy in modelling morphological evolution has not been well explored. Another major hurdle in morphological phylogenetics concerns the process of character coding of discrete characters. The often subjective nature of discrete character coding can generate discordant results that are rooted in individual researchers' subjective interpretations. Employing continuous measurements to infer phylogenies may alleviate some of these issues. Although not widely used in the inference of topology, models describing the evolution of continuous characters have been well examine...
Facebook
Twitterhttps://spdx.org/licenses/CC0-1.0.htmlhttps://spdx.org/licenses/CC0-1.0.html
Coevolution is relentlessly creating and maintaining biodiversity, and therefore has been a central topic in evolutionary biology. Previous theoretical studies have mostly considered coevolution between genetically symmetric traits (i.e., coevolution between two continuous quantitative traits or two discrete Mendelian traits). However, recent empirical evidence indicates that coevolution can occur between genetically asymmetric traits (e.g., between quantitative and Mendelian traits). We examine consequences of antagonistic coevolution mediated by a quantitative predator trait and a Mendelian prey trait, such that predation is more intense with decreased phenotypic distance between their traits (phenotype matching). This antagonistic coevolution produces a complex pattern of bifurcations with bistability (initial state dependence) in a two-dimensional model for trait coevolution. Further, with eco-evolutionary dynamics (so that the trait evolution affects predator-prey population dynamics), we find that coevolution can cause rich dynamics including anti-phase cycles, in-phase cycles, chaotic dynamics, and deterministic predator extinction. Predator extinction is more likely to occur when the prey trait exhibits complete dominance rather than semidominance and when the predator trait evolves very rapidly. Our study illustrates how recognizing the genetic architectures of interacting ecological traits can be essential for understanding the population and evolutionary dynamics of coevolving species.
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
This data is based on the Seshat data release in https://zenodo.org/record/6642230 and aims to dissect the time series of each NGA into culturally and institutionally continuous time series. For both continuity criteria, the central continuous time series is marked in the data (central meaning that this is the time interval during which the NGA has crossed a specified threshold between low-complexity and high-complexity societies). Details can be found in v3 of https://arxiv.org/abs/2212.00563
Facebook
TwitterMany aspects of morphological phylogenetics are controversial in the theoretical systematics literature and yet are often poorly explained and justified in empirical studies. In this paper, I argue that most morphological characters describe variation that is fundamentally quantitative, regardless of whether it is coded qualitatively or quantitatively by systematists. Given this view, three fundamental problems in morphological character analysis (character state definition, delimitation, and ordering) may have a common solution: coding morphological characters as continuous quantitative traits. A new parsimony method (step-matrix gap-weighting, a modification of Thiele's approach) is proposed that allows quantitative traits to be analyzed as continuous variables. The problem of scaling or weighting quantitative characters relative to qualitative characters (and to each other) is reviewed, and three possible solutions are described. The new coding method is applied to data from hoplocer...
Facebook
Twitterhttps://dataintelo.com/privacy-and-policyhttps://dataintelo.com/privacy-and-policy
According to our latest research, the global Quantitative Research AI market size reached USD 1.82 billion in 2024, reflecting robust expansion in the adoption of artificial intelligence for quantitative analysis across industries. The market is expected to grow at a CAGR of 27.6% during the forecast period, with the forecasted market size anticipated to reach USD 16.34 billion by 2033. This significant growth is driven by the increasing demand for advanced data analytics, automation in research processes, and the expanding scope of AI technologies in both academic and commercial quantitative research.
The primary growth driver for the Quantitative Research AI market is the surging volume of data generated across industries and the need for sophisticated tools to extract actionable insights. Organizations are increasingly leveraging AI-powered quantitative research tools to process large datasets efficiently, identify patterns, and predict future trends with higher accuracy. These capabilities are particularly valuable in sectors such as financial services, healthcare, and market research, where data-driven decision-making is critical. The integration of machine learning algorithms and natural language processing further enhances the ability of AI systems to handle complex quantitative tasks, reducing the time and resources required for traditional research methodologies.
Another significant factor contributing to market growth is the rising adoption of cloud-based AI solutions. Cloud deployment offers scalability, flexibility, and cost-effectiveness, enabling organizations of all sizes to access advanced quantitative research tools without the need for substantial upfront investments in infrastructure. The proliferation of AI-as-a-Service (AIaaS) models has democratized access to powerful quantitative research capabilities, allowing even small and medium enterprises (SMEs) to benefit from AI-driven insights. Additionally, continuous advancements in AI hardware, such as specialized processors and accelerators, are further propelling the market by improving the performance and efficiency of AI applications in quantitative research.
The increasing focus on personalized and precision-driven research in industries such as healthcare and finance is also fueling the demand for AI-based quantitative research solutions. In healthcare, for instance, AI-driven quantitative analysis is transforming clinical trials, epidemiological studies, and patient data management, leading to more accurate diagnoses and treatment plans. Similarly, financial institutions are leveraging AI for quantitative trading, risk assessment, and fraud detection. The growing recognition of AI's potential to enhance research accuracy, reduce human error, and accelerate discovery is prompting organizations to invest heavily in quantitative research AI technologies.
From a regional perspective, North America currently dominates the Quantitative Research AI market, accounting for the largest share in 2024, followed closely by Europe and Asia Pacific. The presence of leading technology providers, a mature research ecosystem, and substantial investments in AI R&D are key factors supporting market growth in these regions. Asia Pacific is expected to witness the fastest CAGR during the forecast period, driven by rapid digital transformation, increasing government initiatives to promote AI adoption, and the emergence of innovative startups. Meanwhile, Latin America and the Middle East & Africa are gradually catching up, supported by growing awareness and investments in AI-powered research solutions. These regional dynamics underscore the global nature of the market and the diverse opportunities for growth across different geographies.
The Component segment of the Quantitative Research AI market is broadly categorized into Software, Hardware, and Services, each playing a vital role in the overall ecosystem. Software represents the largest share of the market, as AI-driven quantitative research platforms and analytics tools are fundamental to the digital transformation of research methodologies. These software solutions encompass machine learning frameworks, data visualization tools, statistical analysis packages, and specialized AI algorithms tailored for quantitative research. The continuous evolution of AI software, coupled with advancements i
Facebook
TwitterThe MOD44B Version 6.1 Vegetation Continuous Fields (VCF) yearly product is a global representation of surface vegetation cover as gradations of three ground cover components: percent tree cover, percent non-tree cover, and percent non-vegetated (bare). VCF products provide a continuous, quantitative portrayal of land surface cover at 250 meter (m) pixel resolution, with a sub-pixel depiction of percent cover in reference to the three ground cover components. The sub-pixel mixture of ground cover estimates represents a revolutionary approach to the characterization of vegetative land cover that can be used to enhance inputs to environmental modeling and monitoring applications. The MOD44B data product layers include percent tree cover, percent non-tree cover, percent non-vegetated, cloud cover, and quality indicators. The start date of the annual period for this product begins with day of year (DOY) 65 (March 6 except for leap year which corresponds to March 5). Known Issues For complete information about known issues please refer to the MODIS/VIIRS Land Quality Assessment website.Improvements/Changes from Previous Versions The Version 6.1 Level-1B (L1B) products have been improved by undergoing various calibration changes that include: changes to the response-versus-scan angle (RVS) approach that affects reflectance bands for Aqua and Terra MODIS, corrections to adjust for the optical crosstalk in Terra MODIS infrared (IR) bands, and corrections to the Terra MODIS forward look-up table (LUT) update for the period 2012 - 2017.* A polarization correction has been applied to the L1B Reflective Solar Bands (RSB).
Facebook
TwitterThe MOD44B Version 6 data product was decommissioned on July 31, 2023. Users are encouraged to use the MOD44B Version 6.1 data product.The MOD44B Version 6 Vegetation Continuous Fields (VCF) yearly product is a global representation of surface vegetation cover as gradations of three ground cover components: percent tree cover, percent non-tree cover, and percent non-vegetated (bare). VCF products provide a continuous, quantitative portrayal of land surface cover at 250 meter (m) pixel resolution, with a sub-pixel depiction of percent cover in reference to the three ground cover components. The sub-pixel mixture of ground cover estimates represents a revolutionary approach to the characterization of vegetative land cover that can be used to enhance inputs to environmental modeling and monitoring applications. The MOD44B data product layers include percent tree cover, percent non-tree cover, percent non-vegetated, cloud cover, and quality indicators. The start date of the annual period for this product begins with day of year (DOY) 65 (March 6 except for leap year which corresponds to March 5). Known Issues For complete information about known issues please refer to the MODIS/VIIRS Land Quality Assessment website.Improvements/Changes from Previous Versions The MOD44B Version 6 VCF was produced with the same code and training as the Version 5 products, but improvements to the upstream inputs result in more accurate VCF products.
Facebook
Twitterhttps://www.shibatadb.com/license/data/proprietary/v1.0/license.txthttps://www.shibatadb.com/license/data/proprietary/v1.0/license.txt
Yearly citation counts for the publication titled "357P: Association of TROP2 quantitative continuous scoring (QCS) normalised membrane ratio (NMR) with efficacy in Chinese patients (pts) with advanced/metastatic non-small cell lung cancer (a/mNSCLC) treated with datopotamab deruxtecan (Dato-DXd) in TROPION-PanTumor02".
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
The dataset and replication package of the study "A continuous open source data collection platform for architectural technical debt assessment".
Abstract
Architectural decisions are the most important source of technical debt. In recent years, researchers spent an increasing amount of effort investigating this specific category of technical debt, with quantitative methods, and in particular static analysis, being the most common approach to investigate such a topic.
However, quantitative studies are susceptible, to varying degrees, to external validity threats, which hinder the generalisation of their findings.
In response to this concern, researchers strive to expand the scope of their study by incorporating a larger number of projects into their analyses. This practice is typically executed on a case-by-case basis, necessitating substantial data collection efforts that have to be repeated for each new study.
To address this issue, this paper presents our initial attempt at tackling this problem and enabling researchers to study architectural smells at large scale, a well-known indicator of architectural technical debt. Specifically, we introduce a novel approach to data collection pipeline that leverages Apache Airflow to continuously generate up-to-date, large-scale datasets using Arcan, a tool for architectural smells detection (or any other tool).
Finally, we present the publicly-available dataset resulting from the first three months of execution of the pipeline, that includes over 30,000 analysed commits and releases from over 10,000 open source GitHub projects written in 5 different programming languages and amounting to over a billion of lines of code analysed.
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
An R script which provides a user made function to implement the Q-Gen methodology. The original QuSAGE package is all that is needed. Example code of how to implement the function is provided as well. (TXT 2.61 KB)
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Dataset of tree (.tre) files and R code for running generalized Robinson-Foulds distance (Smith, 2020a;b) analysis.
The .tre files can be read into R (R Core Team., 2023) using the ape::read.tree function (Paradis et al., 2003), full details in R code file.
Paradis, E., Claude, J., & Strimmer, K. (2004). APE: analyses of phylogenetics and evolution in R language. Bioinformatics, 20(2), 289-290.
R Core Team. (2023). R: A Language and Environment for Statistical Computing. (Version 4.2.2). R Foundation for Statistical Computing, Vienna, Austria: https://www.R-project.org/.
Smith, M. R. (2020a). Information theoretic generalized Robinson–Foulds metrics for comparing phylogenetic trees. Bioinformatics, 36(20), 5007-5013. https://doi.org/10.1093/bioinformatics/btaa614
Smith, M. R. (2020b). TreeDist: distances between phylogenetic trees. R package version 2.7.0. doi:10.5281/zenodo.3528124.
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Datasets of Experiments 1 and 2
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
"A protocol for a low-cost quantitative continuous measurement of movements in the extremities of people with PD (McKay, et al., 2019) was administered to people with PD . . . and age- and sex-matched healthy control participants" (Harrigan, et al., Quantitative continuous measurement of movements in the extremities, 2020). "Healthy control participants were matched for age and sex with participants with PD. Participants with PD completed a single test session . . . , a test and a retest session . . . , or a test and two retest sessions . . . . HC participants completed test and retest sessions " (Harrigan, et al., Quantitative continuous measurement of movements in the extremities, 2020). Thirty-two trained raters who were certified in the Movement Disorder Society-Sponsored Revision of the Unified Parkinson’s Disease Rating Scale (MDS-UPDRS) (Goetz, et al., 2008) were presented the output of the ten participants with PD who completed a single test session (Pilot Test and Retest). Raters were presented two sets of 40 quizzes containing five representations for scoring of (A) output signals and FFTs and (B) CWTs (Pilot Test and Retest). Each quiz contained the panels of the x, y, and z representations of the finger and wrist or the toe and ankle of the five repetitive tasks. Each panel to be scored included six images corresponding to the signals of the three dimensions of the two accelerometers on a single extremity. The laterality of the representations was not stated. Raters were presented five sets of six images of the original signal and the fast Fourier transform (FFT) or the continuous wavelet transforms (CWTs). Raters were presented either five panels of output signals and FFTs or CWTs. Panels did not include output signals and FFTs and CWTs simultaneously. Raters were instructed to score (A) output signals and FFTs and (B) CWTs analogously to the clinical coding forms as indicated the the instructions in the data. The raters also completed the output of the ten participants with PD and eight HCs who completed a two test session (CWT Test and Retest). Raters were presented two sets of 72 quizzes containing five representations for scoring of (CWTs (Pilot Test and Retest). Each quiz contained the panels of averaged signals of the x, y, and z representations of the finger and wrist or the toe and ankle of the five repetitive tasks. Each panel to be scored included two images corresponding to the signals of the three dimensions of the two accelerometers on a single extremity. The laterality of the representations was not stated. Raters were asked to complete ratings independently at convenient times during the week.
Facebook
TwitterAttribution-NonCommercial 4.0 (CC BY-NC 4.0)https://creativecommons.org/licenses/by-nc/4.0/
License information was derived automatically
We present a comprehensive and updated Python-based open software to calculate continuous symmetry measures (CSMs) and their related continuous chirality measure (CCM) of molecules across chemistry. These descriptors are used to quantify distortion levels of molecular structures on a continuous scale and were proven insightful in numerous studies. The input information includes the coordinates of the molecular geometry and a desired cyclic symmetry point group (i.e., Cs, Ci, Cn, or Sn). The results include the coordinates of the nearest symmetric structure that belong to the desired symmetry point group, the permutation that defines the symmetry operation, the direction of the symmetry element in space, and a number, between zero and 100, representing the level of symmetry or chirality. Rather than treating symmetry as a binary property by which a structure is either symmetric or asymmetric, the CSM approach quantifies the level of gray between black and white and allows one to follow the course of change. The software can be downloaded from https://github.com/continuous-symmetry-measure/csm or used online at https://csm.ouproj.org.il.
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Research, conservation, and effective natural resource management often depend on maps that characterize vegetation patterns. Quantitative and ecologically specific representations of vegetation pattern better represent observed vegetation patterns than do traditional categorical vegetation maps. They also avoid a human interpretational bias not necessarily shared by or important to plants or wildlife. We developed quantitative continuous foliar cover maps for 16 plant and lichen species or ecologically narrow aggregates in arctic and boreal Alaska and adjacent Yukon (North American Beringia). To provide context to the performance of our continuous foliar cover maps, we compared our results to the performances of three categorical vegetation maps that cover arctic and boreal Alaska: the National Land Cover Database, the coarse classes of the Alaska Vegetation and Wetland Composite, and the fine classes of the Alaska Vegetation and Wetland Composite. We integrated new and existing ground and aerial vegetation observations for arctic and boreal Alaska from three vegetation plots databases. To map patterns of foliar cover, we statistically associated observations of vegetation foliar cover with environmental, multi-season spectral, and surface texture covariates using a Bayesian statistical learning approach. Our maps predicted 33% to 67% of the observed variation in foliar cover per map class at a 10 × 10 m resolution, although the accuracy of each map varied between the Arctic, Southwest, and Interior subregions. We show that while some maps have high noise at the 10 × 10 m resolution, they generally capture vegetation patterns accurately at local to landscape scales. All continuous foliar cover maps performed substantially better than the categorical vegetation maps both for the entire region and for all subregions. The vegetation database and scripted workflow that we developed to create the continuous foliar cover maps will allow consistent future annual or semi-annual updates to include new observations of vegetation patterns and new covariate data. Our scripted workflow will also allow the application of our methods to different areas. Our continuous foliar cover maps extend knowledge of the functional role of vegetation in communities and wildlife habitat in North American Beringia beyond what has been previously available in categorical vegetation maps or quantitative maps of broadly defined vegetation aggregates.
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Mean ± SE of qualitative descriptors continuously recorded through Temporal Dominant Behavioural Expression (duration and score) and quantitative variables continuously recorded through The Observer XT of juvenile goats either reintroduced into their group (group C) or introduced into group C while the mothers were at pasture (group WOM) or introduced into group C while the mothers were in the home pen (group WM).
Facebook
TwitterAs our generation and collection of quantitative digital data increase, so do our ambitions for extracting new insights and knowledge from those data. In recent years, those ambitions have manifested themselves in so-called “Grand Challenge” projects coordinated by academic institutions. These projects are often broadly interdisciplinary and attempt to address to major issues facing the world in the present and the future through the collection and integration of diverse types of scientific data. In general, however, disciplines that focus on the past are underrepresented in this environment – in part because these grand challenges tend to look forward rather than back, and in part because historical disciplines tend to produce qualitative, incomplete data that are difficult to mesh with the more continuous quantitative data sets provided by scientific observation. Yet historical information is essential for our understanding of long-term processes, and should thus be incorporated into our efforts to solve present and future problems. Archaeology, an inherently interdisciplinary field of knowledge that bridges the gap between the quantitative and the qualitative, can act as a connector between the study of the past and data-driven attempts to address the challenges of the future. To do so, however, we must find new ways to integrate the results of archaeological research into the digital platforms used for the modeling and analysis of much bigger data.
Planet Texas 2050 is a grand challenge project recently launched by The University of Texas at Austin. Its central goal is to understand the dynamic interactions between water supply, urbanization, energy use, and ecosystems services in Texas, a state that will be especially affected by climate change and population mobility by the middle of the 21st century. Like many such projects, one of the products of Planet Texas 2050 will be an integrated data platform that will make it possible to model various scenarios and help decision-makers project the results of resent policies or trends into the future. Unlike other such projects, however, PT2050 incorporates data collected from past societies, primarily through archaeological inquiry. We are currently designing a data integration and modeling platform that will allow us to bring together quantitative sensor data related to the present environment with “fuzzier” data collected in the course of research in the social sciences and humanities. Digital archaeological data, from LiDAR surveys to genomic information to excavation documentation, will be a central component of this platform. In this paper, I discuss the conceptual integration between scientific “big data” and “medium-sized” archaeological data in PT2050; the process that we are following to catalog data types, identify domain-specific ontologies, and understand the points of intersection between heterogeneous data sets of varying resolution and precision as we construct the data platform; and how we propose to incorporate digital data from archaeological research into integrated modeling and simulation modules.
Facebook
TwitterCC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
License information was derived automatically
Objectives: In quantitative research, understanding basic parameters of the study population is key for interpretation of the results. As a result, it is typical for the first table ("Table 1") of a research paper to include summary statistics for the study data. Our objectives are 2-fold. First, we seek to provide a simple, reproducible method for providing summary statistics for research papers in the Python programming language. Second, we seek to use the package to improve the quality of summary statistics reported in research papers.
Materials and Methods: The tableone package is developed following good practice guidelines for scientific computing and all code is made available under a permissive MIT License. A testing framework runs on a continuous integration server, helping to maintain code stability. Issues are tracked openly and public contributions are encouraged.
Results: The tableone software package automatically compiles summary statistics into publishable formats such as CSV, HTML, and LaTeX. An executable Jupyter Notebook demonstrates application of the package to a subset of data from the MIMIC-III database. Tests such as Tukey's rule for outlier detection and Hartigan's Dip Test for modality are computed to highlight potential issues in summarizing the data.
Discussion and Conclusion: We present open source software for researchers to facilitate carrying out reproducible studies in Python, an increasingly popular language in scientific research. The toolkit is intended to mature over time with community feedback and input. Development of a common tool for summarizing data may help to promote good practice when used as a supplement to existing guidelines and recommendations. We encourage use of tableone alongside other methods of descriptive statistics and, in particular, visualization to ensure appropriate data handling. We also suggest seeking guidance from a statistician when using tableone for a research study, especially prior to submitting the study for publication.
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Experimental data can broadly be divided in discrete or continuous data. Continuous data are obtained from measurements that are performed as a function of another quantitative variable, e.g., time, length, concentration, or wavelength. The results from these types of experiments are often used to generate plots that visualize the measured variable on a continuous, quantitative scale. To simplify state-of-the-art data visualization and annotation of data from such experiments, an open-source tool was created with R/shiny that does not require coding skills to operate it. The freely available web app accepts wide (spreadsheet) and tidy data and offers a range of options to normalize the data. The data from individual objects can be shown in 3 different ways: (1) lines with unique colors, (2) small multiples, and (3) heatmap-style display. Next to this, the mean can be displayed with a 95% confidence interval for the visual comparison of different conditions. Several color-blind-friendly palettes are available to label the data and/or statistics. The plots can be annotated with graphical features and/or text to indicate any perturbations that are relevant. All user-defined settings can be stored for reproducibility of the data visualization. The app is dubbed PlotTwist and runs locally or online: https://huygens.science.uva.nl/PlotTwist