Facebook
TwitterUnder the new quarterly data summary (QDS) framework departments’ spending data is published every quarter; to show the taxpayer how the government is spending their money.
The QDS grew out of commitments made in the 2011 Budget and the written ministerial statement on business plans. For the financial year 2012 to 2013 the QDS has been revised and improved in line with action 9 of the Civil Service Reform Plan to provide a common set of data that will enable comparisons of operational performance across government so that departments and individuals can be held to account.
The QDS breaks down the total spend of the department in 3 ways:
The QDS template is the same for all departments, though the individual detail of grants and policy will differ from department to department. In using this data:
Please note that the quarter 1 2012 to 2013 return for the Department of Transport (DfT) is for the core department only.
Quarterly data summaries for April 2012 are as follows:
Quarterly data summaries for January 2012 are as follows:
Quarterly data summaries for October 2011 are as follows:
Facebook
TwitterOn 26 January 2012 the Government published the Business Plan Quarterly Data Summaries.
This was published alongside Quarterly Data Summaries from departments across Whitehall to provide a quarterly snapshot on how each department is spending its budget, the results it has achieved and how it is deploying its workforce. The data summary template is the same for all departments, though many of the individual indicators are unique to the department (especially input and impact indicators).
The Quarterly Data Summaries follow commitments made at Budget 2011 and the Written Ministerial Statement on Business Plans. Their primary purpose is to make more of the management information currently held by government available to members of the public on a regular basis.
Before using this data people should ensure they take full note of the caveats noted in each Department’s measurement annex. Please note that this data should not be used to make direct comparisons between departments for several reasons. This is because each department is unique and it does not make sense to compare some measures across all departments. Also, many of the measures are not directly comparable because they do not have common definitions, time periods, or data collection processes.
We intend for an annual version of this information to be formally laid in Parliament in the Annual Report and Accounts from July 2012.
The information is presented in excel format to allow the information to be presented in a re-usable format.
Please note: The Quarterly Data Summary was updated in June 2012 to include HCA and PINs data.
Facebook
TwitterOpen Government Licence 3.0http://www.nationalarchives.gov.uk/doc/open-government-licence/version/3/
License information was derived automatically
The Government has published the Business Plan Quarterly Data Summaries (QDS) The QDS are designed to fit on a single page to provide a quarterly snapshot on how each department is spending its budget, the results it has achieved and how it is deploying its workforce. The QDS follows commitments made at Budget 2011 and the Written Ministerial Statement on Business Plans. Their primary purpose is to make more of the management information currently held by government available to members of the public on a regular basis. This information is not audited and the quality and accuracy of the data needs to dramatically improve. However, over time with improvements in data quality and timeliness the public will be able to judge the performance of each department in a meaningful and understandable manner. We intend for an annual version of this information to be formally laid in Parliament in the Annual Report and Accounts for July 2011/12 onwards. The information is presented in excel format to allow the information to be presented in a re-usable format. The QDS template is the same for all departments, though many of the individual indicators are unique to the department (especially input and impact indicators). This is the first time Government has published this kind of information, and while this is a good start, there is room for improvement. Before using this data people should ensure they take full note of the caveats noted in each Department’s measurement annex and treat with necessary caution. At the moment, people should not be using this data to make direct comparisons between departments for several reasons. Firstly, the business of each department is unique and it does not make sense to compare some measures across all departments. Secondly, many of the measures are not directly comparable because they do not have common definitions, time periods, or data collection processes. We will be updating regularly the QDS each quarter with the next publication following in October 2011.
Facebook
Twitterhttps://digital.nhs.uk/about-nhs-digital/terms-and-conditionshttps://digital.nhs.uk/about-nhs-digital/terms-and-conditions
This is the latest statistical publication of linked HES (Hospital Episode Statistics) and DID (Diagnostic Imaging Dataset) data held by the Health and Social Care Information Centre. The HES-DID linkage provides the ability to undertake national (within England) analysis along acute patient pathways to understand typical imaging requirements for given procedures, and/or the outcomes after particular imaging has been undertaken, thereby enabling a much deeper understanding of outcomes of imaging and to allow assessment of variation in practice. This publication aims to highlight to users the availability of this updated linkage and provide users of the data with some standard information to assess their analysis approach against. The two data sets have been linked using specific patient identifiers collected in HES and DID. The linkage allows the data sets to be linked from April 2012 when the DID data was first collected; however this report focuses on patients who were present in either data set for the period April 2015-February 2016 only. For DID this is provisional 2015/16 data. For HES this is provisional 2015/16 data. The linkage used for this publication was created on 06 June 2016 and released together with this publication on 07 July 2016.
Facebook
TwitterThis technical report documents the acquisition of source data, and calculation of land cover summary statistics datasets for ten National Park Service National Capital Region park units and three custom areas of analysis: Catoctin Mountain Park (CATO), Chesapeake & Ohio Canal National Historical Park (CHOH), George Washington Memorial Parkway (GWMP), Harpers Ferry National Historical Park (HAFE), Manassas National Battlefield Park (MANA), Monocacy National Battlefield (MONO), National Capital Parks - East (NACE), Prince William Forest Park (PRWI), Rock Creek Park (ROCR), Wolf Trap National Park for the Performing Arts (WOTR), and the three custom areas of analysis - National Capital Parks - East: Oxon Cove Park, Oxon Hill Farm, Piscataway Park (NCRN_NACE_OXHI_PISC), National Capital Parks - East: Greenbelt Park and Baltimore-Washington Parkway (NCRN_NACE_GREE_BAWA), and National Capital Parks - East: DC and Suitland Parkway (NCRN_NACE_DC_SUIT). The source data and land cover calculations are available for use within the National Park Service (NPS) Inventory and Monitoring Program. Land cover summary statistics datasets can be calculated for all geographic regions within the extent of the NPS; this report includes statistics calculated for the conterminous United States. The land cover summary statistics datasets are calculated from multiple sources, including Multi-Resolution Land Characteristics Consortium products in the National Land Cover Database (NLCD) and United States Geological Survey’s (USGS) Earth Resources Observation and Science (EROS) Center products in the Land Change Monitoring, Assessment, and Projection (LCMAP) raster dataset. These summary statistics calculate land cover at up to three classification scales: Level 1, modified Anderson Level 2, and Natural versus Converted land cover. The output land cover summary statistics datasets produced here for the ten National Capital Region park units and three custom areas of analysis utilize the most recent versions of the source datasets (NLCD and LCMAP). These land cover summary statistics datasets are used in the NPS Inventory and Monitoring Program, including the NPS Environmental Settings Monitoring Protocol and may be used by networks and parks for additional efforts.
Facebook
TwitterThese tabular data are the summarization of human related variables within the Chesapeake Bay watershed using the xstrm methodology bringing these data to the 1:24,000 scale. Variables being counted as human related include agriculture, barriers, road density and road/stream crossing data. Outputs include tabular comma-separated values files (CSVs), and parquet formatted files for both the local catchment and network summaries linked to the National Hydrography Dataset Plus High-Resolution (NHDPlus HR) catchments by NHDPlus ID. Local catchments are defined as the single catchment the data is summarized within. Network accumulation summaries were completed for each of the local catchments and their network connected upstream and/or downstream catchments. The summarized data tables are structured as a single column representing the catchment id values (ie. nhdplusid) and the remaining columns consisting of the summarized variables. Additionally, for a full description of the variables included within these summaries see xstrm_nhdhr_human_chesapeake_baywide_datadictionary.csv in the attached files. The xstrm local summary methodology takes either raster or point data as input then summarizes those data by "zones" (hereafter referred to as catchment(s)), in this case the NHDPlus HR catchments. The network summaries then take the results from the local summaries and calculates the desired network summary statistic for the local catchment and its respective upstream or downstream catchments. As a note concerning use of these data, any rasters summarized within this process only had their cells included within a catchment if the center of the raster cell fell within the catchment boundary. However, the resolution of the input raster data for these summaries was considered to provide completely adequate coverage of the summary catchments using this option and given computing power limitations. If a confirmed complete coverage of a catchment is desired (even if a raster cell only is minimally included within the catchment) then it is recommended to rerun the xstrm summary process with the "All Touched" option set to "True". These data were updated in September of 2024 where several variables unnecessary to the use of the data summaries were removed, incorrectly calculated area variables and all dependent variables were corrected, and several new variables were added to the dataset. Further information on the Xstrm summary process can be found at the Xstrm software release pages: Xstrm: Wieferich, D.J., Williams, B., Falgout, J.T., Foks, N.L. 2021. xstrm. U.S. Geological Survey software release. https://doi.org/10.5066/P9P8P7Z0. Xstrm Local: Wieferich, D.J., Gressler B., Krause K., Wieczorek M., McDonald, S. 2022. xstrm_local Version-1.1.0. U.S. Geological Survey software release. https://doi.org/10.5066/P98BOGI9.
Facebook
TwitterThis is a dataset downloaded off excelbianalytics.com created off of random VBA logic. I recently performed an extensive exploratory data analysis on it and I included new columns to it, namely: Unit margin, Order year, Order month, Order weekday and Order_Ship_Days which I think can help with analysis on the data. I shared it because I thought it was a great dataset to practice analytical processes on for newbies like myself.
Facebook
TwitterBusiness plan indicators are initial management information to provide an indication of HMRC's performance, and are therefore subject to revision and audit. Final performance figures will be made available when the annual audited accounts and departmental report are published. The quarterly data summary (QDS) is designed to fit on a single page to provide a quarterly snapshot on how each department is spending its budget, the results it has achieved and how it is deploying its workforce. Before using this data people should ensure they take full note of the caveats noted in the measurement annex and treat it with necessary caution. At the moment, people should not be using this QDS data to make direct comparisons between departments for several reasons. Firstly, the business of each department is unique and it does not make sense to compare some measures across all departments. Secondly, many of the measures are not directly comparable because they do not have common definitions, time periods, or data collection processes. The link for Quarter 3, 2023 to 2024 has been updated to reflect Digital Assistant interactions correction. UPDATE : From Q1 2013/14, the Cabinet Office now publish this QDS data on the Government Interrogation Spending Tool (GIST) website http://www.gist.cabinetoffice.gov.uk/.
Facebook
TwitterThe data is published using the Summary Data Template, Version 1.1 as of 05 March 2015.
According to the EITI Standard 5.3.b: "Summary data from each EITI Report should be submitted electronically to the International Secretariat according to the standardised reporting format provided by the International Secretariat" This template should be completed in full and submitted by email by the national secretariat to the International EITI Secretariat following the publication of the report. The data will be used to populate the global EITI data repository, available on the international EITI website.
NB: The data available on ResourceData is republished from the EITI API and covers one section of the Summary Data, Part 3 which is comprised of data on government revenues per revenue stream and company.
Notes for consideration:
Disclaimer: The EITI Secretariat advice that users consult the original reports for detailed information. Where figures are not available in US dollars, the annual average exchange rate is used. Any questions regarding the data collection and Summary Data methodology can be directed to the EITI Secretariat: data@eiti.org or by visiting eiti.org/summary-data.
Facebook
TwitterWelcome to the Cyclistic bike-share analysis case study! In this case study, you will perform many real-world tasks of a junior data analyst. You will work for a fictional company, Cyclistic, and meet different characters and team members. In order to answer the key business questions, you will follow the steps of the data analysis process: ask, prepare, process, analyze, share, and act. Along the way, the Case Study Roadmap tables — including guiding questions and key tasks — will help you stay on the right path. By the end of this lesson, you will have a portfolio-ready case study. Download the packet and reference the details of this case study anytime. Then, when you begin your job hunt, your case study will be a tangible way to demonstrate your knowledge and skills to potential employers.
You are a junior data analyst working in the marketing analyst team at Cyclistic, a bike-share company in Chicago. The director of marketing believes the company’s future success depends on maximizing the number of annual memberships. Therefore, your team wants to understand how casual riders and annual members use Cyclistic bikes differently. From these insights, your team will design a new marketing strategy to convert casual riders into annual members. But first, Cyclistic executives must approve your recommendations, so they must be backed up with compelling data insights and professional data visualizations. Characters and teams ● Cyclistic: A bike-share program that features more than 5,800 bicycles and 600 docking stations. Cyclistic sets itself apart by also offering reclining bikes, hand tricycles, and cargo bikes, making bike-share more inclusive to people with disabilities and riders who can’t use a standard two-wheeled bike. The majority of riders opt for traditional bikes; about 8% of riders use the assistive options. Cyclistic users are more likely to ride for leisure, but about 30% use them to commute to work each day. ● Lily Moreno: The director of marketing and your manager. Moreno is responsible for the development of campaigns and initiatives to promote the bike-share program. These may include email, social media, and other channels. ● Cyclistic marketing analytics team: A team of data analysts who are responsible for collecting, analyzing, and reporting data that helps guide Cyclistic marketing strategy. You joined this team six months ago and have been busy learning about Cyclistic’s mission and business goals — as well as how you, as a junior data analyst, can help Cyclistic achieve them. ● Cyclistic executive team: The notoriously detail-oriented executive team will decide whether to approve the recommended marketing program.
In 2016, Cyclistic launched a successful bike-share offering. Since then, the program has grown to a fleet of 5,824 bicycles that are geotracked and locked into a network of 692 stations across Chicago. The bikes can be unlocked from one station and returned to any other station in the system anytime. Until now, Cyclistic’s marketing strategy relied on building general awareness and appealing to broad consumer segments. One approach that helped make these things possible was the flexibility of its pricing plans: single-ride passes, full-day passes, and annual memberships. Customers who purchase single-ride or full-day passes are referred to as casual riders. Customers who purchase annual memberships are Cyclistic members. Cyclistic’s finance analysts have concluded that annual members are much more profitable than casual riders. Although the pricing flexibility helps Cyclistic attract more customers, Moreno believes that maximizing the number of annual members will be key to future growth. Rather than creating a marketing campaign that targets all-new customers, Moreno believes there is a very good chance to convert casual riders into members. She notes that casual riders are already aware of the Cyclistic program and have chosen Cyclistic for their mobility needs. Moreno has set a clear goal: Design marketing strategies aimed at converting casual riders into annual members. In order to do that, however, the marketing analyst team needs to better understand how annual members and casual riders differ, why casual riders would buy a membership, and how digital media could affect their marketing tactics. Moreno and her team are interested in analyzing the Cyclistic historical bike trip data to identify trends
How do annual members and casual riders use Cyclistic bikes differently? Why would casual riders buy Cyclistic annual memberships? How can Cyclistic use digital media to influence casual riders to become members? Moreno has assigned you the first question to answer: How do annual members and casual rid...
Facebook
TwitterU.S. Government Workshttps://www.usa.gov/government-works
License information was derived automatically
October 31, 2025 (Final DWR Data)
The 2018 Legislation required DWR to provide or otherwise identify data regarding the unique local conditions to support the calculation of an urban water use objective (CWC 10609. (b)(2) (C)). The urban water use objective (UWUO) is an estimate of aggregate efficient water use for the previous year based on adopted water use efficiency standards and local service area characteristics for that year.
UWUO is calculated as the sum of efficient indoor residential water use, efficient outdoor residential water use, efficient outdoor irrigation of landscape areas with dedicated irrigation meter for Commercial, Industrial, and Institutional (CII) water use, efficient water losses, and an estimated water use in accordance with variances, as appropriate. Details of urban water use objective calculations can be obtained from DWR’s Recommendations for Guidelines and Methodologies document (Recommendations for Guidelines and Methodologies for Calculating Urban Water Use Objective - https://water.ca.gov/-/media/DWR-Website/Web-Pages/Programs/Water-Use-And-Efficiency/2018-Water-Conservation-Legislation/Performance-Measures/UWUO_GM_WUES-DWR-2021-01B_COMPLETE.pdf).
The datasets provided in the links below enable urban retail water suppliers calculate efficient outdoor water uses (both residential and CII), agricultural variances, variances for significant uses of water for dust control for horse corals, and temporary provisions for water use for existing pools (as stated in Water Boards’ draft regulation). DWR will provide technical assistance for estimating the remaining UWUO components, as needed. Data for calculating outdoor water uses include:
• Reference evapotranspiration (ETo) – ETo is evaporation plant and soil surface plus transpiration through the leaves of standardized grass surfaces over which weather stations stand. Standardization of the surfaces is required because evapotranspiration (ET) depends on combinations of several factors, making it impractical to take measurements under all sets of conditions. Plant factors, known as crop coefficients (Kc) or landscape coefficients (KL), are used to convert ETo to actual water use by specific crop/plant. The ETo data that DWR provides to urban retail water suppliers for urban water use objective calculation purposes is derived from the California Irrigation Management Information System (CIMIS) program (https://cimis.water.ca.gov/). CIMIS is a network of over 150 automated weather stations throughout the state that measure weather data that are used to estimate ETo. CIMIS also provides daily maps of ETo at 2-km grid using the Spatial CIMIS modeling approach that couples satellite data with point measurements. The ETo data provided below for each urban retail water supplier is an area weighted average value from the Spatial CIMIS ETo.
• Effective precipitation (Peff) - Peff is the portion of total precipitation which becomes available for plant growth. Peff is affected by soil type, slope, land cover type, and intensity and duration of rainfall. DWR is using a soil water balance model, known as Cal-SIMETAW, to estimate daily Peff at 4-km grid and an area weighted average value is calculated at the service area level. Cal-SIMETAW is a model that was developed by UC Davis and DWR and it is widely used to quantify agricultural, and to some extent urban, water uses for the publication of DWR’s Water Plan Update. Peff from Cal-SIMETAW is capped at 25% of total precipitation to account for potential uncertainties in its estimation. Daily Peff at each grid point is aggregated to produce weighted average annual or seasonal Peff at the service area level. The total precipitation that Cal-SIMETAW uses to estimate Peff comes from the Parameter-elevation Regressions on Independent Slopes Model (PRISM), which is a climate mapping model developed by the PRISM Climate Group at Oregon State University.
• Residential Landscape Area Measurement (LAM) – The 2018 Legislation required DWR to provide each urban retail water supplier with data regarding the area of residential irrigable lands in a manner that can reasonably be applied to the standards (CWC 10609.6.(b)). DWR delivered the LAM data to all retail water suppliers, and a tabular summary of selected data types will be provided here. The data summary that is provided in this file contains irrigable-irrigated (II), irrigable-not-irrigated (INI), and not irrigable (NI) irrigation status classes, as well as horse corral areas (HCL_area), agricultural areas (Ag_area), and pool areas (Pool_area) for all retail suppliers.
Facebook
TwitterThis technical report documents the acquisition of source data, and calculation of land cover summary statistics datasets for Antietam National Battlefield. The source data and land cover calculations are available for use within the National Park Service (NPS) Inventory and Monitoring Program. Land cover summary statistics datasets can be calculated for all geographic regions within the extent of the NPS; this report includes statistics calculated for the conterminous United States. The land cover summary statistics datasets are calculated from multiple sources, including Multi-Resolution Land Characteristics Consortium products in the National Land Cover Database (NLCD) and United States Geological Survey’s (USGS) Earth Resources Observation and Science (EROS) Center products in the Land Change Monitoring, Assessment, and Projection (LCMAP) raster dataset. These summary statistics calculate land cover at up to three classification scales: Level 1, modified Anderson Level 2, and Natural versus Converted land cover. The output land cover summary statistics datasets produced here for Antietam National Battlefield utilize the most recent versions of the source datasets (NLCD and LCMAP). These land cover summary statistics datasets are used in the NPS Inventory and Monitoring Program, including the NPS Environmental Settings Monitoring Protocol and may be used by networks and parks for additional efforts.
Facebook
TwitterThis dataset provides detailed insights into daily active users (DAU) of a platform or service, captured over a defined period of time. The dataset includes information such as the number of active users per day, allowing data analysts and business intelligence teams to track usage trends, monitor platform engagement, and identify patterns in user activity over time.
The data is ideal for performing time series analysis, statistical analysis, and trend forecasting. You can utilize this dataset to measure the success of platform initiatives, evaluate user behavior, or predict future trends in engagement. It is also suitable for training machine learning models that focus on user activity prediction or anomaly detection.
The dataset is structured in a simple and easy-to-use format, containing the following columns:
Each row in the dataset represents a unique date and its corresponding number of active users. This allows for time-based analysis, such as calculating the moving average of active users, detecting seasonality, or spotting sudden spikes or drops in engagement.
This dataset can be used for a wide range of purposes, including:
Here are some specific analyses you can perform using this dataset:
To get started with this dataset, you can load it into your preferred analysis tool. Here's how to do it using Python's pandas library:
import pandas as pd
# Load the dataset
data = pd.read_csv('path_to_dataset.csv')
# Display the first few rows
print(data.head())
# Basic statistics
print(data.describe())
Facebook
Twitterhttps://spdx.org/licenses/CC0-1.0.htmlhttps://spdx.org/licenses/CC0-1.0.html
A major objective of plant ecology research is to determine the underlying processes responsible for the observed spatial distribution patterns of plant species. Plants can be approximated as points in space for this purpose, and thus, spatial point pattern analysis has become increasingly popular in ecological research. The basic piece of data for point pattern analysis is a point location of an ecological object in some study region. Therefore, point pattern analysis can only be performed if data can be collected. However, due to the lack of a convenient sampling method, a few previous studies have used point pattern analysis to examine the spatial patterns of grassland species. This is unfortunate because being able to explore point patterns in grassland systems has widespread implications for population dynamics, community-level patterns and ecological processes. In this study, we develop a new method to measure individual coordinates of species in grassland communities. This method records plant growing positions via digital picture samples that have been sub-blocked within a geographical information system (GIS). Here, we tested out the new method by measuring the individual coordinates of Stipa grandis in grazed and ungrazed S. grandis communities in a temperate steppe ecosystem in China. Furthermore, we analyzed the pattern of S. grandis by using the pair correlation function g(r) with both a homogeneous Poisson process and a heterogeneous Poisson process. Our results showed that individuals of S. grandis were overdispersed according to the homogeneous Poisson process at 0-0.16 m in the ungrazed community, while they were clustered at 0.19 m according to the homogeneous and heterogeneous Poisson processes in the grazed community. These results suggest that competitive interactions dominated the ungrazed community, while facilitative interactions dominated the grazed community. In sum, we successfully executed a new sampling method, using digital photography and a Geographical Information System, to collect experimental data on the spatial point patterns for the populations in this grassland community.
Methods 1. Data collection using digital photographs and GIS
A flat 5 m x 5 m sampling block was chosen in a study grassland community and divided with bamboo chopsticks into 100 sub-blocks of 50 cm x 50 cm (Fig. 1). A digital camera was then mounted to a telescoping stake and positioned in the center of each sub-block to photograph vegetation within a 0.25 m2 area. Pictures were taken 1.75 m above the ground at an approximate downward angle of 90° (Fig. 2). Automatic camera settings were used for focus, lighting and shutter speed. After photographing the plot as a whole, photographs were taken of each individual plant in each sub-block. In order to identify each individual plant from the digital images, each plant was uniquely marked before the pictures were taken (Fig. 2 B).
Digital images were imported into a computer as JPEG files, and the position of each plant in the pictures was determined using GIS. This involved four steps: 1) A reference frame (Fig. 3) was established using R2V software to designate control points, or the four vertexes of each sub-block (Appendix S1), so that all plants in each sub-block were within the same reference frame. The parallax and optical distortion in the raster images was then geometrically corrected based on these selected control points; 2) Maps, or layers in GIS terminology, were set up for each species as PROJECT files (Appendix S2), and all individuals in each sub-block were digitized using R2V software (Appendix S3). For accuracy, the digitization of plant individual locations was performed manually; 3) Each plant species layer was exported from a PROJECT file to a SHAPE file in R2V software (Appendix S4); 4) Finally each species layer was opened in Arc GIS software in the SHAPE file format, and attribute data from each species layer was exported into Arc GIS to obtain the precise coordinates for each species. This last phase involved four steps of its own, from adding the data (Appendix S5), to opening the attribute table (Appendix S6), to adding new x and y coordinate fields (Appendix S7) and to obtaining the x and y coordinates and filling in the new fields (Appendix S8).
To determine the accuracy of our new method, we measured the individual locations of Leymus chinensis, a perennial rhizome grass, in representative community blocks 5 m x 5 m in size in typical steppe habitat in the Inner Mongolia Autonomous Region of China in July 2010 (Fig. 4 A). As our standard for comparison, we used a ruler to measure the individual coordinates of L. chinensis. We tested for significant differences between (1) the coordinates of L. chinensis, as measured with our new method and with the ruler, and (2) the pair correlation function g of L. chinensis, as measured with our new method and with the ruler (see section 3.2 Data Analysis). If (1) the coordinates of L. chinensis, as measured with our new method and with the ruler, and (2) the pair correlation function g of L. chinensis, as measured with our new method and with the ruler, did not differ significantly, then we could conclude that our new method of measuring the coordinates of L. chinensis was reliable.
We compared the results using a t-test (Table 1). We found no significant differences in either (1) the coordinates of L. chinensis or (2) the pair correlation function g of L. chinensis. Further, we compared the pattern characteristics of L. chinensis when measured by our new method against the ruler measurements using a null model. We found that the two pattern characteristics of L. chinensis did not differ significantly based on the homogenous Poisson process or complete spatial randomness (Fig. 4 B). Thus, we concluded that the data obtained using our new method was reliable enough to perform point pattern analysis with a null model in grassland communities.
Facebook
TwitterMIT Licensehttps://opensource.org/licenses/MIT
License information was derived automatically
This artifact accompanies the SEET@ICSE article "Assessing the impact of hints in learning formal specification", which reports on a user study to investigate the impact of different types of automated hints while learning a formal specification language, both in terms of immediate performance and learning retention, but also in the emotional response of the students. This research artifact provides all the material required to replicate this study (except for the proprietary questionnaires passed to assess the emotional response and user experience), as well as the collected data and data analysis scripts used for the discussion in the paper.
Dataset
The artifact contains the resources described below.
Experiment resources
The resources needed for replicating the experiment, namely in directory experiment:
alloy_sheet_pt.pdf: the 1-page Alloy sheet that participants had access to during the 2 sessions of the experiment. The sheet was passed in Portuguese due to the population of the experiment.
alloy_sheet_en.pdf: a version the 1-page Alloy sheet that participants had access to during the 2 sessions of the experiment translated into English.
docker-compose.yml: a Docker Compose configuration file to launch Alloy4Fun populated with the tasks in directory data/experiment for the 2 sessions of the experiment.
api and meteor: directories with source files for building and launching the Alloy4Fun platform for the study.
Experiment data
The task database used in our application of the experiment, namely in directory data/experiment:
Model.json, Instance.json, and Link.json: JSON files with to populate Alloy4Fun with the tasks for the 2 sessions of the experiment.
identifiers.txt: the list of all (104) available participant identifiers that can participate in the experiment.
Collected data
Data collected in the application of the experiment as a simple one-factor randomised experiment in 2 sessions involving 85 undergraduate students majoring in CSE. The experiment was validated by the Ethics Committee for Research in Social and Human Sciences of the Ethics Council of the University of Minho, where the experiment took place. Data is shared the shape of JSON and CSV files with a header row, namely in directory data/results:
data_sessions.json: data collected from task-solving in the 2 sessions of the experiment, used to calculate variables productivity (PROD1 and PROD2, between 0 and 12 solved tasks) and efficiency (EFF1 and EFF2, between 0 and 1).
data_socio.csv: data collected from socio-demographic questionnaire in the 1st session of the experiment, namely:
participant identification: participant's unique identifier (ID);
socio-demographic information: participant's age (AGE), sex (SEX, 1 through 4 for female, male, prefer not to disclosure, and other, respectively), and average academic grade (GRADE, from 0 to 20, NA denotes preference to not disclosure).
data_emo.csv: detailed data collected from the emotional questionnaire in the 2 sessions of the experiment, namely:
participant identification: participant's unique identifier (ID) and the assigned treatment (column HINT, either N, L, E or D);
detailed emotional response data: the differential in the 5-point Likert scale for each of the 14 measured emotions in the 2 sessions, ranging from -5 to -1 if decreased, 0 if maintained, from 1 to 5 if increased, or NA denoting failure to submit the questionnaire. Half of the emotions are positive (Admiration1 and Admiration2, Desire1 and Desire2, Hope1 and Hope2, Fascination1 and Fascination2, Joy1 and Joy2, Satisfaction1 and Satisfaction2, and Pride1 and Pride2), and half are negative (Anger1 and Anger2, Boredom1 and Boredom2, Contempt1 and Contempt2, Disgust1 and Disgust2, Fear1 and Fear2, Sadness1 and Sadness2, and Shame1 and Shame2). This detailed data was used to compute the aggregate data in data_emo_aggregate.csv and in the detailed discussion in Section 6 of the paper.
data_umux.csv: data collected from the user experience questionnaires in the 2 sessions of the experiment, namely:
participant identification: participant's unique identifier (ID);
user experience data: summarised user experience data from the UMUX surveys (UMUX1 and UMUX2, as a usability metric ranging from 0 to 100).
participants.txt: the list of participant identifiers that have registered for the experiment.
Analysis scripts
The analysis scripts required to replicate the analysis of the results of the experiment as reported in the paper, namely in directory analysis:
analysis.r: An R script to analyse the data in the provided CSV files; each performed analysis is documented within the file itself.
requirements.r: An R script to install the required libraries for the analysis script.
normalize_task.r: A Python script to normalize the task JSON data from file data_sessions.json into the CSV format required by the analysis script.
normalize_emo.r: A Python script to compute the aggregate emotional response in the CSV format required by the analysis script from the detailed emotional response data in the CSV format of data_emo.csv.
Dockerfile: Docker script to automate the analysis script from the collected data.
Setup
To replicate the experiment and the analysis of the results, only Docker is required.
If you wish to manually replicate the experiment and collect your own data, you'll need to install:
A modified version of the Alloy4Fun platform, which is built in the Meteor web framework. This version of Alloy4Fun is publicly available in branch study of its repository at https://github.com/haslab/Alloy4Fun/tree/study.
If you wish to manually replicate the analysis of the data collected in our experiment, you'll need to install:
Python to manipulate the JSON data collected in the experiment. Python is freely available for download at https://www.python.org/downloads/, with distributions for most platforms.
R software for the analysis scripts. R is freely available for download at https://cran.r-project.org/mirrors.html, with binary distributions available for Windows, Linux and Mac.
Usage
Experiment replication
This section describes how to replicate our user study experiment, and collect data about how different hints impact the performance of participants.
To launch the Alloy4Fun platform populated with tasks for each session, just run the following commands from the root directory of the artifact. The Meteor server may take a few minutes to launch, wait for the "Started your app" message to show.
cd experimentdocker-compose up
This will launch Alloy4Fun at http://localhost:3000. The tasks are accessed through permalinks assigned to each participant. The experiment allows for up to 104 participants, and the list of available identifiers is given in file identifiers.txt. The group of each participant is determined by the last character of the identifier, either N, L, E or D. The task database can be consulted in directory data/experiment, in Alloy4Fun JSON files.
In the 1st session, each participant was given one permalink that gives access to 12 sequential tasks. The permalink is simply the participant's identifier, so participant 0CAN would just access http://localhost:3000/0CAN. The next task is available after a correct submission to the current task or when a time-out occurs (5mins). Each participant was assigned to a different treatment group, so depending on the permalink different kinds of hints are provided. Below are 4 permalinks, each for each hint group:
Group N (no hints): http://localhost:3000/0CAN
Group L (error locations): http://localhost:3000/CA0L
Group E (counter-example): http://localhost:3000/350E
Group D (error description): http://localhost:3000/27AD
In the 2nd session, likewise the 1st session, each permalink gave access to 12 sequential tasks, and the next task is available after a correct submission or a time-out (5mins). The permalink is constructed by prepending the participant's identifier with P-. So participant 0CAN would just access http://localhost:3000/P-0CAN. In the 2nd sessions all participants were expected to solve the tasks without any hints provided, so the permalinks from different groups are undifferentiated.
Before the 1st session the participants should answer the socio-demographic questionnaire, that should ask the following information: unique identifier, age, sex, familiarity with the Alloy language, and average academic grade.
Before and after both sessions the participants should answer the standard PrEmo 2 questionnaire. PrEmo 2 is published under an Attribution-NonCommercial-NoDerivatives 4.0 International Creative Commons licence (CC BY-NC-ND 4.0). This means that you are free to use the tool for non-commercial purposes as long as you give appropriate credit, provide a link to the license, and do not modify the original material. The original material, namely the depictions of the diferent emotions, can be downloaded from https://diopd.org/premo/. The questionnaire should ask for the unique user identifier, and for the attachment with each of the depicted 14 emotions, expressed in a 5-point Likert scale.
After both sessions the participants should also answer the standard UMUX questionnaire. This questionnaire can be used freely, and should ask for the user unique identifier and answers for the standard 4 questions in a 7-point Likert scale. For information about the questions, how to implement the questionnaire, and how to compute the usability metric ranging from 0 to 100 score from the answers, please see the original paper:
Kraig Finstad. 2010. The usability metric for user experience. Interacting with computers 22, 5 (2010), 323–327.
Analysis of other applications of the experiment
This section describes how to replicate the analysis of the data collected in an application of the experiment described in Experiment replication.
The analysis script expects data in 4 CSV files,
Facebook
TwitterCC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
License information was derived automatically
Release Date: 2020-06-09.Release Schedule:.The data in this file come from the 2017 Economic Census data files released on a flow basis starting in September 2019. As such, preliminary U.S. totals released in September 2019 will be superseded with final totals, by sector, once data for all states have been released. Users should be aware that during the release of this consolidated file, data at more detailed North American Industry Classification System (NAICS) and geographic levels may not add to higher-level totals. However, at the completion of the economic census (once all the component files have been released), the detailed data in this file will add to the totals. For more information about economic census planned data product releases, see Economic Census: About: 2017 Release Schedules...Key Table Information:.U.S. totals released in September 2019 will be superseded with final totals, by sector, once data for all states have been released. .Includes only establishments and firms with payroll..Data may be subject to employment- and/or sales-size minimums that vary by industry...Data Items and Other Identifying Records: .Number of firms.Number of establishments.Sales, value of shipments, or revenue ($1,000).Annual payroll ($1,000).First-quarter payroll ($1,000).Number of employees.Range indicating percent of total sales, value of shipments, or revenue imputed.Range indicating percent of total annual payroll imputed.Range indicating percent of total employees imputed..Geography Coverage:.The data are shown for employer establishments and firms at the U.S., State, Combined Statistical Area, Metropolitan and Micropolitan Statistical Area, Metropolitan Division, Consolidated City, County (and equivalent), and Economic Place (and equivalent; incorporated and unincorporated) levels that vary by industry. For information about economic census geographies, including changes for 2017, see Economic Census: Economic Geographies...Industry Coverage:.The data are shown at the 2- through 6-digit, and selected 7-digit 2017 NAICS code levels. For information about NAICS, see Economic Census: Technical Documentation: Code Lists...Footnotes:.106: Railroad transportation and U.S. Postal Service are out of scope...FTP Download:.Download the entire table at: https://www2.census.gov/programs-surveys/economic-census/data/2017/sector48/EC1748BASIC.zip..API Information:.Economic census data are housed in the Census Bureau API. For more information, see Explore Data: Developers: Available APIs: Economic Census..Methodology:.To maintain confidentiality, the U.S. Census Bureau suppresses data to protect the identity of any business or individual. The census results in this file contain sampling and/or nonsampling error. Data users who create their own estimates using data from this file should cite the U.S. Census Bureau as the source of the original data only...To comply with disclosure avoidance guidelines, data rows with fewer than three contributing establishments are not presented. Additionally, establishment counts are suppressed when other select statistics in the same row are suppressed. For detailed information about the methods used to collect and produce statistics, including sampling, eligibility, questions, data collection and processing, data quality, review, weighting, estimation, coding operations, confidentiality protection, sampling error, nonsampling error, and more, see Economic Census: Technical Documentation: Methodology...Symbols:.D - Withheld to avoid disclosing data for individual companies; data are included in higher level totals.N - Not available or not comparable.S - Estimate does not meet publication standards because of high sampling variability, poor response quality, or other concerns about the estimate quality. Unpublished estimates derived from this table by subtraction are subject to these same limitations and should not be attributed to the U.S. Census Bureau. For a description of publication standards and the total quantity response rate, see link to program methodology page..X - Not applicable.A - Relative standard error of 100% or more.r - Revised.s - Relative standard error exceeds 40%.For a complete list of symbols, see Economic Census: Technical Documentation: Data Dictionary.. .Source:.U.S. Census Bureau, 2017 Economic Census.For information about the economic census, see Business and Economy: Economic Census...Contact Information:.U.S. Census Bureau.For general inquiries:. (800) 242-2184/ (301) 763-5154. ewd.outreach@census.gov.For specific data questions:. (800) 541-8345.For additional contacts, see Economic Census: About: Contact Us.
Facebook
TwitterThis dataset comes from the Community Survey questions relating to the Community Health & Well-Being performance measure: "With “10” representing the best possible life for you and “0” representing the worst, how would you say you personally feel you stand at this time?" and "With “10” representing the best possible life for you and “0” representing the worst, how do you think you will stand about five years from now?" – the results of both scores are then used to assess a Cantril Scale which is a way of assessing general life satisfaction. As per the Cantril Self-Anchoring Striving Scale, the three categories of identification are as follows: Thriving – Respondents rate their current life as a 7 or higher AND their future life as an 8 or higher. Suffering – Respondents rate their current life negatively (0 to 4) AND their future life negatively (0 to 4). Struggling – Respondents who do not meet the criteria for Thriving or Suffering. The survey is mailed to a random sample of households in the City of Tempe and has a 95% confidence level. Note on Methodology Update: In 2025, the Cantril classification method was revised to align with Gallup’s official Life Evaluation Index methodology. This change affects only a small number of respondents whose answers did not fit cleanly into the previous custom definition of “Struggling,” which classified respondents who rated their current life moderately (5 or 6) or their future life moderately or negatively (0 to 7). Under the updated approach, respondents who previously fell outside that definition are now appropriately included in the Struggling category. The overall distribution of Thriving, Struggling, and Suffering changed only minimally, and the updated methodology has been applied consistently to all prior years.This page provides data for the Community Health and Well-Being performance measure.The performance measure dashboard is available at 3.34 Community Health and Well-Being.Data Dictionary Additional InformationSource: Community Attitude Survey (Vendor: ETC Institute)Contact: Amber AsburryContact email: amber_asburry@tempe.govPreparation Method: Survey results from two questions are calculated to create a Cantril Scale value that falls into the categories of Thriving, Struggling, and Suffering.Publish Frequency: AnnuallyPublish Method: Manual
Facebook
TwitterCC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
License information was derived automatically
Release Date: 2020-06-09.Release Schedule:.The data in this file come from the 2017 Economic Census data files released on a flow basis starting in September 2019. As such, preliminary U.S. totals released in September 2019 will be superseded with final totals, by sector, once data for all states have been released. Users should be aware that during the release of this consolidated file, data at more detailed North American Industry Classification System (NAICS) and geographic levels may not add to higher-level totals. However, at the completion of the economic census (once all the component files have been released), the detailed data in this file will add to the totals. For more information about economic census planned data product releases, see Economic Census: About: 2017 Release Schedules...Key Table Information:.U.S. totals released in September 2019 will be superseded with final totals, by sector, once data for all states have been released. .Includes only establishments and firms with payroll..Data may be subject to employment- and/or sales-size minimums that vary by industry...Data Items and Other Identifying Records: .Number of firms.Number of establishments.Sales, value of shipments, or revenue ($1,000).Annual payroll ($1,000).First-quarter payroll ($1,000).Number of employees.Range indicating percent of total sales, value of shipments, or revenue imputed.Range indicating percent of total annual payroll imputed.Range indicating percent of total employees imputed..Geography Coverage:.The data are shown for employer establishments and firms at the U.S., State, Combined Statistical Area, Metropolitan and Micropolitan Statistical Area, Metropolitan Division, Consolidated City, County (and equivalent), and Economic Place (and equivalent; incorporated and unincorporated) levels that vary by industry. For information about economic census geographies, including changes for 2017, see Economic Census: Economic Geographies...Industry Coverage:.The data are shown at the 2- through 6-digit 2017 NAICS code levels. For information about NAICS, see Economic Census: Technical Documentation: Code Lists...Footnotes:.Not applicable...FTP Download:.Download the entire table at: https://www2.census.gov/programs-surveys/economic-census/data/2017/sector53/EC1753BASIC.zip..API Information:.Economic census data are housed in the Census Bureau API. For more information, see Explore Data: Developers: Available APIs: Economic Census..Methodology:.To maintain confidentiality, the U.S. Census Bureau suppresses data to protect the identity of any business or individual. The census results in this file contain sampling and/or nonsampling error. Data users who create their own estimates using data from this file should cite the U.S. Census Bureau as the source of the original data only...To comply with disclosure avoidance guidelines, data rows with fewer than three contributing establishments are not presented. Additionally, establishment counts are suppressed when other select statistics in the same row are suppressed. For detailed information about the methods used to collect and produce statistics, including sampling, eligibility, questions, data collection and processing, data quality, review, weighting, estimation, coding operations, confidentiality protection, sampling error, nonsampling error, and more, see Economic Census: Technical Documentation: Methodology...Symbols:.D - Withheld to avoid disclosing data for individual companies; data are included in higher level totals.N - Not available or not comparable.S - Estimate does not meet publication standards because of high sampling variability, poor response quality, or other concerns about the estimate quality. Unpublished estimates derived from this table by subtraction are subject to these same limitations and should not be attributed to the U.S. Census Bureau. For a description of publication standards and the total quantity response rate, see link to program methodology page..X - Not applicable.A - Relative standard error of 100% or more.r - Revised.s - Relative standard error exceeds 40%.For a complete list of symbols, see Economic Census: Technical Documentation: Data Dictionary.. .Source:.U.S. Census Bureau, 2017 Economic Census.For information about the economic census, see Business and Economy: Economic Census...Contact Information:.U.S. Census Bureau.For general inquiries:. (800) 242-2184/ (301) 763-5154. ewd.outreach@census.gov.For specific data questions:. (800) 541-8345.For additional contacts, see Economic Census: About: Contact Us.
Facebook
TwitterU.S. Government Workshttps://www.usa.gov/government-works
License information was derived automatically
Cover crops provide many agroecosystem services, including weed suppression, which is partially exerted through release of allelopathic benzoxazinoid (BX) compounds. This research characterizes (1) changes in concentrations of BX compounds in shoots, roots, and soil at three growth stages (GS) of cereal rye (Secale cereale L.), and (2) their degradation over time following termination. Concentrations of shoot dominant BX compounds, DIBOA-glc and DIBOA, were least at GS 83 (boot). The root dominant BX compound, HMBOA-glc, concentration was least at GS 54 (elongation). Rhizosphere soil BX concentrations were 1000 times smaller than in root tissues. Dominant compounds in soil were HMBOA-glc and HMBOA. Concentrations of BX compounds were similar for soil near root crowns and between-rows. Soil BX concentrations following cereal rye termination declined exponentially over time in three of four treatments: incorporated shoots (S) and roots (R), no-till S+R (cereal rye rolled flat), and no-till R (shoots removed), but not in no-till S. On the day following cereal rye termination, soil concentrations of HMBOA-glc and HMBOA in these three treatments increased above initial concentrations. Concentrations of these two compounds decreased the fastest while DIBOA-glc declined the slowest (half-life of 4 d in no-till S+R soil). Placement of shoots on the surface of an area where cereal rye had not grown (no-till S) did not increase soil concentrations of BX compounds. The short duration and complex dynamics of BX compounds in soil prior to and following termination illustrate the limited window for enhancing weed suppression by cereal rye allelochemicals; valuable information for programs breeding for enhanced weed suppression. In addition to the data analyzed for this article, we also include the R code. Resources in this dataset:Resource Title: BX data following termination. File Name: FinalBXsForMatt-20200908.csvResource Description: For each sample, gives the time, depth, location, and plot treatment, and then the compound concentrations. This is the principal data set analyzed with the R (anal2-cleaned.r) code, see that code for use.Resource Title: BX compounds from 3rd sampling time before termination. File Name: soil2-20201123.csvResource Description: These data are for comparison with the post termination data. They were taken at the 3rd sampling time (pre-termination), a day prior to termination. Each sample is identified with a treatment, date, and plot location, in addition to the BX concentrations. See R code (anal2-cleaned.r) for how this file is used.Resource Title: Soil location (within row versus between row) values of BX compounds. File Name: s2b.csvResource Description: Each row gives the average BX compound for each soil location (within row versus between row) for the second sample for each plot. These data are combined with bx3 (the data set read in from the file , "FinalBXsForMatt-20200908.csv"). See R (anal2-cleaned.r) code for use.Resource Title: R code for analysis of the decay (post-termination) BX data.. File Name: anal2-cleaned.rResource Description: This is the R code used to analyze the termination data. It also creates and writes out some data subsets (used for analysis and plots) that are later read in.Resource Software Recommended: R version 3.6.3,url: https://www.R-project.org/ Resource Title: Tissue BX compounds. File Name: tissues20210728b.csvResource Description: Data file holding results from a tissue analysis for BX compounds, in ug, from shoots and roots, and at various sampling times. Read into the R file, anal1-cleaned.r where it is used in a statistical analysis and to create figures.Resource Title: BX compounds from soil with a live rye cover crop. File Name: soil2-20201214.csvResource Description: BX compounds (in ng/g dry wt), by treatment, sampling time, date, and plot ID. These are data are read into the R program, anal1-cleaned.r, for analysis and to create figures. These are soil samples taken from locations with a live rye plant cover crop.Resource Title: R code for BX analyses of soil under rye and plant tissues. File Name: anal1-cleaned.rResource Description: R code for analysis of the soil BX compounds under a live rye cover crop at different growing stages, and for the analysis of tissue BX compounds. In addition to statistical analyses, code in this file creates figures, also some statistical output that is used to create a file that is later read in for figure creation (s2-CLD20220730-Stage.csv).Resource Software Recommended: R version 3.6.3,url: https://www.R-project.org/ Resource Title: Description of data files for anal2-cleaned.r. File Name: readme2.txtResource Description: Describes the input files used in the R code in anal2-cleaned.r, including descriptions and formats for each field. The file also describes some output (results) files that were uploaded to this site. This is a plain ASCII text file.Resource Title: Estimates produced by anal2-cleaned.r from statistical modeling.. File Name: Estimates20201110.csvResource Description: Estimates produced by anal2-cleaned.r from statistical modeling (see readme2.txt)Resource Title: Summary statistics from anal2-cleaned.r. File Name: CV20210412.csvResource Description: Summary statistics from anal2-cleaned.r, used for plotsResource Title: Data summaries (same as CV20210412.csv), rescaled. File Name: RESCALE-20210412.csvResource Description: Same as "CV20210412.csv" except log of data have been rescaled to minimum at least zero and maximum one, see readme2.txtResource Title: Statistical summaries for different stages. File Name: s2-CLD20220730-Stage.csvResource Description: Statistical summaries used for creating a figure (not used in paper), used in anal1-cleaned.r; data for soil BX under living rye.Resource Title: Description of data files for anal1-cleaned.r. File Name: readme1.txtResource Description: Contains general descriptions of data imported into anal1-cleaned.r, and a description of each field. Also contains some descriptions of files output by anal1-cleaned.r, used to create tables or figures.
Facebook
Twitter
Facebook
TwitterUnder the new quarterly data summary (QDS) framework departments’ spending data is published every quarter; to show the taxpayer how the government is spending their money.
The QDS grew out of commitments made in the 2011 Budget and the written ministerial statement on business plans. For the financial year 2012 to 2013 the QDS has been revised and improved in line with action 9 of the Civil Service Reform Plan to provide a common set of data that will enable comparisons of operational performance across government so that departments and individuals can be held to account.
The QDS breaks down the total spend of the department in 3 ways:
The QDS template is the same for all departments, though the individual detail of grants and policy will differ from department to department. In using this data:
Please note that the quarter 1 2012 to 2013 return for the Department of Transport (DfT) is for the core department only.
Quarterly data summaries for April 2012 are as follows:
Quarterly data summaries for January 2012 are as follows:
Quarterly data summaries for October 2011 are as follows: