https://spdx.org/licenses/CC0-1.0.htmlhttps://spdx.org/licenses/CC0-1.0.html
Professional organizations in STEM (science, technology, engineering, and mathematics) can use demographic data to quantify recruitment and retention (R&R) of underrepresented groups within their memberships. However, variation in the types of demographic data collected can influence the targeting and perceived impacts of R&R efforts - e.g., giving false signals of R&R for some groups. We obtained demographic surveys from 73 U.S.-affiliated STEM organizations, collectively representing 712,000 members and conference-attendees. We found large differences in the demographic categories surveyed (e.g., disability status, sexual orientation) and the available response options. These discrepancies indicate a lack of consensus regarding the demographic groups that should be recognized and, for groups that are omitted from surveys, an inability of organizations to prioritize and evaluate R&R initiatives. Aligning inclusive demographic surveys across organizations will provide baseline data that can be used to target and evaluate R&R initiatives to better serve underrepresented groups throughout STEM. Methods We surveyed 164 STEM organizations (73 responses, rate = 44.5%) between December 2020 and July 2021 with the goal of understanding what demographic data each organization collects from its constituents (i.e., members and conference-attendees) and how the data are used. Organizations were sourced from a list of professional societies affiliated with the American Association for the Advancement of Science, AAAS, (n = 156) or from social media (n = 8). The survey was sent to the elected leadership and management firms for each organization, and follow-up reminders were sent after one month. The responding organizations represented a wide range of fields: 31 life science organizations (157,000 constituents), 5 mathematics organizations (93,000 constituents), 16 physical science organizations (207,000 constituents), 7 technology organizations (124,000 constituents), and 14 multi-disciplinary organizations spanning multiple branches of STEM (131,000 constituents). A list of the responding organizations is available in the Supplementary Materials. Based on the AAAS-affiliated recruitment of the organizations and the similar distribution of constituencies across STEM fields, we conclude that the responding organizations are a representative cross-section of the most prominent STEM organizations in the U.S. Each organization was asked about the demographic information they collect from their constituents, the response rates to their surveys, and how the data were used. Survey description The following questions are written as presented to the participating organizations. Question 1: What is the name of your STEM organization? Question 2: Does your organization collect demographic data from your membership and/or meeting attendees? Question 3: When was your organization’s most recent demographic survey (approximate year)? Question 4: We would like to know the categories of demographic information collected by your organization. You may answer this question by either uploading a blank copy of your organization’s survey (linked provided in online version of this survey) OR by completing a short series of questions. Question 5: On the most recent demographic survey or questionnaire, what categories of information were collected? (Please select all that apply)
Disability status Gender identity (e.g., male, female, non-binary) Marital/Family status Racial and ethnic group Religion Sex Sexual orientation Veteran status Other (please provide)
Question 6: For each of the categories selected in Question 5, what options were provided for survey participants to select? Question 7: Did the most recent demographic survey provide a statement about data privacy and confidentiality? If yes, please provide the statement. Question 8: Did the most recent demographic survey provide a statement about intended data use? If yes, please provide the statement. Question 9: Who maintains the demographic data collected by your organization? (e.g., contracted third party, organization executives) Question 10: How has your organization used members’ demographic data in the last five years? Examples: monitoring temporal changes in demographic diversity, publishing diversity data products, planning conferences, contributing to third-party researchers. Question 11: What is the size of your organization (number of members or number of attendees at recent meetings)? Question 12: What was the response rate (%) for your organization’s most recent demographic survey? *Organizations were also able to upload a copy of their demographics survey instead of responding to Questions 5-8. If so, the uploaded survey was used (by the study authors) to evaluate Questions 5-8.
Pursuant to Local Laws 126, 127, and 128 of 2016, certain demographic data is collected voluntarily and anonymously by persons voluntarily seeking social services. This data can be used by agencies and the public to better understand the demographic makeup of client populations and to better understand and serve residents of all backgrounds and identities. The data presented here has been collected through either electronic form or paper surveys offered at the point of application for services. These surveys are anonymous. Each record represents an anonymized demographic profile of an individual applicant for social services, disaggregated by response option, agency, and program. Response options include information regarding ancestry, race, primary and secondary languages, English proficiency, gender identity, and sexual orientation. Idiosyncrasies or Limitations: Note that while the dataset contains the total number of individuals who have identified their ancestry or languages spoke, because such data is collected anonymously, there may be instances of a single individual completing multiple voluntary surveys. Additionally, the survey being both voluntary and anonymous has advantages as well as disadvantages: it increases the likelihood of full and honest answers, but since it is not connected to the individual case, it does not directly inform delivery of services to the applicant. The paper and online versions of the survey ask the same questions but free-form text is handled differently. Free-form text fields are expected to be entered in English although the form is available in several languages. Surveys are presented in 11 languages. Paper Surveys 1. Are optional 2. Survey taker is expected to specify agency that provides service 2. Survey taker can skip or elect not to answer questions 3. Invalid/unreadable data may be entered for survey date or date may be skipped 4. OCRing of free-form tet fields may fail. 5. Analytical value of free-form text answers is unclear Online Survey 1. Are optional 2. Agency is defaulted based on the URL 3. Some questions must be answered 4. Date of survey is automated
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Open Science in (Higher) Education – data of the February 2017 survey
This data set contains:
Full raw (anonymised) data set (completed responses) of Open Science in (Higher) Education February 2017 survey. Data are in xlsx and sav format.
Survey questionnaires with variables and settings (German original and English translation) in pdf. The English questionnaire was not used in the February 2017 survey, but only serves as translation.
Readme file (txt)
Survey structure
The survey includes 24 questions and its structure can be separated in five major themes: material used in courses (5), OER awareness, usage and development (6), collaborative tools used in courses (2), assessment and participation options (5), demographics (4). The last two questions include an open text questions about general issues on the topics and singular open education experiences, and a request on forwarding the respondent's e-mail address for further questionings. The online survey was created with Limesurvey[1]. Several questions include filters, i.e. these questions were only shown if a participants did choose a specific answer beforehand ([n/a] in Excel file, [.] In SPSS).
Demographic questions
Demographic questions asked about the current position, the discipline, birth year and gender. The classification of research disciplines was adapted to general disciplines at German higher education institutions. As we wanted to have a broad classification, we summarised several disciplines and came up with the following list, including the option "other" for respondents who do not feel confident with the proposed classification:
Natural Sciences
Arts and Humanities or Social Sciences
Economics
Law
Medicine
Computer Sciences, Engineering, Technics
Other
The current job position classification was also chosen according to common positions in Germany, including positions with a teaching responsibility at higher education institutions. Here, we also included the option "other" for respondents who do not feel confident with the proposed classification:
Professor
Special education teacher
Academic/scientific assistant or research fellow (research and teaching)
Academic staff (teaching)
Student assistant
Other
We chose to have a free text (numerical) for asking about a respondent's year of birth because we did not want to pre-classify respondents' age intervals. It leaves us options to have different analysis on answers and possible correlations to the respondents' age. Asking about the country was left out as the survey was designed for academics in Germany.
Remark on OER question
Data from earlier surveys revealed that academics suffer confusion about the proper definition of OER[2]. Some seem to understand OER as free resources, or only refer to open source software (Allen & Seaman, 2016, p. 11). Allen and Seaman (2016) decided to give a broad explanation of OER, avoiding details to not tempt the participant to claim "aware". Thus, there is a danger of having a bias when giving an explanation. We decided not to give an explanation, but keep this question simple. We assume that either someone knows about OER or not. If they had not heard of the term before, they do not probably use OER (at least not consciously) or create them.
Data collection
The target group of the survey was academics at German institutions of higher education, mainly universities and universities of applied sciences. To reach them we sent the survey to diverse institutional-intern and extern mailing lists and via personal contacts. Included lists were discipline-based lists, lists deriving from higher education and higher education didactic communities as well as lists from open science and OER communities. Additionally, personal e-mails were sent to presidents and contact persons from those communities, and Twitter was used to spread the survey.
The survey was online from Feb 6th to March 3rd 2017, e-mails were mainly sent at the beginning and around mid-term.
Data clearance
We got 360 responses, whereof Limesurvey counted 208 completes and 152 incompletes. Two responses were marked as incomplete, but after checking them turned out to be complete, and we added them to the complete responses dataset. Thus, this data set includes 210 complete responses. From those 150 incomplete responses, 58 respondents did not answer 1st question, 40 respondents discontinued after 1st question. Data shows a constant decline in response answers, we did not detect any striking survey question with a high dropout rate. We deleted incomplete responses and they are not in this data set.
Due to data privacy reasons, we deleted seven variables automatically assigned by Limesurvey: submitdate, lastpage, startlanguage, startdate, datestamp, ipaddr, refurl. We also deleted answers to question No 24 (email address).
References
Allen, E., & Seaman, J. (2016). Opening the Textbook: Educational Resources in U.S. Higher Education, 2015-16.
First results of the survey are presented in the poster:
Heck, Tamara, Blümel, Ina, Heller, Lambert, Mazarakis, Athanasios, Peters, Isabella, Scherp, Ansgar, & Weisel, Luzian. (2017). Survey: Open Science in Higher Education. Zenodo. http://doi.org/10.5281/zenodo.400561
Contact:
Open Science in (Higher) Education working group, see http://www.leibniz-science20.de/forschung/projekte/laufende-projekte/open-science-in-higher-education/.
[1] https://www.limesurvey.org
[2] The survey question about the awareness of OER gave a broad explanation, avoiding details to not tempt the participant to claim "aware".
https://www.icpsr.umich.edu/web/ICPSR/studies/29646/termshttps://www.icpsr.umich.edu/web/ICPSR/studies/29646/terms
This data collection is comprised of responses from the March and April installments of the 2008 Current Population Survey (CPS). Both the March and April surveys used two sets of questions, the basic CPS and a separate supplement for each month.The CPS, administered monthly, is a labor force survey providing current estimates of the economic status and activities of the population of the United States. Specifically, the CPS provides estimates of total employment (both farm and nonfarm), nonfarm self-employed persons, domestics, and unpaid helpers in nonfarm family enterprises, wage and salaried employees, and estimates of total unemployment.In addition to the basic CPS questions, respondents were asked questions from the March supplement, known as the Annual Social and Economic (ASEC) supplement. The ASEC provides supplemental data on work experience, income, noncash benefits, and migration. Comprehensive work experience information was given on the employment status, occupation, and industry of persons 15 years old and older. Additional data for persons 15 years old and older are available concerning weeks worked and hours per week worked, reason not working full time, total income and income components, and place of residence on March 1, 2007. The March supplement also contains data covering nine noncash income sources: food stamps, school lunch program, employer-provided group health insurance plan, employer-provided pension plan, personal health insurance, Medicaid, Medicare, CHAMPUS or military health care, and energy assistance. Questions covering training and assistance received under welfare reform programs, such as job readiness training, child care services, or job skill training were also asked in the March supplement.The April supplement, sponsored by the Department of Health and Human Services, queried respondents on the economic situation of persons and families for the previous year. Moreover, all household members 15 years of age and older that are a biological parent of children in the household that have an absent parent were asked detailed questions about child support and alimony. Information regarding child support was collected to determine the size and distribution of the population with children affected by divorce or separation, or other relationship status change. Moreover, the data were collected to better understand the characteristics of persons requiring child support, and to help develop and maintain programs designed to assist in obtaining child support. These data highlight alimony and child support arrangements made at the time of separation or divorce, amount of payments actually received, and value and type of any property settlement.The April supplement data were matched to March supplement data for households that were in the sample in both March and April 2008. In March 2008, there were 4,522 household members eligible, of which 1,431 required imputation of child support data. When matching the March 2008 and April 2008 data sets, there were 170 eligible people on the March file that did not match to people on the April file. Child support data for these 170 people were imputed. The remaining 1,261 imputed cases were due to nonresponse to the child support questions. Demographic variables include age, sex, race, Hispanic origin, marital status, veteran status, educational attainment, occupation, and income. Data on employment and income refer to the preceding year, although other demographic data refer to the time at which the survey was administered.
The Gallup Poll Social Series (GPSS) is a set of public opinion surveys designed to monitor U.S. adults' views on numerous social, economic, and political topics. The topics are arranged thematically across 12 surveys. Gallup administers these surveys during the same month every year and includes the survey's core trend questions in the same order each administration. Using this consistent standard allows for unprecedented analysis of changes in trend data that are not susceptible to question order bias and seasonal effects.
Introduced in 2001, the GPSS is the primary method Gallup uses to update several hundred long-term Gallup trend questions, some dating back to the 1930s. The series also includes many newer questions added to address contemporary issues as they emerge.
The dataset currently includes responses from up to and including 2025.
Gallup conducts one GPSS survey per month, with each devoted to a different topic, as follows:
January: Mood of the Nation
February: World Affairs
March: Environment
April: Economy and Finance
May: Values and Beliefs
June: Minority Rights and Relations (discontinued after 2016)
July: Consumption Habits
August: Work and Education
September: Governance
October: Crime
November: Health
December: Lifestyle (conducted 2001-2008)
The core questions of the surveys differ each month, but several questions assessing the state of the nation are standard on all 12: presidential job approval, congressional job approval, satisfaction with the direction of the U.S., assessment of the U.S. job market, and an open-ended measurement of the nation's "most important problem." Additionally, Gallup includes extensive demographic questions on each survey, allowing for in-depth analysis of trends.
Interviews are conducted with U.S. adults aged 18 and older living in all 50 states and the District of Columbia using a dual-frame design, which includes both landline and cellphone numbers. Gallup samples landline and cellphone numbers using random-digit-dial methods. Gallup purchases samples for this study from Survey Sampling International (SSI). Gallup chooses landline respondents at random within each household based on which member had the next birthday. Each sample of national adults includes a minimum quota of 70% cellphone respondents and 30% landline respondents, with additional minimum quotas by time zone within region. Gallup conducts interviews in Spanish for respondents who are primarily Spanish-speaking.
Gallup interviews a minimum of 1,000 U.S. adults aged 18 and older for each GPSS survey. Samples for the June Minority Rights and Relations survey are significantly larger because Gallup includes oversamples of Blacks and Hispanics to allow for reliable estimates among these key subgroups.
Gallup weights samples to correct for unequal selection probability, nonresponse, and double coverage of landline and cellphone users in the two sampling frames. Gallup also weights its final samples to match the U.S. population according to gender, age, race, Hispanic ethnicity, education, region, population density, and phone status (cellphone only, landline only, both, and cellphone mostly).
Demographic weighting targets are based on the most recent Current Population Survey figures for the aged 18 and older U.S. population. Phone status targets are based on the most recent National Health Interview Survey. Population density targets are based on the most recent U.S. Census.
The year appended to each table name represents when the data was last updated. For example, January: Mood of the Nation - 2025** **has survey data collected up to and including 2025.
For more information about what survey questions were asked over time, see the Supporting Files.
Data access is required to view this section.
The Bangladesh Demographic and Health Survey (BDHS) is the first of this kind of study conducted in Bangladesh. It provides rapid feedback on key demographic and programmatic indicators to monitor the strength and weaknesses of the national family planning/MCH program. The wealth of information collected through the 1993-94 BDHS will be of immense value to the policymakers and program managers in order to strengthen future program policies and strategies.
The BDHS is intended to serve as a source of population and health data for policymakers and the research community. In general, the objectives of the BDHS are to: - asses the overall demographic situation in Bangladesh, - assist in the evaluation of the population and health programs in Bangladesh, and - advance survey methodology.
More specifically, the BDHS was designed to: - provide data on the family planning and fertility behavior of the Bangladesh population to evaluate the national family planning programs, - measure changes in fertility and contraceptive prevalence and, at the same time, study the factors which affect these changes, such as marriage patterns, urban/rural residence, availability of contraception, breastfeeding patterns, and other socioeconomic factors, and - examine the basic indicators of maternal and child health in Bangladesh.
National
Sample survey data
Bangladesh is divided into five administrative divisions, 64 districts (zillas), and 489 thanas. In rural areas, thanas are divided into unions and then mauzas, an administrative land unit. Urban areas are divided into wards and then mahallas. The 1993-94 BDHS employed a nationally-representative, two-stage sample. It was selected from the Integrated Multi-Purpose Master Sample (IMPS), newly created by the Bangladesh Bureau of Statistics. The IMPS is based on 1991 census data. Each of the five divisions was stratified into three groups: 1) statistical metropolitan areas (SMAs) 2) municipalities (other urban areas), and 3) rural areas. In rural areas, the primary sampling unit was the mauza, while in urban areas, it was the mahalla. Because the primary sampling units in the IMPS were selected with probability proportional to size from the 1991 census frame, the units for the BDHS were sub-selected from the IMPS with equal probability to make the BDHS selection equivalent to selection with probability proportional to size. A total of 304 primary sampling units were selected for the BDHS (30 in SMAs, 40 in municipalities, and 234 in rural areas), out of the 372 in the IMPS. Fieldwork in three sample points was not possible, so a total of 301 points were covered in the survey.
Since one objective of the BDHS is to provide separate survey estimates for each division as well as for urban and rural areas separately, it was necessary to increase the sampling rate for Barisal Division und for municipalities relative to the other divisions, SMAs, and rural areas. Thus, the BDHS sample is not self-weighting and weighting factors have been applied to the data in this report.
After the selection of the BDHS sample points, field staffs were trained by Mitra and Associates and conducted a household listing operation in September and October 1993. A systematic sample of households was then selected from these lists, with an average "take" of 25 households in the urban clusters and 37 households in rural clusters. Every second household was identified as selected for the husband's survey, meaning that, in addition to interviewing all ever-married women age 10-49, interviewers also interviewed the husband of any woman who was successfully interviewed. It was expected that the sample would yield interviews with approximately 10,000 ever-married women age 10-49 and 4,200 of their husbands.
Note: See detailed in APPENDIX A of the survey final report.
Data collected for women 10-49, indicators calculated for women 15-49. A total of 304 primary sampling units were selected, but fieldwork in 3 sample points was not possible.
Face-to-face
Four types of questionnaires were used for the BDHS: a Household Questionnaire, a Women's Questionnaire, a Husbands' Questionnaire, and a Service Availability Questionnaire. The contents of these questionnaires were based on the DHS Model A Questionnaire, which is designed for use in countries with relatively high levels of contraceptive use. Additions and modifications to the model questionnaires were made during a series of meetings with representatives of various organizations, including the Asia Foundation, the Bangladesh Bureau of Statistics, the Cambridge Consulting Corporation, the Family Planning Association of Bangladesh, GTZ, the International Centre for Diarrhoeal Disease Research (ICDDR,B), Pathfinder International, Population Communications Services, the Population Council, the Social Marketing Company, UNFPA, UNICEF, University Research Corporation/Bangladesh, and the World Bank. The questionnaires were developed in English and then translated into and printed in Bangla.
The Household Questionnaire was used to list all the usual members and visitors of selected households. Some basic information was collected on the characteristics of each person listed, including his/her age, sex, education, and relationship to the head of the household. The main purpose of the Household Questionnaire was to identify women and men who were eligible for individual interview. In addition, information was collected about the dwelling itself, such as the source of water, type of toilet facilities, materials used to construct the house, and ownership of various consumer goods.
The Women's Questionnaire was used to collect information from ever-married women age 10-49. These women were asked questions on the following topics: - Background characteristics (age, education, religion, etc.), - Reproductive history, - Knowledge and use of family planning methods, - Antenatal and delivery care, - Breastfeeding and weaning practices, - Vaccinations and health of children under age three, - Marriage, - Fertility preferences, and - Husband's background and respondent's work.
The Husbands' Questionnaire was used to interview the husbands of a subsample of women who were interviewed. The questionnaire included many of the same questions as the Women's Questionnaire, except that it omitted the detailed birth history, as well as the sections on maternal care, breastfeeding and child health.
The Service Availability Questionnaire was used to collect information on the family planning and health services available in and near the sampled areas. It consisted of a set of three questionnaires: one to collect data on characteristics of the community, one for interviewing family welfare visitors and one for interviewing family planning field workers, whether government or non-governent supported. One set of service availability questionnaires was to be completed in each cluster (sample point).
All questionnaires for the BDHS were returned to Dhaka for data processing at Mitra and Associates. The processing operation consisted of office editing, coding of open-ended questions, data entry, and editing inconsistencies found by the computer programs. One senior staff member, 1 data processing supervisor, questionnaire administrator, 2 office editors, and 5 data entry operators were responsible for the data processing operation. The data were processed on five microcomputers. The DHS data entry and editing programs were written in ISSA (Integrated System for Survey Analysis). Data processing commenced in early February and was completed by late April 1994.
A total of 9,681 households were selected for the sample, of which 9,174 were successfully interviewed. The shortfall is primarily due to dwellings that were vacant, or in which the inhabitants had left for an extended period at the time they were visited by the interviewing teams. Of the 9,255 households that were occupied, 99 percent were successfully interviewed. In these households, 9,900 women were identified as eligible for the individual interview and interviews were completed for 9,640 or 97 percent of these. In one-half of the households that were selected for inclusion in the husbands' survey, 3,874 eligible husbands were identified, of which 3,284 or 85 percent were interviewed.
The principal reason for non-response among eligible women and men was failure to find them at home despite repeated visits to the household. The refusal rate was very low (less than one-tenth of one percent among women and husbands). Since the main reason for interviewing husbands was to match the information with that from their wives, survey procedures called for interviewers not to interview husbands of women who were not interviewed. Such cases account for about one-third of the non-response among husbands. Where husbands and wives were both interviewed, they were interviewed simultaneously but separately.
Note: See summarized response rates by residence (urban/rural) in Table 1.1 of the survey final report.
The estimates from a sample survey are affected by two types of errors: non-sampling errors and sampling errors. Non-sampling errors are the results of mistakes made in implementing data collection and data processing, such as failure to locate and interview the correct household, misunderstanding of the questions
The 1993 Ghana Demographic and Health Survey (GDHS) is a nationally representative survey of 4,562 women age 15-49 and 1,302 men age 15-59. The survey is designed to furnish policymakers, planners and program managers with factual, reliable and up-to-date information on fertility, family planning and the status of maternal and child health care in the country. The survey, which was carried out by the Ghana Statistical Service (GSS), marks Ghana's second participation in the worldwide Demographic and Health Surveys (DHS) program.
The principal objective of the 1993 GDHS is to generate reliable and current information on fertility, mortality, contraception and maternal and child health indicators. Such data are necessary for effective policy formulation as well as program design, monitoring and evaluation. The 1993 GDHS is, in large measure, an update to the 1988 GDHS. Together, the two surveys provide comparable information for two points in time, thus allowing assessment of changes and trends in various demographic and health indicators over time.
Long-term objectives of the survey include (i) strengthening the capacity of the Ghana Statistical Service to plan, conduct, process and analyze data from a complex, large-scale survey such as the Demographic and Health Survey, and (ii) contributing to the ever-expanding international database on demographic and health-related variables.
National
Sample survey data
The 1993 GDHS is a stratified, self-weighting, nationally representative sample of households chosen from 400 Enumeration Areas (EAs). The 1984 Population Census EAs constituted the sampling frame. The frame was first stratified into three ecological zones, namely coastal, forest and savannah, and then into urban and rural EAs. The EAs were selected with probability proportional to the number of households. Households within selected EAs were subsequently listed and a systematic sample of households was selected for the survey. The survey was designed to yield a sample of 5,400 women age 15-49 and a sub-sample of males age 15-59 systematically selected from one-third of the 400 EAs.
Note: See detailed description of sample design in APPENDIX A of the survey report.
Face-to-face
Survey instruments used to elicit information for the 1993 GDHS are 1) Household Schedule 2) Women's Questionnaire and 3) Men's Questionnaire.
The questionnaires were structured based on the Demographic and Health Survey Model B Questionnaire designed for countries with low levels of contraceptive use. The final version of the questionnaires evolved out of a series of meetings with personnel of relevant ministries, institutions and organizations engaged in activities relating to fertility and family planning, health and nutrition and rehabilitation of persons with disabilities.
The questionnaires were first developed in English and later translated and printed in five major local languages, namely: Akan, Dagbani, Ewe, Ga, and Hausa. In the selected households, all usual members and visitors were listed in the household schedule. Background information, such as age, sex, relationship to head of household, marital status and level of education, was collected on each listed person. Questions on economic activity, occupation, industry, employment status, number of days worked in the past week and number of hours worked per day was asked of all persons age seven years and over. Those who did not work during the reference period were asked whether or not they actively looked for work.
Information on the health and disability status of all persons was also collected in the household schedule. Migration history was elicited from all persons age 15 years and over, as well as information on the survival status and residence of natural parents of all children less than 15 years in the household.
Data on source of water supply, type of toilet facility, number of sleeping rooms available to the household, material of floor and ownership of specified durable consumer goods were also elicited.
Finally, the household schedule was the instrument used to identify eligible women and men from whom detailed information was collected during the individual interview.
The women's questionnaire was used to collect information on eligible women identified in the household schedule. Eligible women were defined as those age 15-49 years who are usual members of the household and visitors who spent the night before the interview with the household. Questions asked in the questionnaire were on the following topics:
All female respondents with at least one live birth since January 1990 and their children born since 1st January 1990 had their height and weight taken.
The men's questionnaire was administered to men in sample households in a third of selected EAs. An eligible man was 15-59 years old who is either a usual household member or a visitor who spent the night preceding the day of interview with the household.
Topics enquired about in the men's questionnaire included the following: - Background Characteristics - Reproductive History - Contraceptive Knowledge and Use - Marriage - Fertility Preferences - Knowledge of AIDS and Other STDs.
Questionnaires from the field were sent to the secretariat at the Head Office for checking and office editing. The office editing, which was undertaken by two officers, involved correcting inconsistencies in the questionnaire responses and coding open-ended questions. The questionnaires were then forwarded to the data processing unit for data entry. Data capture and verification were undertaken by four data entry operators. Nearly 20 percent of the questionnaires were verified. This phase of the survey covered four and a half months - that is, from mid-October, 1993 to the end of February, 1994.
After the data entry, three professional staff members performed the secondary editing of questionnaires that were flagged either because entries were inconsistent or values of specific variables were out of range or missing. The secondary editing was completed on 17th March, 1994 and the tables for the preliminary report were generated on 18th March, 1994. The software package used for the data processing was the Integrated System for Survey Analysis (ISSA).
A sample of 6,161 households was selected, from which 5,919 households were contacted for interview. Interviews were successfully completed in 5,822 households, indicating a household response rate of 98 percent. About 3 percent of selected households were absent during the interviewing period, and are excluded from the calculations of the response rate.
Even though the sample was designed to yield interviews with nearly 5,400 women age 15-49 only 4,700 women were identified as eligible for the individual interview. Individual interviews were successfully completed for 4,562 eligible women, giving a response rate of 97 percent. Similarly, instead of the expected 1,700 eligible men being identified in the households only 1,354 eligible men were found and 1,302 of these were successfully interviewed, with a response rate of 96 percent.
The principal reason for non-response among eligible women and men was not finding them at home despite repeated visits to the households. However, refusal rates for both eligible women and men were low, 0.3 percent and 0.2 percent, respectively.
Note: See summarized response rates in Table 1.1 of the survey report.
The results from sample surveys are affected by two types of errors, non-sampling error and sampling error. Non-sampling error is due to mistakes made in carrying out field activities, such as failure to locate and interview the correct household, errors in the way the questions are asked, misunderstanding on the part of either the interviewer or the respondent, data entry errors, etc. Although efforts were made during the design and implementation of the 1993 GDHS to minimize this type of error, non-sampling errors are impossible to avoid and difficult to evaluate statistically.
Sampling errors, on the other hand, can be measured statistically. The sample of eligible women selected in the 1993 GDHS is only one of many samples that could have been selected from the same population, using the same design and expected size. Each one would have yielded results that differed somewhat from the actual sample selected. The sampling error is a measure of the variability between all possible samples; although it is not known exactly, it can be estimated from the survey results.
Sampling error is usually measured in terms of standard error of a particular statistic (mean, percentage, etc.), which is the square root of the variance of the statistic. The standard error can be used to calculate confidence intervals within which, apart from non-sampling errors, the true value for the population can reasonably be assumed to fall. For example, for any given statistic calculated from a sample survey, the value of that same statistic as measured in 95 percent of all possible samples with the same design (and expected size) will fall within a range
https://borealisdata.ca/api/datasets/:persistentId/versions/7.1/customlicense?persistentId=doi:10.7939/DVN/10004https://borealisdata.ca/api/datasets/:persistentId/versions/7.1/customlicense?persistentId=doi:10.7939/DVN/10004
The Population Research Laboratory (PRL), a member of the Association of Academic Survey Research Organizations (AASRO), seeks to advance the research, education and service goals of the University of Alberta by helping academic researchers and policy makers design and implement applied social science research projects. The PRL specializes in the gathering, analysis, and presentation of data about demographic, social and public issues. The PRL research team provides expert consultation and implementation of quantitative and qualitative research methods, project design, sample design, web-based, paper-based and telephone surveys, field site testing, data analysis and report writing. The PRL follows scientifically rigorous and transparent methods in each phase of a research project. Research Coordinators are members of the American Association for Public Opinion Research (AAPOR) and use best practices when conducting all types of research. The PRL has particular expertise in conducting computer-assisted telephone interviews (referred to as CATI surveys). When conducting telephone surveys, all calls are displayed as being from the "U of A PRL", a procedure that assures recipients that the call is not from a telemarketer, and thus helps increase response rates. The PRL maintains a complement of highly skilled telephone interviewers and supervisors who are thoroughly trained in FOIPP requirements, respondent selection procedures, questionnaire instructions, and neutral probing. A subset of interviewers are specially trained to convince otherwise reluctant respondents to participate in the study, a practice that increases response rates and lowers selection bias. PRL staff monitors data collection on a daily basis to allow any necessary adjustments to the volume and timing of calls and respondent selection criteria. The Population Research Laboratory (PRL) administered the 2012 Alberta Survey B. This survey of households across the province of Alberta continues to enable academic researchers, government departments, and non-profit organizations to explore a wide range of topics in a structured research framework and environment. Sponsors' research questions are asked together with demographic questions in a telephone interview of Alberta households. This data consists of the information from 1207 Alberta residence, interviewed between June 5, 2012 and June 27, 2012. The amount of responses indicates that the response rate, as calculated percentages representing the number of people who participated in the survey divided by the number selected in the eligible sample, was 27.6% for survey B. The subject ares included in the 2012 Alberta Survey B includes socio-demographic and background variables such as: household composition, age, gender, marital status, highest level of education, household income, religion, ethnic background, place of birth, employment status, home ownership, political party support and perceptions of financial status. In addition, the topics of public health and injury control, tobacco reduction, activity limitations and personal directives, unions, politics and health.
The Parent Ping research project began in July 2020 and ended in September 2021. The project aimed to document parental life, particular during the pandemic. Participants were recruited via opportunity sampling or via their children's school. Participants downloaded a mobile phone app which asked them roughly three closed survey questions each day.
Question topics included: lockdown, home learning (due to school closures), attitudes to vaccination, family life, home-school relationships etc. Demographic data were also collected from each participant including: their age, their gender, the number of children they have, the gender of their children, whether their child has a special educational need and whether their child is eligible to receive free school meals. In the time the project has run, data were collected from roughly 1,200 questions and around 3,000 UK parents.
About Parent Ping
Parent Ping was created by Education Intelligence, a company that since 2017 has run a daily survey app for teachers in England called Teacher Tapp. When schools closed for the majority of pupils during the 2020 covid pandemic there was a specific need to understand how parents were coping with home educating their children. Education Intelligence were able to secure a small grant from Big Change to replicate the software used to create Teacher Tapp in order to survey parents. The app they built was call Parent Ping.
Key features of the app were:
Parent Ping was live for just over a year, pivotally collecting data during the second national lockdown in early 2021.
Data analysis
Users should note that the individual questions are all included together in the variable 'questiontext' and the answers are in the variable 'answertext'. In order to analyse questions separately, data will need to be filtered/subset using the 'questionid' and 'answerid' variables. The data file contains 2,412,976 lines, which means the tab-separated format cannot be viewed successfully in Excel.
The National Health and Nutrition Examination Surveys (NHANES) is a program of studies designed to assess the health and nutritional status of adults and children in the United States. The NHANES combines personal interviews and physical examinations, which focus on different population groups or health topics. These surveys have been conducted by the National Center for Health Statistics (NCHS) on a periodic basis from 1971 to 1994. In 1999, the NHANES became a continuous program with a changing focus on a variety of health and nutrition measurements which were designed to meet current and emerging concerns. The sample for the survey is selected to represent the U.S. population of all ages. Many of the NHANES 2007-2008 questions also were asked in NHANES II 1976-1980, Hispanic NHANES 1982-1984, NHANES III 1988-1994, and NHANES 1999-2006. New questions were added to the survey based on recommendations from survey collaborators, NCHS staff, and other interagency work groups. Estimates for previously undiagnosed conditions, as well as those known to and reported by survey respondents, are produced through the survey. In the 2003-2004 wave, the NHANES includes more than 100 datasets. Most have been combined into three datasets for convenience. Each starts with the Demographic dataset and includes datasets of a specific type. 1. National Health and Nutrition Examination Survey (NHANES), Demographic & Examination Data, 2003-2004 (The base of the Demographic dataset + all data from medical examinations). 2. National Health and Nutrition Examination Survey (NHANES), Demographic & Laboratory Data, 2003-2004 (The base of the Demographic dataset + all data from medical laboratories). 3. National Health and Nutrition Examination Survey (NHANES), Demographic & Questionnaire Data, 2003-2004 (The base of the Demographic dataset + all data from questionnaires) Variable SEQN is included for merging files within the waves. All data files should be sorted by SEQN. Additional details of the design and content of each survey are available at the NHANES website.
The Public Service Employee Survey was designed to solicit the views of Public Service employees on their work environment and overall job satisfaction. Employees expressed their opinions on their work units, their communications with their supervisors, skills and career aspirations, client services and labour management relations. General information such as age, gender, years of service and province of work were collected and questions were asked on specific themes such as staffing fairness, official languages, health and safety, harassment and discrimination and retention issues. The results were aggregated at the and Public Service-wide levels. After applied disclosure control methods, 9 demographic variables remains on the file. The survey ensures a measurement capacity between the 1999, 2002 and 2005 questionnaires.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Description
The research employed a mixed methods online survey to understand better the meaning, use, and development of academic research software at the University of Illinois Urbana-Champaign. Other objectives include understanding academic research software support and training needs to make projects successful at Illinois, as well as investigating the use of generative AI tools in using and creating research software.
At the beginning of the survey, all participants gave informed consent. The University of Illinois Urbana-Champaign Institutional Review Board (IRB Protocol no.: Project IRB24-0989) reviewed the study and gave it an exempt determination.
Data collection took place from August 2024 to October 2024. Prior to data analysis, identifiable respondent details were removed during the data cleaning process. Not Applicable and Unsure style responses were used for descriptive statistics, but these responses were excluded for inferential statistics.
Survey design
At the beginning of the online survey, a consent form was provided based on guidelines from the University of Illinois Institutional Review Board to the respondents stating the aims of the study, its benefits and risks, ethical guidelines, being a voluntary survey for participation and withdrawal, privacy and confidentiality, data security, estimated time for survey completion, and contact information of researchers for asking questions. Respondents clicked to indicate their consent. Survey questions were divided into four parts: demographic information, using software for research, creating software for research, and the protocol of citing software for research. The survey had to stop points, whereby not all questions applied to respondents, which led to different sample sizes at the stop points. At the opening of the survey, the number of respondents was 251 with the funding demographic question being answered by all respondents, while other demographic questions had between 225 and 228 respondents answering them. For the first stop question, using research software in their research, the total respondents was 212, and at the last stop question, respondents considering themselves to be research developers, the total number of respondents was 74. The last question of the survey was answered by 71 respondents. Respondents may also have left the survey for other reasons. The questions were primarily closed-type questions with single choice, multiple choice, or Likert scale, as well as a few open-ended questions. Likert scale responses were created utilizing validated scales from Vagias' (2006) Likert Type Scale Response Anchors.
Sampling
Survey Respondents’ Demographics
While most respondents were Tenure Track Faculty (34.7%, f=227), other key categories included Principal Investigator (22.4%, f=227) and Research Scientist (12.1%, f=227). Computer Science, Information Science, Mathematics, and Engineering fields combined for 16% (f=228) of the respondents surveyed, but it should be noted the remaining respondents were from various academic fields across campus from various arts, humanities, and social science fields (25%, f=228) to agriculture (10%, f=228), education (5%, f=228), economics (3%, f=228), medical sciences (4%, f=228), and politics and policy/law (1%, f=228). Most respondents were likely to receive funding from various government agencies. A more detailed breakdown of the demographic information can be found in the supplemental figures. Of the 74 respondents who answered whether they were a research software developer, most respondents did not consider themselves a research software developer, with respondents stating Not at All (39%, n=74) and Slightly (22%, n=74). In addition, open-ended questions asked for further detail about research software titles used in research, research software developer challenges, how generative AI assisted in creating research software, and how research software is preserved (e.g., reproducibility).
Table 1: Survey Respondents’ Demographics
Characteristics
Respondent (%)
Age
18-24
25-34
35-44
45-54
55-64
Over 64
Preferred Not Answer
3%
14%
33%
27%
14%
7%
2%
Gender
Woman
Man
Non-binary / non-conforming
Prefer not to answer
49%
44%
2%
4%
Race
Asian
Black or African American
Hispanic or Latino
Middle Eastern or North African (MENA; new)
White
Prefer not to answer
Other
12%
5%
6%
1%
67%
8%
1%
Highest Degree
Bachelors
Masters
Professional degree (e.g., J.D.)
Doctorate
6%
19%
5%
70%
Professional Title
Tenure Track Faculty
Principal Investigator
Research Scientist
Staff
Research Faculty
Other
Teaching Faculty
Postdoc
Research Assistant
Research Software Engineer
35%
22%
12%
8%
7%
4%
4%
4%
2%
2%
Academic Field
Biological Sciences
Other
Agriculture
Engineering
Psychology
Earth Sciences
Physical Sciences
Education
Medical & Health Sciences
Computer Science
Library
Chemical Sciences
Human Society
Economics
Information Science
Environment
Veterinary
Mathematical Sciences
History
Architecture
Politics and Policy
Law
18%
10%
10%
9%
8%
6%
6%
5%
4%3%
3%
3%
3%
3%
2%
2%
2%
2%
1%
1%
1%
0%
Years Since Last Degree
Less than 1 Year
1-2 Years
3-5 Years
6-9 Years
10-15 Years
More than 15 Years
4%
8%
11%
14%
24%
40%
Receive Funding
Yes
No
73%
27%
Funders for Research
Other
National Science Foundation (NSF)
United States Department of Agriculture (USDA)
National Institute of Health (NIH)
Department of Energy (DOE)
Department of Defense (DOD)
Environmental Protection Agency (EPA)
National Aeronautics and Space Administration (NASA)
Bill and Melinda Gates Foundation
Advanced Research Projects Agency - Energy (ARPA-E)
Institute of Education Sciences
Alfred P. Sloan Foundation
W.M. Keck Foundation
Simons Foundation
Gordon and Betty Moore Foundation
Department of Justice (DOJ)
National Endowment for the Humanities (NEH)
Congressionally Directed Medical Research Programs (CDMRP)
Andrew W. Mellon Foundation
22%
18%
18%
11%
9%
5%
4%
4%
2%
2%
1%
1%
1%
1%
1%
1%
0%
0%
0%
Table 2: Survey Codebook
QuestionID
Variable
Variable Label
Survey Item
Response Options
1
age
Respondent’s Age
Section Header:
Demographics Thank you for your participation in this survey today! Before you begin to answer questions about academic research software, please answer a few demographic questions to better contextualize your responses to other survey questions.
What is your age?
Select one choice.
Years
1-Under 18
2-18-24
3-25-34
4-35-44
5-45-54
6-55-64
7-Over 64
8-Prefer not to answer
2
gender
Respondent’s Gender
What is your gender?
Select one choice.
1-Female
2-Male
3-Transgender
4-Non-binary / non-conforming
5-Prefer not to answer
6-Other:
3
race
Respondent’s Race
What is your race?
Select one choice.
1-American Indian or Alaska Native
2-Asian
3-Black or African American
4-Hispanic or Latino
5-Middle Eastern or North African (MENA; new)
6-Native Hawaiian or Pacific Islander
7-White
8-Prefer not to answer
9-Other:
4
highest_degree
Respondent’s Highest Degree
What is the highest degree you have completed?
Select one choice.
1-None
2-High school
3-Associate
4-Bachelor's
5-Master's
6-Professional degree (e.g., J.D.)
7-Doctorate
8-Other:
5
professional_title
Respondent’s Professional Title
What is your professional title?
Select all that apply.
1-professional_title_1
Principal Investigator
2-professional_title_2
Tenure Track Faculty
3-professional_title_3
Teaching Faculty
4-professional_title_4
Research Faculty
5-professional_title_5
Research Scientist
6-professional_title_6
Research Software Engineer
7-professional_title_7
Staff
8-professional_title_8
Postdoc
9-professional_title_9
Research Assistant
10-professional_title_10
Other:
6
academic_field
Respondent’s most strongly identified Academic Field
What is the academic field or discipline you most strongly identify with (e.g., Psychology, Computer Science)?
Select one choice.
1-Chemical sciences
2-Biological sciences
3-Medical & health sciences
4-Physical sciences
5-Mathematical sciences
6-Earth sciences
7-Agriculture
8-Veterinary
9-Environment
10-Psychology
11-Law
12-Philosophy
13-Economics
14-Human society
15-Journalism
16-Library
17-Education
18-Art & Design Management
19-Engineering
20-Language
21-History
22-Politics and policy
23-Architecture
24-Computer Science
25-Information science
26-Other:
7
years_since_last_degree
Number of years since last respondent’s last degree
How many years since the award of your last completed degree?
Select one choice.
1-Less than 1 year
2-1-2 years
3-3-5 years
4-6-9 years
5-10-15 years
6-More than 15 years
8
receive_funding_for_research
Whether respondent received funding for research
Do you receive funding for your research?
1-Yes
0-No
9
funders_for_research
Respondent’s funding sources if they answered yes in Question 8
Who funds your research or work (e.g., NIH, Gates Foundation)?
Select all that apply.
1-funders_for_research_1
United States Department of Agriculture (USDA)
2-funders_for_research_2
Department of Energy (DOE)
3-funders_for_research_3
National Science
The principal objective of the Republic of the Marshall Islands 2007 Demographic and Health Survey (2007 RMIDHS) is to provide current and reliable data on fertility and family planning behavior, child mortality, adult and maternal mortality, children’s nutritional status, the utilization of maternal and child health services, and knowledge of HIV and AIDS. The specific objectives of the survey are to: • collect data at the national level that will allow the calculation of key demographic rates; • analyze the direct and indirect factors that determine the level and trends of fertility; • measure the level of contraceptive knowledge and practice among women and men by method, urban/rural residence, and region; • collect high-quality data on family health, including immunization coverage among children, prevalence and treatment of diarrhea and other diseases among children under five, and maternity care indicators (including antenatal visits, assistance at delivery, and postnatal care); • collect data on infant and child mortality; • obtain data on child feeding practices, including breastfeeding, and collect ‘observation’ information to use in assessing the nutritional status of women and children; • collect data on knowledge and attitudes of women and men about sexually transmitted infections (STIs), HIV and AIDS and evaluate patterns of recent behavior regarding condom use; and • collect data on support to mentally ill persons and information on the incidence of suicide.
This information is essential for informed policy decisions, planning, monitoring, and evaluation of programs on health in general and reproductive health in particular at both national level and in urban and rural areas. A long-term objective of the survey is to strengthen the technical capacity of government organizations to plan, conduct, process, and analyze data from complex national population and health surveys. Moreover, the 2007 RMIDHS provides national, rural, and urban estimates on population and health that are comparable to data collected in similar surveys in other Pacific DHS pilot countries and other developing countries.
National
Sample survey data [ssd]
The primary focus of the 2007 RMIDHS was to provide estimates of key population and health indicators, including fertility and mortality rates, for the country as a whole and for urban and rural areas separately. The survey used the sampling frame provided by the list of census enumeration areas, with population and household information from the 1999 RMI Census and the 2006 Community Survey.
The survey was designed to obtain completed interviews of 1,070 women aged 15-49. In addition, males aged 15-59 in every second household were interviewed. To take non-response into account, a total of 608 households countrywide were selected: 295 in urban areas and 313 in rural areas.
Face-to-face [f2f]
Three questionnaires were administered for the 2007 RMIDHS: a household questionnaire, a women’s questionnaire, and a men’s questionnaire. These were adapted to reflect population and health issues relevant to the Marshall Islands at a series of meetings with various stakeholders from government ministries and agencies, non-governmental organizations (NGOs) and international donors. The final draft of the questionnaires was discussed at a questionnaire design workshop organized by EPPSO in September 2006 in Majuro. The survey questionnaires were then translated into the local language (Marshallese) and pretested from November 16 to December 13, 2006.
The household questionnaire was used to list all the usual members and visitors in the selected households and to identify women and men who were eligible for the individual interview. Some basic information was collected on the characteristics of each person listed, including age, sex, education, and relationship to the head of the household. For children under age 18, the survival status of their parents was determined. The household questionnaire also collected information on characteristics of the household’s dwelling unit, such as the source of water, type of toilet facilities, materials used for the floor of the house, and ownership of various durable goods. Additionally, it was used to record information on mental illness and suicide experiences of members of the household.
The women’s questionnaire was used to collect information from all women aged 15–49. The women were asked questions on: • characteristics such as education, residential history, and media exposure; • pregnancy history and childhood mortality; • knowledge and use of family planning methods; • fertility preferences; • antenatal, delivery, and postnatal care; • breastfeeding and infant feeding practices; • immunization and childhood illnesses; • marriage and sexual activity; • their own work and their husband’s background characteristics; and • awareness and behavior regarding HIV and other STIs.
The men’s questionnaire was administered to all men aged 15–59 living in every second household in the 2007 RMIDHS sample. It collected much of the same information found in the women’s questionnaire, but was shorter because it did not contain a detailed reproductive history or questions on maternal and child health or nutrition.
The processing of the 2007 RMIDHS results began soon after the start of fieldwork. Completed questionnaires were returned periodically from the field to the EPPSO data processing center in Majuro, where they were entered and edited by four data processing personnel specially trained for this task. The data processing personnel were supervised by EPPSO staff. The concurrent processing of the data was an advantage since field check tables were generated early on to monitor various data quality parameters. As a result, specific and ongoing feedback was given to the field teams to improve performance. The data entry and editing of the questionnaires was completed by June 30, 2007. Data processing was done using CSPro.
A total of 1,141 households were selected for the sample, of which 1,131 were found to be occupied during data collection. Of these existing households, 1,106 were successfully interviewed, giving a household response rate of 98 percent.
In the households, 1,742 women were identified as eligible for the individual interview. Interviews were completed with 1,625 women, yielding a response rate of 93 percent. Of the 1,218 eligible men identified in the selected sub-sample of households, 87 percent were successfully interviewed. Response rates were higher in rural than urban areas, with the rural–urban difference in response rates most marked among eligible men.
Data Quality Tables - Household age distribution - Age distribution of eligible and interviewed women - Age distribution of eligible and interviewed men - Completeness of reporting - Births by calendar years - Reporting of age at death in days - Reporting of age at death in months
Note: See detailed tables in APPENDIX D of the final survey report.
The Bangladesh Demographic and Health Survey (BDHS) is part of the worldwide Demographic and Health Surveys program, which is designed to collect data on fertility, family planning, and maternal and child health.
The BDHS is intended to serve as a source of population and health data for policymakers and the research community. In general, the objectives of the BDHS are to: - assess the overall demographic situation in Bangladesh, - assist in the evaluation of the population and health programs in Bangladesh, and - advance survey methodology.
More specifically, the objective of the BDHS is to provide up-to-date information on fertility and childhood mortality levels; nuptiality; fertility preferences; awareness, approval, and use of family planning methods; breastfeeding practices; nutrition levels; and maternal and child health. This information is intended to assist policymakers and administrators in evaluating and designing programs and strategies for improving health and family planning services in the country.
National
Sample survey data
Bangladesh is divided into six administrative divisions, 64 districts (zillas), and 490 thanas. In rural areas, thanas are divided into unions and then mauzas, a land administrative unit. Urban areas are divided into wards and then mahallas. The 1996-97 BDHS employed a nationally-representative, two-stage sample that was selected from the Integrated Multi-Purpose Master Sample (IMPS) maintained by the Bangladesh Bureau of Statistics. Each division was stratified into three groups: 1 ) statistical metropolitan areas (SMAs), 2) municipalities (other urban areas), and 3) rural areas. 3 In the rural areas, the primary sampling unit was the mauza, while in urban areas, it was the mahalla. Because the primary sampling units in the IMPS were selected with probability proportional to size from the 1991 Census frame, the units for the BDHS were sub-selected from the IMPS with equal probability so as to retain the overall probability proportional to size. A total of 316 primary sampling units were utilized for the BDHS (30 in SMAs, 42 in municipalities, and 244 in rural areas). In order to highlight changes in survey indicators over time, the 1996-97 BDHS utilized the same sample points (though not necessarily the same households) that were selected for the 1993-94 BDHS, except for 12 additional sample points in the new division of Sylhet. Fieldwork in three sample points was not possible (one in Dhaka Cantonment and two in the Chittagong Hill Tracts), so a total of 313 points were covered.
Since one objective of the BDHS is to provide separate estimates for each division as well as for urban and rural areas separately, it was necessary to increase the sampling rate for Barisal and Sylhet Divisions and for municipalities relative to the other divisions, SMAs and rural areas. Thus, the BDHS sample is not self-weighting and weighting factors have been applied to the data in this report.
Mitra and Associates conducted a household listing operation in all the sample points from 15 September to 15 December 1996. A systematic sample of 9,099 households was then selected from these lists. Every second household was selected for the men's survey, meaning that, in addition to interviewing all ever-married women age 10-49, interviewers also interviewed all currently married men age 15-59. It was expected that the sample would yield interviews with approximately 10,000 ever-married women age 10-49 and 3,000 currently married men age 15-59.
Note: See detailed in APPENDIX A of the survey report.
Face-to-face
Four types of questionnaires were used for the BDHS: a Household Questionnaire, a Women's Questionnaire, a Men' s Questionnaire and a Community Questionnaire. The contents of these questionnaires were based on the DHS Model A Questionnaire, which is designed for use in countries with relatively high levels of contraceptive use. These model questionnaires were adapted for use in Bangladesh during a series of meetings with a small Technical Task Force that consisted of representatives from NIPORT, Mitra and Associates, USAID/Bangladesh, the International Centre for Diarrhoeal Disease Research, Bangladesh (ICDDR,B), Population Council/Dhaka, and Macro International Inc (see Appendix D for a list of members). Draft questionnaires were then circulated to other interested groups and were reviewed by the BDHS Technical Review Committee (see Appendix D for list of members). The questionnaires were developed in English and then translated into and printed in Bangla (see Appendix E for final version in English).
The Household Questionnaire was used to list all the usual members and visitors in the selected households. Some basic information was collected on the characteristics of each person listed, including his/her age, sex, education, and relationship to the head of the household. The main purpose of the Household Questionnaire was to identify women and men who were eligible for the individual interview. In addition, information was collected about the dwelling itself, such as the source of water, type of toilet facilities, materials used to construct the house, and ownership of various consumer goods.
The Women's Questionnaire was used to collect information from ever-married women age 10-49. These women were asked questions on the following topics: - Background characteristics (age, education, religion, etc.), - Reproductive history, - Knowledge and use of family planning methods, - Antenatal and delivery care, - Breastfeeding and weaning practices, - Vaccinations and health of children under age five, - Marriage, - Fertility preferences, - Husband's background and respondent's work, - Knowledge of AIDS, - Height and weight of children under age five and their mothers.
The Men's Questionnaire was used to interview currently married men age 15-59. It was similar to that for women except that it omitted the sections on reproductive history, antenatal and delivery care, breastfeeding, vaccinations, and height and weight. The Community Questionnaire was completed for each sample point and included questions about the existence in the community of income-generating activities and other development organizations and the availability of health and family planning services.
A total of 9,099 households were selected for the sample, of which 8,682 were successfully interviewed. The shortfall is primarily due to dwellings that were vacant or in which the inhabitants had left for an extended period at the time they were visited by the interviewing teams. Of the 8,762 households occupied, 99 percent were successfully interviewed. In these households, 9,335 women were identified as eligible for the individual interview (i.e., ever-married and age 10-49) and interviews were completed for 9,127 or 98 percent of them. In the half of the households that were selected for inclusion in the men's survey, 3,611 eligible ever-married men age 15-59 were identified, of whom 3,346 or 93 percent were interviewed.
The principal reason for non-response among eligible women and men was the failure to find them at home despite repeated visits to the household. The refusal rate was low.
Note: See summarized response rates by residence (urban/rural) in Table 1.1 of the survey report.
The estimates from a sample survey are affected by two types of errors: (1) non-sampling errors, and (2) sampling errors. Non-sampling errors are the results of mistakes made in implementing data collection and data processing, such as failure to locate and interview the correct household, misunderstanding of the questions on the part of either the interviewer or the respondent, and data entry errors. Although numerous efforts were made during the implementation of the BDHS to minimize this type of error, non-sampling errors are impossible to avoid and difficult to evaluate statistically.
Sampling errors, on the other hand, can be evaluated statistically. The sample of respondents selected in the BDHS is only one of many samples that could have been selected from the same population, using the same design and expected size. Each of these samples would yield results that differ somewhat from the results of the actual sample selected. Sampling errors are a measure of the variability between all possible samples. Although the degree of variability is not known exactly, it can be estimated from the survey results.
A sampling error is usually measured in terms of the standard error for a particular statistic (mean, percentage, etc.), which is the square root of the variance. The standard error can be used to calculate confidence intervals within which the true value for the population can reasonably be assumed to fall. For example, for any given statistic calculated from a sample survey, the value of that statistic will fall within a range of plus or minus two times the standard error of that statistic in 95 percent of all possible samples of identical size and design.
If the sample of respondents had been selected as a simple random sample, it would have been possible to use straightforward formulas for calculating sampling errors. However, the BDHS sample is the result of a two-stage stratified design, and, consequently, it was necessary to use more complex formulae. The computer software used to calculate sampling errors for the BDHS is the ISSA Sampling Error Module. This module used the Taylor
The 2022 Kenya Demographic and Health Survey (2022 KDHS) is the seventh DHS survey implemented in Kenya. The Kenya National Bureau of Statistics (KNBS) in collaboration with the Ministry of Health (MoH) and other stakeholders implemented the survey. Survey planning began in late 2020 with data collection taking place from February 17 to July 19, 2022. ICF provided technical assistance through The DHS Program, which is funded by the United States Agency for International Development (USAID) and offers financial support and technical assistance for population and health surveys in countries worldwide. Other agencies and organizations that facilitated the successful implementation of the survey through technical or financial support were the Bill & Melinda Gates Foundation, the World Bank, the United Nations Children's Fund (UNICEF), the United Nations Population Fund (UNFPA), Nutrition International, the World Food Programme (WFP), the United Nations Entity for Gender Equality and the Empowerment of Women (UN Women), the World Health Organization (WHO), the Clinton Health Access Initiative, and the Joint United Nations Programme on HIV/AIDS (UNAIDS).
SURVEY OBJECTIVES The primary objective of the 2022 KDHS is to provide up-to-date estimates of demographic, health, and nutrition indicators to guide the planning, implementation, monitoring, and evaluation of population and health-related programs at the national and county levels. The specific objectives of the 2022 KDHS are to: Estimate fertility levels and contraceptive prevalence Estimate childhood mortality Provide basic indicators of maternal and child health Estimate the Early Childhood Development Index (ECDI) Collect anthropometric measures for children, women, and men Collect information on children's nutrition Collect information on women's dietary diversity Obtain information on knowledge and behavior related to transmission of HIV and other sexually transmitted infections (STIs) Obtain information on noncommunicable diseases and other health issues Ascertain the extent and patterns of domestic violence and female genital mutilation/cutting
National coverage
Household, individuals, county and national level
The survey covered sampled households
The sample for the 2022 KDHS was drawn from the Kenya Household Master Sample Frame (K-HMSF). This is the frame that KNBS currently operates to conduct household-based sample surveys in Kenya. In 2019, Kenya conducted a Population and Housing Census, and a total of 129,067 enumeration areas (EAs) were developed. Of these EAs, 10,000 were selected with probability proportional to size to create the K-HMSF. The 10,000 EAs were randomized into four equal subsamples. The survey sample was drawn from one of the four subsamples. The EAs were developed into clusters through a process of household listing and geo-referencing. To design the frame, each of the 47 counties in Kenya was stratified into rural and urban strata, resulting in 92 strata since Nairobi City and Mombasa counties are purely urban.
The 2022 KDHS was designed to provide estimates at the national level, for rural and urban areas, and, for some indicators, at the county level. Given this, the sample was designed to have 42,300 households, with 25 households selected per cluster, resulting into 1,692 clusters spread across the country with 1,026 clusters in rural areas and 666 in urban areas.
Computer Assisted Personal Interview [capi]
Eight questionnaires were used for the 2022 KDHS: 1. A full Household Questionnaire 2. A short Household Questionnaire 3. A full Woman's Questionnaire 4. A short Woman's Questionnaire 5. A Man's Questionnaire 6. A full Biomarker Questionnaire 7. A short Biomarker Questionnaire 8. A Fieldworker Questionnaire.
The Household Questionnaire collected information on: o Background characteristics of each person in the household (for example, name, sex, age, education, relationship to the household head, survival of parents among children under age 18) o Disability o Assets, land ownership, and housing characteristics o Sanitation, water, and other environmental health issues o Health expenditures o Accident and injury o COVID-19 (prevalence, vaccination, and related deaths) o Household food consumption
The Woman's Questionnaire was used to collect information from women age 15-49 on the following topics: o Socioeconomic and demographic characteristics o Reproduction o Family planning o Maternal health care and breastfeeding o Vaccination and health of children o Children's nutrition o Woman's dietary diversity o Early childhood development o Marriage and sexual activity o Fertility preferences o Husbands' background characteristics and women's employment activity o HIV/AIDS, other sexually transmitted infections (STIs), and tuberculosis (TB) o Other health issues o Early Childhood Development Index 2030 o Chronic diseases o Female genital mutilation/cutting o Domestic violence
The Man's Questionnaire was administered to men age 15-54 living in the households selected for long Household Questionnaires. The questionnaire collected information on: o Socioeconomic and demographic characteristics o Reproduction o Family planning o Marriage and sexual activity o Fertility preferences o Employment and gender roles o HIV/AIDS, other STIs, and TB o Other health issues o Chronic diseases o Female genital mutilation/cutting o Domestic violence
The Biomarker Questionnaire collected information on anthropometry (weight and height). The long Biomarker Questionnaire collected anthropometry measurements for children age 0-59 months, women age 15-49, and men age 15-54, while the short questionnaire collected weight and height measurements only for children age 0-59 months.
The Fieldworker Questionnaire was used to collect basic background information on the people who collected data in the field. This included team supervisors, interviewers, and biomarker technicians.
All questionnaires except the Fieldworker Questionnaire were translated into the Swahili language to make it easier for interviewers to ask questions in a language that respondents could understand.
Data were downloaded from the central servers and checked against the inventory of expected returns to account for all data collected in the field. SyncCloud was also used to generate field check tables to monitor progress and flag any errors, which were communicated back to the field teams for correction.
Secondary editing was done by members of the central office team, who resolved any errors that were not corrected by field teams during data collection. A CSPro batch editing tool was used for cleaning and tabulation during data analysis.
A total of 42,022 households were selected for the sample, of which 38,731 (92%) were found to be occupied. Among the occupied households, 37,911 were successfully interviewed, yielding a response rate of 98%. The response rates for urban and rural households were 96% and 99%, respectively. In the interviewed households, 33,879 women age 15-49 were identified as eligible for individual interviews. Interviews were completed with 32,156 women, yielding a response rate of 95%. The response rates among women selected for the full and short questionnaires were the similar (95%). In the households selected for the male survey, 16,552 men age 15-54 were identified as eligible for individual interviews and 14,453 were successfully interviewed, yielding a response rate of 87%.
The National Health and Nutrition Examination Surveys (NHANES) is a program of studies designed to assess the health and nutritional status of adults and children in the United States. The NHANES combines personal interviews and physical examinations, which focus on different population groups or health topics. These surveys have been conducted by the National Center for Health Statistics (NCHS) on a periodic basis from 1971 to 1994. In 1999, the NHANES became a continuous program with a changing focus on a variety of health and nutrition measurements which were designed to meet current and emerging concerns. The sample for the survey is selected to represent the U.S. population of all ages. Many of the NHANES 2007-2008 questions also were asked in NHANES II 1976-1980, Hispanic HANES 1982-1984, NHANES III 1988-1994, and NHANES 1999-2006. New questions were added to the survey based on recommendations from survey collaborators, NCHS staff, and other interagency work groups. Estimates for previously undiagnosed conditions, as well as those known to and reported by survey respondents, are produced through the survey.
In the 2003-2004 wave, the NHANES includes over 100 datasets. Most have been combined into three datasets for convenience. Each starts with the Demographic dataset and includes datasets of a specific type.
1. National Health and Nutrition Examination Survey (NHANES), Demographic & Examination Data, 2003-2004 (The base of the Demographic dataset + all data from medical examinations).
2. National Health and Nutrition Examination Survey (NHANES), Demographic & Laboratory Data, 2003-2004 (The base of the Demographic dataset + all data from medical laboratories).
3. National Health and Nutrition Examination Survey (NHANES), Demographic & Questionnaire Data, 2003-2004 (The base of the Demographic dataset + all data from questionnaires)
Variable SEQN is included for merging files within the waves. All data files should be sorted by SEQN.
Additional details of the design and content of each survey are available at the "https://www.cdc.gov/nchs/nhanes/index.html" Target="_blank">NHANES website.
The 1991 Indonesia Demographic and Health Survey (IDHS) is a nationally representative survey of ever-married women age 15-49. It was conducted between May and July 1991. The survey was designed to provide information on levels and trends of fertility, infant and child mortality, family planning and maternal and child health. The IDHS was carried out as collaboration between the Central Bureau of Statistics, the National Family Planning Coordinating Board, and the Ministry of Health. The IDHS is follow-on to the National Indonesia Contraceptive Prevalence Survey conducted in 1987.
The DHS program has four general objectives: - To provide participating countries with data and analysis useful for informed policy choices; - To expand the international population and health database; - To advance survey methodology; and - To help develop in participating countries the technical skills and resources necessary to conduct demographic and health surveys.
In 1987 the National Indonesia Contraceptive Prevalence Survey (NICPS) was conducted in 20 of the 27 provinces in Indonesia, as part of Phase I of the DHS program. This survey did not include questions related to health since the Central Bureau of Statistics (CBS) had collected that information in the 1987 National Socioeconomic Household Survey (SUSENAS). The 1991 Indonesia Demographic and Health Survey (IDHS) was conducted in all 27 provinces of Indonesia as part of Phase II of the DHS program. The IDHS received financial assistance from several sources.
The 1991 IDHS was specifically designed to meet the following objectives: - To provide data concerning fertility, family planning, and maternal and child health that can be used by program managers, policymakers, and researchers to evaluate and improve existing programs; - To measure changes in fertility and contraceptive prevalence rates and at the same time study factors which affect the change, such as marriage patterns, urban/rural residence, education, breastfeeding habits, and the availability of contraception; - To measure the development and achievements of programs related to health policy, particularly those concerning the maternal and child health development program implemented through public health clinics in Indonesia.
National
Sample survey data [ssd]
Indonesia is divided into 27 provinces. For the implementation of its family planning program, the National Family Planning Coordinating Board (BKKBN) has divided these provinces into three regions as follows:
The 1990 Population Census of Indonesia shows that Java-Bali contains about 62 percent of the national population, while Outer Java-Bali I contains 27 percent and Outer Java-Bali II contains 11 percent. The sample for the Indonesia DHS survey was designed to produce reliable estimates of contraceptive prevalence and several other major survey variables for each of the 27 provinces and for urban and rural areas of the three regions.
In order to accomplish this goal, approximately 1500 to 2000 households were selected in each of the provinces in Java-Bali, 1000 households in each of the ten provinces in Outer Java-Bali I, and 500 households in each of the 11 provinces in Outer Java-Bali II for a total of 28,000 households. With an average of 0.8 eligible women (ever-married women age 15-49) per selected household, the 28,000 households were expected to yield approximately 23,000 individual interviews.
Note: See detailed description of sample design in APPENDIX A of the survey report.
Face-to-face [f2f]
The DHS model "A" questionnaire and manuals were modified to meet the requirements of measuring family planning and health program attainment, and were translated into Bahasa Indonesia.
The first stage of data editing was done by the field editors who checked the completed questionnaires for completeness and accuracy. Field supervisors also checked the questionnaires. They were then sent to the central office in Jakarta where they were edited again and open-ended questions were coded. The data were processed using 11 microcomputers and ISSA (Integrated System for Survey Analysis).
Data entry and editing were initiated almost immediately after the beginning of fieldwork. Simple range and skip errors were corrected at the data entry stage. Secondary machine editing of the data was initiated as soon as sufficient questionnaires had been entered. The objective of the secondary editing was to detect and correct, if possible, inconsistencies in the data. All of the data were entered and edited by September 1991. A brief report containing preliminary survey results was published in November 1991.
Of 28,141 households sampled, 27,109 were eligible to be interviewed (excluding those that were absent, vacant, or destroyed), and of these, 26,858 or 99 percent of eligible households were successfully interviewed. In the interviewed households, 23,470 eligible women were found and complete interviews were obtained with 98 percent of these women.
Note: See summarized response rates by place of residence in Table 1.2 of the survey report.
The results from sample surveys are affected by two types of errors, non-sampling error and sampling error. Non-sampling error is due to mistakes made in carrying out field activities, such as failure to locate and interview the correct household, errors in the way the questions are asked, misunderstanding on the part of either the interviewer or the respondent, data entry errors, etc. Although efforts were made during the design and implementation of the IDHS to minimize this type of error, non-sampling errors are impossible to avoid and difficult to evaluate analytically.
Sampling errors, on the other hand, can be measured statistically. The sample of women selected in the IDHS is only one of many samples that could have been selected from the same population, using the same design and expected size. Each one would have yielded results that differed somewhat from the actual sample selected. The sampling error is a measure of the variability between all possible samples; although it is not known exactly, it can be estimated from the survey results. Sampling error is usually measured in terms of standard error of a particular statistic (mean, percentage, etc.), which is the square root of the variance. The standard error can be used to calculate confidence intervals within which one can reasonably be assured that, apart from non-sampling errors, the true value of the variable for the whole population falls. For example, for any given statistic calculated from a sample survey, the value of that same statistic as measured in 95 percent of all possible samples with the same design (and expected size) will fall within a range of plus or minus two times the standard error of that statistic.
If the sample of women had been selected as a simple random sample, it would have been possible to use straightforward formulas for calculating sampling errors. However, the IDHS sample design depended on stratification, stages and clusters. Consequently, it was necessary to utilize more complex formulas. The computer package CLUSTERS, developed by the International Statistical Institute for the World Fertility Survey, was used to assist in computing the sampling errors with the proper statistical methodology.
Note: See detailed estimate of sampling error calculation in APPENDIX B of the survey report.
Data Quality Tables - Household age distribution - Age distribution of eligible and interviewed women - Completeness of reporting - Births by calendar year since birth - Reporting of age at death in days - Reporting of age at death in months
Note: See detailed tables in APPENDIX C of the survey report.
Abstract copyright UK Data Service and data collection copyright owner.
The Online Time Use Survey (OTUS) was developed by the Office for National Statistics to help improve the measurement of unpaid household production and caring activities that are not captured within traditional economic measures, and to understand better time use from a well-being and quality of life perspective.
The survey collects information from adults aged 18 years and over who are randomly sampled from the NatCen Opinion Panel, which is representative of the UK population. Data collected between March 2020 and March 2021 covers Great Britain and data collected from March 2022 onwards covers the United Kingdom.
Participants were issued with two pre-allocated diary days (one on a weekday and one on a weekend day). They were asked to record their main activities (in 10-minute intervals) and up to five secondary activities (in five-minute intervals) in every 24 hours within an online diary tool. Respondents were able to select activities from a pre-defined list. They were also asked to rate how much they enjoyed different activities. In addition, respondents were asked to complete a demographic questionnaire which records personal and household characteristics.
Latest edition information
For the third edition (August 2024), data and documentation for Wave 8 (9 to 17 March 2024) were added to the study.
The annual data files include the following variables:
The Liberia Demographic and Health Survey (LDHS) was conducted as part of the worldwide Demographic and Health Surveys (DHS) program, in which surveys are being carried out in countries in Africa, Asia, Latin America, and the Middle East. Liberia was the second country to conduct a DHS and the first country in Africa to do so. THe LDHS was a national-level survey conducted from February to July 1986, covering a sample of 5,239 women aged 15 to 49.
The major objective of the LDHS was to provide data on fertility, family planning and maternal and child health to planners and policymakers in Liberia for use in designing and evaluating programs. Although a fair amount of demographic data was available from censuses and surveys, almost no information existed concerning family planning, health, or the determinants of fertility, and the data that did exist were drawn from small-scale, sub-national studies. Thus, there was a need for data to make informed policy choices for family planning and health projects.
A more specific objective was to provide baseline data for the Southeast Region Primary Health Care Project. In order to effectively plan strategies and to eventually evaluate the progress of the project in meeting its goals, there was need for data to indicate the health situation in the two target counties prior to the implementation of the project. Many of the desired topics, such as immunizations, family planning use, and prenatal care, were already incorporated into the model DHS questionnaire; nevertheless, the LDHS was able to better accommodate the needs of this project by adding several questions and by oversampling women living in Sinoe and Grand Gedeh Counties.
Another important goal of the LDHS was to enhance tile skills of those participating in the project for conducting high-quality surveys in the future. Finally, the contribution of Liberian data to an expanding international dataset was also an objective of the LDHS.
National
Sample survey data
The sample for the Liberia Demographic and Health Survey was based on the sampling frame of about 4,500 censal enumeration areas (EAs) that were created for the 1984 Population Census. It was decided to eliminate very remote EAs prior to selecting the sample. The definition of remoteness used was "any EA in which the largest village was estimated to be more than 3-4 hours' walk from a road." According to the 1984 census, the excluded areas represent less than 3 percent of the total number of households in the country. Since the major analytic objective of the LDHS was to adequately estimate basic demographic and health indicators including fertility, mortality, and contraceptive prevalence for the whole country and the two sub-universes (Since and Grand Gedeh Counties), it was decided to oversample these two counties. Consequently, three explicit sub-universes of EAs were created: (1) Since County, (2) Grand Gedeh County, and (3) the rest of the country.
The design provided a self-weighted sample within each sub-universe, but, because of the oversampling in Sinoe and Grand Gedeh Counties, the sample is not self-weighting at the national level. Eligible respondents for the survey were women aged 15-49 years who were present the night before the interview in any of the households included in the sample selected for the LDHS.
The total sample size was expected to be about 6,000 women aged 15-49 with a target by sub-universe of 1,000 each in Sinoe and Grand Gedeh Counties and 4,000 in the rest of the country. It was decided that a sample of approximately 5,500 households selected through a two-stage procedure would be appropriate to reach those objectives. Sampling was carried out independently in each sub-universe. In the rest of the country sub-universe, counties were arranged for selection in serpentine order from the northwest (Cape Mount County) to the southeast (Maryland County). In the first stage EAs were selected systematically with probability proportional to size (size = number of households in 1984). Twenty-four EAs were selected in each of Sinoe and Grand Gedeh Counties and 108 EAs in the rest of the country.
See full sample procedure in the survey final report.
Face-to-face
The Liberia Demographic and Health Survey (LDHS) utilized two questionnaires: One to list members of the selected households (Household Questionnaire) and the other to record information from all women aged 15-49 who were present in the selected households the night before the interview (Individual Questionnaire).
Both questionnaires were produced in Liberian English and were pretested in September 1985. The Individual Questionnaire was an early version of the DHS model questionnaire. It covered three main topics: (1) fertility, including a birth history and questions concerning desires for future childbearing, (2) family planning knowledge and use, and (3) family health, including prevalence of childhood diseases, immunizations for children under age five, and breasffeeding and weaning practices.
Data from the questionnaires were entered onto microcomputers at the Bureau of Statistics office in Monrovia. The data were then subjected to extensive checks for consistency and accuracy.
Errors detected during this operation were resolved either by referring to the original questionnaire, or, in some cases, by logical inference from other information given in the record. Finally, dates were imputed for the small number of cases where complete dates of important events were not given.
Out of the total of 6,1306 households selected, 14.5 percent were found not to be valid households in the field, either because the dwelling had been vacated or destroyed, or the household could not be located or did not exist. Of the 5,609 households that were found to exist, 90 percent were successfully interviewed. In the households that were interviewed, a total of 5,340 women were identified as being eligible for individual interview (that is, they were aged 15-49 and had spent the night before the interview in the selected household). This represents an average of slightly over one eligible woman per household.
The response rate for eligible women was 98 percent. The main reason for nonresponse was the absence of the woman. Similar data are presented by sample subuniverse.
The results from sample surveys are affected by two types of errors: (1) nonsampling error and (2) sampling error. Nonsampling error is due to mistakes made in carrying out field activities, such as failure to locate and interview the correct household, errors in the way questions are asked, misunderstanding of the questions on the part of either the interviewer or the respondent, data entry errors, etc. Although efforts were made during the design and implementation of the Liberia Demographic and Health Survey to minimize this type of error, nonsampling errors are impossible to avoid and difficult to evaluate statistically.
The sample of women selected in the LDHS is only one of many samples of the same size that could have been selected from the same population, using the same design. Each one would have yielded results that differed somewhat from the actual sample selected. The variability observed between all possible samples constitutes sampling error, which, although it is not known exactly, can be estimated from the survey results. Sampling error is usually measured in terms of the "standard error" of a particular statistic (mean, percentage, etc.), which is the square root of the variance of the statistic across all possible samples of equal size and design.
The standard error can be used to calculate confidence intervals within which one can be reasonably assured the true value of the variable for the whole population falls. For example, for any given statistic calculated from a sample survey, the value of that same statistic as measured in 95 percent of all possible samples of identical size and design will fall within a range of plus or minus two times the standard error of that statistic.
If the sample of women had been selected as a simple random sample, it would have been possible to use straightforward formulas for calculating sampling errors. However, the LDHS sample design depended on stratification, stages, and clusters and consequently, it was necessary to utilize more complex formulas. The computer package CLUSTERS was used to assist in computing the sampling errors with the proper statistical methodology.
Information on the completeness of date reporting is of interest in assessing data quality. With regard to dates of birth of individual women, 42 percent of respondents reported both a month and year of birth, 21 percent gave a year of birth in addition to current age, and 37 percent gave only their ages. With regard to children's dates of birth in the birth history, 85 percent of births had both month and year reported, 12 percent had year and age reported, 1 percent had only age reported, and 2 percent had no date information.
Abstract copyright UK Data Service and data collection copyright owner. The Online Time Use Survey (OTUS) was developed by the Office for National Statistics to help improve the measurement of unpaid household production and caring activities that are not captured within traditional economic measures, and to understand better time use from a well-being and quality of life perspective. The survey collects information from adults aged 18 years and over who are randomly sampled from the NatCen Opinion Panel, which is representative of the UK population. Data collected between March 2020 and March 2021 covers Great Britain and data collected from March 2022 onwards covers the United Kingdom. Participants were issued with two pre-allocated diary days (one on a weekday and one on a weekend day). They were asked to record their main activities (in 10-minute intervals) and up to five secondary activities (in five-minute intervals) in every 24 hours within an online diary tool. Respondents were able to select activities from a pre-defined list. They were also asked to rate how much they enjoyed different activities. In addition, respondents were asked to complete a demographic questionnaire which records personal and household characteristics.Latest edition informationFor the third edition (August 2024), data and documentation for Wave 8 (9 to 17 March 2024) were added to the study. Main Topics: The annual data files include the following variables:main activities (in 10-minute periods) up to five secondary activities (in five-minute periods)count of all 5-minute primary activities total in minutes of all primary activitiescount of all 5-minute secondary activitiestotal in minutes of all secondary activitiesenjoyment level (scale 1-7) for all primary activities (in 10-minute periods) enjoyment level (scale 1-7) for all secondary activities (in five-minute periods)basic demographics, including personal well-being rating variables.
https://spdx.org/licenses/CC0-1.0.htmlhttps://spdx.org/licenses/CC0-1.0.html
Professional organizations in STEM (science, technology, engineering, and mathematics) can use demographic data to quantify recruitment and retention (R&R) of underrepresented groups within their memberships. However, variation in the types of demographic data collected can influence the targeting and perceived impacts of R&R efforts - e.g., giving false signals of R&R for some groups. We obtained demographic surveys from 73 U.S.-affiliated STEM organizations, collectively representing 712,000 members and conference-attendees. We found large differences in the demographic categories surveyed (e.g., disability status, sexual orientation) and the available response options. These discrepancies indicate a lack of consensus regarding the demographic groups that should be recognized and, for groups that are omitted from surveys, an inability of organizations to prioritize and evaluate R&R initiatives. Aligning inclusive demographic surveys across organizations will provide baseline data that can be used to target and evaluate R&R initiatives to better serve underrepresented groups throughout STEM. Methods We surveyed 164 STEM organizations (73 responses, rate = 44.5%) between December 2020 and July 2021 with the goal of understanding what demographic data each organization collects from its constituents (i.e., members and conference-attendees) and how the data are used. Organizations were sourced from a list of professional societies affiliated with the American Association for the Advancement of Science, AAAS, (n = 156) or from social media (n = 8). The survey was sent to the elected leadership and management firms for each organization, and follow-up reminders were sent after one month. The responding organizations represented a wide range of fields: 31 life science organizations (157,000 constituents), 5 mathematics organizations (93,000 constituents), 16 physical science organizations (207,000 constituents), 7 technology organizations (124,000 constituents), and 14 multi-disciplinary organizations spanning multiple branches of STEM (131,000 constituents). A list of the responding organizations is available in the Supplementary Materials. Based on the AAAS-affiliated recruitment of the organizations and the similar distribution of constituencies across STEM fields, we conclude that the responding organizations are a representative cross-section of the most prominent STEM organizations in the U.S. Each organization was asked about the demographic information they collect from their constituents, the response rates to their surveys, and how the data were used. Survey description The following questions are written as presented to the participating organizations. Question 1: What is the name of your STEM organization? Question 2: Does your organization collect demographic data from your membership and/or meeting attendees? Question 3: When was your organization’s most recent demographic survey (approximate year)? Question 4: We would like to know the categories of demographic information collected by your organization. You may answer this question by either uploading a blank copy of your organization’s survey (linked provided in online version of this survey) OR by completing a short series of questions. Question 5: On the most recent demographic survey or questionnaire, what categories of information were collected? (Please select all that apply)
Disability status Gender identity (e.g., male, female, non-binary) Marital/Family status Racial and ethnic group Religion Sex Sexual orientation Veteran status Other (please provide)
Question 6: For each of the categories selected in Question 5, what options were provided for survey participants to select? Question 7: Did the most recent demographic survey provide a statement about data privacy and confidentiality? If yes, please provide the statement. Question 8: Did the most recent demographic survey provide a statement about intended data use? If yes, please provide the statement. Question 9: Who maintains the demographic data collected by your organization? (e.g., contracted third party, organization executives) Question 10: How has your organization used members’ demographic data in the last five years? Examples: monitoring temporal changes in demographic diversity, publishing diversity data products, planning conferences, contributing to third-party researchers. Question 11: What is the size of your organization (number of members or number of attendees at recent meetings)? Question 12: What was the response rate (%) for your organization’s most recent demographic survey? *Organizations were also able to upload a copy of their demographics survey instead of responding to Questions 5-8. If so, the uploaded survey was used (by the study authors) to evaluate Questions 5-8.