The JPFHS is part of the worldwide Demographic and Health Surveys Program, which is designed to collect data on fertility, family planning, and maternal and child health. The primary objective of the Jordan Population and Family Health Survey (JPFHS) is to provide reliable estimates of demographic parameters, such as fertility, mortality, family planning, fertility preferences, as well as maternal and child health and nutrition that can be used by program managers and policy makers to evaluate and improve existing programs. In addition, the JPFHS data will be useful to researchers and scholars interested in analyzing demographic trends in Jordan, as well as those conducting comparative, regional or crossnational studies.
The content of the 2002 JPFHS was significantly expanded from the 1997 survey to include additional questions on women’s status, reproductive health, and family planning. In addition, all women age 15-49 and children less than five years of age were tested for anemia.
National
Sample survey data
The estimates from a sample survey are affected by two types of errors: 1) nonsampling errors and 2) sampling errors. Nonsampling errors are the result of mistakes made in implementing data collection and data processing, such as failure to locate and interview the correct household, misunderstanding of the questions on the part of either the interviewer or the respondent, and data entry errors. Although numerous efforts were made during the implementation of the 2002 JPFHS to minimize this type of error, nonsampling errors are impossible to avoid and difficult to evaluate statistically.
Sampling errors, on the other hand, can be evaluated statistically. The sample of respondents selected in the 2002 JPFHS is only one of many samples that could have been selected from the same population, using the same design and expected size. Each of these samples would yield results that differ somewhat from the results of the actual sample selected. Sampling errors are a measure of the variability between all possible samples. Although the degree of variability is not known exactly, it can be estimated from the survey results.
A sampling error is usually measured in terms of the standard error for a particular statistic (mean, percentage, etc.), which is the square root of the variance. The standard error can be used to calculate confidence intervals within which the true value for the population can reasonably be assumed to fall. For example, for any given statistic calculated from a sample survey, the value of that statistic will fall within a range of plus or minus two times the standard error of that statistic in 95 percent of all possible samples of identical size and design.
If the sample of respondents had been selected as a simple random sample, it would have been possible to use straightforward formulas for calculating sampling errors. However, the 2002 JPFHS sample is the result of a multistage stratified design and, consequently, it was necessary to use more complex formulas. The computer software used to calculate sampling errors for the 2002 JPFHS is the ISSA Sampling Error Module (ISSAS). This module used the Taylor linearization method of variance estimation for survey estimates that are means or proportions. The Jackknife repeated replication method is used for variance estimation of more complex statistics such as fertility and mortality rates.
Note: See detailed description of sample design in APPENDIX B of the survey report.
Face-to-face
The 2002 JPFHS used two questionnaires – namely, the Household Questionnaire and the Individual Questionnaire. Both questionnaires were developed in English and translated into Arabic. The Household Questionnaire was used to list all usual members of the sampled households and to obtain information on each member’s age, sex, educational attainment, relationship to the head of household, and marital status. In addition, questions were included on the socioeconomic characteristics of the household, such as source of water, sanitation facilities, and the availability of durable goods. The Household Questionnaire was also used to identify women who are eligible for the individual interview: ever-married women age 15-49. In addition, all women age 15-49 and children under five years living in the household were measured to determine nutritional status and tested for anemia.
The household and women’s questionnaires were based on the DHS Model “A” Questionnaire, which is designed for use in countries with high contraceptive prevalence. Additions and modifications to the model questionnaire were made in order to provide detailed information specific to Jordan, using experience gained from the 1990 and 1997 Jordan Population and Family Health Surveys. For each evermarried woman age 15 to 49, information on the following topics was collected:
In addition, information on births and pregnancies, contraceptive use and discontinuation, and marriage during the five years prior to the survey was collected using a monthly calendar.
Fieldwork and data processing activities overlapped. After a week of data collection, and after field editing of questionnaires for completeness and consistency, the questionnaires for each cluster were packaged together and sent to the central office in Amman where they were registered and stored. Special teams were formed to carry out office editing and coding of the open-ended questions.
Data entry and verification started after one week of office data processing. The process of data entry, including one hundred percent re-entry, editing and cleaning, was done by using PCs and the CSPro (Census and Survey Processing) computer package, developed specially for such surveys. The CSPro program allows data to be edited while being entered. Data processing operations were completed by the end of October 2002. A data processing specialist from ORC Macro made a trip to Jordan in October and November 2002 to follow up data editing and cleaning and to work on the tabulation of results for the survey preliminary report. The tabulations for the present final report were completed in December 2002.
A total of 7,968 households were selected for the survey from the sampling frame; among those selected households, 7,907 households were found. Of those households, 7,825 (99 percent) were successfully interviewed. In those households, 6,151 eligible women were identified, and complete interviews were obtained with 6,006 of them (98 percent of all eligible women). The overall response rate was 97 percent.
Note: See summarized response rates by place of residence in Table 1.1 of the survey report.
The estimates from a sample survey are affected by two types of errors: 1) nonsampling errors and 2) sampling errors. Nonsampling errors are the result of mistakes made in implementing data collection and data processing, such as failure to locate and interview the correct household, misunderstanding of the questions on the part of either the interviewer or the respondent, and data entry errors. Although numerous efforts were made during the implementation of the 2002 JPFHS to minimize this type of error, nonsampling errors are impossible to avoid and difficult to evaluate statistically.
Sampling errors, on the other hand, can be evaluated statistically. The sample of respondents selected in the 2002 JPFHS is only one of many samples that could have been selected from the same population, using the same design and expected size. Each of these samples would yield results that differ somewhat from the results of the actual sample selected. Sampling errors are a measure of the variability between all possible samples. Although the degree of variability is not known exactly, it can be estimated from the survey results.
A sampling error is usually measured in terms of the standard error for a particular statistic (mean, percentage, etc.), which is the square root of the variance. The standard error can be used to calculate confidence intervals within which the true value for the population can reasonably be assumed to fall. For example, for any given statistic calculated from a sample survey, the value of that statistic will fall within a range of plus or minus two times the standard error of that statistic in 95 percent of all possible samples of identical size and design.
If the sample of respondents had been selected as a simple random sample, it would have been possible to use straightforward formulas for calculating sampling errors. However, the 2002 JPFHS sample is the result of a multistage stratified design and, consequently, it was necessary to use more complex formulas. The computer software used to calculate sampling errors for the 2002 JPFHS is the ISSA Sampling Error Module (ISSAS). This module used the Taylor linearization method of variance estimation for survey estimates that are means or proportions. The Jackknife repeated replication method is used for variance estimation of more complex statistics such as fertility and mortality rates.
Note: See detailed
In 2001, the World Bank in co-operation with the Republika Srpska Institute of Statistics (RSIS), the Federal Institute of Statistics (FOS) and the Agency for Statistics of BiH (BHAS), carried out a Living Standards Measurement Survey (LSMS). The Living Standard Measurement Survey LSMS, in addition to collecting the information necessary to obtain a comprehensive as possible measure of the basic dimensions of household living standards, has three basic objectives, as follows:
To provide the public sector, government, the business community, scientific institutions, international donor organizations and social organizations with information on different indicators of the population's living conditions, as well as on available resources for satisfying basic needs.
To provide information for the evaluation of the results of different forms of government policy and programs developed with the aim to improve the population's living standard. The survey will enable the analysis of the relations between and among different aspects of living standards (housing, consumption, education, health, labor) at a given time, as well as within a household.
To provide key contributions for development of government's Poverty Reduction Strategy Paper, based on analyzed data.
The Department for International Development, UK (DFID) contributed funding to the LSMS and provided funding for a further two years of data collection for a panel survey, known as the Household Survey Panel Series (HSPS). Birks Sinclair & Associates Ltd. were responsible for the management of the HSPS with technical advice and support provided by the Institute for Social and Economic Research (ISER), University of Essex, UK. The panel survey provides longitudinal data through re-interviewing approximately half the LSMS respondents for two years following the LSMS, in the autumn of 2002 and 2003. The LSMS constitutes Wave 1 of the panel survey so there are three years of panel data available for analysis. For the purposes of this documentation we are using the following convention to describe the different rounds of the panel survey:
- Wave 1 LSMS conducted in 2001 forms the baseline survey for the panel
- Wave 2 Second interview of 50% of LSMS respondents in Autumn/ Winter 2002
- Wave 3 Third interview with sub-sample respondents in Autumn/ Winter 2003
The panel data allows the analysis of key transitions and events over this period such as labour market or geographical mobility and observe the consequent outcomes for the well-being of individuals and households in the survey. The panel data provides information on income and labour market dynamics within FBiH and RS. A key policy area is developing strategies for the reduction of poverty within FBiH and RS. The panel will provide information on the extent to which continuous poverty is experienced by different types of households and individuals over the three year period. And most importantly, the co-variates associated with moves into and out of poverty and the relative risks of poverty for different people can be assessed. As such, the panel aims to provide data, which will inform the policy debates within FBiH and RS at a time of social reform and rapid change.
National coverage. Domains: Urban/rural/mixed; Federation; Republic
Sample survey data [ssd]
The Wave 3 sample consisted of 2878 households who had been interviewed at Wave 2 and a further 73 households who were interviewed at Wave 1 but were non-contact at Wave 2 were issued. A total of 2951 households (1301 in the RS and 1650 in FBiH) were issued for Wave 3. As at Wave 2, the sample could not be replaced with any other households.
Panel design
Eligibility for inclusion
The household and household membership definitions are the same standard definitions as a Wave 2. While the sample membership status and eligibility for interview are as follows: i) All members of households interviewed at Wave 2 have been designated as original sample members (OSMs). OSMs include children within households even if they are too young for interview. ii) Any new members joining a household containing at least one OSM, are eligible for inclusion and are designated as new sample members (NSMs). iii) At each wave, all OSMs and NSMs are eligible for inclusion, apart from those who move outof-scope (see discussion below). iv) All household members aged 15 or over are eligible for interview, including OSMs and NSMs.
Following rules
The panel design means that sample members who move from their previous wave address must be traced and followed to their new address for interview. In some cases the whole household will move together but in others an individual member may move away from their previous wave household and form a new split-off household of their own. All sample members, OSMs and NSMs, are followed at each wave and an interview attempted. This method has the benefit of maintaining the maximum number of respondents within the panel and being relatively straightforward to implement in the field.
Definition of 'out-of-scope'
It is important to maintain movers within the sample to maintain sample sizes and reduce attrition and also for substantive research on patterns of geographical mobility and migration. The rules for determining when a respondent is 'out-of-scope' are as follows:
i. Movers out of the country altogether i.e. outside FBiH and RS. This category of mover is clear. Sample members moving to another country outside FBiH and RS will be out-of-scope for that year of the survey and not eligible for interview.
ii. Movers between entities Respondents moving between entities are followed for interview. The personal details of the respondent are passed between the statistical institutes and a new interviewer assigned in that entity.
iii. Movers into institutions Although institutional addresses were not included in the original LSMS sample, Wave 3 individuals who have subsequently moved into some institutions are followed. The definitions for which institutions are included are found in the Supervisor Instructions.
iv. Movers into the district of Brcko are followed for interview. When coding entity Brcko is treated as the entity from which the household who moved into Brcko originated.
Face-to-face [f2f]
Questionnaire design
Approximately 90% of the questionnaire (Annex B) is based on the Wave 2 questionnaire, carrying forward core measures that are needed to measure change over time. The questionnaire was widely circulated and changes were made as a result of comments received.
Pretesting
In order to undertake a longitudinal test the Wave 2 pretest sample was used. The Control Forms and Advance letters were generated from an Access database containing details of ten households in Sarajevo and fourteen in Banja Luka. The pretest was undertaken from March 24-April 4 and resulted in 24 households (51 individuals) successfully interviewed. One mover household was successfully traced and interviewed.
In order to test the questionnaire under the hardest circumstances a briefing was not held. A list of the main questionnaire changes was given to experienced interviewers.
Issues arising from the pretest
Interviewers were asked to complete a Debriefing and Rating form. The debriefing form captured opinions on the following three issues:
General reaction to being re-interviewed. In some cases there was a wariness of being asked to participate again, some individuals asking “Why Me?” Interviewers did a good job of persuading people to take part, only one household refused and another asked to be removed from the sample next year. Having the same interviewer return to the same households was considered an advantage. Most respondents asked what was the benefit to them of taking part in the survey. This aspect was reemphasised in the Advance Letter, Respondent Report and training of the Wave 3 interviewers.
Length of the questionnaire. The average time of interview was 30 minutes. No problems were mentioned in relation to the timing, though interviewers noted that some respondents, particularly the elderly, tended to wonder off the point and that control was needed to bring them back to the questions in the questionnaire. One interviewer noted that the economic situation of many respondents seems to have got worse from the previous year and it was necessary to listen to respondents “stories” during the interview.
Confidentiality. No problems were mentioned in relation to confidentiality. Though interviewers mentioned it might be worth mentioning the new Statistics Law in the Advance letter. The Rating Form asked for details of specific questions that were unclear. These are described below with a description of the changes made.
Module 3. Q29-31 have been added to capture funds received for education, scholarships etc.
Module 4. Pretest respondents complained that the 6 questions on "Has your health limited you..." and the 16 on "in the last 7 days have you felt depressed” etc were too many. These were reduced by half (Q38-Q48). The LSMS data was examined and those questions where variability between the answers was widest were chosen.
Module 5. The new employment questions (Q42-Q44) worked well and have been kept in the main questionnaire.
Module 7. There were no problems reported with adding the credit questions (Q28-Q36)
Module 9. SIG recommended that some of Questions 1-12 were relevant only to those aged over 18 so additional skips have been added. Some respondents complained the questionnaire was boring. To try and overcome
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Repeated measures correlation (rmcorr) is a statistical technique for determining the common within-individual association for paired measures assessed on two or more occasions for multiple individuals. Simple regression/correlation is often applied to non-independent observations or aggregated data; this may produce biased, specious results due to violation of independence and/or differing patterns between-participants versus within-participants. Unlike simple regression/correlation, rmcorr does not violate the assumption of independence of observations. Also, rmcorr tends to have much greater statistical power because neither averaging nor aggregation is necessary for an intra-individual research question. Rmcorr estimates the common regression slope, the association shared among individuals. To make rmcorr accessible, we provide background information for its assumptions and equations, visualization, power, and tradeoffs with rmcorr compared to multilevel modeling. We introduce the R package (rmcorr) and demonstrate its use for inferential statistics and visualization with two example datasets. The examples are used to illustrate research questions at different levels of analysis, intra-individual, and inter-individual. Rmcorr is well-suited for research questions regarding the common linear association in paired repeated measures data. All results are fully reproducible.
The Participation Survey started in October 2021 and is the key evidence source on engagement for DCMS. It is a continuous push-to-web household survey of adults aged 16 and over in England.
The Participation Survey provides nationally representative estimates of physical and digital engagement with the arts, heritage, museums & galleries, and libraries, as well as engagement with tourism, major events, live sports and digital.
In 2023/24, DCMS partnered with Arts Council England (ACE) to boost the Participation Survey to be able to produce meaningful estimates at Local Authority level. This has enabled us to have the most granular data we have ever had, which means there were some new questions and changes to existing questions, response options and definitions in the 23/24 survey. The questionnaire for 2023/24 has been developed collaboratively to adapt to the needs and interests of both DCMS and ACE.
The Participation Survey is only asked of adults in England. Currently there is no harmonised survey or set of questions within the administrations of the UK. Data on participation in cultural sectors for the devolved administrations is available in the https://www.gov.scot/collections/scottish-household-survey/" class="govuk-link">Scottish Household Survey, https://gov.wales/national-survey-wales" class="govuk-link">National Survey for Wales and https://www.communities-ni.gov.uk/topics/statistics-and-research/culture-and-heritage-statistics" class="govuk-link">Northern Ireland Continuous Household Survey.
The pre-release access document above contains a list of ministers and officials who have received privileged early access to this release of Participation Survey data. In line with best practice, the list has been kept to a minimum and those given access for briefing purposes had a maximum of 24 hours. Details on the pre-release access arrangements for this dataset are available in the accompanying material.
Our statistical practice is regulated by the OSR. OSR sets the standards of trustworthiness, quality and value in the https://code.statisticsauthority.gov.uk/the-code/" class="govuk-link">Code of Practice for Statistics that all producers of official statistics should adhere to.
You are welcome to contact us directly with any comments about how we meet these standards by emailing evidence@dcms.gov.uk. Alternatively, you can contact OSR by emailing regulation@statistics.gov.uk or via the OSR website.
Patterns were identified in Census 2021 data that suggest that some respondents may not have interpreted the gender identity question as intended, notably those with lower levels of English language proficiency. https://www.scotlandscensus.gov.uk/2022-results/scotland-s-census-2022-sexual-orientation-and-trans-status-or-history/" class="govuk-link">Analysis of Scotland’s census, where the gender identity question was different, has added weight to this observation. Similar respondent error may have occurred during the data collection for these statistics so comparisons between subnational and other smaller group breakdowns should be considered with caution. More information can be found in the ONS https://www.ons.gov.uk/peoplepopulationandcommunity/culturalidentity/sexuality/methodologies/sexualorientationandgenderidentityqualityinformationforcensus2021" class="govuk-link">sexual orientation and gender identity quality information report, and in the National Statistical https://blog.ons.gov.uk/2024/09/12/better-understanding-the-strengths-and-limitations-of-gender-identity-statistics/" class="govuk-link">blog about the strengths and limitations of gender identity statistics.
The responsible statisticians for this release is Donilia Asgill and Ella Bentin. For enquiries on this release, contact participationsurvey@dcms.gov.uk.
Within the frame of PCBS' efforts in providing official Palestinian statistics in the different life aspects of Palestinian society and because the wide spread of Computer, Internet and Mobile Phone among the Palestinian people, and the important role they may play in spreading knowledge and culture and contribution in formulating the public opinion, PCBS conducted the Household Survey on Information and Communications Technology, 2014.
The main objective of this survey is to provide statistical data on Information and Communication Technology in the Palestine in addition to providing data on the following: - Prevalence of computers and access to the Internet. - Study the penetration and purpose of Technology use.
Palestine (West Bank and Gaza Strip), type of locality (urban, rural, refugee camps) and governorate.
All Palestinian households and individuals whose usual place of residence in Palestine with focus on persons aged 10 years and over in year 2014.
Sample survey data [ssd]
Sampling Frame The sampling frame consists of a list of enumeration areas adopted in the Population, Housing and Establishments Census of 2007. Each enumeration area has an average size of about 124 households. These were used in the first phase as Preliminary Sampling Units in the process of selecting the survey sample.
Sample Size The total sample size of the survey was 7,268 households, of which 6,000 responded.
Sample Design The sample is a stratified clustered systematic random sample. The design comprised three phases:
Phase I: Random sample of 240 enumeration areas. Phase II: Selection of 25 households from each enumeration area selected in phase one using systematic random selection. Phase III: Selection of an individual (10 years or more) in the field from the selected households; KISH TABLES were used to ensure indiscriminate selection.
Sample Strata Distribution of the sample was stratified by: 1- Governorate (16 governorates, J1). 2- Type of locality (urban, rural and camps).
Face-to-face [f2f]
The survey questionnaire consists of identification data, quality controls and three main sections: Section I: Data on household members that include identification fields, the characteristics of household members (demographic and social) such as the relationship of individuals to the head of household, sex, date of birth and age.
Section II: Household data include information regarding computer processing, access to the Internet, and possession of various media and computer equipment. This section includes information on topics related to the use of computer and Internet, as well as supervision by households of their children (5-17 years old) while using the computer and Internet, and protective measures taken by the household in the home.
Section III: Data on persons (aged 10 years and over) about computer use, access to the Internet and possession of a mobile phone.
Preparation of Data Entry Program: This stage included preparation of the data entry programs using an ACCESS package and defining data entry control rules to avoid errors, plus validation inquiries to examine the data after it had been captured electronically.
Data Entry: The data entry process started on the 8th of May 2014 and ended on the 23rd of June 2014. The data entry took place at the main PCBS office and in field offices using 28 data clerks.
Editing and Cleaning procedures: Several measures were taken to avoid non-sampling errors. These included editing of questionnaires before data entry to check field errors, using a data entry application that does not allow mistakes during the process of data entry, and then examining the data by using frequency and cross tables. This ensured that data were error free; cleaning and inspection of the anomalous values were conducted to ensure harmony between the different questions on the questionnaire.
Response Rates: 79%
There are many aspects of the concept of data quality; this includes the initial planning of the survey to the dissemination of the results and how well users understand and use the data. There are three components to the quality of statistics: accuracy, comparability, and quality control procedures.
Checks on data accuracy cover many aspects of the survey and include statistical errors due to the use of a sample, non-statistical errors resulting from field workers or survey tools, and response rates and their effect on estimations. This section includes:
Statistical Errors Data of this survey may be affected by statistical errors due to the use of a sample and not a complete enumeration. Therefore, certain differences can be expected in comparison with the real values obtained through censuses. Variances were calculated for the most important indicators.
Variance calculations revealed that there is no problem in disseminating results nationally or regionally (the West Bank, Gaza Strip), but some indicators show high variance by governorate, as noted in the tables of the main report.
Non-Statistical Errors Non-statistical errors are possible at all stages of the project, during data collection or processing. These are referred to as non-response errors, response errors, interviewing errors and data entry errors. To avoid errors and reduce their effects, strenuous efforts were made to train the field workers intensively. They were trained on how to carry out the interview, what to discuss and what to avoid, and practical and theoretical training took place during the training course. Training manuals were provided for each section of the questionnaire, along with practical exercises in class and instructions on how to approach respondents to reduce refused cases. Data entry staff were trained on the data entry program, which was tested before starting the data entry process.
Several measures were taken to avoid non-sampling errors. These included editing of questionnaires before data entry to check field errors, using a data entry application that does not allow mistakes during the process of data entry, and then examining the data by using frequency and cross tables. This ensured that data were error free; cleaning and inspection of the anomalous values were conducted to ensure harmony between the different questions on the questionnaire.
The sources of non-statistical errors can be summarized as: 1. Some of the households were not at home and could not be interviewed, and some households refused to be interviewed. 2. In unique cases, errors occurred due to the way the questions were asked by interviewers and respondents misunderstood some of the questions.
This data is from the first round of a unique, cross-country panel survey conducted in Pakistan by the Secure Livelihoods Research Consortium (SLRC). The Overseas Development Institute (ODI) is the lead organisation of SLRC. SLRC partners who participated in the survey were: the Centre for Poverty Analysis (CEPA) in Sri Lanka, Feinstein International Center (FIC, Tufts University), the Sustainable Development Policy Institute(SDPI) in Pakistan, Humanitarian Aid and Reconstruction, based at Wageningen University (WUR) in the Netherlands, the Nepal Centre for Contemporary Research (NCCR), and the Food and Agriculture Organization (FAO).
This survey generated the first round of data on people's livelihoods, their access to and experience of basic services, and their views of governance actors. SLRC will attempt to re-interview the same respondents in 2015 to find out how the livelihoods and governance perceptions of people shift (or not) over time, and which factors may have contributed towards that change.
Pakistan: Swat and Lower Dir districts of Khyber Pakhtunkhwa (KP) Rural and urban
Some questions are at the level of individuals in household (e.g. livelihood activities, education levels); other questions are at the household level (e.g. assets). A sizeable share of the questionnaire is devoted to perceptions based questions, which are at the individual (respondent) level.
Randomly selected households in purposely sampled sites (sampling procedure varied slightly by country).
Within a selected household, only one household members was interviewed about the household. Respondents were adults and we aimed to interview a fairly even share of men/ women. In some countries this was achieved, but in other countries the share of male respondents is substantially higher (e.g. Pakistan).
Sample survey data [ssd]
The sampling strategy was designed to select households that are relevant to the main research questions and as well as being of national relevance, while also being able to produce statistically significant conclusions at the study and village level. To meet these objectives, purposive and random sampling were combined at different stages of the sampling strategy. The first stages of the sampling process involved purposive sampling, with random sampling only utilized in the last stage of the process. Sampling locations were selected purposely (including districts and locations within districts), and then randomly households were selected within these locations. A rigorous sample is geared towards meeting the objectives of the research. The samples are not representative for the case study countries and cannot be used to represent the case study countries as a whole, nor for the districts. The samples are representative at the village level, with the exception of Uganda.
Sampling locations (sub-regions or districts, sub-districts and villages) were purposively selected, using criteria, such as levels of service provision or levels of conflict, in order to locate the specific groups of interest and to select geographical locations that are relevant to the broader SLRC research areas and of policy relevance at the national level. For instance, locations experienced high/ low levels of conflict and locations with high/ low provision of services were selected and locations that accounted for all possible combinations of selection criteria were included. Survey locations with different characteristics were chose, so that we could explore the relevance of conflict affectedness, access to services and variations in geography and livelihoods on our outcome variables. Depending on the administrative structure of the country, this process involved selecting a succession of sampling locations (at increasingly lower administrative units).
The survey did not attempt to achieve representativeness at the country /or district level, but it aimed for representativeness at the sub-district /or village level through random sampling (Households were randomly selected within villages so that the results are representative and statistically significant at the village level and so that a varied sample was captured. Households were randomly selected using a number of different tools, depending on data availability, such as random selection from vote registers (Nepal), construction of household listings (DRC) and a quasi-random household process that involved walking in a random direction for a random number of minutes (Uganda).
The samples are statistically significant at the survey level and village level (in all countries) and at the district level in Sri Lanka and sub-region level in Uganda. The sample size was calculated with the aim to achieve statistical significance at the study and village level, and to accommodate the available budget, logistical limitations, and to account for possible attrition between 2012-2015. In a number of countries estimated population data had to be used, as recent population data were not available.
The minimum overall sample size required to achieve significance at the study level, given population and average household size across districts, was calculated using a basic sample size calculator at a 95% confidence level and confidence interval of 5. The sample size at the village level was again calculated at the using a 95% confidence level and confidence interval of 5. . Finally, the sample was increased by 20% to account for possible attrition between 2012 and 2015, so that the sample size in 2015 is likely to be still statistically significant.
The overall sample required to achieve the sampling objectives in selected districts in each country ranged from 1,259 to 3,175 households.
Face-to-face [f2f]
One questionnaire per country that includes household level, individual level and respondent level perceptions based questions.
The general structure and content of the questionnaire is similar across all five countries, with about 80% of questions similar, but tailored to the country-specific process. Country-specific surveys were tailored on the basis of a generic survey instrument that was developed by ODI specifically for this survey.
The questionnaires are published in English.
CSPro was used for data entries in most countries.
Data editing took place at a number of stages throughout the processing, including: • Office editing and coding • During data entry • Structure checking and completeness • Extensive secondary editing conducted by ODI
The required sample sizes were achieved in all countries. Response rates were extremely high, ranging from 99%-100%.
No further estimations of sampling error was conducted beyond the sampling design stage.
Done on an ad hoc basis for some countries, but not consistently across all surveys and domains.
CourseKata is a platform that creates and publishes a series of e-books for introductory statistics and data science classes that utilize demonstrated learning strategies to help students learn statistics and data science. The developers of CourseKata, Jim Stigler (UCLA) and Ji Son (Cal State Los Angeles) and their team, are cognitive psychologists interested in improving statistics learning by examining students' interactions with online interactive textbooks. Traditionally, much of the research in how students learn is done in a 1-hour lab or through small-scale interviews with students. CourseKata offers the opportunity to peek into the actions, responses, and choices of thousands of students as they are engaged in learning the interrelated concepts and skills of statistics and coding in R over many weeks or months in real classes.
Questions are grouped into items (item_id). An item can be one of three item_type 's: code, learnosity or learnosity-activity (the distinction between learnosity and learnosity-activity is not important). Code items are a single question and ask for R code as a response. (Responses can be seen in responses.csv.) Learnosity-activities and learnosity items are collections of one or more questions that can be of a variety of lrn_type's: ● association ● choicematrix ● clozeassociation ● formulaV2 ● imageclozeassociation ● mcq ● plaintext ● shorttext ● sortlist
Examples of these question types are provided at the end of this document.
The level of detail made available to you in the responses file depends on the lrn_type. For example, for multiple choice questions (mcq), you can find the options in the responses file in the columns labeled lrn_option_0 through lrn_option_11, and you can see the chosen option in the results variable.
Assessment Types In general, assessments, such as the items and questions included in CourseKata, can be used for two purposes. Formative assessments are meant to provide feedback to the student (and instructor), or to serve as a learning aid to help prompt students improve memory and deepen their understanding. Summative assessments are meant to provide a summary of a student's understanding, often for use in assigning a grade. For example, most midterms and final exams that you've taken are summative assessments.
The vast majority of items in CourseKata should be treated as formative assessments. The exceptions are the end-of-chapter Review questions, which can be thought of as summative. The mean number of correct answers for end-of-chapter review questions is provided within the checkpoints file. You might see that some pages have the word "Quiz" or "Exam" or "Midterm" in them. Results from these items and responses to them are not provided to us in this data set.
The objective of the survey is to obtain feedback from enterprises in client countries on the state of the private sector as well as to help in building a panel of enterprise data that will make it possible to track changes in the business environment over time, thus allowing, for example, impact assessments of reforms. Through interviews with firms in the manufacturing and services sectors, the survey assesses the constraints to private sector growth and creates statistically significant business environment indicators that are comparable across countries.
The standard Enterprise Survey topics include firm characteristics, gender participation, access to finance, annual sales, costs of inputs/labor, workforce composition, bribery, licensing, infrastructure, trade, crime, competition, capacity utilization, land and permits, taxation, informality, business-government relations, innovation and technology, and performance measures. Over 90% of the questions objectively ascertain characteristics of a country’s business environment. The remaining questions assess the survey respondents’ opinions on what are the obstacles to firm growth and performance. The mode of data collection is face-to-face interviews.
National
The primary sampling unit of the study is the establishment. An establishment is a physical location where business is carried out and where industrial operations take place or services are provided. A firm may be composed of one or more establishments. For example, a brewery may have several bottling plants and several establishments for distribution. For the purposes of this survey an establishment must make its own financial decisions and have its own financial statements separate from those of the firm. An establishment must also have its own management and control over its payroll.
The manufacturing and services sectors are the primary business sectors of interest. This corresponds to firms classified with International Standard Industrial Classification of All Economic Activities (ISIC) codes 15-37, 45, 50-52, 55, 60-64, and 72 (ISIC Rev.3.1). Formal (registered) companies with 5 or more employees are targeted for interview. Services firms include construction, retail, wholesale, hotels, restaurants, transport, storage, communications, and IT. Firms with 100% government/state ownership are not eligible to participate in an Enterprise Survey.
Sample survey data [ssd]
The sample for Azerbaijan was selected using stratified random sampling. Three levels of stratification were used in this country: industry, establishment size, and oblast (region).
Industry stratification was designed in the way that follows: the universe was stratified into 23 manufacturing industries, 2 services industries -retail and IT-, and one residual sector. Each sector had a target of 90 interviews.
Size stratification was defined following the standardized definition for the rollout: small (5 to 19 employees), medium (20 to 99 employees), and large (more than 99 employees). For stratification purposes, the number of employees was defined on the basis of reported permanent full-time workers. This seems to be an appropriate definition of the labor force since seasonal/casual/part-time employment is not a common practice, except in the sectors of construction and agriculture.
Regional stratification was defined in eight regions. These regions are Praha, Stredni Cechy, Jihozapad, Severozapad, Severovychod, Jihovychod, Stredni Morava, and Moravskoslezsko.
Given the stratified design, sample frames containing a complete and updated list of establishments for the selected regions were required. Great efforts were made to obtain the best source for these listings. However, the quality of the sample frames was not optimal and, therefore, some adjustments were needed to correct for the presence of ineligible units. These adjustments are reflected in the weights computation.
For most countries covered in BEEPS IV, two sample frames were used. The first was supplied by the World Bank and consisted of enterprises interviewed in BEEPS 2005. The World Bank required that attempts should be made to re-interview establishments responding to the BEEPS 2005 survey where they were within the selected geographical regions and met eligibility criteria. That sample is referred to as the Panel. The second frame for the Czech Republic was an official database known as Albertina data [Creditinfo Czech Republic], which is obtained from the complete Business Register [RES] of the Czech Statistical Office. An extract from that frame was sent to the TNS statistical team in London to select the establishments for interview.
The quality of the frame was assessed at the onset of the project. The frame proved to be useful though it showed positive rates of non-eligibility, repetition, non-existent units, etc. These problems are typical of establishment surveys, but given the impact these inaccuracies may have on the results, adjustments were needed when computing the appropriate weights for individual observations. The percentage of confirmed non-eligible units as a proportion of the total number of contacts to complete the survey was 28% (572 out of 2041 establishments).
Face-to-face [f2f]
The current survey instruments are available: - Core Questionnaire + Manufacturing Module [ISIC Rev.3.1: 15-37] - Core Questionnaire + Retail Module [ISIC Rev.3.1: 52] - Core Questionnaire [ISIC Rev.3.1: 45, 50, 51, 55, 60-64, 72] - Screener Questionnaire.
The “Core Questionnaire” is the heart of the Enterprise Survey and contains the survey questions asked of all firms across the world. There are also two other survey instruments- the “Core Questionnaire + Manufacturing Module” and the “Core Questionnaire + Retail Module.” The survey is fielded via three instruments in order to not ask questions that are irrelevant to specific types of firms, e.g. a question that relates to production and nonproduction workers should not be asked of a retail firm. In addition to questions that are asked across countries, all surveys are customized and contain country-specific questions. An example of customization would be including tourism-related questions that are asked in certain countries when tourism is an existing or potential sector of economic growth.
The standard Enterprise Survey topics include firm characteristics, gender participation, access to finance, annual sales, costs of inputs/labor, workforce composition, bribery, licensing, infrastructure, trade, crime, competition, capacity utilization, land and permits, taxation, informality, business-government relations, innovation and technology, and performance measures. Over 90% of the questions objectively ascertain characteristics of a country’s business environment. The remaining questions assess the survey respondents’ opinions on what are the obstacles to firm growth and performance.
Data entry and quality controls are implemented by the contractor and data is delivered to the World Bank in batches (typically 10%, 50% and 100%). These data deliveries are checked for logical consistency, out of range values, skip patterns, and duplicate entries. Problems are flagged by the World Bank and corrected by the implementing contractor through data checks, callbacks, and revisiting establishments.
Complete information regarding the sampling methodology, sample frame, weights, response rates, and implementation can be found in the document "Description of Czech Republic Implementation 2009.pdf"
analyze the current population survey (cps) annual social and economic supplement (asec) with r the annual march cps-asec has been supplying the statistics for the census bureau's report on income, poverty, and health insurance coverage since 1948. wow. the us census bureau and the bureau of labor statistics ( bls) tag-team on this one. until the american community survey (acs) hit the scene in the early aughts (2000s), the current population survey had the largest sample size of all the annual general demographic data sets outside of the decennial census - about two hundred thousand respondents. this provides enough sample to conduct state- and a few large metro area-level analyses. your sample size will vanish if you start investigating subgroups b y state - consider pooling multiple years. county-level is a no-no. despite the american community survey's larger size, the cps-asec contains many more variables related to employment, sources of income, and insurance - and can be trended back to harry truman's presidency. aside from questions specifically asked about an annual experience (like income), many of the questions in this march data set should be t reated as point-in-time statistics. cps-asec generalizes to the united states non-institutional, non-active duty military population. the national bureau of economic research (nber) provides sas, spss, and stata importation scripts to create a rectangular file (rectangular data means only person-level records; household- and family-level information gets attached to each person). to import these files into r, the parse.SAScii function uses nber's sas code to determine how to import the fixed-width file, then RSQLite to put everything into a schnazzy database. you can try reading through the nber march 2012 sas importation code yourself, but it's a bit of a proc freak show. this new github repository contains three scripts: 2005-2012 asec - download all microdata.R down load the fixed-width file containing household, family, and person records import by separating this file into three tables, then merge 'em together at the person-level download the fixed-width file containing the person-level replicate weights merge the rectangular person-level file with the replicate weights, then store it in a sql database create a new variable - one - in the data table 2012 asec - analysis examples.R connect to the sql database created by the 'download all microdata' progr am create the complex sample survey object, using the replicate weights perform a boatload of analysis examples replicate census estimates - 2011.R connect to the sql database created by the 'download all microdata' program create the complex sample survey object, using the replicate weights match the sas output shown in the png file below 2011 asec replicate weight sas output.png statistic and standard error generated from the replicate-weighted example sas script contained in this census-provided person replicate weights usage instructions document. click here to view these three scripts for more detail about the current population survey - annual social and economic supplement (cps-asec), visit: the census bureau's current population survey page the bureau of labor statistics' current population survey page the current population survey's wikipedia article notes: interviews are conducted in march about experiences during the previous year. the file labeled 2012 includes information (income, work experience, health insurance) pertaining to 2011. when you use the current populat ion survey to talk about america, subract a year from the data file name. as of the 2010 file (the interview focusing on america during 2009), the cps-asec contains exciting new medical out-of-pocket spending variables most useful for supplemental (medical spending-adjusted) poverty research. confidential to sas, spss, stata, sudaan users: why are you still rubbing two sticks together after we've invented the butane lighter? time to transition to r. :D
The requests we receive at the Reference Desk keep surprising us. We'll take a look at some of the best examples from the year on data questions and data solutions.
The Research and Development Survey (RANDS) is a platform designed for conducting survey question evaluation and statistical research. RANDS is an ongoing series of surveys from probability-sampled commercial survey panels used for methodological research at the National Center for Health Statistics (NCHS). RANDS estimates are generated using an experimental approach that differs from the survey design approaches generally used by NCHS, including possible biases from different response patterns and sampling frames as well as increased variability from lower sample sizes. Use of the RANDS platform allows NCHS to produce more timely data than would be possible using traditional data collection methods. RANDS is not designed to replace NCHS’ higher quality, core data collections. Below are experimental estimates of reduced access to healthcare for three rounds of RANDS during COVID-19. Data collection for the three rounds of RANDS during COVID-19 occurred between June 9, 2020 and July 6, 2020, August 3, 2020 and August 20, 2020, and May 17, 2021 and June 30, 2021. Information needed to interpret these estimates can be found in the Technical Notes. RANDS during COVID-19 included questions about unmet care in the last 2 months during the coronavirus pandemic. Unmet needs for health care are often the result of cost-related barriers. The National Health Interview Survey, conducted by NCHS, is the source for high-quality data to monitor cost-related health care access problems in the United States. For example, in 2018, 7.3% of persons of all ages reported delaying medical care due to cost and 4.8% reported needing medical care but not getting it due to cost in the past year. However, cost is not the only reason someone might delay or not receive needed medical care. As a result of the coronavirus pandemic, people also may not get needed medical care due to cancelled appointments, cutbacks in transportation options, fear of going to the emergency room, or an altruistic desire to not be a burden on the health care system, among other reasons. The Household Pulse Survey (https://www.cdc.gov/nchs/covid19/pulse/reduced-access-to-care.htm), an online survey conducted in response to the COVID-19 pandemic by the Census Bureau in partnership with other federal agencies including NCHS, also reports estimates of reduced access to care during the pandemic (beginning in Phase 1, which started on April 23, 2020). The Household Pulse Survey reports the percentage of adults who delayed medical care in the last 4 weeks or who needed medical care at any time in the last 4 weeks for something other than coronavirus but did not get it because of the pandemic. The experimental estimates on this page are derived from RANDS during COVID-19 and show the percentage of U.S. adults who were unable to receive medical care (including urgent care, surgery, screening tests, ongoing treatment, regular checkups, prescriptions, dental care, vision care, and hearing care) in the last 2 months. Technical Notes: https://www.cdc.gov/nchs/covid19/rands/reduced-access-to-care.htm#limitations
This data is from the first round of a unique, cross-country panel survey conducted in Nepal by the Secure Livelihoods Research Consortium (SLRC). The Overseas Development Institute (ODI) is the lead organisation of SLRC. SLRC partners who participated in the survey were: the Centre for Poverty Analysis (CEPA) in Sri Lanka, Feinstein International Center (FIC, Tufts University), the Sustainable Development Policy Institute(SDPI) in Pakistan, Humanitarian Aid and Reconstruction, based at Wageningen University (WUR) in the Netherlands, the Nepal Centre for Contemporary Research (NCCR), and the Food and Agriculture Organization (FAO).
This survey generated the first round of data on people's livelihoods, their access to and experience of basic services, and their views of governance actors. SLRC will attempt to re-interview the same respondents in 2015 to find out how the livelihoods and governance perceptions of people shift (or not) over time, and which factors may have contributed towards that change.
Regional
Households
Randomly selected households in purposely sampled sites (sampling procedure varied slightly by country). Within a selected household, only one household members was interviewed about the household. Respondents were adults and we aimed to interview a fairly even share of men/ women. In some countries this was achieved, but in other countries the share of male respondents is substantially higher (e.g. Pakistan).
Sample survey data [ssd]
The sampling strategy was designed to select households that are relevant to the main research questions and as well as being of national relevance, while also being able to produce statistically significant conclusions at the study and village level. To meet these objectives, purposive and random sampling were combined at different stages of the sampling strategy. The first stages of the sampling process involved purposive sampling, with random sampling only utilized in the last stage of the process. Sampling locations were selected purposely (including districts and locations within districts), and then randomly households were selected within these locations. A rigorous sample is geared towards meeting the objectives of the research. The samples are not representative for the case study countries and cannot be used to represent the case study countries as a whole, nor for the districts. The samples are representative at the village level, with the exception of Uganda.
Sampling locations (sub-regions or districts, sub-districts and villages) were purposively selected, using criteria, such as levels of service provision or levels of conflict, in order to locate the specific groups of interest and to select geographical locations that are relevant to the broader SLRC research areas and of policy relevance at the national level. For instance, locations experienced high/ low levels of conflict and locations with high/ low provision of services were selected and locations that accounted for all possible combinations of selection criteria were included. Survey locations with different characteristics were chose, so that we could explore the relevance of conflict affectedness, access to services and variations in geography and livelihoods on our outcome variables. Depending on the administrative structure of the country, this process involved selecting a succession of sampling locations (at increasingly lower administrative units).
The survey did not attempt to achieve representativeness at the country /or district level, but it aimed for representativeness at the sub-district /or village level through random sampling (Households were randomly selected within villages so that the results are representative and statistically significant at the village level and so that a varied sample was captured. Households were randomly selected using a number of different tools, depending on data availability, such as random selection from vote registers (Nepal), construction of household listings (DRC) and a quasi-random household process that involved walking in a random direction for a random number of minutes (Uganda).
The samples are statistically significant at the survey level and village level (in all countries) and at the district level in Sri Lanka and sub-region level in Uganda. The sample size was calculated with the aim to achieve statistical significance at the study and village level, and to accommodate the available budget, logistical limitations, and to account for possible attrition between 2012-2015. In a number of countries estimated population data had to be used, as recent population data were not available. The minimum overall sample size required to achieve significance at the study level, given population and average household size across districts, was calculated using a basic sample size calculator at a 95% confidence level and confidence interval of 5. The sample size at the village level was again calculated at the using a 95% confidence level and confidence interval of 5. Finally, the sample was increased by 20% to account for possible attrition between 2012 and 2015, so that the sample size in 2015 is likely to be still statistically significant. The overall sample required to achieve the sampling objectives in selected districts in each country ranged from 1,259 to 3,175 households.
Face-to-face [f2f]
CSPro was used for data entries in most countries.
Data editing took place at a number of stages throughout the processing, including: • Office editing and coding • During data entry • Structure checking and completeness • Extensive secondary editing conducted by ODI
Approximately 99 percent
No further estimations of sampling error was conducted beyond the sampling design stage.
Done on an ad hoc basis for some countries, but not consistently across all surveys and domains.
The Associated Press is sharing data from the COVID Impact Survey, which provides statistics about physical health, mental health, economic security and social dynamics related to the coronavirus pandemic in the United States.
Conducted by NORC at the University of Chicago for the Data Foundation, the probability-based survey provides estimates for the United States as a whole, as well as in 10 states (California, Colorado, Florida, Louisiana, Minnesota, Missouri, Montana, New York, Oregon and Texas) and eight metropolitan areas (Atlanta, Baltimore, Birmingham, Chicago, Cleveland, Columbus, Phoenix and Pittsburgh).
The survey is designed to allow for an ongoing gauge of public perception, health and economic status to see what is shifting during the pandemic. When multiple sets of data are available, it will allow for the tracking of how issues ranging from COVID-19 symptoms to economic status change over time.
The survey is focused on three core areas of research:
Instead, use our queries linked below or statistical software such as R or SPSS to weight the data.
If you'd like to create a table to see how people nationally or in your state or city feel about a topic in the survey, use the survey questionnaire and codebook to match a question (the variable label) to a variable name. For instance, "How often have you felt lonely in the past 7 days?" is variable "soc5c".
Nationally: Go to this query and enter soc5c as the variable. Hit the blue Run Query button in the upper right hand corner.
Local or State: To find figures for that response in a specific state, go to this query and type in a state name and soc5c as the variable, and then hit the blue Run Query button in the upper right hand corner.
The resulting sentence you could write out of these queries is: "People in some states are less likely to report loneliness than others. For example, 66% of Louisianans report feeling lonely on none of the last seven days, compared with 52% of Californians. Nationally, 60% of people said they hadn't felt lonely."
The margin of error for the national and regional surveys is found in the attached methods statement. You will need the margin of error to determine if the comparisons are statistically significant. If the difference is:
The survey data will be provided under embargo in both comma-delimited and statistical formats.
Each set of survey data will be numbered and have the date the embargo lifts in front of it in the format of: 01_April_30_covid_impact_survey. The survey has been organized by the Data Foundation, a non-profit non-partisan think tank, and is sponsored by the Federal Reserve Bank of Minneapolis and the Packard Foundation. It is conducted by NORC at the University of Chicago, a non-partisan research organization. (NORC is not an abbreviation, it part of the organization's formal name.)
Data for the national estimates are collected using the AmeriSpeak Panel, NORC’s probability-based panel designed to be representative of the U.S. household population. Interviews are conducted with adults age 18 and over representing the 50 states and the District of Columbia. Panel members are randomly drawn from AmeriSpeak with a target of achieving 2,000 interviews in each survey. Invited panel members may complete the survey online or by telephone with an NORC telephone interviewer.
Once all the study data have been made final, an iterative raking process is used to adjust for any survey nonresponse as well as any noncoverage or under and oversampling resulting from the study specific sample design. Raking variables include age, gender, census division, race/ethnicity, education, and county groupings based on county level counts of the number of COVID-19 deaths. Demographic weighting variables were obtained from the 2020 Current Population Survey. The count of COVID-19 deaths by county was obtained from USA Facts. The weighted data reflect the U.S. population of adults age 18 and over.
Data for the regional estimates are collected using a multi-mode address-based (ABS) approach that allows residents of each area to complete the interview via web or with an NORC telephone interviewer. All sampled households are mailed a postcard inviting them to complete the survey either online using a unique PIN or via telephone by calling a toll-free number. Interviews are conducted with adults age 18 and over with a target of achieving 400 interviews in each region in each survey.Additional details on the survey methodology and the survey questionnaire are attached below or can be found at https://www.covid-impact.org.
Results should be credited to the COVID Impact Survey, conducted by NORC at the University of Chicago for the Data Foundation.
To learn more about AP's data journalism capabilities for publishers, corporations and financial institutions, go here or email kromano@ap.org.
MIT Licensehttps://opensource.org/licenses/MIT
License information was derived automatically
R script for Minimum Credence statistical test for comparing mixed samples. The test evaluates maximum possible population standard deviation from a reference sample and then applies this estimate to mixed samples for comparisons. In case the reference sample is less than 8, chi-squared approximation is used to find confidence interval for population standard deviation. Otherwise, the test bootstraps corrected median absolute deviation of the reference sample to obtain bias-corrected accelerated confidence interval for population standard deviation. This script is adapted for RFA soil analysis, but can be used elswhere if method=0. If method=1, the test extracts GOST metrological data (between-lab analytical error) from the excel file, and must be corrected for every specific analysis. The test implies homogeniety of variances among groups, and the studied parameter must average additively (i. e. arithmetic mean of individual samples is equal to mixed sample) in order to use mixed samples.For any questions/remarks please contact me at: fastovetsilya@yandex.ru
The 1991 Indonesia Demographic and Health Survey (IDHS) is a nationally representative survey of ever-married women age 15-49. It was conducted between May and July 1991. The survey was designed to provide information on levels and trends of fertility, infant and child mortality, family planning and maternal and child health. The IDHS was carried out as collaboration between the Central Bureau of Statistics, the National Family Planning Coordinating Board, and the Ministry of Health. The IDHS is follow-on to the National Indonesia Contraceptive Prevalence Survey conducted in 1987.
The DHS program has four general objectives: - To provide participating countries with data and analysis useful for informed policy choices; - To expand the international population and health database; - To advance survey methodology; and - To help develop in participating countries the technical skills and resources necessary to conduct demographic and health surveys.
In 1987 the National Indonesia Contraceptive Prevalence Survey (NICPS) was conducted in 20 of the 27 provinces in Indonesia, as part of Phase I of the DHS program. This survey did not include questions related to health since the Central Bureau of Statistics (CBS) had collected that information in the 1987 National Socioeconomic Household Survey (SUSENAS). The 1991 Indonesia Demographic and Health Survey (IDHS) was conducted in all 27 provinces of Indonesia as part of Phase II of the DHS program. The IDHS received financial assistance from several sources.
The 1991 IDHS was specifically designed to meet the following objectives: - To provide data concerning fertility, family planning, and maternal and child health that can be used by program managers, policymakers, and researchers to evaluate and improve existing programs; - To measure changes in fertility and contraceptive prevalence rates and at the same time study factors which affect the change, such as marriage patterns, urban/rural residence, education, breastfeeding habits, and the availability of contraception; - To measure the development and achievements of programs related to health policy, particularly those concerning the maternal and child health development program implemented through public health clinics in Indonesia.
National
Sample survey data [ssd]
Indonesia is divided into 27 provinces. For the implementation of its family planning program, the National Family Planning Coordinating Board (BKKBN) has divided these provinces into three regions as follows:
The 1990 Population Census of Indonesia shows that Java-Bali contains about 62 percent of the national population, while Outer Java-Bali I contains 27 percent and Outer Java-Bali II contains 11 percent. The sample for the Indonesia DHS survey was designed to produce reliable estimates of contraceptive prevalence and several other major survey variables for each of the 27 provinces and for urban and rural areas of the three regions.
In order to accomplish this goal, approximately 1500 to 2000 households were selected in each of the provinces in Java-Bali, 1000 households in each of the ten provinces in Outer Java-Bali I, and 500 households in each of the 11 provinces in Outer Java-Bali II for a total of 28,000 households. With an average of 0.8 eligible women (ever-married women age 15-49) per selected household, the 28,000 households were expected to yield approximately 23,000 individual interviews.
Note: See detailed description of sample design in APPENDIX A of the survey report.
Face-to-face [f2f]
The DHS model "A" questionnaire and manuals were modified to meet the requirements of measuring family planning and health program attainment, and were translated into Bahasa Indonesia.
The first stage of data editing was done by the field editors who checked the completed questionnaires for completeness and accuracy. Field supervisors also checked the questionnaires. They were then sent to the central office in Jakarta where they were edited again and open-ended questions were coded. The data were processed using 11 microcomputers and ISSA (Integrated System for Survey Analysis).
Data entry and editing were initiated almost immediately after the beginning of fieldwork. Simple range and skip errors were corrected at the data entry stage. Secondary machine editing of the data was initiated as soon as sufficient questionnaires had been entered. The objective of the secondary editing was to detect and correct, if possible, inconsistencies in the data. All of the data were entered and edited by September 1991. A brief report containing preliminary survey results was published in November 1991.
Of 28,141 households sampled, 27,109 were eligible to be interviewed (excluding those that were absent, vacant, or destroyed), and of these, 26,858 or 99 percent of eligible households were successfully interviewed. In the interviewed households, 23,470 eligible women were found and complete interviews were obtained with 98 percent of these women.
Note: See summarized response rates by place of residence in Table 1.2 of the survey report.
The results from sample surveys are affected by two types of errors, non-sampling error and sampling error. Non-sampling error is due to mistakes made in carrying out field activities, such as failure to locate and interview the correct household, errors in the way the questions are asked, misunderstanding on the part of either the interviewer or the respondent, data entry errors, etc. Although efforts were made during the design and implementation of the IDHS to minimize this type of error, non-sampling errors are impossible to avoid and difficult to evaluate analytically.
Sampling errors, on the other hand, can be measured statistically. The sample of women selected in the IDHS is only one of many samples that could have been selected from the same population, using the same design and expected size. Each one would have yielded results that differed somewhat from the actual sample selected. The sampling error is a measure of the variability between all possible samples; although it is not known exactly, it can be estimated from the survey results. Sampling error is usually measured in terms of standard error of a particular statistic (mean, percentage, etc.), which is the square root of the variance. The standard error can be used to calculate confidence intervals within which one can reasonably be assured that, apart from non-sampling errors, the true value of the variable for the whole population falls. For example, for any given statistic calculated from a sample survey, the value of that same statistic as measured in 95 percent of all possible samples with the same design (and expected size) will fall within a range of plus or minus two times the standard error of that statistic.
If the sample of women had been selected as a simple random sample, it would have been possible to use straightforward formulas for calculating sampling errors. However, the IDHS sample design depended on stratification, stages and clusters. Consequently, it was necessary to utilize more complex formulas. The computer package CLUSTERS, developed by the International Statistical Institute for the World Fertility Survey, was used to assist in computing the sampling errors with the proper statistical methodology.
Note: See detailed estimate of sampling error calculation in APPENDIX B of the survey report.
Data Quality Tables - Household age distribution - Age distribution of eligible and interviewed women - Completeness of reporting - Births by calendar year since birth - Reporting of age at death in days - Reporting of age at death in months
Note: See detailed tables in APPENDIX C of the survey report.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Evidence-based medicine: assessment of
knowledge of basic epidemiological and research methods among medical doctors
Submitted to Venera ma'am by Roshan Shinde Group 32
EVIDENCE BASED MEDICINE is the main source of new knowledge for doctors in this era. The main objectives of EBM are as follows,
To evaluate the knowledge of basic research methods and data analysis among medical doctors. To assess factors such as the country of the medical school graduation profession.
Importance of Research Competence:
1. The study emphasizes that a solid understanding of epidemiology and biostatistics is essential for doctors to critically appraise medical literature and make informed clinical decisions.
2. Previous Findings: Prior studies indicated that many doctors lack proficiency in research methods, with significant gaps in understanding key concepts of evidence-based medicine (EBM).
Materials and Methods
Data collection and study population
The study involved 40 departments and employed around 500 doctors.
A random selection of 15 departments was made for participant recruitment.
Data collection
A supervised, self-administered questionnaire was distributed during morning staff meetings.
The questionnaire consisted of 10 multiple-choice questions focused on basic epidemiology and statistics, along with demographic data.
Participants were divided into two groups based on their country of medical school graduation: those from the former Soviet Union (Eastern education) and those from other countries (Western education).
The questionnaire was completed anonymously, and all participants were efficient in Hebrew.
Questionnaire
1. Sections of the Questionnaire:
Personal Details: This section collected demographic information about the doctors, including:
• Country of graduation
• Year of graduation from medical school
• Professional status (whether they are specialists or residents)
• Reading and writing habits related to medical literature.
Knowledge Assessment: This section consisted of 10 multiple-choice questions focused on basic research methods and statistics, divided as follows:
Statistics: 5 questions
Epidemiology: 5 questions
2. Basis for Statistical Questions:
The questions on statistics were derived from a list of commonly used statistical methods identified by Emerson and Colditz in 1983. This list was previously utilized for quality evaluations of articles published in the New England Journal of Medicine and referenced in a similar study by Horton and Switzee. This approach ensures that the questions are relevant and grounded in established research practices.
3. Scoring Methodology:
• Any missing answers to questions on epidemiological and statistical methods were considered incorrect. This scoring method emphasizes the importance of attempting to answer all questions and reflects a strict approach to assessing knowledge.
• The decision to mark unanswered questions as incorrect may encourage participants to engage more thoughtfully with the questionnaire, although it could also discourage some from attempting to answer if they are unsure
To ensure validity of the questionnaire, the 10 questions assessing knowledge were given to 15 members of the Epidemiology Department, Ben‐Gurion University. All of them correctly answered all the questions.
Results:
Response Rate: Out of 260 eligible doctors, 219 completed the questionnaire (84.2% response rate).
Statistical methods
1. Comparison of Categorical Variables:
Chi-Squared Test (x²): This test was used to examine differences between categorical variables. It assesses whether the observed frequencies in each category differ from what would be expected under the null hypothesis.
Fisher's Exact Test: This test was employed when sample sizes were small or when the assumptions of the chi-squared test were not met. It is particularly useful for 2×2 contingency tables.
2. Comparison of Ordinal Variables:
Mann-Whitney U Test: This non-parametric test was used to compare ordinal variables with multiple values, such as the scores obtained from the questionnaire. It assesses whether the distributions of two independent samples differ.
3. Paired Comparisons:
Wilcoxon's Signed Rank Test: This non-parametric test was used for paired comparisons of scores. It evaluates whether the median of the differences between paired observations is significantly different from zero.
4. Correlation Analysis:
Spearman's Rank Correlation Coefficient: This test was used to estimate the correlation between continuous variables. It assesses how well the relationship between two variables can be described using a monotonic function.
5. Multivariable Analysis:
Linear Regression: This method was used to explain the final score based on multiple variables. The analysis adjusted for all variables that were found to be related in the univariable analysis with a p-value of less than 0.1. This approach helps to identify the independent effects of each variable on the outcome.
6. Significance Level:
A p-value of 0.05 was considered statistically significant, indicating that there is less than a 5% probability that the observed results occurred by chance.
7. Data Presentation:
Normally distributed variables were expressed as mean (standard deviation, SD), while non-normally distributed variables were presented as median and interquartile range (IQR). This distinction is important for accurately representing the data's distribution.
Table 2 depicts doctors' professional characteristics according to the country of medical school graduation. Of 219 participants, 84 (38.4%) graduated from the former Soviet republics. The remaining 135 doctors were distributed by the country of graduation as follows: Israel, 100 (45.7%); West and Central Europe, 22 (10.0%); Italy, 8; Germany, 3; Czech Republic, 3; Hungary, 3; Netherlands, 1; Romania, 4; South America, 10 (4.6%); Argentina, 5; Cuba, 3; Uruguay, 1; Brazil, 1; and North America, 3 (1.4%).
Time Elapsed Since Graduation:
• Doctors from Israel and other countries had a shorter time since graduation compared to those from the former Soviet Union:
• Foreign Graduates: 8 years
(Interquartile Range (IQR) 4-19)
Former Soviet Union Graduates: 10 years (IQR 6-19)
• The difference was statistically significant (p = 0.02), indicating that foreign graduates tended to have graduated more recently.
Professional Status:
There were fewer specialists among foreign graduates compared to those who graduated from Israel
Foreign Graduates: 32.8% were specialists
Israeli Graduates: 48.0% were specialists
This difference was also statistically significant (p = 0.02).
Choice of Residency:
There were notable differences in the choice of residency between the two groups:
Domestic Graduates: 29.3% chose pediatrics or obstetrics and gynecology
Conclusion
The analysis of doctors' professional characteristics based on their country of medical school graduation reveals important insights into the diversity of medical training backgrounds and their implications for specialization and residency choices. These findings underscore the need for ongoing evaluation of medical education and training systems to ensure that all graduates, regardless of their background, are adequately prepared to meet the healthcare needs of the population
Table 3 describes the reading and publishing habits of the participants. A total of 96% of the participants reported reading at least one article per week, whereas 35.2% usually read at least three articles. Specialists read significantly more articles per week—52.3% of them read at least three articles, compared with only 23.8% of the residents; p<0.001. Most of the doctors, 63.6%, participated in the writing of ⩽5 articles. Similar to the reading pattern, only 21.1% of the residents wrote ⩾6 articles compared with 44.0% of the specialists; p<0.001. The Spearman correlation value between reading and writing variables was 0.35; p<0.001
Conclusion
The analysis of reading and publishing habits among the study participants reveals important insights into the professional engagement of doctors with medical literature. The differences between specialists and residents, along with the positive correlation between reading and writing, underscore the need for targeted educational initiatives to enhance research literacy and foster a culture of inquiry within the medical community. Encouraging both reading and writing can contribute to the overall quality of medical practice and the advancement of evidence-based medicine.
Figure 1
The figure describes the average of correct answers to 10 questions in understanding different aspects of basic research methods. Two populations of doctors are compared: those who graduated in the former Soviet Union (Eastern type of education) and those who graduated in Israel, USA, Western and Central Europe,
The survey was conducted in South Sudan between July 2014 and December 2014 as part of Enterprise Surveys roll-out, an initiative of the World Bank. The objective of the survey is to obtain feedback from enterprises on the state of the private sector as well as to help in building a panel of enterprise data that will make it possible to track changes in the business environment over time, thus allowing, for example, impact assessments of reforms. Through interviews with firms in the manufacturing and services sectors, the survey assesses the constraints to private sector growth and creates statistically significant business environment indicators that are comparable across countries.
In South Sudan, data from 738 establishments was analyzed. Stratified random sampling was used to select the surveyed businesses. The data was collected using face-to-face interviews.
The standard Enterprise Survey topics include firm characteristics, gender participation, access to finance, annual sales, costs of inputs and labor, workforce composition, bribery, licensing, infrastructure, trade, crime, competition, capacity utilization, land and permits, taxation, informality, business-government relations, innovation and technology, and performance measures. Over 90 percent of the questions objectively ascertain characteristics of a country’s business environment. The remaining questions assess the survey respondents’ opinions on what are the obstacles to firm growth and performance.
National
The primary sampling unit of the study is an establishment. The establishment is a physical location where business is carried out and where industrial operations take place or services are provided. A firm may be composed of one or more establishments. For example, a brewery may have several bottling plants and several establishments for distribution. For the purposes of this survey an establishment must make its own financial decisions and have its own financial statements separate from those of the firm. An establishment must also have its own management and control over its payroll.
The whole population, or the universe, covered in the Enterprise Surveys is the non-agricultural private economy. It comprises: all manufacturing sectors according to the ISIC Revision 3.1 group classification (group D), construction sector (group F), services sector (groups G and H), and transport, storage, and communications sector (group I). Note that this population definition excludes the following sectors: financial intermediation (group J), real estate and renting activities (group K, except sub-sector 72, IT, which was added to the population under study), and all public or utilities sectors. Companies with 100% government ownership are not eligible to participate in the Enterprise Surveys.
Sample survey data [ssd]
The sample was selected using stratified random sampling. Two levels of stratification were used in this country: industry and region. The size was not available in the sampling frame for most contacts.
For industry stratification, the universe was divided into manufacturing sector and two service sectors (retail and other services).
Regional stratification was defined in four regions: Juba, Nimule, Torit and Yei.
The sampling frame was built using data from the National Bureau of Statistics as well are municipal commercial registries.
The sampling frame was generated with the aim of obtaining interviews at 720 establishments. Establishments with undefined size were included as part of this sample frame for South Sudan in order to ensure a representative sample. Size information collected during the survey process can then be used to categorize these firms. Establishments with undefined location were included as part of this sample frame for Sudan in order to ensure a representative sample. Location information collected during the survey process can then be used to categorize these firms.
Face-to-face [f2f]
The following survey instruments are available: - Manufacturing Module Questionnaire - Services Module Questionnaire
The survey is fielded via manufacturing or services questionnaires in order not to ask questions that are irrelevant to specific types of firms, e.g. a question that relates to production and nonproduction workers should not be asked of a retail firm. In addition to questions that are asked across countries, all surveys are customized and contain country-specific questions. An example of customization would be including tourism-related questions that are asked in certain countries when tourism is an existing or potential sector of economic growth.
The eligible manufacturing industries have been surveyed using the Manufacturing Module Questionnaire (includes a common set of core variables, plus manufacturing specific questions). Eligible service establishments have been covered using the Services Module Questionnaire. Each variation of the questionnaire is identified by the index variable, a0.
Data entry and quality controls are implemented by the contractor and data is delivered to the World Bank in batches (typically 10%, 50% and 100%). These data deliveries are checked for logical consistency, out of range values, skip patterns, and duplicate entries. Problems are flagged by the World Bank and corrected by the implementing contractor through data checks, callbacks, and revisiting establishments.
Survey non-response must be differentiated from item non-response. The former refers to refusals to participate in the survey altogether whereas the latter refers to the refusals to answer some specific questions. Enterprise Surveys suffer from both problems and different strategies were used to address these issues.
Item non-response was addressed by two strategies: a- For sensitive questions that may generate negative reactions from the respondent, such as corruption or tax evasion, enumerators were instructed to collect "Refusal to respond" (-8) as a different option from "Don't know" (-9). b- Establishments with incomplete information were re-contacted in order to complete this information, whenever necessary.
Survey non-response was addressed by maximizing efforts to contact establishments that were initially selected for interview. Attempts were made to contact the establishment for interview at different times/days of the week before a replacement establishment (with similar strata characteristics) was suggested for interview. Survey non-response did occur but substitutions were made in order to potentially achieve strata-specific goals.
The number of interviews per contacted establishments was 0.42. This number is the result of two factors: explicit refusals to participate in the survey, as reflected by the rate of rejection (which includes rejections of the screener and the main survey) and the quality of the sample frame, as represented by the presence of ineligible units. The number of rejections per contact was 0.07.
Most countries collect official statistics on energy use due to its vital role in the infrastructure, economy and living standards.
In Palestine, additional attention is warranted for energy statistics due to a scarcity of natural resources, the high cost of energy and high population density. These factors demand comprehensive and high quality statistics.
In this contest PCBS decided to conduct a special Energy Consumption in Transport Survey to provide high quality data about energy consumption by type, expenditure on maintenance and insurance for vehicles, and questions on vehicles motor capacity and year of production.
The survey aimed to provide data on energy consumption by transport sector and also on the energy consumption by the type of vehicles and its motor capacity and year of production.
Palestine
Vehicles
All the operating vehicles in Palestine in 2014.
Sample survey data [ssd]
Target Population: All the operating vehicles in Palestine in 2014.
2.1Sample Frame A list of the number of the operating vehicles in Palestine in 2014, they are broken down by governorates and vehicle types, this list was obtained from Ministry of transport.
2.2.1 Sample size The sample size is 6,974 vehicles.
2.2.2 Sampling Design it is stratified random sample, and in some of the small size strata the quota sample was used to cover them.
The method of reaching the vehicles sample was through : 1-reaching to all the dynamometers (the centers for testing the vehicles) 2-selecting a random sample of vehicles by type of vehicle, model, fuel type and engine capacity
Face-to-face [f2f]
The design of the questionnaire was based on the experiences of other similar countries in energy statistics subject to cover the most important indicators for energy statistics in transport sector, taking into account Palestine's particular situation.
The data processing stage consisted of the following operations: Editing and coding prior to data entry: all questionnaires were edited and coded in the office using the same instructions adopted for editing in the field.
Data entry: The survey questionnaire was uploaded on office computers. At this stage, data were entered into the computer using a data entry template developed in Access Database. The data entry program was prepared to satisfy a number of requirements: ·To prevent the duplication of questionnaires during data entry. ·To apply checks on the integrity and consistency of entered data. ·To handle errors in a user friendly manner. ·The ability to transfer captured data to another format for data analysis using statistical analysis software such as SPSS. Audit after data entered at this stage is data entered scrutiny by pulling the data entered file periodically and review the data and examination of abnormal values and check consistency between the different questions in the questionnaire, and if there are any errors in the data entered to be the withdrawal of the questionnaire and make sure this data and adjusted, even been getting the final data file that is the final extract data from it. Extraction Results: The extract final results of the report by using the SPSS program, and then display the results through tables to Excel format.
80.7%
Data of this survey may be affected by sampling errors due to use of a sample and not a complete enumeration. Therefore, certain differences are anticipated in comparison with the real values obtained through censuses. The variance was calculated for the most important indicators: the variance table is attached with the final report. There is no problem in the dissemination of results at national and regional level (North, Middle, South of West Bank, Gaza Strip).
The survey sample consisted of around 6,974 vehicles, of which 5,631 vehicles completed the questionnaire, 3,652 vehicles from the West Bank and 1,979 vehicles in Gaza Strip.
http://reference.data.gov.uk/id/open-government-licencehttp://reference.data.gov.uk/id/open-government-licence
Average (mean) rating for 'Life Satisfaction' by County and Unitary Authority in the First ONS Annual Experimental Subjective Wellbeing survey, April 2011 - March 2012.
The Office for National Statistics has included the four subjective well-being questions below on the Annual Population Survey (APS), the largest of their household surveys.
This dataset presents results from the first of these questions, "Overall, how satisfied are you with your life nowadays?" Respondents answer these questions on an 11 point scale from 0 to 10 where 0 is ‘not at all’ and 10 is ‘completely’. The well-being questions were asked of adults aged 16 and older.
Well-being estimates for each unitary authority or county are derived using data from those respondents who live in that place. Responses are weighted to the estimated population of adults (aged 16 and older) as at end of September 2011.
This dataset contains the mean responses: the average reported value for respondents resident in each area. It also contains the standard error, the sample size and lower and upper confidence limits at the 95% level.
The data cabinet also makes available the proportion of people in each county and unitary authority that answer with ‘low wellbeing’ values. For the ‘life satisfaction’ question answers in the range 0-6 are taken to be low wellbeing.
The ONS survey covers the whole of the UK, but this dataset only includes results for counties and unitary authorities in England, for consistency with other statistics available at this website.
At this stage the estimates are considered ‘experimental statistics’, published at an early stage to involve users in their development and to allow feedback. Feedback can be provided to the ONS via this email address.
The APS is a continuous household survey administered by the Office for National Statistics. It covers the UK, with the chief aim of providing between-census estimates of key social and labour market variables at a local area level. Apart from employment and unemployment, the topics covered in the survey include housing, ethnicity, religion, health and education. When a household is surveyed all adults (aged 16+) are asked the four subjective well-being questions.
The 12 month Subjective Well-being APS dataset is a sub-set of the general APS as the well-being questions are only asked of persons aged 16 and above, who gave a personal interview and proxy answers are not accepted. This reduces the size of the achieved sample to approximately 120,000 adult respondents in England.
The original data is available from the ONS website.
Detailed information on the APS and the Subjective Wellbeing dataset is available here.
As well as collecting data on well-being, the Office for National Statistics has published widely on the topic of wellbeing. Papers and further information can be found here.
CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
License information was derived automatically
This data set contains the replication data and supplements for the article "Knowing, Doing, and Feeling: A three-year, mixed-methods study of undergraduates’ information literacy development." The survey data is from two samples: - cross-sectional sample (different students at the same point in time) - longitudinal sample (the same students and different points in time)Surveys were distributed via Qualtrics during the students' first and sixth semesters. Quantitative and qualitative data were collected and used to describe students' IL development over 3 years. Statistics from the quantitative data were analyzed in SPSS. The qualitative data was coded and analyzed thematically in NVivo. The qualitative, textual data is from semi-structured interviews with sixth-semester students in psychology at UiT, both focus groups and individual interviews. All data were collected as part of the contact author's PhD research on information literacy (IL) at UiT. The following files are included in this data set: 1. A README file which explains the quantitative data files. (2 file formats: .txt, .pdf)2. The consent form for participants (in Norwegian). (2 file formats: .txt, .pdf)3. Six data files with survey results from UiT psychology undergraduate students for the cross-sectional (n=209) and longitudinal (n=56) samples, in 3 formats (.dat, .csv, .sav). The data was collected in Qualtrics from fall 2019 to fall 2022. 4. Interview guide for 3 focus group interviews. File format: .txt5. Interview guides for 7 individual interviews - first round (n=4) and second round (n=3). File format: .txt 6. The 21-item IL test (Tromsø Information Literacy Test = TILT), in English and Norwegian. TILT is used for assessing students' knowledge of three aspects of IL: evaluating sources, using sources, and seeking information. The test is multiple choice, with four alternative answers for each item. This test is a "KNOW-measure," intended to measure what students know about information literacy. (2 file formats: .txt, .pdf)7. Survey questions related to interest - specifically students' interest in being or becoming information literate - in 3 parts (all in English and Norwegian): a) information and questions about the 4 phases of interest; b) interest questionnaire with 26 items in 7 subscales (Tromsø Interest Questionnaire - TRIQ); c) Survey questions about IL and interest, need, and intent. (2 file formats: .txt, .pdf)8. Information about the assignment-based measures used to measure what students do in practice when evaluating and using sources. Students were evaluated with these measures in their first and sixth semesters. (2 file formats: .txt, .pdf)9. The Norwegain Centre for Research Data's (NSD) 2019 assessment of the notification form for personal data for the PhD research project. In Norwegian. (Format: .pdf)
The JPFHS is part of the worldwide Demographic and Health Surveys Program, which is designed to collect data on fertility, family planning, and maternal and child health. The primary objective of the Jordan Population and Family Health Survey (JPFHS) is to provide reliable estimates of demographic parameters, such as fertility, mortality, family planning, fertility preferences, as well as maternal and child health and nutrition that can be used by program managers and policy makers to evaluate and improve existing programs. In addition, the JPFHS data will be useful to researchers and scholars interested in analyzing demographic trends in Jordan, as well as those conducting comparative, regional or crossnational studies.
The content of the 2002 JPFHS was significantly expanded from the 1997 survey to include additional questions on women’s status, reproductive health, and family planning. In addition, all women age 15-49 and children less than five years of age were tested for anemia.
National
Sample survey data
The estimates from a sample survey are affected by two types of errors: 1) nonsampling errors and 2) sampling errors. Nonsampling errors are the result of mistakes made in implementing data collection and data processing, such as failure to locate and interview the correct household, misunderstanding of the questions on the part of either the interviewer or the respondent, and data entry errors. Although numerous efforts were made during the implementation of the 2002 JPFHS to minimize this type of error, nonsampling errors are impossible to avoid and difficult to evaluate statistically.
Sampling errors, on the other hand, can be evaluated statistically. The sample of respondents selected in the 2002 JPFHS is only one of many samples that could have been selected from the same population, using the same design and expected size. Each of these samples would yield results that differ somewhat from the results of the actual sample selected. Sampling errors are a measure of the variability between all possible samples. Although the degree of variability is not known exactly, it can be estimated from the survey results.
A sampling error is usually measured in terms of the standard error for a particular statistic (mean, percentage, etc.), which is the square root of the variance. The standard error can be used to calculate confidence intervals within which the true value for the population can reasonably be assumed to fall. For example, for any given statistic calculated from a sample survey, the value of that statistic will fall within a range of plus or minus two times the standard error of that statistic in 95 percent of all possible samples of identical size and design.
If the sample of respondents had been selected as a simple random sample, it would have been possible to use straightforward formulas for calculating sampling errors. However, the 2002 JPFHS sample is the result of a multistage stratified design and, consequently, it was necessary to use more complex formulas. The computer software used to calculate sampling errors for the 2002 JPFHS is the ISSA Sampling Error Module (ISSAS). This module used the Taylor linearization method of variance estimation for survey estimates that are means or proportions. The Jackknife repeated replication method is used for variance estimation of more complex statistics such as fertility and mortality rates.
Note: See detailed description of sample design in APPENDIX B of the survey report.
Face-to-face
The 2002 JPFHS used two questionnaires – namely, the Household Questionnaire and the Individual Questionnaire. Both questionnaires were developed in English and translated into Arabic. The Household Questionnaire was used to list all usual members of the sampled households and to obtain information on each member’s age, sex, educational attainment, relationship to the head of household, and marital status. In addition, questions were included on the socioeconomic characteristics of the household, such as source of water, sanitation facilities, and the availability of durable goods. The Household Questionnaire was also used to identify women who are eligible for the individual interview: ever-married women age 15-49. In addition, all women age 15-49 and children under five years living in the household were measured to determine nutritional status and tested for anemia.
The household and women’s questionnaires were based on the DHS Model “A” Questionnaire, which is designed for use in countries with high contraceptive prevalence. Additions and modifications to the model questionnaire were made in order to provide detailed information specific to Jordan, using experience gained from the 1990 and 1997 Jordan Population and Family Health Surveys. For each evermarried woman age 15 to 49, information on the following topics was collected:
In addition, information on births and pregnancies, contraceptive use and discontinuation, and marriage during the five years prior to the survey was collected using a monthly calendar.
Fieldwork and data processing activities overlapped. After a week of data collection, and after field editing of questionnaires for completeness and consistency, the questionnaires for each cluster were packaged together and sent to the central office in Amman where they were registered and stored. Special teams were formed to carry out office editing and coding of the open-ended questions.
Data entry and verification started after one week of office data processing. The process of data entry, including one hundred percent re-entry, editing and cleaning, was done by using PCs and the CSPro (Census and Survey Processing) computer package, developed specially for such surveys. The CSPro program allows data to be edited while being entered. Data processing operations were completed by the end of October 2002. A data processing specialist from ORC Macro made a trip to Jordan in October and November 2002 to follow up data editing and cleaning and to work on the tabulation of results for the survey preliminary report. The tabulations for the present final report were completed in December 2002.
A total of 7,968 households were selected for the survey from the sampling frame; among those selected households, 7,907 households were found. Of those households, 7,825 (99 percent) were successfully interviewed. In those households, 6,151 eligible women were identified, and complete interviews were obtained with 6,006 of them (98 percent of all eligible women). The overall response rate was 97 percent.
Note: See summarized response rates by place of residence in Table 1.1 of the survey report.
The estimates from a sample survey are affected by two types of errors: 1) nonsampling errors and 2) sampling errors. Nonsampling errors are the result of mistakes made in implementing data collection and data processing, such as failure to locate and interview the correct household, misunderstanding of the questions on the part of either the interviewer or the respondent, and data entry errors. Although numerous efforts were made during the implementation of the 2002 JPFHS to minimize this type of error, nonsampling errors are impossible to avoid and difficult to evaluate statistically.
Sampling errors, on the other hand, can be evaluated statistically. The sample of respondents selected in the 2002 JPFHS is only one of many samples that could have been selected from the same population, using the same design and expected size. Each of these samples would yield results that differ somewhat from the results of the actual sample selected. Sampling errors are a measure of the variability between all possible samples. Although the degree of variability is not known exactly, it can be estimated from the survey results.
A sampling error is usually measured in terms of the standard error for a particular statistic (mean, percentage, etc.), which is the square root of the variance. The standard error can be used to calculate confidence intervals within which the true value for the population can reasonably be assumed to fall. For example, for any given statistic calculated from a sample survey, the value of that statistic will fall within a range of plus or minus two times the standard error of that statistic in 95 percent of all possible samples of identical size and design.
If the sample of respondents had been selected as a simple random sample, it would have been possible to use straightforward formulas for calculating sampling errors. However, the 2002 JPFHS sample is the result of a multistage stratified design and, consequently, it was necessary to use more complex formulas. The computer software used to calculate sampling errors for the 2002 JPFHS is the ISSA Sampling Error Module (ISSAS). This module used the Taylor linearization method of variance estimation for survey estimates that are means or proportions. The Jackknife repeated replication method is used for variance estimation of more complex statistics such as fertility and mortality rates.
Note: See detailed