Facebook
Twitterhttps://creativecommons.org/publicdomain/zero/1.0/https://creativecommons.org/publicdomain/zero/1.0/
This is the large data set as featured in the OCR H240 exam series.
Questions about this dataset will be featured in the statistics paper
The LDS is a .xlsx file containing 5 tables, four data, one information. The data is drawn from the UK censuses from the years 2001 and 2011. It is designed for you to make comparisons and analyses of the changes in demographic and behavioural features of the populace. There is the age structure of each local authority and the method of travel within each local authority.
Facebook
TwitterThis dataset was created by Tom Button
Facebook
Twitterhttps://creativecommons.org/publicdomain/zero/1.0/https://creativecommons.org/publicdomain/zero/1.0/
By Huggingface Hub [source]
The Airoboros-3.1 dataset is the perfect tool to help machine learning models excel in the difficult realm of complicated mathematical operations. This data collection features thousands of conversations between machines and humans, formatted in ShareGPT to maximize optimization in an OS ecosystem. The dataset’s focus on advanced subjects like factorials, trigonometry, and larger numerical values will help drive machine learning models to the next level - facilitating critical acquisition of sophisticated mathematical skills that are essential for ML success. As AI technology advances at such a rapid pace, training neural networks to correspondingly move forward can be a daunting and complicated challenge - but with Airoboros-3.1’s powerful datasets designed around difficult mathematical operations it just became one step closer to achievable!
For more datasets, click here.
- 🚨 Your notebook can be here! 🚨!
To get started, download the dataset from Kaggle and use the train.csv file. This file contains over two thousand examples of conversations between ML models and humans which have been formatted using ShareGPT - fast and efficient OS ecosystem fine-tuning tools designed to help with understanding mathematical operations more easily. The file includes two columns: category and conversations, both of which are marked as strings in the data itself.
Once you have downloaded the train file you can begin setting up your own ML training environment by using any of your preferred frameworks or methods. Your model should focus on predicting what kind of mathematical operations will likely be involved in future conversations by referring back to previous dialogues within this dataset for reference (category column). You can also create your own test sets from this data, adding new conversation topics either by modifying existing rows or creating new ones entirely with conversation topics related to mathematics. Finally, compare your model’s results against other established models or algorithms that are already published online!
Happy training!
- It can be used to build custom neural networks or machine learning algorithms that are specifically designed for complex mathematical operations.
- This data set can be used to teach and debug more general-purpose machine learning models to recognize large numbers, and intricate calculations within natural language processing (NLP).
- The Airoboros-3.1 dataset can also be utilized as a supervised learning task: models could learn from the conversations provided in the dataset how to respond correctly when presented with complex mathematical operations
If you use this dataset in your research, please credit the original authors. Data Source
License: CC0 1.0 Universal (CC0 1.0) - Public Domain Dedication No Copyright - You can copy, modify, distribute and perform the work, even for commercial purposes, all without asking permission. See Other Information.
File: train.csv | Column name | Description | |:------------------|:-----------------------------------------------------------------------------| | category | The type of mathematical operation being discussed. (String) | | conversations | The conversations between the machine learning model and the human. (String) |
If you use this dataset in your research, please credit the original authors. If you use this dataset in your research, please credit Huggingface Hub.
Facebook
TwitterOpen Government Licence 3.0http://www.nationalarchives.gov.uk/doc/open-government-licence/version/3/
License information was derived automatically
Entries and passes for below level 3 English and maths, by qualification type and student characteristics. Includes entries for students triggered for inclusion in performance tables, after discounting of exams at the major qualification level.
Facebook
TwitterApache License, v2.0https://www.apache.org/licenses/LICENSE-2.0
License information was derived automatically
Maths-Grade-School I am releasing large Grade School level Mathematics datatset. This extensive dataset, comprising nearly one million instructions in JSON format, encapsulates a diverse array of topics fundamental to building a strong mathematical foundation. This dataset is in instruction format so that model developers, researchers etc. can easily use this dataset. Following Fields & sub Fields are covered: Calculus Probability Algebra Liner Algebra Trigonometry Differential Equations… See the full description on the dataset page: https://huggingface.co/datasets/pt-sk/Maths-Grade-School.
Facebook
TwitterAttribution-NonCommercial 4.0 (CC BY-NC 4.0)https://creativecommons.org/licenses/by-nc/4.0/
License information was derived automatically
Mathematical reasoning, a fundamental aspect of human cognition, poses significant challenges for artificial intelligence (AI) systems. Despite recent advancements in natural language processing (NLP) and large language models (LLMs), AI's ability to replicate human-like reasoning, generalization, and efficiency remains an ongoing research challenge. In this dissertation, we address key limitations in MWP solving, focusing on the accuracy, generalization ability and efficiency of AI-based mathematical reasoners by applying human-like reasoning methods and principles.
This dissertation introduces several innovative approaches in mathematical reasoning. First, a numeracy-driven framework is proposed to enhance math word problem (MWP) solvers by integrating numerical reasoning into model training, surpassing human-level performance on benchmark datasets. Second, a novel multi-solution framework captures the diversity of valid solutions to math problems, improving the generalization capabilities of AI models. Third, a customized knowledge distillation technique, termed Customized Exercise for Math Learning (CEMAL), is developed to create tailored exercises for smaller models, significantly improving their efficiency and accuracy in solving MWPs. Additionally, a multi-view fine-tuning paradigm (MinT) is introduced to enable smaller models to handle diverse annotation styles from different datasets, improving their adaptability and generalization. To further advance mathematical reasoning, a benchmark, MathChat, is introduced to evaluate large language models (LLMs) in multi-turn reasoning and instruction-following tasks, demonstrating significant performance improvements. Finally, new inference-time verifiers, Math-Rev and Code-Rev, are developed to enhance reasoning verification, combining language-based and code-based solutions for improved accuracy in both math and code reasoning tasks.
In summary, this dissertation provides a comprehensive exploration of these challenges and contributes novel solutions that push the boundaries of AI-driven mathematical reasoning. Potential future research directions are also discussed to further extend the impact of this dissertation.
Facebook
TwitterMIT Licensehttps://opensource.org/licenses/MIT
License information was derived automatically
Challenging the Boundaries of Reasoning: An Olympiad-Level Math Benchmark for Large Language Models
This is the official huggingface repository for Challenging the Boundaries of Reasoning: An Olympiad-Level Math Benchmark for Large Language Models by Haoxiang Sun, Yingqian Min, Zhipeng Chen, Wayne Xin Zhao, Zheng Liu, Zhongyuan Wang, Lei Fang, and Ji-Rong Wen. We have also released the OlymMATH-eval dataset on HuggingFace 🤗, together with a data visualization tool OlymMATH-demo… See the full description on the dataset page: https://huggingface.co/datasets/RUC-AIBOX/OlymMATH.
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
This domain covers statistics and indicators on key aspects of the education systems across Europe. The data show entrants and enrolments in education levels, education personnel and the cost and type of resources dedicated to education.
For a general technical description of the UOE Data Collection see UNESCO OECD Eurostat (UOE) joint data collection – methodology - Statistics Explained (europa.eu).
The standards on international statistics on education and training systems are set by the three international organisations jointly administering the annual UOE data collection:
The following topics are covered:
Data on enrolments in education are disseminated in absolute numbers, with breakdowns available for the following dimensions:
Additionally, the following types of indicators on enrolments are calculated (all indicators using population data use Eurostat’s population database (demo_pjan)):
Data on entrants in education are disseminated in absolute numbers, with breakdowns available for the following dimensions:
Additionally the following indicator on entrants is calculated:
Data on learning mobility is available for degree mobile students, degree mobile graduates and credit mobile graduates. Degree mobility means that students/graduates are/were enrolled as regular students in any semester/term of a programme taught in the country of destination with the intention of graduating from it in the country of destination. Credit mobility is defined as temporary tertiary education or/and study-related traineeship abroad within the framework of enrolment in a tertiary education programme at a "home institution" (usually) for the purpose of gaining academic credit (i.e. credit that will be recognised in that home institution). Further definitions are in Section 2.8 of the UOE manual.
Degree mobile students are referred to as just ‘mobile students’ in UOE learning mobility tables. Data is disseminated for degree mobile students and degree mobile graduates in absolute numbers with breakdowns available for the following dimensions:
Additionally the following types of indicators on degree mobile students and degree mobile graduates are calculated ((all indicators using population data use Eurostat’s population database (demo_pjan)):
For credit mobile graduates, data are disseminated in absolute numbers, with breakdowns available for the following dimensions:
Data on personnel in education are available for classroom teachers/academic staff, teacher aides and school-management personnel. Teachers are employed in a professional capacity to guide and direct the learning experiences of students, irrespective of their training, qualifications or delivery mechanism. Teacher aides support teachers in providing instruction to students. Academic staff are personnel employed at the tertiary level of education whose primary assignment is instruction and/or research. School management personnel covers professional personnel who are responsible for school management/administration (ISCED 0-4) or whose primary or major responsibility is the management of the institution, or a recognised department or subdivision of the institution (tertiary levels). Full definitions of these statistical units are in Section 3.5 of the UOE manual.
Data are disseminated on teachers and academic staff in absolute numbers, with breakdowns available for the following dimensions:
Facebook
TwitterApache License, v2.0https://www.apache.org/licenses/LICENSE-2.0
License information was derived automatically
Lean Workbook
This dataset is about contest-level math problems formalized in Lean 4. Our dataset contains 57231 problems in the split of Lean Workbook and 82893 problems in the split of Lean Workbook Plus. We provide the natural language statement, answer, formal statement, and formal proof (if available) for each problem. These data can support autoformalization model training and searching for proofs. We open-source our code and our data. Our test environment is based on Lean… See the full description on the dataset page: https://huggingface.co/datasets/internlm/Lean-Workbook.
Facebook
Twitterhttps://www.ontario.ca/page/open-government-licence-ontariohttps://www.ontario.ca/page/open-government-licence-ontario
Data includes: board and school information, grade 3 and 6 EQAO student achievements for reading, writing and mathematics, and grade 9 mathematics EQAO and OSSLT. Data excludes private schools, Education and Community Partnership Programs (ECPP), summer, night and continuing education schools.
How Are We Protecting Privacy?
Results for OnSIS and Statistics Canada variables are suppressed based on school population size to better protect student privacy. In order to achieve this additional level of protection, the Ministry has used a methodology that randomly rounds a percentage either up or down depending on school enrolment. In order to protect privacy, the ministry does not publicly report on data when there are fewer than 10 individuals represented.
The information in the School Information Finder is the most current available to the Ministry of Education at this time, as reported by schools, school boards, EQAO and Statistics Canada. The information is updated as frequently as possible.
This information is also available on the Ministry of Education's School Information Finder website by individual school.
Descriptions for some of the data types can be found in our glossary.
School/school board and school authority contact information are updated and maintained by school boards and may not be the most current version. For the most recent information please visit: https://data.ontario.ca/dataset/ontario-public-school-contact-information.
Facebook
Twitterhttps://creativecommons.org/publicdomain/zero/1.0/https://creativecommons.org/publicdomain/zero/1.0/
If this Data Set is useful, and upvote is appreciated. This data approach student achievement in secondary education of two Portuguese schools. The data attributes include student grades, demographic, social and school related features) and it was collected by using school reports and questionnaires. Two datasets are provided regarding the performance in two distinct subjects: Mathematics (mat) and Portuguese language (por). In [Cortez and Silva, 2008], the two datasets were modeled under binary/five-level classification and regression tasks. Important note: the target attribute G3 has a strong correlation with attributes G2 and G1. This occurs because G3 is the final year grade (issued at the 3rd period), while G1 and G2 correspond to the 1st and 2nd-period grades. It is more difficult to predict G3 without G2 and G1, but such prediction is much more useful (see paper source for more details).
Facebook
TwitterThe datasets used to produce this release are available for England, Wales and Northern Ireland.
We welcome your feedback on our publications. Should you have any comments on this statistical release and how to improve it to meet your needs please contact us as statistics@ofqual.gov.uk.
Facebook
TwitterNearly 2.8 million certificates were awarded in between July and September 2017, a decrease of 15.5% on the same period of 2016. The decline is mostly due to a decrease in the number of certificates in the qualifications and credit framework (QCF) and other general qualifications. There are also large decreases in the number of certificates in functional skills, key skills, free-standing maths and entry level. The general decline in overall certification numbers may be caused by a tightening in the availability of funding. This is notable at entry level, level 1, level 2 and level 1/2 qualifications.
Functional skills qualifications continue to replace key skills qualifications leading to a reduction in the number of certificates in the latter.
The reduction in the ‘other general qualifications’ may be an effect of the introduction of the English Baccalaureate and other school performance indicators. For example, the calculation of Progress 8 and Attainment 8 measures can only include a maximum of 3 non-English Baccalaureate qualifications.
The largest increase in number of certificates (59.1%) was seen in vocationally-related qualifications. This is likely caused by awarding organisations re-assigning the qualification type of QCF qualifications to vocationally-related qualification. Following the closure of the QCF unit bank and introduction of the regulated qualifications framework (RQF), Ofqual decided that inclusion of the term ‘QCF’ in qualification titles after 31 December 2017 would be an indicator of non-compliance with Ofqual’s titling rules. As well as amending qualification titles, awarding organisations are therefore likely to be re-assigning the qualification type. A concession to the inclusion of the term “QCF” has been given to applied general qualifications that have similar titles but differing assessment (pre-existing and newly introduced with 40% assessment) allowing differentiation between them.
The sector subject area with notable increase in number of certificates was construction, planning and the built environment.
The sector subject areas with notable decrease in number of certificates were languages, literature and culture, preparation for life and work, information and communication technology, and science and mathematics.
The qualification with the highest number of certificates this quarter was ‘BCS Level 2 ECDL Certificate in IT Application Skills’, followed by ‘Pearson BTEC Level 1/Level 2 First Award in Sport’ and ‘WJEC Foundation/National Skills Challenge Certificate (Welsh Baccalaureate)’.
The datasets used to produce this release are available for England, Wales and Northern Ireland.
Definitions for some of the specific terms used in our statistical bulletins are explained in the ‘Glossary for Ofqual’s statistics’.
We welcome your feedback on our publications. Should you have any comments on this statistical release and how to improve it to meet your needs please contact us as statistics@ofqual.gov.uk.
Facebook
TwitterApache License, v2.0https://www.apache.org/licenses/LICENSE-2.0
License information was derived automatically
Dataset Summary GSM8K (Grade School Math 8K) is a dataset of 8.5K high quality linguistically diverse grade school math word problems. The dataset was created to support the task of question answering on basic mathematical problems that require multi-step reasoning.
These problems take between 2 and 8 steps to solve. Solutions primarily involve performing a sequence of elementary calculations using basic arithmetic operations (+ − ×÷) to reach the final answer. A bright middle school student should be able to solve every problem: from the paper, "Problems require no concepts beyond the level of early Algebra, and the vast majority of problems can be solved without explicitly defining a variable." Solutions are provided in natural language, as opposed to pure math expressions. From the paper: "We believe this is the most generally useful data format, and we expect it to shed light on the properties of large language models’ internal monologues"
Facebook
TwitterThe data cover regulated qualifications in England.
The dataset used to produce this release are available separately.
All our published vocational and other qualifications publications are available at a single collection page.
We welcome your feedback on our publications. Should you have any comments on this statistical release and how to improve it to meet your needs please contact us at statistics@ofqual.gov.uk.
Facebook
TwitterIn 1991 the International Institute for Educational Planning (IIEP) and a number of Ministries of Education in Southern and Eastern Africa began to work together in order to address training and research needs in Education. The focus for this work was on establishing long-term strategies for building the capacity of educational planners to monitor and evaluate the quality of their basic education systems. The first two educational policy research projects undertaken by SACMEQ (widely known as "SACMEQ I" and "SACMEQ II") were designed to provide detailed information that could be used to guide planning decisions aimed at improving the quality of education in primary school systems.
During 1995-1998 seven Ministries of Education participated in the SACMEQ I Project. The SACMEQ II Project commenced in 1998 and the surveys of schools, involving 14 Ministries of Education, took place between 2000 and 2004. The survey was undertaken in schools in Botswana, Kenya, Lesotho, Malawi, Mauritius, Mozambique, Namibia, Seychelles, South Africa, Swaziland, Tanzania, Uganda, Zambia and Zanzibar.
Moving from the SACMEQ I Project (covering around 1100 schools and 20,000 pupils) to the SACMEQ II Project (covering around 2500 schools and 45,000 pupils) resulted in a major increase in the scale and complexity of SACMEQ's research and training programmes.
SACMEQ's mission is to: a) Expand opportunities for educational planners to gain the technical skills required to monitor and evaluate the quality of their education systems; and b) Generate information that can be used by decision-makers to plan and improve the quality of education.
National coverage
The target population for SACMEQ's Initial Project was defined as "all pupils at the Grade 6 level in 1995 who were attending registered government or non-government schools". Grade 6 was chosen because it was the grade level where the basics of reading literacy were expected to have been acquired.
Sample survey data [ssd]
The sample designs used in the SACMEQ II Project were selected so as to meet the standards set down by the International Association for the Evaluation of Educational Achievement. These standards required that sample estimates of important pupil population parameters should have sampling accuracy that was at least equivalent to a simple random sample of 400 pupils (thereby guaranteeing 95 percent confidence limits for sample means of plus or minus one tenth of a pupil standard deviation unit).
Some Constraints on Sample Design Sample designs in the field of education are usually prepared amid a network of competing constraints. These designs need to adhere to established survey sampling theory and, at the same time, give due recognition to the financial, administrative, and socio-political settings in which they are to be applied. The "best" sample design for a particular project is one that provides levels of sampling accuracy that are acceptable in terms of the main aims of the project, while simultaneously limiting cost, logistic, and procedural demands to manageable levels. The major constraints that were established prior to the preparation of the sample designs for the SACMEQ II Project have been listed below.
Target Population: The target population definitions should focus on Grade 6 pupils attending registered mainstream government or non-government schools. In addition, the defined target population should be constructed by excluding no more than 5 percent of pupils from the desired target population.
Bias Control: The sampling should conform to the accepted rules of scientific probability sampling. That is, the members of the defined target population should have a known and non-zero probability of selection into the sample so that any potential for bias in sample estimates due to variations from "epsem sampling" (equal probability of selection method) may be addressed through the use of appropriate sampling weights (Kish, 1965).
Sampling Errors: The sample estimates for the main criterion variables should conform to the sampling accuracy requirements set down by the International Association for the Evaluation of Educational Achievement (Ross, 1991). That is, the standard error of sampling for the pupil tests should be of a magnitude that is equal to, or smaller than, what would be achieved by employing a simple random sample of 400 pupils (Ross, 1985).
Response Rates: Each SACMEQ country should aim to achieve an overall response rate for pupils of 80 percent. This figure was based on the wish to achieve or exceed a response rate of 90 percent for schools and a response rate of 90 percent for pupils within schools.
Administrative and Financial Costs: The number of schools selected in each country should recognize limitations in the administrative and financial resources available for data collection.
Other Constraints: The number of pupils selected to participate in the data collection in each selected school should be set at a level that will maximize validity of the within-school data collection for the pupil reading and mathematics tests.
Note: Detailed descriptions of the sample design, sample selection, and sample evaluation procedures have been presented in the "South Africa Working Report".
Face-to-face [f2f]
The data collection for SACMEQ’s Initial Project took place in October 1995 and involved the administration of questionnaires to pupils, teachers, and school heads. The pupil questionnaire contained questions about the pupils’ home backgrounds and their school life; the teacher questionnaire asked about classrooms, teaching practices, working conditions, and teacher housing; and the school head questionnaire collected information about teachers, enrolments, buildings, facilities, and management. A reading literacy test was also given to the pupils. The test was based on items that were selected after a trial-testing programme had been completed.
Data Checking and Data Entry Data preparation commenced soon after the main data collection was completed. The NRCs had to organize the safe return of all materials to the Ministry of Education where the data collection instruments could be checked, entered into computers, and then "cleaned" to remove errors prior to data analysis. The data-checking involved the "hand editing" of data collection instruments by a team of trained staff. They were required to check that: (i) all questionnaires, tests, and forms had arrived back from the sample schools, (ii) the identification numbers on all instruments were complete and accurate, and (iii) certain logical linkages between questions made sense (for example, the two questions to school heads concerning "Do you have a school library?" and "How many books do you have in your school library?").
The next step was the entry of data into computers using the WINDEM software. A team of 5-10 staff normally undertook this work. In some cases the data were "double entered" in order to monitor accuracy.
The numbers of keystrokes required to enter one copy of each data collection instrument were as follows: pupil questionnaire: 150; pupil reading test: 85; pupil mathematics test: 65; teacher questionnaire: 587; teacher reading test: 51; teacher mathematics test: 43; school head questionnaire: 319; school form: 58; and pupil name form: 51.
This information can be re-expressed to give the total number of keystrokes for the whole body of data for one country by multiplying the above figures by the number of instruments in the final data collection. In the case of South Africa the total number of keystrokes was as follows: pupil questionnaire: 472 450; pupil reading test: 269 855; pupil mathematics test: 205 595; teacher questionnaire: 198 406; school head questionnaire: 62 361; school form: 9 802; and pupil name form: 161 313. That is, a total of 907 332 keystrokes were required to enter all of the data for South Africa.
An experienced keyboard operator can work at a rate of 25 keystrokes per minute (working from multi-paged questionnaires and stopping occasionally to clarify individual questionnaire entries with the supervisor). Assuming that this kind of work rate could be sustained for, say, around a maximum of six hours per day, then the whole data entry operation for South Africa was estimated to amount to around 101 person days of data entry work for South Africa.
Data Cleaning The NRCs received written instructions and follow-up support from IIEP staff in the basic steps of data cleaning using the WINDEM software. This permitted the NRCs to (i) identify major errors in the sequence of identification numbers, (ii) cross-check identification numbers across files (for example, to ensure that all pupils were linked with their own reading and mathematics teachers), (iii) ensure that all schools listed on the original sampling frame also had valid data collection instruments and vice-versa, (iv) check for "wild codes" that occurred when some variables had values that fell outside pre-specified reasonable limits, and (v) validate that variables used as linkage devices in later file merges were available and accurate.
A second phase of data preparation directed efforts towards the identification and correction of "wild codes" (which refer to data values that that fall outside credible limits), and "inconsistencies" (which refer to different responses to the same, or related, questions). There were also some errors in the identification codes for teachers
Facebook
TwitterTIMSS 2007 is the fourth in a cycle of internationally comparative assessments dedicated to improving teaching and learning in mathematics and science for students around the world. Carried out every four years at the fourth and eighth grades, TIMSS provides data about trends in mathematics and science achievement over time.
To inform educational policy in the participating countries, this world-wide assessment and research project also routinely collects extensive background information that addresses concerns about the quantity, quality, and content of instruction.
The survey had international coverage
Individuals and institutions
TIMSS 2007 chose to study achievement in two target populations-the fourth and eighth grade in most countries. Participating countries were free to select either population or both. The formal definitions of the TIMSS target populations make use of UNESCO's International Standard Classification of Education (ISCED) (UNESCO Institute for Statistics, 1999) in identifying the appropriate target grades: - Fourth grade population. This includes all students enrolled in the grade that represents 4 years of formal schooling, counting from the first year of ISCED Level 1, provided that the mean age at the time of testing is at least 9.5 years. For most countries, the target grade should be the fourth grade or its national equivalent.
Sample survey data
A systematic, two-stage probability proportional-to-size (PPS) sampling technique was used, where schools are first sampled and then classes within sampled (and participating) schools. . Because of its large population sizes, it was necessary to include a preliminary sampling stage in the Russian Federation, where regions were sampled first and then schools. Singapore also had a third sampling stage, where students were sampled within classes
Participants could exclude schools from the sampling frame if they were in geographically remote regions, were extremely small, offered curriculum or structure different from the mainstream, or provided instruction only to students in the “within-school” exclusion categories. The general TIMSS rules for defining within-school exclusions can be found in the technical documents.
Face-to-face [f2f]
The study used the following questionnaires: Fourth Grade Student Questionnaire, Fourth Grade Teacher Questionnaire, Fourth Grade School Questionnaire, Eighth Grade Student Questionnaire, Eighth Grade Mathematics Teacher Questionnaire, Eighth Grade Science Teacher Questionnaire, and Eighth Grade School Questionnaire. Information on the variables obtained or derived from questions in the survey is available in the TIMSS 2007 user guide for the international database: Data Supplement3: Variables derived from the Student, Teacher, and School Questionnaire data.
Weighted and unweighted response rates were computed for each participating country by grade, at the school level, and at the student level. Overall response rates (combined school and student response rates) also were computed.
Facebook
TwitterIn 1991 the International Institute for Educational Planning (IIEP) and a number of Ministries of Education in Southern and Eastern Africa began to work together in order to address training and research needs in Education. The focus for this work was on establishing long-term strategies for building the capacity of educational planners to monitor and evaluate the quality of their basic education systems. The first two educational policy research projects undertaken by SACMEQ (widely known as "SACMEQ I" and "SACMEQ II") were designed to provide detailed information that could be used to guide planning decisions aimed at improving the quality of education in primary school systems.
During 1995-1998 seven Ministries of Education participated in the SACMEQ I Project. The SACMEQ II Project commenced in 1998 and the surveys of schools, involving 14 Ministries of Education, took place between 2000 and 2004. The survey was undertaken in schools in Botswana, Kenya, Lesotho, Malawi, Mauritius, Mozambique, Namibia, Seychelles, South Africa, Swaziland, Tanzania, Uganda, Zambia and Zanzibar.
Moving from the SACMEQ I Project (covering around 1100 schools and 20,000 pupils) to the SACMEQ II Project (covering around 2500 schools and 45,000 pupils) resulted in a major increase in the scale and complexity of SACMEQ's research and training programmes.
SACMEQ's mission is to: a) Expand opportunities for educational planners to gain the technical skills required to monitor and evaluate the quality of their education systems; and b) Generate information that can be used by decision-makers to plan and improve the quality of education.
National coverage
The target population for SACMEQ's Initial Project was defined as "all pupils at the Grade 6 level in 1995 who were attending registered government or non-government schools". Grade 6 was chosen because it was the grade level where the basics of reading literacy were expected to have been acquired.
Sample survey data [ssd]
Sampling The "best" sample design for a particular project is one that provides levels of sampling accuracy that are acceptable in terms of the main aims of the project, while simultaneously limiting cost, logistic, and procedural demands to manageable levels. The major constraints that were established prior to the preparation of the sample designs for the SACMEQ II Project have been listed below.
Target Population: The target population definitions should focus on Grade 6 pupils attending registered mainstream government or non-government schools. In addition, the defined target population should be constructed by excluding no more than 5 percent of pupils from the desired target population.
Bias Control: The sampling should conform to the accepted rules of scientific probability sampling. That is, the members of the defined target population should have a known and non-zero probability of selection into the sample so that any potential for bias in sample estimates due to variations from "epsem sampling" (equal probability of selection method) could be addressed through the use of appropriate sampling weights.
Sampling Errors: The sample estimates for the main criterion variables should conform to the sampling accuracy requirements that the standard error of sampling for the pupil tests should be of a magnitude that is equal to, or smaller than, what would be achieved by The Specification of the Target Population employing a simple random sample of 400 pupils.
Response Rates: Each SACMEQ country should aim to achieve an overall response rate for pupils of 80 percent. This figure was based on the wish to achieve or exceed a response rate of 90 percent for schools and a response rate of 90 percent for pupils within schools.
Administrative and Financial Costs: The number of schools selected in each country should recognise limitations in the administrative and financial resources available for data collection.
Other Constraints: The number of learners selected to participate in the data collection in each selected school should be set at a level that will maximise validity of the within-school data collection for the learner reading and mathematics tests.
For Namibia, the desired target population was all learners enrolled in Grade 6 in the ninth month of the school year (i.e. in September 2000). The net enrolment ratio for the age group 7-13 years old who were enrolled in Grades 1 to 7 in Namibia in 2000 was 91.3 percent. However, in Namibia it was decided to exclude certain learners. These were learners in schools having fewer than 15 Grade 6 learners in them, learners in 'inaccessible schools, and learners in special schools. In all 884 learners from 82 schools were excluded but this only amounted to 1.8 percent of all learners. In Namibia there were 849 primary schools having 48,567 learners. After excluding the 1.8 percent of learners the defined population from which a sample had to be drawn consisted of 47,683 learners from 767 schools.
The number of schools required in the sample is in part a function of the intra-class correlation (rho) which is an indicator of the proportion of variation (in achievement in this case) among schools of total variation. The following is the formula often used for estimating the value of rho in situations where two-stage cluster sampling is employed using (approximately) equal sized clusters.
estimated rho = (b. s(a)square - (s)square) / (b - 1)(s)square
where s(a)square is the variance of cluster means, (s)square is the variance of the element values, and b is the cluster size. In SACMEQ I the rho had been 0.60 in Namibia. That is 60 percent of the variation was among schools and only 40 percent within schools. Therefore, in the case of Namibia a rho of 0.60 was used. This meant drawing a sample of 248 schools.
The major aim of the sampling was to have the equivalent of a simple random sample of 400 learners. In Namibia, this was 767 for reading achievement and 810 for mathematics. Hence the sample was a very good one for Namibia. For SACMEQ I it had been 335 which was below the required 400. This was because SACMEQ I was the first sample survey in Namibia and at that time it was assumed that the rho was 0.30. It was not. In SACMEQ II the rhos were 0.60 for reading and 0.53 for mathematics. Thus, in 2000 the variation among schools was slightly lower than in 1995.
Face-to-face [f2f]
The data collection for SACMEQ’s Initial Project took place in October 1995 and involved the administration of questionnaires to pupils, teachers, and school heads. The pupil questionnaire contained questions about the pupils’ home backgrounds and their school life; the teacher questionnaire asked about classrooms, teaching practices, working conditions, and teacher housing; and the school head questionnaire collected information about teachers, enrolments, buildings, facilities, and management. A reading literacy test was also given to the pupils. The test was based on items that were selected after a trial-testing programme had been completed.
Data entry and data cleaning A team of five persons from the University of Namibia Multi-Disciplinary Research Centre computer lab was appointed and trained in the use of WINDEM, a special data entry package to be used in SACMEQ. The numbers of keystrokes required to enter one copy of each data collection instrument were as follows: learner questionnaire: 150; learner reading test: 85; learner mathematics test: 65; teacher questionnaire: 587; teacher reading test: 51; teacher mathematics test: 43; school head questionnaire: 319; school form: 58; and learner name form: 51.
In the case of Namibia the total number of keystrokes was as follows: learner questionnaire: 762,600; learner reading test: 429,080; learner mathematics test: 328,250; teacher questionnaire: 358,657; teacher reading test: 15,504; teacher mathematics test: 14,061; school head questionnaire: 86,130; school form: 39,150; and learner name form: 259,284. That is, a total of 2,292,716 keystrokes were required to enter all of the data for Namibia.
An experienced keyboard operator can work at a rate of 25 keystrokes per minute (working from multi-paged questionnaires and stopping occasionally to clarify individual questionnaire entries with the supervisor). Assuming that this kind of work rate could be sustained for, say, around a maximum of six hours per day, then the whole data entry operation for Namibia was estimated to amount to around 255 person days of data entry work. This implied an estimated 10 weeks of work for the 5-person data entry team that operated in Namibia. However, the work was completed in 7 weeks because the data enterers worked extra hours.
At the end of this procedure the data files were sent by email to the unit 'Monitoring Educational Quality' at the IIEP in Paris. Many consistency checks were made for many variables as well as for the identification codes used. The IIEP team had many queries. The first data files were sent to Paris in May 2001 and after nine to-ings and fro-ings the files were finally declared to be clean on 25 January 2002.
Response rates for pupils and schools respectively were 91.8 percent and 100 percent. The reason for the shortfall in learner numbers was absenteeism by some learners in some of the schools on the day of data collection. However, sampling weights were used to correct for disproportionality among strata in the calculation
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
WirelessMATHBench-XL
Dataset Summary
WirelessMATHBench-XL is a collection of 4,027 graduate-level wireless communications math problems that pair long technical context passages with precise quantitative questions. Each item is derived from recent arXiv preprints in signal processing, networking, and edge intelligence, and is formatted to elicit deliberate step-by-step reasoning from large language models. Problems are tagged as multiple choice or fill-in-the-blank (with… See the full description on the dataset page: https://huggingface.co/datasets/XINLI1997/WirelessMATHBench-XL.
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
ObjectiveTo provide a practical guidance for the analysis of N-of-1 trials by comparing four commonly used models.MethodsThe four models, paired t-test, mixed effects model of difference, mixed effects model and meta-analysis of summary data were compared using a simulation study. The assumed 3-cycles and 4-cycles N-of-1 trials were set with sample sizes of 1, 3, 5, 10, 20 and 30 respectively under normally distributed assumption. The data were generated based on variance-covariance matrix under the assumption of (i) compound symmetry structure or first-order autoregressive structure, and (ii) no carryover effect or 20% carryover effect. Type I error, power, bias (mean error), and mean square error (MSE) of effect differences between two groups were used to evaluate the performance of the four models.ResultsThe results from the 3-cycles and 4-cycles N-of-1 trials were comparable with respect to type I error, power, bias and MSE. Paired t-test yielded type I error near to the nominal level, higher power, comparable bias and small MSE, whether there was carryover effect or not. Compared with paired t-test, mixed effects model produced similar size of type I error, smaller bias, but lower power and bigger MSE. Mixed effects model of difference and meta-analysis of summary data yielded type I error far from the nominal level, low power, and large bias and MSE irrespective of the presence or absence of carryover effect.ConclusionWe recommended paired t-test to be used for normally distributed data of N-of-1 trials because of its optimal statistical performance. In the presence of carryover effects, mixed effects model could be used as an alternative.
Not seeing a result you expected?
Learn how you can add new datasets to our index.
Facebook
Twitterhttps://creativecommons.org/publicdomain/zero/1.0/https://creativecommons.org/publicdomain/zero/1.0/
This is the large data set as featured in the OCR H240 exam series.
Questions about this dataset will be featured in the statistics paper
The LDS is a .xlsx file containing 5 tables, four data, one information. The data is drawn from the UK censuses from the years 2001 and 2011. It is designed for you to make comparisons and analyses of the changes in demographic and behavioural features of the populace. There is the age structure of each local authority and the method of travel within each local authority.