The OECD Programme for International Student Assessment (PISA) surveys collected data on students’ performances in reading, mathematics and science, as well as contextual information on students’ background, home characteristics and school factors which could influence performance. This publication includes detailed information on how to analyse the PISA data, enabling researchers to both reproduce the initial results and to undertake further analyses. In addition to the inclusion of the necessary techniques, the manual also includes a detailed account of the PISA 2006 database and worked examples providing full syntax in SPSS.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
The survey dataset for identifying Shiraz old silo’s new use which includes four components: 1. The survey instrument used to collect the data “SurveyInstrument_table.pdf”. The survey instrument contains 18 main closed-ended questions in a table format. Two of these, concern information on Silo’s decision-makers and proposed new use followed up after a short introduction of the questionnaire, and others 16 (each can identify 3 variables) are related to the level of appropriate opinions for ideal intervention in Façade, Openings, Materials and Floor heights of the building in four values: Feasibility, Reversibility, Compatibility and Social Benefits. 2. The raw survey data “SurveyData.rar”. This file contains an Excel.xlsx and a SPSS.sav file. The survey data file contains 50 variables (12 for each of the four values separated by colour) and data from each of the 632 respondents. Answering each question in the survey was mandatory, therefor there are no blanks or non-responses in the dataset. In the .sav file, all variables were assigned with numeric type and nominal measurement level. More details about each variable can be found in the Variable View tab of this file. Additional variables were created by grouping or consolidating categories within each survey question for simpler analysis. These variables are listed in the last columns of the .xlsx file. 3. The analysed survey data “AnalysedData.rar”. This file contains 6 “SPSS Statistics Output Documents” which demonstrate statistical tests and analysis such as mean, correlation, automatic linear regression, reliability, frequencies, and descriptives. 4. The codebook “Codebook.rar”. The detailed SPSS “Codebook.pdf” alongside the simplified codebook as “VariableInformation_table.pdf” provides a comprehensive guide to all 50 variables in the survey data, including numerical codes for survey questions and response options. They serve as valuable resources for understanding the dataset, presenting dictionary information, and providing descriptive statistics, such as counts and percentages for categorical variables.
The data were collected between 2014 and 2017 and represent interviews of public lands users as they exited the location where they were recreating. Because the data represents human subjects, the data is presented as an aggregate of the original data to protect the participants. Data were aggregated at the party level and specific information about participants removed, as such the analyses represented in the manuscript may not always be represented by this data file as some analyses were conducted at the individual level, not party level, and/or included additional information that was not included here due to concerns over violating human subjects research norms. All analyses were conducted using IBM SPSS Statistics version 25. All data were collected under the approval of the University of Nebraska Institutional Review Board, approval 20120912892EX. For further information about the data or analysis please contact the lead author.
Descriptions of the variables included in the dat...
Some surveys contain multiple units of observation, while others come in many parts. This workshop will give participants hands-on experience using both types of files. The General Social Survey, Cycle 8 and the Canadian Travel Surveys will be used as examples. (Note: Data associated with this presentation is available on the DLI FTP site under folder 1873-216.)
Database of the nation''s substance abuse and mental health research data providing public use data files, file documentation, and access to restricted-use data files to support a better understanding of this critical area of public health. The goal is to increase the use of the data to most accurately understand and assess substance abuse and mental health problems and the impact of related treatment systems. The data include the U.S. general and special populations, annual series, and designs that produce nationally representative estimates. Some of the data acquired and archived have never before been publicly distributed. Each collection includes survey instruments (when provided), a bibliography of related literature, and related Web site links. All data may be downloaded free of charge in SPSS, SAS, STATA, and ASCII formats and most studies are available for use with the online data analysis system. This system allows users to conduct analyses ranging from cross-tabulation to regression without downloading data or relying on other software. Another feature, Quick Tables, provides the ability to select variables from drop down menus to produce cross-tabulations and graphs that may be customized and cut and pasted into documents. Documentation files, such as codebooks and questionnaires, can be downloaded and viewed online.
https://spdx.org/licenses/CC0-1.0.htmlhttps://spdx.org/licenses/CC0-1.0.html
With the rapid development of AIGC (Artificial Intelligence Generated Content) and its expanding role and scope in education and teaching. This study conducted a survey among 394 pre-service STEM teachers enrolled at a university located in Zhejiang Province. Data were collected and a structural model was constructed to examine interplay among psychological stress, anxiety self-efficacy, and learning burnout resulting from the utilization of AIGC. The findings indicate that pre-service STEM teachers may experience psychological stress when applying AIGC, which could exacerbate their anxiety towards artificial intelligence and potentially lead to academic burnout. In order to effectively integrate AIGC in the field of education and enhance the professional development of pre-service teachers, the key lies in the dissemination of artificial intelligence knowledge, enhancing pre-service teachers' understanding of artificial intelligence, and encouraging them to appropriately utilize AIGC as a learning auxiliary tool. Methods This study uses the internationally recognized five-point Likert scale as the main tool to quantitatively assess STEM teachers' psychological pressure, fear of artificial intelligence, self-efficacy and learning fatigue caused by the use of AIGC. The questionnaire design is based on proven and valid scales in published academic literature at home and abroad to ensure the reliability and validity of data collection. After collecting data using the questionnaire method, invalid data were eliminated and SPSS 29.0 was used for statistical analysis.
analyze the current population survey (cps) annual social and economic supplement (asec) with r the annual march cps-asec has been supplying the statistics for the census bureau's report on income, poverty, and health insurance coverage since 1948. wow. the us census bureau and the bureau of labor statistics ( bls) tag-team on this one. until the american community survey (acs) hit the scene in the early aughts (2000s), the current population survey had the largest sample size of all the annual general demographic data sets outside of the decennial census - about two hundred thousand respondents. this provides enough sample to conduct state- and a few large metro area-level analyses. your sample size will vanish if you start investigating subgroups b y state - consider pooling multiple years. county-level is a no-no. despite the american community survey's larger size, the cps-asec contains many more variables related to employment, sources of income, and insurance - and can be trended back to harry truman's presidency. aside from questions specifically asked about an annual experience (like income), many of the questions in this march data set should be t reated as point-in-time statistics. cps-asec generalizes to the united states non-institutional, non-active duty military population. the national bureau of economic research (nber) provides sas, spss, and stata importation scripts to create a rectangular file (rectangular data means only person-level records; household- and family-level information gets attached to each person). to import these files into r, the parse.SAScii function uses nber's sas code to determine how to import the fixed-width file, then RSQLite to put everything into a schnazzy database. you can try reading through the nber march 2012 sas importation code yourself, but it's a bit of a proc freak show. this new github repository contains three scripts: 2005-2012 asec - download all microdata.R down load the fixed-width file containing household, family, and person records import by separating this file into three tables, then merge 'em together at the person-level download the fixed-width file containing the person-level replicate weights merge the rectangular person-level file with the replicate weights, then store it in a sql database create a new variable - one - in the data table 2012 asec - analysis examples.R connect to the sql database created by the 'download all microdata' progr am create the complex sample survey object, using the replicate weights perform a boatload of analysis examples replicate census estimates - 2011.R connect to the sql database created by the 'download all microdata' program create the complex sample survey object, using the replicate weights match the sas output shown in the png file below 2011 asec replicate weight sas output.png statistic and standard error generated from the replicate-weighted example sas script contained in this census-provided person replicate weights usage instructions document. click here to view these three scripts for more detail about the current population survey - annual social and economic supplement (cps-asec), visit: the census bureau's current population survey page the bureau of labor statistics' current population survey page the current population survey's wikipedia article notes: interviews are conducted in march about experiences during the previous year. the file labeled 2012 includes information (income, work experience, health insurance) pertaining to 2011. when you use the current populat ion survey to talk about america, subract a year from the data file name. as of the 2010 file (the interview focusing on america during 2009), the cps-asec contains exciting new medical out-of-pocket spending variables most useful for supplemental (medical spending-adjusted) poverty research. confidential to sas, spss, stata, sudaan users: why are you still rubbing two sticks together after we've invented the butane lighter? time to transition to r. :D
This dataset consists of three data folders including all related documents of the online survey conducted within the NESP 3.2.3 project (Tropical Water Quality Hub) and a survey format document representing how the survey was designed. Apart from participants’ demographic information, the survey consists of three sections: conjoint analysis, picture rating and open question. Correspondent outcome of these three sections are downloaded from Qualtrics website and used for three different data analysis processes.
Related data to the first section “conjoint analysis” is saved in the Conjoint analysis folder which contains two sub-folders. The first one includes a plan file of SAV. Format representing the design suggestion by SPSS orthogonal analysis for testing beauty factors and 9 photoshoped pictures used in the survey. The second (i.e. Final results) contains 1 SAV. file named “data1” which is the imported results of conjoint analysis section in SPSS, 1 SPS. file named “Syntax1” representing the code used to run conjoint analysis, 2 SAV. files as the output of conjoint analysis by SPSS, and 1 SPV file named “Final output” showing results of further data analysis by SPSS on the basis of utility and importance data.
Related data to the second section “Picture rating” is saved into Picture rating folder including two subfolders. One subfolder contains 2500 pictures of Great Barrier Reef used in the rating survey section. These pictures are organised by named and stored in two folders named as “Survey Part 1” and “Survey Part 2” which are correspondent with two parts of the rating survey sections. The other subfolder “Rating results” consist of one XLSX. file representing survey results downloaded from Qualtric website.
Finally, related data to the open question is saved in “Open question” folder. It contains one csv. file and one PDF. file recording participants’ answers to the open question as well as one PNG. file representing a screenshot of Leximancer analysis outcome.
Methods: This dataset resulted from the input and output of an online survey regarding how people assess the beauty of Great Barrier Reef. This survey was designed for multiple purposes including three main sections: (1) conjoint analysis (ranking 9 photoshopped pictures to determine the relative importance weights of beauty attributes), (2) picture rating (2500 pictures to be rated) and (3) open question on the factors that makes a picture of the Great Barrier Reef beautiful in participants’ opinion (determining beauty factors from tourist perspective). Pictures used in this survey were downloaded from public sources such as websites of the Tourism and Events Queensland and Tropical Tourism North Queensland as well as tourist sharing sources (i.e. Flickr). Flickr pictures were downloaded using the key words “Great Barrier Reef”. About 10,000 pictures were downloaded in August and September 2017. 2,500 pictures were then selected based on several research criteria: (1) underwater pictures of GBR, (2) without humans, (3) viewed from 1-2 metres from objects and (4) of high resolution.
The survey was created on Qualtrics website and launched on 4th October 2017 using Qualtrics survey service. Each participant rated 50 pictures randomly selected from the pool of 2500 survey pictures. 772 survey completions were recorded and 705 questionnaires were eligible for data analysis after filtering unqualified questionnaires. Conjoint analysis data was imported to IBM SPSS using SAV. format and the output was saved using SPV. format. Automatic aesthetic rating of 2500 Great Barrier Reef pictures –all these pictures are rated (1 – 10 scale) by at least 10 participants and this dataset was saved in a XLSX. file which is used to train and test an Artificial Intelligence (AI)-based system recognising and assessing the beauty of natural scenes. Answers of the open-question were saved in a XLSX. file and a PDF. file to be employed for theme analysis by Leximancer software.
Further information can be found in the following publication: Becken, S., Connolly R., Stantic B., Scott N., Mandal R., Le D., (2018), Monitoring aesthetic value of the Great Barrier Reef by using innovative technologies and artificial intelligence, Griffith Institute for Tourism Research Report No 15.
Format: The Online survey dataset includes one PDF file representing the survey format with all sections and questions. It also contains three subfolders, each has multiple files. The subfolder of Conjoint analysis contains an image of the 9 JPG. Pictures, 1 SAV. format file for the Orthoplan subroutine outcome and 5 outcome documents (i.e. 3 SAV. files, 1 SPS. file, 1 SPV. file). The subfolder of Picture rating contains a capture of the 2500 pictures used in the survey, 1 excel file for rating results. The subfolder of Open question includes 1 CSV. file, 1 PDF. file representing participants’ answers and one PNG. file for the analysis outcome.
Data Dictionary:
Card 1: Picture design option number 1 suggested by SPSS orthogonal analysis. Importance value: The relative importance weight of each beauty attribute calculated by SPSS conjoint analysis. Utility: Score reflecting influential valence and degree of each beauty attribute on beauty score. Syntax: Code used to run conjoint analysis by SPSS Leximancer: Specialised software for qualitative data analysis. Concept map: A map showing the relationship between concepts identified Q1_1: Beauty score of the picture Q1_1 by the correspondent participant (i.e. survey part 1) Q2.1_1: Beauty score of the picture Q2.1_1 by the correspondent participant (i.e. survey part 2) Conjoint _1: Ranking of the picture 1 designed for conjoint analysis by the correspondent participant
References: Becken, S., Connolly R., Stantic B., Scott N., Mandal R., Le D., (2018), Monitoring aesthetic value of the Great Barrier Reef by using innovative technologies and artificial intelligence, Griffith Institute for Tourism Research Report No 15.
Data Location:
This dataset is filed in the eAtlas enduring data repository at: data esp3\3.2.3_Aesthetic-value-GBR
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
The GAPs Data Repository provides a comprehensive overview of available qualitative and quantitative data on national return regimes, now accessible through an advanced web interface at https://data.returnmigration.eu/.
This updated guideline outlines the complete process, starting from the initial data collection for the return migration data repository to the development of a comprehensive web-based platform. Through iterative development, participatory approaches, and rigorous quality checks, we have ensured a systematic representation of return migration data at both national and comparative levels.
The Repository organizes data into five main categories, covering diverse aspects and offering a holistic view of return regimes: country profiles, legislation, infrastructure, international cooperation, and descriptive statistics. These categories, further divided into subcategories, are based on insights from a literature review, existing datasets, and empirical data collection from 14 countries. The selection of categories prioritizes relevance for understanding return and readmission policies and practices, data accessibility, reliability, clarity, and comparability. Raw data is meticulously collected by the national experts.
The transition to a web-based interface builds upon the Repository’s original structure, which was initially developed using REDCap (Research Electronic Data Capture). It is a secure web application for building and managing online surveys and databases.The REDCAP ensures systematic data entries and store them on Uppsala University’s servers while significantly improving accessibility and usability as well as data security. It also enables users to export any or all data from the Project when granted full data export privileges. Data can be exported in various ways and formats, including Microsoft Excel, SAS, Stata, R, or SPSS for analysis. At this stage, the Data Repository design team also converted tailored records of available data into public reports accessible to anyone with a unique URL, without the need to log in to REDCap or obtain permission to access the GAPs Project Data Repository. Public reports can be used to share information with stakeholders or external partners without granting them access to the Project or requiring them to set up a personal account. Currently, all public report links inserted in this report are also available on the Repository’s webpage, allowing users to export original data.
This report also includes a detailed codebook to help users understand the structure, variables, and methodologies used in data collection and organization. This addition ensures transparency and provides a comprehensive framework for researchers and practitioners to effectively interpret the data.
The GAPs Data Repository is committed to providing accessible, well-organized, and reliable data by moving to a centralized web platform and incorporating advanced visuals. This Repository aims to contribute inputs for research, policy analysis, and evidence-based decision-making in the return and readmission field.
Explore the GAPs Data Repository at https://data.returnmigration.eu/.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
ABSTRACT
The Albero study analyzes the personal transitions of a cohort of high school students at the end of their studies. The data consist of (a) the longitudinal social network of the students, before (n = 69) and after (n = 57) finishing their studies; and (b) the longitudinal study of the personal networks of each of the participants in the research. The two observations of the complete social network are presented in two matrices in Excel format. For each respondent, two square matrices of 45 alters of their personal networks are provided, also in Excel format. For each respondent, both psychological sense of community and frequency of commuting is provided in a SAV file (SPSS). The database allows the combined analysis of social networks and personal networks of the same set of individuals.
INTRODUCTION
Ecological transitions are key moments in the life of an individual that occur as a result of a change of role or context. This is the case, for example, of the completion of high school studies, when young people start their university studies or try to enter the labor market. These transitions are turning points that carry a risk or an opportunity (Seidman & French, 2004). That is why they have received special attention in research and psychological practice, both from a developmental point of view and in the situational analysis of stress or in the implementation of preventive strategies.
The data we present in this article describe the ecological transition of a group of young people from Alcala de Guadaira, a town located about 16 kilometers from Seville. Specifically, in the “Albero” study we monitored the transition of a cohort of secondary school students at the end of the last pre-university academic year. It is a turning point in which most of them began a metropolitan lifestyle, with more displacements to the capital and a slight decrease in identification with the place of residence (Maya-Jariego, Holgado & Lubbers, 2018).
Normative transitions, such as the completion of studies, affect a group of individuals simultaneously, so they can be analyzed both individually and collectively. From an individual point of view, each student stops attending the institute, which is replaced by new interaction contexts. Consequently, the structure and composition of their personal networks are transformed. From a collective point of view, the network of friendships of the cohort of high school students enters into a gradual process of disintegration and fragmentation into subgroups (Maya-Jariego, Lubbers & Molina, 2019).
These two levels, individual and collective, were evaluated in the “Albero” study. One of the peculiarities of this database is that we combine the analysis of a complete social network with a survey of personal networks in the same set of individuals, with a longitudinal design before and after finishing high school. This allows combining the study of the multiple contexts in which each individual participates, assessed through the analysis of a sample of personal networks (Maya-Jariego, 2018), with the in-depth analysis of a specific context (the relationships between a promotion of students in the institute), through the analysis of the complete network of interactions. This potentially allows us to examine the covariation of the social network with the individual differences in the structure of personal networks.
PARTICIPANTS
The social network and personal networks of the students of the last two years of high school of an institute of Alcala de Guadaira (Seville) were analyzed. The longitudinal follow-up covered approximately a year and a half. The first wave was composed of 31 men (44.9%) and 38 women (55.1%) who live in Alcala de Guadaira, and who mostly expect to live in Alcala (36.2%) or in Seville (37.7%) in the future. In the second wave, information was obtained from 27 men (47.4%) and 30 women (52.6%).
DATE STRUCTURE AND ARCHIVES FORMAT
The data is organized in two longitudinal observations, with information on the complete social network of the cohort of students of the last year, the personal networks of each individual and complementary information on the sense of community and frequency of metropolitan movements, among other variables.
Social network
The file “Red_Social_t1.xlsx” is a valued matrix of 69 actors that gathers the relations of knowledge and friendship between the cohort of students of the last year of high school in the first observation. The file “Red_Social_t2.xlsx” is a valued matrix of 57 actors obtained 17 months after the first observation.
The data is organized in two longitudinal observations, with information on the complete social network of the cohort of students of the last year, the personal networks of each individual and complementary information on the sense of community and frequency of metropolitan movements, among other variables.
In order to generate each complete social network, the list of 77 students enrolled in the last year of high school was passed to the respondents, asking that in each case they indicate the type of relationship, according to the following values: 1, “his/her name sounds familiar"; 2, "I know him/her"; 3, "we talk from time to time"; 4, "we have good relationship"; and 5, "we are friends." The two resulting complete networks are represented in Figure 2. In the second observation, it is a comparatively less dense network, reflecting the gradual disintegration process that the student group has initiated.
Personal networks
Also in this case the information is organized in two observations. The compressed file “Redes_Personales_t1.csv” includes 69 folders, corresponding to personal networks. Each folder includes a valued matrix of 45 alters in CSV format. Likewise, in each case a graphic representation of the network obtained with Visone (Brandes and Wagner, 2004) is included. Relationship values range from 0 (do not know each other) to 2 (know each other very well).
Second, the compressed file “Redes_Personales_t2.csv” includes 57 folders, with the information equivalent to each respondent referred to the second observation, that is, 17 months after the first interview. The structure of the data is the same as in the first observation.
Sense of community and metropolitan displacements
The SPSS file “Albero.sav” collects the survey data, together with some information-summary of the network data related to each respondent. The 69 rows correspond to the 69 individuals interviewed, and the 118 columns to the variables related to each of them in T1 and T2, according to the following list:
• Socio-economic data.
• Data on habitual residence.
• Information on intercity journeys.
• Identity and sense of community.
• Personal network indicators.
• Social network indicators.
DATA ACCESS
Social networks and personal networks are available in CSV format. This allows its use directly with UCINET, Visone, Pajek or Gephi, among others, and they can be exported as Excel or text format files, to be used with other programs.
The visual representation of the personal networks of the respondents in both waves is available in the following album of the Graphic Gallery of Personal Networks on Flickr: .
In previous work we analyzed the effects of personal networks on the longitudinal evolution of the socio-centric network. It also includes additional details about the instruments applied. In case of using the data, please quote the following reference:
Maya-Jariego, I., Holgado, D. & Lubbers, M. J. (2018). Efectos de la estructura de las redes personales en la red sociocéntrica de una cohorte de estudiantes en transición de la enseñanza secundaria a la universidad. Universitas Psychologica, 17(1), 86-98. https://doi.org/10.11144/Javeriana.upsy17-1.eerp
The English version of this article can be downloaded from: https://tinyurl.com/yy9s2byl
CONCLUSION
The database of the “Albero” study allows us to explore the co-evolution of social networks and personal networks. In this way, we can examine the mutual dependence of individual trajectories and the structure of the relationships of the cohort of students as a whole. The complete social network corresponds to the same context of interaction: the secondary school. However, personal networks collect information from the different contexts in which the individual participates. The structural properties of personal networks may partly explain individual differences in the position of each student in the entire social network. In turn, the properties of the entire social network partly determine the structure of opportunities in which individual trajectories are displayed.
The longitudinal character and the combination of the personal networks of individuals with a common complete social network, make this database have unique characteristics. It may be of interest both for multi-level analysis and for the study of individual differences.
ACKNOWLEDGEMENTS
The fieldwork for this study was supported by the Complementary Actions of the Ministry of Education and Science (SEJ2005-25683), and was part of the project “Dynamics of actors and networks across levels: individuals, groups, organizations and social settings” (2006 -2009) of the European Science Foundation (ESF). The data was presented for the first time on June 30, 2009, at the European Research Collaborative Project Meeting on Dynamic Analysis of Networks and Behaviors, held at the Nuffield College of the University of Oxford.
REFERENCES
Brandes, U., & Wagner, D. (2004). Visone - Analysis and Visualization of Social Networks. In M. Jünger, & P. Mutzel (Eds.), Graph Drawing Software (pp. 321-340). New York: Springer-Verlag.
Maya-Jariego, I. (2018). Why name generators with a fixed number of alters may be a pragmatic option for personal network analysis. American Journal of
CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
License information was derived automatically
analyze the health and retirement study (hrs) with r the hrs is the one and only longitudinal survey of american seniors. with a panel starting its third decade, the current pool of respondents includes older folks who have been interviewed every two years as far back as 1992. unlike cross-sectional or shorter panel surveys, respondents keep responding until, well, death d o us part. paid for by the national institute on aging and administered by the university of michigan's institute for social research, if you apply for an interviewer job with them, i hope you like werther's original. figuring out how to analyze this data set might trigger your fight-or-flight synapses if you just start clicking arou nd on michigan's website. instead, read pages numbered 10-17 (pdf pages 12-19) of this introduction pdf and don't touch the data until you understand figure a-3 on that last page. if you start enjoying yourself, here's the whole book. after that, it's time to register for access to the (free) data. keep your username and password handy, you'll need it for the top of the download automation r script. next, look at this data flowchart to get an idea of why the data download page is such a righteous jungle. but wait, good news: umich recently farmed out its data management to the rand corporation, who promptly constructed a giant consolidated file with one record per respondent across the whole panel. oh so beautiful. the rand hrs files make much of the older data and syntax examples obsolete, so when you come across stuff like instructions on how to merge years, you can happily ignore them - rand has done it for you. the health and retirement study only includes noninstitutionalized adults when new respondents get added to the panel (as they were in 1992, 1993, 1998, 2004, and 2010) but once they're in, they're in - respondents have a weight of zero for interview waves when they were nursing home residents; but they're still responding and will continue to contribute to your statistics so long as you're generalizing about a population from a previous wave (for example: it's possible to compute "among all americans who were 50+ years old in 1998, x% lived in nursing homes by 2010"). my source for that 411? page 13 of the design doc. wicked. this new github repository contains five scripts: 1992 - 2010 download HRS microdata.R loop through every year and every file, download, then unzip everything in one big party impor t longitudinal RAND contributed files.R create a SQLite database (.db) on the local disk load the rand, rand-cams, and both rand-family files into the database (.db) in chunks (to prevent overloading ram) longitudinal RAND - analysis examples.R connect to the sql database created by the 'import longitudinal RAND contributed files' program create tw o database-backed complex sample survey object, using a taylor-series linearization design perform a mountain of analysis examples with wave weights from two different points in the panel import example HRS file.R load a fixed-width file using only the sas importation script directly into ram with < a href="http://blog.revolutionanalytics.com/2012/07/importing-public-data-with-sas-instructions-into-r.html">SAScii parse through the IF block at the bottom of the sas importation script, blank out a number of variables save the file as an R data file (.rda) for fast loading later replicate 2002 regression.R connect to the sql database created by the 'import longitudinal RAND contributed files' program create a database-backed complex sample survey object, using a taylor-series linearization design exactly match the final regression shown in this document provided by analysts at RAND as an update of the regression on pdf page B76 of this document . click here to view these five scripts for more detail about the health and retirement study (hrs), visit: michigan's hrs homepage rand's hrs homepage the hrs wikipedia page a running list of publications using hrs notes: exemplary work making it this far. as a reward, here's the detailed codebook for the main rand hrs file. note that rand also creates 'flat files' for every survey wave, but really, most every analysis you c an think of is possible using just the four files imported with the rand importation script above. if you must work with the non-rand files, there's an example of how to import a single hrs (umich-created) file, but if you wish to import more than one, you'll have to write some for loops yourself. confidential to sas, spss, stata, and sudaan users: a tidal wave is coming. you can get water up your nose and be dragged out to sea, or you can grab a surf board. time to transition to r. :D
Quantitative data from community observations are stored and managed using SPSS social survey software. The sampling unit used is a harvest event, typically a hunting or fishing event in a particular season. As of 5 September, 2008 we have received and encoded data for 56 harvest events as follows: Harvest type: Mammal (10), Fish (45), Shellfish (1) Community: Gambell (10), Kanchalan (22), Nikolskoye (6), Sandpoint (18) Preliminary SPSS Data structure: Name, Label, Type, Width ID Respondent s Identification Number String 10 INTERNO Interview Number String 2 DATE Date On Which the Interview Took Place Date 8 SEX Gender Numeric 1 YEARBO Year of Birth Numeric 11 VILLAGE Village Where Respndent Resides String 6 LOCATI Respondent Resides in Russia or Alaska Numeric 8 LIVED How Long Respondent Lived in the Area String 100 LANGUAG Language in Which Interiew Conducted Numeric 7 HARVEST Level of Harvester Numeric 4 YEARHU How Many Years Respondent Has Hunted/Fished in the Area Numeric 8 EMPLOY Is the Respondent Employed in a Non-Harvesting Field Numeric 3 TIMEWOR Time Per Week/Month Is Spent in Non-Harvest Work Numeric 8 YEARWOR How Many Years Spent in Non-Harvest Work CATEGORIES Numeric 8 Q1FISHM Is Respondent Hunting Fish or Mammals On Next Trip Numeric 4 SPECIES Species of Fish/Mammal Being Hunted/Fished Numeric 8 Q2RECA Does Respondent Recall When Last Hunt/Fish Trip Occurre Numeric 3 Q2WHEN Date of Last Hunt/Fish Trip String 50 Q2AAGO How Long Ago Was Last Hunt/Fish Trip Numeric 16 Q3FAR How Far Respondent Travelled On Last Hunt/Fish Trip Numeric Q4OFTEN How Often Respondent Hunted/Fished in the Location of Last Trip Numeric 6 Q5AGE Age When Respondent First Went to Location of Last Trip Numeric 18 Q6PROX Prefers Loc. of Last Trip Due to Proximity to Village Numeric 11 Q6ACCES Prefers Location of Last Trip Due to Ease of Access Numeric 11 Q6CATCH Prefers Location of Last Trip Due to Ease of Catching Numeric 11 Q6OTHER Prefers Location of Last Trip Due to Some Other Reason Numeric 11 Q6SPECI Other Reason Prefers Locatin of Last Trip String 200 Q6DONT Respondent Does Not Like Location of Last Trip Numeric 11 Q7RELY Is Location of Last Trip Reliable for Fishing/Hunting Numeric 3 Q8NOTIC In Previous 5-10 Years Has Respondent Noticed Changes at Last Hunt/Fish Location Numeric 3 Q9OTHER Do Others From the Village Also Hunt/Fish at Location of Last Trip Numeric 3 Q10GETA On Last Trip, Was it Easier or More Difficult to Get to Location Numeric 3 Q10GETR On Last Trip Did Respondent Encounter Difficulties Getting to Hunt/Fish Location Numeric 8 Q10ATRA More Difficult to Get to Location of Last Trip Due to Lack of Transportation Numeric 11 Q10AROA More Difficult to Get to Location of Last Trip Due to Poor Road Conditions Numeric 11 Q10AENV More Difficult to Get to Location of Last Trip Due to Poor Environ Conditions Numeric 11 Q10AECO More Diff. to Get to Location of Last Trip Due to Economics Numeric 11 Q10AHEA More Difficult to Get to Location of Last Trip Due to Personal Health Condition Numeric 11 Q10AOTHE More Difficult to Get to Location of Last Trip Due to Other Reasons Numeric 23 Q11TRAD Last Harvest Used for Traditional/Personal Use Numeric 11 Q11CASH Last Harvest Used for Generating Cash or Bartering Numeric 11 Q11REC Last Harvest Used for Recreational Hunting/Fishing Numeric 11 Q11COM Last Harvest Used for Commercial or Business Activity Numeric 11 Q11DOG Last Harvest Used for Feeding Dogs Numeric 11 Q11SHAR Last Harvest Used for Sharing with Friends/Family Numeric 11 Q11OTHE Last Harvest Used for Something Else Numeric 20 Q12QUAN Quantity of XXX Caught on Last Hunt/Fish Trip Numeric 21
Last year we were stretched with complex surveys using the Canadian Travel Survey. This year our data exercise will focus on shaping data from another interesting survey - the Canadian Tobacco Use Monitoring Survey (CTUMS). After last year's exercise using finger puppets, you'll find fitness for more than just your fingers! The objective of this computing exercise is to explore the different types of variables and their functions using the CTUMS 2001 for context. The vocabularies of data and analysis use different labels for the various functions that variables perform. In this exercise, you will identify some of these differences in CTUMS. (Note: Data associated with this presentation is available on the DLI FTP site under folder 1873-203.)
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Data in SPSS formatMeasured language variables across the cultural groups, in SPSS data file format.Data.savData in CSV formatEquivalent data to the SPSS upload, in CSV format.Data.csvAnalysis syntax for SPSSSyntax used to generate the reported results using SPSS.Syntax.sps
https://qdr.syr.edu/policies/qdr-standard-access-conditionshttps://qdr.syr.edu/policies/qdr-standard-access-conditions
This is an Annotation for Transparent Inquiry (ATI) data project. The annotated article can be viewed on the Publisher's Website. Data Generation The research project engages a story about perceptions of fairness in criminal justice decisions. The specific focus involves a debate between ProPublica, a news organization, and Northpointe, the owner of a popular risk tool called COMPAS. ProPublica wrote that COMPAS was racist against blacks, while Northpointe posted online a reply rejecting such a finding. These two documents were the obvious foci of the qualitative analysis because of the further media attention they attracted, the confusion their competing conclusions caused readers, and the power both companies wield in public circles. There were no barriers to retrieval as both documents have been publicly available on their corporate websites. This public access was one of the motivators for choosing them as it meant that they were also easily attainable by the general public, thus extending the documents’ reach and impact. Additional materials from ProPublica relating to the main debate were also freely downloadable from its website and a third party, open source platform. Access to secondary source materials comprising additional writings from Northpointe representatives that could assist in understanding Northpointe’s main document, though, was more limited. Because of a claim of trade secrets on its tool and the underlying algorithm, it was more difficult to reach Northpointe’s other reports. Nonetheless, largely because its clients are governmental bodies with transparency and accountability obligations, some of Northpointe-associated reports were retrievable from third parties who had obtained them, largely through Freedom of Information Act queries. Together, the primary and (retrievable) secondary sources allowed for a triangulation of themes, arguments, and conclusions. The quantitative component uses a dataset of over 7,000 individuals with information that was collected and compiled by ProPublica and made available to the public on github. ProPublica’s gathering the data directly from criminal justice officials via Freedom of Information Act requests rendered the dataset in the public domain, and thus no confidentiality issues are present. The dataset was loaded into SPSS v. 25 for data analysis. Data Analysis The qualitative enquiry used critical discourse analysis, which investigates ways in which parties in their communications attempt to create, legitimate, rationalize, and control mutual understandings of important issues. Each of the two main discourse documents was parsed on its own merit. Yet the project was also intertextual in studying how the discourses correspond with each other and to other relevant writings by the same authors. Several more specific types of discursive strategies were of interest in attracting further critical examination: Testing claims and rationalizations that appear to serve the speaker’s self-interest Examining conclusions and determining whether sufficient evidence supported them Revealing contradictions and/or inconsistencies within the same text and intertextually Assessing strategies underlying justifications and rationalizations used to promote a party’s assertions and arguments Noticing strategic deployment of lexical phrasings, syntax, and rhetoric Judging sincerity of voice and the objective consideration of alternative perspectives Of equal importance in a critical discourse analysis is consideration of what is not addressed, that is to uncover facts and/or topics missing from the communication. For this project, this included parsing issues that were either briefly mentioned and then neglected, asserted yet the significance left unstated, or not suggested at all. This task required understanding common practices in the algorithmic data science literature. The paper could have been completed with just the critical discourse analysis. However, because one of the salient findings from it highlighted that the discourses overlooked numerous definitions of algorithmic fairness, the call to fill this gap seemed obvious. Then, the availability of the same dataset used by the parties in conflict, made this opportunity more appealing. Calculating additional algorithmic equity equations would not thereby be troubled by irregularities because of diverse sample sets. New variables were created as relevant to calculate algorithmic fairness equations. In addition to using various SPSS Analyze functions (e.g., regression, crosstabs, means), online statistical calculators were useful to compute z-test comparisons of proportions and t-test comparisons of means. Logic of Annotation Annotations were employed to fulfil a variety of functions, including supplementing the main text with context, observations, counter-points, analysis, and source attributions. These fall under a few categories. Space considerations. Critical discourse analysis offers a rich method...
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
The Cohort Hip & Cohort Knee (CHECK) is a population-based observational multicenter cohort study of 1002 individuals with early symptomatic osteoarthritis (OA) of knee and/or hip in the Netherlands. The participants were followed for 10 years. The study evaluated clinical, radiographic and biochemical variables in order to establish the course, prognosis and underlying mechanisms of early symptomatic osteoarthritis. The Dutch Artritis Foundation initiated and funded this inception cohort.This dataset covers the data collection of baseline (T0) without the variable 'Subject identification number'. Included is a Kellgren-Lawrence radiographic classification covering T0,T2,T5, T8 and T10. Also X-rays of hips and knees of baseline are available. More information on the variables can be found in the documentation. In the description file you can find an overview of the data belonging to this dataset and more information about the format and kind of view of the X rays.The complete data are available via three separate datasets, each containing again the baseline T0 data of this current dataset. All SPSS data files of these three datasets include the variable 'Subject identification number'.If you wish to make use of the complete CHECK data, please see the see relations for the other CHECK datasets and for the overview 'Thematic collection: CHECK (Cohort Hip & Cohort Knee)'. Date Submitted: 2015-12-09 2019-12-20: a new data file on X-Ray data 'Rontgen_opT10_20191118' was added to the dataset.2017-09-19: A data file on X-Ray ratings has been added and the variable guide is replaced by a new version (6) with information on this data file. Please note the variable names start with 'RontgT10_' in the data file.2017-07-12: Due to an error a data file has been replaced.CHECK_T0_DANS_nsinENG_20151211.sav is now replaced by CHECK_T0_DANS_nsin_ENG_20161128.sav---The informed consent statements of the participants are stored at the participating hospitals.The .dta (STATA) and .por (SPSS) files are conversions of the original .sav (SPSS) files.
Misophonia is a condition characterized by negative affect, intolerance, and functional impairment in response to particular repetitive sounds usually made by others (e.g., chewing, sniffing, pen tapping) and associated stimuli. To date, researchers have largely studied misophonia using self-report measures. As the field is quickly expanding, assessment approaches need to advance to include more objective measures capable of differentiating those with and without misophonia. Although several studies have used sounds as experimental stimuli, few have used standardized stimuli sets with demonstrated reliability or validity. To conduct rigorous research to better understand misophonia, it is important to have an easily accessible, standardized set of acoustic stimuli for use across studies. Accordingly, in the present study, the International Affective Digitized Sounds (IADS-2), developed by Bradley and Lang [1], were used to determine whether participants with misophonia responded to cert..., Group differences in sound ratings were examined using a two-way, mixed analysis of covariance (2 groups x 3 sound types, where “group†corresponds to participants with misophonia or controls, and “sound type†refers to positive, negative, or neutral sounds) on four dependent variables (ratings of valence, arousal, similarity, and avoidance). When statistically significant interactions were observed for sound type, pairwise comparisons were used to determine group differences on each dependent variable, as well as mean differences between sound type on each dependent variable. All analyses were conducted using IBM SPSS27 statistical software. The first step in the data analytic plan included cleaning and screening the dataset by (a) inspecting all variables for data entry errors (none were observed), and (b) examining the normality of distributions across study variables. Next, bivariate correlations were explored to examine the relationships among variables and determine whether it wou..., , # Using a standardized sound set to help characterize misophonia: The international affective digitized sounds
https://doi.org/10.5061/dryad.kh18932fd
MQincluded is the group variable. MQincluded=1 describes all participants who meet misophonia criteria and were included in the dataset. MQincluded=0 describes healthy controls. All variable names for the measures have descriptors in the "label" column of SPSS. Average ratings for the dependent variables are found at the end of the variable view in SPSS, as well as PANAS positive and negative scores and the AIM total score.
Data was derived from the following sources:
SPSS syntax is included with the data upload.Â
This data set contains SPSS syntaxes and simulated and real data for a statistical paper that compares different methods for analyzing a cluster randomized trial with a pretest and a posttest of a quantitative outcome variable. There is also a pdf which states per other file its type (spss system file or syntax file) and its role in the project. Basically, there are three files (one system file, two syntaxes) for doing simulations, and nine files (three system files, six syntaxes) for analyzing real data from a published cluster randomized trial (Kraag et al., J Child Psychology and Psychiatry, 2009). The system files with real data have restricted access in view of EU privacy legislation. All other files are freely accessible.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Balcony as a border between internal and external plays a crucial role in determining the quality of the apartments. The current condition of apartment balconies in Tehran, indicates that balconies have been gradually been forgotten due to their non-functional features for users. In architecture, space quality is measured by user preferences and can serve as a guide for designers in future designs. The objective of the present study was to identify the dominant physical characteristics of balconies in Tehran and improve people’s visual preference from both outside and inside viewpoints. Photographing, Delphi technique and collecting public questionnaire were used to do the research, and quantitative data were gathered through administering questionnaires. Valid responses were gathered, then the collected data were analyzed using SPSS. According to the results, there is a significant difference between the two categories of pleasant and unpleasant. In addition, the characteristics of desirable images from the outside and inside are not always the same. From the outside view of the balcony, the area is the most effective parameter, which in pleasant images was between 50% and 75% of the façade area. Fully solid and completely fences parapet, were preferred from the outside view, while from the user’s point of view, the space from the inside of the balcony with a completely fences parapet was unpleasant and a completely solid or a combination of a solid and fences was preferred.
The OECD Programme for International Student Assessment (PISA) surveys collected data on students’ performances in reading, mathematics and science, as well as contextual information on students’ background, home characteristics and school factors which could influence performance. This publication includes detailed information on how to analyse the PISA data, enabling researchers to both reproduce the initial results and to undertake further analyses. In addition to the inclusion of the necessary techniques, the manual also includes a detailed account of the PISA 2006 database and worked examples providing full syntax in SPSS.