7 datasets found
  1. U

    Statistical Computing: Stata

    • dataverse-staging.rdmc.unc.edu
    pdf +4
    Updated Feb 27, 2014
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Catherine Zimmer; Catherine Zimmer (2014). Statistical Computing: Stata [Dataset]. https://dataverse-staging.rdmc.unc.edu/dataset.xhtml?persistentId=hdl:1902.29/11638
    Explore at:
    pdf(193387), tsv(126), text/plain; charset=us-ascii(146), pdf(318256), tsv(4026), text/plain; charset=us-ascii(443), pdf(352818), text/x-stata-syntax; charset=us-ascii(547), tsv(219941), tsv(75), text/x-stata-syntax; charset=us-ascii(211), tsv(3591), tsv(1152), tsv(2874), tsv(1157), xls(2549)Available download formats
    Dataset updated
    Feb 27, 2014
    Dataset provided by
    UNC Dataverse
    Authors
    Catherine Zimmer; Catherine Zimmer
    License

    https://dataverse-staging.rdmc.unc.edu/api/datasets/:persistentId/versions/4.0/customlicense?persistentId=hdl:1902.29/11638https://dataverse-staging.rdmc.unc.edu/api/datasets/:persistentId/versions/4.0/customlicense?persistentId=hdl:1902.29/11638

    Description

    This is a 3-part short course (held over three afternoons). Stata part 1 will offer an introduction to Stata for Windows. Part 2 will teach entering data in Stata, working with Stata do files, and show how to append, sort, and merge data sets in Stata. Part 3 teaches how to perform basic statistical procedures and how to draw sub samples from large datasets.

  2. H

    pop_dhs: Merging multiple DHS files

    • dataverse.harvard.edu
    Updated May 22, 2014
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Günther Fink (2014). pop_dhs: Merging multiple DHS files [Dataset]. http://doi.org/10.7910/DVN/26098
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    May 22, 2014
    Dataset provided by
    Harvard Dataverse
    Authors
    Günther Fink
    License

    CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
    License information was derived automatically

    Description

    This .do file merges multiple DHS's, and keeps memory requirements to a minimum; what it does it quite simple: 1. Loops through all available survey data files (and counts them) 2. For each file, generates variables with missing values if they are not in the file 3. Generates a partial file for each variable extraction 4. Merges all partial files and deletes them What the user needs to do is to: - paste the name of all surveys on top of the file (after local survey_list) - paste the list of all DHS variables needed into the file (after global myvars)

  3. o

    National Neighborhood Data Archive (NaNDA): Code for merging ZCTA level...

    • openicpsr.org
    Updated Jun 25, 2020
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Megan Chenoweth; Anam Khan (2020). National Neighborhood Data Archive (NaNDA): Code for merging ZCTA level datasets with the UDS Mapper ZIP code to ZCTA crosswalk [Dataset]. http://doi.org/10.3886/E120088V2
    Explore at:
    Dataset updated
    Jun 25, 2020
    Dataset provided by
    University of Michigan. Institute for Social Research
    Authors
    Megan Chenoweth; Anam Khan
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    NaNDA contains some datasets that measure neighborhood context at the ZIP Code Tabulation Area (ZCTA) level. They are intended for use with survey or other individual-level data containing ZIP codes. Because ZIP codes do not exactly match ZIP code tabulation areas, a crosswalk is required to use ZIP-code-level geocoded datasets with ZCTA-level datasets from NaNDA. A ZIP-code-to-ZCTA crosswalk is available on the UDS Mapper website at https://udsmapper.org/zip-code-to-zcta-crosswalk/.The sample SAS and Stata code provided here merges the UDS Mapper crosswalk with NaNDA datasets.

  4. Data from: Code for Merging Waves of the Crime Survey of England and Wales...

    • beta.ukdataservice.ac.uk
    Updated 2025
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Niels Blom (2025). Code for Merging Waves of the Crime Survey of England and Wales and the British Crime Survey, 1982-2024 [Dataset]. http://doi.org/10.5255/ukda-sn-857928
    Explore at:
    Dataset updated
    2025
    Dataset provided by
    DataCitehttps://www.datacite.org/
    UK Data Servicehttps://ukdataservice.ac.uk/
    Authors
    Niels Blom
    Description

    This code merges multiple years of Crime Survey of England and Wales (CSEW) and/or the British Crime Survey (BCS). The current version merges the BCS and CSEW up to the CSEW 2023/2024. The purpose of these code is to help researchers to quickly and easily combine multiple survey sweeps of the CSEW and BCS.

    By combining multiple survey sweeps, people are able to look at, for instance, trends in violence. Furthermore, using such a combined file enables you to look at specific offences, population groups, or consequences, that do not have a high enough frequency if you would use only a single year.

    This is a Stata do file, access to Stata is therefore required, as is access to all the BCS and CSEW that you want to merge. In specifying the code, you can decide which files you want to merge. Namely, which years of the Crime Surveys you want to merge and if you want the bolt-on datasets that provide uncapped codes, the adolescent and young adult panels, and/or if you want to use the ‘non-white’ panel. This code does not harmonize variables that are different between years.

    All original data resources are available via Related Resources.

  5. H

    County FIPS Matching Tool

    • dataverse.harvard.edu
    Updated Jan 20, 2019
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Carl Klarner (2019). County FIPS Matching Tool [Dataset]. http://doi.org/10.7910/DVN/OSLU4G
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Jan 20, 2019
    Dataset provided by
    Harvard Dataverse
    Authors
    Carl Klarner
    License

    CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
    License information was derived automatically

    Description

    This tool--a simple csv or Stata file for merging--gives you a fast way to assign Census county FIPS codes to variously presented county names. This is useful for dealing with county names collected from official sources, such as election returns, which inconsistently present county names and often have misspellings. It will likely take less than ten minutes the first time, and about one minute thereafter--assuming all versions of your county names are in this file. There are about 3,142 counties in the U.S., and there are 77,613 different permutations of county names in this file (ave=25 per county, max=382). Counties with more likely permutations have more versions. Misspellings were added as I came across them over time. I DON'T expect people to cite the use of this tool. DO feel free to suggest the addition of other county name permutations.

  6. H

    Current Population Survey (CPS)

    • dataverse.harvard.edu
    • search.dataone.org
    Updated May 30, 2013
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Anthony Damico (2013). Current Population Survey (CPS) [Dataset]. http://doi.org/10.7910/DVN/AK4FDD
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    May 30, 2013
    Dataset provided by
    Harvard Dataverse
    Authors
    Anthony Damico
    License

    CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
    License information was derived automatically

    Description

    analyze the current population survey (cps) annual social and economic supplement (asec) with r the annual march cps-asec has been supplying the statistics for the census bureau's report on income, poverty, and health insurance coverage since 1948. wow. the us census bureau and the bureau of labor statistics ( bls) tag-team on this one. until the american community survey (acs) hit the scene in the early aughts (2000s), the current population survey had the largest sample size of all the annual general demographic data sets outside of the decennial census - about two hundred thousand respondents. this provides enough sample to conduct state- and a few large metro area-level analyses. your sample size will vanish if you start investigating subgroups b y state - consider pooling multiple years. county-level is a no-no. despite the american community survey's larger size, the cps-asec contains many more variables related to employment, sources of income, and insurance - and can be trended back to harry truman's presidency. aside from questions specifically asked about an annual experience (like income), many of the questions in this march data set should be t reated as point-in-time statistics. cps-asec generalizes to the united states non-institutional, non-active duty military population. the national bureau of economic research (nber) provides sas, spss, and stata importation scripts to create a rectangular file (rectangular data means only person-level records; household- and family-level information gets attached to each person). to import these files into r, the parse.SAScii function uses nber's sas code to determine how to import the fixed-width file, then RSQLite to put everything into a schnazzy database. you can try reading through the nber march 2012 sas importation code yourself, but it's a bit of a proc freak show. this new github repository contains three scripts: 2005-2012 asec - download all microdata.R down load the fixed-width file containing household, family, and person records import by separating this file into three tables, then merge 'em together at the person-level download the fixed-width file containing the person-level replicate weights merge the rectangular person-level file with the replicate weights, then store it in a sql database create a new variable - one - in the data table 2012 asec - analysis examples.R connect to the sql database created by the 'download all microdata' progr am create the complex sample survey object, using the replicate weights perform a boatload of analysis examples replicate census estimates - 2011.R connect to the sql database created by the 'download all microdata' program create the complex sample survey object, using the replicate weights match the sas output shown in the png file below 2011 asec replicate weight sas output.png statistic and standard error generated from the replicate-weighted example sas script contained in this census-provided person replicate weights usage instructions document. click here to view these three scripts for more detail about the current population survey - annual social and economic supplement (cps-asec), visit: the census bureau's current population survey page the bureau of labor statistics' current population survey page the current population survey's wikipedia article notes: interviews are conducted in march about experiences during the previous year. the file labeled 2012 includes information (income, work experience, health insurance) pertaining to 2011. when you use the current populat ion survey to talk about america, subract a year from the data file name. as of the 2010 file (the interview focusing on america during 2009), the cps-asec contains exciting new medical out-of-pocket spending variables most useful for supplemental (medical spending-adjusted) poverty research. confidential to sas, spss, stata, sudaan users: why are you still rubbing two sticks together after we've invented the butane lighter? time to transition to r. :D

  7. d

    Replication Data for: What do cross-country surveys tell us about social...

    • dataone.org
    • dataverse.harvard.edu
    Updated Nov 8, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    David Tannenbaum; Alain Cohn; Christian L. Zünd; Michel A. Maréchal (2023). Replication Data for: What do cross-country surveys tell us about social capital? [Dataset]. http://doi.org/10.7910/DVN/NDDWHJ
    Explore at:
    Dataset updated
    Nov 8, 2023
    Dataset provided by
    Harvard Dataverse
    Authors
    David Tannenbaum; Alain Cohn; Christian L. Zünd; Michel A. Maréchal
    Description

    Code and data to reproduce all results and graphs reported in Tannenbaum et al. (2022). This folder contains data files (.dta files) and a Stata do-file (code.do) that stitches together the different data files and executes all analyses and produces all figures reported in the paper. The do-file uses a number of user-written packages, which are listed below. Most of these can be installed using the ssc install command in Stata. Also, users will need to change the current directory path (at the start of the do-file) before executing the code. List of user written packages (descriptions): revrs (reverse-codes variable) ereplace (extends the egen command to permit replacing) grstyle (changes the settings for the overall look of graphs) spmap (used for graphing spatial data) qqvalue (used for obtaining Benjamini-Hochberg corrected p-values) parmby (creates a dataset by calling an estimation command for each by-group) domin (used to perform dominance analyses) coefplot (used for creating coefficient plots) grc1leg (combine graphs with a single common legend) xframeappend (append data frames to the end of the current data frame)

  8. Not seeing a result you expected?
    Learn how you can add new datasets to our index.

Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
Catherine Zimmer; Catherine Zimmer (2014). Statistical Computing: Stata [Dataset]. https://dataverse-staging.rdmc.unc.edu/dataset.xhtml?persistentId=hdl:1902.29/11638

Statistical Computing: Stata

Explore at:
51 scholarly articles cite this dataset (View in Google Scholar)
pdf(193387), tsv(126), text/plain; charset=us-ascii(146), pdf(318256), tsv(4026), text/plain; charset=us-ascii(443), pdf(352818), text/x-stata-syntax; charset=us-ascii(547), tsv(219941), tsv(75), text/x-stata-syntax; charset=us-ascii(211), tsv(3591), tsv(1152), tsv(2874), tsv(1157), xls(2549)Available download formats
Dataset updated
Feb 27, 2014
Dataset provided by
UNC Dataverse
Authors
Catherine Zimmer; Catherine Zimmer
License

https://dataverse-staging.rdmc.unc.edu/api/datasets/:persistentId/versions/4.0/customlicense?persistentId=hdl:1902.29/11638https://dataverse-staging.rdmc.unc.edu/api/datasets/:persistentId/versions/4.0/customlicense?persistentId=hdl:1902.29/11638

Description

This is a 3-part short course (held over three afternoons). Stata part 1 will offer an introduction to Stata for Windows. Part 2 will teach entering data in Stata, working with Stata do files, and show how to append, sort, and merge data sets in Stata. Part 3 teaches how to perform basic statistical procedures and how to draw sub samples from large datasets.

Search
Clear search
Close search
Google apps
Main menu