59 datasets found
  1. s

    Excel Mapping Template for London Boroughs and Wards

    • ckan.publishing.service.gov.uk
    • data.europa.eu
    Updated Oct 28, 2025
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    (2025). Excel Mapping Template for London Boroughs and Wards [Dataset]. https://ckan.publishing.service.gov.uk/dataset/excel-mapping-template-for-london-boroughs-and-wards
    Explore at:
    Dataset updated
    Oct 28, 2025
    Area covered
    London
    Description

    A free mapping tool that allows you to create a thematic map of London without any specialist GIS skills or software - all you need is Microsoft Excel. Templates are available for London’s Boroughs and Wards. Full instructions are contained within the spreadsheets. Macros The tool works in any version of Excel. But the user MUST ENABLE MACROS, for the features to work. There a some restrictions on functionality in the ward maps in Excel 2003 and earlier - full instructions are included in the spreadsheet. To check whether the macros are enabled in Excel 2003 click Tools, Macro, Security and change the setting to Medium. Then you have to re-start Excel for the changes to take effect. When Excel starts up a prompt will ask if you want to enable macros - click yes. In Excel 2007 and later, it should be set by default to the correct setting, but if it has been changed, click on the Windows Office button in the top corner, then Excel options (at the bottom), Trust Centre, Trust Centre Settings, and make sure it is set to 'Disable all macros with notification'. Then when you open the spreadsheet, a prompt labelled 'Options' will appear at the top for you to enable macros. To create your own thematic borough maps in Excel using the ward map tool as a starting point, read these instructions. You will need to be a confident Excel user, and have access to your boundaries as a picture file from elsewhere. The mapping tools created here are all fully open access with no passwords. Copyright notice: If you publish these maps, a copyright notice must be included within the report saying: "Contains Ordnance Survey data © Crown copyright and database rights." NOTE: Excel 2003 users must 'ungroup' the map for it to work.

  2. c

    Ontario Data Catalogue (Ontario Data Catalogue)

    • catalog.civicdataecosystem.org
    Updated Nov 24, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    (2025). Ontario Data Catalogue (Ontario Data Catalogue) [Dataset]. https://catalog.civicdataecosystem.org/dataset/ontario-data-catalogue-ontario-data-catalogue
    Explore at:
    Dataset updated
    Nov 24, 2025
    Area covered
    Ontario
    Description

    AI Generated Summary: The Ontario Data Catalogue is a data portal providing access to open datasets generated and maintained by the Ontario government. It allows users to search, access, visualize, and download data in various machine-readable formats, often through APIs, while also indicating licensing terms and data update frequencies. The catalogue also provides tools for data visualization and notifications for dataset updates. About: The Ontario government generates and maintains thousands of datasets. Since 2012, we have shared data with Ontarians via a data catalogue. Open data is data that is shared with the public. Click here to learn more about open data and why Ontario releases it. Ontario’s Digital and Data Directive states that all data must be open, unless there is good reason for it to remain confidential. Ontario’s Chief Digital and Data Officer also has the authority to make certain datasets available publicly. Datasets listed in the catalogue that are not open will have one of the following labels: If you want to use data you find in the catalogue, that data must have a licence – a set of rules that describes how you can use it. A licence: Most of the data available in the catalogue is released under Ontario’s Open Government Licence. However, each dataset may be shared with the public under other kinds of licences or no licence at all. If a dataset doesn’t have a licence, you don’t have the right to use the data. If you have questions about how you can use a specific dataset, please contact us. The Ontario Data Catalogue endeavors to publish open data in a machine readable format. For machine readable datasets, you can simply retrieve the file you need using the file URL. The Ontario Data Catalogue is built on CKAN, which means the catalogue has the following features you can use when building applications. APIs (Application programming interfaces) let software applications communicate directly with each other. If you are using the catalogue in a software application, you might want to extract data from the catalogue through the catalogue API. Note: All Datastore API requests to the Ontario Data Catalogue must be made server-side. The catalogue's collection of dataset metadata (and dataset files) is searchable through the CKAN API. The Ontario Data Catalogue has more than just CKAN's documented search fields. You can also search these custom fields. You can also use the CKAN API to retrieve metadata about a particular dataset and check for updated files. Read the complete documentation for CKAN's API. Some of the open data in the Ontario Data Catalogue is available through the Datastore API. You can also search and access the machine-readable open data that is available in the catalogue. How to use the API feature: Read the complete documentation for CKAN's Datastore API. The Ontario Data Catalogue contains a record for each dataset that the Government of Ontario possesses. Some of these datasets will be available to you as open data. Others will not be available to you. This is because the Government of Ontario is unable to share data that would break the law or put someone's safety at risk. You can search for a dataset with a word that might describe a dataset or topic. Use words like “taxes” or “hospital locations” to discover what datasets the catalogue contains. You can search for a dataset from 3 spots on the catalogue: the homepage, the dataset search page, or the menu bar available across the catalogue. On the dataset search page, you can also filter your search results. You can select filters on the left hand side of the page to limit your search for datasets with your favourite file format, datasets that are updated weekly, datasets released by a particular ministry, or datasets that are released under a specific licence. Go to the dataset search page to see the filters that are available to make your search easier. You can also do a quick search by selecting one of the catalogue’s categories on the homepage. These categories can help you see the types of data we have on key topic areas. When you find the dataset you are looking for, click on it to go to the dataset record. Each dataset record will tell you whether the data is available, and, if so, tell you about the data available. An open dataset might contain several data files. These files might represent different periods of time, different sub-sets of the dataset, different regions, language translations, or other breakdowns. You can select a file and either download it or preview it. Make sure to read the licence agreement to make sure you have permission to use it the way you want. A non-open dataset may be not available for many reasons. Read more about non-open data. Read more about restricted data. Data that is non-open may still be subject to freedom of information requests. The catalogue has tools that enable all users to visualize the data in the catalogue without leaving the catalogue – no additional software needed. Get automatic notifications when datasets are updated. You can choose to get notifications for individual datasets, an organization’s datasets or the full catalogue. You don’t have to provide and personal information – just subscribe to our feeds using any feed reader you like using the corresponding notification web addresses. Copy those addresses and paste them into your reader. Your feed reader will let you know when the catalogue has been updated. The catalogue provides open data in several file formats (e.g., spreadsheets, geospatial data, etc). Learn about each format and how you can access and use the data each file contains. A file that has a list of items and values separated by commas without formatting (e.g. colours, italics, etc.) or extra visual features. This format provides just the data that you would display in a table. XLSX (Excel) files may be converted to CSV so they can be opened in a text editor. How to access the data: Open with any spreadsheet software application (e.g., Open Office Calc, Microsoft Excel) or text editor. Note: This format is considered machine-readable, it can be easily processed and used by a computer. Files that have visual formatting (e.g. bolded headers and colour-coded rows) can be hard for machines to understand, these elements make a file more human-readable and less machine-readable. A file that provides information without formatted text or extra visual features that may not follow a pattern of separated values like a CSV. How to access the data: Open with any word processor or text editor available on your device (e.g., Microsoft Word, Notepad). A spreadsheet file that may also include charts, graphs, and formatting. How to access the data: Open with a spreadsheet software application that supports this format (e.g., Open Office Calc, Microsoft Excel). Data can be converted to a CSV for a non-proprietary format of the same data without formatted text or extra visual features. A shapefile provides geographic information that can be used to create a map or perform geospatial analysis based on location, points/lines and other data about the shape and features of the area. It includes required files (.shp, .shx, .dbt) and might include corresponding files (e.g., .prj). How to access the data: Open with a geographic information system (GIS) software program (e.g., QGIS). A package of files and folders. The package can contain any number of different file types. How to access the data: Open with an unzipping software application (e.g., WinZIP, 7Zip). Note: If a ZIP file contains .shp, .shx, and .dbt file types, it is an ArcGIS ZIP: a package of shapefiles which provide information to create maps or perform geospatial analysis that can be opened with ArcGIS (a geographic information system software program). A file that provides information related to a geographic area (e.g., phone number, address, average rainfall, number of owl sightings in 2011 etc.) and its geospatial location (i.e., points/lines). How to access the data: Open using a GIS software application to create a map or do geospatial analysis. It can also be opened with a text editor to view raw information. Note: This format is machine-readable, and it can be easily processed and used by a computer. Human-readable data (including visual formatting) is easy for users to read and understand. A text-based format for sharing data in a machine-readable way that can store data with more unconventional structures such as complex lists. How to access the data: Open with any text editor (e.g., Notepad) or access through a browser. Note: This format is machine-readable, and it can be easily processed and used by a computer. Human-readable data (including visual formatting) is easy for users to read and understand. A text-based format to store and organize data in a machine-readable way that can store data with more unconventional structures (not just data organized in tables). How to access the data: Open with any text editor (e.g., Notepad). Note: This format is machine-readable, and it can be easily processed and used by a computer. Human-readable data (including visual formatting) is easy for users to read and understand. A file that provides information related to an area (e.g., phone number, address, average rainfall, number of owl sightings in 2011 etc.) and its geospatial location (i.e., points/lines). How to access the data: Open with a geospatial software application that supports the KML format (e.g., Google Earth). Note: This format is machine-readable, and it can be easily processed and used by a computer. Human-readable data (including visual formatting) is easy for users to read and understand. This format contains files with data from tables used for statistical analysis and data visualization of Statistics Canada census data. How to access the data: Open with the Beyond 20/20 application. A database which links and combines data from different files or

  3. e

    Teimpléad Mapála Excel do Bhuirgí agus Bardaí Londain

    • data.europa.eu
    Updated Apr 24, 2012
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Greater London Authority (2012). Teimpléad Mapála Excel do Bhuirgí agus Bardaí Londain [Dataset]. https://data.europa.eu/data/datasets/excel-mapping-template-for-london-boroughs-and-wards1?locale=ga
    Explore at:
    Dataset updated
    Apr 24, 2012
    Dataset authored and provided by
    Greater London Authority
    Area covered
    London
    Description

    A free mapping tool that allows you to create a thematic map of London without any specialist GIS skills or software - all you need is Microsoft Excel. Templates are available for London’s Boroughs and Wards. Full instructions are contained within the spreadsheets.

    Macros

    The tool works in any version of Excel. But the user MUST ENABLE MACROS, for the features to work. There a some restrictions on functionality in the ward maps in Excel 2003 and earlier - full instructions are included in the spreadsheet.

    To check whether the macros are enabled in Excel 2003 click Tools, Macro, Security and change the setting to Medium. Then you have to re-start Excel for the changes to take effect. When Excel starts up a prompt will ask if you want to enable macros - click yes.

    In Excel 2007 and later, it should be set by default to the correct setting, but if it has been changed, click on the Windows Office button in the top corner, then Excel options (at the bottom), Trust Centre, Trust Centre Settings, and make sure it is set to 'Disable all macros with notification'. Then when you open the spreadsheet, a prompt labelled 'Options' will appear at the top for you to enable macros.

    To create your own thematic borough maps in Excel using the ward map tool as a starting point, read these instructions. You will need to be a confident Excel user, and have access to your boundaries as a picture file from elsewhere. The mapping tools created here are all fully open access with no passwords.

    Copyright notice: If you publish these maps, a copyright notice must be included within the report saying: "Contains Ordnance Survey data © Crown copyright and database rights."

    NOTE: Excel 2003 users must 'ungroup' the map for it to work.

  4. New York State Unemployment Insurance Average

    • kaggle.com
    zip
    Updated Jan 7, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    The Devastator (2023). New York State Unemployment Insurance Average [Dataset]. https://www.kaggle.com/datasets/thedevastator/new-york-state-unemployment-insurance-average-du
    Explore at:
    zip(92067 bytes)Available download formats
    Dataset updated
    Jan 7, 2023
    Authors
    The Devastator
    Area covered
    New York
    Description

    New York State Unemployment Insurance Average Duration (2002-Present)

    Regional and County Level Trends

    By State of New York [source]

    About this dataset

    This dataset provides crucial insights on the unemployment benefits of New York State residents, unveiling the average duration of unemployment insurance security they receive during their benefit year. From January 2002 to present, discover trends related to ten labor market regions, recapping intricate information gathered from 62 counties and subdivisions. With a simple download of data including columns such as Year, Month, Region, County and Average Duration who insight can be provided with proper understanding and interpretation.

    As each region has distinct characteristics this dataset contains a broad spectrum of data types ranging from regular unemployment insurance (UI) cases not associated with Federal Employees (UCFE), Veterans (UCX), Self Employment Assistance Program (SEAP) or other situations to Shared Work programs including 599.2 training or Federal extensions recipients all adding tremendous value for users leveraging it responsibly. Before using the data make sure you read the Terms of Service in order to understand any legal requirements related executing use right upon installation! Last updated at 2020-09-16 this dataset is an April Fools gift not just for passionate researchers but also community impact leaders seeking direction when addressing prevalent social problems!

    More Datasets

    For more datasets, click here.

    Featured Notebooks

    • 🚨 Your notebook can be here! 🚨!

    How to use the dataset

    This dataset contains data on the average duration of unemployment insurance benefits in New York state from 2002 to present. This data can be useful for analyzing trends in unemployment rates, understanding regional differences, and evaluating labor market changes over time. In this guide we will explore how to use this dataset for your own research and analysis.

    Firstly, you'll need to download the dataset from Kaggle. Once downloaded, you can open it with a spreadsheet program such as Microsoft Excel or Google Sheets to begin exploring the data.

    The columns of the dataset that are available include Year, Month, Region, County, and Average Duration. Year indicates what year the related month's data falls under while Month shows which month that number corresponds with. Region and County represent the geographic areas these numbers are describing within New York State whereas Average Duration provides an indication of how long beneficiaries received their unemployment insurance benefits within their benefit year period on average within each given area.

    Using these columns as your guide you can start analyzing different aspects of state-level unemployment trends in New York over time or compare counties’ benefit level information against each other during any given year or specific month by filtering accordingly using Pivot Tables or Visualizations tools such as Microsoft Power BI and Tableau Desktop/Desktop Server depending on what type of analysis you want to conduct further down (e..g clustering/kmeans algorithms etc). You may also consider combining this with other macroeconomic datasets such as GDP growth rate per county/region etc., if applicable for further insight into factors influencing unemployed benefit duration levels over time etc.. Depending upon your objective make sure to review reference material cited at bottom part & ensure that all applicable terms & conditions have been read & accepted prior to proceeding further on research at hand!

    In conclusion ,this is a comprehensive yet easy-to-use source if you're looking for a detailed overview when examining Unemployment Insurance Average Duration across various geographic regions within New York State between 2002 up until present day! We hope that this guide outlined has been helpful in getting started with understanding insights relevant behind utilizing this powerful yet versatile dataset made available courtesy via Kaggle platform today!

    Research Ideas

    • Comparing current to historical unemployment insurance average duration trends (e.g. year over year, month to month).
    • Analyzing correlations between unemployment insurance average duration and other economic factors such as housing prices or wage growth in a particular county or region.
    • Mapping the distributions of unemployment insurance average duration across different regions and counties in New York State, providing useful insights into regional economic differences within the state that can inform policy decision-making by local governments

    Acknowledgements

    If you use this data...

  5. d

    Commodities Contracts

    • catalog.data.gov
    • demo.jkan.io
    Updated Mar 31, 2025
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    City of Philadelphia (2025). Commodities Contracts [Dataset]. https://catalog.data.gov/dataset/commodities-contracts
    Explore at:
    Dataset updated
    Mar 31, 2025
    Dataset provided by
    City of Philadelphia
    Description

    Commodities contracts are bid and awarded by the Procurement Department, for supplies, equipment, non-professional services, and public works. Each data set includes information regarding contracts that were renewed or received payment during the given quarter. When you click on a file below and it opens in a new tab, simple right click on the page, and choose 'save as.' When the save as dialog box appears, make sure the 'save as type' is Microsoft Excel Comma Separated Values File. You should then be able to open in excel.

  6. e

    PredloĹľak za mapiranje u Excelu za londonske ÄŤetvrti i odjele

    • data.europa.eu
    Updated Apr 9, 2020
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Greater London Authority (2020). PredloĹľak za mapiranje u Excelu za londonske ÄŤetvrti i odjele [Dataset]. https://data.europa.eu/data/datasets/excel-mapping-template-for-london-boroughs-and-wards1?locale=hr
    Explore at:
    Dataset updated
    Apr 9, 2020
    Dataset authored and provided by
    Greater London Authority
    Area covered
    London
    Description

    A free mapping tool that allows you to create a thematic map of London without any specialist GIS skills or software - all you need is Microsoft Excel. Templates are available for London’s Boroughs and Wards. Full instructions are contained within the spreadsheets. Macros The tool works in any version of Excel. But the user MUST ENABLE MACROS, for the features to work. There a some restrictions on functionality in the ward maps in Excel 2003 and earlier - full instructions are included in the spreadsheet. To check whether the macros are enabled in Excel 2003 click Tools, Macro, Security and change the setting to Medium. Then you have to re-start Excel for the changes to take effect. When Excel starts up a prompt will ask if you want to enable macros - click yes. In Excel 2007 and later, it should be set by default to the correct setting, but if it has been changed, click on the Windows Office button in the top corner, then Excel options (at the bottom), Trust Centre, Trust Centre Settings, and make sure it is set to 'Disable all macros with notification'. Then when you open the spreadsheet, a prompt labelled 'Options' will appear at the top for you to enable macros. To create your own thematic borough maps in Excel using the ward map tool as a starting point, read these instructions. You will need to be a confident Excel user, and have access to your boundaries as a picture file from elsewhere. The mapping tools created here are all fully open access with no passwords. Copyright notice: If you publish these maps, a copyright notice must be included within the report saying: "Contains Ordnance Survey data © Crown copyright and database rights." NOTE: Excel 2003 users must 'ungroup' the map for it to work.

  7. i

    Household Health Survey 2012-2013, Economic Research Forum (ERF)...

    • datacatalog.ihsn.org
    • catalog.ihsn.org
    Updated Jun 26, 2017
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Kurdistan Regional Statistics Office (KRSO) (2017). Household Health Survey 2012-2013, Economic Research Forum (ERF) Harmonization Data - Iraq [Dataset]. https://datacatalog.ihsn.org/catalog/6937
    Explore at:
    Dataset updated
    Jun 26, 2017
    Dataset provided by
    Central Statistical Organization (CSO)
    Kurdistan Regional Statistics Office (KRSO)
    Economic Research Forum
    Time period covered
    2012 - 2013
    Area covered
    Iraq
    Description

    Abstract

    The harmonized data set on health, created and published by the ERF, is a subset of Iraq Household Socio Economic Survey (IHSES) 2012. It was derived from the household, individual and health modules, collected in the context of the above mentioned survey. The sample was then used to create a harmonized health survey, comparable with the Iraq Household Socio Economic Survey (IHSES) 2007 micro data set.

    ----> Overview of the Iraq Household Socio Economic Survey (IHSES) 2012:

    Iraq is considered a leader in household expenditure and income surveys where the first was conducted in 1946 followed by surveys in 1954 and 1961. After the establishment of Central Statistical Organization, household expenditure and income surveys were carried out every 3-5 years in (1971/ 1972, 1976, 1979, 1984/ 1985, 1988, 1993, 2002 / 2007). Implementing the cooperation between CSO and WB, Central Statistical Organization (CSO) and Kurdistan Region Statistics Office (KRSO) launched fieldwork on IHSES on 1/1/2012. The survey was carried out over a full year covering all governorates including those in Kurdistan Region.

    The survey has six main objectives. These objectives are:

    1. Provide data for poverty analysis and measurement and monitor, evaluate and update the implementation Poverty Reduction National Strategy issued in 2009.
    2. Provide comprehensive data system to assess household social and economic conditions and prepare the indicators related to the human development.
    3. Provide data that meet the needs and requirements of national accounts.
    4. Provide detailed indicators on consumption expenditure that serve making decision related to production, consumption, export and import.
    5. Provide detailed indicators on the sources of households and individuals income.
    6. Provide data necessary for formulation of a new consumer price index number.

    The raw survey data provided by the Statistical Office were then harmonized by the Economic Research Forum, to create a comparable version with the 2006/2007 Household Socio Economic Survey in Iraq. Harmonization at this stage only included unifying variables' names, labels and some definitions. See: Iraq 2007 & 2012- Variables Mapping & Availability Matrix.pdf provided in the external resources for further information on the mapping of the original variables on the harmonized ones, in addition to more indications on the variables' availability in both survey years and relevant comments.

    Geographic coverage

    National coverage: Covering a sample of urban, rural and metropolitan areas in all the governorates including those in Kurdistan Region.

    Analysis unit

    1- Household/family. 2- Individual/person.

    Universe

    The survey was carried out over a full year covering all governorates including those in Kurdistan Region.

    Kind of data

    Sample survey data [ssd]

    Sampling procedure

    ----> Design:

    Sample size was (25488) household for the whole Iraq, 216 households for each district of 118 districts, 2832 clusters each of which includes 9 households distributed on districts and governorates for rural and urban.

    ----> Sample frame:

    Listing and numbering results of 2009-2010 Population and Housing Survey were adopted in all the governorates including Kurdistan Region as a frame to select households, the sample was selected in two stages: Stage 1: Primary sampling unit (blocks) within each stratum (district) for urban and rural were systematically selected with probability proportional to size to reach 2832 units (cluster). Stage two: 9 households from each primary sampling unit were selected to create a cluster, thus the sample size of total survey clusters was 25488 households distributed on the governorates, 216 households in each district.

    ----> Sampling Stages:

    In each district, the sample was selected in two stages: Stage 1: based on 2010 listing and numbering frame 24 sample points were selected within each stratum through systematic sampling with probability proportional to size, in addition to the implicit breakdown urban and rural and geographic breakdown (sub-district, quarter, street, county, village and block). Stage 2: Using households as secondary sampling units, 9 households were selected from each sample point using systematic equal probability sampling. Sampling frames of each stages can be developed based on 2010 building listing and numbering without updating household lists. In some small districts, random selection processes of primary sampling may lead to select less than 24 units therefore a sampling unit is selected more than once , the selection may reach two cluster or more from the same enumeration unit when it is necessary.

    Mode of data collection

    Face-to-face [f2f]

    Research instrument

    ----> Preparation:

    The questionnaire of 2006 survey was adopted in designing the questionnaire of 2012 survey on which many revisions were made. Two rounds of pre-test were carried out. Revision were made based on the feedback of field work team, World Bank consultants and others, other revisions were made before final version was implemented in a pilot survey in September 2011. After the pilot survey implemented, other revisions were made in based on the challenges and feedbacks emerged during the implementation to implement the final version in the actual survey.

    ----> Questionnaire Parts:

    The questionnaire consists of four parts each with several sections: Part 1: Socio – Economic Data: - Section 1: Household Roster - Section 2: Emigration - Section 3: Food Rations - Section 4: housing - Section 5: education - Section 6: health - Section 7: Physical measurements - Section 8: job seeking and previous job

    Part 2: Monthly, Quarterly and Annual Expenditures: - Section 9: Expenditures on Non – Food Commodities and Services (past 30 days). - Section 10 : Expenditures on Non – Food Commodities and Services (past 90 days). - Section 11: Expenditures on Non – Food Commodities and Services (past 12 months). - Section 12: Expenditures on Non-food Frequent Food Stuff and Commodities (7 days). - Section 12, Table 1: Meals Had Within the Residential Unit. - Section 12, table 2: Number of Persons Participate in the Meals within Household Expenditure Other Than its Members.

    Part 3: Income and Other Data: - Section 13: Job - Section 14: paid jobs - Section 15: Agriculture, forestry and fishing - Section 16: Household non – agricultural projects - Section 17: Income from ownership and transfers - Section 18: Durable goods - Section 19: Loans, advances and subsidies - Section 20: Shocks and strategy of dealing in the households - Section 21: Time use - Section 22: Justice - Section 23: Satisfaction in life - Section 24: Food consumption during past 7 days

    Part 4: Diary of Daily Expenditures: Diary of expenditure is an essential component of this survey. It is left at the household to record all the daily purchases such as expenditures on food and frequent non-food items such as gasoline, newspapers…etc. during 7 days. Two pages were allocated for recording the expenditures of each day, thus the roster will be consists of 14 pages.

    Cleaning operations

    ----> Raw Data:

    Data Editing and Processing: To ensure accuracy and consistency, the data were edited at the following stages: 1. Interviewer: Checks all answers on the household questionnaire, confirming that they are clear and correct. 2. Local Supervisor: Checks to make sure that questions has been correctly completed. 3. Statistical analysis: After exporting data files from excel to SPSS, the Statistical Analysis Unit uses program commands to identify irregular or non-logical values in addition to auditing some variables. 4. World Bank consultants in coordination with the CSO data management team: the World Bank technical consultants use additional programs in SPSS and STAT to examine and correct remaining inconsistencies within the data files. The software detects errors by analyzing questionnaire items according to the expected parameter for each variable.

    ----> Harmonized Data:

    • The SPSS package is used to harmonize the Iraq Household Socio Economic Survey (IHSES) 2007 with Iraq Household Socio Economic Survey (IHSES) 2012.
    • The harmonization process starts with raw data files received from the Statistical Office.
    • A program is generated for each dataset to create harmonized variables.
    • Data is saved on the household and individual level, in SPSS and then converted to STATA, to be disseminated.

    Response rate

    Iraq Household Socio Economic Survey (IHSES) reached a total of 25488 households. Number of households refused to response was 305, response rate was 98.6%. The highest interview rates were in Ninevah and Muthanna (100%) while the lowest rates were in Sulaimaniya (92%).

  8. u

    European Folk Costumes Excel Spreadsheet and Access Database

    • deepblue.lib.umich.edu
    Updated Mar 9, 2017
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    James, David A. (2017). European Folk Costumes Excel Spreadsheet and Access Database [Dataset]. http://doi.org/10.7302/Z2HD7SKC
    Explore at:
    Dataset updated
    Mar 9, 2017
    Dataset provided by
    Deep Blue Data
    Authors
    James, David A.
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Time period covered
    1997
    Description

    An Excel spreadsheet listing the information recorded on each of 18,686 costume designs can be viewed, downloaded, and explored. All the usual Excel sorting possibilities are available, and in addition a useful filter has been installed. For example, to find the number of designs that are Frieze Type #1, go to the top of the frieze type 2 column (column AS), click on the drop-down arrow and unselect every option box except True (i.e. True should be turned on, all other choices turned off). Then in the lower left corner, one reads “1111 of 18686 records found”.

    Much more sophisticated exploration can be carried out by downloading the rich and flexible Access Database. The terms used for this database were described in detail in three sections of Deep Blue paper associated with this project. The database can be downloaded and explored.

    HOW TO USE THE ACCESS DATABASE 1. Click on the Create Cohort and View Math Trait Data button, and select your cohort by clicking on the features of interest (for example: Apron and Blouse).

    Note: Depending on how you exited on your previous visit to the database, there may be items to clear up before creating the cohorts.
    a) (Usually unnecessary) Click on the small box near the top left corner to allow connection to Access. b) (Usually unnecessary) If an undesired window blocks part of the screen, click near the top of this window to minimize it. c) Make certain under Further Filtering that all four Exclude boxes are checked to get rid of stripes and circles, and circular buttons, and the D1 that is trivially associated with shoes.

    1. Click on Filter Records to Form the Cohort button. Note the # of designs, # of pieces, and # of costumes beside Recalculate.

    2. Click on Calculate Average Math Trait Frequency of Cohort button, and select the symmetry types of interest (for example: D1 and D2) .

    3. To view the Stage 1 table, click on Create Stage 1 table. To edit and print this table, click on Create Excel (after table has been created). The same process works for Stages 2, 3.and 4 tables.

    4. To view the matrix listing the math category impact numbers, move over to a button on the right side and click on View Matrix of Math Category Impact Numbers. To edit and print this matrix, click on Create Excel, use the Excel table as usual.

  9. w

    VA System of Records Notices (SORNs)

    • data.wu.ac.at
    • datahub.va.gov
    • +4more
    xls
    Updated Jul 26, 2017
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Department of Veterans Affairs (2017). VA System of Records Notices (SORNs) [Dataset]. https://data.wu.ac.at/schema/data_gov/ZDBhNDYxNjgtN2NjZS00NmE2LTlmYzAtNWVjYTE5ZGMwZTVi
    Explore at:
    xlsAvailable download formats
    Dataset updated
    Jul 26, 2017
    Dataset provided by
    Department of Veterans Affairs
    Description

    A system of records is a file, database, or program from which personal information is retrieved by name or other personal identifier. The Privacy Act provides a number of protections for your personal information. These typically include how information is collected, used, disclosed, stored, and disposed.As part of our privacy policy, VA conducts an annual review of our Privacy Act system of record notices to make sure that they are current and republishes those that require changes or updates.Please select the link to download the excel spreadsheet via the link labeled: 'Privacy Act System of Record'.The spreadsheet contains the following fields: SOR #, PUB DATE, CITATION, HYPERLINK TO FEDERAL REGISTER, SYSTEM TITLE, and POC.

  10. Z

    Data from: Lost in Translation: A Study of Bugs Introduced by Large Language...

    • data.niaid.nih.gov
    • nde-dev.biothings.io
    Updated Jan 25, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Ibrahimzada, Ali Reza (2024). Lost in Translation: A Study of Bugs Introduced by Large Language Models while Translating Code [Dataset]. https://data.niaid.nih.gov/resources?id=zenodo_8190051
    Explore at:
    Dataset updated
    Jan 25, 2024
    Dataset provided by
    University of Illinois Urbana-Champaign
    Authors
    Ibrahimzada, Ali Reza
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Artifact repository for the paper Lost in Translation: A Study of Bugs Introduced by Large Language Models while Translating Code, accepted at ICSE 2024, Lisbon, Portugal. Authors are Rangeet Pan* Ali Reza Ibrahimzada*, Rahul Krishna, Divya Sankar, Lambert Pougeum Wassi, Michele Merler, Boris Sobolev, Raju Pavuluri, Saurabh Sinha, and Reyhaneh Jabbarvand.

    Install

    This repository contains the source code for reproducing the results in our paper. Please start by cloning this repository:

    git clone https://github.com/Intelligent-CAT-Lab/PLTranslationEmpirical

    We recommend using a virtual environment for running the scripts. Please download conda 23.11.0 from this link. You can create a virtual environment using the following command:

    conda create -n plempirical python=3.10.13

    After creating the virtual environment, you can activate it using the following command:

    conda activate plempirical

    You can run the following command to make sure that you are using the correct version of Python:

    python3 --version && pip3 --version

    Dependencies

    To install all software dependencies, please execute the following command:

    pip3 install -r requirements.txt

    As for hardware dependencies, we used 16 NVIDIA A100 GPUs with 80GBs of memory for inferencing models. The models can be inferenced on any combination of GPUs as long as the reader can properly distribute the model weights across the GPUs. We did not perform weight distribution since we had enough memory (80 GB) per GPU.

    Moreover, for compiling and testing the generated translations, we used Python 3.10, g++ 11, GCC Clang 14.0, Java 11, Go 1.20, Rust 1.73, and .Net 7.0.14 for Python, C++, C, Java, Go, Rust, and C#, respectively. Overall, we recommend using a machine with Linux OS and at least 32GB of RAM for running the scripts.

    For running scripts of alternative approaches, you need to make sure you have installed C2Rust, CxGO, and Java2C# on your machine. Please refer to their repositories for installation instructions. For Java2C#, you need to create a .csproj file like below:

    Exe
    net7.0
    enable
    enable
    

    Dataset

    We uploaded the dataset we used in our empirical study to Zenodo. The dataset is organized as follows:

    CodeNet

    AVATAR

    Evalplus

    Apache Commons-CLI

    Click

    Please download and unzip the dataset.zip file from Zenodo. After unzipping, you should see the following directory structure:

    PLTranslationEmpirical ├── dataset ├── codenet ├── avatar ├── evalplus ├── real-life-cli ├── ...

    The structure of each dataset is as follows:

    1. CodeNet & Avatar: Each directory in these datasets correspond to a source language where each include two directories Code and TestCases for code snippets and test cases, respectively. Each code snippet has an id in the filename, where the id is used as a prefix for test I/O files.

    2. Evalplus: The source language code snippets follow a similar structure as CodeNet and Avatar. However, as a one time effort, we manually created the test cases in the target Java language inside a maven project, evalplus_java. To evaluate the translations from an LLM, we recommend moving the generated Java code snippets to the src/main/java directory of the maven project and then running the command mvn clean test surefire-report:report -Dmaven.test.failure.ignore=true to compile, test, and generate reports for the translations.

    3. Real-life Projects: The real-life-cli directory represents two real-life CLI projects from Java and Python. These datasets only contain code snippets as files and no test cases. As mentioned in the paper, the authors manually evaluated the translations for these datasets.

    Scripts

    We provide bash scripts for reproducing our results in this work. First, we discuss the translation script. For doing translation with a model and dataset, first you need to create a .env file in the repository and add the following:

    OPENAI_API_KEY= LLAMA2_AUTH_TOKEN= STARCODER_AUTH_TOKEN=

    1. Translation with GPT-4: You can run the following command to translate all Python -> Java code snippets in codenet dataset with the GPT-4 while top-k sampling is k=50, top-p sampling is p=0.95, and temperature=0.7:

    bash scripts/translate.sh GPT-4 codenet Python Java 50 0.95 0.7 0

    1. Translation with CodeGeeX: Prior to running the script, you need to clone the CodeGeeX repository from here and use the instructions from their artifacts to download their model weights. After cloning it inside PLTranslationEmpirical and downloading the model weights, your directory structure should be like the following:

    PLTranslationEmpirical ├── dataset ├── codenet ├── avatar ├── evalplus ├── real-life-cli ├── CodeGeeX ├── codegeex ├── codegeex_13b.pt # this file is the model weight ├── ... ├── ...

    You can run the following command to translate all Python -> Java code snippets in codenet dataset with the CodeGeeX while top-k sampling is k=50, top-p sampling is p=0.95, and temperature=0.2 on GPU gpu_id=0:

    bash scripts/translate.sh CodeGeeX codenet Python Java 50 0.95 0.2 0

    1. For all other models (StarCoder, CodeGen, LLaMa, TB-Airoboros, TB-Vicuna), you can execute the following command to translate all Python -> Java code snippets in codenet dataset with the StarCoder|CodeGen|LLaMa|TB-Airoboros|TB-Vicuna while top-k sampling is k=50, top-p sampling is p=0.95, and temperature=0.2 on GPU gpu_id=0:

    bash scripts/translate.sh StarCoder codenet Python Java 50 0.95 0.2 0

    1. For translating and testing pairs with traditional techniques (i.e., C2Rust, CxGO, Java2C#), you can run the following commands:

    bash scripts/translate_transpiler.sh codenet C Rust c2rust fix_report bash scripts/translate_transpiler.sh codenet C Go cxgo fix_reports bash scripts/translate_transpiler.sh codenet Java C# java2c# fix_reports bash scripts/translate_transpiler.sh avatar Java C# java2c# fix_reports

    1. For compile and testing of CodeNet, AVATAR, and Evalplus (Python to Java) translations from GPT-4, and generating fix reports, you can run the following commands:

    bash scripts/test_avatar.sh Python Java GPT-4 fix_reports 1 bash scripts/test_codenet.sh Python Java GPT-4 fix_reports 1 bash scripts/test_evalplus.sh Python Java GPT-4 fix_reports 1

    1. For repairing unsuccessful translations of Java -> Python in CodeNet dataset with GPT-4, you can run the following commands:

    bash scripts/repair.sh GPT-4 codenet Python Java 50 0.95 0.7 0 1 compile bash scripts/repair.sh GPT-4 codenet Python Java 50 0.95 0.7 0 1 runtime bash scripts/repair.sh GPT-4 codenet Python Java 50 0.95 0.7 0 1 incorrect

    1. For cleaning translations of open-source LLMs (i.e., StarCoder) in codenet, you can run the following command:

    bash scripts/clean_generations.sh StarCoder codenet

    Please note that for the above commands, you can change the dataset and model name to execute the same thing for other datasets and models. Moreover, you can refer to /prompts for different vanilla and repair prompts used in our study.

    Artifacts

    Please download the artifacts.zip file from our Zenodo repository. We have organized the artifacts as follows:

    RQ1 - Translations: This directory contains the translations from all LLMs and for all datasets. We have added an excel file to show a detailed breakdown of the translation results.

    RQ2 - Manual Labeling: This directory contains an excel file which includes the manual labeling results for all translation bugs.

    RQ3 - Alternative Approaches: This directory contains the translations from all alternative approaches (i.e., C2Rust, CxGO, Java2C#). We have added an excel file to show a detailed breakdown of the translation results.

    RQ4 - Mitigating Translation Bugs: This directory contains the fix results of GPT-4, StarCoder, CodeGen, and Llama 2. We have added an excel file to show a detailed breakdown of the fix results.

    Contact

    We look forward to hearing your feedback. Please contact Rangeet Pan or Ali Reza Ibrahimzada for any questions or comments 🙏.

  11. Energy Consumption of United States Over Time

    • kaggle.com
    zip
    Updated Dec 14, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    The Devastator (2022). Energy Consumption of United States Over Time [Dataset]. https://www.kaggle.com/datasets/thedevastator/unlocking-the-energy-consumption-of-united-state
    Explore at:
    zip(222388 bytes)Available download formats
    Dataset updated
    Dec 14, 2022
    Authors
    The Devastator
    License

    Attribution-ShareAlike 4.0 (CC BY-SA 4.0)https://creativecommons.org/licenses/by-sa/4.0/
    License information was derived automatically

    Area covered
    United States
    Description

    Energy Consumption of United States Over Time

    Building Energy Data Book

    By Department of Energy [source]

    About this dataset

    The Building Energy Data Book (2011) is an invaluable resource for gaining insight into the current state of energy consumption in the buildings sector. This dataset provides comprehensive data on residential, commercial and industrial building energy consumption, construction techniques, building technologies and characteristics. With this resource, you can get an in-depth understanding of how energy is used in various types of buildings - from single family homes to large office complexes - as well as its impact on the environment. The BTO within the U.S Department of Energy's Office of Energy Efficiency and Renewable Energy developed this dataset to provide a wealth of knowledge for researchers, policy makers, engineers and even everyday observers who are interested in learning more about our built environment and its energy usage patterns

    More Datasets

    For more datasets, click here.

    Featured Notebooks

    • 🚨 Your notebook can be here! 🚨!

    How to use the dataset

    This dataset provides comprehensive information regarding energy consumption in the buildings sector of the United States. It contains a number of key variables which can be used to analyze and explore the relations between energy consumption and building characteristics, technologies, and construction. The data is provided in both CSV format as well as tabular format which can make it helpful for those who prefer to use programs like Excel or other statistical modeling software.

    In order to get started with this dataset we've developed a guide outlining how to effectively use it for your research or project needs.

    • Understand what's included: Before you start analyzing the data, you should read through the provided documentation so that you fully understand what is included in the datasets. You'll want to be aware of any potential limitations or requirements associated with each type of data point so that your results are valid and reliable when drawing conclusions from them.

    • Clean up any outliers: You may need to take some time upfront investigating suspicious outliers within your dataset before using it in any further analyses — otherwise, they can skew results down the road if not dealt with first-hand! Furthermore, they could also make complex statistical modeling more difficult as well since they artificially inflate values depending on their magnitude within each example data point (i.e., one outlier could affect an entire model’s prior distributions). Missing values should also be accounted for too since these may not always appear obvious at first glance when reviewing a table or graphical representation - but accurate statistics must still be obtained either way no matter how messy things seem!

    • Exploratory data analysis: After cleaning up your dataset you'll want to do some basic exploring by visualizing different types of summaries like boxplots, histograms and scatter plots etc.. This will give you an initial case into what trends might exist within certain demographic/geographic/etc.. regions & variables which can then help inform future predictive models when needed! Additionally this step will highlight any clear discontinuous changes over time due over-generalization (if applicable), making sure predictors themselves don’t become part noise instead contributing meaningful signals towards overall effect predictions accuracy etc…

    • Analyze key metrics & observations: Once exploratory analyses have been carried out on rawsamples post-processing steps are next such as analyzing metrics such ascorrelations amongst explanatory functions; performing significance testing regression models; imputing missing/outlier values and much more depending upon specific project needs at hand… Additionally – interpretation efforts based

    Research Ideas

    • Creating an energy efficiency rating system for buildings - Using the dataset, an organization can develop a metric to rate the energy efficiency of commercial and residential buildings in a standardized way.
    • Developing targeted campaigns to raise awareness about energy conservation - Analyzing data from this dataset can help organizations identify areas of high energy consumption and create targeted campaigns and incentives to encourage people to conserve energy in those areas.
    • Estimating costs associated with upgrading building technologies - By evaluating various trends in building technologies and their associated costs, decision-makers can determine the most cost-effective option when it comes time to upgrade their structures' energy efficiency...
  12. p

    Hungary Phone Number Data

    • listtodata.com
    .csv, .xls, .txt
    Updated Jul 17, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    List to Data (2025). Hungary Phone Number Data [Dataset]. https://listtodata.com/hungary-number-data
    Explore at:
    .csv, .xls, .txtAvailable download formats
    Dataset updated
    Jul 17, 2025
    Authors
    List to Data
    License

    CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
    License information was derived automatically

    Time period covered
    Jan 1, 2025 - Dec 31, 2025
    Area covered
    Hungary
    Variables measured
    phone numbers, Email Address, full name, Address, City, State, gender,age,income,ip address,
    Description

    Hungary phone number database can bring new sales leads for any business. Most importantly, anyone can choose specific contacts based on their target group. Our expert team also helps you in multiple languages, so you can communicate clearly with customers. Actually, they adhere to a GDPR precise set of procedures for that aim. With this Hungary phone number database, you can share special offers, news, or updates in the language they understand best. In fact, our List To Data team only gathers leads from trustworthy sources. In the end, consider us for these causes and buy this contact number lead in an Excel file. Hungary mobile number data is perfect for your telemarketing campaigns. Similarly, the Hungary mobile number data ensures fast message delivery to keep you connected with your audience. People can also upgrade their products or services straight to their target audience from this number dataset. Moreover, we make sure that your library contains only real B2B and B2C leads. The experts ensure fast results within 24 hours, thus you stay connected. With a 100% prepaid fee, you can take precisely what to desire with no surprises. However, our List To Data website offers customizable packages to fit your needs.

  13. p

    Belize Number Dataset

    • listtodata.com
    • st.listtodata.com
    .csv, .xls, .txt
    Updated Jul 17, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    List to Data (2025). Belize Number Dataset [Dataset]. https://listtodata.com/belize-dataset
    Explore at:
    .csv, .xls, .txtAvailable download formats
    Dataset updated
    Jul 17, 2025
    Authors
    List to Data
    License

    CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
    License information was derived automatically

    Time period covered
    Jan 1, 2025 - Dec 31, 2025
    Area covered
    Belize
    Variables measured
    phone numbers, Email Address, full name, Address, City, State, gender,age,income,ip address,
    Description

    Belize number dataset is now available at List to Data. If anyone wants to increase the sales of their products, they should go after the directory. Our valuable offering can help you improve your marketing campaign in this area. Most importantly, you will receive hundreds of data and valid information if you buy this potent digital product from us. Additionally, the current and functional addresses we’ll offer will assist you and your company in the long run. Belize number dataset is right now the best for telemarketing and SMS marketing. Indeed, we have been in the market for a very long time and dealing with genuine contact lists. We have also developed a solid reputation through providing real service. So, if you purchase Belize number dataset, you can be certain that you will receive full value for your money. Belize phone data contains the best quality data for consumers. You can use these addresses to create effective marketing campaigns. Furthermore, using our list will help you generate a ton of new business and revenue. List to Data is there to create it for you. Consequently, it could benefit your company in every possible way. A well-built list allows you to launch any kind of marketing or sales products or services. Belize phone data can bring new sales leads for any business. However, we make sure that your list contains only genuine leads. We adhere to a precise set of procedures for that aim. Besides, our team exclusively gathers leads from trustworthy sources. Also, they check the Belize phone data several times and discard any that are no longer active. Again, they eliminate all inactive leads and duplicate contacts. Get Belize phone number list by just paying one time. The price of this dataset is very low. In this time, you will receive the precise data you need at a price that fits within your budget. After completing purchasing anyone can download it instantly in Excel or CSV file format. Use this comprehensive database to drive your business properly. In the end, Belize phone number list is the product that you need for your business. To know more about the product or if you are interested in our other products then please contact us. Also, after you purchase this Belize phone number list you will get 24/7 service support from us.

  14. CanadaBuys award notices

    • ouvert.canada.ca
    • canwin-datahub.ad.umanitoba.ca
    • +1more
    csv, html, xml
    Updated Nov 22, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Public Services and Procurement Canada (2025). CanadaBuys award notices [Dataset]. https://ouvert.canada.ca/data/dataset/a1acb126-9ce8-40a9-b889-5da2b1dd20cb
    Explore at:
    html, csv, xmlAvailable download formats
    Dataset updated
    Nov 22, 2025
    Dataset provided by
    Public Services and Procurement Canadahttp://www.pwgsc.gc.ca/
    License

    Open Government Licence - Canada 2.0https://open.canada.ca/en/open-government-licence-canada
    License information was derived automatically

    Description

    This dataset contains information on all Government of Canada award notices published according to the Financial Administration Act. It includes data for all Schedule I, Schedule II and Schedule III departments, agencies, Crown corporations, and other entities (unless specifically exempt) who must comply with the Government of Canada trade agreement obligations. CanadaBuys is the authoritative source of this information. Visit the How procurement works page on CanadaBuys to learn more. All data files in this collection share a common column structure, and the procurement category field (labelled as “*procurementCategory-categorieApprovisionnement*”) can be used to filter by the following four major categories of awards: - Awards for construction, which will have a value of “CNST” - Awards for goods, which will have a value of “GD” - Awards for services, which will have a value of “SRV” - Awards for services related to goods, which will have a value of “SRVTGD” Some award notices may be associated with one or more of the above procurement categories. >Note: Some records contain long award description values that may cause issues when viewed in certain spreadsheet programs, such as Microsoft Excel. When the information doesn’t fit within the cell’s character limit, the program will insert extra rows that don’t conform to the expected column formatting. (Though, all other records will still be displayed properly, in their own rows.) To quickly remove the “spill-over data” caused by this display error in Excel, select the publication date field (labelled as “*publicationDate-datePublication*”), then click the Filter button on the Data menu ribbon. You can then use the filter pull-down list to remove any blank or non-date values from this field, which will hide the rows that only contain “spill-over” description information. --- The following list describes the resources associated with this CanadaBuys award notices dataset. Additional information on Government of Canada award notices can be found on the Award notices tab of the CanadaBuys Tender opportunities page. >NOTE: While the CanadaBuys online portal includes awards notices from across multiple levels of government, the data files in this related dataset only include notices from federal government organizations. --- (1) CanadaBuys data dictionary: This XML file offers descriptions of each data field in the award notices files linked below, as well as other procurement-related datasets CanadaBuys produces. Use this as a guide for understanding the data elements in these files. This dictionary is updated as needed to reflect changes to the data elements. (2) All CanadaBuys award notices, 2022-08-08 onward: This file contains up to date information on all award notices published on CanadaBuys. This includes any award notices that were published on or after August 8, 2022, when CanadaBuys became the system of record for all tender and award notices for the Government of Canada. This file includes any amendments made to these award notices during their lifecycles. It is refreshed each morning, between 7:00 am and 8:30 am (UTC-0500) to include any updates or amendments, as needed. Award notices in this file can have any publication date on or after August 8, 2022 (displayed in the field labelled “*publicationDate-datePublication*”), and can have a status of active, cancelled or expired (displayed in the field labelled “*awardStatus-attributionStatut-eng*”). (3) Legacy award notices, 2012 to 2022-08 (prior to CanadaBuys): This file contains details of the award notices published prior to the implementation of CanadaBuys, which became the system of record for all tender and award notices for the Government of Canada on August 8, 2022. This datafile is refreshed monthly. The over 100,000 awards in this file have publication dates from August 6, 2022 and prior (displayed in the field labelled “*publicationDate-datePublication*”), and have a status of active, cancelled or expired (displayed included in the field labelled “*awardStatus-attributionStatut-eng*”). >Note: Procurement data was structured differently in the legacy applications previously used to administer Government of Canada contracts. Efforts have been made to manipulate these historical records into the structure used by the CanadaBuys data files, to make them easier to analyse and compare with new records. This process is not perfect since simple one-to-one mappings can’t be made in many cases. You can access these historical records in their original format as part of the archived copy of the original tender notices dataset, which contained awards-related data files. You can also refer to the supporting documentation for understanding the new CanadaBuys tender and award notices datasets. (4) Award notices, YYYY-YYYY: These files contain information on all contracts awarded in the specified fiscal year. The current fiscal year's file is refreshed each morning, between 7:00 am and 8:30 am (UTC-0500) to include any updates or amendments, as needed. The files associated with past fiscal years are updated monthly. Awards in these files can have any publication date between April 1 of a given year and March 31 of the subsequent year (displayed in the field labelled “*publicationDate-datePublication*”) and can have an award status of active, cancelled or expired (displayed in the field labelled “*awardStatus-attributionStatut-eng*”). >Note: New award notice data files will be added on April 1 for each fiscal year.

  15. g

    Jacob Kaplan's Concatenated Files: Uniform Crime Reporting (UCR) Program...

    • datasearch.gesis.org
    • openicpsr.org
    Updated Feb 19, 2020
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Kaplan, Jacob (2020). Jacob Kaplan's Concatenated Files: Uniform Crime Reporting (UCR) Program Data: Property Stolen and Recovered (Supplement to Return A) 1960-2017 [Dataset]. http://doi.org/10.3886/E105403V3
    Explore at:
    Dataset updated
    Feb 19, 2020
    Dataset provided by
    da|ra (Registration agency for social science and economic data)
    Authors
    Kaplan, Jacob
    Description

    For any questions about this data please email me at jacob@crimedatatool.com. If you use this data, please cite it.Version 3 release notes:Adds data in the following formats: Excel.Changes project name to avoid confusing this data for the ones done by NACJD.Version 2 release notes:Adds data for 2017.Adds a "number_of_months_reported" variable which says how many months of the year the agency reported data.Property Stolen and Recovered is a Uniform Crime Reporting (UCR) Program data set with information on the number of offenses (crimes included are murder, rape, robbery, burglary, theft/larceny, and motor vehicle theft), the value of the offense, and subcategories of the offense (e.g. for robbery it is broken down into subcategories including highway robbery, bank robbery, gas station robbery). The majority of the data relates to theft. Theft is divided into subcategories of theft such as shoplifting, theft of bicycle, theft from building, and purse snatching. For a number of items stolen (e.g. money, jewelry and previous metals, guns), the value of property stolen and and the value for property recovered is provided. This data set is also referred to as the Supplement to Return A (Offenses Known and Reported). All the data was received directly from the FBI as text or .DTA files. I created a setup file based on the documentation provided by the FBI and read the data into R using the package asciiSetupReader. All work to clean the data and save it in various file formats was also done in R. For the R code used to clean this data, see here: https://github.com/jacobkap/crime_data. The Word document file available for download is the guidebook the FBI provided with the raw data which I used to create the setup file to read in data.There may be inaccuracies in the data, particularly in the group of columns starting with "auto." To reduce (but certainly not eliminate) data errors, I replaced the following values with NA for the group of columns beginning with "offenses" or "auto" as they are common data entry error values (e.g. are larger than the agency's population, are much larger than other crimes or months in same agency): 1000, 2000, 3000, 4000, 5000, 6000, 7000, 8000, 9000, 10000, 20000, 30000, 40000, 50000, 60000, 70000, 80000, 90000, 100000, 99942. This cleaning was NOT done on the columns starting with "value."For every numeric column I replaced negative indicator values (e.g. "j" for -1) with the negative number they are supposed to be. These negative number indicators are not included in the FBI's codebook for this data but are present in the data. I used the values in the FBI's codebook for the Offenses Known and Clearances by Arrest data.To make it easier to merge with other data, I merged this data with the Law Enforcement Agency Identifiers Crosswalk (LEAIC) data. The data from the LEAIC add FIPS (state, county, and place) and agency type/subtype. If an agency has used a different FIPS code in the past, check to make sure the FIPS code is the same as in this data.

  16. Crimes - One year prior to present

    • chicago.gov
    • data.cityofchicago.org
    • +2more
    csv, xlsx, xml
    Updated Nov 24, 2025
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Chicago Police Department (2025). Crimes - One year prior to present [Dataset]. https://www.chicago.gov/city/en/dataset/crime.html
    Explore at:
    xlsx, xml, csvAvailable download formats
    Dataset updated
    Nov 24, 2025
    Dataset authored and provided by
    Chicago Police Departmenthttp://chicagopolice.org/
    Description

    This dataset reflects reported incidents of crime (with the exception of murders where data exists for each victim) that have occurred in the City of Chicago over the past year, minus the most recent seven days of data. Data is extracted from the Chicago Police Department's CLEAR (Citizen Law Enforcement Analysis and Reporting) system. In order to protect the privacy of crime victims, addresses are shown at the block level only and specific locations are not identified. Should you have questions about this dataset, you may contact the Research & Development Division of the Chicago Police Department at 312.745.6071 or RandD@chicagopolice.org. Disclaimer: These crimes may be based upon preliminary information supplied to the Police Department by the reporting parties that have not been verified. The preliminary crime classifications may be changed at a later date based upon additional investigation and there is always the possibility of mechanical or human error. Therefore, the Chicago Police Department does not guarantee (either expressed or implied) the accuracy, completeness, timeliness, or correct sequencing of the information and the information should not be used for comparison purposes over time. The Chicago Police Department will not be responsible for any error or omission, or for the use of, or the results obtained from the use of this information. All data visualizations on maps should be considered approximate and attempts to derive specific addresses are strictly prohibited.

    The Chicago Police Department is not responsible for the content of any off-site pages that are referenced by or that reference this web page other than an official City of Chicago or Chicago Police Department web page. The user specifically acknowledges that the Chicago Police Department is not responsible for any defamatory, offensive, misleading, or illegal conduct of other users, links, or third parties and that the risk of injury from the foregoing rests entirely with the user. The unauthorized use of the words "Chicago Police Department," "Chicago Police," or any colorable imitation of these words or the unauthorized use of the Chicago Police Department logo is unlawful. This web page does not, in any way, authorize such use. Data is updated daily Tuesday through Sunday. The dataset contains more than 65,000 records/rows of data and cannot be viewed in full in Microsoft Excel. Therefore, when downloading the file, select CSV from the Export menu. Open the file in an ASCII text editor, such as Wordpad, to view and search. To access a list of Chicago Police Department - Illinois Uniform Crime Reporting (IUCR) codes, go to http://bit.ly/rk5Tpc.

  17. Prescription Drugs Introduced to Market

    • catalog.data.gov
    • data.ca.gov
    • +3more
    Updated Oct 23, 2025
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Department of Health Care Access and Information (2025). Prescription Drugs Introduced to Market [Dataset]. https://catalog.data.gov/dataset/prescription-drugs-introduced-to-market-b62f7
    Explore at:
    Dataset updated
    Oct 23, 2025
    Dataset provided by
    Department of Health Care Access and Information
    Description

    This dataset provides data for new prescription drugs introduced to market in California with a Wholesale Acquisition Cost (WAC) that exceeds the Medicare Part D specialty drug cost threshold. Prescription drug manufacturers submit information to HCAI within a specified time period after a drug is introduced to market. Key data elements include the National Drug Code (NDC) administered by the FDA, a narrative description of marketing and pricing plans, and WAC, among other information. Manufacturers may withhold information that is not in the public domain. Note that prescription drug manufacturers are able to submit new drug reports for a prior quarter at any time. Therefore, the data set may include additional new drug report(s) from previous quarter(s). There are two types of New Drug data sets: Monthly and Annual. The Monthly data sets include the data in completed reports submitted by manufacturers for calendar year 2025, as of September 9, 2025. The Annual data sets include data in completed reports submitted by manufacturers for the specified year. The data sets may include reports that do not meet the specified minimum thresholds for reporting. The program regulations are available here: https://hcai.ca.gov/wp-content/uploads/2024/03/CTRx-Regulations-Text.pdf The data format and file specifications are available here: https://hcai.ca.gov/wp-content/uploads/2024/03/Format-and-File-Specifications-version-2.0-ada.pdf DATA NOTES: Due to recent changes in Excel capabilities, it is not recommended that you save these files to .csv format. If you do, when importing back into Excel the leading zeros in the NDC number column will be dropped. If you need to save it into a different format other than .xlsx it must be .txt DATA UPDATES: Drug manufacturers may submit New Drug reports to HCAI for prescription drugs which were not initially reported when they were introduced to market. CTRx staff update the posted datasets monthly for current year data and as needed for previous years. Please check the 'Data last updated' date on each dataset page to ensure you are viewing the most current data.

  18. Supplemental data for thesis

    • figshare.com
    txt
    Updated Feb 2, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Alina Cepraga (2022). Supplemental data for thesis [Dataset]. http://doi.org/10.6084/m9.figshare.12445634.v1
    Explore at:
    txtAvailable download formats
    Dataset updated
    Feb 2, 2022
    Dataset provided by
    figshare
    Figsharehttp://figshare.com/
    Authors
    Alina Cepraga
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Supplemental data for thesis submitted to Cardiff University School of Biosciences:Cepraga, A.A. (2020) "An investigation into Mef2-interacting proteins during embryonic muscle development of Drosophila"Each file is provided twice: in a plain-text format such as .tsv , and as an Excel file. Note that .tsv files are tab-delimited. Make sure that they are not treated as space-delimited when imported in spreadsheet software to display them properly.This repository contains:- Table S1: Raw spectral counts and hit ranking of all proteins identified by TAP/MS (total spectral counts for each protein identified by MS in tandem affinity purification from Drosophila embryos with Mef2 as the tagged bait)- Table S2: Mef2 GO biological process terms- Table S3: High-interest candidates identified by each analysis- Table S4: Overlap of TAP datasets with muscleRNAi screen by Schnorrer et al 2010- Data supplementals for individual figures. The file names indicate which figure the file corresponds to. Most of these report the IDs of genes in the different overlap fields of Venn diagrams. The dataset for Fig4.12 contains the functional similarity network including the pairwise functional similarity scores between the candidates.- The file for Fig4.8 contains pmax for all candidates across embryonic stagings. The file for Fig4.10 contains GO enrichment for all Mef2 GO terms in each dataset and subset. These two files are in the appropriate format for use in the Morpheus matrix visualisation tool at https://software.broadinstitute.org/morpheus/ (Download the .gct files from here, open the Morpheus page, click "select file" and select the .gct file from your Downloads folder.)- a Morpheus color scheme file to be used in conjunction with Fig4-10GOenrichment[...].gct. To apply the colour scheme, first download both the Fig4-10.gct and the color-scheme.json files. Open the Fig4-10.gct in Morpheus as described above. Click the Options button in the top-right corner (cogwheel symbol), which opens the Color scheme dialog. Scroll down to find the field "Saved color scheme" and select "My computer" from the dropdown box. Finally, click "Load color scheme" and select the downloaded .json file from your Downloads folder.

  19. US Travel Check-Ins - Analysis

    • kaggle.com
    zip
    Updated Feb 11, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    The Devastator (2023). US Travel Check-Ins - Analysis [Dataset]. https://www.kaggle.com/datasets/thedevastator/us-travel-check-ins-analysis
    Explore at:
    zip(2350764 bytes)Available download formats
    Dataset updated
    Feb 11, 2023
    Authors
    The Devastator
    License

    https://creativecommons.org/publicdomain/zero/1.0/https://creativecommons.org/publicdomain/zero/1.0/

    Area covered
    United States
    Description

    US Travel Check-Ins - Analysis

    In-Depth Study of Location, Date, Temperature, USIndex, and Crime Rates

    By [source]

    About this dataset

    This comprehensive dataset offers an in-depth exploration into US travel check-ins from Instagram. It includes detailed data scraped from Instagram, such as the location of each check-in, the USIndex for each state, average temperature for each state per month, and crime rate per state. In addition to location and time information, this dataset also provides latitude and longitude coordinates for every entry. This extensive collection of data is invaluable for those interested in studying various aspects of movement within the United States. With detailed insights on factors like climate conditions and economic health of a region at a given point in time, this dataset can help uncover fascinating trends regarding how travelers choose their destinations and how they experience their journeys around the country

    More Datasets

    For more datasets, click here.

    Featured Notebooks

    • 🚨 Your notebook can be here! 🚨!

    How to use the dataset

    This Kaggle dataset - US Travel Check-Ins Analysis - provides valuable insights for travel researchers, marketers and businesses in the travel industry. It contains check-in location, USIndex rating (economic health of each state), average temperature, and crime rate per state. Latitude and longitude of each check-ins are also provided with added geographic context to help you visualize the data.

    This guide will show you how to use this dataset for your research or business venture.

    Step 1: Prepare your data First and foremost, it is important to cleanse your data before you can analyze it. Depending on what sort of analysis needs to be conducted (e.g., time series analysis) you will need to select the applicable columns from the dataset that match your needs best and exclude any unnecessary columns such as dates or season related data points as they are not relevant here. Furthermore, variable formatting should be consistent across all instances in a variable/column category as well (elevation is a good example here). You can always double check that everything is formatted correctly by running a quick summary on selected columns using conditional queries like df['var'].describe() command in Python for descriptive results about an entire column’s statistical makeup including mean values, quartile ranges etc..

    Step 2: Explore & Analyze Your Data Graphically Once the data has been prepped properly you can start visualizing it in order to gain better insights into any trends or patterns that may be present within it when compared with other datasets or information sources simultaneously such as weather forecasts or nationwide trend indicators etc.. Grafana dashboards are feasible solutions when multiple dataset need to be compared but depending on what type of graphs/charts being used Excel worksheet formats can offer great customization options flexiblity along with various export file types (.csv; .jpegs; .pdfs). Plotting markers onto map applications like Google Maps API offers more geographical awareness that could useful when analyzing location dependent variables too which means we have one advantage over manual inspection tasks just by leveraging existing software applications alongside publicly available APIs!

    Step 3: Interpretation & Hypothesis Testing
    After generating informative graphical interpretation from exploratory visualizations the next step would involve testing out various hypotheses based on established correlations between different variables derived from overall quantitative estimates vizualizations regarding distribution trends across different regions tends towards geographical areas where certain logistical processes could yeild higher success ratios giving potential customers greater satisfaction than

    Research Ideas

    • Travel trends analysis: Using this dataset, researchers could track which areas of the US are popular destinations based on travel check-ins and spot any interesting trends or correlations in terms of geography, seasonal changes, economic health or crime rates.
    • Predictive Modeling: By using various features from this dataset such as average temperature, US Index and crime rate, predictors could be developed to suggest how safe an area would feel to a tourist based on their current location and other predetermined variables they choose to input into the model.
    • Trip Planning Tool: The dataset can also be used to develop a tool that quickly allows travelers to plan trips according to their preferences in terms of duration and budget as well a...
  20. CanadaBuys tender notices

    • ouvert.canada.ca
    • open.canada.ca
    csv, html, xml
    Updated Nov 26, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Public Services and Procurement Canada (2025). CanadaBuys tender notices [Dataset]. https://ouvert.canada.ca/data/dataset/6abd20d4-7a1c-4b38-baa2-9525d0bb2fd2
    Explore at:
    csv, xml, htmlAvailable download formats
    Dataset updated
    Nov 26, 2025
    Dataset provided by
    Public Services and Procurement Canadahttp://www.pwgsc.gc.ca/
    License

    Open Government Licence - Canada 2.0https://open.canada.ca/en/open-government-licence-canada
    License information was derived automatically

    Description

    This dataset contains information on Government of Canada tender information published according to the Financial Administration Act. It includes data for all Schedule I, Schedule II and Schedule III departments, agencies, Crown corporations, and other entities (unless specifically exempt) who must comply with the Government of Canada trade agreement obligations. CanadaBuys is the authoritative source of this information. Visit the How procurement works page on the CanadaBuys website to learn more. All data files in this collection share a common column structure, and the procurement category field (labelled as “*procurementCategory-categorieApprovisionnement*”) can be used to filter by the following four major categories of tenders: - Tenders for construction, which will have a value of “CNST” - Tenders for goods, which will have a value of “GD” - Tenders for services, which will have a value of “SRV” - Tenders for services related to goods, which will have a value of “SRVTGD” A tender may be associated with one or more of the above procurement categories. >Note: Some records contain long tender description values that may cause issues when viewed in certain spreadsheet programs, such as Microsoft Excel. When the information doesn’t fit within the cell’s character limit, the program will insert extra rows that don’t conform to the expected column formatting. (Though, all other records will still be displayed properly, in their own rows.) To quickly remove the “spill-over data” caused by this display error in Excel, select the publication date field (labelled as “*publicationDate-datePublication*”), then click the Filter button on the Data menu ribbon. You can then use the filter pull-down list to remove any blank or non-date values from this field, which will hide the rows that only contain “spill-over” description information. --- The following list describes the resources associated with this CanadaBuys tender notices dataset. Additional information on Government of Canada tenders can also be found on the Tender notices tab of the CanadaBuys tender opportunities page. >NOTE: While the CanadaBuys online portal includes tender opportunities from across multiple levels of government, the data files in this related dataset only include notices from federal government organizations. --- (1) CanadaBuys data dictionary: This XML file offers descriptions of each data field in the tender notices files linked below, as well as other procurement-related datasets CanadaBuys produces. Use this as a guide for understanding the data elements in these files. This dictionary is updated as needed to reflect changes to the data elements. (2) New tender notices: This file contains up to date information on all new tender notices that are published to CanadaBuys throughout a given day. The file is updated every two hours, from 6:15 am until 10:15 pm (UTC-0500) to include new tenders as they are published. All tenders in this file will have a publication date matching the current day (displayed in the field labelled “*publicationDate-datePublication*”), or the day prior for systems that feed into this file on a nightly basis. (3) Open tender notices: This file contains up to date information on all tender notices that are open for bidding on CanadaBuys, including any amendments made to these tender notices during their lifecycles. The file is refreshed each morning, between 7:00 am and 8:30 am (UTC-0500) to include newly published open tenders. All tenders in this file will have a status of open (displayed in the field labelled “*tenderStatus-tenderStatut-eng*”). (4) All CanadaBuys tender notices, 2022-08-08 onwards: This file contains up to date information on all tender notices published through CanadaBuys. This includes any tender notices that were open for bids on or after August 8, 2022, when CanadaBuys launched as the system of record for all Tender Notices for the Government of Canada. This file includes any amendments made to these tender notices during their lifecycles. It is refreshed each morning, between 7:00 am and 8:30 am (UTC-0500) to include any updates or amendments, as needed. Tender notices in this file can have any publication date on or after August 8, 2022 (displayed in the field labelled “*publicationDate-datePublication*”), and can have a status of open, cancelled or expired (displayed in the field labelled “*tenderStatus-tenderStatut-eng*”). (5) Legacy tender notices, 2009 to 2022-08 (prior to CanadaBuys): This file contains details of the tender notices that were launched prior to the implementation of CanadaBuys, which became the system of record for all tender notices for the Government of Canada on August 8, 2022. This datafile is refreshed monthly. The over 70,000 tenders in this file have publication dates from August 5, 2022 and before (displayed in the field labelled “*publicationDate-datePublication*”) and have a status of cancelled or expired (displayed in the field labelled “*tenderStatus-tenderStatut-eng*”). >Note: Procurement data was structured differently in the legacy applications previously used to administer Government of Canada tender notices. Efforts have been made to manipulate these historical records into the structure used by the CanadaBuys data files, to make them easier to analyse and compare with new records. This process is not perfect since simple one-to-one mappings can’t be made in many cases. You can access these historical records in their original format as part of the archived copy of the original tender notices dataset. You can also refer to the supporting documentation for understanding the new CanadaBuys tender and award notices datasets. (6) Tender notices, YYYY-YYYY: These files contain information on all tender notices published in the specified fiscal year that are no longer open to bidding. The current fiscal year's file is refreshed each morning, between 7:00 am and 8:30 am (UTC-0500) to include any updates or amendments, as needed. The files associated with past fiscal years are refreshed monthly. Tender notices in these files can have any publication date between April 1 of a given year and March 31 of the subsequent year (displayed in the field labelled “*publicationDate-datePublication*”) and can have a status of cancelled or expired (displayed in the field labelled “*tenderStatus-tenderStatut-eng*”). New records are added to these files once related tenders reach their close date, or are cancelled. >Note: New tender notice data files will be added on April 1 for each fiscal year.

Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
(2025). Excel Mapping Template for London Boroughs and Wards [Dataset]. https://ckan.publishing.service.gov.uk/dataset/excel-mapping-template-for-london-boroughs-and-wards

Excel Mapping Template for London Boroughs and Wards

Explore at:
Dataset updated
Oct 28, 2025
Area covered
London
Description

A free mapping tool that allows you to create a thematic map of London without any specialist GIS skills or software - all you need is Microsoft Excel. Templates are available for London’s Boroughs and Wards. Full instructions are contained within the spreadsheets. Macros The tool works in any version of Excel. But the user MUST ENABLE MACROS, for the features to work. There a some restrictions on functionality in the ward maps in Excel 2003 and earlier - full instructions are included in the spreadsheet. To check whether the macros are enabled in Excel 2003 click Tools, Macro, Security and change the setting to Medium. Then you have to re-start Excel for the changes to take effect. When Excel starts up a prompt will ask if you want to enable macros - click yes. In Excel 2007 and later, it should be set by default to the correct setting, but if it has been changed, click on the Windows Office button in the top corner, then Excel options (at the bottom), Trust Centre, Trust Centre Settings, and make sure it is set to 'Disable all macros with notification'. Then when you open the spreadsheet, a prompt labelled 'Options' will appear at the top for you to enable macros. To create your own thematic borough maps in Excel using the ward map tool as a starting point, read these instructions. You will need to be a confident Excel user, and have access to your boundaries as a picture file from elsewhere. The mapping tools created here are all fully open access with no passwords. Copyright notice: If you publish these maps, a copyright notice must be included within the report saying: "Contains Ordnance Survey data © Crown copyright and database rights." NOTE: Excel 2003 users must 'ungroup' the map for it to work.

Search
Clear search
Close search
Google apps
Main menu