Build and customise datasets to match your target audience profile, from a database of 200 million global contacts generated in real-time. Get business contact information that's verified by Leadbook's proprietary A.I. powered data technology.
Our Industry data enables you to reach the prospects and maximize your sales and revenue by offering the most impeccable data. Our data covers several industries that provide result-oriented records to help you build and grow business. Our industry-wise data is a vast repository of verified and opt-in contacts.
Executives and Professionals Contact Data to connect with prospects to effectively market B2B products and services. All of our email addresses come with a 97% deliverability or better guarantee.
Simply specify location, industry, employee headcount, job function and/or seniority attributes, then the platform will verify in real-time their business contact information, and you can download the records in a CSV file.
All records include: - Contact name - Job title - Contact email address - Contact location - Contact LinkedIn URL - Organisation name - Organisation website - Organisation type - Organisation headcount - Primary industry
Additional information like organization phone numbers, organization address, business registration number and secondary industries may be provided where available.
Price starts from USD 0.40 per contact rent & USD 0.80 per contact purchase. Bulk discounts apply.
== Quick starts ==
Batch export podcast metadata to CSV files:
1) Export by search keyword: https://www.listennotes.com/podcast-datasets/keyword/
2) Export by category: https://www.listennotes.com/podcast-datasets/category/
== Quick facts ==
The most up-to-date and comprehensive podcast database available All languages & All countries Includes over 3,500,000 podcasts Features 35+ data fields , such as basic metadata, global rank, RSS feed (with audio URLs), Spotify links, and more Delivered in CSV format
== Data Attributes ==
See the full list of data attributes on this page: https://www.listennotes.com/podcast-datasets/fields/?filter=podcast_only
How to access podcast audio files: Our dataset includes RSS feed URLs for all podcasts. You can retrieve audio for over 170 million episodes directly from these feeds. With access to the raw audio, you’ll have high-quality podcast speech data ideal for AI training and related applications.
== Custom Offers ==
We can provide custom datasets based on your needs, such as language-specific data, daily/weekly/monthly update frequency, or one-time purchases.
We also provide a RESTful API at PodcastAPI.com
Contact us: hello@listennotes.com
== Need Help? ==
If you have any questions about our products, feel free to reach out hello@listennotes.com
== About Listen Notes, Inc. ==
Since 2017, Listen Notes, Inc. has provided the leading podcast search engine and podcast database.
Complete list of all 1984 Apparel & Accessories closings in USA in February 2022 with name, geo-coded address, city, email, phone number etc for download in CSV format or via the API.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
The purpose of this document is to accompany the public release of data collected from OpenCon 2015 applications.Download & Technical Information The data can be downloaded in CSV format from GitHub here: https://github.com/RightToResearch/OpenCon-2015-Application-Data The file uses UTF8 encoding, comma as field delimiter, quotation marks as text delimiter, and no byte order mark.
This data is released to the public for free and open use under a CC0 1.0 license. We have a couple of requests for anyone who uses the data. First, we’d love it if you would let us know what you are doing with it, and share back anything you develop with the OpenCon community (#opencon / @open_con ). Second, it would also be great if you would include a link to the OpenCon 2015 website (www.opencon2015.org) wherever the data is used. You are not obligated to do any of this, but we’d appreciate it!
Unique ID
This is a unique ID assigned to each applicant. Numbers were assigned using a random number generator.
Timestamp
This was the timestamp recorded by google forms. Timestamps are in EDT (Eastern U.S. Daylight Time). Note that the application process officially began at 1:00pm EDT June 1 ended at 6:00am EDT on June 23. Some applications have timestamps later than this date, and this is due to a variety of reasons including exceptions granted for technical difficulties, error corrections (which required re-submitting the form), and applications sent in via email and later entered manually into the form. [a]
Gender
Mandatory. Choose one from list or fill-in other. Options provided: Male, Female, Other (fill in).
Country of Nationality
Mandatory. Choose one option from list.
Country of Residence
Mandatory. Choose one option from list.
What is your primary occupation?
Mandatory. Choose one from list or fill-in other. Options provided: Undergraduate student; Masters/professional student; PhD candidate; Faculty/teacher; Researcher (non-faculty); Librarian; Publisher; Professional advocate; Civil servant / government employee; Journalist; Doctor / medical professional; Lawyer; Other (fill in).
Select the option below that best describes your field of study or expertise
Mandatory. Choose one option from list.
What is your primary area of interest within OpenCon’s program areas?
Mandatory. Choose one option from list. Note: for the first approximately 24 hours the options were listed in this order: Open Access, Open Education, Open Data. After that point, we set the form to randomize the order, and noticed an immediate shift in the distribution of responses.
Are you currently engaged in activities to advance Open Access, Open Education, and/or Open Data?
Mandatory. Choose one option from list.
Are you planning to participate in any of the following events this year?
Optional. Choose all that apply from list. Multiple selections separated by semi-colon.
Do you have any of the following skills or interests?
Mandatory. Choose all that apply from list or fill-in other. Multiple selections separated by semi-colon. Options provided: Coding; Website Management / Design; Graphic Design; Video Editing; Community / Grassroots Organizing; Social Media Campaigns; Fundraising; Communications and Media; Blogging; Advocacy and Policy; Event Logistics; Volunteer Management; Research about OpenCon's Issue Areas; Other (fill-in).
This data consists of information collected from people who applied to attend OpenCon 2015. In the application form, questions that would be released as Open Data were marked with a caret (^) and applicants were asked to acknowledge before submitting the form that they understood that their responses to these questions would be released as such. The questions we released were selected to avoid any potentially sensitive personal information, and to minimize the chances that any individual applicant can be positively identified. Applications were formally collected during a 22 day period beginning on June 1, 2015 at 13:00 EDT and ending on June 23 at 06:00 EDT. Some applications have timestamps later than this date, and this is due to a variety of reasons including exceptions granted for technical difficulties, error corrections (which required re-submitting the form), and applications sent in via email and later entered manually into the form. Applications were collected using a Google Form embedded at http://www.opencon2015.org/attend, and the shortened bit.ly link http://bit.ly/AppsAreOpen was promoted through social media. The primary work we did to clean the data focused on identifying and eliminating duplicates. We removed all duplicate applications that had matching e-mail addresses and first and last names. We also identified a handful of other duplicates that used different e-mail addresses but were otherwise identical. In cases where duplicate applications contained any different information, we kept the information from the version with the most recent timestamp. We made a few minor adjustments in the country field for cases where the entry was obviously an error (for example, electing a country listed alphabetically above or below the one indicated elsewhere in the application). We also removed one potentially offensive comment (which did not contain an answer to the question) from the Gender field and replaced it with “Other.”
OpenCon 2015 is the student and early career academic professional conference on Open Access, Open Education, and Open Data and will be held on November 14-16, 2015 in Brussels, Belgium. It is organized by the Right to Research Coalition, SPARC (The Scholarly Publishing and Academic Resources Coalition), and an Organizing Committee of students and early career researchers from around the world. The meeting will convene students and early career academic professionals from around the world and serve as a powerful catalyst for projects led by the next generation to advance OpenCon's three focus areas—Open Access, Open Education, and Open Data. A unique aspect of OpenCon is that attendance at the conference is by application only, and the majority of participants who apply are awarded travel scholarships to attend. This model creates a unique conference environment where the most dedicated and impactful advocates can attend, regardless of where in the world they live or their access to travel funding. The purpose of the application process is to conduct these selections fairly. This year we were overwhelmed by the quantity and quality of applications received, and we hope that by sharing this data, we can better understand the OpenCon community and the state of student and early career participation in the Open Access, Open Education, and Open Data movements.
For inquires about the OpenCon 2015 Application data, please contact Nicole Allen at nicole@sparc.arl.org.
Complete list of all 334 Village Medical clinic POI locations in the USA with name, geo-coded address, city, email, phone number etc for download in CSV format or via the API.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
LScD (Leicester Scientific Dictionary)April 2020 by Neslihan Suzen, PhD student at the University of Leicester (ns433@leicester.ac.uk/suzenneslihan@hotmail.com)Supervised by Prof Alexander Gorban and Dr Evgeny Mirkes[Version 3] The third version of LScD (Leicester Scientific Dictionary) is created from the updated LSC (Leicester Scientific Corpus) - Version 2*. All pre-processing steps applied to build the new version of the dictionary are the same as in Version 2** and can be found in description of Version 2 below. We did not repeat the explanation. After pre-processing steps, the total number of unique words in the new version of the dictionary is 972,060. The files provided with this description are also same as described as for LScD Version 2 below.* Suzen, Neslihan (2019): LSC (Leicester Scientific Corpus). figshare. Dataset. https://doi.org/10.25392/leicester.data.9449639.v2** Suzen, Neslihan (2019): LScD (Leicester Scientific Dictionary). figshare. Dataset. https://doi.org/10.25392/leicester.data.9746900.v2[Version 2] Getting StartedThis document provides the pre-processing steps for creating an ordered list of words from the LSC (Leicester Scientific Corpus) [1] and the description of LScD (Leicester Scientific Dictionary). This dictionary is created to be used in future work on the quantification of the meaning of research texts. R code for producing the dictionary from LSC and instructions for usage of the code are available in [2]. The code can be also used for list of texts from other sources, amendments to the code may be required.LSC is a collection of abstracts of articles and proceeding papers published in 2014 and indexed by the Web of Science (WoS) database [3]. Each document contains title, list of authors, list of categories, list of research areas, and times cited. The corpus contains only documents in English. The corpus was collected in July 2018 and contains the number of citations from publication date to July 2018. The total number of documents in LSC is 1,673,824.LScD is an ordered list of words from texts of abstracts in LSC.The dictionary stores 974,238 unique words, is sorted by the number of documents containing the word in descending order. All words in the LScD are in stemmed form of words. The LScD contains the following information:1.Unique words in abstracts2.Number of documents containing each word3.Number of appearance of a word in the entire corpusProcessing the LSCStep 1.Downloading the LSC Online: Use of the LSC is subject to acceptance of request of the link by email. To access the LSC for research purposes, please email to ns433@le.ac.uk. The data are extracted from Web of Science [3]. You may not copy or distribute these data in whole or in part without the written consent of Clarivate Analytics.Step 2.Importing the Corpus to R: The full R code for processing the corpus can be found in the GitHub [2].All following steps can be applied for arbitrary list of texts from any source with changes of parameter. The structure of the corpus such as file format and names (also the position) of fields should be taken into account to apply our code. The organisation of CSV files of LSC is described in README file for LSC [1].Step 3.Extracting Abstracts and Saving Metadata: Metadata that include all fields in a document excluding abstracts and the field of abstracts are separated. Metadata are then saved as MetaData.R. Fields of metadata are: List_of_Authors, Title, Categories, Research_Areas, Total_Times_Cited and Times_cited_in_Core_Collection.Step 4.Text Pre-processing Steps on the Collection of Abstracts: In this section, we presented our approaches to pre-process abstracts of the LSC.1.Removing punctuations and special characters: This is the process of substitution of all non-alphanumeric characters by space. We did not substitute the character “-” in this step, because we need to keep words like “z-score”, “non-payment” and “pre-processing” in order not to lose the actual meaning of such words. A processing of uniting prefixes with words are performed in later steps of pre-processing.2.Lowercasing the text data: Lowercasing is performed to avoid considering same words like “Corpus”, “corpus” and “CORPUS” differently. Entire collection of texts are converted to lowercase.3.Uniting prefixes of words: Words containing prefixes joined with character “-” are united as a word. The list of prefixes united for this research are listed in the file “list_of_prefixes.csv”. The most of prefixes are extracted from [4]. We also added commonly used prefixes: ‘e’, ‘extra’, ‘per’, ‘self’ and ‘ultra’.4.Substitution of words: Some of words joined with “-” in the abstracts of the LSC require an additional process of substitution to avoid losing the meaning of the word before removing the character “-”. Some examples of such words are “z-test”, “well-known” and “chi-square”. These words have been substituted to “ztest”, “wellknown” and “chisquare”. Identification of such words is done by sampling of abstracts form LSC. The full list of such words and decision taken for substitution are presented in the file “list_of_substitution.csv”.5.Removing the character “-”: All remaining character “-” are replaced by space.6.Removing numbers: All digits which are not included in a word are replaced by space. All words that contain digits and letters are kept because alphanumeric characters such as chemical formula might be important for our analysis. Some examples are “co2”, “h2o” and “21st”.7.Stemming: Stemming is the process of converting inflected words into their word stem. This step results in uniting several forms of words with similar meaning into one form and also saving memory space and time [5]. All words in the LScD are stemmed to their word stem.8.Stop words removal: Stop words are words that are extreme common but provide little value in a language. Some common stop words in English are ‘I’, ‘the’, ‘a’ etc. We used ‘tm’ package in R to remove stop words [6]. There are 174 English stop words listed in the package.Step 5.Writing the LScD into CSV Format: There are 1,673,824 plain processed texts for further analysis. All unique words in the corpus are extracted and written in the file “LScD.csv”.The Organisation of the LScDThe total number of words in the file “LScD.csv” is 974,238. Each field is described below:Word: It contains unique words from the corpus. All words are in lowercase and their stem forms. The field is sorted by the number of documents that contain words in descending order.Number of Documents Containing the Word: In this content, binary calculation is used: if a word exists in an abstract then there is a count of 1. If the word exits more than once in a document, the count is still 1. Total number of document containing the word is counted as the sum of 1s in the entire corpus.Number of Appearance in Corpus: It contains how many times a word occurs in the corpus when the corpus is considered as one large document.Instructions for R CodeLScD_Creation.R is an R script for processing the LSC to create an ordered list of words from the corpus [2]. Outputs of the code are saved as RData file and in CSV format. Outputs of the code are:Metadata File: It includes all fields in a document excluding abstracts. Fields are List_of_Authors, Title, Categories, Research_Areas, Total_Times_Cited and Times_cited_in_Core_Collection.File of Abstracts: It contains all abstracts after pre-processing steps defined in the step 4.DTM: It is the Document Term Matrix constructed from the LSC[6]. Each entry of the matrix is the number of times the word occurs in the corresponding document.LScD: An ordered list of words from LSC as defined in the previous section.The code can be used by:1.Download the folder ‘LSC’, ‘list_of_prefixes.csv’ and ‘list_of_substitution.csv’2.Open LScD_Creation.R script3.Change parameters in the script: replace with the full path of the directory with source files and the full path of the directory to write output files4.Run the full code.References[1]N. Suzen. (2019). LSC (Leicester Scientific Corpus) [Dataset]. Available: https://doi.org/10.25392/leicester.data.9449639.v1[2]N. Suzen. (2019). LScD-LEICESTER SCIENTIFIC DICTIONARY CREATION. Available: https://github.com/neslihansuzen/LScD-LEICESTER-SCIENTIFIC-DICTIONARY-CREATION[3]Web of Science. (15 July). Available: https://apps.webofknowledge.com/[4]A. Thomas, "Common Prefixes, Suffixes and Roots," Center for Development and Learning, 2013.[5]C. Ramasubramanian and R. Ramya, "Effective pre-processing activities in text mining using improved porter’s stemming algorithm," International Journal of Advanced Research in Computer and Communication Engineering, vol. 2, no. 12, pp. 4536-4538, 2013.[6]I. Feinerer, "Introduction to the tm Package Text Mining in R," Accessible en ligne: https://cran.r-project.org/web/packages/tm/vignettes/tm.pdf, 2013.
Complete list of all 367 Any Lab Test Now POI locations in the USA with name, geo-coded address, city, email, phone number etc for download in CSV format or via the API.
Complete list of all 112 Alta Convenience store POI locations in the the USA with name, geo-coded address, city, email, phone number etc for download in CSV format or via the API.
Premium B2C Consumer Database - 269+ Million US Records
Supercharge your B2C marketing campaigns with comprehensive consumer database, featuring over 269 million verified US consumer records. Our 20+ year data expertise delivers higher quality and more extensive coverage than competitors.
Core Database Statistics
Consumer Records: Over 269 million
Email Addresses: Over 160 million (verified and deliverable)
Phone Numbers: Over 76 million (mobile and landline)
Mailing Addresses: Over 116,000,000 (NCOA processed)
Geographic Coverage: Complete US (all 50 states)
Compliance Status: CCPA compliant with consent management
Targeting Categories Available
Demographics: Age ranges, education levels, occupation types, household composition, marital status, presence of children, income brackets, and gender (where legally permitted)
Geographic: Nationwide, state-level, MSA (Metropolitan Service Area), zip code radius, city, county, and SCF range targeting options
Property & Dwelling: Home ownership status, estimated home value, years in residence, property type (single-family, condo, apartment), and dwelling characteristics
Financial Indicators: Income levels, investment activity, mortgage information, credit indicators, and wealth markers for premium audience targeting
Lifestyle & Interests: Purchase history, donation patterns, political preferences, health interests, recreational activities, and hobby-based targeting
Behavioral Data: Shopping preferences, brand affinities, online activity patterns, and purchase timing behaviors
Multi-Channel Campaign Applications
Deploy across all major marketing channels:
Email marketing and automation
Social media advertising
Search and display advertising (Google, YouTube)
Direct mail and print campaigns
Telemarketing and SMS campaigns
Programmatic advertising platforms
Data Quality & Sources
Our consumer data aggregates from multiple verified sources:
Public records and government databases
Opt-in subscription services and registrations
Purchase transaction data from retail partners
Survey participation and research studies
Online behavioral data (privacy compliant)
Technical Delivery Options
File Formats: CSV, Excel, JSON, XML formats available
Delivery Methods: Secure FTP, API integration, direct download
Processing: Real-time NCOA, email validation, phone verification
Custom Selections: 1,000+ selectable demographic and behavioral attributes
Minimum Orders: Flexible based on targeting complexity
Unique Value Propositions
Dual Spouse Targeting: Reach both household decision-makers for maximum impact
Cross-Platform Integration: Seamless deployment to major ad platforms
Real-Time Updates: Monthly data refreshes ensure maximum accuracy
Advanced Segmentation: Combine multiple targeting criteria for precision campaigns
Compliance Management: Built-in opt-out and suppression list management
Ideal Customer Profiles
E-commerce retailers seeking customer acquisition
Financial services companies targeting specific demographics
Healthcare organizations with compliant marketing needs
Automotive dealers and service providers
Home improvement and real estate professionals
Insurance companies and agents
Subscription services and SaaS providers
Performance Optimization Features
Lookalike Modeling: Create audiences similar to your best customers
Predictive Scoring: Identify high-value prospects using AI algorithms
Campaign Attribution: Track performance across multiple touchpoints
A/B Testing Support: Split audiences for campaign optimization
Suppression Management: Automatic opt-out and DNC compliance
Pricing & Volume Options
Flexible pricing structures accommodate businesses of all sizes:
Pay-per-record for small campaigns
Volume discounts for large deployments
Subscription models for ongoing campaigns
Custom enterprise pricing for high-volume users
Data Compliance & Privacy
VIA.tools maintains industry-leading compliance standards:
CCPA (California Consumer Privacy Act) compliant
CAN-SPAM Act adherence for email marketing
TCPA compliance for phone and SMS campaigns
Regular privacy audits and data governance reviews
Transparent opt-out and data deletion processes
Getting Started
Our data specialists work with you to:
Define your target audience criteria
Recommend optimal data selections
Provide sample data for testing
Configure delivery methods and formats
Implement ongoing campaign optimization
Why We Lead the Industry
With over two decades of data industry experience, we combine extensive database coverage with advanced targeting capabilities. Our commitment to data quality, compliance, and customer success has made us the preferred choice for businesses seeking superior B2C marketing performance.
Contact our team to discuss your specific targeting requirements and receive custom pricing for your marketing objectives.
Hospitals: A full list of health services, including name, address, contact details and Local Health District is available for download in CSV format. Hospitals: A full list of health services, including name, address, contact details and Local Health District is available for download in CSV format.
https://creativecommons.org/publicdomain/zero/1.0/https://creativecommons.org/publicdomain/zero/1.0/
The dataset supports various deep learning applications, including facial anomaly detection, tissue segmentation, and 3D modeling of facial anatomy. With high-resolution sagittal and axial slices, it is ideal for training AI models aimed at accurate facial analysis.
The dataset includes data that showcases the diversity and complexity of facial MRI imaging, suitable for machine learning models and medical analysis. It includes:
All data is anonymized to ensure privacy and complies with publication consent regulations.
The dataset provides a sample from one patient, showcasing the diversity of the full dataset. It contains the following files for exploration:
- DICOM slices with 100 frames
- 3D representation of the facial structure
- CSV file listing the scan characteristics
Complete list of all 1587 Sams Club store POI locations in the USA with name, geo-coded address, city, email, phone number etc for download in CSV format or via the API.
Complete list of all 8927 Napa Auto Parts dealer POI locations in the the USA with name, geo-coded address, city, email, phone number etc for download in CSV format or via the API.
Complete list of all 423 The Perfume Shop store POI locations in the UK with name, geo-coded address, city, email, phone number etc for download in CSV format or via the API.
Complete list of all 252 Cost Plus World Market store POI locations in the USA with name, geo-coded address, city, email, phone number etc for download in CSV format or via the API.
Not seeing a result you expected?
Learn how you can add new datasets to our index.
Build and customise datasets to match your target audience profile, from a database of 200 million global contacts generated in real-time. Get business contact information that's verified by Leadbook's proprietary A.I. powered data technology.
Our Industry data enables you to reach the prospects and maximize your sales and revenue by offering the most impeccable data. Our data covers several industries that provide result-oriented records to help you build and grow business. Our industry-wise data is a vast repository of verified and opt-in contacts.
Executives and Professionals Contact Data to connect with prospects to effectively market B2B products and services. All of our email addresses come with a 97% deliverability or better guarantee.
Simply specify location, industry, employee headcount, job function and/or seniority attributes, then the platform will verify in real-time their business contact information, and you can download the records in a CSV file.
All records include: - Contact name - Job title - Contact email address - Contact location - Contact LinkedIn URL - Organisation name - Organisation website - Organisation type - Organisation headcount - Primary industry
Additional information like organization phone numbers, organization address, business registration number and secondary industries may be provided where available.
Price starts from USD 0.40 per contact rent & USD 0.80 per contact purchase. Bulk discounts apply.