Andrew Wharton's US Consumer Email Databases provide over 650 million current and active email address records in our 36-month Production Email Database, and additionally, over 1.4 billion historical records in our Legacy Email Database. These databases offer a comprehensive look-back at the digital and terrestrial identity information associated with a consumer. This Identity Graph Data has been collected from website registrations and is 100% opted-in for Third Party Uses.
The Email Address Data is fully populated with email addresses, HEMS (MD5, Sha1, Sha256), first name, last name, postal address (primary and secondary), IP Address, and Time Stamps for Last Registration, Verification, and First Seen. Additionally, our email address information assets can be linked with our Date-of-Birth and Phone Number databases to provide a powerful solution for consumer identity recognition and verification platforms through Identity Linkage Data.
As an add-on to our current and historical information, we also offer a database of hard-bounce email addresses. These are email addresses that have hard-bounced during our large-scale email campaign deployments or were identified as hard-bounces during our email verification processes. This database provides over 400 million unproductive email addresses useable as a part of suppression or fraud identification applications.
Our Email Information Assets are utilized by major Identity Graph Data and Identity Linkage platforms due to our comprehensive information that links the email address to consumer identity and IP Address information. This Identity Graph Data provides a robust alternative approach when faced with third-party cookie deprecation in the digital ecosystem.
Our digital advertising partners leverage this information to understand where their clients' customers and prospects are online and align media and content with consumer behavior. The additional Email Address Data, mobile phone numbers, and IP Addresses also work to increase the reach of your Digital Audience Data.
This Identity Graph Data has the scale and depth to help drive the creation of new platforms and products and provide significant enhancements to existing platforms. By utilizing our extensive Email Address Data and Identity Linkage Data, you can ensure precise consumer identity recognition and verification, making your marketing campaigns more effective and far-reaching.
Contact us at successdelivered@andrewswharton.com or visit us at www.andrewswharton.com to learn more about how our Identity Graph Data, Email Address Data, Identity Linkage Data, and Digital Audience Data can meet your marketing needs.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
BIEN data validation and standardization tools.
https://www.datainsightsmarket.com/privacy-policyhttps://www.datainsightsmarket.com/privacy-policy
The Protein Drug Screening Validation Database market is anticipated to reach a valuation of USD 1,855 million by 2033, expanding at a CAGR of 10.8% during the forecast period of 2025-2033. Growth can be attributed to rising prevalence of chronic diseases, increasing investment in pharmaceutical research and development, and expanding use of molecular diagnostics. Hospitals and clinics represent key end-users, seeking accurate and reliable data for diagnosis and treatment decision-making. Cloud-based solutions are gaining traction due to their flexibility, scalability, and accessibility. Key players in the market include Thermo Fisher Scientific, Clarivate, Bio-Rad Laboratories, PerkinElmer, Elsevier, Eurofins Discovery, Abcam, Creative BioMart, Enamine, and Schrödinger. They compete based on factors such as data quality, platform capabilities, and customer service. Asia Pacific is set to witness the fastest growth, driven by increasing healthcare spending and growing prevalence of chronic diseases in countries like China and India. However, data privacy and security concerns may hinder market growth to some extent.
RampedUp helps marketers that are seeing their efforts generate poorer responses over time and do not understand why. Our experience tells us the reasons are mostly due to the impact of contact data decay. We wrote this article to help them understand why that me be case and this post to help them understand how their marketing data became dirty in the first place.
Validation and Enrichment
RampedUp validates email addresses in real-time and provides up to 60 pieces of detailed information on our contacts. This helps for better segmentation and targeted communication. Here are 10 reasons why people validate and enrich their data.
Personal to Professional
We can find professional information from people that complete online forms with their personal email address. This helps identify and qualify inbound leads. Here are 4 additional reasons to bridge the B2C / B2B gap.
Cleansing
By combining email and contact validation – RampedUp can identify harmful contact records within your database. This will improve inbox placement and deliverability. Here is a blog post on the High Risk of Catch All Email servers.
Recovery
RampedUp can identify the old records in your database and inform you where they are working today. This is a great way to find old customers that have moved to a new company. We wrote this blog post on how to engage old customers and get them back in the fold.
Opt-In Compliance
We can help you identify the contacts within your database that are protected by international Opt-In Laws such as the GDPR, CASL, and CCPA. We wrote this article to share how GDPR is impacting sales and marketing efforts.
The Federal Railroad Administration (FRA) sponsored a study of the work schedules and sleep patterns of railroad employees. The purpose of the study was to understand work-schedule related fatigue that affects various categories of railroad employees by documenting a group's work/rest schedules and sleep patterns to ascertain their impact on the level of fatigue/alertness.Employees surveyed include: signalmen, maintenance of way (MOW) workers, dispatchers, and train & engine service workers (in both freight and passenger train service)
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
This paper drives the process of creating VMLA, a language test meant to be used during awake craniotomies. It focuses on step by step process and aims to help other developers to build their own assessment. This project was designed as a prospective study and registered in the Ethic Committee of Educational and Research Institute of Sirio Libanês Hospital. Ethics committee approval number: HSL 2018-37 / CAEE 90603318.9.0000.5461. Images were bought by Shutterstock.com and generated the following receipts: SSTK-0CA8F-1358 and SSTK-0235F-6FC2 VMLA is a neuropsychological assessment of language function, comprising object naming (ON) and semantic. Originally composed by 420 slides, validation among Brazilian native speakers left 368 figures plus fifteen other elements, like numbers, sentences and count. Validation was focused on educational level (EL), gender and age. Volunteers were tested in fourteen different states of Brazil. Cultural differences resulted in improvements to final Answer Template. EL and age were identified as factors that influenced VLMA assessment results. Highly educated volunteers performed better for both ON and semantic. People over 50 and 35 years old had better performance for ON and semantic, respectively. Further validation in unevaluated regions of Brazil, including more balanced number of males and females and more even distribution of age and EL, could confirm our statistical analysis. After validation, ON-VMLA was framed in batteries of 100 slides each, mixing images of six different complexity categories. Semantic-VMLA kept all the original seventy verbal and non-verbal combinations. The validation process resulted in increased confidence during intraoperative test application. We are now able to score and evaluate patient´s language deficits. Currently, VLMA fits its purpose of dynamical application and accuracy during language areas mapping. It is the first test targeted to Brazilians, representing much of our culture and collective imagery. Our experience may be of value to clinicians and researchers working with awake craniotomy who seek to develop their own language test.
The test is available for free use at www.vemotests.com (beginning in February, 2021)
The ckanext-cprvalidation extension for CKAN is designed to validate resources specifically for the Danish national open data platform. According to the documentation, this extension ensures that datasets adhere to specific standards. It appears to be developed for CKAN v2.6, and the documentation stresses that compatibility with other versions is not ensured. Key Features: Resource Validation: Validates resources against specific criteria, presumably related to or mandated by the Danish national open data platform. The exact validation rules are not detailed in the available documentation. Scheduled Scanning: Can be configured to scan resources at regular intervals via a CRON job, enabling automated and ongoing validation. Exception Handling: Allows adding exceptions to the database, potentially to exclude certain resources or validation errors from triggering alerts or blocking publication. Database Integration: Requires a dedicated database user ("cprvalidation") for operation, with database connection settings added to the CKAN configuration file (production.ini). Technical Integration: The extension installs as a CKAN plugin and requires activation in the CKAN configuration. It necessitates database setup, including the creation of a specific database user and corresponding credentials. The extension likely adds functionality through CKAN's plugin interface and may provide custom CLI commands for database initialization. Scheduled tasks are managed through a CRON job, external to CKAN itself, but triggered to interact with the validation logic. It's also evident that the extension makes use of additional database settings to be configured in the production.ini file. Benefits & Impact: The ckanext-cprvalidation extension ensures data quality and compliance with the standards of the Danish national open data platform. By automating validation and enabling scheduled checks, it reduces the manual effort needed to maintain data integrity, ensuring that published resources meet required standards.
Attribution-NonCommercial-ShareAlike 4.0 (CC BY-NC-SA 4.0)https://creativecommons.org/licenses/by-nc-sa/4.0/
License information was derived automatically
Fichier de validation des données issues du retraitement des actes de décès des habitants de Charleville au XIXe siècle dont le patronyme commençait par la lettre B.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Abstract Many biological collections databases feature data quality problems. On the existing computational resources, we present an import tool and data validation. The program applies filters to data submitted through a spreadsheet at the time of data import, streamlining the error-checking process. The validations presented were divided into three categories according to the taxonomic, geographical and general specimen collection data. Its implementation eliminated the errors in the data entry of new vouchers in the Herbarium of the Botanical Garden of Rio de Janeiro.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
This database is a deliverable of the FLOATECH project, funded under the European Union's Horizon 2020 research and innovation program under grant agreement No 101007142. The aim of the accompanying document is to describe the experimental testing campaign C2 at the LHEEA wave-tank facility. The campaign took place between May and June 2023. The objective of the campaign is to test several FOWT control strategy, including a feed forward wave-based control, using the software-in-the-loop SOFTWIND system. The accompanying report on the E.U. portal describes the database created from these experiments and aimed to be shared for model validation.
https://www.datainsightsmarket.com/privacy-policyhttps://www.datainsightsmarket.com/privacy-policy
The global database testing tool market size was valued at USD 2,504.2 million in 2025 and is projected to reach USD 19,405.8 million by 2033, exhibiting a CAGR of 33.6% during the forecast period. The growth of the market is attributed to the rising demand for ensuring the accuracy and reliability of database systems, increasing adoption of cloud-based database testing tools, and growing need for automated database testing solutions to improve efficiency. Key market trends include the advancement of artificial intelligence (AI) and machine learning (ML) technologies in database testing tools, which enables automated test case generation, data validation, and performance optimization. Additionally, the increasing adoption of agile development methodologies and DevOps practices has led to the demand for continuous database testing tools that can integrate seamlessly with CI/CD pipelines. The market is also witnessing the emergence of database testing tools specifically designed for specific database types, such as NoSQL and NewSQL databases, to meet the unique testing requirements of these systems.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
This work evaluates the use of TGA curve markers (similar to genetic markers) to obtain information on biomass. That is, multiple values of percentage of residual weight with respect to the initial one at specific temperatures of the TGA curve are used to identify specie and wood and leaf mixtures. The data obtained from the TGA analyzes for different mixtures of leaves and wood (100% leaf, 50% leaf and 50% wood and 100% wood) of 7 species are presented (poplar (Populus sp.), caper (Euphorbia laurifolia), alder (Alnus acuminata), arupo (Chionanthus pubescens), cypress (Cupressus macrocarpa), eucalyptus (Eucalyptus globulus), linden (Sambucus nigra L.) and pine (Pinus radiata)) From each of the thermogravimetric curves, the residual weights of the sample at fixed temperatures have been selected (100ºC, 175ºC, 200ºC, 325ºC, 400ºC, 475ºC and 550ºC). These weights represent the input to a neural network to identify both the species and the percentage of leaves and wood. The last sheet of the file shows the assignment of values to each species and the final combination of the mixture of leaves and wood, output of the neural network.
https://www.marketresearchintellect.com/privacy-policyhttps://www.marketresearchintellect.com/privacy-policy
Gain in-depth insights into Database Comparison Software Market Report from Market Research Intellect, valued at USD 1.2 billion in 2024, and projected to grow to USD 2.5 billion by 2033 with a CAGR of 10.5% from 2026 to 2033.
This is a new, open, and transparent database of toxicokinetic data supporting EPA decision making. The database has already become the basis of research efforts within EPA to improve HTTK modeling using generic TK models and has facilitated the creation and validation of models for new exposure routes. Publishing the database supports open, transparent science and this database (the largest public database for this domain) will spur improvement and development of TK models by external experts in the field. Future efforts to improving the accessibility of this database (with a graphical user interface) and encouraging crowdsourcing to expand the size and scope of the database will lead to larger validation sets for our modeling efforts and likely lower uncertainties when estimating TK. This dataset is associated with the following publication: Sayre, R., J. Wambaugh, and C. Grulke. Database of pharmacokinetic time-series data and parameters for 144 environmental chemicals. Scientific Data. Springer Nature Group, New York, NY, 7: 122, (2020).
Bats play crucial ecological roles and provide valuable ecosystem services, yet many populations face serious threats from various ecological disturbances. The North American Bat Monitoring Program (NABat) aims to assess status and trends of bat populations while developing innovative and community-driven conservation solutions using its unique data and technology infrastructure. To support scalability and transparency in the NABat acoustic data pipeline, we developed a fully-automated machine-learning algorithm. This dataset includes audio files of bat echolocation calls that were considered to develop V1.0 of the NABat machine-learning algorithm, however the test set (i.e., holdout dataset) has been excluded from this release. These recordings were collected by various bat monitoring partners across North America using ultrasonic acoustic recorders for stationary acoustic and mobile acoustic surveys. For more information on how these surveys may be conducted, see Chapters 4 and 5 of “A Plan for the North American Bat Monitoring Program” (https://doi.org/10.2737/SRS-GTR-208). These data were then post-processed by bat monitoring partners to remove noise files (or those that do not contain recognizable bat calls) and apply a species label to each file. There is undoubtedly variation in the steps that monitoring partners take to apply a species label, but the steps documented in “A Guide to Processing Bat Acoustic Data for the North American Bat Monitoring Program” (https://doi.org/10.3133/ofr20181068) include first processing with an automated classifier and then manually reviewing to confirm or downgrade the suggested species label. Once a manual ID label was applied, audio files of bat acoustic recordings were submitted to the NABat database in Waveform Audio File format. From these available files in the NABat database, we considered files from 35 classes (34 species and a noise class). Files for 4 species were excluded due to low sample size (Corynorhinus rafinesquii, N=3; Eumops floridanus, N =3; Lasiurus xanthinus, N = 4; Nyctinomops femorosaccus, N =11). From this pool, files were randomly selected until files for each species/grid cell combination were exhausted or the number of recordings reach 1250. The dataset was then randomly split into training, validation, and test sets (i.e., holdout dataset). This data release includes all files considered for training and validation, including files that had been excluded from model development and testing due to low sample size for a given species or because the threshold for species/grid cell combinations had been met. The test set (i.e., holdout dataset) is not included. Audio files are grouped by species, as indicated by the four-letter species code in the name of each folder. Definitions for each four-letter code, including Family, Genus, Species, and Common name, are also included as a dataset in this release.
https://user.eumetsat.int/resources/user-guides/data-registration-and-licensing#ID-Data-Licensinghttps://user.eumetsat.int/resources/user-guides/data-registration-and-licensing#ID-Data-Licensing
The SAVS V1.0 (Surface Albedo Validation Sites) is a static database of potential validation sites. It provides information on the characterization of more than 2000 surface sites with respect of their spatial and temporal homogeneity. Statistical measures are provided enabling the user to easily filter the database using their own criteria. It is a powerful tool for providing a traceable approach to characterize potential sites for EO data validation.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
DC voltage
McGRAW’s US B2B Data: Accurate, Reliable, and Market-Ready
Our B2B database delivers over 80 million verified contacts with 95%+ accuracy. Supported by in-house call centers, social media validation, and market research teams, we ensure that every record is fresh, reliable, and optimized for B2B outreach, lead generation, and advanced market insights.
Our B2B database is one of the most accurate and extensive datasets available, covering over 91 million business executives with a 95%+ accuracy guarantee. Designed for businesses that require the highest quality data, this database provides detailed, validated, and continuously updated information on decision-makers and industry influencers worldwide.
The B2B Database is meticulously curated to meet the needs of businesses seeking precise and actionable data. Our datasets are not only extensive but also rigorously validated and updated to ensure the highest level of accuracy and reliability.
Key Data Attributes:
Unlike many providers that rely solely on third-party vendor files, McGRAW takes a hands-on approach to data validation. Our dedicated nearshore and offshore call centers engage directly with data before each delivery to ensure every record meets our high standards of accuracy and relevance.
In addition, our teams of social media validators, market researchers, and digital marketing specialists continuously refine and update records to maintain data freshness. Each dataset undergoes multiple verification checks using internal validation processes and third-party tools such as Fresh Address, BriteVerify, and Impressionwise to guarantee the highest data quality.
Additional Data Solutions and Services
Data Enhancement: Email and LinkedIn appends, contact discovery across global roles and functions
Business Verification: Real-time validation through call centers, social media, and market research
Technology Insights: Detailed IT infrastructure reports, spending trends, and executive insights
Healthcare Database: Access to over 80 million healthcare professionals and industry leaders
Global Reach: US and international GDPR-compliant datasets, complete with email, postal, and phone contacts
Email Broadcast Services: Full-service campaign execution, from testing to live deployment, with tracking of key engagement metrics such as opens and clicks
Many B2B data providers rely on vendor-contributed files without conducting the rigorous validation necessary to ensure accuracy. This often results in outdated and unreliable data that fails to meet the demands of a fast-moving business environment.
McGRAW takes a different approach. By owning and operating dedicated call centers, we directly verify and validate our data before delivery, ensuring that every record is up-to-date and ready to drive business success.
Through continuous validation, social media verification, and real-time updates, McGRAW provides a high-quality, dependable database for businesses that prioritize data integrity and performance. Our Global Business Executives database is the ideal solution for companies that need accurate, relevant, and market-ready data to fuel their strategies.
Anchialine pools are brackish water systems fed by subsurface groundwater (freshwater) and tides (sea water), with no surface connection to the ocean. Although anchialine pools occur around the world, these habitats tend to be small, isolated, and threatened by human development and introduced nonnative species. These pools provide habitat for rare invertebrate species including shrimp, snails, and odonates.The Pacific Island Network (PACN) is monitoring these unique ecosystems to determine status and trends over time. This will provide managers with information to determine how to best protect these unique habitats and the species they support. This dataset contains pilot data collected during the development of the anchialine pool monitoring protocol. Park anchialine pool monitoring data collected by the resource management staff at Kaloko Honokahau National Historical Park are also included in the database. Each sampling event in the database is associated with a project. Those labeled as I&M Pilot Data are data associated with the I&M Monitoring Protocol. This dataset also includes water quality data imported from the PACN water quality monitoring protocol, the project is I&M Water Quality Monitoring Data. Information on the PACN water quality monitoring protocol can be found at https://irma.nps.gov/App/Reference/Profile/2166407. This data went through a vigorous data certification process, including verification and validation of the data. All others are data collected by park staff or outside researchers. The data verification and validation of this data is unknown.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Background: Epidemiological studies for identifying patients with Parkinson's disease (PD) or Parkinsonism (PKM) have been limited by their nonrandom sampling techniques and mainly veteran populations. This reduces their use for health services planning. The purpose of this study was to validate algorithms for the case ascertainment of PKM from administrative databases using primary care patients as the reference standard. Methods: We conducted a retrospective chart abstraction using a random sample of 73,003 adults aged ≥20 years from a primary care Electronic Medical Record Administrative data Linked Database (EMRALD) in Ontario, Canada. Physician diagnosis in the EMR was used as the reference standard and population-based administrative databases were used to identify patients with PKM from the derivation of algorithms. We calculated algorithm performance using sensitivity, specificity, and predictive values and then determined the population-level prevalence and incidence trends with the most accurate algorithms. Results: We selected, ‘2 physician billing codes in 1 year' as the optimal administrative data algorithm in adults and seniors (≥65 years) due to its sensitivity (70.6-72.3%), specificity (99.9-99.8%), positive predictive value (79.5-82.8%), negative predictive value (99.9-99.7%), and prevalence (0.28-1.20%), respectively. Conclusions: Algorithms using administrative databases can reliably identify patients with PKM with a high degree of accuracy.
Andrew Wharton's US Consumer Email Databases provide over 650 million current and active email address records in our 36-month Production Email Database, and additionally, over 1.4 billion historical records in our Legacy Email Database. These databases offer a comprehensive look-back at the digital and terrestrial identity information associated with a consumer. This Identity Graph Data has been collected from website registrations and is 100% opted-in for Third Party Uses.
The Email Address Data is fully populated with email addresses, HEMS (MD5, Sha1, Sha256), first name, last name, postal address (primary and secondary), IP Address, and Time Stamps for Last Registration, Verification, and First Seen. Additionally, our email address information assets can be linked with our Date-of-Birth and Phone Number databases to provide a powerful solution for consumer identity recognition and verification platforms through Identity Linkage Data.
As an add-on to our current and historical information, we also offer a database of hard-bounce email addresses. These are email addresses that have hard-bounced during our large-scale email campaign deployments or were identified as hard-bounces during our email verification processes. This database provides over 400 million unproductive email addresses useable as a part of suppression or fraud identification applications.
Our Email Information Assets are utilized by major Identity Graph Data and Identity Linkage platforms due to our comprehensive information that links the email address to consumer identity and IP Address information. This Identity Graph Data provides a robust alternative approach when faced with third-party cookie deprecation in the digital ecosystem.
Our digital advertising partners leverage this information to understand where their clients' customers and prospects are online and align media and content with consumer behavior. The additional Email Address Data, mobile phone numbers, and IP Addresses also work to increase the reach of your Digital Audience Data.
This Identity Graph Data has the scale and depth to help drive the creation of new platforms and products and provide significant enhancements to existing platforms. By utilizing our extensive Email Address Data and Identity Linkage Data, you can ensure precise consumer identity recognition and verification, making your marketing campaigns more effective and far-reaching.
Contact us at successdelivered@andrewswharton.com or visit us at www.andrewswharton.com to learn more about how our Identity Graph Data, Email Address Data, Identity Linkage Data, and Digital Audience Data can meet your marketing needs.