Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Monte Carlo integration is a commonly used technique to compute intractable integrals and is typically thought to perform poorly for very high-dimensional integrals. To show that this is not always the case, we examine Monte Carlo integration using techniques from the high-dimensional statistics literature by allowing the dimension of the integral to increase. In doing so, we derive nonasymptotic bounds for the relative and absolute error of the approximation for some general classes of functions through concentration inequalities. We provide concrete examples in which the magnitude of the number of points sampled needed to guarantee a consistent estimate varies between polynomial to exponential, and show that in theory arbitrarily fast or slow rates are possible. This demonstrates that the behavior of Monte Carlo integration in high dimensions is not uniform. Through our methods we also obtain nonasymptotic confidence intervals which are valid regardless of the number of points sampled.
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Example data sets and computer code for the book chapter titled "Missing Data in the Analysis of Multilevel and Dependent Data" submitted for publication in the second edition of "Dependent Data in Social Science Research" (Stemmler et al., 2015). This repository includes the computer code (".R") and the data sets from both example analyses (Examples 1 and 2). The data sets are available in two file formats (binary ".rda" for use in R; plain-text ".dat").
The data sets contain simulated data from 23,376 (Example 1) and 23,072 (Example 2) individuals from 2,000 groups on four variables:
ID = group identifier (1-2000) x = numeric (Level 1) y = numeric (Level 1) w = binary (Level 2)
In all data sets, missing values are coded as "NA".
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Sheet 1 (Raw-Data): The raw data of the study is provided, presenting the tagging results for the used measures described in the paper. For each subject, it includes multiple columns: A. a sequential student ID B an ID that defines a random group label and the notation C. the used notation: user Story or use Cases D. the case they were assigned to: IFA, Sim, or Hos E. the subject's exam grade (total points out of 100). Empty cells mean that the subject did not take the first exam F. a categorical representation of the grade L/M/H, where H is greater or equal to 80, M is between 65 included and 80 excluded, L otherwise G. the total number of classes in the student's conceptual model H. the total number of relationships in the student's conceptual model I. the total number of classes in the expert's conceptual model J. the total number of relationships in the expert's conceptual model K-O. the total number of encountered situations of alignment, wrong representation, system-oriented, omitted, missing (see tagging scheme below) P. the researchers' judgement on how well the derivation process explanation was explained by the student: well explained (a systematic mapping that can be easily reproduced), partially explained (vague indication of the mapping ), or not present.
Tagging scheme:
Aligned (AL) - A concept is represented as a class in both models, either
with the same name or using synonyms or clearly linkable names;
Wrongly represented (WR) - A class in the domain expert model is
incorrectly represented in the student model, either (i) via an attribute,
method, or relationship rather than class, or
(ii) using a generic term (e.g., user'' instead ofurban
planner'');
System-oriented (SO) - A class in CM-Stud that denotes a technical
implementation aspect, e.g., access control. Classes that represent legacy
system or the system under design (portal, simulator) are legitimate;
Omitted (OM) - A class in CM-Expert that does not appear in any way in
CM-Stud;
Missing (MI) - A class in CM-Stud that does not appear in any way in
CM-Expert.
All the calculations and information provided in the following sheets
originate from that raw data.
Sheet 2 (Descriptive-Stats): Shows a summary of statistics from the data collection,
including the number of subjects per case, per notation, per process derivation rigor category, and per exam grade category.
Sheet 3 (Size-Ratio):
The number of classes within the student model divided by the number of classes within the expert model is calculated (describing the size ratio). We provide box plots to allow a visual comparison of the shape of the distribution, its central value, and its variability for each group (by case, notation, process, and exam grade) . The primary focus in this study is on the number of classes. However, we also provided the size ratio for the number of relationships between student and expert model.
Sheet 4 (Overall):
Provides an overview of all subjects regarding the encountered situations, completeness, and correctness, respectively. Correctness is defined as the ratio of classes in a student model that is fully aligned with the classes in the corresponding expert model. It is calculated by dividing the number of aligned concepts (AL) by the sum of the number of aligned concepts (AL), omitted concepts (OM), system-oriented concepts (SO), and wrong representations (WR). Completeness on the other hand, is defined as the ratio of classes in a student model that are correctly or incorrectly represented over the number of classes in the expert model. Completeness is calculated by dividing the sum of aligned concepts (AL) and wrong representations (WR) by the sum of the number of aligned concepts (AL), wrong representations (WR) and omitted concepts (OM). The overview is complemented with general diverging stacked bar charts that illustrate correctness and completeness.
For sheet 4 as well as for the following four sheets, diverging stacked bar
charts are provided to visualize the effect of each of the independent and mediated variables. The charts are based on the relative numbers of encountered situations for each student. In addition, a "Buffer" is calculated witch solely serves the purpose of constructing the diverging stacked bar charts in Excel. Finally, at the bottom of each sheet, the significance (T-test) and effect size (Hedges' g) for both completeness and correctness are provided. Hedges' g was calculated with an online tool: https://www.psychometrica.de/effect_size.html. The independent and moderating variables can be found as follows:
Sheet 5 (By-Notation):
Model correctness and model completeness is compared by notation - UC, US.
Sheet 6 (By-Case):
Model correctness and model completeness is compared by case - SIM, HOS, IFA.
Sheet 7 (By-Process):
Model correctness and model completeness is compared by how well the derivation process is explained - well explained, partially explained, not present.
Sheet 8 (By-Grade):
Model correctness and model completeness is compared by the exam grades, converted to categorical values High, Low , and Medium.
Facebook
Twitterhttps://cubig.ai/store/terms-of-servicehttps://cubig.ai/store/terms-of-service
1) Data Introduction • The Sample Sales Data is a retail sales dataset of 2,823 orders and 25 columns that includes a variety of sales-related data, including order numbers, product information, quantity, unit price, sales, order date, order status, customer and delivery information.
2) Data Utilization (1) Sample Sales Data has characteristics that: • This dataset consists of numerical (sales, quantity, unit price, etc.), categorical (product, country, city, customer name, transaction size, etc.), and date (order date) variables, with missing values in some columns (STATE, ADDRESSLINE2, POSTALCODE, etc.). (2) Sample Sales Data can be used to: • Analysis of sales trends and performance by product: Key variables such as order date, product line, and country can be used to visualize and analyze monthly and yearly sales trends, the proportion of sales by product line, and top sales by country and region. • Segmentation and marketing strategies: Segmentation of customer groups based on customer information, transaction size, and regional data, and use them to design targeted marketing and customized promotion strategies.
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
This is a sample of data from a MAYA numerical relativity simulation, primarily for use with the tutorials for the mayawaves python package. This simulation is of binary black holes with a mass ratio of 5 and precessing spins.
Facebook
TwitterDiscover the ultimate resource for your B2B needs with our meticulously curated dataset, featuring 148MM+ highly relevant US B2B Contact Data records and associated company information.
Very high fill rates for Phone Number, including for Mobile Phone!
This encompasses a diverse range of fields, including Contact Name (First & Last), Work Address, Work Email, Personal Email, Mobile Phone, Direct-Dial Work Phone, Job Title, Job Function, Job Level, LinkedIn URL, Company Name, Domain, Email Domain, HQ Address, Employee Size, Revenue Size, Industry, NAICS and SIC Codes + Descriptions, ensuring you have the most detailed insights for your business endeavors.
Key Features:
Extensive Data Coverage: Access a vast pool of B2B Contact Data records, providing valuable information on where the contacts work now, empowering your sales, marketing, recruiting, and research efforts.
Versatile Applications: Leverage this robust dataset for Sales Prospecting, Lead Generation, Marketing Campaigns, Recruiting initiatives, Identity Resolution, Analytics, Research, and more.
Phone Number Data Inclusion: Benefit from our comprehensive Phone Number Data, ensuring you have direct and effective communication channels. Explore our Phone Number Datasets and Phone Number Databases for an even more enriched experience.
Flexible Pricing Models: Tailor your investment to match your unique business needs, data use-cases, and specific requirements. Choose from targeted lists, CSV enrichment, or licensing our entire database or subsets to seamlessly integrate this data into your products, platform, or service offerings.
Strategic Utilization of B2B Intelligence:
Sales Prospecting: Identify and engage with the right decision-makers to drive your sales initiatives.
Lead Generation: Generate high-quality leads with precise targeting based on specific criteria.
Marketing Campaigns: Amplify your marketing strategies by reaching the right audience with targeted campaigns.
Recruiting: Streamline your recruitment efforts by connecting with qualified candidates.
Identity Resolution: Enhance your data quality and accuracy by resolving identities with our reliable dataset.
Analytics and Research: Fuel your analytics and research endeavors with comprehensive and up-to-date B2B insights.
Access Your Tailored B2B Data Solution:
Reach out to us today to explore flexible pricing options and discover how Salutary Data Company Data, B2B Contact Data, B2B Marketing Data, B2B Email Data, Phone Number Data, Phone Number Datasets, and Phone Number Databases can transform your business strategies. Elevate your decision-making with top-notch B2B intelligence.
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
We present a simple, spreadsheet-based method to determine the statistical significance of the difference between any two arbitrary curves. This modified Chi-squared method addresses two scenarios: A single measurement at each point with known standard deviation, or multiple measurements at each point averaged to produce a mean and standard error. The method includes an essential correction for the deviation from normality in measurements with small sample size, which are typical in biomedical sciences. Statistical significance is determined without regard to the functionality of the curves, or the signs of the differences. Numerical simulations are used to validate the procedure. Example experimental data are used to demonstrate its application. An Excel spreadsheet is provided for performing the calculations for either scenario.
Facebook
TwitterSuccess.ai’s Phone Number Data offers direct access to over 50 million verified phone numbers for professionals worldwide, extracted from our expansive collection of 170 million profiles. This robust dataset includes work emails and key decision-maker profiles, making it an essential resource for companies aiming to enhance their communication strategies and outreach efficiency. Whether you're launching targeted marketing campaigns, setting up sales calls, or conducting market research, our phone number data ensures you're connected to the right professionals at the right time.
Why Choose Success.ai’s Phone Number Data?
Direct Communication: Reach out directly to professionals with verified phone numbers and work emails, ensuring your message gets to the right person without delay. Global Coverage: Our data spans across continents, providing phone numbers for professionals in North America, Europe, APAC, and emerging markets. Continuously Updated: We regularly refresh our dataset to maintain accuracy and relevance, reflecting changes like promotions, company moves, or industry shifts. Comprehensive Data Points:
Verified Phone Numbers: Direct lines and mobile numbers of professionals across various industries. Work Emails: Reliable email addresses to complement phone communications. Professional Profiles: Decision-makers’ profiles including job titles, company details, and industry information. Flexible Delivery and Integration: Success.ai offers this dataset in various formats suitable for seamless integration into your CRM or sales platform. Whether you prefer API access for real-time data retrieval or static files for periodic updates, we tailor the delivery to meet your operational needs.
Competitive Pricing with Best Price Guarantee: We provide this essential data at the most competitive prices in the industry, ensuring you receive the best value for your investment. Our best price guarantee means you can trust that you are getting the highest quality data at the lowest possible cost.
Targeted Applications for Phone Number Data:
Sales and Telemarketing: Enhance your telemarketing campaigns by reaching out directly to potential customers, bypassing gatekeepers. Market Research: Conduct surveys and research directly with industry professionals to gather insights that can shape your business strategy. Event Promotion: Invite prospects to webinars, conferences, and seminars directly through personal calls or SMS. Customer Support: Improve customer service by integrating accurate contact information into your support systems. Quality Assurance and Compliance:
Data Accuracy: Our data is verified for accuracy to ensure over 99% deliverability rates. Compliance: Fully compliant with GDPR and other international data protection regulations, allowing you to use the data with confidence globally. Customization and Support:
Tailored Data Solutions: Customize the data according to geographic, industry-specific, or job role filters to match your unique business needs. Dedicated Support: Our team is on hand to assist with data integration, usage, and any questions you may have. Start with Success.ai Today: Engage with Success.ai to leverage our Phone Number Data and connect with global professionals effectively. Schedule a consultation or request a sample through our dedicated client portal and begin transforming your outreach and communication strategies today.
Remember, with Success.ai, you don’t just buy data; you invest in a partnership that grows with your business needs, backed by our commitment to quality and affordability.
Facebook
TwitterThis GIS layer provides detailed information from Pellegrino and Hubbard (1983). It shows the sample locations and provides a summary of the total number of species found at each station (species_richness).
Facebook
TwitterThe considerations in this report A Skateboard are part of the example collection which can be found in http://www3.math.tu-berlin.de/multiphysics/Examples/. The aim is to investigate different formulations, i.e., regularized formulations or also index reduced formulations, of the model equations in combination with different numerical solvers with respect to its applicability, efficiency, accuracy, and robustness.
Facebook
TwitterThe US Consumer Phone file contains phone numbers, mobile and landline, tied to an individual in the Consumer Database. The fields available include the phone number, phone type, mobile carrier, and Do Not Call registry status.
All phone numbers can be processed and cleansed using telecom carrier data. The telecom data, including phone and texting activity, porting instances, carrier scoring, spam, and known fraud activity, comprise a proprietary Phone Quality Level (PQL), which is a data-science derived score to ensure the highest levels of deliverability at a fraction of the cost compared to competitors.
We have developed this file to be tied to our Consumer Demographics Database so additional demographics can be applied as needed. Each record is ranked by confidence and only the highest quality data is used.
Note - all Consumer packages can include necessary PII (address, email, phone, DOB, etc.) for merging, linking, and activation of the data.
BIGDBM Privacy Policy: https://bigdbm.com/privacy.html
Facebook
Twitterhttps://entrepot.recherche.data.gouv.fr/api/datasets/:persistentId/versions/1.0/customlicense?persistentId=doi:10.57745/FMZ9HPhttps://entrepot.recherche.data.gouv.fr/api/datasets/:persistentId/versions/1.0/customlicense?persistentId=doi:10.57745/FMZ9HP
Although it is a widespread phenomenon in nature, turbulence in fluids (gases, liquids) is still very poorly understood. One area of research involves analyzing data from academic flow simulations. To make progress, the scientific community needs a large amount of reliable data in various configurations. Turbulent flows near solid flat or curved walls are very interesting examples. The database is composed of the 3D raw data (velocity, pressure, time derivative of velocity) and statistics (mean, Reynolds stresses, length scales) of a direct numerical simulations of moderate adverse pressure gradient (decelerating) turbulent boundary layer on a flat plate at Reynolds number up to Reθ = 8000 .
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
To create the dataset, the top 10 countries leading in the incidence of COVID-19 in the world were selected as of October 22, 2020 (on the eve of the second full of pandemics), which are presented in the Global 500 ranking for 2020: USA, India, Brazil, Russia, Spain, France and Mexico. For each of these countries, no more than 10 of the largest transnational corporations included in the Global 500 rating for 2020 and 2019 were selected separately. The arithmetic averages were calculated and the change (increase) in indicators such as profitability and profitability of enterprises, their ranking position (competitiveness), asset value and number of employees. The arithmetic mean values of these indicators for all countries of the sample were found, characterizing the situation in international entrepreneurship as a whole in the context of the COVID-19 crisis in 2020 on the eve of the second wave of the pandemic. The data is collected in a general Microsoft Excel table. Dataset is a unique database that combines COVID-19 statistics and entrepreneurship statistics. The dataset is flexible data that can be supplemented with data from other countries and newer statistics on the COVID-19 pandemic. Due to the fact that the data in the dataset are not ready-made numbers, but formulas, when adding and / or changing the values in the original table at the beginning of the dataset, most of the subsequent tables will be automatically recalculated and the graphs will be updated. This allows the dataset to be used not just as an array of data, but as an analytical tool for automating scientific research on the impact of the COVID-19 pandemic and crisis on international entrepreneurship. The dataset includes not only tabular data, but also charts that provide data visualization. The dataset contains not only actual, but also forecast data on morbidity and mortality from COVID-19 for the period of the second wave of the pandemic in 2020. The forecasts are presented in the form of a normal distribution of predicted values and the probability of their occurrence in practice. This allows for a broad scenario analysis of the impact of the COVID-19 pandemic and crisis on international entrepreneurship, substituting various predicted morbidity and mortality rates in risk assessment tables and obtaining automatically calculated consequences (changes) on the characteristics of international entrepreneurship. It is also possible to substitute the actual values identified in the process and following the results of the second wave of the pandemic to check the reliability of pre-made forecasts and conduct a plan-fact analysis. The dataset contains not only the numerical values of the initial and predicted values of the set of studied indicators, but also their qualitative interpretation, reflecting the presence and level of risks of a pandemic and COVID-19 crisis for international entrepreneurship.
Facebook
TwitterVAPOR is the Visualization and Analysis Platform for Ocean, Atmosphere, and Solar Researchers. VAPOR provides an interactive 3D visualization environment that can also produce animations and still frame images. VAPOR runs on most UNIX and Windows systems equipped with modern 3D graphics cards.
VAPOR is a product of the National Center for Atmospheric Research's Computational and Information Systems Lab. Support for VAPOR is provided by the U.S. National Science Foundation and by the Korea Institute of Science and Technology Information
This dataset contains sample files of model outputs from numerical simulations that VAPOR is capable of directly reading. They are not related to each other aside from being sample data for VAPOR.
To unpack the tar.gz files on Linux/OSX, issue the command tar -xzvf [myFile].tar.gz on the file you've downloaded. On Windows, a program like 7-zip can perform that operation. Once unpacked, the files can be directly imported into VAPOR, or converted to VDC. For more information see the "Getting Data Into VAPOR" Related Link below.
Facebook
TwitterThe considerations in this report Slider Crank are part of the example collection which can be found in http://www3.math.tu-berlin.de/multiphysics/Examples/. The aim is to investigate different formulations, i.e., regularized formulations or also index reduced formulations, of the model equations in combination with different numerical solvers with respect to its applicability, efficiency, accuracy, and robustness.
Facebook
Twitterhttps://creativecommons.org/publicdomain/zero/1.0/https://creativecommons.org/publicdomain/zero/1.0/
The perfection ratio of a number is a concept that is related to perfect numbers and how closely a given number approximates the ideal perfection ratio, which is 2.0.
Perfect Numbers:
A perfect number is a positive integer that is equal to the sum of its proper divisors, excluding the number itself. For example: • 6 is a perfect number because its divisors are 1, 2, and 3, and 1 + 2 + 3 = 6 . • 28 is another perfect number because its divisors are 1, 2, 4, 7, and 14, and 1 + 2 + 4 + 7 + 14 = 28 .
Perfection Ratio:
The perfection ratio of a number n is a measure of how close the sum of its divisors (excluding the number itself) is to the number. It is defined as:
\text{Perfection Ratio} = \frac{\text{Sum of Proper Divisors of } n}{n}
• If the perfection ratio is 2.0, the number is considered perfect.
• If the perfection ratio is greater than 2.0, the number is abundant (i.e., the sum of its proper divisors exceeds the number itself).
• If the perfection ratio is less than 2.0, the number is deficient (i.e., the sum of its proper divisors is less than the number itself).
Examples:
1. Perfect Number Example:
• For n = 6 :
• Proper divisors: 1, 2, 3
• Sum of proper divisors: 1 + 2 + 3 = 6
• Perfection ratio: \frac{6}{6} = 1.0
• Since the perfection ratio is 2.0 for a perfect number, we see the idea of perfect numbers where the sum of divisors divides evenly.
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
The complete dataset used in the analysis comprises 36 samples, each described by 11 numeric features and 1 target. The attributes considered were caspase 3/7 activity, Mitotracker red CMXRos area and intensity (3 h and 24 h incubations with both compounds), Mitosox oxidation (3 h incubation with the referred compounds) and oxidation rate, DCFDA fluorescence (3 h and 24 h incubations with either compound) and oxidation rate, and DQ BSA hydrolysis. The target of each instance corresponds to one of the 9 possible classes (4 samples per class): Control, 6.25, 12.5, 25 and 50 µM for 6-OHDA and 0.03, 0.06, 0.125 and 0.25 µM for rotenone. The dataset is balanced, it does not contain any missing values and data was standardized across features. The small number of samples prevented a full and strong statistical analysis of the results. Nevertheless, it allowed the identification of relevant hidden patterns and trends.
Exploratory data analysis, information gain, hierarchical clustering, and supervised predictive modeling were performed using Orange Data Mining version 3.25.1 [41]. Hierarchical clustering was performed using the Euclidean distance metric and weighted linkage. Cluster maps were plotted to relate the features with higher mutual information (in rows) with instances (in columns), with the color of each cell representing the normalized level of a particular feature in a specific instance. The information is grouped both in rows and in columns by a two-way hierarchical clustering method using the Euclidean distances and average linkage. Stratified cross-validation was used to train the supervised decision tree. A set of preliminary empirical experiments were performed to choose the best parameters for each algorithm, and we verified that, within moderate variations, there were no significant changes in the outcome. The following settings were adopted for the decision tree algorithm: minimum number of samples in leaves: 2; minimum number of samples required to split an internal node: 5; stop splitting when majority reaches: 95%; criterion: gain ratio. The performance of the supervised model was assessed using accuracy, precision, recall, F-measure and area under the ROC curve (AUC) metrics.
Facebook
TwitterThe assignment of a symbolic representation to a specific numerosity is a fundamental requirement for humans solving complex mathematical calculations used in diverse applications such as algebra, accounting, physics, and everyday commerce. Here we show that honeybees are able to learn to match a sign to a numerosity, or a numerosity to a sign and subsequently transfer this knowledge to novel numerosity stimuli changed in colour properties, shape, and configuration. While honeybees learnt the associations between two quantities (two; three) and two signs (N-shape; inverted T-shape), they failed at reversing their specific task of sign-to-numerosity-matching to numerosity-to-sign-matching and vice-versa (i.e. a honeybee that learnt to match a sign to a number of elements was not able to invert this learning to match the numerosity of elements to a sign). Thus, while bees could learn the association between a symbol and numerosity, it was linked to the specific task and bees could not spo...
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
This repository includes the data associated to the numerical examples in the 'Applications of an isogeometric indirect boundary element method and the importance of accurate geometrical representation in acoustic problems' article submitted to the Engineering Analysis with Boundary Elements Elsevier journal. The data comprise of the CAD models, NURBS models to be solved by the isogeometric boundary element method (IGiBEM) with the proposed method in the article, meshes to be solved indirect boundary element method (iBEM) with LMS Virtual.Lab 13.7v, and the results of the numerical examples: 1) vibrating cube, 2) car hood and 3) loudspeaker. The IGiBEM software is under private license of KU Leuven and Siemens Industry Software NV, therefore, it is not shared in this repository.
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Prediction comparison for model based on the Lorenz system.
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Monte Carlo integration is a commonly used technique to compute intractable integrals and is typically thought to perform poorly for very high-dimensional integrals. To show that this is not always the case, we examine Monte Carlo integration using techniques from the high-dimensional statistics literature by allowing the dimension of the integral to increase. In doing so, we derive nonasymptotic bounds for the relative and absolute error of the approximation for some general classes of functions through concentration inequalities. We provide concrete examples in which the magnitude of the number of points sampled needed to guarantee a consistent estimate varies between polynomial to exponential, and show that in theory arbitrarily fast or slow rates are possible. This demonstrates that the behavior of Monte Carlo integration in high dimensions is not uniform. Through our methods we also obtain nonasymptotic confidence intervals which are valid regardless of the number of points sampled.