There's a story behind every dataset and here's your opportunity to share yours.
The COVID-19 Government Measures Dataset puts together all the measures implemented by governments worldwide in response to the Coronavirus pandemic. Data collection includes secondary data review. The researched information available falls into five categories:
Social distancing Movement restrictions Public health measures Social and economic measures Lockdowns
Updated last 10/12/2020 The #COVID19 Government Measures Dataset puts together all the measures implemented by governments worldwide in response to the Coronavirus pandemic. Data collection includes secondary data review. The researched information available falls into five categories: - Social distancing - Movement restrictions - Public health measures - Social and economic measures - Lockdowns Each category is broken down into several types of measures.
ID ISO COUNTRY REGION ADMIN_LEVEL_NAME PCODE LOG_TYPE CATEGORY MEASURE_TYPE TARGETED_POP_GROUP COMMENTS NON_COMPLIANCE DATE_IMPLEMENTED SOURCE SOURCE_TYPE LINK ENTRY_DATE ALTERNATIVE SOURCE
We wouldn't be here without the help of others. If you owe any attributions or thanks, include them here along with any citations of past research.
Your data will be in front of the world's largest data science community. What questions do you want to see answered?
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
In this work we present results of all the major global models and normalise the model results by looking at changes over time relative to a common base year value.
We give an analysis of the variability across the models, both before and after normalisation in order to give insights into variance at national and regional level.
A dataset of harmonised results (based on means) and measures of dispersion is presented, providing a baseline dataset for CBCA validation and analysis.
The dataset is intended as a goto dataset for country and regional results of consumption and production based accounts. The normalised mean for each country/region is the principle result that can be used to assess the magnitude and trend in the emission accounts. However, an additional key element of the dataset are the measures of robustness and spread of the results across the source models. These metrics give insight into the amount of trust should be placed in the individual country/region results.
https://dataverse.no/api/datasets/:persistentId/versions/1.0/customlicense?persistentId=doi:10.18710/FVHTFMhttps://dataverse.no/api/datasets/:persistentId/versions/1.0/customlicense?persistentId=doi:10.18710/FVHTFM
Dataset description This dataset contains background data and supplementary material for Sönning (forthcoming), a study that looks at the behavior of dispersion measures when applied to text-level frequency data. For the literature survey reported in that study, which examines how dispersion measures are used in corpus-based work, it includes tabular files listing the 730 research articles that were examined as well as annotations for those studies that measured dispersion in the corpus-linguistic (and lexicographic) sense. As for the corpus data that were used to train the statistical model parameters underlying the simulation study reported in that paper, the dataset contains a term-document matrix for the 49,604 unique word forms (after conversion to lower-case) that occur in the Brown Corpus. Further, R scripts are included that document in detail how the Brown Corpus XML files, which are available from the Natural Language Toolkit (Bird et al. 2009; https://www.nltk.org/), were processed to produce this data arrangement. Abstract: Related publication This paper offers a survey of recent corpus-based work, which shows that dispersion is typically measured across the text files in a corpus. Systematic insights into the behavior of measures in such distributional settings are currently lacking, however. After a thorough discussion of six prominent indices, we investigate their behavior on relevant frequency distributions, which are designed to mimic actual corpus data. Our evaluation considers different distributional settings, i.e. various combinations of frequency and dispersion values. The primary focus is on the response of measures to relatively high and low sub-frequencies, i.e. texts in which the item or structure of interest is over- or underrepresented (if not absent). We develop a simple method for constructing sensitivity profiles, which allow us to draw instructive comparisons among measures. We observe that these profiles vary considerably across distributional settings. While D and DP appear to show the most balanced response contours, our findings suggest that much work remains to be done to understand the performance of measures on items with normalized frequencies below 100 per million words.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Understanding how epidemics spread in a system is a crucial step to prevent and control outbreaks, with broad implications on the system’s functioning, health, and associated costs. This can be achieved by identifying the elements at higher risk of infection and implementing targeted surveillance and control measures. One important ingredient to consider is the pattern of disease-transmission contacts among the elements, however lack of data or delays in providing updated records may hinder its use, especially for time-varying patterns. Here we explore to what extent it is possible to use past temporal data of a system’s pattern of contacts to predict the risk of infection of its elements during an emerging outbreak, in absence of updated data. We focus on two real-world temporal systems; a livestock displacements trade network among animal holdings, and a network of sexual encounters in high-end prostitution. We define the node’s loyalty as a local measure of its tendency to maintain contacts with the same elements over time, and uncover important non-trivial correlations with the node’s epidemic risk. We show that a risk assessment analysis incorporating this knowledge and based on past structural and temporal pattern properties provides accurate predictions for both systems. Its generalizability is tested by introducing a theoretical model for generating synthetic temporal networks. High accuracy of our predictions is recovered across different settings, while the amount of possible predictions is system-specific. The proposed method can provide crucial information for the setup of targeted intervention strategies.
Open Government Licence - Canada 2.0https://open.canada.ca/en/open-government-licence-canada
License information was derived automatically
The guidance identifies core personal and community-based public health measures to mitigate the transmission of coronavirus disease (COVID-19).
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Statistical data on the number of violators of precautionary and preventive measures to limit the spread of the coronavirus in Qatar, categorized by nationality, gender, and type of crime.
https://dataverse.no/api/datasets/:persistentId/versions/1.0/customlicense?persistentId=doi:10.18710/ATCQZWhttps://dataverse.no/api/datasets/:persistentId/versions/1.0/customlicense?persistentId=doi:10.18710/ATCQZW
This dataset contains frequencies for a set of 150 word forms in the BNC. The set of items was compiled by Biber et al. (2016) for the purpose of analyzing the behavior of dispersion measures in different distributional settings. It was therefore assembled to cover a broad range of frequency and dispersion levels. For each form, the dataset lists (i) the number occurrences in each of the 4049 text files in the BNC, including zero counts; and (ii) the length of each text file, i.e. the number of word tokens it contains.
Attribution-ShareAlike 4.0 (CC BY-SA 4.0)https://creativecommons.org/licenses/by-sa/4.0/
License information was derived automatically
The outbreak of the COVID-19 pandemic has prompted the German government and the 16 German federal states to announce a variety of public health measures in order to suppress the spread of the coronavirus. These non-pharmaceutical measures intended to curb transmission rates by increasing social distancing (i.e., diminishing interpersonal contacts) which restricts a range of individual behaviors. These measures span moderate recommendations such as physical distancing, up to the closures of shops and bans of gatherings and demonstrations. The implementation of these measures are not only a research goal for themselves but have implications for behavioral research conducted in this time (e.g., in form of potential confounder biases). Hence, longitudinal data that represent the measures can be a fruitful data source. The presented data set contains data on 14 governmental measures across the 16 German federal states. In comparison to existing datasets, the data set at hand is a fine-grained daily time series tracking the effective calendar date, introduction, extension, or phase-out of each respective measure. Based on self-regulation theory, measures were coded whether they did not restrict, partially restricted or fully restricted the respective behavioral pattern. The time frame comprises March 08, 2020 until May 15, 2020. The project is an open-source, ongoing project with planned continued updates in regular (approximately monthly) intervals. New variables include restrictions on travel and gastronomy. The variable trvl (travel) comprises the following categories: fully restricted (=2) reflecting a potential general ban to travel within Germany (except for sound reasons like health or business); partially restricted (=1): travels are allowed but may be restricted through prohibition of accommodation or entry ban for certain groups (e.g. people from risk areas); free (=0): no travel and accommodation restrictions in place). The variable gastr (gastronomy) comprises: fully restricted (=2): closure of restaurants or bars; partially restricted (=1): Only take-away or food delivery services are allowed; free (=0): restaurants are allowed to open without restrictions). Further, the variables msk (recommendations to wear a mask) and zoo (restrictions of zoo visits) have been adjusted.:
Our target was to predict gender, age and emotion from audio. We found audio labeled datasets on Mozilla and RAVDESS. So by using R programming language 20 statistical features were extracted and then after adding the labels these datasets were formed. Audio files were collected from "Mozilla Common Voice" and “Ryerson AudioVisual Database of Emotional Speech and Song (RAVDESS)”.
Datasets contains 20 feature columns and 1 column for denoting the label. The 20 statistical features were extracted through the Frequency Spectrum Analysis using R programming Language. They are: 1) meanfreq - The mean frequency (in kHz) is a pitch measure, that assesses the center of the distribution of power across frequencies. 2) sd - The standard deviation of frequency is a statistical measure that describes a dataset’s dispersion relative to its mean and is calculated as the variance’s square root. 3) median - The median frequency (in kHz) is the middle number in the sorted, ascending, or descending list of numbers. 4) Q25 - The first quartile (in kHz), referred to as Q1, is the median of the lower half of the data set. This means that about 25 percent of the data set numbers are below Q1, and about 75 percent are above Q1. 5) Q75 - The third quartile (in kHz), referred to as Q3, is the central point between the median and the highest distributions. 6) IQR - The interquartile range (in kHz) is a measure of statistical dispersion, equal to the difference between 75th and 25th percentiles or between upper and lower quartiles. 7) skew - The skewness is the degree of distortion from the normal distribution. It measures the lack of symmetry in the data distribution. 8) kurt - The kurtosis is a statistical measure that determines how much the tails of distribution vary from the tails of a normal distribution. It is actually the measure of outliers present in the data distribution. 9) sp.ent - The spectral entropy is a measure of signal irregularity that sums up the normalized signal’s spectral power. 10) sfm - The spectral flatness or tonality coefficient, also known as Wiener entropy, is a measure used for digital signal processing to characterize an audio spectrum. Spectral flatness is usually measured in decibels, which, instead of being noise-like, offers a way to calculate how tone-like a sound is. 11) mode - The mode frequency is the most frequently observed value in a data set. 12) centroid - The spectral centroid is a metric used to describe a spectrum in digital signal processing. It means where the spectrum’s center of mass is centered. 13) meanfun - The meanfun is the average of the fundamental frequency measured across the acoustic signal. 14) minfun - The minfun is the minimum fundamental frequency measured across the acoustic signal 15) maxfun - The maxfun is the maximum fundamental frequency measured across the acoustic signal. 16) meandom - The meandom is the average of dominant frequency measured across the acoustic signal. 17) mindom - The mindom is the minimum of dominant frequency measured across the acoustic signal. 18) maxdom - The maxdom is the maximum of dominant frequency measured across the acoustic signal 19) dfrange - The dfrange is the range of dominant frequency measured across the acoustic signal. 20) modindx - the modindx is the modulation index, which calculates the degree of frequency modulation expressed numerically as the ratio of the frequency deviation to the frequency of the modulating signal for a pure tone modulation.
Gender and Age Audio Data Souce: Link: https://commonvoice.mozilla.org/en Emotion Audio Data Souce: Link : https://smartlaboratory.org/ravdess/
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Research data for the purpose of reproducing the results presented in the journal publication titled "Single-shot capable surface acoustic wave dispersion measurement of a layered plate"
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Introduction
There are several works based on Natural Language Processing on newspaper reports. Mining opinions from headlines [ 1 ] using Standford NLP and SVM by Rameshbhaiet. Al.compared several algorithms on a small and large dataset. Rubinet. al., in their paper [ 2 ], created a mechanism to differentiate fake news from real ones by building a set of characteristics of news according to their types. The purpose was to contribute to the low resource data available for training machine learning algorithms. Doumitet. al.in [ 3 ] have implemented LDA, a topic modeling approach to study bias present in online news media.
However, there are not many NLP research invested in studying COVID-19. Most applications include classification of chest X-rays and CT-scans to detect presence of pneumonia in lungs [ 4 ], a consequence of the virus. Other research areas include studying the genome sequence of the virus[ 5 ][ 6 ][ 7 ] and replicating its structure to fight and find a vaccine. This research is crucial in battling the pandemic. The few NLP based research publications are sentiment classification of online tweets by Samuel et el [ 8 ] to understand fear persisting in people due to the virus. Similar work has been done using the LSTM network to classify sentiments from online discussion forums by Jelodaret. al.[ 9 ]. NKK dataset is the first study on a comparatively larger dataset of a newspaper report on COVID-19, which contributed to the virus’s awareness to the best of our knowledge.
2 Data-set Introduction
2.1 Data Collection
We accumulated 1000 online newspaper report from United States of America (USA) on COVID-19. The newspaper includes The Washington Post (USA) and StarTribune (USA). We have named it as “Covid-News-USA-NNK”. We also accumulated 50 online newspaper report from Bangladesh on the issue and named it “Covid-News-BD-NNK”. The newspaper includes The Daily Star (BD) and Prothom Alo (BD). All these newspapers are from the top provider and top read in the respective countries. The collection was done manually by 10 human data-collectors of age group 23- with university degrees. This approach was suitable compared to automation to ensure the news were highly relevant to the subject. The newspaper online sites had dynamic content with advertisements in no particular order. Therefore there were high chances of online scrappers to collect inaccurate news reports. One of the challenges while collecting the data is the requirement of subscription. Each newspaper required $1 per subscriptions. Some criteria in collecting the news reports provided as guideline to the human data-collectors were as follows:
The headline must have one or more words directly or indirectly related to COVID-19.
The content of each news must have 5 or more keywords directly or indirectly related to COVID-19.
The genre of the news can be anything as long as it is relevant to the topic. Political, social, economical genres are to be more prioritized.
Avoid taking duplicate reports.
Maintain a time frame for the above mentioned newspapers.
To collect these data we used a google form for USA and BD. We have two human editor to go through each entry to check any spam or troll entry.
2.2 Data Pre-processing and Statistics
Some pre-processing steps performed on the newspaper report dataset are as follows:
Remove hyperlinks.
Remove non-English alphanumeric characters.
Remove stop words.
Lemmatize text.
While more pre-processing could have been applied, we tried to keep the data as much unchanged as possible since changing sentence structures could result us in valuable information loss. While this was done with help of a script, we also assigned same human collectors to cross check for any presence of the above mentioned criteria.
The primary data statistics of the two dataset are shown in Table 1 and 2.
Table 1: Covid-News-USA-NNK data statistics
No of words per headline
7 to 20
No of words per body content
150 to 2100
Table 2: Covid-News-BD-NNK data statistics No of words per headline
10 to 20
No of words per body content
100 to 1500
2.3 Dataset Repository
We used GitHub as our primary data repository in account name NKK^1. Here, we created two repositories USA-NKK^2 and BD-NNK^3. The dataset is available in both CSV and JSON format. We are regularly updating the CSV files and regenerating JSON using a py script. We provided a python script file for essential operation. We welcome all outside collaboration to enrich the dataset.
3 Literature Review
Natural Language Processing (NLP) deals with text (also known as categorical) data in computer science, utilizing numerous diverse methods like one-hot encoding, word embedding, etc., that transform text to machine language, which can be fed to multiple machine learning and deep learning algorithms.
Some well-known applications of NLP includes fraud detection on online media sites[ 10 ], using authorship attribution in fallback authentication systems[ 11 ], intelligent conversational agents or chatbots[ 12 ] and machine translations used by Google Translate[ 13 ]. While these are all downstream tasks, several exciting developments have been made in the algorithm solely for Natural Language Processing tasks. The two most trending ones are BERT[ 14 ], which uses bidirectional encoder-decoder architecture to create the transformer model, that can do near-perfect classification tasks and next-word predictions for next generations, and GPT-3 models released by OpenAI[ 15 ] that can generate texts almost human-like. However, these are all pre-trained models since they carry huge computation cost. Information Extraction is a generalized concept of retrieving information from a dataset. Information extraction from an image could be retrieving vital feature spaces or targeted portions of an image; information extraction from speech could be retrieving information about names, places, etc[ 16 ]. Information extraction in texts could be identifying named entities and locations or essential data. Topic modeling is a sub-task of NLP and also a process of information extraction. It clusters words and phrases of the same context together into groups. Topic modeling is an unsupervised learning method that gives us a brief idea about a set of text. One commonly used topic modeling is Latent Dirichlet Allocation or LDA[17].
Keyword extraction is a process of information extraction and sub-task of NLP to extract essential words and phrases from a text. TextRank [ 18 ] is an efficient keyword extraction technique that uses graphs to calculate the weight of each word and pick the words with more weight to it.
Word clouds are a great visualization technique to understand the overall ’talk of the topic’. The clustered words give us a quick understanding of the content.
4 Our experiments and Result analysis
We used the wordcloud library^4 to create the word clouds. Figure 1 and 3 presents the word cloud of Covid-News-USA- NNK dataset by month from February to May. From the figures 1,2,3, we can point few information:
In February, both the news paper have talked about China and source of the outbreak.
StarTribune emphasized on Minnesota as the most concerned state. In April, it seemed to have been concerned more.
Both the newspaper talked about the virus impacting the economy, i.e, bank, elections, administrations, markets.
Washington Post discussed global issues more than StarTribune.
StarTribune in February mentioned the first precautionary measurement: wearing masks, and the uncontrollable spread of the virus throughout the nation.
While both the newspaper mentioned the outbreak in China in February, the weight of the spread in the United States are more highlighted through out March till May, displaying the critical impact caused by the virus.
We used a script to extract all numbers related to certain keywords like ’Deaths’, ’Infected’, ’Died’ , ’Infections’, ’Quarantined’, Lock-down’, ’Diagnosed’ etc from the news reports and created a number of cases for both the newspaper. Figure 4 shows the statistics of this series. From this extraction technique, we can observe that April was the peak month for the covid cases as it gradually rose from February. Both the newspaper clearly shows us that the rise in covid cases from February to March was slower than the rise from March to April. This is an important indicator of possible recklessness in preparations to battle the virus. However, the steep fall from April to May also shows the positive response against the attack. We used Vader Sentiment Analysis to extract sentiment of the headlines and the body. On average, the sentiments were from -0.5 to -0.9. Vader Sentiment scale ranges from -1(highly negative to 1(highly positive). There were some cases
where the sentiment scores of the headline and body contradicted each other,i.e., the sentiment of the headline was negative but the sentiment of the body was slightly positive. Overall, sentiment analysis can assist us sort the most concerning (most negative) news from the positive ones, from which we can learn more about the indicators related to COVID-19 and the serious impact caused by it. Moreover, sentiment analysis can also provide us information about how a state or country is reacting to the pandemic. We used PageRank algorithm to extract keywords from headlines as well as the body content. PageRank efficiently highlights important relevant keywords in the text. Some frequently occurring important keywords extracted from both the datasets are: ’China’, Government’, ’Masks’, ’Economy’, ’Crisis’, ’Theft’ , ’Stock market’ , ’Jobs’ , ’Election’, ’Missteps’, ’Health’, ’Response’. Keywords extraction acts as a filter allowing quick searches for indicators in case of locating situations of the economy,
Coronavirus disease 2019 (COVID-19) is a contagious disease caused by severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2). The first known case was identified in Wuhan, China, in December 2019. The disease has since spread worldwide, leading to an ongoing pandemic.
Symptoms of COVID-19 are variable, but often include fever, cough, headache, fatigue, breathing difficulties, and loss of smell and taste. Symptoms may begin one to fourteen days after exposure to the virus. At least a third of people who are infected do not develop noticeable symptoms. Of those people who develop symptoms noticeable enough to be classed as patients, most (81%) develop mild to moderate symptoms (up to mild pneumonia), while 14% develop severe symptoms (dyspnea, hypoxia, or more than 50% lung involvement on imaging), and 5% suffer critical symptoms (respiratory failure, shock, or multiorgan dysfunction). Older people are at a higher risk of developing severe symptoms. Some people continue to experience a range of effects (long COVID) for months after recovery, and damage to organs has been observed. Multi-year studies are underway to further investigate the long-term effects of the disease.
COVID-19 transmits when people breathe in air contaminated by droplets and small airborne particles containing the virus. The risk of breathing these in is highest when people are in close proximity, but they can be inhaled over longer distances, particularly indoors. Transmission can also occur if splashed or sprayed with contaminated fluids in the eyes, nose, or mouth, and, rarely, via contaminated surfaces. People remain contagious for up to 20 days and can spread the virus even if they do not develop symptoms.
Several testing methods have been developed to diagnose the disease. The standard diagnostic method is by detection of the virus' nucleic acid by real-time reverse transcription-polymerase chain reaction (rRT-PCR), transcription-mediated amplification (TMA), or by reverse transcription loop-mediated isothermal amplification (RT-LAMP) from a nasopharyngeal swab.
Preventive measures include physical or social distancing, quarantining, ventilation of indoor spaces, covering coughs and sneezes, hand washing, and keeping unwashed hands away from the face. The use of face masks or coverings has been recommended in public settings to minimize the risk of transmissions.
While work is underway to develop drugs that inhibit the virus (and several vaccines for it have been approved and distributed in various countries, which have since initiated mass vaccination campaigns), the primary treatment is symptomatic. Management involves the treatment of symptoms, supportive care, isolation, and experimental measures.
Source - https://en.wikipedia.org/wiki/COVID-19
This Dataset is a collection of records for COVID-19 (World and Continent wise).
https://i.imgur.com/sbvsXhr.png" alt="">
This Dataset is created from: https://www.worldometers.info/. If you want to learn more, you can visit the Website.
Cover Photo by Hakan Nural on Unsplash
https://cubig.ai/store/terms-of-servicehttps://cubig.ai/store/terms-of-service
1) Data Introduction • The COVID-19 India Containment Zone Classification dataset categorizes Indian districts into Red, Orange, and Green Zones based on COVID-19 case metrics as of May 4. This classification aids in understanding the spread and control of COVID-19 across different regions.
2) Data Utilization (1) COVID-19 India Containment Zone data has characteristics that: • It includes detailed district-level information on the zone classification (Red, Orange, Green) based on COVID-19 metrics. This information is crucial for analyzing the spread of the virus, the effectiveness of containment measures, and for planning public health strategies. (2) COVID-19 India Containment Zone data can be used to: • Public Health Management: Assists in resource allocation, planning containment measures, and implementing targeted lockdowns based on zone classification. • Research and Analysis: Supports epidemiological studies, modeling the spread of the virus, and assessing the impact of containment measures in different zones.
There's a story behind every dataset and here's your opportunity to share yours.
As the spread of the novel covid-19 continues to run into countries it is important for us to keep records of every Information on it. Therefore, this dataset is built basically to cover the update from Africa.
What's inside is more than just rows and columns. Make it easy for others to get started by describing how you acquired the data and what time period it represents, too. It contains Information on the dates the cases were recorded across Africa. Detailing the death, confirmed and recovery cases in each country.
We wouldn't be here without the help of others. If you owe any attributions or thanks, include them here along with any citations of past research.
Ethical AI Club John Hopkins University Runmila Institute WHO CDC Ghana Health Service
Your data will be in front of the world's largest data science community. What questions do you want to see answered? We should be able to see contributors answering questions about how Africa should prepare and put in the right measures to contain the spread. A better understanding from the Data scientists.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
These are the 54 Dynamic Spectrum files (.FITS format) behind the paper "A Scintillation Arc Survey of 22 Pulsars with Low to Moderate Dispersion Measures," submitted to the AAS Journals on 2022 April 06. We also include the Tables, Figures, and Text from the work, as well as some explanatory README files.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
PigDetect: AI-Powered Pig Skin Disease Detection & Monitoring PigDetect is an innovative platform designed to assist farmers in the early detection and management of pig skin diseases. By utilizing cutting-edge image processing and machine learning algorithms, the system allows farmers to capture images of their pigs and instantly receive a diagnosis. PigDetect helps farmers make informed decisions about the health of their livestock, reducing potential outbreaks and increasing farm productivity. ** Key Features:**
[ ] Real-Time
Skin Disease Detection: Upload or capture images of pigs, and PigDetect will analyze them for signs of common skin diseases such as mange, ringworm, and erysipelas.
[ ] Geolocation
PigDetect maps the location of reported cases, helping farmers track disease spread and take preventive measures.
[ ] Notifications:
The system sends real-time notifications to farmers when a skin disease outbreak is detected within a certain radius of their farm. Admin Controls: Farm administrators can manage blog posts, monitor disease data, and interact with PigDetect's community of users.
[ ] PigDetect
provides farmers with a powerful tool to improve the health and welfare of their pigs, ultimately boosting productivity and preventing potential losses.
Synthetic and real dispersion measurements for paths across the Pacific, consists of 2 datasets; SS3DPacific_new - This is a data set of surface-wave dispersion measurements. The dispersion is measured between a synthetic reference seismogram (computed with normal-mode summation using the MINEOS software in the radial model stw105 from Kustowski et al., 2008), and a real observed seismogram. This data set is used by Latallerie et al. (2024) to build a Vs model of the Pacific upper-mantle with full 3D resolution and uncertainty using SOLA inversion (Zaroli 2016) and finite-frequency theory (Zhou 2009). Data are for a set of source-receiver pairs for frequencies ranging from 6 to 21 mHz, every 1mHz. The measurement algorithm uses the multi-taper technique (Thompson 1982). The first 5 Slepians are used (Slepian 1978). A datum is the average of measurements over these tapers, and the uncertainty is the standard deviation. SS3DPacificSyn_new - This is a data set of surface-wave dispersion measurements. The dispersion is measured between a synthetic reference seismogram (computed with normal-mode summation using the MINEOS software in the radial model stw105 from Kustowski et al., 2008), and a synthetic seismogram computed using the spectral element method software Specfem in the 3D model S362ANI from Kustowski etl al. (2018). This data set is used by Latallerie et al. (2024) in a synthetic tomography study to retrieve the Vs structure of the input 3D model S362ANI in the Pacific upper-mantle with full 3D resolution and uncertainty using SOLA inversion (Zaroli 2016) and finite-frequency theory (Zhou 2009). Measurements are provided for source-receiver pairs for frequencies ranging from 6 to 21 mHz, every 1mHz. The measurement algorithm uses the multi-taper technique (Thompson 1982). The first 5 Slepians (Slepian 1978) are used. A datum is the average of measurements over these tapers, and the uncertainty is the standard deviation.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
The Japanese beetle Popillia japonica was introduced on Terceira Island (Azores) early in the 1970s. Mild temperatures, high relative humidity, and heavy rain created the perfect conditions for the beetle's establishment and rapid spread. Despite initial control efforts, the beetle quickly spread to the island's interior agricultural regions and threatened the local plants and horticultural lands. Since 1974, adult populations have been monitored in Terceira Island using pheromone and floral lure traps distributed across the island. The data revealed a distribution pattern across three circular zones with decreasing population densities and a movement of the infestation's central core to the island's interior to more conducive zones for the beetle's development. In 1989, 16 years after the first insects were discovered on the island, the pest had taken over all the available space. A contingency plan was drawn up to establish protective measures to prevent the spread of the Popillia japonica to Madeira and Portugal mainland in 1985 (Decreto Legislativo Regional 11/85/A, de 23 de Agosto). Later, it was actualized to comply with legislation of the European Union (EU), paying particular attention to categorizing this insect as a priority pest. Although these preventive measures were applied, the pest spread to other islands over the years; currently, eight of the nine islands of the Archipelago are infested. Although preventive measures have been applied, the pest has spread to other islands over the years, and currently, eight of the nine islands of the Archipelago are infested. In 1996, the Japanese beetle was detected in Faial; in 2003, on the island of São Miguel; in 2006, in the island of Pico; in 2007, on Flores and São Jorge islands; in 2013, in Corvo; and 2017, in Graciosa. Only Santa Maria has not recorded the pest's presence. The Japanese beetle completes its life cycle in a year, with individuals starting to emerge from the ground at the end of May and reaching their peak densities in early August. The last beetles were seen as late as the end of October. The first and second larval instars typically have a brief lifespan, and by early October, most of the population has reached the third instar. The third instar grubs stop feeding and pupate at the beginning of May. The pupal stage lasts less than a month, and no pupae were seen after late July. Adults eat the foliage, floral parts, and occasionally, the fruits of various agricultural plants and ornamentals. At the same time, the grubs live off the roots of the pastures that make up most of the island. It is important to clarify that the adult beetle pest can damage around 414 host plants belonging to 94 families, which may cause elevated crop damage, which makes this a priority pest to maintain under control. The data presented here is related to the Popillia japonica captured in the Azores from 2008 to 2023, which resulted from the work of the operational services on each island of the Secretaria Regional da Agricultura e Alimentação. It is a compilation of the officials’ records from the local authorities who contributed to this data from their fieldwork monitoring of Popillia japonica during these 16 years
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
This is a spectral dataset of natural objects and daylights collected in Japan.
We collected 359 natural objects and measured the reflectance of all objects and the transmittance of 75 leaves. We also measured daylights from dawn till dusk on four different days using a white plate placed (i) under the direct sun and (ii) under the casted shadow (in total 359 measurements). We also separately measured daylights at five different locations (including a sports ground, a space between tall buildings and a forest) with minimum time intervals to reveal the influence of surrounding environments on the spectral composition of daylights reaching the ground (in total 118 measurements).
If you use this dataset in your research, please cite the following publication.
Dataset contains following Excel spread sheets and csv files:
(A) Surface properties of natural objects
(A-1) Reflectance_ver1-2.xlsx and .csv
(A-2) Transmittance_FrontSideUp_ver1-2.xlsx and .csv
(A-2) Transmittance_BackSideUp_ver1-2.xlsx and .csv
(B) Daylight measurements
(B-1) Daylight_TimeLapse_v1-2.xlsx and .csv
(B-2) Daylight_DifferentLocations_v1-2.xlsx and .csv
Data description
(A) Surface properties
(A-1) Reflectance_ver1-2.xlsx and .csv
This file contains surface spectral reflectance data (380 - 780 nm, 5 nm step) of 359 natural objects, including 200 flowers, 113 leaves, 23 fruits, 6 vegetables, 8 barks, and 9 stones measured by a spectrophotometer (SR-2A, Topcon, Tokyo, Japan). Photos of all samples are included in the .xlsx file.
For the analysis presented in the paper, we identified reflectance pairs that have a Pearson’s correlation coefficient across 401 spectral channels of more than 0.999 and removed one of reflectances from each pair. The column 'Used in analysis' indicates whether or not each sample is used for the analysis (TRUE indicates used and FALSE indicate not used).
At the time of collection, we noted the scientific names of flowers, leaves and barks from a name board provided by the Tokyo Institute of Technology in which samples are collected. If not available, we used a smartphone software which automatically identifies the scientific name from an input image (PictureThis - Plant Identifier developed by Glority Global Group Ltd.). The names of 2 flowers and 9 stones whose name could not be identified through either method were left blank.
(A-2) Transmittance_FrontSideUp_v1-2.xlsx and .csv
This file contains surface spectral transmittance data (380 - 780 nm, 5 nm step) for 75 leaves measured by a spectrophotometer (SR-2A, Topcon, Tokyo, Japan). Photos of all samples are included in the .xlsx file.
For this data, the transmittance was measured with the front-side of leaves up (the light was transmitted from the back side of the leaves). This is the data presented in the associated article.
(A-3) Transmittance_BackSideUp_v1-2.xlsx and .csv
Spectral transmittance data of the same leaves presented in (A-2).
For this data, the transmittance was measured with the back-side of leaves up (the light was transmitted from the front side of the leaves).
(B) Daylight measurements
(B-1) Daylight_TimeLapse_ver1-2.xlsx and .csv
This file contains daylight spectra from sunrise to sunset on four different days (2013/11/20, 2013/12/24, 2014/07/03 and 2014/10/27) measured by a spectrophotometer (SR-LEDW, Topcon, Tokyo, Japan) with a wavelength range from 380 nm to 780 nm with 1 nm step. We measured the reflected light from the white calibration plate placed either under a direct sunlight or under a casted shadow.
The column 'Cloud cover' provides visual estimate of percentage of cloud cover across the sky at the time of each measurement. The column 'Red lamp' indicates whether an aircraft warning lamp at the measurement site was on (circle) or off (blank).
(B-2) Daylight_DifferentLocations_ver1-2.xlsx and .csv
This file includes daylight spectra measured at five different sites within the Suzukakedai Campus of Tokyo Institute of Technology with minimum time gap on 2014/07/08, using a spectroradiometer (IM-1000, Topcon) from 380 nm to 780 nm with 1 nm step. The instrument was oriented either towards the sun or towards the zenith sky. When the instrument was oriented to the sun, we measured spectra in two ways: (i) one using a black cylinder covering the photodetector and (ii) the other without using a cylinder.
The column 'Cylinder' indicates whether the black cylinder was used (circle) or not (cross). The column 'Cloud cover' shows the visual estimate of percentage of cloud cover at the time of each measurement. The column 'Sun hidden in clouds' denotes whether the measurement was taken when the sun was covered by clouds (circle) or not (blank).
Lucror Analytics: Proprietary Risk Modelling Data for Credit Quality & Bond Valuation
At Lucror Analytics, we provide cutting-edge corporate data solutions tailored to fixed income professionals and organizations implementing quant-driven strategies. Our risk modelling data encompasses issuer and issue-level credit quality, bond fair value metrics, and proprietary scores designed to offer nuanced, actionable insights into global bond markets that help you stay ahead of the curve. Covering over 3,300 global issuers and over 80,000 bonds, we empower our clients with robust risk modelling data to make data-driven decisions with confidence and precision.
By leveraging our proprietary C-Score, V-Score, and V-Score I models, which utilize CDS and OAS data, we provide unparalleled granularity in credit analysis and valuation. Whether you are a portfolio manager, credit analyst, or institutional investor, Lucror’s risk modelling data solutions deliver actionable insights to enhance strategies, identify mispricing opportunities, and assess credit risk.
What Makes Lucror’s Risk Modelling Data Unique?
Proprietary Credit and Valuation Models Developed for Risk Modelling Our proprietary C-Score, V-Score, and V-Score I are designed to provide a deeper understanding of credit quality and bond valuation:
C-Score: A composite score (0-100) reflecting an issuer's credit quality based on market pricing signals such as CDS spreads. Responsive to near-real-time market changes, the C-Score offers granular differentiation within and across credit rating categories, helping investors identify mispricing opportunities.
V-Score: Measures the deviation of an issue’s option-adjusted spread (OAS) from the market fair value, indicating whether a bond is overvalued or undervalued relative to the market.
V-Score I: Similar to the V-Score but benchmarked against industry-specific fair value OAS, offering insights into relative valuation within an industry context.
These models provide foundational risk modelling data for fixed income strategies aimed at outperforming benchmarks.
Comprehensive Global Coverage Our risk modelling data covers over 3,300 issuers and 80,000 bonds across global markets, ensuring 90%+ overlap with prominent IG and HY benchmark indices. This extensive coverage provides valuable insights into issuers across sectors and geographies, enabling users to analyze issuer credit risk and market dynamics comprehensively.
Risk Modelling Data Customization and Flexibility We recognize that different users have unique requirements. Lucror Analytics offers tailored datasets delivered in customizable formats, frequencies, and levels of granularity, ensuring that our risk modelling data integrates seamlessly into your workflows.
High-Frequency, High-Quality Risk Modeling Data Our C-Score, V-Score, and V-Score I models and metrics are updated daily using end-of-day (EOD) data from S&P. This ensures that users have access to current and accurate risk modelling data, empowering timely and informed decision-making.
How Is the Risk Modelling Data for Sourced? Lucror Analytics employs a rigorous methodology to source, structure, transform and process data, ensuring reliability and actionable insights:
Proprietary Quantitative Risk Models: Our scores are derived from proprietary quant algorithms based on CDS spreads, OAS, and other issuer and bond data.
Global Data Partnerships: Our collaborations with S&P and other reputable data providers ensure comprehensive and accurate risk modelling datasets.
Cleaning and Structuring of Risk Modelling Data: Advanced processes ensure data integrity, transforming raw inputs into actionable credit risk insights.
Primary Use Cases
Quant-driven Portfolio Construction & Rebalancing Lucror’s C-Score provides a granular view of issuer credit quality, allowing portfolio managers to evaluate risks and identify mispricing opportunities. With CDS-driven insights and daily updates, clients can incorporate near-real-time issuer/bond movements into their credit assessments using risk modelling data.
Portfolio Optimization The V-Score and V-Score I allow portfolio managers to identify undervalued or overvalued bonds, supporting strategies that optimize returns relative to credit risk. By benchmarking valuations against market and industry standards, users can uncover potential mean-reversion opportunities and enhance portfolio performance with risk modelling data.
Strategic Decision-Making Our comprehensive risk modelling data enables financial institutions to make informed strategic decisions. Whether it’s assessing the fair value of bonds, analyzing industry-specific credit risk, or underst...
There's a story behind every dataset and here's your opportunity to share yours.
The COVID-19 Government Measures Dataset puts together all the measures implemented by governments worldwide in response to the Coronavirus pandemic. Data collection includes secondary data review. The researched information available falls into five categories:
Social distancing Movement restrictions Public health measures Social and economic measures Lockdowns
Updated last 10/12/2020 The #COVID19 Government Measures Dataset puts together all the measures implemented by governments worldwide in response to the Coronavirus pandemic. Data collection includes secondary data review. The researched information available falls into five categories: - Social distancing - Movement restrictions - Public health measures - Social and economic measures - Lockdowns Each category is broken down into several types of measures.
ID ISO COUNTRY REGION ADMIN_LEVEL_NAME PCODE LOG_TYPE CATEGORY MEASURE_TYPE TARGETED_POP_GROUP COMMENTS NON_COMPLIANCE DATE_IMPLEMENTED SOURCE SOURCE_TYPE LINK ENTRY_DATE ALTERNATIVE SOURCE
We wouldn't be here without the help of others. If you owe any attributions or thanks, include them here along with any citations of past research.
Your data will be in front of the world's largest data science community. What questions do you want to see answered?