Apache License, v2.0https://www.apache.org/licenses/LICENSE-2.0
License information was derived automatically
This dataset was created by TheViewlytics
Released under Apache 2.0
MIT Licensehttps://opensource.org/licenses/MIT
License information was derived automatically
YouTube Music Hits Dataset
A collection of YouTube music video data sourced from Wikidata, focusing on videos with significant viewership metrics.
Dataset Description
Overview
24,329 music videos View range: 1M to 5.5B views Temporal range: 1977-2024
Features
youtubeId: YouTube video identifier itemLabel: Video/song title performerLabel: Artist/band name youtubeViews: View count year: Release year genreLabel: Musical genre(s)
View… See the full description on the dataset page: https://huggingface.co/datasets/akbargherbal/youtube-music-hits.
The global number of Youtube users in was forecast to continuously increase between 2024 and 2029 by in total 232.5 million users (+24.91 percent). After the ninth consecutive increasing year, the Youtube user base is estimated to reach 1.2 billion users and therefore a new peak in 2029. Notably, the number of Youtube users of was continuously increasing over the past years.User figures, shown here regarding the platform youtube, have been estimated by taking into account company filings or press material, secondary research, app downloads and traffic data. They refer to the average monthly active users over the period.The shown data are an excerpt of Statista's Key Market Indicators (KMI). The KMI are a collection of primary and secondary indicators on the macro-economic, demographic and technological environment in up to 150 countries and regions worldwide. All indicators are sourced from international and national statistical offices, trade associations and the trade press and they are processed to generate comparable data sets (see supplementary notes under details for more information).Find more key insights for the number of Youtube users in countries like Africa and South America.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
News dissemination plays a vital role in supporting people to incorporate beneficial actions during public health emergencies, thereby significantly reducing the adverse influences of events. Based on big data from YouTube, this research study takes the declaration of COVID-19 National Public Health Emergency (PHE) as the event impact and employs a DiD model to investigate the effect of PHE on the news dissemination strength of relevant videos. The study findings indicate that the views, comments, and likes on relevant videos significantly increased during the COVID-19 public health emergency. Moreover, the public’s response to PHE has been rapid, with the highest growth in comments and views on videos observed within the first week of the public health emergency, followed by a gradual decline and returning to normal levels within four weeks. In addition, during the COVID-19 public health emergency, in the context of different types of media, lifestyle bloggers, local media, and institutional media demonstrated higher growth in the news dissemination strength of relevant videos as compared to news & political bloggers, foreign media, and personal media, respectively. Further, the audience attracted by related news tends to display a certain level of stickiness, therefore this audience may subscribe to these channels during public health emergencies, which confirms the incentive mechanisms of social media platforms to foster relevant news dissemination during public health emergencies. The proposed findings provide essential insights into effective news dissemination in potential future public health events.
Summary: This dataset encapsulates the vibrant community interaction surrounding MrBeast's most viewed YouTube video, "$456,000 Squid Game In Real Life!" By meticulously compiling 100,000 comments, this collection offers a unique window into the public discourse, engagement, and sentiments of one of YouTube's most significant viral phenomena. Researchers, linguists, marketers, and sociocultural analysts can delve into the depth and diversity of viewer conversations, drawing insights and patterns from a broad swath of public opinion. Privacy Preservation: Recognizing the importance of privacy and ethical data handling, we've employed SHA-256 encryption to anonymize user names. Each username is first salted with a unique sequence and then passed through the SHA-256 algorithm, turning it into a distinct and irreversible hash. This approach ensures the confidentiality of individuals while retaining the integrity and utility of the data for analysis. Dataset Usage: This dataset is a rich resource for various analytical pursuits, including but not limited to sentiment analysis, linguistic trends, engagement patterns, and sociocultural research. Academics can explore the dynamics of viral content and public interaction, while marketers and content creators can gauge audience response and engagement strategies. The dataset is also invaluable for training machine learning models in natural language processing and understanding, providing real-world text data in a diverse, dynamic context. Acknowledgements: We extend our gratitude to YouTube and MrBeast for the platform and content that made this dataset possible. Their commitment to fostering a robust and engaging digital community is at the heart of this collection. We also acknowledge the YouTube API for enabling access to this rich vein of public discourse, and we commend the countless viewers whose comments have collectively shaped this dataset into a mirror of public sentiment and interaction. Column Descriptors:
Comment: The full text of the user's comment, reflecting their thoughts, reactions, or interactions with the content. Anonymized Author: The SHA-256 hashed representation of the user's name, ensuring anonymity while maintaining a unique identifier for each commenter. Published At: The ISO 8601 timestamp marking when the comment was originally posted, offering insights into the timing and relevance of user interactions. Likes: The number of likes attributed to the comment, serving as an indicator of its resonance or approval among the community. Reply Count: The count of replies to the comment, reflecting its capacity to engage and provoke discussion within the community.
CC-BY-NC
Original Data Source: Mr Beast: Most Viewed YT Video 100K Comments
This dataset provides estimated YouTube RPM (Revenue Per Mille) ranges for different niches in 2025, based on ad revenue earned per 1,000 monetized views.
Attribution-ShareAlike 4.0 (CC BY-SA 4.0)https://creativecommons.org/licenses/by-sa/4.0/
License information was derived automatically
This dataset is for object detection task of the blue-ringed octopus (one of the most venomous animals in the world). With this dataset, I hope people can become more familiar with the blue ringed octopus and be aware of its dangers
I collected the images and labeled them myself (for a competition). I have less experience in collecting datasets, so I cannot guarantee the quality of this dataset. I trained a yolov7 object detection model with this data and got a mean average precision of 0.987 (with an IoU threshold of 0.5) .
I didn't go into the field to take these images, instead I took them from Google, some also from screenshots of some Youtube videos: - https://www.youtube.com/watch?v=MBHjo6UaHzk&t=62s - https://www.youtube.com/watch?v=c4BoYORmgSM - https://www.youtube.com/watch?v=DSdq8XFQdKo - https://www.youtube.com/watch?v=64mY1klkf4I&t=215s - https://www.youtube.com/watch?v=C0DOusbGWbU - https://www.youtube.com/watch?v=mTnmw5o4vRI - https://www.youtube.com/watch?v=bejKAB2Eazw&t=317s - https://www.youtube.com/watch?v=emisZUHJAEA - https://www.youtube.com/watch?v=6b_UYwyWI6E - https://www.youtube.com/watch?v=vVamzP52qwA - https://www.youtube.com/watch?v=3Bt1LvpZ1Oo
I also played around with the ai text to image generator to create multiple images and manually choose which one is acceptable (r_blue_ringed_octopus_100
-
r_blue_ringed_octopus_110
, you can remove it if you want). After collecting the images, I do the labeling my self.
Attribution-NonCommercial-ShareAlike 4.0 (CC BY-NC-SA 4.0)https://creativecommons.org/licenses/by-nc-sa/4.0/
License information was derived automatically
Description
The Ryerson Audio-Visual Database of Emotional Speech and Song (RAVDESS) contains 7356 files (total size: 24.8 GB). The dataset contains 24 professional actors (12 female, 12 male), vocalizing two lexically-matched statements in a neutral North American accent. Speech includes calm, happy, sad, angry, fearful, surprise, and disgust expressions, and song contains calm, happy, sad, angry, and fearful emotions. Each expression is produced at two levels of emotional intensity (normal, strong), with an additional neutral expression. All conditions are available in three modality formats: Audio-only (16bit, 48kHz .wav), Audio-Video (720p H.264, AAC 48kHz, .mp4), and Video-only (no sound). Note, there are no song files for Actor_18.
The RAVDESS was developed by Dr Steven R. Livingstone, who now leads the Affective Data Science Lab, and Dr Frank A. Russo who leads the SMART Lab.
Citing the RAVDESS
The RAVDESS is released under a Creative Commons Attribution license, so please cite the RAVDESS if it is used in your work in any form. Published academic papers should use the academic paper citation for our PLoS1 paper. Personal works, such as machine learning projects/blog posts, should provide a URL to this Zenodo page, though a reference to our PLoS1 paper would also be appreciated.
Academic paper citation
Livingstone SR, Russo FA (2018) The Ryerson Audio-Visual Database of Emotional Speech and Song (RAVDESS): A dynamic, multimodal set of facial and vocal expressions in North American English. PLoS ONE 13(5): e0196391. https://doi.org/10.1371/journal.pone.0196391.
Personal use citation
Include a link to this Zenodo page - https://zenodo.org/record/1188976
Commercial Licenses
Commercial licenses for the RAVDESS can be purchased. For more information, please visit our license page of fees, or contact us at ravdess@gmail.com.
Contact Information
If you would like further information about the RAVDESS, to purchase a commercial license, or if you experience any issues downloading files, please contact us at ravdess@gmail.com.
Example Videos
Watch a sample of the RAVDESS speech and song videos.
Emotion Classification Users
If you're interested in using machine learning to classify emotional expressions with the RAVDESS, please see our new RAVDESS Facial Landmark Tracking data set [Zenodo project page].
Construction and Validation
Full details on the construction and perceptual validation of the RAVDESS are described in our PLoS ONE paper - https://doi.org/10.1371/journal.pone.0196391.
The RAVDESS contains 7356 files. Each file was rated 10 times on emotional validity, intensity, and genuineness. Ratings were provided by 247 individuals who were characteristic of untrained adult research participants from North America. A further set of 72 participants provided test-retest data. High levels of emotional validity, interrater reliability, and test-retest intrarater reliability were reported. Validation data is open-access, and can be downloaded along with our paper from PLoS ONE.
Contents
Audio-only files
Audio-only files of all actors (01-24) are available as two separate zip files (~200 MB each):
Audio-Visual and Video-only files
Video files are provided as separate zip downloads for each actor (01-24, ~500 MB each), and are split into separate speech and song downloads:
File Summary
In total, the RAVDESS collection includes 7356 files (2880+2024+1440+1012 files).
File naming convention
Each of the 7356 RAVDESS files has a unique filename. The filename consists of a 7-part numerical identifier (e.g., 02-01-06-01-02-01-12.mp4). These identifiers define the stimulus characteristics:
Filename identifiers
Filename example: 02-01-06-01-02-01-12.mp4
License information
The RAVDESS is released under a Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License, CC BY-NC-SA 4.0
Commercial licenses for the RAVDESS can also be purchased. For more information, please visit our license fee page, or contact us at ravdess@gmail.com.
Related Data sets
We learn high fidelity human depths by leveraging a collection of social media dance videos scraped from the TikTok mobile social networking application. It is by far one of the most popular video sharing applications across generations, which include short videos (10-15 seconds) of diverse dance challenges as shown above. We manually find more than 300 dance videos that capture a single person performing dance moves from TikTok dance challenge compilations for each month, variety, type of dances, which are moderate movements that do not generate excessive motion blur. For each video, we extract RGB images at 30 frame per second, resulting in more than 100K images. We segmented these images using Removebg application, and computed the UV coordinates from DensePose.
Download TikTok Dataset:
Please use the dataset only for the research purpose.
The dataset can be viewed and downloaded from the Kaggle page. (you need to make an account in Kaggle to be able to download the data. It is free!)
The dataset can also be downloaded from here (42 GB). The dataset resolution is: (1080 x 604)
The original YouTube videos corresponding to each sequence and the dance name can be downloaded from here (2.6 GB).
https://spdx.org/licenses/CC0-1.0.htmlhttps://spdx.org/licenses/CC0-1.0.html
Animal-related content on social media is hugely popular but is not always appropriate in terms of how animals are portrayed or how they are treated. This has potential implications beyond the individual animals involved, for viewers, for wild animal populations, and for societies and their interactions with animals. Whilst social media platforms usually publish guidelines for permitted content, enforcement relies at least in part on viewers reporting inappropriate posts. Currently, there is no external regulation of social media platforms. Based on a set of 241 "fake animal rescue" videos that exhibited clear signs of animal cruelty and strong evidence of being deliberately staged (i.e. fake), we found little evidence that viewers disliked the videos and an overall mixed response in terms of awareness of the fake nature of the videos, and their attitudes towards the welfare of the animals involved. Our findings suggest, firstly, that despite the narrowly defined nature of the videos used in this case study, exposure rates can be extremely high (one of the videos had been viewed over 100 million times), and, secondly, that many YouTube viewers cannot identify (or are not concerned by) animal welfare or conservation issues within a social media context. In terms of the current policy approach of social media platforms, our findings raise questions regarding the value of their current reliance on consumers as watch dogs.
Methods
Data collection
The dataset pertains to 241YouTube videos identified using the search function in YouTube and the search terms "primitive man saves" and "primitive boy saves" between May and July 2021; supplemented with additional similar videos held in a database collated by Animals for Asia (www.asiaforanimals.com). Video metrics were extracted automatically between 24.06.21 and 02.08.21 using the "tuber" package in R (Sood 2020, https://cran.r-project.org/web/packages/tuber/tuber.pdf ). Additional information (e.g. on animal taxa) was obtained manually by screening the videos. For five of the videos that received > 1,000 comments, comment text was also extracted using the tuber package. Only publicly available videos were accessed.
Data processing
Users (video posters and commenters) have been de-identified. For each video for which comment text was analysed, the text was converted into a list of the most frequently used words and emojis. Please refer to the manuscript for further details on the methods and approach used to identify and define the most frequently used words/emojis, and to assign sentiment scores.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Update — December 7, 2014. – Evidence-based medicine (EBM) is not working for many reasons, for example: 1. Incorrect in their foundations (paradox): hierarchical levels of evidence are supported by opinions (i.e., lowest strength of evidence according to EBM) instead of real data collected from different types of study designs (i.e., evidence). http://dx.doi.org/10.6084/m9.figshare.1122534 2. The effect of criminal practices by pharmaceutical companies is only possible because of the complicity of others: healthcare systems, professional associations, governmental and academic institutions. Pharmaceutical companies also corrupt at the personal level, politicians and political parties are on their payroll, medical professionals seduced by different types of gifts in exchange of prescriptions (i.e., bribery) which very likely results in patients not receiving the proper treatment for their disease, many times there is no such thing: healthy persons not needing pharmacological treatments of any kind are constantly misdiagnosed and treated with unnecessary drugs. Some medical professionals are converted in K.O.L. which is only a puppet appearing on stage to spread lies to their peers, a person supposedly trained to improve the well-being of others, now deceits on behalf of pharmaceutical companies. Probably the saddest thing is that many honest doctors are being misled by these lies created by the rules of pharmaceutical marketing instead of scientific, medical, and ethical principles. Interpretation of EBM in this context was not anticipated by their creators. “The main reason we take so many drugs is that drug companies don’t sell drugs, they sell lies about drugs.” ―Peter C. Gøtzsche “doctors and their organisations should recognise that it is unethical to receive money that has been earned in part through crimes that have harmed those people whose interests doctors are expected to take care of. Many crimes would be impossible to carry out if doctors weren’t willing to participate in them.” —Peter C Gøtzsche, The BMJ, 2012, Big pharma often commits corporate crime, and this must be stopped. Pending (Colombia): Health Promoter Entities (In Spanish: EPS ―Empresas Promotoras de Salud).
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
This dataset is designed for hospital assistive robots to detect people and their mobility aids Sources: A lot of the images in the dataset has been token from this paper dataset : [1] A. Vasquez, M. Kollmitz, A. Eitel, and W. Burgard, “Deep detection of people and their mobility aids for a hospital robot,” in 2017 European Conference on Mobile Robots (ECMR), pp. 1–7, IEEE, 2017. and some others collected from YouTube videos: [2] https://www.youtube.com/watch?v=f-1e-m7WnVo [3] https://www.youtube.com/watch?v=m4VzCiTvQlA [4] https://www.youtube.com/watch?v=QG8zq90zi0A [5] https://www.youtube.com/watch?v=mYNEpGDZJjc [6] https://www.youtube.com/watch?v=X_KMEJ7rYLc [7] https://www.youtube.com/watch?v=2nW78NxWpag [8] https://www.youtube.com/watch?v=ghudpH_d5Io
Attribution-NonCommercial 4.0 (CC BY-NC 4.0)https://creativecommons.org/licenses/by-nc/4.0/
License information was derived automatically
https://www.youtube.com/watch?v=21qqFM-XPVU" alt="YT Video">
The dataset includes the following columns:
rank
→ The global ranking of the player (e.g., #1
, #2
, #3
).user link
→ Archived URL to the player's osu! profile.country
→ The nationality of the player.ranking-page-table_user-link href 2
→ Another archived profile link (potential duplicate).players name
→ The username of the player.performance points
→ The pp (performance points), which measure player skill in osu!.Here’s a preview of the dataset:
Rank | Player | Country | Performance Points |
---|---|---|---|
#1 | FlyingTuna | 🇰🇷 South Korea | 15,321 pp |
#2 | idke | 🇺🇸 USA | 15,297 pp |
#3 | Mathi | 🇩🇪 Germany | 14,994 pp |
#4 | Freddie Benson | 🇺🇸 USA | 14,977 pp |
#5 | nathan on osu | 🇦🇺 Australia | 14,690 pp |
Attribution 1.0 (CC BY 1.0)https://creativecommons.org/licenses/by/1.0/
License information was derived automatically
This dataset represents a 4.7 km long tour (odometry path shown in htwddKogRob-TSDReal_path.png) in an environment whose representation (see map htwddKogRob-TSDReal.png | 1px (\widehat{=}) 0.1m) is now obsolete. Several static objects have been moved or removed, and there are varying numbers of dynamic obstacles (people walking around).
The work was first presented in:
A Fuzzy-based Adaptive Environment Model for Indoor Robot Localization
Authors: Frank Bahrmann, Sven Hellbach, Hans-Joachim Böhme
Date of Publication: 2016/10/6
Conference: Telehealth and Assistive Technology / 847: Intelligent Systems and Robotics
Publisher: ACTA Press
Additionally, we present a video with the proposed algorithm and an insight of this dataset under:
youtube.com/AugustDerSmarte
https://www.youtube.com/watch?v=26NBFN_XeQg
Instructions for use
The zip archive contains ascii files, which hold the log files of the robot observations and robot poses. Since this data set was recorded in a real environment, the logfiles hold only the odometry based robot poses. For further information, please refer to the header of the logfiles. To simplify the parsing of the files, you can use these two Java snippets:
Laser Range Measurements:
List ranges = new ArrayList<>(numOfLaserRays);
List errors = new ArrayList<>(numOfLaserRays);
String s = line.substring(4);
String delimiter = "()";
StringTokenizer tokenizer = new StringTokenizer(s, delimiter);
while(tokenizer.hasMoreElements()){
String[] arr = tokenizer.nextToken().split(";");
boolean usable = (arr[0].equals("0")?false:true);
double range = Double.parseDouble(arr[1]);
ranges.add(range);
errors.add(usable?Error.OKAY:Error.INVALID_MEASUREMENT);
}
Poses:
String poseString = line.split(":")[2];
String[] elements = poseString.substring(1, poseString.length()-1).split(";");
double x = Double.parseDouble(elements[0]);
double y = Double.parseDouble(elements[1]);
double phi = Double.parseDouble(elements[2]);
Amyotrophic Lateral Sclerosis (ALS) is a devastating neurological disease that affects 1 in 400 people. Patients suffer a progressive loss of voluntary muscle control leading to paralysis, difficulty speaking, swallowing and ultimately, breathing. Over 60% of patients with ALS die within 3 years of clinical presentation and 90% will die within 10 years. But there is reason for hope.
We stand at a very important time and place in our mission to end ALS. Biotechnologies are rapidly evolving to produce new sources of data and change the way we learn about the brain in health and disease. Answer ALS has generated an unprecedented amount of clinical and biological data from ALS patients and healthy controls. We need your help to analyze that data, increase our understanding of ALS and bring clarity to potential therapeutic targets.
Answer ALS, EverythingALS and Roche Canada's Artificial Intelligence Centre of Excellence is requesting the collaborative effort of the AI community to fight ALS. This challenge presents a curated collection of datasets from a number of Answer ALS sources and asks you to model solutions to key questions that were developed and evaluated by ALS neurologists, researchers, and patient communities. See Task Detail Pages for more detail.
In 2014, Jay Fishman, chairman and CEO of The Travelers Companies, was diagnosed with ALS. He lost his battle with the disease just two years later. Although Jay realized that medical research would not move quickly enough to help him, he was determined to find and support a research effort that might someday change the prognosis for people diagnosed with ALS. The Answer ALS Research Program began in 2015. Learn more about Jay, ALS and Answer ALS in the video below.
The tasks associated with this dataset were developed and evaluated by ALS neurologists, researchers, and patient communities. They represent key research questions, for which insights developed by the Kaggle community can be most impactful in identifying at-risk ALS populations and possible early treatments.
To participate in this challenge, review the research questions posed in the dataset tasks and submit solutions in the form of Kaggle Notebooks.
We encourage participants to use the presented data and if needed, additional public datasets, to create their submissions.
The goal of this challenge is to connect the AI community to explore the ALS datasets, surface insights, and findings:
- May 15, 2021, 11:59PM UTC
On this date, each task will have one submission identified as the best response to the research question posed in the task. That submission will be marked as the “accepted solution” to that task. Submissions will be reviewed on a rolling basis, so participants are encouraged to work publicly and collaboratively to accelerate the research available for each task.
Prizes: - The award will be given to the participant who is identified as best meeting the evaluation criteria. To be eligible for a prize you will need to make a submission using the submission form. - A multi-disciplinary group of subject matter experts derived from named collaborators will evaluate solutions and identify a winner two weeks after the deadline submission. A team or user is eligible to receive awards for multiple tasks. - All judgments are considered final. - The winner will be awarded an opportunity to pitch to a panel of industry and VC investors (i.e. General Catalyst - Hemant Taneja, Google - Graham Spencer, and others) who have a proven track record in digital solution development, implementation and scale.
Overview of biology - The simplified central dogma of molecular biology for an organism is: DNA -> RNA -> Protein. Nature Scitable explanation.
Data Description: This dataset is composed of a curated collection of 134 people with ALS or motor neuron disease (M...
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Concise comparison of the top 10 YouTube alternatives for content creators in 2025. Covers monetization, audience size, and ideal use cases.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
The Chemotion Files is a series of datasets that can be used to get to know the electronic lab notebook Chemotion ELN. This is the second lesson in the series. With these datasets, people can get familiar with some of the features of the electronic lab notebook in a mystery, escape-room-like style "game".
How to use The Chemotion Files:
Download the .zip file and go to the Chemotion ELN Demo hosted by the KIT (https://demo.chemotion.ibcs.kit.edu/home), this is the official demo version of the electronic lab notebook. If it is your first time there, then you need to sign up. After signing up, you can directly log into the electronic lab notebook and upload the .zip with "import collection". After the import is finished, you will see one new collection in the ELN called "The_Chemotion_Files" with a sub folder called Case_002-The-Missing_Spectra. Now you can start and try to solve the mystery of the Hidden sample.
You can check the following video for details on how to "play" The Chemotion Files - Case 002":
https://www.youtube.com/watch?v=Soqtw4KTT0I
The dataset was tested with the Chemotion ELN version v1.8.1
Comment: The data used in this riddle was published by Simone Gräßle (see cited works). I recommend not to check out the original dataset before solving the riddle; otherwise you might spoil your fun with the riddle.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
:fa-spacer:
https://i.imgur.com/ED4jpM3.png" alt="Mountain Dew">
During Super Bowl LV, Mountain Dew sponsored an ad that encourages viewers to count all unique occurrences of Mountain Dew bottles. You can watch the full ad here. The first person to tweet the exactly correct count at Mountain Dew is eligible to win $1 million (see rules here).
Counting things is a perfect place for where computer vision can help.
We uploaded the Mountain Dew video to Roboflow, created three images per each second of the commercial (91 images from ~30 seconds of commercial), and annotated all bottles we could see. This dataset is the result.
We trained a model to recognize the Mountain Dew bottles, and then ran the original commercial back through this model. This helps identify Mountain Dew bottles that the human eye may have missed when completing counts.
https://i.imgur.com/rjZCS2a.png" alt="Image example">
Click "Fork" in the upper right hand corner or download the raw annotations in your desired format.
Note that while the images are property of PepsiCo, we are using them here as fair-use for educational purposes and have released the annotations under a Creative Commons license.
Roboflow enables teams to use computer vision. :fa-spacer: Our end-to-end platform enables developers to collect, organize, annotate, train, deploy, and improve their computer vision models -- all without needing to hire a new ML engineering team. :fa-spacer:
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
This study investigates the features of fake news networks and how they spread during the 2020 South Korean election. Using Actor-Network Theory (ANT), we assessed the network's central players and how they are connected. Results reveal the characteristics of the videoclips and channel networks responsible for the propagation of fake news. Analysis of the videoclip network reveals a high number of detected fake news videos and a high density of connections among users. Assessment of news videoclips on both actual and fake news networks reveals that the real news network is more concentrated. However, the scale of the network may play a role in these variations. Statistics for network centralization reveal that users are spread out over the network, pointing to its decentralized character. A closer look at the real and fake news networks inside videos and channels reveals similar trends. We find that the density of the real news videoclip network is higher than that of the fake news network, whereas the fake news channel networks are denser than their real news counterparts, which may indicate greater activity and interconnectedness in their transmission. We also found that fake news videoclips had more likes than real news videoclips, whereas real news videoclips had more dislikes than fake news videoclips. These findings strongly suggest that fake news videoclips are more accepted when people watch them on YouTube. In addition, we used semantic networks and automated content analysis to uncover common language patterns in fake news which helps us better understand the structure and dynamics of the networks involved in the dissemination of fake news. The findings reported here provide important insights on how fake news spread via social networks during the South Korean election of 2020. The results of this study have important implications for the campaign against fake news and ensuring factual coverage.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Outline
Details
Specification of Ontology
Related Resources
Apache License, v2.0https://www.apache.org/licenses/LICENSE-2.0
License information was derived automatically
This dataset was created by TheViewlytics
Released under Apache 2.0