Facebook
TwitterThe Visual Storytelling Dataset (VIST) consists of 10,117 Flickr albums and 210,819 unique images. Each sample is one sequence of 5 photos selected from the same album paired with a single human constructed story, where each story is comprised of mostly one sentence per image.
Facebook
TwitterWe introduce the first dataset for sequential vision-to-language, and explore how this data may be used for the task of visual storytelling. The dataset includes 81,743 unique photos in 20,211 sequences, aligned to descriptive and story language. VIST is previously known as "SIND", the Sequential Image Narrative Dataset (SIND).
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
This version of the Bloom Library data is developed specifically for the Visual Story Telling (VIST) task. It includes data from 363 languages across 36 language families, with many of the languages represented being extremely low resourced languages.
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Video Storytelling is a dataset for generating text story/summarization for videos containing social events. It consists of 105 videos from four categories: birthday, camping, Christmas and wedding. For each video, we provide at least 5 human-written stories.
Please cite the following paper if you use the Video Storytelling dataset in your work (papers, articles, reports, books, software, etc):
Facebook
TwitterVisual Story-Telling dataset (VIST) is the only publicly accessible dataset for storytelling problems. It comprises 210,819 distinct images that can be found in 10,117 different albums on Flickr and is arranged in sets of five different images.
Facebook
Twitterhttps://choosealicense.com/licenses/cc/https://choosealicense.com/licenses/cc/
BLOOM VIST is a visual storytelling of books that consists of 62 languages indigenous to SEA. This dataset is owned by Bloom, a free, open-source software developed by SIL International and associated with Bloom Library, app, and services. This dataset is released with the LICENSE family of Creative Commons (although each story datapoints has its licensing in more detail, e.g cc-by, cc-by-nc, cc-by-nd, cc-by-sa, cc-by-nc-nd, cc-by-nc-sa). Before using this dataloader, please accept the… See the full description on the dataset page: https://huggingface.co/datasets/SEACrowd/bloom_vist.
Facebook
TwitterVisual storytelling refers to the manner of describing a set of images rather than a single image, also known as multi-image captioning. Visual Storytelling Task (VST) takes a set of images as input and aims to generate a coherent story relevant to the input images. In this dataset, we bridge the gap and present a new dataset for expressive and coherent story creation. We present the Sequential Storytelling Image Dataset (SSID), consisting of open-source video frames accompanied by story-like annotations. In addition, we provide four annotations (i.e., stories) for each set of five images. The image sets are collected manually from publicly available videos in three domains: documentaries, lifestyle, and movies, and then annotated manually using Amazon Mechanical Turk. In summary, SSID dataset is comprised of 17,365 images, which resulted in a total of 3,473 unique sets of five images. Each set of images is associated with four ground truths, resulting in a total of 13,892 unique ground truths (i.e., written stories). And each ground truth is composed of five connected sentences written in the form of a story.
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Story maps have emerged as a popular storytelling device in recent years with cartographers and journalists leveraging geospatial web technologies to create unique spatial narratives. However, empirical research analyzing the design of story maps remains limited. Two recently proposed design frameworks provide promising avenues to characterize story maps in terms of elements of vivid cartography and techniques of map-based storytelling. In this article, I conducted a quantitative content analysis on 117 story maps of COVID-19 to operationalize map-based storytelling and vividness frameworks and to identify common design traits in contemporary story maps. My findings indicated that most story maps are longform infographics that use scrolling to advance the narrative. Stories applied a variety of attention, dosing, and mood techniques to enrich the storytelling experience. Story maps were primarily vivid through their use of color and novelty. Overall, most story maps utilized only a fraction of the map-based storytelling framework techniques. This research also demonstrated that it is challenging to analyze story maps based on these frameworks. Finally, this article improves the frameworks by proposing two new story map techniques and suggesting avenues of refinement.
Facebook
TwitterCC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
License information was derived automatically
Supporting data for the article "The Illustrated Page: Analyzing Illustrations of Historical Children’s Books Using Citizen Science" (CHR 2025)
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Comprehensive analysis and strategic guide for mortgage lending professionals seeking to implement advanced content marketing strategies in 2025. This dataset provides detailed insights into video marketing, interactive content, visual storytelling, and specialized SEO techniques specifically tailored for the mortgage industry. The content addresses the evolution from generic blog posts to personalized, engaging content that builds trust with modern borrowers who expect authentic, helpful guidance through their home-buying journey.
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
This dataset includes responses from 51 students who participated in a survey evaluating a short film used in Computer Science education, that portrayed three algorithmic approaches: ad-hoc, greedy, and dynamic programming. Using a 7-point Likert scale (-3 to 3), students rated statements about the film's characteristics and potential benefits. The questionnaire aimed to thoroughly capture students' perspectives on the film's attributes and educational impact.Items used in the survey.EF.Entertainment - The short film provided a high entertainment value.EF.ProductionValue - The short film had a high production value.EF.Premise - The premise (escape room) was intriguing.EF.Expressive - The short film was expressive.EF.Immersive - The short film was immersive.EF.Creative - The short film was creative.EF.Pacing The pacing of the story was appropriate.FA.Story-plot - I appreciate as important the presence of the story-plot.FA.LiveAction - I appreciate as important the use of live-action performances.FA.CameraWork - I appreciate as important the cut and switch of camera angles.FA.Atmosphere - I appreciate as important the mood and atmosphere.FA.Choreography - I appreciate as important the choreography.FA.Cinematography - I appreciate as important the depicted cinematography.FA.NonVerbal - I appreciate as important the facial expressions, body language of the actors.FA.SoundDesign I appreciate as important the sound design, narration and sound effects present.CB.Educational. The short film provided a high educational value.CB.Understanding The learning experience deepened my understanding of the subject.CB.Clarity The algorithmic strategies were clearly depicted.EB.Attention - The movie engaged my attention.EB.Curiosity - The movie engaged my curiosity.PU.Quicker - Using such short films during a class would enable me to learn and deepen algorithmic concepts more quickly.PU.Performance - Using such short films during a class would improve my learning performance and grades.PU.Efficiency - Using such short films could help me get the most out of my time while learning.PU.Knowledge - Using such short films may improve my knowledge.PU.Easier - Using such short films would make it easier to accomplish my learning tasks.PU.Overall Using such short films would be overall beneficial.PE.Enjoyable - The learning experience was enjoyable.PE.Exciting - The learning experience was exciting.PE.Pleasant - The learning experience was pleasant.PE.Interesting - The learning experience was interesting.PE.Immersive - The learning experience was immersive.C.Changes - The use of such short films may imply major changes in how I learn.C.Incorporation - It would be easy to incorporate such short films in my learning process.A.Worthwhile - Using similar educational short films to learn algorithmic concepts is a good idea.A.Positivity - I am positive towards using visual media to better understand algorithmic concepts.A.Appreciate - I would appreciate the availabilty of similar short films as learning instuments.A.WouldUse - If available, I would use such short films in my learning process.Eval.Use - I often use/used existing AlgoRythmics videos in my learning process.Eval.Comp - Overall, the short film approach (story-line, live-acting etc.) provides a richer and more valuable learning experience than the viewing of simple videos or animations.
Facebook
Twitterhttps://www.datainsightsmarket.com/privacy-policyhttps://www.datainsightsmarket.com/privacy-policy
Discover the booming digital visual content market! This in-depth analysis reveals key trends, growth projections (CAGR), major players (Shutterstock, Getty Images, Adobe), and regional insights from 2019-2033. Learn about the driving forces and challenges shaping this multi-billion dollar industry.
Facebook
Twitterhttps://www.datainsightsmarket.com/privacy-policyhttps://www.datainsightsmarket.com/privacy-policy
Discover the booming digital storytelling platform market! Our analysis reveals a $3712 million market in 2025, growing at 7.8% CAGR through 2033. Learn about key drivers, trends, and top players like Adobe and Canva. Explore market segmentation and regional insights for informed business decisions.
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Visual narratives are promising tools for science and health communication, especially for broad audiences in times of public health crisis, such as during the COVID-19 pandemic. In this study, we used the Lifeology illustrated “flashcard” course platform to construct visual narratives about COVID-19, and then assessed their impact on behavioral intentions. We conducted a survey experiment among 1,775 health app users. Participants viewed illustrated (sequential art) courses about: 1) sleep, 2) what COVID-19 is and how to protect oneself, 3) mechanisms of how the virus works in the body and risk factors for severe disease. Each participant viewed one of these courses and then answered questions about their understanding of the course, how much they learned, and their perceptions and behavioral intentions toward COVID-19. Participants generally evaluated “flashcard” courses as easy to understand. Viewing a COVID-19 “flashcard” course was also associated with improved self-efficacy and behavioral intentions toward COVID-19 disease prevention as compared to viewing a “flashcard” course about sleep science. Our findings support the use of visual narratives to improve health literacy and provide individuals with the capacity to act on health information that they may know of but find difficult to process or apply to their daily lives.
Facebook
TwitterAttribution-NonCommercial 3.0 (CC BY-NC 3.0)https://creativecommons.org/licenses/by-nc/3.0/
License information was derived automatically
This dataset is introduced by the paper "Understanding, Categorizing and Predicting Semantic Image-Text Relations". If you are using this dataset it in your work, please cite: @inproceedings{otto2019understanding, title={Understanding, Categorizing and Predicting Semantic Image-Text Relations}, author={Otto, Christian and Springstein, Matthias and Anand, Avishek and Ewerth, Ralph}, booktitle={In Proceedings of ACM International Conference on Multimedia Retrieval (ICMR 2019)}, year={2019} } To create the full tar use the following command in the command line: cat train.tar.part* > train_concat.tar Then simply untar it via tar -xf train_concat.tar The jsonl files contain metadata of the following format: id, origin, CMI, SC, STAT, ITClass, text, tagged text, image_path License Information: This dataset is composed of various open access sources as described in the paper. We thank all the original authors for their work. Pitt Image Ads Dataset: http://people.cs.pitt.edu/~kovashka/ads/ Image-Net challenge: http://image-net.org/ Visual Storytelling Dataset (VIST): http://visionandlanguage.net/VIST/ Wikipedia: https://www.wikipedia.org/ Microsoft COCO: http://cocodataset.org/#home
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
In visual narratives like comics, the most overt form of perspective-taking comes in panels that directly depict the viewpoints of characters in the scene. We therefore examined these subjective viewpoint panels (also known as point-of-view panels) in a corpus of over 300 annotated comics from Asia, Europe, and the United States. In line with predictions that Japanese manga use a more “subjective” storytelling style than other comics, we found that more manga use subjective panels than other comics, with high proportions of subjective panels also found in Chinese, French, and American comics. In addition, panels with more “focal” framing, i.e. micro panels showing close ups and/or amorphic panels showing views of the environment, had higher proportions of subjective panels than panels showing wider views of scenes. These findings further show that empirical corpus analyses provide evidence of cross-cultural variation and reveal relationships across structures in the visual languages of comics.
Facebook
Twitterhttps://www.datainsightsmarket.com/privacy-policyhttps://www.datainsightsmarket.com/privacy-policy
Explore the dynamic Online Comic Generator market's growth, key drivers, and future trends. Discover opportunities in personal, educational, and advertising comic creation.
Facebook
Twitterhttps://dataverse.nl/api/datasets/:persistentId/versions/2.1/customlicense?persistentId=doi:10.34894/DTBW7Mhttps://dataverse.nl/api/datasets/:persistentId/versions/2.1/customlicense?persistentId=doi:10.34894/DTBW7M
This project examined the processing of bridging inferences in visual narratives, i.e. when readers need to infer information missing in a previous panel in a comic sequence. Rather than omitting the key event, this study replaced the climax of the scene with a variety of five inferential techniques, which implicitly express the unseen event while each balancing several underlying features that describe their informativeness. The main question asked to what extent the processing of these techniques differed. Two self-paced reading experiments measured viewing times as well as comprehensibility ratings; experiment 1 directly compared the five types and experiment 2 explored the effect of combining techniques. Additionally, this project explored the underlying features as predictors for viewing times and ratings.
Facebook
Twitterhttps://www.marketreportanalytics.com/privacy-policyhttps://www.marketreportanalytics.com/privacy-policy
The social media design app market is booming, projected to reach $701.3 million by 2033 with a 9.6% CAGR. Learn about key drivers, trends, and leading players like Canva and Adobe in this in-depth market analysis. Discover regional market shares and growth opportunities in this rapidly evolving sector.
Facebook
TwitterMIT Licensehttps://opensource.org/licenses/MIT
License information was derived automatically
DATABIRD VISUALS
Description
This dataset contains 5755 entries focused on understanding color theory principles and their application in various fields including complimentary colors, working with color contrasts, color psychology, color and branding, color in fashion design, color grading and visual storytelling, associating colors with emotions, building a color palette, and the impact of black and white vs. color imagery. The data is provided in JSON format. Each… See the full description on the dataset page: https://huggingface.co/datasets/theprint/databird-visuals.
Facebook
TwitterThe Visual Storytelling Dataset (VIST) consists of 10,117 Flickr albums and 210,819 unique images. Each sample is one sequence of 5 photos selected from the same album paired with a single human constructed story, where each story is comprised of mostly one sentence per image.