Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
The Webis-Web-Archive-17 comprises a total of 10,000 web page archives from mid-2017 that were carefully sampled from the Common Crawl to involve a mixture of high-ranking and low-ranking web pages. The dataset contains the web archive files, HTML DOM, and screenshots of each web page, as well as per-page annotations of visual web archive quality.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
The Webis-Web-Archive-Quality-22 comprises a total of 6,500 pairs of screenshots from web pages as they were archived and as they were reproduced from that archive, along with archive quality annotations and information of DOM elements on the screenshot.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
The Webis-Web-Errors-19 comprises various annotations for the 10,000 web page archives of the Webis-Web-Archive-17. The annotations are whether the page is (1) mostly advertisement, (2) cut off, (3) still loading, (4) pornographic; and whether it shows (not/a bit/ very) (5) pop-ups, (6) CAPTCHAs, or (7) error messages. If you use this dataset in your research, please cite it using this paper.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Dataset of crowdsourced annotations for web page segmentations.
Web pages are taken from the webis-web-archive-17.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
The Webis-WebSeg-20 dataset comprises 42,450 crowdsourced segmentations for 8,490 web pages from the Webis-Web-Archive-17. Segmentations were fused from the segmentations of five crowd workers each.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
The Webis Clickbait Corpus 2017 (Webis-Clickbait-17) comprises a total of 38,517 Twitter posts from 27 major US news publishers. In addition to the posts, information about the articles linked in the posts are included. The posts had been published between November 2016 and June 2017. To avoid publisher and topical biases, a maximum of ten posts per day and publisher were sampled. All posts were annotated on a 4-point scale [not click baiting (0.0), slightly click baiting (0.33), considerably click baiting (0.66), heavily click baiting (1.0)] by five annotators from Amazon Mechanical Turk. A total of 9,276 posts are considered clickbait by the majority of annotators. In terms of its size, this corpus outranges the Webis Clickbait Corpus 2016 by one order of magnitude. The corpus is divided into two logical parts, a training and a test dataset. The training dataset has been released in the course of the Clickbait Challenge and a download link is provided below. To allow for an objective evaulatuion of clickbait detection systems, the test dataset is available only through the Evaluation-as-a-Service platform TIRA at the moment. On TIRA, developers can deploy clickbait detection systems and execute them against the test dataset. The performance of the submitted systems can be viewed on the TIRA page of the Clickbait Challenge.To make working with the Webis Clickbait Corpus 2017 convenient, and to allow for its validation and replication, we are developing and sharing a number of software tags:
Corpus Viewer. Our Django web service for exploring corpora. For importing the Webis Clickbait Corpus 2017 into the corpus viewer, we provide an appropriate configuration file.MTurk Manager. Our Django web service for conducting sophisticated crowd sourcing tasks on Amazon Mechanical Turk. The service allows to manage projects, upload batches of HITS, apply custom reviewing interfaces, and more. To make the clickbait crowd-sourcing task replicable, we share the worker template that we used to instruct the workers and to display the tweets. Also shared is a reviewing template that can be used to accept/reject assignments and to assess the quality of the received annotations quickly.Web Archiver. Software for archiving web pages as WARC files and reproducing them later on. This software can be used to open the WARC archives provided above.
In addition to the corpus ""clickbait17-train-170630.zip"", we provide the original WARC archives of the articles that are linked in the posts. They are split in 5 archives that can be extracted separately.
Archive of the Geohistory/Géohistoire website and related files. Captured July 17, 2020.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
The MS COCO (Microsoft Common Objects in Context) dataset is a large-scale object detection, segmentation, key-point detection, and captioning dataset. The dataset consists of 164K images.
This is the original version from 2014 made available here for easy access in Kaggle and because it does not seem to be still available on the COCO Dataset website. This has been retrieved from the mirror that Joseph Redmon has setup on this own website.
The 2014 version of the COCO dataset is an excellent object detection dataset with 80 classes, 82,783 training images and 40,504 validation images. This dataset contains all this imagery on two folders as well as the annotation with the class and location (bounding box) of the objects contained in each image.
The initial split provides training (83K), validation (41K) and test (41K) sets. Since the split between training and validation was not optimal in the original dataset, there is also two text (.part) files with a new split with only 5,000 images for validation and the rest for training. The test set has no labels and can be used for visual validation or pseudo-labelling.
This is mostly inspired by Erik Linder-Norén and [Joseph Redmon](https://pjreddie.com/darknet/yolo
http://www.nationalarchives.gov.uk/doc/non-commercial-government-licence/non-commercial-government-licence.htmhttp://www.nationalarchives.gov.uk/doc/non-commercial-government-licence/non-commercial-government-licence.htm
Data from Transnationalizing Modern Languages (09-2018)
Transnationalizing Modern Languages: Mobility, Identity and Translation in Modern Italian Cultures (TML) (funded by the AHRC under the ‘Translating Cultures’ theme, 2014-17)
PI Charles Burdett, University of Bristol. CIs Jenny Burns (Warwick), Loredana Polezzi (Warwick/Cardiff), Derek Duncan (St Andrews), Margaret Hills de Zarate (QMU)
RAs: Barbara Spadaro (Bristol), Carlo Pirozzi (St Andrews), Marco Santello (Warwick), Naomi Wells (Warwick), Luisa Percopo (Cardiff)
PhD students: Iacopo Colombini (St Andrews), Georgia Wall (Warwick)
Below is a short description of the project. Within the repository, there is a longer description of TML and each folder is accompanied by an explanatory text.
The project investigates practices of linguistic and cultural interchange within communities and individuals and explores the ways in which cultural translation intersects with linguistic translation in the everyday lives of people. The project has used as its primary object of enquiry the 150-year history of Italy as a nation state and its patterns of emigration and immigration. TML has concentrated on a series of exemplary cases, representative of the geographic, historical and linguistic map of Italian mobility. Focussing on the cultural associations that each community has formed, it examines the wealth of publications and materials that are associated with these organizations.
Working closely with researchers from across Modern Languages, the project has sought to demonstrate the principle that language is most productively apprehended in the frame of translation and the national in the frame of the transnational. TML is contributing to the development of a new framework for the disciplinary field of MLs, one which puts the interaction of languages and cultures at its core.
The principles of co-production and co-research lie at the core of the project and TML has worked closely with a very extensive range of partners. It has worked closely with Castlebrae and Drummond Community High Schools and with cultural associations across the world. The project exhibition, featuring the research of the project and including the work of photographer Mario Badagliacca, was curated by Viviana Gravano and Giulia Grechi of Routes Agency. Project events in the UK have drawn on the expertise of Rita Wilson (Monash), the writer Shirin Ramzanali Fazel and all members of the Advisory Board. The project, in close collaboration with the University of Namibia (UNAM) and the Phoenix Project (Cardiff), has been followed by ‘TML: Global Challenges’.
Not seeing a result you expected?
Learn how you can add new datasets to our index.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
The Webis-Web-Archive-17 comprises a total of 10,000 web page archives from mid-2017 that were carefully sampled from the Common Crawl to involve a mixture of high-ranking and low-ranking web pages. The dataset contains the web archive files, HTML DOM, and screenshots of each web page, as well as per-page annotations of visual web archive quality.