A dataset consisting of stereo thermal, stereo color, and cross-modality image pairs with high accuracy ground truth (< 2mm) generated from a LiDAR. The authors scanned 100 cluttered indoor and 80 outdoor scenes featuring challenging environments and conditions. CATS contains approximately 1400 images of pedestrians, vehicles, electronics, and other thermally interesting objects in different environmental conditions, including nighttime, daytime, and foggy scenes.
Open Database License (ODbL) v1.0https://www.opendatacommons.org/licenses/odbl/1.0/
License information was derived automatically
Humans have long struggled to discern between these majestic big cats species, and now we invite data scientists, researchers, and enthusiasts to unleash the power of artificial intelligence. Your mission is to craft and fine-tune models that transcend human perception, differentiating between diverse big cat species with unrivaled accuracy.
Our meticulously curated dataset lays the foundation for this remarkable undertaking. With images meticulously sourced from various habitats, the dataset forms a comprehensive compendium of big cat diversity. As a participant, you'll harness this trove of data to create models that decipher the intricate features distinguishing lions, tigers, cheetahs, and more. Through your innovative approach and algorithmic prowess, the challenge aims to crown the model that can elegantly navigate the spectrum of big cat species.
A large set of images of cats and dogs.
Homepage: https://www.microsoft.com/en-us/download/details.aspx?id=54765
Source code: tfds.image_classification.CatsVsDogs
Versions:
4.0.0 (default): New split API (https://tensorflow.org/datasets/splits) Download size: 786.68 MiB
Source: https://www.tensorflow.org/datasets/catalog/cats_vs_dogs
https://choosealicense.com/licenses/unknown/https://choosealicense.com/licenses/unknown/
Dataset Card for Cats Vs. Dogs
Dataset Summary
A large set of images of cats and dogs. There are 1738 corrupted images that are dropped. This dataset is part of a now-closed Kaggle competition and represents a subset of the so-called Asirra dataset. From the competition page:
The Asirra data set Web services are often protected with a challenge that's supposed to be easy for people to solve, but difficult for computers. Such a challenge is often called a CAPTCHA… See the full description on the dataset page: https://huggingface.co/datasets/microsoft/cats_vs_dogs.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
This dataset was created by exporting the Oxford Pets dataset from Roboflow Universe, generating a version with Modify Classes to drop all of the classes for the labeled dog breeds and consolidating all cat breeds under the label, "cat." The bounding boxes were also modified to incude the entirety of the cats within the images, rather than only their faces/heads.
https://i.imgur.com/3IEzlCf.png" alt="Annotated image of a cat from the dataset">
The Oxford Pets dataset (also known as the "dogs vs cats" dataset) is a collection of images and annotations labeling various breeds of dogs and cats. There are approximately 100 examples of each of the 37 breeds. This dataset contains the object detection portion of the original dataset with bounding boxes around the animals' heads.
Origin: This dataset was collected by the Visual Geometry Group (VGG) at the University of Oxford.
Apache License, v2.0https://www.apache.org/licenses/LICENSE-2.0
License information was derived automatically
The dataset contains two files in h5 format: 1. test_catvnoncat.h5: It contains 50 test examples of cat and non-cat images 2. train_catvnoncat.h5: It contains 209 train examples of cat and non-cat images
The dataset contains images of size 64x64. The task is to classify an image as a cat (1) or not a cat (0). I am going to publish a series of notebooks for this dataset that would demonstrate neural networks from very basic level. Each notebook will build upon the previous one. Stay tuned to learn neural networks with the help of those notebooks!
You can use the below code snippet to load and visualize the dataset. ```python import numpy as np import matplotlib.pyplot as plt import h5py import os
for dirname, _, filenames in os.walk('/kaggle/input'): for filename in filenames: print(os.path.join(dirname, filename))
def load_data(): train_dataset = h5py.File('/kaggle/input/cat-images-classification-dataset/train_catvnoncat.h5', "r") train_set_x_orig = np.array(train_dataset["train_set_x"][:]) # your train set features train_set_y_orig = np.array(train_dataset["train_set_y"][:]) # your train set labels
test_dataset = h5py.File('/kaggle/input/cat-images-classification-dataset/test_catvnoncat.h5', "r")
test_set_x_orig = np.array(test_dataset["test_set_x"][:]) # your test set features
test_set_y_orig = np.array(test_dataset["test_set_y"][:]) # your test set labels
classes = np.array(test_dataset["list_classes"][:]) # the list of classes
train_set_y_orig = train_set_y_orig.reshape((1, train_set_y_orig.shape[0]))
test_set_y_orig = test_set_y_orig.reshape((1, test_set_y_orig.shape[0]))
return train_set_x_orig, train_set_y_orig, test_set_x_orig, test_set_y_orig, classes
train_x_orig, train_y, test_x_orig, test_y, classes = load_data()
index = 10 plt.imshow(train_x_orig[index]) print ("y = " + str(train_y[0,index]) + ". It's a " + classes[train_y[0,index]].decode("utf-8") + " picture.") ```
https://choosealicense.com/licenses/other/https://choosealicense.com/licenses/other/
DALL-E-Cats is a dataset meant to produce a synthetic animal dataset. This is a successor to DALL-E-Dogs. DALL-E-Dogs and DALL-E-Cats will be fed into an image classifier to see how it performs. This is under the BirdL-AirL License.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
## Overview
BIG CATS is a dataset for object detection tasks - it contains Cheetah Tiger Jaguar annotations for 1,498 images.
## Getting Started
You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
## License
This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
The Controlled Anomalies Time Series (CATS) Dataset consists of commands, external stimuli, and telemetry readings of a simulated complex dynamical system with 200 injected anomalies.
The CATS Dataset exhibits a set of desirable properties that make it very suitable for benchmarking Anomaly Detection Algorithms in Multivariate Time Series [1]:
Multivariate (17 variables) including sensors reading and control signals. It simulates the operational behaviour of an arbitrary complex system including:
4 Deliberate Actuations / Control Commands sent by a simulated operator / controller, for instance, commands of an operator to turn ON/OFF some equipment.
3 Environmental Stimuli / External Forces acting on the system and affecting its behaviour, for instance, the wind affecting the orientation of a large ground antenna.
10 Telemetry Readings representing the observable states of the complex system by means of sensors, for instance, a position, a temperature, a pressure, a voltage, current, humidity, velocity, acceleration, etc.
5 million timestamps. Sensors readings are at 1Hz sampling frequency.
1 million nominal observations (the first 1 million datapoints). This is suitable to start learning the "normal" behaviour.
4 million observations that include both nominal and anomalous segments. This is suitable to evaluate both semi-supervised approaches (novelty detection) as well as unsupervised approaches (outlier detection).
200 anomalous segments. One anomalous segment may contain several successive anomalous observations / timestamps. Only the last 4 million observations contain anomalous segments.
Different types of anomalies to understand what anomaly types can be detected by different approaches. The categories are available in the dataset and in the metadata.
Fine control over ground truth. As this is a simulated system with deliberate anomaly injection, the start and end time of the anomalous behaviour is known very precisely. In contrast to real world datasets, there is no risk that the ground truth contains mislabelled segments which is often the case for real data.
Suitable for root cause analysis. In addition to the anomaly category, the time series channel in which the anomaly first developed itself is recorded and made available as part of the metadata. This can be useful to evaluate the performance of algorithm to trace back anomalies to the right root cause channel.
Affected channels. In addition to the knowledge of the root cause channel in which the anomaly first developed itself, we provide information of channels possibly affected by the anomaly. This can also be useful to evaluate the explainability of anomaly detection systems which may point out to the anomalous channels (root cause and affected).
Obvious anomalies. The simulated anomalies have been designed to be "easy" to be detected for human eyes (i.e., there are very large spikes or oscillations), hence also detectable for most algorithms. It makes this synthetic dataset useful for screening tasks (i.e., to eliminate algorithms that are not capable to detect those obvious anomalies). However, during our initial experiments, the dataset turned out to be challenging enough even for state-of-the-art anomaly detection approaches, making it suitable also for regular benchmark studies.
Context provided. Some variables can only be considered anomalous in relation to other behaviours. A typical example consists of a light and switch pair. The light being either on or off is nominal, the same goes for the switch, but having the switch on and the light off shall be considered anomalous. In the CATS dataset, users can choose (or not) to use the available context, and external stimuli, to test the usefulness of the context for detecting anomalies in this simulation.
Pure signal ideal for robustness-to-noise analysis. The simulated signals are provided without noise: while this may seem unrealistic at first, it is an advantage since users of the dataset can decide to add on top of the provided series any type of noise and choose an amplitude. This makes it well suited to test how sensitive and robust detection algorithms are against various levels of noise.
No missing data. You can drop whatever data you want to assess the impact of missing values on your detector with respect to a clean baseline.
Change Log
Version 2
Metadata: we include a metadata.csv with information about:
Anomaly categories
Root cause channel (signal in which the anomaly is first visible)
Affected channel (signal in which the anomaly might propagate) through coupled system dynamics
Removal of anomaly overlaps: version 1 contained anomalies which overlapped with each other resulting in only 190 distinct anomalous segments. Now, there are no more anomaly overlaps.
Two data files: CSV and parquet for convenience.
[1] Example Benchmark of Anomaly Detection in Time Series: “Sebastian Schmidl, Phillip Wenig, and Thorsten Papenbrock. Anomaly Detection in Time Series: A Comprehensive Evaluation. PVLDB, 15(9): 1779 - 1797, 2022. doi:10.14778/3538598.3538602”
About Solenix
Solenix is an international company providing software engineering, consulting services and software products for the space market. Solenix is a dynamic company that brings innovative technologies and concepts to the aerospace market, keeping up to date with technical advancements and actively promoting spin-in and spin-out technology activities. We combine modern solutions which complement conventional practices. We aspire to achieve maximum customer satisfaction by fostering collaboration, constructivism, and flexibility.
MIT Licensehttps://opensource.org/licenses/MIT
License information was derived automatically
This dataset was created by Sarim Javed
Released under MIT
https://choosealicense.com/licenses/cc0-1.0/https://choosealicense.com/licenses/cc0-1.0/
Dataset Summary
A dataset from kaggle with duplicate data removed.
Data Fields
The data instances have the following fields:
image: A PIL.Image.Image object containing the image. Note that when accessing the image column: dataset[0]["image"] the image file is automatically decoded. Decoding of a large number of image files might take a significant amount of time. Thus it is important to first query the sample index before the "image" column, i.e. dataset[0]["image"]… See the full description on the dataset page: https://huggingface.co/datasets/Bingsu/Cat_and_Dog.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
## Overview
Different Cats is a dataset for object detection tasks - it contains Cats annotations for 2,970 images.
## Getting Started
You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
## License
This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).
huggan/cats dataset hosted on Hugging Face and contributed by the HF Datasets community
Open Government Licence 3.0http://www.nationalarchives.gov.uk/doc/open-government-licence/version/3/
License information was derived automatically
This dataset is a modelled dataset, describing an upper estimate of cats per square kilometre across GB. The figures are aligned to the British national grid, with a population estimate provided for each 1km square. These data were generated as part of the delivery of commissioned research. The data contained within this dataset are modelled figures, based on upper 95th percentile national estimates for pet population, and available information on Veterinary activity across GB. The data are accurate as of 01/01/2015. The data provided are summarised to the 1km level. Further information on this research is available in a research publication by James Aegerter, David Fouracre & Graham C. Smith, discussing the structure and density of pet cat and dog populations across Great Britain.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
## Overview
COCO 2017 Cats is a dataset for object detection tasks - it contains Cats annotations for 4,112 images.
## Getting Started
You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
## License
This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).
The estimated number of cats owned by households in Turkey constantly increased in the observed years from 2012 to 2023. The cat population in Turkey peaked at over 4.7 million in 2023.
https://academictorrents.com/nolicensespecifiedhttps://academictorrents.com/nolicensespecified
Abstract This dataset, composed of 440 sounds, contains meows emitted by cats in different contexts. Specifically, 21 cats belonging to 2 breeds (Maine Coon and European Shorthair) have been repeatedly exposed to three different stimuli that were expected to induce the emission of meows: Brushing - Cats were brushed by their owners in their home environment for a maximum of 5 minutes; Isolation in an unfamiliar environment - Cats were transferred by their owners into an unfamiliar environment (e.g., a room in a different apartment or an office). Distance was minimized and the usual transportation routine was adopted so as to avoid discomfort to animals. The journey lasted less than 30 minutes and cats were allowed 30 minutes with their owners to recover from transportation, before being isolated in the unfamiliar environment, where they stayed alone for maximum 5 minutes; Waiting for food - The owner started the routine operations that preceded food delivery in the usual environment the cat was familiar with. Food was given at most 5 minutes after the beginning of the experiment. The dataset has been produced and employed in the context of an interdepartmental project of the University of Milan (for further information, please refer to this doi). The content of the dataset has been described in detail in a scientific work currently under review; the reference will be provided as soon as the paper is published. File naming conventions Files containing meows are in the dataset.zip archive. They are PCM streams (.wav). Naming conventions follow the pattern C_NNNNN_BB_SS_OOOOO_RXX, which has to be exploded as follows: C = emission context (values: B = brushing; F = waiting for food; I: isolation in an unfamiliar environment); NNNNN = cat’s unique ID; BB = breed (values: MC = Maine Coon; EU: European Shorthair); SS = sex (values: FI = female, intact; FN: female, neutered; MI: male, intact; MN: male, neutered); OOOOO = cat owner’s unique ID; R = recording session (values: 1, 2 or 3) XX = vocalization counter (values: 01..99) Extra content The extra.zip archive contains excluded recordings (sounds other than meows emitted by cats) and uncut sequences of close vocalizations. Terms of use The dataset is open access for scientific research and non-commercial purposes. The authors require to acknowledge their work and, in case of scientific publication, to cite the most suitable reference among the following entries: Ntalampiras, S., Ludovico, L.A., Presti, G., Prato Previde, E., Battini, M., Cannas, S., Palestrini, C., Mattiello, S.: Automatic Classification of Cat Vocalizations Emitted in Different Contexts. Animals, vol. 9(8), pp. 543.1–543.14. MDPI (2019). ISSN: 2076-2615 Ludovico, L.A., Ntalampiras, S., Presti, G., Cannas, S., Battini, M., Mattiello, S.: CatMeows: A Publicly-Available Dataset of Cat Vocalizations. In: Li, X., Lokoč, J., Mezaris, V., Patras, I., Schoeffmann, K., Skopal, T., Vrochidis, S. (eds.) MultiMedia Modeling. 27th International Conference, MMM 2021, Prague, Czech Republic, June 22–24, 2021, Proceedings, Part II, LNCS, vol. 12573, pp. 230–243. Springer International Publishing, Cham (2021). ISBN: 978-3-030-67834-0 (print), 978-3-030-67835-7 (online) ISSN: 0302-9743 (print), 1611-3349 (online) {"references": ["Ntalampiras, S., Ludovico, L.A., Presti, G., Prato Previde, E., Battini, M., Cannas, S., Palestrini, C., Mattiello, S.: Automatic Classification of Cat Vocalizations Emitted in Different Contexts. Animals, vol. 9(8), pp. 543.1\u2013543.14. MDPI (2019). ISSN: 2076-2615."]} This project was supported by "Piano di Sostegno alla Ricerca, Linea 2-Mattiello" of the University of Milan (2016).
CATS-ISS_L1B_D-M7.1-V3-00 is the Cloud-Aerosol Transport System (CATS) International Space Station (ISS) Level 1B Day Mode 7.1 Version 3-00 data product. The collection spans from February 10, 2015 through March 21, 2015. CATS, which was launched on January 10, 2015, was a lidar remote sensing instrument that provided range-resolved profile measurements of atmospheric aerosols and clouds from the ISS. CATS was intended to operate on-orbit for up to three years. CATS provides vertical profiles at three wavelengths, orbiting between ~230 and ~270 miles above the Earth's surface at a 51-degree inclination with nearly a three-day repeat cycle. For the first time, scientists were able to study diurnal (day-to-night) changes in cloud and aerosol effects from space by observing the same spot on Earth at different times each day. Level 1B data have been calibrated and annotated with ancillary meteorological data and processed to sensor units.
The reports include data on licensed dogs and cats residing in Toronto after year 2023, including primary breed, Forward Sortation Area (FSA), and year of licensing. For data before 2023, please visit: Licensed Dog and Cat Names: https://open.toronto.ca/dataset/licensed-dog-and-cat-names/ Licensed Dogs and Cats Reports: https://open.toronto.ca/dataset/licensed-dogs-and-cats-reports/
A dataset consisting of stereo thermal, stereo color, and cross-modality image pairs with high accuracy ground truth (< 2mm) generated from a LiDAR. The authors scanned 100 cluttered indoor and 80 outdoor scenes featuring challenging environments and conditions. CATS contains approximately 1400 images of pedestrians, vehicles, electronics, and other thermally interesting objects in different environmental conditions, including nighttime, daytime, and foggy scenes.