100+ datasets found
  1. f

    UoS Buildings Image Dataset for Computer Vision Algorithms

    • salford.figshare.com
    application/x-gzip
    Updated Jan 23, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Ali Alameer; Mazin Al-Mosawy (2025). UoS Buildings Image Dataset for Computer Vision Algorithms [Dataset]. http://doi.org/10.17866/rd.salford.20383155.v2
    Explore at:
    application/x-gzipAvailable download formats
    Dataset updated
    Jan 23, 2025
    Dataset provided by
    University of Salford
    Authors
    Ali Alameer; Mazin Al-Mosawy
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    The dataset for this project is represented by photos, photos for the buildings of the University of Salford, these photos are taken by a mobile phone camera from different angels and different distances , even though this task sounds so easy but it encountered some challenges, these challenges are summarized below: 1. Obstacles. a. Fixed or unremovable objects. When taking several photos for a building or a landscape from different angels and directions ,there are some of these angels blocked by a form of a fixed object such as trees and plants, light poles, signs, statues, cabins, bicycle shades, scooter stands, generators/transformers, construction barriers, construction equipment and any other service equipment so it is unavoidable to represent some photos without these objects included, this will raise 3 questions. - will these objects confuse the model/application we intend to create meaning will that obstacle prevent the model/application from identifying the designated building? - Or will the photos be more precise with these objects and provide the capability for the model/application to identify these building with these obstacles included? - How far is the maximum length for detection? In other words, how far will the mobile device with the application be from the building so it could or could not detect the designated building? b. Removable and moving objects. - Any University is crowded with staff and students especially in the rush hours of the day so it is hard for some photos to be taken without a personnel appearing in that photo in a certain time period of the day. But, due to privacy issues and showing respect to that person, these photos are better excluded. - Parked vehicles, trollies and service equipment can be an obstacle and might appear in these images as well as it can block access to some areas which an image from a certain angel cannot be obtained. - Animals, like dogs, cats, birds or even squirrels cannot be avoided in some photos which are entitled to the same questions above.
    2. Weather. In a deep learning project, more data means more accuracy and less error, at this stage of our project it was agreed to have 50 photos per building but we can increase the number of photos for more accurate results but due to the limitation of time for this project it was agreed for 50 per building only. these photos were taken on cloudy days and to expand our work on this project (as future works and recommendations). Photos on sunny, rainy, foggy, snowy and any other weather condition days can be included. Even photos in different times of the day can be included such as night, dawn, and sunset times. To provide our designated model with all the possibilities to identify these buildings in all available circumstances.

    1. The selected buildings. It was agreed to select 10 buildings only from the University of Salford buildings for this project with at least 50 images per building, these selected building for this project with the number of images taken are:
    2. Chapman: 74 images
    3. Clifford Whitworth Library: 60 images
    4. Cockcroft: 67 images
    5. Maxwell: 80 images
    6. Media City Campus: 92 images
    7. New Adelphi: 93 images
    8. New Science, Engineering & Environment: 78 images
    9. Newton: 92 images
    10. Sports Centre: 55 images
    11. University House: 60 images Peel building is an important figure of the University of Salford due to its distinct and amazing exterior design but unfortunately it was excluded from the selection due to some maintenance activities at the time of collecting the photos for this project as it is partially covered with scaffolding and a lot of movement by personnel and equipment. If the supervisor suggests that this will be another challenge to include in the project then, it is mandatory to collect its photos. There are many other buildings in the University of Salford and again to expand our project in the future, we can include all the buildings of the University of Salford. The full list of buildings of the university can be reviewed by accessing an interactive map on: www.salford.ac.uk/find-us

    12. Expand Further. This project can be improved furthermore with so many capabilities, again due to the limitation of time given to this project , these improvements can be implemented later as future works. In simple words, this project is to create an application that can display the building’s name when pointing a mobile device with a camera to that building. Future featured to be added: a. Address/ location: this will require collection of additional data which is the longitude and latitude of each building included or the post code which will be the same taking under consideration how close these buildings appear on the interactive map application such as Google maps, Google earth or iMaps. b. Description of the building: what is the building for, by which school is this building occupied? and what facilities are included in this building? c. Interior Images: all the photos at this stage were taken for the exterior of the buildings, will interior photos make an impact on the model/application for example, if the user is inside newton or chapman and opens the application, will the building be identified especially the interior of these buildings have a high level of similarity for the corridors, rooms, halls, and labs? Will the furniture and assets will be as obstacles or identification marks? d. Directions to a specific area/floor inside the building: if the interior images succeed with the model/application, it would be a good idea adding a search option to the model/application so it can guide the user to a specific area showing directions to that area, for example if the user is inside newton building and searches for lab 141 it will direct him to the first floor of the building with an interactive arrow that changes while the user is approaching his destination. Or, if the application can identify the building from its interior, a drop down list will be activated with each floor of this building, for example, if the model/application identifies Newton building, the drop down list will be activated and when pressing on that drop down list it will represent interactive tabs for each floor of the building, selecting one of the floors by clicking on its tab will display the facilities on that floor for example if the user presses on floor 1 tab, another screen will appear displaying which facilities are on that floor. Furthermore, if the model/application identifies another building, it should activate a different number of floors as buildings differ in the number of floors from each other. this feature can be improved with a voice assistant that can direct the user after he applies his search (something similar to the voice assistant in Google maps but applied to the interior of the university’s buildings. e. Top View: if a drone with a camera can be afforded, it can provide arial images and top views for the buildings that can be added to the model/application but these images can be similar to the interior images situation , the buildings can be similar to each other from the top with other obstacles included like water tanks and AC units.

    13. Other Questions:

    14. Will the model/application be reproducible? the presumed answer for this question should be YES, IF, the model/application will be fed with the proper data (images) such as images of restaurants, schools, supermarkets, hospitals, government facilities...etc.

  2. Data from: AI & Computer Vision Dataset

    • kaggle.com
    zip
    Updated Apr 1, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Khushi Yadav (2025). AI & Computer Vision Dataset [Dataset]. https://www.kaggle.com/datasets/khushikyad001/ai-and-computer-vision-dataset
    Explore at:
    zip(128433 bytes)Available download formats
    Dataset updated
    Apr 1, 2025
    Authors
    Khushi Yadav
    License

    MIT Licensehttps://opensource.org/licenses/MIT
    License information was derived automatically

    Description

    This dataset contains metadata related to three categories of AI and computer vision applications:

    Handwritten Math Solutions: Metadata on images of handwritten math problems with step-by-step solutions.

    Multi-lingual Street Signs: Road sign images in various languages, with translations.

    Security Camera Anomalies: Surveillance footage metadata distinguishing between normal and suspicious activities.

    The dataset is useful for machine learning, image recognition, OCR (Optical Character Recognition), anomaly detection, and AI model training.

  3. D

    COCO-style geographically unbiased image dataset for computer vision...

    • dataverse.ird.fr
    pdf, txt, zip
    Updated Jan 13, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Theophile Bayet; Theophile Bayet (2023). COCO-style geographically unbiased image dataset for computer vision applications [Dataset]. http://doi.org/10.23708/N2UY4C
    Explore at:
    zip(176316624), zip(218991), pdf(57252), txt(1731), pdf(83345), zip(308454)Available download formats
    Dataset updated
    Jan 13, 2023
    Dataset provided by
    DataSuds
    Authors
    Theophile Bayet; Theophile Bayet
    License

    https://dataverse.ird.fr/api/datasets/:persistentId/versions/1.0/customlicense?persistentId=doi:10.23708/N2UY4Chttps://dataverse.ird.fr/api/datasets/:persistentId/versions/1.0/customlicense?persistentId=doi:10.23708/N2UY4C

    Time period covered
    Jan 1, 2022 - Apr 1, 2022
    Description

    There are already a lot of datasets linked to computer vision tasks (Imagenet, MS COCO, Pascal VOC, OpenImages, and numerous others), but they all suffer from important bias. One bias of significance for us is the data origin: most datasets are composed of data coming from developed countries. Facing this situation, and the need of data with local context in developing countries, we try here to adapt common data generation process to inclusive data, meaning data drawn from locations and cultural context that are unseen or poorly represented. We chose to replicate MS COCO's data generation process, as it is well documented and easy to implement. Data was collected from January to April 2022 through Flickr platform. This dataset contains the results of our data collection process, as follows : 23 text files containing comma separated URLs for each of the 23 geographic zones identified in the UN M49 norm. These text files are named according to the names of the geographic zones they cover. Annotations for 400 images per geographic zones. Those annotations are COCO-style, and inform on the presence or absence of 91 categories of objects or concepts on the images. They are shared in a JSON format. Licenses for the 400 annotations per geographic zones, based on the original licenses of the data and specified per image. Those licenses are shared under CSV format. A document explaining the objectives and methodology underlying the data collection, also describing the different components of the dataset.

  4. Low-quality image dataset

    • kaggle.com
    zip
    Updated May 15, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Po-Chih Wu (2025). Low-quality image dataset [Dataset]. https://www.kaggle.com/datasets/pochihwu/low-quality-image-dataset
    Explore at:
    zip(11113072986 bytes)Available download formats
    Dataset updated
    May 15, 2025
    Authors
    Po-Chih Wu
    License

    MIT Licensehttps://opensource.org/licenses/MIT
    License information was derived automatically

    Description

    Description

    Due to the scarcity of suitable image datasets online related to low-quality images, we created a new dataset specifically for this purpose. The dataset can be used to develop or train models aimed at improving image quality, or serve as a benchmark dataset for evaluating the performance of computer vision on low-quality images. The image image processing code in this dataset is available at https://github.com/pochih-code/Low-quality-image-dataset

    Content

    Low-quality image dataset is based on the MS COCO 2017 validation images, with images processed into four categories, including lossy compression, image intensity, image noise and image blur. In total, the dataset comprises 100,000 processed images and is modified by humans to ensure that images are valid in the real world.

    Format

    • coco2017val image blur (25K images)
      • coco_average_4 (5K images)
      • coco_average_6 (5K images)
      • coco_average_8 (5K images)
      • coco_average_10 (5K images)
      • coco_average_12 (5K images)
    • coco2017val lossy compression (25K images)
      • coco_compressed_0 (5K images)
      • coco_compressed_20 (5K images)
      • coco_compressed_40 (5K images)
      • coco_compressed_60 (5K images)
      • coco_compressed_80 (5K images)
    • coco2017val gamma correction (25K images)
      • coco_gamma_4 (5K images)
      • coco_gamma_8 (5K images)
      • coco_gamma_12 (5K images)
      • coco_gamma_16 (5K images)
      • coco_gamma_20 (5K images)
    • coco2017val image noise (25K images)
      • coco_gaussian_10 (5K images)
      • coco_gaussian_20 (5K images)
      • coco_gaussian_30 (5K images)
      • coco_gaussian_40 (5K images)
      • coco_gaussian_50 (5K images) # Acknowledgements "https://cocodataset.org/#home">Microsoft COCO: Common Objects in Context
  5. Chemistry Lab Image Dataset Covering 25 Apparatus Categories

    • figshare.com
    application/x-rar
    Updated Aug 3, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Md. Sakhawat Hossain; Md. Sadman Haque; Md. Mostafizur Rahman; Md. Mosaddik Mashrafi Mousum; Zobaer Ibn Razzaque; Robiul Awoul Robin (2025). Chemistry Lab Image Dataset Covering 25 Apparatus Categories [Dataset]. http://doi.org/10.6084/m9.figshare.29110433.v3
    Explore at:
    application/x-rarAvailable download formats
    Dataset updated
    Aug 3, 2025
    Dataset provided by
    Figsharehttp://figshare.com/
    Authors
    Md. Sakhawat Hossain; Md. Sadman Haque; Md. Mostafizur Rahman; Md. Mosaddik Mashrafi Mousum; Zobaer Ibn Razzaque; Robiul Awoul Robin
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    This dataset contains 4,599 high-quality, annotated images of 25 commonly used chemistry lab apparatuses. The images, each containing structures in real-world settings, have been captured from different angles, backgrounds, and distances, while also undergoing variations in lighting to aid in the robustness of object detection models. Every image has been labeled using bounding box annotation in TXT (YOLO) format, alongside the class IDs and normalized bounding box coordinates, making object detection more precise. The annotations and bounding boxes have been built using the Roboflow platform.To achieve a better learning procedure, the dataset has been split into three sub-datasets: training, validation, and testing. The training dataset constitutes 70% of the entire dataset, with validation and testing at 20% and 10% respectively. In addition, all images undergo scaling to a standard of 640x640 pixels while being auto-oriented to rectify rotation discrepancies brought about by the EXIF metadata. The dataset is structured in three main folders - train, valid, and test, and each contains images/ and labels/ subfolders. Every image contains a label file containing class and bounding box data corresponding to each detected object.The whole dataset features 6,960 labeled instances per 25 apparatus categories including beakers, conical flasks, measuring cylinders, test tubes, among others. The dataset can be utilized for the development of automation systems, real-time monitoring and tracking systems, tools for safety monitoring, alongside AI educational tools.

  6. m

    iCubWorld28 - Full Images

    • data.mendeley.com
    Updated Aug 29, 2016
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Giulia Pasquale (2016). iCubWorld28 - Full Images [Dataset]. http://doi.org/10.17632/5txngd6td6.1
    Explore at:
    Dataset updated
    Aug 29, 2016
    Authors
    Giulia Pasquale
    License

    http://www.gnu.org/licenses/gpl-3.0.en.htmlhttp://www.gnu.org/licenses/gpl-3.0.en.html

    Description

    A complete description of this dataset is available at https://robotology.github.io/iCubWorld .

  7. m

    Human Faces and Objects Mix Image Dataset

    • data.mendeley.com
    Updated Mar 13, 2025
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Bindu Garg (2025). Human Faces and Objects Mix Image Dataset [Dataset]. http://doi.org/10.17632/nzwvnrmwp3.1
    Explore at:
    Dataset updated
    Mar 13, 2025
    Authors
    Bindu Garg
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Dataset Description: Human Faces and Objects Dataset (HFO-5000) The Human Faces and Objects Dataset (HFO-5000) is a curated collection of 5,000 images, categorized into three distinct classes: male faces (1,500), female faces (1,500), and objects (2,000). This dataset is designed for machine learning and computer vision applications, including image classification, face detection, and object recognition. The dataset provides high-quality, labeled images with a structured CSV file for seamless integration into deep learning pipelines.

    Column Description: The dataset is accompanied by a CSV file that contains essential metadata for each image. The CSV file includes the following columns: file_name: The name of the image file (e.g., image_001.jpg). label: The category of the image, with three possible values: "male" (for male face images) "female" (for female face images) "object" (for images of various objects) file_path: The full or relative path to the image file within the dataset directory.

    Uniqueness and Key Features: 1) Balanced Distribution: The dataset maintains an even distribution of human faces (male and female) to minimize bias in classification tasks. 2) Diverse Object Selection: The object category consists of a wide variety of items, ensuring robustness in distinguishing between human and non-human entities. 3) High-Quality Images: The dataset consists of clear and well-defined images, suitable for both training and testing AI models. 4) Structured Annotations: The CSV file simplifies dataset management and integration into machine learning workflows. 5) Potential Use Cases: This dataset can be used for tasks such as gender classification, facial recognition benchmarking, human-object differentiation, and transfer learning applications.

    Conclusion: The HFO-5000 dataset provides a well-structured, diverse, and high-quality set of labeled images that can be used for various computer vision tasks. Its balanced distribution of human faces and objects ensures fairness in training AI models, making it a valuable resource for researchers and developers. By offering structured metadata and a wide range of images, this dataset facilitates advancements in deep learning applications related to facial recognition and object classification.

  8. s

    Landmark Image Dataset

    • shaip.com
    • lb.shaip.com
    • +4more
    json
    Updated Nov 26, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Shaip (2024). Landmark Image Dataset [Dataset]. https://www.shaip.com/offerings/environment-scene-segmentation-datasets/
    Explore at:
    jsonAvailable download formats
    Dataset updated
    Nov 26, 2024
    Dataset authored and provided by
    Shaip
    License

    CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
    License information was derived automatically

    Description

    Images of landmarks within the context of their environment

  9. i

    Arm Image Classification Dataset

    • images.cv
    zip
    Updated Dec 19, 2021
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    (2021). Arm Image Classification Dataset [Dataset]. https://images.cv/dataset/arm-image-classification-dataset
    Explore at:
    zipAvailable download formats
    Dataset updated
    Dec 19, 2021
    License

    https://images.cv/licensehttps://images.cv/license

    Description

    Labeled Arm images suitable for training and evaluating computer vision and deep learning models.

  10. Face Re-identification Image Dataset

    • kaggle.com
    zip
    Updated Jul 7, 2025
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Unidata (2025). Face Re-identification Image Dataset [Dataset]. https://www.kaggle.com/datasets/unidpro/face-re-identification-image-dataset
    Explore at:
    zip(17758297 bytes)Available download formats
    Dataset updated
    Jul 7, 2025
    Authors
    Unidata
    License

    Attribution-NonCommercial-NoDerivs 4.0 (CC BY-NC-ND 4.0)https://creativecommons.org/licenses/by-nc-nd/4.0/
    License information was derived automatically

    Description

    Dataset of face images with different angles and head positions

    Dataset contains 23,110 individuals, each contributing 28 images featuring various angles and head positions, diverse backgrounds, and attributes, along with 1 ID photo. In total, the dataset comprises over 670,000 images in formats such as JPG and PNG. It is designed to advance face recognition and facial recognition research, focusing on person re-identification and recognition systems.

    By utilizing this dataset, researchers can explore various recognition applications, including face verification, face identification. - Get the data

    Metadata for the dataset

    https://www.googleapis.com/download/storage/v1/b/kaggle-user-content/o/inbox%2F22059654%2Fed374cc92b935b209749cb7b32fd41da%2FFrame%201%20(10).png?generation=1743160276352983&alt=media" alt=""> The accuracy of labels of face pose is more than 97%, ensuring reliable data for training and testing recognition algorithms.

    💵 Buy the Dataset: This is a limited preview of the data. To access the full dataset, please contact us at https://unidata.pro to discuss your requirements and pricing options.

    Dataset includes high-quality images that capture human faces in different poses and expressions, allowing for comprehensive analysis in recognition tasks. It is particularly valuable for developing and evaluating deep learning models and computer vision techniques.

    🌐 UniData provides high-quality datasets, content moderation, data collection and annotation for your AI/ML projects

  11. Mohamed Salah Image Dataset for Computer Vision

    • kaggle.com
    zip
    Updated Jun 5, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Selman Yalçın (2025). Mohamed Salah Image Dataset for Computer Vision [Dataset]. https://www.kaggle.com/datasets/selmanyaln/mohamed-salah-image-dataset-for-computer-vision
    Explore at:
    zip(961309 bytes)Available download formats
    Dataset updated
    Jun 5, 2025
    Authors
    Selman Yalçın
    License

    MIT Licensehttps://opensource.org/licenses/MIT
    License information was derived automatically

    Description

    This dataset contains 100 images of Mohamed Salah, curated for a computer vision project focused on YOLO-based object detection. The images showcase Salah in various contexts and angles, enabling robust model training and testing. This dataset can help anyone working on computer vision projects who wants to experiment with custom object detection tasks using YOLO. It is ideal for those exploring sports-related image recognition or fine-tuning YOLO models with domain-specific data.

  12. i

    Monitor Labeled Image Dataset

    • images.cv
    zip
    Updated Jan 20, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    (2022). Monitor Labeled Image Dataset [Dataset]. https://images.cv/dataset/monitor-image-classification-dataset
    Explore at:
    zipAvailable download formats
    Dataset updated
    Jan 20, 2022
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Labeled Monitor images suitable for AI and computer vision.

  13. i

    Desktop computer Labeled Image Dataset

    • images.cv
    zip
    Updated Dec 14, 2021
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    (2021). Desktop computer Labeled Image Dataset [Dataset]. https://images.cv/dataset/desktop-computer-image-classification-dataset
    Explore at:
    zipAvailable download formats
    Dataset updated
    Dec 14, 2021
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Labeled Desktop computer images suitable for AI and computer vision.

  14. Image Data (Object Detection and Captioning)

    • kaggle.com
    Updated Apr 15, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Arunesh (2024). Image Data (Object Detection and Captioning) [Dataset]. https://www.kaggle.com/datasets/aruneshhh/object-detection-images
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Apr 15, 2024
    Dataset provided by
    Kagglehttp://kaggle.com/
    Authors
    Arunesh
    License

    MIT Licensehttps://opensource.org/licenses/MIT
    License information was derived automatically

    Description

    🌟 Unlock the potential of advanced computer vision tasks with our comprehensive dataset comprising 15,000 high-quality images. Whether you're delving into segmentation, object detection, or image captioning, our dataset offers a diverse array of visual data to fuel your machine learning models.

    🔍 Our dataset is meticulously curated to encompass a wide range of streams, ensuring versatility and applicability across various domains. From natural landscapes to urban environments, from wildlife to everyday objects, our collection captures the richness and diversity of visual content.

    📊 Dataset Overview:

    Total ImagesTraining Set (70%)Testing Set (30%)
    15,00010,5004,500

    🔢 Image Details:

    • Format: JPG
    • Size Range: Approximately 150 to 300 KB per image

    Embark on your computer vision journey and leverage our dataset to develop cutting-edge algorithms, advance research, and push the boundaries of what's possible in visual recognition tasks. Join us in shaping the future of AI-powered image analysis.

  15. m

    iCubWorld Transformations - Full Images - 15x10 - Parts 2a and 2b

    • data.mendeley.com
    Updated Aug 30, 2016
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Giulia Pasquale (2016). iCubWorld Transformations - Full Images - 15x10 - Parts 2a and 2b [Dataset]. http://doi.org/10.17632/wgcdpj9pn9.1
    Explore at:
    Dataset updated
    Aug 30, 2016
    Authors
    Giulia Pasquale
    License

    http://www.gnu.org/licenses/gpl-3.0.en.htmlhttp://www.gnu.org/licenses/gpl-3.0.en.html

    Description

    A complete description of this dataset is available at https://robotology.github.io/iCubWorld .

  16. D

    A Dataset Documenting Representations of Machine Vision Technologies in...

    • dataverse.no
    • dataverse.azure.uit.no
    • +1more
    csv +3
    Updated Jun 2, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Jill Walker Rettberg; Jill Walker Rettberg; Linda Kronman; Linda Kronman; Ragnhild Solberg; Ragnhild Solberg; Marianne Gunderson; Marianne Gunderson; Stein Magne Bjørklund; Stein Magne Bjørklund; Linn Heidi Stokkedal; Gabriele de Seta; Gabriele de Seta; Kurdin Jacob; Annette Markham; Annette Markham; Linn Heidi Stokkedal; Kurdin Jacob (2025). A Dataset Documenting Representations of Machine Vision Technologies in Artworks, Games and Narratives [Dataset]. http://doi.org/10.18710/2G0XKN
    Explore at:
    txt(51965), csv(10398), csv(7553), csv(163630), type/x-r-syntax(13711), csv(54897), csv(590517), csv(39545396), csv(883829), text/comma-separated-values(13673), csv(9360), csv(102018)Available download formats
    Dataset updated
    Jun 2, 2025
    Dataset provided by
    DataverseNO
    Authors
    Jill Walker Rettberg; Jill Walker Rettberg; Linda Kronman; Linda Kronman; Ragnhild Solberg; Ragnhild Solberg; Marianne Gunderson; Marianne Gunderson; Stein Magne Bjørklund; Stein Magne Bjørklund; Linn Heidi Stokkedal; Gabriele de Seta; Gabriele de Seta; Kurdin Jacob; Annette Markham; Annette Markham; Linn Heidi Stokkedal; Kurdin Jacob
    License

    CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
    License information was derived automatically

    Time period covered
    Jan 1, 1897 - Oct 30, 2021
    Dataset funded by
    European Research Council
    Description

    This dataset captures cultural attitudes towards machine vision technologies as they are expressed in art, games and narratives. The dataset includes records of 500 creative works (including 77 digital games, 191 digital artworks and 236 movies, novels and other narratives) that use or represent machine vision technologies like facial recognition, deepfakes, and augmented reality. The dataset is divided into three main tables, relating to the works, to specific situations in each work involving machine vision technologies, and to the characters that interact with the technologies. Data about each work includes title, author, year and country of publication; types of machine vision technologies featured; topics the work addresses, and sentiments associated with that machine vision usage in the work. In the various works we identified 884 specific situations where machine vision is central. The dataset includes detailed data about each of these situations that describes the actions of human and non-human agents, including machine vision technologies. The dataset is the product of a digital humanities project and can be also viewed as a database at http://machine-vision.no. Data was collected by a team of topic experts who followed an analytical model developed to explore relationships between humans and technologies, inspired by posthumanist and feminist new materialist theories. The project team identified relevant works by searching databases, visiting exhibitions and conferences, reading scholarship, and consulting other experts. The inclusion criteria were creative works( art, games, narratives (movies, novels, etc)) where one of the following machine vision technologies was used in or represented by the work: 3D scans, AI, Augmented reality, Biometrics, Body scans, Camera, Cameraphone, Deepfake, Drones, Emotion recognition, Facial recognition, Filtering, Holograms, Image generation, Interactive panoramas Machine learning, MicroscopeOrTelescope Motion tracking, Non-Visible Spectrum Object recognition, Ocular implant, Satellite images, Surveillance cameras, UGV, Virtual reality, and Webcams. The dataset as well as the more detailed database can be viewed, searched, extracted, or otherwise used or reused and is considered particularly useful for humanities and social science scholars interested in the relationship between technology and culture, and by designers, artists, and scientists developing machine vision technologies.

  17. g

    Wild Animals Image Dataset

    • gts.ai
    json
    Updated Jul 1, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    GTS (2024). Wild Animals Image Dataset [Dataset]. https://gts.ai/dataset-download/wild-animals-image-dataset/
    Explore at:
    jsonAvailable download formats
    Dataset updated
    Jul 1, 2024
    Dataset provided by
    GLOBOSE TECHNOLOGY SOLUTIONS PRIVATE LIMITED
    Authors
    GTS
    License

    CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
    License information was derived automatically

    Description

    Explore our Wild Animals Dataset designed for AI development, featuring resized images of various species for classification, detection, and GAN applications.

  18. s

    Damaged Car Image Dataset

    • shaip.com
    json
    Updated Nov 26, 2024
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Shaip (2024). Damaged Car Image Dataset [Dataset]. https://www.shaip.com/offerings/machine-industry-datasets/
    Explore at:
    jsonAvailable download formats
    Dataset updated
    Nov 26, 2024
    Dataset authored and provided by
    Shaip
    License

    CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
    License information was derived automatically

    Description

    490+ cars and 3958 car photos with annotated images (along with metadata) of damaged cars. Covers all sides of the car (8 photos for each car) - Insurance Claim Process Use Cases.

  19. g

    Google reCAPTCHA Image Dataset

    • gts.ai
    json
    Updated Sep 2, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    GLOBOSE TECHNOLOGY SOLUTIONS PRIVATE LIMITED (2024). Google reCAPTCHA Image Dataset [Dataset]. https://gts.ai/dataset-download/page/66/
    Explore at:
    jsonAvailable download formats
    Dataset updated
    Sep 2, 2024
    Dataset authored and provided by
    GLOBOSE TECHNOLOGY SOLUTIONS PRIVATE LIMITED
    License

    CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
    License information was derived automatically

    Description

    Explore the Google reCAPTCHA V2 Image Dataset containing diverse object images for training object detection and classification models, including YOLO and other ML frameworks.

  20. d

    Selfie photo Dataset | 10M+ images | Global Coverage | Face Detection |...

    • datarade.ai
    .jpg, .jpeg, .png
    Updated Jul 25, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    FileMarket (2025). Selfie photo Dataset | 10M+ images | Global Coverage | Face Detection | Computer vision Data [Dataset]. https://datarade.ai/data-products/selfie-photo-dataset-10m-images-global-coverage-face-d-filemarket
    Explore at:
    .jpg, .jpeg, .pngAvailable download formats
    Dataset updated
    Jul 25, 2025
    Dataset authored and provided by
    FileMarket
    Area covered
    Senegal, South Georgia and the South Sandwich Islands, Uganda, Saint Kitts and Nevis, Djibouti, Honduras, New Zealand, Myanmar, Martinique, Chile
    Description

    Total Users 10,229,822 Total Pictures 10M+ (mostly 1 per ID)

    Gender: - Male 60% - Female 40%

    Ethnicity: - Asian 9% - African Decent 13% - East Indian 3% - Latino Hispanic 28% - Caucasian 47%

    Age Group: - 0-17 3% - 18-24 62% - 25-34 21% - 35-44 10% - 45-54 3% - 55+ 1%

    Top Phone Models: - iPhone 6s 9% - iPhone XR 6% - iPhone 6 6% - iPhone 7 (US/CDMA) 6% - iPhone 11 5% - iPhone 8 (US/CDMA) 4% (Total 141 device)

    Top Countries: - US 48.84% - GB 10.57% - CA 4.26% - AU 3.48% - FR 2.80% - SA 2.17% (Total 131 countries)

    Average resolution 5761024 px.

    All photos are collected with the consent of users.

Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
Ali Alameer; Mazin Al-Mosawy (2025). UoS Buildings Image Dataset for Computer Vision Algorithms [Dataset]. http://doi.org/10.17866/rd.salford.20383155.v2

UoS Buildings Image Dataset for Computer Vision Algorithms

Explore at:
application/x-gzipAvailable download formats
Dataset updated
Jan 23, 2025
Dataset provided by
University of Salford
Authors
Ali Alameer; Mazin Al-Mosawy
License

Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically

Description

The dataset for this project is represented by photos, photos for the buildings of the University of Salford, these photos are taken by a mobile phone camera from different angels and different distances , even though this task sounds so easy but it encountered some challenges, these challenges are summarized below: 1. Obstacles. a. Fixed or unremovable objects. When taking several photos for a building or a landscape from different angels and directions ,there are some of these angels blocked by a form of a fixed object such as trees and plants, light poles, signs, statues, cabins, bicycle shades, scooter stands, generators/transformers, construction barriers, construction equipment and any other service equipment so it is unavoidable to represent some photos without these objects included, this will raise 3 questions. - will these objects confuse the model/application we intend to create meaning will that obstacle prevent the model/application from identifying the designated building? - Or will the photos be more precise with these objects and provide the capability for the model/application to identify these building with these obstacles included? - How far is the maximum length for detection? In other words, how far will the mobile device with the application be from the building so it could or could not detect the designated building? b. Removable and moving objects. - Any University is crowded with staff and students especially in the rush hours of the day so it is hard for some photos to be taken without a personnel appearing in that photo in a certain time period of the day. But, due to privacy issues and showing respect to that person, these photos are better excluded. - Parked vehicles, trollies and service equipment can be an obstacle and might appear in these images as well as it can block access to some areas which an image from a certain angel cannot be obtained. - Animals, like dogs, cats, birds or even squirrels cannot be avoided in some photos which are entitled to the same questions above.
2. Weather. In a deep learning project, more data means more accuracy and less error, at this stage of our project it was agreed to have 50 photos per building but we can increase the number of photos for more accurate results but due to the limitation of time for this project it was agreed for 50 per building only. these photos were taken on cloudy days and to expand our work on this project (as future works and recommendations). Photos on sunny, rainy, foggy, snowy and any other weather condition days can be included. Even photos in different times of the day can be included such as night, dawn, and sunset times. To provide our designated model with all the possibilities to identify these buildings in all available circumstances.

  1. The selected buildings. It was agreed to select 10 buildings only from the University of Salford buildings for this project with at least 50 images per building, these selected building for this project with the number of images taken are:
  2. Chapman: 74 images
  3. Clifford Whitworth Library: 60 images
  4. Cockcroft: 67 images
  5. Maxwell: 80 images
  6. Media City Campus: 92 images
  7. New Adelphi: 93 images
  8. New Science, Engineering & Environment: 78 images
  9. Newton: 92 images
  10. Sports Centre: 55 images
  11. University House: 60 images Peel building is an important figure of the University of Salford due to its distinct and amazing exterior design but unfortunately it was excluded from the selection due to some maintenance activities at the time of collecting the photos for this project as it is partially covered with scaffolding and a lot of movement by personnel and equipment. If the supervisor suggests that this will be another challenge to include in the project then, it is mandatory to collect its photos. There are many other buildings in the University of Salford and again to expand our project in the future, we can include all the buildings of the University of Salford. The full list of buildings of the university can be reviewed by accessing an interactive map on: www.salford.ac.uk/find-us

  12. Expand Further. This project can be improved furthermore with so many capabilities, again due to the limitation of time given to this project , these improvements can be implemented later as future works. In simple words, this project is to create an application that can display the building’s name when pointing a mobile device with a camera to that building. Future featured to be added: a. Address/ location: this will require collection of additional data which is the longitude and latitude of each building included or the post code which will be the same taking under consideration how close these buildings appear on the interactive map application such as Google maps, Google earth or iMaps. b. Description of the building: what is the building for, by which school is this building occupied? and what facilities are included in this building? c. Interior Images: all the photos at this stage were taken for the exterior of the buildings, will interior photos make an impact on the model/application for example, if the user is inside newton or chapman and opens the application, will the building be identified especially the interior of these buildings have a high level of similarity for the corridors, rooms, halls, and labs? Will the furniture and assets will be as obstacles or identification marks? d. Directions to a specific area/floor inside the building: if the interior images succeed with the model/application, it would be a good idea adding a search option to the model/application so it can guide the user to a specific area showing directions to that area, for example if the user is inside newton building and searches for lab 141 it will direct him to the first floor of the building with an interactive arrow that changes while the user is approaching his destination. Or, if the application can identify the building from its interior, a drop down list will be activated with each floor of this building, for example, if the model/application identifies Newton building, the drop down list will be activated and when pressing on that drop down list it will represent interactive tabs for each floor of the building, selecting one of the floors by clicking on its tab will display the facilities on that floor for example if the user presses on floor 1 tab, another screen will appear displaying which facilities are on that floor. Furthermore, if the model/application identifies another building, it should activate a different number of floors as buildings differ in the number of floors from each other. this feature can be improved with a voice assistant that can direct the user after he applies his search (something similar to the voice assistant in Google maps but applied to the interior of the university’s buildings. e. Top View: if a drone with a camera can be afforded, it can provide arial images and top views for the buildings that can be added to the model/application but these images can be similar to the interior images situation , the buildings can be similar to each other from the top with other obstacles included like water tanks and AC units.

  13. Other Questions:

  14. Will the model/application be reproducible? the presumed answer for this question should be YES, IF, the model/application will be fed with the proper data (images) such as images of restaurants, schools, supermarkets, hospitals, government facilities...etc.

Search
Clear search
Close search
Google apps
Main menu