Android maintained its position as the leading mobile operating system worldwide in the first quarter of 2025 with a market share of about ***** percent. Android's closest rival, Apple's iOS, had a market share of approximately ***** percent during the same period. The leading mobile operating systems Both unveiled in 2007, Google’s Android and Apple’s iOS have evolved through incremental updates introducing new features and capabilities. The latest version of iOS, iOS 18, was released in September 2024, while the most recent Android iteration, Android 15, was made available in September 2023. A key difference between the two systems concerns hardware - iOS is only available on Apple devices, whereas Android ships with devices from a range of manufacturers such as Samsung, Google and OnePlus. In addition, Apple has had far greater success in bringing its users up to date. As of February 2024, ** percent of iOS users had iOS 17 installed, while in the same month only ** percent of Android users ran the latest version. The rise of the smartphone From around 2010, the touchscreen smartphone revolution had a major impact on sales of basic feature phones, as the sales of smartphones increased from *** million units in 2008 to **** billion units in 2023. In 2020, smartphone sales decreased to **** billion units due to the coronavirus (COVID-19) pandemic. Apple, Samsung, and lately also Xiaomi, were the big winners in this shift towards smartphones, with BlackBerry and Nokia among those unable to capitalize.
The number of smartphone users in the United States was forecast to continuously increase between 2024 and 2029 by in total 17.4 million users (+5.61 percent). After the fifteenth consecutive increasing year, the smartphone user base is estimated to reach 327.54 million users and therefore a new peak in 2029. Notably, the number of smartphone users of was continuously increasing over the past years.Smartphone users here are limited to internet users of any age using a smartphone. The shown figures have been derived from survey data that has been processed to estimate missing demographics.The shown data are an excerpt of Statista's Key Market Indicators (KMI). The KMI are a collection of primary and secondary indicators on the macro-economic, demographic and technological environment in up to 150 countries and regions worldwide. All indicators are sourced from international and national statistical offices, trade associations and the trade press and they are processed to generate comparable data sets (see supplementary notes under details for more information).Find more key insights for the number of smartphone users in countries like Mexico and Canada.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
CSV file with code smell occurrences per application. One file for iOS and one for Android. Analysis of open source applications.
Context Imagine you have to process bug reports about an application your company is developing, which is available for both Android and iOS. Could you find a way to automatically classify them so you can send them to the right support team?
Content The dataset contains data from two StackExchange forums: Android Enthusiasts and Ask Differently (Apple). I pre-processed both datasets from the raw XML files retrieved from Internet Archive in order to only contain useful information for building Machine Learning classifiers. In the case of the Apple forum, I narrowed down to the subset of questions that have one of the following tags: "iOS", "iPhone", "iPad".
Think of this as a fun way to learn to build ML classifiers! The training, validation and test sets are all available, but in order to build robust models please try to use the test set as little as possible (only as a last validation for your models).
Acknowledgements The image was retrieved from unsplash and made by @thenewmalcolm. Link to image here.
The data was made available for free under a CC-BY-SA 4.0 license by StackExchange and hosted by Internet Archive. Find it here.
CC-BY-SA
Original Data Source: Question Classification: Android or iOS?
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
This is a public release of Beiwe-generated data. The Beiwe Research Platform collects high-density data from a variety of smartphone sensors such as GPS, WiFi, Bluetooth, gyroscope, and accelerometer in addition to metadata from active surveys. A description of passive and active data streams, and a documentation concerning the use of Beiwe can be found here. This data was collected from an internal test study and is made available solely for educational purposes. It contains no identifying information; subject locations are de-identified using the noise GPS feature of Beiwe.
As part of the internal test study, data from 6 participants were collected from the start of March 21, 2022 to the end of March 28, 2022. The local time zone of this study is Eastern Standard Time. Each participant was notified to complete a survey at 9am EST on Monday, Thursday, and Saturday of the study week. An additional survey was administered on Tuesday at 5:15pm EST. For each survey, subjects were asked to respond to the prompt "How much time (in hours) do you think you spent at home?".
Attribution-ShareAlike 4.0 (CC BY-SA 4.0)https://creativecommons.org/licenses/by-sa/4.0/
License information was derived automatically
Casual smartphone 3D scans using three devices: iPhone 15 Pro, Samsung Galaxy S20 FE and Google Pixel 5. The raw data (prefixed "spectacular-rec-") has been captured with the Spectacular Rec applications for Android and iOS, and contains time-synchronized video and IMU data. The extras file also contains AprilGrid calibration sequences, as well as pre-computed calibration results, for the Android devices.
The data is captured in non-ideal lighting conditions and has a moderate amount of motion blur and rolling shutter artefacts. The included metadata also contains the exposure times and the (Android) rolling shutter readout times, as well as the built-in calibration data, as reported by the devices.
The dataset also contains three different processed variants (prefixed with "colmap-"), which are directly trainable with Nerfstudio. In the processed variants, suitable minimally blurry video frames have been selected as key frames and their poses have been registered with COLMAP. In addition, the local linear and angular velocities of each key frame has been estimated using VIO with Spectacular AI Mapping Tools. The "calib-intrinsics" and "orig-intrinsics" variants include manually calibrated and built-in intrinsics, respectively. They depend on the third variant with symbolic links. The third variant has COLMAP-estimated intrinsics, which are relatively inaccurate for the Android data with high levels rolling shutter deformation.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
App Permission data of 2.2 million android applications from Google Play store. Backup repo: https://github.com/gauthamp10/android-permissions-dataset
I've collected the data with the help of Python and Scrapy running on a cloud virtual machine with the United States as geolocation. The data was collected on June 2021.
Also checkout:
I couldn't have build this dateset without the help of Digitalocean and github. Switched to facundoolano/google-play-scraper for sane reasons.
Took inspiration from: https://www.kaggle.com/gauthamp10/google-playstore-apps to build a big database for students and researchers who are interested to analyze and find insights on mobile application privacy.
Gautham Prakash
My other projects: github.com/gauthamp10
Website: gauthamp10.github.io
This is a GPS dataset acquired from Google.
Google tracks the user’s device location through Google Maps, which also works on Android devices, the iPhone, and the web.
It’s possible to see the Timeline from the user’s settings in the Google Maps app on Android or directly from the Google Timeline Website.
It has detailed information such as when an individual is walking, driving, and flying.
Such functionality of tracking can be enabled or disabled on demand by the user directly from the smartphone or via the website.
Google has a Take Out service where the users can download all their data or select from the Google products they use the data they want to download.
The dataset contains 120,847 instances from a period of 9 months or 253 unique days from February 2019 to October 2019 from a single user.
The dataset comprises a pair of (latitude, and longitude), and a timestamp.
All the data was delivered in a single CSV file.
As the locations of this dataset are well known by the researchers, this dataset will be used as ground truth in many mobility studies.
Please cite the following papers in order to use the datasets:
T. Andrade, B. Cancela, and J. Gama, "Discovering locations and habits from human mobility data," Annals of Telecommunications, vol. 75, no. 9, pp. 505–521, 2020.
10.1007/s12243-020-00807-x (DOI)
and
T. Andrade, B. Cancela, and J. Gama, "From mobility data to habits and common pathways," Expert Systems, vol. 37, no. 6, p. e12627, 2020.
10.1111/exsy.12627 (DOI)
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
A large-scale mobile typing dataset contains 46 755 participants typing sentences in English and 8661 participants in Finnish on their own mobile devices. Participants used various iPhone and Android devices with different operation system versions. The data was collected between 2019 and 2020 by the Computational Behaviour Lab of Aalto University. The user's typing operations and use of Intelligent Text Entry (ITE) methods (Autocorrection and Suggestion Bar) are labelled on a keystroke level. The dataset enables analysis of the effects of the user demographics and the usage and accuracy of ITE methods on typing. The dataset also has a separate table for all ITE corrected and predicted words e.g. for the ITE error analysis.
Code repository: https://github.com/aalto-speech/ite-typing-dataset/
Citation:
Leino, Katri, Markku Laine, Mikko Kurimo, and Antti Oulasvirta. Mobile Typing with Intelligent Text Entry: A Large-Scale Dataset and Results. 2024. https://doi.org/10.21203/rs.3.rs-4654512/v1
Infant Crying Smartphone speech dataset, collected by Android smartphone and iPhone, covering infant crying. Our dataset was collected from extensive and diversify speakers(201 people in total, with balanced gender distribution), geographicly speaking, enhancing model performance in real and complex tasks. Quality tested by various AI companies. We strictly adhere to data protection regulations and privacy standards, ensuring the maintenance of user privacy and legal rights throughout the data collection, storage, and usage processes, our datasets are all GDPR, CCPA, PIPL complied.
Infant Crying smartphone speech dataset, collected by Android smartphone and iPhone, covering infant crying. Our dataset was collected from extensive and diversify speakers(201 people in total, with balanced gender distribution), geographicly speaking, enhancing model performance in real and complex tasks. Quality tested by various AI companies. We strictly adhere to data protection regulations and privacy standards, ensuring the maintenance of user privacy and legal rights throughout the data collection, storage, and usage processes, our datasets are all GDPR, CCPA, PIPL complied.
Population distribution : race distribution: Asians, Caucasians, black people; gender distribution: gender balance; age distribution: from child to the elderly, the young people and the middle aged are the majorities
Collection environment : indoor scenes, outdoor scenes
Collection diversity : various postures, expressions, light condition, scenes, time periods and distances
Collection device : iPhone, android phone, iPad
Collection time : daytime,night
Image Parameter : the video format is .mov or .mp4, the image format is .jpg
Accuracy : the accuracy of actions exceeds 97%
https://www.futurebeeai.com/policies/ai-data-license-agreementhttps://www.futurebeeai.com/policies/ai-data-license-agreement
Introducing the English Product Image Dataset - a diverse and comprehensive collection of images meticulously curated to propel the advancement of text recognition and optical character recognition (OCR) models designed specifically for the English language.
Dataset Contain & Diversity:Containing a total of 2000 images, this English OCR dataset offers diverse distribution across different types of front images of Products. In this dataset, you'll find a variety of text that includes product names, taglines, logos, company names, addresses, product content, etc. Images in this dataset showcase distinct fonts, writing formats, colors, designs, and layouts.
To ensure the diversity of the dataset and to build a robust text recognition model we allow limited (less than five) unique images from a single resource. Stringent measures have been taken to exclude any personally identifiable information (PII) and to ensure that in each image a minimum of 80% of space contains visible English text.
Images have been captured under varying lighting conditions – both day and night – along with different capture angles and backgrounds, to build a balanced OCR dataset. The collection features images in portrait and landscape modes.
All these images were captured by native English people to ensure the text quality, avoid toxic content and PII text. We used the latest iOS and Android mobile devices above 5MP cameras to click all these images to maintain the image quality. In this training dataset images are available in both JPEG and HEIC formats.
Metadata:Along with the image data, you will also receive detailed structured metadata in CSV format. For each image, it includes metadata like image orientation, county, language, and device information. Each image is properly renamed corresponding to the metadata.
The metadata serves as a valuable tool for understanding and characterizing the data, facilitating informed decision-making in the development of English text recognition models.
Update & Custom Collection:We're committed to expanding this dataset by continuously adding more images with the assistance of our native English crowd community.
If you require a custom product image OCR dataset tailored to your guidelines or specific device distribution, feel free to contact us. We're equipped to curate specialized data to meet your unique needs.
Furthermore, we can annotate or label the images with bounding box or transcribe the text in the image to align with your specific project requirements using our crowd community.
License:This Image dataset, created by FutureBeeAI, is now available for commercial use.
Conclusion:Leverage the power of this product image OCR dataset to elevate the training and performance of text recognition, text detection, and optical character recognition models within the realm of the English language. Your journey to enhanced language understanding and processing starts here.
This is the dataset that I used in my iOS and Android plant disease detection app, PlantifyDr. You can check out my full open-source project here: https://github.com/lavaman131/PlantifyDr
The dataset contains over 125,000 jpg images of 10 different plant types: Apple, Bell pepper, Cherry, Citrus, Corn, Grape, Peach, Potato, Strawberry, and Tomato. The total number of plant diseases is 37. Augmentations have already been applied to the data, but feel free to add your own augmentations if you like.
Special thanks to: https://data.mendeley.com/datasets/tywbtsjrjv/1 https://www.kaggle.com/vipoooool/new-plant-diseases-dataset https://github.com/pratikkayal/PlantDoc-Dataset https://data.mendeley.com/datasets/3f83gxmv57/2
for the data.
The Food and Agriculture Organization of the United Nations (FAO) estimates that annually between 20 to 40 percent of global crop production is lost. Each year, plant diseases cost the global economy around $220 billion. I hoped to use deep learning to solve this problem and be able to better educate farmers and the public with the necessary knowledge to treat their plants.
Attribution-NonCommercial 4.0 (CC BY-NC 4.0)https://creativecommons.org/licenses/by-nc/4.0/
License information was derived automatically
Data abstract: This Zenodo upload contains the ADVIO data for benchmarking and developing visual-inertial odometry methods. The data documentation is available on Github: https://github.com/AaltoVision/ADVIO
Paper abstract: The lack of realistic and open benchmarking datasets for pedestrian visual-inertial odometry has made it hard to pinpoint differences in published methods. Existing datasets either lack a full six degree-of-freedom ground-truth or are limited to small spaces with optical tracking systems. We take advantage of advances in pure inertial navigation, and develop a set of versatile and challenging real-world computer vision benchmark sets for visual-inertial odometry. For this purpose, we have built a test rig equipped with an iPhone, a Google Pixel Android phone, and a Google Tango device. We provide a wide range of raw sensor data that is accessible on almost any modern-day smartphone together with a high-quality ground-truth track. We also compare resulting visual-inertial tracks from Google Tango, ARCore, and Apple ARKit with two recent methods published in academic forums. The data sets cover both indoor and outdoor cases, with stairs, escalators, elevators, office environments, a shopping mall, and metro station.
Attribution: If you use this data set in your own work, please cite this paper:
Santiago Cortés, Arno Solin, Esa Rahtu, and Juho Kannala (2018). ADVIO: An authentic dataset for visual-inertial odometry. In European Conference on Computer Vision (ECCV). Munich, Germany.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
The dataset to be published was generated through exploratory case studies conducted on wrist-worn devices from three vendors: Huawei, Amazfit, and Xiaomi. The specific devices investigated include the Huawei Fit 2 Smartwatch and Band 7, Amazfit Band 7, and Xiaomi Watch 3. These devices operate on different operating systems, namely Android Wear, Zepp OS, and Wear OS.
The data collection period for each device varies, with Huawei having approximately one year of data collected, while the other devices have shorter durations. All wrist-wear devices from different vendors were connected to an iPhone 11 mobile device, which acted as the host device. The iPhone facilitated data synchronization and provided access to the data through the respective health applications provided by the vendors.
To extract the data, MD-NEXT was employed, and the extracted data was further analyzed using the MD-RED tool. These tools were chosen due to their recognized forensically sound capabilities. As a result, the dataset contains data that is considered suitable for use in digital forensics fields.
Overall, the dataset provides valuable information obtained from wrist-worn devices, covering multiple vendors, operating systems, and data collection periods. Researchers in the digital forensics field can utilize this dataset for various investigative and analytical purposes.
ScreenSpot Evaluation Benchmark ScreenSpot is an evaluation benchmark for GUI grounding, comprising over 1,200 instructions from various environments, including iOS, Android, macOS, Windows, and Web. Each data point includes annotated element types (Text or Icon/Widget). For more details and examples, please refer to our paper.
Test Sample Details Each test sample includes:
img_filename: The interface screenshot file. instruction: Human-provided instruction. bbox: The bounding box of the target element corresponding to the instruction. data_type: The type of the target element, either "icon" or "text". data_source: The interface platform, which could be iOS, Android, macOS, Windows, or Web (e.g., GitLab, Shop, Forum, Tool).
https://www.zionmarketresearch.com/privacy-policyhttps://www.zionmarketresearch.com/privacy-policy
Global Cloud Mobile Backend as a Service (BaaS) Market size was $3.0 Billion in 2022 and is slated to hit $7.3 Billion by the end of 2030 with a CAGR of nearly 24.1%.
The SYRCityline requests is a very large dataset This includes requests for city services which have been made by residents through SYRCityline, which utilizes SeeClickFix software. Service requests can be made at:https://seeclickfix.com/syracuseData DictionaryCreated at local - When this complaint or service was requested. (This is also in the format of MM/DD/YYYY - HH:MM(AM/PM)).Address - Address of the service request or complaint, provided by the community member.Summary - String that users select to categorize the nature of their complaint. Can be either Large or Bulk Items, Illegal Setouts, Sewer Back-ups, Weekly Trash Pickup, Large or Bulk Items - Skipped Pickup, Home & Building Maintenance, Sewer-related Concerns, Recyclling, Other Housing & Property Maintenance Concern, Streeth Lights, or Other.Rating - The number of followers on the Request in SeeClickFix.Description - Write up of the service request or complaint, provided by the community member.Agency Name - What type of City Department was this complaint assigned to. These include:Streets, Sidewalks & TransportationGarbage, Recycling & GraffitiHousing & Property MaintenanceFeedback to the CityParking & VehiclesGreen Spaces, Trees & Public UtilitiesWater & SewageAnimalsURL - Unique website address (url) that the complaint as well as comments from the City personnel can be viewed at.Latitude - Latitude GPS coordinate where the address is.Longitude - Longitude GPS coordinate where the address is.Export tagged places - Which quadrant of the city is this address matched to (Northeast, Southeast, Northwest, or Southwest).Acknowledged at local - When this complaint or service request was acknowledged by the City department.Closed at local - When this complaint or service request was marked as being closed by the City department.Minutes to acknowledged - The amount of time, in minutes, after it was Created at Local to being marked Acknowledged at local.Minutes to closed - The amount of time, in minutes, after service request was created at local to when it was marked as Closed at local.Assignee name - Which city Department was assigned to this request.Category - How was this request categorized. This can be Potholes, Large or Bulk Items, Water-related Concerns, Home & Building Maintenance, Street Lights, Weekly Trash Pickup, Public Trash Can, Yard Waste, Report Litter on Private Land, among other categories.SLA Limit - This is the limit assigned by the City of Syracuse, that puts a limit on how a request can stay in the list of tasks untouched. That amount of time, in hours, SeeClickFix will forward the request to the department head as well as an administrator to help ensure that requests are addressed in a timely manner.Report source - How this service request was obtained: Either Web-Mobile, iPhone, Portal, Web-Desktop, Android, or Request Form.Dataset OwnerOrganization: Department of Public Works (DPW)Position: Data Program ManagerCity: Syracuse, NYE-Mail Address:opendata@syrgov.net
Recording environment : quiet indoor environment, without echo Recording content (read speech) : general category; human-machine interaction category
Demographics : Speakers are evenly distributed across all age groups, covering children, teenagers, middle-aged, elderly, etc.
Device : Android mobile phone, iPhone;
Language : English-Korean, English-Japanese, German-English, Hong Kong Cantonese-English, Taiwanese-English,
Application scenarios : speech recognition; voiceprint recognition.
Accuracy rate : 97%
Android maintained its position as the leading mobile operating system worldwide in the first quarter of 2025 with a market share of about ***** percent. Android's closest rival, Apple's iOS, had a market share of approximately ***** percent during the same period. The leading mobile operating systems Both unveiled in 2007, Google’s Android and Apple’s iOS have evolved through incremental updates introducing new features and capabilities. The latest version of iOS, iOS 18, was released in September 2024, while the most recent Android iteration, Android 15, was made available in September 2023. A key difference between the two systems concerns hardware - iOS is only available on Apple devices, whereas Android ships with devices from a range of manufacturers such as Samsung, Google and OnePlus. In addition, Apple has had far greater success in bringing its users up to date. As of February 2024, ** percent of iOS users had iOS 17 installed, while in the same month only ** percent of Android users ran the latest version. The rise of the smartphone From around 2010, the touchscreen smartphone revolution had a major impact on sales of basic feature phones, as the sales of smartphones increased from *** million units in 2008 to **** billion units in 2023. In 2020, smartphone sales decreased to **** billion units due to the coronavirus (COVID-19) pandemic. Apple, Samsung, and lately also Xiaomi, were the big winners in this shift towards smartphones, with BlackBerry and Nokia among those unable to capitalize.