42 datasets found
  1. SQLite Sakila Sample Database

    • kaggle.com
    zip
    Updated Mar 14, 2021
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Atanas Kanev (2021). SQLite Sakila Sample Database [Dataset]. https://www.kaggle.com/atanaskanev/sqlite-sakila-sample-database
    Explore at:
    zip(4495190 bytes)Available download formats
    Dataset updated
    Mar 14, 2021
    Authors
    Atanas Kanev
    Description

    SQLite Sakila Sample Database

    Database Description

    The Sakila sample database is a fictitious database designed to represent a DVD rental store. The tables of the database include film, film_category, actor, customer, rental, payment and inventory among others. The Sakila sample database is intended to provide a standard schema that can be used for examples in books, tutorials, articles, samples, and so forth. Detailed information about the database can be found on the MySQL website: https://dev.mysql.com/doc/sakila/en/

    Sakila for SQLite is a part of the sakila-sample-database-ports project intended to provide ported versions of the original MySQL database for other database systems, including:

    • Oracle
    • SQL Server
    • SQLIte
    • Interbase/Firebird
    • Microsoft Access

    Sakila for SQLite is a port of the Sakila example database available for MySQL, which was originally developed by Mike Hillyer of the MySQL AB documentation team. This project is designed to help database administrators to decide which database to use for development of new products The user can run the same SQL against different kind of databases and compare the performance

    License: BSD Copyright DB Software Laboratory http://www.etl-tools.com

    Note: Part of the insert scripts were generated by Advanced ETL Processor http://www.etl-tools.com/etl-tools/advanced-etl-processor-enterprise/overview.html

    Information about the project and the downloadable files can be found at: https://code.google.com/archive/p/sakila-sample-database-ports/

    Other versions and developments of the project can be found at: https://github.com/ivanceras/sakila/tree/master/sqlite-sakila-db

    https://github.com/jOOQ/jOOQ/tree/main/jOOQ-examples/Sakila

    Direct access to the MySQL Sakila database, which does not require installation of MySQL (queries can be typed directly in the browser), is provided on the phpMyAdmin demo version website: https://demo.phpmyadmin.net/master-config/

    Files Description

    The files in the sqlite-sakila-db folder are the script files which can be used to generate the SQLite version of the database. For convenience, the script files have already been run in cmd to generate the sqlite-sakila.db file, as follows:

    sqlite> .open sqlite-sakila.db # creates the .db file sqlite> .read sqlite-sakila-schema.sql # creates the database schema sqlite> .read sqlite-sakila-insert-data.sql # inserts the data

    Therefore, the sqlite-sakila.db file can be directly loaded into SQLite3 and queries can be directly executed. You can refer to my notebook for an overview of the database and a demonstration of SQL queries. Note: Data about the film_text table is not provided in the script files, thus the film_text table is empty. Instead the film_id, title and description fields are included in the film table. Moreover, the Sakila Sample Database has many versions, so an Entity Relationship Diagram (ERD) is provided to describe this specific version. You are advised to refer to the ERD to familiarise yourself with the structure of the database.

  2. classicmodels

    • kaggle.com
    zip
    Updated Dec 10, 2022
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Marta Tavares (2022). classicmodels [Dataset]. https://www.kaggle.com/datasets/martatavares/classicmodels
    Explore at:
    zip(72431 bytes)Available download formats
    Dataset updated
    Dec 10, 2022
    Authors
    Marta Tavares
    Description

    MySQL Classicmodels sample database

    The MySQL sample database schema consists of the following tables:

    • Customers: stores customer’s data.
    • Products: stores a list of scale model cars.
    • ProductLines: stores a list of product line categories.
    • Orders: stores sales orders placed by customers.
    • OrderDetails: stores sales order line items for each sales order.
    • Payments: stores payments made by customers based on their accounts.
    • Employees: stores all employee information as well as the organization structure such as who reports to whom.
    • Offices: stores sales office data.

    https://www.googleapis.com/download/storage/v1/b/kaggle-user-content/o/inbox%2F8652778%2Fefc56365be54c0e2591a1aefa5041f36%2FMySQL-Sample-Database-Schema.png?generation=1670498341027618&alt=media" alt="">

  3. Employees

    • kaggle.com
    zip
    Updated Nov 12, 2021
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Sudhir Singh (2021). Employees [Dataset]. https://www.kaggle.com/datasets/crepantherx/employees
    Explore at:
    zip(31992550 bytes)Available download formats
    Dataset updated
    Nov 12, 2021
    Authors
    Sudhir Singh
    Description

    Dataset

    This dataset was created by Sudhir Singh

    Released under Data files © Original Authors

    Contents

  4. Z

    Rediscovery Datasets: Connecting Duplicate Reports of Apache, Eclipse, and...

    • nde-dev.biothings.io
    Updated Aug 3, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Miranskyy, Andriy V. (2024). Rediscovery Datasets: Connecting Duplicate Reports of Apache, Eclipse, and KDE [Dataset]. https://nde-dev.biothings.io/resources?id=zenodo_400614
    Explore at:
    Dataset updated
    Aug 3, 2024
    Dataset provided by
    Bener, Ayse Basar
    Miranskyy, Andriy V.
    Sadat, Mefta
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    We present three defect rediscovery datasets mined from Bugzilla. The datasets capture data for three groups of open source software projects: Apache, Eclipse, and KDE. The datasets contain information about approximately 914 thousands of defect reports over a period of 18 years (1999-2017) to capture the inter-relationships among duplicate defects.

    File Descriptions

    apache.csv - Apache Defect Rediscovery dataset

    eclipse.csv - Eclipse Defect Rediscovery dataset

    kde.csv - KDE Defect Rediscovery dataset

    apache.relations.csv - Inter-relations of rediscovered defects of Apache

    eclipse.relations.csv - Inter-relations of rediscovered defects of Eclipse

    kde.relations.csv - Inter-relations of rediscovered defects of KDE

    create_and_populate_neo4j_objects.cypher - Populates Neo4j graphDB by importing all the data from the CSV files. Note that you have to set dbms.import.csv.legacy_quote_escaping configuration setting to false to load the CSV files as per https://neo4j.com/docs/operations-manual/current/reference/configuration-settings/#config_dbms.import.csv.legacy_quote_escaping

    create_and_populate_mysql_objects.sql - Populates MySQL RDBMS by importing all the data from the CSV files

    rediscovery_db_mysql.zip - For your convenience, we also provide full backup of the MySQL database

    neo4j_examples.txt - Sample Neo4j queries

    mysql_examples.txt - Sample MySQL queries

    rediscovery_eclipse_6325.png - Output of Neo4j example #1

    distinct_attrs.csv - Distinct values of bug_status, resolution, priority, severity for each project

  5. d

    TWIW database dump

    • data.dtu.dk
    txt
    Updated Jul 10, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Sidsel Nag; Gunhild Larsen; Judit Szarvas; Laura Elmlund Kohl Birkedahl; Gábor Máté Gulyás; Wojciech Jakub Ciok; Timmie M. R. Lagermann; Silva Tafaj; Susan Bradbury; Peter Collignon; Denise Daley; Victorien Dougnon; Kafayath Fabiyi; Boubacar Coulibaly; René Dembélé; Georgette Nikiema; Natama Magloire; Isidore Juste Ouindgueta; Zenat Zebin Hossain; Anowara Begum; Deyan Donchev; Mathew Diggle; LeeAnn Turnbull; Simon Lévesque; Livia Berlinger; Kirstine Kobberoe Søgaard; Paula Diaz Guevara; Carolina Duarte Valderrama; Panagiota Maikanti; Jana Amlerova; Pavel Drevinek; Jan Tkadlec; Milica Dilas; Achim J. Kaasch; HenrikTorkil Westh; Mohamed Azzedine Bachtarzi; Wahiba Amhis; Carolina Elizabeth Satán Salazar; José Eduardo Villacis; Mária Angeles Dominguez Lúzon; Dàmaris Berbel Palau; Claire Duployez; Maxime Paluch; Solomon Asante-Sefa; Mie Møller; Margaret Ip; Ivana Marecović; Agnes Pál-Sonnevend; Clementiza Elvezia Cocuzza; Asta Dambrauskiene; Alexandre Macanze; Anelsio Cossa; Inácio Mandomando; Philip Nwajiobi-Princewill; Iruka N. Okeke; Aderemi O. Kehinde; Ini Adebiyi; Ifeoluwa Akintayo; Oluwafemi Popoola; Anthony Onipede; Anita Blomfeldt; Nora Elisabeth Nyquist; Kiri Bocker; James Ussher; Amjad Ali; Nimat Ullah; Habibullah Khan; Natalie Weiler Gustafson; Ikhlas Jarrar; Arif Al-Hamad; Viravarn Luvira; Wantana Paveenkittiporn; Irmak Baran; James C. L. Mwansa; Linda Sikakwa; Kaunda Yamba; Rene Sjøgren Hendriksen; Frank Møller Aarestrup (2023). TWIW database dump [Dataset]. http://doi.org/10.11583/DTU.21758456.v2
    Explore at:
    txtAvailable download formats
    Dataset updated
    Jul 10, 2023
    Dataset provided by
    Technical University of Denmark
    Authors
    Sidsel Nag; Gunhild Larsen; Judit Szarvas; Laura Elmlund Kohl Birkedahl; Gábor Máté Gulyás; Wojciech Jakub Ciok; Timmie M. R. Lagermann; Silva Tafaj; Susan Bradbury; Peter Collignon; Denise Daley; Victorien Dougnon; Kafayath Fabiyi; Boubacar Coulibaly; René Dembélé; Georgette Nikiema; Natama Magloire; Isidore Juste Ouindgueta; Zenat Zebin Hossain; Anowara Begum; Deyan Donchev; Mathew Diggle; LeeAnn Turnbull; Simon Lévesque; Livia Berlinger; Kirstine Kobberoe Søgaard; Paula Diaz Guevara; Carolina Duarte Valderrama; Panagiota Maikanti; Jana Amlerova; Pavel Drevinek; Jan Tkadlec; Milica Dilas; Achim J. Kaasch; HenrikTorkil Westh; Mohamed Azzedine Bachtarzi; Wahiba Amhis; Carolina Elizabeth Satán Salazar; José Eduardo Villacis; Mária Angeles Dominguez Lúzon; Dàmaris Berbel Palau; Claire Duployez; Maxime Paluch; Solomon Asante-Sefa; Mie Møller; Margaret Ip; Ivana Marecović; Agnes Pál-Sonnevend; Clementiza Elvezia Cocuzza; Asta Dambrauskiene; Alexandre Macanze; Anelsio Cossa; Inácio Mandomando; Philip Nwajiobi-Princewill; Iruka N. Okeke; Aderemi O. Kehinde; Ini Adebiyi; Ifeoluwa Akintayo; Oluwafemi Popoola; Anthony Onipede; Anita Blomfeldt; Nora Elisabeth Nyquist; Kiri Bocker; James Ussher; Amjad Ali; Nimat Ullah; Habibullah Khan; Natalie Weiler Gustafson; Ikhlas Jarrar; Arif Al-Hamad; Viravarn Luvira; Wantana Paveenkittiporn; Irmak Baran; James C. L. Mwansa; Linda Sikakwa; Kaunda Yamba; Rene Sjøgren Hendriksen; Frank Møller Aarestrup
    License

    Attribution-ShareAlike 4.0 (CC BY-SA 4.0)https://creativecommons.org/licenses/by-sa/4.0/
    License information was derived automatically

    Description

    Two Weeks in the World is a global research collaboration which seeks to shed light on various aspects of antimicrobial resistance. The research project has resulted in a dataset of 3100 clinically relevant bacterial genomes with pertaining metadata. “Clinically relevant” refers to the fact that the bacteria from which the genomes were obtained, were all concluded as being a cause of clinical manifestations of infection. The metadata refers to the data describing the infection from which the bacteria was obtained, like geographic origin and approximate collection date. The bacteria were collected from 59 microbiological diagnostic units in 35 countries around the world during 2020. The data from the project consists of tabular data and genomic sequence data. The tabular data is available as a mysql dump (relational database) and as csv files. The tabular data includes the infection metadata, the results from bioinformatic analyses (species prediction, identification of acquired resistance genes and phylogenetic analysis) as well as the pertaining accession numbers of the individual genomic sequence data, which are available through the European Nucleotide Archive (ENA). At time of submission, the project also has a dedicated web app, from which data can be browsed and downloaded: https://twiw.genomicepidemiology.org/ This complete dataset is created and shared according to the FAIR principles and has large reuse potential within the research fields of antimicrobial resistance, clinical microbiology and global health.

    .v2: Author list and readme has been updated. And a file containing column descriptions, for the database dump, has been added: TWIW_dbcolumns_explained.csv.

  6. Most popular database management systems worldwide 2026

    • statista.com
    Updated Jan 5, 2026
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Statista (2026). Most popular database management systems worldwide 2026 [Dataset]. https://www.statista.com/statistics/809750/worldwide-popularity-ranking-database-management-systems/
    Explore at:
    Dataset updated
    Jan 5, 2026
    Dataset authored and provided by
    Statistahttp://statista.com/
    Time period covered
    Jan 2026
    Area covered
    Worldwide
    Description

    As of January 2026, the most popular database management system (DBMS) worldwide was Oracle, with a ranking score of *****; MySQL and Microsoft SQL Server rounded out the top three. Although the database management industry contains some of the largest companies in the tech industry, such as Microsoft, Oracle and IBM, a number of free and open-source DBMSs such as PostgreSQL and MariaDB remain competitive. Database Management Systems As the name implies, DBMSs provide a platform through which developers can organize, update, and control large databases. Given the business world’s growing focus on big data and data analytics, knowledge of SQL programming languages has become an important asset for software developers around the world, and database management skills are seen as highly desirable. In addition to providing developers with the tools needed to operate databases, DBMS are also integral to the way that consumers access information through applications, which further illustrates the importance of the software.

  7. Chinook Database

    • kaggle.com
    zip
    Updated Nov 7, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Rana Sabry (2023). Chinook Database [Dataset]. https://www.kaggle.com/datasets/ranasabrii/chinook
    Explore at:
    zip(448874 bytes)Available download formats
    Dataset updated
    Nov 7, 2023
    Authors
    Rana Sabry
    Description

    The Chinook database was created as an alternative to the Northwind database. It represents a digital media store, including tables for artists, albums, media tracks, invoices and customers.

    The Chinook database is available on GitHub. It’s available for various DBMSs including MySQL, SQL Server, SQL Server Compact, PostgreSQL, Oracle, DB2, and of course, SQLite.

  8. Z

    Data from: SQL Injection Attack Netflow

    • data.niaid.nih.gov
    • portalcientifico.unileon.es
    • +2more
    Updated Sep 28, 2022
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Ignacio Crespo; Adrián Campazas (2022). SQL Injection Attack Netflow [Dataset]. https://data.niaid.nih.gov/resources?id=zenodo_6907251
    Explore at:
    Dataset updated
    Sep 28, 2022
    Authors
    Ignacio Crespo; Adrián Campazas
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Introduction

    This datasets have SQL injection attacks (SLQIA) as malicious Netflow data. The attacks carried out are SQL injection for Union Query and Blind SQL injection. To perform the attacks, the SQLMAP tool has been used.

    NetFlow traffic has generated using DOROTHEA (DOcker-based fRamework fOr gaTHering nEtflow trAffic). NetFlow is a network protocol developed by Cisco for the collection and monitoring of network traffic flow data generated. A flow is defined as a unidirectional sequence of packets with some common properties that pass through a network device.

    Datasets

    The firts dataset was colleted to train the detection models (D1) and other collected using different attacks than those used in training to test the models and ensure their generalization (D2).

    The datasets contain both benign and malicious traffic. All collected datasets are balanced.

    The version of NetFlow used to build the datasets is 5.

        Dataset
        Aim
        Samples
        Benign-malicious
        traffic ratio
    
    
    
    
        D1
        Training
        400,003
        50%
    
    
        D2
        Test
        57,239
        50%
    

    Infrastructure and implementation

    Two sets of flow data were collected with DOROTHEA. DOROTHEA is a Docker-based framework for NetFlow data collection. It allows you to build interconnected virtual networks to generate and collect flow data using the NetFlow protocol. In DOROTHEA, network traffic packets are sent to a NetFlow generator that has a sensor ipt_netflow installed. The sensor consists of a module for the Linux kernel using Iptables, which processes the packets and converts them to NetFlow flows.

    DOROTHEA is configured to use Netflow V5 and export the flow after it is inactive for 15 seconds or after the flow is active for 1800 seconds (30 minutes)

    Benign traffic generation nodes simulate network traffic generated by real users, performing tasks such as searching in web browsers, sending emails, or establishing Secure Shell (SSH) connections. Such tasks run as Python scripts. Users may customize them or even incorporate their own. The network traffic is managed by a gateway that performs two main tasks. On the one hand, it routes packets to the Internet. On the other hand, it sends it to a NetFlow data generation node (this process is carried out similarly to packets received from the Internet).

    The malicious traffic collected (SQLI attacks) was performed using SQLMAP. SQLMAP is a penetration tool used to automate the process of detecting and exploiting SQL injection vulnerabilities.

    The attacks were executed on 16 nodes and launch SQLMAP with the parameters of the following table.

        Parameters
        Description
    
    
    
    
        '--banner','--current-user','--current-db','--hostname','--is-dba','--users','--passwords','--privileges','--roles','--dbs','--tables','--columns','--schema','--count','--dump','--comments', --schema'
        Enumerate users, password hashes, privileges, roles, databases, tables and columns
    
    
        --level=5
        Increase the probability of a false positive identification
    
    
        --risk=3
        Increase the probability of extracting data
    
    
        --random-agent
        Select the User-Agent randomly
    
    
        --batch
        Never ask for user input, use the default behavior
    
    
        --answers="follow=Y"
        Predefined answers to yes
    

    Every node executed SQLIA on 200 victim nodes. The victim nodes had deployed a web form vulnerable to Union-type injection attacks, which was connected to the MYSQL or SQLServer database engines (50% of the victim nodes deployed MySQL and the other 50% deployed SQLServer).

    The web service was accessible from ports 443 and 80, which are the ports typically used to deploy web services. The IP address space was 182.168.1.1/24 for the benign and malicious traffic-generating nodes. For victim nodes, the address space was 126.52.30.0/24. The malicious traffic in the test sets was collected under different conditions. For D1, SQLIA was performed using Union attacks on the MySQL and SQLServer databases.

    However, for D2, BlindSQL SQLIAs were performed against the web form connected to a PostgreSQL database. The IP address spaces of the networks were also different from those of D1. In D2, the IP address space was 152.148.48.1/24 for benign and malicious traffic generating nodes and 140.30.20.1/24 for victim nodes.

    To run the MySQL server we ran MariaDB version 10.4.12. Microsoft SQL Server 2017 Express and PostgreSQL version 13 were used.

  9. r

    Fully Calculated Samples for Discrete Manufacturing Simulation Environment

    • resodate.org
    Updated Sep 8, 2022
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    David Heik (2022). Fully Calculated Samples for Discrete Manufacturing Simulation Environment [Dataset]. https://resodate.org/resources/aHR0cHM6Ly96ZW5vZG8ub3JnL3JlY29yZHMvNzA1NDMzMg==
    Explore at:
    Dataset updated
    Sep 8, 2022
    Dataset provided by
    Zenodo
    Authors
    David Heik
    Description

    This dataset is a MySQL file and contains fully calculated samples for different baseline situations:

        Env-Version
        ammountOfCarriers
        uncertainty
        onlyForTraining
        quantity
    
    
    
    
        1
        4
        1
        0
        10000
    
    
        1
        4
        1
        1
        90786
    
    
        1
        4
        3
        0
        10000
    
    
        1
        4
        3
        1
        92254
    
    
        1
        6
        1
        0
        20000
    
    
        1
        6
        1
        1
        210000
    
    
        1
        6
        3
        0
        20000
    
    
        1
        6
        3
        1
        214000
    
    
        1
        7
        3
        1
        30000
    
    
        1
        8
        3
        1
        34188
    
    
        1
        9
        3
        1
        30000
    
    
        1
        10
        3
        1
        34000
    
    
        1
        12
        3
        1
        3272
    
    
        2
        4
        1
        0
        10000
    
    
        2
        4
        1
        1
        94791
    
    
        2
        4
        3
        0
        10000
    
    
        2
        4
        3
        1
        100000
    
    
        2
        6
        1
        0
        10000
    
    
        2
        6
        1
        1
        90423
    
    
        2
        6
        3
        0
        10000
    
    
        2
        6
        3
        1
        249154
    

    How do you use the files? You need a MySQL instance on your computer (e.g. via XAMPP). If you do not already have one, create a database there with the name: rlskipping. The command for this is:

    CREATE DATABASE IF NOT EXISTSrlskipping;

    On Windows:

    1) Open a new CMD-Shell. 2) Navigate to the path where the sql file is located. (e.g.: cd C:\Users\MrScience\Desktop\Work\myProject\Paper\SQL) 3) Import the Files. (e.g.: mysql -u root -p rlskipping evaluierung.sql) (e.g.: mysql -u root -p rlskipping calculatedsamples.sql)

  10. s

    Orphan Drugs - Dataset 1: Twitter issue-networks as excluded publics

    • orda.shef.ac.uk
    txt
    Updated Oct 22, 2021
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Matthew Hanchard (2021). Orphan Drugs - Dataset 1: Twitter issue-networks as excluded publics [Dataset]. http://doi.org/10.15131/shef.data.16447326.v1
    Explore at:
    txtAvailable download formats
    Dataset updated
    Oct 22, 2021
    Dataset provided by
    The University of Sheffield
    Authors
    Matthew Hanchard
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    This dataset comprises of two .csv format files used within workstream 2 of the Wellcome Trust funded ‘Orphan drugs: High prices, access to medicines and the transformation of biopharmaceutical innovation’ project (219875/Z/19/Z). They appear in various outputs, e.g. publications and presentations.

    The deposited data were gathered using the University of Amsterdam Digital Methods Institute’s ‘Twitter Capture and Analysis Toolset’ (DMI-TCAT) before being processed and extracted from Gephi. DMI-TCAT queries Twitter’s STREAM Application Programming Interface (API) using SQL and retrieves data on a pre-set text query. It then sends the returned data for storage on a MySQL database. The tool allows for output of that data in various formats. This process aligns fully with Twitter’s service user terms and conditions. The query for the deposited dataset gathered a 1% random sample of all public tweets posted between 10-Feb-2021 and 10-Mar-2021 containing the text ‘Rare Diseases’ and/or ‘Rare Disease Day’, storing it on a local MySQL database managed by the University of Sheffield School of Sociological Studies (http://dmi-tcat.shef.ac.uk/analysis/index.php), accessible only via a valid VPN such as FortiClient and through a permitted active directory user profile. The dataset was output from the MySQL database raw as a .gexf format file, suitable for social network analysis (SNA). It was then opened using Gephi (0.9.2) data visualisation software and anonymised/pseudonymised in Gephi as per the ethical approval granted by the University of Sheffield School of Sociological Studies Research Ethics Committee on 02-Jun-201 (reference: 039187). The deposited dataset comprises of two anonymised/pseudonymised social network analysis .csv files extracted from Gephi, one containing node data (Issue-networks as excluded publics – Nodes.csv) and another containing edge data (Issue-networks as excluded publics – Edges.csv). Where participants explicitly provided consent, their original username has been provided. Where they have provided consent on the basis that they not be identifiable, their username has been replaced with an appropriate pseudonym. All other usernames have been anonymised with a randomly generated 16-digit key. The level of anonymity for each Twitter user is provided in column C of deposited file ‘Issue-networks as excluded publics – Nodes.csv’.

    This dataset was created and deposited onto the University of Sheffield Online Research Data repository (ORDA) on 26-Aug-2021 by Dr. Matthew S. Hanchard, Research Associate at the University of Sheffield iHuman institute/School of Sociological Studies. ORDA has full permission to store this dataset and to make it open access for public re-use without restriction under a CC BY license, in line with the Wellcome Trust commitment to making all research data Open Access.

    The University of Sheffield are the designated data controller for this dataset.

  11. o

    FooDrugs database: A database with molecular and text information about food...

    • explore.openaire.eu
    • resodate.org
    • +1more
    Updated Jun 13, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Marco Garranzo; Óscar Piette Gómez; Blanca Lacruz Pleguezuelos; David Pérez; Teresa Laguna Lobo; Enrique Carrillo de Santa Pau (2022). FooDrugs database: A database with molecular and text information about food - drug interactions [Dataset]. http://doi.org/10.5281/zenodo.8192515
    Explore at:
    Dataset updated
    Jun 13, 2022
    Authors
    Marco Garranzo; Óscar Piette Gómez; Blanca Lacruz Pleguezuelos; David Pérez; Teresa Laguna Lobo; Enrique Carrillo de Santa Pau
    Description

    FooDrugs database is a development done by the Computational Biology Group at IMDEA Food Institute (Madrid, Spain), in the context of the Food Nutrition Security Cloud (FNS-Cloud) project. Food Nutrition Security Cloud (FNS-Cloud) has received funding from the European Union's Horizon 2020 Research and Innovation programme (H2020-EU.3.2.2.3. – A sustainable and competitive agri-food industry) under Grant Agreement No. 863059 – www.fns-cloud.eu (See more details about FNS-Cloud below) FooDrugs stores information extracted from transcriptomics and text documents for foo-drug interactiosn and it is part of a demonstrator to be done in the FNS-Cloud project. The database was built using MySQL, an open source relational database management system. FooDrugs_V2 host information for a total of 161 transcriptomics GEO series with 585 conditions for food or bioactive compounds (see below changes in versions V3 and V4). Each condition is defined as a food/biocomponent per time point, per concentration, per cell line, primary culture or biopsy per study. FooDrugs includes information about a bipartite network with 510 nodes and their similarity scores (tau score; https://clue.io/connectopedia/connectivity_scores) related with possible drug interactions with drugs assayed in conectivity map (https://www.broadinstitute.org/connectivity-map-cmap). The information is stored in eight tables: Table “study” : This table contains basic information about study identifiers from GEO, pubmed or platform, study type, title and abstract Table “sample”: This table contains basic information about the different experiments in a study, like the identifier of the sample, treatment, origin type, time point or concentration. Table “misc_study”: This table contains additional information about different attributes of the study. Table “misc_sample”: This table contains additional information about different attributes of the sample. Table “cmap”: This table contains information about 70895 nodes, compromising drugs, foods or bioactives, overexpressed and knockdown genes (see section 3.4). The information includes cell line, compound and perturbation type. Table “cmap_foodrugs”: This table contains information about the tau score (see section 3.4) that relates food with drugs or genes and the node identifier in the FooDrugs network. Table “topTable”: This table contains information about 150 over and underexpressed genes from each GEO study condition, used to calculate the tau score (see section 3.4). The information stored is the logarithmic fold change, average expression, t-statistic, p-value, adjusted p-value and if the gene is up or downregulated. Table “nodes”: This table stores the information about the identification of the sample and the node in the bipartite network connecting the tables “sample”, “cmap_foodrugs” and “topTable”. In addition, FooDrugs_V2 database stores a total of 6422 food/drug interactions from 2849 text documents, obtained from three different sources: 2312 documents from PubMed, 285 from DrugBank, and 252 from drugs.com. These documents describe potential interactions between 1464 food/bioactive compounds and 3009 drugs (see below changes in versions V3 and V4). The information is stored in two tables: Table “texts”: This table contains all the documents with its identifiers where interactions have been identified with strategy described in section 4. Table “TM_interactions”: This table contains information about interaction identifiers, the food and drug entities, and the start and the end positions of the context for the interaction in the document. FNS-Cloud will overcome fragmentation problems by integrating existing FNS data, which is essential for high-end, pan-European FNS research, addressing FNS, diet, health, and consumer behaviours as well as on sustainable agriculture and the bio-economy. Current fragmented FNS resources not only result in knowledge gaps that inhibit public health and agricultural policy, and the food industry from developing effective solutions, making production sustainable and consumption healthier, but also do not enable exploitation of FNS knowledge for the benefit of European citizens. FNS-Cloud will, through three Demonstrators; Agri-Food, Nutrition & Lifestyle and NCDs & the Microbiome to facilitate: (1) Analyses of regional and country-specific differences in diet including nutrition, (epi)genetics, microbiota, consumer behaviours, culture and lifestyle and their effects on health (obesity, NCDs, ethnic and traditional foods), which are essential for public health and agri-food and health policies; (2) Improved understanding agricultural differences within Europe and what these means in terms of creating a sustainable, resilient food systems for healthy diets; and (3) Clear definitions of boundaries and how these affect the compositions of foods and consumer choices and, ultimately, personal and public health in the future. Long-term sustainability of the FNS-Clou...

  12. c

    Data Base Management Systems market size was USD 50.5 billion in 2022 !

    • cognitivemarketresearch.com
    pdf,excel,csv,ppt
    Updated Dec 15, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Cognitive Market Research (2025). Data Base Management Systems market size was USD 50.5 billion in 2022 ! [Dataset]. https://www.cognitivemarketresearch.com/data-base-management-systems-market-report
    Explore at:
    pdf,excel,csv,pptAvailable download formats
    Dataset updated
    Dec 15, 2025
    Dataset authored and provided by
    Cognitive Market Research
    License

    https://www.cognitivemarketresearch.com/privacy-policyhttps://www.cognitivemarketresearch.com/privacy-policy

    Time period covered
    2022 - 2034
    Area covered
    Global
    Description

    The global Data Base Management Systems market was valued at USD 50.5 billion in 2022 and is projected to reach USD 120.6 Billion by 2030, registering a CAGR of 11.5 % for the forecast period 2023-2030. Factors Affecting Data Base Management Systems Market Growth

    Growing inclination of organizations towards adoption of advanced technologies like cloud-based technology favours the growth of global DBMS market
    

    The cloud-based data base management system solutions offer the organizations with an ability to scale their database infrastructure up or down as per requirement. In a crucial business environment data volume can vary over time. Here, the cloud allows organizations to allocate resources in a dynamic and systematic manner, thereby, ensuring optimal performance without underutilization. In addition, these cloud-based solutions are cost-efficient. As, these cloud-based DBMS solutions eliminate the need for companies to maintain and invest in physical infrastructure and hardware. It helps in reducing ongoing operational costs and upfront capital expenditures. Organizations can choose pay-as-you-go pricing models, where they need to pay only for the resources they consume. Therefore, it has been a cost-efficient option for both smaller businesses and large-enterprises. Moreover, the cloud-based data base management system platforms usually come with management tools which streamline administrative tasks such as backup, provisioning, recovery, and monitoring. It allows IT teams to concentrate on more of strategic tasks rather than routine maintenance activities, thereby, enhancing operational efficiency. Whereas, these cloud-based data base management systems allow users to remote access and collaboration among teams, irrespective of their physical locations. Thus, in regards with today's work environment, which focuses on distributed and remote workforces. These cloud-based DBMS solution enables to access data and update in real-time through authorized personnel, allowing collaboration and better decision-making. Thus, owing to all the above factors, the rising adoption of advanced technologies like cloud-based DBMS is favouring the market growth.

    Availability of open-source solutions is likely to restrain the global data base management systems market growth
    

    Open-source data base management system solutions such as PostgreSQL, MongoDB, and MySQL, offer strong functionality at minimal or no licensing costs. It makes open-source solutions an attractive option for companies, especially start-ups or smaller businesses with limited budgets. As these open-source solutions offer similar capabilities to various commercial DBMS offerings, various organizations may opt for this solutions in order to save costs. The open-source solutions may benefit from active developer communities which contribute to their development, enhancement, and maintenance. This type of collaborative environment supports continuous innovation and improvement, which results into solutions that are slightly competitive with commercial offerings in terms of performance and features. Thus, the open-source solutions create competition for commercial DBMS market, they thrive in the market by offering unique value propositions, addressing needs of organizations which prioritize professional support, seamless integration into complex IT ecosystems, and advanced features. Introduction of Data Base Management Systems

    A Database Management System (DBMS) is a software which is specifically designed to organize and manage data in a structured manner. This system allows users to create, modify, and query a database, and also manage the security and access controls for that particular database. The DBMS offers tools for creating and modifying data models, that define the structure and relationships of data in a database. This system is also responsible for storing and retrieving data from the database, and also provide several methods for searching and querying the data. The data base management system also offers mechanisms to control concurrent access to the database, in order to ensure that number of users may access the data. The DBMS provides tools to enforce security constraints and data integrity, such as the constraints on the value of data and access controls that restricts who can access the data. The data base management system also provides mechanisms for recovering and backing up the data when a system failure occurs....

  13. g

    Meta-Information des Samples der Media-Analyse Daten: IntermediaPlus...

    • search.gesis.org
    • datacatalogue.cessda.eu
    • +1more
    Updated Nov 3, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Brentel, Inga; Kampes, Céline Fabienne; Jandura, Olaf (2023). Meta-Information des Samples der Media-Analyse Daten: IntermediaPlus (2014-2016) [Dataset]. https://search.gesis.org/research_data/SDN-10.7802-2030
    Explore at:
    Dataset updated
    Nov 3, 2023
    Dataset provided by
    GESIS search
    GESIS, Köln
    Authors
    Brentel, Inga; Kampes, Céline Fabienne; Jandura, Olaf
    License

    https://www.gesis.org/en/institute/data-usage-termshttps://www.gesis.org/en/institute/data-usage-terms

    Description

    Bei dem aufbereiteten Längsschnitt-Datensatzes 2014 bis 2016 handelt es sich um „Big-Data“, weshalb der Gesamtdatensatz nur in Form einer Datenbank (MySQL) verfügbar sein wird. In dieser Datenbank liegt die Information verschiedener Variablen eines Befragten untereinander. Die vorliegende Publikation umfasst eine SQL-Datenbank mit den Meta-Daten des Sample des Gesamtdatensatzes, das einen Ausschnitt der verfügbaren Variablen des Gesamtdatensatzes darstellt und die Struktur der aufbereiteten Daten darlegen soll, und eine Datendokumentation des Samples. Für diesen Zweck beinhaltet das Sample alle Variablen der Soziodemographie, dem Freizeitverhalten, der Zusatzinformation zu einem Befragten und dessen Haushalt sowie den interviewspezifischen Variablen und Gewichte. Lediglich bei den Variablen bezüglich der Mediennutzung des Befragten, handelt es sich um eine kleine Auswahl: Für die Onlinemediennutzung wurden die Variablen aller Gesamtangebote sowie der Einzelangebote der Genre Politik und Digital aufgenommen. Die Mediennutzung von Radio, Print und TV wurde im Sample nicht berücksichtigt, da deren Struktur anhand der veröffentlichten Längsschnittdaten der Media-Analyse MA Radio, MA Pressemedien und MA Intermedia nachvollzogen werden kann.
    Die Datenbank mit den tatsächlichen Befragungsdaten wäre auf Grund der Größe des Datenmaterials bereits im kritischen Bereich der Dateigröße für den normalen Up- und Download. Die tatsächlichen Befragungsergebnisse, die zur Analyse nötig sind, werden dann 2021 in Form des Gesamtdatensatzes der Media-Analyse-Daten: IntermediaPlus (2014-2016) im DBK bei GESIS veröffentlicht werden.

    Die Daten sowie deren Datenaufbereitung sind ein Vorschlag eines Best-Practice Cases für Big-Data Management bzw. den Umgang mit Big-Data in den Sozialwissenschaften und mit sozialwissenschaftlichen Daten. Unter Verwendung der GESIS Software CharmStats, die im Rahmen dieses Projektes um Big-Data Features erweitert wurde, erfolgt die Dokumentation und Herstellung der Transparenz der Harmonisierungsarbeit. Durch ein Python-Skript sowie ein html-Template wurde der Arbeitsprozess um und mit CharmStats zudem stärker automatisiert.

    Der aufbereitete Längsschnitt des Gesamtdatensatzes der MA IntermediaPlus für 2014 bis 2016 wird 2021 in Kooperation mit GESIS herausgegeben werden und den FAIR-Prinzipien (Wilkinson et al. 2016) entsprechend verfügbar gemacht werden. Ziel ist es durch die Harmonisierung der einzelnen Querschnitte die Datenquelle der Media-Analyse, die im Rahmen des Dissertationsprojektes „Angebots- und Publikumsfragmentierung online“ durch Inga Brentel und Céline Fabienne Kampes erfolgt, für Forschung zum sozialen und medialen Wandel in der Bundesrepublik Deutschland zugänglich zu machen.

    Künftige Studiennummer des Gesamtdatensatzes der IndermediaPlus im DBK der GESIS: ZA5769 (Version 1-0-0) und der doi: https://dx.doi.org/10.4232/1.13530

    ****************English Version****************

    The prepared Longitudinal IntermediaPlus dataset 2014 to 2016 is a "big data", which is why the entire dataset will only be available in the form of a database (MySQL). In this database, the information of different variables of a respondent is organized in one column, one below the other. The present publication includes a SQL-Database with the meta data of a sample of the full database, which represents a section of the available variables of the total data set and is intended to show the structure of the prepared data and the data-documentation (codebook) of the sample. For this purpose, the sample contains all variables of sociodemography, free-time activities, additional information on a respondent and his household as well as the interview-specific variables and weights. Only the variables concerning the respondent's media use are a small selection: For online media use, the variables of all overall offerings as well as the individual offerings of the genres politics and digital were included. The media use of radio, print and TV was not included in the sample because its structure can be traced using the published longitudinal data of the media analysis MA Radio, MA Pressemedien and MA Intermedia.
    Due to the size of the datafile, the database with the actual survey data would already be in the critical range of the file size for the common upload and download. The actual survey result...

  14. Data from: A Fast Hop-Biased Approximation Algorithm for the Quadratic Group...

    • zenodo.org
    Updated Sep 14, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Anonymous; Anonymous (2023). A Fast Hop-Biased Approximation Algorithm for the Quadratic Group Steiner Tree Problem [Dataset]. http://doi.org/10.5281/zenodo.7784147
    Explore at:
    Dataset updated
    Sep 14, 2023
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Anonymous; Anonymous
    Description

    The dataset for our paper 'A Fast Hop-Biased Approximation Algorithm for the Quadratic Group Steiner Tree Problem'. It consists of 5 real KGs (Mondial, OpenCyc, LinkedMDB, YAGO, DBpedia) and 5 synthetic KGs (LUBM-10U, LUBM-50U, LUBM-250U, LUBM-2U, DBP-50K). Each KG is compressed in one file, which including (for example, in LUBM-2U):

    • lubm_2u_nodes.sql: the id, the name and the weight of a node,

    • lubm_2u_edges.sql: the ids of two nodes an edge connects,

    • lubm_2u_queries.sql: a query consists of some keywords,

    • lubm_2u_keymap.sql: a keyword maps to a set of nodes,

    • lubm_2u_nodevec.sql: the vector of a node, used to compute quadratic function qw,

    • lubm_2u_hub_hop.sql: the hub labeling index to compute in Section 4.1,

    • lubm_2u_hub_mix_1.sql: the hub labeling index to compute in Section 4.1 where α=0.1,

    • lubm_2u_hub_mix_5.sql: the hub labeling index to compute in Section 4.1 where α=0.5,

    • lubm_2u_hub_mix_9.sql: the hub labeling index to compute in Section 4.1 where α=0.9.

    You can dump the data into MySQL database. For example,

    create database lubm_2u;
    use lubm_2u;
    source lubm_2u_nodes.sql;
    …

    Unfortunately, due to the limit of space, for large KGs (DBpedia and LUBM-250U), we don't directly provide the data of hub labeling, i.e., these two compressed files only contains the first 5 sql files. You should generate hub labeling by yourself where the process is detailed in README of our project.

  15. CHINOOK Music

    • kaggle.com
    zip
    Updated Sep 19, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    willian oliveira (2024). CHINOOK Music [Dataset]. https://www.kaggle.com/datasets/willianoliveiragibin/chinook-music
    Explore at:
    zip(9603 bytes)Available download formats
    Dataset updated
    Sep 19, 2024
    Authors
    willian oliveira
    License

    https://creativecommons.org/publicdomain/zero/1.0/https://creativecommons.org/publicdomain/zero/1.0/

    Description

    The Chinook Database is a sample database designed for use with multiple database platforms, such as SQL Server, Oracle, MySQL, and others. It can be easily set up by running a single SQL script, making it a convenient alternative to the popular Northwind database. Chinook is widely used in demos and testing environments, particularly for Object-Relational Mapping (ORM) tools that target both single and multiple database servers.

    Supported Database Servers Chinook supports several database servers, including:

    DB2 MySQL Oracle PostgreSQL SQL Server SQL Server Compact SQLite Download Instructions You can download the SQL scripts for each supported database server from the latest release assets. The appropriate SQL script file(s) for your database vendor are provided, which can be executed using your preferred database management tool.

    Data Model The Chinook Database represents a digital media store, containing tables that include:

    Artists Albums Media tracks Invoices Customers Sample Data The media data in Chinook is derived from a real iTunes Library, providing a realistic dataset for users. Additionally, users can generate their own SQL scripts using their personal iTunes Library by following specific instructions. Customer and employee details in the database were manually crafted with fictitious names, addresses (mappable via Google Maps), and well-structured contact information such as phone numbers, faxes, and emails. Sales data is auto-generated and spans a four-year period, using random values.

    Why is it Called Chinook? The Chinook Database's name is a nod to its predecessor, the Northwind database. Chinooks are warm, dry winds found in the interior regions of North America, particularly over southern Alberta in Canada, where the Canadian Prairies meet mountain ranges. This natural phenomenon inspired the choice of name, reflecting the idea that Chinook serves as a refreshing alternative to the Northwind database.

  16. n

    Heparome

    • neuinfo.org
    • dknet.org
    • +2more
    Updated Oct 11, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    (2025). Heparome [Dataset]. http://identifiers.org/RRID:SCR_008615
    Explore at:
    Dataset updated
    Oct 11, 2025
    Description

    THIS RESOURCE IS NO LONGER IN SERVICE, documented on July 17, 2013. A database which contains the information of heparin-binding proteins of E. coli K-12 MG1655 cells. Heparin affinity columns were applied to enrich and fractionate proteins. Identification of proteins was done via the collaboration with David Russell''s lab. Because heparin is negatively charged sulfated glucosaminoglycan, polyamion binding proteins, which contain nucleic acid-binding proteins, are expected to bind to heparin columns. Study of the expression pattern of heparin-binding proteins will help to study the nucleic acid-binding proteins, most of which are related to regulation. Moreover, heparin affinity columns will also erich low abundance proteins. Heparome database is constructed using MySQL. Website interface is built using HTML and PHP. Queries between MySQL database and website interface are executed using PHP. Besides including information of identified proteins, such as swiss accession number, gene name, molecular weight, isoelectric point, condon adaptation index (CAI), functional classification, et. al. , it also includes information of experiments, such as sample preparation, heparin-HPLC chromatography, SDS-PAGE gel separation and MALDI-MS.

  17. p

    Royal Institute for Cultural Heritage Radiocarbon and stable isotope...

    • pandora.earth
    • pandoradata.earth
    Updated Jul 12, 2011
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    (2011). Royal Institute for Cultural Heritage Radiocarbon and stable isotope measurements [Dataset]. https://pandora.earth/dataset/royal-institute-for-cultural-heritage-radiocarbon-and-stable-isotope-measurements
    Explore at:
    Dataset updated
    Jul 12, 2011
    Description

    The Radiocarbon dating laboratory of IRPA/KIK was founded in the 1960s. Initially dates were reported at more or less regular intervals in the journal Radiocarbon (Schreurs 1968). Since the advent of radiocarbon dating in the 1950s it had been a common practice amongst radiocarbon laboratories to publish their dates in so-called ‘date-lists’ that were arranged per laboratory. This was first done in the Radiocarbon Supplement of the American Journal of Science and later in the specialised journal Radiocarbon. In the course of time the latter, with the added subtitle An International Journal of Cosmogenic Isotope Research, became a regular scientific journal shifting focus from date-lists to articles. Furthermore the world-wide exponential increase of radiocarbon dates made it almost impossible to publish them all in the same journal, even more so because of the broad range of applications that use radiocarbon analysis, ranging from archaeology and art history to geology and oceanography and recently also biomedical studies.The IRPA/KIK database From 1995 onwards IRPA/KIK’s Radiocarbon laboratory started to publish its dates in small publications, continuing the numbering of the preceding lists in Radiocarbon. The first booklet in this series was “Royal Institute for Cultural Heritage Radiocarbon dates XV” (Van Strydonck et al. 1995), followed by three more volumes (XVI, XVII, XVIII). The next list (XIX, 2005) was no longer printed but instead handed out as a PDF file on CD-rom. The ever increasing number of dates and the difficulties in handling all the data, however, made us look for a more permanent and easier solution. In order to improve data management and consulting, it was thus decided to gather all our dates in a web-based database. List XIX was in fact already a Microsoft Access database that was converted into a reader friendly style and could also be printed as a PDF file. However a Microsoft Access database is not the most practical solution to make information publicly available. Hence the structure of the database was recreated in Mysql and the existing content was transferred into the corresponding fields. To display the records, a web-based front-end was programmed in PHP/Apache. It features a full-text search function that allows for partial word-matching. In addition the records can be consulted in PDF format. Old records from the printed date-lists as well as new records are now added using the same Microsoft Acces back-end, which is now connected directly to the Mysql database. The main problem with introducing the old data was that not all the current criteria were available in the past (e.g. stable isotope measurements). Furthermore since all the sample information is given by the submitter, its quality largely depends on the persons willingness to contribute as well as on the accuracy and correctness of the information he provides. Sometimes problems arrive from the fact that a certain investigation (like an excavation) is carried out over a relatively long period (sometimes even more than ten years) and is directed by different people or even institutions. This can lead to differences in the labeling procedure of the samples, but also in the interpretation of structures and artifacts and in the orthography of the site’s name. Finally the submitter might change address, while the names of institutions or even regions and countries might change as well (e.g.Zaire - Congo)

  18. E

    Data from: ChEssBase

    • erddap.eurobis.org
    • obis.org
    Updated Aug 12, 2025
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Perry, Hall, Baker, Ramirez-Llodra (2025). ChEssBase [Dataset]. https://erddap.eurobis.org/erddap/info/chessbase/index.html
    Explore at:
    Dataset updated
    Aug 12, 2025
    Dataset authored and provided by
    Perry, Hall, Baker, Ramirez-Llodra
    Area covered
    Variables measured
    aphia_id, latitude, longitude, MaximumDepth, MinimumDepth, BasisOfRecord, ScientificName, InstitutionCode
    Description

    ChEssBase is a dynamic relational database for all deep-water species from chemosynthetic ecosystems (hydrothermal vents, cold seeps and other reducing environments such as whale carcasses, sunken wood or OMZs) being constructed from the ChEss project (Biogeography of Deep-Water Chemosynthetic Ecosystems) within the Census of Marine Life initiative. AccConID=21 AccConstrDescription=This license lets others distribute, remix, tweak, and build upon your work, even commercially, as long as they credit you for the original creation. This is the most accommodating of licenses offered. Recommended for maximum dissemination and use of licensed materials. AccConstrDisplay=This dataset is licensed under a Creative Commons Attribution 4.0 International License. AccConstrEN=Attribution (CC BY) AccessConstraint=Attribution (CC BY) AccessConstraints=None Acronym=None added_date=2013-06-12 15:21:34.517000 BrackishFlag=0 CDate=2004-06-24 cdm_data_type=Other CheckedFlag=0 Citation=Ramirez-Llodra, E., Blanco, 2005. ChEssBase: an online information system on biodiversity and biogeography of deep-sea fauna from chemosynthetic ecosystems. Version 2. World Wide Web electronic publications, http://www.noc.soton.ac.uk/chess/database/db_home.php Comments=None ContactEmail=None Conventions=COARDS, CF-1.6, ACDD-1.3 CurrencyDate=None DasID=212 DasOrigin=Literature research DasType=Data DasTypeID=1 DateLastModified={'date': '2025-08-12 01:34:46.196267', 'timezone_type': 1, 'timezone': '+02:00'} DescrCompFlag=0 DescrTransFlag=0 Easternmost_Easting=179.8 EmbargoDate=None EngAbstract=ChEssBase is a dynamic relational database for all deep-water species from chemosynthetic ecosystems (hydrothermal vents, cold seeps and other reducing environments such as whale carcasses, sunken wood or OMZs) being constructed from the ChEss project (Biogeography of Deep-Water Chemosynthetic Ecosystems) within the Census of Marine Life initiative. EngDescr=The aim of ChEssBase is to provide taxonomical, biological, ecological and distributional data for all species described from deep-water chemosynthetic ecosystems, as well as information on available samples, images, bibliography and information on the habitats.These habitats include hydrothermal vents, cold seeps, whale falls, sunken wood and areas of minimum oxygen that intersect with the continental margin or seamounts. Since the discovery of hydrothermal vents in 1977 and of cold seep communities in 1984, over 590 species from vents and over 230 species from seeps have been described. Chemosynthetically fueled communities have now also been found on large organic falls to the deep-sea floor such as whale falls and sunken wood, as well as on benthic zones of oxygen minimum.The data gathered in the last 30 years has shown that some species are shared amongst these ecosystems and our knowledge of their phylogeography improves with every new discovery. New species are continuously being discovered and described from research programmes around the globe and therefore ChEssBase is in active development and new data are being entered regularly. At present, ChEssBase includes data on 1740 species from 193 chemosynthetic sites around the globe. These data contain information (when available) on the taxonomy, morphology, trophic level, reproduction, endemicity, habitat type and distribution. There are now 1880 papers in our reference database.The first version of ChEssBase was available online in December 2004. In summer 2005, ChEssBase and the InterRidge biological database (www.interridge.org) were fused into a single source of information for biological data from chemosynthetic ecosystems. This second version of ChEssBase is available online since August 2005, with new records as well as new search and download options. Since December 2005, ChEssBase is integrated in the Ocean Biogeographic Information System (OBIS, www.iobis.org).ChEssBase is supported by a species-based relational database in MySQL. The database includes 3 major components:Taxonomy (from kingdom to subspecies)Distribution (from site to major geographic area)Samples (including sample, cruise and institution information)ChEssBase is regularly updated with new information available in the literature. In order to quickly obtain accurate new data and help maintain the database up to date, we would be very grateful if you could send us any new publications with data relevant to ChEssBase, which we would add to the database, together with the relevant references. FreshFlag=0 geospatial_lat_max=72.0 geospatial_lat_min=-55.1 geospatial_lat_units=degrees_north geospatial_lon_max=179.8 geospatial_lon_min=-158.1 geospatial_lon_units=degrees_east infoUrl=None InputNotes=None institution=COML, SOTON-NOC, SOTON-SOES License=https://creativecommons.org/licenses/by/4.0/ Lineage=Prior to publication data undergo quality control checked which are described in https://github.com/EMODnet/EMODnetBiocheck?tab=readme-ov-file#understanding-the-output MarineFlag=1 modified_sync=2021-02-05 00:00:00 Northernmost_Northing=72.0 OrigAbstract=None OrigDescr=None OrigDescrLang=None OrigDescrLangNL=None OrigLangCode=None OrigLangCodeExtended=None OrigLangID=None OrigTitle=None OrigTitleLang=None OrigTitleLangCode=None OrigTitleLangID=None OrigTitleLangNL=None Progress=In Progress PublicFlag=1 ReleaseDate=Jun 12 2013 12:00AM ReleaseDate0=2013-06-12 RevisionDate=None SizeReference=1740 species from 193 sites sourceUrl=(local files) Southernmost_Northing=-55.1 standard_name_vocabulary=CF Standard Name Table v70 StandardTitle=ChEssBase StatusID=1 subsetVariables=ScientificName,BasisOfRecord,aphia_id TerrestrialFlag=0 UDate=2025-03-26 VersionDate=Jun 3 2004 12:00AM VersionDay=23 VersionMonth=10 VersionName=2 VersionYear=2007 VlizCoreFlag=1 Westernmost_Easting=-158.1

  19. Popularity distribution of database management systems worldwide 2024, by...

    • statista.com
    Updated Nov 28, 2025
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Statista (2025). Popularity distribution of database management systems worldwide 2024, by model [Dataset]. https://www.statista.com/statistics/1131595/worldwide-popularity-database-management-systems-category/
    Explore at:
    Dataset updated
    Nov 28, 2025
    Dataset authored and provided by
    Statistahttp://statista.com/
    Time period covered
    Jun 2024
    Area covered
    Worldwide
    Description

    As of December 2022, relational database management systems (RDBMS) were the most popular type of DBMS, accounting for a ** percent popularity share. The most popular RDBMS in the world has been reported as Oracle, while MySQL and Microsoft SQL server rounded out the top three.

  20. Data from: MWSTAT: A MODULATED WEB-BASED STATISTICAL SYSTEM

    • scielo.figshare.com
    jpeg
    Updated Jun 1, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Francisco Louzada; Anderson Ara (2023). MWSTAT: A MODULATED WEB-BASED STATISTICAL SYSTEM [Dataset]. http://doi.org/10.6084/m9.figshare.6967682.v1
    Explore at:
    jpegAvailable download formats
    Dataset updated
    Jun 1, 2023
    Dataset provided by
    SciELOhttp://www.scielo.org/
    Authors
    Francisco Louzada; Anderson Ara
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    ABSTRACT In this paper we present the development of a modulated web based statistical system, hereafter MWStat, which shifts the statistical paradigm of analyzing data into a real time structure. The MWStat system is useful for both online storage data and questionnaires analysis, as well as to provide real time disposal of results from analysis related to several statistical methodologies in a customizable fashion. Overall, it can be seem as a useful technical solution that can be applied to a large range of statistical applications, which needs of a scheme of devolution of real time results, accessible to anyone with internet access. We display here the step-by-step instructions for implementing the system. The structure is accessible, built with an easily interpretable language and it can be strategically applied to online statistical applications. We rely on the relationship of several free languages, namely, PHP, R, MySQL database and an Apache HTTP server, and on the use of software tools such as phpMyAdmin. We expose three didactical examples of the MWStat system on institutional evaluation, statistical quality control and multivariate analysis. The methodology is also illustrated in a real example on institutional evaluation. A MWStat module was specifically built for providing a real time poll for teacher evaluation at the Federal University of São Carlos (Brazil).

Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
Atanas Kanev (2021). SQLite Sakila Sample Database [Dataset]. https://www.kaggle.com/atanaskanev/sqlite-sakila-sample-database
Organization logo

SQLite Sakila Sample Database

SQLite Port of the Original MySQL Sakila Sample Database

Explore at:
zip(4495190 bytes)Available download formats
Dataset updated
Mar 14, 2021
Authors
Atanas Kanev
Description

SQLite Sakila Sample Database

Database Description

The Sakila sample database is a fictitious database designed to represent a DVD rental store. The tables of the database include film, film_category, actor, customer, rental, payment and inventory among others. The Sakila sample database is intended to provide a standard schema that can be used for examples in books, tutorials, articles, samples, and so forth. Detailed information about the database can be found on the MySQL website: https://dev.mysql.com/doc/sakila/en/

Sakila for SQLite is a part of the sakila-sample-database-ports project intended to provide ported versions of the original MySQL database for other database systems, including:

  • Oracle
  • SQL Server
  • SQLIte
  • Interbase/Firebird
  • Microsoft Access

Sakila for SQLite is a port of the Sakila example database available for MySQL, which was originally developed by Mike Hillyer of the MySQL AB documentation team. This project is designed to help database administrators to decide which database to use for development of new products The user can run the same SQL against different kind of databases and compare the performance

License: BSD Copyright DB Software Laboratory http://www.etl-tools.com

Note: Part of the insert scripts were generated by Advanced ETL Processor http://www.etl-tools.com/etl-tools/advanced-etl-processor-enterprise/overview.html

Information about the project and the downloadable files can be found at: https://code.google.com/archive/p/sakila-sample-database-ports/

Other versions and developments of the project can be found at: https://github.com/ivanceras/sakila/tree/master/sqlite-sakila-db

https://github.com/jOOQ/jOOQ/tree/main/jOOQ-examples/Sakila

Direct access to the MySQL Sakila database, which does not require installation of MySQL (queries can be typed directly in the browser), is provided on the phpMyAdmin demo version website: https://demo.phpmyadmin.net/master-config/

Files Description

The files in the sqlite-sakila-db folder are the script files which can be used to generate the SQLite version of the database. For convenience, the script files have already been run in cmd to generate the sqlite-sakila.db file, as follows:

sqlite> .open sqlite-sakila.db # creates the .db file sqlite> .read sqlite-sakila-schema.sql # creates the database schema sqlite> .read sqlite-sakila-insert-data.sql # inserts the data

Therefore, the sqlite-sakila.db file can be directly loaded into SQLite3 and queries can be directly executed. You can refer to my notebook for an overview of the database and a demonstration of SQL queries. Note: Data about the film_text table is not provided in the script files, thus the film_text table is empty. Instead the film_id, title and description fields are included in the film table. Moreover, the Sakila Sample Database has many versions, so an Entity Relationship Diagram (ERD) is provided to describe this specific version. You are advised to refer to the ERD to familiarise yourself with the structure of the database.

Search
Clear search
Close search
Google apps
Main menu