Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
This dataset is used in Master thesis on topic "The impact of upholding environmental, social and governance principles on the market value of capital-intensive companies". The full dataset consits data on Public american companies included in S&P 500 index, traded on the New York Stock Exchange. There are data on ESG-score and its components (E, S, G), as well as components of Envitonmental pillar score. Additionaly dataset includes financial data, like market capitalization, leverage, ROCE, Capex and etc. The main sources of data are Thomson Reuters Eikon and Bloomberg terminals, along with Form 10-k by SEC. The final sample consists of 52 capital-intensive companies, time horizon: 2012-2021 [520 observations in total].
CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
License information was derived automatically
This thesis-mpc-dataset-public-readme.txt file was generated on 2020-10-20 by Masud Petronia
GENERAL INFORMATION
1. Title of Dataset: Data underlying the thesis: Multiparty Computation: The effect of multiparty computation on firms' willingness to contribute protected data
2. Author Information A. Principal Investigator Contact Information Name: Masud Petronia Institution: TU Delft, Faculty of Technology, Policy and Management Address: Mekelweg 5, 2628 CD Delft, Netherlands Email: masud.petronia@gmail.com ORCID: https://orcid.org/0000-0003-2798-046X
3: Description of dataset: This dataset contains perceptual data of firms' willingness to contribute protected data through multi party computation (MPC). Petronia (2020, ch. 6) draws several conclusions from this dataset and provides recommendations for future research Petronia (2020, ch. 7.4).
4. Date of data collection: July-August 2020
5. Geographic location of data collection: Netherlands
6. Information about funding sources that supported the collection of the data: Horizon 2020 Research and Innovation Programme, Grant Agreement no 825225 – Safe Data Enabled Economic Development (SAFE-DEED), from the H2020-ICT-2018-2
SHARING/ACCESS INFORMATION
1. Licenses/restrictions placed on the data: CC 0
2. Links to publications that cite or use the data: Petronia, M. N. (2020). Multiparty Computation: The effect of multiparty computation on firms' willingness to contribute protected data (Master's thesis). Retrieved from http://resolver.tudelft.nl/uuid:b0de4a4b-f5a3-44b8-baa4-a6416cebe26f
3. Was data derived from another source? No
4. Citation for this dataset: Petronia, M. N. (2020). Multiparty Computation: The effect of multiparty computation on firms' willingness to contribute protected data (Master's thesis). Retrieved from https://data.4tu.nl/. doi:10.4121/13102430
DATA & FILE OVERVIEW
1. File List: thesis-mpc-dataset-public.xlsxthesis-mpc-dataset-public-readme.txt (this document)
2. Relationship between files: Dataset metadata and instructions
3. Additional related data collected that was not included in the current data package: Occupation and role of respondents (traceable to unique reference), removed for privacy reasons.
4. Are there multiple versions of the dataset? No
METHODOLOGICAL INFORMATION
1. Description of methods used for collection/generation of data: A pre- and post test experimental design. For more information; see Petronia (2020, ch. 5)
2. Methods for processing the data: Full instructions are provided by Petronia (2020, ch. 6)
3. Instrument- or software-specific information needed to interpret the data: Microsoft Excel can be used to convert the dataset to other formats.
4. Environmental/experimental conditions: This dataset comprises three datasets collected through three channels. These channels are Prolific (incentive), LinkedIn/Twitter (voluntarily), and respondents in a lab setting (voluntarily). For more information; see Petronia (2020, ch. 6.1)
5. Describe any quality-assurance procedures performed on the data: A thorough examination of consistency and reliability is performed. For more information; see Petronia (2020, ch. 6).
6. People involved with sample collection, processing, analysis and/or submission: See Petronia (2020, ch. 6)
DATA-SPECIFIC INFORMATION
1. Number of variables: see worksheet experiment_matrix of thesis-mpc-dataset-public.xlsx
2. Number of cases/rows: see worksheet experiment_matrix of thesis-mpc-dataset-public.xlsx
3. Variable List: see worksheet labels of thesis-mpc-dataset-public.xlsx
4. Missing data codes: see worksheet comments of thesis-mpc-dataset-public.xlsx
5. Specialized formats or other abbreviations used: Multiparty computation (MPC) and Trusted Third Party (TTP).
INSTRUCTIONS
1. Petronia (2020, ch. 6) describes associated tests and respective syntax.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Availability of data, code, and plot creation for various figures throughout my PhD thesis. Rough organisation currently. Pertains to Figures 5.4, 5.8, 6.11, 6.18, 7.3, 7.12, and Table 6.1.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
This codebook was used to analyze the interview data (from 11 interviews) in the master thesis project titled "Enhancing Open Research Data Sharing and Reuse via Infrastructural and Institutional Instruments: a Case Study in Epidemiology" which is openly available on TU Delft Education Repository.
The data in this collection comprises the bibliographic metadata for all UK doctoral theses listed in EThOS, the UK's national thesis service. We estimate the data covers around 98% of all PhDs ever awarded by UK Higher Education institutions, dating back to 1787. Thesis metadata from every PhD-awarding university in the UK is included. You can investigate and re-use this unique collection of UK universities' PhD thesis data to analyse trends in postgraduate research, make connections between researchers, apply large data analysis, improve citation of theses and many more applications.
http://inspire.ec.europa.eu/metadata-codelist/LimitationsOnPublicAccess/noLimitationshttp://inspire.ec.europa.eu/metadata-codelist/LimitationsOnPublicAccess/noLimitations
Collection of data from the PhD Thesis 'Development of coupled processes numerical models of tracer, colloid and radionuclide tranpsort in field migration experiments', submitted as part of the RATE HydroFrame WP5. This collection of data includes blank model files in COMSOL Multiphysics and PHREEQC, as described in the PhD thesis. Also included in this data package are different spreadsheets with model outputs from the model files that describe the transport of conservative tracers, colloids and radionuclides in experiments carried out at the Grimsel Test Site, Switzerland as part of the Colloid Radionuclide and Retardation (CRR) and the Colloid Formation and Migration (CFM) experiments (www.grimsel.com).
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Description not specified.........................
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Data supporting the Master thesis "Monitoring von Open Data Praktiken - Herausforderungen beim Auffinden von Datenpublikationen am Beispiel der Publikationen von Forschenden der TU Dresden" (Monitoring open data practices - challenges in finding data publications using the example of publications by researchers at TU Dresden) - Katharina Zinke, Institut für Bibliotheks- und Informationswissenschaften, Humboldt-Universität Berlin, 2023
This ZIP-File contains the data the thesis is based on, interim exports of the results and the R script with all pre-processing, data merging and analyses carried out. The documentation of the additional, explorative analysis is also available. The actual PDFs and text files of the scientific papers used are not included as they are published open access.
The folder structure is shown below with the file names and a brief description of the contents of each file. For details concerning the analyses approach, please refer to the master's thesis (publication following soon).
## Data sources
Folder 01_SourceData/
- PLOS-Dataset_v2_Mar23.csv (PLOS-OSI dataset)
- ScopusSearch_ExportResults.csv (export of Scopus search results from Scopus)
- ScopusSearch_ExportResults.ris (export of Scopus search results from Scopus)
- Zotero_Export_ScopusSearch.csv (export of the file names and DOIs of the Scopus search results from Zotero)
## Automatic classification
Folder 02_AutomaticClassification/
- (NOT INCLUDED) PDFs folder (Folder for PDFs of all publications identified by the Scopus search, named AuthorLastName_Year_PublicationTitle_Title)
- (NOT INCLUDED) PDFs_to_text folder (Folder for all texts extracted from the PDFs by ODDPub, named AuthorLastName_Year_PublicationTitle_Title)
- PLOS_ScopusSearch_matched.csv (merge of the Scopus search results with the PLOS_OSI dataset for the files contained in both)
- oddpub_results_wDOIs.csv (results file of the ODDPub classification)
- PLOS_ODDPub.csv (merge of the results file of the ODDPub classification with the PLOS-OSI dataset for the publications contained in both)
## Manual coding
Folder 03_ManualCheck/
- CodeSheet_ManualCheck.txt (Code sheet with descriptions of the variables for manual coding)
- ManualCheck_2023-06-08.csv (Manual coding results file)
- PLOS_ODDPub_Manual.csv (Merge of the results file of the ODDPub and PLOS-OSI classification with the results file of the manual coding)
## Explorative analysis for the discoverability of open data
Folder04_FurtherAnalyses
Proof_of_of_Concept_Open_Data_Monitoring.pdf (Description of the explorative analysis of the discoverability of open data publications using the example of a researcher) - in German
## R-Script
Analyses_MA_OpenDataMonitoring.R (R-Script for preparing, merging and analyzing the data and for performing the ODDPub algorithm)
https://heidata.uni-heidelberg.de/api/datasets/:persistentId/versions/1.1/customlicense?persistentId=doi:10.11588/DATA/9JKAVWhttps://heidata.uni-heidelberg.de/api/datasets/:persistentId/versions/1.1/customlicense?persistentId=doi:10.11588/DATA/9JKAVW
This dataset contains source code and system output used in the PhD thesis "Aspects of Coherence for Entity Analysis". This dataset is split into three parts corresponding to the chapters describing the three main contributions of the thesis: chapter3.tar.gz: Java source code for the entity linking system based on interleaved multitasking, system results, system output. Java and Python source code for automatic verification of entity linking results. Java source code for the Visual Entity Explorer. chapter4.tar.gz: Java and Scala source code for extracting pairs of terms and their dependency context from GigaWord and Wikilinks. chapter5.tar.gz: Python code used to run entity typing experiments.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
This dataset contains data used in the publication "Identifying publications of cumulative dissertation theses by bilingual text similarity. Evaluation of similarity methods on a new short text task". It included bibliographical data for German PhD theses (dissertations) and associated publications for cumulative dissertations. Not included is content from Elsevier's Scopus database used in the study, except item identifiers. Users with access to the data can use these for matching.
File diss_data.csv contains bibliographic data of dissertation theses obtained from German National Library and cleaned and postprocessed The columns are: REQUIZ_NORM_ID: Identifier for the thesis TITLE: Cleaned thesis title HEADING: Descriptor terms (German) AUTO_LANG: Language, either from original record or automatically derived from title
File ground_truth_pub_metadata.csv contains bibliographic data for identified consitutive publications of theses. If columns 2 to 7 are empty, the thesis did not include any publications ("stand-alone" or monograph thesis).
The columns are: REQUIZ_NORM_ID: Identifier for the thesis, for matching with the data in file SCOPUS_ID: Scopus ID for the identified publication AUTORS: Author names of the publication as in the original thesis citation YEAR: Publication year of the publication as in the original thesis citation TITLE: Publication title as in the original thesis citation SOURCETITLE: Source title as in the original thesis citation PAGES: Page information of the publication as in the original thesis citation
Scopus identifiers are published with permission by Elsevier.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
This data set contains the research data for the master's thesis: Integrating Explainability into Federated Learning: A Non-functional Requirement Perspective. The master's thesis was written by Nicolas Sebastian Schuler at the Computer Science Department at Karlsruhe Institute for Technology (KIT) in Germany. The data set contains: - Associate Jupyter notebooks for reproducing the figures in the master's thesis. - Generated experiment data by the federated learning simulations. - Results of the user survey conducted for the master's thesis. - Used Python Libraries. It also includes the submitted final thesis. Notice: The research data is split into multiple chunks and can be combined via the following command after downloading: $ cat thesis-results-part-* > thesis-results.tar.zst and extracted via: $ tar --zstd -xvf thesis-results.tar.zst
MIT Licensehttps://opensource.org/licenses/MIT
License information was derived automatically
This repository contains code and data related to the underlying PhD thesis: Data-driven methods to design, learn, and interpret complex materials across scales. The repository is divided into the individual codes and datasets of each chapter. Chapter 2 explores the inverse design of 2D metamaterials for elastic properties, utilizing machine learning techniques to optimize material structure and performance. Chapter 3 focuses on learning hyperelastic material models without relying on stress data, employing data-driven approaches to predict material behavior under large strains. Chapter 4 extends this by developing interpretable hyperelastic material models, ensuring both accuracy and physical consistency without stress data. Chapter 5 explores the inverse design of 3D metamaterials under finite strains and applies novel ML frameworks to design these complex material structures. Chapter 6 investigates the use of deep learning to uncover key predictors of thermal conductivity in covalent organic frameworks (COFs) and reveals new insights into the relationship between molecular structure and thermal transport. Chapter 7 introduces a graph grammar-based approach for generating novel polymers in data-scarce settings, thus combines computational design with minimal data.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Files containing the data of the synhtetic and experimental tests run during my PhD at Nantes Université, Ecole Centrale Nantes, CNRS, GeM, UMR 6183, F-44000 Nantes, France.
Short documentation can be found in the PDF file.
More information about the tests can be found in my PhD manuscript (link to be added soon).
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Supplementary Tables for the PhD thesis 'Interpreting Genetic Variants in the Context of Autism' by Catriona Miller. Tables are uploaded as Excel files containing all supplementary tables for each chapter (2-4) separately.
The script named "Trip_initial_script" contains the code for loading the Xsens mvnx files and calculates the percentage of collision. The script named "Trip_calculating" contains the calculations of the vertical position of center of mass, the linear momentum, the recovery strategies and the Statistical nonParametric Mapping. To run the script "Trip_calculating" - Please change the path in the function "Saveas" after each figure! There are six places in the script. Remember to unzip the files! This data is from a master thesis in Sports technology.
This dataset was created by Shifat Habib
Attribution-NonCommercial 4.0 (CC BY-NC 4.0)https://creativecommons.org/licenses/by-nc/4.0/
License information was derived automatically
To identify relevant actors for the governance of co-produced forest nature's contributions to people (NCP) the researchers conducted a social-network analysis based on 39 semi-structured interviews with foresters and conservation managers. These interviews were conducted across three case study sites in Germany: Schorfheide-Chorin in the Northeast, Hainich-Dün in the Centre, and Schwäbische Alb in the Southwest. All three case study sites belong to the large-scale and long-term research platform Biodiversity Exploratories. The researchers employed a predefined coding set to analyse the interviews and grasp the relationships between different actors based on the anthropogenic capitals they used to co-produce forest nature's contributions to people (NCP). To secure the interviewees anonymity this coding cannot be published. Therefore, this data set is limited to this coding set.
https://dataverse.nl/api/datasets/:persistentId/versions/1.1/customlicense?persistentId=doi:10.34894/NFUKLQhttps://dataverse.nl/api/datasets/:persistentId/versions/1.1/customlicense?persistentId=doi:10.34894/NFUKLQ
This dataset includes a selection of the interview transcripts used in the PhD thesis The Many Faces of Ravana. Ravanisation: The Revitalisation of Ravana among Sinhalese Buddhists in Post-War Sri Lanka. The fieldwork for this PhD thesis was conducted in the years 2016, 2017, and 2018 in Sri Lanka to investigate the popularity of the mythological figure Ravana among the Sinhalese Buddhist majority. The main question of the thesis is: What kinds of representations of Ravana have emerged among Sinhalese Buddhists in post-war Sri Lanka (2009 onwards), how do they take shape on local levels, and how does the interest in Ravana – including these Ravana representations – relate to the hegemonic Sinhalese Buddhist ethno-nationalist perception of Sri Lanka as a Sinhalese Buddhist country?
https://heidata.uni-heidelberg.de/api/datasets/:persistentId/versions/1.0/customlicense?persistentId=doi:10.11588/DATA/RAS7U7https://heidata.uni-heidelberg.de/api/datasets/:persistentId/versions/1.0/customlicense?persistentId=doi:10.11588/DATA/RAS7U7
This dataset contains source code and data used in the PhD thesis "Metrics of Graph-Based Meaning Representations with Applications from Parsing Evaluation to Explainable NLG Evaluation and Semantic Search". The dataset is split into five repositories: S3BERT: Source code to run experiments for chapter 9 "Building efficient and effective similarity models from MR metrics". amr-metric-suite, weisfeiler-leman-amr-metrics: Source code to run metric experiments for chapters 4, 5, 6. amr-argument-sim: Source code to run experiments for chapter 8 "Exploring argumentation with MR metrics". bamboo-amr-benchmark: Benchmark for testing and developing metrics (chapter 5).
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
This dataset is used in Master thesis on topic "The impact of upholding environmental, social and governance principles on the market value of capital-intensive companies". The full dataset consits data on Public american companies included in S&P 500 index, traded on the New York Stock Exchange. There are data on ESG-score and its components (E, S, G), as well as components of Envitonmental pillar score. Additionaly dataset includes financial data, like market capitalization, leverage, ROCE, Capex and etc. The main sources of data are Thomson Reuters Eikon and Bloomberg terminals, along with Form 10-k by SEC. The final sample consists of 52 capital-intensive companies, time horizon: 2012-2021 [520 observations in total].