9 datasets found
  1. Z

    Data from: Reliability Analysis of Random Telegraph Noisebased True Random...

    • data.niaid.nih.gov
    • zenodo.org
    Updated Sep 30, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Ranjan, Alok (2024). Reliability Analysis of Random Telegraph Noisebased True Random Number Generators [Dataset]. https://data.niaid.nih.gov/resources?id=zenodo_13169457
    Explore at:
    Dataset updated
    Sep 30, 2024
    Dataset provided by
    Pey, Kin Leong
    Ranjan, Alok
    O'Shea, Sean J.
    Thamankar, Dr. Ramesh
    Zanotti, Tommaso
    Raghavan, Nagarajan
    PUGLISI, Francesco Maria
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description
    • Repository author: Tommaso Zanotti* email: tommaso.zanotti@unimore.it or francescomaria.puglisi@unimore.it * Version v1.0

    This repository includes MATLAB files and datasets related to the IEEE IIRW 2023 conference proceeding:T. Zanotti et al., "Reliability Analysis of Random Telegraph Noisebased True Random Number Generators," 2023 IEEE International Integrated Reliability Workshop (IIRW), South Lake Tahoe, CA, USA, 2023, pp. 1-6, doi: 10.1109/IIRW59383.2023.10477697

    The repository includes:

    The data of the bitmaps reported in Fig. 4, i.e., the results of the simulation of the ideal RTN-based TRNG circuit for different reseeding strategies. To load and plot the data use the "plot_bitmaps.mat" file.

    The result of the circuit simulations considering the EvolvingRTN from the HfO2 device shown in Fig. 7, for two Rgain values. Specifically, the data is contained in the following csv files:

    "Sim_TRNG_Circuit_HfO2_3_20s_Vth_210m_no_Noise_Ibias_11n.csv" (lower Rgain)

    "Sim_TRNG_Circuit_HfO2_3_20s_Vth_210m_no_Noise_Ibias_4_8n.csv" (higher Rgain)

    The result of the circuit simulations considering the temporary RTN from the SiO2 device shown in Fig. 8. Specifically, the data is contained in the following csv files:

    "Sim_TRNG_Circuit_SiO2_1c_300s_Vth_180m_Noise_Ibias_1.5n.csv" (ref. Rgain)

    "Sim_TRNG_Circuit_SiO2_1c_100s_200s_Vth_180m_Noise_Ibias_1.575n.csv" (lower Rgain)

    "Sim_TRNG_Circuit_SiO2_1c_100s_200s_Vth_180m_Noise_Ibias_1.425n.csv" (higher Rgain)

  2. OpenCon Application Data

    • figshare.com
    txt
    Updated Jun 4, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    OpenCon 2015; SPARC; Right to Research Coalition (2023). OpenCon Application Data [Dataset]. http://doi.org/10.6084/m9.figshare.1512496.v1
    Explore at:
    txtAvailable download formats
    Dataset updated
    Jun 4, 2023
    Dataset provided by
    Figsharehttp://figshare.com/
    figshare
    Authors
    OpenCon 2015; SPARC; Right to Research Coalition
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    OpenCon 2015 Application Open Data

    The purpose of this document is to accompany the public release of data collected from OpenCon 2015 applications.Download & Technical Information The data can be downloaded in CSV format from GitHub here: https://github.com/RightToResearch/OpenCon-2015-Application-Data The file uses UTF8 encoding, comma as field delimiter, quotation marks as text delimiter, and no byte order mark.

    License and Requests

    This data is released to the public for free and open use under a CC0 1.0 license. We have a couple of requests for anyone who uses the data. First, we’d love it if you would let us know what you are doing with it, and share back anything you develop with the OpenCon community (#opencon / @open_con ). Second, it would also be great if you would include a link to the OpenCon 2015 website (www.opencon2015.org) wherever the data is used. You are not obligated to do any of this, but we’d appreciate it!

    Data Fields

    Unique ID

    This is a unique ID assigned to each applicant. Numbers were assigned using a random number generator.

    Timestamp

    This was the timestamp recorded by google forms. Timestamps are in EDT (Eastern U.S. Daylight Time). Note that the application process officially began at 1:00pm EDT June 1 ended at 6:00am EDT on June 23. Some applications have timestamps later than this date, and this is due to a variety of reasons including exceptions granted for technical difficulties, error corrections (which required re-submitting the form), and applications sent in via email and later entered manually into the form. [a]

    Gender

    Mandatory. Choose one from list or fill-in other. Options provided: Male, Female, Other (fill in).

    Country of Nationality

    Mandatory. Choose one option from list.

    Country of Residence

    Mandatory. Choose one option from list.

    What is your primary occupation?

    Mandatory. Choose one from list or fill-in other. Options provided: Undergraduate student; Masters/professional student; PhD candidate; Faculty/teacher; Researcher (non-faculty); Librarian; Publisher; Professional advocate; Civil servant / government employee; Journalist; Doctor / medical professional; Lawyer; Other (fill in).

    Select the option below that best describes your field of study or expertise

    Mandatory. Choose one option from list.

    What is your primary area of interest within OpenCon’s program areas?

    Mandatory. Choose one option from list. Note: for the first approximately 24 hours the options were listed in this order: Open Access, Open Education, Open Data. After that point, we set the form to randomize the order, and noticed an immediate shift in the distribution of responses.

    Are you currently engaged in activities to advance Open Access, Open Education, and/or Open Data?

    Mandatory. Choose one option from list.

    Are you planning to participate in any of the following events this year?

    Optional. Choose all that apply from list. Multiple selections separated by semi-colon.

    Do you have any of the following skills or interests?

    Mandatory. Choose all that apply from list or fill-in other. Multiple selections separated by semi-colon. Options provided: Coding; Website Management / Design; Graphic Design; Video Editing; Community / Grassroots Organizing; Social Media Campaigns; Fundraising; Communications and Media; Blogging; Advocacy and Policy; Event Logistics; Volunteer Management; Research about OpenCon's Issue Areas; Other (fill-in).

    Data Collection & Cleaning

    This data consists of information collected from people who applied to attend OpenCon 2015. In the application form, questions that would be released as Open Data were marked with a caret (^) and applicants were asked to acknowledge before submitting the form that they understood that their responses to these questions would be released as such. The questions we released were selected to avoid any potentially sensitive personal information, and to minimize the chances that any individual applicant can be positively identified. Applications were formally collected during a 22 day period beginning on June 1, 2015 at 13:00 EDT and ending on June 23 at 06:00 EDT. Some applications have timestamps later than this date, and this is due to a variety of reasons including exceptions granted for technical difficulties, error corrections (which required re-submitting the form), and applications sent in via email and later entered manually into the form. Applications were collected using a Google Form embedded at http://www.opencon2015.org/attend, and the shortened bit.ly link http://bit.ly/AppsAreOpen was promoted through social media. The primary work we did to clean the data focused on identifying and eliminating duplicates. We removed all duplicate applications that had matching e-mail addresses and first and last names. We also identified a handful of other duplicates that used different e-mail addresses but were otherwise identical. In cases where duplicate applications contained any different information, we kept the information from the version with the most recent timestamp. We made a few minor adjustments in the country field for cases where the entry was obviously an error (for example, electing a country listed alphabetically above or below the one indicated elsewhere in the application). We also removed one potentially offensive comment (which did not contain an answer to the question) from the Gender field and replaced it with “Other.”

    About OpenCon

    OpenCon 2015 is the student and early career academic professional conference on Open Access, Open Education, and Open Data and will be held on November 14-16, 2015 in Brussels, Belgium. It is organized by the Right to Research Coalition, SPARC (The Scholarly Publishing and Academic Resources Coalition), and an Organizing Committee of students and early career researchers from around the world. The meeting will convene students and early career academic professionals from around the world and serve as a powerful catalyst for projects led by the next generation to advance OpenCon's three focus areas—Open Access, Open Education, and Open Data. A unique aspect of OpenCon is that attendance at the conference is by application only, and the majority of participants who apply are awarded travel scholarships to attend. This model creates a unique conference environment where the most dedicated and impactful advocates can attend, regardless of where in the world they live or their access to travel funding. The purpose of the application process is to conduct these selections fairly. This year we were overwhelmed by the quantity and quality of applications received, and we hope that by sharing this data, we can better understand the OpenCon community and the state of student and early career participation in the Open Access, Open Education, and Open Data movements.

    Questions

    For inquires about the OpenCon 2015 Application data, please contact Nicole Allen at nicole@sparc.arl.org.

  3. P

    darpa_sd2_perovskites Dataset

    • paperswithcode.com
    Updated May 25, 2020
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Ian M. Pendleton; Mary K. Caucci; Michael Tynes; Aaron Dharna; Mansoor Ani Najeeb Nellikkal; Zhi Li; Emory M. Chan; Alexander J. Norquist; and Joshua Schrier (2020). darpa_sd2_perovskites Dataset [Dataset]. https://paperswithcode.com/dataset/darpa-sd2-perovskites
    Explore at:
    Dataset updated
    May 25, 2020
    Authors
    Ian M. Pendleton; Mary K. Caucci; Michael Tynes; Aaron Dharna; Mansoor Ani Najeeb Nellikkal; Zhi Li; Emory M. Chan; Alexander J. Norquist; and Joshua Schrier
    Description

    Included in this content:

    0045.perovksitedata.csv - main dataset used in this article. A more detailed description can be found in the “dataset overview” section below Chemical Inventory.csv - the hand curated file of all chemicals used in the construction of the perovskite dataset. This file includes identifiers, chemical properties, and other information. ExcessMolarVolumeData.xlsx - record of experimental data, computations, and final dataset used in the generation of the excess molar volume plots. MLModelMetrics.xlsx - all of the ML metrics organized in one place (excludes reactant set specific breakdown, see ML_Logs.zip for those files). OrganoammoniumDensityDataset.xlsx - complete set of the data used to generate the density values. Example calculations included. model_matchup_main.py - python pipeline used to generate all of the ML runs associated with the article. More detailed instructions on the operation of this code is included in the “ML Code” Section below. This file is also hosted on GIT: https://github.com/ipendlet/MLScripts/blob/master/temp_densityconc/model_matchup_main_20191231.py

    SolutionVolumeDataset - complete set of 219 solutions in the perovskite dataset. Tabs include the automatically generated reagent information from ESCALATE, hand curated reagent information from early runs, and the generation of the dataset used in the creation of Figure 5. error_auditing.zip - code and historical datasets used for reporting the dataset auditing. “AllCode.zip” which contains: model_matchup_main_20191231.py - python pipeline used to generate all of the ML runs associated with the article. More detailed instructions on the operation of this code is included in the “ML Code” Section below. This file is also hosted on GIT: https://github.com/ipendlet/MLScripts/blob/master/temp_densityconc/0045.perovskitedata.csv VmE_CurveFitandPlot.py - python code for generating the third order polynomial fit to the VmE vs mole fraction of FAH included in the main text. Requires the ‘MolFractionResults.csv’ to function (also included). Calculation_Vm_Ve_CURVEFITTING.nb - mathematica code for generating the third order polynomial fit to the VmE vs mole fraction of FAH included in the main text.
    Covariance_Analysis.py - python code for ingesting and plotting the covariance of features and volumes in the perovskite dataset. Includes renaming dictionaries used for the publication. FeatureComparison_Plotting.py - python code for reading in and plotting features for the ‘GBT’ and ‘OHGBT’ folders in this directory. The code parses the contents of these folders and generates feature comparison metrics used for Figure 9 and the associated Figure S8. Some assembly required. Requirements.txt - all of the packages used in the generation of this paper 0045.perovskitedata.csv - the main dataset described throughout the article. This file is required to run some of the code and is therefore kept near the code.

    “ML_Logs.zip” which contains: A folder describing every model generated for this article. In each folder there are a number of files: Features_named_important.csv and features_value_importance.csv - these files are linked together and describe the weighted feature contributions from features (only present for GBT models) AnalysisLog.txt - Log file of the run including all options, data curation and model training summaries
    LeaveOneOut_Summary.csv - Results of the leave-one-reactant set-out studies on the model (if performed) LOOModelInfo.txt - Hyperparameter information for each model in the study (associated with the given dataset, sometimes includes duplicate runs). STTSModelInfo.txt - Hyperparameter information for each model in the study (associated with the given dataset, sometimes includes duplicate runs). StandardTestTrain_Summary.csv - Results of the 6 fold cross validation ML performance (for the hold out case) LeaveOneOut_FullDataset_ByAmine.csv - Results of the leave-one-reactant set-out studies performed on the full dataset (all experiments) specified by reactant set (delineated by the amine) LeaveOneOut_StratifiedData_ByAmine.csv - Results of the leave-one-reactant set-out studies performed on a random stratified sample (96 random experiments) specified by reactant set (delineated by the amine) model_matchup_main_*.py - code used to generate all of the runs contained in a particular folder. The code is exactly what was used at run time to generate a given dataset (requires 0045.perovskitedata.csv file to run).

  4. Z

    Data from: Tough Tables: Carefully Evaluating Entity Linking for Tabular...

    • data.niaid.nih.gov
    • zenodo.org
    Updated Jan 14, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Bianchi, Federico (2023). Tough Tables: Carefully Evaluating Entity Linking for Tabular Data [Dataset]. https://data.niaid.nih.gov/resources?id=ZENODO_3840646
    Explore at:
    Dataset updated
    Jan 14, 2023
    Dataset provided by
    Palmonari, Matteo
    Jiménez-Ruiz, Ernesto
    Cutrona, Vincenzo
    Bianchi, Federico
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Tough Tables (2T) is a dataset designed to evaluate table annotation approaches in solving the CEA and CTA tasks. The dataset is compliant with the data format used in SemTab 2019, and it can be used as an additional dataset without any modification. The target knowledge graph is DBpedia 2016-10. Check out the 2T GitHub repository for more details about the dataset generation.

    New in v2.0: We release the updated version of 2T_WD! The target knowledge graph is Wikidata (online instance) and the dataset complies with the SemTab 2021 data format.

    This work is based on the following paper:

    Cutrona, V., Bianchi, F., Jimenez-Ruiz, E. and Palmonari, M. (2020). Tough Tables: Carefully Evaluating Entity Linking for Tabular Data. ISWC 2020, LNCS 12507, pp. 1–16.

    Note on License: This dataset includes data from the following sources. Refer to each source for license details: - Wikipedia https://www.wikipedia.org/ - DBpedia https://dbpedia.org/ - Wikidata https://www.wikidata.org/ - SemTab 2019 https://doi.org/10.5281/zenodo.3518539 - GeoDatos https://www.geodatos.net - The Pudding https://pudding.cool/ - Offices.net https://offices.net/ - DATA.GOV https://www.data.gov/

    THIS DATA IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.

    Changelog:

    v2.0

    New GT for 2T_WD

    A few entities have been removed from the CEA GT, because they are no longer represented in WD (e.g., dbr:Devonté points to wd:Q21155080, which does not exist)

    Tables codes and values differ from the previous version, because of the random noise.

    Updated ancestor/descendant hierarchies to evaluate CTA.

    v1.0

    New Wikidata version (2T_WD)

    Fix header for tables CTRL_DBP_MUS_rock_bands_labels.csv and CTRL_DBP_MUS_rock_bands_labels_NOISE2.csv (column 2 was reported with id 1 in target - NOTE: the affected column has been removed from the SemTab2020 evaluation)

    Remove duplicated entries in tables

    Remove rows with wrong values (e.g., the Kazakhstan entity has an empty name "''")

    Many rows and noised columns are shuffled/changed due to the random noise generator algorithm

    Remove row "Florida","Floorida","New York, NY" from TOUGH_WEB_MISSP_1000_us_cities.csv (and all its NOISE1 variants)

    Fix header of tables:

    CTRL_WIKI_POL_List_of_current_monarchs_of_sovereign_states.csv

    CTRL_WIKI_POL_List_of_current_monarchs_of_sovereign_states_NOISE2.csv

    TOUGH_T2D_BUS_29414811_2_4773219892816395776_videogames_developers.csv

    TOUGH_T2D_BUS_29414811_2_4773219892816395776_videogames_developers_NOISE2.csv

    v0.1-pre

    First submission. It contains only tables, without GT and Targets.

  5. Z

    ATM: Black-box Test Case Minimization based on Test Code Similarity and...

    • data.niaid.nih.gov
    • zenodo.org
    Updated Mar 25, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Rongqi Pan (2023). ATM: Black-box Test Case Minimization based on Test Code Similarity and Evolutionary Search – Replication Package [Dataset]. https://data.niaid.nih.gov/resources?id=zenodo_7455765
    Explore at:
    Dataset updated
    Mar 25, 2023
    Dataset provided by
    Taher A. Ghaleb
    Lionel Briand
    Rongqi Pan
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    This is the replication package associated with the paper "ATM: Black-box Test Case Minimization based on Test Code Similarity and Evolutionary Search" accepted at the 45th IEEE/ACM International Conference on Software Engineering (ICSE 2023) – Technical Track. Cite this paper using the following:

    @inproceedings{pan2023atm, title={ATM: Black-box Test Case Minimization based on Test Code Similarity and Evolutionary Search}, author={Pan, Rongqi and Ghaleb, Taher A. and Briand, Lionel}, booktitle={Proceedings of the 45th IEEE/ACM International Conference on Software Engineering}, year={2023}, pages={1--12} }

    Replication Package Contents: The replication package contains all the necessary data and code required to reproduce the results reported in the paper. We also provide the results for other minimization budgets, and detailed FDR, execution time, and statistical test results. In addition, we provide the data and code required to reproduce the results of baselines techniques: FAST-R and random minimization.

    Data: We provide in the Data directory the data used in our experiments, which is based on 16 projects from Defects4J, whose characteristics can be found in Data/subject_projects.csv.

    Code: We provide in the Code directory the code and scripts (Java, Python, and Bash) required to run the experiments and reproduce the results.

    Results: We provide in the Results directory the results for each technique independently, and also a summary of all results together for comparison purposes. The source code for this step is in the Code/ATM/CodeToAST directory. The source code for this step is in the Code/ATM/Similarity directory.

    ATM - Code to AST transformation:

    Requirements: * Eclipse IDE (we used 2021-12) * The libraries (the .jar files in the Code/ATM/CodeToAST/lib directory)

    Input: All zipped data files should be unzipped before running each step. * Data/test_suites/all_test_cases.zip → Data/test_suites/all_test_cases * Data/test_suites/changed_test_cases.zip → Data/test_suites/changed_test_cases * Data/test_suites/relevant_test_cases.zip → Data/test_suites/relevant_test_cases

    Output: * Data/ATM/ASTs/all_test_cases * Data/ATM/ASTs/changed_test_cases

    Running the experiment: To generate ASTS for all test cases in the project test suites, the Code/ATM/CodeToAST/src/CodeToAST.java file should be compiled and run using the Eclipse IDE by including all the required .jar files in the Code/ATM/CodeToAST/lib directory as part of the classpath. A bash script is provided along with a pre-generated .jar file in the Code/ATM/CodeToAST/bin directory to run this step, as follows:

    cd Code/ATM/CodeToAST bash transform_code_to_ast.sh

    Each test file in the Data/test_suites/all_test_cases and Data/test_suites/changed_test_cases directories is parsed to generate a corresponding AST for each test case method (saved in an XML format in Data/ATM/ASTs/all_test_cases and Data/ATM/ASTs/changed_test_cases for each project version)

    ATM - Similarity Measurement:

    Requirements: * Eclipse IDE (we used 2021-12) * The libraries (the .jar files in the Code/ATM/Similarity/lib directory)

    Input: * Data/test_suites/all_test_cases * Data/test_suites/changed_test_cases

    Output: * Data/ATM/similarity_measurements

    Running the experiment: To measure the similarity between each pair of test cases, the Code/ATM/Similarity/src/SimilarityMeasurement.java file should be compiled and run using the Eclipse IDE by including all the required .jar files in the Code/ATM/Similarity/lib directory as part of the classpath. A bash script is provided along with a pre-generated .jar file in the Code/ATM/Similarity/bin directory to run this step, as follows:

    cd Code/ATM/Similarity bash measure_similarity.sh

    ASTs of each project in the Data/ATM/ASTs/all_test_cases and Data/ATM/ASTs/changed_test_cases directories are parsed to create pairs of ASTs containing one test case from the Data/ATM/ASTs/all_test_cases directory with another test case from the Data/ATM/ASTs/changed_test_cases directory (redundant pairs are discarded). Then, all similarity measurements are saved in the Data/ATM/similarity_measurements.zip file.

    Search-based Minimization Algorithms: The source code for this step is in the Code/ATM/Search directory.

    Requirements: To run this step, Python 3 is required (we used Python 3.10). Also, the libraries in the Code/AMT/Search/requirements.txt file should be installed, as follows:

    cd Code/ATM/Search pip install -r requirements.txt

    Input: * Data/ATM/similarity_measurements

    Output: * Results/ATM/minimization_results

    Running the experiment: To minimize the test suites in our dataset, the following bash script should be executed:

    bash minimize.sh

    All similarity measurements are parsed for each version of the projects, independently. Each version is run 10 times using three minimization budgets (25%, 50%, and 75%). Genetic Algorithm (GA) is run using four similarity measures, namely top-down, bottom-up, combined, and tree edit distance. NSGA-II is run using two combinations of similarity measures: top-down & bottom-up and combined & tree edit distance. The minimization results are generated in the Results/ATM/minimization_results directory.

    Evaluate results: To evaluate and summarize the minimization results, run the following:

    cd Code/ATM/Evaluation bash evaluate.sh

    This will generate summarized FDR and execution time results (per-project and per-version) for each minimization budget, which can all be found in Results/ATM. In this replication package, we provide the final, merged FDR with execution time results.

    Running FAST-R experiments ATM was compared to FAST-R, a state-of-the-art baseline, which is a set of test case minimization techniques called: FAST++, FAST-CS, FAST-pw, and FAST-all, which we adapted to our data and experimental setup.

    Requirements: To run this step, Python 3.7 is required. Also, the libraries in the Code/FAST-R/requirements.txt file should be installed, as follows:

    cd Code/FAST-R pip install -r requirements.txt

    Input: * Data/FAST-R/test_methods * Data/FAST-R/test_classes

    Output: * Results/FAST-R/test_methods/FDR_and_Exec_Time_Results_[budget]%_budget.csv * Results/FAST-R/test_classes/FDR_and_Exec_Time_Results_[budget]%_budget.csv

    To run FAST-R experiments, the following bash script should be executed:

    bash fast_r.sh test_methods #method level bash fast_r.sh test_classes #class level

    Results are generated in .csv files for each budget. For example, for the 50% budget, results are saved in FDR_and_Exec_Time_Results_50%_budget.csv in the Results/FAST-R/test_methods and Results/FAST-R/test_classes directories.

    Running the random minimization experiments ATM was also compared to random minimization as a standard baseline.

    Requirements: To run this step, Python 3 is required (we used Python 3.10). Also, the libraries in the Code/RandomMinimization/requirements.txt file should be installed, as follows:

    cd Code/RandomMinimization pip install -r requirements.txt

    Input: N/A

    Output: * Results/RandomMinimization/FDR_and_Exec_Time_Results_[budget]%_budget.csv

    To run the random selection experiments, the following bash script should be executed:

    bash random_minimization.sh

    Results are generated in .csv files for each budget. For example, for the 50% budget, results are saved in FDR_and_Exec_Time_Results_50%_budget.csv in the Results/RandomMinimization directory.

  6. n

    Statistical data of a clinical study of the severity of bronchial asthma in...

    • narcis.nl
    • data.mendeley.com
    Updated Dec 31, 2020
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Kozhyna, O (via Mendeley Data) (2020). Statistical data of a clinical study of the severity of bronchial asthma in children of the Kharkov region, 2017 [Dataset]. http://doi.org/10.17632/9wf2hhd4yf.2
    Explore at:
    Dataset updated
    Dec 31, 2020
    Dataset provided by
    Data Archiving and Networked Services (DANS)
    Authors
    Kozhyna, O (via Mendeley Data)
    Description

    2017 Part 1 Common statistical data.csv_ Bronchial Asthma (BA) is one of the most common chronic respiratory diseases among children. 70 children aged 6-17 years old were observed. The control group consisted of 20 virtually healthy children, randomized by age. Standard methods of examination in accordance with the unified clinical protocol “Bronchial Asthma in Children” (order of the Ministry of Healthcare of Ukraine on 08.10.2013, No 868) were applied. This file contains the results of the survey, which are grouped by factors. Each side corresponds to a specific patient, individual data (full name) which is encoded. The number before the factor corresponds to the category which name in accordance with the number is presented separately (see section classification of factors).

    Factors 14.age, 5.Complete blood count, 6.Urinalysis Test, 11.Spirogram, 12.Immunological status, 13: TSLP are presented in quantitative terms.

    Factors 1.SEVERE PERSISTENT, 1.MODERATE PERSISTENT are encoded with yes / no values ​​(1/0).

    From the table, 12 factors are removed, which occur as a result of the survey once or twice.

    Classification of factors id : name 0 : Description 1 : Course type 2 : Principal diagnosis (case taking) 3 : Anamnesis of disease 4 : Anamnesis of life 5 : Complete blood count 6 : Urinalysis Test 7 : Skin allergy test (pollen allergens) 8 : Skin allergy test(household allergens) 9 : Food allergy test 10 : Ig E 11 : Spirogram 12 : Immunological status 13 : TSLP 14 : Age 15 : Gender

    2017 Part 2 Separeted statistical data.csv_ This paper presents selected data of clinical and paraclinical studies of 70 children in the Kharkiv region, suffering from bronchial asthma, and 20 children in the control group. Individual patient data coded. The study group of patients (Table #1 in the file "2017 Part 1 Common statistical data.csv") was divided into 2 parts: 1)a separated group (Table #1 in the file "2017 Part 2 Separated statistical data.csv") 2)a test group (Table #1 in the file "2017 Part 3 Test statistical data.csv").

    In the formation of the test group used a random number generator and the algorithm Random random = new Random(31); ... nextNumber = random.nextInt(size-1); size = 90 - the number of patients who participated in clinical trials.

    2017 Part 3 Test statistical data.csv_ This paper presents selected data of clinical and paraclinical studies of 70 children in the Kharkiv region, suffering from bronchial asthma, and 20 children in the control group. Individual patient data coded. The study group of patients (Table #1 in the file "2017 Part 1 Common statistical data.csv") was divided into 2 parts: 1)a separated group (Table #1 in the file "2017 Part 2 Separated statistical data.csv") 2)a test group (Table #1 in the file "2017 Part 3 Test statistical data.csv").

  7. RailEnV-PASMVS: a dataset for multi-view stereopsis training and...

    • zenodo.org
    • data.niaid.nih.gov
    bin, csv, png, txt +1
    Updated Jul 18, 2024
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    André Broekman; André Broekman; Petrus Johannes Gräbe; Petrus Johannes Gräbe (2024). RailEnV-PASMVS: a dataset for multi-view stereopsis training and reconstruction applications [Dataset]. http://doi.org/10.5281/zenodo.5233840
    Explore at:
    bin, csv, txt, zip, pngAvailable download formats
    Dataset updated
    Jul 18, 2024
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    André Broekman; André Broekman; Petrus Johannes Gräbe; Petrus Johannes Gräbe
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    A Perfectly Accurate, Synthetic dataset featuring a virtual railway EnVironment for Multi-View Stereopsis (RailEnV-PASMVS) is presented, consisting of 40 scenes and 79,800 renderings together with ground truth depth maps, extrinsic and intrinsic camera parameters and binary segmentation masks of all the track components and surrounding environment. Every scene is rendered from a set of 3 cameras, each positioned relative to the track for optimal 3D reconstruction of the rail profile. The set of cameras is translated across the 100-meter length of tangent (straight) track to yield a total of 1,995 camera views. Photorealistic lighting of each of the 40 scenes is achieved with the implementation of high-definition, high dynamic range (HDR) environmental textures. Additional variation is introduced in the form of camera focal lengths, random noise for the camera location and rotation parameters and shader modifications of the rail profile. Representative track geometry data is used to generate random and unique vertical alignment data for the rail profile for every scene. This primary, synthetic dataset is augmented by a smaller image collection consisting of 320 manually annotated photographs for improved segmentation performance. The specular rail profile represents the most challenging component for MVS reconstruction algorithms, pipelines and neural network architectures, increasing the ambiguity and complexity of the data distribution. RailEnV-PASMVS represents an application specific dataset for railway engineering, against the backdrop of existing datasets available in the field of computer vision, providing the precision required for novel research applications in the field of transportation engineering.

    File descriptions

    • RailEnV-PASMVS.blend (227 Mb) - Blender file (developed using Blender version 2.8.1) used to generate the dataset. The Blender file packs only one of the HDR environmental textures to use as an example, along with all the other asset textures.
    • RailEnV-PASMVS_sample.png (28 Mb) - A visual collage of 30 scenes, illustrating the variability introduced by using different models, illumination, material properties and camera focal lengths.
    • geometry.zip (2 Mb) - Geometry CSV files used for scenes 01 to 20. The Bezier curve defines the geometry of the rail profile (10 mm intervals).
    • PhysicalDataset.7z (2.0 Gb) - A smaller, secondary dataset of 320 manually annotated photographs of railway environments; only the railway profiles are annotated.
    • 01.7z-40.7z (2.0 Gb each) - Archive of every scene (01 through 40).
    • all_list.txt, training_list.txt, validation_list.txt - Text files containing the all the scene names, together with those used for validation (validation_list.txt) and training (training_list.txt), used by MVSNet.
    • index.csv - CSV file provides a convenient reference for all the sample files, linking the corresponding file and relative data path.

    Steps to reproduce

    The open source Blender software suite (https://www.blender.org/) was used to generate the dataset, with the entire pipeline developed using the exposed Python API interface. The camera trajectory is kept fixed for all 40 scenes, except for small perturbations introduced in the form of random noise to increase the camera variation. The camera intrinsic information was initially exported as a single CSV file (scene.csv) for every scene, from which the camera information files were generated; this includes the focal length (focalLengthmm), image sensor dimensions (pixelDimensionX, pixelDimensionY), position, coordinate vector (vectC) and rotation vector (vectR). The STL model files, as provided in this data repository, were exported directly from Blender, such that the geometry/scenes can be reproduced. The data processing below is written for a Python implementation, transforming the information from Blender's coordinate system into universal rotation (R_world2cv) and translation (T_world2cv) matrices.

    import numpy as np
    from scipy.spatial.transform import Rotation as R
    
    #The intrinsic matrix K is constructed using the following formulation:
    focalLengthPixel = focalLengthmm x pixelDimensionX / sensorWidthmm
    K = [[focalLengthPixel, 0, dimX/2],
       [0, focalPixel, dimY/2],
       [0, 0, 1]]
    
    #The rotation vector as provided by Blender was first transformed to a rotation matrix:
    r = R.from_euler('xyz', vectR, degrees=True)
    matR = r.as_matrix()
    
    #Transpose the rotation matrix, to find matrix from the WORLD to BLENDER coordinate system:
    R_world2bcam = np.transpose(matR)
    
    #The matrix describing the transformation from BLENDER to CV/STANDARD coordinates is:
    R_bcam2cv = np.array([[1, 0, 0],
                   [0, -1, 0],
                   [0, 0, -1]])
    
    #Thus the representation from WORLD to CV/STANDARD coordinates is:
    R_world2cv = R_bcam2cv.dot(R_world2bcam)
    
    #The camera coordinate vector requires a similar transformation moving from BLENDER to WORLD coordinates:
    T_world2bcam = -1 * R_world2bcam.dot(vectC)
    T_world2cv = R_bcam2cv.dot(T_world2bcam)

    The resulting R_world2cv and T_world2cv matrices are written to the camera information file using exactly the same format as that of BlendedMVS developed by Dr. Yao. The original rotation and translation information can be found by following the process in reverse. Note that additional steps were required to convert from Blender's unique coordinate system to that of OpenCV; this ensures universal compatibility in the way that the camera intrinsic and extrinsic information is provided.

    Equivalent GPS information is provided (gps.csv), whereby the local coordinate frame is transformed into equivalent GPS information, centered around the Engineering 4.0 campus, University of Pretoria, South Africa. This information is embedded within the JPG files as EXIF data.

  8. Experimental Data Set for the study "Exploratory Landscape Analysis is...

    • zenodo.org
    • explore.openaire.eu
    csv, text/x-python +1
    Updated Jan 28, 2021
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Quentin Renau; Carola Doerr; Carola Doerr; Johann Dreo; Johann Dreo; Benjamin Doerr; Benjamin Doerr; Quentin Renau (2021). Experimental Data Set for the study "Exploratory Landscape Analysis is Strongly Sensitive to the Sampling Strategy" [Dataset]. http://doi.org/10.5281/zenodo.3886816
    Explore at:
    text/x-python, csv, zipAvailable download formats
    Dataset updated
    Jan 28, 2021
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Quentin Renau; Carola Doerr; Carola Doerr; Johann Dreo; Johann Dreo; Benjamin Doerr; Benjamin Doerr; Quentin Renau
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    This are the feature values used in the study "Exploratory Landscape Analysis is Strongly Sensitive to the Sampling Strategy".

    The dataset regroups feature values for every "cheap" features available in the R package flacco and are computed using 5 sampling strategies and in dimension \($d=5$\):

    1. Random: the classical Mersenne-Twister algorithm;
    2. Randu: a random number generator that is notoriously bad;
    3. LHS: a centered Latin Hypercube Design;
    4. iLHS: an improved Latin Hypercube Design;
    5. Sobol: points extracted from a Sobol' low-discrepancy sequence.

    The csv file features_summury_dim_5_ppsn.csv regroups 100 values for every features whereas features_summury_dim_5_ppsn_median.csv regroups for every feature the median of the 100 values.

    In the folder PPSN_feature_plots are the histograms of feature values on the 24 COCO functions for 3 sampling strategies: Random, LHS and Sobol.

    The Python file sampling_ppsn.py is the code used to generate the sample points from which the feature values are computed.

    The file stats50_knn_dt.csv provide the raw data of median and IQR (inter quartile interval) for the heatmaps and boxplots available in the paper.

    Finally, the files results_classif_knn100.csv (resp. dt) provide the accuracy of 100 classifications for every settings.

  9. 7+ Million Company Dataset

    • kaggle.com
    zip
    Updated May 10, 2019
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    People Data Labs (2019). 7+ Million Company Dataset [Dataset]. https://www.kaggle.com/datasets/peopledatalabssf/free-7-million-company-dataset
    Explore at:
    zip(291957415 bytes)Available download formats
    Dataset updated
    May 10, 2019
    Authors
    People Data Labs
    Description

    Dataset

    This dataset was created by People Data Labs

    Contents

  10. Not seeing a result you expected?
    Learn how you can add new datasets to our index.

Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
Ranjan, Alok (2024). Reliability Analysis of Random Telegraph Noisebased True Random Number Generators [Dataset]. https://data.niaid.nih.gov/resources?id=zenodo_13169457

Data from: Reliability Analysis of Random Telegraph Noisebased True Random Number Generators

Related Article
Explore at:
Dataset updated
Sep 30, 2024
Dataset provided by
Pey, Kin Leong
Ranjan, Alok
O'Shea, Sean J.
Thamankar, Dr. Ramesh
Zanotti, Tommaso
Raghavan, Nagarajan
PUGLISI, Francesco Maria
License

Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically

Description
  • Repository author: Tommaso Zanotti* email: tommaso.zanotti@unimore.it or francescomaria.puglisi@unimore.it * Version v1.0

This repository includes MATLAB files and datasets related to the IEEE IIRW 2023 conference proceeding:T. Zanotti et al., "Reliability Analysis of Random Telegraph Noisebased True Random Number Generators," 2023 IEEE International Integrated Reliability Workshop (IIRW), South Lake Tahoe, CA, USA, 2023, pp. 1-6, doi: 10.1109/IIRW59383.2023.10477697

The repository includes:

The data of the bitmaps reported in Fig. 4, i.e., the results of the simulation of the ideal RTN-based TRNG circuit for different reseeding strategies. To load and plot the data use the "plot_bitmaps.mat" file.

The result of the circuit simulations considering the EvolvingRTN from the HfO2 device shown in Fig. 7, for two Rgain values. Specifically, the data is contained in the following csv files:

"Sim_TRNG_Circuit_HfO2_3_20s_Vth_210m_no_Noise_Ibias_11n.csv" (lower Rgain)

"Sim_TRNG_Circuit_HfO2_3_20s_Vth_210m_no_Noise_Ibias_4_8n.csv" (higher Rgain)

The result of the circuit simulations considering the temporary RTN from the SiO2 device shown in Fig. 8. Specifically, the data is contained in the following csv files:

"Sim_TRNG_Circuit_SiO2_1c_300s_Vth_180m_Noise_Ibias_1.5n.csv" (ref. Rgain)

"Sim_TRNG_Circuit_SiO2_1c_100s_200s_Vth_180m_Noise_Ibias_1.575n.csv" (lower Rgain)

"Sim_TRNG_Circuit_SiO2_1c_100s_200s_Vth_180m_Noise_Ibias_1.425n.csv" (higher Rgain)

Search
Clear search
Close search
Google apps
Main menu