This text files aims to explain the procedure to perform both the data handling and the analyses in the paper: Description of the files: 1. Datasets PhenoAsreml.txt contains the observed values for the phenotypes described in the paper PhenoAsreml_scaled.txt contains the scaled (mean of zero and standard deviation of one) of the same phenotypes. THIS DATASET IS USED FOR BIVARIATE ANALYSES WITH 2x2 STRATA PhenoAsreml_scaledLong.txt contains the same information, organized differently because the dataset is reshaped from wide to long format. PhenoAsreml_scaledLong_3x3strata.txt contains the information for BackFat and BodyWeight, THIS DATASET IS USED FOR BIVARIATE ANALYSES WITH 3x3 STRATA PhenoAsreml_scaledLong_3x2strata.txt contains the information for BackFat and Adiponectin, THIS DATASET IS USED FOR BIVARIATE ANALYSES WITH 3x2 STRATA. It is only an example of dataset that needs to be generated in order to obtain all the results presented in the paper BEDERE_2023_Data_PedigreeHens.txt is the pedigree file (individual/sire/dam) traced back over 5 generations 2. Codes and parameter files BEDERE_2023_RScript_handlingdata_Long.R is an R code to reshape PhenoAsreml_scaled into PhenoAsreml_scaledLong and to subset it to generate PhenoAsreml_scaledLong_3x3strata for instance BEDERE_2023_ASREMLScript_bivariate_2x2strata.as (as well as ...3x2strata.as and ...3x3strata.as) are ASReml parameter files used to state the data, model specification and post-hoc calculation to ASReml Please, note that some variance components have been fixed in some analyses when the algorithm was struggling to converge. BEDERE_2023_RScript_BartlettTest.R is an R code to perform the Bartlett test. 2. Results examples Some output files of ASReml are provided to give an example of results for each type of bivariate analysis. The .asr file is the log of the program, explaining how the program ran The .res file is describing the residuals The .pvc file describes the variance components and provides the genetic parameters with their associated standard errors.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Complete dataset of “Film Circulation on the International Film Festival Network and the Impact on Global Film Culture”
A peer-reviewed data paper for this dataset is in review to be published in NECSUS_European Journal of Media Studies - an open access journal aiming at enhancing data transparency and reusability, and will be available from https://necsus-ejms.org/ and https://mediarep.org
Please cite this when using the dataset.
Detailed description of the dataset:
1 Film Dataset: Festival Programs
The Film Dataset consists a data scheme image file, a codebook and two dataset tables in csv format.
The codebook (csv file “1_codebook_film-dataset_festival-program”) offers a detailed description of all variables within the Film Dataset. Along with the definition of variables it lists explanations for the units of measurement, data sources, coding and information on missing data.
The csv file “1_film-dataset_festival-program_long” comprises a dataset of all films and the festivals, festival sections, and the year of the festival edition that they were sampled from. The dataset is structured in the long format, i.e. the same film can appear in several rows when it appeared in more than one sample festival. However, films are identifiable via their unique ID.
The csv file “1_film-dataset_festival-program_wide” consists of the dataset listing only unique films (n=9,348). The dataset is in the wide format, i.e. each row corresponds to a unique film, identifiable via its unique ID. For easy analysis, and since the overlap is only six percent, in this dataset the variable sample festival (fest) corresponds to the first sample festival where the film appeared. For instance, if a film was first shown at Berlinale (in February) and then at Frameline (in June of the same year), the sample festival will list “Berlinale”. This file includes information on unique and IMDb IDs, the film title, production year, length, categorization in length, production countries, regional attribution, director names, genre attribution, the festival, festival section and festival edition the film was sampled from, and information whether there is festival run information available through the IMDb data.
2 Survey Dataset
The Survey Dataset consists of a data scheme image file, a codebook and two dataset tables in csv format.
The codebook “2_codebook_survey-dataset” includes coding information for both survey datasets. It lists the definition of the variables or survey questions (corresponding to Samoilova/Loist 2019), units of measurement, data source, variable type, range and coding, and information on missing data.
The csv file “2_survey-dataset_long-festivals_shared-consent” consists of a subset (n=161) of the original survey dataset (n=454), where respondents provided festival run data for films (n=206) and gave consent to share their data for research purposes. This dataset consists of the festival data in a long format, so that each row corresponds to the festival appearance of a film.
The csv file “2_survey-dataset_wide-no-festivals_shared-consent” consists of a subset (n=372) of the original dataset (n=454) of survey responses corresponding to sample films. It includes data only for those films for which respondents provided consent to share their data for research purposes. This dataset is shown in wide format of the survey data, i.e. information for each response corresponding to a film is listed in one row. This includes data on film IDs, film title, survey questions regarding completeness and availability of provided information, information on number of festival screenings, screening fees, budgets, marketing costs, market screenings, and distribution. As the file name suggests, no data on festival screenings is included in the wide format dataset.
3 IMDb & Scripts
The IMDb dataset consists of a data scheme image file, one codebook and eight datasets, all in csv format. It also includes the R scripts that we used for scraping and matching.
The codebook “3_codebook_imdb-dataset” includes information for all IMDb datasets. This includes ID information and their data source, coding and value ranges, and information on missing data.
The csv file “3_imdb-dataset_aka-titles_long” contains film title data in different languages scraped from IMDb in a long format, i.e. each row corresponds to a title in a given language.
The csv file “3_imdb-dataset_awards_long” contains film award data in a long format, i.e. each row corresponds to an award of a given film.
The csv file “3_imdb-dataset_companies_long” contains data on production and distribution companies of films. The dataset is in a long format, so that each row corresponds to a particular company of a particular film.
The csv file “3_imdb-dataset_crew_long” contains data on names and roles of crew members in a long format, i.e. each row corresponds to each crew member. The file also contains binary gender assigned to directors based on their first names using the GenderizeR application.
The csv file “3_imdb-dataset_festival-runs_long” contains festival run data scraped from IMDb in a long format, i.e. each row corresponds to the festival appearance of a given film. The dataset does not include each film screening, but the first screening of a film at a festival within a given year. The data includes festival runs up to 2019.
The csv file “3_imdb-dataset_general-info_wide” contains general information about films such as genre as defined by IMDb, languages in which a film was shown, ratings, and budget. The dataset is in wide format, so that each row corresponds to a unique film.
The csv file “3_imdb-dataset_release-info_long” contains data about non-festival release (e.g., theatrical, digital, tv, dvd/blueray). The dataset is in a long format, so that each row corresponds to a particular release of a particular film.
The csv file “3_imdb-dataset_websites_long” contains data on available websites (official websites, miscellaneous, photos, video clips). The dataset is in a long format, so that each row corresponds to a website of a particular film.
The dataset includes 8 text files containing the script for webscraping. They were written using the R-3.6.3 version for Windows.
The R script “r_1_unite_data” demonstrates the structure of the dataset, that we use in the following steps to identify, scrape, and match the film data.
The R script “r_2_scrape_matches” reads in the dataset with the film characteristics described in the “r_1_unite_data” and uses various R packages to create a search URL for each film from the core dataset on the IMDb website. The script attempts to match each film from the core dataset to IMDb records by first conducting an advanced search based on the movie title and year, and then potentially using an alternative title and a basic search if no matches are found in the advanced search. The script scrapes the title, release year, directors, running time, genre, and IMDb film URL from the first page of the suggested records from the IMDb website. The script then defines a loop that matches (including matching scores) each film in the core dataset with suggested films on the IMDb search page. Matching was done using data on directors, production year (+/- one year), and title, a fuzzy matching approach with two methods: “cosine” and “osa.” where the cosine similarity is used to match titles with a high degree of similarity, and the OSA algorithm is used to match titles that may have typos or minor variations.
The script “r_3_matching” creates a dataset with the matches for a manual check. Each pair of films (original film from the core dataset and the suggested match from the IMDb website was categorized in the following five categories: a) 100% match: perfect match on title, year, and director; b) likely good match; c) maybe match; d) unlikely match; and e) no match). The script also checks for possible doubles in the dataset and identifies them for a manual check.
The script “r_4_scraping_functions” creates a function for scraping the data from the identified matches (based on the scripts described above and manually checked). These functions are used for scraping the data in the next script.
The script “r_5a_extracting_info_sample” uses the function defined in the “r_4_scraping_functions”, in order to scrape the IMDb data for the identified matches. This script does that for the first 100 films, to check, if everything works. Scraping for the entire dataset took a few hours. Therefore, a test with a subsample of 100 films is advisable.
The script “r_5b_extracting_info_all” extracts the data for the entire dataset of the identified matches.
The script “r_5c_extracting_info_skipped” checks the films with missing data (where data was not scraped) and tried to extract data one more time to make sure that the errors were not caused by disruptions in the internet connection or other technical issues.
The script “r_check_logs” is used for troubleshooting and tracking the progress of all of the R scripts used. It gives information on the amount of missing values and errors.
4 Festival Library Dataset
The Festival Library Dataset consists of a data scheme image file, one codebook and one dataset, all in csv format.
The codebook (csv file “4_codebook_festival-library_dataset”) offers a detailed description of all variables within the Library Dataset. It lists the definition of variables, such as location and festival name, and festival categories,
Timeseries data from 'BIG SUR R NR BIG SUR CA (USGS 11143000)' (gov_usgs_nwis_11143000)
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Supporting Information of 'forceX and forceR: a mobile setup and R package to measure and analyse a wide range of animal closing forces'
This dataset contains the Supporting Information of the publication
Rühr PT & Blanke A (2022): 'forceX and forceR: a mobile setup and R package to measure and analyse a wide range of animal closing forces'.
It includes
validation measurements the forceX setups (1 Ruehr Blanke 2022 validation measurements.zip)
all CAD files to build the forceX setup (3D-printed or metal-turned) (2 Ruehr Blanke 2022 forceX CAD files.zip)
forceX assembly instructions in HTML format, including schematics of custom electronics (3 Ruehr Blanke 2022 forceX Assembly instructions.html)
forceX assembly instructions as video (4 Ruehr Blanke 2022 forceX assembly video 03.mp4)
R code that produced all validation-related figures used in the original publication and that functions as a forceR v.1.0.13 example workflow (5 Ruehr Blanke 2022 forceR_workflow_example.R)
Python code to take videos of force measurements using the forceX camera module (6 Ruehr Blanke 2022 forceX_RPi_camera_code.py)
bundled version of forceR v.1.0.13 (forceR_1.0.13.tar.gz)
The CAD files and assembly instructions are also available on Thingiverse. The forceR package is available on CRAN (stable version) and GitHub (development version).
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Aytes, A.*, Mitrofanova, A.*, Lefebvre, C.*, Alvarez, M. J., Castillo-Martin, M., Zheng, T., Eastham, J. A., Gopalan, A., Pienta, K. J., Shen, M. M., Califano, A., and Abate-Shen, C. (2014). Cross-species analysis of genome-wide regulatory networks identifies a synergistic interaction between FOXM1 and CENPF that drives prostate cancer malignancy. Cancer Cell. *Equal contributions.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Additional file 1. R code of StepLMM. The data contains the R function of StepLMM, an example data and users’ guide.
Not seeing a result you expected?
Learn how you can add new datasets to our index.
This text files aims to explain the procedure to perform both the data handling and the analyses in the paper: Description of the files: 1. Datasets PhenoAsreml.txt contains the observed values for the phenotypes described in the paper PhenoAsreml_scaled.txt contains the scaled (mean of zero and standard deviation of one) of the same phenotypes. THIS DATASET IS USED FOR BIVARIATE ANALYSES WITH 2x2 STRATA PhenoAsreml_scaledLong.txt contains the same information, organized differently because the dataset is reshaped from wide to long format. PhenoAsreml_scaledLong_3x3strata.txt contains the information for BackFat and BodyWeight, THIS DATASET IS USED FOR BIVARIATE ANALYSES WITH 3x3 STRATA PhenoAsreml_scaledLong_3x2strata.txt contains the information for BackFat and Adiponectin, THIS DATASET IS USED FOR BIVARIATE ANALYSES WITH 3x2 STRATA. It is only an example of dataset that needs to be generated in order to obtain all the results presented in the paper BEDERE_2023_Data_PedigreeHens.txt is the pedigree file (individual/sire/dam) traced back over 5 generations 2. Codes and parameter files BEDERE_2023_RScript_handlingdata_Long.R is an R code to reshape PhenoAsreml_scaled into PhenoAsreml_scaledLong and to subset it to generate PhenoAsreml_scaledLong_3x3strata for instance BEDERE_2023_ASREMLScript_bivariate_2x2strata.as (as well as ...3x2strata.as and ...3x3strata.as) are ASReml parameter files used to state the data, model specification and post-hoc calculation to ASReml Please, note that some variance components have been fixed in some analyses when the algorithm was struggling to converge. BEDERE_2023_RScript_BartlettTest.R is an R code to perform the Bartlett test. 2. Results examples Some output files of ASReml are provided to give an example of results for each type of bivariate analysis. The .asr file is the log of the program, explaining how the program ran The .res file is describing the residuals The .pvc file describes the variance components and provides the genetic parameters with their associated standard errors.