Summary data for the studies used in the meta-analysis of local adaptation (Table 1 from the publication)This table contains the data used in this published meta-analysis. The data were originally extracted from the publications listed in the table. The file corresponds to Table 1 in the original publication.tb1.xlsSAS script used to perform meta-analysesThis file contains the essential elements of the SAS script used to perform meta-analyses published in Hoeksema & Forde 2008. Multi-factor models were fit to the data using weighted maximum likelihood estimation of parameters in a mixed model framework, using SAS PROC MIXED, in which the species traits and experimental design factors were considered fixed effects, and a random between-studies variance component was estimated. Significance (at alpha = 0.05) of individual factors in these models was determined using randomization procedures with 10,000 iterations (performed with a combination of macros in SAS), in which effect sizes a...
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
It is commonly believed that if a two-way analysis of variance (ANOVA) is carried out in R, then reported p-values are correct. This article shows that this is not always the case. Results can vary from non-significant to highly significant, depending on the choice of options. The user must know exactly which options result in correct p-values, and which options do not. Furthermore, it is commonly supposed that analyses in SAS and R of simple balanced experiments using mixed-effects models result in correct p-values. However, the simulation study of the current article indicates that frequency of Type I error deviates from the nominal value. The objective of this article is to compare SAS and R with respect to correctness of results when analyzing small experiments. It is concluded that modern functions and procedures for analysis of mixed-effects models are sometimes not as reliable as traditional ANOVA based on simple computations of sums of squares.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
SAS Code for Spatial Optimization of Supply Chain Network for Nitrogen Based Fertilizer in North America, by type, by mode of transportation, per county, for all major crops, using Proc OptModel. the code specifies set of random values to run the mixed integer stochastic spatial optimization model repeatedly and collect results for each simulation that are then compiled and exported to be projected in GIS (geographic information systems). Certain supply nodes (fertilizer plants) are specified to work at either 70 percent of their capacities or more. Capacities for nodes of supply (fertilizer plants), demand (county centroids), transhipment nodes (transfer points-mode may change), and actual distance travelled are specified over arcs.
This publication provides all the information required to understand the PISA 2003 educational performance database and perform analyses in accordance with the complex methodologies used to collect and process the data. It enables researchers to both reproduce the initial results and to undertake further analyses. The publication includes introductory chapters explaining the statistical theories and concepts required to analyse the PISA data, including full chapters on how to apply replicate weights and undertake analyses using plausible values; worked examples providing full syntax in SAS®; and a comprehensive description of the OECD PISA 2003 international database. The PISA 2003 database includes micro-level data on student educational performance for 41 countries collected in 2003, together with students’ responses to the PISA 2003 questionnaires and the test questions. A similar manual is available for SPSS users.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Example of the code used to account for statistical significances for phenotype and other variables.
CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
License information was derived automatically
Sequential designs and competing risks methodology are both well established. Their combined use has recently received some attention from a theoretical perspective, but their joint application in practice has been discussed less. The aim of this paper is to provide the applied statistician with a basic understanding of both sequential design theory and competing risks methodology and how to combine them in practice. Relevant references to more detailed theoretical discussions are provided and all discussions are illustrated using a real case study. Extensive R and SAS code is provided in the online supplementary material.
Emerson Process Management Sas Company Export Import Records. Follow the Eximpedia platform for HS code, importer-exporter records, and customs shipment details.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
SAS code to reproduce the simulation study and the analysis of the urine osmolarity example. (ZIP)
https://search.gesis.org/research_data/datasearch-httpwww-da-ra-deoaip--oaioai-da-ra-de455875https://search.gesis.org/research_data/datasearch-httpwww-da-ra-deoaip--oaioai-da-ra-de455875
Abstract (en): The purpose of this data collection is to provide an official public record of the business of the federal courts. The data originate from district and appellate court offices throughout the United States. Information was obtained at two points in the life of a case: filing and termination. The termination data contain information on both filing and terminations, while the pending data contain only filing information. ICPSR data undergo a confidentiality review and are altered when necessary to limit the risk of disclosure. ICPSR also routinely creates ready-to-go data files along with setups in the major statistical software formats as well as standard codebooks to accompany the data. In addition to these procedures, ICPSR performed the following processing steps for this data collection: Performed consistency checks.; Standardized missing values.; Checked for undocumented or out-of-range codes.. All federal court cases in the United States in 2002. Smallest Geographic Unit: county 2015-09-18 Six data files were created with docket numbers blanked for Parts 1, 3, and 5, and with docket numbers containing original values for Parts 2, 4, and 6.2012-06-26 All parts are being moved to restricted access and will be available only using the restricted access procedures.2005-04-29 Data files for Part 3, Criminal Data, 2002, Part 4, Civil Pending Data, 2002, and a Civil Pending Restricted Data, 2002 file have been added to the data collection along with corresponding SAS and SPSS setup files and codebooks in PDF formats.2005-01-07 A restricted data file for Part 1, Civil Terminations, 2002, has been added to the data collection. The public use data file for Part 1 and its corresponding SAS and SPSS setup files have been updated. The codebook has been modified to reflect these changes. Funding insitution(s): United States Department of Justice. Office of Justice Programs. Bureau of Justice Statistics. record abstractsStarting with the year 2001, each year of data for Federal Court Cases is released by ICPSR as a separate study number. Federal Court Cases data for the years 1970-2000 can be found in FEDERAL COURT CASES: INTEGRATED DATA BASE, 1970-2000 (ICPSR 8429).
We compiled macroinvertebrate assemblage data collected from 1995 to 2014 from the St. Louis River Area of Concern (AOC) of western Lake Superior. Our objective was to define depth-adjusted cutoff values for benthos condition classes (poor, fair, reference) to provide tool useful for assessing progress toward achieving removal targets for the degraded benthos beneficial use impairment in the AOC. The relationship between depth and benthos metrics was wedge-shaped. We therefore used quantile regression to model the limiting effect of depth on selected benthos metrics, including taxa richness, percent non-oligochaete individuals, combined percent Ephemeroptera, Trichoptera, and Odonata individuals, and density of ephemerid mayfly nymphs (Hexagenia). We created a scaled trimetric index from the first three metrics. Metric values at or above the 90th percentile quantile regression model prediction were defined as reference condition for that depth. We set the cutoff between poor and fair condition as the 50th percentile model prediction. We examined sampler type, exposure, geographic zone of the AOC, and substrate type for confounding effects. Based on these analyses we combined data across sampler type and exposure classes and created separate models for each geographic zone. We used the resulting condition class cutoff values to assess the relative benthic condition for three habitat restoration project areas. The depth-limited pattern of ephemerid abundance we observed in the St. Louis River AOC also occurred elsewhere in the Great Lakes. We provide tabulated model predictions for application of our depth-adjusted condition class cutoff values to new sample data. This dataset is associated with the following publication: Angradi, T., W. Bartsch, A. Trebitz, V. Brady, and J. Launspach. A depth-adjusted ambient distribution approach for setting numeric removal targets for a Great Lakes Area of Concern beneficial use impairment: Degraded benthos. JOURNAL OF GREAT LAKES RESEARCH. International Association for Great Lakes Research, Ann Arbor, MI, USA, 43(1): 108-120, (2017).
The simulated synthetic aperture sonar (SAS) data presented here was generated using PoSSM [Johnson and Brown 2018]. The data is suitable for bistatic, coherent signal processing and will form acoustic seafloor imagery. Included in this data package is simulated sonar data in Generic Data Format (GDF) files, a description of the GDF file contents, example SAS imagery, and supporting information about the simulated scenes. In total, there are eleven 60 m x 90 m scenes, labeled scene00 through scene10, with scene00 provided with the scatterers in isolation, i.e. no seafloor texture. This is provided for beamformer testing purposes and should result in an image similar to the one labeled "PoSSM-scene00-scene00-starboard-0.tif" in the Related Data Sets tab. The ten other scenes have varying degrees of model variation as described in "Description_of_Simulated_SAS_Data_Package.pdf". A description of the data and the model is found in the associated document called "Description_of_Simulated_SAS_Data_Package.pdf" and a description of the format in which the raw binary data is stored is found in the related document "PSU_GDF_Format_20240612.pdf". The format description also includes MATLAB code that will effectively parse the data to aid in signal processing and image reconstruction. It is left to the researcher to develop a beamforming algorithm suitable for coherent signal and image processing. Each 60 m x 90 m scene is represented by 4 raw (not beamformed) GDF files, labeled sceneXX-STARBOARD-000000 through 000003. It is possible to beamform smaller scenes from any one of these 4 files, i.e. the four files are combined sequentially to form a 60 m x 90 m image. Also included are comma separated value spreadsheets describing the locations of scatterers and objects of interest within each scene. In addition to the binary GDF data, a beamformed GeoTIFF image and a single-look complex (SLC, science file) data of each scene is provided. The SLC data (science) is stored in the Hierarchical Data Format 5 (https://www.hdfgroup.org/), and appended with ".hdf5" to indicate the HDF5 format. The data are stored as 32-bit real and 32-bit complex values. A viewer is available that provides basic graphing, image display, and directory navigation functions (https://www.hdfgroup.org/downloads/hdfview/). The HDF file contains all the information necessary to reconstruct a synthetic aperture sonar image. All major and contemporary programming languages have library support for encoding/decoding the HDF5 format. Supporting documentation that outlines positions of the seafloor scatterers is included in "Scatterer_Locations_Scene00.csv", while the locations of the objects of interest for scene01-scene10 are included in "Object_Locations_All_Scenes.csv". Portable Network Graphic (PNG) images that plot the location of objects of all the objects of interest in each scene in Along-Track and Cross-Track notation are provided.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
dfgdfgdsfgfsdfsdfg
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
SAS PROC used to evaluate SSMT data
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
This paper aims to quantify the uncertainty in the SAS-SST simulation of a prism bluff-body flow due to varying the higher-wavenumber damping factor (). Instead of performing the uncertainty quantification on the CFD simulation directly, a surrogate modelling approach is adopted. The mesh sensitivity is first studied and the numerical error due to the mesh is approximated accordingly. The Gaussian processes/Kriging method is used to generate surrogate models for quantities of interest (QoIs). The suitability of the surrogate models is assessed using the leave-one-out cross-validation tests (LOO-CV). The stochastic tests are then performed using the cross-validated surrogate models to quantify the uncertainty of QoIs by varying Cs. Four prior probability density functions (such as U(0,1), N(0.5, 0.1^2), Beta (2,2) and Beta (5,1.5)) of Cs are considered.
It is demonstrated in this study that the uncertainty of a predicted QoI due to varying Cs is regionally dependent. The flow statistics in the near wake of the prism body are subject to larger variance due to the uncertainty in Cs. The influence of Cs rapidly decays as the location moves downstream. The response of different QoIs to the changing Cs varies greatly. Therefore, the calibration of Cs only using observations of one variable may bias the results. Last but not least, it is important to consider different sources of uncertainties within the numerical model when scrutinising a turbulence model, as ignoring the contributions to the total error may lead to biased conclusions.
Multienvironment trials (METs) enable the evaluation of the same genotypes under a v ariety of environments and management conditions. We present META (Multi Environment Trial Analysis), a suite of 31 SAS programs that analyze METs with complete or incomplete block designs, with or without adjustment by a covariate. The entire program is run through a graphical user interface. The program can produce boxplots or histograms for all traits, as well as univariate statistics. It also calculates best linear unbiased estimators (BLUEs) and best linear unbiased predictors for the main response variable and BLUEs for all other traits. For all traits, it calculates variance components by restricted maximum likelihood, least significant difference, coefficient of variation, and broad-sense heritability using PROC MIXED. The program can analyze each location separately, combine the analysis by management conditions, or combine all locations. The flexibility and simplicity of use of this program makes it a valuable tool for analyzing METs in breeding and agronomy. The META program can be used by any researcher who knows only a few fundamental principles of SAS.
CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
License information was derived automatically
Users are able to access data related discharge information on all emergency department visits. Data is focused on but not limited to emergency room diagnoses, procedures, demographics, and payment source. Background The State Emergency Department Databases (SEDD) is focused on capturing discharge information on all emergency department visits that do not result in an admission, (Information on patients initially seen in the emergency room and then admitted to the hospital is included in the State Inpatient Databases (SID)). The SEDD contains emergency department information from 27 states. The SEDD contain more than 100 clinical and non-clinical variables included in a hospital dis charge abstract, such as: diagnoses, procedures, patient demographics, expected payment source and total charges. User functionality Users must pay to access the SEDD database. SEDD files from 1999-2009 are available through the HCUP Central Distributor. The SEDD data set can be run on desktop computers with a CD-ROM reader, and comes in ASCII format. The data on the CD set require a statistical software package such as SAS or SPSS to use for analytic purposes. The data set comes with full documentation. SAS and SPSS users are provided programs for converting ASCII files. Data Notes Data is available from 1999-2009. The website does not indicate when new data will be updated. Twenty-seven States now currently participate in the SEDD including Arizona, California, Connecticut, Florida, Georgia, Hawaii, Indiana, Iowa, Kansas, Maine, Maryland, Massachusetts, Minnesota, Missouri, Nebraska, New Hampshire, New Jersey, New York, North Carolina, Ohio, Rhode Island, South Carolina, South Dakota, Tennessee, Utah, Vermont, and Wisconsin.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
FunCoup network information for gene sas in Drosophila melanogaster. SAS_DROME Putative epidermal cell surface receptor
Novasep Process Solutions Sas Company Export Import Records. Follow the Eximpedia platform for HS code, importer-exporter records, and customs shipment details.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
FunCoup network information for gene SAS in Homo sapiens. SIAS_HUMAN Sialic acid synthase
no abstract provided