100+ datasets found
  1. e

    Geodatabase for the Baltimore Ecosystem Study Spatial Data

    • portal.edirepository.org
    • search.dataone.org
    application/vnd.rar
    Updated May 4, 2012
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Jarlath O'Neal-Dunne; Morgan Grove (2012). Geodatabase for the Baltimore Ecosystem Study Spatial Data [Dataset]. http://doi.org/10.6073/pasta/377da686246f06554f7e517de596cd2b
    Explore at:
    application/vnd.rar(29574980 kilobyte)Available download formats
    Dataset updated
    May 4, 2012
    Dataset provided by
    EDI
    Authors
    Jarlath O'Neal-Dunne; Morgan Grove
    Time period covered
    Jan 1, 1999 - Jun 1, 2014
    Area covered
    Description

    The establishment of a BES Multi-User Geodatabase (BES-MUG) allows for the storage, management, and distribution of geospatial data associated with the Baltimore Ecosystem Study. At present, BES data is distributed over the internet via the BES website. While having geospatial data available for download is a vast improvement over having the data housed at individual research institutions, it still suffers from some limitations. BES-MUG overcomes these limitations; improving the quality of the geospatial data available to BES researches, thereby leading to more informed decision-making.

       BES-MUG builds on Environmental Systems Research Institute's (ESRI) ArcGIS and ArcSDE technology. ESRI was selected because its geospatial software offers robust capabilities. ArcGIS is implemented agency-wide within the USDA and is the predominant geospatial software package used by collaborating institutions.
    
    
       Commercially available enterprise database packages (DB2, Oracle, SQL) provide an efficient means to store, manage, and share large datasets. However, standard database capabilities are limited with respect to geographic datasets because they lack the ability to deal with complex spatial relationships. By using ESRI's ArcSDE (Spatial Database Engine) in conjunction with database software, geospatial data can be handled much more effectively through the implementation of the Geodatabase model. Through ArcSDE and the Geodatabase model the database's capabilities are expanded, allowing for multiuser editing, intelligent feature types, and the establishment of rules and relationships. ArcSDE also allows users to connect to the database using ArcGIS software without being burdened by the intricacies of the database itself.
    
    
       For an example of how BES-MUG will help improve the quality and timeless of BES geospatial data consider a census block group layer that is in need of updating. Rather than the researcher downloading the dataset, editing it, and resubmitting to through ORS, access rules will allow the authorized user to edit the dataset over the network. Established rules will ensure that the attribute and topological integrity is maintained, so that key fields are not left blank and that the block group boundaries stay within tract boundaries. Metadata will automatically be updated showing who edited the dataset and when they did in the event any questions arise.
    
    
       Currently, a functioning prototype Multi-User Database has been developed for BES at the University of Vermont Spatial Analysis Lab, using Arc SDE and IBM's DB2 Enterprise Database as a back end architecture. This database, which is currently only accessible to those on the UVM campus network, will shortly be migrated to a Linux server where it will be accessible for database connections over the Internet. Passwords can then be handed out to all interested researchers on the project, who will be able to make a database connection through the Geographic Information Systems software interface on their desktop computer. 
    
    
       This database will include a very large number of thematic layers. Those layers are currently divided into biophysical, socio-economic and imagery categories. Biophysical includes data on topography, soils, forest cover, habitat areas, hydrology and toxics. Socio-economics includes political and administrative boundaries, transportation and infrastructure networks, property data, census data, household survey data, parks, protected areas, land use/land cover, zoning, public health and historic land use change. Imagery includes a variety of aerial and satellite imagery.
    
    
       See the readme: http://96.56.36.108/geodatabase_SAL/readme.txt
    
    
       See the file listing: http://96.56.36.108/geodatabase_SAL/diroutput.txt
    
  2. s

    Spatial Multimodal Analysis (SMA) - Spatial Transcriptomics

    • figshare.scilifelab.se
    • researchdata.se
    json
    Updated Jan 15, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Marco Vicari; Reza Mirzazadeh; Anna Nilsson; Patrik Bjärterot; Ludvig Larsson; Hower Lee; Mats Nilsson; Julia Foyer; Markus Ekvall; Paulo Czarnewski; Xiaoqun Zhang; Per Svenningsson; Per Andrén; Lukas Käll; Joakim Lundeberg (2025). Spatial Multimodal Analysis (SMA) - Spatial Transcriptomics [Dataset]. http://doi.org/10.17044/scilifelab.22778920.v1
    Explore at:
    jsonAvailable download formats
    Dataset updated
    Jan 15, 2025
    Dataset provided by
    KTH Royal Institute of Technology, Science for Life Laboratory
    Authors
    Marco Vicari; Reza Mirzazadeh; Anna Nilsson; Patrik Bjärterot; Ludvig Larsson; Hower Lee; Mats Nilsson; Julia Foyer; Markus Ekvall; Paulo Czarnewski; Xiaoqun Zhang; Per Svenningsson; Per Andrén; Lukas Käll; Joakim Lundeberg
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    This dataset contains Spatial Transcriptomics (ST) data matching with Matrix Assisted Laser Desorption/Ionization - Mass Spetrometry Imaging (MALDI-MSI). This data is complementary to data contained in the same project. FIles with the same identifiers in the two datasets originated from the very same tissue section and can be combined in a multimodal ST-MSI object. For more information about the dataset please see our manuscript posted on BioRxiv (doi: https://doi.org/10.1101/2023.01.26.525195). This dataset includes ST data from 19 tissue sections, including human post-mortem and mouse samples. The spatial transcriptomics data was generated using the Visium protocol (10x Genomics). The murine tissue sections come from three different mice unilaterally injected with 6-OHDA. 6-OHDA is a neurotoxin that when injected in the brain can selectively destroy dopaminergic neurons. We used this mouse model to show the applicability of the technology that we developed, named Spatial Multimodal Analysis (SMA). Using our technology on these mouse brain tissue sections we were able to detect both dopamine with MALDI-MSI and the corresponding gene expression with ST. This dataset includes also one human post-mortem striatum sample that was placed on one Visium slide across the four capture areas. This sample was analyzed with a different ST protocol named RRST (Mirzazadeh, R., Andrusivova, Z., Larsson, L. et al. Spatially resolved transcriptomic profiling of degraded and challenging fresh frozen samples. Nat Commun 14, 509 (2023). https://doi.org/10.1038/s41467-023-36071-5), where probes capturing the whole transcriptome are first hybridized in the tissue section and then spatially detected. Each tissue section contained in the dataset has been given a unique identifier that is composed of the Visium array ID and capture area ID of the Visium slide that the tissue section was placed on. This unique identifier is included in the file names of all the files relative to the same tissue section, including the MALDI-MSI files published in the other dataset included in this project. In this dataset you will find the following files for each tissue section: - raw files: these are the read one fastq files (containing the pattern *R1*fastq.gz in the file name), read two fastq files (containing the pattern *R1*fastq.gz in the file name) and the raw microscope images (containing the pattern Spot.jpg in the file name). These are the only files needed to run the Space Ranger pipeline, which is freely available for any user (please see the 10x Genomics website for information on how to install and run Space Ranger); - processed data files: we provide processed data files of two types: a) Space Ranger outputs that were used to produce the figures in our publication; b) manual annotation tables in csv format produced using Loupe Browser 6 (csv tables with file names ending _RegionLoupe.csv, _filter.csv, _dopamine.csv, _lesion.csv, _region.csv patterns); c) json files that we used as input for Space Ranger in the cases where the automatic tissue detection included in the pipeline failed to recognize the tissue or the fiducials. Using these processed files the user can reproduce the figures of our publication without having to restart from the raw data files. The MALDI-MSI analyses preceding ST was performed with different matrices in different tissue section. We used 1) 9-aminoacridine (9-AA) for detection of metabolites in negative ionization mode, 2) 2,5-dihydroxybenzoic acid (DHB) for detection of metabolites in positive ionization mode, 3) 4-(anthracen-9-yl)-2-fluoro-1-ethylpyridin-1-ium iodide (FMP-10), which charge-tags molecules with phenolic hydroxyls and/or primary amines, including neurotransmitters. The information about which matrix was sprayed on the tissue sections and other information about the samples is included in the metadata table. We also used three types of control samples: - standard Visium: samples processed with standard Visium (i.e. no matrix spraying, no MALDI-MSI, protocol as recommended by 10x Gemomics with no exeptions) - internal controls (iCTRL): samples not sprayed with any matrix, neither processed with MALDI-MSI, but located on the same Visium slide were other samples were processed with MALDI-MSI - FMP-10-iCTRL: sample sprayed with FMP-10, and then processed as an iCTRL. This and other information is provided in the metadata table.

  3. f

    IMCOMA-example-datasets

    • figshare.com
    xml
    Updated Feb 12, 2021
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Nowosad (2021). IMCOMA-example-datasets [Dataset]. http://doi.org/10.6084/m9.figshare.13379228.v1
    Explore at:
    xmlAvailable download formats
    Dataset updated
    Feb 12, 2021
    Dataset provided by
    figshare
    Authors
    Nowosad
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Datasets- simple_land_cover1.tif - an example land cover dataset presented in Figures 1 and 2- simple_landform1.tif - an example landform dataset presented in Figures 1 and 2- landcover_europe.tif - a land cover dataset with nine categories for Europe - landcover_europe.qml - a QGIS color style for the landcover_europe.tif dataset- landform_europe.tif - a landform dataset with 17 categories for Europe - landform_europe.qml - a QGIS color style for the landform_europe.tif dataset- map1.gpkg - a map of LTs in Europe constructed using the INCOMA-based method- map1.qml - a QGIS color style for the map1.gpkg dataset- map2.gpkg - a map of LTs in Europe constructed using the COMA method to identify and delineate pattern types in each theme separately- map2.qml - a QGIS color style for the map2.gpkg dataset- map3.gpkg - a map of LTs in Europe constructed using the map overlay method- map3.qml - a QGIS color style for the map3.gpkg dataset

  4. m

    GeoStoryTelling

    • data.mendeley.com
    Updated Apr 21, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Manuel Gonzalez Canche (2023). GeoStoryTelling [Dataset]. http://doi.org/10.17632/nh2c5t3vf9.1
    Explore at:
    Dataset updated
    Apr 21, 2023
    Authors
    Manuel Gonzalez Canche
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Database created for replication of GeoStoryTelling. Our life stories evolve in specific and contextualized places. Although our homes may be our primarily shaping environment, our homes are themselves situated in neighborhoods that expose us to the immediate “real world” outside home. Indeed, the places where we are currently experiencing, and have experienced life, play a fundamental role in gaining a deeper and more nuanced understanding of our beliefs, fears, perceptions of the world, and even our prospects of social mobility. Despite the immediate impact of the places where we experience life in reaching a better understanding of our life stories, to date most qualitative and mixed methods researchers forego the analytic and elucidating power that geo-contextualizing our narratives bring to social and health research. From this view then, most research findings and conclusions may have been ignoring the spatial contexts that most likely have shaped the experiences of research participants. The main reason for the underuse of these geo-contextualized stories is the requirement of specialized training in geographical information systems and/or computer and statistical programming along with the absence of cost-free and user-friendly geo-visualization tools that may allow non-GIS experts to benefit from geo-contextualized outputs. To address this gap, we present GeoStoryTelling, an analytic framework and user-friendly, cost-free, multi-platform software that enables researchers to visualize their geo-contextualized data narratives. The use of this software (available in Mac and Windows operative systems) does not require users to learn GIS nor computer programming to obtain state-of-the-art, and visually appealing maps. In addition to providing a toy database to fully replicate the outputs presented, we detail the process that researchers need to follow to build their own databases without the need of specialized external software nor hardware. We show how the resulting HTML outputs are capable of integrating a variety of multi-media inputs (i.e., text, image, videos, sound recordings/music, and hyperlinks to other websites) to provide further context to the geo-located stories we are sharing (example https://cutt.ly/k7X9tfN). Accordingly, the goals of this paper are to describe the components of the methodology, the steps to construct the database, and to provide unrestricted access to the software tool, along with a toy dataset so that researchers may interact first-hand with GeoStoryTelling and fully replicate the outputs discussed herein. Since GeoStoryTelling relied on OpenStreetMap its applications may be used worldwide, thus strengthening its potential reach to the mixed methods and qualitative scientific communities, regardless of location around the world. Keywords: Geographical Information Systems; Interactive Visualizations; Data StoryTelling; Mixed Methods & Qualitative Research Methodologies; Spatial Data Science; Geo-Computation.

  5. H

    Replication Data for the Turnout Example in Chapter 6 of Spatial Analysis...

    • dataverse.harvard.edu
    Updated Jun 28, 2015
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    David Darmofal (2015). Replication Data for the Turnout Example in Chapter 6 of Spatial Analysis for the Social Sciences [Dataset]. http://doi.org/10.7910/DVN/NGVKUS
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Jun 28, 2015
    Dataset provided by
    Harvard Dataverse
    Authors
    David Darmofal
    License

    CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
    License information was derived automatically

    Description

    Replication data for the turnout example in Chapter 6 of Spatial Analysis for the Social Sciences.

  6. f

    fdata-02-00044_Parallel Processing Strategies for Big Geospatial Data.pdf

    • frontiersin.figshare.com
    pdf
    Updated Jun 3, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Martin Werner (2023). fdata-02-00044_Parallel Processing Strategies for Big Geospatial Data.pdf [Dataset]. http://doi.org/10.3389/fdata.2019.00044.s001
    Explore at:
    pdfAvailable download formats
    Dataset updated
    Jun 3, 2023
    Dataset provided by
    Frontiers
    Authors
    Martin Werner
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    This paper provides an abstract analysis of parallel processing strategies for spatial and spatio-temporal data. It isolates aspects such as data locality and computational locality as well as redundancy and locally sequential access as central elements of parallel algorithm design for spatial data. Furthermore, the paper gives some examples from simple and advanced GIS and spatial data analysis highlighting both that big data systems have been around long before the current hype of big data and that they follow some design principles which are inevitable for spatial data including distributed data structures and messaging, which are, however, incompatible with the popular MapReduce paradigm. Throughout this discussion, the need for a replacement or extension of the MapReduce paradigm for spatial data is derived. This paradigm should be able to deal with the imperfect data locality inherent to spatial data hindering full independence of non-trivial computational tasks. We conclude that more research is needed and that spatial big data systems should pick up more concepts like graphs, shortest paths, raster data, events, and streams at the same time instead of solving exactly the set of spatially separable problems such as line simplifications or range queries in manydifferent ways.

  7. a

    GEOSPATIAL DATA Progress Needed on Identifying Expenditures, Building and...

    • arc-gis-hub-home-arcgishub.hub.arcgis.com
    • hub.arcgis.com
    Updated Jun 11, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    GeoPlatform ArcGIS Online (2024). GEOSPATIAL DATA Progress Needed on Identifying Expenditures, Building and Utilizing a Data Infrastructure, and Reducing Duplicative Efforts [Dataset]. https://arc-gis-hub-home-arcgishub.hub.arcgis.com/documents/c0cef9e4901143cbb9f15ddbb49ca3b4
    Explore at:
    Dataset updated
    Jun 11, 2024
    Dataset authored and provided by
    GeoPlatform ArcGIS Online
    Description

    Progress Needed on Identifying Expenditures, Building and Utilizing a Data Infrastructure, and Reducing Duplicative Efforts The federal government collects, maintains, and uses geospatial information—data linked to specific geographic locations—to help support varied missions, including national security and natural resources conservation. To coordinate geospatial activities, in 1994 the President issued an executive order to develop a National Spatial Data Infrastructure—a framework for coordination that includes standards, data themes, and a clearinghouse. GAO was asked to review federal and state coordination of geospatial data. GAO’s objectives were to (1) describe the geospatial data that selected federal agencies and states use and how much is spent on geospatial data; (2) assess progress in establishing the National Spatial Data Infrastructure; and (3) determine whether selected federal agencies and states invest in duplicative geospatial data. To do so, GAO identified federal and state uses of geospatial data; evaluated available cost data from 2013 to 2015; assessed FGDC’s and selected agencies’ efforts to establish the infrastructure; and analyzed federal and state datasets to identify duplication. What GAO Found Federal agencies and state governments use a variety of geospatial datasets to support their missions. For example, after Hurricane Sandy in 2012, the Federal Emergency Management Agency used geospatial data to identify 44,000 households that were damaged and inaccessible and reported that, as a result, it was able to provide expedited assistance to area residents. Federal agencies report spending billions of dollars on geospatial investments; however, the estimates are understated because agencies do not always track geospatial investments. For example, these estimates do not include billions of dollars spent on earth-observing satellites that produce volumes of geospatial data. The Federal Geographic Data Committee (FGDC) and the Office of Management and Budget (OMB) have started an initiative to have agencies identify and report annually on geospatial-related investments as part of the fiscal year 2017 budget process. FGDC and selected federal agencies have made progress in implementing their responsibilities for the National Spatial Data Infrastructure as outlined in OMB guidance; however, critical items remain incomplete. For example, the committee established a clearinghouse for records on geospatial data, but the clearinghouse lacks an effective search capability and performance monitoring. FGDC also initiated plans and activities for coordinating with state governments on the collection of geospatial data; however, state officials GAO contacted are generally not satisfied with the committee’s efforts to coordinate with them. Among other reasons, they feel that the committee is focused on a federal perspective rather than a national one, and that state recommendations are often ignored. In addition, selected agencies have made limited progress in their own strategic planning efforts and in using the clearinghouse to register their data to ensure they do not invest in duplicative data. For example, 8 of the committee’s 32 member agencies have begun to register their data on the clearinghouse, and they have registered 59 percent of the geospatial data they deemed critical. Part of the reason that agencies are not fulfilling their responsibilities is that OMB has not made it a priority to oversee these efforts. Until OMB ensures that FGDC and federal agencies fully implement their responsibilities, the vision of improving the coordination of geospatial information and reducing duplicative investments will not be fully realized. OMB guidance calls for agencies to eliminate duplication, avoid redundant expenditures, and improve the efficiency and effectiveness of the sharing and dissemination of geospatial data. However, some data are collected multiple times by federal, state, and local entities, resulting in duplication in effort and resources. A new initiative to create a national address database could potentially result in significant savings for federal, state, and local governments. However, agencies face challenges in effectively coordinating address data collection efforts, including statutory restrictions on sharing certain federal address data. Until there is effective coordination across the National Spatial Data Infrastructure, there will continue to be duplicative efforts to obtain and maintain these data at every level of government.https://www.gao.gov/assets/d15193.pdfWhat GAO Recommends GAO suggests that Congress consider assessing statutory limitations on address data to foster progress toward a national address database. GAO also recommends that OMB improve its oversight of FGDC and federal agency initiatives, and that FGDC and selected agencies fully implement initiatives. The agencies generally agreed with the recommendations and identified plans to implement them.

  8. n

    Data from: A new digital method of data collection for spatial point pattern...

    • data.niaid.nih.gov
    • zenodo.org
    • +1more
    zip
    Updated Jul 6, 2021
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Chao Jiang; Xinting Wang (2021). A new digital method of data collection for spatial point pattern analysis in grassland communities [Dataset]. http://doi.org/10.5061/dryad.brv15dv70
    Explore at:
    zipAvailable download formats
    Dataset updated
    Jul 6, 2021
    Dataset provided by
    Chinese Academy of Agricultural Sciences
    Inner Mongolia University of Technology
    Authors
    Chao Jiang; Xinting Wang
    License

    https://spdx.org/licenses/CC0-1.0.htmlhttps://spdx.org/licenses/CC0-1.0.html

    Description

    A major objective of plant ecology research is to determine the underlying processes responsible for the observed spatial distribution patterns of plant species. Plants can be approximated as points in space for this purpose, and thus, spatial point pattern analysis has become increasingly popular in ecological research. The basic piece of data for point pattern analysis is a point location of an ecological object in some study region. Therefore, point pattern analysis can only be performed if data can be collected. However, due to the lack of a convenient sampling method, a few previous studies have used point pattern analysis to examine the spatial patterns of grassland species. This is unfortunate because being able to explore point patterns in grassland systems has widespread implications for population dynamics, community-level patterns and ecological processes. In this study, we develop a new method to measure individual coordinates of species in grassland communities. This method records plant growing positions via digital picture samples that have been sub-blocked within a geographical information system (GIS). Here, we tested out the new method by measuring the individual coordinates of Stipa grandis in grazed and ungrazed S. grandis communities in a temperate steppe ecosystem in China. Furthermore, we analyzed the pattern of S. grandis by using the pair correlation function g(r) with both a homogeneous Poisson process and a heterogeneous Poisson process. Our results showed that individuals of S. grandis were overdispersed according to the homogeneous Poisson process at 0-0.16 m in the ungrazed community, while they were clustered at 0.19 m according to the homogeneous and heterogeneous Poisson processes in the grazed community. These results suggest that competitive interactions dominated the ungrazed community, while facilitative interactions dominated the grazed community. In sum, we successfully executed a new sampling method, using digital photography and a Geographical Information System, to collect experimental data on the spatial point patterns for the populations in this grassland community.

    Methods 1. Data collection using digital photographs and GIS

    A flat 5 m x 5 m sampling block was chosen in a study grassland community and divided with bamboo chopsticks into 100 sub-blocks of 50 cm x 50 cm (Fig. 1). A digital camera was then mounted to a telescoping stake and positioned in the center of each sub-block to photograph vegetation within a 0.25 m2 area. Pictures were taken 1.75 m above the ground at an approximate downward angle of 90° (Fig. 2). Automatic camera settings were used for focus, lighting and shutter speed. After photographing the plot as a whole, photographs were taken of each individual plant in each sub-block. In order to identify each individual plant from the digital images, each plant was uniquely marked before the pictures were taken (Fig. 2 B).

    Digital images were imported into a computer as JPEG files, and the position of each plant in the pictures was determined using GIS. This involved four steps: 1) A reference frame (Fig. 3) was established using R2V software to designate control points, or the four vertexes of each sub-block (Appendix S1), so that all plants in each sub-block were within the same reference frame. The parallax and optical distortion in the raster images was then geometrically corrected based on these selected control points; 2) Maps, or layers in GIS terminology, were set up for each species as PROJECT files (Appendix S2), and all individuals in each sub-block were digitized using R2V software (Appendix S3). For accuracy, the digitization of plant individual locations was performed manually; 3) Each plant species layer was exported from a PROJECT file to a SHAPE file in R2V software (Appendix S4); 4) Finally each species layer was opened in Arc GIS software in the SHAPE file format, and attribute data from each species layer was exported into Arc GIS to obtain the precise coordinates for each species. This last phase involved four steps of its own, from adding the data (Appendix S5), to opening the attribute table (Appendix S6), to adding new x and y coordinate fields (Appendix S7) and to obtaining the x and y coordinates and filling in the new fields (Appendix S8).

    1. Data reliability assessment

    To determine the accuracy of our new method, we measured the individual locations of Leymus chinensis, a perennial rhizome grass, in representative community blocks 5 m x 5 m in size in typical steppe habitat in the Inner Mongolia Autonomous Region of China in July 2010 (Fig. 4 A). As our standard for comparison, we used a ruler to measure the individual coordinates of L. chinensis. We tested for significant differences between (1) the coordinates of L. chinensis, as measured with our new method and with the ruler, and (2) the pair correlation function g of L. chinensis, as measured with our new method and with the ruler (see section 3.2 Data Analysis). If (1) the coordinates of L. chinensis, as measured with our new method and with the ruler, and (2) the pair correlation function g of L. chinensis, as measured with our new method and with the ruler, did not differ significantly, then we could conclude that our new method of measuring the coordinates of L. chinensis was reliable.

    We compared the results using a t-test (Table 1). We found no significant differences in either (1) the coordinates of L. chinensis or (2) the pair correlation function g of L. chinensis. Further, we compared the pattern characteristics of L. chinensis when measured by our new method against the ruler measurements using a null model. We found that the two pattern characteristics of L. chinensis did not differ significantly based on the homogenous Poisson process or complete spatial randomness (Fig. 4 B). Thus, we concluded that the data obtained using our new method was reliable enough to perform point pattern analysis with a null model in grassland communities.

  9. Codes in R for spatial statistics analysis, ecological response models and...

    • zenodo.org
    bin
    Updated Apr 24, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    D. W. Rössel-Ramírez; D. W. Rössel-Ramírez; J. Palacio-Núñez; J. Palacio-Núñez; S. Espinosa; S. Espinosa; J. F. Martínez-Montoya; J. F. Martínez-Montoya (2025). Codes in R for spatial statistics analysis, ecological response models and spatial distribution models [Dataset]. http://doi.org/10.5281/zenodo.7603557
    Explore at:
    binAvailable download formats
    Dataset updated
    Apr 24, 2025
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    D. W. Rössel-Ramírez; D. W. Rössel-Ramírez; J. Palacio-Núñez; J. Palacio-Núñez; S. Espinosa; S. Espinosa; J. F. Martínez-Montoya; J. F. Martínez-Montoya
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    In the last decade, a plethora of algorithms have been developed for spatial ecology studies. In our case, we use some of these codes for underwater research work in applied ecology analysis of threatened endemic fishes and their natural habitat. For this, we developed codes in Rstudio® script environment to run spatial and statistical analyses for ecological response and spatial distribution models (e.g., Hijmans & Elith, 2017; Den Burg et al., 2020). The employed R packages are as follows: caret (Kuhn et al., 2020), corrplot (Wei & Simko, 2017), devtools (Wickham, 2015), dismo (Hijmans & Elith, 2017), gbm (Freund & Schapire, 1997; Friedman, 2002), ggplot2 (Wickham et al., 2019), lattice (Sarkar, 2008), lattice (Musa & Mansor, 2021), maptools (Hijmans & Elith, 2017), modelmetrics (Hvitfeldt & Silge, 2021), pander (Wickham, 2015), plyr (Wickham & Wickham, 2015), pROC (Robin et al., 2011), raster (Hijmans & Elith, 2017), RColorBrewer (Neuwirth, 2014), Rcpp (Eddelbeuttel & Balamura, 2018), rgdal (Verzani, 2011), sdm (Naimi & Araujo, 2016), sf (e.g., Zainuddin, 2023), sp (Pebesma, 2020) and usethis (Gladstone, 2022).

    It is important to follow all the codes in order to obtain results from the ecological response and spatial distribution models. In particular, for the ecological scenario, we selected the Generalized Linear Model (GLM) and for the geographic scenario we selected DOMAIN, also known as Gower's metric (Carpenter et al., 1993). We selected this regression method and this distance similarity metric because of its adequacy and robustness for studies with endemic or threatened species (e.g., Naoki et al., 2006). Next, we explain the statistical parameterization for the codes immersed in the GLM and DOMAIN running:

    In the first instance, we generated the background points and extracted the values of the variables (Code2_Extract_values_DWp_SC.R). Barbet-Massin et al. (2012) recommend the use of 10,000 background points when using regression methods (e.g., Generalized Linear Model) or distance-based models (e.g., DOMAIN). However, we considered important some factors such as the extent of the area and the type of study species for the correct selection of the number of points (Pers. Obs.). Then, we extracted the values of predictor variables (e.g., bioclimatic, topographic, demographic, habitat) in function of presence and background points (e.g., Hijmans and Elith, 2017).

    Subsequently, we subdivide both the presence and background point groups into 75% training data and 25% test data, each group, following the method of Soberón & Nakamura (2009) and Hijmans & Elith (2017). For a training control, the 10-fold (cross-validation) method is selected, where the response variable presence is assigned as a factor. In case that some other variable would be important for the study species, it should also be assigned as a factor (Kim, 2009).

    After that, we ran the code for the GBM method (Gradient Boost Machine; Code3_GBM_Relative_contribution.R and Code4_Relative_contribution.R), where we obtained the relative contribution of the variables used in the model. We parameterized the code with a Gaussian distribution and cross iteration of 5,000 repetitions (e.g., Friedman, 2002; kim, 2009; Hijmans and Elith, 2017). In addition, we considered selecting a validation interval of 4 random training points (Personal test). The obtained plots were the partial dependence blocks, in function of each predictor variable.

    Subsequently, the correlation of the variables is run by Pearson's method (Code5_Pearson_Correlation.R) to evaluate multicollinearity between variables (Guisan & Hofer, 2003). It is recommended to consider a bivariate correlation ± 0.70 to discard highly correlated variables (e.g., Awan et al., 2021).

    Once the above codes were run, we uploaded the same subgroups (i.e., presence and background groups with 75% training and 25% testing) (Code6_Presence&backgrounds.R) for the GLM method code (Code7_GLM_model.R). Here, we first ran the GLM models per variable to obtain the p-significance value of each variable (alpha ≤ 0.05); we selected the value one (i.e., presence) as the likelihood factor. The generated models are of polynomial degree to obtain linear and quadratic response (e.g., Fielding and Bell, 1997; Allouche et al., 2006). From these results, we ran ecological response curve models, where the resulting plots included the probability of occurrence and values for continuous variables or categories for discrete variables. The points of the presence and background training group are also included.

    On the other hand, a global GLM was also run, from which the generalized model is evaluated by means of a 2 x 2 contingency matrix, including both observed and predicted records. A representation of this is shown in Table 1 (adapted from Allouche et al., 2006). In this process we select an arbitrary boundary of 0.5 to obtain better modeling performance and avoid high percentage of bias in type I (omission) or II (commission) errors (e.g., Carpenter et al., 1993; Fielding and Bell, 1997; Allouche et al., 2006; Kim, 2009; Hijmans and Elith, 2017).

    Table 1. Example of 2 x 2 contingency matrix for calculating performance metrics for GLM models. A represents true presence records (true positives), B represents false presence records (false positives - error of commission), C represents true background points (true negatives) and D represents false backgrounds (false negatives - errors of omission).

    Validation set

    Model

    True

    False

    Presence

    A

    B

    Background

    C

    D

    We then calculated the Overall and True Skill Statistics (TSS) metrics. The first is used to assess the proportion of correctly predicted cases, while the second metric assesses the prevalence of correctly predicted cases (Olden and Jackson, 2002). This metric also gives equal importance to the prevalence of presence prediction as to the random performance correction (Fielding and Bell, 1997; Allouche et al., 2006).

    The last code (i.e., Code8_DOMAIN_SuitHab_model.R) is for species distribution modelling using the DOMAIN algorithm (Carpenter et al., 1993). Here, we loaded the variable stack and the presence and background group subdivided into 75% training and 25% test, each. We only included the presence training subset and the predictor variables stack in the calculation of the DOMAIN metric, as well as in the evaluation and validation of the model.

    Regarding the model evaluation and estimation, we selected the following estimators:

    1) partial ROC, which evaluates the approach between the curves of positive (i.e., correctly predicted presence) and negative (i.e., correctly predicted absence) cases. As farther apart these curves are, the model has a better prediction performance for the correct spatial distribution of the species (Manzanilla-Quiñones, 2020).

    2) ROC/AUC curve for model validation, where an optimal performance threshold is estimated to have an expected confidence of 75% to 99% probability (De Long et al., 1988).

  10. Geostatistical Analysis of SARS-CoV-2 Positive Cases in the United States

    • zenodo.org
    • data.niaid.nih.gov
    Updated Sep 17, 2020
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Peter K. Rogan; Peter K. Rogan (2020). Geostatistical Analysis of SARS-CoV-2 Positive Cases in the United States [Dataset]. http://doi.org/10.5281/zenodo.4032708
    Explore at:
    Dataset updated
    Sep 17, 2020
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Peter K. Rogan; Peter K. Rogan
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Geostatistics analyzes and predicts the values associated with spatial or spatial-temporal phenomena. It incorporates the spatial (and in some cases temporal) coordinates of the data within the analyses. It is a practical means of describing spatial patterns and interpolating values for locations where samples were not taken (and measures the uncertainty of those values, which is critical to informed decision making). This archive contains results of geostatistical analysis of COVID-19 case counts for all available US counties. Test results were obtained with ArcGIS Pro (ESRI). Sources are state health departments, which are scraped and aggregated by the Johns Hopkins Coronavirus Resource Center and then pre-processed by MappingSupport.com.

    This update of the Zenodo dataset (version 6) consists of three compressed archives containing geostatistical analyses of SARS-CoV-2 testing data. This dataset utilizes many of the geostatistical techniques used in previous versions of this Zenodo archive, but has been significantly expanded to include analyses of up-to-date U.S. COVID-19 case data (from March 24th to September 8th, 2020):

    Archive #1: “1.Geostat. Space-Time analysis of SARS-CoV-2 in the US (Mar24-Sept6).zip” – results of a geostatistical analysis of COVID-19 cases incorporating spatially-weighted hotspots that are conserved over one-week timespans. Results are reported starting from when U.S. COVID-19 case data first became available (March 24th, 2020) for 25 consecutive 1-week intervals (March 24th through to September 6th, 2020). Hotspots, where found, are reported in each individual state, rather than the entire continental United States.

    Archive #2: "2.Geostat. Spatial analysis of SARS-CoV-2 in the US (Mar24-Sept8).zip" – the results from geostatistical spatial analyses only of corrected COVID-19 case data for the continental United States, spanning the period from March 24th through September 8th, 2020. The geostatistical techniques utilized in this archive includes ‘Hot Spot’ analysis and ‘Cluster and Outlier’ analysis.

    Archive #3: "3.Kriging and Densification of SARS-CoV-2 in LA and MA.zip" – this dataset provides preliminary kriging and densification analysis of COVID-19 case data for certain dates within the U.S. states of Louisiana and Massachusetts.

    These archives consist of map files (as both static images and as animations) and data files (including text files which contain the underlying data of said map files [where applicable]) which were generated when performing the following Geostatistical analyses: Hot Spot analysis (Getis-Ord Gi*) [‘Archive #1’: consecutive weeklong Space-Time Hot Spot analysis; ‘Archive #2’: daily Hot Spot Analysis], Cluster and Outlier analysis (Anselin Local Moran's I) [‘Archive #2’], Spatial Autocorrelation (Global Moran's I) [‘Archive #2’], and point-to-point comparisons with Kriging and Densification analysis [‘Archive #3’].

    The Word document provided ("Description-of-Archive.Updated-Geostatistical-Analysis-of-SARS-CoV-2 (version 6).docx") details the contents of each file and folder within these three archives and gives general interpretations of these results.

  11. spatial temporal sample data

    • kaggle.com
    zip
    Updated Dec 14, 2018
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Takao Shibamoto (2018). spatial temporal sample data [Dataset]. https://www.kaggle.com/takaoo/spatial-temporal-data-examples
    Explore at:
    zip(12020 bytes)Available download formats
    Dataset updated
    Dec 14, 2018
    Authors
    Takao Shibamoto
    Description

    Dataset

    This dataset was created by Takao Shibamoto

    Contents

    It contains the following files:

  12. d

    Data from: CrimeMapTutorial Workbooks and Sample Data for ArcView and...

    • catalog.data.gov
    • icpsr.umich.edu
    • +1more
    Updated Mar 12, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    National Institute of Justice (2025). CrimeMapTutorial Workbooks and Sample Data for ArcView and MapInfo, 2000 [Dataset]. https://catalog.data.gov/dataset/crimemaptutorial-workbooks-and-sample-data-for-arcview-and-mapinfo-2000-3c9be
    Explore at:
    Dataset updated
    Mar 12, 2025
    Dataset provided by
    National Institute of Justice
    Description

    CrimeMapTutorial is a step-by-step tutorial for learning crime mapping using ArcView GIS or MapInfo Professional GIS. It was designed to give users a thorough introduction to most of the knowledge and skills needed to produce daily maps and spatial data queries that uniformed officers and detectives find valuable for crime prevention and enforcement. The tutorials can be used either for self-learning or in a laboratory setting. The geographic information system (GIS) and police data were supplied by the Rochester, New York, Police Department. For each mapping software package, there are three PDF tutorial workbooks and one WinZip archive containing sample data and maps. Workbook 1 was designed for GIS users who want to learn how to use a crime-mapping GIS and how to generate maps and data queries. Workbook 2 was created to assist data preparers in processing police data for use in a GIS. This includes address-matching of police incidents to place them on pin maps and aggregating crime counts by areas (like car beats) to produce area or choropleth maps. Workbook 3 was designed for map makers who want to learn how to construct useful crime maps, given police data that have already been address-matched and preprocessed by data preparers. It is estimated that the three tutorials take approximately six hours to complete in total, including exercises.

  13. d

    The significance of the world map in video games for the interpretation of...

    • search.dataone.org
    Updated Sep 25, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Horbiński, Tymoteusz (2024). The significance of the world map in video games for the interpretation of spatial situations using the example of The Witcher 3: Wild Hunt [Dataset]. http://doi.org/10.7910/DVN/ZVBNY0
    Explore at:
    Dataset updated
    Sep 25, 2024
    Dataset provided by
    Harvard Dataverse
    Authors
    Horbiński, Tymoteusz
    Description

    This article examines the significance of the world map in video games for the interpretation of spatial situations. An example is the popular role-playing game The Witcher 3: Wild Hunt. Nowadays, most video games are characterized by the presence of a spatial aspect. The game world map is the most important navigational element of the game that the gamer can use. To this end, the authors decided to test the importance of the game world map in the context of analyzing different examples of spatial situations that appear in The Witcher 3: Wild Hunt by the respondents. Eye movement tracking was chosen as the research method. The analysis was conducted using statistical tests. Both gamers and non-gamers of The Witcher 3: Wild Hunt, gamers and non-gamers in general, and people who identified themselves as women or men participated in the survey. Each subject was shown 5 movies (1 introductory movie, 4 movies in the main part of the study) from the gameplay of the game, in which the game world map was opened. After each video, a question was asked about both the gameplay and the game world map. It was found that familiarity with The Witcher 3: Wild Hunt, frequency of playing video games and gender influenced the correctness and time of answering the questions asked. In addition, it was found that the game world map and gameplay segments do not cognitively burden the users. Differences in visual strategy were observed between the groups of test subjects. The authors emphasized the importance of conducting further research on video games in relation to the analysis of spatial situations.

  14. d

    (HS 5) Large Extent Spatial Datasets in Virginia

    • search.dataone.org
    • hydroshare.org
    • +1more
    Updated Oct 19, 2024
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Young-Don Choi (2024). (HS 5) Large Extent Spatial Datasets in Virginia [Dataset]. http://doi.org/10.4211/hs.96794cc564894ffa9f8fb4e419a8720f
    Explore at:
    Dataset updated
    Oct 19, 2024
    Dataset provided by
    Hydroshare
    Authors
    Young-Don Choi
    Area covered
    Description

    This HydroShare resource was created to share large extent spatial (LES) datasets in Virginia on GeoServer (https://geoserver.hydroshare.org/geoserver/web/wicket/bookmarkable/org.geoserver.web.demo.MapPreviewPage) and THREDDS (https://thredds.hydroshare.org/thredds/catalog/hydroshare/resources/catalog.html).

    Users can access the uploaded LES datasets on HydroShare-GeoServer and THREDDS using this HS resource id. This resource was created using HS 2.

    Then, through the RHESSys workflows, users can subset LES datasets using OWSLib and xarray.

  15. Data from: The Effects of Spatial Reference Systems on the Predictive...

    • data.wu.ac.at
    • data.gov.au
    pdf
    Updated Jun 24, 2017
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Geoscience Australia (2017). The Effects of Spatial Reference Systems on the Predictive Accuracy of Spatial Interpolation Methods [Dataset]. https://data.wu.ac.at/schema/data_gov_au/MDk3MDczYmUtOGJiNy00ZTZjLTg5ZDEtOTJjOTFjZTY4ZDc3
    Explore at:
    pdfAvailable download formats
    Dataset updated
    Jun 24, 2017
    Dataset provided by
    Geoscience Australiahttp://ga.gov.au/
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Area covered
    cdb4de486436d3d9ac634ede7971967692d8235f
    Description

    Geoscience Australia has been deriving raster sediment datasets for the continental Australian Exclusive Economic Zone (AEEZ) using existing marine samples collected by Geoscience Australia and external organisations. Since seabed sediment data are collected at sparsely and unevenly distributed locations, spatial interpolation methods become essential tools for generating spatially continuous information. Previous studies have examined a number of factors that affect the performance of spatial interpolation methods. These factors include sample density, data variation, sampling design, spatial distribution of samples, data quality, correlation of primary and secondary variables, and interaction among some of these factors. Apart from these factors, a spatial reference system used to define sample locations is potentially another factor and is worth investigating. In this study, we aim to examine the degree to which spatial reference systems can affect the predictive accuracy of spatial interpolation methods in predicting marine environmental variables in the continental AEEZ. Firstly, we reviewed spatial reference systems including geographic coordinate systems and projected coordinate systems/map projections, with particular attention paid to map projection classification, distortion and selection schemes; secondly, we selected eight systems that are suitable for the spatial prediction of marine environmental data in the continental AEEZ. These systems include two geographic coordinate systems (WGS84 and GDA94) and six map projections (Lambert Equal-area Azimuthal, Equidistant Azimuthal, Stereographic Conformal Azimuthal, Albers Equal-Area Conic, Equidistant Conic and Lambert Conformal Conic); thirdly, we applied two most commonly used spatial interpolation methods, i.e. inverse distance squared (IDS) and ordinary kriging (OK) to a marine dataset projected using the eight systems. The accuracy of the methods was assessed using leave-one-out cross validation in terms of their predictive errors and, visualization of prediction maps. The difference in the predictive errors between WGS84 and the map projections were compared using paired Mann-Whitney test for both IDW and OK. The data manipulation and modelling work were implemented in ArcGIS and R. The result from this study confirms that the little shift caused by the tectonic movement between WGS84 and GDA94 does not affect the accuracy of the spatial interpolation methods examined (IDS and OK). With respect to whether the unit difference in geographical coordinates or distortions introduced by map projections has more effect on the performance of the spatial interpolation methods, the result shows that the accuracies of the spatial interpolation methods in predicting seabed sediment data in the SW region of AEEZ are similar and the differences are considered negligible, both in terms of predictive errors and prediction map visualisations. Among the six map projections, the slightly better prediction performance from Lambert Equal-Area Azimuthal and Equidistant Azimuthal projections for both IDS and OK indicates that Equal-Area and Equidistant projections with Azimuthal surfaces are more suitable than other projections for spatial predictions of seabed sediment data in the SW region of AEEZ. The outcomes of this study have significant implications for spatial predictions in environmental science. Future spatial prediction work using a data density greater than that in this study may use data based on WGS84 directly and may not have to project the data using certain spatial reference systems. The findings are applicable to spatial predictions of both marine and terrestrial environmental variables.

    You can also purchase hard copies of Geoscience Australia data and other products at http://www.ga.gov.au/products-services/how-to-order-products/sales-centre.html

  16. J

    Estimation of sample selection models with spatial dependence (replication...

    • jda-test.zbw.eu
    • journaldata.zbw.eu
    csv, pdf, txt
    Updated Jul 22, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Alfonso Flores-Lagunes; Kurt E. Schnier; Alfonso Flores-Lagunes; Kurt E. Schnier (2024). Estimation of sample selection models with spatial dependence (replication data) [Dataset]. https://jda-test.zbw.eu/dataset/estimation-of-sample-selection-models-with-spatial-dependence
    Explore at:
    pdf(319310), csv(704598), csv(50715), txt(1446)Available download formats
    Dataset updated
    Jul 22, 2024
    Dataset provided by
    ZBW - Leibniz Informationszentrum Wirtschaft
    Authors
    Alfonso Flores-Lagunes; Kurt E. Schnier; Alfonso Flores-Lagunes; Kurt E. Schnier
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    We consider the estimation of a sample selection model that exhibits spatial autoregressive errors (SAE). Our methodology is motivated by a two-step strategy where in the first step we estimate a spatial probit model and in the second step (outcome equation) we include an estimated inverse Mills ratio (IMR) as a regressor to control for selection bias. Since the appropriate IMR under SAE depends on a parameter from the second step, both steps are jointly estimated employing the generalized method of moments. We explore the finite sample properties of the estimator using simulations and provide an empirical illustration.

  17. d

    Replication Data for the Poverty Rates Example in Chapter 4 of Spatial...

    • search.dataone.org
    Updated Nov 21, 2023
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Darmofal, David (2023). Replication Data for the Poverty Rates Example in Chapter 4 of Spatial Analysis for the Social Sciences [Dataset]. http://doi.org/10.7910/DVN/OCINEV
    Explore at:
    Dataset updated
    Nov 21, 2023
    Dataset provided by
    Harvard Dataverse
    Authors
    Darmofal, David
    Description

    Replication data for the poverty rates example in Chapter 4 of Spatial Analysis for the Social Sciences.

  18. Z

    Dataset of Mordvin GOAL-cases

    • data.niaid.nih.gov
    Updated Feb 2, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Riku Erkkilä (2022). Dataset of Mordvin GOAL-cases [Dataset]. https://data.niaid.nih.gov/resources?id=zenodo_5940174
    Explore at:
    Dataset updated
    Feb 2, 2022
    Dataset authored and provided by
    Riku Erkkilä
    License

    Attribution-NonCommercial-NoDerivs 4.0 (CC BY-NC-ND 4.0)https://creativecommons.org/licenses/by-nc-nd/4.0/
    License information was derived automatically

    Description

    This open access dataset contains examples of the goal-oriented cases illative and lative in Mordvin languages Erzya and Moksha. The dataset consists of two parts: senses of the goal-cases (files with prefix sense) and type of the landmark noun (files with prefix LM). There is currently one dataset that includes 200 analyzed examples of each case in each language (800 examples in total). All the data is analyzed according to sense and landmark type. The data is collected from MokshEr corpus maintained by University of Turku.

    Sense means the semantic content that is expressed by the goal-case in a clause. The following values are found in the data: direction, location, part, place, purpose, reason, result, staying, target, and temporal. In addition, contextual variants of the senses are shown in brackets.

    Landmark type means what kind of entity the referent of the landmark is. There are eight categories in the data: 1D object, 2D bounded landmark, 2D unbounded landmark, 3D bounded landmark, 3D unbounded landmark, abstract landmark, institution, and temporal landmark.

    The senses-data is annotated for following information:

    The noun or relational noun phrase inflected in goal-case.

    The predicate as inflected in the data.

    Translations of both (mainly in citation form, but in predicate sometimes with some grammatical information)

    The sense of the goal-case in the utterance.

    The prototypicality of the example as a member of its sense on a scale from 1 (non- prototypical) to 5 (prototypical) NB! This assessment is based on the authors language competence and on general semantic principles, and as such should be considered only as directive.

    The original sentence from the corpus.

    Free translation. Some of the translations are done following the lexical meanings and syntactic structures of Mordvin languages, so the English is unidiomatic from time to time.

    The file name with which the original sentence can be located in the corpus.

    The landmark-data is annotated for the same information, except that there is additional information of the landmark type, and the prototypicality score is a value from 1 (non-prototypical) to 4 (prototypical) showing how prototypical the referent of the landmark is in its landmark type. Score 0 means that the prototypicality scale is not applied to the landmark type in question. Unlike in sense-data, the prototypicality in the landmark-data is based on prelinguistic spatial primitives of containment and support, as well as the human ability to recognize boundaries.

    The author of this dataset is Riku Erkkilä and it is published under CC-BY-NC-ND licence. The data is originally collected in the framework of Descriptive Grammar of Mordvin project (University of Helsinki) funded by Kone Foundation. If used in a publication, please refer to this publication as well as mention the original source:

    MokshEr V.3 (2010). Corpus of Mordvin languages. University of Turku.

    This dataset has been used in following publications:

  19. d

    Replication Data for the Senate Roll-Call Voting Example in Chapter 4 of...

    • search.dataone.org
    • dataverse.harvard.edu
    Updated Nov 21, 2023
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Darmofal, David (2023). Replication Data for the Senate Roll-Call Voting Example in Chapter 4 of Spatial Analysis for the Social Sciences [Dataset]. http://doi.org/10.7910/DVN/7QX2J2
    Explore at:
    Dataset updated
    Nov 21, 2023
    Dataset provided by
    Harvard Dataverse
    Authors
    Darmofal, David
    Description

    Replication data for the Senate roll-call voting example in Chapter 4 of Spatial Analysis for the Social Sciences.

  20. H

    Replication Data for the New Deal Realignment GWR Model Example in Chapter 7...

    • dataverse.harvard.edu
    doc, docx, tsv
    Updated Jun 28, 2015
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Harvard Dataverse (2015). Replication Data for the New Deal Realignment GWR Model Example in Chapter 7 of Spatial Analysis for the Social Sciences [Dataset]. http://doi.org/10.7910/DVN/HY7HGC
    Explore at:
    tsv(291458), doc(25600), docx(13024)Available download formats
    Dataset updated
    Jun 28, 2015
    Dataset provided by
    Harvard Dataverse
    License

    CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
    License information was derived automatically

    Description

    Replication data for the New Deal realignment GWR model example in Chapter 7 of Spatial Analysis for the Social Sciences.

Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
Jarlath O'Neal-Dunne; Morgan Grove (2012). Geodatabase for the Baltimore Ecosystem Study Spatial Data [Dataset]. http://doi.org/10.6073/pasta/377da686246f06554f7e517de596cd2b

Geodatabase for the Baltimore Ecosystem Study Spatial Data

Explore at:
287 scholarly articles cite this dataset (View in Google Scholar)
application/vnd.rar(29574980 kilobyte)Available download formats
Dataset updated
May 4, 2012
Dataset provided by
EDI
Authors
Jarlath O'Neal-Dunne; Morgan Grove
Time period covered
Jan 1, 1999 - Jun 1, 2014
Area covered
Description

The establishment of a BES Multi-User Geodatabase (BES-MUG) allows for the storage, management, and distribution of geospatial data associated with the Baltimore Ecosystem Study. At present, BES data is distributed over the internet via the BES website. While having geospatial data available for download is a vast improvement over having the data housed at individual research institutions, it still suffers from some limitations. BES-MUG overcomes these limitations; improving the quality of the geospatial data available to BES researches, thereby leading to more informed decision-making.

   BES-MUG builds on Environmental Systems Research Institute's (ESRI) ArcGIS and ArcSDE technology. ESRI was selected because its geospatial software offers robust capabilities. ArcGIS is implemented agency-wide within the USDA and is the predominant geospatial software package used by collaborating institutions.


   Commercially available enterprise database packages (DB2, Oracle, SQL) provide an efficient means to store, manage, and share large datasets. However, standard database capabilities are limited with respect to geographic datasets because they lack the ability to deal with complex spatial relationships. By using ESRI's ArcSDE (Spatial Database Engine) in conjunction with database software, geospatial data can be handled much more effectively through the implementation of the Geodatabase model. Through ArcSDE and the Geodatabase model the database's capabilities are expanded, allowing for multiuser editing, intelligent feature types, and the establishment of rules and relationships. ArcSDE also allows users to connect to the database using ArcGIS software without being burdened by the intricacies of the database itself.


   For an example of how BES-MUG will help improve the quality and timeless of BES geospatial data consider a census block group layer that is in need of updating. Rather than the researcher downloading the dataset, editing it, and resubmitting to through ORS, access rules will allow the authorized user to edit the dataset over the network. Established rules will ensure that the attribute and topological integrity is maintained, so that key fields are not left blank and that the block group boundaries stay within tract boundaries. Metadata will automatically be updated showing who edited the dataset and when they did in the event any questions arise.


   Currently, a functioning prototype Multi-User Database has been developed for BES at the University of Vermont Spatial Analysis Lab, using Arc SDE and IBM's DB2 Enterprise Database as a back end architecture. This database, which is currently only accessible to those on the UVM campus network, will shortly be migrated to a Linux server where it will be accessible for database connections over the Internet. Passwords can then be handed out to all interested researchers on the project, who will be able to make a database connection through the Geographic Information Systems software interface on their desktop computer. 


   This database will include a very large number of thematic layers. Those layers are currently divided into biophysical, socio-economic and imagery categories. Biophysical includes data on topography, soils, forest cover, habitat areas, hydrology and toxics. Socio-economics includes political and administrative boundaries, transportation and infrastructure networks, property data, census data, household survey data, parks, protected areas, land use/land cover, zoning, public health and historic land use change. Imagery includes a variety of aerial and satellite imagery.


   See the readme: http://96.56.36.108/geodatabase_SAL/readme.txt


   See the file listing: http://96.56.36.108/geodatabase_SAL/diroutput.txt
Search
Clear search
Close search
Google apps
Main menu