Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
In this course, you will explore a variety of open-source technologies for working with geosptial data, performing spatial analysis, and undertaking general data science. The first component of the class focuses on the use of QGIS and associated technologies (GDAL, PROJ, GRASS, SAGA, and Orfeo Toolbox). The second component of the class introduces Python and associated open-source libraries and modules (NumPy, Pandas, Matplotlib, Seaborn, GeoPandas, Rasterio, WhiteboxTools, and Scikit-Learn) used by geospatial scientists and data scientists. We also provide an introduction to Structured Query Language (SQL) for performing table and spatial queries. This course is designed for individuals that have a background in GIS, such as working in the ArcGIS environment, but no prior experience using open-source software and/or coding. You will be asked to work through a series of lecture modules and videos broken into several topic areas, as outlined below. Fourteen assignments and the required data have been provided as hands-on opportunites to work with data and the discussed technologies and methods. If you have any questions or suggestions, feel free to contact us. We hope to continue to update and improve this course. This course was produced by West Virginia View (http://www.wvview.org/) with support from AmericaView (https://americaview.org/). This material is based upon work supported by the U.S. Geological Survey under Grant/Cooperative Agreement No. G18AP00077. The views and conclusions contained in this document are those of the authors and should not be interpreted as representing the opinions or policies of the U.S. Geological Survey. Mention of trade names or commercial products does not constitute their endorsement by the U.S. Geological Survey. After completing this course you will be able to: apply QGIS to visualize, query, and analyze vector and raster spatial data. use available resources to further expand your knowledge of open-source technologies. describe and use a variety of open data formats. code in Python at an intermediate-level. read, summarize, visualize, and analyze data using open Python libraries. create spatial predictive models using Python and associated libraries. use SQL to perform table and spatial queries at an intermediate-level.
Open Government Licence - Canada 2.0https://open.canada.ca/en/open-government-licence-canada
License information was derived automatically
Have you ever wanted to create your own maps, or integrate and visualize spatial datasets to examine changes in trends between locations and over time? Follow along with these training tutorials on QGIS, an open source geographic information system (GIS) and learn key concepts, procedures and skills for performing common GIS tasks – such as creating maps, as well as joining, overlaying and visualizing spatial datasets. These tutorials are geared towards new GIS users. We’ll start with foundational concepts, and build towards more advanced topics throughout – demonstrating how with a few relatively easy steps you can get quite a lot out of GIS. You can then extend these skills to datasets of thematic relevance to you in addressing tasks faced in your day-to-day work.
Public Domain Mark 1.0https://creativecommons.org/publicdomain/mark/1.0/
License information was derived automatically
This dataset holds all materials for the Inform E-learning GIS course
Through the Department of the Interior-Bureau of Indian Affairs Enterprise License Agreement (DOI-BIA ELA) program, BIA employees and employees of federally-recognized Tribes may access a variety of geographic information systems (GIS) online courses and instructor-led training events throughout the year at no cost to them. These online GIS courses and instructor-led training events are hosted by the Branch of Geospatial Support (BOGS) or offered by BOGS in partnership with other organizations and federal agencies. Online courses are self-paced and available year-round, while instructor-led training events have limited capacity and require registration and attendance on specific dates. This dataset does not any training where the course was not completed by the participant or where training was cancelled or otherwise not able to be completed. Point locations depict BIA Office locations or Tribal Office Headquarters. For completed trainings where a participant location was not provided a point locations may not be available. For more information on the Branch of Geospatial Support Geospatial training program, please visit:https://www.bia.gov/service/geospatial-training.
Through the Department of the Interior-Bureau of Indian Affairs Enterprise License Agreement (DOI-BIA ELA) program, BIA employees and employees of federally-recognized Tribes may access a variety of geographic information systems (GIS) online courses and instructor-led training events throughout the year at no cost to them. These online GIS courses and instructor-led training events are hosted by the Branch of Geospatial Support (BOGS) or offered by BOGS in partnership with other organizations and federal agencies. Online courses are self-paced and available year-round, while instructor-led training events have limited capacity and require registration and attendance on specific dates. This dataset does not any training where the course was not completed by the participant or where training was cancelled or otherwise not able to be completed. Point locations depict BIA Office locations or Tribal Office Headquarters. For completed trainings where a participant _location was not provided a point locations may not be available. For more information on the Branch of Geospatial Support Geospatial training program, please visit:https://www.bia.gov/service/geospatial-training.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
This seminar is an applied study of deep learning methods for extracting information from geospatial data, such as aerial imagery, multispectral imagery, digital terrain data, and other digital cartographic representations. We first provide an introduction and conceptualization of artificial neural networks (ANNs). Next, we explore appropriate loss and assessment metrics for different use cases followed by the tensor data model, which is central to applying deep learning methods. Convolutional neural networks (CNNs) are then conceptualized with scene classification use cases. Lastly, we explore semantic segmentation, object detection, and instance segmentation. The primary focus of this course is semantic segmenation for pixel-level classification. The associated GitHub repo provides a series of applied examples. We hope to continue to add examples as methods and technologies further develop. These examples make use of a vareity of datasets (e.g., SAT-6, topoDL, Inria, LandCover.ai, vfillDL, and wvlcDL). Please see the repo for links to the data and associated papers. All examples have associated videos that walk through the process, which are also linked to the repo. A variety of deep learning architectures are explored including UNet, UNet++, DeepLabv3+, and Mask R-CNN. Currenlty, two examples use ArcGIS Pro and require no coding. The remaining five examples require coding and make use of PyTorch, Python, and R within the RStudio IDE. It is assumed that you have prior knowledge of coding in the Python and R enviroinments. If you do not have experience coding, please take a look at our Open-Source GIScience and Open-Source Spatial Analytics (R) courses, which explore coding in Python and R, respectively. After completing this seminar you will be able to: explain how ANNs work including weights, bias, activation, and optimization. describe and explain different loss and assessment metrics and determine appropriate use cases. use the tensor data model to represent data as input for deep learning. explain how CNNs work including convolutional operations/layers, kernel size, stride, padding, max pooling, activation, and batch normalization. use PyTorch, Python, and R to prepare data, produce and assess scene classification models, and infer to new data. explain common semantic segmentation architectures and how these methods allow for pixel-level classification and how they are different from traditional CNNs. use PyTorch, Python, and R (or ArcGIS Pro) to prepare data, produce and assess semantic segmentation models, and infer to new data.
A community's permit file must have an official record that shows new buildings and substantial improvements in all identified Special Flood Hazard Areas (SFHAs)are properly elevated. This elevation information is needed to show compliance with the floodplain management ordinance. FEMA encourages communities to use the Elevation Certificate developed by FEMA to fulfill this requirement since it also can be used by the property owner to obtain flood insurance. Communities participating in the Community Rating System (CRS) are required to use the FEMA Online Elevation Certificate, FEMA Form FF-206-FY-22-152 (formerly 086-0-33).
Seattle Parks and Recreation Golf Course locations. SPR Golf Courses are managed by contractors.Refresh Cycle: WeeklyFeature Class: DPR.GolfCourse
This web map shows the cumulative data of certificate of compliance issued from 2018 onwards in Hong Kong. It is a set of data made available by the Lands Department under the Government of Hong Kong Special Administrative Region (the "Government") at https://portal.csdi.gov.hk ("CSDI Portal"). The source data has been processed and converted into Esri File Geodatabase format and uploaded to Esri's ArcGIS Online platform for sharing and reference purpose. The objectives are to facilitate our Hong Kong ArcGIS Online users to use the data in a spatial ready format and save their data conversion effort.For details about the data, source format and terms of conditions of usage, please refer to the website of Hong Kong CSDI Portal at https://portal.csdi.gov.hk.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
In successful geoinformatics education, students’ active role in the learning process, e.g. through applying self-assessment, show an increasing interest but the evidence of benefits and challenges of self-assessment are sporadic. In this article, we examine the usefulness of an online self-assessment tool developed for geoinformatics education. We gathered data in two Finnish universities on five courses (n = 11–73 students/course) between 2019 and 2021. We examined 1) how the students’ self-assessed knowledge and understanding in geoinformatics subject topics changed during a course, 2) how the competencies at the end of a course changed between the years in different courses, and 3) what was the perceived usefulness of the self-assessment approach among the students. The results indicate support for the implementation of self-assessment, both as a formative and summative assessment. However, it is crucial to ensure that the students understand the contents of the self-assessment subject topics. To increase students’ motivation to take a self-assessment, it is crucial that the teacher actively highlights how it supports their studying and learning. As the teachers of the examined courses, we discuss the benefits and challenges of the self-assessment approach and the applied tool for the future development of geoinformatics education.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
This resource contains data inputs and a Jupyter Notebook that is used to introduce Hydrologic Analysis using Terrain Analysis Using Digital Elevation Models (TauDEM) and Python. TauDEM is a free and open-source set of Digital Elevation Model (DEM) tools developed at Utah State University for the extraction and analysis of hydrologic information from topography. This resource is part of a HydroLearn Physical Hydrology learning module available at https://edx.hydrolearn.org/courses/course-v1:Utah_State_University+CEE6400+2019_Fall/about
In this activity, the student learns how to (1) derive hydrologically useful information from Digital Elevation Models (DEMs); (2) describe the sequence of steps involved in mapping stream networks, catchments, and watersheds; and (3) compute an approximate water balance for a watershed-based on publicly available data.
Please note that this exercise is designed for the Logan River watershed, which drains to USGS streamflow gauge 10109000 located just east of Logan, Utah. However, this Jupyter Notebook and the analysis can readily be applied to other locations of interest. If running the terrain analysis for other study sites, you need to prepare a DEM TIF file, an outlet shapefile for the area of interest, and the average annual streamflow and precipitation data. - There are several sources to obtain DEM data. In the U.S., the DEM data (with different spatial resolutions) can be obtained from the National Elevation Dataset available from the national map (http://viewer.nationalmap.gov/viewer/). Another DEM data source is the Shuttle Radar Topography Mission (https://www2.jpl.nasa.gov/srtm/), an international research effort that obtained digital elevation models on a near-global scale (search for Digital Elevation at https://www.usgs.gov/centers/eros/science/usgs-eros-archive-products-overview?qt-science_center_objects=0#qt-science_center_objects). - If not already available, you can generate the outlet shapefile by applying basic terrain analysis steps in geospatial information system models such as ArcGIS or QGIS. - You also need to obtain average annual streamflow and precipitation data for the watershed of interest to assess the annual water balance and calculate the runoff ratio in this exercise. In the U.S., the streamflow data can be obtained from the USGS NWIS website (https://waterdata.usgs.gov/nwis) and the precipitation from PRISM (https://prism.oregonstate.edu/normals/). Note that using other datasets may require preprocessing steps to make data ready to use for this exercise.
This geospatial vector data set represents as described in solid waste collection companies Certificate of Public Convenience and Necessity issued to regulated Solid Waste Transportation Company by the Washington Utilities and Transportation Commission.
Detailed street center lines for Baltimore City. No metadata was provided with this dataset; the UVM Spatial Analysis Lab has attempted to evaluate this dataset and generate metadata. This dataset depicts the linear boundaries for street and paved areas in Baltimore City and has an extremely high degree of positional accuracy. For the best available transportation data use the Roads_GDT_MSA dataset.
This is part of a collection of 221 Baltimore Ecosystem Study metadata records that point to a geodatabase.
The geodatabase is available online and is considerably large. Upon request, and under certain arrangements, it can be shipped on media, such as a usb hard drive.
The geodatabase is roughly 51.4 Gb in size, consisting of 4,914 files in 160 folders.
Although this metadata record and the others like it are not rich with attributes, it is nonetheless made available because the data that it represents could be indeed useful.
This is part of a collection of 221 Baltimore Ecosystem Study metadata records that point to a geodatabase.
The geodatabase is available online and is considerably large. Upon request, and under certain arrangements, it can be shipped on media, such as a usb hard drive.
The geodatabase is roughly 51.4 Gb in size, consisting of 4,914 files in 160 folders.
Although this metadata record and the others like it are not rich with attributes, it is nonetheless made available because the data that it represents could be indeed useful.
(See USGS Digital Data Series DDS-69-H) A geographic information system focusing on the Upper Cretaceous Taylor and Navarro Groups was developed for the U.S. Geological Survey's (USGS) 2003 assessment of undiscovered, technically recoverable oil and natural gas resources of the Gulf Coast Region. The USGS Energy Resources Science Center has developed map and metadata services to deliver the 2003 assessment results GIS data and services online. The Gulf Coast assessment is based on geologic elements of a total petroleum system (TPS) as described in Condon and Dyman (2005). The estimates of undiscovered oil and gas resources are within assessment units (AUs). The hydrocarbon assessment units include the assessment results as attributes within the AU polygon feature class (in geodatabase and shapefile format). Quarter-mile cells of the land surface that include single or multiple wells were created by the USGS to illustrate the degree of exploration and the type and distribution of production for each assessment unit. Other data that are available in the map documents and services include the TPS and USGS province boundaries. To easily distribute the Gulf Coast maps and GIS data, a web mapping application has been developed by the USGS, and customized ArcMap (by ESRI) projects are available for download at the Energy Resources Science Center Gulf Coast website. ArcGIS Publisher (by ESRI) was used to create a published map file (pmf) from each ArcMap document (.mxd). The basemap services being used in the GC map applications are from ArcGIS Online Services (by ESRI), and include the following layers: -- Satellite imagery -- Shaded relief -- Transportation -- States -- Counties -- Cities -- National Forests With the ESRI_StreetMap_World_2D service, detailed data, such as railroads and airports, appear as the user zooms in at larger scales.
Access the file geodatabase source data in SC State Plane coordinate system
This layer shows the location of Clubs with Certificate of Compliance in Hong Kong. It is a set of data made available by the Home Affairs Department under the Government of Hong Kong Special Administrative Region (the "Government") at https://GEODATA.GOV.HK/ ("Hong Kong Geodata Store"). The source data is in CSV format and has been processed and converted into Esri File Geodatabase format and uploaded to Esri's ArcGIS Online platform for sharing and reference purpose. The objectives are to facilitate our Hong Kong ArcGIS Online users to use the data in a spatial ready format and save their data conversion effort.For details about the data, source format and terms of conditions of usage, please refer to the website of Hong Kong Geodata Store at https://geodata.gov.hk/.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Important Note: This item is in mature support as of February 2023 and will be retired in December 2025. A new version of this item is available for your use. Esri recommends updating your maps and apps to use the new version. This layer displays change in pixels of the Sentinel-2 10m Land Use/Land Cover product developed by Esri, Impact Observatory, and Microsoft. Available years to compare with 2021 are 2018, 2019 and 2020. By default, the layer shows all comparisons together, in effect showing what changed 2018-2021. But the layer may be changed to show one of three specific pairs of years, 2018-2021, 2019-2021, or 2020-2021.Showing just one pair of years in ArcGIS Online Map ViewerTo show just one pair of years in ArcGIS Online Map viewer, create a filter. 1. Click the filter button. 2. Next, click add expression. 3. In the expression dialogue, specify a pair of years with the ProductName attribute. Use the following example in your expression dialogue to show only places that changed between 2020 and 2021:ProductNameis2020-2021By default, places that do not change appear as a
transparent symbol in ArcGIS Pro. But in ArcGIS Online Map Viewer, a transparent
symbol may need to be set for these places after a filter is
chosen. To do this:4. Click the styles button. 5. Under unique values click style options. 6. Click the symbol next to No Change at the bottom of the legend. 7. Click the slider next to "enable fill" to turn the symbol off.Showing just one pair of years in ArcGIS ProTo show just one pair of years in ArcGIS Pro, choose one of the layer's processing templates to single out a particular pair of years. The processing template applies a definition query that works in ArcGIS Pro. 1. To choose a processing template, right click the layer in the table of contents for ArcGIS Pro and choose properties. 2. In the dialogue that comes up, choose the tab that says processing templates. 3. On the right where it says processing template, choose the pair of years you would like to display. The processing template will stay applied for any analysis you may want to perform as well.How the change layer was created, combining LULC classes from two yearsImpact Observatory, Esri, and Microsoft used artificial intelligence to classify the world in 10 Land Use/Land Cover (LULC) classes for the years 2017-2021. Mosaics serve the following sets of change rasters in a single global layer: Change between 2018 and 2021Change between 2019 and 2021Change between 2020 and 2021To make this change layer, Esri used an arithmetic operation
combining the cells from a source year and 2021 to make a change index
value. ((from year * 16) + to year) In the example of the change between 2020 and 2021, the from year (2020) was multiplied by 16, then added to the to year (2021). Then the combined number is served as an index in an 8 bit unsigned mosaic with an attribute table which describes what changed or did not change in that timeframe. Variable mapped: Change in land cover between 2018, 2019, or 2020 and 2021 Data Projection: Universal Transverse Mercator (UTM)Mosaic Projection: WGS84Extent: GlobalSource imagery: Sentinel-2Cell Size: 10m (0.00008983152098239751 degrees)Type: ThematicSource: Esri Inc.Publication date: January 2022What can you do with this layer?Global LULC maps provide information on conservation planning, food security,
and hydrologic modeling, among other things. This dataset can be used to
visualize land cover anywhere on Earth. This
layer can also be used in analyses that require land cover input. For
example, the Zonal Statistics tools allow a user to understand the
composition of a specified area by reporting the total estimates for
each of the classes. Land Cover processingThis map was produced by a deep learning model trained using over 5 billion hand-labeled Sentinel-2 pixels, sampled from over 20,000 sites distributed across all major biomes of the world. The underlying deep learning model uses 6 bands of Sentinel-2 surface reflectance data: visible blue, green, red, near infrared, and two shortwave infrared bands. To create the final map, the model is run on multiple dates of imagery throughout the year, and the outputs are composited into a final representative map. Processing platformSentinel-2 L2A/B data was accessed via Microsoft’s Planetary Computer and scaled using Microsoft Azure Batch.Class definitions1. WaterAreas
where water was predominantly present throughout the year; may not
cover areas with sporadic or ephemeral water; contains little to no
sparse vegetation, no rock outcrop nor built up features like docks;
examples: rivers, ponds, lakes, oceans, flooded salt plains.2. TreesAny
significant clustering of tall (~15-m or higher) dense vegetation,
typically with a closed or dense canopy; examples: wooded vegetation,
clusters of dense tall vegetation within savannas, plantations, swamp or
mangroves (dense/tall vegetation with ephemeral water or canopy too
thick to detect water underneath).4. Flooded vegetationAreas
of any type of vegetation with obvious intermixing of water throughout a
majority of the year; seasonally flooded area that is a mix of
grass/shrub/trees/bare ground; examples: flooded mangroves, emergent
vegetation, rice paddies and other heavily irrigated and inundated
agriculture.5. CropsHuman
planted/plotted cereals, grasses, and crops not at tree height;
examples: corn, wheat, soy, fallow plots of structured land.7. Built AreaHuman
made structures; major road and rail networks; large homogenous
impervious surfaces including parking structures, office buildings and
residential housing; examples: houses, dense villages / towns / cities,
paved roads, asphalt.8. Bare groundAreas
of rock or soil with very sparse to no vegetation for the entire year;
large areas of sand and deserts with no to little vegetation; examples:
exposed rock or soil, desert and sand dunes, dry salt flats/pans, dried
lake beds, mines.9. Snow/IceLarge
homogenous areas of permanent snow or ice, typically only in mountain
areas or highest latitudes; examples: glaciers, permanent snowpack, snow
fields. 10. CloudsNo land cover information due to persistent cloud cover.11. Rangeland Open
areas covered in homogenous grasses with little to no taller
vegetation; wild cereals and grasses with no obvious human plotting
(i.e., not a plotted field); examples: natural meadows and fields with
sparse to no tree cover, open savanna with few to no trees, parks/golf
courses/lawns, pastures. Mix of small clusters of plants or single
plants dispersed on a landscape that shows exposed soil or rock;
scrub-filled clearings within dense forests that are clearly not taller
than trees; examples: moderate to sparse cover of bushes, shrubs and
tufts of grass, savannas with very sparse grasses, trees or other
plants.CitationKarra,
Kontgis, et al. “Global land use/land cover with Sentinel-2 and deep
learning.” IGARSS 2021-2021 IEEE International Geoscience and Remote
Sensing Symposium. IEEE, 2021.AcknowledgementsTraining
data for this project makes use of the National Geographic Society
Dynamic World training dataset, produced for the Dynamic World Project
by National Geographic Society in partnership with Google and the World
Resources Institute.For questions please email environment@esri.com
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
NOTE: This dataset is an older dataset that we have removed from the SGID and 'shelved' in ArcGIS Online. There may (or may not) be a newer vintage of this dataset in the SGID.
(See USGS Digital Data Series DDS-69-E) A geographic information system focusing on the Cretaceous Travis Peak and Hosston Formations was developed for the U.S. Geological Survey's (USGS) 2002 assessment of undiscovered, technically recoverable oil and natural gas resources of the Gulf Coast Region. The USGS Energy Resources Science Center has developed map and metadata services to deliver the 2002 assessment results GIS data and services online. The Gulf Coast assessment is based on geologic elements of a total petroleum system (TPS) as described in Dyman and Condon (2005). The estimates of undiscovered oil and gas resources are within assessment units (AUs). The hydrocarbon assessment units include the assessment results as attributes within the AU polygon feature class (in geodatabase and shapefile format). Quarter-mile cells of the land surface that include single or multiple wells were created by the USGS to illustrate the degree of exploration and the type and distribution of production for each assessment unit. Other data that are available in the map documents and services include the TPS and USGS province boundaries. To easily distribute the Gulf Coast maps and GIS data, a web mapping application has been developed by the USGS, and customized ArcMap (by ESRI) projects are available for download at the Energy Resources Science Center Gulf Coast website. ArcGIS Publisher (by ESRI) was used to create a published map file (pmf) from each ArcMap document (.mxd). The basemap services being used in the GC map applications are from ArcGIS Online Services (by ESRI), and include the following layers: -- Satellite imagery -- Shaded relief -- Transportation -- States -- Counties -- Cities -- National Forests With the ESRI_StreetMap_World_2D service, detailed data, such as railroads and airports, appear as the user zooms in at larger scales. This map service shows the structural configuration of the top of the Travis Peak or Hosston Formations in feet below sea level. The map was produced by calculating the difference between a datum at the land surface (either the Kelly bushing elevation or the ground surface elevation) and the reported depth of the Travis Peak or Hosston. This map service also shows the thickness of the interval from the top of the Travis Peak or Hosston Formations to the top of the Cotton Valley Group.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Degree days are based on the assumption that when the outside temperature is 65°F, we don't need heating or cooling to be comfortable. Cooling Degree Days (CDD) are the difference between the daily temperature mean (high temperature plus low temperature divided by two) and 65°F. In essence, it tells us how many degrees we need to cool our houses/buildings by each day to achieve that "comfortable" level. More information on CCDs can be found here. This layer shows the total number of CDDs needed per year over the average period of 2036-2065. This information is sourced from the high resolution LOCA climate models used in the 4th National Climate Assessment. Specifically, we are showing CDDs under a high CO2 emissions scenario (RCP 8.5), which is, at this point, the most realistic scenario. Time Extent: Annual average from 2036-2065Units: degree daysCell Size: 1/16th degree (~6 km)Source Type: StretchedPixel Type: 32 Bit floating pointData Projection: GCS WGS84Extent: United States plus some of Canada and MexicoSource: CMIP5 Localized Constructed Analogs (LOCA)What can this layer be used for?In addition to mapping, this ArcGIS Imagery for ArcGIS Online tile imagery layer supports spatial analysis, and contains 32-bit floating point values for CDD. Original data can be downloaded from the LOCA-Viewer.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
In this course, you will explore a variety of open-source technologies for working with geosptial data, performing spatial analysis, and undertaking general data science. The first component of the class focuses on the use of QGIS and associated technologies (GDAL, PROJ, GRASS, SAGA, and Orfeo Toolbox). The second component of the class introduces Python and associated open-source libraries and modules (NumPy, Pandas, Matplotlib, Seaborn, GeoPandas, Rasterio, WhiteboxTools, and Scikit-Learn) used by geospatial scientists and data scientists. We also provide an introduction to Structured Query Language (SQL) for performing table and spatial queries. This course is designed for individuals that have a background in GIS, such as working in the ArcGIS environment, but no prior experience using open-source software and/or coding. You will be asked to work through a series of lecture modules and videos broken into several topic areas, as outlined below. Fourteen assignments and the required data have been provided as hands-on opportunites to work with data and the discussed technologies and methods. If you have any questions or suggestions, feel free to contact us. We hope to continue to update and improve this course. This course was produced by West Virginia View (http://www.wvview.org/) with support from AmericaView (https://americaview.org/). This material is based upon work supported by the U.S. Geological Survey under Grant/Cooperative Agreement No. G18AP00077. The views and conclusions contained in this document are those of the authors and should not be interpreted as representing the opinions or policies of the U.S. Geological Survey. Mention of trade names or commercial products does not constitute their endorsement by the U.S. Geological Survey. After completing this course you will be able to: apply QGIS to visualize, query, and analyze vector and raster spatial data. use available resources to further expand your knowledge of open-source technologies. describe and use a variety of open data formats. code in Python at an intermediate-level. read, summarize, visualize, and analyze data using open Python libraries. create spatial predictive models using Python and associated libraries. use SQL to perform table and spatial queries at an intermediate-level.