Michigan

Legacy ID: 
25
State Code: 
MI
Country Code: 
USA
Area: 
57 899.40
Latitude: 
44.34
Longitude: 
-85.44
Publication Title: 

Bathythermograph Data, Lake Michigan, 1954

Authors: 
Patterson, E.L., Dabrowski, S.A., Bell, T.M., Nelson, S.R., Everette, A.L., Faundeen, J.L.
Related Staff: 
Lance Everette
Publication Date: 
2017-04-17
Parent Publication Title: 
Publication Type: 
Archive number: 

Pub Abstract: 

In 1954 researchers at the USGS Great Lakes Science Center conducted 11 research cruises on Lake Michigan during which 779 bathythermographs were cast to collect temperature profile data (temperature at depth). Bathythermographs of that era recorded water pressure and temperature data by mechanically etching them as a curve on a glass slide. Data was collected from the glass slide by projecting the image of the curve, superimposing a grid, and taking a photo of it, thereby creating a  bathythermogram.  Data collection personnel could then read the data from the curve. This USGS data release is a digitized set of those original bathythermogram print photos and the temperature and depth data the project team collected from them using the open-source software, Web Plot Digitizer, as well as metadata describing each. In addition, because of their historical value and potential future use, this data release includes the cruise logs, which include nautical and research notes beyond the logical scope of this data release.

Publication Title: 

Riparian spiders as sentinels of polychlorinated biphenyl contamination across heterogeneous aquatic ecosystems

FORT Contact: 
Johanna Kraus
Authors: 
Johanna M. Kraus, Polly P. Gibson, David M. Walters, and Marc A. Mills
Related Staff: 
Johanna Kraus
David Walters
Publication Date: 
2017
Parent Publication Title: 
Environmental Toxicology and Chemistry
Publication Type: 
Archive number: 
States: 

Pub Abstract: 
Riparian spiders are being used increasingly to track spatial patterns of contaminants in and fluxing from aquatic ecosystems.However, our understanding of the circumstances under which spiders are effective sentinels of aquatic pollution is limited. The present study tests the hypothesis that riparian spiders may be effectively used to track spatial patterns of sediment pollution by polychlorinated biphenyls (PCBs) in aquatic ecosystems with high habitat heterogeneity. The spatial pattern of ΣPCB concentrations in 2 common families of riparian spiders sampled in 2011 to 2013 generally tracked spatial variation in sediment ΣPCBs across all sites within the Manistique River Great Lakes Area of Concern (AOC), a rivermouth ecosystem located on the south shore of the Upper Peninsula, Manistique (MI,USA) that includes harbor, river, backwater, and lake habitats. Sediment ΣPCB concentrations normalized for total organic carbon explained 41% of the variation in lipid-normalized spider ΣPCB concentrations across 11 sites. Furthermore, 2 common riparian spider taxa (Araneidae and Tetragnathidae) were highly correlated (r2> 0.78) and had similar mean ΣPCB concentrations when averaged acrossall years. The results indicate that riparian spiders may be useful sentinels of relative PCB availability to aquatic and riparian food webs in heterogeneous aquatic ecosystems like rivermouths where habitat and contaminant variability may make the use of aquatic taxa lesseffective. Furthermore, the present approach appears robust to heterogeneity in shoreline development and riparian vegetation that support different families of large web-building spiders. Environ Toxicol Chem 2016;9999:1–9. Published 2016 Wiley Periodicals, Inc. on behalf of SETAC. This article is a US government work and, as such, is in the public domain in the United States of America.

Preserving and Publishing USGS Legacy Data at Risk: Historical Bathythermograph Data

The USGS Fort Collins Science Center’s Data at Risk (DaR) project has publicly released “Bathythermograph Data, Lake Michigan, 1954" (https://doi.org/10.5066/F7DV1H4B), the second of five USGS legacy data products identified by the project as at risk of loss or damage in 2016. The project’s third legacy data release, “River Channel Survey Data, Redwood Creek and Mill Creek, California, 1973-2013” is expected to be available to the public in the coming weeks.

Preserving and Publishing USGS Legacy Data at Risk: Historical Bathythermograph Data

The USGS Fort Collins Science Center’s Data at Risk (DaR) project has publicly released “Bathythermograph Data, Lake Michigan, 1954" (https://doi.org/10.5066/F7DV1H4B), the second of five USGS legacy data products identified by the project as at risk of loss or damage in 2016. The project’s third legacy data release, “River Channel Survey Data, Redwood Creek and Mill Creek, California, 1973-2013” is expected to be available to the public in the coming weeks.

Developing a USGS Legacy Data Inventory to Preserve and Release Historical USGS Data

Code: 
EJ10EWN.45
Abstract: 

Legacy data (n) - Information stored in an old or obsolete format or computer system that is, therefore, difficult to access or process. (Business Dictionary, 2016)

For over 135 years, the U.S. Geological Survey has collected diverse information about the natural world and how it interacts with society. Much of this legacy information is one-of-a-kind and in danger of being lost forever through decay of materials, obsolete technology, or staff changes. Several laws and orders require federal agencies to preserve and provide the public access to federally collected scientific information. The information is to be archived in a manner that allows others to examine the materials for new information or interpretations. Data-at-Risk is a systematic way for the USGS to continue efforts to meet the challenge of preserving and making accessible enormous amount of information locked away in inaccessible formats. Data-at-Risk efforts inventory and prioritize inaccessible information and assist with the preservation and release of the information into the public domain. Much of the information the USGS collects has permanent or long-term value to the Nation and the world through its contributions to furthering scientific discovery, public policies, or decisions. These information collections represent observations and events that will never be repeated and warrant preservation for future generations to learn and benefit from them.

Goal: Expand the USGS contribution to scientific discovery and knowledge by demonstrating a long-term approach to inventorying, prioritizing and releasing to the public the wealth of USGS legacy scientific data.

Objectives:

  1. Implement a systematic workflow to create a USGS Legacy Data Inventory that catalogs and describes known USGS legacy data sets.
  2. Develop a methodology to evaluate and prioritize USGS legacy data sets based on USGS mission and program objectives and potential of successful release within USGS records management and open data policies.
  3. Preserve and release select, priority legacy data sets through the USGS IPDS data release workflow
  4. Analyze the time and resources required to preserve/release legacy data and develop estimates to inform future legacy data inventory efforts.

Scope

As one of the largest and oldest earth science organizations in the world, the scientific legacy of the USGS is its data, to include, but not limited to images, video, audio files, physical samples, etc., and the scientific knowledge derived from them, gathered over 130 years of research. However, it is widely understood that high-quality data collected and analyzed as part of now completed projects are hidden away in case files, file cabinets and hard drives housed in USGS facilities. Therefore, despite their potential significance to current USGS mission and program research objectives, these “legacy data” are unavailable. In addition, legacy data are by definition at risk of permanent loss or damage because they pre-date current, open-data policies, standards and formats. Risks to legacy data can be technical, such as obsolescence of the data’s storage media and format, or they can be organizational, such as a lack of funding or facility storage. Conveniently, addressing legacy data risks such as these generally results in the science data becoming useable by modern data tools, as well as accessible to the broader scientific community.

Building on past USGS legacy data inventory and preservation projects

USGS has long history of proactively researching and developing solutions to data management needs, including legacy data inventory and preservation. For example, in 1994 USGS was instrumental in establishing the FGDC-CSDGM metadata standard for geospatial scientific data that is still part of the foundation of USGS data management. Today, USGS is a lead agency in establishing meaningful and actionable policies that facilitate data release to the greater, public scientific community. In recent years, CDI has invested in several legacy data inventory and preservation projects, including the “Legacy Data Inventory” project (aka, “Data Mine” 2013-present), which examined the time, resources and workflows needed for science centers to inventory legacy data. Another CDI project, the “North American Bat Data Recovery and Integration” project (2014-present), is preserving previously unavailable bat banding data (1932-1972) and white-nose syndrome disease data and making them available via APIs. Both of these CDI projects were forward-thinking legacy data initiatives, several years ahead of Federal open data policies and mandates.

However, one of the most comprehensive, Bureau-level legacy data preservation efforts was the   USGS Data Rescue project, which provided funding, tools, and support to USGS scientists to preserve legacy data sets at imminent risk of permanent loss or damage. A small sample of USGS science data rescued over those eight fiscal years included:

  • Inventoried, catalogued, indexed, and preserved Famine Early Warning one-of-a-kind, hardcopy maps.
  • Landsat orphan scenes, totaling over 146,000 were retrieved and processed, allowing the land research community to access previously unavailable satellite records.
  • Through a partnership with the Alaska State Division of Geological and Geophysical Surveys, the Alaska Water Science Center scanned, added metadata to, and included in a database volcano imagery dating from the 1950s to 2004.
  • 20,000 original, historical stream flow measurements from Kentucky dating from the early 1900s to the late 1980s were scanned and entered into NWIS.
  • Central Mineral and Environmental Resources Science Center geochemical data conversion totaling approximately 250,000 primary documents from paper to electronic format were completed.
  • California Water Science Center migrated paper well schedules and other groundwater records dating back more than 100 years old. The records define historical climate variability, geologic conditions where natural hazards occur, and the extents of freshwater resources.

Over 100 projects were supported in the 8 years the Data Rescue project was in operation (2006-2013), while an additional 300 projects went unfunded, providing a glimpse of the potential trove of USGS legacy data at risk of damage or loss. The urgency of and strategies for preserving USGS legacy data have been discussed at length at the 2014 CSAS&L Data Management Workshop and the 2015 CDI Workshop, further emphasizing a Bureau-wide recognition of the importance of legacy data preservation and release. During the 2015 CDI Workshop, legacy data preservation was rated a top-rated FY16 priority by the Data Management Working Group, laying the groundwork for this proposal, which intends to apply the legacy data inventory and evaluation methods developed through the CDI Legacy Data Inventory project to formalize and extend the inventory successfully started through the Data Rescue Program. By creating a formal method to submit, document and evaluate legacy data known to be in need of preservation, USGS would have a tool that USGS scientists, science centers, and mission areas can use to identify significant historical legacy data that can inform, new, data-intensive scientific efforts.

Challenges and improvements for USGS legacy data preservation and release

Based on our experiences managing and preserving USGS legacy data, we have seen two challenges that often undermine legacy data preservation and release:

  1. The most scientifically significant legacy data aren’t always the most recoverable: Legacy data by definition are “dated” because there is some length of time that has passed since the data were collected, the project completed and recovery efforts begin. The longer the time that’s passed, the more likely project staff aren’t available and supporting project and data documents are lost. Lacking this knowledge and/or documentation, metadata may not be completed, resulting in preserved data that aren’t useable - a critical element of the USGS data release peer review and approval process. If data is not useable, it is more difficult to release. Critically evaluating legacy data for their “release potential,” not just their scientific significance, increases the likelihood of selecting legacy data that will be successfully released.
  2. Research scientists may not have data science skills/expertise/resources: Traditionally, legacy data efforts provide funding directly to the data owner, who is generally a principal investigator and knows the data intimately, but may lack the data science experience, time and tools to preserve and release data in an open format with complete, compliant metadata. In our experience, this can lead to delays in preserving and releasing legacy data. Data scientists can/should not replace data owners, but they can provide a significant level of assistance to data owners, by applying their data and metadata development experience and tools.

We believe that each of challenges have good solutions that can improve the efficiency and predictability of preservation and release efforts:

  1. Make “potential for successful release” a primary evaluation factor in prioritizing and selecting legacy data for preservation and release. By developing a method of estimating the feasibility and cost of preserving and releasing data and incorporating it into the evaluation and priority criteria, we can better select and prioritize data sets.
  2. Provide funding to a USGS data scientist to collaborate with data owners and ensure preservation and releases are consistently produced and of the highest quality.

Technical Approach

Each objective of this proposal will be addressed in a sequence of 3 phases:

  1. Legacy Data Inventory Submission Period
  2. Evaluation and prioritization of the Legacy Data Inventory; selection of data sets for preservation and release.
  3. Preservation and release of selected datasets.

Phase I: Identification and inventory of USGS data at risk

Data owners will document their legacy data sets electronically, providing the primary project and data set metadata elements needed to score, evaluate and prioritize the legacy data inventory. The core of these metadata elements will be derived from the established “USGS Metadata 20 Questions” form, which has proven effective at gathering metadata from research scientists with little/no data science experience. Narrative fields will be used for evaluating need. Categorical fields will be used to calculate feasibility scores used to determine level of effort required to successfully rescue the proposed data.

Phase II: Evaluation and prioritization of the USGS data at risk requests

The CDI Data Management Working Group’s Data at Risk sub-group will facilitate the evaluation and prioritization of the legacy data inventory. Mission Areas will be engaged to verify inventory submissions are supported programmatically and meet mission objectives.  The USGS Records Management Program, Enterprise Publishing Program, and Sciencebase will be consulted to verify submitted legacy data inventory submissions can be released within Bureau records management and data release policies. Once these checkpoints have been verified the Data at Risk sub-group and data scientist will score and prioritize the legacy data inventory based on the following criteria:

  • Scientific value/significance to USGS mission area and program objectives.
  • Potential of successfully preserving and releasing the data by the data scientist.
  • Severity/Imminence of loss or damage to data based on identified risk factors.

Phase III: Preservation and Release of Select, Priority Legacy Data

Working in order of priority as set in Phase II, the data scientist(s) will collaborate with the data owner and work with them to complete the process of  preserving and releasing their legacy data. Through this data owner/scientist collaboration, the data scientist will create and validate the FGDC-CSDGM metadata and develop the data set in an open-format as documented in the metadata. By process, the data scientist will act as an agent of the data owner, coordinating and completing all steps in each workflow until the the IPDS record approved and disseminated by the Bureau and the Sciencebase data release item(s) are approved, locked and made public by the Sciencebase team. However, while the data scientist is responsible for ensuring all preservation and release tasks are completed consistently and within policies and best practices, the data owner retains all approval of final metadata attribution (e.g., title, authorship), as well as disposition of their legacy data (e.g., pre/post processing methods; derivative data architectures).

At the completion of Phase III, each legacy data release will have the following created by the data scientist:

  • complete, compliant FGDC-CSDGM metadata
  • legacy data set(s) in an open-format, publicly discoverable and available from Sciencebase.
  • a USGS highlight submitted through the SW Region to Reston.
  • a CDI update describing the data set(s) released and a summary of time and resources required to complete the release.
Publication Title: 

Community for Data Integration 2016 Annual Report

FORT Contact: 
Lance Everette
Authors: 
Langseth, M.L., Hsu, Leslie, Amberg, Jon, Bliss, Norman, Bock, A.R., Bolus, R.T., Bristol, R.S., Chase, K.J., Crimmins, T.M., Earle, P.S., Erickson, Richard, Everette, A.L., Falgout, Jeff, Faundeen, J.L., Fienen, Michael, Griffin, Rusty, Guy, M.R., Henry, K.D., Hoebelheinrich, N.J., Hunt, Randall, Hutchison, V.B., Ignizio, D.A., Infante, D.M., Jarnevich, Catherine, Jones, J.M., Kern, Tim, Leibowitz, Scott, Lightsom, F.L., Marsh, R.L., McCalla, S.G., McNiff, Marcia, Morisette, J.T., Nelson, J.C., Norkin, Tamar, Preston, T.M., Rosemartin, Alyssa, Sando, Roy, Sherba, J.T., Signell, R.P., Sleeter, B.M., Sundquist, E.T., Talbert, C.B., Viger, R.J., Weltzin, J.F., Waltman, Sharon, Weber, Marc, Wieferich, D.J., Williams, Brad, Windham-Myers, Lisamarie
Related Staff: 
Lance Everette
Drew Ignizio
Catherine Jarnevich
Tim Kern
Jeff Morisette
Colin Talbert
Publication Date: 
2017
Parent Publication Title: 
Publication Type: 
Archive number: 
Topics: 

Pub Abstract: 

The Community for Data Integration (CDI) represents a dynamic community of practice focused on advancing science data and information management and integration capabilities across the U.S. Geological Survey and the CDI community. This annual report describes the various presentations, activities, and outcomes of the CDI monthly forums, working groups, virtual training series, and other CDI-sponsored events in fiscal year 2016. The report also describes the objectives and accomplishments of the 13 CDI-funded projects in fiscal year 2016.

Social Scientists Debuts HDgov at the International Symposium on Society and Resource Management Conference

Dr. Rudy Schuster and colleagues will present a poster at the International Symposium on Society and Resource Management conference in Houghton, Michigan on June 24, 2016, to highlight HDgov. They will provide a live demonstration navigating through website resources and collaborative functions. Schuster will answer questions about why HDgov was created, how resources and tools are collected, and the ways in which the web portal can be used to help professionals and land managers incorporate human dimensions into policy decisions and land management practices.

Interdisciplinary Climate Change Study Presented at the International Symposium on Society and Resource Management

Representing the Fort Collins Science Center at the International Symposium on Society and Resource Management (ISSRS) is Dr. Rudy Schuster and NinaBurkardt, M.S., who will be presenting on a project that develops integrated adaptation planning tools for climate change scenarios. The study offers an opportunity for stakeholders, including scientists, land managers, tribal members, and planners to cooperate and identify actions to reduce the negative impacts of climate change on southwestern Colorado landscapes.

Publication Title: 

Effects of hypoxia on consumption, growth, and RNA:DNA ratios of young Yellow Perch

FORT Contact: 
James Roberts
Authors: 
Roberts, J.J., S.B. Brandt, D. Fanslow, S.A. Ludsin, S.A. Pothoven, D. Scavia, T.O. Höök
Related Staff: 
James Roberts
Publication Date: 
2011
Updated Date (text): 
2013-05-07
Parent Publication Title: 
Transactions of the American Fisheries Society
Publication Type: 
Archive number: 
NONCTR/00395
States: 

Pub Abstract: 
Publication Title: 

Seasonal and interannual effects of hypoxia on fish habitat quality in central Lake Erie

FORT Contact: 
James Roberts
Authors: 
Arend, K.K., D. Beletsky, J.V. Depinto, S.A. Ludsin, J.J. Roberts, D.K. Rucinski, D. Scavia, D.J. Schwab, and T.O. Höök
Related Staff: 
James Roberts
Publication Date: 
2011
Updated Date (text): 
2013-03-18
Parent Publication Title: 
Freshwater Biology
Publication Type: 
Archive number: 
NONCTR/00391
States: 

Pub Abstract: 

Summary

  1. Hypoxia occurs seasonally in many stratified coastal marine and freshwater ecosystems when bottom dissolved oxygen (DO) concentrations are depleted below 2–3 mg O2 L−1.
  2. We evaluated the effects of hypoxia on fish habitat quality in the central basin of Lake Erie from 1987 to 2005, using bioenergetic growth rate potential (GRP) as a proxy for habitat quality. We compared the effect of hypoxia on habitat quality of (i) rainbow smelt, Osmerus mordax mordax Mitchill (young-of-year, YOY, and adult), a cold-water planktivore, (ii) emerald shiner, Notropis atherinoides Rafinesque (adult), a warm-water planktivore, (iii) yellow perch, Perca flavescens Mitchill (YOY and adult), a cool-water benthopelagic omnivore and (iv) round goby Neogobius melanostomus Pallas (adult) a eurythermal benthivore. Annual thermal and DO profiles were generated from 1D thermal and DO hydrodynamics models developed for Lake Erie’s central basin.
  3. Hypoxia occurred annually, typically from mid-July to mid-October, which spatially and temporally overlaps with otherwise high benthic habitat quality. Hypoxia reduced the habitat quality across fish species and life stages, but the magnitude of the reduction varied both among and within species because of the differences in tolerance to low DO levels and warm-water temperatures.
  4. Across years, trends in habitat quality mirrored trends in phosphorus concentration and water column oxygen demand in central Lake Erie. The per cent reduction in habitat quality owing to hypoxia was greatest for adult rainbow smelt and round goby (mean: −35%), followed by adult emerald shiner (mean: −12%), YOY rainbow smelt (mean: −10%) and YOY and adult yellow perch (mean: −8.5%).
  5. Our results highlight the importance of differential spatiotemporally interactive effects of DO and temperature on relative fish habitat quality and quantity. These effects have the potential to influence the performance of individual fish species as well as population dynamics, trophic interactions and fish community structure.

Pages