Marine Imaging Workshop/presentations/talks
A Visualization Tool for Reproducible Image Preprocessing in Ecological Research
Tobias Ferreira , National Oceanography Centre - UK
Eric Orenstein, Jennifer Durden, Colin Sauze, Loic Van Audenhaege, Mojtaba Masoudi . All co-authors have the same filiation: National Oceanography Centre - UK
Image preprocessing is essential to prepare data for AI models and ensuring trustworthy biodiversity metrics. Properly selecting and documenting preprocessing steps is crucial for reproducible biodiversity research. However, current software solutions are fragmented and application-specific, requiring multiple tools and platforms, and are not designed for ecologists. To address this gap, the PAIDIVER project are developing an image processing workflow builder and visualization tool specifically designed for ecologists. Our workflow manager enables users to create preprocessing workflows and ensures reproducibility by meticulously recording each step in a standardized, interoperable format. Workflow visualization is provided through a lightweight web application, accessible online for testing and constructing image preprocessing workflows. The web application features an intuitive drag-and-drop interface, allowing users to easily add, remove, and rearrange processing steps while observing the output on test images. Each step can be customized with parameters and settings, providing flexibility while maintaining simplicity. Our tool emphasizes a user-friendly interface and experience, allowing users to visualize, adjust, and preview intermediate and final outputs at each stage before applying them to larger datasets. We will eventually release a Dockerized version of the environment, enabling users to run the software on any available local machine or high-performance computing (HPC) system. Furthermore, the tool provides comprehensive documentation and tutorials, making it accessible even to those with limited technical expertise. We will also release two curated image sets from contrasting environments (pelagic and benthic), complete with metadata, hosted at BODC/CEDA and bundled with the software tool. By combining an intuitive interface with powerful functionality, PAIDIVER aims to make image preprocessing accessible, reproducible, and efficient for ecological research.
A review of seafloor annotation methods used in deep-sea benthic diversity studies
Bárbara de Moura Neves , Fisheries and Oceans Canada
Bárbara de Moura Neves 1, Marion Boulard 1, Rylan Command 1, Merlin Best 2, Katleen Robert 3, Marilyn Thorne 4, Chris Yesson 5 1 Fisheries and Oceans Canada, Newfoundland and Labrador Region 2 Fisheries and Oceans Canada, Pacific Region 3 School of Ocean Technology, Fisheries and Marine Institute of Memorial University of Newfoundland, 4 Fisheries and Oceans Canada, Québec Region 5 Institute of Zoology, Zoological Society of London
The use of seafloor imagery to study deep-sea benthic environments has increased considerably in the past decade. Methods and tools used to perform imagery annotation (i.e., extracting data from the imagery) have also evolved and depend on study objectives and available equipment. Here, we review seafloor imagery annotation protocols and methods used in deep-sea benthic diversity studies. Our objective is to produce a centralized resource for researchers working in the field. The literature review consisted of searching Google Scholar, Web of Science, and Scopus for 12 keyword combinations that include “benthic”, “camera”, “deep sea”, “diversity”, “epifauna”, “imagery”, “fauna”, “field of view”, and “megafauna”. We have restricted our search to the years 2000-2023 and to the first 40 publications for each of the keyword combinations. Manuscripts were imported into the reference manager software Mendeley. We then scanned the list obtained from the literature search and used the authors’ expert knowledge to complete relevant gaps in the results and to remove papers that fell outside of the review’s scope (e.g., SCUBA surveys). The literature is currently being scanned for >20 parameters, including: 1) whether an annotation software was used (if so, which one), 2) whether images or videos were used, 3) how taxa were identified (i.e., is it specified, are there specific ID guides?), 4) presence/absence of lasers, 5) field of view calculation method, 6) sampling unit selected (i.e., temporal vs. spatial), 7) whether classification schemes were used, 8) whether both biota and substrate were annotated, etc. As expected, our preliminary observations indicate an array of different methods and terms used, which is particularly challenging for new projects/programs. Our review will provide a summary of the different methods and strategies used in the field and hopefully better inform decision-making related to seafloor imagery annotation projects.
Advancing Marine Imaging Technologies for Sustainable Shellfish Distribution in the Atlantic Waters
Isa ELEGBEDE , Lasu
The sustainable management of shellfish populations in the Atlantic waters is crufor both ecological balance and economic viability. This research aims to leverage advanced marine imaging technologies to map and monitor the distribution of shellfish species, providing a comprehensive understanding of their spatial dynamics and habitat preferences. Utilizing high-resolution underwater cameras, autonomous underwater vehicles (AUVs), and satellite imagery, we will conduct extensive surveys across diverse Atlantic marine environments. The collected data will be analyzed using machine learning algorithms to identify patterns and predict potential changes in shellfish distribution due to environmental factors and human activities. By integrating this innovative approach with traditional ecological methods, our study seeks to develop a robust framework for the sustainable management of shellfish resources, ensuring the preservation of marine biodiversity and supporting the livelihoods of coastal communities dependent on shellfish fisheries. This research will also contribute to the broader understanding of marine ecosystems and the impacts of climate change on marine life.
Aquascope: automated imaging and machine learning to understand and predict plankton community dynamics in lakes
Francesco Pomati , Eawag: Swiss Federal Institute of Water Science and Technology, Aquatic Ecology
Ewa Merz (Eawag & UCSD), Pinelopi Ntetsika (Eawag), Marta Reyes (Eawag), Stefanie Merkli (Eawag), Luis Gilarranz (Eawag), Stuart Dennis (Eawag), Marco Baity-Jesi (Eawag)
Plankton communities are highly diverse and dynamic and influence important ecosystem properties at the local and global scale. They are sensitive indicators of ecosystem health and global change's effects on ecosystem services. Algal blooms, for example, are an emergent property of a complex interaction network of producers and consumers, often triggered by climate change and eutrophication. Traditional plankton monitoring based on sampling and microscopy fails to capture high-frequency dynamics and complexity in plankton communities. Here we present automated in-situ tracking and modelling of plankton interaction networks in lakes based on automated underwater microscopy, machine learning and time-series analysis. The imaging approach is embedded into a floating platform, comprising sensors for profiling water column physics and chemistry, and meteorological information. Sensor data are integrated with lake hydrodynamic models, and the whole platform is suitable for stable long-term deployments for research and water quality monitoring. A dual magnification dark-field microscope and associated image processing and classification provide real-time, high-frequency plankton data from ~10 μm to ~ 1 cm, covering virtually all the components of the planktonic food web and their morphological diversity. We find that vision transformers, with targeted augmentation, constitute the most robust models for high-accuracy plankton classification and low sensitivity to temporal dataset shifts. Comparing imaging data from the field to traditional sampling and microscopy revealed a general overall agreement in plankton diversity and abundances, despite some limitations in detection (e.g. small phytoplankton) and classification (e.g. rare and morphologically similar taxa). Using high-frequency time-series of plankton abundances, or their daily growth rates, environmental conditions and state-space reconstruction methods for modelling chaotic dynamics, our data allow us to: map interactions between taxa, reverse engineer mechanisms that drive algal blooms or biodiversity change, and forecast ecosystem properties.
Combining in situ imagery with acoustic surveys for a comprehensive view of pelagic ecosystems
Ben Grassian , Woods Hole Oceanographic Institute
Heidi Sosik, Andone Lavery, Megan Ferguson, Sidney Batchelder, E. Taylor Crockford (all WHOI)
Ocean midwaters are the largest habitat on Earth but are among the least understood marine environments due to sampling challenges. This project focuses on combining concurrently collected deep towed imagery and shipboard acoustic measurements in the epi and mesopelagic environment to interpret zooplankton and nekton distributions within the detailed environment and scattering from mixed assemblages. We have trained a machine learning detector model for the automated detection of 13 zooplankton functional groups from a towed shadowgraph imaging system. The zooplankton detection model currently achieves 81% F1 scores on our validation image set. We present results of taxa-specific zooplankton distributions derived from the image data and examine their distributions with respect to scattering layers detected at different frequencies within the upper 1000m of a pelagic ecosystem in the slope waters offshore of the New England Shelf. We derived zooplankton biometrics from the image data useful for feature-based (e.g. size) discriminations and to generate forward scattering predictions. We describe animal zonation within the environment and varying diel habits across species and size classes from the acoustic, optical, and environmental data. Collectively this work will further enable and automate the use of optical techniques for midwater surveys and the interpretation of acoustic scattering returns from mixed zooplankton populations.
Deep sea imaging considerations for geometric and color consistent environmental monitoring
Dennis Giaya , Northeastern University
Aniket Gupta (1), Pushyami Kaveti (1), Jasen Levoy (1), Victoria Preston (1), Zhiyong Zhang (1), Dan Fornari (2), Hanumant Singh (1) (1): Northeastern University (2): Woods Hole Oceanographic Institution
Deep-sea marine ecosystems are increasingly impacted by environmental stressors, including rising sea temperatures, plastic pollution, and habitat erosion. To quantify bulk and rate changes of an environment from a historical baseline, we must preserve a snapshot of the 3D structure and color for spatiotemporal analysis. Modern reconstruction methods, including state-of-the-art structure-from-motion techniques, neural radiance fields, and Gaussian Splatting, are generally benchmarked against in-air scenes. We demonstrate that for underwater scenes, these methods are insufficient for geometric- and color-consistent reconstructions due to outstanding challenges in robust in situ camera calibration and lighting pattern recovery. Extracting quantitative spatial data from imagery requires precise camera calibration to ensure accurate geometric and color measurements. We focus on calibration challenges specific to deep-sea underwater imaging involving practical integration of cameras in depth-capable housings and lighting fixtures onto imaging platforms. First, we show how intrinsic and extrinsic camera calibration parameters can change over the course of field missions, using calibration imagery acquired in a tank as well as benthic seafloor imagery from a stereo camera pair. Intrinsic calibration parameters can change between platform deployments whenever cameras need to be serviced within underwater housings. Extrinsic calibration parameters may change over the course of a single deployment due to flexing of the underwater platform, requiring online calibration methods to compensate. Second, using hydrothermal vent imagery acquired at the northern East Pacific Rise, we show how color rendering is impacted by different lighting patterns and how uncertainty in lighting can lead to color inconsistency across deployments. Geometric- and color-consistency challenges are critical areas for hardware, data-processing, and methodological innovations to translate imagery into scientifically viable data products. This is especially pressing as novel scientific expeditions aim to track both natural and man-made changes. From our analyses, we suggest several best practices to consider for image collection, processing, and curation.
Deep sea invertebrate bodies viewed through Microcomputed Tomography (microCT)
Freya Goetz , Smithsonian Institution
Marine invertebrates exhibit broad diversity in anatomy and adaptations to their environment, especially in the midwater and deep-sea floor. Darkness and the viscosity of water foster physical adaptations not found elsewhere. This has resulted in vast soft body variations ideal for explorations that could inspire new soft robotic designs. Microcomputed Tomography (microCT) is an exciting tool for studying these adaptations, especially now that microCT scanners are becoming more accessible and resolution capabilities are improving. Animals best suited for these analyses range in size from roughly similar to a 5 gallon bucket down to a few millimeters. Resolution depends on animal size, where the smallest animals can yield voxel sizes around 1.6 micrometers, approximately the thickness of a red blood cell. In practice, we cannot see individual cells but can differentiate between tissue types like muscle and nerve. Data can be generated exponentially faster than with traditional histology and without distorting anatomical structures. By staining with compounds containing metals such as tungsten, osmium and iodine, soft tissue contrast can be enhanced to visualize internal anatomy. This means that animals can be viewed in 3D without dissection, maintaining the most life-like state. The resulting dataset of each scan is a stack of grayscale image slices which one can virtually fly through the animal along all three axes. Data analysis possibilities are vast and open to questions involving physiology, 3D modeling and anything structural that can be visualized or measured. I will highlight practical approaches of specimen preparation, acquisition, data analysis and data management along with examples of various study systems, including leech internal anatomy and hyperiid amphipod visual systems.
Developing a Portable, Cost-Effective Multimodal Imaging System for Underwater Mapping and Monitoring with Remotely Operated Vehicles
Giancarlo Troni , Monterey Bay Aquarium Research Institute (MBARI)
Sebastián Rodríguez-Martínez, MBARI Eric Martin, MBARI Paul Roberts, MBARI Kevin Barnard, MBARI Bastian Muñoz
Underwater exploration is crucial in advancing our understanding of marine environments and ecosystems. Employing multimodal imaging sensors enables us to conduct comprehensive studies with multidimensional capabilities. This work presents the ongoing developments of the CoMPAS Lab at the Monterey Bay Aquarium Research Institute (MBARI). Our focus lies in utilizing cost-effective sensors to improve perception, particularly in seafloor mapping and navigation with remotely operated vehicles (ROVs) and autonomous underwater vehicles (AUVs). A multimodal sensing platform has been developed for benthic surveys, incorporating downward-looking stereo cameras, a forward-looking sonar (FLS) for acoustic imaging, a push-broom hyperspectral camera, a laser-sheet projector, and a suite of low-cost navigation sensors. Each component provides an additional dimension to our understanding of the underwater eenviorment. Leveraging deep learning techniques, we are improving our perception based on sensors such as FLSs and laser scanners to enhance the quality and resolution of seafloor mapping data and enable long-term monitoring of those areas. The hyperspectral system shows promise in advancing our understanding of underwater ecosystems through enhanced imaging capabilities with more than one hundred short-bandwidth images per experiment. Additionally, the stereo cameras enable the creation of photomosaics and point cloud representations of the seabed while improving navigation through a simultaneous localization and mapping (SLAM) framework. The performance and capabilities of these systems have been validated through multiple deployments with platforms such as the MiniROV from MBARI, as well as in the exploration of the Salas and Gomez Ridge, Chile, with the Schmidt Ocean Institute’s ROV, SuBastian. Through this work, we aim to validate and demonstrate the potential for scalable approaches using cost-effective sensors, setting the foundations for their integration in scalable autonomous platforms for widespread underwater mapping and monitoring, contributing to a greater understanding of marine ecosystems, and fostering advancements in oceanographic research and conservation.
Economies of Scale: The GEOMAR Seafloor Photogrammetry Workflow
Tom Kwasnitschka , GEOMAR Helmholtz Centre for Ocean Research Kiel
Armin Bernstetter, Jacob Eichler, Marco Rohleder, Oliver Jahns, Malte Eggersglüss, Carl von Brandis, Markus Schlüter, Jan Fleer - all affiliated at GEOMAR Helmholtz Centre for Ocean Research Kiel
Photogrammetry is a transformative methodology in seafloor research both in terms of its potential for situational awareness, and for the quantitative assessment of otherwise inacessible outcrops and habitats. Recent work has been focused on the economies of scale connected to particularly large, synoptic and holistic surveys of study areas covering up to several acres, and the challenges connected to such work. We will report on recent developments at GEOMAR in this field, including the finalizing of a suite of immersive fisheye cameras, metrologic water-compensating survey cameras, and a corresponding 48-channel LED video strobe lighting system. We will elaborate on the construction and first field trials of an all-new dedicated underwater vehicle to carry this sensor suite. We will show progress on immersive virtual fieldwork and quantitative measurements on photogrammetric models using the Unreal Game Engine in our ARENA2 visualization lab. Lastly, we will give an overview on the progress applying this methodology to a host of ongoing marine geoscientific mapping projects.
Enhancing Pelagic Ecosystem Research with a Standardized Image Preprocessing Software Tool
Mojtaba Masoudi , Ocean BioGeosciences, National Oceanography Centre, Southampton, UK
Eric Orenstein1, Loic Van Audenhaege1, Tobias Ferreira 2, Colin Sauze2, Jennifer Durden1 1 Ocean BioGeosciences, National Oceanography Centre, Southampton, UK 2 Ocean Infomatics, National Oceanography Centre, Liverpool, UK
When studying pelagic ecosystems, marine scientists often grapple with the challenge of managing and analyzing large volumes of imagery. These datasets are collected in diverse habitats and, often, from several imaging systems. Each of these sampling regions and instruments may require unique preprocessing protocols to effectively prepare the data for downstream ecological analyses. Practitioners are often left to piece these pipelines together from previous projects and disparate software tools, often resulting in disjointed and poorly document workflows. Unlike benthic imaging, which often deals with more static environments, pelagic imaging must contend with the dynamic nature of open water habitats, where rapidly moving organisms and fluctuating light conditions complicate data capture and analysis. This can lead to inconsistencies, inefficiencies, potential errors in ecological data analysis, and an inability to compare biodiversity indicators between studies. To address these issues, we have developed a suite of software tools designed to standardize and streamline the selection of image preprocessing steps crucial for pelagic research. Our toolkit, enables researchers to effortlessly automate crucial tasks like adjusting brightness variations, detecting edges, and more. Furthermore, its detailed documentation helps users grasp the range of preprocessing choices available, letting them make well-informed decisions that suit their research requirements. By ensuring consistent preprocessing, the toolkit provides robust support for addressing a range of ecological inquiries in midwater, enabling more reliable ecological computations, and supporting reproducible biodiversity assessments. This unified approach is expected to enhance the efficiency and reliability of data analyses in marine ecosystem studies.
Enhancing Situational Awareness in Underwater Robotics with Real-Time Multi-Camera Mapping
Pushyami Kaveti , Norwegian University of Science and Technology (NTNU) and Northeastern University, Boston
Ambjørn Grimsrud Waldum (NTNU), Oscar Pizarro (NTNU), Hanumant Singh (NU), Martin Ludvigsen (NTNU)
Technological advances in robotics, particularly Autonomous Underwater Vehicles (AUVs) and Remotely Operated Vehicles (ROVs), have revolutionized deep ocean exploration, enabling access to unseen, inaccessible, and hazardous environments. Robust and reliable state estimation and situational awareness are crucial for both, autonomous and remotely piloted operations, facilitating informed decision-making in real-time. Visual imaging systems are integral to this advancement, with AUVs and ROVs typically equipped with multiple cameras. Typically, ROV pilots must manage multiple camera feeds to make operational decisions, which can be cumbersome, lead to fatigue, and require substantial training. Integrating vision-based navigation and mapping into these platforms enhances operator support and enables various downstream tasks, including online planning, virtual environment creation, scene comprehension, and potential semantic analysis. It also significantly impacts broader marine applications such as habitat monitoring, environmental sampling, and understanding ocean ecosystems. However, automating visual state estimation in underwater environments presents challenges such as visual degradation due to backscatter, illumination variability, and featureless surfaces. Other challenges include lack of synchronization among cameras, difficulty in processing multiple streams rapidly due to computational constraints, and the variability and sometimes unknown nature of controllable camera parameters such as zoom, pan, and tilt. Existing visual mapping systems are often designed for monocular or stereo cameras and do not scale well for multi-camera systems. Our work addresses these needs with a comprehensive solution encompassing sensor synchronization and advanced algorithms. We effectively fuse multi-camera data within a Simultaneous Localization and Mapping (SLAM) framework for robust and accurate state estimation. We achieve real-time dense 3D mapping of underwater environments by leveraging modern reconstruction and rendering techniques. We present mapping results from ROV operations for subsea infrastructure inspection, showcasing the sensor setup, collected data, and generated dense maps, highlighting the practical significance and the role of real-time mapping capabilities for improving marine exploration and situational awareness.
FathomVerse: Gaming for ocean exploration
Kakani Katija , Bioinspiration Lab, Research and Development, MBARI
Lilli Carlsen (MBARI), Emily Clark (MBARI), Giovanna Sainz (MBARI), Joost Daniels (MBARI), Kevin Barnard (MBARI), Ellemieke Berings (&ranj Serious Games), Meggy Pepelanova (&ranj Serious Games), GAF van Baalen (&ranj Serious Games)
In order to fully explore our ocean and effectively steward the life that lives there, we need to increase our capacity for biological observations; massive disparities in effort between visual data collection and annotation make it prohibitively challenging to process this information. State-of-the-art approaches in automation and machine learning cannot solve this problem alone, and we must aggressively build an integrated community of educators, taxonomists, scientists, and enthusiasts to enable effective collaboration between humans and AI. FathomVerse, a mobile game designed to inspire a new wave of ocean explorers, teaches casual gamers about ocean life while improving machine learning models and expanding annotated datasets (FathomNet). Of the three billion gamers worldwide, up to 70% say they care about the environment, and FathomVerse taps into this engaged community with innovative gameplay and rich graphics that draw players into the captivating world of underwater imagery and cutting-edge ocean science. Here we will share our process of designing FathomVerse, discuss early successes from the v1 launch on May 1, 2024, and highlight some areas of future focus. In less than one month after v1 launch, 8k players from 100 different countries generated >3M annotations that are currently being used to generate consensus labels and retrain machine learning models. Through FathomVerse, we hope to activate global audiences in high school and up, with the goal of increasing public awareness and inspiring empathy for ocean life.
Ghana Marine Species Diversity App
Peter Teye Busumprah , Ministry of Fisheries and Aquaculture Development
This oral presentation addresses the UN Ocean Decade Challenges 8.9 &10 which focuses on digital representation of the Ocean, create skills, knowledge and technology for all and change human relationships with the Ocean. It focuses on SDGs 9 & 14 which is centered o industry, innovation and infrastructure and life below water In Ghana, the fisheries sector is a crucial contributor to the national economy, but the country's fish stocks are facing declining trends due to overfishing and degradation. Accurate monitoring and analysis of fish species are essential to inform sustainable fishing practices and conservation efforts. However, traditional methods of data collection are often limited by inadequate resources and expertise. To address this challenge, we developed "Sea Rock Base App", a mobile-based citizen science app designed specifically for local fishermen in Ghana The app enables fishermen to collect and submit data on fish species they encounter during their daily activities. The app includes a user-friendly interface for recording species information, photographs, and geolocation data. The submitted data is then analyzed using machine learning algorithms to identify patterns and trends in fish species distribution and abundance. The app also provides real-time feedback to fishermen on the most common species found in their fishing grounds, allowing them to make informed decisions about their fishing activities. Pilot testing of the app in four coastal communities in Ghana revealed high user adoption rates and accurate data collection. The app has also led to the identification of new fish species and areas of high conservation concern. By empowering local fishermen with data analysis tools, Sea Rock Base App aims to promote sustainable fishing practices, improve fisheries management, and support the conservation of Ghana's marine biodiversity. This citizen science approach has the potential to be replicated in other fisheries contexts worldwide, contributing to the global effort to safeguard marine ecosystems and ensure food security for future generations.
Harnessing 3D Geometrical Features for Automated Coral Habitat Classification
Larissa M. C de Oliveira , 1. Earth & Ocean Lab, Department of Geography, University College Cork, Ireland 2. Environmental Research Institute (ERI), University College Cork, Ireland
Patricia Schontag 3, Judith Fischer 3, Timm Schoening 3 3. Data Science Unit (DSU), GEOMAR Helmholtz Centre for Ocean Research Kiel, Germany
Cold-water coral reefs are among the most structurally complex marine ecosystems, supporting vast biodiversity. Recent advancements in photogrammetry, such as Structure-from-Motion (SfM) and gaussian splatting, have facilitated the creation of high-resolution 3D models of these environments. However, integrating 3D structural complexity into artificial intelligence (AI)-based classification methods remains an emerging area of research. While established 3D classification methodologies have proven effective in terrestrial applications, their potential for underwater contexts has yet to be fully explored. Measures of 3D complexity provide ecological indicators for coral reef assessments. But their use beyond the ecological landscape remains uncertain. This study investigates the application of geometrical features to enhance the automated classification of 3D reconstructions of coral reefs. By computing geometrical features at multiple scales, seabed details are captured at varying spatial resolutions, allowing for a data-driven optimisation of 3D scene classification parameters. We developed four feature sets incorporating geometric features derived from point cloud covariance matrices and primary features such as 3D coordinates and colour. These feature sets were passed through a gradient-boosted tree model to assess performance and feature importances. Our results indicate that primary features alone are insufficient for accurate classification, whereas incorporating geometrical features significantly enhances performance. Furthermore, the combination of geometrical features with colour information provided only marginal improvements. These findings underscore the robustness of geometrical features for precise 3D classification and habitat mapping in complex 3D environments like coral reefs.
High-resolution bathymetry coupled with 3D models of hydrothermal vents at the Aurora vent field - Gakkel ridge, Arctic Ocean
Tea Isler , Alfred Wegener Institute Helmholtz-Center for Polar and Marine Research Bremerhaven
Christopher R. German [Woods Hole Oceanographic Institution], Vera Schlindwein [Alfred Wegener Institute Helmholtz-Center for Polar and Marine Research Bremerhaven; University of Bremen], Tom Kwasnitschka [GEOMAR Helmholtz Centre for Ocean Research Kiel], Michael Jakuba [Woods Hole Oceanographic Institution]
At the slowest-spreading ridge on earth, the Gakkel ridge, evidence of hydrothermal activity was first observed in 2001 and later confirmed in 2014 by video recordings using the deep-tow camera survey system OFOS (Ocean Floor Observing System). This provided the evidence for submarine venting under ice-covered oceans, an analog for possible habitability on ice-covered ocean worlds such as Saturn's moon Enceladus or Jupiter's moon Europa. In July 2023, the AUV/ROV Nereid Under Ice (NUI) was deployed from the RV Polarstern and discovered seven new submarine vents at the Aurora hydrothermal field. Bathymetry and optical data acquired during the dives allowed for a deeper understanding of this topographically complex environment. The dives primarily targeted the sampling of rocks and fluids, with cameras mainly used for navigation and the identification of new vents and species. From the bathymetry map we are now able to identify morphological features at a higher resolution than was previously possible, while the opportunistic video footage is used to create a scaled 3D model of the newly discovered hydrothermal vents using structure-from-motion techniques. This approach allows for further investigations (e.g., habitat mapping) which would otherwise not be possible using classic ship based multibeam technologies. We present a new high-resolution bathymetric map of the Aurora vent field which, coupled with 3D models, provides a deeper understanding of the morphologically complex study area. This study highlights the usefulness of opportunistic data, especially when surveying in extreme environments, where data collection requires time consuming operations, expensive devices and experienced operators. The final product shows the results achieved during the most recent expedition to the Aurora vent field and also gaps for areas that are yet to be explored.
High-resolution time series of plankton and carbon from underwater imaging
Klas Ove Möller , Institute of Carbon Cycles, Helmholtz-Zentrum Hereon
Ankita Vaswani, Katharina Kordubel, Daniel Blandfort, Götz Flöser (all Institute of Carbon Cycles, Helmholtz-Zentrum Hereon) and Saskia Rühl (Plymouth Marine Laboratory)
High-resolution snapshots of plankton biodiversity, behavior and carbon flux from an underwater imaging time series: Plankton, organic particles and aggregates ("marine snow") form the basis of marine ecosystems and play a fundamental role in the food webs and the biological carbon cycle of the ocean. For these reasons, it is necessary to understand the processes that influence the spatial and temporal distribution as well as the origin, characteristics and biodiversity of these organisms and particles. However, these processes are still insufficiently resolved and quantified, since marine ecosystems are characterized by immense variability and traditional measurement methods are limited in their resolution. To overcome this limitation we deployed an underwater high-resolution imaging observatory in the North Sea to investigate the dynamics and characterization of plankton and particles and to quantify the balance between biological productivity and carbon export at an unprecedented resolution (from seconds to months). To unlock the full potential of our observations, we developed new AI based image analysis pipelines that go beyond classifying organisms by taxonomic groups, but also quantify functional traits and biological phenomena from images. We here present highlights and snaphots of our unique high-resolution time series, including a one year time series of carbon flux in a coastal environment and the underlying drivers, the biodiversity, seasonal variation and vertical migration patterns of plankton organisms, and the increase and impact of Noctiluca on the food web dynamics. These high-resolution observations of the interactions between physics, biology and geochemistry at unprecedented spatial and temporal scales contribute significantly to our understanding of the productivity, resilience and carbon storage capacity of marine ecosystems as a function of climatic changes (rise in temperature, acidification, oxygen minimum zones) and anthropogenic influences and their future developments (nutrients, economic use, loss of biodiversity).
Imaging and taxonomic identification of meiofauna: progress on the BLUEREVOLUTION project
Catherine Borremans , IFREMER (Biology and Ecology of dEEP marine ecosystems Unit (BEEP)/Deep Sea Lab)
Daniela Zeppilli (IFREMER), Abdesslam Benzinou (ENIB), Kamal Nasreddine (ENIB), Valentin Foulon (ENIB), Anthonin Martinel (ENIB), Edwin Dache (IFREMER)
The Earth’s Ocean is the largest three-dimensional living space on our Planet. It is crucial to life as we know it, yet we know less about the entire seafloor than we do of the surface of the moon. Infaunal benthic communities (meiofauna) comprise some of the most diverse groups of organisms on Earth, and only a very small amount of this diversity has been described by science. Despite our partial exploration of this vast domain, all marine habitats, including the deepest trenches, experience direct or indirect human impacts. Thus, the scientific community faces the need for fast and accurate monitoring studies and building baseline datasets to measure future changes and biodiversity losses. In this context the “Biodiversity underestimation in our bLUe planEt: artificial intelligence (AI) REVOLUTION in benthic taxonomy” (BLUEREVOLUTION) project aims at developing in- and ex-situ methods using automated or semi-automated imaging method on meiofauna linked with taxonomic classification tools (AI) for high throughput analysis of benthic diversity. Different imaging techniques -from low to high resolution- were assessed (e.g. flow imaging, focus stacking brightfield microscopy, 3D-fluorescence imaging) and developed (‘MeioCam’) to identify the best configuration(s) according to resolution (capacity to taxonomic identification), throughput, quantitative power, data volume (Foulon et al., 2024). In addition, computer vision-based methods were investigated to support the need of faster recognition and morphometric measurement of meiofauna (e.g. nematods), as well as to compensate for the lack of available imaging dataset to train artificial intelligence models. Approaches based on deep learning and GAN are proposed in this work. The presentation will give an updated overview of the developed imaging pipeline that will ultimately allow for quantitative and functional data of benthic communities being generated at speeds unseen before. It will produce a standardized method for building open-access reference databases together with fast and reliable tools for impact assessments and biodiversity surveys. Ref : Foulon, V., et al. 2024. Taxonomic identification of meiofauna through imaging – a game of compromise. Submitted in Limnology and Oceanography: Methods
Improving Video Annotation Quality Assurance and Quality Control
Ashley N. Marranzino , University Corporation for Atmospheric Research | NOAA Ocean Exploration Affiliate
Sarah Groves ( University Corporation for Atmospheric Research | NOAA Ocean Exploration Affiliate), Samuel Candio (NOAA Ocean Exploration), Vanessa Steward (Ocean Networks Canada)
The increasing volume of video data on deepwater habitats provides scientists with critical information needed to close gaps in our understanding of these poorly understood environments. However, analysis of these videos requires appropriate annotations of features of interest, with the quality of the data analysis dependent on the quality of video annotations. NOAA Ocean Exploration is dedicated to exploring the unknown oceans and routinely collects video data using remotely operated vehicles (ROVs) during expeditions aboard NOAA Ship Okeanos Explorer to provide scientists with a view of unexplored deepwater habitats. All video collected during these expeditions is archived with the National Centers for Environmental Information and made publicly accessible through the Ocean Exploration Video Portal. NOAA Ocean Exploration also partners with Ocean Networks Canada (ONC), using the cloud-based annotation platform SeaTube V3 to host video and ROV sensor data from expeditions. NOAA Ocean Exploration and its partners use SeaTube V3 to annotate ROV videos for biological organisms, geological features, engineering events, and other features of interest during dives and post-expedition. Annotations can also be used to flag content for education and public outreach purposes. While the collaborative features of SeaTube V3 can allow for a more thorough and accurate annotation of video data, it also raises the questions of the level of accuracy and completeness of annotations. NOAA Ocean Exploration and ONC have been working together to develop and test sets of features to help determine the robustness of a set of annotations and develop workflows for the analysis and archival of annotations when they are deemed “complete”. This presentation will highlight some of these features along with next steps the programs plan to take.
Improving marine imaging development by acknowledging contrasting values of data
Jennifer M. Durden , National Oceanography Centre
Key recurring themes in many Marine Imaging Workshop discussion sessions are the desire to share and align imagery data. Considerable effort is being put towards enabling practicalities, such as hosting and transferring large volumes of imagery, and developing metadata standards, yet a lack of discussions about our contrasting values towards imagery data across the marine imaging community hold us back from realising the desired collaboration. The ways that we value data influence how, with whom and why we share and collaborate, and what we expect from those actions. Marine imaging is a multidisciplinary community, made up of engineers, software developers, scientists, data managers and other users, operating in research, government and industry; each of these groups values data (and metadata) differently. For example, a group that spent significant effort generating human-derived annotations may value those image-derived data differently than a software developer looking to train an algorithm. Other drivers of the ways that we value data include research funders, relationships with industry, contractual obligations and academic publishing requirements. I explore the different ways that data are valued across members of the marine imaging community and where conflicts arise that prevent us from achieving the desired sharing and data alignment and governance, and suggest ways to surmount these issues.
Interactive machine learning: reflections on how a human-in-the-loop approach can increase model training efficiency.
Dr Laurence H. De Clippele , School of Biodiversity, One Health & Veterinary Medicine, University of Glasgow
De Clippele L.H., Dominguez Carrió C., Vad J., Vlaar T., Balsa J., Beckmann L.M., Boolukos C., Burgues I., Clark H.P., Cook E., de la Torriente Diez A., Easton. B., Ferreira T. C., Guzzi A., Harris L., Kaufmann M., Mendoza J.C., Riley T.G., Villalobos V.I.
Due to technical advances in imaging technologies and cost-effective data storage solutions, there is a global increase in the amount of underwater image data. Since it is too time-consuming to analyse this data manually, marine scientists are increasingly interested in using convolutional neural networks to automate their annotations. However, deep learning is considered to be a “black box” because it is difficult for the user to understand how deep neural networks make their decisions. In addition, if the user lacks an understanding of the machine learning fundamentals, it is unlikely that they will be able to develop models that are effective and produced in a time-efficient manner. Interactive, “human-in-the-loop”, machine learning is here suggested as a novel approach to increase the user’s understanding of fundamental machine learning concepts, increase model training efficiency and accuracy drastically, and make deep learning less of a black box. In June 2024, a Marine Animal Forest COST action training school was organised at the University of Glasgow. This training school brought together early career researchers from across the UK and European institutions. Through training models on a variety of Remotely Operated Vehicle and timelapse image datasets, reflections were gathered on how interactive machine learning can contribute to increasing the efficiency of model training. The main conclusions were regarding the importance of user expertise of the target object (i.e. the species), the complexity of the habitat, the quality of the data, the abundance of the species, and the psychological effect of observing a change in model performance. Other interesting observations were made regarding a lack of model improvement when trying to “recover” from mistakes during the training process and dealing with the uncertainty when creating a training dataset. A practical workflow is proposed to encourage more efficient model training decisions by users new to machine learning.
Inverse Physically Faithful Underwater Imaging
David Nakath , Kiel University and GEOMAR Helmholtz Centre for Ocean Research Kiel
Kevin Köser (Kiel University and GEOMAR Helmholtz Centre for Ocean Research Kiel)
All work presented at Marine Imaging Workshop depends in some way on underwater imagery which is governed by a unique image formation model. Its complexity often makes automatic processing a very complex task. Hence, knowing the basics is greatly beneficial for practitioners defining downstream tasks on underwater data. This talk will detail geometric as well as radiometric distortions, due to cameras operating directly water. The former emerge when a light ray passes interfaces between media with different optical densities, specifically: air--glass--water. While the latter are caused by attenuation and scattering effects inside the medium itself. Furthermore, homogeneous illumination by the Sun, inhomogeneous artificial illumination, or even a mixture of both contribute another dimension of complexity to be explored. Physically based rendering approaches are particularly well suited to tackle problems in underwater vision, due to the dualism between models originally devised in physical oceanography and the medium models which are nowadays typically employed in physically-based raytracing. This enables us to capture underwater imagery and simultaneously measure optical properties of the water using an established sensor suite. We can then directly synthesize images with the same medium-properties and verify our rendering-systems. This enables us to provide reliable synthetic image data focusing on specific problems to train, develop, and test algorithms with. With the advent of massively parallel computations on GPUs in conjunction with inverse physically-based rendering methods it, vice versa, became possible to infer inherent optical properties of the water body directly from images using an analysis-by-synthesis approach. The same approach can be applied to flat ports, dome ports as well as light sources. Being able to calibrate and simulate refraction, light and the optical properties of the water directly enables a wide range of applications such as image restoration, shadow removal and light removal directly on submerged 3D models.
Learning Ultra-high-throughput Fluorescence Microscopy-in-flow for Marine Phytoplankton Observation
Jianping Li, Claude , 1.Shenzhen Institute of Advanced Technology, CAS, Shenzhen, China; 2.University of CAS, Beijing, China.
Zhisheng Zhou 1,2; Zhenping Li 2,1. 1.Shenzhen Institute of Advanced Technology, CAS, Shenzhen, China; 2.University of CAS, Beijing, China.
The development of automated in-situ analysis technologies of photosynthetically active phytoplankton cells and colonies in natural seawater is of great significance for biological oceanography and marine ecology monitoring. However, the composition of natural seawater is highly complex. The size range of phytoplankton spans at least 3 orders of magnitude, from single cells <1μm to large diatoms or colonies >500μm. In addition, seawater also contains countless non-phytoplankton particles. These facts present enormous challenges in terms of specificity, sensitivity, and spatial resolution for imaging flow cytometers (IFC) such as CytoSense and IFCB to observe phytoplankton in situ. Obviously, ocean observation prefers high-throughput methods in analyzing more seawater within less time to extract more realistic phytoplankton information. Since most phytoplankton are tiny, IFCs usually adopt slow flow with high magnification to obtain sufficient resolution for imaging phytoplankton. However, to enhance imaging throughput, IFCs should use higher flow rates with lower magnifications, though this may be very likely at the cost of imaging resolution and quality sacrifice, to gain increased seawater sampling capability. The compromise between imaging resolution and observation accuracy of current IFCs essentially limits their net throughput. To balance this trade-off, we are trying to combine "low-magnification imaging plus computational super-resolution" in a light-sheet fluorescence microscopy-in-flow system named FluoSieve. By building up a large-scale phytoplankton fluorescence image dataset, we are training a super-resolution CNN network called IfPhytoSR (in-flow phytoplankton super-resolution). The preliminary results indicate that the IfPhytoSR model can restore the poorer resolution and quality images acquired by a 5× objective lens into much better counterparts as if were acquired by the FluoSieve system equipped with a 20× lens for downstream task such as recognition or measurement, theoretically achieving a ~40-fold imaging throughput. In this talk, we will report on the progress of this research.
Marimba: Efficient Marine Image Data Processing with FAIR Principles
Chris Jackett , CSIRO
Chris Jackett - CSIRO Environment Kevin Barnard - MBARI Nick Mortimer - CSIRO Environment David Webb - CSIRO NCMI Aaron Tyndall - CSIRO NCMI Franzis Althaus - CSIRO Environment Bec Gorton - CSIRO Environment Ben Scoulding - CSIRO Environment
Advancements in underwater imaging technologies have greatly enhanced marine exploration and monitoring, resulting in substantial volumes of image data. Efficient management, processing, standardization, and dissemination of this data remain challenging. Marimba is a Python framework co-developed by CSIRO and MBARI that addresses these challenges by facilitating the FAIR (Findable, Accessible, Interoperable, and Reusable) processing of marine image datasets. Marimba is built around three core constructs: Project, Pipeline, and Collection. A Project encompasses the entire data processing workflow; Pipelines provide isolated environments for each data processing stage, encapsulating all necessary logic and operations; Collections aggregate groups of data for isolated processing. Marimba offers a robust CLI and API for flexible user interaction and scripting in scientific imaging projects. It supports project structuring, file and metadata management, integrates with the iFDO standard, and provides a standard library to manage common image and video processing tasks. Dataset packaging capabilities include comprehensive processing logs that capture the full dataset processing provenance, file manifest generation, and dataset statistical summaries, with support for distribution to S3 storage buckets. Managing the entire lifecycle of marine image datasets, Marimba upholds FAIR principles, transforming raw data into structured, usable, and shareable assets for marine environmental research.
Marine snow particle imaging resolves the mechanism and magnitude of coastal carbon export
Colleen Durkin , MBARI
Danelle Cline (MBARI), Duane Edgington (MBARI), Sachithma Edirisinghe (U. Maine), Margaret Estapa (U. Maine), Nils Haentjens (U. Maine), Christine Huffard (MBARI), Paul Roberts (MBARI), Henry Ruhl (MBARI), Fernanda Lecaros Saavedra (MBARI), Melissa Omand (URI), Sebastian Sudek (MBARI)
Ocean waters are filled with marine snow: particles made of fecal pellets, aggregates, and other detritus generated by marine life. These particles are the primary food source for many deep-sea ecosystems, and are a key mechanism for ocean carbon uptake and sequestration. Constraining the spatial, temporal, and compositional variability of sinking marine snow is a fundamental challenge for both ocean biogeochemistry and deep-sea biology. We image individually resolved particles collected in samplers and also by in situ instruments, to estimate the quantity of carbon export in the ocean and its ecological drivers. We apply methods in computer vision and machine learning to automate the classification of large imaging datasets into ecologically meaningful categories of marine snow. These particle classes can then be used to estimate carbon biomass and export. Coordinated observations made in Monterey Bay, California demonstrate how a suite of imaging systems can expand the scale of particle flux measurements while also resolving mechanisms driving its variability. We deployed traditional sediment traps to collect physical samples of sinking particles, and used chemical measurements of carbon flux to ground-truth estimates generated by particle imaging. These ground truth measurements were related to in situ imagers of particles, including camera systems attached to long-range autonomous underwater vehicles, remotely operated vehicles, neutrally buoyant floats, and moored to the seafloor. Together, these temporally sustained and/or spatially distributed observations will enable us to quantify carbon export in dynamic coastal environments where the ecological source and fate of sinking carbon is difficult to trace.
Object Detection for quantitative analysis of deep-sea benthic megafauna distribution: making the best of video transects
Nils Piechaud , Institute of Marine Research, Bergen, Norway
Heidi Meyer¹, Pål Buhl-Mortensen¹, Genoveva Gonzalez-Mirelis¹, Gjertrud Jensen¹, Yngve Klungseth Johansen¹, Anne Kari Sveistrup¹, Rebecca Ross¹ 1: Institute of Marine Research, Bergen, Norway
The MAREANO program maps habitats on the Norwegian seabed by analysing data from a large dataset of underwater imagery collected over many years. Counting individual objects and animals in videos is a common technique to sample the seabed and its automation (partial or complete) with Computer Vision (CV) and Machine Learning (ML) would greatly help quantitative benthic ecology. Yet, many practical problems must be solved before it can reliably generate data on species distribution and abundance. An important drawback of automated annotations is the difficulty for Object Detectors (OD) models to account for contextual information in the video where the behaviour of the object over multiple frames can inform its identification. We investigated how recent innovations and open source libraries could help address this challenge and enable an OD to quickly generate the data required for mapping. Seabed images were analysed manually in Biigle to get bounding boxes annotations which were used as a training set for various OD models (V8m, V8X, Fathomnet and Fathomnet’s Megalodon). Predictions from these models were used to count the number of individuals of selected taxa in videos which were compared to the manual counts. We used additional techniques (tracking, deflector classes, zonal counting) to refine and filter the predictions and maximize performances. F1 scores in training can reach 0.95 for target taxa. Automated Individual counts per videos can also be within 95% of the manual counts. Performances can, however, drop dramatically in challenging visibility conditions prompting the need to manually review and filter the predictions to ensure consistency and reliability. These results show that, in the right conditions, CV can perform well and be useful to enumerate target taxa. However, the need to manually review the results remains a bottleneck and we discuss the possible ways around it.
Ocean on a table top: A virtual reality arena for longterm migration/mapping of plankton behavior in current and future oceans
Manu Prakash , Stanford University
Gravity Machine team, Prakash Lab, Stanford University
Marine plankton exhibit a Diel Vertical Migration with vertical displacement scales from several tens to hundreds of meters. Even at the scale of small phytoplankton and zooplankton (100 μm to a few mm) the interaction of this vertical swimming behavior with hydrodynamics affects large scale distribution of populations in the ocean and is thus an important component of understanding ocean ecology. However, concurrently observing organismal physiology and behavior is challenging due to the vast separation of scales involved. Resolving physiological processes involves sub-cellular (micron) resolution while tracking freely swimming organisms implies vertical displacements of several meters. We present a simple solution to this problem in the form of a “hydrodynamic treadmill” incorporated into a table-top scale-free vertical tracking microscope. We use this method to study the behavior of freely swimming marine plankton, both in lab and on-board a research vessel, revealing a rich space of dynamic behavioral states in marine micro-organism. We will present data sets from five different expeditions from all major oceans of the world. This effort has culminated in one of the largest behavioral plankton dataset on record till today.
Online Learning of Appearance Models to Enable Real-Time Visual Tracking of Arbitrary Marine Animals
Levi Cai , MIT and WHOI
Nathan McGuire (WHOI), Roger Hanlon (MBL), T. Aran Mooney (WHOI), Yogesh Girdhar (WHOI)
Collection of video and co-located telemetry information about various marine animals is crucial for understanding their behaviors and even establishing conservation practices. Divers or tags are the two most commonly used approaches to gathering this type of data, but is mostly limited to surface dwelling species and difficult to scale. Autonomous underwater vehicles (AUVs) equipped with cameras and edge computers directly integrated into their control systems are showing increased promise for enabling collection of this type of data without the use of tags or divers. However, there is a dearth of well-labeled visual data accessible in the marine domain for many species. In this work, we propose the use of a class of visual algorithms known as semi-supervised learning and tracking approaches to estimate online appearance models of never-before-seen visual objects. These approaches require a scientist to initially provide a bounding box or visual template of an animal, and then tracking can be done autonomously. We demonstrate how to incorporate these models into AUV platforms to enable tracking of marine animals, in-situ and in real-time, for which labeled data may not exist. Additionally, this allows us to simultaneously gather video and co-located telemetry information about them. Furthermore, we discuss strategies for deploying these types of systems on a larger-scale and argue the importance of continued innovation and development of AUV platforms that are equipped with modern edge devices and stereo-vision that are deeply integrated into their control systems.
ROVs for Science & Communications: The Making of the Documentary “All Too Clear: Beneath the Surface of the Great Lakes”
Yvonne Drebert , Inspired Planet Productions & Boxfish Robotics
Zach Melnick, Inspired Planet Productions & Boxfish Robotics
Remotely-operated vehicles (ROVs) enable us to explore previously inaccessible underwater environments and observe animal behaviours in unprecedented ways, all while incurring zero animal mortalities. ROVs also facilitate the study of hard-to-access underwater archaeological sites, thereby enriching our cultural heritage. Recent advances in ROV and camera technology allow us to “swim” further, deeper, and longer than with older models, while viewing 4k results in real time from the surface. After more than 20 years creating documentary film and television, our team has taken a deep dive into the world of ROVs as tools for both filmmaking and research. For our most recent documentary series, “All Too Clear: Beneath the Surface of the Great Lakes,” we spent more than 150 days underwater with the Boxfish Robotics “Luna” ROV. The results are a compelling underwater tv series, as well as an archive of wildlife footage that has provided scientists with new insights into the spawning, feeding, and schooling behaviours of North American Great Lakes fishes, including the first-ever recording of lake whitefish spawning in the wild. As an added bonus, in June 2023, we discovered the wreck of the steamer "Africa," lost to the depths in 1895. The story received international media attention, further raising the profile of the freshwater species and environments under the waves of the Great Lakes. We’re excited to share the successes and challenges of this work, as well as our learnings around techniques, gear, post-processing, and other tools to optimize our chances of capturing animal behaviours. Footage example: https://vimeo.com/871465906/4caed4807a
Rate-A-Skate – Deep learning for individual recognition of Flapper Skates
John Halpin , The Scottish Association for Marine Science
Joe Marlow1, Steven Benjamins1, Jane Dodd2, Thomas Wilding1 1 - Scottish Association for Marine Science, Oban, Argyll, PA37 1QA 2 - NatureScot
Flapper skates (Dipturus intermedius), once common in the Northeast Atlantic, are currently classified as Critically Endangered by the International Union for the Conservation of Nature (IUCN). To assess the efficacy of management measures, such as dedicated marine protected areas on skate populations, novel monitoring approaches are needed. Flapper skates can be accurately identified by their distinctive dorsal spot patterns. This has led to the development of the SkateSpotter database, managed by SAMS and NatureScot. The database collates images submitted by anglers for identification/tracking of individual skates, enabling study of skate migration and home-ranges, important considerations in their management. SkateSpotter presently relies on manual identification of skates which, while accurate, is time consuming. Here we present a deep learning method for skate re-identification, using a custom 2-stage deep learning algorithm. The algorithm can re-identify skate with a high degree of certainty (for a test set, unseen during training, 80% of images had correct individual as 1st match) and deal with the variation in the angler submitted images, for example, glare and occlusion. The algorithm is being incorporated into SkateSpotter, enabling a step-change in the speed at which new skate images can be checked for a match. This will enable the cost-effective extension of the database (in time and geography) enhancing scientific value. The combination of Skate Spotter and Rate-A-Skate represent best practice for future animal re-identification challenges, with the success of the deep learning algorithm being proportional to the quality of the images in the database.
Seabed field truth by imaging to protect biodiversity
Paula Vieira Castellões , CLS Brasil
Aline Wyllie Rodrigues (TotalEnergies EP Brasil); João Régis Santos-Filho (Lagemar/Universidade Federal Fluminense -UFF); Leonardo Maturo Marques da Cruz (CLS Brasil); Anderson Catarino (TotalEnergies EP Brasil); Fernanda Ramos de Lima (CLS Brasil); Cleverson Guizán (Lagemar/Universidade Federal Fluminense - UFF).
To safeguard seabed biogenic structures and diverse fauna, and to conserve these ecosystems and their biodiversity, it’s crucial to conduct initial studies to foresee and prevent potential impacts from exploratory activities. This is a best practice and one of the requirements by the Brazilian environmental agency, IBAMA, for environmental licensing and is also in line with TotalEnergies commitments of defining voluntary exclusion zones and managing local biodiversity in its new projects. These restrictions act as a catalyst for the use of techniques and tools that not only fulfill these requirements but also surpass them. To perform these studies, TotalEnergies EP Brasil adopted the approach developed by CLS Brasil, in partnership with Lagemar/UFF. Such approach is based on the integrated analysis of regional and local data, with subsequent verification of field truth through ROV (remotely operated vehicle) imaging, for mapping seabed surface. This case study focus on Blocks S-M-1711/S-M-1815 (1,296 km² in total), an area in Santos Basin, Brazil, with no prior exploratory history. The integrated and georeferenced analysis was conducted to map submarine relief and surface sediments compiled from seismic, mono and multibeam bathymetric, and sedimentological data. Seabed imaging aimed to identify biogenic environments, especially formed by benthic bioconstructive communities such as beds of calcareous algae and deep-water corals (known to occur in Santos Basin), and to recognize other organisms, colonies or biological structures, as well as relevant physical environmental features, including exudations and oil seeps, that could expand knowledge about the seabed, enabling more accurate and detailed mapping. The imaging was conducted at the same nine stations used for water and sediment quality characterization. If the integrated mapping had identified areas of interest for E&P activities resembling locations with bioconstructions, TotalEnergies would have applied the mitigation hierarchy (avoidance) and new imaging would have been performed.
Sweating the small stuff: Annotation methods to improve small and low cover organism estimates
Emma J. Curtis , School of Ocean and Earth Science, National Oceanography Centre, University of Southampton Waterfront Campus, Southampton, SO14 3ZH, UK; Centre for In Situ and Remote Intelligent Sensing, University of Southampton, Southampton, SO16 7QF, UK
Jennifer M. Durden, Ocean BioGeosciences, National Oceanography Centre, European Way, Southampton, SO14 3ZH, UK; Blair Thornton, Centre for In Situ and Remote Intelligent Sensing, University of Southampton, Southampton, SO16 7QF, UK, Institute of Industrial Science, The University of Tokyo, 4-6-1 Komaba Meguro-ku, Tokyo 153-8505, Japan; Brian J. Bett, Ocean BioGeosciences, National Oceanography Centre, European Way, Southampton, SO14 3ZH, UK
Small and low cover organisms can make up a large proportion of community structure and provide important insight into the resilience of ecosystems. Manually extracting measurements used to monitor these organisms with image annotation can be a high effort process and subject to human error and bias. Automating image annotation of marine images can reduce manual annotation effort and some human error, but supervised techniques and evaluating predictions rely on robust manual annotations. We present considerations for annotating small and low cover organisms from seafloor images, using images annotated for sparsely distributed cold-water coral by 11 researchers as an example case. Using at least two annotators with some overlap in annotation allows for an evaluation of annotation success and uncertainty in derived measurements. Segmenting distinct coral colonies reduced the standard deviation of cover estimates threefold compared to a grid-based cover estimation method, with no significant difference in annotation time. Size bias in manual annotation can lead to missing or incomplete annotations of small organisms and can be propagated by supervised machine learning techniques when incomplete annotations are used as training data, as demonstrated in our example case. By modelling missing sizes of colonies, we reduced the error in our coral cover estimates by up to 23%. When these modelled sizes were used to generate segments for incomplete annotations, trained machine learning predictions significantly improved. We discuss some additional methods which are either more computationally expensive or require more human effort to reduce size bias in image annotations, giving a summary of the impacts and limitations of the outlined methods to provide researchers with guidance on how to more efficiently and reliably monitor small and low cover organisms using seafloor images.
Three-dimensional movements of deep-sea octopus crawling quantified using the EyeRIS plenoptic imaging system
Joost Daniels , MBARI
Crissy L. Huffard, Paul L.D. Roberts, Alana D. Sherman, Henry Ruhl, and Kakani Katija (MBARI)
Ocean animals have adapted to their underwater environment in a myriad of ways, developing sensory systems and biomechanical strategies unlike those found on land. Octopuses in particular, use their flexible arms to swim and crawl in ways not seen in other species. Their movements have garnered considerable interest from roboticists, who attempt to emulate this highly complex yet adaptable system. However, studies have not yet quantified three-dimensional arm movements during crawling in-situ, due in part to the complexity of obtaining these 3D data. We used the plenoptic imaging system EyeRIS, deployed by a remotely operated vehicle, to record the three-dimensional arm kinematics of Muusoctopus robustus at a depth of 3200 meters off the coast of Central California, USA. Using semi-automatic, markerless tracking of the arms, we quantified strain, as well as the radius and location of bends in the arms. Preliminary analysis supports the hypothesis that these octopuses prioritize simple, rotation-based arm movements to generate locomotion. This demonstrates a novel use of plenoptic cameras in underwater environments, where the captured high-resolution 3D data can be used both for detailed point tracking as well as error reduction. Additional perspective cameras on the remotely operated vehicle helped reveal conserved attributes in whole-animal gait kinematics as the octopus crawled across the terrain. Despite moving over rocks of irregular sizes, their strides showed remarkable consistency, even across individuals. This represents the first in-depth study of deepwater octopus crawling kinematics, and significantly expands the existing data on in-situ octopus movement, which thus far – despite significant interest of the robotics community – has been limited.
Underwater Image Formation Ground-Truth
Patricia Schöntag , Geomar Helmholtz Centre for Ocean Research AND Christian-Albrechts Universität zu Kiel (CAU Kiel))
David Nakath (CAU Kiel), Mengkun She (CAU Kiel), Yifan Song (CAU Kiel), Gabriel Nolte (Geomar), Rüdiger Röttgers (Helmholtz Centre Hereon Geesthacht), Kevin Köser (CAU Kiel)
Underwater data acquisition usually demands high effort in equipment, logistics, time and cost. Particularly, acquiring images for developing underwater computer vision comes with a higher dimension of complexity as in air. An essential foundation for computer vision development are image formation models that link physical properties to image appearance. For the underwater case this necessitates to consider also complex optical properties of water as spectral attenuation and volume scattering in addition to light, camera and scene parameters. Thus, development and evaluation of those models require comprehensive data sets which have not been available yet. However, physically accurate underwater image formation models can enable various applications as generating training images for deep learning approaches, deduction of environment parameters from the image or true color estimation. Therefore, we address this gap in evaluation data with two image sets, one from a controlled tank environment and another in the open water. Light and camera parameters were acquired both times by using methods based on calibration images, the water parameters with spectrophotometers and volume scattering meters. To determine the underlying scene geometry and water and light-free texture in the open water, we deployed a known target beneath the camera. A more intricate scene was arranged in the tank that allows to reconstruct geometry and texture in air before filling water. Both sets combine images that show the same scene in varying turbidity and color. The open water image set was acquired in 10 near- and offshore spots in the western Baltic Sea. The tank image set was in turn created by adding colorants and scattering agents to the water according to literature knowledge about Jerlov water type properties. In this way, we obtain two diverse sets of underwater images with all the corresponding image formation parameters that are necessary to explain their appearance.
Using seafloor imagery to study short-term temporal variation in megafaunal communities in the Clarion-Clipperton Zone
Bethany Fleming , University of Southampton; National Oceanography Centre, UK
Erik Simon- Lledό, Noelie Benoist, Alejandra Mejía-Saenz, Daniel O. B. Jones. Affiliation: National Oceanography Centre, UK
Biological communities change over time. This can be in response to environmental variability, disturbance, biological or neutral processes. Studying temporal variability in the deep sea is challenging as the opportunity to revisit study areas regularly or establish continuous monitoring is relatively rare. Long-term monitoring sites have demonstrated that abyssal benthic communities change in response to natural variation, such as changes in POC flux (seasonal and interannual) and decadal climatic events such as ENSO. However, temporal change in the Clarion-Clipperton Zone (CCZ) has not been extensively studied. Photo transects were collected in the NORI-D area over 3 consecutive years. This provides a unique opportunity to study short-term temporal variation of megafaunal communities in the CCZ. However, differences in platforms and camera systems between surveys can make whole community comparisons across years challenging. We developed a framework to be able to standardise across datasets that differ in image quality, thereby ensuring that the results are a reflection of true temporal patterns and not differences in methodologies. To standardise across datasets, we focused on individuals >20mm (large enough to be detected in all years); assigned morphospecies to a detectability category (reliably detectable across all datasets; detectable but abundances likely to be skewed by differences in image quality between years; not likely detectable across all datasets i.e. small, indistinct), morphospecies in the third category were excluded from analyses. We also only retained morphospecies that had high enough density to be unlikely to be missed by chance. This framework was applied on image data from NORI-D. Megafaunal communities were expected to show relatively little variation in density, diversity and community composition between years, although temporal variation may vary between functional groups. It is important to understand how benthic communities vary over time so that future monitoring strategies can adequately take into account natural variability.
Validating an AI-based holographic imaging platform (Aqusens) for monitoring harmful algal bloom species
Holly Bowers , Moss Landing Marine Laboratories
Maxim Batalin, Lucendi, Inc. Alexis Pasulka, California Polytechnic State University
Harmful algal bloom (HAB) species can have detrimental effects on ecosystems, public health, fisheries, drinking water, subsistence harvesting, aquatic recreation and tourism. The HAB field is working to expand detection capabilities within a nationally networked system for long-term monitoring and early warning (National HAB Observing Network; NHABON). To meet this need, instruments demonstrating portability, robustness, ease-of-use, and the ability to produce real-time data are highly desired. Our project compares and validates a new cost-effective automated cell imaging platform (Aqusens) with the Imaging Flow CytoBot (IFCB). The Aqusens is based on holographic imaging principles coupled with machine learning for optimizing computation and object characterization. The Aqusens does not contain expensive and complex to maintain optical or mechanical components. Most of the hardware is off-the-shelf and mass-produced, allowing for scalable production and cost reduction. Accompanying software is used to control Aqusens operations and analyze collected data. It is user friendly and does not require substantial training or maintenance. Validation of the Aqusens is being carried out through three objectives: 1) laboratory experiments with a variety of cultured HAB and non-HAB species to provide foundational platform comparisons; 2) deployments at the Santa Cruz Wharf monitoring site to test performance during varied conditions (e.g. algal blooms, upwelled sediment); and 3) in San Francisco Bay cruise transects (slated for 2025) targeting seasonal succession of phytoplankton populations. These cruises, along with the Santa Cruz Wharf monitoring station, provide avenues for new platform integration into well-established programs that offer publicly available long-term data sets. Furthermore, platform portability is allowing us to support a broad variety of stakeholders (e.g. aquaculture, educational programs). These efforts are assessing functionality of the Aqusens in various settings and evaluating integration of data into existing public portals (California Ocean Observing Systems Data Portal [CalOOS] / Harmful Algal Bloom Data Assembly Center [HABDAC]).
Marine Imaging Workshop/presentations/posters
"Promoting Quality and Inclusive Education at the Faculty of Natural Sciences in Namibe, Angola: A Multifaceted Approach"
Onelis Portuondo Savón , Ana Paula Sarmento Dos Santos
Ubaldo Jorje Augusto de Filipe André
This article proposes a comprehensive strategy to promote quality and inclusive education at the Faculty of Natural Sciences in Namibe, Angola, aiming at the personal and professional development of students. The strategy addresses various aspects, such as assessing students' needs, implementing inclusion policies, ensuring physical and digital accessibility, establishing student support programs, developing inclusive curricula, teacher training, community partnerships, and implementing evaluation and monitoring systems. By adopting this multifaceted approach, the aim is to empower the college to play a significant role in preparing students to contribute positively to their communities and society at large, thus promoting quality and inclusive education. This strategy is crucial, especially in a context where distance education is becoming increasingly relevant, as during the COVID-19 pandemic, and where the adaptation of educational practices is essential to ensure access and quality teaching. Keywords: Quality education, Inclusion, Inclusive curricula, Professional development, Multifaceted approach
AESA 2: Homogenised Multi Application Camera System for Marine Science
Daniel Roper , National Oceanography Centre
Building on the long term success of the Autonomous Ecological Survey Abyss (AESA ) camera systems developed by National Oceanography Centre. Developed in 2011 specifically for marine ecological imaging from a moving Autonomous underwater Vehicle (AUV). The original AESA camera system utilised the FLIR grasshopper 2 machine vision camera using a Sony ICX625 CCD sensor. A machine vision camera was selected over digital single light reflector (DSLR) cameras because the simplicity and lack of moving parts, improves reliability, and the additional connectivity, allows storage and access to the images off the camera. The compact formfactor of machine vision camera is also a significant advantage for packaging for the subsea environment. Due to obsolescence of the FLIR grasshopper 2 series, for the latest iteration of the camera system the decision was made to update the sensor to a CMOS sensors, utilising the Sony IMX 183 from the Starvis low light imaging sensor series, offering improvements in both absolute sensitivity, and power consumption. The updated system still uses the Xenon high speed flash, which enables taking of high-resolution stills from moving platform. The goal is now to utilise the same camera system over multiple observing platforms. By homogenising camera and flash across multiple platforms, AUV, ROV and Bathysnap time-lapse lander, and the long range AUV fleet developed at NOC. It is hoped that this will result in streamlined workflows, cross comparable images, and a reduction in engineering overheads. The use of machine vision cameras also allows the possibility of real time image processing and analysis onboard the platform. Realtime image quality metrics could feasibly be fed into platform behaviour algorithms, to automatically adjust the platforms altitude to optimise image quality.
AI-Derived Krill Detection and Quantification from ROV Video
Savana Stallsmith , Oregon State University
Chris Sullivan Oregon State University: Astrid Leitner Oregon State University and MBARI
Artificial Intelligence is becoming a popular solution for solving a variety of complex problems in a way that allows us to work smarter and not harder. One important and evolving application of AI is in marine imaging, specifically helping to widen the bottleneck that the time consuming annotation of marine video creates for oceanographic research. While rapid advances are being made in AI for benthic marine images, the pelagic system is much more challenging because animals are highly mobile and low contrast with flexible, delicate morphologies. Krill are a key component of the California upwelling system food web and are highly mobile and often highly abundant from the surface to 200m depths. Quantifying krill from video is extremely difficult and time consuming. To solve this issue, we trained an A.I. model to automatically detect and quantify krill in pelagic ROV video transects from the Monterey Bay. We compare the AI generated quantifications with pre-existing hand quantified videos. We show that it is possible to train an A.I. model to detect krill and to export those detections to generate counts, though comparison to manual counts remains a challenge and requires tracking across frames. Several solutions to tracking individuals were tested, and with continuing development this approach could make video a viable method for quantifying these important prey organisms as well as other challenging midwater species.
AI-based Identification of Ocean Organisms in Imagery and Enhanced Human-Machine Collaboration for improved Video Analysis
Patrick Cooper , University Corporation for Atmospheric Research / NOAA Ocean Exploration Affiliate
Mashkoor Malik_2, Ashley N. Marranzino_1, Philip Hoffman_2 (1-University Corporation for Atmospheric Research / NOAA Ocean Exploration Affiliate, 2- NOAA Ocean Exploration )
Marine imagery analysis is essential for understanding and exploring underwater environments. However, manually annotating and characterizing captured footage and images is a time-consuming task for human experts. NOAA Ocean Exploration currently leverages SeaTube - A tool built by Ocean Networks Canada for human-based video annotations. To address these limitations, NOAA Ocean Exploration is scoping and developing several tools including “Seabot” and “Critter Detector” to increase annotation efficiency and evaluate the potential for automated annotations. The unique data distribution of marine imagery and sparse sampling compared to terrestrial domains requires human-in-the-loop systems and our efforts are focused on improving the efficiency of human-based annotations. Seabot uses Vision Transformers (ViTs) to categorize and localize organisms within individual images, relying on the Ocean Vision AI / FathomNet training dataset. Critter Detector operates on video footage and identifies distinct video segments of interest, entities, organisms, and debris within arbitrary video clips. Critter Detector utilizes deep learning, specifically convolutional neural networks (CNNs) and long short-term memory (LSTM) networks, to detect significant events and visually interesting content within underwater video footage leveraging SeaTube footage and annotations to train the Critter Detector. While Seabot emphasizes organism identification on individual images, Critter Detector is designed to utilize the recurrence relation of an RNN architecture to allow for identification of significant temporal regions within video. Both of these proposed models, trained on a diverse dataset of annotated images and videos, achieve useful accuracy, reducing the time and effort required for manual analysis. As the technology advances, incorporating features such as species recognition and behavior analysis, realtime deployment, it may provide new opportunities for marine research, targeted sampling operations and conservation efforts.
AQUA-VIS: Autonomous QUantitative Analysis and Visual Identification System
Judith Fischer , GEOMAR
Autonomous Underwater Vehicle (AUV) missions have revolutionized the exploration of the underwater world and enabled high-resolution seafloor imaging. This imagery is instrumental for monitoring benthic ecosystems, assessing biodiversity, or detecting objects such as unexploded ordnance. However, post-processing the acquired data from AUV surveys is often time-consuming, hindering immediate decision-making for subsequent mapping locations and effective environmental analysis. This work addresses this issue by developing an advanced image processing workflow tailored for AUVs that leverages open-access image processing tools. The proposed workflow aims to process incoming images in near real-time, extract features for 3D reconstruction, and detect objects from seafloor images. The workflow consists of two main steps: i) feature extraction and ii) object detection. Feature extraction is carried out using the open-source software COLMAP, which is a "general-purpose Structure-from-Motion (SfM) and Multi-View Stereo (MVS) pipeline". In the second part of the workflow, pre-trained tracking models such as YOLO are utilized to create an object detection framework that can detect objects from incoming images. Our current focus is on training YOLO using seafloor image datasets. We aim to train models for different locations, such as the Atlantic, Pacific, and Baltic Sea, thereby creating trained models that the wider community can use. These workflows are integrated into the AUV system to ensure they can operate in near real-time conditions. Processing time is optimized to handle the large number of incoming images while maintaining high accuracy. We anticipate that the results of this framework will support the accessible use of open-source workflows to the research community. Scientists can use this tool to gain valuable insights into underwater ecosystems and make informed and fast decisions about further dive sites based on immediate reports generated by the AUV at the end of each dive.
Accelerating Marine UAV Drone Image Analysis with Sliced Detection and Clustering (MBARI SDCAT)
Danelle E. Cline , Monterey Bay Aquarium Research Institute (MBARI), Moss Landing, CA, USA
Duane R. Edgington, Thomas O'Reilly, Steven H.D. Haddock, John Phillip Ryan, Bryan Touryan-Schaefer, William J. Kirkwood, Paul R. McGill, Enoch Nicholson, and Rob S. McEwen. Monterey Bay Aquarium Research Institute (MBARI), Moss Landing, CA, USA
Uncrewed Aerial Vehicles (UAVs) can be a cost-effective solution for capturing a comprehensive view of surface ocean phenomena to study marine population dynamics and ecology. UAVs have several advantages, such as quick deployment from shore, low operational costs, and the ability to be equipped with various sensors, including visual imaging systems and thermal imaging sensors. However, analyzing high-resolution images captured from UAVs can be challenging and time-consuming, especially when identifying small objects or anomalies. Therefore, we developed a method to quickly identify a diverse range of targets in UAV images. We will discuss our workflow for accelerating the analysis of high-resolution visual images captured from a Trinity F90+ Vertical Take-Off and Landing (VTOL) drone in near-shore habitats around the Monterey Bay region in California at approximately 60 meters altitude. Our approach uses a state-of-the-art self-distillation with knowledge (DINO) transformer foundation model and multi-scale, sliced object detection (SAHI) methods to locate a wide range of objects, from small to large, such as schools or individual jellyfish, flocks of birds, kelp forests or kelp fragments, small debris, occasional cetaceans, and pinnipeds. To make the data analysis more efficient, we create clusters of similar objects based on visual similarity, which can be quickly examined through a web-based interface. This approach eliminates the need for previously labeled objects to train a model, optimizing limited human resources. Our work demonstrates the useful application of state-of-the-art techniques to assist in the rapid analysis of images and how this can be used to develop a recognition system based upon machine-learning for the rapid detection and classification of UAV images. All of our work is freely available as open-source code. (https://github.com/mbari-org/sdcat)
Addressing challenges in using object detection for small datasets of imagery with a data-centric approach
Caroline Chin , National Institute of Water and Atmospheric Research (NIWA), Wellington, New Zealand; Te Herenga Waka - Victoria University of Wellington, Wellington, New Zealand
Ashley Rowden (National Institute of Water and Atmospheric Research (NIWA), Wellington, New Zealand; Te Herenga Waka - Victoria University of Wellington, Wellington, New Zealand), Alan Tan (National Institute of Water and Atmospheric Research (NIWA), Auckland, New Zealand); Daniel Leduc (National Institute of Water and Atmospheric Research (NIWA), Wellington, New Zealand
In marine imagery research, real-world applications using object detection models face challenges such as class imbalance due to heterogeneity within benthic communities, class/labelling ambiguity, and image quality (e.g., illumination). With these challenges, it is difficult to determine the optimal quantity and quality of research-derived training data – that are typically small datasets – thus impacting model performance. Furthermore, traditional object detection performance metrics such as mean Average Precision (mAP), F1-score, precision, and recall may not comprehensively evaluate a model’s efficiency nor inform the researcher on where errors or gaps may exist within the training data that can be targeted for improvement. In this study, our aim was to understand benthic community patterns within the Kermadec Trench by utilising object detection models for video imagery. We created training datasets by annotating benthic megafauna from random subsets of submersible dive imagery from abyssal and hadal depths using the BIIGLE 2.0 annotation tool. Using these datasets, we evaluated the performance of YOLOv8 object detection models trained on both single- and multi-class datasets. Our methods included analysis of performance metrics and data visualisations provided within the YOLOv8 framework. Our results demonstrated the importance of analysis within the training dataset to identify issues such as bounding box size, orientation, and localisation within an image that can be resolved with methods such as data augmentation. Furthermore, we observed that random sampling of annotated imagery resulted in a skewed dataset biased towards abyssal megafaunal classes, subsequently increasing class imbalance. To mitigate this issue, we undertook stratified random sampling of training imagery by depth that balanced and improved dataset quality, particularly as faunal classes within abyssal and hadal zones exhibited heterogeneity. Overall, our study demonstrates that by using a data-centric approach towards deep-learning, researchers can relatively easily infer how small-scale training data can be improved for object detection performance.
Advanced seagrass monitoring using automated image processing on underwater drones
Rod Connolly , Global Wetlands Project, Griffith University, Australia
Cèsar Herrera(1), Jasmine Rasmussen(1), Alison McNamara(1), Michael Sievers(1), Bela Stantic(2) 1. Global Wetlands Project, Griffith University, Queensland, Australia 2. Big Data Lab, Griffith University, Queensland, Australia
Advances in capacity for mapping and monitoring seagrass habitat are overcoming deficiencies in local monitoring and global distribution maps, and enabling rapid condition reporting. Underwater vehicles and drones reliably scan the seabed collecting vast amounts of imagery, in waters of any depth without putting divers at risk. We show how automated data extraction has overcome the processing bottleneck of manual data extraction. Using a combination of deep learning architectures (CCN and Transformer), we provide computer vision software that detects seagrass and records plant characteristics rapidly and reliably, anywhere in the world. Robust testing against large databases of manual records of seagrass presence and percentage cover shows high accuracy of automated measures. Seagrass presence data are ideal for detecting meadow size and location, and percentage cover data can be used as an indicator of seagrass condition, paired with automated records of seagrass species or morphological type. The combination of underwater drone imagery and automated data extraction substantially increases the scale of seagrass surveys, even in deep or turbid waters, and improves reproducibility from sequential visits.
Algorithmic construction of topologically complex biomineral lattices via cellular syncytia
Manu Prakash , Stanford University (BioE, Biology, Océans)
Pranav Vyas, Charlotte Brannon, Laurent Formery, Christopher J. Lowe, Manu Prakash (Stanford University)
Biomineralization is ubiquitous in both unicellular and multicellular living systems [1, 2] and has remained elusive due to a limited understanding of physicochemical and biomolecular processes [3]. Echinoderms, identified with diverse architectures of calcite-based structures in the dermis[4], present an enigma of how cellular processes control shape and form of individual structures. Specifically, in holothurians (sea cucumbers), multi-cellular clusters construct discrete single-crystal calcite ‘ossicles’ (∼ 100 µm length scale), with diverse morphologies both across species and even within an individual animal [5]. The local rules that might encode these unique morphologies in calcite ossicles in holothurians remain largely unknown. Here we show how transport processes in a cellular syncytium impart a top-down control on ossicle geometry via symmetry breaking, branching, and fusion in finite cellular clusters. As a unique example of cellular masonary, we show how coordination within a small cluster of cells builds calcite structures about an order of magnitude larger than any individual participating cell. We establish live imaging of ossicle growth in Apostichopus parvimensis juveniles revealing how individual crystalline seeds (∼ 1 − 2 µm) grow inside a multi-cellular syncytial complex with the biomineral completely wrapped within a membrane-bound cytoplasmic sheath. Constructing a topological description of ossicle geometries from 3D micro-CT (computational tomography) data reveals the hidden growth history and conserved patterns across ossicle types. We further demonstrate vesicle transport on the surface of the ossicle, rather than cell motility, regulates material transport to the ossicle tips via a unique cytoskeletal architecture. Finally, using reduced order models of conserved transport on self-closing active branching networks, we highlight the hidden universality in the growth process of distinct ossicles. The system presented here serves as a unique playground merging top-down cellular physiology and classical branching morphogenesis [6] with bottom-up non-equilibrium mineralization [7] processes at the interface of living and non-living matter [8].
Annotating Benthic Megafauna of the Cobalt-Rich Manganese Seamounts of the Mid-Pacific Mountains, Hawaiian and Necker Ridges with BIIGLE
Sierra Landreth , Department of Earth, Ocean, and Atmospheric Science, Florida State University
E. Brendan Roark (Department of Geography, Texas A&M University), Virginia Biede, Nicole B. Morgan, Amy R. Baco (Department of Earth, Ocean, and Atmospheric Science, Florida State University)
Seamounts are habitats for a diverse array of benthic deep-sea megafauna. Generally, seamounts are poorly studied in the central and western Pacific (CWP), an area that is also targeted for cobalt-manganese crust mining. The Mid-Pacific Mountains (MPM) have cobalt-rich crusts but are virtually unexplored. Necker Ridge has been hypothesized to be a key stepping-stone for deep-water faunal dispersal between the MPM and the seamounts of the Northwestern Hawaiian Islands (NWHI). The goal of this project was to test this hypothesis by comparing the species composition of CWP seamount fauna among the MPM, NWHI, and Necker Ridge while also observing the communities at potential risk from future mining activities. The remotely operated vehicle (ROV) SuBastian was used to collect video transects across 10 sites. Replicate 500m long transects, were taken at 500 m depth intervals from 1,500 to 3,500m and annotated for morphology and taxonomy of benthic megafauna using the BioImage Indexing Graphical Labeling and Exploration (BIIGLE) web service. Video transects were converted to screen grabs collected every 10 seconds. Organisms were labeled with the Standardised Marine Taxon Reference Image Database (SMarTaR-ID) morphology and taxonomy label trees. Species composition, abundance, and diversity of the benthic community were calculated at site MPM3 to establish depth zonation. Based on preliminary results, octocorals were the most abundant taxon at all depths, with the two deepest depths dominated by unbranched isidids. There was a peak in overall faunal abundance at 2000 m, with approximately 2000 megafaunal individuals across 3 transects. The annotation process was repeated across all survey sites to begin comparisons among the 3 regions. Across all sites, faunal abundance was lowest at the deeper depths of 3000 to 3500 m. Characterizing seamount communities and their connectivity can be essential to informing the conservation and management of vital deep-sea habitats targeted for mining.
Applications of the YOLOv8 Computer Vision Model for Long-term Underwater Ecological Monitoring.
Talen Rimmer , Department of Biology, University of Victoria
Bates, C.R. - University of British Columbia McIntosh, D., Zhang, T., A., Branzan, Albu A., & Juanes, F. - University of Victoria
The world's oceans are undergoing rapid changes due to climate change and other anthropogenic impacts, affecting marine species' distribution, abundance, and behavior. Traditional ecological monitoring methods struggle to keep pace with these transformations, especially in underwater habitats. Recently, computer vision techniques have emerged to enhance the efficiency of video and image-based underwater monitoring. These methods facilitate the detection and classification of objects in visual data, potentially streamlining the process of counting and classifying marine organisms. However, the adoption of computer vision in marine ecology has been slow, partly due to its inaccessibility to ecologists and the lack of easily adaptable tools for ecological monitoring. This study investigates the application and validation of computer vision techniques for monitoring underwater pelagic macrofaunal diversity, using a case study from coastal British Columbia. Over 9000 hours of underwater video were collected from four sites over 18 months, using mounted cameras programmed to record five minutes of video every hour. A supervised computer vision model (YOLOv8) was trained with approximately 240,000 annotations to assess the presence and abundance of marine species at these sites. This approach provided high-resolution temporal data on the diversity and abundance of pelagic fish and gelatinous zooplankton at these sites. We detail the process of employing computer vision techniques for long-term underwater ecological monitoring, emphasizing accessibility for ecologists. A stepwise method for adapting computer vision techniques to achieve biodiversity monitoring objectives is presented, highlighting the strengths and limitations of our approach. Finally, we propose future directions for integrating these technologies into new and existing monitoring programs, and suggest priority areas for future research to advance the use of computer vision in underwater ecological monitoring.
Aqusens: AI-powered holographic imaging flow cytometer for cost-effective high-throughput plankton monitoring and identification
Maxim Batalin , Lucendi, Inc.
Zoltan Gorocs (Lucendi, Inc.), Dhruvin Naik (Lucendi, Inc.), Jessica Keszey (Lucendi, Inc.), Link Li (Lucendi, Inc.), Andrew Jones (Lucendi, Inc.), Maria Arredondo (Lucendi, Inc.)
Aqusens is an AI-powered cost-effective, high-throughput platform for automated monitoring and identification of microobjects, such as plankton, in water. Unlike existing technologies, Aqusens is a portable device, weighing under 3 lbs, that can be used for in-lab or in-field monitoring. Instead of relying on lenses, Aqusens uses holographic microscopy coupled with deep learning. This approach allows the platform to provide a large object identification size range (3 micron – 3 mm) with high-throuhgput of 100 mL/hour, which can be further adjusted to several liters per hour. In Aqusens, the sample is pumped through a monitoring volume of the device, which is 4.6mm x 6.1mm x 0.8mm, and is imaged 3 times per second to ensure every object is captured at least once. Such a large monitoring volume allows the system to be less susceptible to clogging and supports a variety of different applications from imaging cyanobacteria to larger zooplankton organisms without the need to modify the setup. The system is robust and does not require any calibration or adjustments for day-to-day operations. For every object in its observational volume, Aqusens creates an intensity and phase image pair. Phase images allow to capture additional data, such as parts of the object that are transparent, or to capture changes in density or even material/media. Furthermore, intensity and phase image data provides many additional features for development of machine learning algorithms capable of accurate object identification and characterization. These capabilities are part of Aqusens analytics engine that can be customized for the automated detection of specific types/classes of objects for a given application. The Aqusens platform has been experimentally used for a variety of applications, including identification of harmful algae, characterization of copepods, rotifers and artemia, as well as for quantification of microplastics in water samples. Aqusens is developed to expand the opportunities for the use of high-throughput automated microscopy and can be easily adapted and supported for custom applications.
Automated Imaging in Long-Term Ecological Research Reveals Taxon- and Group-Specific Bloom Anomalies in Phytoplankton
Heidi M. Sosik , Woods Hole Oceanographic Institution
E. Taylor Crockford, Emily E. Peacock, and Miraflor Santos; Woods Hole Oceanographic Institution
Coastal pelagic waters support complex and dynamic ecosystems that depend on planktonic food webs. These systems are challenging to observe at the appropriate spatial, temporal, and taxonomic scales. We are working to meet this challenge with automated microscopic imaging at the Northeast U.S. Shelf Long-Term Ecological Research (NES-LTER) site. For over a decade, we have deployed Imaging FlowCytobot both at a nearshore cabled observatory and on-board research vessels surveying across the shelf seasonally. With this vast dataset (>1.5 billion images) and machine-learning for image classification, we are now able to document where and when unusual blooms of specific taxa or groups of phytoplankton occur. These atypical blooms are surprisingly common in the dataset, but also highly variable both in which taxa occur and in spatial and temporal location and extent. This presentation will highlight examples of decade-scale outlier abundances of specific diatoms, dinoflagellates, and haptophytes. We will show that some events appear to be nearshore phenomena with timescales of weeks while others are localized in mid- or outer- shelf waters over similar timeframes, and furthermore that certain massive, unprecedented blooms can persist for months and extend over 100s of kilometers along the NES. While we have an evolving understanding about factors promoting some of these types of blooms, in many cases they remain mysterious to explain.
BathyBot: The deep-sea crawler to see the unseen of the NW Mediterranean Sea
Séverine Martini , Mediterranean Institut of Oceanography (MIO)
Séverine Martini (1), Carl Gojak (2), Christian Tamburini (1), Dominique Lefèvre (1), Karim Bernardet (2), Karim Mahiouz (2), Céline Laus (2), Marc Picheral (3), Camille Catalano (3) 1- Mediterranean Institute of Oceanography 2- Division Technique de l'INSU 3- Laboratoire d'Oceanographie de Villefranche
Increasing exploration and industrial exploitation of the vast and fragile deep-ocean environment for a wide range of resources (e.g., oil, gas, fisheries, new molecules, and soon, minerals) raises global concerns about potential ecological impacts. BathyBot is a multi-instrumented deep-sea crawler deployed from a dock, at 2500m depth, 40 km off the French coast (Mediterranean Sea), at the EMSO-LO station. BathyBot is connected to the LSPM, a deep sea cabled observatory allowing real-time observations of the deep sea. The deployment of this benthic vehicle complements the ALBATROSS-MII pelagic line, instrumented with oceanographic sensors. Two cameras are installed on BathyBot for real time imaging of the deep sea, as well as an Underwater Video Profiler (UVP6) on its BathyDock, and a biocamera closeby the vehicle. Such instrumentation allows: 1) to better understand the biodiversity in the deep Mediterranean sea over time, 2) to to study the dynamic of exchanges at the water column and sediment interface, 3) to involve citizen through the diffusion of images acquired by BathyBot through the platform “Ocean Spy – Meditteranean Spy”.
Benthic imagery data for UK statutory reporting: working towards consistent and repeatable methods and delivering high quality data
Kate Wade , Joint Nature Conservation Committee
Lucy Hirschle (JNCC)
Marine imagery surveys offer a valuable means of gathering high-quality information about benthic communities. The Joint Nature Conservation Committee (JNCC) lead the UK offshore MPA Monitoring Programme, collecting empirical data to monitor, assess, and understand the health of the UK offshore seabed. The imagery collected in the process of monitoring the UK’s offshore MPA network has grown almost exponentially as the usefulness and accessibility of this technology has increased. We now routinely capture thousands of still images for the enumeration of epifauna communities and video footage to map broadscale habitat features. Often imagery is used in tandem with more traditional extractive methods such as benthic grab sampling, but in some cases, surveys are exclusively imagery-based – particularly in locations where grabbing and coring is impractical, or impossible. Compared to physical sampling, imagery can be deployed more effectively across larger and more physically variable study areas, it’s non-destructive, and often more cost-effective. Analysis of benthic community data from imagery has significant challenges with respect to data resolution and quality and it is vital that we have consistent and repeatable methods. This poster will explore these challenges, showcasing JNCC’s work towards producing more useful, accessible, and accurate imagery data, provide recommendations, and invite collaboration. We will detail recent work with the University of Plymouth that builds on the SMarTaR-ID platform, which offers a standardised approach to the identification of taxa from imagery, enabling a more accurate and complete capture of epibiotic diversity in UK waters. We will show JNCC led work on the UK Morophotaxa System (based on the popular CATAMI classification scheme) and the Epibiota Identification Protocol with NMBAQC, a guide for the standardisation of taxon identification linked to image quality. Looking to the future, we will present ideas on the implementation of training datasets to be used with machine learning algorithms.
Boosting Standardised Conservation Reporting and Public Awareness: The Game-Changing Impacts of Collaborative Platforms like SQUIDLE+ and Seamap Australia
Jacquomo Monk , Institute for Marine and Antarctic Studies, University of Tasmania
Ariell Friedman2, Leo Güenzel1, Neville Barrett1, Nicholas Perkins1, Emma Flukes1, Peter Walsh1, Vanessa Lucieer1 1 Institute for Marine and Antarctic Studies, University of Tasmania, Australia 2 Greybits Engineering, Australia
Marine imagery has become an essential non-destructive tool for studying changes in benthic ecosystems across all depths. Thanks to advancements in platform design, there has been a significant increase in the number, resolution, and spatial coverage of images globally. This technology not only underpins scientific research and management but also raises awareness about the importance of marine biodiversity conservation. Meeting the demands of national and international marine environment reporting increasingly requires extracting standardised indicators, like the Essential Ocean Variables (EOVs) set by the Global Ocean Observing System (GOOS). Initially, these indicators were derived from marine imagery using combined datasets from manual annotation efforts. However, integrating artificial intelligence (AI) into processing pipelines has the potential to revolutionise this historically laborious process. AI, when paired with advanced translation frameworks like those in SQUIDLE+, significantly streamlines the creation of these indicators, enhancing our capacity to monitor and comprehend large-scale, long-term shifts in marine ecosystems. Moreover, public interest in the marine environment has also grown significantly, driven by initiatives such as the UN Ocean Decade and philanthropic efforts from organisations like the Schmidt Ocean Institute and Ocean Exploration Trust. Marine imagery plays a crucial role in this process, captivating audiences and highlighting the wonders of the underwater world. In this conference presentation, we will showcase the development and integration of AI within SQUIDLE+ to enable streamlined reporting for indicators for Australia. We will discuss the current strengths and limitations of AI in developing these indicators. Additionally, we will highlight how we are using open marine imagery for public outreach, showcasing collaborations between SQUIDLE+ and Seamap Australia to create immersive content. This aims to engage the public and non-specialists in marine conservation, raise awareness about marine ecosystems, and promote sustainable management practices.
Cold-water coral biodiversity and distribution across oxygen minimum zones in the Galapagos Islands and La Isla del Coco: preliminary findings
Ana Belen Yanez Suarez , Memorial University of Newfoundland
Katleen Robert, Memorial University of Newfoundland
Climate change is altering the chemical and physical properties of the ocean, leading to ocean warming, acidification, and oxygen loss. In some areas of the deep sea, dissolved oxygen levels are already extremely low, and the expansion of these areas may threaten vulnerable marine ecosystems such as cold-water coral (CWC) reefs. CWC can form oases of life in the deep sea, providing habitat, food, and shelter to marine species, but are sensitive to increased temperature and low oxygen. The Eastern Tropical Pacific (ETP) presents one of the world's most pronounced oxygen minimum zones (OMZ), as well as some of the most biodiverse marine areas, such as the Galapagos, Ecuador, and La Isla del Coco, Costa Rica. These Islands are surrounded by large marine protected areas (MPA) where bottom trawling has never occurred, and their deep sea is considered pristine. Given the expected expansion of the OMZ in these critical areas, it is essential to understand the current distribution pattern of CWC and its implications to the local biodiversity. This study aims to investigate whether CWC biodiversity and abundance across OMZ differ within and between Islands and identify the environmental factors that may support CWC resilience under low oxygen conditions. In September 2023, the R/V Falkor (too) collected video, coral samples and environmental data using the ROV SuBastian in several sites across the Islands. Preliminary observations reveal varying patterns of coral distribution, diversity, and community structure related to abiotic oceanographic factors as well as large differences in the OMZ extension between the Galapagos and La Isla del Coco. Our findings will provide essential information to strengthen the development of a regional marine conservation corridor in the ETP currently under implementation by bordering nations as well as improve our current knowledge on CWC adaptation and vulnerability in the Pacific.
Comparing AUV and ROV Imaging in Exploring Deep-Sea Fish Communities of the Papahānaumokuākea Marine National Monument
Beatriz E. Mejía-Mercado , Coastal and Marine Laboratory, Florida State University
Amy R. Baco. Florida State University
In the exploration of deep-sea ecosystems, the comparison between images acquired from Autonomous Underwater Vehicles (AUVs) and Remote Operated Vehicles (ROVs) emerges as a critical aspect, each presenting unique advantages and limitations. AUV imagery excels in capturing high-resolution visuals, providing intricate details on fish morphology, behavior, and distribution, while its autonomy enables efficient surveying of vast areas, including remote and challenging deep-sea locations. However, challenges such as lighting issues and the angle of the field of view may impede accurate fish identification. Despite these limitations, recent studies demonstrated the utility of AUVs in unveiling variations in the structure and spatial distribution of deep-sea fish assemblages in locations such as Necker Island, Pioneer Bank, and seamounts within the Papahānaumokuākea Marine National Monument. On the other hand, ROVs, although tethered, offer real-time control and precise maneuverability, ensuring optimal image capture from desired perspectives. However, the labor-intensive nature and high costs may limit their broad use, exacerbated by restricted mobility in complex terrains. Despite these hurdles, both AUV and ROV imagery are indispensable in advancing our understanding of deep-sea fish communities, providing complementary insights crucial for comprehensive research and conservation efforts in these remote ecosystems. By judiciously leveraging the strengths of each technology while mitigating their respective weaknesses, researchers can significantly enhance our knowledge of deep-sea ecosystems, facilitating more effective conservation strategies.
Crowd-powered precision: Enhancing image segmentation with collaborative annotations
Emma J. Curtis , School of Ocean and Earth Science, National Oceanography Centre, University of Southampton Waterfront Campus, Southampton, SO14 3ZH, UK; Centre for In Situ and Remote Intelligent Sensing, University of Southampton, Southampton, SO16 7QF, UK
Jennifer M. Durden, Ocean BioGeosciences, National Oceanography Centre, European Way, Southampton, SO14 3ZH, UK; Blair Thornton, Centre for In Situ and Remote Intelligent Sensing, University of Southampton, Southampton, SO16 7QF, UK, Institute of Industrial Science, The University of Tokyo, 4-6-1 Komaba Meguro-ku, Tokyo 153-8505, Japan; Brian J. Bett, Ocean BioGeosciences, National Oceanography Centre, European Way, Southampton, SO14 3ZH, UK
Instance segmentation of organisms from marine images can be used for biodiversity estimation and can be used as training data for supervised machine learning techniques. However, manual segmentation can be subject to human error found in other visual data extraction techniques; including misidentification, imperfect detection and inconsistent boundary delineation; which can be propagated through trained machine learning predictions. User variability can lead to low segmentation agreement (intersection over union scores; IoU; of 29%) as seen in our case example of 11 researchers segmenting cold-water coral colonies. We outline two methods for combining users’ segments together to produce less variable annotations. The first method establishes a user consensus on the detection and identification of the annotation and then combines the distance scenes of user drawn segments together to generate an average consensus segment, which can be used to estimate organism properties or as training data for machine learning. The second method overlays the user drawn segments over each other to create a heatmap mask with an associated user certainty for the detection, identification and delineation of the organism. These non-binary masks can be used for organism estimates, or, by adapting the loss function of Mask R-CNN as an example, can be used to train machine learning. Both methods reduce the variability of drawn segments in our example, with a 12% and 25% increase in average IoU scores by combining two and three users’ segments, respectively. The value of combining user annotations at the expense of increased annotation effort and the ease of implementing each combination technique are summarised to advise researchers on how to create more reliable segment annotations and higher quality training data for supervised machine learning techniques.
Deep-sea plastic debris detection using deep learning with a large scale image dataset
Takaki Nishio , Japan Agency for Marine-Earth Science and Technology (JAMSTEC)
Hideaki Saito (Japan Agency for Marine-Earth Science and Technology (JAMSTEC)), Daisuke Matsuoka (Japan Agency for Marine-Earth Science and Technology (JAMSTEC)), Ryota Nakajima (Japan Agency for Marine-Earth Science and Technology (JAMSTEC))
Marine pollution by plastic debris is a global issue, and monitoring of the debris distribution is required to formulate policies for the solution and verify their effectiveness. In particular, Japan has the world's largest seafloor area deeper than 6,000 m in the exclusive economic zone (EEZ), and it is believed that there are a lot of "sinks" of debris on the seafloor. Accordingly, Japan plays an important role in continuous wide-area monitoring of deep-sea debris. In order to make the monitoring of deep-sea debris more efficient, the Japan Agency for Marine-Earth Science and Technology (JAMSTEC) is developing a system that uses deep learning methods to automatically detect plastic debris from deep-sea videos. A deep learning model was trained to detect macro-sized debris such as plastic bags and plastic bottles. The training dataset contains over 22,000 images sliced from videos recorded during JAMSTEC's deep sea surveys conducted over the past 30 years, and debris information was collected from the Deep-sea Debris Database (https://www.godac.jamstec.go.jp/dsdebris/j/index.html). The trained model was applied to newly recorded deep-sea videos to verify whether it could detect plastic debris. In addition, the trained model was tested on videos under multiple conditions, such as the size of the detection target and turbidity in order to determine the shooting conditions for videos suitable for automatic detection. As a result, the automatic detection rate of visually detectable plastic debris achieved over 60%. In this presentation, we report on the latest trial results of deep-sea plastic debris detection, current issues, and future prospects.
Defining the target population to make marine image-based biological data FAIR
Jennifer M. Durden , National Oceanography Centre, Southampton, UK
Timm Schoening (DeepSea Monitoring Group, GEOMAR Helmholtz Centre for Ocean Research, Kiel, Germany), Emma J. Curtis (Ocean and Earth Science, University of Southampton, UK), Anna Downie (Centre for Environment, Fisheries and Aquaculture Science, Lowestoft, UK), Andrew R. Gates (National Oceanography Centre, Southampton, UK), Daniel O.B. Jones (National Oceanography Centre, Southampton, UK), Alexandra Kokkinaki (British Oceanographic Data Centre, National Oceanography Centre, Southampton, UK), Erik Simon-Lledó (National Oceanography Centre, Southampton, UK), Danielle Wright (British Oceanographic Data Centre, National Oceanography Centre, Southampton, UK), Brian J. Bett (National Oceanography Centre, Southampton, UK)
Marine imaging studies have unique constraints on the data collected requiring a tool for defining the biological scope to facilitate data discovery, quality evaluation, sharing and reuse. Defining the ‘target population’ is way of scoping biological sampling or observations by setting the pool of organisms to be observed or sampled. It is used in survey design and planning, to determine statistical inference, and is critical for data interpretation and reuse (both images and derived data). We designed a set of attributes for defining and recording the target population in biological studies using marine photography, incorporating ecological and environmental delineation and marine imaging method constraints. We describe how this definition may be altered and recorded at different phases of a project. The set of attributes records the definition of the target population in a structured metadata format to enhance data FAIRness. It is designed as an extension to the image FAIR Digital Objects metadata standard, and we map terms to other biological data standards where possible. This set of attributes serves a need to update ecological metadata to align with new remotely-sensed data, and can be applied to other remotely-sensed ecological image data.
Developing automated multi-modal monitoring strategies of vulnerable marine ecosystems (VMEs)
Chloe Game , University of Bergen, Norway
Associate Professor Pekka Parviainen, University of Bergen, Norway Dr. Pål Buhl-Mortensen, Institute of Marine Research, Norway Associate Professor Ketil Malde, University of Bergen, Norway and Institute of Marine Research, Norway Dr. Pedro A. Rebeiro, University of Bergen, Norway Dr. Rebecca Ross, Institute of Marine Research, Norway;
Anthropogenic impacts on the marine environment are increasing as growing resource demands must be met. Recently (2024), the Norwegian government voted to open an area of the Norwegian Continental Shelf for seabed mining, putting the deep-sea bed, which harbours a rich diversity of ecologically and economically valuable, yet vulnerable, ecosystems under serious threat. It is critical that extensive accurate maps of the seafloor are created to establish baselines and support monitoring of impacts and recovery. Given the region’s vastness and isolation, such monitoring is logistically challenging and too slow to meet the requirements. It must therefore be conducted with photography and machine learning (ML) used to automatically identify and quantify seafloor organisms; avoiding labour-intensive manual analysis. This project aims to design and develop a multimodal DL model for automated identification of VMEs and VME indicator-species in Norwegian waters. This model will help to optimize mapping efforts, by promoting efficiency, consistency and quality of ecological data extraction. Crucially, this will better enable sustainable management of VMEs. Current monitoring efforts do not provide the scale and resolution of data required to protect and conserve these valuable seabed assets. Specifically, we are investigating how effective can off-the-shelf Vision Transformers (ViT) automate classification of VMEs from seabed imagery compared to current approaches (CNNs). We are also exploring whether the addition of environmental data (from other sensors such as local topography and water temperature), in a new multimodal architecture can increase accuracy of automated analysis and refine the level of biological details used to make predictions, which is currently not possible for ML with images alone. Importantly, we will consider explainability of model decisions, opening up the ‘black-box’, and seek to generalize the approach across target seabed communities and datasets.
Development of illuminated stereo-BRUVS for benthic marine monitoring in the NW Atlantic: Investigation of technological and operational methodologies for best practice
Jessica Sajtovich , Dalhousie University
Dr. Craig J. Brown (craig.brown@dal.ca), Dalhousie University
Imaging systems are valuable tools for sustainable oceanographic survey methods, enabling non-destructive monitoring of marine ecosystems and their associated fauna. Baited Remote Underwater Video Systems (BRUVS) represent an innovative and cost-effective monitoring technology, suitable for a wide variety of marine research, including Marine Protected Area (MPA) and fisheries monitoring. Since their development in the 1990’s, the use of BRUVS for marine monitoring has expanded considerably due to advancements in high-quality, low-cost digital cameras, however, an overwhelming majority of BRUVS surveys have been conducted in the photic zone within the southern hemisphere. BRUVS utilized in these regions typically lack integrated lights and record continuous video footage over short deployment durations (15-60 minutes). Significant knowledge gaps surround the use of BRUVS with integrated lighting for use in poor-visibility, low light, or aphotic environments in more northern latitudes such as the North Atlantic. Considering the potential benefits of BRUVS as a marine monitoring technology, research is essential to determine the effectiveness of BRUVS in these regions. Additionally, in such locations, including the coastal waters of Canada, the use of BRUVS with integrated lights are desirable for MPA monitoring, due to the non-destructive nature of the technology, and suitability for monitoring the diverse range of benthic environments included in Canadian MPAs, many of which are below the photic zone. This research describes the development and application of a stereo-BRUVS with integrated lights for marine monitoring in low-visibility coastal waters surrounding Nova Scotia. The stereo-BRUVS were constructed using in-house designed and consumer grade elements to develop a cost-effective system and capable of producing high-quality images. Depth rated to 500m, the systems record staggered videos, maximizing battery life and video storage, thereby providing advantages for long-term monitoring efforts.
DigiSteert - Digital System for Efficient Evaluation and Regional Tracking of Plankton Image Data
Veit Dausmann , GEOMAR Helmholtz-Center for Ocean Research Kiel
Jan Taucher, GEOMAR Helmholtz-Center for Ocean Research Kiel
A critical challenge in plankton imaging analysis is the contextual isolation of individual plankton objects. Image analytics often result in these entities being studied in isolation, devoid of their inherent spatial and semantic relationships with nearby objects. This isolation overlooks the specific conditions under which the images were captured, which could provide valuable insights into the broader marine ecosystem and its complex interactions. We present a novel software tool designed to enhance the utility and accessibility of in-situ pelagic image data, particularly plankton imagery. The software aims to provide a comprehensive approach to data linkage and semantic context creation within pelagic in-situ imaging datasets. The tool not only establishes a semantic and regional context between these individual objects on a full frame image, but also places them in the larger context of subsequent and preceding full frames and their respective objects. Additionally, the tool further expands this context by linking the crop to oceanographic and image acquisition metadata tied to the entire dataset. This allows for the integration of data into regional and global contexts by connecting it to nearby datasets. In essence, our software tool aims to transform isolated plankton objects within an image into integral parts of a complex, interconnected digital marine ecosystem. Beyond data linkage, the software incorporates features for swift editing and annotation, creating an interface for both image data and linked metadata (including image class, image features, deep learning features, particle statistics, etc.) that is aesthetically pleasing and user-friendly. Importantly, the tool is versatile and compatible with data from various imaging devices.
Discovery of a mud-burrowing squid evidences the complex life-habits of abyssal Cephalopoda
Alejandra Mejia-Saenz , Interdisciplinary Centre of Marine and Environmental Research (CIIMAR), University of Porto, Matosinhos, Portugal
Bethany Fleming 2,3, Daniel O.B. Jones 2, Loïc Van Audenhaege 2, Henk-Jan Hoving 4, Erik Simon-Lledó 2 2 Ocean BioGeosciences, National Oceanography Centre, Southampton, UK 3 Ocean and Earth Science, University of Southampton, Southampton, UK 4 GEOMAR, Helmholtz Centre for Ocean Research Kiel, Kiel, Germany
Cephalopods are a conspicuous component of marine ecosystems, present across all ocean depths, but little is known about the distribution, diversity, and life-habits of those dwelling the deep-sea, especially at remote abyssal depths. Here, we present the unexpected mud-burrowing behaviour of an undescribed species of whiplash squid (Mastigotheutidae) observed from ROV videos at a depth of 4100 m within the Clarion Clipperton Zone, an area in the abyssal central Pacific that is targeted for seabed mining. The specimen was detected motionless in the soft sediment, positioned upside-down with its elongated tentacles extending rigidly towards the water column, resembling a biogenic structure. This behaviour and orientation are in marked contrast to the head-down turning fork posture typically exhibited by Mastigotheutis species. As biogenic structures such as sponge stalks are biodiversity hotspots in the region, we hypothesise this behaviour to be a form of aggressive mimicry. Using disguise to lure prey is possibly efficient in the abyss, where food may be scarce. Alternatively, the observed behaviour may be a defence strategy against predators. This finding highlights the importance of imagery in the understanding of cephalopod behaviour and further expands the diversity of their complex adaptations to life in the abyss. Being one of the few groups thought to be purely predatory in the abyss, with an important role in benthopelagic trophic food-webs, further investigations must specifically target cephalopods. This knowledge is urgently required to protect the whole functional complexity and biodiversity of the abyssal Pacific seabed from human impacts such as ocean acidification from climate change and seabed mining, which are likely to reach these remote habitats in the years to come.
Emergency Response in Rocky Intertidal Ecosystems Using Drone Imagery: A MARINe Initiative
Karah Cox-Ammann , University of CA at Santa Cruz
Alexis Necarsulmer UCSC
The Multi-Agency Rocky Intertidal Network (MARINe) collaborates to monitor 200+ rocky intertidal sites along the West Coast of North America. Recently, MARINe has integrated drone imagery with on-the-ground monitoring to bolster surveying and mapping capabilities, particularly for emergency response. Given the increasing frequency of emergencies in ecological systems due to climate change effects, our objective is to showcase the considerable utility of drones as a potent tool for enhancing emergency response capabilities. By focusing on rocky intertidal ecosystems, we aim to highlight the versatility and effectiveness of drone technology in swiftly and effectively responding to ecological emergencies. We have employed drones to respond to various emergency scenarios, including oil spills, landslides, wildfire-induced debris flows, and sea level rise impacts. Drones enable us to capture baseline imagery, track sediment movement, access unsafe areas under landslides, and create high-resolution orthomosaics and 3-D models of affected sites. Our surveys have demonstrated the effectiveness of drones in capturing baseline data before impacts and tracking alterations to habitat over time. The high-resolution imagery produced by drones allows for habitat classification and assessment of impact zones, thus informing resource managers and guiding mitigation efforts. The integration of drones into emergency response strategies in rocky intertidal ecosystems offers several benefits, including rapid and repeatable monitoring, access to remote and otherwise inaccessible areas, and detailed assessment of impacts. This initiative has implications for agencies, such as NOAA, CDFW, and National Marine Sanctuaries, involved in coastal management and conservation efforts, particularly those involved in mitigating the impacts of emergencies within critical habitats and affecting sensitive species, such as the endangered black abalone (Haliotis cracherodii).
Enabling Large Scale Coral Reef 3D Models with Accurate Color Representation using DeepSeeColor
Yogesh Girdhar , Woods Hole Oceanographic Institution
Daniel Yang, John Walsh, Stewart Jamieson
Successful applications of complex vision-based behaviours underwater have lagged behind progress in terrestrial and aerial domains. This is largely due to the degraded image quality resulting from the physical phenomena involved in underwater image formation. Spectrally-selective light attenuation drains some colors from underwater images while backscattering adds others, making it challenging to perform vision-based tasks underwater. These distortions make image observation depend on distance, beyond simply resolution and scale change. We present DeepSeeColor algorithm, which is a state-of-the-art underwater image formation model with the computational efficiency enabled by modern deep learning. The proposed approach enables efficient color correction for underwater imagery enabling novel applications such as accurate in-situ imaging using AUVs and and scaling up building of underwater 3d models with accurate color representation. We show the application of the approach to coral reef survey imagery collected by an underwater robot.
End-to-end processing of water column video data
Ashley N. Marranzino , University Corporation for Atmospheric Research | NOAA Ocean Exploration Affiliate
Adrienne Copeland (NOAA Ocean Exploration), Michael Ford (NOAA), Jason Gronich (Western Washington University)
The National Oceanic and Atmospheric Administration (NOAA)’s office of Ocean Exploration is dedicated to exploring the unknown ocean, unlocking its potential through scientific discovery, technological advancements, partnerships, and data delivery. One of the ways in which NOAA Ocean Exploration leads national efforts to fill gaps in our basic understanding of the marine environment is by conducting exploration missions with NOAA Ship Okeanos Explorer and the remotely operated vehicles (ROVs) Deep Discoverer and Seirios. In 2014, NOAA Ocean Exploration began using ROV dives to explore the poorly documented pelagic fauna of the deep oceanic water column, working with water column biologists and taxonomic experts to determine exploration priorities and strategies. Over the past decade, 51 water column exploration dives have been conducted aboard NOAA Ship Okeanos Explorer, resulting in more than 175 hours of video data archived with the National Centers for Environmental Information (NCEI) and made available to the public through the Ocean Exploration Video Portal and SeaTube V.3. These videos can give scientists an unprecedented view of water column fauna - providing valuable information on species distributions, abundances, and behaviors. However, reviewing the raw video footage is time consuming and can require specific taxonomic expertise, thereby limiting the ultimate usability and accessibility of the data to the larger scientific community. For over 6 years, NOAA Ocean Exploration has trained undergraduate interns and partnered with taxonomic experts to develop a workflow for annotating, verifying, and analyzing water column exploration video data. Here we present the end-to-end workflow that NOAA Ocean Exploration has developed for water column exploration - including data collection during ROV dives, real-time and post-cruise annotation processes, data analysis and publication, and ongoing efforts using these data to feed into machine learning algorithms.
Exploring Oceans of Visual Data: Integrating Deep Learning Computer Vision and Information Visualization for Large-Scale Ecological Assessment of Cold-Water Coral Reefs
Daniel Langenkämper , 1. Biodata Mining Group, Faculty of Technology, Bielefeld University, Germany
Ingunn Nilssen Aksel A. Mogstad (Environmental Technology, Technology, Digital & Innovation, Equinor ASA, Norway), Tim W. Nattkemper (Biodata Mining Group, Faculty of Technology, Bielefeld University, Germany)
Growing amounts of marine video and image data are collected using different kinds of camera setups and technologies (e.g., ROV, AUV, stationary platforms) to explore and monitor marine habitats. Concurrently, ambitious ocean digital twin (ODT) projects aim for the integrative collection and management of multi-modal sensor data in large data bases. The aim on ODT development is to make use of this data for analysis, visualization and modeling, together with establishing feedback loops into the marine system by providing documentation for decision makers. However, there are still few clearly defined workflows from raw marine image (or video) data to real data products required for such ODTs. Progresses in artificial intelligence and computer vision (like convolutional neural networks) have increased the potential of computational approaches to (semi-) automatic marine image analysis significantly and the number of papers describing their applications for marine object classification or segmentation is growing year by year. Usually, the works end up with an accuracy assessment and a comparison with competing algorithms, but a gap to a real data product remains. In this contribution, we address the problem of large-scale assessments of cold-water coral reefs based on ROV-collected video data. These assessments are still performed by human operators evaluating hundreds of video files, sometimes with sub-optimal visibility. We present a workflow to estimate the coverage of Desmophyllum pertusum, Paragorgia arborea, other gorgonians and sponges on reefs based on semi-automatic annotation of less than 0.0003% of the available data, training a segformer b3 model and applying it to automatically segment 855 ROV videos from the Haltenbanken area (Norwegian Sea). The results are displayed in a web-based customized visual interface that supports an integrative analysis of the computed coverage data together with data and manual evaluations derived from operator reports.
Framing image preprocessing to build capacity and efficiency in benthic ecology
Loïc Van Audenhaege , Ocean BioGeosciences, National Oceanography Centre, European Way, Southampton, UK
Eric Orenstein, Ocean Informatics, National Oceanography Centre, European Way, Southampton, UK Mojtaba Masoudi, Ocean BioGeosciences, National Oceanography Centre, European Way, Southampton, UK Tobias Ferreira, Ocean Informatics, National Oceanography Centre, European Way, Southampton, UK Colin Sauze, Ocean Informatics, National Oceanography Centre, European Way, Southampton, UK Jennifer Durden, Ocean BioGeosciences, National Oceanography Centre, European Way, Southampton, UK
For the past decades, benthic ecologists have been leveraging images for quantitative surveys. Image-based benthic surveys are typically conducted with static or mobile cameras collecting images at a fixed rate over various types of seabed habitat. The images are then manually or automatically annotated, from which descriptors of density and biodiversity are derived. Image preprocessing is a critical step to consider in imaging workflows; it has an enormous impact on downstream ecological insights. However, preprocessing is an underappreciated element of imaging studies, especially as more attention has been paid to AI-based annotation techniques. There are no widely used set of best practices nor a decision framework supporting the definition of an adequate image preprocessing workflow, constrained by research goals and image sample quality. As a result, image sample preparation is usually done on an ad hoc basis and is conducted with a set of discontinuous software tools for image colour correction, overlap management and standardisation of the sampling effort (among many: manual, ImageJ, Python, R, MATLAB, Photoshop). The complexity of decision making, coupled with inconsistent preprocessing tools, undermines a user’s ability to design optimal workflows to return robust and reproducible ecological measurements. The lack of a holistic, codified approach introduces bias and makes comparison between studies difficult or impossible, undermining efforts to assess ecological dynamics at regional to global scales. The PAIDIVER project aims to tackle this issue by providing decision making resources and software tools to support image preprocessing, specifically targeting consistent computation of biodiversity indicators. The tools are designed to support reproducibility by outputting processing steps as metadata. PAIDIVER will help benthic image users design and implement an image preprocessing workflow by supporting data standardisation (e.g. seabed area calculation) and optimising image and sample selection (e.g. quality and overlap management).
From Research to Operational and Sustained Marine Imaging for Place-Based Management
Henry A Ruhl , Monterey Bay Aquarium Research Institute
HA Ruhl, Monterey Bay Aquarium Research Institute; JH Adelaars, Monterey Bay Aquarium Research Institute; CR Anderson, University of California San Diego; FL Bahr, Monterey Bay Aquarium Research Institute; R Bochenek, Axiom Data Science; FP Chavez, Monterey Bay Aquarium Research Institute; P Daniel, University of California Santa Cruz; CA Durkin, Monterey Bay Aquarium Research Institute; J Erickson, Monterey Bay Aquarium Research Institute; SHD Haddock, Monterey Bay Aquarium Research Institute; B Jones, Monterey Bay Aquarium Research Institute; K Katija, Monterey Bay Aquarium Research Institute; D Klimov, Monterey Bay Aquarium Research Institute; RM Kudela, University of California Santa Cruz; T Maughan, Monterey Bay Aquarium Research Institute; PLD Roberts, Monterey Bay Aquarium Research Institute; A Schnittger, Monterey Bay Aquarium Research Institute; AE West, Monterey Bay Aquarium Research Institute.
Biology and ecosystem information can now be readily collected across a wide range of organism sizes in conjunction with physical and biogeochemical oceanographic data. Biology and ecosystem variables are at the heart of many statutory monitoring requirements for the marine environment including for protected species and understanding natural hazards, to managing fisheries, offshore energy, and mineral industries. Synchro—launched in 2023—accelerates technology solutions for ocean research and monitoring. Its aim is to propel promising ocean observing technology from prototype to broader use and involve users of the technology from the get-go. Marine imaging is one of the focus areas of Synchro, where we are testing and evaluating various systems. These evaluations importantly include perspectives from not only scientists, but also tech developers and the users of resulting information for decision making. Additionally, many regions of the Integrated Ocean Observing System (IOOS) are working to transition imaging and artificial intelligence tools into sustained observing systems that meet Federal accreditation standards. Here we will cover cases of such transitions from research to operations (R2O) and the data lifecycle elements needed to realize broader imaging tech use. For example, a system in mid-stage technology readiness (TRL) is the Planktivore camera operating on a Long-Range Autonomous Underwater Vehicle (LRAUV). An advanced TRL system example will be discussed with the California Imaging FlowCytobot (IFCB) Network. Such examples will be given in the context of embedding such systems into sustained ocean observing efforts to improve coastal and climate resilience and inform place-based management in changing ecosystems. Cyberinfrastructure tools from FathomNet and Axiom Data Science contribute to timely access to this information. Applications include information to underpin Large Marine Ecosystem - Integrated Ecosystem Assessments, National Marine Sanctuary - Condition Reports, baseline and impact assessment for development of the offshore wind energy industry, and providing real-time information on harmful algal blooms.
Gradients of habitat complexity and diversity shape the megafaunal communities associated with cold-water coral reefs of the Blake Plateau (Southeast USA)
Ryan Gasbarro , UC Santa Cruz; NOAA Fisheries
Erik Cordes (Temple University); Morgan Will (Temple University)
The Blake Plateau of the southeast United States continental margin contains over 50,000 carbonate mounds formed primarily by skeletal build-up of the cold-water coral Lophelia pertusa, making it the world’s largest deep-sea coral reef province. These reefs contain invertebrate assemblages that rival the biodiversity and density of tropical shallow-water reefs, but little is known how the diversity or abundance of assemblages are modulated by gradients of habitat complexity created by live and dead-standing coral cover. Here, we characterize the functional and taxonomic diversity of invertebrate megafaunal and macrofaunal assemblages present from 11 submersible dives transecting L. pertusa reefs on the Blake Plateau using high-resolution submersible imagery and physical collections, respectively. Then, we assess patterns in alpha- and beta-diversity metrics across bathymetric, temperature, and biocomplexity gradients present across the samples to test the following hypotheses: (i) Taxonomic diversity has a positive but saturating relationship with live coral cover, necessitating VME designation for areas with coral cover above a threshold value. (ii) In addition to habitat complexity as represented by live and dead standing coral cover, faunal diversity will increase with habitat diversity (i.e. the presence of multiple substrata with varying densities). (iii) Diversity has a negative relationship with depth in accordance with macroecological theory that posits food availability from surface export as a driver of biodiversity in the deep ocean. Overall, this study augments baseline information on the CWC reefs of the SEUS and tests for mechanisms underlying their biodiversity and community structure, which will aid in conservation efforts that seek to assess and map CWC reef VME status in this region and throughout the global ocean.
High-resolution multispectral imaging of deep sea corals
Irene Hu , MBARI
Steve Litvin, MBARI (USA), litvin@mbari.org Paul Roberts, MBARI (USA), proberts@mbari.org Aaron Schnittger, MBARI (USA), aschnittger@mbari.org Jim Barry, MBARI (USA), barry@mbari.org
Multispectral imaging, which provides spectral reflectance information for each pixel of a two dimensional image, is a promising technique for studying the health and functionality of marine organisms. However, despite growing use in a variety of marine applications, its use in deep sea studies has been limited due to technical challenges. Here, we present the development of a macro-multispectral camera designed for the study of deep sea corals. High resolution reflectance information at multiple spectral bands can provide information on coral health, including potentially aiding in delineation and quantification of gametes, or providing information on the degradation state of corals. Thus, this approach presents a new tool for advancing research and monitoring of deep sea corals. The system uses high powered LEDs for illumination and a monochrome camera as a detector, to provide high-resolution images of reflectance at eight different wavelengths ranging from 370 nm to 940 nm. Two iterations of the system have been developed: a laboratory version which can be used in controlled studies of known targets, and a deep sea version for in situ deployment by ROV. The laboratory system utilizes a 25 MP camera capable of streaming images at 17 fps, and a total of 80 LEDs. The ROV system utilizes a 65 MP camera capable of streaming images at 35 fps, and a total of 500 LEDS. Both systems provide a field of view on the order of 5 cm and resolution on the order of 5-10 um/px, and image at a distance of 20-30 cm from the object. In this poster, we share instrument details as well as examples of collected data.
Image curation in action – from the deep sea to the community
Sophie V. Schindler , ^1GEOMAR Helmholtz Centre for Ocean Research Kiel
Henk-Jan Hoving^1, Jan Dierking^1, Julian B. Stauffer^1, Karl Heger^1 & Timm Schöning^1
Imaging has become a powerful marine scientific technique in the past decades to observe and monitor ocean environments like the sea floor or the open water column. Modern optical gears enable us to access the most remote ocean areas and continuous technological advances lead to exciting new possibilities and discoveries – but also to new challenges in handling the increasingly larger volumes of collected data. Marine imaging often involves multi-terabyte dataset volumes which require adequate storage and server hardware and can limit curation and management during field work. Other challenges include structuring and standardizing of on-site workflows and efficient and sustainable curation of gathered images. Here, we visualize a successful and efficient on-site image curation protocol, and identify challenges and suggestions for improvement. During the research expedition MSM126 “JellyWeb Madeira ” focusing on biodiversity and habitat exploration in the waters surrounding Madeira (Feb-March 2024), we gathered more than 14 TB of still images and video data, which were collected by ROV PHOCA and the two towed camera platforms XOFOS (sea floor surveys) and PELAGIOS (midwater surveys). Pre-cruise and pre-deployment preparation covered an assessment of used gear and the usage and ongoing development of Best Practice Protocols ensuring the quality of technical settings as well as metadata collection. After each deployment, back-ups were created by transferring image data to a transportable MAMS (media access storage system) and archiving data on the ship-internal server. The ELEMENTS media software system was then used to manage the images on the MAMS. The GEOMAR Data Science Unit developed software for sustainable image curation, including a standardized workflow structure and creation tool for FAIR (findable, accessible, interoperable and reusable) digital objects for images (iFDOs ). These efforts aimed to have image data ready-to-use after fieldwork and make the curated image data quickly accessible and publishable for researchers.
Imaging as an emergency need to understand jellyfish ecology and population dynamics along the coastal region of Kribi
Gisele Flodore Ghepdeu Youbouni , Specialized Research Station for Marine Ecosystems
Specialized Research Station for Marine Ecosystems
Jellyfish proliferations are increasing over the world with important socio-economic, environmental and health impacts. In many regions like the West/Central Africa, insufficient data are available on jellyfish which stand among the least studied marine zooplankton group. In Cameroon, jellyfish research started in 2017 under an ongoing PhD research work has yielded first reports about jellyfish taxonomy, diversity and ecology. This research has enabled the identification of four main jellyfish taxa including Catostylus, Chimaerus, Chrysaora and Cyanea, as well as their seasonality and the physicochemical characteristics of their living environments. However, jellyfish are species with metagenetic life cycle and our research focused only on the mobile phase. The vision of this research was to improve jellyfish literacy and establish a jellyfish industry in the region as a solution to the huge jellyfish discards in the region. Therefore, for the sake of sustainability, information regarding the polyp stage of jellyfish, jellyfish habitat as well as their relationship with other marine species especially fish and sea turtles is highly recommended. To achieve this purpose, marine imaging, which also represents the main method to study the polyp stage in their natural environment, is highly recommended. This talk is going to give a preview of jellyfish taxonomy, diversity, ecology, and presents the challenges link to the study of their habitat and of a better understanding of and role in the food web.
Imaging marine snow: Hidden Comet-Tails of Marine Snow Impede Ocean-based Carbon Sequestration
Manu Prakash , Stanford university
Rahul Chajwa (Stanford), Eliott Flaum (Stanford), Kay D. Bidle (Rutgers), Benjamin Van Mooy (WHOI), Manu Prakash (Stanford)
Global carbon-cycle on our planet ties together the living and the non-living world, coupling ecosystem function to our climate. Gravity driven downward flux of carbon in our oceans in the form of marine snow, commonly referred to as biological pump directly regulates our climate. Multi-scale nature of this phenomena, biological complexity of the marine snow particles and lack of direct observations of sedimentation fundamentally limits a mechanistic understanding of this downward flux. The absence of a physics based understanding of sedimentation of these multi-phase particles in a spatially and temporally heterogeneous ocean adds significant uncertainty in our carbon flux predictions. Using a newly invented scale-free vertical tracking microscopy, we measure for the first time, the microscopic sedimentation and detailed fluid-structure dynamics of marine snow aggregates in field settings. The microscopically resolved in-situ PIV of large number of field-collected marine snow reveals a comet tail like flow morphology that is universal across a range of hydrodynamic fingerprints. Based on this dataset, we construct a reduced order model of Stokesian sedimentation and viscoelastic distortions of mucus to understand the sinking speeds and tail lengths of marine snow dressed in mucus. We find that the presence of these mucus-tails doubles the mean residence time of marine snow in the upper ocean, reducing overall carbon sequestration due to microbial remineralization. We set forth a theoretical framework within which to understand marine snow sinking flux, paving the way towards a predictive understanding of this crucial transport phenomena in the open ocean.
Innovative Approaches to Great Lakes Benthic Mapping: Integrating AUV, ROV, and Drop Camera Data with Tator
Ayman Mabrouk , NOAA
Charles Menza and Tim Batista/NOAA
The Great Lakes, a vital freshwater resource, require precise and comprehensive benthic mapping to support ecological research, conservation efforts, and resource management. This presentation explores the innovative methodologies employed in the Great Lakes benthic mapping project, focusing on the integration of Autonomous Underwater Vehicles (AUVs), Remotely Operated Vehicles (ROVs), and drop camera systems. By leveraging the powerful video annotation platform Tator, we have enhanced the accuracy and efficiency of lakebed image analysis. Our approach combines high-resolution imagery from multiple sensors to create a detailed and comprehensive map of the lakebed. The AUVs provide extensive coverage and high-resolution data, while the ROVs offer maneuverability and detailed close-up imaging. Drop cameras complement these systems by capturing additional visual data in areas that are challenging to access with AUVs and ROVs. Tator’s robust annotation capabilities enable us to efficiently process and analyze vast amounts of video data, facilitating the identification and classification of benthic habitats and features. This multi-sensor integration, combined with advanced annotation techniques, offers a new paradigm in lakebed mapping, significantly improving our understanding of the Great Lakes' underwater environment. In this presentation, I will discuss the methodologies used, the challenges encountered, and the solutions implemented to overcome these challenges. I will also showcase preliminary results and highlight the potential applications of this innovative approach in future benthic mapping and marine research projects.
Integrating Deep Learning for Multi-Species Detection in Marine Conservation: Combining Acoustic and Visual Data
Artash Nath , MonitorMyOcean.com
Increased marine traffic, overfishing, and climate change have caused the marine mammal population to decline. The International Union for the Conservation of Nature (IUCN) has identified 25% of marine mammals at risk of extinction. Measures to reverse their declining numbers require updated data about their population, shifting habitats and migration patterns. Marine mammals use sound to communicate and navigate. Analysis of hydrophone data can be used to detect the presence of marine mammals and improve animal abundance estimates. Unfortunately, current algorithms for identifying marine mammal calls are limited to a singular species. It slows the analysis and provides an incomplete measure of marine mammal diversity. This research establishes a 2-stage model to simultaneously detect multiple marine mammal species. The first stage uses supervised learning to identify known stereotyped contact calls. The second uses unsupervised neural networks to identify probable marine mammal calls on which the model has not been trained. The supervised learning stage uses a deep convolutional neural network currently trained on over 5000 labelled spectrogram samples of known calls from blue and humpback whales. The algorithm is modular and can be expanded by training it on the collection of marine life sounds gathered by the upcoming Global Library of Underwater Biological Sounds (GLUBS) initiative. The unsupervised stage involves training Convolutional Variational Autoencoders on spectral regions of interest (ROI). The trained encoder is used to transform ROIs into low-dimensional feature space representations, group possible bioacoustic signals, and identify outliers. The resulting groupings of bioacoustic signals can be annotated, enabling the model to re-train itself for higher accuracy. In addition to acoustic data, this research integrates visual data from the FathomNet underwater camera image dataset. By training custom Region-Based Convolutional Neural Networks (RCNNs) on this dataset, the model can detect various classes of both known and unknown aquatic creatures with over 85% accuracy weighed across 18 aquatic creature classes. This combined approach of acoustic and visual data enhances the detection accuracy and provides a comprehensive understanding of marine biodiversity. When used on the Monterey Bay Aquarium Research Institute's hydrophone data pipeline, the model could scan 24 hours of hydrophone data in under 15 minutes on a single GPU and identified blue and humpback whale calls with a 98.7% accuracy. This research can accelerate data analysis, allowing policymakers to develop informed conservation strategies, such as altering shipping lanes, adjusting fishing quotas, and creating protected areas. By linking visual and acoustic data, this integrated approach offers a more robust and comprehensive method for monitoring and protecting marine mammals, ensuring the conservation of both well-known and lesser-known species.
Lessons learned in the application of deep-learning AI to benthic image analysis
Kerry Howell , University of Plymouth
Amelia Bridges (University of Plymouth), Kyran Graves (University of Plymouth), Giulia La Bianca, David Moffat (Plymouth Marine Laboratory), Nils Piechaud (Norwegian Institute of Marine Research), Antony Raymont (University of Plymouth), other members of the Big Picture Group.
Image based platforms are now firmly cemented into the future of marine benthic biodiversity survey and monitoring. The datasets delivered by these platforms are vast and there is a widely recognised problem in the time taken to process such data to extract biological information. Deep learning artificial intelligence holds significant promise in helping to break the bottleneck of manual biological image analysis. But what exactly can it deliver in practical terms? how should training data be organised to deliver the best performing models, and what are some of the challenges in the outputs produced? We report on multiple studies within the Plymouth University and Plymouth Marine Laboratory teams that assess the performance of CNNs trained on benthic annotated datasets using different approaches to animal classification. Specifically, we focus on two studies that have considered single classes (biota), vs taxonomy at the level of phyla, vs morphological classification. The first study used imagery from a single platform and a single research cruise. The second collated annotated imagery from across UK academic and government organisations to provide a large and highly diverse training dataset. Multiple models were trained on 70% of the annotated data and validated on 30%. Precision and recall were compared between models to assess performance. Results suggest that taxonomic based groupings may not be optimal for delivering the best performance from models. Model performance metrics were distorted due to issues with models detecting the same animal multiple times and detecting animals missed by human observers. Our research highlights the need to carefully consider the classification system applied in model training as well as the need to manually check model validation results to better understand where models fail / succeed. Our research will inform the UK government future monitoring strategy and the role AI could play in this area.
Low-Cost Stereo Imaging for In Situ Particle Sinking Speed Measurement on Autonomous Platforms
Fernanda Lecaros , MBARI (Monterey Bay Aquarium Research Institute)
Paul Roberts, MBARI (USA) Joost Daniels, MBARI (USA) Melissa Omand, University of Rhode Island (USA Kakani Katija, MBARI (USA) Colleen Durkin, MBARI (USA)
The sinking speed of carbon particles in aquatic environments plays a crucial role in the Earth's carbon cycle and affects the efficiency of the ocean carbon pump, which transports carbon dioxide from the atmosphere to the deep ocean. Changes in sinking speed could impact this pump's effectiveness, ultimately influencing atmospheric carbon dioxide levels and, by extension, global climate dynamics. Measuring in situ particle sinking speed remains a significant challenge due to the small size and slow movement of particles, potential platform motion, and power limitations for camera and lighting operations. We present a low-cost imaging system designed for attachment to an isopycnal-following float, capable of capturing images of sinking particles in the water column. This dual-camera, dual-strobe system uses stereo imaging techniques, tracking the motion of sinking particles, and precisely calculates their 3-D velocities within a specified volume of the water column. The imaging constraints have been extensively tested in simulations. Controlled tank experiments were used to image both synthetic particles and natural aggregates, demonstrating the instrument’s capability to quantify both particle size and sinking speed using a free-space imaging approach. Field measurements in a particle-rich coastal environment and results will be discussed. Through this innovation, we aim to enhance our understanding of particle transport and their significance in ecosystem dynamics and the global carbon cycle.
MINUET: Harmony from the deep - an educational and machine learning development project that generates scientific data
Deanna Soper (dsoper@udallas.edu), University of Dallas)
Abigail Fritz (afritz@udallas.edu, University of Dallas), Jana Rocha (jrocha@udallas.edu, University of Dallas), Benjamin Woodward (benjamin.woodward@cvisionai.com, CVision AI), Kakani Katija (kakani@mbari.org, MBARI)
Recent efforts of the NOAA Office of Oceanic Exploration and Research (OER) have yielded substantial quantities of underwater footage recorded by Remotely Operated Vehicles (ROVs), which currently takes a significant amount of time in order analyze. The application of artificial intelligence (AI) has the potential to increase efficiency, reduce costs, and curate large biological data sets. In order to develop software that has the capability of organism detection and identification we utilized undergraduate students to assist in AI training data set development and biological analysis. Undergraduate students used previously expertly annotated ROV video footage to localized organisms by drawing a box around the individual and then connected the expert taxonomic identification to the organism. This had two outcomes: the generation of a biological data set that has revealed deep-sea coral abundance, distribution, and epibiont association patterns and resulted in a training data set for AI development. Here we show that instructors, software developers, and scientists at research institutes can form close partnerships to provide technological development, education, and hands-on training for undergraduate students. These important experiences for students early in their academic careers provide them with the skills needed to engage them in the fields of oceanic research and data science.
Making use of unlabeled data: Comparing strategies for marine animal detection in long-tailed datasets using self-supervised and semi-supervised pre-training
Tarun Sharma , Caltech
Danelle E. Cline (MBARI) and Duane Edgington (MBARI)
This paper discusses strategies for object detection in marine images from a practitioner’s perspective working with real-world long-tail distributed datasets with a large amount of additional unlabeled data on hand. The paper discusses the benefits of separating the localization and classification stages, making the case for robustness in localization through the amalgamation of additional datasets inspired by a widely used approach by practitioners in the camera-trap literature. For the classification stage, the paper compares strategies to use additional unlabeled data, comparing supervised, supervised iteratively, self-supervised, and semi-supervised pre-training approaches. Our findings reveal that semi-supervised pre-training, followed by supervised fine-tuning, yields a significantly improved balanced performance across the long-tail distribution, albeit occasionally with a trade-off in overall accuracy. These insights are validated through experiments on two real-world long-tailed underwater datasets collected by MBARI.
Marine Image Analysis with RootPainter
H. Poppy Clark , University of Aberdeen
Abraham George Smith (Department of Computer Science, University of Copenhagen, Universitetsparken 1, 2100 Copenhagen, Denmark) Daniel Mckay Fletcher (Rural Economy, Environment and Society, Scotland’s Rural College, West Mains Road, Edinburgh, EH9 3JG, Scotland, United Kingdom) Ann I. Larsson (Tjärnö Marine Laboratory, Department of Marine Sciences, University of Gothenburg, Strömstad, Sweden) Marcel Jaspars (Marine Biodiscovery Centre, Department of Chemistry, University of Aberdeen, Old Aberdeen, AB24 3DT, Scotland, United Kingdom) Laurence H. De Clippele (School of Biodiversity, One Health & Veterinary Medicine, University of Glasgow, Bearsden Road, G61 1QH, Scotland, United Kingdom)
RootPainter is an open-source and user-friendly interactive machine learning tool that is capable of analysing large marine image datasets quickly and accurately. This study tested the ability of RootPainter to extract the presence and surface area of the cold-water coral reef associate sponge species, Mycale lingua, in two datasets: 18,346 time-lapse images and 1,420 remotely operated vehicle video frames. New corrective annotation metrics integrated with RootPainter allow objective assessment of when to stop model training and reduce the need for manual model validation. Three highly accurate Mycale lingua models were created using RootPainter, with an average dice score of 0.94 ± 0.06. Transfer learning aided production of two of the models, increasing analysis efficiency from 6 to 16 times faster than manual annotation for time-lapse images. Surface area measurements were extracted from both datasets allowing future investigation of sponge behaviours and distributions. The ability of RootPainter to accurately segment the cold-water coral Lophelia pertusa was also tested in 3,681 ROV video frames, with model dice scores surpassing 0.80. Moving forward, interactive machine learning tools and model sharing could dramatically increase image analysis speeds, collaborative research, and our understanding of spatiotemporal patterns in biodiversity.
Microscopy in Motion in the Deep Ocean
Thom Maughan , MBARI
Paul L.D. Roberts, Brent Jones, Aaron Schnittger, Jon Erickson, Denis Klimov, Henry Ruhl, Steve Haddock, Francisco Chavez
Plankton are the foundation of the ocean food web, but observations are difficult to sustain at ecologically relevant time and space scales. Advances in in-situ imaging are expanding the toolkit of biological oceanographers, enabling better quantification of plankton communities as well as detrital particles that together drive the ocean’s carbon cycle. This paper/talk presents the engineering of "Planktivore," a deep-sea microscope for autonomous underwater vehicles (AUVs) and lessons learned in our first half year deploying as an ocean instrument for science on the MBARI Long Range Autonomous Underwater Vehicle (LRAUV). Planktivore is a free space imaging dual microscope that spans a size range from 10s of microns to centimeter size with vehicle speeds up to 1 meter per second. Storage of images is optimized using computer vision to identify and store regions of interest (ROI) from the full image frame. Weeklong LRAUV deployments in Monterey Bay typically record 10 Million to 30 Million plankton/particle ROIs. After deployment the images are uploaded and processed on a GPU enabled server and software tools are used to tag plankton images with their scientific name for use in machine learning for automated plankton identification.
North Sea 3D – Image analysis for measuring biofouling biomass from industry ROV footage
John Halpin , Scottish Association for Marine Science
Joseph Marlow1, and Tom A. Wilding1 1Scottish Association of Marine Science
Continued developments in marine infrastructure have given rise to an increasing number of Offshore man-made structures (MMS). This has become known as ‘ocean sprawl’. These structures are quickly colonised, becoming artificial reefs. To understand the role of these structure in marine systems, the colonising organisms need to be quantified, both by species and biomass. To date, this quantification has been limited by the difficulties in performing the studies on the scales needed. In the North Sea 3D (NS3D) project we use existing footage collected by offshore energy operators with remotely operated vehicles (ROVs). This footage already exists in abundance since structures are routinely surveyed by ROV for maintenance purposes. We use recent advances in image/video processing - Structure from Motion (SfM) Photogrammetry and semantic segmentation by convolutional neural networks (CNNs) to generate 3D images of marine growth classified by taxa from standard 2D-video ROV footage. These are then calibrated and used for biomass quantification. In this talk, the methodology of the associated image analysis is discussed, including the challenges of underwater photogrammetry and species auto identification, particularly when ROV footage is not taken specifically for this task. A 3D headset will accompany the talk, letting attendees ‘walk around’ the 3D dataset’s we have generated.
OCEAN SPY’s citizen annotation data
Catherine Borremans , IFREMER (Biology and Ecology of dEEP marine ecosystems Unit (BEEP)/Deep Sea Lab)
Marjolaine Matabos (IFREMER), Pierre Cottais (IFREMER), Julie Tourolle (IFREMER), Antoine Carlier (IFREMER), Séverine Martini (MIO)
For the purpose of assessing and monitoring conservation state of marine ecosystems, scientists are deploying seabed observatories and using mobile devices to acquire temporal and spatialized biodiversity data using imagery, to complement traditional 'stationary' sampling approaches. These monitoring programs make it possible to gather a large amount of data, particularly underwater images, which represents huge volume of datasets that are difficult to process. Artificial intelligence (AI) has enabled the development of algorithms to facilitate the processing of large datasets. However, the ability of machines to detect and classify objects automatically for scientific purposes depends on a learning phase based on large reference databases that have been built manually by human brains, a highly time-consuming task. In this context, involving citizens in collecting training data is an interesting solution, thanks to the important observation power that crowd represents (Matabos et al., 2024, Preprint). Launched in 2023, the Ocean Spy project aims at engaging citizen in the process of marine images annotation. It provides a web platform allowing general public to access images collected in various marine habitats, from shallow to deep waters. Different tools and functionalities were designed to guide users for locating and identifying animals (or other subjects of interest) in the images, through the annotation interface. Beyond the development of the online portal and the associated database, such an initiative requires methods for pre-processing and validating the data generated by citizen annotation. Particularly, to identify common organisms (resulting from multiple annotations) and to clean the database according to a threshold of agreement between participants. These are essential steps in the production of reliable, high-quality scientific knowledge, and in boosting the performance of AI methods for identifying taxa. General statistics on this first year of operation and the data analysis pipeline being fine-tuned will be presented at the workshop. Ref: Matabos, Marjolaine and Cottais, Pierre and Leroux, Riwan and Cenatiempo, Yannick and Gasne-Destaville, Charlotte and Roullet, Nicolas and Sarrazin, Jozée and Tourolle, Julie and Borremans, Catherine, Deep Sea Spy: An Online Citizen Science Annotation Platform for Science and Ocean Literacy. Available at SSRN: https://ssrn.com/abstract=4848325 or http://dx.doi.org/10.2139/ssrn.4848325
Object Detection of Benthic Morphospecies using Synthetic Data and Self-Training Domain Adaptation
Heather Doig , University of Sydney
Oscar Pizarro (Norwegian University of Science and Technology, University of Sydney); Stefan Williams (University of Sydney)
Scientists can observe benthic morphospecies using images taken from underwater robotic vehicles. The presence and quantity of morphospecies like endangered handfish or invasive sea urchins can be collected using neural network object detectors on high-resolution images. The detector is trained with labelled images, but its performance can suffer from two issues. First, the detector can degrade when used on images captured from different cameras, water conditions, locations or time compared to the source training images, known as domain shift. Secondly, very few labelled examples of the morphospecies of interest needed to train a supervised model may exist. We propose a framework to train a benthic morphospecies detector using synthetic underwater images followed by unsupervised domain adaptation to bridge the domain gap from synthetic to real. We generated synthetic images with a customised version of Infinigen (infinigen.org) from Princeton Vision & Learning Lab that creates natural scenes with annotation data for training. We have customised Infinigen to mimic images from underwater robots, modelling the attenuation of light and including species of interest, such as invasive sea urchins. We train an object detector using synthetic source images followed by an adaptation step with unlabeled target images. We tested our approach on sea urchins and other species of interest and successfully detected these species in visually complex seafloor images. The framework offers a novel approach for detecting rare and endangered morphospecies using increasingly popular 3D modelling software.
Ocean Observatories Initiative Imaging Options and Expansion Capabilities
Michael F. Vardaro , University of Washington, School of Oceanography
Katharine T. Bigham & Deborah S. Kelley, University of Washington, School of Oceanography
The NSF’s Ocean Observatories Initiative (OOI) Regional Cabled Array (RCA) has been streaming real-time data to shore from a diverse array of 150 instruments since 2014. The electro-optical cable network spans the Juan de Fuca tectonic plate. One backbone cable extends ~480 km west of the Oregon coast to Axial Seamount, an active undersea volcano, and a southern branch runs 208 km along the base of the Cascadia Subduction Zone (2900 m depth) and then turns east towards the Oregon Shelf (80 m depth), crossing the Cascadia Margin. Among the cabled instruments are six SubC digital cameras that capture 3 images every 30 minutes at various seafloor and water column sites. An HD Video camera is aimed at the 300 °C Mushroom vent in the ASHES hydrothermal field within the highly dynamic caldera of Axial Seamount, located 300 miles offshore (1500 m depth). The HD camera alone has collected 47,000 hours of video, imaging the entire venting edifice every 3 hours. All OOI-RCA imagery is open access and available for viewing and downloading via the recently upgraded OOI Data Explorer image and video gallery (dataexplorer.oceanobservatories.org). Since 2016, the OOI-RCA submarine observatory has also hosted a diverse array of cabled instruments through externally funded PI projects, including multibeam sonars and a 4K camera installed at Southern Hydrate Ridge by a group from MARUM to study methane bubble plume dynamics, environmental impacts on venting, and chemosynthetic communities (Marcon et al., G3, 2021). The RCA power and bandwidth support extensive expansion capabilities, and NSF encourages the addition of instruments and platforms onto the observatory. We continue to improve our image gallery annotation and quality control procedures, and several active projects are using automated event detection, machine learning, and computer vision techniques to sort, label, and sift data out of the OOI image archive.
PAMS, a low-cost, simple-to-assemble seafloor imagery solution for monitoring corals
Thierry Baussant , NORCE Norwegian Research Centre
Christian Andreas Hansen, Alan Le Tressoler, Junyong You, Gro Fonnes, Xue-Cheng Tai, Morten Kompen*, Christina Rørvik - NORCE; * NORCE and Compelling AS
NORCE has developed PAMS (Polyp Activity Monitoring System) as an innovative camera-based sensing platform and environmental effect methodology to monitor coral welfare in their natural environment and impacts resulting from use of the ocean space. The system comprises a hardware and software delivery, with the main value attributed to expert and machine learning (ML)-based interpretation of data. Based on off-the-shelf components, PAMS can be regarded as an effective, affordable, simple-to-assemble yet modular and customizable tool for users who want to monitor corals (shallow and deep) and assess their status. Whilst there exist solutions to measure coral endpoints in shallow waters, there is still a technology gap for deepwater corals and in general - it is difficult to find an appropriate parameter for monitoring corals in a non-intrusive manner which reflect any influences from their surrounding environment. High resolution still photos can support the non-intrusive and specific identification of polyp behavioral change on corals exposed to particle plumes or change of conditions in the water. The PAMS technology is based on years of lab research and observations dedicated to specific analysis of coral endpoints and more specifically coral behavioral endpoints with polyp activity. It is currently developed using specific ML modules for automatic coral recognition and classification of the coral status according to their polyp activity using a high-throughput process independent of user expertise. Currently at Technology Readiness Level (TRL) 5-6, the vision for PAMS encompasses both full development and qualification for industrial applications as well as the advancement of research on marine seafloor environments where coral is found. In this workshop, we will present a detailed description of the current hardware and the status of the software, including an open-source data visualization platform for displaying and analyzing coral images. The PAMS project is funded by the Research Council of Norway.
Planktwin -- Physically Faithful Synthetic Imagery from Digital Twins of Plankton
David Nakath , Kiel University and GEOMAR Helmholtz Centre for Ocean Research Kiel
Xiangyu Weng(GEOMAR Helmholtz Centre for Ocean Research Kiel), Jan Taucher (GEOMAR Helmholtz Centre for Ocean Research Kiel), Elisa Wendt(Kiel University), Veit Dausmann (GEOMAR Helmholtz Centre for Ocean Research Kiel)
Plankton observation is a ubiquitous task in underwater imaging. A multitude of imaging systems exists, covering several optical and acoustic imaging modalities from macroscopic to microscopic, dark to bright field and perspective to orthographic projections. The resultant versatility of the imaging data hinders standardized workflows in downstream tasks like species classification, trait classification, and continuous-valued trait regression. The key challenge remains to disentangle the semantic information of the individual plankton objects from the underlying imaging domain-specific information. To overcome this, we present a rendering-based system which can synthesize physically faithful image data of plankton in different underwater environments and with different camera-types. Specifically, our pipeline features a model learning step which disentangles plankton 3D-representations from the light and optical conditions it was taken in. This model can then be submerged in standardized Jerlov-Water bodies, as defined by measurable attenuation and scattering parameters, in different light fields like homogeneous sun illumination, highly heterogeneous artificial setups like light sheets, shadowgraphy, or even a mixture of all. Finally, the resultant scene can be captured with different simulated optical systems. This approach allows for the training of AI networks to understand the information inherent to plankton, which can then be transferred to other imaging domains or used as a digital twins. It is also possible to test newly developed approaches against image data with full known ground truth - which is barely available in underwater scenarios or has to be obtained by annotation approaches. It also enables the design of new plankton imaging systems by virtually submerging hypothetical setups and inspect the resultant imagery in different simulated water columns. Finally, as the system is a physically-based implementation in an autodiff-framework, the model parameters can be optimized with respect to e.g., image quality criteria and implemented in an actual system.
Publicly available and freely accessible Imagery AI Training Data set repository
errol.ronje@noaa.gov , NOAA National Centers for Environmental Information
Megan Cromwell (NOAA National Ocean Service), Vidhya Gondle, Yee Lau, Kirsten Larsen, Hassan Moustahfid, Rachel Medley (NOAA Ocean Exploration), Mashkoor Malik (NOAA Ocean Exploration), Patrick Cooper (University Corporation for Atmospheric Research UCAR / NOAA Ocean Exploration Affiliate), Kakani Katija, Brian Schlining (Monterey Bay Aquarium Research Institute)
Efforts to advance artificial intelligence (AI) applications, particularly in underwater image data analysis, are hindered by the absence of a curated, long-term repository for AI training data. The lack of such resources constrains collaboration and knowledge sharing across various entities, including governmental organizations, research institutions, and industries. In order to address this community-wide need for a robust standardized and openly accessible training data set, NOAA National Centers for Environmental Information (NCEI), NOAA Ocean Exploration, and the Monterey Bay Aquarium Research Institute have agreed to co-design and develop a public and freely available underwater imagery training data repository at NCEI using curated data from MBARI’s FathomNet project. This initiative will allow data and appropriate metadata submitted to FathomNet to be transferred to NCEI for web-accessible hosting and continuous community access. Dedicated pipeline development has begun with the image and metadata extraction automation from FathomNet image banks for hosting on a partner managed web-accessible server. The FathomNet data and metadata will be combined into a version-controlled bi-annual archival package within NCEI and assigned a digital object identifier (DOI). This facilitates citability and replication of the database and image collection over-time. This process is a pioneering effort to create, collect, and maintain AI-ready data for the oceanographic imaging community. The goal is to foster collaboration and innovation, empowering NOAA and its partners to leverage AI technologies effectively for advancing environmental research and stewardship.
Quantifying the Efficiency of Auto-Classification using Towed Underwater Optical Imagery
Karah M. Kniola , MS Student (University of South Florida - College of Marine Science)
Sarah Grasty, Chad Lembke, Alexander Ilich, Alex Silverman, Steven Murawski (ALL USF-CMS)
There is an ever-increasing need for non-invasive survey techniques in marine resource management due to the modern prevalence of no-take marine managed areas (e.g., Marine Protected Areas and offshore wind farms). Traditional sampling gear, such as trawls, can also damage hard bottom habitats where many commercially and recreationally important species congregate. The increased use camera-based sampling methods provides an important alternative to traditional approaches. However, manually processing the imagery from these surveys is laborious, subject to operator error and fatigue, and time-intensive thereby delaying delivery of important information to management. In this project we use a series of optical imagery from a towed video array (C-BASS – Camera Based Assessment Survey System) to investigate the feasibility and reliability of supervised and unsupervised autoclassification applications to detect and classify organisms. The imagery used in testing was collected among several habitat types on the West Florida (USA) continental shelf which targeted snappers (Lutjanidae sp.), groupers (Epinephelinae sp.) and other demersal fishes. Two machine learning software applications were used: VIAME (Video and Image Analytics for Marine Environments, NOAA) and UAA (Undersea Biological/Geological Auto-Identification Application), a custom software developed by SRI International under contract to our project) to process monochrome high-definition video imagery from the towed camera system. Metrics collected from the autoclassification analyses include precision, accuracy, and variability of results as well as time-to-completion of each video transect. Utilization of transects in the same region but not included in the training sets allows calculation of the accuracy of naïve classifications relative to a known library of stored imagery for verified species. The output from the autoclassification methods were compared to manually (human) classified observations. These comparisons can help steer the development and direction of future analytical approaches for fisheries sampling data which are less biased, more precise, and more efficient.
Returning to the Deep: Maximizing Insights from Deep-Sea Coral and Sponge Imagery from the U.S. Pacific Islands
Savannah Goode , Deep Sea Coral Research & Technology Program, NOAA
Kiki Kamakura (Barnard College, NOAA Pacific Islands Fisheries Science Center), Beth Lumsden (NOAA Pacific Islands Fisheries Science Center), Heather Coleman (Deep Sea Coral Research & Technology Program, NOAA)
One major disadvantage to seafloor image analysis is the bottleneck created by the time required to manually annotate imagery for scientific study. It is thus uncommon for imagery to be re-analysed after the original study is completed. Consequently, it is difficult to combine imagery datasets to conduct broader-scale investigations, which can be particularly important when studying habitats that are difficult to survey, such as the deep-sea benthos. A second issue that has arisen from limited re-annotation of seafloor imagery is that many existing image datasets are composed primarily of images with non-localized annotations (i.e., records of different taxa and/or individuals not differentiated spatially within a single image). Such images cannot be readily incorporated as training data for automated annotation pipelines, nor can they be easily used by non-experts for identification purposes. We sought to address these issues of non-standardized, non-localized annotations using photos obtained from [NOAA’s National Database of Deep-Sea Corals and Sponges](https://www.ncei.noaa.gov/maps/deep-sea-corals/mapSites.htm), focusing on the U.S. Pacific Islands region due to the abundance of high-resolution images available. Using BIIGLE 2.0 (an online image annotation platform), 14,226 images were initially assessed to determine whether they comprised a single taxon or a wider ‘community’. Images labeled ‘community’ (3,230 total) were selected for re-analysis. Each image was annotated to characterize the visible substrate (using the Coastal and Marine Ecological Classification Standard ([CMECS](https://iocm.noaa.gov/standards/cmecs-home.html)) classification) and to localize the taxon of interest (using rectangular bounding boxes). This dataset will be used to: (1) investigate taxon-substrate relationships in this region, (2) provide representative _in situ_ taxon images for non-experts, (3) train automated classification tools, and (4) inform a national guide to identify important and vulnerable deep-sea habitats, currently in development by the NOAA Deep Sea Coral Research & Technology Program and the Bureau of Ocean Energy Management.
SQSIZE - Stereographic Quantification and Shadowgraph Imaging of Zooplankton and their Environment
Mehul Sangekar , X-STAR, JAMSTEC
Ariell Friedman, Greybits Engineering (Australia), ariell@greybits.com.au ; Dhugal Lindsay, X-STAR, JAMSTEC (Japan), dhugal@jamstec.go.jp
The SQSIZE (Stereographic Quantification and Shadowgraph Imaging of Zooplankton and their Environment) system is comprised of two autonomous imaging systems and is designed to be deployed simultaneously on a 36-bottle CTD rosette system. This system was designed primarily for gathering data on the particle field in near real time at areas being assessed for their suitability or currently being exploited for deep sea mineral extraction. We describe the system hardware and the software pipelines that are tightly integrated to minimize the time necessary between retrieval of the CTD system and visualization of the quantitative data on particles, plankton and larger nekton imaged by the system.
SeagrassFinder: Deep Learning Models for Eelgrass Detection and Coverage Estimation in the Wild
Jannik Elsässer , DHI A/S and former IT University of Copenhagen Student
Associate Professor Veronika Cheplygina and PhD Fellow Laura Weihl, vech@itu.dk & lawe@itu.dk
Seagrass meadows play a crucial role in marine ecosystems, providing important services such as carbon sequestration, water quality improvement, and habitat provision. Monitoring the distribution and abundance of seagrass is essential for environmental impact assessments and conservation efforts. However, the current manual methods of analyzing underwater video transects to assess seagrass coverage are time-consuming and subjective. This work explores the use of deep learning models to automate the process of seagrass detection and coverage estimation from underwater video data. A dataset of over 8,300 annotated underwater images was created, and several deep learning architectures, including ResNet, InceptionNetV3, DenseNet, and Vision Transformer, were evaluated for the task of binary classification of "Eelgrass Present" and "Eelgrass Absent" images. The results demonstrate that deep learning models, particularly the Vision Transformer, can achieve high performance in predicting eelgrass presence, with AUROC scores exceeding 0.95 on the final test dataset. The use of transfer learning and the application of the Deep WaveNet underwater image enhancement model further improved the models' capabilities. The proposed methodology allows for the efficient processing of large volumes of video data, enabling the acquisition of much more detailed information on seagrass distributions compared to current manual methods. This information is crucial for environmental impact assessments and monitoring programs, as seagrasses are important indicators of coastal ecosystem health. Overall, this project demonstrates the value that deep learning can bring to the field of marine ecology and environmental monitoring.
Spatial distribution and hotspots of habitat forming coral and sponge communities of the Hawaiian – Emperor Seamount Chain: Use of marine imagery for spatial analyses
Virginia Biede , Florida State University
Nicole B Morgan, NOAA ; Amy R Baco, Florida State University
Deep-sea benthic community spatial distribution is a fundamental parameter for understanding communities at multiple scales. Spatial analyses inform spatial management of vulnerable ecosystems, like sponge and coral communities on seamounts exposed to bottom contact fishing. To gain a better understanding of patch sizes on North Pacific seamounts, an existing dataset derived from surveys with the AUV Sentry, was reanalyzed. The Sentry data included annotations from surveys of seafloor communities between 200m–800m depth on seamounts of the Hawaiian-Emperor Seamount Chain. Sentry images were annotated for coral, sponge, crinoid, and brisingid presence, and categorized for density as sparse, medium, or abundant for each taxon. 4627 individual patches were found when the taxa were analyzed individually, 3860 of those being Coral patches. For all taxa combined, 4399 patches were found. For most patch analyses, the largest patch was found on Koko Guyot, with a linear length of over 2km for all fauna combined, over 1km for the largest coral patch, and 988m for Crinoids. For Sponges the largest patch was 78m on Bank 11, and for Brisingids maximum patch size was 109m on Kammu. Patch size of Coral, Crinoid, and All Fauna groups varied by seamount (Coral and Crinoids: p< 0.001; All Fauna: p <0.01) and depth (p<0.0001 for all three) but no significant pattern was found for Brisingids or Sponges. A second scale of analyses used Getis-Ord to find hotspots of faunal abundance. Using the average nearest neighbor distance and median distance between patches, neighborhood size was found for each fauna group. A total of 850 hotspots were found across the seven seamounts and both neighborhood size and hotspot size varied substantially by taxon; from 500m2 to 95,284m2. Future discussions of Marine Imaging should incorporate needs for spatial analyses on multiple scales.
Sustainable preservation of underwater archaeological sites: marine imaging supporting heritage management
Chris McGonigle , School of Geography and Environmental Sciences, Ulster University. Cromore Road, Coleraine. Northern Ireland. BT55 7GB
David Gregory (National Museum of Denmark) Anne Marie Høier Eriksen (National Museum of Denmark) Knud Jakobsen (Sea War Museum) Rasmus Normann Andersen (JD-Contractor A/S) Henning Matthiesen (National Museum of Denmark) Gert Normann Andersen (Sea War Museum) Christian Toftgaard (Sea War Museum) Mogens Dam (JD-Contractor A/S) Rory Quinn (Ulster University)
Preserving the remains of the past for the benefit of future generations is a common approach in international heritage policy. Current management practice advocates preserving underwater cultural heritage (UCH) where it lies on the seabed, in situ. However, this practice is questioned due to a lack of understanding of the entangled threats posed by multiple natural and anthropogenic drivers. In a rapidly changing ocean, and with increasing human exploitation of the marine environment, it is necessary to develop new concepts for assessing and preserving this resource. With over 3,000,000 shipwrecks and thousands of submerged prehistoric sites lying on the floors of the world’s oceans, our work aims to disentangle both natural and anthropogenic decay processes, determine their cumulative and interactive effects on UCH and propose a novel conceptual framework to preserve this heritage based on site entropy. Marine imaging methodologies represent one of a suite of techniques utilised for investigating shipwrecks, both from archaeological and ecological perspectives. This integrated approach can develop mutually beneficial understanding around environmental change and archaeological site formation processes, whilst informing natural resource management about the ecological significance of the sites themselves. One established technique of relevance is Structure from Motion (SfM) and the development of photogrammetric models which have been shown to compliment more conventional remotely sensed data at a variety of spatial and ecological scales. This can facilitate the detection, visualisation and interpretation of the products of natural and anthropogenic decay of shipwrecks and submerged prehistoric sites, integrated with natural and anthropogenic variables. This work in development will present a series of exemplar case studies from the Baltic, North Sea and north-west European margin that explore the physical, chemical, biological and anthropogenic drivers of change at a range of sites and contexts of archaeological and ecological interest.
Synthesize Large-scale in situ Images for Training Deep Plankton Detection Algorithms
Jianping Li, Claude , 1.Shenzhen Institute of Advanced Technology, CAS, Shenzhen, China; 2.University of CAS, Beijing, China.
Zhenping Li 2,1 1.Shenzhen Institute of Advanced Technology, CAS, Shenzhen, China; 2.University of CAS, Beijing, China.
The content distribution of full-frame raw images captured by in situ plankton imaging instruments is highly complex and variable, easily affected by multiple factors such as the properties of seawater, the quantity, taxonomy, posture, and morphology of underwater objects, as well as changes in camera focusing status and working settings. Therefore, most plankton observation methods based on in situ imaging have divided the recognition task into two stages: the first is to locate the objects from the raw full-frame images and crop out the region of interest (ROI) vignettes containing only one target; the second is to classify the ROI images to achieve the ultimate automatic plankton recognition. However, the raw image complexity can significantly impact the performance of the two-stage plankton target cropping and recognition algorithms, and may even cause serious errors for the observation of interested plankton. Although existing in situ imaging instruments have already collected massive full-frame image data under different sea conditions, their content variety poses great challenges for manual annotation, making it almost impossible to obtain large-scale high-quality datasets to train usable end-to-end deep CNN models for plankton detection. We note that it is possible to utilize deep generative models to perform large-scale raw image synthesis using a limited amount of full-frame images and annotated ROI data. Based on the images collected by our dark-field underwater imager IPP (Imaging Plankton Probe) across different sea regions in China and Australia, we propose a method that can quickly build a massive full-frame image dataset IsPlanktonGC with definite object identity ground truth. The preliminary experimental results show that the imitated images have very realistic visual similarity to the true in-situ collected ones. This strategy is expected to be very helpful in providing big data foundations for developing robust downstream object detection algorithms to better observe marine plankton.
Tackling exponential increases in deep-sea video data: VARS + AI
Lonny Lundsten , Monterey Bay Aquarium Research Institute
Nancy Jacobsen Stout (MBARI), Kyra Schlining (MBARI), Kristine Walz (MBARI), Larissa Lemon (MBARI), Megan Bassett (MBARI), Brian Schlining (MBARI), Kevin Barnard (MBARI), Danelle Cline (MBARI), Duane Edgington (MBARI)
The Monterey Bay Aquarium Research Institute (MBARI) has made significant advancements towards incorporating machine learning (ML) into the annotation and analysis of our 37-year archive of deep-sea video. Central to these efforts is the integration of the Video Annotation and Reference System (VARS) with advanced ML tools and the development of a new machine-assisted annotation workflow. VARS is a software system developed by MBARI to annotate and manage video data collected from remotely operated vehicles (ROVs), autonomous underwater vehicles (AUVs), and other camera platforms. Using VARS, researchers create detailed annotations of marine organisms, geological features, and habitats from deep-sea video footage using a structured format within a searchable relational database. Visual observations are merged with additional data collected during vehicle deployments, such as salinity, temperature, position, and depth. Historically, MBARI’s Video Lab staff have manually annotated all deep-sea videos recorded by MBARI's ROVs in a consistent manner. However, the recent, exponential increase in camera platforms used for conducting surveys has necessitated the development of innovative workflows to handle this massive influx of visual data, which is rapidly outpacing the feasibility of manual processing. To address this, we have developed new tools and workflows for 1) localizing ML model training data, 2) generating ML proposals rapidly and inexpensively, and 3) validating and editing ML generated proposals. To date, we have produced 500,000 deep-sea localizations for ML training and used the subsequent ML models to generate detections on 200 hours of ROV video and 800+ hours of AUV video (PiscivoreCam, i2MAP). These efforts underscore MBARI's commitment to utilizing cutting-edge technology to advance marine research, ultimately providing critical insights into ocean health and ecosystems while fostering collaboration among scientists.
Through their eyes: adaptations of hyperiid amphipod eyes for the midwater
Karen J. Osborn , Smithsonian National Museum of Natural History
Jan M. Hemmi, University of Western Australia Oceans Institute
Vision in the midwater (the open ocean below the photic zone) is challenging given the relatively high attenuation of light over short distances that results in a highly structured light field, which is periodically interrupted by bioluminescence. This unfamiliar light field has led to visual adaptations in many midwater animals that can provide inspiration for underwater and low-light imaging. Hyperiid amphipods are a small group of midwater crustaceans notable for the diversity of their visual systems and their importance in the midwater food web, where they are both abundant prey and voracious predators. Having compound eyes allows hyperiids to drape their photoreceptor sheets over any shape imaginable creating local, optimize regions in visual space for different tasks. This flexibility has led to a unique array of visual adaptations. By studying the anatomical and functional differences between the different eye designs, we learn what it takes to see in the midwater and subsequently, how successful visual predators image this challenging habitat. To determine their visual field composition, we reconstruct the 3D eye anatomy of hyperiids from microCT scans. A custom neural network allows us to map a compound eye’s visual field in a matter of hours to days, rather than months. Combining the resulting “map” information with physiological measurements from live animals, we model what each hyperiid eye sees and how their vision changes with through the light field. We test and interpret these models against what we are learning of hyperiid behavior from in situ observations and a newly designed, immersive, visual experimental chamber, to understand what each eye design is “optimized” for. This work provides a native view of the midwater and a better understanding of how animals optimize vision in the midwater under extreme size and energy constraints.
Time-lapse observations reveal daily to seasonal changes in deep-sea biological systems
Jim Barry , MBARI
Alana Sherman (MBARI), Paul McGill (MBARI), Fanny Girard (UH), Steve Litvin (MBARI), Chris Lovera (MBARI), Aaron Schnittger (MBARI), Andrew DeVogelaere (NOAA-MBNMS)
Our understanding of adaptions animals depend upon to thrive in seemingly harsh deep sea environments is limited by our ability to observe them over scales in time and space that affect their survival, growth, and reproduction. For deep-sea organisms and communities, logistic challenges typically limit studies to rare, short-term visits using deep-diving submersibles or sampling systems. To expand our observations of deep-sea coral ecosystems, we developed two time-lapse camera systems to measure daily to seasonal variation in coral behavior. The mega-coralcam and macro-coralcam can capture images once per hour for at least a year. The mega-coralcam images a large scene (ca. 5 m across) of up to several coral colonies. The macro-coralcam is used for close-up scenes (ca. 25 cm across), providing close view of coral polyps. Deployments of these camera systems has revealed daily to seasonal variation in feeding behavior for bubblegum corals (Paragorgia arborea), as well the seasonal pattern of reproduction in deep-sea bamboo corals (Isidella tentacula).
Towards Species-Level Detection of Caribbean Reef Fish
Levi Cai , WHOI
Austin Greene (WHOI), Daniel Yang (MIT and WHOI), Nadege Aoki (MIT and WHOI), Sierra Jarriel (WHOI), T. Aran Mooney (WHOI), Yogesh Girdhar (WHOI)
Visual detectors and classifiers have shown remarkable utility for animal conservation and monitoring tasks by providing rapid estimates of species abundance and biodiversity or can also reduce labeling efforts during data preprocessing tasks. These approaches have been proposed for animals in underwater environments, such as the deep ocean with FathomNet [1]. However, their deployment in coral reef environments still faces many obstacles. Like other marine environments, there is a lack of previously annotated data and the ecological nature of the data faces a long-tail distribution problem where many less common species are difficult to obtain observations about. Distinct from the more controlled environment of imaging animals in the deep ocean, coral reefs present a visually rich and complex environment. Their location in shallow waters requires visual detectors to cope with significant variations in lighting while corals themselves provide visually confounding characteristics to ecologically interesting species. We have built a new dataset for training and evaluating visual detectors and classifiers at the species-level in coral reef environments. It consists of fish-species labels from diver transect videos across 5 different reef sites spanning over 8 years in the U.S. Virgin Islands. In total there are 162 videos, of which we label 14K images. We also provide tracking information for every fish across each video, enabling novel insights into tracking challenges. Furthermore, we examine the entire pipeline, from data pre-processing, labeling, and training, to deployment. We investigate how best to utilize pre-existing fish detection datasets when applied to our specific domain, either as supplemental data or in transfer learning. We further show results of detectors trained on our dataset that are subsequently deployed on stationary camera or AUV data from the same or nearby sites, to demonstrate the utility and challenges that still remain in deploying these systems even in similar locations. [1] K. Katija et al. “FathomNet: A global image database for enabling artificial intelligence in the ocean” Nature Scientific Reports, 2022.
Towards an AIoT-based Marine Plankton Imaging Network
Jianping Li, Claude , 2. Shenzhen Institute of Advanced Technology, CAS, Shenzhen, China; 3. University of CAS, Beijing, China;4. Xcube Technology Co, Ltd., Shenzhen, China;
Zekai Zheng 1,2, Zhenping Li 2,3, Chi Liu 2, Junqiang Jiang 2,4, Peng Liu 2,3, Liangpei Chen 2,3, Shunming Chen 2,4, Zhisheng Zhou 2,3, Ming Zhu 2,3, Haifeng Gu 5, Jiande Sun 1, Jixin Chen 6. 1. Shandong Normal University, Jinan, China; 2. Shenzhen Institute of Advanced Technology, CAS, Shenzhen, China; 3. University of CAS, Beijing, China; 4. Xcube Technology Co, Ltd., Shenzhen, China; 5. Third Institute of Oceanography, Ministry of Natural Resource, Xiamen, China; 6. Xiamen University, Xiamen, China.
In situ observation of plankton based on underwater imaging should theoretically have real-time and continuous advantages. However, current methods rely almost entirely on post-processing to convert image data into observational information. This results in significant delays in information acquisition and exerts enormous pressure on transmission network bandwidth and real-time processing capabilities required for the massive amounts of data collected from remote instruments. To address these issues, we have developed an Artificial Intelligence of Things (AIoT) marine plankton imaging system as a proof-of-concept to combine edge-computing and cloud-computing to network multiple underwater dark-field imagers (imaging plankton probe, IPP) distributed across various locations, aiming to leverage the high temporal resolution advantage of in situ observations while expanding the spatial coverage of ocean observations. At the edge, a client imager first captures in situ images and then processes them through operations such as object detection, DOF extension, feature extraction, etc. On the cloud side, the server supports complex plankton data management and AI retrieval functions, with an intelligent application layer providing user management, device control and monitoring, and real-time data analysis and visualization functions. Compared to traditional methods, this AIoT-based multi-location underwater imaging approach has significant advantages. The system optimizes computing and storage resources, enhances real-time data processing and security, improves transmission reliability, and reduces deployment and maintenance costs. By deploying multiple IPPs in various locations in China and Australia for periods ranging from one to eight months, we have preliminarily demonstrated the capability of this innovative AIoT-based paradigm to monitor complex marine planktonic dynamics, such as the bloom and decay of Noctiluca scintillans. This offers a more cost-effective solution for expanding global marine plankton observation capabilities in the future.
Towards standardised reporting of macro-litter and biotic interactions from imagery
Jaime S Davies , University of Gibraltar
Alice L Bruemmer (University of Gibraltar) Awantha Dissanayake (University of Gibraltar)
The use of marine imagery as a tool to monitor species and habitats is commonplace, but can also provide insightful information of marine litter, and its potential impacts. Monitoring and quantification of marine litter often do not include reporting of interactions between fauna and litter, meaning impacts are largely unconsidered and unknown. It is important to be able to quantify anthropogenic effects on vulnerable deep-sea habitat to ensure adequate protection. A standardised, comprehensive framework for the reporting of Litter-Fauna interactions from imagery was created from a literature review and includes 6 major categories: entanglement, ingestion, smothering, habitat provision, adaptive behaviour, and encountering (entanglement and smothering occur on abiotic features as well). Litter can cause long term impacts such as smoothing of corals or even snagging of discarded fishing gear; the L-F framework allows extraction of additional information from imagery. The framework gives clear definition and examples to allow standardised, repeated implementation. This work identifies the gaps in knowledge and presents a standardised litter-fauna interaction framework for deep-sea imagery data, which was applied to a case study to test the applicability.
Understanding the Current State of Southern Ocean Benthic Ecosystems Using Deep Computer Vision
Cameron Trotter , British Antarctic Survey
Huw Griffiths (hjg@bas.ac.uk) & Rowan Whittle (roit@bas.ac.uk), British Antarctic Survey
Loss of marine biodiversity is a key issue facing the modern world. The removal of species from an environment can have profound effects on the overall ecosystem structure, though to what degree any species contributes to ecosystem stability is often unknown until they are removed. Due to its remoteness, relatively little is known about the benthic ecosystems situated in the Southern Ocean and around Antarctica. This region is also one of the most vulnerable to climate change, and is currently one of the fastest warming areas on the planet. Traditionally, our understanding of Southern Ocean biodiversity has relied on nets or devices to bring benthic organisms to the surface, though this is intrinsically destructive and fails to detail community structure. The development of underwater imaging technologies has allowed for the capturing of this information non-destructively in-situ, though localising and classifying the organisms within these images can be time-consuming and requires specialist expertise, given that many organisms found here are seen nowhere else on Earth. This has resulted in a data bottleneck, greatly limiting our understanding of the Southern Ocean’s benthic ecosystems and the effect of climate change on them. To help combat this, we present the development of a deep computer vision model trained to detect key taxa captured in Southern Ocean benthic imagery. This model is trained using only a small subset of labelled images captured from a downward facing towed camera. Once trained, the model is capable of processing unlabelled imagery in an autonomous manner, requiring only human verification of system output. This allows analysis to be performed faster and over a larger spatio-temporal range when compared to a fully-human approach, providing a clearer picture of the current state of the Southern Ocean’s benthic ecosystems. Ultimately this allows for the development of more comprehensive protection strategies for these ecosystems.
Unveiling a hidden sanctuary: first description of mesophotic gorgonian assemblages thriving in a coralligenous reef, close to a tuna fish farm in l’Ametlla de Mar (Spain)
Judith Camps-Castella , Departament de Biologia Evolutiva, Ecologia i Ciències Ambientals, Facultat de Biologia, Universitat de Barcelona, Avda. Diagonal 643, 08028 Barcelona, Spain
Patricia Prado. Instituto de Investigación en Medio Ambiente y Ciencia Marina (IMEDMAR-UCV), Universidad Católica de Valencia SVM, C/Explanada del Puerto S/n, 03710 Calpe, Alicante, Spain.
Mesophotic ecosystems (30-120 m) may be less affected by thermal stress from climate change and provide crucial refuge to vulnerable species. This study provides the first characterization of a Mediterranean coralligenous reef locally referred as “El Ramonet” located near a large tuna farm in l’Ametlla de Mar (Southern Catalonia, Spain), hosting gorgonians such as Paramuricea clavata and Eunicella spp. at 50 m depth. The biodiversity of the megabenthic assemblages, including the demographic structure of P. clavata, was described in six video transects taken at different distances from the fish farm using Remotely Operated Vehicle (ROV) video imaging. Also, we calculated MAES Index to evaluate the ecological status of coralligenous bioconstructions. Our results show that the reef host one of the largest colony sizes ever recorded for P. clavata in the Mediterranean Sea (ca. 113 cm), with the population mostly dominated by large colonies. Video analysis revealed 387 gorgonian colonies with highest abundances recorded in a transect near the fish farm (mean: 1.88 ± 0.42 colonies · m2). For Eunicella spp., the highest abundance was found away from the farm (1.12 ± 0.12 colonies · m2). The MAES Index revealed that transects closest to the fish farm were in the worst condition, resulting from the presence of branches with necrosis and discarded fishing gear. Further, Shannon-Wiener diversity index for sessile fauna showed homogeneous patterns across transects (0.28-0.17) except for significantly lower traits in a site near the fish farm (0.09 ± 0.03). This study proclaims the significance of this unique reef as a hotspot for the conservation of P. clavata and associated species. The unique features associated to mesophotic environments might provide an advantage for the species in the face of climate change, but management and conservation initiatives are also necessary to protect the site from other local human activities.
Using Computer Vision to Objectively Quantify Features in Geological Drill Core
Lewis Grant , University of Southampton
Rosalind M. Coggon (University of Southampton), Blair Thornton (University of Southampton), and Damon A. H. Teagle (University of Southampton)
Ocean crust paves two thirds of Earth’s surface and its physiochemical exchanges with seawater circulating through it exert a primary control on oceanic and atmospheric composition through time. Quantifying the rate and nature of this exchange of heat, elements and nutrients is imperative if we are to understand long term climate, changing ocean conditions, and possible sites for the origin of life. Drilling into in situ ocean crust, as well as ancient crustal slabs obducted onto land (ophiolites) provides valuable records of these interactions. Geologists painstakingly document the distributions of petrographic features in recovered cores, based on visual assessments. Core description is therefore limited to the time and resources available during core recovery and prone to human bias. As cores are often digitally imaged, we have explored the potential benefits computer vision provides to the expert geologist tasked with describing the recovered material. We have demonstrated that self-supervised machine learning methods are capable of accurately identifying and quantifying features within core images, and that the addition of spatial metadata during network training improves accuracy by as much as 50 %. Using the classification output of a spatially-guided contrastive learning model (GeoCLR), important geological information has been quantified through 400 m of core at a resolution impossible for a human expert to generate in the same time. This work highlights the promising potential of using machine learning to extract valuable information from large unutilised image datasets, although improvements in core imaging best-practices are needed to improve the machine readability of future core imagery. Furthermore, automated core description has wide-reaching potential to improve the efficiency of numerous industries such as geoengineering, remote sensing and mineral exploration.
Using machine learning to investigate temporal dynamics of methane seep fauna at the Ocean Observatories Initiative Regional Cabled Array
Katharine T. Bigham , University of Washington
Michael F. Vardaro, University of Washington; Deborah S. Kelley, University of Washington
Methane hydrate seeps are unique chemosynthetic marine environments where organisms use subsurface-derived chemicals and volatiles rather than sunlight to create energy. Seeps are found on continental margins worldwide; >800 gas emission sites have been imaged along the Cascadia Margin. The seeps provide important services such as energy storage in the form of methane hydrates, habitats for commercially important fish stocks, and contribute to regional biodiversity and productivity. One of these methane seeps, Southern Hydrate Ridge (SHR), has hosted a diverse suite of instrumentation on NSF’s Ocean Observatories Initiative (OOI) Regional Cabled Array (RCA) that has been streaming real-time data to shore for a decade. Among the instruments is a SubC digital still camera, which takes images every half an hour resulting in >50,000 images in a year. Additionally, ROV imagery (still, HD, 4k) are collected at the site annually as part of the RCA operations and maintenance cruises. In concert, this has amounted to a collection of >37 TB of visual data that has the potential to provide important insights into the evolution of these seeps over time, as well as into the dynamics of the benthic fauna community that thrives there. Due to the time-consuming nature of manually annotating the large amount of SHR imagery that the RCA has collected and continues to collect, machine learning pipelines are being developed to annotate imagery as it is captured. This effort will serve as an important contribution, by making the OOI-RCA imagery more accessible to researchers and the public, and product integration with a wealth of available stored and real-time environmental data. We will present preliminary results from those pipelines, discuss methods being used to integrate human verification in the pipeline, give an example of the type of data produced by the pipeline, and discuss the ecologic questions that we aim to answer.