Monograph |
Corresponding author: Georg F. Tschan ( g.tschan@leibniz-lib.de ) © 2024 J. Wolfgang Wägele, Georg F. Tschan.
This is an open access book distributed under the terms of the Creative Commons Attribution License (CC BY 4.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Citation:
Wägele JW, Tschan GF (2024) Weather stations for biodiversity: a comprehensive approach to an automated and modular monitoring system. Advanced Books, Pensoft, Sofia, 1-218. https://doi.org/10.3897/ab.e119534
|
The editors, in the name of all authors and other participants in the project, would like to express their thanks to the German Federal Ministry of Education and Research (BMBF) for funding and to both the DLR and the VDI/VDE-IT for administering the AMMOD project.
Furthermore, we are grateful to the personnel at the three test sites for allowing access, helping with the maintenance, and facilitating our visits: the Melbgarten of the University Botanic Gardens in Bonn, the Ecological research station in Britz, and the Energieberg Georgswerder site in Hamburg.
Last but not least special thanks to all the colleagues who helped us with the administration of the project, as well as various students, all who are rarely seen or mentioned, but have been essential to its success.
J. Wolfgang Wägele and Georg F. Tschan
Nobody would doubt that it is good to have weather stations. They provide measurements of precipitation and temperature, and often also of humidity and wind speed. Tens of thousands of weather stations worldwide form a Global Observing System that has proven indispensable for weather forecasting, prediction of extreme events such as floods or droughts, and monitoring of climatic changes.
Weather and climate are things that everyone experiences in their everyday lives: Will it rain tomorrow? I should better bring a raincoat. Will it be hot in two days? I should better bring a pair of sandals.
The situation is less clear when it comes to biodiversity. Coined in 1988 by world-famous biologist Edward O. Wilson, the term “biodiversity” is used to describe “the greatest wonder on the planet” – the diversity of life on Earth in all its facets. While we are still struggling to discover how many species there are on Earth, we are already losing thousands of them forever at unprecedented rates, much higher than the background extinction rate. Yet, we are lacking “weather stations” to continuously assess the status of biodiversity on the planet. How can we take conservation decisions, how can we manage landscapes more sustainably, if we don’t know anything about the ups and downs of biodiversity? Nobody remembers how many birds were singing in the 18th century, or how many bumblebees were buzzing, how many butterflies were in the air.
Advances in technology now make it possible to change all this. Even if the past is hidden in the dark or in natural history collections, we can still make a difference in the future – showing, hopefully, how biodiversity recovers from anthropogenic pressure and how we can shape a biodiversity-friendly future.
The present book, edited by Wolfgang Wägele and Georg Tschan, provides an as yet unprecedented collection of technological advances that could form the basis for future biodiversity weather stations. The book describes a range of approaches that could make biodiversity assessment as easy as measuring temperature or rainfall in a reliable, reproducible manner. Of course, there are still quite some more steps to go – but the combination of approaches shown in the book would make it worthwhile to consider carefully when designing the next level of Global Observing Systems.
The spectrum of approaches covered is simply amazing: From assessing plant diversity using volatile organic compounds and pollen traces, to automated insect trapping systems, multi-channel bioacoustics and depth-aware visual monitoring and non-destructive DNA metabarcoding, the authors provide insights into the most advanced approaches to assess biodiversity using as much automation as possible. The book concludes with a description of a base station integrating the inputs from multiple sensors, including a web-based data portal for long-term data management.
Clearly, we still don’t have weather stations for biodiversity. But it can be hoped that the contents of this book will serve as a basis for eventually arriving there – so that measuring biodiversity will one time be as straightforward as measuring temperature. The greatest wonder on the planet deserves it.
Sincerely,
Christoph Scherber
Head of the Centre for Biodiversity Monitoring and Conservation Science
Vice director of the Leibniz Institute for the Analysis of Biodiversity Change
Paul Baggenstoss • Fraunhofer FKIE, Fraunhoferstr. 20, D-53343 Wachtberg, Germany
Sarah J. Bourlat • Centre for Biodiversity Monitoring and Conservation Science, Leibniz Institute for the Analysis of Biodiversity Change (LIB), Museum Koenig, Adenauerallee 127, D-53113 Bonn, Germany https://orcid.org/0000-0003-0218-0298
Karl-Heinz Frommolt • Museum für Naturkunde – Leibniz Institute for Evolution and Biodiversity Science, Invalidenstr. 43, D-10115 Berlin https://orcid.org/0000-0002-5157-7358
Birgit Gemeinholzer • Institute of Biology, Botany, University of Kassel, Heinrich-Plett-Str. 40, D-34132 Kassel, Germany https://orcid.org/0000-0002-9145-9284
Frank Oliver Glöckner • Alfred Wegener Institute, Am Handelshafen 12, D-27570 Bremerhaven, Germany https://orcid.org/0000-0001-8528-9023
Timm Haucke • Institute of Computer Science 4, University of Bonn, D-53115 Bonn, Germany https://orcid.org/0000-0003-1696-6937
Olaf Jahn • Museum für Naturkunde – Leibniz Institute for Evolution and Biodiversity Science, Invalidenstr. 43, D-10115 Berlin https://orcid.org/0000-0001-7936-033X
Ameli Kirse • Centre for Biodiversity Monitoring and Conservation Science, Leibniz Institute for the Analysis of Biodiversity Change (LIB), Museum Koenig, Adenauerallee 127, D-53113 Bonn, Germany
Ivaylo Kostadinov • GFBio – German Federation for Biological Data, Unicom 2 Haus 2-4, Mary-Somerville-Str. 2–4, D-28359 Bremen, Germany https://orcid.org/0000-0003-4476-6764
Frank Kurth • Fraunhofer FKIE, Fraunhoferstr. 20, D-53343 Wachtberg, Germany
Kathrin Langen • Centre for Biodiversity Monitoring and Conservation Science, Leibniz Institute for the Analysis of Biodiversity Change (LIB), Museum Koenig, Adenauerallee 127, D-53113 Bonn, Germany https://orcid.org/0000-0002-4519-2689
Mario Lasseck • Museum für Naturkunde – Leibniz Institute for Evolution and Biodiversity Science, Invalidenstr. 43, D-10115 Berlin
Sascha Liedtke • ION-GAS GmbH, Konrad-Adenauer-Allee 11, D-44263 Dortmund, Germany
Florian Losch • Nees Institute for Biodiversity of Plants, University of Bonn, D-53115 Bonn, Germany https://orcid.org/0000-0001-7519-440X
Deniss Marinuks • Jacobs University Bremen gGmbH, Campus Ring 1, D-28759 Bremen, Germany https://orcid.org/0000-0003-2350-0449
Mario Paja • Hamburg University of Technology, Institute of High-Frequency Technology, Denickestr. 22, D-21073 Hamburg, Germany
Krzysztof Piotrowski • IHP GmbH – Innovations for High Performance Microelectronics / Leibniz-Institut für innovative Mikroelektronik, Im Technologiepark 25, D-15236 Frankfurt (Oder) https://orcid.org/0000-0002-7231-6704
Hanna Raus • Institute of Biology, Botany, University of Kassel, Heinrich-Plett-Str. 40, D-34132 Kassel, Germany https://orcid.org/0009-0009-0493-6600
Lukas Reinhold • Hamburg University of Technology, Institute of High-Frequency Technology, Denickestr. 22, D-21073 Hamburg, Germany
Alice M. Scherges • Centre for Biodiversity Monitoring and Conservation Science, Leibniz Institute for the Analysis of Biodiversity Change (LIB), Museum Koenig, Adenauerallee 127, D-53113 Bonn, Germany https://orcid.org/0009-0002-0824-7991
Pierre-Louis Sixdenier • University of Erlangen–Nuremberg, Department of Computer Science, Chair of Computer Science 12 (Hardware-Software-Co-Design), Cauerstr. 11, D-91058 Erlangen
Volker Steinhage • Institute of Computer Science 4, University of Bonn, D-53115 Bonn, Germany https://orcid.org/0000-0002-3172-3645
Stephanie J. Swenson • Institute of Biology, Botany, University of Kassel, Heinrich-Plett-Str. 40, D-34132 Kassel, Germany https://orcid.org/0000-0002-7550-6693
Georg F. Tschan • Centre for Biodiversity Monitoring and Conservation Science, Leibniz Institute for the Analysis of Biodiversity Change (LIB), Museum Koenig, Adenauerallee 127, D-53113 Bonn, Germany https://orcid.org/0000-0002-5108-9602
Wolfgang Vautz • ION-GAS GmbH, Konrad-Adenauer-Allee 11, D-44263 Dortmund, Germany https://orcid.org/0000-0002-6193-3766
Domenico Velotto • MARUM – Center for Marine Environmental Sciences, University of Bremen, D-28359 Bremen, Germany https://orcid.org/0000-0002-8592-0652
Maximilian Weigend • Nees Institute for Biodiversity of Plants, University of Bonn, D-53115 Bonn, Germany https://orcid.org/0000-0003-0813-6650
Benjamin Werner • Museum für Naturkunde – Leibniz Institute for Evolution and Biodiversity Science, Invalidenstr. 43, D-10115 Berlin https://orcid.org/0009-0001-2796-012X
J. Wolfgang Wägele • Centre for Biodiversity Monitoring and Conservation Science, Leibniz Institute for the Analysis of Biodiversity Change (LIB), Museum Koenig, Adenauerallee 127, D-53113 Bonn, Germany
Vera M. A. Zizka • Centre for Biodiversity Monitoring and Conservation Science, Leibniz Institute for the Analysis of Biodiversity Change (LIB), Museum Koenig, Adenauerallee 127, D-53113 Bonn, Germany
Biodiversity is one of the most valuable resources of our planet. There are 10+ Million extant species and most of these are still unknown to science (
More than 30 years ago, large-scale destruction of habitats and loss of biodiversity alarmed researchers and policy makers (
It is a fact that the disappearance of a large part of the insect fauna in Central Europe (
Unfortunately, there are six major obstacles impeding biodiversity monitoring at the species level:
The workload implies high costs (e. g.
Large-scale and long-term automated monitoring of biodiversity (as established for climate research) does not exist, among other reasons, because the required technology is not yet currently available, however, the technical prerequisites for this are in place. It is therefore crucial to adapt existing technologies for the development of automated, reliable, and verifiable biodiversity monitoring. Similar to climate research, we need ‘weather stations for species monitoring’ in addition to remote sensing.
It is possible to construct automatized multisensor stations for monitoring species diversity (AMMODs) using available technologies (see
The goal of the first phase of the AMMOD project, funded by the German Federal Ministry for Research and Education, was to build single prototypes of automated sensors and base stations, and to develop data workflows. After this proof of concept, we intended to build a small network of stations with more cost-efficient hardware. Here we summarize results from the first three years of the project. Unfortunately, the project came to a halt, because of the changed priorities for the use of public funds, a consequence of the Russian invasion of Ukraine and the subsequent support of the Ukrainian military by other countries, including Germany.
Schematic view of the AMMOD experimental setup. ① Bioacoustics, microphones of the four-channel recording device (Chapters 6 and 7). ② ‘Smellscapes’, stationary module for the detection of volatile organic compounds (VOCs) emitted by plants and fungi (Chapter 2). ③ ‘Moth scanner’, photographic imaging and identification of mostly-nocturnal flying insects (not presented here). ④ Camera trap, representing the automated visual monitoring of vertebrate animals, especially mammals (Chapter 8). ⑤ ‘Base station’, the communication and processing block for the sensor data and (partly) the electricity supply (Chapter 9). ⑥ Data management, the connection of the base station to the data portal (Chapter 10). ⑦ Automated ‘Malaise’ trap for catching flying insects (Chapters 4 and 5). ⑧ Automated pollen trap (Chapter 3). The material from both the insect and the pollen trap is subsequently analysed in the laboratory using Metabarcoding techniques. – Graphic design and creation: J. W. Wägele.
We hope that the detailed descriptions of technologies, laboratory procedures and data workflows included herein will help those who have the chance to develop networks of biodiversity monitoring stations.
Since the technologies had to be tested in the field, three test sites were selected together with all project participants. These sites had to meet the following requirements:
It should be noted that the last point, ‘Biodiversity’ – although central to the project’s objective – was of less importance during the pilot phase, because the technological, logistical and informatics basis had to be developed first. For biodiversity research, significantly more stations need to be set up based on clearly defined ecological questions, which are based on the AMMOD technique.
The above requirements also clearly limited the potential sites where the stations could be assembled. However, in collaboration with local research institutes, we could select three test sites that could be reached relatively easily (by bicycle and public transport) during the project duration. That the locations were well chosen became particularly evident during the Corona pandemic, when access was severely limited. The three test sites were:
The three test sites are geographically well distributed across Germany (
(A) The AMMOD test sites in ① Bonn, ② Britz and ③ Hamburg. (B) Natura 2000 sites in Germany and neighbouring countries. The high degree of fragmentation is evident. (C) Biogeographical regions and urban areas in relation to the sites. – Maps: G. F. Tschan, created with QGIS (
The Melbgarten is a branch of the University Botanical Gardens and is located on the Venusberg in Bonn (① in
The Melbgarten site in November 2020, looking west. The control cabinet of the base station is approximately in the centre of the picture, to the left of it the automated insect trap (but without the Malaise net) and its solar panels can be seen. – Photograph: G. F. Tschan.
For this test phase of the AMMOD design, an external power supply was planned, as the actual consumption of electricity first had to be assessed before a completely autonomous supply would be possible. However, after an initial site visit in December 2019, the first component to be permanently installed in July 2020 was an automated Malaise trap (Barcoding module) that fully autonomously collects insect samples with the help of two solar panels (
The automated malaise trap of the Metabarcoding module, July 2020. Note the two solar panels, each facing in a different direction. The garden’s greenhouse can be seen in the background on the right. – Photograph: G. F. Tschan.
Between 2020 and 2022, all sensors were successively installed at the Melgarten site (Malaise and pollen trap, Smellscapes module, wildlife cameras and animal sound recorders). Many of the components received updates during the course of the project, especially the acoustics and visual monitoring devices, as well as the base station. Some devices such as the pollen trap were still in operation in 2023 (
The automated pollen trap of the Metabarcoding module, February 2022. In the beginning, and at the end of the sampling season, the external power supply was used, when the sunlight was not strong enough to enable adequate supply via the solar panels. – Photograph: G. F. Tschan.
An important contribution to the understanding and subsequent acceptance of the use of the AMMOD technologies remained outreach, more specifically the explanation of the devices used on-site to interested visitors. Since the Melbgarten site is not open to the public, specific events were used to promote the project locally in 2022 and 2023, when the collections can be visited during the ‘Spring Festival’ (cf.
The ecological research station Britz of the Thünen Institute for Forest Ecosystems (
Habitat structures at the ecological research station Britz. (A) Coniferous forest. (B) Open space. – Photographs: K.-H. Frommolt.
Due to the relative closeness of the Museum für Naturkunde Berlin, regular visits of project personnel could be made to the site. In contrast to the Melbgarten site, however, not all devices were permanently employed here. As a benefit, the bioacoustic sensors received special attention at Britz (
Through the Museum of Nature Hamburg (formerly CeNak), contact was established with the operators of the so-called ‘Energieberg’ (literally meaning ‘energy mountain’) in Georgswerder, which is located northeast of Hamburg’s Wilhelmsburg quarter (
The industrial environment of the Energieberg, on the other hand, posed a challenge for the sensors. In particular, the acoustic sensors could only be used to a limited extent due to the noise pollution and were also only used for short time periods. The site was not commissioned for the project until the first experiences and results from the Melbgarten site were available, which was at the end of 2021. A base station and some of the sensors were then set up. The site was used for the project until May 2023.
Plant volatile organic compounds (pVOCs) are emitted by plants into the atmosphere. The emissions are influenced by a variety of abiotic and biotic factors (e. g. Herbivory, drought, heat, etc.) and can therefore provide information about the physiological status of plants within an ecosystem. However, ambient air is a complex and humid mixture and the concentrations of pVOCs are very low. Thus, highly sensitive and selective analytical tools are required for a continuous monitoring. In the AMMOD project, we installed an ion mobility spectrometer with coupled gas chromatographic pre-separation combined with an in-line preconcentration systems in the field (ppq-tec-GC-IMS). This allowed automated monitoring, with minimal maintenance and good results in terms of the robustness of the device in the field. Based on this, annual courses of emissions could be analysed from the ambient with high time resolution, revealing a clear seasonal course of emissions. Furthermore, 15 compounds have already been identified in reference experiments and assigned to plant origin, including typical green leave volatiles such as (Z)-3-hexenyl acetate or (Z)-3-hexen-1-ol and monoterpenes such as α-pinene, β-pinene and camphene.
In addition to seasonal changes, the temporal resolution was sufficient to record detailed diurnal concentration differences of individual volatile compounds. In our data, especially monoterpenes such as α-pinene showed a maximum in the morning hours, while other substances showed an early afternoon peak. Furthermore, correlations with abiotic factors could also be identified by comparing the data with weather data, whereby temperature seems to be the main driver.
The volatile organic compounds (VOC) released by living organisms carry information about their identity and physiological status.
VOC emissions from vegetation (plant volatile compounds = pVOC) are primarily determined by the taxonomic composition, the relative abundance of individual taxa and their phenology (
Plant abundance and phenology are of crucial importance for the consumers in a habitat and have a dramatic influence on the entire food chain. Thus, detailed real-time data on vegetation development would permit conclusions on ecosystem integrity and trends. A simultaneous documentation of anthropogenic compounds released into the ambient air (e. g. pesticides) would be of particular interest for understanding ecosystem reactions to xenobiotics. However, environmental VOC (eVOC; Sum of the VOCs that are in the ambient air at a given point in time) or more precisely plant VOC (pVOC) concentrations in ambient air are in the lower ppb down to the ppt level. The expected mixture of compounds is extremely complex and humid, and could be influenced by biogenic and anthropogenic emissions. Therefore, a sensitive and selective analytical tool is required for a continuous monitoring of characteristic volatiles. The most common and well-established analytical technique is mass spectrometry combined with GC pre-separation or GC combined with flame ionisation detectors (FID). However, broader time resolved measurements of VOC-fluxes using GC-MS or GC-FID often rely on highly sophisticated and extremely costly experimental set-ups, i. e. the Amazonian tall tower observatory or the prophet tower (Yáñez-Serrano et al. 2018; Fischer et al. 2021). Therefore, a rapid and automated analytical tool providing the necessary high selectivity and sensitivity is the ideal alternative.
Ion mobility spectrometry (IMS) is an analytical tool for detecting traces of ionised molecules in the gas phase after separation for size and shape. It is an extremely sensitive method that can detect molecules in the low ppb to ppt concentration range.
In general, a measurement can be divided into two phases, first the ionisation phase and second the separation phase. In the first phase, molecules are ionised in the ionisation chamber. Most commonly this is achieved via a ß-radiation source, which ionizes the drift gas flushing through the spectrometer, thus forming the so-called reactant ions in the ionization region. If a sample with other gas-phase compounds is introduced in ionization region, analyte molecules are ionized mainly by proton or charge transfer. In the second phase, separation of ionised molecules occurs. For this, the ions are accelerated in a weak electric field towards the detector, but only clouds of ions are introduced periodically into the drift region by a Bradbury-Nilsson ion grid. During their drift from the grid to the detector, the ions collide with the molecules of the drift gas counterflow (
Schematic illustration of an ion mobility spectrometer. Analytes are ionized and then accelerated in an electric field towards the detector. The drift time is influenced by the collision frequency of analytes with an opposing drift gas.
Collision frequency depends on size and shape of the ions, thus influencing drift velocity. Drift velocity can be determined by measuring the drift time of the particular ions under known drift length and the ion mobility is obtained by normalization to electric field strength. Further normalization of the measured drift time to the position of the reactant ion peak (RIP) is leading to reproducible relative ion mobilities, which are independent of instrumentation and environmental parameters such as temperature and air pressure (
Such separation and detection can be operated in the positive and negative ion mode just by changing the polarity of the electric field. Thus, alternating analysis of positive and negative ions is carried out. For details of ionisation and separation in ion mobility spectrometry see Eiceman et al. (2016).
Coupling fast gas-chromatographic (GC) pre-separation to IMS provides the characteristic retention time as additional measure for the identification of the analyte and furthermore avoids clustering of different analytes in very complex mixtures in the ionisation and drift region of the IMS. Analysis time of a full GC-IMS run is in the range of 10–30 minutes, depending on the experimental setup of the GC. Additionally, innovative in-line MEMS-based (microelectromechanical systems) in-line pre-concentration systems are applied. This further increases the sensitivity by ca. 1–3 orders of magnitude without increasing the total analysis time and allows automation of pre-concentration and analysis without separating these steps in time and space, enabling a high sample throughput (
After construction of the stationary GC-IMS for continuous measurements and optimization for the detection of VOCs in the ambient air, the prototype was implemented for long-term, time resolved measurements in the study area. Furthermore, a mobile version of the GC-IMS with the similar experimental setup was provided for measurements in the lab or in the field, e. g. directly to characterize different plants and plant parts. Concomitantly, a reference database was compiled. This includes an inventory of frequently occurring species in the Melbgarten, as well as a substance database for identification and characteristic emission patterns of common plant species. The aim is to use this reference database to clearly identify plant-related volatiles and, if possible, to assign signals from the long-term measurements to individual plants. In addition, trends are to be analyzed via the time-resolved measurements with regard to seasonality, diurnal and weekly variations and on the influence of particular weather conditions.
The resulting initial questions were “Are there differences in the emission patterns of different plants?”, “Can we detect pVOC signals in the ambient air?”, “Can we assign detected signals to plant sources?”, “Are there substances that show a diurnal/weekly cycle?”, “How do the emission data correlate with abiotic factors (temperature, humidity)?”.
In the following, we want to go into more detail about the methods and the structure of the station, as well as the first results of long-term time resolved measurements.
The sampling site is located at the Melbgarten (Bonn, Germany) at 50.71297°N, 7.09035°E (
Two identical ion mobility spectrometers are used for the measurements in the Smellscapes module. The first device is designated to stationary use and is installed in the monitoring station at the Melbgarten. The second device is additionally equipped with batteries and is used for reference measurements in the laboratory or mobile measurements in target areas.
The ion mobility spectrometers are ppq-tec-GC-IMS from ION-GAS GmbH (Dortmund, Germany) based on hardware provided by STEP GmbH (Pockau-Lengefeld, Germany). The default sample loop was substituted with a MEMS-based in-line pre-concentration chip filled with Carbograph 4 as adsorbent (CNR-IMM, Bologna, Italy, for detailed description see
The internal gas flow is purified via a filter system. In total, the GC-IMS has two external filters (Molsieve and activated carbon) and four larger internal filters filled with molsieve 9 Å. Filters need to be replaced approx. after 12–18 months. The filters are regenerated externally by the STEP GmbH.
Ionization is carried out with a tritium source of β-radiation (100 MBq). Drift length of the IMS is 5.61 cm. The drift tube is operated isothermally at 70 °C at a field strength of 300 V cm-1. The instrument provides an automatic polarity switch between measurements, which allows measurements to be made in both positive and negative modes.
The ppq-tec-GC-IMSs have built-in computers that permit on-site control, but also permit remote access via the internet (e. g. via TeamViewer). Data is stored on a built-in 256 GB hard disk. A user manual for detailed technical information and operating instructions of the ppq-tec-GC-IMS is available from the Nees-Institute and ION-GAS GmbH.
To protect the GC-IMS from harmful influences (weather, vandalism etc.), it is housed in a safety cabinet (SciCab12,
Time-resolved measurements of VOCs have been performed at the sampling site in the Melbgarten since 2021-03-22 (including pauses for optimization and maintenance). The main intervals of measurements were from 2021-06-15 to 2021-08-13 and from 2022-03-22 ongoing.
The samples for the GC-IMS are taken through the T-junction at a sampling rate of ~ 60 mL min-1 from the constant flow of ambient air. For the long-term monitoring, a total sample volume of 1000 mL was used for the enrichment (sampling time approx. 20 minutes, enrichment temperature = 43 °C). Subsequently, substances were released via heat desorption (50 °C for 5 s, then temperature ramp to 290 °C in 5 s, 290 °C for 12 s) and transferred to the GC-IMS for separation and detection. To eliminate carry-over, the pre-concentrator chip is purged at 300 °C for 2 minutes prior to each measurement. Furthermore, the IMS and GC should be baked out at regular intervals to prevent the accumulation of impurities. For this purpose, there is an internal bake-out procedure in which the IMS and GC column are heated above the usual sample temperature for 24 hours.
The duration of the GC pre-separation is 25 min, followed by rapid mobility separation and detection in the IMS (a few µs). After each measurement, the polarity of the IMS switches automatically for the next measurement to cover a wider range of compounds (positive and negative ionisation modes address different substance classes). Combined with a short sampling break (3 min) after each measurement, we are able to record one measurement per hour. In-line enrichment and analysis are fully automated and run 24/7. The data generated, including telemetry data, are directly uploaded into the AMMOD-cloud and are additionally stored on the built-in hard disk.
For signal identification we generated a reference database of common pVOCs and some characteristic anthropogenic compounds. Substances of interest were selected based on literature research and GC/MS reference measurements. Measurements of pure reference substance were conducted for validation. Furthermore, some substances could be identified by comparing Kovats retention indices against a series of n-ketones combined with GC/MS reference measurements. However, this is only suitable for less complex mixtures of volatile compounds.
The database contains signals for 26 aliphatic hydrocarbons, 27 aromatic compounds, 27 Monoterpenes and two nitrogen containing compounds. Most of these substances are biogenic, but they also include some predominantly anthropogenic substances such as toluene or xylenes. A list of all the substances identified is provided in Appendix 6.2 (
Reference measurements of individual plants were carried out in the laboratory and in the field. For measurements in the field, a means of transport (pram) was converted as required and equipped with additional storage area, suspension and fixing options (
Exemplary images for measurements of flower volatiles in the genus Narcissus in the field (A) and in the laboratory (B).
The sample inlet is positioned directly in front of the relevant plant part (e. g. the flower) and an appropriate sample volume (between 10 and 100 mL, depending on pVOC concentration) is taken for enrichment and subsequent analysis. In the case of unknown or unpredictable samples, it is advisable to proceed slowly from low sample volumes to higher sample volumes in order to avoid overloading the GC-IMS.
Alternatively, the plant parts of interest may be enclosed with inert material (e. g. frying hose) and the samples are taken from the enclosed volume. This static headspace is particularly beneficial for plant parts that are difficult to reach or for low emitting parts (e. g. foliage), as volatile compounds accumulate in the enclosed gas space. Under laboratory conditions, higher concentration of pVOCs compared to environmental measurements were expected. Therefore, the valve opening times of the PreConcentrator could be reduced with no influence on ion mobility and retention time, but leading to a reduction of signal intensity, and thus avoiding GC-IMS overload. The particular setup with regard to sensitivity was adapted to the specific plants and measurement conditions. Blank measurements and background measurements are conducted prior to sample measurements, to identify pre-concentrator background and background signals from ambient air.
Based on the procedure described above, over 40 samples of typical plants or plant parts, respectively, where already analysed. This included commonly distributed plants such as Elder (Sambucus nigra), beech (Fagus sylvatica), maple (Acer pseudoplatanus), hornbeam (Carpinus betulus), birch (Betula papyrifera), different species of rowan (Sorbus spp.), walnut (Juglans regia), cornel (Cornus mas), different species of cherry (Prunus spp.), apple (Malus sylvestris), Scots pine (Pinus sylvestris), snowdrop (Galanthus spp.), ramps (Allium ursinum), dandelion (Taraxacum spp.), daisy (Bellis perennis), daffodil (Narcissus), crocus (Crocus), lilac (Syringa vulgaris), herb-rober (Geranium robertianum), creeping thistle (Cirsium arvense), clover (Trifolium spp.) and greater stinging nettle (Urticia dioica). Furthermore, differences in floral scents within individual genera were investigated using the example of Narcissus.
IONysos, a custom-made software developed by ION-GAS GmbH, Dortmund, Germany is used for processing, evaluation and visualization. In a first step, drift times are normalized to obtain the relative ion mobility and a gaussian smoothing algorithm is applied to reduce noise. Next, the retention times of each measurement are aligned on the basis of ubiquitous occurring signal with known retention time, allowing comparison between different measurements. This is necessary, as small changes in gas flow can lead to small fluctuations in retention time. After blurring, normalization of drift time and an alignment of retention times, emission patterns are visualized as two-dimensional heatmaps. Substance-specific signal position is determined by the retention time in the GC and the relative ion mobility from the IMS. The peak volume of individual signals permits semi-quantitative statements of the abundance for individual compounds (
Exemplary visualization of the raw data as a 2-dimensional heat map (left) and in a 3-dimensional plot (right) using the example of the emission pattern of green daffodil (Narcissus viridiflorus). Three main components could be identified based on the retention time and the relative drift time as eucalyptol (1), benzyl acetate (2) and phenethyl acetate (3).
Substances are identified by a comparison of the retention times and relative ion mobilities with those of reference substances in our database. Unidentified substances are characterised by retention time and relative ion mobility and the peak data is stored for future identification. Peak volumes of the signals detected can be calculated across all measurements for semi-quantitative evaluation. Furthermore, telemetry data (e. g. GC temperature, IMS temperature, gas flow, etc.) is collected in addition to the actual measurements.. This permits (remote-) monitoring of the performance of the instrument.
The collected data is stored in a database based on MySQL (Oracle, Texas), a relational database management system (RDBMS). The database is composed of different tables These tables contain a list of species from the Melbgarten (SpeciesList), a list of reference substances (refcomp), information on sampled plants (samplestatus), emission patterns of sampled plants (emissionpattern) and information on the equipment used (device). These parts are linked to each other via so-called keys, allowing information from different tables to be accessed quickly (
Scheme of the relational database for plant emission patterns. The database contains various parameters that make it possible to identify the presence of certain signals in species that have already been measured individually.
Individual plant names are recorded with an ID in the SpeciesList table, together with taxonomic information. Data on substances that have already been identified as well as unknown signals are stored in the RefComp table. An ID is assigned to each signal or each substance, respectively with the corresponding retention time and reduced ion mobility. For identified substances, there is additional information on substance name, CAS number, retention index, mass weight and signal type (monomers or dimers).
Device-specific information is stored in the device table. Since only two devices are currently in use, there are only these two entries here.
The last two tables refer to the sample status, which contains information of reference measurements performed (e. g. plant part, plant status, polarity of the measurement, number of identified signals, origin of the sample) and the actual emission patterns recorded. The table emissionpattern contains information about the relative abundance of individual signals in reference measurements, but also information about the nominal signal strength.
Information can thus be extracted from the measurements employing different queries, while the keys guarantee database integrity. The aim is to make this database publicly available. The basic framework and functionality have already been implemented, but the expansion with measurements is still an ongoing work process.
Meteorological data were provided initially by the Meteorological Institute of Bonn University, whose weather station is located in Endenich, approx. 2.5 km from the sampling site. Later on, they were obtained directly from the weather station in the Melbgarten, which was implemented during the AMMOD project. The data considered here include temperature, air pressure, relative humidity, precipitation and global radiation, as well as wind data (wind direction and wind speed).
The year 2022 was an overall warm year with low precipitation. The summer months of July and August, as well as March and April were particularly dry. Even in October and November there were still high temperatures with rather little rainfall. The rainiest month was September, the driest month was August. The coldest month was March, the warmest August (
Monthly summary of weather conditions in Bonn for the measurement series in 2022.
Month | Mean temperature (°C) | Precipitation (mL) | Mean rel. Humidity (%) | Mean air pressure (hPa) |
---|---|---|---|---|
Mar | 8.3± 5.1 | 22.1 | 66.9 ± 20.5 | 1014.9 ± 8.7 |
Apr | 10.4 ± 5.2 | 45.4 | 82.2 ± 19.4 | 1005.0 ± 9.6 |
May | 16.6 ± 5.1 | 54.9 | 64.9 ± 19.0 | 1009.3 ± 5.1 |
Jun | 19.6 ± 5.1 | 89.8 | 68.9 ± 22.2 | 1007.3 ± 4.6 |
Jul | 21.1 ± 4.9 | 19.3 | 81.6 ± 24.8 | 1011.7 ± 4.1 |
Aug | 22.6 ± 5.0 | 8.6 | 69.7 ± 26.3 | 1008.3 ± 5.0 |
Sep | 15.7 ± 5.3 | 105.8 | 75.3 ± 18.6 | 1004.2 ± 6.8 |
Oct | 14.5 ± 3.8 | 47.8 | 81.7 ± 13.8 | 1009.7 ± 6.0 |
Nov | 9.9 ± 3.1 | 43.9 | 81.8 ± 12.1 | 1004.7 ± 10.5 |
The conditions in the summer of 2021 differed considerably from those of the following year (
Monthly summary of weather conditions in Bonn for the summer months of 2021. Long-term measurements in that year were limited to June to August.
Month | Mean temperature (°C) | Precipitation (mL) | Mean rel. Humidity (%) | Mean air pressure (hPa) |
---|---|---|---|---|
Jun | 20.7 ± 4.9 | 97.3 | 74.2 ± 19.5 | 1008.5 ± 4.8 |
Jul | 19.3 ± 3.4 | 206.4 | 78.9 ± 16.8 | 1006.2 ± 5.2 |
Aug | 17.9 ± 3.4 | 91.3 | 81.0 ± 14.3 | 1007.4 ± 5.1 |
Conventional descriptive parameters (e. g. mean, median, standard deviation) are calculated for the eVOC and weather data. Signal intensities are plotted against time for time-resolved immission patterns. Correlations are identified using Pearson correlation coefficient.
Shapiro-Wilk test and Levene’s test are used to test for normal distribution and the homogeneity of the variance. In case of non-normal distribution data, we use the non-parametric Kruskal-Wallis test as significance test and the Wilcox-test for pairwise testing of significant differences between two or more groups. Statistical tests and data evaluation are conducted with the open source software R (version 4.0.4;
Key points in the recording of long-term data are reliable automation and system stability to generate a high throughput of comparable measurements. We are able to achieve an automated measurement for environmental volatiles using GC-IMS by combining in-line enrichment and subsequent GC-IMS analysis. Minimal maintenance is required. However, sensor failure can still lead to occasional measurement gaps. Therefore, the functionality of the device needs to be monitored at regular intervals. Critical parameters for an overview of performance parameters of the GC-IMS are drift, sample and carrier gas flow, detector temperature, GC temperature, preconcentration temperature (precon-chip), Baseline, RIP position and RIP width at half height (WHM).
The mean sample gas flow during the measurement series was 60.74 ± 2.8 mL min-1 and the mean carrier gas flow was 21.53 ± 0.72 mL min-1 across measurements for both polarities (
Telemetry data of gas flow rates, GC and IM temperatures over a selected time period. Carrier gas flow and temperatures in the GC and IMS are very stable. The sample gas flow is subject to greater fluctuations, which can be attributed to the viscosity of air at different temperatures.
The mean enrichment temperature for the PreCon-Chip was 38.1 ± 3.93 °C. Enrichment temperature was slightly sensitive to ambient temperature, especially at high ambient temperatures. This can cause problems, especially in summer. Conversely, the enrichment temperature was broadly stable in low ambient temperature. The mean reduced inverted ion mobility of the RIP was 0.446 ± 0.002 Vs cm-2 for negative polarity and 0.487 ± 0.002 with a mean WHM of 0.20 ± 0.01. The baseline was constant at 4.97 V. We thus find that all critical parameters are reasonably constant around the target values. Ion mobilities and retention times can be aligned to known signals and a comparability of measurements over longer time periods is ensured.
We were able to record emission data for different plants and observed differences between the detected pVOC patterns. However, there were also plants where we could not detect volatile compounds under undamaged conditions (e. g. Trifolium pratense or Convolvulus arvensis). Individual substances could be identified on the basis of the retention time and the relative ion mobility (
Emission patterns of different plant species or plant parts, respectively, showing characteristic emissions of certain compounds.
There are distinct differences in the emission patterns of flowers and vegetative tissues. Flowers in particular stand out due to their diversity and, in some cases, high intensity of individual signals. Benzaldehyde, benzyl alcohol, (E)-β-ocimene, methyl salicylate, linalool, methyl benzoate or lilac aldehydes could be identified as typical components of floral odours. We also found as yet unidentified substances e. g. in different species from the genus Sorbus (
Comparison of emission pattern. (A) Characteristic patterns of different plants and plant parts, respectively. (B) Comparison of leaf emission patterns of Fagus sylvatica for undamaged and damaged leaves.
For vegetative tissue or green plant parts (e. g. leaves), respectively, little or no emissions where observed under undamaged condition and there is hardly any diversity in the composition of the substances. 6-methyl-5-hepten-2-one (e. g. Fagus sylvatica, Acer pseudoplatanus, Sambucus nigra, Quercus robur etc.) or (Z)-3-hexenyl acetate (Urtica dioica) were detected as characteristic substances of green plant parts. Pinus sylvestris represents a notable exception with several typical monoterpenes such as α-pinene, camphene or β-pinene detected. These common components of essential oils are widespread in, e. g. conifers. Damaged leaves, on the other hand, show significantly stronger emissions of substances such as (Z)-3-hexen-1-ol or (Z)-3-hexenyl acetate. In Sambucus nigra, even aromatic substances such as benzyl alcohol were emitted by damaged leaf tissue. In general, a comparatively small number of common and widespread substances are released from mechanically damaged leaf tissue.
During the first measurement phase in the summer of 2021 (n = 539 for positive ions and n = 541 for negative ions), a total of 64 positive ions and 32 negative ions were detected, amounting to a total of 96 unique signals. The measurement series for 2022 was noticeably longer (n = 2152 for positive ions and n = 1913 for negative ions) and 146 unique signals could be detected. The overall number of ambient VOCs present is likely higher, but some signal overlap compromises resolution. Despite these limitations, considerable sensitivity and selectivity could be demonstrated for both negative and positive ions. For positive ions, we observed some superimpositions for highly volatile compounds (retention time < 50 s), due to similar characteristics in retention time and ion mobility. Some of these signal clusters could not be resolved. Many short-chain carbon bodies (< 6 C-atoms) such as ethanol, ethylene, acetone, methanol and also the ubiquitous isoprene are included in this region. Separation was sufficient to identify individual signals at retention times > 50 s. Negative ionisation provides overall fewer and weaker signals, but selectivity is higher due to the lower degree of superimposition. Only very few substances are retrieved after 300 seconds, with the last peaks observed at ca. 700 s (both polarities). While we were not able to identify substances in the negative range, we could identify 15 substances amongst the detected signals for positive ionisation, including monoterpenes (MTs) such as α-pinene, β-pinene, camphene, DL-limonene, p-cymene, (E)-β-ocimene and Eucalyptol, as well as aromatic compounds (e. g. Benzaldehyde, Anisole, p-cresol) and aliphatic compounds such as 6-methyl-5-hepten-2-one or (Z)-3-hexen-1-ol. Surprisingly, we failed to identify any typical anthropogenic compounds (e. g. xylene or toluene). Floral volatiles were also less strongly represented than expected.
The data obtained in this way contain information on changes over time such as weekly averages for the sum of signal intensities, but also for individual substances (
Time resolved data as weekly average for the total signal intensity (Positive & negative ions) and α-pinene for the period of 2022.
In general, the measurement period of 2022, covers a wide range of seasonal variation and thus allows a more detailed view, compared to the rather short series of 2021. June was again the month with the highest average signal intensities. However, patterns across the summer months are less divergent, which may be due to the overall less variable weather patterns in 2022.
The data can also be utilized to look at weekly or diurnal variation, respectively. In principle, intra-week variation could be an indication of the anthropogenic origin of substances, because emission of air pollutants tends to be higher on weekdays (e. g. NOx and VOCs from traffic) Especially for the first period (March/April) it seems that there are differences between weekends and weekdays (
Weekly variation of total signal intensity and α-pinene during three seasonal intervals for 2022.
The diurnal changes can provide information about the emission behaviour of plants. Fine-grained analysis revealed that signal intensities across both polarities display significant diurnal patterns, which is particularly pronounced during the main period of activity from May to August 2022 (
Correlations with abiotic factors can also be investigated, in addition to time patterns. A comparison of pVOC data to meteorological data is relevant since meteorological factors are expected to affect pVOC-emissions.
Over the entire measurement period of 2022 (n = 2152), we obtained two more or less separate clusters with regard to the correlation to temperature, which reflect the abovementioned seasonality quite well. Almost all substances show a positive correlation to temperature across the year, which is reflected by total signal intensity. We find a strong positive correlation (Pearson correlation) to temperature, with R = 0.57 (p < 0.001) for the total signal intensity for positive ions and with R = 0.57 (p < 0.001) for negative ions (
Overall, we achieved good results in terms of the robustness of the measurements. The measurements were automated and maintenance was minimal, in particular as remote quality control was available. Ion mobility and the retention time from gas chromatographic pre-separation are essential for the identification or comparability of substances across a series of measurements. These parameters are heavily influenced by the temperature of the GC pre-separation, detector temperature and carrier gas flow. It was therefore important that these parameters were successfully kept sufficiently constant over the course of the measurement series. The automated normalization of ion mobility to the RIP (or RIN for negative polarity) also worked excellently and no deviations were observed.
Fluctuations occurred mainly for the enrichment temperature of the PreCon-chip. For the adsorption of VOCs, the chip is cooled to approx. 40 °C using air cooling. Especially at very high and very low ambient temperatures, a deviation from the target temperature can occur, as the temperature gradient is no longer sufficient for cooling. This is reflected in an enrichment temperature of ~ 50 °C during the hottest days in June (ambient temperature 39 °C). In theory, a higher enrichment temperature is associated with lower performance for adsorption. However, we observed no drops in signal intensity due to high temperatures, indicating that sufficient adsorption of volatiles also takes place at higher temperatures and/or that the effect of increased emissions of volatiles at higher temperature overcompensates the effect of reduced adsorption efficiency. In order to further improve reproducibility of the measurements, a cooling system could be implemented in the station in future.
Slight fluctuations were also observed in the sample gas flow. This is related to the viscosity of gases at different temperatures and air pressures. As the sample volume is fixed to 1000 mL, this factor only affects the duration of sample collection and should not affect signal intensity. In summary, the continuous stationary field measurements can be considered a complete success in technical terms.
We have here demonstrated for the first time a novel method for automatic long-term monitoring of volatile organic compounds in environmental air using a ppq-tec-GC-IMS as alternative to existing monitoring techniques such as PTR-MS, zNose® or relaxed eddy accumulation (
The actual number may well be higher, but a clear assignment of additional individual signals is difficult, especially for highly volatile substances, due to superimpositions. The technique, as currently implemented, thus reaches its technical limits especially for substances with < 6 C atoms (including isoprene). The time resolution of the measurement permits the identification of seasonal changes and a correlation to abiotic factors (weather) and in the long-term even interannual trends. Furthermore, it was possible to record time-resolved changes of pVOC-imission and observing even more fine-grained diurnal variation of certain signals.
Based on the parameters (retention time and ion mobility), it was possible to identify certain signals that can be directly assigned to a biogenic or plant origin, respectively. These substances comprise typical plant monoterpenes such as α-pinene, β-pinene, camphene, DL-limonene and eucalyptol. In addition, aromatic compounds such as benzaldehyde or p-cresol, but also aliphatic compounds such as 6-methyl-5-hepten-2-ones were identified. The location of the monitoring station and the fact that we found concentrations to be essentially uniform across the week, suggest that the vast majority of volatiles detected can be considered as biogenic. However, the substances that could be identified represent more or less ubiquitous plant volatiles and cannot at present be assigned to any particular plant species or plant group. Eucalyptol represents a notable exception and can be assigned to the emissions of Juglans regia at our specific sampling site with reasonable certainty.
Furthermore, anthropogenic air pollution, in particular from traffic, would usually be expected to have a significantly lower impact at weekends (
We have first promising results regarding the seasonal and diurnal variation of volatile organic compounds in ambient air, permitting the identification of trends. Many signals have not so far been identified as to chemical identity - the identification of additional signals and their inclusion in the reference database are therefore certainly major issues that need to be addressed in future. Due to the nature of GC-IMS, the physical signal parameters also permit a future identification of currently unidentified substances. An expansion of the sampling of plants from the target area to identify characteristic emitters will hopefully reveal characteristic source species for particular signals, permitting a clearer biological interpretation of the patterns observed.
However, one of the most important points should be the evaluation of the raw data. At present, the spectra are visually revised and the signals segregated and identified. With more than 4000 measurements per year from a single sampling locality, the volume of data generated is enormous. If this is extrapolated over several years, and possibly with an increasing number of stations, a visual approach for comparing individual emission patterns does not appear to be expedient. We are therefore planning to automate pattern recognition in cooperation with AI specialists.
The statistical analysis of the selected data only scratches the surface of what is possible. In fact, the interplay of abiotic and biotic factors can involve a complex network of interactions (
During the project, we were able to show that a GC-IMS is suitable for automated, continuous and time-resolved pVOC monitoring in the ambient air. Concentrations can be recorded in great detail and some of the signals could already be successfully assigned to plant sources. Significant progress was made in the area of the database, although at present it shows a strong bias towards floral volatiles. In the ambient air, however, these floral volatiles were found to represent only minor components of the ambient air, with other volatile compounds dominating.
All in all, it could be demonstrated that the sensitivity of the PreCon-GC-IMS is sufficient to detect VOCs even at the very low concentrations found in ambient air. The first clear seasonal and daily variations could be identified with statistically significant results. This is an encouraging starting point for future studies into the interaction between phytosphere and atmosphere, inviting a range of investigations on plant physiology and ecosystem function.
List of plants that can be found in direct vicinity of the measuring station.
Entry | Name | Vernacular name | Familiy |
---|---|---|---|
1 | Sambucus nigra | black elderberry | Adoxaceae |
2 | Viburnum opulus | guelder-rose | Adoxaceae |
3 | Allium ursinum | wild garlic | Amaryllidaceae |
4 | Galanthus sp. | snowdrop | Amaryllidaceae |
5 | Narcissus sp. | daffodil | Amaryllidaceae |
6 | Aegopodium podagraria | ground elder | Apiaceae |
7 | Heracleum sphondylium | hogweed | Apiaceae |
8 | Arum italicum | italian arum | Araceae |
9 | Arum maculatum | wild arum | Araceae |
10 | Hedera helix | common ivy | Araliaceae |
11 | Ornithogalum umbellatum | grass lily | Asparagaceae |
12 | Scilla bifolia | alpine squill | Asparagaceae |
13 | Bellis perennis | daisy | Asteraceae |
14 | Crepis sp. | hawksbeard | Asteraceae |
15 | Hiracium sp. | hawkweed | Asteraceae |
16 | Jacobaea vulgaris | stinking willie | Asteraceae |
17 | Sonchus asper | prickly sow-thistle | Asteraceae |
18 | Taraxacum sect. Ruderalia | dandelion | Asteraceae |
19 | Betula papyrifera | paper birch | Betulaceae |
20 | Corylus avellana | common hazel | Betulaceae |
21 | Carpinus betulus | common hornbeam | Betulaceae |
22 | Pulmonaria officinalis | common lungwort | Boraginaceae |
23 | Cardamine heptaphylla | pinnate coralroot | Brassicaceae |
24 | Cardamine bulbifera | coral root | Brassicaceae |
25 | Dispsacus fullonum | wild teasel | Caprifoliaceae |
26 | Convolvulus arvensis | field bindweed | Convolvulaceae |
27 | Cornus kousa | - | Cornaceae |
28 | Cornus mas | cornel | Cornaceae |
29 | Cornus officinalis | japanese cornel | Cornaceae |
30 | Carex sp. | true sedges | Cyperaceae |
31 | Lathyrus pratensis | meadow vetchling | Fabaceae |
32 | Medicago lupulina | black medick | Fabaceae |
33 | Trifolium pratense | red clover | Fabaceae |
34 | Trifolium repens | white clover | Fabaceae |
35 | Vicia sepium | bush vetch | Fabaceae |
36 | Fagus sylvatica | common beech | Fagaceae |
37 | Quercus robur | pedunculate oak | Fagaceae |
38 | Quercus rubra | northern red oak | Fagaceae |
39 | Geranium robertianum | herb-robert | Geraniaceae |
40 | Hypericum maculatum | spotted St. Johnswort | Hypericaceae |
41 | Juglans regia | english walnut | Juglandaceae |
42 | Glechoma hederacea | ground ivy | Lamiaceae |
43 | Lamium galeobdolon | yellow archangel | Lamiaceae |
44 | Prunella vulgaris | common self-heal | Lamiaceae |
45 | Stachys sylvatica | hedge woundwort | Lamiaceae |
46 | Malva sylvestris | common mallow | Malvaceae |
47 | Tilia cordata | small-leaved linden | Malvaceae |
48 | Tilia platyphyllos | large-leaved linden | Malvaceae |
49 | Circaea lutetiana | broad-leaved enchanter’s nightshade | Onagraceae |
50 | Corydalis cava | - | Papaveraceae |
51 | Fumaria officinalis | common fumitory | Papaveraceae |
52 | Plantago media | hoary plantain | Plantaginaceae |
53 | Veronica persica | birdeye | Plantaginaceae |
54 | Veronica sublobata | false ivy-leaved speedwell | Plantaginaceae |
55 | Arrhenatherum elatius | centaury | Poaceae |
56 | Brachypodium sylvaticum | false-brome | Poaceae |
57 | Bromus sterilis | barren brome | Poaceae |
58 | Dactylis glomerata | cock’s-foot | Poaceae |
59 | Elymus caninus | bearded wheatgrass | Poaceae |
60 | Festuca sp. | - | Poaceae |
61 | Holcus lanatus | tufted grass | Poaceae |
62 | Lolium perenne | perennial ryegrass | Poaceae |
63 | Poa annua | annual meadow-grass | Poaceae |
64 | Poa compressa | flattened meadow-grass | Poaceae |
65 | Poa trivialis | rough meadow-grass | Poaceae |
66 | Reynoutria japonica | asian knotweed | Polygonaceae |
67 | Rumex acetosa | common sorrel | Polygonaceae |
68 | Cyclamen coum | Eastern snowbread | Primulaceae |
69 | Lysimachia nummularia | creeping jenny | Primulaceae |
70 | Anemonoides nemorosa | wood anemone | Ranunculaceae |
71 | Ranunculus acris | common buttercup | Ranunculaceae |
72 | Ranunculus repens | creeping buttercup | Ranunculaceae |
73 | Geum urbanum | colewort | Rosaceae |
74 | Prunus avium | wild cherry | Rosaceae |
75 | Prunus spinosa | blackthorn | Rosaceae |
76 | Pyrus pyrifolia var. culta | asian pear | Rosaceae |
77 | Rubus caesius | european dewberry | Rosaceae |
78 | Malus baccata | siberian crab apple | Rosaceae |
79 | Malus sylvestris | european crab apple | Rosaceae |
80 | Prunus sp. | - | Rosaceae |
81 | Sorbus alnifolia | alder-leaved whitebeam | Rosaceae |
82 | Sorbus lancastriensis | lancshire whitebeam | Rosaceae |
83 | Sorbus aucuparia subsp. pohuashanensis | - | Rosaceae |
84 | Galium album | white bedstraw | Rubiaceae |
85 | Acer pseudoplatanus | sycamore maple | Sapindaceae |
86 | Urtica dioica | stinging nettle | Urticaceae |
87 | Viola odorata | sweet violet | Violaceae |
Substance | CAS-No. | Molar Mass | Rel. ion mobility | Retention Time in s | Retention Index | Signal Typea |
---|---|---|---|---|---|---|
Ammonia | 7664-41-7 | 17.0305 | 0.836 | 32 | 131 | M |
Butanone | 78-93-3 | 72.1057 | 1.085 | 40.4 | 350 | M |
Butanone | 78-93-3 | 72.1057 | 1.374 | 40.8 | 350 | D |
1-pentanol | 71-41-0 | 88.1482 | 1.333 | 50.1 | 760 | M |
2-pentanone | 107-87-9 | 86.1323 | 1.541 | 54 | 691 | D |
2-pentanone | 107-87-9 | 86.1323 | 1.162 | 54.3 | 691 | M |
Toluene | 108-88-3 | 92.1384 | 1.071 | 55.5 | 770 | M |
Toluene | 108-88-3 | 92.1384 | 1.291 | 55.5 | 770 | D |
1-pentanol | 71-41-0 | 88.1482 | 1.333 | 56.7 | 760 | M |
2-hexanone | 591-78-6 | 100.1589 | 1.715 | 69.1 | 790 | D |
2-hexanone | 591-78-6 | 100.1589 | 1.245 | 69.3 | 790 | M |
p-xylene | 106-42-3 | 106.165 | 1.135 | 73.6 | 859 | M |
o/m-xylene | 108-38-3 | 106.165 | 1.263 | 73.6 | 859 | M |
1-hexanol | 111-27-3 | 102.1748 | 1.864 | 75.5 | 869 | D |
1-hexanol | 111-27-3 | 102.1748 | 1.424 | 75.7 | 869 | M |
(Z)-3-hexen-1-ol | 928-96-1 | 100.1589 | 1.121 | 78.4 | 858 | M |
(Z)-3-hexen-1-ol | 928-96-1 | 100.1589 | 1.244 | 78.4 | 858 | M |
(Z)-3-hexen-1-ol | 928-96-1 | 100.1589 | 1.719 | 78.5 | 858 | D |
(-)-α-pinene | 7785-26-4 | 136.234 | 1.338 | 88.0 | 937 | M |
(-)-α-pinene | 7785-26-4 | 136.234 | 1.429 | 88.0 | 937 | M |
Cumene | 98-82-8 | 120.1916 | 1.250 | 92.0 | 912 | M |
Cumene | 98-82-8 | 120.1916 | 1.355 | 92.0 | 912 | D |
Camphene | 79-92-5 | 136.234 | 1.338 | 101.7 | 955 | M |
Anisol | 100-66-3 | 108.1378 | 1.1299 | 102.6 | 915 | M |
n-amylacetat | 628-63-7 | 130.1849 | 1.416 | 107.3 | 908 | M |
β-pinene | 127-91-3 | 136.238 | 1.341 | 112 | 980 | M |
Myrcen | 123-35-3 | 136.234 | 1.342 | 113 | 985 | M |
Myrcen | 123-35-3 | 136.234 | 1.994 | 113 | 985 | D |
2-heptanone | 110-43-0 | 114.1855 | 1.331 | 116 | 889 | M |
2-methyl-2-cyclopenten-1-one | 1120-73-6 | 96.1271 | 1.155 | 130.7 | 915 | M |
2-methyl-2-cyclopenten-1-one | 1120-73-6 | 96.1271 | 1.602 | 130.7 | 915 | D |
DL-limonene | 138-86-3 | 136.234 | 1.341 | 139.6 | 1031 | M |
DL-limonene | 138-86-3 | 136.234 | 1.433 | 139.8 | 1031 | M |
Ocimene(Isomer) | - | 136.234 | 1.343 | 145 | - | M |
p-cymene | 1195-32-0 | 132.2023 | 1.342 | 145.4 | 1025 | M |
Benzaldehyd | 100-52-7 | 106.1219 | 1.210 | 159.2 | 936 | M |
Eucalyptol | 470-82-6 | 154.2493 | 1.432 | 159.4 | 1035 | M |
Benzaldehyd | 100-52-7 | 106.1219 | 1.664 | 159.5 | 936 | D |
(E)-β-ocimene | 13877-91-3 | 136.234 | 1.336 | 160.6 | 1037 | M |
(E)-β-ocimene | 13877-91-3 | 136.234 | 1.948 | 161.4 | 1037 | D |
β-ocimene-nebenpeak(Isomerengemisch) | - | 136.234 | 1.3779 | 161.5 | - | M |
2-ethyl-1-hexanol | 104-76-7 | 130.2279 | 1.266 | 169.3 | 1030 | M |
2-ethyl-1-hexanol | 104-76-7 | 130.2279 | 2.103 | 169.3 | 1030 | D |
Hexenyl-acetat | 3681-71-8 | 142.1956 | 1.128 | 176.6 | 1009 | M |
Hexenyl-acetat | 3681-71-8 | 142.1956 | 1.346 | 176.6 | 1009 | D |
6-methyl-5-hepten-2-one | 110-93-0 | 126.1962 | 1.291 | 181.1 | 985 | M |
Benzyl alcohol | 100-51-6 | 108.1378 | 1.055 | 193.9 | 1032 | M |
Benzyl alcohol | 100-51-6 | 109.1378 | 1.305 | 193.9 | 1032 | D |
Benzyl alcohol | 100-51-6 | 110.1378 | 1.524 | 193.9 | 1032 | T |
2-octanon | 111-13-7 | 128.212 | 1.421 | 195.1 | 999 | M |
2-octanon | 111-13-7 | 128.212 | 2.045 | 197.2 | 999 | D |
p-cresol | 106-44-5 | 108.1378 | 1.188 | 222 | 1084 | M |
Linalool | 78-70-6 | 154.2493 | 1.338 | 254.0 | 1098 | M |
Acetophenone | 98-86-2 | 120.1485 | 1.238 | 305 | 1078 | M |
Acetophenone | 98-86-2 | 120.1485 | 1.804 | 305 | 1078 | D |
Methyl benzoate | 93-58-3 | 136.1479 | 1.275 | 310.1 | 1100 | M |
Methyl benzoate | 93-58-3 | 136.1479 | 1.851 | 310.1 | 1100 | D |
α-thujone | 546-80-5 | 152.2334 | 1.411 | 337.6 | 1100 | M |
α-thujone | 546-80-5 | 152.2334 | 2.105 | 337.6 | 1100 | D |
2-nonanone | 821-55-6 | 142.2386 | 1.5089 | 355.5 | 1093 | M |
2-nonanone | 821-55-6 | 142.2386 | 2.206 | 355.7 | 1093 | D |
β-thujone | 471-15-8 | 152.2334 | 1.488 | 400.7 | 1115 | M |
β-thujone | 471-15-8 | 152.2334 | 2.192 | 400.7 | 1115 | D |
DL-α-terpineol | 98-55-5 | 154.2493 | 1.344 | 466.7 | 1150 | M |
DL-α-terpineol | 98-55-5 | 154.2493 | 1.431 | 467.8 | 1150 | M |
Lilac aldehyde B | 53447-45-3 | 168.2328 | 1.416 | 469.4 | 1154 | M |
Benzyl acetate | 140-11-4 | 150.1745 | 2.062 | 487.9 | 1165 | T |
Benzyl acetate | 140-11-4 | 150.1745 | 1.48 | 488.1 | 1165 | D |
Benzyl acetate | 140-11-4 | 150.1745 | 1.05 | 488.3 | 1165 | M |
Lilac aldehyde A | 53447-46-4 | 168.2328 | 1.424 | 516.5 | 1155 | M |
Methyl salicylate | 119-36-8 | 152.1473 | 1.303 | 558.9 | 1198 | M |
Lilac aldehyde D | 53447-47-5 | 168.2328 | 1.426 | 580.8 | 1169 | M |
Ketoisophorone | 1125-21-9 | 152.1904 | 1.307 | 643.9 | 1142 | M |
Umbellulone | 24545-81-1 | 150.2176 | 1.386 | 691.0 | 1186 | M |
3,5-dimethoxytoluene | 4179-19-5 | 152.1904 | 1.337 | 808 | 1270 | M |
Anethole | 4180-23-8 | 148.2017 | 1.333 | 846 | 1284 | M |
Anethole | 4180-23-8 | 148.2017 | 2.049 | 846 | 1284 | D |
Indole | 120-72-9 | 117.1479 | 1.235 | 963.6 | 1292 | M |
Neral | 106-26-3 | 152.2334 | 1.114 | 1016.0 | 1240 | M |
Neral | 106-26-3 | 152.2334 | 1.310 | 1016.0 | 1240 | D |
Phenetyl acetate* | 103-45-7 | 164.2011 | 1.404 | 1047.3 | 1229 | M |
Cinnamaldehyde | 455-522-2 | 132.1592 | 1.316 | 1199.4 | 1265 | M |
Geranial | 141-27-5 | 152.2334 | 1.514 | 1246.2 | 1275 | M |
Eugenol | 97-53-0 | 164.2011 | 1.428 | 1504.9 | 1356 | M |
GLV Green leaf volatiles
VOC Volatile organic compound
pVOV plant volatile organic compound
eVOC environmental volatile organic compound
IMS Ion mobility spectrometry
GC Gas chromatography
MEMS Microelectromechanical system
MS Mass spectrometry
ppb parts per billion
ppt parts per trillion
The development of high-throughput and quality techniques to monitor plant biodiversity and plant-insect interactions has become critical with the documented decline in biodiversity and abundance of both plants and insects. Metabarcoding has the potential to outperform conventional methods for large-scale biomonitoring due to lower cost, potential for automation, and lack of technician bias. Here we present the optimization of plant trace monitoring within the AMMOD stations via metabarcoding, with two different collection methods, airborne pollen from Hirst traps and plant traces from the preservative ethanol of Malaise traps. A new wind pollen trap was developed that allows autonomous operations in the field over a longer period than previous models, and offers the more options for sampling intervals. Recommendations for laboratory processing are made, and an optimized data analysis workflow are described.
In the age of accelerated changes to biodiversity due to anthropogenic influences and climate fluctuations, increased monitoring of biodiversity, as well as increased speed in data retrieval are critical. Traditional methods of pollen identification as well as monitoring plant-insect interactions require expert knowledge with a large amount of training and are extremely time consuming. Pollen of several plant taxa can only be identified to family or genus level, such as Poaceae, a major allergen, for which the pollen is usually only identifiable to family level with morphological characters. In addition, monitoring plant-insect interactions traditionally has involved many hours in the field with direct observation.
Airborne pollen monitoring via the Hirst spore trap is used worldwide. In Germany, the Stiftung Deutscher Polleninformationsdienst (PID) is using this method since 1983 to monitor pollen and fungal allergens. Malaise traps are a common trap used to collect flying insects and to monitor their populations across the world. Both the Malaise and the Hirst traps were optimized for single sample intervals and require human intervention after each interval. Optimizations for multiple autonomous sampling in the field and a considerable extension of the necessary service intervals for metabarcoding purposes were addressed in this project.
Here, we have developed and optimized tools and workflows for metabarcoding of plants. Metabarcoding allows for automated high throughput species identification, even though the methods require further development, and at present it is still necessary to transport the samples to a laboratory for processing. Future miniaturization in laboratory equipment is emerging, such as Nanopore sequencing with a MinION sequencer, which is quite small in size (105 mm × 23 mm × 33 mm) and can be attached to a laptop for sequencing outside of traditional molecular labs. With further developments, it is conceivable that in a few years monitoring and species identification can be automated in the field. In this study we developed, tested, and optimized preparatory steps for long term, continuous monitoring of plants and insects via passive collection and DNA metabarcoding.
Collection sites for the airborne traps should be in an open habitat, as bushes and trees, as well as boulders, buildings, streets and railroads in the near vicinity can lead to air turbulation and will affect the pollen collection. The trap should be equipped with a built-in spirit level and height-adjustable feet to level the device even on uneven ground and make it easier to turn depending on the wind direction. When a trap is not able to rotate in the direction of the wind, it results in non-optimal catch conditions, which influence the comparability of the results. The trap is designed to operate in all weather conditions in the field and can be powered by a solar panel and battery. However, the currently available solar modules do not yet guarantee year-round continuous operation and therefore require connection of the trap to a power grid. The control module and the battery are located in a lockable compartment of the trap, where a GPS tracker could also be attached. To reduce rodent infestation, all openings were reduced in size compared to the original prototype.
We made changes to the design of the Hirst spore trap (
Interior of the A1 Volumetric Air Sampler, with the cover removed, revealing the carousel for 24 sample tubes (Photograph: Gulzar Khan, test site Britz).
The sampling intervals for the wind-dispersed pollen was coordinated with the sampling of the Malaise (insect) trap to compare airborne pollen with the pollen transported by insects (fan suction time: 45 min/h; fan cycle: 75 %; fan duration: 1350; day mode only; sampling period: 7 days per tube).
DNA extraction and PCR were performed in a laboratory designed to minimize external contamination. Best practice recommendations for highest quality assurance in the laboratory are the following:
Following removal of insects, the preservative ethanol from the Malaise trap samples were vacuum filtered using a cellulose nitrate (CN) membrane (GVS Filter Technology, Sanford, USA; diameter 47 mm and 0.22 µl pore size,
Vacuum filtration system for the preservative ethanol from the Malaise trap insect samples. (Compare with Figure 5).
DNA isolation was performed with the NucleoMag Plant kit (Macherey-Nagel, Düren, Germany) for all samples and two DNA extraction blanks. Prior to DNA isolation three sterile tungsten carbide beads (Qiagen, Hilden, Germany; diameter 3 mm) and 600 µl of lysis buffer MC1 were added to the filter paper in the microcentrifuge tube. These samples were disrupted in a Retsch MM400 bead mill (Retsch GmbH, Haan, Deutschland) for 2.5 minutes at a frequency of 30 Hz. 10 µl of Proteinase K (20 mg/ml) and 5 µl of RNase A were added to the ground material and the samples were incubated at 60 °C for 45 minutes with constant shaking in a ThermoMixer C (Eppendorf, Hamburg, Germany; 400 rpm). The incubation step was followed by centrifugation of the samples (12,000 × g; 10 min). 400 µl of the cleared lysate were transferred to a clean 2 ml microcentrifuge tube and 400 µl of binding buffer MC2 and 10 µl of magnetic beads were added. In the further steps of the manufacturer’s protocol, only 25 % of the specified amounts of MC3, MC4, 80 % ethanol and MC5 were used. The washing step with ethanol was performed twice. 35 µl of MC6 elution buffer (55 °C) was added to the samples and incubated at room temperature for 5 minutes. Up to 30 µl was removed after application to magnets for a final elution volume.
For DNA isolation of pollen from air, we recommend the use of the DNeasy PowerSoil Pro Kit (Qiagen, Hilden, Germany), which seems to overcome the humic acid components as well as other PCR inhibitors contained in these sample types. Some changes to the standard protocol are recommended. The Power Beads provided in the kit were transferred from the Power Beads Pro Tubes to the 2 ml micro tube containing the airborne pollen. This step is necessary because it is not possible to transfer the collected pollen directly. After adding 800 µl of the lysis buffer CD1, the tubes were placed in a Retsch MM400 bead mill for 2.5 minutes at a frequency of 30 Hz. A subsequent centrifugation step was carried out at 15,000 × g for 5 minutes. All other steps were performed as described in the manufacturer’s protocol, except for the elution step; here only 50 µl of the C6 solution was used.
Several publications have addressed the quality of plant specific barcoding regions (
Here, we used the nuclear internal transcribed spacer region (ITS) for plant identification which has strongly conserved primer binding sites in adjacent mRNA genes. Kolter and Gemeinholzer (
PCR allows for the amplification of specific barcoding regions as well as tagging specific samples with unique labels, as metabarcoding is achieved via parallel sequencing of multiple samples in one run (
PCR was performed with three replicates per sample. Two DNA extraction blanks and two PCR blanks were added to every 96 well PCR plate. The Canadian Centre for Barcoding PlatinumTaq Protocol (
After PCR, 5 µl of each of the three PCR replicates of a sample were combined to a total volume of 15 µl and purified with 1.5 µl Exonuclease 1 (Thermo Fisher Scientific, Waltham, USA) with 37 °C incubation for 30 minutes followed by 80 °C incubation for 15 minutes. Illumina sequencing was performed at LCG Genomics GmbH (Berlin) on a MiSeq (2 x 300 bp) platform with 12 additional cycles. For the additional PCR cycling MyTaqTM Red Mix polymerase (Meridian Bioscience, Cincinnati, USA) was used and it consisted of three cycles with a low annealing temperature (15 sec 96 °C, 30 sec 50 °C, 90 sec 70 °C), followed by nine cycles with increased annealing temperature (15 sec 96 °C, 30 sec 58 °C, 90 sec 70 °C).
We recommend double-stranded sequencing on a MiSeq or NextSeq platform (2 × 300 bp, Illumina, San Diego, USA), depending on the amount of parallel sequencing and sequence read depth. This platform has currently the highest sequence accuracy and quality score. Long reads can also be performed with a MinION-tool (Oxford Nanopore Technologies, Oxford, United Kingdom) which has the benefit of being small and transportable, but due to the high error rates, species level identification via metabarcoding is currently complicated. Another long-read sequencing platform is the Revio or Sequel platform (Pacific Biosciences, Menlo Park, USA). This PacBio sequencing platform has a shorter sequencing time than the MinION, but the sequencing equipment is very expensive, and the sequences have high error rates. In addition, this sequencing platform is significantly larger than the MinION and can therefore not be taken into the field (
The sequencing data were demultiplexed by index sequences into three different files, and primers and the index sequences were trimmed with Cutadapt (
Several studies have demonstrated metabarcoding to provide identifications of pollen to a lower taxonomic level than traditional morphological methods, and identification of plant traces from digested food material or plant fragments is nearly impossible with morphological characters (
We have developed a volumetric air sampler that is powered via an external connection or a solar cell and battery, but delivers samples that are compatible with historical pollen counts. There are several passive pollen traps worldwide that do not require electricity, for example the Durham pollen trap (
In the workflow presented here, molecular processing and sequencing of samples must be performed in the laboratory. However, analysis of environmental samples in the field will be feasible in the near future. Miniaturization of next generation sequencing (NGS) in the field using the MinION (Oxford Nanopore Technologies) is already possible. However, nanopore sequencing with the MinION currently still struggles with high error rates in sequences, especially those with a high GC bias (e. g. Delahaye and Nicolas et al. 2021) which is an issue in several plant families. For this reason, we implemented the higher-quality MiSeq sequencing technology, which can only be performed in the lab. Portable thermal cyclers for on-site PCR experiments (e. g. miniPCR, https://www.minipcr.com/portable-pcr-testing-the-minipcr-for-dna-sequencing-in-the-field/) and also DNA extraction kits are already available on the market, but these are not yet optimized for plant traces and pollen.
Caution must be taken in pollen metabarcoding and we stress the importance of a sterile laboratory infrastructure. Contamination from airborne pollen is extremely high in insect-mediated pollen studies (e. g.
The best practice recommendations during laboratory procedures presented here are critical due to the high sensitivity of the sequence detection method.
For the analysis of metabarcoding data, complete reference databases are needed, both regarding the taxa represented and replicates of these, in order to accurately identify all species in a given sample. Universally available databases that are frequently updated would increase the comparability and utility for plant metabarcoding projects worldwide.
Despite optimized metabarcoding workflows and reference databases, species-level identification will not be possible for some rapidly evolving plant groups due to hybridization, polyploidization, apomixis and adaptive radiation. The incorporation of additional molecular DNA barcodes poses an assignment problem and is expensive. For species complexes, this is unlikely to be a solution. Due to the highly variable size of plant genomes, whole genome sequencing (WHS) or reduced complexity WHS is also currently unrealistic. However, complementing metabarcoding results with existing vegetation information on local occurrences and biological processes with deep learning is promising, but still requires the development of algorithms and the crosslinking of already existing databases. Furthermore, complementing metabarcoding with multispectral flow cytometry could enable not only species identification but also abundance estimation (
Further development of plant metabarcoding is promising and also serves to achieve UN development goals by monitoring not only species and ecosystems, but also genetic diversity. Swenson and Gemeinholzer (
The technology developed here can be used to detect changes in biodiversity, set up early warning systems that provide a basis for further preciseness as technology develops. Plant-insect interactions and aerial pollen flight times can be monitored in a high-throughput process with unprecedented accuracy and speed, and new insights into biodiversity interactions can be gained, which was previously not possible.
In the AMMOD project, flying insects are collected using Malaise traps equipped with automated bottle changers (see Chapter 5). Insect catches are identified using DNA metabarcoding, a promising tool for biodiversity assessments, especially targeting highly diverse groups such as the arthropods. Non-destructive DNA isolation methods are highly desirable for the preservation of sample integrity and subsequent morphological analysis of specimens. In this chapter, we present a comprehensive step-by-step laboratory protocol for the non-destructive DNA extraction of insect bulk samples from lysis buffer followed by all subsequent amplicon library preparation steps required for the sequencing of insect bulk samples on Illumina platforms.
DNA isolation, lysis buffer, fixative, sample integrity, non-destructive, metabarcoding, amplicon library preparation
While the morphological identification of every individual specimen in complex arthropod bulk samples is extremely time consuming and even impossible for certain taxonomic groups, DNA metabarcoding represents a fast and reliable tool, enabling high-resolution biodiversity assessments. The method is based on the bulk isolation of genomic DNA and the subsequent amplification of a specific marker gene fragment from mixed samples (
The metabarcoding workflow introduced here is a two-step PCR protocol based on
(See Note 3)
PCR 1 (amplicon PCR):
PCR 2 (index PCR): Incorporation of Illumina index adapters:
(See Note 5)
(See Note 6)
Proceed with a left side size selection of the sample pool using magnetic beads at a ratio of 1 : 0.7 (PCR product : beads) to remove primer dimers (SPRIselect, Beckman Coulter) using a slightly modified protocol (see below). Depending on the sample and fragment size, other PCR product to beads ratios may be considered (see Users Guide SPRIselect, Beckman Coulter).
Preparation steps:
Clean up:
Note 1: The primer pair used here, from
Note 2: Different Illumina indexing strategies exist, including unique dual indexing and combinatorial indexing. In general, unique dual indexing, which requires distinct, unrelated index sequences for each of the i5 and i7 index reads, will allow the highest accuracy and allow detection of index hopping (https://support.illumina.com/bulletins/2018/08/understanding-unique-dual-indexes--udi--and-associated-library-p.html). A list of Illumina Nextera i5 and i7 indexes is available here (https://support-docs.illumina.com/SHARE/AdapterSeq/Content/SHARE/AdapterSeq/Nextera/DNAIndexesNXT.htm).
Note 3: Here, insects are size sorted into two size fractions (< 4 mm, ≥ 4 mm) using a wire mesh sieve prior to DNA extraction and library preparation to ensure detection of smaller sized specimens. Mesh size and number of fractions needed will depend on the range of insect sizes in the sample which is highly variable depending on sampling location (temperate versus tropical). Size fractions can be either processed separately, or pooled again at the lysate stage in a ratio favouring the small size fraction and thereby enriching the sample with the smaller specimens (
Note 4: Note on the use of experimental controls: In general the use of DNA extraction negative controls as well as positive controls is recommended. See
Note 5: Normalisation of the samples is required to ensure equal sequencing depth across samples. Formerly, this was done manually, but this process can be highly simplified for large sample numbers using a limited binding capacity solid phase kit in plate format. The SequalPrep kit (Thermo Fisher Scientific) yields up to 25 ng of DNA per well when starting with at least 250 ng of PCR product.
Note 6: Most metabarcoding projects are sequenced on Illumina platforms due to low error rates and long read lengths. The number of samples that can be pooled depends on throughput of the selected Illumina sequencing platform and desired sequencing depth for each sample (e. g. species rich samples with higher biomass should have a higher sequencing depth, of at least 200,000 reads per sample). Similarly, amplicon length will determine which sequencing kit should be used (e. g. MiSeq 250PE, HiSeq 150PE or 250PE). Currently, the NovaSeq 6000 system provides the highest sequencing depth and throughput of all Illumina platforms, and is set to replace the HiSeq 2500 which is now discontinued (https://www.illumina.com/systems/sequencing-platforms/novaseq/specifications.html).
While sample integrity is fully given when DNA is isolated from sample fixative (e. g. through ethanol filtration or evaporation), the incubation time in lysis buffer solution directly influences the quality of sampled specimen preservation. After two to four hours of incubation, no decrease in sample integrity was observed but the abdomen of especially small specimens with soft body structures started to dissolve after eight hours of lysis (
The non-destructive DNA metabarcoding protocol from commercial lysis buffer solution presented here reveals comparable biodiversity estimates (species number and composition) of a complex arthropod mixture as destructive extraction from homogenised sample tissue (
The development and refinement of automated biodiversity assessment technologies could be a game changer for the identification of drivers of ongoing insect declines. In the AMMOD (“Automated Multisensor Stations for Monitoring of BioDiversity”) project, experts from various disciplines joined forces to develop an autonomous multi-sensor station for the assessment of biodiversity, which can provide species occurrence data across trophic levels and taxonomic groups (
In times of climate change and biodiversity declines the protection of our ecosystems is of major importance. Insects, one of the most diverse groups on earth, are indispensable for the function of ecosystems as they provide fundamental services such as pollination (
Metabarcoding, a genetic approach developed in the last 10 years with the emergence of next-generation sequencing, enables the rapid identification of several thousand of specimens in parallel, partly circumventing the increasing shortage of taxonomic experts. Despite these technological advances in the lab, permanent biodiversity surveys still continue to require human intervention to capture specimens for monitoring on a regular basis. Many of the traditionally used insect traps like Malaise traps, vane traps and pitfall traps are usually emptied in a bi-weekly cycle, depending on weather conditions, season and chosen fixative, to ensure specimen quality. Especially in rural areas or in extreme environments, which are often only poorly accessible, permanent insect surveys are often excessively difficult or virtually impossible, unless they rely on a team of local volunteers to empty the traps. Here, we present a prototype for an automated insect trap which can sample up to 12 bulk samples without human intervention.
As a passive flight interception trap, the Malaise trap has a broad target spectrum, including all insects actively flying through the habitat but also some ground dwelling taxa (
The automated AMMOD multisampler replaces the traditionally used collection bottle. The AMMOD multisampler has a housing measuring 66 cm × 66 cm × 35.5 cm which is mounted onto a tripod stand (Nedo, heavy-duty Aluminium Tripod, Ref.-No. 200204). The height of the tripod stand must be manually adapted to the height of the apex of the trap, which should be around 180–190 cm above ground. Here, we used a Malaise trap which was designed by the Entomological Society Krefeld (
The AMMOD multisampler consists of six subsystems: (1) the rotary mechanism, (2) the user interface, (3) the external sensors and modules, (4) the stopping mechanism, (5) the energy supply and (6) the microcontroller (
Setup of a Townes Malaise trap (Krefeld model) equipped with an automated AMMOD multisampler.
The entire system is controlled by an Arduino Mega 2560 microcontroller which is connected to each of the five other subsystems. Like all microcontrollers the Arduino comes with a central processing unit (CPU), input/output (I/O), memory, and peripherals. The programming language used is ‘Arduino language’, which is based on the C programming language. The Arduino board is comparatively cheap and comes with the freely available development tool ‘Arduino Integrated Development Environment’ (IDE) used to write code and upload it to the board. The Arduino IDE can be freely downloaded from https://arduino.cc/en/software and is available for all major platforms (Windows, Linux and Mac). Another great advantage of Arduino is the wide variety of available libraries, boards and extensions, making the system modular and easily adaptable for future requirements.
The Arduino Mega 2560 was chosen because of its low power consumption, paired with a high number of pins enabling the connection of several modules. Additionally, it comes with a pull-up resistor, allowing the connection of low-power sensors and modules. In contrast to other microcontrollers, the Arduino Mega 2560 supports up to 12 V DC input power due to a built-in voltage regulator (see section 5.7, ‘Energy supply’).
The code of the controller is separated into 4 main sections (
Additionally to the controller code, there are two more codes which are used to assist the development.
The rotation plate system (RPS) is installed under the upper cover plate. At first sight, the rotation plate system consists of three matching round plexiglass plates which are vertically aligned and fixed by a steel pole through the center of the plates. The lower and the middle plate are located approximately 10.5 cm from each other (from here on referred to as lower RPS). 13 plexiglas dividers ensure equal distance and alignment of the two plates. The upper plate is located approximately 2.5 cm above the middle plate and is fixated with plexiglas dividers, again to ensure equal distance between the two plates (from here on referred to as upper RPS).
The lower RPS is divided into 13 equal sized sections by the plexiglas dividers, each providing a screw-in connector consisting of a S65 thread fitting for commercial wide mouth bottles (Kautex 1000 ml) and an opening towards the upper cover plate. Openings are located on a circle around the center of the rotation plate with equal distance to each other. When rotating, each opening is pushed under the connection adapter, forming a passage between the trapping bottle and associated collection bottle. To prevent insects from escaping through the small gap between the upper side of the rotation plate and lower side of the lid plate, openings are surrounded by Teflon foil on the upper site. A total of 12 collection bottles can be screwed into the multisampler using screw-in connector positions 1 to 12. Position 13 should remain empty to ensure that insects can escape from the trap when the AMMOD multisampler is in standby mode (e. g. between, before, and after sample trials).
The upper RPS is surrounded by a toothed belt which is driven by a 24 V DC electric wiper motor. The motor is controlled by a H-bridge, hooked up with the microcontroller. As the system uses a 12 V battery for power supply, the motor operates at half speed allowing enough time for signal processing between stopping mechanism sensors (see section 5.4) and microcontroller.
To align trapping and collection bottle two sensors are implemented: (1) a stopping switch and (2) a magnetic switch (
To allow for a wider variety of research questions, but also to monitor the system itself, the AMMOD multisampler is equipped with several sensors. The sensors can be divided into two categories: environmental sensors and system sensors.
The environmental sensors periodically collect air and soil temperature, relative soil moisture, relative air humidity, and illuminance. All sensors are located outside of the control system box but are physically connected to the control system with (long) cables. On the one hand this allows for more flexibility in choice of sensor position, but on the other hand bears the risk that wildlife gets entangled in the cables. Therefore, sensor position should be thoroughly considered. While there are two different sensors implemented for soil moisture (Capacitive Soil Moisture Sensor v.1.2) and soil temperature (DS18B20), a single sensor measures air temperature and moisture (DHT-22). However, this sensor is not waterproof and should under all circumstances be located beneath the AMMOD multisampler. To measure illumination a BH1750 light sensor has been implemented. To fulfill its purpose the sensor must be exposed to sunlight, although it is not waterproof. To avoid sensor failure, the sensor is placed inside an acrylic glass box on top of the multisampler.
The system sensors are implemented to monitor system-related data to check for eventual errors and to ensure measured environmental data quality. In detail two parameters are monitored: controller enclosure data and system stand.
The system stand data checks for the upright position of the system. As the AMMOD multisampler is comparatively heavy and mounted to a tripod stand, strong winds could topple the multisampler or the tripod can tilt in softened rain-sodden soils. The ADXL335 accelerometer is attached to the downside of the controller box lid. The sensor measures its position in a three dimensional room sending three acceleration values to the microcontroller. If the sensor position changes, the communication system is triggered and an error SMS is sent to the user and the running sampling program is immediately ended. The importance of the sensor should not be underestimated as a collapse of the system will result in data loss and could possibly damage the system.
The second important system sensor controls the enclosure conditions. The controller box houses the highly sensitive electronic control unit which connects and partly contains the 6 subsystems of the AMMOD multisampler. To control ambient temperature and humidity in the controller box a BM280 air sensor is used. The sensor monitors operating conditions and informs the user in the case of exceptionally high temperature which could possibly damage the system and the samples. Additionally, the sensor monitors unexpected increases in humidity. Although the controller box is waterproof, it cannot be guaranteed that no plug or cable fitting becomes loose over time, allowing water to enter the box.
Next to a real-time clock (RTC), which provides the actual time and date, the system comes with an SD card and a communication module. The SD card module equipped with an 8 GB SD card is used as a hard disk drive to log information about trap performance in a simple text file (log file). The log file is an important tool as it allows to keep track of sensor data, but also system failures. As soon as the system is turned on, booting information is collected including local time, machine ID, UTC time and geographical parameters (longitude, latitude). With the start of a new sampling cycle (program start) every 5 minutes the measured sensor values but also the in-use collection bottle is logged. Additionally, the log file keeps track of system failures, which will be indicated with an “E” next to any questionable sensor value sensor value.
The implemented communication module functions on European frequencies and consists of a Plug and Play Arduino shield. The shield comes with a GPRS/LTE and GNSS antenna. Like all external modules is the communication module placed inside the metallic AMMOD multisampler box which can reduce signal quality. To amplify the signal, an external antenna is installed outside the AMMOD multisampler and connected via an antenna cable using an adapter connector to connect to the SMA cable extender. The GNSS system can use navigational satellites from other networks beyond the GPS. Here the GNSS module is used to record the geo location but also the exact UTC-time. The GPRS is used to send SMS notifications to the users e. g. in case of an error. After inserting a new SIM card into the shield, the user has to activate the modem and register the network via AT-command prompt.
The AMMOD multisampler is capable of performing time-controlled, temperature-controlled and also illuminance-controlled sampling cycles. The user can directly program the trap on-site by using either a manual or a serial connection.
The main Menu provides four different submenus: ‘Manual Program’, ‘Auto Program’, ‘Config Data’ and ‘Info’. By using the built-in 1×5 keypad the user can choose between menus. The initial screen shows the actual date and time that future programs will be based on. If the provided information is incorrect the user can change settings in the ‘Config Data’ menu. Additionally, the user can retrieve additional information from the system by selecting the ‘Info’ menu. This menu contains information about the AMMOD multisampler-ID and readings of implemented sensors: air temperature, air humidity, soil temperature, soil moisture, light intensity.
The ‘Auto Program’ menu provides two built-in standard programs which the user can choose from: (1) ambient illuminance program and (2) soil temperature program. Collection bottles will be turned under the trapping bottle depending on either measured illuminance or soil temperature. Thus, the trap can be used to monitor insect activity patterns depending on abiotic factors (
Selectable sensor-dependent standard settings for automated sampling programs. The user can choose out of four standard sampling programs for which the conditions and associated sampling bottles have been predefined. Depending on program sampling bottle is chosen based on measured illuminance, soil temperature, relative air humidity or relative soil moisture.
Bottle | Illuminance (lux) | soil temperature (°C) | air humidity (%) | soil moisture (%) |
---|---|---|---|---|
1 | 0–200 | < 4 °C | 0.0–5.0 | 0.0–5.0 |
2 | 201–400 | 5.0–10.0 | 5.1–10.0 | 5.1–10.0 |
3 | 401–800 | 10.1–12.0 | 10.1–15.0 | 10.1–15.0 |
4 | 801–2200 | 12.1–14.0 | 15.1–20.0 | 15.1–20.0 |
5 | 2201–5000 | 14.1–16.0 | 20.1–30.0 | 20.1–30.0 |
6 | 5001–7000 | 16.1–18.0 | 30.1–40.0 | 30.1–40.0 |
7 | 7001–10000 | 19.0 | 40.1–50.0 | 40.1–50.0 |
8 | 10001–15000 | 20.0 | 50.1–60.0 | 50.1–60.0 |
9 | 15001–20000 | 21.0 | 60.1–70.0 | 60.1–70.0 |
11 | 20001–25000 | 22.0 | 70.1–80.0 | 70.1–80.0 |
11 | 25001–30000 | 23.0 | 80.1–90.0 | 80.1–90.0 |
12 | 30001–35000 | 24.0 | 90.1–100.0 | 90.1–100.0 |
13 | < 0 and > 35000 | ≥ 25.0 | < 0.0 or > 100.0 | < 0.0 or > 100.0 |
Next to built-in programs users have the possibility to design their own time-controlled program in the ‘Manual Program’ menu. This way, the sampling interval can be easily adapted to the respective research question.
First the user must enter the planned starting time and date. In the next menu the user is prompted to enter the total number of intervals (number of bottle changes) including pauses (position 13). In the last step each interval must be exactly defined in terms of duration lasting from 15 minutes to up to 99 days. After entering data for the last interval, the system will turn position 13 under the trapping bottle before starting with the program at the defined starting time.
Depending on the number of collection intervals, defining the sampling program manually can be very time consuming. With the Graphical User Interface (GUI) ArduinoGUI.exe a desired sampling program can be defined in advance. The GUI is only available for windows machines on which the Arduino IDE is installed. To use all functions of the serial communication, the microcontroller must be connected to the windows machine using the USB ports of the two systems. Like the ‘Info’ and ‘Config Data’ menu of the in-built manual control system, the user can now read information but also change data via serial communication by clicking on ‘Info’ and ‘Config Data’ respectively (
Four-channel audio sensor BRITZ02. The ultrasound sensor of BRITZ03 was placed at the tower in the background.
With either program (manual and automated) the system will turn position 13 under the trapping bottle and will be in stand by until the defined starting time.
As already mentioned, the system is based on a low-power consumption microcontroller which comes with internal pull-up resistors and a voltage regulator, meaning that the system can be run on a 12 V DC battery. Over nighttime and in the absence of sunlight the system relies on power from the rechargeable battery. To allow for continuous running of the system, the AMMOD multisampler is equipped with an integrated automatic battery charger, which is connected to two solar panels each providing up to 30 W under perfect conditions.
The trap presented here has been used in various field experiments with good success. The system enables new study designs e. g. continuous insect activity tracking depending on illuminance, temperature, but also to study circadian rhythms. Thereby the AMMOD sampler can significantly contribute to broadening our understanding of underlying causes for variations in insect activity patterns which in the long term could provide the basis for models calculating the effect of climate change.
This chapter presents a solution for automated long-term acoustic monitoring. The four-channel sensor is based on available commercial components. The configuration allows the recording of acoustic signals in the field. It is suitable for on-site data preprocessing and transmission to a base station for in-depth analysis.
The classifier BirdID-Europe254 is presented. It provides probability values for 254 common European bird species in sound recordings and audio streams. Furthermore, we describe a web interface for further analysis of raw classification results. By setting species-specific thresholds for recording sites, it is possible to obtain information on the presence–absence and acoustic activity patterns of species in large audio datasets. The tool also supports manual validation of classification results by providing full or random samples of audio snippets with confidence values above the selected threshold.
Numerous animal species use acoustic signals to communicate with each other, especially for the delimitation of territories and in the context of reproductive behaviour (
First attempts to use acoustic recordings to document nocturnal bird migration date back to the 1950s (
The use of autonomous sound recorders for continuous acoustic monitoring began in the 1990s. This method is of particular importance for bats. Not surprisingly, in the period 1992–2018, 50 % of all publications on passive acoustic monitoring (PAM) were related to bats (
Four-channel recordings were used for bittern monitoring (
Various attempts have been made to transfer acoustic recordings wirelessly to a remote server, e. g. ARBIMON project (
Algorithms for automated analysis of animal sound recordings have undergone rapid development in recent decades. Encouraging results have spawned an active field of research (
In recent years deep-learning methods based on training artificial neural networks are gaining importance and increasingly replace classical pattern recognition and machine learning methods like logistic regression, support vector machines, hidden Markov models and decision trees; see
Deep-learning methods have several advantages over conventional methods. Single models can be trained in an almost end-to-end fashion where features are learned directly from the data rather than being hand-crafted by experts. They also show a significant performance increase if large numbers of species or classes of audio events need to be identified. A recent review on deep-learning methods for computational bioacoustics is given by
The sensor components have been selected with the following goals in mind:
Since the area to be monitored grows quadratically with the range, we decided to use a high-quality sensor array, even though this solution is more cost-intensive than standard audio recording devices. The main advantage of using a single long-range sensor over a larger number of close-range sensors is that significantly less data needs to be stored for virtually the same information. It also facilitates maintenance and data collection in the field.
To obtain a good spatial resolution, we chose a setup with four cardioid microphones arranged in a plane and each microphone oriented in one of the four cardinal directions. The plate above the microphones protects them from rain (
Four-channel audio sensor BRITZ02. The ultrasound sensor of BRITZ03 was placed at the tower in the background.
AMMOD station | Microphones | Distance between neighbouring microphone membranes | Audio interface | Power supply |
---|---|---|---|---|
BRITZ02 | 4× Sennheiser ME64 with K6 | 350 mm | Behringer UMC 404 HD | Photovoltaic / mains |
BRITZ03 | Dodotronic Ultramic 384 | Not applicable | USB | mains |
BRITZ04 | 4× Kortwich NierePro | 265 mm (adjustable) | Behringer UMC 404 HD | mains |
BRITZ07 | Pettersson M500-384 | Not applicable | USB | mains |
MGB02 | 4× Sennheiser ME64 with K6P | 270 mm (adjustable) | Behringer UMC 404 HD | Photovoltaic |
As a computing plat
form we chose the widely used Raspberry Pi (
Recording Operation
The control software was implemented in Python as system services based on the Raspbian operating system. All settings were made via the ‘config.yaml’ file. An example configuration file can be found in ‘config-default.yaml’. Setup instructions are available in the repository (https://code.naturkundemuseum.berlin/tsa/ammod-acoustic-sensor).
ammod-audio.service
Takes care of audio recording and, if requested, the recording is also analysed. There is the option to adjust the recording regime to the needs of the project, including continuous recording. Likewise, the recording schedule can be adjusted to the time of sunrise and sunset.
ammod-autossh.service
A small service that starts ‘autossh’ to enable the ‘ssh’ connection to the server, providing administrator access to the sensor for maintenance. The server address, port, and user can be changed in the service file. The Readme file in the repository provides details on the setup.
ammod-connection.service
Monitors the stability of the internet connection. If the connection fails, it tries to re-establish it by restarting the network adapter. This process helps to maintain the connection to the base station in case of very weak or strongly fluctuating WLAN signals.
ammod-send-report.service
Sends a daily report to one or more email addresses of your choice. The report contains a list of recently recorded files and the most important system health information. The configuration is done via ‘config.yaml’.
ammod-log.service
Logs system temperature and internet connection status.
ammod-base-station-client.service
Generates the JSON files needed for the AMMOD Cloud, both for telemetry data and audio recording data. The former are provided to the base station for transfer via COAP API. The service is also configured in the ‘config.yaml’ file. Note that the serial number and device ID for the AMMOD Cloud must be set.
On-site inference
If on-site data analysis is desired, a 64-Bit based system is required, such as a Raspberry Pi 4 with at least 4 GB of RAM. Species-specific thresholds can be set in a separate YAML file, see the example configuration in the ‘species_threshold_example.yaml’ file. The classifier results are stored in the result folder of the connected USB storage. The inference results for each recording are stored as a JSON file with the following structure:
The classifiers for bird species identification are based on models described in
Usage
Run ‘inference.py’ to analyse audio files. To set the input path, either select a file or a folder with several audio files. If set to a folder, audio files in subfolders are also analysed. Optionally, an output path to save result files may be set, e. g.:
python inference.py -i /path/to/audio/folder/or/file -o / path/to/output/folder
If no output path is set, result files are saved to the folder of the input path.
Audio analysis can be customised in different ways. In most cases default parameters work across a wide range of application scenarios. Default parameters for inference can be set in the ‘config.py’ file or changed via command line arguments. To see a list of all arguments use the command (
-h, --help | Show help message and exit. |
-d, --debug | Show debug info. |
-i , --inputPath | Path to input audio file or folder. Defaults to ‘example/’. |
-s , --startTime | Start time of audio segment to analyse in seconds. Defaults to ‘0’. |
-e , --endTime | End time of audio segment to analyse in seconds. Defaults to ‘duration of audio file’. |
--mono | Mix all audio channels to mono before inference. |
--channels [...] | Audio channels to process. List of values in [1, 2, ..., #channels]. Defaults to ‘None’ (using all channels). |
-sd , --segmentDuration | Duration of audio segments to analyse in seconds. Value between 1 and 5. Defaults to ‘5’. |
-ov , --overlapInPerc | Overlap of audio segments to analyse in percent. Value between 0 and 80. Defaults to ‘60’. |
-m , --modelSize | Model size. Value in [small, medium, large]. Defaults to ‘medium’. |
-f , --batchSizeFiles | Number of files to preprocess in parallel (if input path is a directory). Defaults to ‘16’. |
-b , --batchSizeInference | Number of segments to process in parallel (by GPU). Defaults to ‘16’. |
-o , --outputDir | Directory for result output file(s). Defaults to ‘directory of input path’. |
--fileOutputFormats [...] | Format of output file(s). List of values in [raw_pkl, raw_excel, raw_csv, labels_excel, labels_csv, labels_audacity, labels_raven]. Defaults to ‘raw_pkl, raw_excel, labels_excel’. |
--minConfidence | Minimum confidence threshold. Value between 0.01 and 0.99. Defaults to ‘0.75’. |
--channelPooling | Pooling method to aggregate predictions from different channels. Value in [max, mean]. Defaults to ‘max’. |
--mergeLabels | Merge overlapping/adjacent species labels. Defaults to ‘NO’. |
--nameType | Type or language for species names. Value in [sci, en, de, ...]. Defaults to ‘de’. |
--useFloat16InPkl | Reduce pkl file size by casting prediction values to float16. |
--outputPrecision | Number of decimal places for values in text output files. Defaults to ‘5’. |
--sortSpecies | Sort order of species columns in raw data files or rows in label files by max value. Defaults to ‘NO’. |
--csvDelimiter | Delimiter used in CSV files. Defaults to ‘;’. |
--speciesPath | Path to custom species metadata file or folder. If a folder is provided, the file needs to be named ‘species.csv’. Defaults to ‘species.csv’. |
--errorLogPath | Path to error log file. Defaults to ‘error-log.txt in outputDir’. |
--terminalOutputFormat | Format of terminal output. Value in [summary, summaryJson, ammodJson, resultDictJson]. Defaults to ‘summary’. |
python inference.py -h
Set start and end time in seconds by passing ‘-s’ and ‘-e’ (or ‘--startTime’ and ‘--endTime’) to select a certain part of the recording for inference, e. g. first 10 seconds:
python inference.py -i example/ -o example/ -s 0.0 -e 10.0
To mix a stereo or multi-channel audio file to mono before analysing it, pass ‘--mono’. Alternatively pass a list of channels, so inference is performed only on the selected channels. For instance, to select only the first and last channel of a 4-channel recording use ‘--channels’:
python inference.py -i example/ -o example/ --channels 1 4
Usually, inference is successively done on 5-second intervals because audio segments of this duration were originally used for training. Optionally segment duration can be set to smaller values (between 1 and 5 seconds). This leads to a higher time resolution of output results and usually to more accurate onset/offset times of detected sound events. Smaller segment durations can also increase identification performance for some species, especially in soundscapes with many different birds calling at the same time. However, in some cases, performance might decrease because certain birds and song types need longer intervals for reliable identification. For example, to set segment duration to 3 seconds (default duration in BirdNET) use argument ‘-sd’ or ‘--segmentDuration’:
python inference.py -i example/ -o example/ -sd 3
Overlap of analysed segments can be set in percent via ‘-ov’ or ‘--overlapInPerc’. For instance, to analyse 5-second segments with a step size of two seconds use an overlap of 60 %:
python inference.py -i example/ -o example/ -ov 60
The repository includes three models of different sizes. All models are trained on the same data but differ in the number of layers and parameters. Larger models usually give better identification results but need more computing resources and time for inference. If run on small devices like Raspberry Pi or in real-time, the small model might be the better or even only option. Results of different models can be assembled in a post-processing step (late fusion) to further improve identification performance. The small model uses an EfficientNet B0, the medium model an EfficientNet B2 and the large model an EfficientNet V2 backbone. To switch model size use ‘-m’ or ‘--modelSize’:
python inference.py -i example/ -o example/ -m small
If input is a folder with several audio files, analysis can be accelerated by preprocessing multiple files in parallel. Use ‘-f’ or ‘--batchSizeFile’ to specify how many files to read and preprocess at the same time. If one or multiple GPUs are used to accelerate inference, the number of batches to be processed in parallel by the GPUs may be passed via ‘-b’ or ‘--batchSizeInference’. The maximum number of batches depends on the selected model size and the available memory (RAM) of the GPUs. Choose the number of CPU threads to prepare the audio segments in parallel for inference with ‘-c’ or ‘--nCpuWorkers’. For single short audio files, small values for ‘batchSizeInference’ and ‘nCpuWorkers’ should be chosen (if only a single file is passed, ‘batchSizeFile’ is set to 1 by default). If files with large durations or a folder with many files are passed, batch sizes and number of CPU workers may be set as high as computing resources allow to increase processing speed.
Analysis results can be customised in various ways. Different output and file formats can be selected. Output files have the same name as the original audio file, but with different extensions and/or file types, depending on output type and format. A list of desired output files can be passed via ‘--fileOutputFormats’. The following formats can be selected:
‘raw_pkl’
Raw results can be saved in a dictionary as a binary (pickle) file for further post-processing in Python. The result dictionary holds information and data accessible via keys, e. g. ‘modelId’, ‘fileId’, ‘filePath’, ‘startTime’, ‘endTime’, ‘channels’, ‘classIds’, ‘classNamesScientific’, ‘classNamesGerman’, ‘startTimes’, ‘endTimes’ and ‘probs’. With ‘startTimes’ and ‘endTimes’ the start and end times for each analysed audio segment can be accessed. Likewise, use ‘probs’ to access a three-dimensional NumPy array that holds prediction probabilities for all channels, audio segments, and species. It has the shape: [number of channels, number of time intervals, number of species]. Raw results can also be saved as Excel and/or CSV files:
‘raw_excel’ / ‘raw_csv’
In Excel files, results for each channel are saved in separate sheets. For CSV, results for each channel are saved in separate files with the channel information added to the filename (e. g. filename_c1.csv for first channel results). Output files consist of a header line and rows for each time interval. Each row has two columns for start and end time of the audio segment and additional 254 columns for each species, listing their prediction probability per time interval. A total of four output formats can be selected for species labels:
‘labels_excel’ / ‘labels_csv’ / ‘labels_audacity’ / ‘labels_raven’
Besides saving raw data, results can also be post-processed and aggregated to allow user-friendly access to the more relevant information on what species was identified at what time within the audio recording. So instead of outputting probabilities for all species and time intervals, labels are created only for those species and time periods where the model’s prediction probability exceeds a minimal confidence threshold. Resulting label files can be saved in the following formats: Excel, CSV, Audacity label track and Raven selection table. The following is an example of saving the results as raw data and aggregated labels in Excel format:
python inference.py -i example/ -o example/ --fileOutputFormats raw_excel labels_excel
The minimum confidence value (prediction probability threshold) necessary to decide if a species was identified can be set by passing a value between 0.01 and 0.99 to ‘--minConfidence’. Classifications below the threshold are not included in the output label file. Higher confidence values lead to better precision, lower values to better recall rates.
For label files, predictions of multi-channel audio files are pooled or aggregated by taking either the mean or maximum value of each channel. The pooling method can be selected by passing ‘mean’ or ‘max’ to ‘--channelPooling’.
Species labels are provided for each time interval analysed. With ‘--mergeLabels’ adjacent or overlapping time intervals with labels of the same species are merged.
How species are named can be customised by passing a name type via ‘--nameType’. Possible types/languages are: Scientific (‘sci’), English (‘en’), German (‘de’), short identifier (‘id’), or index number (‘ix’).
Output in result files can be further customised by passing additional arguments. To save storage space, the size of binary pickel files can be reduced by passing ‘--useFloat16InPkl’ to store 16 Bit instead of 32-Bit floats. For float values in text output files the number of decimal places can be changed via ‘--outputPrecision’. The columns in raw data text files and the label rows in label files within the same time interval can be sorted in descending order regarding species prediction confidence by passing ‘--sortSpecies’. With ‘--csvDelimiter’ select the delimiter used in CSV files.
By default, all 254 species are predicted. The species are listed in the file ‘species.csv’ (including scientific and common names in different languages). A custom species list can be created to filter output results by modifying the original CSV file (
Custom species CSV file. Only the species listed are considered in the classification process.
ix | id | sci | de | en | minConfidence |
---|---|---|---|---|---|
0 | ParMaj0 | Parus major | Kohlmeise | Great Tit | |
1 | FriCoe0 | Fringilla coelebs | Buchfink | Common Chaffinch | |
3 | TurMer0 | Turdus merula | Amsel | Common Blackbird | |
10 | TurPhi0 | Turdus philomelos | Singdrossel | Song Thrush | 0.85 |
The custom species CSV file can also be used to assign individual minimum confidence thresholds to certain species (
Errors during analysis are saved to an error log file. Its path is specified via ‘--errorLogPath’. If no path is assigned, the file is named ‘error-log.txt’ and saved to the output directory.
With ‘--terminalOutputFormat’ the terminal output is controlled. By default, a summary is printed for each input file listing the top three species with the highest confidence scores identified in the entire recording.
On site, audio signals are recorded and optionally analysed. The raw data, sensor information, and preliminary results are sent to the base station. Any updates can be made through a remote maintenance connection to the station. A backup of the raw data is also saved on local storage. Depending on the size of the storage and recording schedule, the data is periodically collected and taken to a data centre for further analysis. The in-depth analysis can lead to new insights and improvements, by fine-tuning species-specific detection thresholds or by using larger versions of the classifier models. The results and metadata are uploaded to the AMMOD Cloud (
Passive acoustic monitoring projects often result in big collections of audio data that are difficult to handle. To address this challenge, a service with two components was developed to make the process more efficient (
The software components of the service can be found in the Git repository (https://code.naturkundemuseum.berlin/tsa/monitoring-data-analyze-service).
The web service consists of two parts. The front end is accessible via a web browser. It displays all background tasks on the landing page (
Clicking on one of the entries in the navigation menu switches to the view of a collection (
Collection view, showing statistics and eligible query parameters for the selected dataset “BRITZ01: 2019”.
This view comprises five components.
In this pane the user can set parameters for different queries (
Requirements are a Linux system with docker and docker-compose installed. The docker-compose file allows easy setup of the service. For detailed instructions, look for the Readme file in the repository (https://code.naturkundemuseum.berlin/tsa/monitoring-data-analyze-service/-/blob/main/README.md).
For better management of monitoring data, it is best to organise them by recording location and year. Creating separate collections for each year keeps the number of entries in the database tables manageable, thus making it easier to search and retrieve data.
To import the data, run an import script. The corresponding settings need to be specified in a config file written in YAML format, cf. the Template in the repository (https://code.naturkundemuseum.berlin/tsa/monitoring-data-analyze-service/-/blob/main/import/config/config-template.yaml). The following settings are available:
Up to this date, 104 bird species in Britz and 54 in Melbgarten have been documented by sound recordings. For several other bird species, no reliably identifiable voucher recordings have been found yet (Britz = 21, Melbgarten = 6 species), mostly due to poor signal quality. However, we anticipate that audio signals from numerous other species could be found in the AMMOD soundscape recordings through systematic screening.
The species lists were complemented by means of classical observation. Systematic breeding bird surveys were carried out in 2021 (Melbgarten) and 2022 (Britz). In addition, unscheduled observations up to a distance of 5000 m from the AMMOD stations were taken into account. For Britz, reliable observations by third parties, some dating back several years, were considered in addition to project-related records. In the near vicinity of Melbgarten, i. e. up to a distance of 500 m from the recording site, observations took place in 2021 and 2022. For the wider area of the Melbgarten, reliable observations obtained from 2010 to 2022 were also considered.
In Britz, 49 bird species were detected by territory mapping of a 16-ha area with 7 visits. Together with the unscheduled observations, a total of 74 bird species were detected in the close vicinity of the recording station. Visual records of two additional species were reported by third parties. Only 11 of the species observed within a 500 m radius have not been found in sound recordings to date. Observations within a radius of up to 5000 m revealed 47 additional bird species. Of these, verifiable voucher recordings could not yet be found for 24 species. In total, 139 bird species could be reliably identified for Britz, 16 of them so far only in the sound recordings, but not by classical audiovisual observations.
In the Melbgarten area, 44 bird species were found by territory mapping of a 25-ha plot during 5 visits. Together with the unscheduled observations in 2021 and 2022, the number of species observed within a 500 m radius increases to 46, of which 10 species have not yet been found in sound recordings. Within a 5000 m radius, at least 68 additional bird species have been detected in the last 13 years, 53 of which have not yet been found in the sound recordings. Most of the additional bird species found through classical observation were migrants, roaming non-breeders, and species with territories outside the range of the AMMOD recording site. In total, 117 bird species could be reliably detected in the Melbgarten and its wider vicinity by the end of 2022, three of them so far only in the sound recordings but not by classical audiovisual observations.
Natural soundscapes are often complex. Four-channel recordings with high-quality directional microphones make it easier for the listener to distinguish between sound sources, since the signal amplitude is highest at the microphone pointing in the direction from which the sound originated (
Spectrogram of a four-channel recording. BRITZ02, 13 May 2022, 08:02 h; spectrogram settings – downsampling to 24 kHz, high-pass filter up to 950 Hz, DFT size 512 samples, grid spacing 46.9 Hz, Window Hann, overlap 50 %.
Oscillogram of the recording presented in Figure 13. Since each microphone faces in another cardinal direction, the amplitude of audio signals differs considerably between channels.
The difference in signal amplitude between channels can be exploited for computing directional spectrograms (D-SPEC) in which colour represents direction of arrival (DOA) (
D-SPEC of the recording shown in Figures 13 and 14 before (top) and after spatial processing (bottom).
Source separation into directional clusters, exemplarily shown for the recording presented in Figures 13 through 15 (
Sound sources associated with the directional clusters shown in Figure 16. Some clusters contain audio signals of more than one species because they roughly sang in the same direction.
Cluster | Species / sound source | Angle | Weight |
---|---|---|---|
1 | Tree Pipit Anthus trivialis, song | 235.5 | 33 % |
2 | Common Chaffinch Fringilla coelebs (individual 1), song; Common Redstart Phoenicurus phoenicurus (individuals 1 and 3), song | 35.8 | 31 % |
3 | Common Redstart Phoenicurus phoenicurus (individual 2), song; European Pied Flycatcher Ficedula hypoleuca, song | 98.2 | 21 % |
4 | Common Chaffinch Fringilla coelebs (individual 2), song | 311.7 | 11 % |
5 | Clicking noise caused by falling twigs, water drops, leaves, or seeds | 195.7 | 1 % |
The ultrasound sensor of AMMOD station BRITZ03 was deployed in a small forest glade. It was operated only between 14 and 26 July 2021, since hail destroyed the protective membrane of the microphone, which led to a device failure. A total of 686 minutes were recorded, of which 293 minutes contained bat activity found by manual screening (
Spectrogram of an ultrasonic recording made with the microphone Dodotronic Ultramic 384. Besides echolocation and social calls of Leisler’s Bat Nyctalus leisleri there is a frequency band around 8 kHz, representing stridulation sounds of Great Green Bush-cricket Tettigonia viridissima. BRITZ03, 16 July 2021, 21:50 h; spectrogram settings – downsampling to 192 kHz, FFT size 1024, Frame size 50 %, Window Hann, overlap 87.5 %.
Another ultrasound device, BRITZ07, was deployed inside the forest, starting on 1 September 2022. In total 610 minutes of ultrasound recordings were revised manually, of which only 54 minutes contained bat activity. In general, the signals were weaker than those recorded at station BRITZ03 and more affected by reverberation.
Manually annotated recordings were used for the performance assessment of the classification model BirdID-Europe254. In total 19,844 audio signals were labelled in 99 soundscape recordings, most of which were collected at the AMMOD study sites Britz and Melbgarten. Since it was not possible to reliably identify all the signals to species level, in the end 14,683 annotated signals were available for performance tests (
Number of annotated signals per animal class and identification level (ID1 = certain, ID2 = probably correct, ID3 = uncertain). Only those signals that could be safely identified to species level (ID1) were used for validation. The training dataset was explicitly not used for building the BirdID-Europe254 classification model.
Class | Training signals | Validation signals | ||||||
---|---|---|---|---|---|---|---|---|
Species | ID1 | ID2 | ID3 | Species | ID1 | ID2 | ID3 | |
Amphibians | 2 | 13 | 1 | 7 | 1 | 0 | 10 | 13 |
Birds | 85 | 7128 | 1143 | 902 | 72 | 6960 | 1201 | 1316 |
Insects | 4 | 385 | 192 | 266 | 2 | 25 | 6 | 30 |
Mammals | 5 | 80 | 22 | 28 | 2 | 92 | 11 | 13 |
Total | 96 | 7606 | 1358 | 1203 | 77 | 7077 | 1228 | 1372 |
All visible and audible signals were labelled with bounding boxes in the spectrograms, even very faint sounds (
A fully annotated stereo recording used for performance assessments. The bounding box #638, highlighted in red in the lower spectrogram, does not refer to a labelled audio signal, but to the segment shown in Figure 20. BRITZ01, 19 May 2019, 08:45 h.
Detail of the spectrogram area highlighted red in Figure 19. Two 5-second segments are shown to illustrate the length of BirdID-Europe254 classification windows in relation to common bird songs and calls. Since in practice the analysis is computed with a step size of two seconds, we have omitted the respective overlapping classification windows for the sake of simplicity. Spectrogram settings – sample rate 48 kHz, DFT size 1024 samples, grid spacing 46.9 Hz, Window Hann, overlap 50 %. The spectrogram was truncated above 14 kHz because only noise was present in the high-frequency range.
Robust monitoring of bird vocalisations requires reliable species records, making high precision of classification results crucial. However, there always is a trade-off between precision and recall rate (
Precision-recall curves for 24 representative species based on AMMOD validation data, computed for the classification model BirdID-Europe254. The average precision (AP) stated in the legend facilitates the correct assignment of similarly coloured curves to the correct species.
The raw classification results are the basis for all further analysis with respect to robust acoustic monitoring. In the first step, we made example plots for selected species, showing the distribution of classifications per day and confidence level (
Raw classifications for Common Redstart Phoenicurus phoenicurus at BRITZ01 in April 2019 plotted by date and confidence level. The onset of vocal activity of the local breeding population on 15 April 2019 is clearly visible.
In the next step, we plotted the bin-size distribution for selected species, i. e. the number of classifications over bins of confidence values, exemplarily shown here for Common Redstart (
(A–D) Bin-size distribution of raw classifications for Common Redstart Phoenicurus phoenicurus at Britz (BRITZ01) (A, B) and Melbgarten (MGB01) (C, D) in spring 2022 (1 March to 30 June). X-axis: confidence level with a resolution of 0.02 (2 % steps), y-axis: number of classifications. Top (A, C): Bin-size distribution when the entire confidence interval is considered (≤ 0.02 to 1.0). Total number of classifications ‒ BRITZ01, n = 1,639,352; MGB01, n = 1,671,813. Bottom (B, D): Zoom-in on bins with confidence levels ≥ 0.2.
Passive acoustic monitoring (PAM) is a complex task, and thus accompanied by formidable challenges. With the aim to provide some information on how to address these challenges, we decided to focus our discussion on the lessons learned during the implementation of AMMOD.
Team composition is crucial for the success of projects aiming at the development of hardware and software tools for PAM. Since acoustic monitoring of biodiversity concerns diverse aspects of life sciences, engineering, and logistics, multidisciplinary teams need to closely cooperate for data acquisition, device maintenance, software development, database management, performance assessment of classification models as well as analysis, interpretation, and validation of classification results. For instance, only due to the permanent interchange of ideas among computer engineers and bioacousticians it was possible to develop the web interface for efficient analysis of raw classification results presented here.
Custom-built hardware has some advantages over standard solutions. For example, it provides flexibility regarding the usage of high-quality directional microphones and wireless data transmission. Additional functionality is easily implemented, such as remote-control options and on-site data processing. However, long-term PAM studies must also consider other important aspects such as power supply, protection against vandalism, and the cost of purchasing and maintaining equipment and software. Hence, there are also many arguments in favour of using commercial audio recorders. If standard equipment is chosen, we strongly recommend energy-efficient stereo recorders with weatherproof electret microphones, for two reasons: Compared to mono recordings, stereo recordings provide some spatial information that facilitates data processing in many ways, whether for manual or computer-aided analysis. Equally important, electret microphones produce less self-noise and deliver much better audio quality than MEMS microphones, increasing the range monitored by each sensor unit.
In bioacoustic research, microphone arrays have been used to localise and track animals in the field (
A critical bottleneck for the development of robust pattern recognition models has been the availability of training and validation data (
We emphasise that even if fully annotated validation files are available, this does not guarantee objective performance evaluation of classification models. The output may be biassed by the effect of long-tailed distribution of audio signals in soundscapes (
In multilabel classification, each classification window is assigned a confidence value for the presence of each species included in the model. This means that for a given number of audio recordings, the number of classifications is the same for all species. Because classification windows in which a particular species is highly unlikely to occur receive much lower confidence values than those in which the species is likely to occur, the bin-size distribution typically takes the form of an L lying on its long side (
This work was sponsored by the German Federal Ministry of Education and Research (Grants 01LC1903F).
This software description is based on the technical paper that provides motivation and a complete technical and mathematical description of the processing (
The purpose of the D-SPEC algorithm is to create a spectrogram with a color dimension that approximates the direction of arrival of the signal. It operates by frequency-domain beam-forming. A beam amplitude response is calculated at each and every time/frequency bin, and then this beam amplitude response is converted to a 3-dimensional color.
The following are the key parameters that affect the calculation of D-DSPEC.
Parameter description: Variable name:
Number of microphones. M, nmic
FFT size (window size in samples). N
processing window shift (samples). shft = N/3 (i. e. 2/3 overlap)
angular resolution (for beam-forming). ang_res (degrees)
speed of sound. c (m/s)
sampling rate. fs (samples per second)
microphone separation. d (m)
break frequency for high-pass filter. fmin (Hz)
start and end times for processing. (tstart, tend) (seconds from start of recording)
number of angles (beams). na
In the setup phase, the delays are calculated based on the array geometry. These time delays are then used to compute the steering vectors. We will use the case of 4 microphones to illustrate the ideas in
pos =(-d/2, d/2, d/2, d/2, d/2, -d/2, -d/2, -d/2)
pos = np.asarray(pos).reshape((nmic,2))
am = np.asarray ((-pi/4, pi/4, 3*pi/4, -3*pi/4))
calculates the positions of microphones 1 through 4, as illustrated in
The vector of angles in radians, for which a pre-formed beam will be computed is created. There are na angles and there will be na pre-formed beams in the beam response:
# beamformer beam angles
angs=np.asarray(range(0,360,ang_res))*(pi/180)
maxang=360
na=len(angs)
Next, one imagines that sources located a distance r from the array center (indicated as microphone 0 in the Figure) with angle angs[1...na] (See
This is implemented by the code:
#
# dst[i,ia] is distance from microphone i to a distant point at
# angle ia minus dctr, where dctr is distance to center of array
dst=np.zeros((nmic,na))
r=1000 # target at this range
for ia in range(na):
ry=np.cos(angs[ia])*r
rx=np.sin(angs[ia])*r
dctr = np.sqrt(np.square(rx) + np.square(ry))
for i in range(nmic):
dst[i,ia] = np.sqrt(np.square(rx-pos[i,0]) + \
np.square(ry-pos[i,1])) – dctr
The steering vectors are 4-dimensional (4 microphones) vectors that describe (the conjugate of) the complex value of a frequency-domain signal that would appear at the array from a source at a distant location with specified angle, as illustrated in
# set up steering vectors
n=N//2+1
sfrqs=np.asarray (range(n))/N*fs
s=[None]*nmic
for i in range(nmic):
# ft = (f)requency times delay (t)time
ft = np.outer(sfrqs, dst[i,:]/c)
if nmic==4:
# microphone cardioid pattern
mr=(1+np.cos(angs-am[i]))/2
#mr=np.exp (-np.power((angs-am[i])/(np.pi),4))
else:
# omnidirec
mr=np.ones((na,))
# steering vector
s[i] = np.exp(1j*2*pi*ft) * mr
High-pass filtering is needed to remove low-frequency noise and is accomplished in the frequency-domain. The coefficients of a Butterworth filter are computed and the frequency response is calculated. This is done by the code:
# high-pass filtering
Bf,Af = scipy.signal.butter(2,cfg.fmin/fs*2,btype=’high’)
w,h=scipy.signal.freqz(Bf, Af, worN=N, whole=True, plot=None)
frsp = np.square(np.abs(h))
A color map is initialized. This will be later used to convert beam response to RGB color. It assigns a 3-dimensional RGB color to each of the na beam angles.
# create hsv color map of size equal to number of angles
mp = matplotlib.cm.get_cmap(‘hsv’)
co=np.zeros((3,na))
for ia in range(na):
ct = mp(1.0*ia/na)
co[:,ia] = ct[0:3]
This code creates a color map array that corresponds to the angle array angs described above, which is of length na and for ang_res=3 runs from 0 to 357 degrees (but has values in radians). For reference, it can be deduced from the code that Red (color 1,0,0) corresponds to 0 degrees, green (color 0,1,0) corresponds to 120 degrees (2*pi/3 radians), and Blue (color 0,0,1) corresponds to 240 degrees (4*pi/3 radians).
Data is read in for a specified time range (tstart,tend) from the .wav files, which contains 4-channel data. The function read_wave.read_wav() is reads just one channel (microphone) at a time from the specified time range. This time-series is passed to the spectrogram function that calculates a spectrogram with FFT size N and overlap 2/3. The FFT size N must be divisible by 3. High-pass filtering is done by multiplying by the filter frequency response, which is stored in the variable frsp. Some scaling is done so that the Python code has the same values as the MATLAB/OCTAVE code. The spectrogram data for channel ch is stored in the list variable B[ch].
The code to do beamforming and calculating the D-Spec is provided in the function ammod_utils.get_cspect(). Memory for the D-Spec output is allocated using the variable I, which has shape (n,NSEG,3), where n is the number of FFT bins and NSEG is the total number of spectrogram time segments (called nseg elsewhere). The third dimension is the RGB color space.
The code loops over the NSEG segments, processing one segment at a time. For a given time step „ismp“, the following steps are made:
The beams are formed for each of the angles angs[i] by taking the inner product of the spectrogram output with the steering vectors, and summing over the microphone channels:
for i in range(M):
bt = B[i][:,ismp]
xbf = xbf + bt * s[i].T
To suppress side-lobes, the magnitude of the resulting complex beam outputs (xbf) is computed and then this is raised to the power 16 by repeatedly squaring. The reasons for this are explained in the paper referred to above (cf. section 7.1), section III-B. The result is stored in the variable b. Then, this is normalized so that it sums to 1 over the na beams and stored in variable bn.
There are two approaches, selectable by the operator, to convert the normalized beam response bn into a RGB color, and the method can be chosen using the variable cstyp, which is settable on the GUI. The „MAX“ approach just finds the location of the maximum response and assigns the color corresponding to that direction (na possible directions). The normal approach (default and recommended) is to form the dot product of bn with the color map, which is described above, resulting in an RGB color. After computing the dot-product, the result is multiplied by an estimate of the combined spectrogram amplitide, stored in variable bs. This is needed because the color vector after inner product with the color map is normalized and has no amplitude information. The final result is stored in the D-SPEC output vector I. This can be directly plotted as a D-SPEC image using MATLAB/OCTAVE or Python.
The purpose of the clustering algorithm is to assign each of the n*nseg D-SPEC bins to a cluster, with the hope that each cluster should represent a bird individual. This is helped by the tendency of birds, which are territorial in nature, to sing from different directions. The clustering algorithm is implemented by the Python function ammod_util.cluster_C(). In the following sections, the code is explained in the order in which it appears in the function. The input variable C corresponds to the D-Spec image variable I in Section 7.2.4 and has dimension (n,nseg,3), where nseg is the same as NSEG.
Relevant parameters are listed as follows, along with default values.
Parameter: Default value: Description:
expfac # .5 # (make larger to give more weight to higher signal strength)
lfac # 1.0 # (make larger to give more weight to color (direction))
minwt # 1e-3 # mixture weight assigned to noise
radius # 6.0 # search radius for neighborhoods (pixels)
dfac # 30.0 # Make larger to give more weight to spatial effects
P # 2 # GMM mixture components per cluster
nit_gmm # 90 # number of GMM iterations to execute
nit_spatial # 10 # number of spatial iterations to execute
minvar # .00015 # Minimum GMM variance. Increase for fewer clusters
merge_thresh # -.2 # Threshold to merge clusters. Decrease for fewer clusters
nhbd_type # 2. # 1 = r^2/d^2 2 = exp (- d^2/r^2)
The following steps are required prior to clustering iterations.
The color image C is converted to an amplitude image Cs of dimension (n,nseg) by taking the root-mean-square of C over the color dimension.
Location indexes are integers that determine the time and frequency locations within the D-SPEC to make spatial processing more efficient. Once the D-SPEC is thresholded, and the bins that passed the threshold are gathered into a single vector, the index of these bins are no longer organized as a 2-dimensional array. This makes it difficult to determine the relative positions betweeen two D-SPEC bins. To make this more efficient, we need to create the location indexes, and then when thresholding, select the subset of indexes for bins that exceed the threshold. .The D-SPEC matrix C has dimension (n,nseg,3) and the D-SPEC amplitide matrix Cs has dimension (n,nseg). If we combine the first two dimensions, we can view these matrices as dimension (n*nseg,3) and (n*nseg). The vectors ipos and jpos are also of dimension n*nseg and are equal to the time and frequency bin indexes within the spectrogram. Having these index vectors makes it much more efficient to compute, for example, the euclidean distance (in bins) between two spectrogram bins. For example, let k, l be two arbitrary locations within the thresholded D-SPEC. Then, the eucliden distance between k and l is simply
d=sqrt ((ipos[k]-ipos[l])^2+(jpos[k]-jpos[l])^2)
the index vector ijpos points back to the location within the original array. So, ijpos[k]=k, etc. It is necessary to keep this pointer into the original arrays, once the D-SPEC gets thresholded.
To determine the hue values, the input D-spec array C is normalized to produce array which has a color dimension that adds to 1. This is then converted to a ‘feature’ using function rgb2feat(), but all this does is extract the hue value, which is the first dimension of the HSV color representaion. The color ‘feature’ dimension, dim, is then equal to 1. The result is hue array Cn, which has dimension (n,nseg,dim), where dim=1. The correspondence of hue to RGB color and direction is
hue: RGB: Direction(deg): Direction(rad):
0 (1,0,0) 0 0
0.333 (0,1,0) 120 2*pi/3
0.667 (0,0,1) 240 4*pi/3
It would not be practical to do clustering using all the n*nseg bins of the D-SPEC. For this reason, it is necessary to cluster only the bins with higher amplitude. The user can select two thresholding methods defined by the variable rel_thr. If rel_thr is true, a relative threshold is found by sorting the amplitude values in matrix Cs, and choosing the threshold that passes the desired fraction of the bins. The user sets the desired fraction using the threshold parameter thr. For example, if thr=0.1, then 10 percent of the data will be passed. To make it more efficient, the threshold is found using just 1/50 of the bins. If rel_thr is true, the manually entered threshold thr is used directly as an ampltude threshold. Once the bins that pass the threshold are located, all the important arrays such as C, Cs, ipos, jpos, ijpos, are also compressed.
In clustering the D-SPEC bins, each bin is assigned a weight (higher amplitude bins are considered more important). The variable dwts is the weight value which is calculated as the amplitude Cs raised to the power expfac. Normally, a power less than one is used. Default is expfac=0.5.
The D-SPEC bins that exceeded a threshold are clustered using the k-means clustering method. Clustering is based only on the hue value.
Each cluster is assumed to have a separate Gaussian mixture distribution with P components (by default, P = 2); refer to the paper mentioned above (cf. section 7.1), section IV. Assuming there are M clusters, and P components to each cluster, the GMM variables are:
Variable Dimension Description
cmean (dim,P,M) means of the GMM component
cvar (dim*dim,P,M) covariance of the GMM component. Since dim=1, this is just a variance.
gwts (M,P) GMM weights that add to 1 in the dimension P
The mean and variance of the data assigned to each cluster by the k-means algorithm is calculated. Then, to initialize the P-component GMM for each cluster, some equally-spaced locations are found within the cluster and these are used as the component means.
A neighborhood of a given D-SPEC bin is the collection of other bins that are near a given bin. Within a neighborhood, the ‘other’ bins are weighted according to how far away they are. By default, we use an exponential weight w = exp (- d^2/r^2) where r is the radius and d is the euclidean distance in pixels. To make spatial processing more efficient, the neighborhoods of each D-SPEC bin (that passed threshold) is computed in advance and the weights are computed. Suppose there are nc D-SPEC bins that exceeded the threshold. The neighborhood indexes and weights for D-SPEC bin i are:
D_i[i] a list of D-DPEC bin indexes (within the nc bins) that are in the neighborhood.
D_w[i] weights for the above.
The clustering algorithm has two phases. The first phase is essentially a Gaussian mixture model estimation algorithm, which is a type of E-M algorithm. The GMM is applied to 1-dimensional features (just hue), so is unable to apply spatial information. Without spatial information, the pixels assigned to a cluster can be distributed randomly across the spectrogram. A second spatial processing phase is applied once the GMM has converged. In the spatial phase, we take a look at the pixels around a given pixel. The cluster that these neigboring pixels are assigned to will then influence the assignment of the given pixel.
The GMM estimation algorithm is explained in the paper (cf. section 7.1), section IV-B-D. To help understand the code, we provide the correspondence between the mathematical notation in the paper, and some of the variables in the code. Vectors and matrices are shown with indexes that are the size of each dimension.
Ns nc number of D-SPEC bins exceeding threshhold
M M number of clusters
C P number of GMM components per cluster
Wi,j Ns × M Wc(nc,M) Equation (2)
Li,j,c Ns × M × C Lcp(nc,P,M) Equation (1)
αj,c M × C gwts(M,P) Equation (3)
P(j) wts Equation (4)
δµj,c M × C dmu Equation (5)
γ dfac
η(i, i′) D_w
Ps(j|i) W(nc,M+1) Equation (8)
The clustering algorithm is initialized with a larger number of clusters than required, then un-needed clusters are removed. Two clusters are merged into one if they get too close to one another. Closeness is measured by the function cluster_dist(). Only one merging can occur per iteration.
To use a recording for the D-SPEC software, it needs to be 4-channel in .wav format. Optimal sampling rate is 24 Khz. If the file is in 48 kHz sampling rate, a utility program is provided to convert it to 24 kHz. Using MATLAB or OCTAVE, invoke the program dsamp2.m:
>> dsamp2(‘file1.wav’,’file1_ds2.wav’)
This will down-sample the file file1.wav and save it as file1_ds2.wav. To enter the file into the file list, open the file file_list.txt in an editor, we enter a text line with 7 fields. Using the example of file1_ds2.wav), the line looks like
File46 24000 0.27 343 4 “file1_ds2.wav” “None”
The first entry is a ‘nickname’, a short string to identify the file. The convention is to use FileXX, and make XX a number that has not already been assigned. The second entry is the file sampling rate. The third entry is the array microphone separation in meters. The fourth is the speed of sound in m/s when the recording was made. The fifth entry is the number of microphones. The sixth entry is the file name in quotes. The last entry is the name of an annotation file in quotes, or ‘None’ if there is none.
The D-SPEC software is launched simply by invoking python on the program ammod_gui.py. In Linux:
$ python ammod_gui.py
Launching the software from Windows might differ slightly.
The graphical interface comes up with default parameter settings. These can be changed by the operator, and saved in a parameter file by specifying the file name and pressing SAVE on the top of the window, or LOAD to retrieve the saved parameters.
To load a data set, use the File: drop-down manu to select the file. On the left of the drop-down menu appears the nick-name that was entered in the file list. Once this has been selected, the information stored in the file list (microphone separation d, speed of sound c, sample rate fs, number of microphones nmic) is transferred to the parameters displayed in the GUI.
Before computing the D-SPEC, the operator can specify a desired time range (T1, T2). The total time processed at a time can be up to 30 seconds, although the time to process and cluster increases with time range. The time window can be moved forward or backwards by 2 seconds using the (+2, -2) keys.
Once the time range has been set, press the CSPEC button to compute the D-SPEC. The D-SPEC will not be displayed until the requested by the PLT button. The brightness (brt) and contrast (contr) can be changed to optimize the display, then only the PLT button needs to be pressed. Note that these values will be saved or loaded when using the SAVE/LOAD buttons. Note: never ‘kill’ the D-SPEC graphics window. This will cause errors.
Many of the parameters that affect D-SPEC are pre-determined by the array and recording, but the following can be changed: N, ang_res, fmin, T1, T2 Some details of these are listed in Section 7.2.1. The FFT size N can be experimentally changed. Generally, for a 24000 Hz sample rate, 384 is a good choice, and this would change proportional to sample rate. The highpass-filter cuttoff ‘fmin’ should be set experimentally for each recording to eliminate low-frequency noise.
Once the D-SPEC has been computed, the clustering can be started by pressing ‘Cluster’. Information is displayed in the system console. Once the clustering operation ahs stopped, the cluster-specific spectrograms can be displayed using the ‘PLOT Clusters’ button.
Operator-settable parameters include the following:
thr, rel_thr Threshold. When the rel_thr box is checked, then thr defines the fraction of D-SPEC pixels that are going to exceed the threshold. Default is 0.1 for keeping 10 % of the pixels. When the rel_thr box is not checked, then thr defines the actual amplitude threshold. A good starting value is 0.3.
expfac. Exponential factor. Default value is 0.5. There is little need to change this.
nclus. Default value of 9. This is the starting number of clusters. There is little need to change this.
P. Number of GMM components per cluster. P=2 is default and there is little need to change this.
lfac. This is the factor multiplying the GMM log-likelihood. This is experimental, and there is no need to change this.
minwt. Minimum weight, with a default value of 0.001. It is used for the ‘dummy’ cluster, a cluster that is just there to gather un-assigned pixels. This is experimental, and there is no need to change this.
radius. This defines the neighborhood radius for spatial processing. Larger neighborhood should promote smoother-looking results. Default is 6 pixels. There is no need to change this.
dfac. This determines the amount of spatial influence to the clustering. Default is 30. Higher will result in clustering that will produce more local consistency.
nit_gmm, nit_spatial. Number of GMM and spatial iterations. Default is 40 and 5. Making nit_gmm smaller makes tthe clustering run faster, but if time is not an issue, use 100 iterations. nit_spatial should be changed with care. Too many spatial iterations might result in poor results.
wts_exp. Default is 0.01. Making this higher will promote faster convergence. This is experimental, and there is no need to change this.
minvar. This parameters is important in determining the final number of clusters. Default is .00015. If two signals are coming from nearby angles, they will produce nearby clusters in the hue space. Minvar defines the minimum GMM variance. By making this large, the GMM cannot resolve two closely lying clusters in the hue space, so is likely to group them into one cluster. However, making this too small can result in a single signal being split into multiple clusters.
merge_thr. This is the threshold for cluster merging, a negative number. If the log-distance between clusters is above this, they will get merged. To promote fewer clusters, make this lower (more negative).
nplot. Defines the number of sub-plots when plotting clusters, so is effectively the largest number of clusters to display.
To generate cluster-specific time-series, press the ‘Gen Timeseries’ button. This causes the files ‘cluster1.wav’, ‘cluster2.wav’, etc. to be written. These files can be used for external purposes, such as classifying. Or, they can be played by selecting the desired cluster ‘CLUS’ and pressing ‘Play’.
An experimental version of a neural network implementation of D-SPEC (not including clustering) has been developed. This explains, for purposes of documentation, how it works and points to the required software.
NN-DSPEC processing is the same as D-SPEC up to the formation of beams. Instead of forming 120 beams in 3-degree increments, NN-DSPEC operates by creating 32 pre-formed beams in 11.25 degree increments, but is otherwise the same as laid out in Section 2.4. The amplitude of these beams is computed, resulting in a beam output with dimension (nseg,n,32). This is a spectrogram with 32-dimensional bins. This 32-dimensional feature is passed to a neural network to convert the 32-dimensional bind to 3-dimensional (RGB) bins. From the perspective of the neural network, it can be seen as just a 32-dimensional feature with nseg*n samples. The neural network itself is very simple. There are 5 network layers, each has 32 neurons except for the last that has 3 neurons. Each layer has the soft-plus activation function. The network is simply trained to re-create the ground-truth which is an artificial RGB D-SPEC.
Since the implementation is spread across mutiple experimental programs, we cannot give a complete description here. Instead, we point to the required programs and their purpose, so that the process can be repeated in the future.
The program ‘collect_calls.py’ is a multi-purpose Python program. By running it with the task ‘circle’, it reads in a specified 4-channel recording (the same files that the GUI reads) for a specified time range. It then calculates the spectrogram on all microphones, then sums the mic-specific spectrograms to get a ‘sum beam’. The ‘sum beam’ is then converted into an artificial set of 4-channel spectrograms with the signal arriving from a specified direction. From the artificial 4-channel recording, the 32 beams are extracted and saved to a file along with the ground truth, which is the desired RGB color values that correspond to the requested direction (an artificial D-SPEC). This is repeated for 120 directions in 3-degree increments. There are in the end 120 files, each for a different direction. This forms a data set called ‘circle32’ which is made available to the neural network toolkit (PBN Toolkit). As mentioned, any arbitrary data set can be used to produce the ‘circle32’ training data. But, it is a good idea to use a data set with a bird call having a loud wide-band vocalization, so that all frequency bins are excited. We used the file ‘BRITZ02_20210331_121000_d2.wav’ in the time range 1–4 seconds.
The PBN Toolkit (http://class-specific.com/pbntk) is used to train the neural network using the network model ‘cspec32’. The data set ‘circle32’, which is in streaming format, is read in by the toolkit, and the network is trained to re-produce the ground-truth RGB color. Once this is trained, the parameters of the network are saved to files ‘cspec32_circle32_lyrX.mat’, where ‘X’ runs from 1 to 5.
The program ‘collect_calls.py’ is used to complete the process. Using the ‘bfm’ function, 32-beam responses are created for arbitrary data sets, and the result is converted to D-SPEC using the trained network parameters.
The visual monitoring of animals using automatically triggered cameras (so-called camera traps) has a long history, even dating back to the late nineteenth century (
In contrast to prior works, we designed a novel StereO CameRA Trap for monitoring of biodivErSity (SOCRATES), which is optimized for:
In the following sections, we first describe our final implementation and how we addressed the requirements above. We then discuss the challenges and limitations of our approach.
We first address the stereo camera design (cameras and Baseline, design goals 1 and 3). The raw data produced by the cameras is processed and stored by the control unit (design goals 2 and 3). Weather-resistance (design goal 4) is provided by the case. Infrared motion Detection and illumination facilitate energy efficiency (design goal 2) and operability at night time (design goal 1a). We additionally describe in detail the power supply, how we obtain animal-camera distances using stereo correspondence and how the captured data may be transferred using different connectivity options.
Part | Article Name | Article Number | Supplier | Unit Price | Count | Total Price |
---|---|---|---|---|---|---|
Cameras | Raspberry Pi High Quality Camera | 633696492738 | Sertronics GmbH | 59.90 € | 1 | 108.90 € |
Lenses | 6 mm Wide Angle CS-Mount Lens | Sertronics GmbH | 28.90 € | 2 | 57.80 € | |
SOC | NVIDIA Jetson Nano Developer Kit B01 | 812674024356 | Sertronics GmbH | 108.90 € | 1 | 108.90 € |
Camera Cables | Flex Cable for Raspberry Pi Camera | 4251266700159 | Sertronics GmbH | 2.90 € | 2 | 5.80 € |
PIR Sensor | HC-SR501 | 4251266700715 | Sertronics GmbH | 2.15 € | 1 | 2.15 € |
Storage | SanDisk Extreme microSDXC UHS-I 128GB | 0619659188467 | Sertronics GmbH | 16.15 € | 1 | 16.15 € |
Batteries | SLSXT160004120 | RC Multistore | 170.99 € | 2 | 341.98 € | |
BMS | Naltronic | 9.90 € | 1 | 9.90 € | ||
MOSFET | IRLZ44NPBF | Sertronics GmbH | 1.29 € | 1 | 1.29 € | |
Battery Bags | Extron LiPo-Safety-Bag | X6670 | Conrad Electronic SE | 8.49 € | 2 | 16.98 € |
Battery Charger | Voltcraft V-Charge Eco LiPo 4000 | 1406443 | Conrad Electronic SE | 29.99 € | 1 | 29.99 € |
5V Regulator | Mean Well SCW20A-05 | SCW20A-05 | Conrad Electronic SE | 22.99 € | 1 | 22.99 € |
12V Regulator | Mean Well SCW12A-12 | SCW12A-12 | Conrad Electronic SE | 17.99 € | 1 | 17.99 € |
740.82 € |
Cameras and Baseline: A pair of Raspberry Pi High Quality Cameras were chosen for their cost-effectiveness and the high sensitivity of their Sony IMX477 sensor (Sony Semiconductor Solutions Corporation, no date). Interchangeable lenses allow adaptation to specific scenarios (i. e. shorter focal lengths for close-up scenes, higher focal lengths for more distant objects). Removal of the infrared filter allows sufficient exposure at night using artificial infrared illumination. The cameras have an additional Bayer filter above the sensor, which is usually responsible for filtering different wavelengths to create a color image. We leave this filter intact to not risk damaging the sensor itself. As near-infrared illumination (either from the environment or the illuminator) illuminates all color bands, we do not try to recover any color information and instead average all bands to obtain a grayscale image. The cameras are mounted on a long U-shaped aluminum rail with holes drilled at regular intervals to allow configuration of different baseline distances between both cameras. Both cameras are connected through long ribbon cables to the two MIPI CSI-2 interfaces of an NVIDIA Jetson Nano Developer Kit. Both design aspects, i. e., the interchangeable high quality lenses as well as the configurable baseline construction, allow for adaptation to specific scenarios, namely free fields, feeding places, animal crosses, green bridges, etc., where animals are observable at different distances.
Control: We use an NVIDIA Jetson Nano Developer Kit as the central control and storage unit. It is responsible for taking motion detection signals from the PIR sensor, turning on the power to the IR illuminator, capturing, encoding, and archiving image material from the cameras. We decided on the Jetson Nano for the following reasons: (1) compared to most single-board computers, it provides two MIPI CSI-2 interfaces for the two cameras, (2) it provides a powerful GPU that can be used for encoding video efficiently, and (3) it supports a power-efficient hibernation mode. The Jetson Nano uses a 128GB microSDXC card for persistent storage.
Case: To make SOCRATES as weather-resistant as possible, most components are placed inside a single weather-proof case. The case is made of 0.8 cm thick birch plywood and is 80 cm wide, 11.6 cm high and 20 cm deep. We decided for a very wide case to be able to adapt the baseline of the stereo camera to different configurations. The front of the case is shielded by a piece of acrylic glass. In the bottom, we add a 4 cm wide, circular hole for ventilation, which is covered from the inside with an insect screen. The battery is mounted via Velcro strip onto a hatch in the bottom of the case, to allow quick replacement. We add two further holes for the wiring of the IR illuminator and motion detector, respectively, both of which are sealed using silicone. The top of the case is sealed using a silicone strip and secured by screws, which can be loosened to take it off for maintenance. All exposed wooden parts are further treated with marine varnish for weather resistance.
Motion Detection: Like most camera traps, we use a pyroelectric infrared (PIR) sensor for detecting motion and thereby triggering capture. We choose an HC-SR501 PIR sensor due to its compatibility with the 3.3 V GPIO pins of the Jetson Nano. We initially mounted the PIR sensor inside the case, just behind the acrylic glass. However, we found that this severely impaired the ability of the sensor to detect any kind of motion outside the case. This is because acrylic glass is opaque around wavelengths of 10 μm (
Illumination: We employ a simple 12 W, 850 nm infrared illuminator to ensure properly exposed images at night without disturbing most animals. The illuminator has a weatherproof case and is mounted on the bottom of the main case. The 12 V power supply is switched by a Jetson Nano GPIO pin using an IRLZ44NPBF MOSFET.
Power supply: All components are powered by a lithium ion polymer battery, which has a high power density. We employ a battery with a theoretical capacity of 236.8 Wh (1600 mAh at 14.8 V). A generic 4S balancer circuit board provides over-discharge protection. The variable voltage of the battery is then regulated to 5 V for the Jetson Nano and 12 V for the infrared illuminator by Mean Well SCW20A-05 and SCW12A-12 converters, respectively.
Connectivity: SOCRATES may transmit the recorded data via three different means: wired ethernet cable, wireless LAN (Edimax EW-7811UN) or cellular connection (Huawei E3372H). If no basestation is available, we use the cellular connection to manually download the captured data. Otherwise, we connect via wireless LAN and the CoAP protocol (
A significant drawback of the custom SOCRATES hardware is the difficulty of producing a high volume of units, which is primarily due to the manually assembled case. Compared to commercially available monocular camera traps, mean power consumption is relatively high at around 1W, which necessitates relatively short battery exchange intervals of around 8 days. Finally, if possible, the wavelength of the infrared illuminator should be increased to around the widespread 940 nm. We did not notice any adverse reaction of observed individuals to the current 850 nm illumination, however, it might scare some other species with wider spectral sensitivity.
The software of SOCRATES is freely available (github.com/timmh/socrates) and is divided into three parts. The control software is responsible for controlling the camera hardware and runs directly on the device (cf. section 8.2.1). The stereo processing and animal localization run on a dedicated GPU-accelerated server (cf. section 8.4). The following sections describe the role and implementation of each software part in detail.
The control software of SOCRATES is responsible for reacting to signals from the PIR motion detector (cf. section 8.2.1), triggering capture, and communicating with the AMMOD basestation (cf. section 8.4). It is implemented in Python and runs on Linux via the NVIDIA JetPack SDK. We furthermore adjust the device tree to allow the PIR to trigger a wake up signal via the GPIO16 pin. This allows the control software to put the SOC into the power-efficient SC7 mode during periods of inactivity and resume once motion is detected. Once motion is detected, video material is encoded on the Jetson Nano’s GPU by synchronizing the image streams from the left and right cameras, concatenating them horizontally, and compressing the resulting video of resolution 2 × 1920 × 1080 using the HEVC video codec (
The central goal of SOCRATES is to infer depth information through stereo vision. In the natural world, as well as in computer vision, this is achieved by solving the stereo correspondence problem. To solve the stereo correspondence problem efficiently, the left and right images must be rectified. To obtain an accurate rectification, the intrinsic (internal camera parameters) and extrinsic (rotation and translation between the cameras) parameters have to be obtained by a calibration procedure. For the calibration of the intrinsic parameters, a calibration object (e. g. checkerboard pattern printed on cardboard) has to be captured by the camera(s) to be able to associate 3D points in the scene with 2D points in the resulting image. To obtain the extrinsic parameters, eight or more correspondences between images of points in the projections of both cameras must be established (
To detect and localize animals in the images produced by SOCRATES, we use the mmdetection (
A central goal of the AMMOD project is to automatically collect all observed data in a central repository (the AMMOD Portal, https://data.ammod.de), which will eventually be accessible to biologists and the general public. For SOCRATES, we ensure this by uploading the captured raw data via the CoAP protocol (
where NT is equal to the number of frames in the input video, NP is the number of pixels in a single frame, D(x,y,n) is the scalar disparity at some pixel (x,y) at time n, and mx, my is the optical flow from frame n to frame n − 1, calculated using
Samples of the data collected. The photographs on the left show the grayscale image of the left camera, the right image the colorcoded depth map obtained using stereo correspondence.
As can be seen, the temporal error is low for the vast majority of observations. Like with regular camera traps, at night time, some regions in the field of view might be insufficiently lit and therefore underexposed in the resulting images. In these regions, insufficient image information is available to perform successful stereo correspondence, leading to the outliers with poor temporal error apparent in
Due to the manual assembly, SOCRATES is hard to produce in large volumes. We plan to address this by working with commercial manufacturers and modifying existing camera traps models to include an optional wired trigger synchronization cable. By synchronizing the triggers of two or more camera traps in this way, we would be able to build flexible stereo setups with commercially available hardware and their superior power efficiency. The SOCRATES stereo calibration and correspondence procedures could be adapted to support this new hardware with minimal effort.
An AMMOD measuring station consists of the base station and a set of sensors that capture and deliver the measurements to the base station. The measurements are then preprocessed at the base station and can be forwarded to the cloud storage. The architecture of an AMMOD measurement station and the interactions between its components is depicted in
An AMMOD base station with external sensors, connected via wired (solid lines) or via wireless (dashed lines) connections.
As shown in the figure, the sensors are connected to the base station communication and processing block either via wired or wireless connection. In the former case, the sensor can be also provided with energy from the base station internal power supply. In the latter, the sensor has to be also equipped with external power supply.
The base station computer is currently deployed as a Raspberry Pi 4 single-board Computer running a Debian Linux operating system. Communication with the AMMOD cloud is done via an LTE modem that is connected to the Raspberry Pi via an Ethernet cable and, forwards collected data from sensors to the Internet.
The power supply module developed for the AMMOD project in the first phase consists of three elements (see
The power supply functions as a sensor as well, so that the parameters of the power supply (i. e. drawn currents and telemetry data of the power supply) are monitored and provided to the base station for further processing and maintenance. These measurements can be used to monitor the state of the power supply, such as state of the battery, charge state, power consumption and detect potential outages, but it may also be used to control other sensors and the base station, in order to adapt to the available energy. The power supply sensors are connected to the base station via low power, long range communication means that can also be used to connect external sensors that do not require high data throughput. Based on the developed module there were research works done towards predicting the available energy and the time remaining for the given consumer to run on it (
The currently deployed power supply modules are defined to support energy to constant load of up to 10 W. The energy storage options used range between 500 Wh and 2000 Wh, while the PV power is rated at 200 W peak nominal.
The main piece of software running on the board is written in C++. The code is currently hosted on a private repository on a Gitlab instance. The code can be moved to another repository once the current phase of AMMOD is finished.
The application goal is to receive files from the sensors and to upload them to the cloud. Due to the fluctuating power availability, the software can buffer files on a disk upon their transmission to the cloud. An internal component of the program can decide whether or not to transmit data to the cloud, compress it or not, etc. by applying a policy. A way to find an optimal policy (along with hardware components dimensions) for a given deployment site is given in Section 9.3.5.
The software context is divided into modules, each running on its own thread, which can belong to three classes:
As of the current time of writing, all sensors are connected to an Ethernet network (except for the sensing functions in the power supply, as already mentioned and the weather station, which is directly connected via USB). To unify their interfacing, it has been decided to write a library that handles the transmission to the base station, this library is referred to throughout this chapter as SensorAPI. The library uses the Constrained Application Protocol (CoAP), which is more lightweight, but similar, to HTTP, to transmit the given files to the base station.
The resource discovery feature is used for the automated discovery of the sensors, which means they can be added and removed from a site without touching the configuration file. The observation feature allows for sensors to trigger a GET request from the base station when a new file is produced. This is more energy-efficient than active waiting, and its embedding in the library is more convenient for a developer integrating a sensor.
The library is written in C++ and a Python binding is provided. It is available on a publicly accessible Gitlab repository
The base station software is configured via a JSON file that follows a predefined schema. The configuration has three main properties:
An exhaustive description of the configuration file is available in the git repository.
To access the base station from the Internet for maintenance purposes, we deployed a VPN running on a server accessible from the outside. For security reasons, it is possible to open an SSH connection to a base station only when inside this VPN.
The VPN server is currently hosted on a computer on the FAU campus. In the production phase, it could be hosted alongside the other AMMOD services.
From a research-oriented perspective, we tackled the problem of designing a base station (e. g. finding the right battery size, the right PV module, and an optimal energy management policy) by leveraging Design Space Exploration techniques. An overview of this process is shown in
An analytical parametric model of a typical AMMOD deployment site (a base station and its sensors) was designed and run inside a custom simulator. The parameters of this model are divided into a site-specific part (e. g. number of sensors, their type, their power consumption, etc.) and a base station part (e. g. the battery size, the size and number of hard drives, etc.) At the end of a simulation, the simulator outputs some metrics of the system: uptime of the base station, total amount of data transmitted, etc. These metrics, alongside the financial cost of the components, can be used as objectives for a Multi-Objective Evolutionary Algorithm (MOEA), which optimises the base station parameters.
Our results show that it can be financially interesting to have different base stations that have a different cost range to be deployed in different locations. For example, a base station deployed in a sunny region of Germany requires cheaper components than one deployed in a rainy region. We also use our tool to evaluate different energy management policies and to optimise the parameters of such policies. This methodology, which is named SIDAM, and its results were published in
For the communication links two areas with different requirements will be distinguished. The local link to the sensors and the internet link to the cloud. Both fields have different requirements which need to be addressed. All the communication links discussed can be found in
An overview of the communication links of the base station. Dashed lines signify wireless links.
All the data collected by the sensors and send to the base station is preprocessed before it is forwarded on to the AMMOD cloud. Erroneous data and false positives are discarded and data is compressed. Depending on the configuration of the base station and the activity of the sensors still many gigabytes of useful data may be generated each day. To get any generated files to the AMMOD cloud where it can be stored for the long-term, fully analysed and presented to the public.
To transfer this amount of data different technologies can be utilised. Wired solutions like Ethernet or DSL are the most efficient approach but usually not available as the AMMOD stations need to be placed where animals roam freely, not where the man-made infrastructure is best.
The wireless technology that provides the best compromise between data rates, power consumption, availability of parts and long range network coverage within Germany is 4G/LTE. Any Internet of Things (IoT) focused long range technology like LoRa usually does not allow for a high enough data throughput. Satellite communication links are an interesting option to look at for very remote locations where 4G/LTE coverage is insufficient. Those however usually require more expensive hardware, a more complex setup and have a higher power consumption. For the default AMMOD station these are too expensive.
Another alternative is the implementation of a directional radio link between the AMMOD base station and either an 4G base station or a 4G client specifically set up for the usage in AMMOD in a location with better cell reception, possibly via multiple hops. This is always a more expensive setup than just implementing a 4G modem and antenna in the AMMOD base station itself but can be very valuable for the setup in for example a remote valley. The directional link itself can be more efficient in transmitting the sending power to to next communication partner, but that is only an advantage if that device does not need to supply a 4G modem with battery power as well.
The huge advantage 4G/LTE has over most any competitive technology at the moment of writing is the availability of systems. A 4G modem can be bought relatively cheaply and be effectively enhanced with a directional antenna on a mast a few meters high. Such an additional antenna allows for a very efficient transmission of signals and therefore the required sending power can be greatly reduced. The mast assures, that the antenna is as close as possible to having a direct line of sight to the nearest 4G base station. Such a static setup prevents any mobility of the system but this is in line with the requirements for the stationary AMMOD base station. A precise orientation of the antenna needs only to be done once when the station is assembled first.
For the test sites that where set up within the AMMOD Project, a few different configurations where examined. The following is the configuration of the ones used at the current test sites. Different hardware can be used in a similar manner to achieve comparable results.
In this example implementation a RUT240 Industrial Cellular Router by Teltonika Networks is used for the 4G/LTE radio link. It is specifically designed for Machine to Machine Communication, provides good interfaces which can be accessed and modified easily externally and a low power consumption. The modem accepts a comparatively wide range of DC voltages which allows it to be used with either a 12 V or a 24 V supply voltage from the base stations power lines. The two external antennas for MIMO-4G communication were exchanged for a WB 23 antenna by Wittenberg, which is shown in
For the test site in Bonn, a mast-setup was used that has minimal impact on the setup area and is thus suitible to be set up in natural reserves. Four ground anchors are used in conjunction with one steel wire each to keep the mast upright. This setup is depicted in
The sensor links allowing for data transfer between the sensors and the base station need to be very flexible as they are required to work for many different scenarios. The positioning of the sensors is determined through its function. Visual sensors need to have a clear view onto places of interest, for example an animal trail or a den. Audio sensors work best if they are shielded from unwanted noise, and insect traps will be placed where insects are commonly found. Sensors may be placed directly at the base station or 100 m away. There might also be a free line of sight or many obstacles between base station and sensor. Therefore the path between the sensors and the base station will usually be suboptimal for the transfer of data. To provide these data links, we investigated different technologies.
As with the internet link, a wired connection will always provide the best reliability in data transfer and the highest power efficiency compared to wireless solutions. This is always the preferred solution for sensors close or directly attached to the base station. An Ethernet connection is usually the most commonly available option and can be integrated into many systems easily. While it also allows for power transfer via Power over Ethernet, it does impose a certain amount of overhead and is not really adequate for minimal systems. A lower level wired link that is commonly used, especially for industrial applications, would be a serial interface like RS-232 of RS-485 over which protocols like Modbus can be transferred. This is a simpler and more efficient approach that does not require full internet capabilities from the sensor but is limited in its bandwidth when compared to the Ethernet standard. This is the preferred option when connecting a close sensor up top 10 m away from the base station, which does not generate a large data volume. When connecting a sensor that generates high data volumes of 1 GB per day or more, like video sensors, to the base station, a wired connection is usually the only option as any wireless solution with a high enough data rate consumes also a very high amount of data. In these cases, Ethernet will usually be chosen as such sensors are also fairly powerful in terms of computational capabilities and Ethernet does not provide much of a burden. Also any sensors that draw power from the base station need a wired power connection. Along the same routing, a data link cable can also be placed which makes a wired solution realisable with very little effort.
The most powerful wireless technology in terms of data rate, flexibility, and compatibility is WiFi. It does, however, also consume more power if compared to other technologies, which are more focused on an application within the Internet of Things (IoT). It also has a range limit of 100 m, which may not be enough for some sensors. The same problem of range limitation has Bluetooth which builds upon a lot of WiFi technologies and does not provide any real benefit for the AMMOD scenario. ZigBee also focuses on short range communication usually below 100 m. It is capable of communication over longer distances in good conditions and might be a good fit for specific AMMOD scenarios but is limited to low data rates and is not tested further in this AMMOD phase as there are alternatives that offer a better feature set. WiFi is the preferred option for sensors that generate considerable amounts of data between about 100 MB and 1 GB per day and need to be placed too far away from the base station to actually be connected with a cable.
Some sensors are spread out over a large area to collect data optimally. For example insect traps may be set up across a large field or around a wooded area. In a constellation like this, distances between 100 m and 1 km are common. As these traps actually collect biological samples and do not generate large amounts of data, a long-range, low throughput wireless link is ideal to realise an efficient link. An approach with multipath-propagation might also be interesting to research but is not within the scope of this project. LoRa is a long range protocol useful for Internet of Things (IoT) application which is suitable for the aforementioned scenario. It allows for communication over up to 10 km at data rates of up to 50 kbit per second. It is also widely available. The LoRa-WAN network is a large area network which covers wide parts of Germany. It provides an alternative access method to the internet for base stations or stand-alone sensors which produce only limited amounts of data. It covers mostly areas which are also covered but LTE but could be expanded upon for specific AMMOD-scenarios and is considerably cheaper to access than a contract with an LTE service provider. LoRa is the solution of choice for the AMMOD project to connect any sensor that is placed far away from the base station above 50 m.
For the LoRa connectivity of the sensors, different approaches are possible. If the sensor utilises a Raspberry Pi or an Arduino compatible system, there are extension boards available that integrate a LoRa modem on a low level. There are also USB solutions which can be utilised for the base station connectivity and can be an option for the sensors as well, if other methods of integration are not viable.
To analyse the large amount of data collected by all AMMOD systems and correlate this data correctly, a precise time synchronisation over all systems is necessary. That way all deployed sensors as well as all AMMOD base stations assign the correct time stamp to each measurement. The external timer used within the AMMOD project is derived from the Global Positioning System (GPS). LTE/GSM could also be used as a reference. Their coverage is however not as universal in very remote areas. A GPS receiver can also be integrated into every wireless sensor to achieve time synchronisation independent of the AMMOD base station. Sensors directly connected to the base station with a wire can have their clocks updated from the base station through NTP and do not need to invest in their own GPS receiver. The GPS modem also provides location data which can be used to map the layout of an AMMOD site automatically and possibly be useful in the recovery of lost or stolen equipment.
The time synchronisation can be used to connect two or more local measurements. For example if an animal is detected by a video camera and a microphone, movements and sounds can be correlated if they are registered at the same moment. Also data from multiple base station can combined. For example reactions to a lightning strike may be observed at multiple stations within a certain range. Also this synchronisation remediates clock drift. For example, events like a sunset can be reliably targeted with scheduled measurement runs.
Just like the time of day, or more practically the position of the sun, weather data is very useful to give context to animal and plant behaviour. Rain and wind for example can greatly influence the flight of insects and birds. Certain temperatures or values of air humidity may trigger the expulsion of pollen which can be detected. Also certain weather conditions may prevent reliable measurements by some types of sensors. Heavy movement of leaves through strong winds may falsely trigger motion sensors for video sensors. The noise of heavy rainfall may drown out any animal sounds detectable by audio sensors. Thus deactivating these sensors in adverse weather can save a lot of power and reduce the amount of unusable data collected.
Data drawn from weather services may also be an option for a minimal AMMOD station were it is not feasible to invest into a local weather station. A dedicated local solution is however preferable as it provides a far better spacial and chronological resolution. It should be emphasised that only a local station can actually measure local effects such as the wind direction at a forest boundary or the air temperature and humidity on a river bank.
A good weather station measures all the necessary weather data as well as GPS data needed for the operation of an AMMOD station. For the test sites of the AMMOD project, the Weather Station Compact WSC11 from Thies Clima was used. It is a compact system that was designed for local weather monitoring. It measures a wide range of values like temperature, humidity, air pressure, wind speed and direction, precipitation, and illumination. A GPS receiver is also integrated. The precision of the data provided and the measurement setup do not allow for a usage in a meteorological application but is more than enough for the kind of context data required.
The WSC11 can be mounted to the same mast as the antennas of the base station. Ideally the weather station is mounted to the very top as any components above it may obstruct its sensors for sun and rain detection. Many adaptors and extensions are available from the manufacturer to facilitate this. The WSC11 can give analogue outputs which can be digitised externally. It can also be controlled and read out via Modbus, which was implemented for the AMMOD project. The WSC11 was bought with the optional, ready-for-use 5-core cable through which the Modbus and power lines run. These can be split in the base station housing and connected to the base station’s power supply and a RS485-to-USB adaptor. The Modbus calls for the base station where implemented as a dedicated library. This way a modular integration into the base station as described in Section 9.3.2 was realised.
As shown above, the power management, the communication links and the software of the base station could be implemented so that the AMMOD test sites can be operated to specification. For optimal operation, power efficiency can be increased further by integrating components and evaluating collected data about power consumption. Alternative communication technologies can be tested to improve the adaptability of the setup for different scenarios. The operation requires further field tests to standardise the connection links throughout the AMMOD platform. To realise a feature-rich and efficient computational unit as the core of the base station, an FPGA-board would be ideal. It can be programmed in software and hardware, and thus, would enable power-efficient hardware acceleration for sensor data processing, compression, etc., at the site. However, these boards require a waterproof housing that still allows adequate cooling. Availability of such boards were poor for the duration of the project and could therefore not be investigated further.
This chapter provides all the necessary steps to connect the AMMOD base station to the AMMOD data portal. The AMMOD data portal is the web service developed in the first phase of the AMMOD project to meet the project’s data management requirements. The AMMOD data portal consists of: (1) a backend for storing and managing the data, (2) a web-based user interface and (3) an Representative State Transfer (REST) Application Programming Interface (API) for programmatic upload, search and download of data. The data is either raw, processed or telemetry. The website provides additional support for data and sensor management tasks like: visualizing the status of the network of deployed sensors and base stations, a dashboard for displaying and plotting the telemetry data, a metadata schema validation tool and access to documentation. The backend and frontend of the AMMOD data portal are containerized and deployed in a cloud environment. The system has a dependency to the third-party web service, https://sensor.awi.de, which is highlighted in the requirements section.
Since AMMOD stations are designed for modularity and flexibility of both sensors and data processing, a large amount of heterogeneous data is produced. The data management workflow put in place is designed to accommodate the needs of the AMMOD station and sensors developed in the pilot phase. Automated transfer, preservation, lineage and accessibility of the data produced are ensured provided that the requirements explained in Section 10.2.1 are met.
The AMMOD data portal is the cloud-based web service developed as part of the automatization of the AMMOD data management and provides:
Currently, the AMMOD data portal consists of two separate systems, a production system and a staging system. Both run the same code base and offer the same functionality with some significant differences, described below.
The system is deployed in a cloud environment using docker-compose, allowing it to work with multi-container applications. Backend and frontend both have their dockerfiles, so that a docker image can be created and used anywhere. All required environmental variables are included in the repository’s readme file.
In the remainder of the chapter, the essential steps needed to use the system as-is are listed. It is highly recommended to comply with these requirements and connect future AMMOD stations to the existing systems.
An AMMOD station acts as a local hub for the network of sensors deployed nearby it. It can be considered a middleware between the sensors and the cloud solution (see
Base station and sensor produced (meta-)data and telemetry files are stored first at the base station before being automatically transferred to the AMMOD data portal. When and how often data is automatically uploaded to the AMMOD data portal is managed by the base station.
Base stations need to be registered in the sensor registry and assigned a valid access token (see Section 10.2.1 for more details) before they can upload data to the AMMOD data portal. Authorized base stations upload (meta-)data and telemetry to the dedicated end-points of the API. The API is providing back a message response which is used by the base station for housekeeping and error handling operations. All models of the responses are included in the API documentation.
Please note that for cases where the data is not physically suitable for being stored at the base station and need to be processed in a lab or facility first, it can be uploaded manually to the AMMOD data portal using the same API.
The requirements for connecting base stations to the AMMOD data portal are grouped as follows:
Base stations and sensors need to be registered and assigned a unique identifier (deviceID) before the integration with the AMMOD data portal. This is done using the AWI sensor registry, (which instance of the sensor registry is used for which instance of the AMMOD data portal is detailed in the introduction to this chapter). Upon filling up the registration form with all mandatory information of the base station or sensor, the deviceID is assigned by the sensor registry manager. The AMMOD data portal is automatically pulling the base stations and sensors-related information from the registry to visualize all the deployed AMMOD devices in a map and tabular view. Only uploads with valid deviceID are accepted.
It is highly recommended to nominate a sensor registry manager, who is responsible for the registration process and updating sensor’s metadata. The AMMOD base station operator/s is/are responsible for the sensor and data for one or more collection sites.
The sensor registry manager needs to create an account for the AWI sensor registry. This is currently done by sending the request via e-mail to o2a-support@awi.de. To get familiar with the functionalities of the service is recommended to visit the tutorial webpage of the AWI web service registry https://sensor.awi.de/?site=tutorial.
Please refer to
Please refer to
The AMMOD data portal uses the API of the AWI web service registry to retrieve all information of the AMMOD collection. The collection AMMOD has a persistent ID named “collectionID” which is included in the environmental variables of the AMMOD data portal.
An AMMOD base station is considered a non-movable piece of hardware which keeps the same geographical location till it is dismissed. In contrast, AMMOD sensors can be:
Metadata files are a mandatory requirement for any data file upload. Each REST API POST request to the AMMOD data portal metadata endpoint must have a minimum of 2 files, one of which must be the metadata file. Only the telemetry endpoint of the API accepts just one file which contains the base station or sensor telemetry record.
To ease the integration and development process, JSON schemas for the metadata and telemetry file are publicly accessible at the following repository: https://gitlab-pe.gwdg.de/gfbio/ammod-examples-schemas. The repository provides the full schema documentation with examples of data and their metadata.
The API access of the AMMOD data portal is restricted to authenticated base stations and users. The authentication procedure is token-based.
Once the registration and assignment procedure of a new base station is complete (see Subsection 10.2.1.1, ‘Base station and sensor requirements’) the sensor registry manager will provide a token pair (access and refresh token) to the base station operator. As default, base stations get a token only with upload permission.
Users that want to consume the API for upload, search and download data, either via command line or using the web graphical interface, need to login in the AMMOD data portal and request the desired permissions. Users can login in the production and/or stage system using the GFBio Single-Sign-On-Service (SSO) service by entering the credentials of:
The portal will automatically detect for which service you don’t have permissions to interact with the data (you should see a lock on the services where access is not granted). By clicking any of the locked services it pop-ups a message with the instructions to request the specific access. After a first login, the user is available in the website administration page and the admin can then set the required permissions. Please note that as the production and staging systems are two identical but independent systems, the login and permissions parameters are not synchronised among them.
AMMOD data portal, Home section. Devices overview in map (A) and table (B) format. (URL: https://data.ammod.de, accessed 24 January 2023).
AMMOD data portal - Data and metadata section. Search results from the applied filter settings (A) and upload data GUI interface for manual transfer of data to the Cloud (B). (URL: https://data.ammod.de, accessed 24 January 2023).
The first access to the Search will list all the data available in the database with a pagination of 20 results per page. The list of results can be refined by providing user-defined values for the parameters in the filter page: time frame, location (base station based), data type (raw, processed) and sensors (
Users can upload data collected manually using the Upload tool (
AMMOD data portal – Tools section. API documentation in Swagger (A) and the user token management interface (B). (URL: https://data.ammod.de, accessed 24 January 2023).
In this section some key factors related to the current limitations and possible extension for future usage of the AMMOD data portal are discussed.
The storage component of the AMMOD data portal has been sized according to the estimated volume of data being generated during the project pilot phase. Accordingly, the storage is set up to an initial amount of 5 TB. This is also backed up in two different locations making a total size of 15 TB. These limitations are given by the cloud service providers hosting the application. For the runtime of the AMMOD project, the data portal is deployed using the cloud computing services offered by the German Network for Bioinformatics Infrastructure (de.NBI) and by the storage in the GWDG. The current storage space is not a hard limit, and the cloud service providers offer the possibility to attach new volume to the running instance. Nevertheless, this is not an automated process and the used space needs to be monitored and managed accordingly.
As a long term solution, the data uploaded to the AMMOD portal will be transferred to the NFDI4Biodiversity core storage. Dedicated APIs to upload/download the data from the core storage are already prepared and tested. They can be deployed once a stable version of the NFDI4Biodiversity core storage is available.
Although the system has been proofed and tested under operational circumstances, bugs and errors can show up unexpectedly anytime. To continue to operate the AMMOD data portal in a resource-saving way, the performance monitoring and error tracking application Sentry (https://sentry.io/) is integrated in the AMMOD data portal. Errors are automatically collected and sent to the Sentry instance, and the admin gets informed via email immediately when something is wrong. Moreover, users can report bugs/errors using the “contact us” feature on the AMMOD website and an issue is automatically created in the software development environment.
The system has been designed to be ready to scale in case the number of users and base stations increases. At the moment the workload in terms of requests for data uploads is not considered a critical point. Indeed the workload in terms of requests for data download could increase exponentially in case the data is public. Therefore, users have no limits on the amount of files they can download, but only a 1 GB data volume per request. Once the archive with the download request is generated and the link to download the data has been sent to the user, the link is alive for 24 h. Afterwards, the link expires and the archive file is deleted from the server. These limitations might change in the future when the AMMOD data portal will be integrated with the NFDI Core Storage.
The AMMOD Data Portal allows data to be shared mainly between project members and possibly authorized collaborators. Only the status and telemetry data of the stations is publicly available. The data portal is neither a long-term archive, nor a data publication platform. To make any of the collected scientific data public, it is recommended to deposit it on one or more of the long-term GFBio data centers (https://www.gfbio.org/data-centers). This can be done using the GFBio Data Submission System (https://submissions.gfbio.org).
We thank our colleagues from the German Federation of Biological data (GFBio e.V.) and Alfred Wegener Institute (AWI) who provided insight and expertise on Research Data Management (RDM), metadata standardization and sensor management.