the Creative Commons Attribution 4.0 License.
the Creative Commons Attribution 4.0 License.
Coastal Atmosphere and Sea Time Series (CoASTS) and Bio-Optical mapping of Marine Properties (BiOMaP): the CoASTS-BiOMaP dataset
Giuseppe Zibordi
Jean-François Berthon
Download
- Final revised paper (published on 29 Nov 2024)
- Preprint (discussion started on 26 Jun 2024)
Interactive discussion
Status: closed
-
RC1: 'Comment on essd-2024-240', Jaime Pitarch, 04 Jul 2024
-
AC1: 'Reply on RC1', Giuseppe Zibordi, 27 Jul 2024
Reply to the Review by Jaime Pitarch
Below are the replies from the Authors to the Reviewer’s comments.
General comment
I am very pleased to have been given the opportunity to review this manuscript as I am aware of the lifetime work of the authors in defining the highest standards and producing high quality reference data in the field of satellite ocean color. The monitoring programs CoASTS and BiOMaP have generated lots of publications, and a remaining question was where all the data was going to be after the finalization of such programs. So now it appears that a circle is closed.
I have read the paper and downloaded the dataset. Before publication, I have a number of comments of varying importance that, to my understanding, need attention.
Reply
The Reviewer comments are duly considered and itemized. A reply and clear actions are provided for each one (for the benefit of conciseness, the figures provided by the Reviewer are omitted from the reply).
Major comments
Comment #1
Absorption from water samples is only provided at the Satlantic bands, which is regretful, as it was measured hyperspectrally. I do not know the reason to downgrade the data, and it definitely reduces its value for optical studies, also considering the growing interest in hyperspectral data (e.q., PACE). The authors are encouraged to submit the hyperspectral data.
Reply
The CoASTS-BiOMaP data set provided through PANGAEA was conceived to support bio-optical investigations with comprehensive multi-parametric near-surface quantities. Because of this, the laboratory absorption measurements were only provided at the center-wavelengths of the related multi-spectral field radiometric data.
It is definitively appreciated that i. the CoASTS and BiOMaP measurements can support a number bio-optical, methodological and instrumental applications beyond the strict and obvious bio-optical ones and that ii. hyperspectral measurements (when available) or full profiles instead of the sole near-surface data are relevant and desirable data. But this is something that goes beyond the objectives of the current work. An expansion of the shared data set could be considered as a future task, but not for the current data submission.
The objective of the work and of the related data set will be strengthened in the introduction.
Comment #2
Paragraph from line 175 to 182: on the above-water reference sensor, I see the correction for the imperfect non-cosine response. What about other uncertainty sources such as temperature and non-linearity, as it is recommended in above-water radiometry (e.g., Trios)? And are any of these corrections made to the in-water sensors?
Reply
In agreement with common know-how, multi-spectral radiometers rely on a much simpler design and technology with respect to the hyper-spectral ones. Because of this, they exhibit lower sources of significant uncertainties: the temperature dependence is negligible within the 410-700 spectral range (Zibordi et al. JTECH 2017); stray-lights are negligible assuming interference filters are of high quality and their out-of-band response is within specifications for ocean color applications (Johnson et al. AO 2021). Because of this, some uncertainties due to the potential non-ideal performance of multi-spectral radiometers are commonly not included in uncertainty budgets. Exception is the non-cosine response of irradiance sensors, which depends on the manufacturing and material of individual irradiance collectors.
These elements will be mentioned in the relevant section.
Comment #3
Line 193-196: the interval 0.3 m – 5 m looks arbitrary. Any comments on why this choice is appropriate? Does it relate to the unphysical 𝐾𝑑 values that I report below?
Reply
The determination of subsurface radiometric values from profile data always requires the identification of a suitable near surface “extrapolation interval” exhibiting linear dependence of log-transformed radiometric data with depth. In the case of the CoASTS-BiOMaP data, the most appropriate extrapolation intervals were determined within the 0.3 and 5 m depth limits (these limits do not identify the extrapolation intervals themselves, but the values within which the extrapolation interval is generally located).
How the extrapolation intervals are determined is well stated in the manuscript. The process is definitively subjective (i.e., the extrapolation interval is chosen by an analyst with the aid of a number of ancillary information), but for sure, it is not arbitrary.
Some further clarifications will be added together with references to avoid any misinterpretation on the actual extrapolation intervals and the depth limits within which they are commonly located.
Comment #4
The transmission of upwelling radiance through the surface to form the water-leaving radiance is made with that 0.544 factor, which is not up to date with today’s knowledge. I suspect that the reason to choose this value is because the difference in the final product would be minimal when using another one. However, I report evidence that this is not the case.
For a flat surface, the relationship between in-water and in-air upwelling radiance is: 𝐿𝑤=𝜏𝑤,𝐿𝑢(0−)
Where 𝜏𝑤,=1−𝜌𝑛𝑤2
𝜌 is the Fresnel reflectance of the air-sea interface. Assuming unpolarized light, it has an analytical expression 𝜌=12|sin2(𝜃𝑎−𝜃𝑤)sin2(𝜃𝑎+𝜃𝑤)+tan2(𝜃𝑎−𝜃𝑤)tan2(𝜃𝑎+𝜃𝑤)|
𝜃𝑎 and 𝜃𝑤 are the wave propagation angles in air and in water, respectively, and are related by sin(𝜃𝑎)=𝑛𝑤sin(𝜃𝑤)
For 𝜃𝑎=𝜃𝑤=0, there is a singularity. One can apply the small angle approximations for the trigonometric functions, so in the limit, it is: (𝜃𝑎=𝜃𝑤=0)=(𝑛𝑤−𝑛𝑎𝑛𝑤+𝑛𝑎)2
It is commonly accepted now that it is inaccurate to assume a constant 𝜏𝑤,𝑎 for a given geometry due to the spectral dependence of 𝑛𝑤 (and secondary influence by temperature and salinity too). Such dependences are taken from the state-of-the-art values by Roettgers et al. (2016). Therefore, the theoretical curve for 𝜏𝑤, can be seen in Fig. 1. In addition, I have made some Hydrolight simulations, in which the same 𝑛𝑤 values are used, but also, the transmission is affected by the surface roughness depending on the wind speed. What emerges from Fig. 1 is that increasing wind speeds reduces light transmission. In terms of the total error made by assuming 𝜏𝑤,𝑎=0.544, it may not seem much, but in reality, they are in the order of 1-2%, which accounts for about 20% of the total uncertainty reported for the final 𝑅𝑟𝑠 product. Therefore, to reduce total uncertainty I encourage the authors to consider updated look up tables for 𝜏𝑤.
Reply
The spectral dependence of the water-air transmission factor for a flat sea surface in the interval of interest for the CoASTS-BiOMaP data set is well within +/-1%: this is confirmed by the data provided by the Reviewer and also by the work of Voss and Flora (JTECH, 2017). Because of this, neglecting the spectral dependence of the water-air transmission factor does not appreciably affect the uncertainty budget of the derived radiometric quantities.
The inclusion of the wind speed dependence in such a transmission factor adds the detrimental dependence on wind speed to Lw. In fact a properly determined subsurface Lu is marginally affected by sea state (and consequently by the wind speed). Thus introducing a wind speed dependence on the water-air transmission factor would add such a dependence to Lw. This is not a desirable dependence for data envisaged to support bio-optical modelling.
A mention to spectral dependence of the water-air transmittance will be added.
Comment #5
On the fit of simple analytical functions to variables like the Q factor and possibly others like 𝑏𝑏, I did not understand if the actual data were replaced by the fits. If that is the case, I prefer then to have both data and uncertainty rather than their surrogate analytical forms.
Reply
The fit of Qn data is introduced to minimize the impact of any uncertainty affecting intra-band radiometric calibrations of multi-spectral radiometers or the extrapolation process. In practice, any deviation indicated by the fit with respect to the actual Qn values is used to monitor the performance of the Lu and Eu sensors in the field (deviations of +/-1% are typical, variations exceeding +/-2% are a warning). The fitted Qn data are those saved and included in the shared dataset. Still, both Lu and Eu data are provided, any data user may re-compute the data at his preference .
More details will be provided on Qn fitting, but no additional action is taken. Quality controlled and “smoothed” Qn values are those expected to best serve the community.
Comment #6
On the bidirectional correction in lines 233-241, it is a bit disturbing to read in line 241 that in case 2 waters “this correction may be affected by large uncertainties”. There are significant parts of the dataset in case 2 waters. How large are those uncertainties? Ongoing research has proven that is it better to apply Morel than not to apply any correction at all, and Morel has shown to provide surprisingly good results in case 2 waters, not because of the qualities of the model itself, but because all bidirectional correction models underestimate the correction to be made, but chlorophyll is overestimated in case 2 waters with the band ratio of Morel, which produces a higher correction, that ends up being beneficial. In any case, I believe that this part of the processing will need update to be in line with latest developments in bidirectional studies, knowing the interest of the authors in keeping the uncertainty budget as low as possible.
Reply
Sorry if the Reviewer feels a bit disturbed by a reasonable sentence such as “this correction may be affected by large uncertainties” addressed to the Morel et al. (AO, 2002) correction for bidirectional effects applied to non-Case 1 waters. Clarifying that the implementation of this correction for CoASTS–BiOMaP data relies on actual Chla values from HPLC analysis and not from any algorithm as suggested by the Reviewer, the sentence is certainly well supported by the work of Talone et al. (2018).
It is emphasized that all the fundamental data for producing alternative corrections for bidirectional effects are available: any user can thus implement its own ignoring the one applied for the shared data.
The potential for producing high level radiometric data products with alternative corrections for bidirectional effects benefitting of the basic radiometric quantities included in the data set, will be mentioned.
Comment #7
On the ac-9 measurements, I have several comments that follow.
How regular were the factory calibrations? It is said that instruments have to be calibrated before and after any campaign. Is this the case with the ac-9?
Reply
The two AC9s used during the CoASTS and BiOMaP campaigns were factory calibrated on a yearly basis (obviously with a number of exceptions over almost three decades). Definitively, the instruments were sent to the manufacturer for maintenance and calibration each time there was evidence of sensitivity decay in a single band (implying the replacement of the related filter and detector). The pre- and post-campaign “calibrations” correspond to the milli-Q water offset measurements performed by the JRC team on board and with the instrument in his deployment configuration. These measurements were intended to detect and correct any minor bias affecting the factory calibration coefficients of individual bands over time (i.e., between successive factory calibrations).
Factory and field calibrations will be better detailed.
Comment #8
The Zaneveld method does not correct the non-finite acceptance angle of the c detectors as it is stated (note that the “c” is missing in line 277), and in fact it is rarely corrected by anybody. To do that, one should have a guess of the VSF between 0 and 0.93 degrees, but in any case, the “real” 𝑐𝑡−𝑤 is higher than the measured than the factor that varies a lot, mostly between 1 and 2.
Reply
Thanks you for catching the inappropriateness of the statement on the correction for the non-finite acceptance angle. Definitively, Boss et al., (2009) determined Ct-w(AC-9)/Ct-w(LISST-F) = 0.56 (0.40-0.73). But again, it would have been speculative any correction not supported by specific VSF measurements.
The text will be revised declaring that corrections are not applied for the non-finite acceptance angle of the c detector.
Comment #9
On the scattering correction method of the absorption data from the “a” tube, I also believe that the Zaneveld method questionable. Zaneveld overcorrects the absorption data, which leads to an underestimation. I see indirect evidence of it in Figure 5 from the manuscript, where the absorption comparison at 443 nm almost always shows negative biases with respect to the laboratory measurements (although the ac-9 provides better closure of 𝑅𝑟𝑠 than the water samples as I show below, so this is puzzling and needs to be addressed by the authors). I suggest using the method by Roettgers et al. (2013), that, if applied, is supposed to perform much better. This choice should be in line with authors approach of using only consolidated methods, approved by the two very good assessments by Stockley et al. (2017) and Kostakis et al. (2021).
Reply
Roettgers et al. (2013) showed that the Zaneveld et al. (1994) underestimates the absorption for the wavelength greater than 550 nm. In the blue and blue-green, and in particular at 443 nm, the agreement was shown quite good between the AC9 and the “true” absorption from a PSICAM. Stockley et al. (2017) overserved relative errors lower than 20% for the “Zaneveld et al. (1994)” correction in the spectral range 412-550 nm (lower than 10% for wavelengths 412-488nm). Thus, the negative biases observed at 443 nm documented in the manuscript and mentioned by the reviewer, cannot be explained only by the scattering correction method “Zaneveld et al. 1994”.
Also, the hypothesis of negligible non-water absorption in the NIR was shown to be questionable for highly turbid waters (e.g., Elbe River, Baltic Sea and North Sea) but acceptable for the oligotrophic Mediterranean Sea (Stockley et al., 2017).
It is agreed that the correction method proposed by Roettgers et al. (2013) and verified in Stockley et al. (2017) is definitively a progress with respect to Zaneveld et al (1994), in particular in the green and red spectral regions, but his universal applicability is not assured. An excerpt from Stockley et al. (2017) states: “The performance of the empirical approach is encouraging as it relies only on the ac meter measurement and may be readily applied to historical data, although there are inevitably some inherent assumptions about particle composition that hinder universal applicability.”
Also from Stockley et al. (2017): “Methods experience the greatest difficulty providing accurate estimates in highly absorbing waters and at wavelengths greater than about 600 nm. In fact, residual errors of 20% or more were still observed with the best performing scattering correction methods.”
Considering the above findings, the AC9 data are provided with the correction originally proposed by Zaneveld et al. 1994, still appreciating it is far from being the most accurate. In the manuscript this explicitly acknowledged through the comparison of absorption measurements from the AC9 with those from laboratory measurements performed on discrete water samples.
Some of the above elements will be included in the manuscript to support the preference to process the AC9 data applying the correction scheme proposed by Zaneveld et al. (1994).
Comment #10
In fact, for research purposes, it is recommended that the authors share the absorption coefficient uncorrected for residual scattering, so it can be useful material to further investigate this matter.
Reply
As clearly stated in the section on ‘Data availability’, CoASTS-BiOMaP data not included in the current dataset are accessible upon reasonable request. The objective of the work is to provide open access to processed near-surface data accompanied by a comprehensive description of the field and data handling methods.
Comment #11
On the quantification of the uncertainties coming from the ac-9, certainly the value 0.005 𝑚−1 is not a proper estimate. That is a rule of thumb estimate of the instrument precision in the user manual, which is accompanied by the 0.01 𝑚−1 accuracy, also in the manual. There is no mention of uncertainty sources related to instrument absolute calibration, non-linearity, determination of the pure water measurement, correction of the temperature and salinity differences and correction of the residual scatter, some others related to the measurement protocol and the individual operator, and even some others that I may have missed. All these sources are likely to result in something bolder than the manufacturer user manual. The authors are expected and encouraged to investigate and comment on these aspects. Otherwise, how does one explain the differences that the authors find in their Figure 5?
Reply
Based on theoretical Monte Carlo computation, Leymarie et al. (2010) provided estimations of the relative errors of 10 to 40% for ct-w and generally lower than 25% for at-w (5-10% when absorption by in water optically active components is high) but up to 100% for waters showing high scattering.
Stockley et al. (2017) observed relative errors lower than 20% for the “Zaneveld et al. 1994” correction for wavelengths 412-550nm (lower than 10% for wavelengths 412-488nm) and more than 50% for wavelengths greater 600nm. Twardowski et al. (2018) provided an estimate of the “operational” uncertainty (for example, considering 2 calibrated AC9 close one to the other) as low as 0.004 m-1 (not taking into account errors associated to the scattering corrections).
Considering these results the manuscript will be revised indicating that the uncertainties in AC9 absorption are larger than 0.005 m-1, and can reach several ten percent in highly scattering waters with more pronounced values in in the blue-green spectral regions.
Comment #12
I also have a few concerns about the Hydroscat backscattering data. First, in lines 331 and 332, what is exactly meant with the annual factory calibration “complemented” by pre-field calibration, in terms of determining the scale factor and the dark offset of the measurement?
Reply
Equivalent to the procedure put in place for the two AC9s used within the framework of the CoASTS-BiOMaP campaigns, also for the two HydroScat-6 there were regular factory calibrations tentatively performed on a yearly basis. The pre-field and post-field calibrations (determination of the spectral “Mu” response curve coefficients and gain ratios) performed in laboratory by the JRC team with a “calibration cube” and a spectralon reference plaque allowed to detect and correct sensitivity changes between successive factory calibrations.
The difference between factory and pre-field calibrations will be clarified.
Comment #13
Equation (4) is the correction for absorption along the pathlength recommended by the manufacturer. However, after investigating on it, Doxaran et al. (2016) investigated on it and found that the “0.4” is a totally arbitrary number. They proposed a more accurate expression instead.
Reply
Doxaran et al. (2016) provided findings on the basis of measurements performed in: i. Río de la Plata turbid waters (Argentina, with total scattering coefficient at 550 nm greater than 20 m-1 (average around 50 m-1 ?) and ii. Bay of Bourgneuf Waters (France, with total scattering coefficient at 550 nm greater than 10 m-1 (average around 40 m-1 ?)
Because of this, the empirical relationship by Doxaran et al. (2016) indicating “Kbb-anw=4.34*bb” (see their figure 5b for the HS-6) refers to values of bb spanning between “0.0” and 2.5 m-1.
The BiOMaP values of bb roughly range between 0.0005 and 0.1 m-1 (with values of anw < 1.0 m-1). In this interval of bb values, figure 5 by Doxaran et al. (2016) shows that simulated values follow a relationship with a much high slope than the empirical fit resulting from the whole range of simulated bb. Thus, is that empirical fit really more appropriate for the low bb values found in the data set than the standard relationship used here? For sure, the problem is an open one.
In the manuscript it will be stated that in the absence of any consolidated processing for HdroScat-6 data, the CoASTS-BiOMaP processing was made relying on the equations provided by the manufacturer.
Comment #14
Removal of pure water data is made after tabulated data by either salt water or fresh water by Morel, but the state of the art values are those given by Zhang et al. (2009). Their model is analytical and has an explicit dependency on salinity, so that one may use concurrent CTD data for obtain 𝑏𝑏𝑤 accurately. Again here, the differences on the final products are likely to be small, but it is preferrable to replace old and biased values with updated ones at zero cost.
Reply
Zhang analytical values are for sure a general improvement. However, in the majority of cases it would have an almost negligible effect on the retrieval of CoASTS-BiOMaP bbp. In the oligotrophic clear waters of the eastern Mediterranean Sea showing salinity values around 38.0-39.0, the difference in Beta(90degrees) between Morel (1974) and Zhang et al. (2009) is very low, i.e., approximately 0.00004 m-1.
The manuscript will make mention to the alternative of applying Zhang et al. (2009) instead of Morel (1974).
Comment #15
As for the ac-9 data, estimating an uncertainty of 0.0007 𝑚−1 for 𝑏𝑏𝑝 is wishful thinking. True uncertainties are much larger than that and are the result of a number of factors like those listed above. Can the authors look for a more realistic value based on their own research or in literature?
Reply
The value of 0.0007 m-1 was estimated by Whitmire et al (2007). The actual uncertainty is expected to be higher and dependent on many factors related to processing hypotheses (like the correction for attenuation along the pathlength evoked above). An further uncertainty source is that related to the choice of the “chi” value for converting Beta140 into bb: the standard value used here was 1.08 but Berthon et al. (2007) found that, for the Adriatic Sea, a more appropriate value (based on VSF measurements) was 1.15(+/- 0.04). Also in this case it can be said that more work would be needed.
In the manuscript the uncertainty of 0.0007 𝑚−1 will be stated to be a minimum value, but likely to be much larger due to variability of some of the processing hypothesis.
Comment #16
On the absorption from water samples, the paragraph of lines 378-380 is confusing to me. Probably it needs rephrasing. Maybe the authors mean that the absorption of particulate material between 0.2 and 0.7 micron is negligible with respect to the fraction larger than 0.7 micron? If so, is there some evidence of that in data or literature?
Reply
The text simply states that the absorption budget misses some components that cannot be captured due to difference in pore-size of the filters used produce samples for dissolved and particulate matter absorption analysis. It is also added that likely the missing contribution is not big.
The text will be slightly revised and a citation to Morel and Ahn (J. Mar. Res. 1990) will be added.
Comment #17
CDOM measurements - usage of a 10 cm cuvette inside of a spectrometer is known to be suboptimal in oligotrophic areas like the Mediterranean Sea, even the western basin and in winter. Water is simply too clear to provide a clean spectrum at visible wavelengths. I understand that there is nothing that the authors can do to overcome this issue in case they did not use better suited instruments (like Ultrapath), so at least, an acknowledgement is needed that measurements were performed in suboptimal conditions.
Reply
It will be acknowledged that the accuracy of CDOM in oligotrophic clear water is definitively challenged by the short path-lengths of the laboratory spectrophotometers used for absorbance measurements.
Comment #18
Next type of comments is on the data present in the dataset. It is written (lines 507-511) that basic quality control criteria, like 𝐾𝑑 to be higher than the clear water theoretical value (𝑎𝑤+𝑏𝑏𝑤?), were required for a measurement to be included in the dataset, but I have plotted all 𝐾𝑑 values and I see that many spectra are less than such value, and some even negative, see Fig. 2. I have repeated the analysis for 𝐾𝐿 and 𝐾𝑢 and I have found the same issue (not shown). Same for some absorption data. Regarding 𝑏𝑏, all values are positive, but when removing the water contribution following Zhang et al. (2009), many derived 𝑏𝑏𝑝 values are negative. Although the number of bad spectra may be marginal, this reduces the confidence that this dataset aspires to; so this needs attention before making the public release.
Reply
The two quality indices provided for Kd and bb spectra are obtained from the subtraction of a constant Kw value at 490 nm (0.0212 by Smith and Baker, AO 1981) and a constant bbw value at 488 nm (0.000161 m-1 for salty water by Morel, Optical Aspects of Oceanography, 1974) from the corresponding Kd(490) and bb(488) values. These indices do not have any impact on the data themselves, their negative value simply suggests some caution.
These indices were mostly introduced to support the use of data from highly oligotrophic clear waters by identifying questionable spectra challenged by the water type and the applied measurement method. Any user can ignore, use or re-compute those indices and consequently drop whatever spectrum is later judged ‘bad’. Still, the relatively small number of these spectra challenged by measurements methods applied in a critical measurement condition, cannot become the reason to question the data set.
The values of Kw and bbw applied to determine the quality indices will be provided and some additional detail will be added.
Comment #19
On the phytoplankton absorption data and the chlorophyll concentration, I have plotted one against the other in Fig. 3 at 665 nm, with a highlight on the Eastern Mediterranean data. What I see is that there is the expected tight relationship, but I am concerned about a drop in sensitivity that I see in the lower end. The chlorophyll data has an evident trend towards saturation at about 0.03 – 0.04 𝑚𝑔 𝑚−3, which is too high to resolve the variability in the oligotrophic oceans. I have overplotted the public data by Valente et al. (2022) and, for the few dots in the lower part, I see that the general linear trend is continued. So, authors may try to explain, and if possible, solve this issue.
Reply
As already stated, the highly oligotrophic clear waters of the Eastern Med sea challenge the absorption and scattering methods applied. Clearly the same water type may also affects the accuracy of the derived Chla concentrations. This may certainly explain why a few (4-6 points) in the aph(665) versus Chla plot suggest saturation for the lowest Chla values. This is what the data set can provide for highly oligotrophic clear waters. Still, away from arguing with the Reviewer, his plot including an additional open access data set, only shows 3 points out of thousands below the questioned Chla values.
A statement on the challenging measurement conditions offered by highly oligotrophic clear water conditions of the Med Sea will be restated for Chla data too.
Comment #20
The dataset is optically complete, and therefore something that I am missing in the paper is an 𝑅𝑟𝑠 closure exercise. A high degree of closure helps to increase the confidence on the dataset. In the case that large differences appear, the individual sources have to be inspected. The authors have provided a closure exercise for absorption, which is appreciated, and where significant differences appeared. For 𝑅𝑟𝑠, I have done the closure exercises myself for absorption both from the ac-9 and from the water samples. This is done in Figure 4, for the ac-9 and in Fig. 5, for the water samples. To calculate 𝑅𝑟𝑠 in both cases, Lee et al. (2011) model was used. Considering the radiometric data as reference, results seem to indicate that absorption from ac-9 delivers quite clean data and closure seems very good in general. On the other hand, there are clear differences when absorption from the water samples are used. The plot suggests that absorption from the water samples is much noisier at blue wavelengths and tends to underestimate the real value.
Reply
The Reviewer is acknowledged for his effort to produce closure exercises using the CoASTS-BiOMaP data.
The authors consider this further analysis beyond the objectives of the manuscript.
Comment #21
Final comment is related to the data presentation in the article. It is nice to see the spectra and the ternary plots, and readers can have an idea of the water types that are represented. There are many ways to present the dataset, here just a few that might be of interest to the reader:
- Crossed relationships among IOPs
- 𝑅𝑟𝑠 of chlorophyll, compared to the global relationship
- 𝐾𝑑 of chlorophyll, compared to the relationship by Morel
- Chlorophyll vs. the other two water constituents
- One 𝑅𝑟𝑠 band ratio vs. another one
- TSS vs. 𝑅𝑟𝑠(665)
Reply
Thanks for all the suggestions, clearly feasible, desirable and hopefully interesting. However, the manuscript aims at presenting the data set with some analysis, and not exploiting its content in any possible direction: major extended analyses are not requested for a manuscript submitted to ESSD with the objective to introduce a data set.
- Minor comments
I think it is a requirement that the link to the dataset is shown in the abstract too.
Line 21: “applied equal” → used equally.
Line 39: “benefited of” → benefited from.
Line 54: “moderately” →moderate
Line 91: “attempting”: very vague term. What does it mean in this context, precisely?
Line 96: probably a link to the IOCCG protocol will help here, for those interested.
Table 2: the two-letter country code chosen by the authors looks arbitrary. There is a standardized one named ISO 3166-1 alpha-2, which I advise to follow.
Line 124: talking about in situ vs. laboratory measurements is confusing. Laboratory measurements are made on part of the in situ data. I prefer to talk about field instrumentation vs. laboratory measurement of field samples.
Line 209: no need to say “so called”, as this name is well consolidated and known by everybody.
Line 343. “Wattman” → Whatman
Reply
All relevant corrections will be made. Thanks.
-
AC1: 'Reply on RC1', Giuseppe Zibordi, 27 Jul 2024
-
RC2: 'Comment on essd-2024-240', Michael Twardowski, 29 Jul 2024
The comment was uploaded in the form of a supplement: https://essd.copernicus.org/preprints/essd-2024-240/essd-2024-240-RC2-supplement.pdf
-
AC2: 'Reply on RC2', Giuseppe Zibordi, 01 Aug 2024
Reply to the comments by Mike Twardowski
Overall recommendation: Minor Revision
Comment 1
Public release of CoASTS and BiOMaP is exciting for the ocean color community as these comprehensive data sets have strong value for algorithm development and validation activities. This paper provides an overview of the data sets, methods used, an assessment of errors, and serves as a quick reference guide.
A question is how much of these data have been made publicly available previously in other compendiums such as the Vicente et al. (2019, 2022) and NASA SeaBASS. It is important to specify, moreover to ensure data is not duplicated in any future analyses.
Reply
Some early Lwn and Chla (only) data were submitted to SeaBASS for SeaWIFS validation. However, those data are outside the temporal interval considered for the CoASTS-BiOMaP dataset matter of this manuscript. A few BiOMaP parameters (Rrs, Chla, aph, adg, bbp and Kd) were submitted to MEREMAID for MERIS validation. Some of those data, 33 stations out of the overall 695 performed in the Black Sea, were later included in the dataset assembled by Valente et al. in 2016 and successive versions.
A note on the data included in Valente et al. (2016) and successive versions, will be added in the manuscript.
Comment 2
Data sets extending 2 decades can be relevant to climate change studies. It would be useful to state this and use the term in key words.
Reply
The keyword ‘Climate change’ will be added.
Comment 3
Comments below are intended to compliment, not duplicate, the excellent comments by reviewer J Pitarch (JP), which I have read. I have also read the authors’ replies. On these specific comments and replies, I only comment here where there may perhaps be some disagreement and I have a strong opinion.
The authors discuss some data being consistent with Case 1 or Case 2 waters. Since the authors mention the topic and it is relevant to intended applications, it would be useful to include some estimate of %Case 1 vs Case2 in Tables 1 and 2. While the practical application of Case 1 v Case 2 designations can be ambiguous, there are published quantitative metrics for this that would be very straightforward to implement. Even if approximate, providing these general water type estimates would be useful to many who will use these data.
Reply
The Case-1 / Case-2 index, which is always determined according to Loisel and Morel (1998) during data processing, was not included among the CoASTS BiOMaP quantities because its value is questionable in some marine regions such as the Baltic Sea. Still, any potential user will have the possibility to determine its own index considering the comprehensiveness of the CoASTS-BiOMaP dataset.
Comment 4
Lines 189-208: the extrapolation and derivation of slopes for irradiance profiles is described as taking the log and fitting a line. The most accurate method is fitting the nonlinear exp relationship to the profile data. Derived slopes will be different between the two methods because assumed error distributions get skewed after taking the log, which is inaccurate. The authors appear (understandably) reluctant to revisit processing procedures for these very large data sets, but it would be a small effort to select a representative smattering of profiles from each campaign and apply both methods so an estimate of related biases in derived parameters such as Kd could be given.
Reply
The classical extrapolation method based on the linear fit of log-transformed data was specifically chosen for the CoASTS-BiOMaP data publication to preserve consistency with any other similar dataset. The two extrapolation methods were already and comprehensively investigated in D’Alimonte, D., Shybanov, E. B., Zibordi, G., & Kajiyama, T. (OE, 2013). Not being aware of any previous or successive equivalent investigation, that paper already provides the basis for satisfying curiosities on the method relying on actual exponential extrapolation of profile data (which was specifically applied to some BiOMaP radiometric profiles).
Comment 5
It is also stated that spikes above 3 std due to wave focusing were rejected from radiometer profile data. However, there is nothing wrong with this radiometric data and it should be included in any fit; these spikes can make a significant difference. If a time series was collected at depth we would absolutely want to include the full time series in deriving average
radiometric intensities. These spikes can be orders magnitude greater than average intensity at a particular depth (see Stramski’s work on this). If light is being focused by a wave at any moment during a profile, surrounding data points will be affected by defocusing and thus be deficient in intensity relative to a time series average at that depth. Spikes due to focusing should be included. Again, maybe an analysis can be carried out on a subset of the data to gauge potential associated biases.
Reply
Definitively Dera and Olszewski (1978), Dera and Stramski (1986), Dera et al.(1993) investigated flashes with specific instrumentation allowing to detect their intensity and duration. The same features determined by Dera and colleagues for light flashes in the water, however, cannot equivalently affect Satlantic multispectral data because of the larger diameter of the diffuser and longer integration time with respect to that applied by Dera and colleagues. Because of this, findings from the above studies on downward flashes are not fully applicable to the Ed measurements included in the CoASTS and BiOMaP dataset.
The 3-sigma filter only affects the determination of the slope when the number of points per unit depth is low. This was prevented through the application of the multicast profiling method which increases the number of points to several hundred in the few meter extrapolation interval and consequently increases the precision of the regression (see Zibordi et al. JAOT 2004). In conclusion, the 3-sigma filter allows to detect and remove very few outliers without any appreciable impact on the extrapolation process.
Some more details on the filtering scheme and the number of points per unit depth will be added.
Comment 6
Similarly, some “tuned” automated outlier removal algorithm was apparently used for all the IOP data, removing measurements “exhibiting poor spectral and spatial (i.e., vertical) consistency” but neither the “filtering process,” criteria for “consistency” or “extreme differences,” or the approach to “tuning” are provided. These details are needed for a reader to understand how the data was processed. It is furthermore stated in line 293 that the filtering removed spikes from bubbles and large particles. If effects of bubbles are removed due incomplete air evacuation in water, this is absolutely appropriate and typically only occurs at the very beginning of data records, as the plumbing soon clears of air. However, if the filtering is also removing spikes during profiles of “large particles” and data “exhibiting pronounced differences with respect to those characterizing the mean of profile spectra,” this can be highly problematic. There is no justification for removing spikes in IOPs from large particles. In some particle fields comprised mostly of large detrital aggregates or large colonial plankton, almost all the IOP signal can come from significant spikes associated with numerous large particles. These large particles are inevitably undersampled by the relatively small sample volumes of AC devices and bb sensors, so there is likely residual bias in our measurements relative to the GSD of a satellite unless long in-water time series were recorded, but removing spikes of good data from large particles would certainly exacerbate any bias. Similarly, significant work was done in the 1990’s and 2000’s on the optical properties of thin layers, which can be intense (order of magnitude higher than background) layers of particles less than a meter thick and have strong effects on ocean color (Petrenko et al. 1998; Zaneveld and Pegau 1998). These layers are common throughout the coastal and open ocean. Would your filtering approach remove these effects?
Reply
The relevance of characterizing detrital aggregates or large colonial particles is appreciated. However, this was not one of the objectives considered for CoASTS-BiOMaP measurements. The AC9 measurements were always performed using the inlet filters provided by the manufacturer. This solution destroys any aggregate or colony.
The filtering discussed in the manuscript acts on ‘spikes’: perturbations that generally affect one single value in the profile and always just the ‘a’ or ‘c’ measurements. Spikes often occur in coastal waters and near the surface where, regardless of any effort to get read of the air in the measurement tubes, sometimes occasional bubbles or big particles in the surface layer affect the measurements. Without removing these spikes, the average of the AC9 data collected near the surface and included in the CoASTS-BiOMaP data set would not be representative of the typical water at the station and more than this could exhibit inconsistencies between ‘a’ and ‘c’ values.
Some more details on the measurement methodology will be added in the revised manuscript.
Comment 7
In section 4.2, it is stated that measurements were processed in accordance with guidance from the manufacturer (WET Labs 1996), but this guidance has always been insufficient and antiquated relative to the best methods agreed upon by the community. These best practices have been maintained in published IOCCG Protocols that have recently been updated. Methods here should cite relevant chapters from the Protocols and provide detail on any deviations with related impacts to data quality.
I strongly agree with JP that the a_nw(715) value should be published in these datasets. As JP states, many would argue the method for the scattering correction applied here is not the most accurate. Including a_nw(715) enables the community to apply other published scattering corrections and possibly other scattering corrections developed in the future.
Reply
Full agreement with the Reviewer(s).
The limits of the applied processing will be recognized.
As per the AC9 data at 715 nm, their relevance is appreciated. The plan was to not provide those (ancillary) data. Still an effort will be made to include those data in the CoASTS-BiOMaP dataset if allowed by PANGEA and the review time (a full reprocessing of CoASTS-BiOMaP AC9 data and their successive verification is actually needed).
Comment 8
As JP states, the 0.4 factor for the Hydroscat correction is problematic, but any value is guessing really. There is also no separation of a constant water background in Eq 4, which has always
been inherently problematic. The only thing we can really do however is acknowledge what the realistic errors for this sensor are.
Reply
Also in this case the limits of the applied processing will be recognized.
Comment 9
Was there replication for the TSM measurements? It looks like there was in some cases but was this standard practice? Please clarify.
Reply
Duplicates were always collected and analysed. The average of the two sample values was commonly taken as the final value for each station. Occasionally, one of the two samples was excluded when the duplicates were showing differences tentatively exceeding 20%. Often a look at the filter allowed to identify the problem. If not, TSM values from temporally and spatially close stations were used to subjectively choose what sample to keep. Rarely, AC9 profile data were required to identify the affected sample.
Some more details will provided.
Comment 10
I strongly suggest including histogram plots of c(490 or 532) and SPM as was done for Chl in Fig. 7. These are quick diagnostics for water types for your reader and contribute to the objective of this paper as an overview and guide for the data set.
Reply
The plots will be created and later, if considered relevant, added to the manuscript.
Comment 11
I agree with JP that the inclusion of negative values for parameters such as bbp suggests a lack of rigorous QA/QC. I suggest if you choose to include, add a statement this is a conscious decision and that such negative values “remain within expected errors reported herein” (if you agree with that statement).
Reply
There is no lack of QA/QC. This is witnessed by the continuous efforts put in instrument calibration, verification of performance, inter-comparisons and data curation over almost 3 decades. Definitively negative bbp indicate limits. But measurements are affected by uncertainties and in the case of the HydroScat-6 and AC9 data the impact of uncertainties is enhanced in highly oligotrophic waters. The objective of the flags was to put this forward for critical measurement conditions: those mostly performed in the Eastern Med.
It is quite singular that an almost obvious and needed detail can become matter of major criticism. Actual values close to ‘zero’ of any quantity could be determined as negative due to measurement uncertainties. Nobody stated that negative bbp make sense. They simply indicates the impact of measurement uncertainties, and may provide some indication for their value.
A sentence will be added to state that any negative index is expected to be explained by measurement uncertainties.
Comment 12
Regarding the plots, an IOP plot I find is a strong diagnostic of the quality of a data set while also being a strong proxy for particle composition is bb/b. This parameter incorporates a and c measurements from the AC device as well as bb from the Hydroscat and falls within a relatively narrow range of about 0.04 to 0.3. I would suggest the authors add this plot.
Reply
The scatter plots will be created and, if considered relevant, included in the manuscript. Still, this cannot be done for each center-wavelength.
Comment 13
Moreover, more attention could/should be given to the robustness of the data, QA/QC, and error assessments here. In my opinion, addressing the quality of the data set in a rigorous manner is what elevates this paper to a peer-reviewed contribution as opposed to a simple introduction and guide to these data sets, which could just be posted as a readme online with the data sets.
Reply
QA indicates any action taken to ensure proper execution of measurements. QC is any effort addressed to quality check the quality of data products. This is what was done for each individual quantity included in the dataset as documented in several publications. Still for some uncertainties, the authors could only refer to literature.
When looking at equivalent datasets published in the recent years, the perceived efforts on QA/QC is often insignificant when compared to what implemented over decades for CoASTS and BiOMaP. Definitively, further extended data analysis may strengthen QC. But ESSD papers are specifically intended to support dataset shared with the community. They are not considered research articles (see also the reply to next comment).
Comment 14
Reviewer JP suggests a closure analysis would be a straightforward means of assessing the inherent robustness of the data sets – I thought the same thing in reading the manuscript and strongly agree, this is a super idea. Such an analysis effectively boils down all disparate bias and random errors in the entire data set down to one error number. As such I disagree with the authors’ comment such an assessment is beyond the scope of the paper. Closure results can also be directly compared to a handful of other closure analyses with high quality data such as Pitarch et al. (2016) and Tonizzo et al. (2017) and would provide an immediate comprehensive gauge of quality. But not only did J Pitarch suggest such an analysis, I believe we are all indebted to JP for actually doing the assessment in his review! I was not able to access the figures from his review online, but he states the results appear good. At the very least, the authors should reference JP’s closure assessment in the online ESSD Discussion (I assume these stay online indefinitely?), provide the salient results, and make a statement as to how these results compare with previous closure assessments from the literature. Well done, Jaime, we all thank you, this is an important contribution! If the Editor is looking for Reviewer awards, you get my vote .
Reply
Below is an excerpt from the ESSD web page (https://www.earth-system-science-data.net/about/manuscript_types.html)
Although examples of data outcomes may prove necessary to demonstrate data quality, extensive interpretations of data – i.e. detailed analysis as an author might report in a research article – remain outside the scope of this data journal. ESSD data descriptions should instead highlight and emphasize the quality, usability, and accessibility of the dataset, database, or other data product and should describe extensive carefully prepared metadata and file structures at the data repository.
When the Authors state that a closure investigation is out of the scope of this manuscript (which is not a research article), they are simply following the journal's indications. It is felt that the quality of the data is already proven through the basic elements provided in the manuscript and the papers published in the former decades by the authors.
Comment 14
Section 3 title: suggest “Measurements” should be “Measurements overview”
Reply
The title of the section will be changed.
Comment 15
Section 3.f: I believe a_p, a_ph, and a_dt were measured. This sentence should be reworded to be precise.
Reply
The sentence will be rewritten.
Comment 16
Section 3.i: “Total suspended matter (TSM)” is not precise since a filter was used with some pore size cutoff, thus “total” particles were not assessed. The convention that is often used is “Suspended particulate matter (SPM).
Reply
PANGAEA uses the term Total Suspended Particulate (TSP). For consistency, TSM will be replaced with TSP in the manuscript. SPM will be mentioned.
Citation: https://doi.org/10.5194/essd-2024-240-AC2
-
AC2: 'Reply on RC2', Giuseppe Zibordi, 01 Aug 2024