Third Revision of the Global Surface Seawater Dimethyl Sulfide Climatology (DMS-Rev3)
- 1Indian Institute of Tropical Meteorology, Ministry of Earth Sciences, Pune, India
- 2Institut de Ciències del Mar (CSIC), Barcelona, Catalonia, Spain
- 3Barcelona Supercomputing Center (BSC)
- 4Plymouth Marine Laboratory (PML), Plymouth, UK
- 5Institut Mediterrani d’Estudis Avançats (IMEDEA, UIB-CSIC), Esporles, Balearic Islands, Spain
- 6Institute of Environment and Sustainable Development, Banaras Hindu University, Varanasi, India
- 7College of Life and Environmental Sciences, University of Exeter, Exeter, UK
- 1Indian Institute of Tropical Meteorology, Ministry of Earth Sciences, Pune, India
- 2Institut de Ciències del Mar (CSIC), Barcelona, Catalonia, Spain
- 3Barcelona Supercomputing Center (BSC)
- 4Plymouth Marine Laboratory (PML), Plymouth, UK
- 5Institut Mediterrani d’Estudis Avançats (IMEDEA, UIB-CSIC), Esporles, Balearic Islands, Spain
- 6Institute of Environment and Sustainable Development, Banaras Hindu University, Varanasi, India
- 7College of Life and Environmental Sciences, University of Exeter, Exeter, UK
Abstract. This paper presents an updated estimation of the bottom-up global surface seawater dimethyl sulfide (DMS) climatology. This update, called DMS-Rev3, is the third of its kind and includes five significant changes from the last climatology, ‘L11’ (Lana et al., 2011) that was released about a decade ago. The first change is the inclusion of new observations that have become available over the last decade, creating a database of 872,427 observations leading to a ~18-fold increase in raw data as compared to the last estimation The second is significant improvements in data handling, processing, and filtering, to avoid biases due to different observation frequencies which results from different measurement techniques. Thirdly, we incorporate the dynamic seasonal changes observed in the geographic boundaries of the ocean biogeochemical provinces. The fourth change involves the refinement of the interpolation algorithm used to fill in the missing data. And finally, an upgraded smoothing algorithm based on observed DMS variability length scales (VLS) helps to reproduce a more realistic distribution of the DMS concentration data. The results show that DMS-Rev3 estimates the global annual mean DMS concentration to be ~1.87 nM (2.35 nM without a sea-ice mask), i.e., about 4 % lower than the previous bottom-up ‘L11’ climatology. However, significant regional differences of more than 100 % as compared to L11 are observed. The global sea to air flux of DMS is estimated at ~27 TgS yr−1 which is about 4 % lower than L11, although, like the DMS distribution, large regional differences were observed. The largest changes are observed in high concentration regions such as the polar oceans, although oceanic regions that were under-sampled in the past also show large differences between revisions of the climatology. Finally, DMS-Rev3 reduces the previously observed patchiness in high productivity regions.
- Preprint
(4372 KB) -
Supplement
(5030 KB) - BibTeX
- EndNote
Shrivardhan Hulswar et al.
Status: closed
-
RC1: 'Comment on essd-2021-236', Giuseppe M.R. Manzella, 21 Oct 2021
Comments to:
Third Revision of the Global Surface Seawater Dimethyl Sulfide Climatology (DMSRev3)
by Shrivardhan Hulswar et al.
The paper is well written and easy to follow. The description of mechanisms producing DMS is very clear also for non expert people. The new climatology includes new data and this is a paper added value.
Weak points are related to statistics, number of useful data for climatology, spatial and temporal distribution. These problems that are presented by the authors but not resolved.
The initial data set consisted of 872,427 data points of which only 48,567 are used after post processing. Therefore, the spatial and temporal coverage is worse than that shown in figure 1. Hence a first series of questions: in each month how many data are available in all geographical areas of 1 ° x 1 °? Does each square of 1x1 have a statistically significant number of data points? Are the data for each month and 1x1 areas statistically sufficient or should authors examine them seasonally?
The climatology obtained in data-poor areas with similarity-estimated VLS is not convincing, by taking into account that the phenomena under investigation are occurring at high frequency and varying from place to place.
Amother point that the authors present but do not investigate is related to methodologies and technologies for data collection. Over the years they have changed and so has the data accuracy. In the paper there should be an indication of what is the final accuracy of the climatology in the various areas.
The authors should also provide information on calibration standard if exists and if used in data selection.
The problem of ecological provinces is well posed, and the authors refer to previously published articles. The 'geographically homogeneous' data can be identified with a cluster analysis. In a non-static environment it is possible that geographic homogeneity may vary over time. Authors should discuss this.
The new data included in the paper makes it interesting and publishable after major revision.
Special comments.
Figure 1 should shows the total raw data (1a) and those used for climatology (1b).
An indication of errors or accuracy in the various regions would be desirable
-
RC2: 'Comment on essd-2021-236', Patricia Matrai, 08 Nov 2021
This manuscript presents the 3rd version of the climatology of marine DMS concentration and flux estimates.
Having sampled DMS since the 1980s and provided data to the NOAA PML database over the past 20+ years, it is both rewarding and sobering to see what the spatial and temporal data distribution is; still scarce and often nonexistent for a compound that affects ocean biogeochemistry and ocean-air-clouds-climate processes. Even if DMS's piece of the climate action continues to decrease w/r to the initial CLAW *hypothesis*, DMS continues to provide the main path for CCN growth.
The ms is well written, clear and to the point. The main change is in polar regions; while the sea ice mask results in a better representation of DMS concentrations in the Southern Ocean, the 50% sea ice mask "eliminates" most of the Arctic- sub-Arctic transition from the climatology. In satellite oceanography, one uses a 15% sea ice cut off; even then, one loses all the coastal regions. It would be good to have a bit more discussion of what happens at varying levels of sea ice mask (rather than none, 100%, and the 50% middle).
Please note that I included comments pertinent to the SI file in the main ms file to clarify several figure legends.
-
RC3: 'Comment on essd-2021-236', Murat Aydin, 12 Nov 2021
This is the third revision of global surface seawater DMS climatology. DMS is an important natural source of atmospheric sulfur. The new climatology includes a substantial amount of new data and represents a significant upgrade over the previous version; it is likely to be widely used in various modeling applications. Overall, it is a reasonably well-written manuscript with sound methodology, although some aspects of the data processing were difficult to understand. My concerns, questions, and suggestions are explained in detail below, followed by more specific comments, most of which are fairly trivial.
One overarching concern I have is with regards to the decisions made during the initial stages of the data processing. I understand that subjective decisions are unavoidable when it comes to data filtering, but proper justification is important and the impact of these decisions should be quantified if possible.
First data “clean-up” decision is the omission of the near zero and negative measurements. I find it difficult to believe that the negative values are incorrectly reported numbers (Ln 182). Negative measurements, unless they are extreme negative numbers, commonly occur when working at or below detection limits due to uncertainties in the calibration and/or a system blank that is being subtracted from the signal. I would expect there to be roughly an equal number of small positive measurements that balance out the negatives so that the average is equal to zero within the uncertainties. I do not see a reason to throw out these data as doing so will introduce a positive bias to the data set, especially if data filling schemes are implemented after data removal. In the current case, I realize that a very small fraction of the data is thrown out and the impact of this data clean-up is likely insignificantly small. Still, I do not find the justification convincing. Do the negative numbers cause problems with the data processing that follows? A more elegant solution might be to replace all data below detection limit with zeros, or a very small positive number if zeros cause problems too.
The second item is filtering out high numbers. I was glad to see they did not go with the L11 method of using a 99.9% filter on the entire data set. I think the biggest issue with that approach is it will filter out much more than 0.1% of the data from provinces with high DMS concentrations. The question I have is, is there really a need to filter out any data at all? If you think that the highest 0.1% from each province do not represent surface ocean concentrations, it is necessary to filter out these measurements, I agree. However, if you think these high end measurements might represent blooms (Ln 186), I don’t think it is justified to throw these data out, even if there are repeat samplings of the same bloom. My guess is blooms might be severely underrepresented in these data sets because I suspect they occupy a small fraction of the surface ocean, or is there evidence to the contrary?
The next critical data handling decision is how to combine the newer high resolution mass spectrometer (HRMS) data sets with the older low temporal resolution GC data. Rather, the 1 hr averaging applied to the HRMS data before data unification. This decision should be primarily data driven and the related discussions should include two important criteria: 1) The fidelity of the HRMS data to real variability in the surface ocean. How much temporal averaging to HRMS data is needed to filter out the noise from issues related to these instruments and sampling methods to capture the variability in the surface ocean? 2) The relative representativeness of HRMS vs. GC data. Are the GC data representative of a larger volume of seawater averaged over a longer period than the HRMS? The paper does not discuss these points. Instead, there is an emphasis on the need to balance the influence of historical vs. modern data on the climatology. What is the reason to worry about this balance? This is not adequately explained. It is possible that the quantitative impact of this 1 hr averaging is not all that large, but it would be good to know because HRMS data will become even more dominant in the future.
Is it possible to demonstrate how these decision impacts the final climatology by repeating what was done for testing the sensitivity of results to using dynamic vs. static boundaries for provinces and VLS?
I didn't understand parts of section 2.5. It is not clear how the 1degx1deg product was created and how it was merged with the province averages-based product? Are you doing a 1x1 bin averaging of the underlying raw data? It is stated that the province-based averages and the binned data are superimposed to merge them. I suspect some kind of weighted averaging is done; this needs to be explained. In the Barnes filter step, the ROI is not explained (in one sentence the average distance between a grid point and the data points that influence it can be mentioned), but the sensitivity of results to this parameter is discussed using patchiness as a measure of accuracy. This is problematic because the climatology itself should be the best representation of reality, yet this makes it sound like you already know what reality should look like. VLS, which is actually based on data although it is not clear from which reference(s) the underlying VLS data are taken, is used as ROI in the Barnes filter. I’m not an expert in this kind of thing, but this made sense to me. However, I struggled to reconcile short VLS values with the starting point of this climatology, which is the notion that DMS concentration should more or less be the same within the same biogeochemical provinces. Perhaps a sentence or two are needed to bring everything together here.
A final suggestion: It would be useful if the new DMS flux estimates are used to calculate DMS emissions. The emissions could be presented in tabular form for each province and for 30deg latitude bands, and also as a graph vs. latitude, which could include the L11-based emissions for comparison. This would be helpful for atmospheric scientists studying atmospheric oxidation products of DMS.
The specific comments below are listed in the order they appear in the paper, not in the order of importance.
Specific comments:
Ln 31: Missing full stop.
Ln 91-93: I don’t understand the sentence. Delete “incorporating” perhaps?
Ln 94: No need for comma.
Ln 109: Which database?
LN 139-141: It is stated here that data from highly productive regions lead to patchiness in L11. These sentences, and some others later, seem to imply DMS emissions from blooms should be ignored because the climatology should not look patchy. I agree that patchiness is not good when introduced during data analysis, for example due to introduction of provinces and related assumptions to come up with first guess estimates od DMS concentrations. However, if there is patchiness inherent to the underlying data, don’t we need to accept that as reality?
Ln 178: It is not clear what is meant by data should be within a range of 25%.
Ln 195: I think the 0.001% is a typo. Should be 0.1% for 99.9% threshold.
Ln 209-210: Should be less than or equal to 24 hours?
Ln 211: Fig. 2a does not show any data collected less frequently than 24 hours. If lower frequencies are bundled into the 24 hour bin, this should be noted in the caption.
Ln 214: Fig. 2b shows high-res campaigns are 15% of the total.
Ln 215: No need for “which is based on observations.”
Ln 218-219: I don’t understand what is meant by “matching?” Is 1 hour chosen to have roughly equal contribution from CIMS/MIMS vs. GC data?
Ln 235: “Computing a mean value across the provinces” is confusing to me, I think you mean calculating a mean value that is representative of the entire province?
Ln 280: I think this is the first instance a province acronym being used. It would be easier to read the manuscript if the number associated with the provinces accompanied the acronyms every time an acronym is used because the maps only show the numbers.
Ln 300-304: I suggest seasonal cycle or month-to-month variability instead of annual trend.
Ln 350: Is there really data from under the sea-ice, or is this a result of calculating average values for provinces?
Ln 373-375: Would the use of wind data from more recent periods change the fluxes appreciably? My guess is no. A one-to-one comparison to L11 argument is not that convincing because many aspects of the methods have been changed.
Ln 418-419: Do you mean southwest coast of Africa and the west coast of South America?
Ln 441: You switched what number is reported in the parenthesis, I think.
Ln 583: Fig. S11b color bar label is not in percentage.
Ln 640: Should be Fig. S13.
Ln 646-648: Fig. S14 shows L11 annual average is slightly less than Rev3. Obviously, this depends on what is subtracted from what to make those figures. Caption should include this information.
- AC1: 'Comment on essd-2021-236', Anoop Mahajan, 10 May 2022
Status: closed
-
RC1: 'Comment on essd-2021-236', Giuseppe M.R. Manzella, 21 Oct 2021
Comments to:
Third Revision of the Global Surface Seawater Dimethyl Sulfide Climatology (DMSRev3)
by Shrivardhan Hulswar et al.
The paper is well written and easy to follow. The description of mechanisms producing DMS is very clear also for non expert people. The new climatology includes new data and this is a paper added value.
Weak points are related to statistics, number of useful data for climatology, spatial and temporal distribution. These problems that are presented by the authors but not resolved.
The initial data set consisted of 872,427 data points of which only 48,567 are used after post processing. Therefore, the spatial and temporal coverage is worse than that shown in figure 1. Hence a first series of questions: in each month how many data are available in all geographical areas of 1 ° x 1 °? Does each square of 1x1 have a statistically significant number of data points? Are the data for each month and 1x1 areas statistically sufficient or should authors examine them seasonally?
The climatology obtained in data-poor areas with similarity-estimated VLS is not convincing, by taking into account that the phenomena under investigation are occurring at high frequency and varying from place to place.
Amother point that the authors present but do not investigate is related to methodologies and technologies for data collection. Over the years they have changed and so has the data accuracy. In the paper there should be an indication of what is the final accuracy of the climatology in the various areas.
The authors should also provide information on calibration standard if exists and if used in data selection.
The problem of ecological provinces is well posed, and the authors refer to previously published articles. The 'geographically homogeneous' data can be identified with a cluster analysis. In a non-static environment it is possible that geographic homogeneity may vary over time. Authors should discuss this.
The new data included in the paper makes it interesting and publishable after major revision.
Special comments.
Figure 1 should shows the total raw data (1a) and those used for climatology (1b).
An indication of errors or accuracy in the various regions would be desirable
-
RC2: 'Comment on essd-2021-236', Patricia Matrai, 08 Nov 2021
This manuscript presents the 3rd version of the climatology of marine DMS concentration and flux estimates.
Having sampled DMS since the 1980s and provided data to the NOAA PML database over the past 20+ years, it is both rewarding and sobering to see what the spatial and temporal data distribution is; still scarce and often nonexistent for a compound that affects ocean biogeochemistry and ocean-air-clouds-climate processes. Even if DMS's piece of the climate action continues to decrease w/r to the initial CLAW *hypothesis*, DMS continues to provide the main path for CCN growth.
The ms is well written, clear and to the point. The main change is in polar regions; while the sea ice mask results in a better representation of DMS concentrations in the Southern Ocean, the 50% sea ice mask "eliminates" most of the Arctic- sub-Arctic transition from the climatology. In satellite oceanography, one uses a 15% sea ice cut off; even then, one loses all the coastal regions. It would be good to have a bit more discussion of what happens at varying levels of sea ice mask (rather than none, 100%, and the 50% middle).
Please note that I included comments pertinent to the SI file in the main ms file to clarify several figure legends.
-
RC3: 'Comment on essd-2021-236', Murat Aydin, 12 Nov 2021
This is the third revision of global surface seawater DMS climatology. DMS is an important natural source of atmospheric sulfur. The new climatology includes a substantial amount of new data and represents a significant upgrade over the previous version; it is likely to be widely used in various modeling applications. Overall, it is a reasonably well-written manuscript with sound methodology, although some aspects of the data processing were difficult to understand. My concerns, questions, and suggestions are explained in detail below, followed by more specific comments, most of which are fairly trivial.
One overarching concern I have is with regards to the decisions made during the initial stages of the data processing. I understand that subjective decisions are unavoidable when it comes to data filtering, but proper justification is important and the impact of these decisions should be quantified if possible.
First data “clean-up” decision is the omission of the near zero and negative measurements. I find it difficult to believe that the negative values are incorrectly reported numbers (Ln 182). Negative measurements, unless they are extreme negative numbers, commonly occur when working at or below detection limits due to uncertainties in the calibration and/or a system blank that is being subtracted from the signal. I would expect there to be roughly an equal number of small positive measurements that balance out the negatives so that the average is equal to zero within the uncertainties. I do not see a reason to throw out these data as doing so will introduce a positive bias to the data set, especially if data filling schemes are implemented after data removal. In the current case, I realize that a very small fraction of the data is thrown out and the impact of this data clean-up is likely insignificantly small. Still, I do not find the justification convincing. Do the negative numbers cause problems with the data processing that follows? A more elegant solution might be to replace all data below detection limit with zeros, or a very small positive number if zeros cause problems too.
The second item is filtering out high numbers. I was glad to see they did not go with the L11 method of using a 99.9% filter on the entire data set. I think the biggest issue with that approach is it will filter out much more than 0.1% of the data from provinces with high DMS concentrations. The question I have is, is there really a need to filter out any data at all? If you think that the highest 0.1% from each province do not represent surface ocean concentrations, it is necessary to filter out these measurements, I agree. However, if you think these high end measurements might represent blooms (Ln 186), I don’t think it is justified to throw these data out, even if there are repeat samplings of the same bloom. My guess is blooms might be severely underrepresented in these data sets because I suspect they occupy a small fraction of the surface ocean, or is there evidence to the contrary?
The next critical data handling decision is how to combine the newer high resolution mass spectrometer (HRMS) data sets with the older low temporal resolution GC data. Rather, the 1 hr averaging applied to the HRMS data before data unification. This decision should be primarily data driven and the related discussions should include two important criteria: 1) The fidelity of the HRMS data to real variability in the surface ocean. How much temporal averaging to HRMS data is needed to filter out the noise from issues related to these instruments and sampling methods to capture the variability in the surface ocean? 2) The relative representativeness of HRMS vs. GC data. Are the GC data representative of a larger volume of seawater averaged over a longer period than the HRMS? The paper does not discuss these points. Instead, there is an emphasis on the need to balance the influence of historical vs. modern data on the climatology. What is the reason to worry about this balance? This is not adequately explained. It is possible that the quantitative impact of this 1 hr averaging is not all that large, but it would be good to know because HRMS data will become even more dominant in the future.
Is it possible to demonstrate how these decision impacts the final climatology by repeating what was done for testing the sensitivity of results to using dynamic vs. static boundaries for provinces and VLS?
I didn't understand parts of section 2.5. It is not clear how the 1degx1deg product was created and how it was merged with the province averages-based product? Are you doing a 1x1 bin averaging of the underlying raw data? It is stated that the province-based averages and the binned data are superimposed to merge them. I suspect some kind of weighted averaging is done; this needs to be explained. In the Barnes filter step, the ROI is not explained (in one sentence the average distance between a grid point and the data points that influence it can be mentioned), but the sensitivity of results to this parameter is discussed using patchiness as a measure of accuracy. This is problematic because the climatology itself should be the best representation of reality, yet this makes it sound like you already know what reality should look like. VLS, which is actually based on data although it is not clear from which reference(s) the underlying VLS data are taken, is used as ROI in the Barnes filter. I’m not an expert in this kind of thing, but this made sense to me. However, I struggled to reconcile short VLS values with the starting point of this climatology, which is the notion that DMS concentration should more or less be the same within the same biogeochemical provinces. Perhaps a sentence or two are needed to bring everything together here.
A final suggestion: It would be useful if the new DMS flux estimates are used to calculate DMS emissions. The emissions could be presented in tabular form for each province and for 30deg latitude bands, and also as a graph vs. latitude, which could include the L11-based emissions for comparison. This would be helpful for atmospheric scientists studying atmospheric oxidation products of DMS.
The specific comments below are listed in the order they appear in the paper, not in the order of importance.
Specific comments:
Ln 31: Missing full stop.
Ln 91-93: I don’t understand the sentence. Delete “incorporating” perhaps?
Ln 94: No need for comma.
Ln 109: Which database?
LN 139-141: It is stated here that data from highly productive regions lead to patchiness in L11. These sentences, and some others later, seem to imply DMS emissions from blooms should be ignored because the climatology should not look patchy. I agree that patchiness is not good when introduced during data analysis, for example due to introduction of provinces and related assumptions to come up with first guess estimates od DMS concentrations. However, if there is patchiness inherent to the underlying data, don’t we need to accept that as reality?
Ln 178: It is not clear what is meant by data should be within a range of 25%.
Ln 195: I think the 0.001% is a typo. Should be 0.1% for 99.9% threshold.
Ln 209-210: Should be less than or equal to 24 hours?
Ln 211: Fig. 2a does not show any data collected less frequently than 24 hours. If lower frequencies are bundled into the 24 hour bin, this should be noted in the caption.
Ln 214: Fig. 2b shows high-res campaigns are 15% of the total.
Ln 215: No need for “which is based on observations.”
Ln 218-219: I don’t understand what is meant by “matching?” Is 1 hour chosen to have roughly equal contribution from CIMS/MIMS vs. GC data?
Ln 235: “Computing a mean value across the provinces” is confusing to me, I think you mean calculating a mean value that is representative of the entire province?
Ln 280: I think this is the first instance a province acronym being used. It would be easier to read the manuscript if the number associated with the provinces accompanied the acronyms every time an acronym is used because the maps only show the numbers.
Ln 300-304: I suggest seasonal cycle or month-to-month variability instead of annual trend.
Ln 350: Is there really data from under the sea-ice, or is this a result of calculating average values for provinces?
Ln 373-375: Would the use of wind data from more recent periods change the fluxes appreciably? My guess is no. A one-to-one comparison to L11 argument is not that convincing because many aspects of the methods have been changed.
Ln 418-419: Do you mean southwest coast of Africa and the west coast of South America?
Ln 441: You switched what number is reported in the parenthesis, I think.
Ln 583: Fig. S11b color bar label is not in percentage.
Ln 640: Should be Fig. S13.
Ln 646-648: Fig. S14 shows L11 annual average is slightly less than Rev3. Obviously, this depends on what is subtracted from what to make those figures. Caption should include this information.
- AC1: 'Comment on essd-2021-236', Anoop Mahajan, 10 May 2022
Shrivardhan Hulswar et al.
Data sets
Third Revision of the Global Surface Seawater Dimethyl Sulfide Climatology (DMS-Rev3) Anoop Mahajan https://doi.org/10.17632/hyn62spny2.1
Model code and software
Third Revision of the Global Surface Seawater Dimethyl Sulfide Climatology (DMS-Rev3) Anoop Mahajan https://doi.org/10.17632/hyn62spny2.1
Shrivardhan Hulswar et al.
Viewed
HTML | XML | Total | Supplement | BibTeX | EndNote | |
---|---|---|---|---|---|---|
717 | 188 | 26 | 931 | 72 | 16 | 19 |
- HTML: 717
- PDF: 188
- XML: 26
- Total: 931
- Supplement: 72
- BibTeX: 16
- EndNote: 19
Viewed (geographical distribution)
Country | # | Views | % |
---|
Total: | 0 |
HTML: | 0 |
PDF: | 0 |
XML: | 0 |
- 1