Investigating interhemispheric precipitation changes over the past millennium¶
Overview¶
This CookBook demonstrates how to compare paleoclimate model output and proxy observations using EOF to identify large-scale spatio-temporal patterns in the data. It is inspired from a study by Steinman et al. (2022) although it reuses different datasets.
Prerequisites¶
Concepts | Importance | Notes |
---|---|---|
Intro to Cartopy | Necessary | |
Intro to Matplotlib | Necessary | |
Intro to Pandas | Necessary | |
Understanding of NetCDF | Necessary | |
Using xarray | Necessary | Familiarity with understanding opening multiple files and merging |
EOF (PCA) Analysis - See Chapter 12 of this book | Helpful | Familiarity with the concepts is helpful for interpretation of the results |
eofs package | Helpful | A good introduction on the package can be found in this notebook |
Using Pyleoclim for Paleoclimate Data | Helpful | |
SPARQL | Familiarity | Query language for graph database |
- Time to learn: 40 min.
Imports¶
#To deal with model data
import s3fs
import fsspec
import xarray as xr
import glob
#To deal with proxy data
import pandas as pd
import numpy as np
import json
import requests
import pandas as pd
import io
import ast
#To deal with analysis
import pyleoclim as pyleo
from eofs.xarray import Eof
#Plotting and mapping
import cartopy.crs as ccrs
import matplotlib.pyplot as plt
import nc_time_axis
from matplotlib import gridspec
from matplotlib.colors import Normalize
PCA analysis on proxy observations¶
After looking at the model data, let’s have a look at the proxy datasets.
Query to remote database¶
The first step is the query the LinkedEarth Graph Database for relevant datasets for comparison with the model.
The database uses the SPARQL language for queries. We are filtering the database for the following criteria:
- Datasets bounded by 27°S-27°N and 70°W-150°W
- Datasets from the Pages2k, Iso2k, CoralHydro2k and SISAL working groups. These working groups identified archived datasets that were sensitive to temperature and the isotopic composition of precipication (precipitation O) and curated them for use in a standardized database.
- Timeseries within these datasets representing precipitation.
We asked for the following information back:
- The name of the dataset
- Geographical Location of the record expressed in latitude and longitude
- The type of archive (e.g., speleothem, Lake sediment) the measurements were made on
- The name of the variable
- The values and units of the measurements
- The time information (values and units) associated with the variable of interest.
The following cell points to the query API and creates the query itself.
url = 'https://linkedearth.graphdb.mint.isi.edu/repositories/LiPDVerse-dynamic'
query = """PREFIX le: <http://linked.earth/ontology#>
PREFIX wgs84: <http://www.w3.org/2003/01/geo/wgs84_pos#>
PREFIX rdfs: <http://www.w3.org/2000/01/rdf-schema#>
SELECT distinct?varID ?dataSetName ?lat ?lon ?varname ?interpLabel ?val ?varunits ?timevarname ?timeval ?timeunits ?archiveType where{
?ds a le:Dataset .
?ds le:hasName ?dataSetName .
OPTIONAL{?ds le:hasArchiveType ?archiveTypeObj .
?archiveTypeObj rdfs:label ?archiveType .}
?ds le:hasLocation ?loc .
?loc wgs84:lat ?lat .
FILTER(?lat<26 && ?lat>-26)
?loc wgs84:long ?lon .
FILTER(?lon<-70 && ?lon>-150)
?ds le:hasPaleoData ?data .
?data le:hasMeasurementTable ?table .
?table le:hasVariable ?var .
?var le:hasName ?varname .
VALUES ?varname {"d18O"} .
?var le:partOfCompilation ?comp .
?comp le:hasName ?compName .
VALUES ?compName {"iso2k" "Pages2kTemperature" "CoralHydro2k" "SISAL-LiPD"} .
?var le:hasInterpretation ?interp .
?interp le:hasVariable ?interpVar .
?interpVar rdfs:label ?interpLabel .
FILTER (REGEX(?interpLabel, "precipitation.*", "i"))
?var le:hasVariableId ?varID .
?var le:hasValues ?val .
OPTIONAL{?var le:hasUnits ?varunitsObj .
?varunitsObj rdfs:label ?varunits .}
?table le:hasVariable ?timevar .
?timevar le:hasName ?timevarname .
VALUES ?timevarname {"year" "age"} .
?timevar le:hasValues ?timeval .
OPTIONAL{?timevar le:hasUnits ?timeunitsObj .
?timeunitsObj rdfs:label ?timeunits .}
}"""
The following cell sends the query to the database and returns the results in a Pandas Dataframe.
response = requests.post(url, data = {'query': query})
data = io.StringIO(response.text)
df_res = pd.read_csv(data, sep=",")
df_res['val']=df_res['val'].apply(lambda row : json.loads(row) if isinstance(row, str) else row)
df_res['timeval']=df_res['timeval'].apply(lambda row : json.loads(row) if isinstance(row, str) else row)
df_res.head()
We have retrieved the following number of proxy records:
len(df_res)
752
Make sure we have unique timeseries (some may be found across compilations):
df = df_res[~df_res['varID'].duplicated()]
len(df)
45
The first step is to make sure that everything is on the same representation of the time axis. Year is considered prograde while age is considered retrograde:
df['timevarname'].unique()
array(['year', 'age'], dtype=object)
Since we have records expressed in both year and age, let’s convert everything to year. First let’s have a look at the units:
df['timeunits'].unique()
array(['yr AD', 'yr BP'], dtype=object)
The units for age are expressed in BP (before present), if we assume the present to be 1950 by convention, then we can transform:
df['timeval'] = df['timeval'].apply(np.array)
def adjust_timeval(row):
if row['timevarname'] == 'age':
return 1950 - row['timeval']
else:
return row['timeval']
# Apply the function across the DataFrame rows
df['timeval'] = df.apply(adjust_timeval, axis=1)
It is obvious that some of the timeseries do not have correct time information (e.g., row 2). Let’s filter the dataframe to make sure that the time values are within 0-2000 and that there is at least 1500 years of record:
def range_within_limits(array, lower = 0, upper = 2000, threshold = 1500):
filtered_values = array[(array >= lower) & (array <= upper)]
if filtered_values.size > 0: # Check if there are any values within the range
return np.ptp(filtered_values) >= threshold # np.ptp() returns the range of values
return False # If no values within the range, filter out the row
# Apply the function to filter the DataFrame
filtered_df = df[df['timeval'].apply(range_within_limits)]
We are now left with:
len(filtered_df)
18
Let’s also make sure that the records are long enough (i.e., more than 1500 years long):
def array_range_exceeds(array, threshold=1500):
return np.max(array) - np.min(array) > threshold
filt_df = filtered_df[filtered_df['timeval'].apply(array_range_exceeds)]
This leaves us with the following number of datasets:
len(filt_df)
17
Let’s filter for records with a resolution finer or equal to 60 years:
def min_resolution(array, min_res=60):
if len(array) > 1: # Ensure there are at least two points to calculate a difference
# Calculate differences between consecutive elements
differences = np.mean(np.diff(array))
# Check if the minimum difference is at least 50
return abs(differences) <= min_res
return False # If less than two elements, can't compute difference
# Apply the function and filter the DataFrame
filtered_df2 = filt_df[filt_df['timeval'].apply(min_resolution)]
This leaves us with the following number of datasets:
len(filtered_df2)
14
ts_list = []
for _, row in filtered_df2.iterrows():
ts_list.append(pyleo.GeoSeries(time=row['timeval'],value=row['val'],
time_name='year',value_name=row['varname'],
time_unit='CE', value_unit=row['varunits'],
lat = row['lat'], lon = row['lon'],
archiveType = row['archiveType'], verbose = False,
label=row['dataSetName']+'_'+row['varname']))
#print(row['timeval'])
Now let’s use a MultipleGeoSeries object for visualization and analysis:
mgs = pyleo.MultipleGeoSeries(ts_list, label='HydroAm2k', time_unit='year CE')
Let’s first map the location of the records by the type of archive:
mgs.map()
(<Figure size 1600x600 with 2 Axes>,
{'map': <GeoAxes: xlabel='lon', ylabel='lat'>, 'leg': <Axes: >})

Let’s have a look at the records, sliced for the 0-2000 period
fig, ax = mgs.sel(time=slice(0,2000)).stackplot(v_shift_factor=1.2)
plt.show(fig)

Run PCA Analysis¶
Let’s place them on a common time axis for analysis and standardize:
mgs_common = mgs.sel(time=slice(850,2000)).common_time().standardize()
pca = mgs_common.pca()
Let’s have a look at the screeplot:
pca.screeplot()
The provided eigenvalue array has only one dimension. UQ defaults to NB82
(<Figure size 600x400 with 1 Axes>,
<Axes: title={'center': 'HydroAm2k PCA eigenvalues'}, xlabel='Mode index $i$', ylabel='$\\lambda_i$'>)

As is nearly always the case with geophysical timeseries, the first few of eigenvalues trully overwhelm the rest. In this case, let’s have a look at the first three.
pca.modeplot()
(<Figure size 800x800 with 4 Axes>,
{'pc': <Axes: xlabel='Time [year CE]', ylabel='$PC_1$'>,
'psd': <Axes: xlabel='Period [year]', ylabel='PSD'>,
'map': {'cb': <Axes: ylabel='EOF'>,
'map': <GeoAxes: xlabel='lon', ylabel='lat'>}})

Let’s have a look at the second mode:
pca.modeplot(index=1)
(<Figure size 800x800 with 4 Axes>,
{'pc': <Axes: xlabel='Time [year CE]', ylabel='$PC_2$'>,
'psd': <Axes: xlabel='Period [year]', ylabel='PSD'>,
'map': {'cb': <Axes: ylabel='EOF'>,
'map': <GeoAxes: xlabel='lon', ylabel='lat'>}})

Finally, let’s have a look at the third mode:
pca.modeplot(index=2)
(<Figure size 800x800 with 4 Axes>,
{'pc': <Axes: xlabel='Time [year CE]', ylabel='$PC_3$'>,
'psd': <Axes: xlabel='Period [year]', ylabel='PSD'>,
'map': {'cb': <Axes: ylabel='EOF'>,
'map': <GeoAxes: xlabel='lon', ylabel='lat'>}})

As you can see, the first three modes explain 60% of the variance.
PCA analysis on the CESM Last Millennium Ensemble¶
The first step is the calculate precipitation for the all forcings simulation. The following section demonstrates how to get the data from JetStream2 and pre-process each file to save the needed variable and place them into a new xarray.Dataset
. This process can be time-consuming.
Get the CESM Last Millennium Ensemble data from JetStream2¶
Let’s open the needed files for this analysis, which consists of various precipitation isotopes. All data have been made available on NSF JetStream2.
URL = 'https://js2.jetstream-cloud.org:8001/' #Locate and read a file
path = f'pythia/cesmLME' # specify data location
fs = fsspec.filesystem("s3", anon=True, client_kwargs=dict(endpoint_url=URL))
pattern = f's3://{path}/*.nc'
files = sorted(fs.glob(pattern))
Let’s open relevant files for the all forcing simulations (‘LME.002’) for the 850-1850 period:
base_name = 'pythia/cesmLME/b.ie12.B1850C5CN.f19_g16.LME.002.cam.h0.'
time_period = '085001-184912'
names = [name for name in files if base_name in name and time_period in name]
names
['pythia/cesmLME/b.ie12.B1850C5CN.f19_g16.LME.002.cam.h0.PRECRC_H216Or.085001-184912.nc',
'pythia/cesmLME/b.ie12.B1850C5CN.f19_g16.LME.002.cam.h0.PRECRC_H218Or.085001-184912.nc',
'pythia/cesmLME/b.ie12.B1850C5CN.f19_g16.LME.002.cam.h0.PRECRL_H216OR.085001-184912.nc',
'pythia/cesmLME/b.ie12.B1850C5CN.f19_g16.LME.002.cam.h0.PRECRL_H218OR.085001-184912.nc',
'pythia/cesmLME/b.ie12.B1850C5CN.f19_g16.LME.002.cam.h0.PRECSC_H216Os.085001-184912.nc',
'pythia/cesmLME/b.ie12.B1850C5CN.f19_g16.LME.002.cam.h0.PRECSC_H218Os.085001-184912.nc',
'pythia/cesmLME/b.ie12.B1850C5CN.f19_g16.LME.002.cam.h0.PRECSL_H216OS.085001-184912.nc',
'pythia/cesmLME/b.ie12.B1850C5CN.f19_g16.LME.002.cam.h0.PRECSL_H218OS.085001-184912.nc']
fileset = [fs.open(file) for file in names]
Next, let’s open these datasets and extract the needed variables into another xarray.Dataset
.
Warning
Note that this cell may take some time to run! On a 2024 MacBook pro with an M3 chip, it took about 7minutes.%%time
for idx,item in enumerate(fileset):
ds_u = xr.open_dataset(item)
var_name = names[idx].split('.')[-3] #This uses the file name to obtain the variable name.
da = ds_u[var_name]
try:
ds[var_name]= da
except:
ds = da.to_dataset()
ds.attrs = ds_u.attrs
ds_u.close()
da.close()
CPU times: user 13.6 ms, sys: 3.51 ms, total: 17.1 ms
Wall time: 117 ms
---------------------------------------------------------------------------
ValueError Traceback (most recent call last)
Cell In[31], line 1
----> 1 get_ipython().run_cell_magic('time', '', "for idx,item in enumerate(fileset):\n ds_u = xr.open_dataset(item)\n var_name = names[idx].split('.')[-3] #This uses the file name to obtain the variable name. \n da = ds_u[var_name]\n try:\n ds[var_name]= da\n except:\n ds = da.to_dataset()\n ds.attrs = ds_u.attrs \n ds_u.close()\n da.close()\n")
File ~/micromamba/envs/paleoPCA-Cookbook/lib/python3.11/site-packages/IPython/core/interactiveshell.py:2565, in InteractiveShell.run_cell_magic(self, magic_name, line, cell)
2563 with self.builtin_trap:
2564 args = (magic_arg_s, cell)
-> 2565 result = fn(*args, **kwargs)
2567 # The code below prevents the output from being displayed
2568 # when using magics with decorator @output_can_be_silenced
2569 # when the last Python token in the expression is a ';'.
2570 if getattr(fn, magic.MAGIC_OUTPUT_CAN_BE_SILENCED, False):
File ~/micromamba/envs/paleoPCA-Cookbook/lib/python3.11/site-packages/IPython/core/magics/execution.py:1470, in ExecutionMagics.time(self, line, cell, local_ns)
1468 if interrupt_occured:
1469 if exit_on_interrupt and captured_exception:
-> 1470 raise captured_exception
1471 return
1472 return out
File ~/micromamba/envs/paleoPCA-Cookbook/lib/python3.11/site-packages/IPython/core/magics/execution.py:1434, in ExecutionMagics.time(self, line, cell, local_ns)
1432 st = clock2()
1433 try:
-> 1434 exec(code, glob, local_ns)
1435 out = None
1436 # multi-line %%time case
File <timed exec>:2
File ~/micromamba/envs/paleoPCA-Cookbook/lib/python3.11/site-packages/xarray/backends/api.py:696, in open_dataset(filename_or_obj, engine, chunks, cache, decode_cf, mask_and_scale, decode_times, decode_timedelta, use_cftime, concat_characters, decode_coords, drop_variables, create_default_indexes, inline_array, chunked_array_type, from_array_kwargs, backend_kwargs, **kwargs)
693 kwargs.update(backend_kwargs)
695 if engine is None:
--> 696 engine = plugins.guess_engine(filename_or_obj)
698 if from_array_kwargs is None:
699 from_array_kwargs = {}
File ~/micromamba/envs/paleoPCA-Cookbook/lib/python3.11/site-packages/xarray/backends/plugins.py:194, in guess_engine(store_spec)
186 else:
187 error_msg = (
188 "found the following matches with the input file in xarray's IO "
189 f"backends: {compatible_engines}. But their dependencies may not be installed, see:\n"
190 "https://docs.xarray.dev/en/stable/user-guide/io.html \n"
191 "https://docs.xarray.dev/en/stable/getting-started-guide/installing.html"
192 )
--> 194 raise ValueError(error_msg)
ValueError: did not find a match in any of xarray's currently installed IO backends ['h5netcdf', 'scipy']. Consider explicitly selecting one of the installed engines via the ``engine`` parameter, or installing additional IO dependencies, see:
https://docs.xarray.dev/en/stable/getting-started-guide/installing.html
https://docs.xarray.dev/en/stable/user-guide/io.html
And we’re done! Let’s have a look at the data we will be working with:
ds
Select the tropical Central and South America region¶
Let’s look at the region bounded by 27°S-27°N and 70°W-150°W:
ds_geo_all = ds.sel(lat=slice(-27,27), lon=slice(250,330))
Since the loading and pre-processing of the files take a long time, let’s save a version of this dataset in netCDF for further use:
#ds_geo.to_netcdf(path='../data/LME.002.cam.h0.precip_iso.085001-184912.nc')
For model-data comparison, it might be useful to resample the model data on the proxy scales:
print("The minimum time is: "+str(np.min(mgs_common.series_list[0].time)))
print("The maximum time is: "+str(np.max(mgs_common.series_list[0].time)))
print("The resolution is: "+str(np.mean(np.diff(mgs_common.series_list[0].time))))
ds_geo_time = ds_geo_all.sel(time=slice("0910-01-01 00:00:00" ,"1642-12-01 00:00:00"))
And resample to the proxy resolution (~20yr as calculated above):
ds_geo = ds_geo_time.resample(time='20A').mean()
ds_geo
Calculate Precipitation O¶
%%time
p16O = ds_geo['PRECRC_H216Or'] + ds_geo['PRECSC_H216Os'] + ds_geo['PRECRL_H216OR'] + ds_geo['PRECSL_H216OS']
p18O = ds_geo['PRECRC_H218Or'] + ds_geo['PRECSC_H218Os'] + ds_geo['PRECRL_H218OR'] + ds_geo['PRECSL_H218OS']
p16O = p16O.where(p16O > 1e-18, 1e-18)
p18O = p18O.where(p18O > 1e-18, 1e-18)
d18Op = (p18O / p16O - 1)*1000
Run PCA analysis¶
Let’s first standardize the data:
d18Oa = (d18Op - d18Op.mean(dim='time'))/d18Op.std(dim='time')
Create an EOF solver to do the EOF analysis.
solver = Eof(d18Oa, weights=None)
Retrieve the leading EOF, expressed as the covariance between the leading PC time series and the input d18O anomalies at each grid point.
eof1 = solver.eofsAsCovariance(neofs=3)
Plot the leading EOF expressed covariance
clevs = np.linspace(-1, 1, 20)
proj = ccrs.PlateCarree(central_longitude=290)
fig, ax = plt.subplots(figsize=[10,4], subplot_kw=dict(projection=proj))
ax.coastlines()
eof1[0].plot.contourf(ax=ax, levels = clevs, cmap=plt.cm.RdBu_r,
transform=ccrs.PlateCarree(), add_colorbar=True)
fig.axes[1].set_ylabel('')
fig.axes[1].set_yticks(np.arange(-1,1.2,0.2))
ax.set_title('EOF1 expressed as covariance', fontsize=16)
plt.show()
Let’s have a look at the first three PCs:
pcs = solver.pcs(npcs=3, pcscaling=1)
fig, ax = plt.subplots(figsize=[20,4])
pcs[:, 0].plot(ax=ax, linewidth=1)
ax = plt.gca()
ax.axhline(0, color='k')
ax.set_ylim(-3, 3)
ax.set_xlabel('Year')
ax.set_ylabel('Normalized Units')
ax.set_title('PC1 Time Series', fontsize=16)
Model-Data Comparison¶
Let’s map the first EOF for the model and data and the first PC.
# Create a figure
fig = plt.figure(figsize=[20,8])
# Define the GridSpec
gs = gridspec.GridSpec(1, 2, figure=fig)
# Add a geographic map in the first subplot using Cartopy
ax1 = fig.add_subplot(gs[0, 0], projection=ccrs.PlateCarree(central_longitude=290))
ax1.coastlines() # Add coastlines to the map
# Plot the model results
norm = Normalize(vmin=-1, vmax=1)
eof1[0].plot.contourf(ax=ax1, levels = clevs, cmap=plt.cm.RdBu_r,
transform=ccrs.PlateCarree(), add_colorbar=True, norm=norm)
ax1.set_title('EOF1 expressed as covariance', fontsize=16)
fig.axes[1].set_ylabel('')
fig.axes[1].set_yticks(np.arange(-1,1.2,0.2))
#Now let's scatter the proxy data
EOF = pca.eigvecs[:, 0]
ax1.scatter(filtered_df2['lon'],filtered_df2['lat'], c =EOF, cmap=plt.cm.RdBu_r, transform=ccrs.PlateCarree(), norm=norm, s=400, edgecolor='k', linewidth=3)
## Let's plot the PCS!
PC = pca.pcs[:, 0]
ax2 = fig.add_subplot(gs[0, 1:],)
time_model = np.arange(910,1660,20)
ts1 = pyleo.Series(time = time_model, value = pcs[:, 0], time_name = 'Years', time_unit = 'CE', value_name='PC1', label = 'CESM-LEM',verbose=False)
ts2 = pyleo.Series(time = mgs_common.series_list[0].time, value = PC, time_name = 'Years', time_unit = 'CE', value_name='PC1', label = 'Proxy Data', verbose = False)
ts1.plot(ax=ax2, legend = True)
ts2.plot(ax=ax2, legend = True)
ax2.set_ylim([-1,1])
ax2.legend()
# Layout adjustments and display the figure
plt.tight_layout()
plt.show()
Summary¶
In this Cookbook, you learned to run PCA analysis on gridded and point datasets to allow for model-data comparison. Although this was focued on the paleoclimate domain, the technique is broadly applicable to the instrumental era.
Resources and references¶
Model Output¶
- CESM LME: Otto-Bliesner, B.L., E.C. Brady, J. Fasullo, A. Jahn, L. Landrum, S. Stevenson, N. Rosenbloom, A. Mai, G. Strand. Climate Variability and Change since 850 C.E. : An Ensemble Approach with the Community Earth System Model (CESM), Bulletin of the American Meteorological Society, 735-754 (May 2016 issue)
Proxy Compilations¶
Iso2k: Konecky, B. L., McKay, N. P., Churakova (Sidorova), O. V., Comas-Bru, L., Dassié, E. P., DeLong, K. L., Falster, G. M., Fischer, M. J., Jones, M. D., Jonkers, L., Kaufman, D. S., Leduc, G., Managave, S. R., Martrat, B., Opel, T., Orsi, A. J., Partin, J. W., Sayani, H. R., Thomas, E. K., Thompson, D. M., Tyler, J. J., Abram, N. J., Atwood, A. R., Cartapanis, O., Conroy, J. L., Curran, M. A., Dee, S. G., Deininger, M., Divine, D. V., Kern, Z., Porter, T. J., Stevenson, S. L., von Gunten, L., and Iso2k Project Members: The Iso2k database: a global compilation of paleo-δ18O and δ2H records to aid understanding of Common Era climate, Earth Syst. Sci. Data, 12, 2261–2288, Konecky et al. (2020), 2020.
PAGES2kTemperature: PAGES2k Consortium. A global multiproxy database for temperature reconstructions of the Common Era. Sci. Data 4:170088 doi: 10.1038/sdata.2017.88 (2017).
CoralHydro2k: Walter, R. M., Sayani, H. R., Felis, T., Cobb, K. M., Abram, N. J., Arzey, A. K., Atwood, A. R., Brenner, L. D., Dassié, É. P., DeLong, K. L., Ellis, B., Emile-Geay, J., Fischer, M. J., Goodkin, N. F., Hargreaves, J. A., Kilbourne, K. H., Krawczyk, H., McKay, N. P., Moore, A. L., Murty, S. A., Ong, M. R., Ramos, R. D., Reed, E. V., Samanta, D., Sanchez, S. C., Zinke, J., and the PAGES CoralHydro2k Project Members: The CoralHydro2k database: a global, actively curated compilation of coral δ18O and Sr ∕ Ca proxy records of tropical ocean hydrology and temperature for the Common Era, Earth Syst. Sci. Data, 15, 2081–2116, Walter et al. (2023), 2023.
SISAL: Comas-Bru, L., Rehfeld, K., Roesch, C., Amirnezhad-Mozhdehi, S., Harrison, S. P., Atsawawaranunt, K., Ahmad, S. M., Brahim, Y. A., Baker, A., Bosomworth, M., Breitenbach, S. F. M., Burstyn, Y., Columbu, A., Deininger, M., Demény, A., Dixon, B., Fohlmeister, J., Hatvani, I. G., Hu, J., Kaushal, N., Kern, Z., Labuhn, I., Lechleitner, F. A., Lorrey, A., Martrat, B., Novello, V. F., Oster, J., Pérez-Mejías, C., Scholz, D., Scroxton, N., Sinha, N., Ward, B. M., Warken, S., Zhang, H., and SISAL Working Group members: SISALv2: a comprehensive speleothem isotope database with multiple age–depth models, Earth Syst. Sci. Data, 12, 2579–2606, Comas-Bru et al. (2020), 2020.
Software¶
xarray: Hoyer, S., & Joseph, H. (2017). xarray: N-D labeled Arrays and Datasets in Python. Journal of Open Research Software, 5(1). Hoyer & Hamman (2017)
Khider, D., Emile-Geay, J., Zhu, F., James, A., Landers, J., Ratnakar, V., & Gil, Y. (2022). Pyleoclim: Paleoclimate timeseries analysis and visualization with Python. Paleoceanography and Paleoclimatology, 37, e2022PA004509. Khider et al. (2022)
Khider, D., Emile-Geay, J., Zhu, F., James, A., Landers, J., Kwan, M., Athreya, P., McGibbon, R., & Voirol, L. (2024). Pyleoclim: A Python package for the analysis and visualization of paleoclimate data (Version v1.0.0) [Computer software]. Khider et al. (2025)
- eofs: Dawson, A. (2016) ‘eofs: A Library for EOF Analysis of Meteorological, Oceanographic, and Climate Data’, Journal of Open Research Software, 4(1), p. e14. Available at: Dawson (2016).
- Steinman, B. A., Stansell, N. D., Mann, M. E., Cooke, C. A., Abbott, M. B., Vuille, M., Bird, B. W., Lachniet, M. S., & Fernandez, A. (2022). Interhemispheric antiphasing of neotropical precipitation during the past millennium. Proceedings of the National Academy of Sciences, 119(17). 10.1073/pnas.2120015119
- Konecky, B. L., McKay, N. P., Churakova (Sidorova), O. V., Comas-Bru, L., Dassié, E. P., DeLong, K. L., Falster, G. M., Fischer, M. J., Jones, M. D., Jonkers, L., Kaufman, D. S., Leduc, G., Managave, S. R., Martrat, B., Opel, T., Orsi, A. J., Partin, J. W., Sayani, H. R., Thomas, E. K., … von Gunten, L. (2020). The Iso2k database: a global compilation of paleo-δ18O and δ2H records to aid understanding of Common Era climate. Earth System Science Data, 12(3), 2261–2288. 10.5194/essd-12-2261-2020
- Emile-Geay, J., McKay, N. P., Kaufman, D. S., von Gunten, L., Wang, J., Anchukaitis, K. J., Abram, N. J., Addison, J. A., Curran, M. A. J., Evans, M. N., Henley, B. J., Hao, Z., Martrat, B., McGregor, H. V., Neukom, R., Pederson, G. T., Stenni, B., Thirumalai, K., Werner, J. P., … Zinke, J. (2017). A global multiproxy database for temperature reconstructions of the Common Era. Scientific Data, 4(1). 10.1038/sdata.2017.88
- Walter, R. M., Sayani, H. R., Felis, T., Cobb, K. M., Abram, N. J., Arzey, A. K., Atwood, A. R., Brenner, L. D., Dassié, É. P., DeLong, K. L., Ellis, B., Emile-Geay, J., Fischer, M. J., Goodkin, N. F., Hargreaves, J. A., Kilbourne, K. H., Krawczyk, H., McKay, N. P., Moore, A. L., … Zinke, J. (2023). The CoralHydro2k database: a global, actively curated compilation of coral δ18O and Sr ∕ Ca proxy records of tropical ocean hydrology and temperature for the Common Era. Earth System Science Data, 15(5), 2081–2116. 10.5194/essd-15-2081-2023
- Comas-Bru, L., Rehfeld, K., Roesch, C., Amirnezhad-Mozhdehi, S., Harrison, S. P., Atsawawaranunt, K., Ahmad, S. M., Brahim, Y. A., Baker, A., Bosomworth, M., Breitenbach, S. F. M., Burstyn, Y., Columbu, A., Deininger, M., Demény, A., Dixon, B., Fohlmeister, J., Hatvani, I. G., Hu, J., … Zhang, H. (2020). SISALv2: a comprehensive speleothem isotope database with multiple age–depth models. Earth System Science Data, 12(4), 2579–2606. 10.5194/essd-12-2579-2020
- Hoyer, S., & Hamman, J. (2017). xarray: N-D labeled Arrays and Datasets in Python. Journal of Open Research Software, 5(1), 10. 10.5334/jors.148
- Khider, D., Emile‐Geay, J., Zhu, F., James, A., Landers, J., Ratnakar, V., & Gil, Y. (2022). Pyleoclim: Paleoclimate Timeseries Analysis and Visualization With Python. Paleoceanography and Paleoclimatology, 37(10). 10.1029/2022pa004509
- Khider, D., Emile-Geay, J., Zhu, F., James, A., Landers, J., Kwan, M., Athreya, P., McGibbon, R., Voirol, L., & Stiller, D. (2025). Pyleoclim: A Python package for the analysis and visualization of paleoclimate data. Zenodo. 10.5281/ZENODO.1205661
- Dawson, A. (2016). eofs: A Library for EOF Analysis of Meteorological, Oceanographic, and Climate Data. Journal of Open Research Software, 4(1), 14. 10.5334/jors.122