This shows you the differences between two versions of the page.
Both sides previous revision Previous revision Next revision | Previous revision Next revision Both sides next revision | ||
other:uvcdat:cdat_conda:cdat_8_2_1 [2023/01/31 14:32] jypeter [Extra packages list] Added uxarray |
other:uvcdat:cdat_conda:cdat_8_2_1 [2023/12/15 10:26] jypeter [Extra packages list] Added D-Tale |
||
---|---|---|---|
Line 266: | Line 266: | ||
* [[https://cmor.llnl.gov/|CMOR]]: CMOR (//Climate Model Output Rewriter//) is used to produce CF-compliant netCDF files | * [[https://cmor.llnl.gov/|CMOR]]: CMOR (//Climate Model Output Rewriter//) is used to produce CF-compliant netCDF files | ||
* ''conda install -n cdatm19_py3 -c conda-forge cmor'' | * ''conda install -n cdatm19_py3 -c conda-forge cmor'' | ||
- | * Get version number with: ''python -c 'from cmor import *; print( (CMOR_VERSION_MAJOR, CMOR_VERSION_MINOR, CMOR_VERSION_PATCH) )' '' | ||
- | * <wrap hi>Warning!</wrap> [[https://github.com/PCMDI/cmor/issues/449|CMOR currently requires Python 2.7]] | ||
* [[http://scitools.org.uk/cartopy/|cartopy]]: a Python package designed for geospatial data processing in order to produce maps and other geospatial data analyses | * [[http://scitools.org.uk/cartopy/|cartopy]]: a Python package designed for geospatial data processing in order to produce maps and other geospatial data analyses | ||
Line 341: | Line 339: | ||
* ''python-cdo'' will install the ''cdo'' package (providing the ''cdo ''executable) as a dependency | * ''python-cdo'' will install the ''cdo'' package (providing the ''cdo ''executable) as a dependency | ||
* see also [[https://code.mpimet.mpg.de/projects/cdo/wiki/Cdo%7Brbpy%7D|Using CDO from python or ruby]] | * see also [[https://code.mpimet.mpg.de/projects/cdo/wiki/Cdo%7Brbpy%7D|Using CDO from python or ruby]] | ||
+ | * [[https://cds.climate.copernicus.eu/api-how-to|cdsapi]]: The Climate Data Store (CDS) Application Program Interface (API) is a service providing programmatic access to CDS data | ||
+ | * CDS = Copernicus [[https://cds.climate.copernicus.eu/|Climate Data Store]] | ||
+ | * Example: [[https://cds.climate.copernicus.eu/cdsapp#!/dataset/reanalysis-era5-pressure-levels?tab=form|ERA5 hourly data on pressure levels from 1959 to present]] | ||
* [[https://github.com/ecmwf/cfgrib|cfgrib]]: Python interface to map GRIB files to the Unidata's Common Data Model v4 following the CF Conventions | * [[https://github.com/ecmwf/cfgrib|cfgrib]]: Python interface to map GRIB files to the Unidata's Common Data Model v4 following the CF Conventions | ||
* see also ''eccodes'' and ''pygrib'' | * see also ''eccodes'' and ''pygrib'' | ||
* [[https://unidata.github.io/cftime/|cftime]]: Python library for decoding time units and variable values in a netCDF file conforming to the Climate and Forecasting (CF) netCDF conventions | * [[https://unidata.github.io/cftime/|cftime]]: Python library for decoding time units and variable values in a netCDF file conforming to the Climate and Forecasting (CF) netCDF conventions | ||
+ | * Used with ''xarray'' | ||
+ | * [[https://cf-xarray.readthedocs.io/|cf_xarray]]: provides an accessor (''DataArray.cf'' or ''Dataset.cf'') that allows you to interpret //Climate and Forecast// metadata convention attributes present on ''xarray'' objects | ||
* Used with ''xarray'' | * Used with ''xarray'' | ||
* [[https://clustergram.readthedocs.io/|clustergram]]: visualization and diagnostics for cluster analysis | * [[https://clustergram.readthedocs.io/|clustergram]]: visualization and diagnostics for cluster analysis | ||
Line 351: | Line 354: | ||
* [[https://dash.plotly.com/|dash]] and [[https://github.com/plotly/jupyter-dash|jupyter-dash]]: the original low-code framework for rapidly building data apps in Python, R, Julia, and F# | * [[https://dash.plotly.com/|dash]] and [[https://github.com/plotly/jupyter-dash|jupyter-dash]]: the original low-code framework for rapidly building data apps in Python, R, Julia, and F# | ||
* see also ''plotly'' | * see also ''plotly'' | ||
+ | * [[https://github.com/man-group/dtale|D-Tale]] brings you an easy way to view & analyze Pandas data structures. It integrates seamlessly with ipython notebooks & python/ipython terminals. | ||
+ | * Install with: ''conda install dtale -c conda-forge'' | ||
* [[https://dynamictimewarping.github.io/|python-dtw]]: implementation of Dynamic Time Warping-type (DTW) | * [[https://dynamictimewarping.github.io/|python-dtw]]: implementation of Dynamic Time Warping-type (DTW) | ||
* Note: older installations provided [[https://github.com/pierre-rouanet/dtw|dtw]]: DTW (Dynamic Time Warping) python module | * Note: older installations provided [[https://github.com/pierre-rouanet/dtw|dtw]]: DTW (Dynamic Time Warping) python module | ||
Line 356: | Line 361: | ||
* need to install ''eccodes'' and ''python-eccodes'' ([[https://github.com/ecmwf/eccodes-python/issues/56|details]]) | * need to install ''eccodes'' and ''python-eccodes'' ([[https://github.com/ecmwf/eccodes-python/issues/56|details]]) | ||
* see also ''cfgrib'' and ''pygrib'' | * see also ''cfgrib'' and ''pygrib'' | ||
+ | * [[https://ajdawson.github.io/eofs/|eofs]]: a Python package for EOF analysis of spatial-temporal data | ||
* [[https://docs.esmvaltool.org/en/latest/|ESMValTool]]: Earth System Model Evaluation Tool | * [[https://docs.esmvaltool.org/en/latest/|ESMValTool]]: Earth System Model Evaluation Tool | ||
* not installed yet (March 2022) because it required download of 302 Mb extra packages and a downgrade of gdal | * not installed yet (March 2022) because it required download of 302 Mb extra packages and a downgrade of gdal | ||
+ | * [[https://flox.readthedocs.io/|flox]]: fast & furious //GroupBy// reductions for dask.array | ||
+ | * See also | ||
+ | * [[https://xarray.dev/blog/flox|flox: Faster GroupBy reductions with Xarray]] | ||
+ | * [[https://flox.readthedocs.io/en/latest/implementation.html|Parallel Algorithms]] | ||
* [[http://ferret.pmel.noaa.gov/Ferret/documentation/pyferret|pyferret]] and ''ferret_datasets'': Ferret encapsulated in Python | * [[http://ferret.pmel.noaa.gov/Ferret/documentation/pyferret|pyferret]] and ''ferret_datasets'': Ferret encapsulated in Python | ||
* [[https://github.com/PBrockmann/ipython_ferretmagic|ipython_ferretmagic]]: IPython notebook extension for ferret | * [[https://github.com/PBrockmann/ipython_ferretmagic|ipython_ferretmagic]]: IPython notebook extension for ferret | ||
Line 365: | Line 375: | ||
* [[https://github.com/TEOS-10/GSW-python|gsw]]: Python implementation of the Thermodynamic Equation Of Seawater | * [[https://github.com/TEOS-10/GSW-python|gsw]]: Python implementation of the Thermodynamic Equation Of Seawater | ||
* see also //seawater// | * see also //seawater// | ||
+ | * [[https://noaa-orr-erd.github.io/gridded/index.html|gridded]]: a single API for accessing / working with gridded model results on multiple grid types | ||
+ | * Supports the [[http://cfconventions.org/|CF]], [[https://ugrid-conventions.github.io/ugrid-conventions/|UGRID]] and [[http://sgrid.github.io/sgrid/|SGRID]] conventions | ||
+ | * [[https://icclim.readthedocs.io/|icclim]]: icclim (Index Calculation for CLIMate) is a Python library to compute climate indices | ||
* [[https://intake.readthedocs.io/|intake]]: a lightweight package for finding, investigating, loading and disseminating data | * [[https://intake.readthedocs.io/|intake]]: a lightweight package for finding, investigating, loading and disseminating data | ||
* [[https://intake-esm.readthedocs.io/|intake-esm]]: data cataloging utility built on top of intake, pandas, and xarray | * [[https://intake-esm.readthedocs.io/|intake-esm]]: data cataloging utility built on top of intake, pandas, and xarray | ||
Line 375: | Line 388: | ||
* [[https://jupyterlab.readthedocs.io/|jupyterlab]]: the next-generation web-based user interface for Project Jupyter | * [[https://jupyterlab.readthedocs.io/|jupyterlab]]: the next-generation web-based user interface for Project Jupyter | ||
* [[https://github.com/plotly/Kaleido|python-kaleido]]: a cross-platform library for generating static images (e.g. png, svg, pdf, etc.) for web-based visualization libraries | * [[https://github.com/plotly/Kaleido|python-kaleido]]: a cross-platform library for generating static images (e.g. png, svg, pdf, etc.) for web-based visualization libraries | ||
+ | * [[https://unidata.github.io/MetPy/latest/|MetPy]]: a collection of tools in Python for reading, visualizing, and performing calculations with weather data | ||
* [[https://mpltern.readthedocs.io/|mpltern]]: a Python plotting library based on Matplotlib specifically designed for ternary plots | * [[https://mpltern.readthedocs.io/|mpltern]]: a Python plotting library based on Matplotlib specifically designed for ternary plots | ||
* [[https://nc-time-axis.readthedocs.io/en/stable/|nc-time-axis]]: a package that enables making plots in matplotlib with axes made up of ''cftime.datetime'' dates with any calendar type | * [[https://nc-time-axis.readthedocs.io/en/stable/|nc-time-axis]]: a package that enables making plots in matplotlib with axes made up of ''cftime.datetime'' dates with any calendar type | ||
Line 380: | Line 394: | ||
* can probably also be done with [[https://code.mpimet.mpg.de/projects/cdo/embedded/index.html#x1-580002.1.3|cdo -v diffn file1.nc file2.nc]] | * can probably also be done with [[https://code.mpimet.mpg.de/projects/cdo/embedded/index.html#x1-580002.1.3|cdo -v diffn file1.nc file2.nc]] | ||
* [[https://opencv.org/|OpenCV]]: OpenCV (Open Source Computer Vision Library) is an open-source library that includes several hundreds of computer vision algorithms. See also [[https://www.geeksforgeeks.org/opencv-python-tutorial/|OpenCV Python Tutorial]] and **scikit-image** | * [[https://opencv.org/|OpenCV]]: OpenCV (Open Source Computer Vision Library) is an open-source library that includes several hundreds of computer vision algorithms. See also [[https://www.geeksforgeeks.org/opencv-python-tutorial/|OpenCV Python Tutorial]] and **scikit-image** | ||
+ | * [[https://openpyxl.readthedocs.io/|openpyxl]]: a Python library to read/write Excel 2010 ''xlsx''/''xlsm'' files | ||
* [[https://jiffyclub.github.io/palettable/|Palettable]]: Color palettes for Python ''[color]'' | * [[https://jiffyclub.github.io/palettable/|Palettable]]: Color palettes for Python ''[color]'' | ||
* [[http://pandas.pydata.org/|pandas]]: Python Data Analysis Library | * [[http://pandas.pydata.org/|pandas]]: Python Data Analysis Library | ||
Line 389: | Line 404: | ||
* [[https://www.fatiando.org/pooch/|pooch]]: a friend to fetch your data files (makes it easy to download a file, without messing with ''requests'' and ''urllib'') | * [[https://www.fatiando.org/pooch/|pooch]]: a friend to fetch your data files (makes it easy to download a file, without messing with ''requests'' and ''urllib'') | ||
* [[https://proplot.readthedocs.io/en/latest/|proplot]]: a lightweight **matplotlib wrapper** for making beautiful, publication-quality graphics | * [[https://proplot.readthedocs.io/en/latest/|proplot]]: a lightweight **matplotlib wrapper** for making beautiful, publication-quality graphics | ||
+ | * [[https://psyplot.github.io/|psyplot]]: Interactive Data Visualization from Python and GUIs | ||
* [[https://github.com/jswhit/pygrib|pygrib]]: high-level interface to the ECWMF ECCODES C library for reading GRIB files | * [[https://github.com/jswhit/pygrib|pygrib]]: high-level interface to the ECWMF ECCODES C library for reading GRIB files | ||
* see also ''eccodes'' and ''cfgrib'' | * see also ''eccodes'' and ''cfgrib'' | ||
Line 407: | Line 423: | ||
* [[https://anaconda.org/anaconda/tensorflow-mkl|tensorflow-mkl]] will install the **CPU**-based (**not GPU**) version | * [[https://anaconda.org/anaconda/tensorflow-mkl|tensorflow-mkl]] will install the **CPU**-based (**not GPU**) version | ||
* [[https://github.com/tqdm|tqdm]]: make your loops show a smart progress meter | * [[https://github.com/tqdm|tqdm]]: make your loops show a smart progress meter | ||
- | * [[https://uxarray.readthedocs.io/|uxarray]]: provide xarray styled functionality for unstructured grid datasets following [[https://ugrid-conventions.github.io/|UGRID Conventions]] | + | * [[https://uxarray.readthedocs.io/|uxarray]]: provide xarray styled functionality for unstructured grid datasets following [[https://ugrid-conventions.github.io/ugrid-conventions/|UGRID Conventions]] |
* [[https://docs.xarray.dev/en/stable/|xarray]]: Xarray makes working with labelled multi-dimensional arrays in Python simple, efficient, and fun! | * [[https://docs.xarray.dev/en/stable/|xarray]]: Xarray makes working with labelled multi-dimensional arrays in Python simple, efficient, and fun! | ||
- | * [[https://xcdat.readthedocs.io/|xcdat]]: Xarray Extended with Climate Data Analysis Tools | + | * See also: ''flox'', ''xcdat'', ... |
+ | * [[https://xcdat.readthedocs.io/|xcdat]]: Xarray extended with Climate Data Analysis Tools | ||
+ | * [[https://xclim.readthedocs.io/|xclim]]: an operational Python library for climate services, providing numerous climate-related indicator tools with an extensible framework for constructing custom climate indicators, statistical downscaling and bias adjustment of climate model simulations, as well as climate model ensemble analysis tools. | ||
+ | * [[https://xesmf.readthedocs.io/|xESMF]]: Universal Regridder for Geospatial Data | ||
+ | * [[https://xgrads.readthedocs.io/|xgrads]]: parse and read binary dataset described by a ''.ctl'' file commonly used by [[http://cola.gmu.edu/grads/|GrADS]] or [[http://www.opengrads.org/|openGrADS]] | ||
+ | * [[https://xlrd.readthedocs.io/|xlrd]]: a library for reading data and formatting information from Excel files in the historical .xls format | ||
* [[https://xoa.readthedocs.io/en/latest/|xoa]]: xarray-based ocean analysis library | * [[https://xoa.readthedocs.io/en/latest/|xoa]]: xarray-based ocean analysis library | ||
* ''xoa'' is the successor of [[http://www.ifremer.fr/vacumm/|vacumm]] (vacumm does **not** support Python3) | * ''xoa'' is the successor of [[http://www.ifremer.fr/vacumm/|vacumm]] (vacumm does **not** support Python3) | ||
+ | * [[https://xskillscore.readthedocs.io/|xskillscore]]: metrics for verifying forecasts | ||
+ | * [[https://docs.profiling.ydata.ai/|ydata-profiling]]: a leading package for data profiling, that automates and standardizes the generation of detailed reports, complete with statistics and visualizations. | ||
==== Removed packages ==== | ==== Removed packages ==== |