From ef5fcf962b55c188932fcf6fe72eb8fd792637b9 Mon Sep 17 00:00:00 2001 From: Jennifer Oxelson Date: Mon, 13 May 2024 16:48:14 -0600 Subject: [PATCH] ftp --> resources --- INSTALL.md | 13 +++++++------ cmake/dependencies.cmake | 2 +- configure.ac | 6 +++--- docs/attribute_conventions.md | 2 +- docs/static-pages/usage.html | 2 +- hdf4_test/run_get_hdf4_files.sh | 4 ++-- nc_perf/run_knmi_bm.sh | 2 +- 7 files changed, 16 insertions(+), 15 deletions(-) diff --git a/INSTALL.md b/INSTALL.md index b9cf104d7d..47b5c384c5 100644 --- a/INSTALL.md +++ b/INSTALL.md @@ -103,9 +103,10 @@ libraries. (And, optionally, the szlib library). Versions required are at least HDF5 1.8.9, zlib 1.2.5, and curl 7.18.0 or later. (Optionally, if building with szlib, get szip 2.0 or later.) -HDF5 1.8.9 and zlib 1.2.7 packages are available from the netCDF-4 ftp -site. If you wish to use the remote data client code, then you +These packages are available at: +https://resources.unidata.ucar.edu/netcdf/netcdf-4/ + +If you wish to use the remote data client code, then you will also need libcurl, which can be obtained from the curl website. @@ -314,7 +315,7 @@ $ make check install If parallel I/O access to netCDF classic, 64-bit offset, CDF-5 files is also needed, the PnetCDF library should also be installed. (Note: the previously recommended replacement +href="https://resources.unidata.ucar.edu/netcdf/contrib/pnetcdf.h">replacement pnetcdf.h should no longer be used.) Then configure netCDF with the "--enable-pnetcdf" option. @@ -361,7 +362,7 @@ Note: --disable prefix indicates that the option is normally enabled. --enable-netcdf-4build with netcdf-4HDF5 and zlib --enable-netcdf4synonym for enable-netcdf-4 --enable-hdf4build netcdf-4 with HDF4 read capabilityHDF4, HDF5 and zlib ---enable-hdf4-file-teststest ability to read HDF4 filesselected HDF4 files from Unidata ftp site +--enable-hdf4-file-teststest ability to read HDF4 filesselected HDF4 files from Unidata resources site --enable-pnetcdfbuild netcdf-4 with parallel I/O for classic, 64-bit offset, and CDF-5 files using PnetCDF --enable-extra-example-testsRun extra example tests--enable-netcdf-4,GNU sed --enable-parallel-tests run extra parallel IO tests--enable-netcdf-4, parallel IO support @@ -384,7 +385,7 @@ Note: --disable prefix indicates that the option is normally enabled. The benchmarks are a bunch of extra tests, which are timed. We use these tests to check netCDF performance. - sample data files from the Unidata ftp site + sample data files from the Unidata resources site --disable-extreme-numbers don't use extreme numbers during testing, such as MAX_INT - 1 --enable-dllbuild a win32 DLLmingw compiler diff --git a/cmake/dependencies.cmake b/cmake/dependencies.cmake index 2294c272bc..db8b903d50 100644 --- a/cmake/dependencies.cmake +++ b/cmake/dependencies.cmake @@ -69,7 +69,7 @@ if(NETCDF_ENABLE_HDF4) message(STATUS "Found JPEG libraries: ${JPEG_LIB}") # Option to enable HDF4 file tests. - option(NETCDF_ENABLE_HDF4_FILE_TESTS "Run HDF4 file tests. This fetches sample HDF4 files from the Unidata ftp site to test with (requires curl)." ON) + option(NETCDF_ENABLE_HDF4_FILE_TESTS "Run HDF4 file tests. This fetches sample HDF4 files from the Unidata resources site to test with (requires curl)." ON) if(NETCDF_ENABLE_HDF4_FILE_TESTS) find_program(PROG_CURL NAMES curl) if(PROG_CURL) diff --git a/configure.ac b/configure.ac index b8b544e871..90479654b0 100644 --- a/configure.ac +++ b/configure.ac @@ -354,9 +354,9 @@ AC_MSG_RESULT([$enable_dynamic_loading]) # Does the user want to turn on extra HDF4 file tests? -AC_MSG_CHECKING([whether to fetch some sample HDF4 files from Unidata ftp site to test HDF4 reading (requires wget)]) +AC_MSG_CHECKING([whether to fetch some sample HDF4 files from Unidata resources site to test HDF4 reading (requires wget)]) AC_ARG_ENABLE([hdf4-file-tests], [AS_HELP_STRING([--enable-hdf4-file-tests], - [get some HDF4 files from Unidata ftp site and test that they can be read])]) + [get some HDF4 files from Unidata resources site and test that they can be read])]) test "x$enable_hdf4" = xyes -a "x$enable_hdf4_file_tests" = xyes || enable_hdf4_file_tests=no if test "x$enable_hdf4_file_tests" = xyes; then AC_DEFINE([USE_HDF4_FILE_TESTS], 1, [If true, use use wget to fetch some sample HDF4 data, and then test against it.]) @@ -1096,7 +1096,7 @@ fi AC_MSG_CHECKING([whether benchmarks should be run]) AC_ARG_ENABLE([benchmarks], [AS_HELP_STRING([--enable-benchmarks], - [Run benchmarks. This will cause sample data files from the Unidata ftp + [Run benchmarks. This will cause sample data files from the Unidata resources site to be fetched. The benchmarks are a bunch of extra tests, which are timed. We use these tests to check netCDF performance.])]) test "x$enable_benchmarks" = xyes || enable_benchmarks=no diff --git a/docs/attribute_conventions.md b/docs/attribute_conventions.md index b13ec4d693..ed75a88f81 100644 --- a/docs/attribute_conventions.md +++ b/docs/attribute_conventions.md @@ -115,7 +115,7 @@ It is strongly recommended that applicable conventions be followed unless there `Conventions` -> If present, 'Conventions' is a global attribute that is a character array for the name of the conventions followed by the dataset. Originally, these conventions were named by a string that was interpreted as a directory name relative to the directory /pub/netcdf/Conventions/ on the now defunct host ftp.unidata.ucar.edu. The web page https://www.unidata.ucar.edu/netcdf/conventions.html is now the preferred and authoritative location for registering a URI reference to a set of conventions maintained elsewhere. Authors of new conventions should submit a request to support-netcdf@unidata.ucar.edu for listing on the Unidata conventions web page. +> If present, 'Conventions' is a global attribute that is a character array for the name of the conventions followed by the dataset. Originally, these conventions were named by a string that was interpreted as a directory name relative to the directory /pub/netcdf/Conventions/ on the now defunct ftp host. The web page https://www.unidata.ucar.edu/netcdf/conventions.html is now the preferred and authoritative location for registering a URI reference to a set of conventions maintained elsewhere. Authors of new conventions should submit a request to support-netcdf@unidata.ucar.edu for listing on the Unidata conventions web page.

diff --git a/docs/static-pages/usage.html b/docs/static-pages/usage.html index 2c159cfaf4..6acc16269b 100644 --- a/docs/static-pages/usage.html +++ b/docs/static-pages/usage.html @@ -28,7 +28,7 @@

Where is NetCDF Used?

recently made some use of netCDF, based on
  1. - downloads from the Unidata site (ftp and http) + downloads from the Unidata downloads site
  2. subscribers and posters to netCDF mailing lists diff --git a/hdf4_test/run_get_hdf4_files.sh b/hdf4_test/run_get_hdf4_files.sh index faf1206a67..df286d48ea 100755 --- a/hdf4_test/run_get_hdf4_files.sh +++ b/hdf4_test/run_get_hdf4_files.sh @@ -14,7 +14,7 @@ if test "x$srcdir" = x ; then srcdir=`pwd`; fi # Get a file from the resources site; retry several times getfile() { - DATAFILE="https://resources.unidata.ucar.edu/sample_data/hdf4/$1.gz" + DATAFILE="https://resources.unidata.ucar.edu/netcdf/sample_data/hdf4/$1.gz" for try in 1 2 3 4 ; do # try 4 times @@ -30,7 +30,7 @@ getfile() { set -e echo "" -echo "Getting HDF4 sample files from Unidata FTP site..." +echo "Getting HDF4 sample files from Unidata resources site..." file_list="AMSR_E_L2_Rain_V10_200905312326_A.hdf AMSR_E_L3_DailyLand_V06_20020619.hdf \ MYD29.A2009152.0000.005.2009153124331.hdf MYD29.A2002185.0000.005.2007160150627.hdf \ diff --git a/nc_perf/run_knmi_bm.sh b/nc_perf/run_knmi_bm.sh index 36d4ad590f..5d33e274a1 100755 --- a/nc_perf/run_knmi_bm.sh +++ b/nc_perf/run_knmi_bm.sh @@ -15,7 +15,7 @@ echo "Getting KNMI test files $file_list" for f1 in $file_list do if ! test -f $f1; then - wget https://resources.unidata.ucar.edu/sample_data/$f1.gz + wget https://resources.unidata.ucar.edu/netcdf/sample_data/$f1.gz gunzip $f1.gz fi done