v4.5
Overview:
The main features of this release are Harvesting, Metadata Export, Private URL, and Federated Login Mode.
- Harvesting allows the dataset metadata from another site to be imported so that these files appear to be local, though data files remain on the remote site. This allows Dataverse installations and other repositories to share metadata with each other to create a data sharing community and to provide more access to the datasets stored in each repository. Harvesting is implemented using the standard OAI-PMH protocol - any Dataverse installation can be configured as an OAI client and server.
- Metadata Export makes it easier to share and discover from Dataverse. Another key function of the export is to store the metadata in the file system in a format for preservation. In this release, the metadata is exported in Dublin Core, Data Documentation Initiative (DDI), and native JSON. Metadata export will support additional standards in the future.
- Private URL provides a simple way to generate and send a special URL to an unpublished dataset for a group of anonymous peer reviewers to review the data.
- Federated Login Mode allows users from 200+ different organizations to sign in using his or her organizational credentials.
Big thanks to all the community members who participated in usability testing, provided feedback in Github issues, and worked with us in other ways to make sure that this release was successful.
For the complete list of issues, see the 4.5 milestone in Github.
For help with upgrading, installing, or general questions please email support@dataverse.org.
Installation:
If this is a new installation, please see our Installation Guide.
Upgrade:
If you are upgrading from v4.x, you must upgrade to each intermediate version before installing this version. When upgrading from the previous version, you will need to do the following:
- Undeploy the previous version.
- /glassfish4/bin/asadmin list-applications
- /glassfish4/bin/asadmin undeploy dataverse
- Stop glassfish and remove the generated directory
- service glassfish stop
- remove the generated directory: rm -rf /usr/local/glassfish4/glassfish/domains/domain1/generated
- Add a new jvm option to /usr/local/glasssfish4/glassfish/domains/domain1/config/domain.xml to support timers, start glassfish
- -Ddataverse.timerServer=true (note, see http://guides.dataverse.org/en/latest/admin/timers.html if using more than one server)
- service glassfish start
- Deploy this version.
- /glassfish4/bin/asadmin deploy dataverse-4.5.war
- Run the database update script.
"psql -U -d -f upgrade_v4.4_to_4.5.sql" - Replace the Solr schema.xml file
- stop solr (kill -9 pid)
- replace schema.xml
- start solr (java -jar start.jar &)
- Run the incremental index to prepare for harvesting facet.
- Remove last index times from the database:
update dvobject set indextime=null;
update dvobject set permissionindextime=null; - From the glassfish console, run the incremental index:
curl -X GET "http://localhost:8080/api/admin/index/continue?numPartitions=1&partitionIdToProcess=0"
- Remove last index times from the database:
If you are upgrading from v3.x, you will need to perform a migration to v4.x since our application was redesigned and the database schema are completely different. This is a significant undertaking. Please contact us (support at dataverse.org) before beginning. Also refer to our migration google group for additional support and information: https://groups.google.com/d/msgid/dataverse-migration-wg
Please note: v4.x does not currently support creating new handles though it will support existing ones. We intend to add this feature but have not yet scheduled this work.
IMPORTANT: If you are running TwoRavens with your dataverse:
Make sure the two applications are using the same version of the "pre-processed statistics" R code. Compare the 2 files:
On the TwoRavens side:
.../dataexplore/rook/preprocess/preprocess.R
On the Dataverse side:
.../applications/dataverse-4.5/WEB-INF/classes/edu/harvard/iq/dataverse/rserve/scripts/preprocess.R
If they are different, replace the Dataverse copy with the TwoRavens copy (i.e., the TwoRavens version wins!).
And, also, remove all the already-generated pre-processed fragments in your Dataverse file directory, for example:
cd [files directory]
rm -f find . -name '*.prep'
If the two copies are the same, you don't need to do any of this.
Please note that this is a temporary measure, we are working on a fix that will make the two applications resolve code version conflicts like this automatically.