ScholarSphere Version Build Status Dependency StatusStories in Ready

ScholarSphere is Penn State's self- and proxy-deposit repository for access to and preservation of scholarly works and data. It is built atop Sufia, a Hydra/Rails-based component.

ScholarSphere is being developed as part of Penn State's Digital Stewardship Program. Development on ScholarSphere began as part of the prototype CAPS project. Code and documentation are freely available via Github.

For more information, read the ScholarSphere development docs.

License

ScholarSphere is available under the Apache 2.0 license. Read the copyright statement and license.

Install

Infrastructural components

  • Ruby 2.0 (we use RVM and rbenv to manage our Rubies)
  • Fedora (if you don't have access to an instance, use the built-in hydra-jetty)
  • Solr (if you don't have access to an instance, use the built-in hydra-jetty)
  • A relational database (SQLite and MySQL have been tested)
  • Redis (for activity streams and background jobs)

Install system dependencies

  • libmysqlclient-dev (if running MySQL as RDBMS)
  • libsqlite3-dev (if running SQLite as RDBMS)
  • libmagick-dev (or libmagickcore-dev on Ubuntu 12.10)
  • libmagickwand-dev
  • clamav
  • clamav-daemon
  • libclamav-dev
  • ghostscript (required to create thumbnails from pdfs)
  • FITS -- put it in a directory on your PATH, or just use the included git submodule
  • phantomjs -- if you're running the test suite, you'll need phantomjs (headless webkit browser) on your PATH for the feature specs

Get the ScholarSphere code

git clone https://github.com/psu-stewardship/scholarsphere.git

Install gems

bundle install

Copy config samples

cp config/devise.yml.sample config/devise.yml
cp config/database.yml.sample config/database.yml
cp config/fedora.yml.sample config/fedora.yml
cp config/solr.yml.sample config/solr.yml
cp config/redis.yml.sample config/redis.yml
cp config/hydra-ldap.yml.sample config/hydra-ldap.yml

If you're using SQLite, a vanilla Redis installation, and the Hydra-Jetty Solr and Fedora components (see below), you should not need to tweak the database.yml, fedora.yml, solr.yml, or redis.yml files.

If you're planning to use LDAP for user account information and groups, you will need to know some information about your LDAP service, which will go into hydra-ldap.yml.

Create database

rake db:create

Generate a new secret token

rake scholarsphere:generate_secret

Migrate relational database

rake db:migrate

To use the built-in Fedora and Solr instances, get the bundled hydra-jetty, configure it, & fire it up

rake jetty:clean
rake sufia:jetty:config
rake jetty:start

Start the resque-pool workers (needed for characterization, audit, and resolrization services)

resque-pool --daemon --environment development start

Run the app server (the bundled app server is Unicorn)

rails server

Browse to http://localhost:3000/ and you should see ScholarSphere!

Usage Notes

Enabling Zotero integration

To enable integration with Zotero (more about that feature), here are the required steps:

  1. Register an OAuth client for ScholarSphere with Zotero. Note the client key and secret for later.
  2. Install and start arkivo-sufia in a server environment. Note the server hostname and IP address arkivo-sufia is running on for later.
  3. Create Arkivo tokens for all existing users in your application database via (assuming production environment) RAILS_ENV=production rake sufia:user:tokens
  4. Set environment variables for the Zotero OAuth client key and secret you generated in step 1 above, called ZOTERO_CLIENT_KEY and ZOTERO_CLIENT_SECRET. The config/zotero.yml file depends on these environment variables. (If you'd prefer not to manage these via env vars, you are also welcome to handle zotero.yml in a different way, e.g., the way we handle our other configs that are hidden from version control.) Make sure that these variables are available to the user running the ScholarSphere rails server.
  5. Copy config/arkivo.yml.sample to config/arkivo.yml and set it up with the hostname and IP address you noted in step 2 above. (This is how we handle most of our production configs already, so this just mirrors current practice.)
  6. Edit config/initializers/arkivo_constraint.rb to allow connections to the Arkivo API. If you're just testing, you can have the matches? method return true but do not do this in production! This effectively allows any client unauthenticated access to an API that permits adding, modifying, and removing content. In production, you can use the routing constraint to ensure that the API is accessible only to the specific IP address of the host running the Arkivo service. You can do that by having the matches? method return something like request.remote_ip == '10.0.0.3'. (If you're not comfortable having a back-end IP address stored in a file that is under version control, you can also make use of an environment variable here, e.g., request.remote_ip == ENV['ARKIVO_HOST_IP'], in which case you'll need to make sure the server environment has that variable set to the proper value.)
  7. Restart the Rails server and all background jobs, and you should now be able to OAuth to Zotero via the Edit Profile screen, at which point the magic should start happening.

Auditing All Datastreams

To audit the digital signatures of every version of every object in the repository, run the following command

script/audit_repository

You'll probably want to schedule this regularly (e.g., via cron) in production environments. Note that this does not force an audit -- it respects the value of max_days_between_audits in application.rb. Also note that if you want to run this on any environment other than development, you will need to call the script with RAILS_ENV=environment in front.

Re-solrize All Objects

If for some reason you need to force all objects to be re-solrizer, perhaps because you have updated which fields are facetable and which are not, ScholarSphere contains a rake task that kicks off a re-solrization asynchronously via a Resque job.

 rake scholarsphere:resolrize

Note that if you want to run this on any environment other than development, you will need to call the script with RAILS_ENV=environment in front.

Characterize All Uncharacterized Datastreams

In the event that some objects have not undergone characterization (for whatever reason), there is a rake task that sweeps through the entire repository looking for objects that lack a characterization datastream. For each object that lacks this datastream, a CharacterizationJob that will characterize and thumbnailize the object is queued up.

 rake scholarsphere:characterize

Note that if you want to run this on any environment other than development, you will need to call the script with RAILS_ENV=environment in front.

Export Metadata as RDF/XML

There is a rake task that exports the metadata of every object that is readable by the public to the RDF/XML format. This might be useful as an export mechanism, e.g., to Summon or a similar discovery system.

 rake scholarsphere:export:rdfxml

Note that if you want to run this on any environment other than development, you will need to call the script with RAILS_ENV=environment in front.

Harvesting Authorities Locally

ScholarSphere supports "authority suggestion," a feature that links controlled vocabularies to descriptive metadata elements. This provides functionality both for mapping string values to URIs and for populating dropdowns in metadata form fields, e.g., if a user types "Cro" into a subject field, they might see a list that includes "Croatian independence," the subject they were going to type out.

In order to avoid network latency, these vocabularies are harvested in advance and stuffed into the ScholarSphere relational database for easy and quick lookups.

To get a sense for how this works, pull in database fixtures containing pre-harvested authorities

rake db:data:load

To harvest more authorities:

  1. Harvest the authority (See available harvest tasks via rake -T scholarsphere:harvest) -- N.B. depending on the size of the vocabulary, this may take a very long time, especially if you're using a slower database such as SQLite.
  2. (OPTIONAL) Generate fixtures so other instances don't need to re-harvest (See available database tasks via rake -T db)
  3. Register the vocabulary with a domain term in generic_file_rdf_datastream.rb (See the bottom of the file for examples)

Contribute