3. Installation guide


The following installation guide for the Goobi viewer refers to Ubuntu Linux 20.04. It is written as a step-by-step guide from top to bottom, meaning that settings and configurations build on each other. If the order is not followed, certain commands may fail.

VIEWER.EXAMPLE.ORG is used as the domain name in this manual, please adapt this to your own DNS name.

For productive use, a virtual machine with at least 4 CPUs and 8GB RAM is recommended.


First, log on to the server on which the Goobi viewer is to be installed and obtain root rights:

sudo -i

Then generate a password for the Goobi viewer database and a token and store them as a variable in the session. The DNS name is also stored there:

export TOKEN=$(uuidgen)
export install=/tmp/install

Now install the following packages:

apt -y install git maven patch openjdk-11-jdk-headless tomcat9 mariadb-server apache2 ttf-mscorefonts-installer unzip zookeeperd

Finally, a temporary installation directory must be created and the Goobi viewer Connector, Goobi viewer Indexer and Goobi viewer Core Config Repository cloned. This includes the files required for the installation:

mkdir -p $install
cd $install
git clone https://github.com/intranda/goobi-viewer-connector.git
git clone https://github.com/intranda/goobi-viewer-indexer.git
git clone https://github.com/intranda/goobi-viewer-core-config.git

It is recommended to have set a DNS entry for the server at this time.

Variables and aliases

Add the following aliases to /root/.bash_aliases:

cat << "EOF" >>/root/.bash_aliases
alias cata='journalctl -u tomcat9 -n 1000 -f'
alias ct='chown tomcat:tomcat *'
alias ctr='chown -R tomcat:tomcat *'
alias ind='tail -n 1000 -f /opt/digiverso/logs/indexer.log'
alias vl='tail -n 1000 -f /opt/digiverso/logs/viewer.log'

Adopt the changes in the current session:

. /root/.bashrc

Install packages

Install the following packages including all dependencies:

apt -y install openjdk-11-jdk-headless tomcat9 mariadb-server apache2 ttf-mscorefonts-installer unzip

Create directory structure

The following commands create the necessary folder structure:

mkdir -p /opt/digiverso/{config/bin,indexer,logs,tomcat-lib,viewer/{abbyy,cmdi,deleted_mets,hotfolder,media,orig_lido,orig_denkxweb,success,ugc,alto,cms_media,error_mets,indexed_lido,mix,pdf,tei,updated_mets,cache,config/{PDFTitlePage,watermark},fulltext,indexed_mets,oai/token,ptif,themes,wc,bin}}
chown -R tomcat. /opt/digiverso

Configure services


SYSTEMD_EDITOR=tee systemctl edit tomcat9 << "EOF"

In /etc/default/tomcat9 adjust the memory under -Xmx to the available machine memory, choose reasonable garbage collector options and use urandom for faster Tomcat startup:

patch /etc/default/tomcat9 << "EOF"
@@ -11,6 +11,15 @@
# To enable remote debugging uncomment the following line.
# You will then be able to use a Java debugger on port 8000.
#JAVA_OPTS="${JAVA_OPTS} -agentlib:jdwp=transport=dt_socket,address=8000,server=y,suspend=n"
+JAVA_OPTS="-Djava.awt.headless=true -Xmx4g -Xms2g"
+JAVA_OPTS="${JAVA_OPTS} -XX:+ParallelRefProcEnabled"
+JAVA_OPTS="${JAVA_OPTS} -XX:+DisableExplicitGC"
+JAVA_OPTS="${JAVA_OPTS} -XX:+CMSClassUnloadingEnabled"
+JAVA_OPTS="${JAVA_OPTS} -Djava.security.egd=file:/dev/./urandom"
+JAVA_OPTS="${JAVA_OPTS} -Dfile.encoding='utf-8'"
# Java compiler to use for translating JavaServer Pages (JSPs). You can use all
# compilers that are accepted by Ant's build.compiler property.
@@ -20,4 +29,4 @@
# Whether to compress logfiles older than today's

Further configurations are made in /etc/tomcat9/server.xml. This will disable the appContextProtection for the JreMemoryLeakPreventionListener, set up a HTTP connector (with correct proxyName) and an AJP connector on localhost. A Crawler Session Manager Valve is also activated:

sed -e "s|VIEWER.EXAMPLE.ORG|${VIEWER_HOSTNAME}|g" << "EOF" | patch /etc/tomcat9/server.xml
@@ -66,59 +66,22 @@
APR (HTTP/AJP) Connector: /docs/apr.html
Define a non-SSL/TLS HTTP/1.1 Connector on port 8080
- <Connector port="8080" protocol="HTTP/1.1"
+ <Connector address="" port="8080" protocol="HTTP/1.1"
- redirectPort="8443" />
- <!-- A "Connector" using the shared thread pool-->
- <!--
- <Connector executor="tomcatThreadPool"
- port="8080" protocol="HTTP/1.1"
+ redirectPort="8443"
+ maxThreads="400"
+ URIEncoding="UTF-8"
+ enableLookups="false"
+ disableUploadTimeout="true"
+ proxyPort="80" />
+ <Connector address="" port="8009" protocol="AJP/1.3"
+ secretRequired="false"
- redirectPort="8443" />
- -->
- <!-- Define an SSL/TLS HTTP/1.1 Connector on port 8443
- This connector uses the NIO implementation. The default
- SSLImplementation will depend on the presence of the APR/native
- library and the useOpenSSL attribute of the
- AprLifecycleListener.
- Either JSSE or OpenSSL style configuration may be used regardless of
- the SSLImplementation selected. JSSE style configuration is used below.
- -->
- <!--
- <Connector port="8443" protocol="org.apache.coyote.http11.Http11NioProtocol"
- maxThreads="150" SSLEnabled="true">
- <SSLHostConfig>
- <Certificate certificateKeystoreFile="conf/localhost-rsa.jks"
- type="RSA" />
- </SSLHostConfig>
- </Connector>
- -->
- <!-- Define an SSL/TLS HTTP/1.1 Connector on port 8443 with HTTP/2
- This connector uses the APR/native implementation which always uses
- OpenSSL for TLS.
- Either JSSE or OpenSSL style configuration may be used. OpenSSL style
- configuration is used below.
- -->
- <!--
- <Connector port="8443" protocol="org.apache.coyote.http11.Http11AprProtocol"
- maxThreads="150" SSLEnabled="true" >
- <UpgradeProtocol className="org.apache.coyote.http2.Http2Protocol" />
- <SSLHostConfig>
- <Certificate certificateKeyFile="conf/localhost-rsa-key.pem"
- certificateFile="conf/localhost-rsa-cert.pem"
- certificateChainFile="conf/localhost-rsa-chain.pem"
- type="RSA" />
- </SSLHostConfig>
- </Connector>
- -->
+ maxThreads="400"
+ URIEncoding="UTF-8" />
- <!-- Define an AJP 1.3 Connector on port 8009 -->
- <!--
- <Connector protocol="AJP/1.3"
- address="::1"
- port="8009"
- redirectPort="8443" />
- -->
<!-- An Engine represents the entry point (within Catalina) that processes
every request. The Engine implementation for Tomcat stand alone
@@ -161,9 +124,14 @@
<!-- Access log processes all example.
Documentation at: /docs/config/valve.html
Note: The pattern used is equivalent to using pattern="common" -->
+ <!--
<Valve className="org.apache.catalina.valves.AccessLogValve" directory="logs"
prefix="localhost_access_log" suffix=".txt"
pattern="%h %l %u %t &quot;%r&quot; %s %b" />
+ -->
+ <Valve className="org.apache.catalina.valves.CrawlerSessionManagerValve"
+ crawlerUserAgents=".*[bB]ot.*|.*Yahoo! Slurp.*|.*Feedfetcher-Google.*|.*Apache-HttpClient.*|.*[Ss]pider.*|.*[Cc]rawler.*|.*nagios.*|.*Yandex.*"
+ sessionInactiveInterval="60"/>

Disable session persistence in /etc/tomcat9/context.xml by running the following patch:

patch /etc/tomcat9/context.xml << "EOF"
@@ -25,7 +25,5 @@
<!-- Uncomment this to disable session persistence across Tomcat restarts -->
- <!--
<Manager pathname="" />
- -->


Apache must be set up to make the viewer available externally. Activate the following modules for this:

a2enmod proxy_ajp
a2enmod proxy_http
a2enmod proxy_wstunnel
a2enmod rewrite
a2enmod expires
a2enmod headers

Place the following file at /etc/apache2/sites-available/${VIEWER_HOSTNAME}.conf:

sed -e "s|VIEWER.EXAMPLE.ORG|${VIEWER_HOSTNAME}|g" << "EOF" >/etc/apache2/sites-available/${VIEWER_HOSTNAME}.conf
<VirtualHost *:80>
ServerAdmin [email protected]
DocumentRoot /var/www
## make sure rewrite is enabled
RewriteEngine On
## search engines: do not follow certain urls
RewriteCond %{HTTP_USER_AGENT} ^.*bot.*$ [NC,OR]
RewriteCond %{HTTP_USER_AGENT} ^.*Yandex.*$ [NC,OR]
RewriteCond %{HTTP_USER_AGENT} ^.*spider.*$ [NC,OR]
RewriteCond %{HTTP_USER_AGENT} ^.*rawler.*$ [NC]
ReWriteRule ^(.*);jsessionid=[A-Za-z0-9]+(.*)$ $1$2 [L,R=301]
RewriteCond %{HTTP_USER_AGENT} ^.*bot.*$ [NC,OR]
RewriteCond %{HTTP_USER_AGENT} ^.*Yandex.*$ [NC,OR]
RewriteCond %{HTTP_USER_AGENT} ^.*spider.*$ [NC,OR]
RewriteCond %{HTTP_USER_AGENT} ^.*rawler.*$ [NC]
ReWriteRule ^(.*)viewer/!(.*)$ $1viewer/$2 [L,R=301]
RewriteCond %{HTTP_USER_AGENT} ^.*bot.*$ [NC,OR]
RewriteCond %{HTTP_USER_AGENT} ^.*Yandex.*$ [NC,OR]
RewriteCond %{HTTP_USER_AGENT} ^.*spider.*$ [NC,OR]
RewriteCond %{HTTP_USER_AGENT} ^.*rawler.*$ [NC]
ReWriteRule ^(.*)/[Ll][Oo][Gg]_(.*)$ $1/ [L,R=301]
## compress output
<IfModule mod_deflate.c>
AddOutputFilterByType DEFLATE text/plain text/html text/xml
AddOutputFilterByType DEFLATE text/css text/javascript
AddOutputFilterByType DEFLATE application/xml application/xhtml+xml
AddOutputFilterByType DEFLATE application/rss+xml
AddOutputFilterByType DEFLATE application/javascript application/x-javascript
## general proxy settings
ProxyPreserveHost On
SetEnv force-proxy-request-1.0 1
SetEnv proxy-nokeepalive 1
<Proxy *>
Require local
## CORS for IIIF
Header set Access-Control-Allow-Origin "*"
Header always set Access-Control-Allow-Methods "GET, OPTIONS"
Header always set Access-Control-Max-Age "600"
Header always set Access-Control-Allow-Headers "Authorization, Content-Type"
Header always set Access-Control-Expose-Headers "Content-Security-Policy, Location"
RewriteRule ^(.*)$ $1 [R=200,L]
# make sure ETag headers are forwarded correctly
# Post Apache 2.4 have a look at
# https://httpd.apache.org/docs/trunk/mod/mod_deflate.html#deflatealteretag
RequestHeader edit "If-None-Match" '(.*)-gzip"$' '$1", $1-gzip"'
## Enable WebSockets to check concurrent access
RewriteCond %{HTTP:Upgrade} websocket [NC]
RewriteCond %{HTTP:Connection} upgrade [NC]
RewriteRule /?(.*) ws://localhost:8080/$1 [P,L]
## Viewer
redirect 301 /index.html http://VIEWER.EXAMPLE.ORG/viewer/
redirect 301 /viewer http://VIEWER.EXAMPLE.ORG/viewer/
RewriteRule ^/viewer/oai/oai2.xsl$ http://VIEWER.EXAMPLE.ORG/viewer/oai2.xsl
ProxyPassMatch ^/viewer/(oai.*)$ ajp://localhost:8009/M2M/$1 retry=0
<LocationMatch ^/viewer/(oai.*)$>
Forcetype text/xml
ProxyPassReverse ajp://localhost:8009/M2M/$1
ProxyPassMatch ^/viewer/(sru.*)$ ajp://localhost:8009/M2M/$1 retry=0
<LocationMatch ^/viewer/(sru.*)$>
Forcetype text/xml
ProxyPassReverse ajp://localhost:8009/M2M/$1
ProxyPassMatch ^/viewer/(.*)$ ajp://localhost:8009/viewer/$1 retry=0
<LocationMatch ^/viewer/(.*)$>
ProxyPassReverse ajp://localhost:8009/viewer/$1
<IfModule mod_expires.c>
ExpiresActive on
ExpiresByType image/jpg "access plus 1 months"
ExpiresByType image/gif "access plus 1 months"
ExpiresByType image/jpeg "access plus 1 months"
ExpiresByType image/png "access plus 1 months"
ExpiresByType font/ttf "access plus 1 year"
ExpiresByType application/x-font-woff "access plus 1 year"
ExpiresByType application/vnd.ms-fontobject "access plus 1 year"
Require all granted
## solr
redirect 301 /solr http://VIEWER.EXAMPLE.ORG/solr/
<Location /solr/>
Require local
ProxyPass ajp://localhost:8009/solr/ retry=0
ProxyPassReverse ajp://localhost:8009/solr/
## logging
CustomLog /var/log/apache2/VIEWER.EXAMPLE.ORG_access.log combined
ErrorLog /var/log/apache2/VIEWER.EXAMPLE.ORG_error.log
# Possible values include: debug, info, notice, warn, error, crit, alert, emerg.
LogLevel warn

Enable the Goobi viewer vhost, disable the default vhost and restart the Apache web server:

a2dissite 000-default
a2ensite ${VIEWER_HOSTNAME}.conf
systemctl restart apache2.service

Finally create a robots.txt:

sed -e "s|VIEWER.EXAMPLE.ORG|${VIEWER_HOSTNAME}|g" << "EOF" >>/var/www/robots.txt
User-agent: *
Disallow: /viewer/content*action=pdf
Disallow: /viewer/search/
Disallow: /viewer/tags/
Disallow: /viewer/term/
Disallow: /viewer/oai
Disallow: /viewer/nextHit/
Disallow: /viewer/prevHit/
#Sitemap: http://VIEWER.EXAMPLE.ORG/viewer/sitemap_index.xml
Crawl-delay: 10

If UFW is used, Apache must be enabled:

ufw allow Apache\ Full

MySQL / MariaDB

Goobi viewer requires a database and its own user. This is created with the following command:

mysql -e "CREATE DATABASE viewer;
GRANT ALL PRIVILEGES ON viewer.* TO 'viewer'@'localhost' WITH GRANT OPTION;

The database schema is created automatically the first time the application is started.


This point is only relevant if Goobi workflow is/is also installed and this is not done on the same machine.

Then the /opt/digiverso/viewer folder must be exported to the Goobi workflow server. NFS is used for this. The adjustments for this are here:

apt install nfs-kernel-server -y

Export of the hotfolder:

export IP_GOOBI= # IP-Adresse des Goobi workflow Servers
echo "/opt/digiverso/viewer/hotfolder ${IP_GOOBI}/,sync,no_subtree_check,all_squash,anonuid=$(id -u tomcat),anongid=$(id -g tomcat))" >> /etc/exports
exportfs -av

If UFW is used, TCP 2049 must be enabled for NFSv4:

ufw allow from $IP_GOOBI proto tcp to any port 2049

The adjustments for Goobi workflow can be found in the installation instructions there:


The installation of the required components is described below.

Apache Solr

Download the current Solr 8.x version and unzip the installation script:

cd $install
wget http://archive.apache.org/dist/lucene/solr/8.5.2/solr-8.5.2.tgz
tar -xzf solr-8.5.2.tgz solr-8.5.2/bin/install_solr_service.sh --strip-components=2

Now create the folder /opt/digiverso/solr and install it there:

mkdir -p /opt/digiverso/solr/
./install_solr_service.sh solr-8.5.2.tgz -i /opt/digiverso/solr -d /opt/digiverso/solr -u solr -s solr -p 8983 -n

Now adjust the limits for the user solr:

cat << "EOF" >/etc/security/limits.d/solr.conf
solr hard nofile 65535
solr soft nofile 65535
solr hard nproc 65535
solr soft nproc 65535

To use streaming expressions Solr is not installed in the standalone version but in the SolrCloud variant. However, we only use one node. For the configuration Zookeeper is used:

Make sure that Zookeeper only listens on localhost:

patch /etc/zookeeper/conf/zoo.cfg << "EOF"
@@ -15,6 +15,7 @@
# the port at which the clients will connect
# specify all zookeeper servers
# The fist port is used by followers to connect to the leader
systemctl restart zookeeper

Then make the installation known to Solr and adjust the memory, the garbage collector options and the log level. Furthermore, the Solr should only listen on localhost:

patch /etc/default/solr.in.sh << "EOF"
@@ -30,3 +30,3 @@
# Increase Java Heap as needed to support your indexing / query needs
@@ -48,15 +48,15 @@
# These GC settings have shown to work well for a number of common Solr workloads
-#GC_TUNE=" \
-#-XX:SurvivorRatio=4 \
-#-XX:TargetSurvivorRatio=90 \
-#-XX:MaxTenuringThreshold=8 \
-#-XX:+UseConcMarkSweepGC \
-#-XX:ConcGCThreads=4 -XX:ParallelGCThreads=4 \
-#-XX:+CMSScavengeBeforeRemark \
-#-XX:PretenureSizeThreshold=64m \
-#-XX:+UseCMSInitiatingOccupancyOnly \
-#-XX:CMSInitiatingOccupancyFraction=50 \
-#-XX:CMSMaxAbortablePrecleanTime=6000 \
-#-XX:+CMSParallelRemarkEnabled \
-#-XX:+ParallelRefProcEnabled \
+GC_TUNE=" \
+-XX:SurvivorRatio=4 \
+-XX:TargetSurvivorRatio=90 \
+-XX:MaxTenuringThreshold=8 \
+-XX:+UseConcMarkSweepGC \
+-XX:ConcGCThreads=4 -XX:ParallelGCThreads=4 \
+-XX:+CMSScavengeBeforeRemark \
+-XX:PretenureSizeThreshold=64m \
+-XX:+UseCMSInitiatingOccupancyOnly \
+-XX:CMSInitiatingOccupancyFraction=50 \
+-XX:CMSMaxAbortablePrecleanTime=6000 \
+-XX:+CMSParallelRemarkEnabled \
#-XX:-OmitStackTraceInFastThrow etc.
@@ -66,3 +66,3 @@
# Leave empty if not using SolrCloud
@@ -73,3 +73,3 @@
# for production SolrCloud environments to control the hostname exposed to cluster state
@@ -115,3 +115,3 @@
# This is an alternative to changing the rootLogger in log4j2.xml
@@ -212,3 +212,3 @@
# host checking can be disabled by using the system property "solr.disable.shardsWhitelist"
-#SOLR_OPTS="$SOLR_OPTS -Dsolr.shardsWhitelist=http://localhost:8983,http://localhost:8984"
+SOLR_OPTS="$SOLR_OPTS -Dsolr.disable.shardsWhitelist"
@@ -234,2 +234,3 @@
+SOLR_OPTS="$SOLR_OPTS -Djetty.host="

Now make the Solr script executable:

chmod 755 /opt/digiverso/solr/solr/bin/solr

Now the configuration is stored in a new configset:

cd /opt/digiverso/solr/solr/server/solr/configsets/
cp -a _default/ goobiviewer
cd goobiviewer/conf/
rm managed-schema
wget -O schema.xml https://raw.githubusercontent.com/intranda/goobi-viewer-indexer/master/goobi-viewer-indexer/src/main/resources/other/schema.xml
cp lang/stopwords_de.txt lang/stopwords.txt
wget https://raw.githubusercontent.com/intranda/goobi-viewer-indexer/master/goobi-viewer-indexer/src/main/resources/other/mapping-ISOLatin1Accent.txt
patch solrconfig.xml << "EOF"
@@ -21,6 +21,8 @@
this file, see http://wiki.apache.org/solr/SolrConfigXml.
+ <schemaFactory class="ClassicIndexSchemaFactory"/>
<!-- In all configuration below, a prefix of "solr." for class names
is an alias that causes solr to search appropriate packages,
including org.apache.solr.(search|update|request|core|analysis)
@@ -57,6 +59,10 @@
<lib dir="./lib" />
+ <lib dir="${solr.install.dir:../../../..}/contrib/analysis-extras/lib" regex=".*\.jar" />
+ <lib dir="${solr.install.dir:../../../..}/contrib/analysis-extras/lucene-libs" regex=".*\.jar" />
+ <lib dir="${solr.install.dir:../../../..}/dist/" regex="solr-analysis-extras-\d.*\.jar" />
<!-- A 'dir' option by itself adds any files found in the directory
to the classpath, this is useful for including all jars in a
@@ -773,7 +779,7 @@
<initParams path="/update/**,/query,/select,/spell">
<lst name="defaults">
- <str name="df">_text_</str>
+ <str name="df">DEFAULT</str>
@@ -1108,7 +1114,7 @@
<!-- The update.autoCreateFields property can be turned to false to disable schemaless mode -->
- <updateRequestProcessorChain name="add-unknown-fields-to-the-schema" default="${update.autoCreateFields:true}"
+ <updateRequestProcessorChain name="add-unknown-fields-to-the-schema" default="${update.autoCreateFields:false}"
<processor class="solr.LogUpdateProcessorFactory"/>
<processor class="solr.DistributedUpdateProcessorFactory"/>
chown -R solr. *

Upload the new configset to Zookeeper now:

cd /opt/digiverso/solr/solr/
sudo -u solr bin/solr zk upconfig -n goobiviewer -d server/solr/configsets/goobiviewer/

The JTS library is still needed for the search for geocoordinates:

wget -O /opt/digiverso/solr/solr/server/solr-webapp/webapp/WEB-INF/lib/jts-core-1.17.0.jar https://github.com/locationtech/jts/releases/download/1.17.0/jts-core-1.17.0.jar
chown solr. /opt/digiverso/solr/solr/server/solr-webapp/webapp/WEB-INF/lib/jts-core-1.17.0.jar

Finally, start the service and create a new collection for the Goobi viewer with its configuration files:Solr in Tomcat:

systemctl start solr
sudo -u solr bin/solr create -c collection1 -n goobiviewer

Goobi viewer Indexer

First the application is compiled. Then copy the required files to /opt/digiverso/indexer/ and rename the indexerconfig_solr.xml to solr_indexerconfig.xml, control the Tomcat port and replace C: with /opt if necessary. Finally the systemd service unit is activated:

mvn -f $install/goobi-viewer-indexer/goobi-viewer-indexer package
mv $install/goobi-viewer-indexer/goobi-viewer-indexer/target/solr-Indexer.jar /opt/digiverso/indexer/solrIndexer.jar
sed -e 's|<solrUrl>.*</solrUrl>|<solrUrl>http://localhost:8983/solr/collection1</solrUrl>|' -e 's|C:|/opt|g' $install/goobi-viewer-indexer/goobi-viewer-indexer/target/classes/indexerconfig_solr.xml >/opt/digiverso/indexer/solr_indexerconfig.xml
mv $install/goobi-viewer-indexer/goobi-viewer-indexer/src/main/resources/other/solrindexer.service /etc/systemd/system/
systemctl enable solrindexer.service

Goobi viewer

Stop the Tomcat service:

systemctl stop tomcat9

Then put the following file to /etc/tomcat9/Catalina/localhost/viewer.xml:

sed -e "s|PW_SQL_VIEWER|${PW_SQL_VIEWER}|g" << "EOF" >/etc/tomcat9/Catalina/localhost/viewer.xml
<?xml version='1.0' encoding='UTF-8'?>
webAppMount="/WEB-INF/lib" />
webAppMount="/resources/themes" />
validationQuery="SELECT SQL_NO_CACHE 1"
url="jdbc:mysql://localhost/viewer?characterEncoding=UTF-8&amp;autoReconnect=true&amp;autoReconnectForPools=true" />

Now adjust the file rights so that the file is not deleted with an update:

chown -R root.tomcat /etc/tomcat9/Catalina
chmod -R g-w /etc/tomcat9/Catalina

The theme is integrated as an external theme:

mkdir -p /opt/digiverso/viewer/themes/
cd /opt/digiverso/viewer/themes/
git clone https://github.com/intranda/goobi-viewer-theme-reference.git

Now compile the Goobi viewer and move it to the expected location in the file system:

mvn -f /opt/digiverso/viewer/themes/goobi-viewer-theme-reference/goobi-viewer-theme-reference package
mv /opt/digiverso/viewer/themes/goobi-viewer-theme-reference/goobi-viewer-theme-reference/target/viewer.war /var/lib/tomcat9/webapps/viewer.war

Finally move the last configuration files and restart the Tomcat service:

mv $install/goobi-viewer-core-config/goobi-viewer-core-config/src/main/resources/install/* /opt/digiverso/viewer/config/
chown -R tomcat: /opt/digiverso/viewer/
systemctl start tomcat9

Connector - (OAI & SRU)

For the installation of the OAI and SRU interface the application has to be compiled first:

mvn -f $install/goobi-viewer-connector/goobi-viewer-connector package

After that the war file and the configuration files have to be moved to the expected location:

mv $install/goobi-viewer-connector/goobi-viewer-connector/target/M2M.war /opt/digiverso/viewer/bin/M2M.war
mv $install/goobi-viewer-connector/goobi-viewer-connector/target/M2M/WEB-INF/classes/*.xsl /opt/digiverso/viewer/oai/
chown -R tomcat: /opt/digiverso/viewer/bin/
chown -R tomcat: /opt/digiverso/viewer/oai/

Then put the following file to /etc/tomcat9/Catalina/localhost/M2M.xml:

cat << "EOF" >/etc/tomcat9/Catalina/localhost/M2M.xml
<?xml version="1.0" encoding="UTF-8"?>

Extraction of config_oai.xml:

unzip -p /opt/digiverso/viewer/bin/M2M.war WEB-INF/classes/config_oai.xml > /opt/digiverso/viewer/config/config_oai.xml

Adjust the settings to the correct URL so that the resolver links work and ensure that the port of the tomcat on which the Solr service is running is 8080:

sed -i -e "s|http://localhost:8080/viewer/oai|http://${VIEWER_HOSTNAME}/viewer/oai|g" -e "s|http://localhost:8080/viewer/resolver|http://${VIEWER_HOSTNAME}/viewer/resolver|g" -e "s|http://localhost:8080/viewer/piresolver|http://${VIEWER_HOSTNAME}/viewer/piresolver|g" -e "s|http://localhost:8081/solr|http://localhost:8983/solr|g" /opt/digiverso/viewer/config/config_oai.xml



A cronjob must be set up for regular tasks:

sed -e "s|VIEWER.EXAMPLE.ORG|${VIEWER_HOSTNAME}|g" -e "s|TOKEN|${TOKEN}|g" << "EOF" >/etc/cron.d/intranda-goobiviewer
# Regular cron jobs for the Goobi viewer
## This REST call triggers the email notification about new search hits for users,
## that enabled notifications for saved searches
42 8,12,17 * * * root curl -s -H "Content-Type: application/json" -H "token:TOKEN" -d '{"type":"NOTIFY_SEARCH_UPDATE"}' http://localhost:8080/viewer/api/v1/tasks/ 1>/dev/null
## This REST call creates an XML sitemap for the Goobi viewer instance. Please always
## call it on it's external URL because otherwise the protocol (http/https) might not
## be detected correctly
18 1 * * * root curl -s -H "Content-Type: application/json" -H "token:TOKEN" -d '{"type":"UPDATE_SITEMAP"}' http://localhost:8080/viewer/api/v1/tasks/ 1>/dev/null
## These two scripts pull the theme git repository regulary. The @daily part is only
## a reminder for the 1-minute schedule
#*/1 * * * * root cd /opt/digiverso/viewer/themes/goobi-viewer-theme-reference; git pull | grep -v -e "Already up.to.date." -e "Bereits aktuell."
#@daily root echo "Please look at the git checkout interval for the Goobi viewer theme" | mail -s "Reference: Theme repository is checked out every minute" [email protected]
## Optimize the Solr search index once a month
@monthly root curl -s 'http://localhost:8080/solr/collection1/update?optimize=true&waitFlush=false'


In the local config_viewer.xml different settings have to be stored:

sed -e "s|VIEWER.EXAMPLE.ORG|${VIEWER_HOSTNAME}/viewer|g" -e "s|TOKEN|${TOKEN}|g" << "EOF" >/opt/digiverso/viewer/config/config_viewer.xml
<?xml version="1.0" encoding="UTF-8" ?>
<theme mainTheme="reference" discriminatorField="">
<title>Goobi viewer RSS Feed</title>
<description>new items</description>
<copyright>(c) Goobi viewer using institution </copyright>

Create user account

The Goobi viewer has a backend. The following command inserts a test account with the user name [email protected] and the password viewer into the database:

mysql -e 'USE viewer;
INSERT INTO users (active,email,password_hash,score,superuser) VALUES (1,"[email protected]","$2a$10$Z5GTNKND9ZbuHt0ayDh0Remblc7pKUNlqbcoCxaNgKza05fLtkuYO",0,1);'

Optionally, you can generate your own password hash as follows:

python3 -c 'from passlib.hash import bcrypt; print(bcrypt.using(rounds=10, ident="2a").hash("PASSWORD"))'