Elastic Path Search Server provides product discovery (searching and browsing) for Cortex and catalog management for Commerce Manager. This section provides advice for how to scale a Search Server deployment in a production environment. For staging environments, it is usually sufficient to have a single search master and no slaves.
Search Server is implemented using a master-slave topology. Only a single master can run at a time, otherwise database conflicts occur updating the
Two common deployment topologies:
Small Deployment Topology
The simplest topology suitable for small deployments is to deploy a Search slave on every Cortex node. Then access Search Server through the
Large Deployment Topology
Large deployments typically have a separate cluster of Search slaves behind a stateless load balancer. The cluster is accessed through the load balancer URL.
As a starting point, Elastic Path recommends deploying one Search slave for every three Cortex servers to support catalog searching and browsing. This ratio may vary depending on the searching and browsing patterns for the store and customizations that change search server usage. A large deployment example is shown below:
The Search master and slave are deployed using the same WAR file. By default, the WAR is configured to run as a master node with replication disabled. The steps below configure both master and slave.
solrHomedirectory (master & slave)
solrHomedirectory from the Search WAR to an external location. Elastic Path recommends to include the
solrHomein the deployment package as a separate artifact and then deploy the artifact to the desired location on the server. The benefits of this method are:
- You can add replication configuration without invading the WAR
- Search indexes are not deleted when a new WAR is deployed
solrHomeis deployed to the desired location, add the following JVM (Java Virtual Machine) parameters to the appserver startup script:
/home/ep/solrHome/data/index-category =/home/ep/solrHome/data/index-cmuser =/home/ep/solrHome/data/index-customer =/home/ep/solrHome/data/index-product =/home/ep/solrHome/data/index-promotion =/home/ep/solrHome/data/index-shippingservicelevel =/home/ep/solrHome/data/index-sku=
Configure master replication
Each config file in
<xi:include>tag that points to a corresponding replication configuration file.
<!-- include master/slave replication settings --> <xi:include href="file:/etc/ep/product.solr.replication.config.xml" xmlns:xi="http://www.w3.org/2001/XInclude"> <xi:fallback> <!-- Currently this reference is relative to the CWD. As of SOLR 3.1 it will be relative to the current XML file's directory --> <xi:include href="../replication/product.solr.replication.config.xml" xmlns:xi="http://www.w3.org/2001/XInclude"> <!-- Silent fallback if no external configuration found --> <xi:fallback /> </xi:include> </xi:fallback> </xi:include>
In this case, Solr first looks for
/etc/epand then in the
../replicationrelative path.The relative path is relative to where the Search server shell script is run.
If necessary, you can change the
<xi:include>block in the Solr config files if
/etc/epis not an acceptable location.
Here is an example of the
<requestHandler name="/replication" class="solr.ReplicationHandler"> <lst name="master"> <!--Replicate on 'startup' and 'commit'. 'optimize' is also a valid value for replicateAfter. --> <str name="replicateAfter">startup</str> <str name="replicateAfter">commit</str> <!-- <str name="replicateAfter">optimize</str> --> <!--Create a backup after 'optimize'. Other values can be 'commit', 'startup'. It is possible to have multiple entries of this config string. Note that this is just for backup, replication does not require this. --> <!-- <str name="backupAfter">optimize</str> --> <!--If configuration files need to be replicated give the names here, separated by comma --> <str name="confFiles">schema.xml,stopwords.txt,elevate.xml</str> <!--The default value of reservation is 10 secs.See the documentation below . Normally , you should not need to specify this --> <str name="commitReserveDuration">00:00:10</str> </lst> </requestHandler>
Create separate config files for each index you want replicated. You don’t have to replicate all indexes; just the ones that are actually needed. Here is the full list of config files:
Index Replication Config File Index Used By Category
Cortex, CM CM User
Cortex, CM Promotion
For each replication configuration file on the master, create a corresponding configuration file on the slave. Here is an example of the
<requestHandler name="/replication" class="solr.ReplicationHandler" > <lst name="slave"> <str name="enable">true</str> <!--fully qualified url for the replication handler of master . It is possible to pass on this as a request param for the fetchindex command--> <str name="masterUrl">http://venus.elasticpath.net:8080/searchserver/product/replication</str> <!--Interval in which the slave should poll master .Format is HH:mm:ss . If this is absent slave does not poll automatically. But a fetchindex can be triggered from the admin or the http API --> <str name="pollInterval">00:00:20</str> <!-- THE FOLLOWING PARAMETERS ARE USUALLY NOT REQUIRED--> <!--to use compression while transferring the index files. The possible values are internal|external if the value is 'external' make sure that your master Solr has the settings to honour the accept-encoding header. see here for details http://wiki.apache.org/solr/SolrHttpCompression If it is 'internal' everything will be taken care of automatically. USE THIS ONLY IF YOUR BANDWIDTH IS LOW . THIS CAN ACTUALLY SLOWDOWN REPLICATION IN A LAN--> <str name="compression">internal</str> <!--The following values are used when the slave connects to the master to download the index files. Default values implicitly set as 5000ms and 10000ms respectively. The user DOES NOT need to specify these unless the bandwidth is extremely low or if there is an extremely high latency--> <str name="httpConnTimeout">5000</str> <str name="httpReadTimeout">10000</str> <!-- If HTTP Basic authentication is enabled on the master, then the slave can be configured with the following <str name="httpBasicAuthUser">username</str> <str name="httpBasicAuthPassword">password</str> --> </lst> </requestHandler>
Disable indexing on slaves
Add the following line to the
ep.propertiesfile on each slave server to stop the slaves from indexing themselves.
Enable the Search Master
Add the following line to
ep.propertieson the server containing the search master to identify which server contains the Search master.
Configure the search server URLs through the system configuration setting
COMMERCE/SYSTEM/SEARCH/searchHost. You can also use the
-Dep.search.master.urlJava system properties. The URLs you set depends on your deployment topology.
In a simple deployment topology, you only need to configure the default value to point to
Slave cluster topology
If the search slaves are clustered behind a load balancer, then you need two setting values or Java system properties. The default value identifies the slave cluster and an additional value identifies the search master.
Setting Context Path Java System Property Value
URL of the load balancer for the slave cluster master
URL of the search master
Updating Search Server Cluster
Halt any user updates to the system and ensure no changes are in progress through Commerce Manager, the Import/Export tool, or the Data Sync tool.
In Commerce Manager, ensure the indexing activity is complete for all indexes. You can find the index status in the Configuration section under Search Indexes.
Back up all existing indexes by calling the following links. These links will call the Solr replication API:
http://master_host:port/searchserver/category/replication?command=backup http://master_host:port/searchserver/cmuser/replication?command=backup http://master_host:port/searchserver/customer/replication?command=backup http://master_host:port/searchserver/dynamicContentDelivery/replication?command=backup http://master_host:port/searchserver/product/replication?command=backup http://master_host:port/searchserver/promotion/replication?command=backup http://master_host:port/searchserver/shippingservicelevel/replication?command=backup http://master_host:port/searchserver/sku/replication?command=backup
Disable replication on the slave servers by calling the following links. Make sure this is done for all slaves.
http://slave_host:port/searchserver/category/replication?command=disablepoll http://slave_host:port/searchserver/cmuser/replication?command=disablepoll http://slave_host:port/searchserver/customer/replication?command=disablepoll http://slave_host:port/searchserver/dynamicContentDelivery/replication?command=disablepoll http://slave_host:port/searchserver/product/replication?command=disablepoll http://slave_host:port/searchserver/promotion/replication?command=disablepoll http://slave_host:port/searchserver/shippingservicelevel/replication?command=disablepoll http://slave_host:port/searchserver/sku/replication?command=disablepoll
Shut down, update, and restart the master search server:
- Shut down the application server
- Redeploy the updated search server war
- Restart the application server.
If your search indexes are normally located within the
searchserver/WEB-INFfolder, you may need to restore them from your back up location or do a full rebuild
Using Commerce Manager, rebuild any indexes that would be affected by the update.
For step-by-step instructions on how to do this, see Requesting Search Index Rebuilds.
Verify the master indexes.
Since the master server is a full search instance, you can check its state by doing one of the following:
- Issuing manual Solr queries.
- By redirecting Cortex server instance to point to the master instead of a slave.
Do one of the following:
If you are applying code updates to the slave servers, do the following for all slaves:
Remove the slave from the search cluster
Shutdown the server.
Redeploy the updated search server war.
Restart the server.
The restarted slave will immediately begin polling the master for index updates on startup.
Restore the slave to the search cluster.
If you are not updating the slave servers, resume index replication by calling the following links for all slaves:
http://slave_host:port/searchserver/category/replication?command=enablepoll http://slave_host:port/searchserver/cmuser/replication?command=enablepoll http://slave_host:port/searchserver/customer/replication?command=enablepoll http://slave_host:port/searchserver/dynamicContentDelivery/replication?command=enablepoll http://slave_host:port/searchserver/product/replication?command=enablepoll http://slave_host:port/searchserver/promotion/replication?command=enablepoll http://slave_host:port/searchserver/shippingservicelevel/replication?command=enablepoll http://slave_host:port/searchserver/sku/replication?command=enablepoll
Search Server Cluster Checklist
To verify if your search server cluster is running properly:
- Check if the indexes are being replicated to the slave server
- Ensure requests are being redirected to the slave
- Make sure the update requests are only redirected to the master server
- Trace requests in the Apache HTTP server’s log and Tomcat server’s access logs
- See if requests are redirected to another server if a server is unavailable