lucene-solr-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Shawn Heisey <apa...@elyograg.org>
Subject Re: Memory requirements for TLOGs (7.3.1)
Date Fri, 20 Jul 2018 13:17:34 GMT
On 7/18/2018 6:33 PM, Ash Ramesh wrote:
> Thanks for the quick responses Shawn & Erick! Just to clarify another few
> points:
>  1. Does having a larger heap size impact ingesting additional documents to
> the index (all CRUD operations) onto a TLOG?

It's extremely difficult, maybe even impossible, for anyone on this list
to predict whether performance will be improved by increasing the heap,
at least not without some really concrete information from the system. 
If you shared your GC log and whatever activity you want to improve was
happening during that log creation, I could probably answer that
question for your specific server.

>  2. Does having a larger ram configured machine (in this case 32gb) affect
> ingestion on TLOGS also?

Having more memory for the OS disk cache does not usually improve
indexing performance.  The only kind of memory that is likely to matter
for that is heap memory.  Once you reach a sufficient heap size,
increasing it further won't help and might actually hurt performance.

>  3. We are currently routing queries via Amazon ASG / Load Balancer. Is
> this one of the recommended ways to set up SOLR infrastructure?

If your client software is not cloud-aware, you'll want an external load
balancer.  The only cloud-aware client that I know for sure exists is
the Java client, which is part of Solr itself as well as a standalone
client.  I did hear once about a cloud-aware client under development
for Python, but I do not know the status of that client -- it would be
third-party software.

Because you're using an external load balancer, you could list only the
PULL replicas in the load balancer back end configuration, and include
the preferLocalShards parameter on the request, so that SolrCloud will
not load balance the requests further.

Thanks,
Shawn


Mime
View raw message