Return-Path: X-Original-To: apmail-cassandra-user-archive@www.apache.org Delivered-To: apmail-cassandra-user-archive@www.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 4418E178FD for ; Tue, 17 Mar 2015 03:07:47 +0000 (UTC) Received: (qmail 97620 invoked by uid 500); 17 Mar 2015 03:07:44 -0000 Delivered-To: apmail-cassandra-user-archive@cassandra.apache.org Received: (qmail 97566 invoked by uid 500); 17 Mar 2015 03:07:44 -0000 Mailing-List: contact user-help@cassandra.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: user@cassandra.apache.org Delivered-To: mailing list user@cassandra.apache.org Received: (qmail 97556 invoked by uid 99); 17 Mar 2015 03:07:44 -0000 Received: from athena.apache.org (HELO athena.apache.org) (140.211.11.136) by apache.org (qpsmtpd/0.29) with ESMTP; Tue, 17 Mar 2015 03:07:44 +0000 X-ASF-Spam-Status: No, hits=2.7 required=5.0 tests=FREEMAIL_ENVFROM_END_DIGIT,FREEMAIL_REPLYTO_END_DIGIT,HTML_MESSAGE,RCVD_IN_DNSWL_NONE X-Spam-Check-By: apache.org Received-SPF: error (athena.apache.org: local policy) Received: from [106.10.151.62] (HELO nm35-vm7.bullet.mail.sg3.yahoo.com) (106.10.151.62) by apache.org (qpsmtpd/0.29) with ESMTP; Tue, 17 Mar 2015 03:07:38 +0000 DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=yahoo.co.in; s=s2048; t=1426561616; bh=y14UfvmwfeoUCuyNfuZx2iZSG89rG0LOeYOAwyg9+24=; h=Date:From:Reply-To:To:In-Reply-To:References:Subject:From:Subject; b=ZraCRrOzEin4WWQ1lPvBqxypQrqN+sbNzXUdO0109DlF5VKABPqvydJfH+qE6BVfT50lmO6FIgQL7SG4bi8mBKHlELReoFcoIzq9v7z5WAEcyedBstjQ0kJh10SY3q7p9WXKwEMVaJCHwyzXFId7/YX4tPpXrxDl3geaJhXD2eSLZYVu9GgND+lFwv4JdeExOUFxDshL+dr5t64QkwllfjgobvftygG0TAMl6LoTRoHArxW/E1ZQoA+9WFjDov+3uh5u8AItaI70xYRlsxEnxSsT8VvSRZLXZbZ2C1xPw2btUx4OjqUN25z5G8YZgsy7iS4QxDCG22IBAQOsWD/0Og== Received: from [106.10.166.120] by nm35.bullet.mail.sg3.yahoo.com with NNFMP; 17 Mar 2015 03:06:56 -0000 Received: from [106.10.150.29] by tm9.bullet.mail.sg3.yahoo.com with NNFMP; 17 Mar 2015 03:06:56 -0000 Received: from [127.0.0.1] by omp1030.mail.sg3.yahoo.com with NNFMP; 17 Mar 2015 03:06:56 -0000 X-Yahoo-Newman-Property: ymail-3 X-Yahoo-Newman-Id: 423259.46838.bm@omp1030.mail.sg3.yahoo.com X-YMail-OSG: fgXPu4sVM1kGE2f9khIqzI.okveGCY__kSK0yxa.ZGdmUFFh6eKiYLB3hdoaSaM P9.MGpqKP9ooFYZdEvHQtzfmjSw4U0eE5Lv3vwcRFkxJx4N3osjpUCPXmM5U.ryJrW.lEigZitsm d16hOQ1x6NtL9iEXiYBmmFAB69vgCq_OFqsQ_Ki2W7r6rSjkyugXckf7V0O8faF8f6PWphZvcUYj kJ6zkFF.46oKeZEuhSEyrvrVukjp7cVyK4E_8c2711M3rI069BWSq0Gju_.MBudOAy.8OeB8Ab89 OZN6B7uZC9Zk7H3rwYgLGNZFtdjefHVhIHJwKLK_f1WviKErSalHQTAMQ8RSKCNodilu0TmIO1VX xeVe4EMJEd6r6aVBhGbxh4c935jf6mrYdzu9zwpYtAQWDrzzbGOVo9HEAj5p5J8Oth.sqNuFmP6g YBjvlq49B8SAV2JlxDUuvllwq.SFGRfPOFq2o3gZ3VNuUEIL2otwptAbe2D_g4tEeREH8NHjHrqv zbm5olaonSu0- Received: by 106.10.196.176; Tue, 17 Mar 2015 03:06:55 +0000 Date: Tue, 17 Mar 2015 03:06:55 +0000 (UTC) From: Anuj Wadehra Reply-To: Anuj Wadehra To: "user@cassandra.apache.org" Message-ID: <663448416.574648.1426561615202.JavaMail.yahoo@mail.yahoo.com> In-Reply-To: <1426558849.87177.YahooMailMobile@web192904.mail.sg3.yahoo.com> References: <1426558849.87177.YahooMailMobile@web192904.mail.sg3.yahoo.com> Subject: Run Mixed Workload using two instances on one node MIME-Version: 1.0 Content-Type: multipart/alternative; boundary="----=_Part_574647_1333801493.1426561615198" X-Virus-Checked: Checked by ClamAV on apache.org ------=_Part_574647_1333801493.1426561615198 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: quoted-printable Hi, We are trying to Decouple our Reporting DB from OLTP. Need urgent help on t= he feasibility of proposed solution for PRODUCTION. Use Case: Currently, our OLTP and Reporting application and DB are same. So= me CF are used for both OLTP and Reporting while others are solely used for= Reporting.Every business transaction synchronously updates the main OLTP C= F and asynchronously updates other Reporting CFs. Problem Statement: 1. Decouple Reporting and OLTP such that Reporting load can't impact=C2=A0 = OLTP performance. 2. Scaling of Reporting=C2=A0 and OLTP modules must be independent 3. OLTP client should not update all Reporting CFs. We generate Data Record= s on File sytem/shared disk.Reporting should use these Records to create Re= porting DB. 4. Small customers may do OLTP and Reporting on same 3-node cluster. Bigger= customers can be given an option to have dedicated OLTP and Reporting node= s. So, standard Hardware box should be usable for 3 deployments (OLTP,Repor= ting or OLTP+Reporting) Note: Reporting is ad-hoc, may involve full table scans and does not involv= e Analytics. Data size is huge 2TB (OLTP+Reporting) per node. Hardware : Standard deployment -3 node cluster with each node having 24 cor= es, 64GB RAM, 400GB * 6 SSDs in RAID5 Proposed Solution: 1. Split OLTP and Reporting clients into two application components. 2. For small deployments where more than 3 nodes are not required: =C2=A0 =C2=A0 A. Install 2 Cassandra instances on each node one for OLTP an= d other for Reporting =C2=A0 =C2=A0 B. To distribute I/O load in 2:1 --Remove RAID5 (as Cassandra= offers replication) and assign 4 disks as JBod for OLTP and 2 disks for Re= porting =C2=A0 =C2=A0 C. RAM is abundant and often under-utilized , so assign 8GB e= ach for 2 Cassandra instance =C2=A0 =C2=A0 D. To make sure that Reporting is not able to overload CPU, t= une concurrent_reads,concurrent_writes=20 OLTP client will only write to OLTP DB and generate DB record. Reporting c= lient will poll FS and populate Reporting DB in required format. 3. Larger customers can have Reporting clients and DB on dedicated physical= nodes with all resources. Key Questions: Is it ok to run 2 Cassandra instances on one node in Production system and = limit CPU Usage,Disk I/O and RAM as suggested above? Any other solution for above mentioned problem statement? Thanks Anuj ------=_Part_574647_1333801493.1426561615198 Content-Type: text/html; charset=UTF-8 Content-Transfer-Encoding: quoted-printable


Hi,

We are trying to Decoup= le our Reporting DB from OLTP. Need urgent help on the feasibility of propo= sed solution for PRODUCTION.

Use Case: Currently, our OLTP and Repor= ting application and DB are same. Some CF are used for both OLTP and Report= ing while others are solely used for Reporting.Every business transaction s= ynchronously updates the main OLTP CF and asynchronously updates other Repo= rting CFs.

Problem Statement:
1. Decouple Reporting and OLTP such= that Reporting load can't impact  OLTP performance.
2. Scaling of = Reporting  and OLTP modules must be independent
3. OLTP client shou= ld not update all Reporting CFs. We generate Data Records on File sytem/sha= red disk.Reporting should use these Records to create Reporting DB.
4. S= mall customers may do OLTP and Reporting on same 3-node cluster. Bigger cus= tomers can be given an option to have dedicated OLTP and Reporting nodes. S= o, standard Hardware box should be usable for 3 deployments (OLTP,Reporting= or OLTP+Reporting)

Note: Reporting is ad-hoc, may involve full tabl= e scans and does not involve Analytics. Data size is huge 2TB (OLTP+Reporti= ng) per node.

Hardware : Standard deployment -3 node cluster with ea= ch node having 24 cores, 64GB RAM, 400GB * 6 SSDs in RAID5

Proposed = Solution:
1. Split OLTP and Reporting clients into two application compo= nents.
2. For small deployments where more than 3 nodes are not required= :
    A. Install 2 Cassandra instances on each node one for OL= TP and other for Reporting
    B. To distribute I/O load in 2:= 1 --Remove RAID5 (as Cassandra offers replication) and assign 4 disks as JB= od for OLTP and 2 disks for Reporting
    C. RAM is abundant a= nd often under-utilized , so assign 8GB each for 2 Cassandra instance
&n= bsp;   D. To make sure that Reporting is not able to overload CPU, tun= e concurrent_reads,concurrent_writes
OLTP client will only write to OL= TP DB and generate DB record. Reporting client will poll FS and populate Re= porting DB in required format.
3. Larger customers can have Reporting cl= ients and DB on dedicated physical nodes with all resources.

Key Que= stions:
Is it ok to run 2 Cassandra instances on one node in Production = system and limit CPU Usage,Disk I/O and RAM as suggested above?
Any othe= r solution for above mentioned problem statement?



Thanks
= Anuj


------=_Part_574647_1333801493.1426561615198--