Return-Path: X-Original-To: apmail-hadoop-mapreduce-user-archive@minotaur.apache.org Delivered-To: apmail-hadoop-mapreduce-user-archive@minotaur.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 6B723D80F for ; Tue, 28 Aug 2012 16:31:06 +0000 (UTC) Received: (qmail 66587 invoked by uid 500); 28 Aug 2012 16:31:01 -0000 Delivered-To: apmail-hadoop-mapreduce-user-archive@hadoop.apache.org Received: (qmail 66508 invoked by uid 500); 28 Aug 2012 16:31:01 -0000 Mailing-List: contact user-help@hadoop.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: user@hadoop.apache.org Delivered-To: mailing list user@hadoop.apache.org Received: (qmail 66500 invoked by uid 99); 28 Aug 2012 16:31:01 -0000 Received: from athena.apache.org (HELO athena.apache.org) (140.211.11.136) by apache.org (qpsmtpd/0.29) with ESMTP; Tue, 28 Aug 2012 16:31:01 +0000 X-ASF-Spam-Status: No, hits=1.5 required=5.0 tests=FSL_RCVD_USER,HTML_MESSAGE,RCVD_IN_DNSWL_LOW,SPF_PASS X-Spam-Check-By: apache.org Received-SPF: pass (athena.apache.org: domain of modemide@gmail.com designates 209.85.214.48 as permitted sender) Received: from [209.85.214.48] (HELO mail-bk0-f48.google.com) (209.85.214.48) by apache.org (qpsmtpd/0.29) with ESMTP; Tue, 28 Aug 2012 16:30:55 +0000 Received: by bkty5 with SMTP id y5so1846908bkt.35 for ; Tue, 28 Aug 2012 09:30:34 -0700 (PDT) Received: by 10.204.152.211 with SMTP id h19mr5363818bkw.45.1346171434159; Tue, 28 Aug 2012 09:30:34 -0700 (PDT) Received: from mail-lpp01m010-f48.google.com (mail-lpp01m010-f48.google.com [209.85.215.48]) by mx.google.com with ESMTPS id n5sm13202350bkv.14.2012.08.28.09.30.31 (version=SSLv3 cipher=OTHER); Tue, 28 Aug 2012 09:30:32 -0700 (PDT) Received: by lagr15 with SMTP id r15so3810122lag.35 for ; Tue, 28 Aug 2012 09:30:30 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20120113; h=mime-version:in-reply-to:references:from:date:message-id:subject:to :content-type; bh=VVXNH4HK3wGWrRS8PFg39qP9fIDqUeHoyELmLuaLDQI=; b=u4j2osaB2FE/PFxjHyu3tQpjwD7oyUfpzWnL8w5sE+QFBbfKdxa0gBL+ry7tHsvHTE 1+gzULBqcbn1Ouwvqp27vrWrj4RS5UlbNCQVyXUQYn+VCc5S8r9rMRFwb4TqWZAe7vz9 dOWH160AuhL5NiSdPf/eVU95yVoo9ZMhCOM8I6cFeBqbkURfRrxeuZPjwHdsc2sqw9cg JAIVqU1fD7OiAyJufUYj9Vg1wHs1QEuLxBr3Bt4e/JbAGzLTuMrnuW8pOerUipISrxth 3RSdiaYC/M6KfPOP8qCB5HirNT4BVNQ9UQ8Q/kdOI8wM3wkWy1fdUn62cW34uu3UVvjU QLWQ== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=20120113; h=mime-version:in-reply-to:references:from:date:message-id:subject:to :content-type:x-gm-message-state; bh=VVXNH4HK3wGWrRS8PFg39qP9fIDqUeHoyELmLuaLDQI=; b=P5KTzEda9jNN4ulG2c71djJF9EPPd8X/NYe0ApAPVXl58fku3ZKY8gqFd44qjAkIfX biBJp5Xg0v9LzPjMkKOw5rwYMHe3fCOJevdH5uusvclf6UoxwoFbQ4ZtzdnZvZbhZOKk eFX1c8ZD67xdCpmcsuRFL628B6p5ZBXrXMTBxeyFdhDsqzOLtUXJFJQVtmCkQUDmhpeb ePQOm5WUNcZ3YUvu+xgJ6pA2VHi+UOYTyNjT/Z0951V8my+j26HAivNuZErG5iieH8ym CAHKTs5+gDP1rZf4LcqrsAuXi3nIrEDyAOTCaV7RBJ7dEoJ8JUzLY7rOihDo6KUEFPRs tHIg== Received: by 10.112.82.170 with SMTP id j10mr4901936lby.8.1346171430768; Tue, 28 Aug 2012 09:30:30 -0700 (PDT) MIME-Version: 1.0 Received: by 10.112.147.225 with HTTP; Tue, 28 Aug 2012 09:30:10 -0700 (PDT) In-Reply-To: References: From: modemide Date: Tue, 28 Aug 2012 12:30:10 -0400 Message-ID: Subject: Re: Hadoop and MainFrame integration To: user@hadoop.apache.org Content-Type: multipart/alternative; boundary=f46d04016c23f00f7304c855f551 X-Gm-Message-State: ALoCoQleZnVSc1BvnOd/w93mq3c+UKoZcQqyuPuYtWna8Lurect1R4UD1jv96a2WdGFzDSrY532a X-Virus-Checked: Checked by ClamAV on apache.org --f46d04016c23f00f7304c855f551 Content-Type: text/plain; charset=windows-1252 Content-Transfer-Encoding: quoted-printable At some point in the work flow you're going to have to transfer the file from the mainframe to the Hadoop cluster for processing, and then send it back for storage on the mainframe. You should be able to automate the process of sending the files back and forth. It's been my experience that it's often faster to process and sort large files on a Hadoop cluster even while factoring in the cost to transfer to/from the mainframe. Hopefully that answers your question. If not, are you looking to actually use Hadoop to process files in place on the mainframe? That concept conflicts with my understanding of Hadoop. On Tue, Aug 28, 2012 at 12:24 PM, Siddharth Tiwari < siddharth.tiwari@live.com> wrote: > Hi Users. > > We have flat files on mainframes with around a billion records. We need t= o > sort them and then use them with different jobs on mainframe for report > generation. I was wondering was there any way I could integrate the > mainframe with hadoop do the sorting and keep the file on the sever itsel= f > ( I do not want to ftp the file to a hadoop cluster and then ftp back the > sorted file to Mainframe as it would waste MIPS and nullify the advantage > ). This way I could save on MIPS and ultimately improve profitability. > > Thank you in advance > > > **------------------------** > *Cheers !!!* > *Siddharth Tiwari* > Have a refreshing day !!! > *"Every duty is holy, and devotion to duty is the highest form of worship > of God.=94 * > *"Maybe other people will try to limit me but I don't limit myself"* > --f46d04016c23f00f7304c855f551 Content-Type: text/html; charset=windows-1252 Content-Transfer-Encoding: quoted-printable At some point in the work flow you're going to have to transfer the fil= e from the mainframe to the Hadoop cluster for processing, and then send it= back for storage on the mainframe.

You should be able t= o automate the process of sending the files back and forth. =A0

It's been my experience that it's often faster = to process and sort large files on a Hadoop cluster even while factoring in= the cost to transfer to/from the mainframe.

Hopefully that answers your question. =A0If not, are you looking to actuall= y use Hadoop to process files in place on the mainframe? =A0That concept co= nflicts with my understanding of Hadoop.

On Tue, Aug 28, 2012 at 12:24 PM, Siddh= arth Tiwari <siddharth.tiwari@live.com> wrote:
Hi Users.

We have flat files= on mainframes with around a billion records. We need to sort them and then= use them with different jobs on mainframe for report generation. I was won= dering was there any way I could integrate the mainframe with hadoop do the= sorting and keep the file on the sever itself ( I do not want to ftp the f= ile to a hadoop cluster and then ftp back the sorted file to Mainframe as i= t would waste MIPS and nullify the advantage ). This way I could save on MI= PS and ultimately improve profitability.

Thank you in advance


= *------------------------*
Cheers != !!
Siddh= arth Tiwari
Have a refreshing d= ay !!!
"Every duty is holy, a= nd devotion to duty is the highest form of worship of God.=94 "Maybe= other people will try to limit me but I don't limit myself"


--f46d04016c23f00f7304c855f551--