Return-Path: X-Original-To: apmail-hadoop-mapreduce-user-archive@minotaur.apache.org Delivered-To: apmail-hadoop-mapreduce-user-archive@minotaur.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 5855317C44 for ; Sat, 28 Feb 2015 05:36:50 +0000 (UTC) Received: (qmail 50680 invoked by uid 500); 28 Feb 2015 05:36:45 -0000 Delivered-To: apmail-hadoop-mapreduce-user-archive@hadoop.apache.org Received: (qmail 50563 invoked by uid 500); 28 Feb 2015 05:36:45 -0000 Mailing-List: contact user-help@hadoop.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: user@hadoop.apache.org Delivered-To: mailing list user@hadoop.apache.org Received: (qmail 50553 invoked by uid 99); 28 Feb 2015 05:36:45 -0000 Received: from nike.apache.org (HELO nike.apache.org) (192.87.106.230) by apache.org (qpsmtpd/0.29) with ESMTP; Sat, 28 Feb 2015 05:36:45 +0000 X-ASF-Spam-Status: No, hits=1.5 required=5.0 tests=HTML_MESSAGE,RCVD_IN_DNSWL_LOW,SPF_PASS X-Spam-Check-By: apache.org Received-SPF: pass (nike.apache.org: domain of xeonmailinglist@gmail.com designates 209.85.215.43 as permitted sender) Received: from [209.85.215.43] (HELO mail-la0-f43.google.com) (209.85.215.43) by apache.org (qpsmtpd/0.29) with ESMTP; Sat, 28 Feb 2015 05:36:20 +0000 Received: by labgm9 with SMTP id gm9so21377157lab.2 for ; Fri, 27 Feb 2015 21:35:34 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20120113; h=mime-version:in-reply-to:references:date:message-id:subject:from:to :content-type; bh=XjOCFKJotGTezcycxBqUFsyCg/AEFmvfXBhbcutB3YA=; b=RrD/V+Jnpz7QhCofay6YZerHsTjJLSn7M7q0R2C8CM3yf1e1NtV6dzvmh5k5YDQ1D5 Du+FUUUpfBsxQrXvWlu6eF/CgijeO8Ar2WPS0puaP1mT3wUS3oB8At1F1qPbnhGs8dGj YkkFVAI0i1uSQ0n3Pzcfp7JSqjLMNDOCm8r9v0yDHedzaidhH7aIADyxYWlmplZll36s cgQ26ksjmfmIOgt8AwIXJ8TAuIQuMXtvIxKLWi/xMR+rK9dpRe9s3yZ/EdHKBCbQZVeZ lpWO8BVu2phdwvFLQrnOZuulwV2xd5Lx2DURmX8EJSm/sAY4fLAW+se3yinfa66t7mqN yCgg== MIME-Version: 1.0 X-Received: by 10.152.4.5 with SMTP id g5mr15345751lag.119.1425101734154; Fri, 27 Feb 2015 21:35:34 -0800 (PST) Received: by 10.25.31.82 with HTTP; Fri, 27 Feb 2015 21:35:34 -0800 (PST) In-Reply-To: <117B7176-B732-4864-B798-349B10BB81FE@hortonworks.com> References: <54F08BC6.7080100@gmail.com> <117B7176-B732-4864-B798-349B10BB81FE@hortonworks.com> Date: Sat, 28 Feb 2015 05:35:34 +0000 Message-ID: Subject: Re: 1 job with Input data from 2 HDFS? From: xeon Mailinglist To: "user@hadoop.apache.org" Content-Type: multipart/alternative; boundary=089e01419f86a219c305101f5a3c X-Virus-Checked: Checked by ClamAV on apache.org --089e01419f86a219c305101f5a3c Content-Type: text/plain; charset=UTF-8 Hi, I don't understand this part of your answer: "read the other as a side-input directly by creating a client.". If I consider both inputs through the InputFormat, this means that a job will contain both input path in its configuration, and this is enough to work. So, what is the "other"? Is is the second input? Can you please explain what you have meant? On Friday, February 27, 2015, Vinod Kumar Vavilapalli < vinodkv@hortonworks.com> wrote: > It is entirely possible. You should treat one of them as the primary > inputs through the InputFormat/Mapper and read the other as a side-input > directly by creating a client. > > +Vinod > > On Feb 27, 2015, at 7:22 AM, xeonmailinglist > wrote: > > > Hi, > > > > I would like to have a mapreduce job that reads input data from 2 HDFS. > Is this possible? > > > > Thanks, > > --089e01419f86a219c305101f5a3c Content-Type: text/html; charset=UTF-8 Content-Transfer-Encoding: quoted-printable Hi,

I don't understand this part of your answer: &qu= ot;read the othe= r as a side-input directly by creating a client.".=C2=A0=

If I consider = both inputs through the InputFormat, this means that a job will contain bot= h input path in its configuration, and this is enough to work. So, what is = the "other"? Is is the second input? Can you please explain what = you have meant?

On Friday, February 27, 2015, Vinod Kumar Vavilapalli <vinodkv@hortonworks.com> wrote:
It is entirely possible. You should treat one of t= hem as the primary inputs through the InputFormat/Mapper and read the other= as a side-input directly by creating a client.

+Vinod

On Feb 27, 2015, at 7:22 AM, xeonmailinglist <x= eonmailinglist@gmail.com> wrote:

> Hi,
>
> I would like to have a mapreduce job that reads input data from 2 HDFS= . Is this possible?
>
> Thanks,

--089e01419f86a219c305101f5a3c--