Return-Path: Delivered-To: apmail-hadoop-chukwa-dev-archive@minotaur.apache.org Received: (qmail 81041 invoked from network); 18 Jun 2009 16:48:22 -0000 Received: from hermes.apache.org (HELO mail.apache.org) (140.211.11.3) by minotaur.apache.org with SMTP; 18 Jun 2009 16:48:22 -0000 Received: (qmail 45169 invoked by uid 500); 18 Jun 2009 16:48:33 -0000 Delivered-To: apmail-hadoop-chukwa-dev-archive@hadoop.apache.org Received: (qmail 45149 invoked by uid 500); 18 Jun 2009 16:48:33 -0000 Mailing-List: contact chukwa-dev-help@hadoop.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: chukwa-dev@hadoop.apache.org Delivered-To: mailing list chukwa-dev@hadoop.apache.org Received: (qmail 45134 invoked by uid 99); 18 Jun 2009 16:48:32 -0000 Received: from nike.apache.org (HELO nike.apache.org) (192.87.106.230) by apache.org (qpsmtpd/0.29) with ESMTP; Thu, 18 Jun 2009 16:48:32 +0000 X-ASF-Spam-Status: No, hits=-2000.0 required=10.0 tests=ALL_TRUSTED X-Spam-Check-By: apache.org Received: from [140.211.11.140] (HELO brutus.apache.org) (140.211.11.140) by apache.org (qpsmtpd/0.29) with ESMTP; Thu, 18 Jun 2009 16:48:29 +0000 Received: from brutus (localhost [127.0.0.1]) by brutus.apache.org (Postfix) with ESMTP id 53A66234C044 for ; Thu, 18 Jun 2009 09:48:07 -0700 (PDT) Message-ID: <1877420291.1245343687337.JavaMail.jira@brutus> Date: Thu, 18 Jun 2009 09:48:07 -0700 (PDT) From: "Jerome Boulon (JIRA)" To: chukwa-dev@hadoop.apache.org Subject: [jira] Commented: (CHUKWA-306) Standalone (non-daemon) Chukwa operation In-Reply-To: <516723908.1245276847477.JavaMail.jira@brutus> MIME-Version: 1.0 Content-Type: text/plain; charset=utf-8 Content-Transfer-Encoding: 7bit X-JIRA-FingerPrint: 30527f35849b9dde25b450d4833f0394 X-Virus-Checked: Checked by ClamAV on apache.org [ https://issues.apache.org/jira/browse/CHUKWA-306?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12721338#action_12721338 ] Jerome Boulon commented on CHUKWA-306: -------------------------------------- The solution to this is to use the backfilling tool. This tool will produce your dataSink files (on per input file) then you just have to run demux like a standard M/R job. By configuring chukwa-collector-conf.xml to use the local filesystem then you don't have to worry about hadoop config at all. > Standalone (non-daemon) Chukwa operation > ---------------------------------------- > > Key: CHUKWA-306 > URL: https://issues.apache.org/jira/browse/CHUKWA-306 > Project: Hadoop Chukwa > Issue Type: Wish > Reporter: Jiaqi Tan > Priority: Critical > > This is an articulation of a possible alternative use of Chukwa as a standalone log analysis pipeline. This would enable users to read in existing logs from files, process (Demux) and perform analysis (e.g. current SALSA/Mochi toolchain) on them, and visualize them, without requiring the user to setup or run any daemons, nor database servers. > This can be presented as an alternative interface to Chukwa for the user, where the main architectural parts (Chunks, post-Demux SequenceFiles of ChukwaRecords, post-Demux-processing SequenceFiles of ChukwaRecords, and finally time-aggregated database entries for fast visualization) remain unchanged, and Chukwa is manifest as a set of files in HDFS. The main value that Chukwa then provides to users is 1. centralized one-stop-shop for log processing+analysis+anomaly detection, 2. the ability to use MapReduce to process logs, regardless of whether they had used Chukwa to collect the logs. > That way, the ability to process logs and analyze/do diagnosis is not tied to having to run the entire Chukwa daemon infrastructure, since many users who use Hadoop clusters may not have superuser access to those machines, e.g. users at universities using shared clusters. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.