Return-Path: X-Original-To: apmail-hadoop-common-commits-archive@www.apache.org Delivered-To: apmail-hadoop-common-commits-archive@www.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 63CFD6A4F for ; Fri, 24 Jun 2011 05:24:43 +0000 (UTC) Received: (qmail 80581 invoked by uid 500); 24 Jun 2011 05:24:43 -0000 Delivered-To: apmail-hadoop-common-commits-archive@hadoop.apache.org Received: (qmail 80277 invoked by uid 500); 24 Jun 2011 05:24:27 -0000 Mailing-List: contact common-commits-help@hadoop.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: common-dev@hadoop.apache.org Delivered-To: mailing list common-commits@hadoop.apache.org Received: (qmail 80267 invoked by uid 500); 24 Jun 2011 05:24:22 -0000 Delivered-To: apmail-hadoop-core-commits@hadoop.apache.org Received: (qmail 80263 invoked by uid 99); 24 Jun 2011 05:24:21 -0000 Received: from nike.apache.org (HELO nike.apache.org) (192.87.106.230) by apache.org (qpsmtpd/0.29) with ESMTP; Fri, 24 Jun 2011 05:24:21 +0000 X-ASF-Spam-Status: No, hits=-2000.0 required=5.0 tests=ALL_TRUSTED X-Spam-Check-By: apache.org Received: from [140.211.11.131] (HELO eos.apache.org) (140.211.11.131) by apache.org (qpsmtpd/0.29) with ESMTP; Fri, 24 Jun 2011 05:24:17 +0000 Received: from eos.apache.org (localhost [127.0.0.1]) by eos.apache.org (Postfix) with ESMTP id 00958C67; Fri, 24 Jun 2011 05:23:55 +0000 (UTC) MIME-Version: 1.0 Content-Type: text/plain; charset="utf-8" Content-Transfer-Encoding: quoted-printable From: Apache Wiki To: Apache Wiki Date: Fri, 24 Jun 2011 05:23:55 -0000 Message-ID: <20110624052355.44518.74802@eos.apache.org> Subject: =?utf-8?q?=5BHadoop_Wiki=5D_Update_of_=22Hbase/PoweredBy=22_by_OtisGospod?= =?utf-8?q?netic?= Auto-Submitted: auto-generated X-Virus-Checked: Checked by ClamAV on apache.org Dear Wiki user, You have subscribed to a wiki page or wiki category on "Hadoop Wiki" for ch= ange notification. The "Hbase/PoweredBy" page has been changed by OtisGospodnetic: http://wiki.apache.org/hadoop/Hbase/PoweredBy?action=3Ddiff&rev1=3D70&rev2= =3D71 [[http://www.adobe.com|Adobe]] - We currently have about 30 nodes running= HDFS, Hadoop and HBase in clusters ranging from 5 to 14 nodes on both pro= duction and development. We plan a deployment on an 80 nodes cluster. We ar= e using HBase in several areas from social services to structured data and = processing for internal use. We constantly write data to HBase and run mapr= educe jobs to process then store it back to HBase or external systems. Our = production cluster has been running since Oct 2008. = - [[http://caree.rs|Caree.rs]] - Accelerated hiring platform for HiTech com= panies. We use HBase and Hadoop for all aspects of our backend - job and co= mpany data storage, analytics processing, machine learning algorithms for o= ur hire recommendation engine. Our live production site is directly served = from HBase. We use cascading for running offline data processing jobs. = + [[http://caree.rs|Caree.rs]] - Accelerated hiring platform for HiTech com= panies. We use HBase and Hadoop for all aspects of our backend - job and co= mpany data storage, analytics processing, machine learning algorithms for o= ur hire recommendation engine. Our live production site is directly served = from HBase. We use cascading for running offline data processing jobs. = [[http://www.drawntoscaleconsulting.com|Drawn to Scale Consulting]] consu= lts on HBase, Hadoop, Distributed Search, and Scalable architectures. = @@ -16, +16 @@ = [[http://gumgum.com|GumGum]] is an In-Image ad network. We use HBase on a= n 8-node Amazon EC2 High-CPU Extra Large (c1.xlarge) cluster for both real-= time data and analytics. Our production cluster has been running since June= 2010. = - [[http://www.impetus.com/ |Impetus]] - With a strong focus, established t= hought leadership and open source contributions in the area of Big Data ana= lytics and consulting services, Impetus uses its Global Delivery Model to h= elp technology businesses and enterprises evaluate and implement solutions = tailored to their specific context, without being biased towards a particul= ar solution. [[http://bigdata.impetus.com/# | More info about BigData @Impe= tus]] + [[http://www.impetus.com/|Impetus]] - With a strong focus, established th= ought leadership and open source contributions in the area of Big Data anal= ytics and consulting services, Impetus uses its Global Delivery Model to he= lp technology businesses and enterprises evaluate and implement solutions t= ailored to their specific context, without being biased towards a particula= r solution. [[http://bigdata.impetus.com/#|More info about BigData @Impetus= ]] = - [[http://www.infolinks.com/ |Infolinks]] - Infolinks is an In-Text ad pro= vider. We use HBase to process advertisement selection and user events for = our In-Text ad network. The reports generated from HBase are used as feedba= ck for our production system to optimize ad selection. = + [[http://www.infolinks.com/|Infolinks]] - Infolinks is an In-Text ad prov= ider. We use HBase to process advertisement selection and user events for o= ur In-Text ad network. The reports generated from HBase are used as feedbac= k for our production system to optimize ad selection. = [[http://www.kalooga.com|Kalooga]] is a discovery service for image galle= ries. We use Hadoop, HBase and Pig on a 20-node cluster for our crawling, a= nalysis and events processing. = @@ -40, +40 @@ = [[http://www.readpath.com/|ReadPath]] uses HBase to store several hundred= million RSS items and dictionary for its RSS newsreader. Readpath is curre= ntly running on an 8 node cluster. = - [[http://resu.me/|resu.me]] - Career network for the net generation. We u= se HBase and Hadoop for all aspects of our backend - user and resume data s= torage, analytics processing, machine learning algorithms for our job recom= mendation engine. Our live production site is directly served from HBase. W= e use cascading for running offline data processing jobs. = + [[http://resu.me/|resu.me]] - Career network for the net generation. We u= se HBase and Hadoop for all aspects of our backend - user and resume data s= torage, analytics processing, machine learning algorithms for our job recom= mendation engine. Our live production site is directly served from HBase. W= e use cascading for running offline data processing jobs. = [[http://www.runa.com/|Runa Inc.]] offers a SaaS that enables online merc= hants to offer dynamic per-consumer, per-product promotions embedded in the= ir website. To implement this we collect the click streams of all their vis= itors to determine along with the rules of the merchant what promotion to o= ffer the visitor at different points of their browsing the Merchant website= . So we have lots of data and have to do lots of off-line and real-time ana= lytics. HBase is the core for us. We also use Clojure and our own open sour= ced distributed processing framework, Swarmiji. The HBase Community has bee= n key to our forward movement with HBase. We're looking for experienced dev= elopers to join us to help make things go even faster! + = + [[http://www.sematext.com/|Sematext]] runs [[http://www.sematext.com/sear= ch-analytics/index.html|Search Analytics]], a service that uses HBase to st= ore search activity and MapReduce to produce reports showing user search be= haviour and experience. + = + [[http://www.sematext.com/search-analytics/index.html|Sematext]] runs [[h= ttp://www.sematext.com/spm/index.html|Scalable Performance Monitoring]] (SP= M), a service that uses HBase to store performance data over time, crunch i= t with the help of MapReduce, and display it in a visually rich browser-bas= ed UI. Interestingly, SPM features [[http://www.sematext.com/spm/hbase-per= formance-monitoring/index.html|SPM for HBase]], which is specifically desig= ned to monitor all HBase performance metrics. = [[http://www.socialmedia.com/|SocialMedia]] uses HBase to store and proce= ss user events which allows us to provide near-realtime user metrics and re= porting. HBase forms the heart of our Advertising Network data storage and = management system. We use HBase as a data source and sink for both realtime= request cycle queries and as a backend for mapreduce analysis. = @@ -70, +74 @@ = [[http://www.yahoo.com/|Yahoo!]] uses HBase to store document fingerprint= for detecting near-duplications. We have a cluster of few nodes that runs = HDFS, mapreduce, and HBase. The table contains millions of rows. We use thi= s for querying duplicated documents with realtime traffic. = - [[http://h50146.www5.hp.com/products/software/security/icewall/eng/|HP Ic= eWall SSO]] - is a web-based single sign-on solution and uses HBase to stor= e user data to authenticate users. We have supported RDB and LDAP previousl= y but have newly supported HBase with a view to authenticate over tens of m= illions of users and devices. = + [[http://h50146.www5.hp.com/products/software/security/icewall/eng/|HP Ic= eWall SSO]] - is a web-based single sign-on solution and uses HBase to stor= e user data to authenticate users. We have supported RDB and LDAP previousl= y but have newly supported HBase with a view to authenticate over tens of m= illions of users and devices. =20