hawq-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Wen Lin <w...@pivotal.io>
Subject Re: [INTRODUCTIONS] Hi I'm Greg and I'm part of the Apache HAWQ community
Date Tue, 20 Dec 2016 10:03:05 GMT
Hi, All,

My name is LinWen. I joined Pivotal HAWQ Beijing team since November, 2014.
Before that, I am an engineer of VMware.
I have implemented some features for HAWQ, like new fault tolerance
service, libyarn, etc.
I believe HAWQ can be the best SQL on Hadoop engine with our joint effort.

On Thu, Dec 15, 2016 at 11:03 PM, Lili Ma <lilima@apache.org> wrote:

> Hello everyone,
> Glad to know everybody here:)
> I'm Lili Ma, from Pivotal HAWQ R&D team in Beijing. I have been focusing
> on HAWQ development and product management since 2012 when I joined
> Pivotal. I experienced and contributed in HAWQ's all growth path, from
> birth, Alpha, 1.X, 2.0...
> My main covering fields about HAWQ include three parts: 1) Storage such as
> internal table storage, HAWQ Input/OutputFormat, hawq extract/register,etc
> 2) Dispatcher and interconnect 3) Security including Ranger integration,
> Kerberos and LDAP.
> Before Pivotal, I worked at IBM for more than 2 years and focused on
> providing data service inside of our public cloud provision. The data
> service includes RDS(relational data service) which can provision a
> distributed relational database based on DB2 Federation, and NOSQL service
> which is based on HBase.
> I believe HAWQ can become more successful with our joint effort!  Welcome
> to reach me or this mail list for any HAWQ or other kinds of issues :)
> Thanks
> Lili
> 2016-12-15 4:45 GMT+08:00 Dan Baskette <dbbaskette@gmail.com>:
>> I will add to the email flow…
>> I am Dan Baskette, I am the Director of Tech Marketing for Pivotal and
>> cover Pivotal HDB/Apache HAWQ, Pivotal Greenplum Database, and Apache
>> MADlib.   I started my career at Sun Microsystems, and have been working
>> for EMC/Greenplum and now Pivotal since 2000….a LONG time in quite a number
>> of roles.   I was part of the team that launched Greenplum’s first Hadoop
>> distribution and was around for the birth of HAWQ or as we called it when
>> it was in it’s infancy…. GOH or Greenplum on Hadoop.   I have been actively
>> running some webcasts on various HAWQ how-to topics for Hortonworks, so you
>> can check those out on their site.
>> Hoping this community really takes off in a big way!
>> Dan
>> On December 14, 2016 at 10:09:34 AM, Ruilong Huo (rhuo@pivotal.io) wrote:
>> Hi All,
>> Great for Gregory to start the thread that people can know each other
>> much better, at least in Apache HAWQ community!
>> I am Ruilong Huo and I am from HDB/HAWQ engineering team in Pivotal. I am
>> from Teradata and joined Pivotal after that. It's my honor to be part of
>> HAWQ project at its early stage. I am a fan of RDBMS (especially MPP
>> database), big data, and cloud technology that changes the IT
>> infrastructure of the enterprises and helps to do information
>> transformation in a very large extent.
>> I hope that with joint effort from hawq community, it will become even
>> greater product in big data area, especially in SQL-on-Hadoop category.
>> Best regards,
>> Ruilong Huo
>> On Wed, Dec 14, 2016 at 2:27 PM, Bob Glithero <rglithero@pivotal.io>
>> wrote:
>>> Hello all,
>>> I'm Bob, and I'm doing product marketing for HDB/HAWQ at Pivotal.  I'm
>>> new-ish here, and not so much from a coding background as from networking.
>>> I'm from Cisco Systems, where I focused on analytics use cases in
>>> telecommunications, particularly for mobile network operators, for service
>>> assurance, customer care, and customer profiling.  (also, as you're
>>> introducing yourselves, we'd love to hear what use cases you're involved
>>> with, too).
>>> About a year before I left my group at Cisco acquired an MPP database of
>>> its own -- ParStream -- for its IoT and fog computing use cases, so it's
>>> interesting to come here and learn about the architecture and applications
>>> of HAWQ.
>>> I hope to help make your experience with HAWQ a good one.  If I can help
>>> in any way, please reach out to me directly or on the list.
>>> Cheers,
>>> Bob
>>> Bob Glithero | Product Marketing
>>> Pivotal, Inc.
>>> rglithero@pivotal.io | m: 415.341.5592
>>> On Sun, Dec 11, 2016 at 6:56 PM, Roman Shaposhnik <roman@shaposhnik.org>
>>> wrote:
>>>> Greg, thanks for kicking off the roll call. Getting to know each other
>>>> is super
>>>> useful (and can be fun! ;-)). I'll go next:
>>>> I am Roman (your friendly neighborhood mentor). I hang around a lot of
>>>> ASF
>>>> big data projects (as a committer and a PMC member), but lately I've
>>>> been
>>>> gravitating towards IoT as well (Apache Mynewt). I started my career at
>>>> Sun
>>>> microsystems back at a time when Linux  wasn't even 1.x and I've been
>>>> doing
>>>> enterprise software ever since. I was lucky enough to get to work on
>>>> the original
>>>> Hadoop team at Yahoo! and fall in love with not one but two elephants
>>>> (Hadoop
>>>> and Postgres). Recently I've assumed a position of VP of Technology at
>>>> ODPi
>>>> and I'm still hoping to MHGA! My secret weapon is Apache Bigtop (which
>>>> co-founded)
>>>> and I'm not afraid to use it!
>>>> I'm here to help as much as I can to make sure that this community
>>>> evolves into
>>>> a vibrant, self-governed, exciting place worthy of being a top level
>>>> project (TLP)
>>>> at ASF. If you have any questions or ideas that you may want to bounce
>>>> off of
>>>> me -- please don't hesitate to reach out directly or on the mailing
>>>> list.
>>>> Thanks,
>>>> Roman.
>>>> On Fri, Dec 9, 2016 at 11:53 AM, Gregory Chase <gchase@pivotal.io>
>>>> wrote:
>>>> >
>>>> > Dear HAWQs,
>>>> >
>>>> > I thought it would be fun to get to know some of the other people in
>>>> the community.
>>>> >
>>>> > My name is Greg Chase and I run community development for Pivotal for
>>>> big data open source communities that Pivotal contributes to.
>>>> >
>>>> > Some of you may have seen my frequent emails about virtual events I
>>>> help organize for user and contributor education.
>>>> >
>>>> > Not so long ago, I was in charge of product marketing for an
>>>> in-memory data warehouse named after a Hawaiian town from a three-letter
>>>> acronymed German Company. We treated Hadoop as an external table, and
>>>> returning results from these queries was both slow and brittle due to the
>>>> network transfer rates.
>>>> >
>>>> > So I have a special appreciation of the innovation that has gone into
>>>> creating Hadoop-native HAWQ out of PostgreSQL and Greenplum.
>>>> >
>>>> > These days I'm much more of a marketer than a coder, but I still love
>>>> hearing about the kinds of projects that HAWQ users are involved in.
>>>> >
>>>> > I know we'd all love to hear more about everyone else's projects, and
>>>> how you became a HAWQ user.  So please introduce yourselves!
>>>> >
>>>> > --
>>>> > Greg Chase
>>>> >
>>>> > Global Head, Big Data Communities
>>>> > http://www.pivotal.io/big-data
>>>> >
>>>> > Pivotal Software
>>>> > http://www.pivotal.io/
>>>> >
>>>> > 650-215-0477
>>>> > @GregChase
>>>> > Blog: http://geekmarketing.biz/
>>>> >

View raw message