Hi, All, My name is LinWen. I joined Pivotal HAWQ Beijing team since November, 2014. Before that, I am an engineer of VMware. I have implemented some features for HAWQ, like new fault tolerance service, libyarn, etc. I believe HAWQ can be the best SQL on Hadoop engine with our joint effort.
On Thu, Dec 15, 2016 at 11:03 PM, Lili Ma <[email protected]> wrote: > Hello everyone, > > Glad to know everybody here:) > > I'm Lili Ma, from Pivotal HAWQ R&D team in Beijing. I have been focusing > on HAWQ development and product management since 2012 when I joined > Pivotal. I experienced and contributed in HAWQ's all growth path, from > birth, Alpha, 1.X, 2.0... > > My main covering fields about HAWQ include three parts: 1) Storage such as > internal table storage, HAWQ Input/OutputFormat, hawq extract/register,etc > 2) Dispatcher and interconnect 3) Security including Ranger integration, > Kerberos and LDAP. > > Before Pivotal, I worked at IBM for more than 2 years and focused on > providing data service inside of our public cloud provision. The data > service includes RDS(relational data service) which can provision a > distributed relational database based on DB2 Federation, and NOSQL service > which is based on HBase. > > I believe HAWQ can become more successful with our joint effort! Welcome > to reach me or this mail list for any HAWQ or other kinds of issues :) > > Thanks > Lili > > > 2016-12-15 4:45 GMT+08:00 Dan Baskette <[email protected]>: > >> I will add to the email flow… >> >> I am Dan Baskette, I am the Director of Tech Marketing for Pivotal and >> cover Pivotal HDB/Apache HAWQ, Pivotal Greenplum Database, and Apache >> MADlib. I started my career at Sun Microsystems, and have been working >> for EMC/Greenplum and now Pivotal since 2000….a LONG time in quite a number >> of roles. I was part of the team that launched Greenplum’s first Hadoop >> distribution and was around for the birth of HAWQ or as we called it when >> it was in it’s infancy…. GOH or Greenplum on Hadoop. I have been actively >> running some webcasts on various HAWQ how-to topics for Hortonworks, so you >> can check those out on their site. >> >> Hoping this community really takes off in a big way! >> >> Dan >> >> >> On December 14, 2016 at 10:09:34 AM, Ruilong Huo ([email protected]) wrote: >> >> Hi All, >> >> Great for Gregory to start the thread that people can know each other >> much better, at least in Apache HAWQ community! >> >> I am Ruilong Huo and I am from HDB/HAWQ engineering team in Pivotal. I am >> from Teradata and joined Pivotal after that. It's my honor to be part of >> HAWQ project at its early stage. I am a fan of RDBMS (especially MPP >> database), big data, and cloud technology that changes the IT >> infrastructure of the enterprises and helps to do information >> transformation in a very large extent. >> >> I hope that with joint effort from hawq community, it will become even >> greater product in big data area, especially in SQL-on-Hadoop category. >> >> Best regards, >> Ruilong Huo >> >> On Wed, Dec 14, 2016 at 2:27 PM, Bob Glithero <[email protected]> >> wrote: >> >>> Hello all, >>> >>> I'm Bob, and I'm doing product marketing for HDB/HAWQ at Pivotal. I'm >>> new-ish here, and not so much from a coding background as from networking. >>> I'm from Cisco Systems, where I focused on analytics use cases in >>> telecommunications, particularly for mobile network operators, for service >>> assurance, customer care, and customer profiling. (also, as you're >>> introducing yourselves, we'd love to hear what use cases you're involved >>> with, too). >>> >>> About a year before I left my group at Cisco acquired an MPP database of >>> its own -- ParStream -- for its IoT and fog computing use cases, so it's >>> interesting to come here and learn about the architecture and applications >>> of HAWQ. >>> >>> I hope to help make your experience with HAWQ a good one. If I can help >>> in any way, please reach out to me directly or on the list. >>> >>> Cheers, >>> Bob >>> >>> >>> >>> Bob Glithero | Product Marketing >>> Pivotal, Inc. >>> [email protected] | m: 415.341.5592 >>> >>> >>> On Sun, Dec 11, 2016 at 6:56 PM, Roman Shaposhnik <[email protected]> >>> wrote: >>> >>>> Greg, thanks for kicking off the roll call. Getting to know each other >>>> is super >>>> useful (and can be fun! ;-)). I'll go next: >>>> >>>> I am Roman (your friendly neighborhood mentor). I hang around a lot of >>>> ASF >>>> big data projects (as a committer and a PMC member), but lately I've >>>> been >>>> gravitating towards IoT as well (Apache Mynewt). I started my career at >>>> Sun >>>> microsystems back at a time when Linux wasn't even 1.x and I've been >>>> doing >>>> enterprise software ever since. I was lucky enough to get to work on >>>> the original >>>> Hadoop team at Yahoo! and fall in love with not one but two elephants >>>> (Hadoop >>>> and Postgres). Recently I've assumed a position of VP of Technology at >>>> ODPi >>>> and I'm still hoping to MHGA! My secret weapon is Apache Bigtop (which >>>> co-founded) >>>> and I'm not afraid to use it! >>>> >>>> I'm here to help as much as I can to make sure that this community >>>> evolves into >>>> a vibrant, self-governed, exciting place worthy of being a top level >>>> project (TLP) >>>> at ASF. If you have any questions or ideas that you may want to bounce >>>> off of >>>> me -- please don't hesitate to reach out directly or on the mailing >>>> list. >>>> >>>> Thanks, >>>> Roman. >>>> >>>> On Fri, Dec 9, 2016 at 11:53 AM, Gregory Chase <[email protected]> >>>> wrote: >>>> > >>>> > Dear HAWQs, >>>> > >>>> > I thought it would be fun to get to know some of the other people in >>>> the community. >>>> > >>>> > My name is Greg Chase and I run community development for Pivotal for >>>> big data open source communities that Pivotal contributes to. >>>> > >>>> > Some of you may have seen my frequent emails about virtual events I >>>> help organize for user and contributor education. >>>> > >>>> > Not so long ago, I was in charge of product marketing for an >>>> in-memory data warehouse named after a Hawaiian town from a three-letter >>>> acronymed German Company. We treated Hadoop as an external table, and >>>> returning results from these queries was both slow and brittle due to the >>>> network transfer rates. >>>> > >>>> > So I have a special appreciation of the innovation that has gone into >>>> creating Hadoop-native HAWQ out of PostgreSQL and Greenplum. >>>> > >>>> > These days I'm much more of a marketer than a coder, but I still love >>>> hearing about the kinds of projects that HAWQ users are involved in. >>>> > >>>> > I know we'd all love to hear more about everyone else's projects, and >>>> how you became a HAWQ user. So please introduce yourselves! >>>> > >>>> > -- >>>> > Greg Chase >>>> > >>>> > Global Head, Big Data Communities >>>> > http://www.pivotal.io/big-data >>>> > >>>> > Pivotal Software >>>> > http://www.pivotal.io/ >>>> > >>>> > 650-215-0477 >>>> > @GregChase >>>> > Blog: http://geekmarketing.biz/ >>>> > >>>> >>> >>> >> >
