+1

Pratyaksh Sharma <pratyaks...@gmail.com> 于2021年7月30日周五 上午1:47写道:

> Guess we should rebrand Hudi on README.md file as well -
> https://github.com/apache/hudi#readme?
>
> This page still mentions the following -
>
> "Apache Hudi (pronounced Hoodie) stands for Hadoop Upserts Deletes and
> Incrementals. Hudi manages the storage of large analytical datasets on
> DFS (Cloud stores, HDFS or any Hadoop FileSystem compatible storage)."
>
> On Sat, Jul 24, 2021 at 6:31 AM Vinoth Chandar <vin...@apache.org> wrote:
>
>> Thanks Vino! Got a bunch of emoticons on the PR as well.
>>
>> Will land this monday, giving it more time over the weekend as well.
>>
>>
>> On Wed, Jul 21, 2021 at 7:36 PM vino yang <yanghua1...@gmail.com> wrote:
>>
>> > Thanks vc
>> >
>> > Very good blog, in-depth and forward-looking. Learned!
>> >
>> > Best,
>> > Vino
>> >
>> > Vinoth Chandar <vin...@apache.org> 于2021年7月22日周四 上午3:58写道:
>> >
>> > > Expanding to users@ as well.
>> > >
>> > > Hi all,
>> > >
>> > > Since this discussion, I started to pen down a coherent strategy and
>> > convey
>> > > these ideas via a blog post.
>> > > I have also done my own research, talked to (ex)colleagues I respect
>> to
>> > get
>> > > their take and refine it.
>> > >
>> > > Here's a blog that hopefully explains this vision.
>> > >
>> > > https://github.com/apache/hudi/pull/3322
>> > >
>> > > Look forward to your feedback on the PR. We are hoping to land this
>> early
>> > > next week, if everyone is aligned.
>> > >
>> > > Thanks
>> > > Vinoth
>> > >
>> > > On Wed, Apr 21, 2021 at 9:01 PM wei li <lw309637...@gmail.com> wrote:
>> > >
>> > > > +1 , Cannot agree more.
>> > > >  *aux metadata* and metatable, can make hudi have large preformance
>> > > > optimization on query end.
>> > > > Can continuous develop.
>> > > > cache service may the necessary component in cloud native
>> environment.
>> > > >
>> > > > On 2021/04/13 05:29:55, Vinoth Chandar <vin...@apache.org> wrote:
>> > > > > Hello all,
>> > > > >
>> > > > > Reading one more article today, positioning Hudi, as just a table
>> > > format,
>> > > > > made me wonder, if we have done enough justice in explaining what
>> we
>> > > have
>> > > > > built together here.
>> > > > > I tend to think of Hudi as the data lake platform, which has the
>> > > > following
>> > > > > components, of which - one if a table format, one is a
>> transactional
>> > > > > storage layer.
>> > > > > But the whole stack we have is definitely worth more than the sum
>> of
>> > > all
>> > > > > the parts IMO (speaking from my own experience from the past 10+
>> > years
>> > > of
>> > > > > open source software dev).
>> > > > >
>> > > > > Here's what we have built so far.
>> > > > >
>> > > > > a) *table format* : something that stores table schema, a metadata
>> > > table
>> > > > > that stores file listing today, and being extended to store column
>> > > ranges
>> > > > > and more in the future (RFC-27)
>> > > > > b) *aux metadata* : bloom filters, external record level indexes
>> > today,
>> > > > > bitmaps/interval trees and other advanced on-disk data structures
>> > > > tomorrow
>> > > > > c) *concurrency control* : we always supported MVCC based log
>> based
>> > > > > concurrency (serialize writes into a time ordered log), and we now
>> > also
>> > > > > have OCC for batch merge workloads with 0.8.0. We will have
>> > multi-table
>> > > > and
>> > > > > fully non-blocking writers soon (see future work section of
>> RFC-22)
>> > > > > d) *updates/deletes* : this is the bread-and-butter use-case for
>> > Hudi,
>> > > > but
>> > > > > we support primary/unique key constraints and we could add foreign
>> > keys
>> > > > as
>> > > > > an extension, once our transactions can span tables.
>> > > > > e) *table services*: a hudi pipeline today is self-managing -
>> sizes
>> > > > files,
>> > > > > cleans, compacts, clusters data, bootstraps existing data - all
>> these
>> > > > > actions working off each other without blocking one another. (for
>> > most
>> > > > > parts).
>> > > > > f) *data services*: we also have higher level functionality with
>> > > > > deltastreamer sources (scalable DFS listing source, Kafka, Pulsar
>> is
>> > > > > coming, ...and more), incremental ETL support, de-duplication,
>> commit
>> > > > > callbacks, pre-commit validations are coming, error tables have
>> been
>> > > > > proposed. I could also envision us building towards streaming
>> egress,
>> > > > data
>> > > > > monitoring.
>> > > > >
>> > > > > I also think we should build the following (subject to separate
>> > > > > DISCUSS/RFCs)
>> > > > >
>> > > > > g) *caching service*: Hudi specific caching service that can hold
>> > > mutable
>> > > > > data and serve oft-queried data across engines.
>> > > > > h) t*imeline metaserver:* We already run a metaserver in spark
>> > > > > writer/drivers, backed by rocksDB & even Hudi's metadata table.
>> Let's
>> > > > turn
>> > > > > it into a scalable, sharded metastore, that all engines can use to
>> > > obtain
>> > > > > any metadata.
>> > > > >
>> > > > > To this end, I propose we rebrand to "*Data Lake Platform*" as
>> > opposed
>> > > to
>> > > > > "ingests & manages storage of large analytical datasets over DFS
>> > (hdfs
>> > > or
>> > > > > cloud stores)." and convey the scope of our vision,
>> > > > > given we have already been building towards that. It would also
>> > provide
>> > > > new
>> > > > > contributors a good lens to look at the project from.
>> > > > >
>> > > > > (This is very similar to for e.g, the evolution of Kafka from a
>> > pub-sub
>> > > > > system, to an event streaming platform - with addition of
>> > > > > MirrorMaker/Connect etc. )
>> > > > >
>> > > > > Please share your thoughts!
>> > > > >
>> > > > > Thanks
>> > > > > Vinoth
>> > > > >
>> > > >
>> > >
>> >
>>
>

Reply via email to