Skip to content
HenryCaiHaiying edited this page Dec 2, 2014 · 22 revisions

Gobblin Image

Gobblin is an open source unified data ingestion framework to bring significant amount of data from internal data sources (data generated on premise) and external data sources (data sourced from external web sites) into one central repository (HDFS) for analysis.

As companies are moving more and more towards a data-driven decision making business model, increasing number of business products are driven by business insights from data generated both internally on premise or sourced externally from public web sites or web services. Gobblin is developed to address ingesting those big data with ease:

  • Centralized data lake: standardized data formats, directory layouts;
  • Standardized catalog of lightweight transformations: security filters, schema evolution, type conversion, etc;
  • Data quality measurements and enforcement: schema validation, data audits, etc;
  • Scalable ingest: auto-scaling, fault-tolerance, etc;
  • Ease of operations: centralized monitoring, enforcement of SLA-s etc;
  • Ease of use: self-serve on-boarding of new datasets to minimize time and involvement of engineers.

Features

Support Matrix

Gobblin supports the following combination of data sources and protocols:

  • The types of data sources: RDBMS, distributed NoSQL, event streams, log files, etc.;
  • The types of data transport protocols: file copies over HTTP or SFTP, JDBC, REST, Kafka, Databus, vendor-specific APIs, etc.;
  • The semantics of the data bundles: increments, appends, full dumps, change stream, etc.;
  • The types of the data flows: batch, streaming.

Community

Requirements

  • Java 1.6+
  • Gradle 1.12+
  • Hadoop 1.2.1+ or Hadoop 2.3.0+

Quickstart Guides and Examples

Clone this wiki locally