Serializing the complete payload object instead of serializing just the GenericRecord
Removing Converter hierarchy as we now depend purely on JavaSerialization and require the payload to be java serializable
14 files changed
tree: 213e5db02f91d5c96e37ac25be3aa7564d054187
  1. deploy/
  2. docker/
  3. docs/
  4. hoodie-cli/
  5. hoodie-client/
  6. hoodie-common/
  7. hoodie-hadoop-mr/
  8. hoodie-hive/
  9. hoodie-integ-test/
  10. hoodie-spark/
  11. hoodie-utilities/
  12. packaging/
  13. style/
  14. .gitignore
  15. .travis.yml
  16. _config.yml
  17. CHANGELOG.md
  18. LICENSE.txt
  19. pom.xml
  20. README.md
  21. RELEASE_NOTES.md
README.md

Hudi

Hudi (pronounced Hoodie) stands for Hadoop Upserts anD Incrementals. Hudi manages storage of large analytical datasets on HDFS and serve them out via two types of tables

  • Read Optimized Table - Provides excellent query performance via purely columnar storage (e.g. Parquet)
  • Near-Real time Table (WIP) - Provides queries on real-time data, using a combination of columnar & row based storage (e.g Parquet + Avro)

For more, head over here