THIS IS A TEST INSTANCE. ALL YOUR CHANGES WILL BE LOST!!!!
...
Code Block |
---|
public enum DataSetType { HDFS, HBASE, HIVE } // use this enum to determine what dataset kite needs to create underneath @Input public DataSetType datasetType or // parse this to figure out the data set @Input(size = 255, validators = {@Validator(DatasetURIValidator.class)}) public String uri |
- Piggy back on config annotations ( conditions that we are intending to add since ages! ) to show only relevant config subsequently. For instance
hdfsHostAndPort may not be relevant for HIVE or HBase
Pros : - No code duplication
- No weird build dependency of KiteHbaseConnector depending on KiteConnector that might make independent connector upgrade complicated
...
- Add support for Hbase related configs
- Add support to create hbase dataset in the Kite
New code
- for column mapping and paritioning
- KiteExtractor to support creating Hbase datasets via Kite SDK and reading records
- KiteLoader to support creating Hbase datasets via Kite SDK and writing records ( merge temp data sets), this needs to be investigated more.
- How will the Hbase write happen? How different is it from HDFS write or HIVE write?
- If we support DFM, add relevant DFM configs and code in KiteConnector
...