Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.

...

As one mission for Griffin is to reduce MTTD(mean-time-to-detect)

  • During define phase, next generation architecture should use more expressive rules to define data quality requirements. SQL based rule is a good candidate for defining data quality, it is abstract but also concrete. It is abstract so that we can dispatch data quality rules to different query engines, it is concrete that all data quality stakeholders can understand the rules and align easily.
  • During define phase, the data quality should be uniformly defined among different scenarios such as batch, near realtime and realtime.
  • During measure phase, the next generation Griffin should standardize measure pipelines to different stages as recording stage, checking stage and alerting stage. It is easily for different data platform teams to integration with Griffin during different stages.
  • During measure phase, the next generation Griffin should not couple with any particular query engine, so it should able to dispatch requests to different query engine(spark, hive, flink, presto) upon different data quality rules.
  • During measure phase, the next generation Griffin should support different schedule strategies such as event trigger or time-based trigger.
  • During analyze phase, the next generation Griffin should provide standardize solutions as anomaly detection algorithm to detect anomaly, since in most cases, related stakeholders need support to define anomaly.
  • Last but not least, the next generation Griffin should provides data quality reports/scorecards for different levels requirements. 

...