Data Governance Strategy Course: Zurich, Switzerland
I had the opportunity to make a trip to Switzerland and instruct the Hortonworks University Data Governance course. I had an excellent experience traveling to Switzerland. In addition to having the opportunity to work with a very talented team. We covered how to use Hortonworks Data Platform 3.0 to create a data governance solution strategy for enterprise data. In addition to HDP 3.0 we discussed Data Plane Services and how Data Steward Studio can be used to profile data from HDP to give insight on data collections and assets. We spent a lot of time discussing the integration of Atlas classifications to entities in HDP and how to assign Ranger security policies to those classifications. There was a lot of interest in designing custom Apache Atlas integrations not only with the HDP ecosystem but other external data entities.
In the discussion, we discussed some topics to develop a data governance solution. Out of the discussion I want to mention a few points to consider when designing a data governance solution with HDP.
- When creating classifications in Apache Atlas try to match them to your business taxonomy structure.
- When integrating 3rd party data stores with Apache Atlas utilized the import/export APIs to bring in data to Atlas to create entities.( You can utilize custom adapters for some 3rd party vendors.)
- Consider using tag based policies in Ranger for metadata driven security policies.
- Enable HDP Hooks for metadata ingestion from HDP components.
- Use Atlas Reporting task controller service in HDF Nifi for Atlas integration to retrieve metadata throughout your NiFi data pipeline.
While in Switzerland I had the opportunity to go to Old town and Bahnhofstrasse street.