As organizations pursue Hadoop initiatives to capture new opportunities for data-driven insights, data governance and data security requirements can pose a key challenge. Hortonworks created an Apache Hadoop Data Governance Initiative to address the need for open source governance solution to manage data classification, data lineage, security and data lifecycle management.
Uma boa gestão e controle de dados não pode ser passiva ou simplesmente forense. O controle de acesso centralizado alimentado por classificação de dados consistente é a base da segurança dinâmica e um requisito fundamental para o Open Enterprise Hadoop. Para atingir esse objetivo, a Hortonworks anuncia o lançamento de novos recursos de visualização pública com o Apache Atlas e o Apache Ranger, que combinam classificação de dados com aplicação de políticas de segurança.
Apache Atlas, created as part of the Hadoop data governance initiative, empowers organizations to apply consistent data classification across the data ecosystem. Apache Ranger provides centralized security administration for Hadoop. By integrating Atlas with Ranger, Hortonworks empowers enterprises to institute dynamic access policies at run time that proactively prevents violations from occurring.
The Atlas/ Ranger integration represents a paradigm shift for big data governance and data security in Apache Hadoop. By integrating Atlas with Ranger enterprises can now implement dynamic classification-based security policies, in addition to role-based security. Ranger’s centralized platform empowers data administrators to define security policy based on Atlas metadata tags or attributes and apply this policy in real-time to the entire hierarchy of data assets including databases, tables and columns.
Hortonworks empowers data managers to ensure the transparency, reproducibility, auditability and consistency of the Data Lake and the assets it contains. Apache Atlas now provides the ability to visualize cross-component lineage, delivering a complete view of data movement across a number of analytic engines such as Apache Storm, Kafka, Falcon and Hive. Hadoop operations, stewards, operations, and compliance personnel now have the ability to visualize a data set’s lineage and then drill down into operational, security and provenance-related details. As this tracking is done at the platform level, any application that uses multiple engines will be natively tracked. This allows for extended visibility beyond a single application view.