dataguise presents 10 best practices for securing sensitive data in hadoop: start early! determine the data privacy protection strategy during theplanning phase of a deployment, preferably before moving any data intohadoop. this will preve
dataguise presents 10 best practices for securing sensitive data in hadoop: start early! determine the data privacy protection strategy during theplanning phase of a deployment, preferably before moving any data intohadoop. this will prevent the possibility of damaging compliance exposurefor the company and avoid unpredictability in the roll out schedule.
identify what data elements are defined as sensitive within yourorganization. consider company privacy policies, pertinent industryregulations and governmental regulations.
discover whether sensitive data is embedded in the environment, assembledor will be assembled in hadoop.
determine the compliance exposure risk based on the informationcollected.
determine whether business analytic needs require access to real data orif desensitized data can be used. then, choose the right remediationtechnique (masking or encryption). if in doubt, remember that maskingprovides the most secure remediation while encryption provides the mostflexibility, should future needs evolve.
ensure the data protection solutions under consideration support bothmasking and encryption remediation techniques, especially if the goal is tokeep both masked and unmasked versions of sensitive data in separate hadoopdirectories.
ensure the data protection technology used implements consistent maskingacross all data files (joe becomes dave in all files) to preserve theaccuracy of data analysis across every data aggregation dimensions.
determine whether a tailored protection for specific data sets isrequired and consider dividing hadoop directories into smaller groups wheresecurity can be managed as a unit.?
ensure the selected encryption solution interoperates with the company’saccess control technology and that both allow users with differentcredentials to have the appropriate, selective access to data in the hadoopcluster.
ensure that when encryption is required, the proper technology (java,pig, etc.) is deployed to allow for seamless decryption and ensure expeditedaccess to data.
wait… where’s point 11, buy dataguise?
original title and link: dataguise presents 10 best practices for securing sensitive data in hadoop (nosql database?mynosql)
原文地址:dataguise presents 10 best practices for securing , 感谢原作者分享。