Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.


Table of Contents

Document datasets

Image RemovedImage Added

  1. Click the Datasets tab.
  2. Select the dataset you want to classify.

Image RemovedImage Added

  1. Classify the dataset in this window.

...

Add a comment to the dataset

Document personal data

Image RemovedImage Added

To access the personal data panel :

...

Tag dataset containing personal data

Image RemovedImage Added

  1. Check "contains personal data" box if the dataset contains personal data

Document the consent and / or configure anonymization process

Image RemovedImage Added

  1. Click to Edit Settings to document the consent and / or configure anonymization process

...

Select fields to be anonymized

Image RemovedImage Added

  1. Select fields to be anonymized

...

Warning

Csv and parquet files on table TABLE are supported in default anonymization process. More technical formats will be released at a later date (FILE and DIRECTORY are not currently supported).


Warning

Anonymization process substitute the existing data by a generated value respecting the same format. 

This operation is not reversible.

Anonymization job

An default anonymization process in scala is proposed by Saagie. The code source is available here : https://github.com/saagie/outis.

You can use it as is, change it or replace it with one of your process.

To use it, build jar and create Spark processing job on platform with command line : 


Code Block
spark-submit \
--conf "spark.executor.extraJavaOptions='-Dlog4j.configuration=log4j.xml'" \
--conf spark.ui.showConsoleProgress=false \
--driver-java-options "-Dlog4j.configuration=log4j.xml" \
{file} -u hdfs_user -t metastore_url -d datagov_user -p $ENV_VAR_PASSWORD datasetsToAnonymized_url callback_url

where :

  • hdfs_user = user to launch job

...

  • - user must have right to write in hdfs
  • metastore_url = url of the hive metastore (exp : thrift://nn1:9083)
  • datagov_user = user to access to Data Governance on the platform with right "Access all datasets" (may be the same as hdfs_user)
  • datasetsToAnonymized_url = url to obtain datasets to anonymized (exp : http://{IP_DATAGOVERNANCE}:{PORT}/api/v1/datagovernance/platform/{PLATFORM_ID}/privacy/datasets)
  • callback_url = url to inform dataset is anonymized (exp : http://{IP_DATAGOVERNANCE}:{PORT}/api/v1/datagovernance/platform/{PLATFORM_ID}/privacy/events/datasetAnonymized)
  • $ENV_VAR_PASSWORD : environment variable for password


You can dowload the last version of the jar here :  

outis-link-1.1.0.jar

Exceptions handling

No dataset anonymization if : 
  • You don't provide an entry date
  • You don't provide a list of fields to anonymize
  • If you provide a string field as an entry date without a pattern to parse the data
  • Dataset isn't csv or parquet files on table
No record anonymization if : 
  • The value of the field to anonymize has a NULL value
  • You try to replace a string field mark as a date without providing a pattern
  • An error occured


Managed types

String anonymization : 
The strings fields are anonymized by substitution (character by character) 
  • If the character is a digit, he is substituted by an another digit
  • If the character is a letter, he is substituted by an another letter
  • otherwise the character remains unchanged
Date anonymization : 
The date fields are anonymized randomly between January 1, 1920 and now :
  • If the field is a String type and tagged as a Date type, a randomized date in String format with the same pattern is generated 

  • If the field is a Timestamp type, a randomized Timestamp is generated

  • if the field is a Date type, a randomized Date is generated

Numeric anonymization : 
All numeric types are anonymized randomly. The generated value can not exceed the type max value.
This covers these types : Byte, Short, Int, Long, Float, Double and BigDecimal.

Error

You can find execution errors of the job in Error Output part of Job Logs.