Skip to main content

Data Masking When Ingesting Into Databricks

 

Photo by Alba Leader

Data masking is a data security technique that involves hiding data by changing its original numbers and letters. It's a way to create a fake version of data that's similar enough to the actual data, while still protecting it. This fake data can then be used as a functional alternative when the real data isn't needed. 



Unity Catalog is not a feature within Databricks. Instead, Databricks provides the Delta Lake feature, which includes data governance capabilities such as row filters and column masking.

Unity Catalog in Databricks allows you to apply data governance policies such as row filters and column masks to sensitive data. Let’s break it down:

  1. Row Filters:

    • Row filters enable you to apply a filter to a table so that subsequent queries only return rows for which the filter predicate evaluates to true.
    • To create a row filter, follow these steps:
      1. Write a SQL user-defined function (UDF) to define the filter policy.
      • CREATE FUNCTION <function_name> (<parametergoog_1380099708_name> <parameter_type>, ...) RETURN {filter clause whobe a boolean};
  2. Apply the row filter to an existing table using the following syntax:
    ALTER TABLE <table_name> SET ROW FILTER <function_name> ON (<column_name>, ...);
      1. You can also specify a row filter during the initial table creation.
    • Each table can have only one row filter, and it accepts input parameters that bind to specific columns of the table.
  3. Column Masks:

    • Column masks allow you to transform or mask specific column values before returning them in query results.
    • To apply column masks:
      1. Create a function that defines the masking logic.
      2. Apply the masking function to a table column using an ALTER TABLE statement.
      3. Alternatively, you can apply the masking function during table creation.
  4. Unity Catalog Best Practices:

  5. When setting up Unity Catalog, consider assigning a location to a catalog level. For example:
    CREATE CATALOG hr_prod
    LOCATION 'abfss://mycompany-hr-prod@storage-account.dfs.core.windows.net/unity-catalog';

You can apply column masks to transform or conceal specific column values before returning them in query results. Here’s how you can achieve this:

  1. Create a Masking Function:

    • Define a function that specifies the masking logic. This function will be used to transform the column values.
    • For example, let’s say you want to mask the last four digits of a credit card number. You can create a masking function that replaces the last four digits with asterisks.
  2. Apply the Masking Function to a Column:

    • Use an ALTER TABLE statement to apply the masking function to a specific column.
    • For instance, if you have a column named credit_card_number, you can apply the masking function to it:
      ALTER TABLE my_table SET COLUMN MASK credit_card_number USING my_masking_function;
      
  3. Example Masking Function:

    • Suppose you want to mask the last four digits of a credit card number with asterisks. You can create a masking function like this:
      CREATE FUNCTION my_masking_function AS
      BEGIN
          RETURN CONCAT('************', RIGHT(credit_card_number, 4));
      END;
      
  4. Query the Table:

    • When querying the table, the masked values will be returned instead of the original values.

Let’s focus on how you can achieve column masking in Databricks using Delta Lake:

  1. Column Masking:

    • Delta Lake allows you to apply column-level transformations or masks to sensitive data.
    • You can create custom masking functions to modify specific column values before returning them in query results.
  2. Creating a Masking Function:

    • Define a user-defined function (UDF) that specifies the masking logic. For example, you can create a function that masks the last four digits of a credit card number.
    • Here’s an example of a masking function that replaces the last four digits with asterisks:
      def mask_credit_card(card_number):
          return "************" + card_number[-4:]
      
  3. Applying the Masking Function:

    • Use the withColumn method to apply the masking function to a specific column in your DataFrame.
    • For instance, if you have a DataFrame named my_table with a column named credit_card_number, you can apply the masking function as follows:
      from pyspark.sql.functions import udf
      from pyspark.sql.types import StringType
      
      # Register the UDF
      spark.udf.register("mask_credit_card", mask_credit_card, StringType())
      
      # Apply the masking function to the column
      masked_df = my_table.withColumn("masked_credit_card", udf("credit_card_number"))
      
  4. Querying the Masked Data:

    • When querying the masked_df, the transformed (masked) values will be returned for the masked_credit_card column.

You can find different related articles here kindly search.


Comments

Popular posts from this blog

Financial Engineering

Financial Engineering: Key Concepts Financial engineering is a multidisciplinary field that combines financial theory, mathematics, and computer science to design and develop innovative financial products and solutions. Here's an in-depth look at the key concepts you mentioned: 1. Statistical Analysis Statistical analysis is a crucial component of financial engineering. It involves using statistical techniques to analyze and interpret financial data, such as: Hypothesis testing : to validate assumptions about financial data Regression analysis : to model relationships between variables Time series analysis : to forecast future values based on historical data Probability distributions : to model and analyze risk Statistical analysis helps financial engineers to identify trends, patterns, and correlations in financial data, which informs decision-making and risk management. 2. Machine Learning Machine learning is a subset of artificial intelligence that involves training algorithms t...

Wholesale Customer Solution with Magento Commerce

The client want to have a shop where regular customers to be able to see products with their retail price, while Wholesale partners to see the prices with ? discount. The extra condition: retail and wholesale prices hasn’t mathematical dependency. So, a product could be $100 for retail and $50 for whole sale and another one could be $60 retail and $50 wholesale. And of course retail users should not be able to see wholesale prices at all. Basically, I will explain what I did step-by-step, but in order to understand what I mean, you should be familiar with the basics of Magento. 1. Creating two magento websites, stores and views (Magento meaning of website of course) It’s done from from System->Manage Stores. The result is: Website | Store | View ———————————————— Retail->Retail->Default Wholesale->Wholesale->Default Both sites using the same category/product tree 2. Setting the price scope in System->Configuration->Catalog->Catalog->Price set drop-down to...

How to Prepare for AI Driven Career

  Introduction We are all living in our "ChatGPT moment" now. It happened when I asked ChatGPT to plan a 10-day holiday in rural India. Within seconds, I had a detailed list of activities and places to explore. The speed and usefulness of the response left me stunned, and I realized instantly that life would never be the same again. ChatGPT felt like a bombshell—years of hype about Artificial Intelligence had finally materialized into something tangible and accessible. Suddenly, AI wasn’t just theoretical; it was writing limericks, crafting decent marketing content, and even generating code. The world is still adjusting to this rapid shift. We’re in the middle of a technological revolution—one so fast and transformative that it’s hard to fully comprehend. This revolution brings both exciting opportunities and inevitable challenges. On the one hand, AI is enabling remarkable breakthroughs. It can detect anomalies in MRI scans that even seasoned doctors might miss. It can trans...