<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Introduction on</title><link>https://docs.protegrity.com/anonymization/1.4.0/docs/introduction/</link><description>Recent content in Introduction on</description><generator>Hugo</generator><language>en</language><lastBuildDate>Fri, 06 Feb 2026 09:26:52 +0000</lastBuildDate><atom:link href="https://docs.protegrity.com/anonymization/1.4.0/docs/introduction/index.xml" rel="self" type="application/rss+xml"/><item><title>Business cases</title><link>https://docs.protegrity.com/anonymization/1.4.0/docs/introduction/hide_intro_buscase/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>https://docs.protegrity.com/anonymization/1.4.0/docs/introduction/hide_intro_buscase/</guid><description>&lt;p>Consider the following business cases:&lt;/p>
&lt;ul>
&lt;li>&lt;strong>Case 1&lt;/strong>: A hospital wants to share patient data with a third-party research lab. The privacy of the patient, however, must be preserved.&lt;/li>
&lt;li>&lt;strong>Case 2&lt;/strong>: An organization requires customer data from several credit unions to create training data. The data will be used to train machine learning models looking for new insights. The customers, however, have not agreed for their data to be used.&lt;/li>
&lt;li>&lt;strong>Case 3&lt;/strong>: An organization which must be compliant with GDPR, CCPA, or other privacy regulations requires to keep some information beyond the period that meets regulations.&lt;/li>
&lt;li>&lt;strong>Case 4&lt;/strong>: An organization requires raw data to train their software for machine learning.&lt;/li>
&lt;/ul>
&lt;p>In all these cases, data forms an integral part of the source for continuing the business process or analysis. Additionally, only &lt;em>what was done&lt;/em> is required in all the cases, &lt;em>who did it&lt;/em> does not have any value in the data. In this case, the personal information about the individual users can be removed from the dataset. This removes the personal factor from the data and at the same time retains the value of the data from the business point of view. This data, since it does not have any private information, is also pulled from the legal requirements governing the data.&lt;/p></description></item><item><title>Data security and data privacy</title><link>https://docs.protegrity.com/anonymization/1.4.0/docs/introduction/hide_data_priv/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>https://docs.protegrity.com/anonymization/1.4.0/docs/introduction/hide_data_priv/</guid><description>&lt;p>Most organizations understand the need to secure access to personally identifiable information. Sensitive values in records are often protected at rest (storage), in transit (network) and in use (fine-grained access control), through a process known as &lt;em>de-identification&lt;/em>. De-Identification is a spectrum, where data security and data privacy issues must be balanced with data usability.&lt;/p>
&lt;p>&lt;img src="https://docs.protegrity.com/anonymization/1.4.0/docs/images/hide/hide_intromap.jpg" alt="" title="Data Protection Spectrum">&lt;/p>
&lt;h2 id="pseudonymization">Pseudonymization&lt;/h2>
&lt;p>Pseudonymization is the process of de-identification by substituting sensitive values with a consistent, non-sensitive value. This is most often accomplished through encryption, tokenization, or dynamic data masking. Access to the process for re-identification (decryption, detokenization, unmasking) is controlled, so that only users with a business requirement will see the sensitive values.&lt;/p></description></item><item><title>Importance and types of data</title><link>https://docs.protegrity.com/anonymization/1.4.0/docs/introduction/hide_intro_datatypes/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>https://docs.protegrity.com/anonymization/1.4.0/docs/introduction/hide_intro_datatypes/</guid><description>&lt;p>These records might be linked with other records, such as, income statements or medical records to provide valuable information. The various fields as a whole, called a record, is private and is user-centric. However, the individual fields may or may not be personal. Accordingly, based on the privacy level, the following data classifications are available:&lt;/p>
&lt;ul>
&lt;li>&lt;strong>Direct Identifier&lt;/strong>: Identity Attributes can identify an individual with the value alone. These attributes are unique to an individual in a dataset and at times even in the world. It is personal and private to the user. For example, name, passport, Social Security Number (SSN), mobile number, and so on.&lt;/li>
&lt;li>&lt;strong>Quasi-Identifier or Indirect Identifier&lt;/strong>: Quasi-Identifying Attributes are identifying characteristic about a data subject. However, you cannot identify an individual with the quasi-identifier alone. For example, date of birth or an address. Moreover, the individual pieces of data in a quasi-identifier might not be enough to identify a single individual. Take the example of date of birth, the year might be common to many individuals and would be difficult to narrow down to a single individual. However, if the dataset is small, then it might be easy to identify an individual using this information.&lt;/li>
&lt;li>&lt;strong>Data about data subject&lt;/strong>: Data about the data subject is typically the data that is being analyzed. This data might exist in the same table or a different related table of the dataset. It provides valuable information about the dataset and is very helpful for analysis. This data may or might not be private to an individual. For example, salary, account balance, or credit limit. However, like quasi-identifiers, in a small dataset, this data might be unique to an individual. Additionally, this data can be classified as follows:
&lt;ul>
&lt;li>&lt;strong>Sensitive Attributes&lt;/strong>: This data may disclose something like a health condition which in a small result set may identify a single individual.&lt;/li>
&lt;li>&lt;strong>Insensitive Attributes&lt;/strong>: This data is not associated with a privacy risk and is common information, such as, the type of bank accounts in a bank, individual or business.&lt;/li>
&lt;/ul>
&lt;/li>
&lt;/ul>
&lt;p>A sample dataset is shown in the following figure:&lt;/p></description></item><item><title>Data anonymization techniques</title><link>https://docs.protegrity.com/anonymization/1.4.0/docs/introduction/hide_priv_types/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>https://docs.protegrity.com/anonymization/1.4.0/docs/introduction/hide_priv_types/</guid><description>&lt;h2 id="important-terminology">Important terminology&lt;/h2>
&lt;ul>
&lt;li>&lt;strong>De-identification&lt;/strong>:
General term for any process of removing the association between a set of identifying data and the data subject.&lt;/li>
&lt;li>&lt;strong>Pseudonymization&lt;/strong>:
Particular type of data de-identification that both removes the association with a data subject and adds an association between a particular set of characteristics relating to the data subject and one or more pseudonyms.&lt;/li>
&lt;li>&lt;strong>Anonymization&lt;/strong>:
Process that removes the association between the identifying dataset and the data subject.
Anonymization is another subcategory of de-identification. Unlike pseudonymization, it does not provide a means by which the information may be linked to the same person across multiple data records or information systems.
Hence reidentification of anonymized data is not possible.&lt;/li>
&lt;/ul>
&lt;p>&lt;strong>Note&lt;/strong>: As defined in ISO/TS 25237:2008.&lt;/p></description></item><item><title>How Protegrity Anonymization Works</title><link>https://docs.protegrity.com/anonymization/1.4.0/docs/introduction/hide_work_how/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>https://docs.protegrity.com/anonymization/1.4.0/docs/introduction/hide_work_how/</guid><description>&lt;!--
It processes the data, removing personal information and summarizing or generalizing the remaining information. It then provides an output with the processed data that can be used for analysis, this output has generalized data using which individuals cannot be identified. The same can be seen in the following figure:
-->
&lt;p>Protegrity Anonymization is a software solution that processes data by removing personal information and transforming the remaining details to protect privacy.&lt;/p></description></item></channel></rss>