<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Introduction on</title><link>https://docs.protegrity.com/anonymization/1.4.1/docs/introduction/</link><description>Recent content in Introduction on</description><generator>Hugo</generator><language>en</language><lastBuildDate>Mon, 06 Apr 2026 11:24:23 +0000</lastBuildDate><atom:link href="https://docs.protegrity.com/anonymization/1.4.1/docs/introduction/index.xml" rel="self" type="application/rss+xml"/><item><title>Business cases</title><link>https://docs.protegrity.com/anonymization/1.4.1/docs/introduction/hide_intro_buscase/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>https://docs.protegrity.com/anonymization/1.4.1/docs/introduction/hide_intro_buscase/</guid><description>&lt;p>Consider the following business cases:&lt;/p>
&lt;ul>
&lt;li>&lt;strong>Case 1&lt;/strong>: A hospital wants to share patient data with a third-party research lab. The privacy of the patient, however, must be preserved.&lt;/li>
&lt;li>&lt;strong>Case 2&lt;/strong>: An organization requires customer data from several credit unions to create training data. The data will be used to train machine learning models looking for new insights. The customers, however, have not agreed to their data to be used.&lt;/li>
&lt;li>&lt;strong>Case 3&lt;/strong>: An organization which must be compliant with GDPR, CCPA, or other privacy regulations requires to keep some information beyond the period that meets regulations.&lt;/li>
&lt;li>&lt;strong>Case 4&lt;/strong>: An organization requires raw data to train their software for machine learning.&lt;/li>
&lt;/ul>
&lt;p>In all these cases, data forms an integral part of the source for continuing the business process or analysis. Additionally, only &lt;em>what was done&lt;/em> is required in all the cases, &lt;em>who did it&lt;/em> does not have any value in the data. In this case, personal information about individual users can be removed from the dataset. This removes the personal factor from the data and at the same time retains the value of the data from the business point of view. This data, since it does not have any private information, is also pulled from the legal requirements governing the data.&lt;/p></description></item><item><title>Data security and data privacy</title><link>https://docs.protegrity.com/anonymization/1.4.1/docs/introduction/hide_data_priv/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>https://docs.protegrity.com/anonymization/1.4.1/docs/introduction/hide_data_priv/</guid><description>&lt;p>Most organizations understand the need to secure access to personally identifiable information. Sensitive values in records are often protected at rest (storage), in transit (network) and in use (fine-grained access control), through a process known as &lt;em>de-identification&lt;/em>. De-Identification is a spectrum where data security and data privacy issues must be balanced with data usability.&lt;/p>
&lt;p>&lt;img src="https://docs.protegrity.com/anonymization/1.4.1/docs/images/hide/hide_intromap.jpg" alt="" title="Data Protection Spectrum">&lt;/p>
&lt;h2 id="pseudonymization">Pseudonymization&lt;/h2>
&lt;p>Pseudonymization is the process of de-identification by substituting sensitive values with a consistent, non-sensitive value. This is most often accomplished through encryption, tokenization, or dynamic data masking. Access to the process for re-identification (decryption, detokenization, unmasking) is controlled, so that only users with a business requirement will see the sensitive values.&lt;/p></description></item><item><title>Importance and types of data</title><link>https://docs.protegrity.com/anonymization/1.4.1/docs/introduction/hide_intro_datatypes/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>https://docs.protegrity.com/anonymization/1.4.1/docs/introduction/hide_intro_datatypes/</guid><description>&lt;p>A record consists of all the information pertaining to a user. This record consists of different fields of information, such as first name, last name, address, telephone number, age, and so on. These records might be linked with other records, such as income statements or medical records to provide valuable information. A record is made up of various fields and is private and user-centric. However, the individual fields may or may not be personal. Accordingly, based on the privacy level, the following data classifications are available:&lt;/p></description></item><item><title>Data Anonymization Techniques</title><link>https://docs.protegrity.com/anonymization/1.4.1/docs/introduction/hide_priv_types/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>https://docs.protegrity.com/anonymization/1.4.1/docs/introduction/hide_priv_types/</guid><description>&lt;h2 id="important-terminology">Important terminology&lt;/h2>
&lt;ul>
&lt;li>&lt;strong>De-identification&lt;/strong>:
A general term for any process of removing the association between a set of identifying data and the data subject.&lt;/li>
&lt;li>&lt;strong>Pseudonymization&lt;/strong>:
A particular type of data de‑identification that removes the direct association with a data subject. It replaces that association by linking a specific set of characteristics to one or more pseudonyms.&lt;/li>
&lt;li>&lt;strong>Anonymization&lt;/strong>:
A process that removes the association between the identifying dataset and the data subject.
Anonymization is another subcategory of de-identification. Unlike pseudonymization, it does not provide a means by which the information may be linked to the same person across multiple data records or information systems.
Hence reidentification of anonymized data is not possible.&lt;/li>
&lt;/ul>
&lt;p>&lt;strong>Note&lt;/strong>: As defined in ISO/TS 25237:2008.&lt;/p></description></item><item><title>How Protegrity Anonymization Works</title><link>https://docs.protegrity.com/anonymization/1.4.1/docs/introduction/hide_work_how/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>https://docs.protegrity.com/anonymization/1.4.1/docs/introduction/hide_work_how/</guid><description>&lt;p>Protegrity Anonymization is a software solution that processes data by removing personal information and transforming the remaining details to protect privacy.&lt;/p>
&lt;p>In simple terms, it takes raw data as input, applies techniques like generalization and summarization, and outputs anonymized data that can still be used for analysis—without revealing individual identities. The following figure illustrates this process.&lt;/p>
&lt;p>&lt;img src="https://docs.protegrity.com/anonymization/1.4.1/docs/images/hide/hide_intro_sampleprocess.jpg" alt="" title="Basics of Protegrity Anonymization">&lt;/p>
&lt;p>As shown in the above image, a sample table is fed as input into Protegrity Anonymization. The private data that can be used to identify a particular individual is removed from the table. The final table with anonymized information is provided as output. The output table shows data loss due to column and row removals during Protegrity Anonymization. This data loss is necessary to mitigate the risk of de-identification.&lt;/p></description></item></channel></rss>