ACHIEVING k-ANONYMITY PRIVACY PROTECTION USING GENERALIZATION AND SUPPRESSION
Often a data holder, such as a hospital or bank, needs to share person-specific records in such a way that the identities of the individuals who are the subjects of the data cannot be determined. One way to achieve this is to have the released records adhere to k-anonymity, which means each released...
Saved in:
| Published in | International journal of uncertainty, fuzziness, and knowledge-based systems Vol. 10; no. 5; pp. 571 - 588 |
|---|---|
| Main Author | |
| Format | Journal Article |
| Language | English |
| Published |
World Scientific Publishing Company
01.10.2002
|
| Subjects | |
| Online Access | Get full text |
| ISSN | 0218-4885 1793-6411 |
| DOI | 10.1142/S021848850200165X |
Cover
| Summary: | Often a data holder, such as a hospital or bank, needs to share
person-specific records in such a way that the identities of
the individuals who are the subjects of the data cannot be determined.
One way to achieve this is to have the released records adhere to
k-anonymity, which means each released record has at least
(k-1) other records in the release whose values are indistinct
over those fields that appear in external data. So, k-anonymity
provides privacy protection by guaranteeing that each released record will
relate to at least k individuals even if the records are directly linked
to external information. This paper provides a formal presentation of combining
generalization and suppression to achieve k-anonymity. Generalization
involves replacing (or recoding) a value with a less specific but semantically
consistent value. Suppression involves not releasing a value at all. The
Preferred Minimal Generalization Algorithm (MinGen), which is a theoretical
algorithm presented herein, combines these techniques to provide
k-anonymity protection with minimal distortion. The
real-world algorithms Datafly and μ-Argus are compared
to MinGen. Both Datafly and μ-Argus use heuristics to make
approximations, and so, they do not always yield optimal results. It
is shown that Datafly can over distort data and μ-Argus can
additionally fail to provide adequate protection. |
|---|---|
| Bibliography: | This paper significantly amends and expands the earlier paper "Protecting privacy when disclosing information: k-anonymity and its enforcement through generalization and suppression" (with Samarati) submitted to IEEE Security and Privacy 1998, and extends parts of my Ph.D. thesis [10]. |
| ISSN: | 0218-4885 1793-6411 |
| DOI: | 10.1142/S021848850200165X |