some observations :

Method 1 suggests, just anonymising the incremental updates and adding it to the table, this could over-generalise the incremental data hence lose data quality → BAD IDEA

Thus any incremental data must be added to the entire database and then the anonymisation technique must be applied

Thus, the major drawback of the approach anonymizing incremental updates is twofold: it over generalizes the incremental updates, and cannot use the incremental data to reduce the information loss in the anonymization of the existing data.

Untitled

Interesting paper, talks about incremental insertions in a dataset and how k -anonymity is maintained at every release of the incremental additions of the data

Releases F(T), F1(T ∪ ∆T1), F2(T ∪ ∆T1 ∪ ∆T2), . . . are k-anonymous against incremental updates if the following three conditions are always satisfied. (1) F(T) is k-anonymous; (2) for each i ≥ 1, Fi(Ti), the re- lease after the i-th incremental update, is k-anonymous; and (3) for each nonempty set of positive integers {i1, . . . , in}, I(Fi1 (T1), . . . , Fin (Tn)), the inference table using the re- leases, is still k-anonymous.

Most specific information from all releases is called the inference table

Untitled

All the tables must maintain k-anonymity, this is a better method than ensuring the generalisatio level of a tuple is always greater than the current generalisation level, can get very difficult to track

Untitled