isp-data-pollution VS amazon-s3-find-and-forget

Compare isp-data-pollution vs amazon-s3-find-and-forget and see what are their differences.

isp-data-pollution

ISP Data Pollution to Protect Private Browsing History with Obfuscation (by essandess)

amazon-s3-find-and-forget

Amazon S3 Find and Forget is a solution to handle data erasure requests from data lakes stored on Amazon S3, for example, pursuant to the European General Data Protection Regulation (GDPR) (by awslabs)
Our great sponsors
  • InfluxDB - Power Real-Time Data Analytics at Scale
  • WorkOS - The modern identity platform for B2B SaaS
  • SaaSHub - Software Alternatives and Reviews
isp-data-pollution amazon-s3-find-and-forget
2 3
566 232
- 2.6%
0.0 7.2
about 1 year ago 8 days ago
Python Python
MIT License Apache License 2.0
The number of mentions indicates the total number of mentions that we've tracked plus the number of user suggested alternatives.
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.

isp-data-pollution

Posts with mentions or reviews of isp-data-pollution. We have used some of these posts to build our list of alternatives and similar projects.

amazon-s3-find-and-forget

Posts with mentions or reviews of amazon-s3-find-and-forget. We have used some of these posts to build our list of alternatives and similar projects.
  • Deleting particular data from S3 External Tables
    1 project | /r/dataengineering | 31 Oct 2022
    Take a look at this: https://github.com/awslabs/amazon-s3-find-and-forget We use it for GDPR compliance; it will open a file, delete a row and pack it back. It will modify the file so watch out if you are using Glue job bookmarks. Because you are using external tables, the manifest file will also have to be updated with a proper lenght for the new, updated file. If you have hundreds of tables and thousands of files, and you need to do this on a regular basis this would be the scalable solution, but if you have few files honestly I would do it manually
  • Update S3 Files
    1 project | /r/aws | 27 Jan 2022
    Have a look at S3 Find and Forget
  • How to handle GDPR requests for data stored in S3 ?
    1 project | /r/dataengineering | 22 Nov 2021
    S3 Find and Forget is probably worth looking into, even if just to get ideas on how to implement a similar solution for yourself

What are some alternatives?

When comparing isp-data-pollution and amazon-s3-find-and-forget you can also consider the following projects:

Social-Amnesia - Forget the past. Social Amnesia makes sure your social media accounts only show your posts from recent history, not from "that phase" 5 years ago.

DataEngineeringProject - Example end to end data engineering project.

gretel-python-client - The Gretel Python Client allows you to interact with the Gretel REST API.

awesome-aws - A curated list of awesome Amazon Web Services (AWS) libraries, open source repos, guides, blogs, and other resources. Featuring the Fiery Meter of AWSome.

tracardi - TRACARDI is a new HOME for your customer data. TRACARDI is an Composable API-first solution for any company that need unexpensive CDP to intergrate with.

data-toolset - Upgrade from avro-tools and parquet-tools jars to a more user-friendly Python package.

tracardi - TRACARDI is a new HOME for your customer data. TRACARDI is an API-first solution, low-code / no-code platform aimed at any e-commerce business that wants to start using user data for marketing purposes.

s3-credentials - A tool for creating credentials for accessing S3 buckets

ReTube - ReImagine Tubing

Differential-Privacy-Guide - Differential Privacy Guide