Skip to content

Our project detects a hash poisoning attack, in which the poison image having the same hash of the innocent image is uploaded and causes a denial of service to the legitimate users trying to upload, but because it has the same hash as that of poison image, the system compares the hash with the hash table and deletes the innocent image. We aim to…

Notifications You must be signed in to change notification settings

ankitd3/poisoning_attack_perceptual_hash

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

10 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Hash Poisoning Attack and Steganography Attack with mitigation using MD5 Hash

Use Case (Reference link)

A system that allows users to submit photos to a database of images to ban. A human reviews the image to ensure it is an image that deserves banning (and that the image is say, not the Coca-Cola logo). If approved, this hash gets added to the database and is checked against whenever a new image is uploaded. If this new image’s hash collides with the banned hash, the image is prevented from being uploaded.

Block

Usage

  1. Run the (Server side) - "app.py" (For phase 1 use case) and "app_steg.py" for steganography check (Not simultaneously)
  2. Run test.py alog with app.py
  3. Change the IP Address to your IP address in both app.py and app_steg.py for access to clients in your network

About

Our project detects a hash poisoning attack, in which the poison image having the same hash of the innocent image is uploaded and causes a denial of service to the legitimate users trying to upload, but because it has the same hash as that of poison image, the system compares the hash with the hash table and deletes the innocent image. We aim to…

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published