So, I’m selfhosting immich, the issue is we tend to take a lot of pictures of the same scene/thing to later pick the best, and well, we can have 5~10 photos which are basically duplicates but not quite.
Some duplicate finding programs put those images at 95% or more similarity.

I’m wondering if there’s any way, probably at file system level, for the same images to be compressed together.
Maybe deduplication?
Have any of you guys handled a similar situation?

  • @ShortN0te@lemmy.ml
    link
    fedilink
    English
    010 months ago

    Would be surprised if there is any AI involved. Finding duplicates is a solved problem.

    AI is only involved in object detection and face recognition.

    • @tehnomad@lemm.ee
      link
      fedilink
      English
      010 months ago

      I wasn’t sure if it was AI or not. According to the description on GitHub:

      Utilizes state-of-the-art algorithms to identify duplicates with precision based on hashing values and FAISS Vector Database using ResNet152.

      Isn’t ResNet152 a neural network model? I was careful to say neural network instead of AI or machine learning.