Most, if not all of these images are hosted through i.reddit.com or imgur, which will stop working if the original reddit or imgur, etc post is taken down as well.
That way, OP on Reddit still remains in control even if unaware of the bot.
It doesn’t stop someone from saving the image or video and republish somewhere else without OP consent (either here or on Reddit, etc). But that’s not what the bot is doing, it’s simply linking to the original source.
Looks like some CSAM fuzzy hashing would go a long way to catch someone trying to submit that kind of content if each uploaded image is scanned.
https://blog.cloudflare.com/the-csam-scanning-tool/
Not saying to go with CloudFlare (just showing how the detection works overall), but some kind of builtin detection system coded into Lemmy that grabs an updated hash table periodically
MAC address is a Level 2 addressing system (OSI model) and will not leave the local network / stay within the broadcast domain. The web browser will not expose this kind of information to a web server.
Seems like they did no longer than 24 hours ago.
https://web.archive.org/web/20230814154704/https://lemmy.ca/instances
Depending on the website, there might be some tools specifically tailored for that website you could use that will extract the content you’re looking for, but they’re likely going to be command-line based, and you’ll likely have to extract your cookies so that the tools can work as if you were logged in your account from outside your browser.
Is it too much to ask which website?
I went to a wedding last week and I was one of the very few who was actually watching the procession with my own eyes rather through a camera app.
Exactly why that during our wedding we made a PSA to not take pictures or videos, and we have a dedicated professional taking care of that and we’ll give access to those photos after.
Moral of the story: when in doubt, pixelize any potential personally identifiable information (aka PII).