Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

This one? https://www.youtube.com/watch?v=Vz0oQ0v0W10

This comment is pretty depressing but it seems to be the path we're headed to:

> It's bad enough that people think fake videos are real, but they also now think real videos are fake. My channel is all wildlife that I filmed myself in my own yard, and I've had people leaving comments that it's AI, because the lighting is too pretty or the bird is too cute. The real world is pretty and cute all the time, guys! That's why I'm filming it!

Combine this with selecting only what you want to believe in and you can say that video/image that goes against your "facts" is "fake AI". We already have some people in pretty powerful positions doing this to manipulate their bases.



> We already have some people in pretty powerful positions doing this to manipulate their bases.

You don't have to be vague. Let's be specific. The President of the United States implied a very real voiceover of President Reagan was AI. Reagan was talking about the fallacy of tariffs as engines of economic growth, and it was used in an ad by the government of Ontario to sow divide within Republicans. It worked, and the President was nakedly mad at being told by daddy Reagan.


We are heading to an apocalyptic level of psychosis where human beings won't even believe the things they see with their own eyes are real anymore because of being flooded with AI slop 24/7/365.


We desperately need a technological solution to be able to somehow "sign" images and videos as being real and not generated or manipulated by AI.

I have no idea how such a thing would work.


It won't work, because most people do not understand what a digital signature is and they will just say that has been faked as well.


Journalists will know how to check it in high profile cases.

And annoyed and suspicious techies can use it to check other people's content and report them as fake.

Yeah, there are a lot of dumb people who want to be deceived. But would be good for the rest of us to have some tools.


This will just create a black market for AI generated video content that doesn't have a signature. Which I'm sure that China, Russia, hell even the US governments would not have a problem with because that would be extremely useful for them.


So everything without a signature should just be assumed to be AI.


Honestly - not gonna be feasible to sign things that are not AI-generated. For example, my friend's phone uses generative AI for the insane zoom capability it has - so every photo my friend takes would not be able to be signed.

And this is the crux of the issue - we are beelining to a world where EVERYTHING gets an AI filter in front of it. In a few years there will be no authentic content at all.


Enter “analog deep-fakers” who create misleading images using the lost arts of our ancestors. I guess it would at least be more expensive than AI, so might not be as prevalent and damaging.


Such a thing exists currently, at least for images created by Google and other major AI companies.


We desperately need a high-tech solution to disengage from the torrent of AI slop. I have found a form of photosynthesizing plant that grows in the soil of many places where humans live. Apparently, touching this with the palm of your hand creates a kind of protective endorphin effect that coalesces conflicting versions of reality into the one you're actually experiencing. It'll take a lot of money to advertise this to the general population, but I'm fundraising as hard as I can: this approach will be called "Touch Grass" and I hope to see you all when we do it together.


I feel bad for the guy but I think this confusion will be extraordinary and get people off the internet.


There was a discussion on here recently about a new camera that could prove images taken with it weren't AI fakes, and most of the comments were skeptical anyone would care about such things.

This is an example of how people viscerally hate anyone passing off AI generated images and video as real.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: