So because you “make” AI generated images you are saying that they are magical and don’t follow the rules of their generation?
They are based on noise maps and inferred forwards from there. They leave a history in the pixels it’s how lots of people are detecting them.
Just because it’s trained on real photos does not mean it’s still a real photo. Just cause it looks fine doesn’t mean there isn’t stuff true beneath it.
In the video I linked they even talk about how the red blue green maps have the same values cause it started with a colorless pixel anyways. A real sensor doesn’t do that.
I worked with photographers and in Photoshop and did what you think you are doing. Working with images and pixels are not just pixels. That means nothing. Dogs are just dogs. There are still different breeds and types of dogs.
So because you “make” AI generated images you are saying that they are magical and don’t follow the rules of their generation?
That’s what you got from what I wrote?
There’s nothing “magical,” but the variety of AI images that can be produced belies the simplicity of their detection. Which has been my point this whole time.
They are based on noise maps and inferred forwards from there.
There are an infinite number of methods to diffuse noise into an image, and changes to any one of a wild number of variables produces a different image. Even with the same seed and model, different noise samplers can produce entirely different types of images. And there are a LOT of different samplers. And thousands of models.
Then there are millions of LORAs that can add or remove concepts or styles. There are ControlNets that let a generator adjust other features of the image generation, from things like poses to depth mapping to edge smoothing to color noise offsets and many many many more.
The number of tweaks that can be made by someone trying to generate a specific concept is insanely high. And the outputs are wildly different.
I don’t pretend to be an expert in this subject, I’ve barely scratched the surface.
In the video I linked they even talk about how the red blue green maps have the same values cause it started with a colorless pixel anyways. A real sensor doesn’t do that.
No, they give an extremely simple explanation of how noise maps work, and then speak as if it were law, “You’ll never see an AI image that’s mostly dark with a tiny little bit of light or mostly light with a tiny little bit of dark.” Or “You won’t have an AI photo of a flat sunny field with no dark spots.”
But that’s simply not true. It’s nonsense that sounds simple enough to be believable, but the reality isn’t that simple. Each step diffuses further from the initial noise maps. And you can adjust how that happens, whether it focuses more in lighter or darker areas, or in busier or smoother areas.
Just because someone on YouTube says something with confidence doesn’t mean they’re right. YouTubers often scratch the surface of whatever they’re researching to give an overview of the subject because that’s their job. I don’t fault them for it. But they aren’t experts.
(Neither am I, but I know enough to know how insanely much there is that I—and they—don’t know.)
None of the things they say in that video as though they are law or fact are things that haven’t already been thought of by people who know far more about the subject than these YouTubers (or me).
I did mention earlier that this sort of thing might be true for Dall-E or Midjourney or other cheap/free online services with no settings the user can tweak. AI images generated with as few steps as possible, with as little machine use as possible. They will be easier to spot, more uniform. But those aren’t all there is of AI images.
Another thing to consider: this technology is, at any given moment, at the worst it’s going to be going forward. The leaps and bounds that have been made in image diffusion even in the last year is remarkable. It is currently, sometimes, difficult to detect AI images. As time goes on, it will become harder.
Obviously it doesn’t apply to everything but nothing does. But I asked a question about using a different method of detecting AI images using the fact that color brightness does still average out and base values are usually identical and was met with condescension and incorrect information from you as well as to how color in pixel math work.
You started with dismissal and haven’t gotten better. It’s been an argument and an uphill battle to point out that this is true and yet you push it off because it’s easier to hold your position.
I wanted a conversation and you wanted to punch down. You still want to be from the pulpit of right because you like your toy. I’m done talking to you.
Why on earth are you taking this so personally? We’re talking about AI image generation, why is your pride involved?
I asked a question about using a different method of detecting AI images using the fact that color brightness does still average out and base values are usually identical and was met with condescension and incorrect information from you as well as to how color in pixel math work.
You asked a question about why tools don’t use an extremely simple method of detecting AI images. I said that wouldn’t work. Initially I misunderstood your question and my response was overly simple, but it wasn’t wrong. Simple methods of detecting AI images don’t work for all AI images.
You started with dismissal and haven’t gotten better.
I didn’t dismiss you. If I had I wouldn’t have bothered to respond. You hadn’t presented much besides a vague question initially, and I disagreed with it.
When you came back with more I presented my position, that AI image generation is much more varied and complicated than your question and YouTube video assume. Just because I’m disagreeing with you and providing context doesn’t mean I’m dismissing you. Dismissing you would be to say, “No, you’re wrong, go away.” Not to explain why the simple method you’re talking about isn’t feasible, broadly, for the entirety of AI images.
If I wanted to dismiss you, I wouldn’t bother wasting my time on a response.
It’s been an argument and an uphill battle to point out that this is true
And you’re accusing me of clinging to my position. 🙄
I wanted a conversation and you wanted to punch down. You still want to be from the pulpit of right because you like your toy.
Where on earth did you get the impression that I want to be right because I like AI image generation? Or that I wanted to punch down?
Someone disagreeing with you and responding to your argument without accepting it isn’t dismissal, it isn’t punching down, it isn’t condescension. It’s engagement with what you’re saying. Just because I don’t agree with you doesn’t mean I think I’m better than you or smarter than you or anything like that, it just means I think I’m right.
So because you “make” AI generated images you are saying that they are magical and don’t follow the rules of their generation?
They are based on noise maps and inferred forwards from there. They leave a history in the pixels it’s how lots of people are detecting them.
Just because it’s trained on real photos does not mean it’s still a real photo. Just cause it looks fine doesn’t mean there isn’t stuff true beneath it.
In the video I linked they even talk about how the red blue green maps have the same values cause it started with a colorless pixel anyways. A real sensor doesn’t do that.
I worked with photographers and in Photoshop and did what you think you are doing. Working with images and pixels are not just pixels. That means nothing. Dogs are just dogs. There are still different breeds and types of dogs.
That’s what you got from what I wrote?
There’s nothing “magical,” but the variety of AI images that can be produced belies the simplicity of their detection. Which has been my point this whole time.
There are an infinite number of methods to diffuse noise into an image, and changes to any one of a wild number of variables produces a different image. Even with the same seed and model, different noise samplers can produce entirely different types of images. And there are a LOT of different samplers. And thousands of models.
Then there are millions of LORAs that can add or remove concepts or styles. There are ControlNets that let a generator adjust other features of the image generation, from things like poses to depth mapping to edge smoothing to color noise offsets and many many many more.
The number of tweaks that can be made by someone trying to generate a specific concept is insanely high. And the outputs are wildly different.
I don’t pretend to be an expert in this subject, I’ve barely scratched the surface.
No, they give an extremely simple explanation of how noise maps work, and then speak as if it were law, “You’ll never see an AI image that’s mostly dark with a tiny little bit of light or mostly light with a tiny little bit of dark.” Or “You won’t have an AI photo of a flat sunny field with no dark spots.”
But that’s simply not true. It’s nonsense that sounds simple enough to be believable, but the reality isn’t that simple. Each step diffuses further from the initial noise maps. And you can adjust how that happens, whether it focuses more in lighter or darker areas, or in busier or smoother areas.
Just because someone on YouTube says something with confidence doesn’t mean they’re right. YouTubers often scratch the surface of whatever they’re researching to give an overview of the subject because that’s their job. I don’t fault them for it. But they aren’t experts.
(Neither am I, but I know enough to know how insanely much there is that I—and they—don’t know.)
None of the things they say in that video as though they are law or fact are things that haven’t already been thought of by people who know far more about the subject than these YouTubers (or me).
I did mention earlier that this sort of thing might be true for Dall-E or Midjourney or other cheap/free online services with no settings the user can tweak. AI images generated with as few steps as possible, with as little machine use as possible. They will be easier to spot, more uniform. But those aren’t all there is of AI images.
Another thing to consider: this technology is, at any given moment, at the worst it’s going to be going forward. The leaps and bounds that have been made in image diffusion even in the last year is remarkable. It is currently, sometimes, difficult to detect AI images. As time goes on, it will become harder.
(Which your video example even says.)
Obviously it doesn’t apply to everything but nothing does. But I asked a question about using a different method of detecting AI images using the fact that color brightness does still average out and base values are usually identical and was met with condescension and incorrect information from you as well as to how color in pixel math work.
You started with dismissal and haven’t gotten better. It’s been an argument and an uphill battle to point out that this is true and yet you push it off because it’s easier to hold your position.
I wanted a conversation and you wanted to punch down. You still want to be from the pulpit of right because you like your toy. I’m done talking to you.
Why on earth are you taking this so personally? We’re talking about AI image generation, why is your pride involved?
You asked a question about why tools don’t use an extremely simple method of detecting AI images. I said that wouldn’t work. Initially I misunderstood your question and my response was overly simple, but it wasn’t wrong. Simple methods of detecting AI images don’t work for all AI images.
I didn’t dismiss you. If I had I wouldn’t have bothered to respond. You hadn’t presented much besides a vague question initially, and I disagreed with it.
When you came back with more I presented my position, that AI image generation is much more varied and complicated than your question and YouTube video assume. Just because I’m disagreeing with you and providing context doesn’t mean I’m dismissing you. Dismissing you would be to say, “No, you’re wrong, go away.” Not to explain why the simple method you’re talking about isn’t feasible, broadly, for the entirety of AI images.
If I wanted to dismiss you, I wouldn’t bother wasting my time on a response.
And you’re accusing me of clinging to my position. 🙄
Where on earth did you get the impression that I want to be right because I like AI image generation? Or that I wanted to punch down?
Someone disagreeing with you and responding to your argument without accepting it isn’t dismissal, it isn’t punching down, it isn’t condescension. It’s engagement with what you’re saying. Just because I don’t agree with you doesn’t mean I think I’m better than you or smarter than you or anything like that, it just means I think I’m right.