Uhhh I don’t think so.

Reply to this note

Please Login to reply.

Discussion

Why not? Our attention is separating good from bad pretty effectively. So I can really see the nonsense, garbage, hallucinations, badly generated images, and all of the other things that can potentially poison future models will be filtered out by humans.

Humans already unable to tell the difference between generated and real. You get compounding of training on ever more fake data. You don’t see this as problematic?

I don't see it problematic for future AI training, not at all. And for us as people? You are talking about data as real and fake. What does it even mean? There was so much unreal data before AI ever went mainstream. For example in architecture randers are used for decades.

I really don't see a problem in the source of the data. If it is good, inspirational, sane, realistic, creative, pretty, etc. enough I'm going to use it and I really don't see a reason why I shouldn't. Just because it was generated by AI?

Fake = generated. Even architectural renders are based on real objects. AI generated fakery can do all sorts of unrealistic things that have no parallels to what exists. I already see issues with various searches and seeing results that are unlikely to be real. You get a bunch of this idiotic data into your model and get idiotic output. Not limited to images either. Written word will be even worse. Trying to identify what’s factual and what’s hallucinated must be a nightmare job.

even architect plans change when applied to actual build on ground/they just can't see all builder