• barsoap@lemm.ee
    link
    fedilink
    arrow-up
    5
    ·
    1 year ago

    Images, yes, but mixing concepts is a mixed bag. Just because the model can draw, say, human faces and dog faces doesn’t mean it has the understanding necessary to blend those concepts. Without employing specialised models (and yes of course the furries have been busy) the best you’ll get is facepaint. The pope at a beach bar doesn’t even come close to exercising that kind of capability: The pope is still the pope and the beach bar is still the beach bar, and a person is still sitting there slurping a caipirinha.

    • Amju Wolf@pawb.social
      link
      fedilink
      English
      arrow-up
      11
      ·
      1 year ago

      I mean if you train a model on porn with adult actors and on regular photos with children, it shouldn’t be hard to generate the combination.

      You probably wouldn’t even need any fancy training data but if you really wanted you could pick adult actors that look young or in other ways similar to the children to help the process.

      • barsoap@lemm.ee
        link
        fedilink
        arrow-up
        3
        ·
        1 year ago

        Knowing what a nude adult looks like doesn’t mean that the model knows what a nude child looks like. I’m quite sure it’s easy to generate disturbing images like that, but actual paedophiles I think won’t be satisfied with child faces on small adult bodies.

        Ordinary deepfakes actually have a very similar problem: Sure you can take a picture of a celebrity and tell the AI to undress them – but it won’t be their actual body. The AI is going to be able to approximate their overall build but it’s going to be a generic adult body, not the celebrity’s body. Or, differently put, AI models aren’t any better at undressing people with their eyes than teenagers.

        • Amju Wolf@pawb.social
          link
          fedilink
          arrow-up
          3
          ·
          1 year ago

          I see where you’re coming from but that’s a technical issue that will probably be solved in time.

          It’s also really not a black and white; sure maybe you can see it isn’t perfect but you’d still prefer it to content where you know no one was actually harmed.

          Despite what reputation people like that have (due to the simple fact of how reporting works), most are harmless like me and you and don’t actually want to see innocent people suffer and would never act on their desires. So having a safe and harmless outlet might help.

          • barsoap@lemm.ee
            link
            fedilink
            arrow-up
            4
            ·
            edit-2
            1 year ago

            I see where you’re coming from but that’s a technical issue that will probably be solved in time.

            You cannot create information from nothing.

            So having a safe and harmless outlet might help.

            Psychologists/Psychiatrists are still on the fence on that one, I wouldn’t be surprised if it depends on the person. And yes the external harm produced by AI images is definitely lower than that produced from actual CSAM, doubly so newly produced CSAM, but that doesn’t mean that therapy, even in its current early stages, couldn’t do even better.

            Differently put: We may be again falling into the trap of trying to find technological solutions to societal problems (well, this is /c/technology…). Which isn’t to say that we shouldn’t care at all about models trained on CSAM, but that’s addressing symptoms, not causes. Ultimately addressing root causes is more important: The vast majority of paedophiles are not exclusive paedophiles, often they’re not even really attracted to kids at all beyond having developed a fetish, they’re rapists focussing on the most vulnerable, often due to having been victims of sexual abuse themselves.

      • barsoap@lemm.ee
        link
        fedilink
        arrow-up
        1
        ·
        1 year ago

        That’s not concept mixing, also, it’s not proper origami (paper doesn’t fold like that). The AI knows “realistic swan” and “origami swan”, meaning it has a gradient from “realistic” to “origami”, crucially: Not changing the subject, only the style. It also knows “realistic human”, now follow the gradient down to “origami human” and there you are. It’s the same capability that lets it draw a realistic mickey mouse.

        It having understanding of two different subjects, say, “swan” and “human”, however, doesn’t mean that it has a gradient between the two, much less a usable one. It might be able to match up the legs and blend that a bit because the anatomy somehow matches, and well a beak is a protrusion and it might try to match it with the nose. Wings and arms? Well it has probably seen pictures of angels, and now we’re nowhere close to a proper chimera. There’s a model specialised on chimeras (gods is that ponycat cute) but when you flick through the examples you’ll see that it’s quite limited if you don’t happen to get lucky: You often get properties of both chimera ingredients but they’re not connected in any reasonable way. Which is different from the behaviour of base sdxl, which is way more prone to bail out and put the ingredients next to each other. If you want it to blend things reliably you’ll have to train a specialised model using appropriate input data, like e.g. this one.