boatsnhos931 ,

[Thread, post or comment was deleted by the moderator]

  • Loading...
  • uriel238 ,
    @uriel238@lemmy.blahaj.zone avatar

    FOSTA is still in effect and still causing harm to sex workers while actually protecting human traffickers from investigation (leaving victims stuck as captive labor / sex slaves for longer). And it'd still regarded by our federal legislators as a win, since they don't know any better and can still spin it as a win.

    I don't believe our legislators can actually write a bill that won't be used by the federal Department of Justice merely to funnel kids for the sake of filling prison cells with warm bodies.

    We've already seen DoJ's unnuanced approach to teen sexting which convicts teens engaging in normal romantic intercourse as professional producers of CSAM.

    Its just more fuel for the US prison industrial complex. It is going to heavily affect impoverished kids caught in the crossfire while kids in richer families will get the Brock Turner treatment.

    This bill is wholly for political points and has nothing to do with serving the public or addressing disruption due to new technology.

    Until we reform or even abolish the law enforcement state, anything we criminalize will be repurposed to target poor and minorities and lock them up in unconscionable conditions.

    Blaidd ,
    @Blaidd@lemm.ee avatar

    Creating fake child porn of real people using things like Photoshop is already illegal in the US, I don't see why new laws are required?

    General_Effort ,

    This is not at all about protecting children. That's just manipulation. In truth, kids are more likely to prosecuted than protected by this bill.

    There are already laws that could be used against teen bullies but it's rarely done. (IMHO it would create more harm than good, anyway.)

    This is part of an effort to turn the likenesses of people into intellectual property. Basically, it is about more money for the rich and famous.

    This bill would even apply to anyone who shares a movie with a sex scene in it. It's enough that the "depiction" is "realistic" and "created or altered using digital manipulation". Pretty much any photo nowadays, and certainly any movie, can be said to "altered using digital manipulation". There's no mention of age, deception, AI, or anything that the PR bullshit suggests.

    Coskii ,
    @Coskii@lemmy.blahaj.zone avatar

    There is no world where a law aimed at this type of thing will ever be used for its intended purpose.

    hydration9806 ,

    I feel we are in need of a societal shift here, just like another commenter said about the printing press. When that first came out, the pushback was from the worry that the words would be attributed to someone who never said them (reverse plaigerism). The societal adjustment to this was the universal doubt that anyone said that thing without proof.

    For generative AI, when it becomes widespread, photos will be generateable for literally everyone, not just minors but every person with photos online. It will be a societal shift; images will be assumed to be AI generated, making any guilt or shame about a nude photo existing obselete.

    Just a matter of time so may as well start now!

    wetferret ,
    @wetferret@lemmy.world avatar

    I agree that this is probably the inevitable end result of the proliferation of the technology. The journey society is going to have to take to get to that point is going to be pretty uncomfortable though I think.

    Meowoem ,

    Think of the children being used to push an agenda that helps the very wealthy? Well I'll be, what a totally new and not at all predictable move.

    Ban all ai that aren't owned by rich people, make open source impossible, restrict everything that might allow regular people to compete with the corporations - only then will you children be safe!

    joyjoy ,

    You kind of have to be rich in order to run these image generation AIs. The RTX 4090 TI isn't cheap.

    JackGreenEarth ,
    @JackGreenEarth@lemm.ee avatar

    The NVIDIA 1660 ti is perfectly adequate and not only the property of rich people.

    PatFussy ,

    Links to deep fakes? For science ofcourse

    GBU_28 ,

    Victim is underage. Why don't you take a seat over there?

    PatFussy ,

    Would it help if I wrote /s like some ledditor m'sir?

    dong ,

    🤮 bruh she's 14

    PatFussy ,

    She's 18 somewhere am I right chat

    dong ,

    No. You are not funny in the slightest.

    It would be weird and creepy if this was an adult woman, but your "joke" about a situation involving children is just disgusting.
    You're sick, please seek help

    PatFussy ,

    "I'm sure you are fun at parties" - Le heckin funnerino

    dong ,

    At least I'm going to parties with people my own age and not creeping on literal high schoolers.

    Jesus fuck dude stop acting like a pedophile

    PatFussy ,
    DR_Hero ,

    Excuse me but, the fuck is wrong with you?

    PatFussy ,

    I'm just living my best life

    PatFussy ,

    This comment got rule 3'd for a joke.

    ShellMonkey ,
    @ShellMonkey@lemmy.socdojo.com avatar

    Just wait until them tech savvy folks in Congress try to understand the difference between 'deepfakes' in the sense of pasting a new face on existing footage and whole cloth generative AI creating the entire scene, and then someone tells them that the latter is derived from multiple existing media sources. Gonna be some smoke pouring out of their ears like in the cartoons trying to slice up all the specifics.

    trackcharlie ,

    There are already laws against creating false content about people, so adding more laws isn't going to make the previous laws more or less valid and it's only going to waste time and money.

    Of course it's being pushed by a "teen" since this teen clearly doesn't have any understanding of the issues at hand, the technology at hand nor the laws that already exist to help them with this issue.

    It was up to the adults around this teen to help her navigate the issue and instead the incompetent pieces of worthless shit choose to push a new bill against AI rather than use the current legal framework that exists to actually help this girl.

    Anything to abuse a child or teens situation for their political gain. Worthless trash.

    Laticauda ,

    It's being pushed by someone who was a victim of deep fake ai porn, so I think they understand the issues at hand just fine, you don't have to agree with her, but don't be a patronizing asshole about it.

    trackcharlie ,

    Thanks for not actually reading my comment and making it clear to everyone who did that you're either illiterate or a dishonest asshole.

    Laticauda ,

    I read your comment just fine, and the way you spoke about the teen in question was incredibly patronizing.

    trackcharlie ,

    Then you should work on your reading comprehension.

    GBU_28 ,

    Disagreeing strongly isn't patronizing.

    LWD , (edited )

    [Thread, post or comment was deleted by the author]

  • Loading...
  • GBU_28 ,

    This is an actual law proceeding, with lawyers and adults involved. The teen is just the face of it.

    LWD , (edited )

    [Thread, post or comment was deleted by the author]

  • Loading...
  • trackcharlie ,
    1. Pretending abortion is in the same realm as AI tool abuse is ridiculous and completely dishonest.

    .

    1. Scare quotes? Are you for real? Since you don't understand, quotes denote emphasis or specificity, not emotion. "Teen", the emphasis here, was used because teens are barely educated in legal matters and it's their responsibility to seek assistance, not push legislation.

    .

    1. If the adults in that area had actually responded properly, there wouldn't be an article about a new bill against AI, instead there'd be an article about defamation and debasement of a minor (likely by another minor, but that doesn't mean the other minors parents are infallible in this situation, you are either parents or you're not, if you are, you're responsible for your brats actions, if you're not, then the state will take the child and destroy them mentally, and likely physically, for your failures as a parent).

    .

    1. Not a single thing I said dismissed this, and I would go even further to say anyone pursuing the AI angle is the one dismissing this, especially given the laws that already exist that can be used to assist and protect the teen in question.

    .

    1. We've had adobe premier for more than 2 decades before generative AI and all of the issues surrounding deepfakes then cover every issue with generative AI now and wasting peoples time and money on this does not help improve the situation, instead it makes a mockery of the victim for the purposes of pushing an agenda.
    nybble41 ,

    Since you don't understand, quotes denote emphasis or specificity, not emotion.

    Actually quotes denote quotations. When used casually around an individual word or short phrase they generally indicate that the writer is emphasizing that these are someone else's words, and that the writer would have chosen a different description. As in: These people are described as "teens" but are probably not only/mostly teenagers. That may not be what you meant, but it's how that text will be read.

    If you just want emphasis you might consider using bold or italics rather than quotes.

    Infiltrated_ad8271 ,
    @Infiltrated_ad8271@kbin.social avatar

    If (as it seems) the point is not impersonation but damage to the person's honor/image, where exactly is the line?

    If realism is the determining factor, what about a hyperrealistic human work? And if it is under human interpretation how realistic it should be, could a sketch be included?

    guyrocket ,
    @guyrocket@kbin.social avatar

    How different is photoshopped fakes from AI fakes? Are we going to try to bad that too?

    ETA: *ban that too. Thx phone kb.

    OsrsNeedsF2P ,

    More laws that will only affect people who follow laws

    henfredemars ,

    It's like how DRM only hurts people who purchase content legally.

    It's been very illegal to pirate games for decades, and still pirated content is quite common in the wild. What's banning it (defamation) harder going to practically achieve?

    Phil_in_here ,

    Isn't that... all laws?

    henfredemars ,

    This is just the tip of the iceberg of the threat AI poses to human social structures. We have yet to appreciate the gravity of what these new technologies enable. It's incredibly dangerous yet equally naive to think that AI-generated porn laws will keep us safe.

    Firstly, the cat's out of the bag. We can ban the technology or its misuse all we like, but can we really practically stop people from computing mathematical functions? Legal or not, generative AI can and will be used to generate content that hurts people. We need better planning for identifying, authenticating, and responding when this misuse happens.

    Secondly, we have an already huge, huge problem with fake news and disinformation. What is such a law for this special case of AI porn going to do for our inability to address harmful content?

    It's a shame, but it strikes me as more feel-good than actually doing something effective.

    ZILtoid1991 ,
    @ZILtoid1991@kbin.social avatar

    Just ban any instance of generative unconsentual generative nudity, AI or not.

    henfredemars ,

    I just drew a stick figure of Johnny Depp and it's indeed naked. And I'd do it again, too.

    fine_sandy_bottom ,

    I really wonder whether this is the right move.

    This girl, and many others, are victims and I don't want to diminish that, but I for better or worse I just don't see how legislation can resolve this.

    Surely deepfakes will be just different enough to the subject to create reasonable doubt that it depicts the subject.

    I wonder whether, as deep fakes become commonplace, people might be more willing to just ignore it like any other form of trolling.

    galoisghost ,
    @galoisghost@aussie.zone avatar

    It’s not trolling it’s bullying. You need to think beyond this being about “porn”. This is a reputational attack that makes the victim more likely to be further victimised via date rape, stalking, murder. These things already happen based on rumours, deepfakes images/videos will only make it worse. The other problem is that it’s almost impossible to erase once it’s on the internet, so the victim will likely never be free of the trauma or danger as the images/videos resurface.

    fine_sandy_bottom ,

    Trolling / bullying is just semantics, which I don't think will help us very much.

    I think the heightened risk of other crimes is... dubious. Is that conjecture?

    Your position seems to be framed in the reality of several years ago, where if you saw a compromising video of someone it was likely real, while in 2024 the opposite is true.

    Were headed towards a reality where someone can say "assistant, show me a deepfake of a fictitious person who looks a bit like that waitress at the Cafe getting double teamed by two black guys". I don't claim to know all the ethical considerations, but I do think that changing social norms are part of the picture.

    I don't have any authority to assert when anyone else should feel victimised. All I know is that in my own personal case, a few years ago I would've felt absolutely humiliated if someone saw a compromising video of me, but with the advent of deep fakes I just wouldn't care very much. If someone claimed to have seen it I would ask them why they were watching it, and why in the world they would want to tell me about their proclivities.

    Overzeetop ,
    @Overzeetop@kbin.social avatar

    I think it doesn't go far enough. Straight up, no one should be permitted to create or transmit the likeness of anyone [prior to, say, 20 years following their death] without their explicit, written permission. Make the fine $1,000,000 or 10% of the offender's net worth, whichever is greater; same penalty and corporate revocation for any corporation involved. Everyone involved from the prompt writer to the work-for-hire people should be liable for the full penalty. I can't think of a valid, non-entertainment (parody/humor), reason for non-consensual impersonation - and using it for humor or parody is a slippery slope to propaganda weaponization. There is no baby in this tub of bathwater.

    fine_sandy_bottom ,

    I'm not sure this is practically possible.

    A $1m penalty is more or less instant bankruptcy for 99% of the population. It's probably not much of a deterrent for, say an 18 year old. In my jurisdiction I don't think there are criminal penalties higher than a few thousand dollaridoos. It doesn't matter whether you think this act is so aggregious that it deserves a penalty 1000 time higher than any other, my point is that it would be unenforceable ineffective.

    Secondly, how do you determine whether an image is someone's likeness? Create any random image and surely it will look like someone, but that doesn't mean that creating that image violates that someone.

    TimeSquirrel ,
    @TimeSquirrel@kbin.social avatar

    Yeah, just like the FBI warnings on VHS tapes about massive fines and jail time stopped us from copying them in the 80s and 90s...

    flipht ,

    I think you're right if the goal is to stop them all together.

    But what we can do is stop people from sending them around and saying that it's true/actually the person.

    Once they've turned it from a art project into a weapon, it should have similar consequences to "revenge porn."

    HubertManne ,
    @HubertManne@kbin.social avatar

    I would think this would be covered by libel, slander, defamation type laws. The crime is basically lying about a persons actions and character.

    AlteredStateBlob ,
    @AlteredStateBlob@kbin.social avatar

    My dude there are people out there thinking they're in a relationship with Johnny fucking Depp because some Nigerian scammer sent them five badly photoshopped pictures. Step out of your bubble, maybe. This shit isn't easy to spot for the vaaaaaast majority of people and why would this lie with the victim to sort of clear their name or hope that idiots realize it's fake?

    Especially with and around teenagers who can barely think further than their next meal?

    Good lord.

    fine_sandy_bottom ,

    WDYM "step out of your bubble"?

    It's not a question of being able to detect whether or not a video is fake. When deepfakes become so prevalent that everyone's grandma understands that they're prevalent, it won't matter whether you can identify the video as fake.

  • All
  • Subscribed
  • Moderated
  • Favorites
  • random
  • technology@lemmy.world
  • test
  • worldmews
  • mews
  • All magazines