If AI and deep fakes can listen to a video or audio of a person and then are able to successfully reproduce such person, what does this entail for trials?

It used to be that recording audio or video would give strong information which often would weigh more than witnesses, but soon enough perfect forgery could enter the courtroom just as it’s doing in social media (where you’re not sworn to tell the truth, though the consequences are real)

I know fake information is a problem everywhere, but I started wondering what will happen when it creeps in testimonies.

How will we defend ourselves, while still using real videos or audios as proof? Or are we just doomed?

  • SirEDCaLot@lemmy.today
    link
    fedilink
    arrow-up
    1
    ·
    edit-2
    33 minutes ago

    Eventually, we will just have to accept that photographic proof is no longer proof.

    There are ways that you could guarantee an image is valid. You would need a hardware security module inside the camera, which signs a hash of the picture with its own built-in security key that can’t be extracted and a serial number that it generates. That can prove that an image came from a particular camera, and if you change even one pixel of that image the signature won’t match anymore. I don’t see this happening anytime soon. Not mainstream at least. There are one or two camera manufacturers that offer this as a feature, but it’s not on things like surveillance cameras or cell phones nor will it be anytime soon.

  • Call me Lenny/Leni@lemm.ee
    link
    fedilink
    English
    arrow-up
    1
    ·
    59 minutes ago

    A camera can only show us what it sees. It doesn’t objectively necessitate a viewer’s interpretation of it. I remember some of us being called down to the principal’s office (before the age of footage-based scandals, which if anything imply shortcoming in the people progressing the rulings to be in so much awe at, sadly a common occurrence, adding to the “normal people distaste” I have, and something authorities have made sure I’m no stranger to) who may say “we saw you on the camera doing something against the rules” only to be responded to with “that’s not me, I have an alibi” or “that’s not me, I wouldn’t wear that jacket” or “that’s not me, I can’t do that person’s accent” (aforementioned serial slander of me serving as a prime example where this would be the case). In connection to the process, you might say it’s witness testimony from a machine and that they’ve “just started” to get into the habit of not being very honest to the humans in thw court. I remember my first lie.

  • LesserAbe@lemmy.world
    link
    fedilink
    arrow-up
    21
    ·
    7 hours ago

    I think other answers here are more essential - chain of custody, corroborating evidence, etc.

    That said, Leica has released a camera that digitally signs its images, and other manufacturers are working on similar things. That will allow people to verify whether the image is original or has been edited. From what I understand Leica has some scheme where you can sign images when you update them too, so there’s a whole chain of documentation. Here’s a brief article

    • andrew_bidlaw@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      6
      ·
      7 hours ago

      It’s an interesting experiment, but why would we trust everything that Leica supposedly verified? The same shit with digital signatures and blockchain stuff. We are at the gates of the world where we have zero trust by default and would only intentionally outsource verification to third parties we trust, because penalties for mistakes are growing each day.

      • BrianTheeBiscuiteer@lemmy.world
        link
        fedilink
        arrow-up
        4
        ·
        6 hours ago

        I don’t think we should inherently. I’ve thought about the idea of digitally signed photos and it seems sound unless someone is quite clever with electronics. I’m guessing there’s some embedded key on the camera that is hard but maybe not impossible to access. If people can hack Teslas for “full autopilot” or run Doom on an ATM machine I’m not confident that this kind of encryption will never be cracked. However, I would hope an expert witness would also examine the camera that supposedly took the picture. I would think it to be impossible for someone to acquire the key without a 3rd party detecting the intrusion.

        • Passerby6497@lemmy.world
          link
          fedilink
          English
          arrow-up
          1
          ·
          edit-2
          4 hours ago

          They make Hardware Security Modules (HSMs) that are very difficult to crack, to the point that it is unbreakable at our current technology level. With a strong HSM, a high-bit per-device certificate signed by the company’s private key gives you authenticity and validation until the root key or HSM are broken, which is probably good enough for today while we try to figure out something better IMO.

        • andrew_bidlaw@sh.itjust.works
          link
          fedilink
          English
          arrow-up
          2
          ·
          6 hours ago

          Today we have EXIFs and it’s better to wipe them all of these for privacy reasons. Because every picture you take otherwise contains a lot of your data like geoloc, model, exposuer, etc. That’s the angle they are yet to tackle - because most of these things are also leave us vulnerable.

  • lemmefixdat4u@lemmy.world
    link
    fedilink
    arrow-up
    20
    arrow-down
    1
    ·
    10 hours ago

    When video or audio evidence is submitted, it will be questioned as to its authenticity. Who recorded it? On what device? Then we’ll look for other corroborating evidence. Are there other videos that captured the events in the background of the evidence video? Are there witnesses? Is there contradictory evidence?

    Say there’s a video depicting a person committing murder in an alley. The defense will look for video from the adjoining streets that show the presence or absence of the murderer before or after. If those videos show cars driving by with headlights on, they will look for corresponding changes in the luminosity of the crime video. If the crime happened in the daytime, they will check that the shadows correspond to Sun’s position at that moment. They’ll see if the reflections of objects match the scene. They’ll look for evidence that the murderer was not at the scene. Perhaps a neighbor’s surveillance camera shows they were at home or their cell phone indicated they were someplace else.

    But if all these things indicate the suspect was in the alley and the video is legitimate, that’s powerful evidence toward a conviction.

    • BrianTheeBiscuiteer@lemmy.world
      link
      fedilink
      arrow-up
      3
      ·
      6 hours ago

      Are there other videos that captured the events in the background of the evidence video?

      I think this is key in a trial setting. A published picture might be unique but to think the photographer snapped just one picture while nobody else was present or also photographing is a bit of a stretch.

  • logos@sh.itjust.works
    link
    fedilink
    arrow-up
    79
    ·
    12 hours ago

    Fake evidence, e.g. forged documents, are not not new things. They take things like origin, chain of custody etc into account.

    • ColeSloth@discuss.tchncs.de
      link
      fedilink
      arrow-up
      2
      arrow-down
      2
      ·
      5 hours ago

      Sure, but if you meet up with someone and they later have an audio recording that is completely fabricated from the real audio, there’s nothing for chain of anything. Audio used to be damning evidence and was fairly easily discoverable if it was hacked together to try to sound different. If that goes away, then it just becomes useless as evidence.

      • hypna@lemmy.world
        link
        fedilink
        arrow-up
        3
        arrow-down
        1
        ·
        4 hours ago

        It becomes useless as evidence unless you can establish authenticity. It just makes audio recordings more in a class with text documents; perfectly fakeable, but admissible with the right supporting information. So I agree it’s a change, but it’s not the end of audio evidence, and it’s a change in a direction which courts already have experience.

  • tlou3please@lemmy.world
    link
    fedilink
    arrow-up
    55
    ·
    12 hours ago

    As someone who works in the field of criminal law (in Europe, and I would be shocked if it wasn’t the same in the US) - I’m not actually very worried about this. By that I don’t mean to say it’s not a problem, though.

    The risk of evidence being tampered with or outright falsified is something that already exists, and we know how to deal with it. What AI will do is lower the barrier for technical knowledge needed to do it, making the practice more common.

    While it’s pretty easy for most AI images to be spotted by anyone with some familiarity with them, they’re only going to get better and I don’t imagine it will take very long before they’re so good the average person can’t tell.

    In my opinion this will be dealt with via two mechanisms:

    • Automated analysis of all digital evidence for signatures of AI as a standard practice. Whoever can be the first person to land contracts with police departments to provide bespoke software for quick forensic AI detection is going to make a lot of money.

    • A growth in demand for digital forensics experts who can provide evidence on whether something is AI generated. I wouldn’t expect them to be consulted on all cases with digital evidence, but for it to become standard practice where the defence raises a challenge about a specific piece of evidence during trial.

    Other than that, I don’t think the current state of affairs when it comes to doctored evidence will particularly change. As I say, it’s not a new phenomenon, so countries already have the legal and procedural framework in place to deal with it. It just needs to be adjusted where needed to accommodate AI.

    What concerns me much more than the issue you raise is the emergence of activities which are uniquely AI dependent and need legislating for. For example, how does AI generated porn of real people fit into existing legislation on sex offences? Should it be an offence? Should it be treated differently to drawing porn of someone by hand? Would this include manually created digital images without the use of AI? If it’s not decided to be illegal generally, what about when it depicts a child? Is it the generation of the image that should be regulated, or the distribution? That’s just one example. What about AI enabled fraud? That’s a whole can of worms in itself, legally speaking. These are questions that in my opinion are beyond the remit of the courts and will require direction from central governments and fresh, tailor made legislation to deal with.

  • j4k3@lemmy.world
    link
    fedilink
    English
    arrow-up
    21
    arrow-down
    3
    ·
    12 hours ago

    Science has proven that the entire model of human memory as factual testimony is a fallacy. That came out long before AI in the public space. I don’t think anyone has addressed that revelation. I doubt anyone will address this one. Hell, there are still people sketching the courtroom like cameras don’t exist. A president can stage a failed coup and a SC judge can fly the traitor’s flag and there are no consequences for either.

    So what will be done, absolutely nothing, unless some billionaires stage a proxy war over it.

  • Mio@feddit.nu
    link
    fedilink
    arrow-up
    11
    arrow-down
    1
    ·
    12 hours ago

    Fake does not change what actually happened. Just look for facts in the real world that support the theory. Remember, photoshop existed before AI. We have DNA checks today.

    • fart_pickle@lemmy.world
      link
      fedilink
      arrow-up
      1
      arrow-down
      6
      ·
      10 hours ago

      Fake does not change what actually happened

      Seriously? You want to charge people just because the fake photo/video could have incriminate them? Welcome to the minority report…

  • JaggedRobotPubes@lemmy.world
    link
    fedilink
    English
    arrow-up
    3
    ·
    12 hours ago

    It’s a scary question, made a lot less scary by whoever it was that said “you know, I guess we’ve had text deepfakes a long time”

    Eventually people just know it could be fake, so they look for other ways of verifying. The inevitability and the scale of it mean that, at the very least, we’ll have all our brainpower on it eventually.

    It’s the meantime where shit could get wild.

  • AbouBenAdhem@lemmy.world
    link
    fedilink
    English
    arrow-up
    6
    arrow-down
    3
    ·
    12 hours ago

    Maybe each camera could have a unique private key that it could use to watermark keyframes with a hash of the frames themselves.

    • SavvyWolf
      link
      fedilink
      English
      arrow-up
      5
      ·
      11 hours ago

      How would you prove that the camera itself is real, is the only device with access to the private key and isn’t falsifying it’s video feed?

      • AbouBenAdhem@lemmy.world
        link
        fedilink
        English
        arrow-up
        2
        arrow-down
        1
        ·
        10 hours ago

        The sort of case I was thinking of is if different parties present different versions of an image or video and you want to establish which version is altered and which is original.

        • SavvyWolf
          link
          fedilink
          English
          arrow-up
          1
          ·
          8 hours ago

          You still have the same problem though. You can produce a camera in court and reject one of the images, but you still need to prove that the camera wasn’t tampered with and it was the one at the scene of the crime.

            • SavvyWolf
              link
              fedilink
              English
              arrow-up
              1
              ·
              6 hours ago

              The camera can sign things however it wishes, but that doesn’t automatically make the camera trustworthy.

              In the same sense, I can sign any number of documents claiming to have seen a crime take place but that doesn’t make it sufficient evidence.

    • OsrsNeedsF2P@lemmy.ml
      link
      fedilink
      arrow-up
      5
      arrow-down
      1
      ·
      12 hours ago

      Usually I see non-technical people throw ideas like this and they’re stupid, but I’ve been thinking about this for a few minutes and it’s actually kinda smart

    • MoonManKipper@lemmy.world
      link
      fedilink
      English
      arrow-up
      4
      arrow-down
      1
      ·
      12 hours ago

      I think that’s exactly how it’s going to work - you can’t force all ‘fake’ sources to have signatures- it’s too easy to make one without one for malicious reasons. Instead you have to create trusted sources of real images. Much easier and more secure

  • balderdash@lemmy.zip
    link
    fedilink
    arrow-up
    2
    ·
    13 hours ago

    I’m not a tech person, so I’ll take the lowest hanging fruit. The obvious answer is to write a program that can detect AI. Then there will be a competition between AI fakes and AI detection. This is similar to what we have in sports. There are forbidden enhancement procedures (e.g., steroids, blood doping, etc.) that have to keep improving in subtlety so not to be detected by Anti cheating measures.

  • GrymEdm@lemmy.world
    link
    fedilink
    arrow-up
    2
    arrow-down
    1
    ·
    edit-2
    11 hours ago

    Disclaimer: I’m not an expert, just an interested amateur wanting to chat and drawing comparisons from past leaps in tech and other conversations/videos.

    For a time expert analysis will probably work. For instance, the “click here to prove your not a robot” boxes can definitely be clicked by robots, but for now the robot moves in detectably different ways. My guess is that, for at least a while, AI content will be different from actual video in ways like code. There will probably be an arms race of sorts between AI and methods to detect AI.

    Other forms of evidence like DNA, eyewitness accounts, cell phone tracking etc. will likely help mitigate deceitful AI somewhat. My guess is that soon video/audio will no longer be considered as ironclad as it was even a few years ago. Especially if it comes from an unverified source.

    There are discussions about making AI tools have a digital “watermark” than can be used to identify AI-generated content. Of course this won’t help with black market-type programs, but it will keep most people out of the “deep fake for trials” game.

    When it comes to misinformation on social media though, well…it’s probably going to get crazy. The last decade or so has been a race at an unprecedented scale to try and keep up with BS “proof”, psuedoscience, etc. Sadly those on the side of truth haven’t always won. The only answer I have for that is making sure people are educated about how to deal with misinformation and deepfakes - eg. awareness they exist, identifying reputable sources and expert consensus, and so on.

  • Zos_Kia@lemmynsfw.com
    link
    fedilink
    arrow-up
    1
    arrow-down
    1
    ·
    11 hours ago

    I doubt these tools will ever get to a level of quality that can confuse a court. They’ll get better, sure, but they’ll never really get there.

  • webghost0101@sopuli.xyz
    link
    fedilink
    arrow-up
    3
    arrow-down
    5
    ·
    edit-2
    12 hours ago

    For the longest time now, from before AI, before NFT was a thing i had an idea to incorporating blockchain tech into real life media footage to combat the rise of misinformation.

    The metadata, original author would be stored on this chain the moment footage is recorded. The biggest challenge is that this means the devices themselves need to be connected.

    Adoption would be slow but i imagined news and official channels make use of this tech first. Eventually all footage outside of this will be seen as not trustworthy

    Then NFTBros came along and people have shit on this idea ever since. Some days i feel that was a conspiracy to ruim out perception of potential but more likely humans where just greedy.

    I still believe this could work. Detailed example below:

    The system works with a fair amount of transparency, verifiable digital signatures for recording devices and their owners. Professional cameras and organizations would have publicly known IDs, while individuals could choose to remain pseudonymous authors but would need to build credibility over time.

    Let’s say BBC records an interview. When viewers watch this content on any platform, they can access blockchain verification through an embedded interface (perhaps a small icon in the corner). This shows the complete chain of custody from recording to broadcast.

    The system verifies content through computational comparisons. When a raw interview is edited into a final piece:

    • Each original clip has a unique blockchain signature
    • The final edited version’s signature can be compared against source material
    • Automated analysis shows what percentage of original footage matches
    • Modifications like color correction or audio adjustments are detected through signature differences
    • Additional elements like station logos or intro sequences have their own verified identifiers
    • conciselyverbose@sh.itjust.works
      link
      fedilink
      arrow-up
      4
      ·
      10 hours ago

      Because it’s insanely idiotic. Signing videos is one thing.

      Hooking it into blockchain bullshit is entirely deranged. It adds a bunch of complexity to provide literally zero benefit in any possible context.

      • webghost0101@sopuli.xyz
        link
        fedilink
        arrow-up
        2
        arrow-down
        1
        ·
        edit-2
        7 hours ago

        I am not sure what you think blockchain actually is but in essence its a decentralized ledger of signatures.

        Not coins, no sellable goods. Just that. Computers connected in a network to verify the correctness of a cloud ledger.

        So if you say signing footage is one thing how do you propose a laymen can verify that signature without centralized databank.

        I understand some people may not mind centralized authority but i prefer against it.

        I am willing to hear peoples thoughts on this. I am not pro or against blockchain or any form of technology. With the information i have this just seems like a reasonable and practical solution.

        • conciselyverbose@sh.itjust.works
          link
          fedilink
          arrow-up
          2
          ·
          7 hours ago

          I am well aware of what it is. It serves no purpose and provides no benefit.

          Ignoring the fact that hardware signing doesn’t validate inputs as “real”, because it’s entirely possible to replicate the actual signals entering the camera, and the fact that the entire premise by definition would be a terrible power grab by big hardware/software tools, the very obvious way to implement such an approach would be the exact same system as certificate authorities. You have to have actual root certificate signers.

          Blockchain is horseshit and serves no purpose.

          • webghost0101@sopuli.xyz
            link
            fedilink
            arrow-up
            1
            arrow-down
            1
            ·
            7 hours ago

            That hardware inputs can be faked is part of my reasoning here because there would be transparency of the source of footage.

            If a reputable journalists fake their own footage and it would be found out their credibility would be gone.

            If they often rely on borrowing footage and don’t fact check it. Credibility will degrade as well.

            Journalist media that does their work and only uses credible sources will thrive.

            My solution isn’t about who or how signature gets created but how ordinary people can check for themselves where a clip within footage originates from.

            I am fine with inventing a new system that does this and call it something else than blockchain. But my understanding is that it does pretty much provide this functionality in a robust manner.

            Also typing these comments on the go caused me to lose something dear to me on public transport. I am very sad now and probably wont engage further.

            • conciselyverbose@sh.itjust.works
              link
              fedilink
              arrow-up
              2
              ·
              edit-2
              6 hours ago

              Again, you have to completely ignore that the core premise is evil intended to give big players even stronger monopoly control. It’s anti-free in every sense, and as an added bonus, would very certainly make possession of specific hardware sufficient to be executed in some countries, because everything it has ever captured would be tracked to it.

              But if you do that, there is already a system that does exactly what you’re asking. You don’t need to invent anything. It’s certificate authorities.

              I’m not actually trying to be an asshole, though I’m sure I’m coming off as one. But the only thing blockchain actually does is validate transactions. It’s a shared ledger.

              • webghost0101@sopuli.xyz
                link
                fedilink
                arrow-up
                1
                arrow-down
                1
                ·
                6 hours ago

                Sure i’ll have a look at decentralized certificate authorities options.

                Very possibles to adapt my idea to whatever technology provides those function honestly.

                The only actual connection i have with blockchain is that reading about it when it was new directly inspired in me a possible way to combat fake news.