fhd2 3 days ago

Sounds like reasonable changes.

Generally speaking, I think evidence tampering is not a new problem, and even though it's easy in some cases, I don't think it's _that_ widespread. Just like it's possible to lie on the stand, but people usually think twice before they do it, because _if_ they are found to have lied, they're in trouble.

My main concern is rather that legit evidence can now easily be called into question. That seems to me like a much higher risk than fake evidence, considering the overall dynamics.

But ultimately: Humanity has coped without photo, audio or video evidence for most of its existence. I suppose it will cope again.

  • SoftTalker 3 days ago

    Have you ever heard some of the wiretap or hidden microphone recordings used to convict mafia bosses back in the 1980s? It was so bad I can't believe it was accepted. It could easily have been faked. The only thing that made it work was the sworn statements of authenticity from the people who did the recording, and the chain of custody thereafter.

    • qingcharles 3 days ago

      I know a lawyer who was convicted on one of those. The detectives had the mic + transmitter taped to his groin so it would get through a pat down. Then the undercover just spoke both sides of the conversation (the defendant's side was whispered and muffled). Someone testified it sounded like the defendant. Conviction was overturned on appeal on a simpler issue unrelated to the evidence.

      In fact, I remember a drug dealer I was helping with his defense. Hidden mic on undercover was taped under his armpit. He arrived to defendant's hotel room for a deal and defendant made him undress. The recording is hilarious because defendant is like "You fucking snitch, what's that under your armpit?" and the undercover says "It's my .. er .. MP3 player?" LOL

      • siva7 2 days ago

        So what happened to him?

        • NoToP 2 days ago

          I refer you to the precedent of Snitches v. Stitches

        • qingcharles 2 days ago

          The first one got a get-out-of-jail-free card. The second one got away because the drug dealer didn't want to do anything to him while he had a mic + transmitter attached. If anything happened to him later, I wasn't privy to that..!

  • mitthrowaway2 3 days ago

    I agree. The article didn't touch on this aspect, but we're now at the point where even authentic recordings could be plausibly denied and claimed to be fake. So the entire usage of recordings as evidence will suffer a hit. We may essentially be knocked back to an 18th century level of reliance on eyewitness testimony. One wonders what the consequences for justice will be.

    • Ukv 3 days ago

      I wouldn't say we'd be quite back to pre-photo evidence days. I feel a lot if not most of the value in a video/audio recording is not just that the medium has traditionally been difficult to edit, but that it's attesting to a lot of details with high specificity. There's a lot to potentially get caught out on with not a lot of wiggle room when inconsistencies are spotted (compared to recalling from memory). Document scans and static images are still useful despite having long been trivial to edit, for instance.

      • dartos 2 days ago

        I think forensic analysis of photographs would reveal pretty quickly if they’re AI anyway.

        AI images of real life still don’t pass intense scrutiny.

        • wongarsu 2 days ago

          And we have been able to edit photos nearly as long as we have been able to make them. Yet photo evidence continues being useful

          • philipov 2 days ago

            I won't deny that genAI is still a long way from passing the strictest scrutiny, but I don't like the above argument. Ease of use is not a trivial matter, and extreme advances in accessibility can cause a phase-change. Photo evidence continues to be useful because editing photos is below the critical threshold for becoming a problem. If it becomes easy, low-skill, and reliable, at that point we will cross into the realm of Reasonable Doubt. It's the difference between possible, and practical, and pervasive.

        • ykonstant 2 days ago

          Right... forensics, the famously reliable discipline.

          • dartos 2 days ago

            You can entirely remove the word “forensics” and my statement would be the same…

        • throwway120385 2 days ago

          It's just another thing you'll have to pay for as the defendant.

    • thereddaikon 3 days ago

      There's already a process for this, its called chain of custody. If you cant prove the evidence has a solid chain of custody then it was potentially tampered with and isn't reliable.

      • AnthonyMouse 3 days ago

        The usual chain of custody goes something like: The store has a video surveillance system which the police collect the footage from, so the chain of custody goes through the store and the police which implies that nobody other than those two have tampered with it.

        But then you have an inside job where the perpetrators work for the store and have doctored the footage before the police come to pick it up, or a corrupt cop who wants to convict someone without proving their case or is accepting bribes to convict the wrong person and now has easy access to forgeries. Chain of custody can't help you in either of these cases, and both of those things definitely happen in real life, so how do you determine when they happen or don't?

        • iamacyborg 3 days ago

          Surely chain of custody applies if the accused has access to the evidence? Perhaps I’m missing your point or I’m overly optimistic about the legal system.

          • AnthonyMouse 3 days ago

            Suppose the store manager is having a dispute with a kid who keeps skateboarding in the parking lot, so the store manager decides to commit insurance fraud by robbing the store herself and then submits forged video of the kid doing it to the police.

            The store manager is in the chain of custody but isn't a suspect, the accused is the kid. The kid doesn't even know who actually committed the crime. How is the kid supposed to prove this?

            • condiment 3 days ago

              In this case, chain of custody needs to extend to the capture device itself, and to any software that exists in the supply chain for the video content.

              There are some experimental specifications that exist to provide attestation as to the authenticity of media. But most of what I’ve seen so far is a “perjury based” approach that just requires a human to say that something is authentic.

              • _DeadFred_ 3 days ago

                Chain of custody isn't real as long as the judiciary gives the government a 'good faith' pass when chain of custody isn't maintained/documentable in court. Go into Lexus Nexus and look up 'good faith' related to 'chain of custody'. Any 'protections' that can be waived away at the judges whim when the standard isn't met by the government are not actually real but pure theater to lend legitimacy to the American judicial system that it doesn't deserve.

              • AnthonyMouse 3 days ago

                > In this case, chain of custody needs to extend to the capture device itself, and to any software that exists in the supply chain for the video content.

                There are two major problems with this.

                First, is all footage from existing surveillance systems going to be thrown out because it doesn't use this technology? Answer: No, because it would be impractical. But then nobody cares to adopt the technology because using it isn't required. How's that IPv6 transition going?

                Second, that sort of thing doesn't actually work anyway. Surveillance cameras are made by the lowest bidder. Their security record is appalling. They're going to publish their private keys on github and expose buffer overflows to the public internet and leave a telnet server running on the camera that gives you a root shell with no password. Does it sound like hyperbole? Those are all things that have actually happened.

                There is only one known way to prevent this from happening: Do not allow the hardware vendor to write the software. Any of the software. Instead, demand hardware documentation so that the firmware can be written by open source software people instead of lowest bidder hardware companies. This is incompatible with using the hardware vendor as the root of trust, which is a natural consequence because the hardware vendors are completely untrustworthy.

                But let's suppose we find some way to do it. We'll pass a law imposing a $100 fine on any company that has a security vulnerability. Then there will never be a security vulnerability again because security vulnerabilities will be illegal; I'm assured this is how laws work. At that point the forger takes the camera and points it a a high resolution playback of the forgery, and the camera records and signs the forgery.

                I kind of wish people would stop suggesting this. It's completely useless but it creates the false impression that it can be solved this way and then people stop trying to find a real solution.

      • mrandish 3 days ago

        Yep, "chain of custody." Came here hoping to see that concept discussed since it's how the system already deals with cases of potential evidence tampering. If the evidence is of material importance and there's no sufficiently credible chain of custody, then its validity can be questioned. The concept started around purely physical evidence but applies to image, audio and video. The good thing about the ubiquity of deepfake memes on social media is that it familiarizes judges and juries with how easy it now is to create plausible fake media.

        • dragonwriter 3 days ago

          Chain of custody only covers from when the evidence came into the hands of the police; the real issue here is original provenance, which chain of custody doesn't solve.

          Evidence of provenance is already important, to be sure, but the the ability to have some degree of validation of the contents has itself provided some evidence of provenance; lose that and there is a real challenge.

      • LPisGood 3 days ago

        This is unironically a usecase for blockchain.

        • adiabatichottub 3 days ago

          Who needs a whole blockchain? Just basic public-key cryptography would do the job.

          Imagine if you will, that the NVR (recording system) has a unique private key flashed in during manufacturing, with the corresponding public key printed on it's nameplate. The device can sign a video clip and its related meta-data before exporting. Now, any decent hacker could see possible holes in this system, but it could be made tamper-resistant enough that any non-expert wouldn't be able to fabricate a signed video. Then the evidence becomes the signed video and the NVR's serial number and public key. Not perfect, but probably good enough.

          • hippari2 3 days ago

            Unfortunately consumer devices often have weak cryptography built into them. The one properly implemented are just out of reach for average consumer.

      • _DeadFred_ 3 days ago

        This is such BS. The government is ALWAYS deferred to when the chain of custody is broken because 'good faith' is applied. As long as 'good faith' is rountely dispensed 'chain of custody' is nothing but propaganda for the justice system not an actual tool used for justice.

        As long as chain of custody ca be discarded because 'good faith' whenever it becomes inconvenient it is not a real thing.

    • nine_k 3 days ago

      I can easily imagine a future where video evidence is only acceptable in the form of chemically developed analog film, at resolutions that are prohibitively expensive to model, and audio recordings of any kind are not admissible as evidence at all. Signatures on paper, faxes, etc are, of course, inadmissible, too.

      • pmontra 3 days ago

        The point of a signature on paper is that (at least in my country) you can summon somebody in court and ask "is this your signature"? If they say it is, it is, even if it does not look much like any of their other signatures. Then there might be a suspect of false testimony or being blackmailed etc but that's not always the most important issue in a case. E.g.: if it is a contract and all the signers state that they agree with its terms, then there is no much else to discuss.

    • datadrivenangel 3 days ago

      Digital forensics will continue to be an in-demand skill!

  • trhway 3 days ago

    >But ultimately: Humanity has coped without photo, audio or video evidence for most of its existence.

    which allowed to burn witches based on testimony of citizens in good standing.

    And that leads us to using Neuralink and similar tech and the next gen lie detectors (like say the defendant's fMRI (most probably interpreted by AI too:)) to look into the brain and extract confession. No need for evidence, deposition and all that expensive time consuming stuff standing in the way of truth especially given that it can't be trusted anymore in the face of AI capabilities.

  • EGreg 3 days ago

    I kind of want to have an LLM which takes absolutely any criticism of AI or news of it doing something bad and then generates a plausible HN comment that basically goes "I don't think it's a new problem, X has always been possible, there's nothing really new"... because that comment always appears like clockwork beneath it :)

    • fhd2 3 days ago

      Strangely, I'm not even offended by the notion that an AI can replace this work I apparently do :D Though my thought here was just pure optimism in the face of something bad, not an attempt to frame a bad side effect of something good as not a big deal.

      When it comes to generative AI, I personally don't see a lot of good applications, but a plethora of bad ones. The only solution I could imagine would be regulation to the degree that using or distributing models with certain capabilities is just illegal. Judging from the war on file sharing a few decades ago, probably very difficult to enforce, even if it is perhaps still worth doing.

      But I don't see any governments line up to do it. Given that, this particular (semi) new development that generative AI is effective for evidence tampering, I think we'll manage to deal with it.

    • throwway120385 2 days ago

      The nuance that's usually missing is that new technology invariably enables the same things but more of them and faster.

      • consteval 2 days ago

        Everyone misses this all the time.

        Most problems can be overlooked because they’re not that prolific. There’s a chain of events that really hinges on that assumption.

        For example, piracy is really not a problem. In fact I’m pro-piracy most of the time. But if everyone could pirate with no thought and no down time, I don’t think our economy will survive. Just because something is okay in small doses doesn’t mean it scales well. We understand this intuitively with substances, but not with technology like LLMs.

        • EGreg 2 days ago

          Do you think the people who keep commenting "X was always a problem" are really unaware of this? Or they are just trying to dampen any alarm on purpose?

          • consteval a day ago

            I think they’re genuinely unaware of this because many, maybe most, people who analyze risk casually don’t think about ease or probability. Because it’s easier that way, and typically allows them to maintain some beliefs about personal freedoms.

      • pixl97 2 days ago

        A snowflake isn't an avalanche, but if you let them build up, there is potential.

    • dragonwriter 3 days ago

      I mean, you are already getting the result for free, so why do you need an LLM for it?

      In any case, you can probably do this with a fairly simple prompt with any LLM.

  • RataNova 3 days ago

    That's a great point. The ability to undermine real evidence by claiming it's AI-generated could be just as (if not more) damaging than fake evidence itself

    • pixl97 2 days ago

      Well, and to generate fake evidence to lead cops off the trail for the longest period of time possible. Evidence could be said to contain a half life. The longer (real) evidence is not gathered the more it can decay in one way or another. For example security footage gets overwritten. Actual witness memory gets foggy. Physical items get thrown away.

      If the cops spend the first month going after some poor Mark that had nothing to do with it, by the time they realize they've been had it will likely be much more difficult to catch the actual perp.

      Also, that's saying nothing about the court of public opinion.

  • kenjackson 2 days ago

    > I don't think it's _that_ widespread. Just like it's possible to lie on the stand, but people usually think twice before they do it, because _if_ they are found to have lied, they're in trouble.

    I don't have data, but I suspect a LOT of people lie on the stand. This is mostly based on what I see on reality court shows and true crime type shows, so admittedly not a great sample, but I figure once something gets to trial things are going to be contentious.

  • HPsquared 3 days ago

    At this stage it's more a risk for people who have their likeness out in the public domain where it can be copied. But that's just about everyone these days.

  • vasco 3 days ago

    > But ultimately: Humanity has coped without photo, audio or video evidence for most of its existence. I suppose it will cope again.

    Same argument for electricity, the internet, sanitation, democracy... doesn't seem like a great test for stuff that we just didn't have it before and survived.

    • fhd2 3 days ago

      Well, I'm not arguing "good riddance". I just optimistically think we'll manage. I wouldn't want to miss any of the things you listed, actually. I'd also prefer evidence tampering to be impossible or at least very difficult. But that's not a call I can make. All I can decide about things out of my control, is how I think about it. And here I'm carefully optimistic.

      • vasco 2 days ago

        But your argument for why we can live without audio or video evidence was that we lived like that before. All I'm saying is that's not a great argument. You can probably come up with other arguments and they might be perfectly valid.

        • fhd2 2 days ago

          Fair point. I didn't mean it too sound that dismissive. I'd rather not live without electricity, but if something was happening that eliminated electricity, I'd probably say "we did it before, we can do it again", too. And yeah, I'd probably say it in that case, not post to HN, for technical reasons.

        • pixl97 2 days ago

          I mean, technically, a lot of murders did get away with their crimes in the days before video evidence, so there is that.

treetalker 3 days ago

I question the wisdom of setting the judge up as a superjury / gatekeeper for this kind of situation. This seems like a reliability / weight of the evidence scenario, not a reliability / qualification of the witness scenario (as with an expert witness).

Why would the judge be better qualified to determine whether the voice was authentic, as opposed to the witness? And why should the judge effectively determine the witness's credibility or ability to discern, when that's what juries are for?

All that said, emulated voices do pose big problems for litigation.

  • _DeadFred_ 3 days ago

    You mean like expert witness polygraphers that were treated as fact by the courts for years and still used to re-incarcerate people on parole/probation?

    Or gun matching (ballistics) that are no longer considered conclusive but subjective? https://www.mdcourts.gov/data/opinions/coa/2023/10a22.pdf

    Hair and Fiber expert analysis that was wrong? https://innocenceproject.org/fbi-agents-gave-erroneous-testi...

    Or do you mean bite mark analysis that was again wrong?

    Many of these forensic methods were used for decades and presented/treated as conclusive evidence before being challenged leading to wrongful convictions. But yes, let's have 'certified' voice experts whose living is based on being hired by the government and giving testimony to convict people. Surely this time it will be altruistic and scientific.

    • klipklop 3 days ago

      What always worried me is that nobody ever challenged these methods. We just accepted it as fact. Endless crime tv shows reenforce that these methods are infallible. People that watch these shows then become jurors. Scary.

      • SoftTalker 3 days ago

        Cops looking at blurry 240p video frame. "Enhance!" and suddenly it's 4k.

  • bluGill 3 days ago

    I don't know what the latest is, but often judges are supposed to not allow "expert testimony" without checking that the person is an expert. However this is a really complex area. Judges don't want to be the one deciding a case, but not allowing some "expert" is in a way deciding the case.

    • zusammen 3 days ago

      There’s a built-in design paradox. How does a judge assess an expert in a field where he is not one? There’s probably some improvement that comes from experience but it’s not perfect.

      • stogot 3 days ago

        Qualifications=academics usually & career experience. No way to know if they actually learned something and aren’t a fraud. Even corporations that do thousands of interviews get duped

        • aziaziazi 2 days ago

          Some thought

          - perhaps the judge make less experts interviews than the corporations, leading to less experience in that but also takes each of them more seriously.

          - one way to remove/add some credit to someone claim is to ask some of their peers opinion and see if there’s a strong majority.

          - the judge personal expertise may help him forge a precise opinion but that wouldn’t clear him of making a mistake. For important matters it’s always a good idea to ask for peers reviews. Academics knows that too.

      • bluGill 2 days ago

        Opposing lawyers have an incentive to help the judge. Of course the lawyers will lie, but they still will point out important details for the judge to look at.

    • hiatus 3 days ago

      Isn't it up to the opposing party, and not the judge, to impeach an expert?

      • ARandumGuy 3 days ago

        If it was solely up to the opposing party, then they'd strike down every single expert. The judge still is the ultimate decider on what evidence (including expert testimony) is admissible.

        • SoftTalker 3 days ago

          It's the jury that ultimately decides whether the evidence is convincing, even if the judge allows it. The defense will try to cast doubt on the evidence when they make their arguments.

          This also reminds me of one of Norm Macdonald's bits. He says (very seriously) that if he were on a jury he would not convict someone on the basis of DNA evidence. "What do I know about DNA?" he says.

  • RataNova 3 days ago

    But I think the argument here is less about judges making definitive calls on authenticity and more about ensuring that clearly questionable evidence isn't automatically admitted just because a witness vouches for it.

  • whywhywhywhy 3 days ago

    Gell-Mann Amnesia effect comes to mind.

  • NoMoreNicksLeft 3 days ago

    Excluding fake evidence is very much a responsibility of the judge. In the age of Fox News, letting the jury decide for themselves whether or not made-up bullshit is actual evidence seems like a recipe for disaster, and not necessarily one that errs on the side of caution.

    • treetalker 3 days ago

      To the contrary, in US courts the jury determines whether evidence (documentary, testimonial) is credible or not, and what weight (if any) to assign it. (Experts, à la Daubert etc., are a different matter because they give expert opinions, not factual evidence based on personal witnessing of the events in the case, so the judge does perform a gatekeeping function, essentially to ensure the underlying field/science is reliable.)

      While certainly Fox News headlines would not reach the jury in most instances, that is on account of hearsay, lack of qualification, materiality, relevance, and similar rules. It is not a prior credibility or weight determination by the judge, as I understand TFA to be advocating. So: did the witness hear a voice that he believed to be the one in question? If so, jury gets to decide (unless unfairly prejudicial or some other overriding rule comes into play).

      • pharrington 3 days ago

        IANAL, and I am assuming you are a lawyer - I interpreted @NoMoreNicksLeft as saying that it's a judge's responsibility to determine whether or not evidence is admissible - before it gets to a jury. And that's also what the article is talking about - "The examples should illustrate circumstances that may satisfy the authentication requirement while still leaving judges discretion to exclude an item of evidence if there is other proof that it is a fake. "

        • cperciva 3 days ago

          Judges have a role in excluding evidence which is more prejudicial than probative. In the case of a fake voice recording, hearing someone who sounds like the accused participating in a crime may prejudice the jury even if they later hear evidence that the recording is fake. (This is probably even more true for faked videos.)

      • NoMoreNicksLeft 3 days ago

        >While certainly Fox News headlines would not reach the jury in most instances, that is on account of hearsay,

        Nor should obviously fake evidence reach the jury. They can judge for themselves whether testimony is credible, but this is far different than admitting faked evidence. And if you can't see the difference, I'm not sure I'm qualified to explain it to you.

        • pixl97 2 days ago

          Exactly how is one supposed to determine what evidence is faked versus what evidence is not? Especially in the case where high technology has been used to mask that it is a fake?

          • NoMoreNicksLeft 2 days ago

            >Exactly how is one supposed to determine what evidence is faked versus what evidence is not?

            It's pretty simple. If the "evidence" can be created by software, it's not evidence. Perhaps some specially designed recording hardware might digitally sign a voice recording, and one might be inclined to trust that it was a real recording if the design of that hardware/software system was vetted.

            But just for a recording? There's no point. Allowing a jury to decide that this recording is bad, but this other one is okay when they have no expertise to be able to determine if it is fake or not and none of the technical details either is just asking for prejudicial and even superstitious deliberation.

            We are at the point where audio (and probably video) recordings no longer count as evidence of real world events. It's been this way for photographs for a long time already.

dghlsakjg 3 days ago

How is AI voice faking any different than any other type of faking? How is it different than a manipulated recording, or a recording where someone is imitating another?

It is just as easy to fake many paper documents, and we have accepted documents as evidence for centuries.

Photos can be faked, video can be edited or faked, witnesses lie or misremember.

Is this just about telling lawyers that unvetted audio recordings can be unreliable? Because that shouldn't be news.

Edit: this is a good faith question. I'm legitimately just curious. Splicing and editing have been around since recording was invented, I was legitimately curious why voice recordings would have been given extra evidential weight when manipulating recordings is a known possibility.

  • rout39574 3 days ago

    Presuming good faith here, faking recordings has been harder to do, easier to detect, and less equivocal in the past than it is now.

    If it takes an FX house to generate a plausible recording of me saying something I didn't say, that's a risky enterprise with a lot of witnesses.

    If my enemy can do it in their basement with an hour of research, the exposure risk goes way down, and consequently the expectation you'll see it in real life goes way up.

  • ceejayoz 3 days ago

    Nuclear weapons require nation-state levels of resources. Now imagine you can build one in your basement for $10k. Does that change things at all?

    • dghlsakjg 3 days ago

      I understand what you are saying, but my point is that I could make plausible sounding recordings that did not reflect reality by, for example, cutting recordings up with freeware like Audacity, or even using a consumer level double tapedeck before that. It wasn't Manhattan project levels of effort before this.

      This seems more like people losing their minds over 3d printed guns, when hobbyists with a drill press have been making guns in the garage for decades.

      Yeah, its easier now to fake voice, but its not as if what this article warns against wasn't possible before the latest AI hype cycle. And it is also worth noting that voice cloning/changing technology is not particularly new either (I've been able to sound like Morgan Freeman using a phone app for at least half a decade).

      I agree that courts should be cautious around accepting voice recording evidence, I just don't think that the ability to do this is new.

      • Majromax 3 days ago

        > I could make plausible sounding recordings that did not reflect reality by, for example, cutting recordings up with freeware like Audacity,

        Cutting up what recordings, exactly? If you're mixing-and-matching, you need a pretty broad corpus of source material with fairly consistent recording quality (background noise, etc), and even still you're limited to reproducing words that are already present. You can't cut-and-splice together a recording of me saying 'Yes, dghlsakjg, I embezzled $1 million and blew it on the ponies' because there will be no recording of me saying your username, 'embezzled', or 'ponies.'

        The problem comes with the voice-cloning technology that can construct entirely new sentences based on relatively short voice profile samples.

        > I've been able to sound like Morgan Freeman using a phone app for at least half a decade

        You've been able to sound like Morgan Freeman because of specific, hard tuning work put into the voice changer. Now, you can sound like your boss, or your neighbour, or your ex.

        • gus_massa 3 days ago

          I agree.

          I edited the audio of a few math videos to upload to YouTube and I had to fix "typos" like "then one plus zero is zero". The problem is that there are no spaces in the voice, so it sounds like "thenonepluszeroiszero."

          So I had to look for a "one" that is in a similar context, after an "s" and before a ".", or at least a similar one. And hope the speed matches, and adjust the volume.

          They were free videos, so it was not necesary to get it perfect, but at least reduce the distraction caused by mistakes.

        • dghlsakjg 3 days ago

          I'm giving limited examples, not a conclusive list of how audio could be manipulated. I understand that there are circumstances where my examples don't work.

          My broader point is that audio evidence has never been perfect, just like every other medium that evidence can exist in.

          Generative audio technology is not the first time that audio has been corruptible as evidence, and any lawyer who believed otherwise was naive at best.

          • pixl97 2 days ago

            When I was younger I never had a solicited call on my landline. Simply put any place that would have called to try and sell me something would have been long distance and it wasn't worth the cost. Back then I always answered the phone because it was going to be something important.

            Then the cost of long distance started to drop to zero and the commercial calls started coming in. After it was zero, the outright illegal scams showed up. Now, probably one out of every fifty calls to me is an actual legitimate call. My phone goes to automatic voicemail if you're not already in my contacts.

            Were talking about both quantitative changes and qualitative changes. Things like this can have large impacts on society.

      • ceejayoz 3 days ago

        > I could make plausible sounding recordings that did not reflect reality by, for example, cutting recordings up with freeware like Audacity, or even using a consumer level double tapedeck before that.

        Sure, and https://en.wikipedia.org/wiki/David_Hahn managed to get quite a bit of nuclear material.

        Being able to do it at scale, convincingly, in real-time, for any arbitrary text, with just 30s or so of someone's voice as a sample, changes the calculus a lot.

  • gdulli 3 days ago

    "We've had projectile murder for centuries with arrows, how are these machine guns and missiles any different?"

Lammy 3 days ago

In the future this stuff will get so good that the public will beg to be surveilled at all times because it will be the only way to prove what you didn't do. You will learn to love Total Information Awareness. Consent status: manufactured :)

  • nostrademons 3 days ago

    It could easily go the other way, where the public doesn't care what people think they did or didn't do and just does whatever they want, because they don't respect the state and believe the social contract has been broken. "Fuck justice, talk to my AR-15."

    There's ample evidence that this is already happening, eg. recent headlines about kids being radicalized at increasingly younger ages, groups like No Lives Matter that embrace violent nihilism, increased domestic terrorism, record high gun ownership across both sides of the political spectrum, authority figures that just do whatever they want and ignore any form of law or accountability, etc.

    • bilbo0s 3 days ago

      I don't know man?

      We're already at a place where most people don't care what other people think of them.

      Issue is, as long as the government has the big guns, what the government thinks of you will still matter in a major way.

      In such an environment, most people are going to choose to have some kind of way to prove to the government what they did and what they didn't do. Not because they care what other people think as you're implying, but rather because they very much care that the government not get the wrong idea about them. Because the government getting the wrong idea about you can be fatal.

      • nostrademons 3 days ago

        Government is based on the threat of the use of force, not the actual use of force. If you're a government that is regularly using force against a significant proportion of your citizens, you have problems, and probably will not remain the government for long.

        I suspect that we're saying the same thing but with reversed causality. Both of us agree that non-deterministic enforcement breaks down the incentives needed for pro-social behavior. You're saying that this will cause people to demand ways to improve the governments enforcement abilities. I'm saying that this will cause people to adapt their behavior to the new, lessened enforcement abilities. In defense of my point, I'd point out that changes to government are a coordination problem while adaptation of behavior is an individual-only response, and it is much easier to effect changes to your own behavior than it is to convince 300 million people to agree on a solution and implement it, particularly when the root problem is a lack of enforcement ability.

        • bilbo0s 3 days ago

          Keep in mind that being able to demonstrate your innocence to the government is also just an individual change in behavior. A person might not necessarily care if you use the tracking technologies or submit yourself to all the cameras in society. But they'll choose to do it themselves just so that there's that record out there.

          The government coming up with one way to track everyone is not really necessary to get people to submit to tracking. In fact, most of the law enforcement "watcher" types wouldn't want that anyway. Not only is having a myriad number of ways to track and surveil people is far preferable to law enforcement, but it also allows people to individually choose to set up all the Ring doorbells and security cameras and GPS trackers and smart glasses based body cams etc etc all on their own.

          And they'll happily choose to buy all that stuff voluntarily and without regard to what everyone else is doing.

          • nostrademons 3 days ago

            I think the issue demonstrated by this article is that technology is getting such that demonstrating your innocence to the government is not an individual change of behavior, and that there are ways to abuse the demonstration mechanisms that effectively feed false information to the government.

            Imagine that the populace supports widespread surveillance techniques, and so cameras are setup everywhere. Some hacker group figures out how to hack into the cameras and insert deepfakes in them. Now members of that hacker group have a government-proof alibi whenever they want it, and can commit crimes at will, and get it blamed on others. Justice goes out the window.

            • bilbo0s 3 days ago

              Which speaks to why the law enforcement types prefer multiple surveillance and tracking channels. A hacker group compromising one may be possible, but to get away with a crime, it would be necessary to compromise the feeds coming from all of the channels at the same time. Extremely unlikely. The deep fake of someone else robbing a store that you put on the store's security camera system is nice. But the police are likely to be more convinced by the other ten thousand smart phones, smart glasses, and security and door bell cameras in the neighborhood that recorded you and your buddies running out of that store with guns at the time of the robbery.

              This is how the "watcher" types are trained. Information is only valid if they can get it from multiple independent sources. So they love when new tracking and surveillance channels are released. (Social media apps, or smart glasses, or doorbell cameras or what have you.)

              The bar would be much higher than compromising a single channel. With multiplying channels, the task of compromising them all approaches impossibility.

        • pixl97 2 days ago

          > If you're a government that is regularly using force against a significant proportion of your citizens

          Yea. Russia is going to collapse any day now........... um, I'm still waiting.

PolieBotics 3 days ago

I've developed a novel approach to creating tamper-evident video via cryptographic feedback loops between projectors and cameras. The process works as follows:

1. A projector displays a challenge pattern (Perlin noise derived from of a hash) 2. A camera captures this projection 3. The system hashes the captured image concatenated with the previous hash and uses it to derive the next projection 4. This chain demonstrates true temporal sequentiality that's difficult to forge

By incorporating random noise derived from Byzantine Fault Tolerant networks and using these networks as timestamping servers, the proofs inherit the network's decentralization properties. ML then confirms that the feature distributions in projection-photograph pairs match expected patterns from the training dataset.

Demo video and GitHub repo available here: https://www.reddit.com/r/PoliePals/comments/1j8qm2j/truth_be...

  • o11c 3 days ago

    There are actually (at least) 3 different places where cryptography is needed here:

    * Proof that this starts after a given time. Traditionally this has used methods like "this is the headline of a major newspaper today", which is limited to 1-day granularity and has problems if you can just generate a large number of expected headlines and use them in parallel. But with crypto, we can just query any random-number-timestamp-signing server, and a network of such servers can mutually sign each other's previous packets so it's very reliable both against downtime and against attacks.

    * Proof of sequencing. This is trivial with a chain of hashes, though it does prevent recompression.

    * Proof that this ends before a given time. This requires actively submitting your signature data to a timestamp server for additional signing, which is a much more complicated task than the initial half. It is still possible to eliminate the single source of vulnerability, but much more work.

    "Camera looks at monitor" is going to be a much cheaper way to make this air-gapped than adding a projector. And this doesn't strictly need to be continuous; most things are tolerant of one-day granularity and almost everything of 15-minute granularity.

    • PolieBotics 3 days ago

      Largely true, although submitting timestamp hashes to the blockchain is probably the easiest bit.

      "Camera looking at a monitor": While that might be simpler in some setups, it doesn’t really solve my main issue. I want the signal to permeate the entire scene, not just appear in the corner of a display or overlaid on the video. By projecting the challenge onto all visible surfaces, we create a physical environment that’s difficult to fake (since you’d have to convincingly generate or remove those patterns in real time). Air-gapping isn’t really the goal right now.

      Finally, we're need much finer granularity than 15 minutes! The point is to lower the generation time below what is achievable via generative model.

      Thank you for the comment, and I hope these clarifications are useful. It's a new concept, so please forgive the clumsiness with which I may be communicating it.

      • o11c 3 days ago

        "The blockchain" is just a wasteful way of doing things compared to the plain crypto.

        "Overlay the entire scene" doesn't actually appear to add any information-theoretic value compared to simply bounding the timestamp at which the video was made. Nothing either of us is talking about will actually prevent fakes (before the camera signs it), only constrain the time at which the fakes are made.

        Slower-than-real-time generation of fakes is still significantly inhibited by the fact that the hash sequence can be checked for continuity across long lengths of time whose bounds are verified using the other steps.

        • PolieBotics 2 days ago

          I’m using a blockchain because it’s currently the most practical Byzantine Fault Tolerant (BFT) network available for timestamping the Initialization Vector and final hash. If you prefer a web-of-trust or another cryptographic framework, feel free to use that instead.

          By projecting across the entire scene, we significantly increase the complexity of any real-time forgery attempt. A single display that barely interacts with the surroundings is comparatively easy to spoof and not particularly convenient to deploy. Of course, if you think that simpler setup is worth exploring, go ahead and experiment!

          Keep in mind, this method differs from merely signing an image. It creates an optical puzzle that reality solves faster than a simulation can. The physical interaction with the environment adds complexity that’s difficult to replicate artificially. I tried placing a display in front of the camera about a decade ago, but found that approach too “analytically solvable,” so it lost its appeal for robust authentication.

  • asddubs 3 days ago

    pretty clever, kind of like cipher block chaining

    • PolieBotics 3 days ago

      Thank you! It is indeed a little like a signature based on proof-of-projection.

      As they say, once you have a signature, you have a most of a cryptosystem. I've been experimenting with those and other applications of non-linear functions in projector-camera systems.

      https://github.com/poliebotics/PolieBotics

recursive 3 days ago

Here's my plan.

1. Cryptographically hash each piece of media when it's recorded.

2. Submit the hash to a "trusted" authority.

3. It will add a timestamp and sign the result.

4. Now, as long as you keep the original, without re-compressing, and you trust the authority, you have some evidence that the media existed at a timestamp. On or before.

This doesn't prove authenticity, but in many cases, establishing a timestamp would be enough. Forgeries probably wouldn't be created until later, after the shit hit the fan.

Or maybe this doesn't work at all.

  • AnthonyMouse 3 days ago

    Thieves are planning an inside job. They forge the surveillance video ahead of time, do the theft and submit the forgery to be timestamped while it's happening.

    Also, a lot of surveillance systems are purposely kept offline to prevent them from being compromised, but your system doesn't allow that because they would need external connectivity to get signatures.

    • lolc 3 days ago

      Sure when you're controlling the source, you can fake it. But requiring the fakes be prepared ahead of time locks the faker into one story that may be contradicted by other evidence.

      • AnthonyMouse 3 days ago

        That's pretty much what happens anyway. The police are going to come collect the footage as soon as the crime is reported and you can't change it after they do.

    • Edman274 2 days ago

      This technique is known as an Ella Fitzgerald to those in the heisting industry, as popularized in the documentary Oceans 11.

  • coppsilgold 3 days ago

    You can do this already by yourself or use a service which you need not even trust.

    Every x seconds, collect all the pieces of data that need to be notarized as existing in the world prior to some time t. It can be an arbitrary amount of data.

    Construct a Merkle Tree[1] of all the hashes (or HMACs) of all the data to be notarized. Compute the Merkle Root. Make sure that everyone gets their Merkle Proof (path from leaf to root) or publish the Merkle Tree publicly.

    Embed the Merkle Root into a one or more cryptocurrency blockchains to exploit their immutability guarantees. By either including it as "additional data" to some transaction or just straight up as a fictional cryptocurrency address.

    Every piece of data processed in this way will have a Merkle Proof (cryptographic path from hash of the data to the Merkle Root) that proves it existed prior to the creation of the Merkle Root. The Merkle Root will have its creation time bounded by the proof of work conducted on the cryptocurrency blockchain.

    [1] <https://en.wikipedia.org/wiki/Merkle_tree>

    • recursive 2 days ago

      There are some tradeoffs here. You need to do currency transfers get a timestamp, with all the constraints that comes with. But you get probably a more trustworthy authority.

  • HPsquared 3 days ago

    Isn't this basically achieved by standard cloud storage on phones? Photos and videos get uploaded automatically (with authentication for the user account) and presumably that action is logged somewhere the user can't edit. Just need to prove those logs are secure, and there you go.

    • echoangle 3 days ago

      Unless you’re they police in a murder case, you’re probably not going to get google or apple to give you something that certifies that a file was uploaded before a specific time and not modified after. And just showing me the modification date in the UI wouldn’t convince me.

      • HPsquared 3 days ago

        Much easier with the account holder's consent, it's probably in the "data download" thing you can request. And probably available by subpoena. It's like phone records.

        I don't mean "date modified" in the file metadata when it's uploaded (of course that could easily be spoofed before upload) but the actual "date uploaded". You know something must have been made before a certain time.

tgv 3 days ago

I'll say it again, even though it is rather unpopular here: there has never been a need to develop these tools, nor one to make them easy to deploy, nor one to make them easy to use. Yet all this has happened, and now it may occur that someone is acquitted because AI generated media is so good, the evidence might be artificial. If that happens, and the suspect commits another crime, it's on the conscience of the people that contributed to this. You cannot create something and pretend its use has nothing to do with you.

The tools aren't perfect yet, so it's not too late to stop. Stop the ridiculous image and audio generation tools before it's too late. Nothing of value is lost when these models are made private again, and research is simply halted.

  • cootsnuck 3 days ago

    It actually is too late for that. For anyone unaware, the open source models are already more than sufficient enough for imitations and deepfakes. For better or worse there's no going back.

    Personally, I'd rather we all know this tech is out there and develop defense mechanisms rather than thinking hiding it away will prevent harm.

    The cyber security industry exists because of all the privacy and security issues posed by all the tech we already have had for the past several decades.

    I'm confident the same will happen for AI simply because it is a business opportunity and other businesses and institutions are already talking about these issues.

  • bdangubic 3 days ago

    You cannot create something and pretend its use has nothing to do with you.

    has always worked with guns so it'll always work with anything else :) and guns will kill more people that AI-generated shit for foreseeable future

    • consteval 2 days ago

      Text, images, and things of that class - human communication tools - are much higher level than guns.

      Human communication tools are responsible for, say, the holocaust. At its core, you can distill the holocaust to that. Or choose any tragedy, really.

      I think what we’re dealing with is much worse than we are giving it credit. I’m envisioning the death of trust as a concept itself. Such has never before been faced by humanity, and I don’t know if even our biological social circuits can handle it.

      We’re already half way there. It’s trivial to convince large amount of people of something that isn’t true, and then get them to act on that belief. It just requires some time and some money. Remove the time and the money requirement and I don’t know where we land.

  • qwertox 3 days ago

    I do want a computer which talks to me like a person, with that agility. I am used to listening to voice.

    I am tired of reading so much stuff which could well be spoken to me. Like this comment here, which you now need to read.

    It's only very recently that living beings on this world have learned to read and write, it's not normal. It's normal to communicate through sound.

    • Loughla 2 days ago

      >It's only very recently that living beings on this world have learned to read and write, it's not normal. It's normal to communicate through sound.

      It's normal for humans to communicate via sight and sound combined.

      That being said, we invest a lot of calories and brain power into vision. It's our primary mode of interacting with the planet. This is what we have evolved to rely on.

      I think, therefore, you're wrong. We rely heavily on eyesight, naturally. It's why our eyes are focused on the front, allowing precise binocular vision, whereas our ears are on the side, allowing for broader coverage, but less specific information received. Reading and writing (or at least some form of image communication) is probably more natural for communicating ideas than talking, for humans.

      Even if you are still trying to average us out across other life on earth (which doesn't really make a bit of sense), I think vision is the way to go. Colors and visual attractants and other visual forms of communication are common even in plants, as well as animals.

    • krainboltgreene 3 days ago

      What an incredible line to end on. “It’s not normal to read/write”.

      • qwertox 3 days ago

        I'm not aware of any other species which is capable of reading and writing. But many can certainly hear and make sounds.

        • krainboltgreene 3 days ago

          What a truly fascinating perspective. I wonder if you yearn to build complex tunnel systems underground at the behest of a queen? I’m of course describing the most common habit on the planet.

        • MichaelDickens 3 days ago

          I'm not aware of any other species with a >10,000 word vocabulary, either.

        • consteval 2 days ago

          Yes, return to monkey. And, somehow, this will make life… better?

    • triceratops 3 days ago

      Reading and writing isn't "normal" but using a computer is? Humans have been doing the former for way longer.

    • gowld 3 days ago

      You don't need writing, if you want to live in a cave or wander a prairie, I guess?

  • Ukv 3 days ago

    > Nothing of value is lost when these models are made private again

    To list a few uses I see for voice-generation/cloning tools:

    * Real-time translation of a user's voice, maintaining emotion and intonation

    * Professional-quality audio from cheap microphone setups (for video tutorials, indie games, etc.)

    * Allowing those with speech impairments to communicate using their natural voice again, or:

    * Allowing those uncomfortable with their natural voice to communicate closer to how they wish to be perceived

    * Customization of voice assistants, such as to use a native accent/dialect

    * Movies, podcasts, audiobooks, news broadcasts, etc. made available in a huge range of languages

    * And of course: memes, satire, and parody

    If it exists but isn't widely accessible, it's likely in the hands of Musk/Zuck and and various state actors. To me that seems possibly the worst alternative - to have the public generally unaware that it's possible and receiving few of the benefits listed, yet still having it available as a tool for competent disinformation.

  • educasean 3 days ago

    Fatalities due to automobile accidents are a major drag. We should've stuck with horses

  • a2128 3 days ago

    There is some use to bringing deepfakes to mass adoption. The thing is that since the tech exists, powerful actors with lots of resources will develop these tools for their own use either way. The question is whether they'll be able to fool the masses who are unaware that such realistic deepfakes can exist, or whether they will have no effect as everyone and their mom have already seen similar AI slop on their Facebook feed

  • pixl97 2 days ago

    >The tools aren't perfect yet, so it's not too late to stop

    This is some serious level of cope for HN.

    Simply put, it's not going to happen.

gortok 3 days ago

What shocks (and irritates!) me is that Charles Schwab keeps wanting me to set up voice ID. Why would I want to set up a voice ID for something that is now trivially spoofed?

  • alamortsubite 3 days ago

    When was the last time you encountered this? I remember getting nags up until around the end of last year, but not lately. I like to think they dropped the program because I expressed concerns about it to so many reps, but more likely I've just been dialing in on a different number.

  • MichaelDickens 3 days ago

    Schwab used to have an 8 character maximum on passwords (although at least they changed that). They have never been a paragon of good security practices.

    • SoftTalker 3 days ago

      They've presumably been good enough that their losses are less costly than better security.

carra 3 days ago

Wait some more time and photos or even video recordings will be deemed just as dangerous. And then what? Even if there is real evidence, it will have to be discarded unless it can't be sufficiently validated. It will get very hard to prove anything.

  • bluGill 3 days ago

    Or perhaps camera manufactures will start putting traces in. Anyone who makes surveillance systems (like stores use) should put some end to end things in so they can say "this camera took that recording and we can see that it wasn't tampered with by...". (if anyone works for a surveillance company please run with this!) With encryption we can verify a lot of things, but it sets the bar higher than someone took a picture.

    Of course in a darkroom someone skilled could always make a fake photo - but the bar is a lot lower with AI.

    • AnthonyMouse 3 days ago

      Cryptography doesn't really fix it. There are a zillion camera makers and all it takes is one of them to have poor security and leak the keys. Then the forger uses any of the cameras with extractable keys to sign their forgery.

      Or they just point a camera at a high resolution playback of a forged video.

      This also assumes you can trust all the camera makers, because by doing this you're implicitly giving them each authorization to produce forged videos. Recall that many of these companies are based in China.

      • gs17 3 days ago

        > Or they just point a camera at a high resolution playback of a forged video.

        Exactly, it's just like DRM or cheating in video games. Even if everything in software is blocked, there's always the analogue hole.

      • bluGill 3 days ago

        The point is the camera maker certifies in court under perjury penalty that their cameras are not compromised and that is their image. "Other camera systems are compromised, but ours is not...".

        • AnthonyMouse 3 days ago

          Cameras are an IoT device. The S is for security. What good is having the company certify something that isn't true?

          It seems like it would just be a mechanism to lend undeserved credibility to something that still isn't trustworthy.

          • bluGill 2 days ago

            Here the camera manufacture is asserting under oath that they (not their competitors) are secure. Just because IoT is almost always insecure doesn't mean it 100% is, being the exception can allow you to charge a premium price at times - but you better really be the exception.

            • pixl97 2 days ago

              This isn't how security works. Devices like IoT stay in the field for decades quite often, and generally without updates (or if they do have updates, the updater itself is a security risk).

      • DrFalkyn 3 days ago

        Maybe go back to old school analog. Like they did in Altered Carbon

    • gblargg 3 days ago

      I think the best proof of authenticity is already there, in the quirks each particular camera model has in the recording encoding. It's likely that a forgery would easily be shown to have differences in encoding when compared with a video file from the same camera. It would be more difficult with an audio recorder, as you could just record the AI voice in the room. An audio expert could probably show that the acoustics of the fakery don't match where it was claimed to be recorded.

  • john2x 3 days ago

    I wonder if film photography (slide film in particular) will become viable again and that photojournalists will once again become necessary for society.

    Rather than cryptography which could be difficult to grok for a non-technical jury, a physical slide of film would be the source of truth.

    It can still be faked by photographing a manufactured/generated scene though.

  • paulnpace 3 days ago

    If it goes that direction, one silver lining could be that they also realize video conference tools used for court proceedings should also go away.

  • Lanolderen 3 days ago

    The smart encryption people will fix it. At some point it had to be fixed anyway since LLMs have only made it less effort.

nottorp 3 days ago

So... those talking head "influencers" who leave multiple hours of voice and video samples on social networking for anyone to download and clone are the most at risk for an attack like this?

  • lblume 2 days ago

    Not necessarily. Recent advancements suggest that just a few soundbites of a typical person's voice already allow modern AI to synthesize it fairly accurately.

exe34 3 days ago

On a related note, why oh why does Lloyds Bank insist on grabbing my voice for login every time I call them? I have to keep saying "no, fcuk off!" a dozen times until it gives up.

RataNova 3 days ago

This is a pretty terrifying problem, especially considering how easily accessible AI voice cloning tech has become. The fact that courts still operate under rules written before AI-generated evidence was even a concept is concerning

tiahura 3 days ago

Why can't objecting party ask to voir dire the witness and find out if they can distinguish between AI and real recordings?

TeeMassive 3 days ago

I think we're going to see C2PA (https://c2pa.org/) being mandatory for cellphones. At least when there's at least one implementation and that all digital cameras has an integrated TPM.

WalterBright 2 days ago

Add to that AI generated photographs and video.

BTW, I've always wondered about the validity of email evidence. Email is text on a computer. You can edit it to say whatever you want.

  • A4ET8a8uTh0_v2 2 days ago

    It is true in a very general kind of way, but we no longer live in a world when one admin controls every aspect of company's IT ( with maybe exception of small mom and pop shops ). As a result, default controls in place have gotten a little more granular ( and annoying ). All that said, the evidence by itself is not as useful as the hired expert, who explains what the evidence is, how we know it is good and so on.

  • kragen 2 days ago

    You may be interested the decision in the COPA vs. Wright case https://www.judiciary.uk/judgments/copa-v-wright/ where Craig Wright falsely claimed to be Satoshi Nakamoto and forged many kinds of text on computers in order to support his baseless claims, including emails. The COPA's experts, among others, found lots of cases where he slipped up, and the court's decision and its appendix go into a great deal of detail about how those slipups were detected.

    • WalterBright 2 days ago

      Thank you. Interesting! His mistake was faking the destination address, which did not exist at the time of the timestamp.

      • kragen 2 days ago

        In one case! He made lots of different mistakes.

  • mholt 2 days ago

    In a lawsuit in the early 2000s where my family member was the defendant, the plaintiff had forged emails. In the end, my family member won the lawsuit, so whatever they did, worked! I wish I could remember the details. I will have to go back and ask them how they did it.

    • kragen 2 days ago

      I'm interested to see whether they attempted to show that the email was forged. Are the filings in the case a matter of public record? Usually they are in countries like the US and UK.

  • michaelbuckbee 2 days ago

    I'm aware of using DKIM signing records to prove that certain emails were actually sent through systems at certain times as there are headers that change based on them in the actual emails.

    Note: pretty sure I got the concept right and misused a few terms here.

TriangleEdge 3 days ago

AI threatens all digital perceptions, not just voice. Images, videos, recordings, ... I think soon enough proving things in court beyond a reasonable doubt when the evidence is digital media will be difficult/impossible.

belter 3 days ago

To avoid the scenarios like the one described in the court case, setup a family password or keyword. The bot will not know it.