Leaked Docs Show What Phones Cellebrite Can (and Can’t) Unlock

The leaked April 2024 documents, obtained and verified by 404 Media, show Cellebrite could not unlock a large chunk of modern iPhones.

Cellebrite, the well-known mobile forensics company, was unable to unlock a sizable chunk of modern iPhones available on the market as of April 2024, according to leaked documents verified by 404 Media.

The documents, which also show what various Android handsets and operating system versions Cellebrite can access, provide granular insight into the very recent state of mobile forensic technology. Mobile forensics companies typically do not release details on what specific models their tools can or cannot penetrate, instead using vague terms in marketing materials. The documents obtained by 404 Media, which are given to customers but not published publicly, show how fluid and fast moving the success, or failure, of mobile forensic tools can be, and highlights the constant cat and mouse game between hardware and operating manufacturers like Apple and Google, and the hacking companies looking for vulnerabilities to exploit.

Analysis of the documents also comes after the FBI announced it had successfully gained access to the mobile phone used by Thomas Matthew Crooks, the suspected shooter in the attempted assassination of former President Donald Trump. The FBI has not released details on what brand of phone Crooks used, and it has not said how it was able to unlock his phone.

The documents are titled “Cellebrite iOS Support Matrix” and “Cellebrite Android Support Matrix” respectively. An anonymous source recently sent the full PDFs to 404 Media, who said they obtained them from a Cellebrite customer. GrapheneOS, a privacy and security focused Android-based operating system, previously published screenshots of the same documents online in May, but the material did not receive wider attention beyond the mobile forensics community.

For all locked iPhones able to run 17.4 or newer, the Cellebrite document says “In Research,” meaning they cannot necessarily be unlocked with Cellebrite’s tools. For previous iterations of iOS 17, stretching from 17.1 to 17.3.1, Cellebrite says it does support the iPhone XR and iPhone 11 series. Specifically, the document says Cellebrite recently added support to those models for its Supersonic BF [brute force] capability, which claims to gain access to phones quickly. But for the iPhone 12 and up running those operating systems, Cellebrite says support is “Coming soon.”

A SECTION OF THE IOS DOCUMENT. IMAGE: 404 MEDIA.

The iPhone 11 was released in 2019. The iPhone 12 was launched the following year. In other words, Cellebrite was only able to unlock iPhones running the penultimate version of iOS that were released nearly five years ago.

The most recent version of iOS in April 2024 was 17.4.1, which was released in March 2024. Apple then released 17.5.1 in May. According to Apple’s own publicly released data from June, the vast majority of iPhone users have upgraded to iOS 17, with the operating system being installed on 77 percent of all iPhones, and 87 percent of iPhones introduced in the last four years. The data does not break what percentage of those users are on each iteration of iOS 17, though.

Cellebrite offers a variety of mobile forensics tools. That includes the UFED, a hardware device that can extract data from a physically connected mobile phone. The UFED is a common sight in police departments across the country and world, and is sometimes used outside of law enforcement too. Cellebrite also sells Cellebrite Premium, a service that either gives the client’s UFED more capabilities, is handled in Cellebrite’s own cloud, or comes as an “offline turnkey solution,” according to a video on Cellebrite’s website.

That video says that Cellebrite Premium is capable of obtaining the passcode for “nearly all of today’s mobile devices, including the latest iOS and Android versions.”

That claim does not appear to be reflected in the leaked documents, which show that, as of April, Cellebrite could not access from locked iOS phones running 17.4.

The second document shows that Cellebrite does not have blanket coverage of locked Android devices either, although it covers most of those listed. Cellebrite cannot, for example, brute force a Google Pixel 6, 7, or 8 that has been turned off to get the users’ data, according to the document. The most recent version of Android at the time of the Cellebrite documents was Android 14, released October 2023. The Pixel 6 was released in 2021.

A SECTION OF THE ANDROID DOCUMENT. IMAGE: 404 MEDIA.

Cellebrite confirmed the authenticity of the documents in an emailed statement to 404 Media. “Similar to any other software company, the documents are designed to help our customers understand Cellebrite’s technology capabilities as they conduct ethical, legally sanctioned investigations—bound by the confines of a search warrant or an owner’s consent to search. The reason we do not openly advertise our updates is so that bad actors are not privy to information that could further their criminal activity,” Victor Ryan Cooper, senior director of corporate communications and content at Cellebrite, wrote.

“Cellebrite does not sell to countries sanctioned by the U.S., EU, UK or Israeli governments or those on the Financial Action Task Force (FATF) blacklist. We only work with and pursue customers who we believe will act lawfully and not in a manner incompatible with privacy rights or human rights,” the email added. In 2021 Al Jazeera and Haaretz reported that a paramilitary force in Bangladesh was trained to use Cellebrite’s technology.

Cellebrite is not the only mobile forensics company targeting iOS devices. Grayshift makes a product called the GrayKey, which originally was focused on iOS devices before expanding to Android phones too. It is not clear what the GrayKey’s current capabilities are. Magnet Forensics, which merged with Grayshift in 2023, did not immediately respond to a request for comment.

Cellebrite’s Android-focused document also explicitly mentions GrapheneOS in two tables. As well as being an operating system that the privacy-conscious might use, 404 Media has spoken to multiple people in the underground industry selling secure phones to drug traffickers who said some of their clients have moved to using GrapheneOS in recent years.

Daniel Micay, founder of GrapheneOS, told 404 Media that GrapheneOS joined a Discord server whose members include law enforcement officials and which is dedicated to discussions around mobile forensics. “We joined and they approved us, with our official GrapheneOS account, but it seems some cops got really mad and got a mod to ban us even though we didn’t post anything off topic or do anything bad,” Micay said.

There is intense secrecy around the community of mobile forensics experts that discuss the latest unlocking tricks and shortcomings with their peers. In 2018 at Motherboard, I reported that law enforcement officials were trying to hide their emails about phone unlocking tools. At the time, I was receiving leaks of emails and documents from inside mobile forensics groups. In an attempt to obtain more information, I sent public records requests for more emails.

“Just a heads up, my department received two public records request[s] from a Joseph Cox at Motherboard.com requesting 2 years of my emails,” a law enforcement official wrote in one email to other members. I learned of this through a subsequent leak of that email. (404 Media continues to receive leaks, including a recent set of screenshots from a mobile forensics Discord group).

Google did not respond to a request for comment. Apple declined to comment.

    • disguy_ovahea@lemmy.world
      link
      fedilink
      arrow-up
      26
      ·
      4 months ago

      Exploiting security vulnerabilities is always a game of cat-and-mouse.

      It’s best to set your phone to automatically update to the newest software available to keep it secure.

      • poVoq@slrpnk.net
        link
        fedilink
        arrow-up
        7
        arrow-down
        12
        ·
        4 months ago

        This assumes the official upstream is not compromised. Apple absolutely will push new vulnerabilities in updates when requested so by higher level law enforcement and/or the NSA etc. The FBI is also known to have placed moles in app vendor teams.

        • disguy_ovahea@lemmy.world
          link
          fedilink
          arrow-up
          14
          arrow-down
          1
          ·
          edit-2
          4 months ago

          I’ve read about the NSA exploiting third-party push notification systems from Snowden. I’ve only read the opposite of your claim about Apple providing a backdoor for the government. Do you have a source so I can read more about it?

          • poVoq@slrpnk.net
            link
            fedilink
            arrow-up
            10
            arrow-down
            10
            ·
            4 months ago

            Apple is a US based company. They are legally required to comply and not openly speak about it when a court asks them to do so.

            Same for Google obviously, but since their software on end-devices is much more diverse, it is much harder to do.

            • disguy_ovahea@lemmy.world
              link
              fedilink
              arrow-up
              17
              arrow-down
              3
              ·
              edit-2
              4 months ago

              Apple’s legal policy is very clear. They do not have access to data on customer devices, and therefore cannot provide said data to law enforcement or government officials.

              They’ve fought the backdoor request in court 11 times and won. They are not mandated to comply by law if they themselves do not have access.

              They can and will provide iCloud data in response to a warrant.

              https://www.apple.com/legal/privacy/law-enforcement-guidelines-us.pdf

                • disguy_ovahea@lemmy.world
                  link
                  fedilink
                  arrow-up
                  14
                  ·
                  edit-2
                  4 months ago

                  You claimed that Apple intentionally places vulnerabilities in updates to allow governmental access, right?

                  They’ve repeatedly claimed in court that they do not have access to on-device data or transmissions, as stated in their legal privacy policy. If they created a vulnerability, then they could exploit it, making the above claims both perjury and breach of contract.

                  Please clarify if I misunderstood your comment.

              • notanaltaccount@lemmy.world
                link
                fedilink
                arrow-up
                3
                arrow-down
                9
                ·
                4 months ago

                This is an irrelevant and stupid comment. If Apple is required by a gag order not to disclose an exploit, their legal policy would not violate the gag order.

                You think if they did build an exploit after a demand they would write about it in their legal notices? That’s just dumb.

          • unexposedhazard@discuss.tchncs.de
            link
            fedilink
            arrow-up
            7
            arrow-down
            10
            ·
            edit-2
            4 months ago

            Its not a matter of “have they done it” its just that the US feds can legally force any corporation to do as they please under threat of violence while binding them with gag orders.

            Even if there is no proof so far, there is no reason to believe that they havent done so, because they legally can.

            Its also specifically disallowed to deploy any kind of system that automatically notifies the public in case law enforcement compromises a companies servers as part of an investigation. Apple used to have such a warrant canary but they removed it some years ago.

            See Warrant Canary

              • unexposedhazard@discuss.tchncs.de
                link
                fedilink
                arrow-up
                9
                arrow-down
                9
                ·
                edit-2
                4 months ago

                Its a closed source operating system dude, they can do whatever they want. Just because there are lawsuits, doesnt mean they dont have a backdoor to your phone.

                They can literally force apple to pretend that they didnt give them a backdoor. There are basically no limits to law enforcement powers as long as they say the word “terorrism”.

                • disguy_ovahea@lemmy.world
                  link
                  fedilink
                  arrow-up
                  9
                  ·
                  4 months ago

                  I’m not for trusting corporations either, but you don’t have a source, only theories.

                  Snowden is pretty much the authority on NSA vulnerabilities, and he hasn’t released any proof that Apple has a backdoor on their devices. The only thing he’s demonstrated is how the NSA has used MITM on third-party push notifications, which Apple has since encrypted and relayed to obfuscate.

                  If you have a source, I’m down to read it. Otherwise, you’re just speculating.

                • jet@hackertalks.com
                  link
                  fedilink
                  English
                  arrow-up
                  6
                  arrow-down
                  2
                  ·
                  edit-2
                  4 months ago

                  The conversation below this message goes really off the rails.

                  If anyone wants to look into the actual legal theory of compelling businesses to do things, look for a national security letters.

                  And yes, the legal framework exists, to compel businesses to do something for national security purposes, and not disclose that they’ve done that.

                  I just want to remind everybody fighting below, security is about capabilities, not intentions. It doesn’t matter if a company wouldn’t do something if they had a choice, just that they can do it. That’s how you model your threats by what people can do not what you think they will do

  • narc0tic_bird@lemm.ee
    link
    fedilink
    arrow-up
    35
    ·
    4 months ago

    So in short newer Pixel and iPhone models seem to be the most resilient to these attacks, with every iPhone able to run iOS 17.4 (XS/XR or newer) currently not attackable.

    But obviously an attacker in possession of the device can wait for an exploit to be found on whatever OS version the device is running.

    The by far best protection then is to set a strong passphrase so even if/when your device/OS have known vulnerabilities to allow brute force attacks, the passphrase is too complex to be brute forced in a realistic amount of time.

      • narc0tic_bird@lemm.ee
        link
        fedilink
        arrow-up
        4
        ·
        4 months ago

        I was talking about the “Brute-Force Password to get the user data (CE) decrypted” column, which is probably the more interesting part.

    • umami_wasabi@lemmy.ml
      link
      fedilink
      arrow-up
      8
      ·
      4 months ago

      Can we create a situation where brute force is unfeasible while using 6 digits PIN? That would be nice to have security and convenience at the sametime. Using passphrase isn’t that feasible for me as my Samsung phone randomly prompts for PIN when I need the most.

      • asap@lemmy.world
        link
        fedilink
        English
        arrow-up
        15
        ·
        4 months ago

        Can we create a situation where brute force is unfeasible while using 6 digits PIN?

        According to this comment from GrapheneOS, the latest Pixels and iPhones are not brute forceable with a 6+ digit PIN:

        Pixel 6 and later or the latest iPhones are the only devices where a random 6 digit PIN can’t be brute forced in practice due to the secure element.

      • narc0tic_bird@lemm.ee
        link
        fedilink
        arrow-up
        8
        ·
        edit-2
        4 months ago

        Not really. A 6-digit PIN gives you ~20 bits of entropy so it’ll be cracked in no time. The only protection you’re relying on is the hardware and the OS, and according to the Cellebrite compatibility table it’s mostly a question of when a vulnerability is found, not if.

        So it’s a trade-off between security and usability.

          • nforminvasion@lemmy.world
            link
            fedilink
            arrow-up
            10
            ·
            4 months ago

            If you live in the USA, the cops can now legally force you to unlock your phone using biometrics. And I mean I guess they can also illegally (and whose gonna stop them) force you to open your phone with a passcode or pattern as well. But it’s a lot easier for them to hold your finger on a scanner or your phone to your face than make you type in the code.

            • disguy_ovahea@lemmy.world
              link
              fedilink
              arrow-up
              6
              ·
              4 months ago

              Absolutely. If you have an iPhone, press the side button 5 times to disable Face ID or Touch ID. We should all be aware of how to disable biometrics on our devices.

            • helenslunch@feddit.nl
              link
              fedilink
              arrow-up
              2
              ·
              4 months ago

              And I mean I guess they can also illegally (and whose gonna stop them) force you to open your phone with a passcode or pattern as well.

              You would stop them. By not giving it to them.

              Of course there’s always torture, which cops have been known to do, but is probably “unlikely”.

        • Possibly linux@lemmy.zip
          link
          fedilink
          English
          arrow-up
          2
          ·
          edit-2
          4 months ago

          Couldn’t you just add time to the equation? Just make the hash really hard

          6 digits is still kind of weak but you could bump it up but not so much that it takes forever to enter

          • FierySpectre@lemmy.world
            link
            fedilink
            arrow-up
            1
            ·
            4 months ago

            Simply making the hash really hard is not a good option. All most people will notice is that their underpowered phone suddenly takes way longer to unlock compared to before. Cracking the hash on very powerful hardware is then ‘trivial’

            As the other comment mentioned, a hardware solution seems to be the only one.

            • Possibly linux@lemmy.zip
              link
              fedilink
              English
              arrow-up
              1
              ·
              4 months ago

              Hardware solutions involve a lot of trust and assumptions. All it takes is a little hardware flaw or simple bus injection and you are toast.

              You could offload cryptography to the dedicated hardware but that is already pretty common

    • helenslunch@feddit.nl
      link
      fedilink
      arrow-up
      3
      ·
      4 months ago

      newer Pixel and iPhone models seem to be the most resilient to these attacks

      Pretty much exactly what one would expect, honestly. Especially with so many Android OEMs lacking critical security updates.

    • Not_mikey@slrpnk.net
      link
      fedilink
      arrow-up
      21
      arrow-down
      1
      ·
      4 months ago

      By a truly unbelievable coincidence, I was recently out for a walk when I saw a small package fall off a truck ahead of me. … Inside, we found the latest versions of the Cellebrite software, a hardware dongle designed to prevent piracy (tells you something about their customers I guess!), and a bizarrely large number of cable adapters.

      I guess it makes sense signal works with the mafia

  • jet@hackertalks.com
    link
    fedilink
    English
    arrow-up
    20
    arrow-down
    1
    ·
    4 months ago

    The main takeaway from this, is not that these phones are absolutely secure now and forever, it’s just that the newer phones are secure… For now.

    Most security just buys you time. Might be months, might be years. But it just buys you time.

    This is demonstrated that basically any phone older than 2 years is completely breakable. It’s an arms race, and if you own a phone you become a static target.

    So if data is absolutely critical, it shouldn’t rely on a low entropy password dependent on the secure element of a phone. That could be fine for things that are ephemeral, things you can change from the cloud, but not say the KFC secret recipe

    • disguy_ovahea@lemmy.world
      link
      fedilink
      arrow-up
      20
      arrow-down
      2
      ·
      edit-2
      4 months ago

      That’s simply not true. The chart clearly states that Cellebrite cannot break encryption on any iPhone that can run the current version of iOS. That includes any iPhone made in the last six years.

      What this illustrates is the importance keeping your phone’s software updated for maximum security. New hardware is not necessary.

      • jet@hackertalks.com
        link
        fedilink
        English
        arrow-up
        15
        ·
        4 months ago

        Scenario: You have the super secret KFC recipe on your phone. I steal your phone and wait until the cracking tools have caught up to the version that was on your phone. All I have to do is keep it in storage. While your phone is in storage its not getting updates.

        By keeping your phone updated you have EXTENDED the time phone security can keep an attacker out of your phone, but in time they will be able to get into it.

        • disguy_ovahea@lemmy.world
          link
          fedilink
          arrow-up
          7
          ·
          edit-2
          4 months ago

          I see your point. The attacker would need to be far more savvy than police.

          They’d need to keep it charged to prevent powering off and activating the connection access denial of the Secure Enclave, while keeping the iPhone in a Mylar bag to prevent any nearby iOS devices from relaying the Find My remote erase request.

          With enough time, a vulnerability may be found for that version of iOS.

          • jet@hackertalks.com
            link
            fedilink
            English
            arrow-up
            3
            ·
            4 months ago

            Probably not powered on for years. The battery would die at some point. But yes the rest of your statement I agree with, there are protocols for cold case phones, they’re just stored waiting for the cracking to catch up

            • disguy_ovahea@lemmy.world
              link
              fedilink
              arrow-up
              3
              ·
              4 months ago

              Cellebrite uses Lightning/USB-C port access to bypass passcode security. Once iPhone is powered off, the Secure Enclave will deny port access until the passcode is entered. It’s not impossible, but it’ll certainly take far longer than if they keep the iPhone charging while they wait for an exploit.

              • jet@hackertalks.com
                link
                fedilink
                English
                arrow-up
                1
                arrow-down
                1
                ·
                4 months ago

                Do they remove/disconnect the battery in this mode? If not wont the battery eventually balloon?

                • disguy_ovahea@lemmy.world
                  link
                  fedilink
                  arrow-up
                  2
                  ·
                  4 months ago

                  It’s fine if it happens. Balooning is just a chemical breakdown of the battery. It won’t explode. Disconnecting the battery while keeping the device connected to power runs a high risk of shorting, potentially causing component damage.

                • BearOfaTime@lemm.ee
                  link
                  fedilink
                  arrow-up
                  2
                  arrow-down
                  1
                  ·
                  4 months ago

                  Simply not charging to 100% constantly will prevent this.

                  I’ve had phones on a charger for years, using a simple timer, so it never gets to 100%. Pretty easy to do.

        • helenslunch@feddit.nl
          link
          fedilink
          arrow-up
          2
          ·
          4 months ago

          That’s really interesting. For example it says GOS phones can’t be accessed past 2022. But if you had confiscated a GOS phone in, say, 2020, and just waited for that particular vulnerability to rear it’s head, you could access it and all of its’ contents.

          • jet@hackertalks.com
            link
            fedilink
            English
            arrow-up
            2
            ·
            4 months ago

            100%

            Just like door locks and safes, they just buy time. Hopefully whatever data was seized in 2020 isn’t going to be relevant in 2024

        • Possibly linux@lemmy.zip
          link
          fedilink
          English
          arrow-up
          1
          ·
          4 months ago

          Chances are they don’t need to just randomly guess keys. They can use stored data and known data such as birthdays and passwords to online services

    • qprimed@lemmy.ml
      link
      fedilink
      English
      arrow-up
      3
      ·
      4 months ago

      I know this may be pedantic, but

      it’s just that the newer phones are secure… For now.

      that statement suggests that other recovery techniques (e.g. hardware decaping, state zero-days, etc), dont already make absolutely current devices insecure.

      it would not surprise me that a TLA with physical access could not recover enclave information and completely expose naked storage in 2 days or less - its just a matter of how urgent is. a former president nearly loosing their head might count as pretty damn urgent. meddle with the us power structure and it will protect itself - if not the particular individuals.

      if you noticed any power dips in your area right after the attempt on trump, that was probably the local NSA cluster firing up.

      • jet@hackertalks.com
        link
        fedilink
        English
        arrow-up
        4
        ·
        edit-2
        4 months ago

        In the context of what we know is commercially available, the new phones are reasonably secure, for now

        If whatever your doing is worth decapping the chip and scanning the underlying electron state, then I’d be more worried about your kneecaps then your phone.

        dont already make absolutely current devices insecure.

        We don’t have any evidence for this statement, and we can never prove the negative (that a device is absolutely secure).

        Depends on your threat model of course, but given the data we have available, the advice still stands, use a modern phone directly from google/apple and keep it updated.

        • Socsa@sh.itjust.works
          link
          fedilink
          arrow-up
          2
          ·
          edit-2
          4 months ago

          Right, there is no lazy replacement for good opsec. If you are facing a state actor then you need to assume your device storage can be accessed.

        • qprimed@lemmy.ml
          link
          fedilink
          English
          arrow-up
          1
          ·
          edit-2
          4 months ago

          We don’t have any evidence for this statement, and we can never prove the negative (that a device is absolutely secure).

          does anyone have info on crooks’s phone make/model (and an OS version guess)? I have not seen this anywhere and, if it is not already in the public domain, why?

          FBI says they got the phone contents in the clear. that gives some boundsfor both known (easy) potential compromises and “oh, fuck.”

          edit: typo

          edit 2: still no detail but I found this…

          The phone was a relatively new model, which can be harder for law enforcement to access than old phones because of newer software, according to technology experts. In many federal investigations, it can take hours, weeks or months to open a suspect’s phone.

          https://www.washingtonpost.com/nation/2024/07/16/thomas-crooks-phone-motive-parents-trump-shooting/

          unless this information is now a state secret, we may learn more.

            • qprimed@lemmy.ml
              link
              fedilink
              English
              arrow-up
              3
              ·
              4 months ago

              couldnt read the whole article, but I would say, this is an “oh, fuck.” moment for a lot of people

              The FBI tries and cannot unlock a recent phone. they allegedly turn to a private company for emergency access and they are in within hours(?)

              I think many people should reevaluate their threat model and mitigations after this. perhaps this changes nothing for you. perhaps it changes everything for you.

  • Possibly linux@lemmy.zip
    link
    fedilink
    English
    arrow-up
    8
    ·
    edit-2
    4 months ago

    It would be nice if it would focus on Android and AOSP from a publication perspective.

    However, most devices are going to be highly vulnerable as they don’t have boot loader relock. Also Androids encryption is weak and not well implemented.

  • ShortN0te@lemmy.ml
    link
    fedilink
    arrow-up
    7
    ·
    4 months ago

    Actually worse than I thought.

    So basically the USB is the point of entry? When you could permanently fuse the data connection in the SoC it would be a huge improvement in security. OFC you could only use the port for charging anymore.

    • jet@hackertalks.com
      link
      fedilink
      English
      arrow-up
      13
      ·
      4 months ago

      Yes. The entire risk surface increases every time you add a new protocol. Bluetooth, NFC, charging, it’s a real problem because features keep increasing.

      There are some crazy vendors who will remove a lot of that from your phone, but it’s pretty rare.

  • krolden@lemmy.ml
    link
    fedilink
    arrow-up
    6
    ·
    4 months ago

    GOS is spooked time to switch to carrier pigeons and live under a bridge with a dolphin.

  • bdonvr@thelemmy.club
    link
    fedilink
    arrow-up
    3
    ·
    4 months ago

    So it seems most of this revolves around brute forcing. So using a properly long and secure password should make it nearly impossible for them right?

    • jet@hackertalks.com
      link
      fedilink
      English
      arrow-up
      7
      ·
      4 months ago

      Kinda. If the phone is off completely, then having a very long key that is cryptographically sound would be sufficient. Unless, there was an exploit in the secure element itself. Even if your user key is very secure, most phones use a secure element taking the user key going to a super secure cryptographic key. So if that lookup can get exploited in the hardware, game over

      • bdonvr@thelemmy.club
        link
        fedilink
        arrow-up
        3
        ·
        4 months ago

        Figured it had to be something like that seeing as changing your password doesn’t involve a long reencryption process