After the $285M Drift hack, new Solana scare shows crypto’s next security risk may already be inside

6 hours ago 2

The Drift exploit and Stabble’s precautionary warning point to a difficult crypto security problem: the next major breach may begin long before funds move on-chain.

That is what makes these incidents more than isolated alarms. They suggest that some protocols may still be looking for smart contract flaws, while the real exposure lies in hiring, access, governance, and trusted relationships.

On Apr. 1, Drift suspended deposits and withdrawals and told users it was under an active attack.

By Apr. 5, the team said with medium-high confidence that the same threat actors behind the October 2024 Radiant Capital hack had executed the operation.

TRM Labs estimated the drain at approximately $285 million, and the Drift post-mortem described a complex scheme in which individuals used $1 million of their own capital and met in person with Drift team members to infiltrate the protocol's structure.

On the technical side, TRM identified the critical weakness as social engineering of multisig signers combined with a zero-timelock Security Council migration. This governance design enabled attackers to execute privileged actions without the delays intended to catch unauthorized changes.

Elliptic said the laundering patterns and network indicators matched those of prior DPRK-attributed operations and pointed to a probable compromise of administrator keys that enabled privileged withdrawals and administrative control.

Attackers earned enough trust to convert ordinary access into a 12-minute, $285 million drain.

New vector of attack for cryptoA timeline shows the Drift exploit unfolded across months of social engineering before a 12-minute, $285 million drain on Apr. 1.

On Apr. 7, the Solana-based liquidity protocol Stabble told its liquidity providers to withdraw funds as a precaution.

The new team that recently acquired the protocol said it had discovered that a former CTO appeared to be the same person ZachXBT had publicly flagged as a North Korean IT worker.

The protocol promised new audits before resuming operations. What Stabble demonstrated was that alleged insider exposure now moves users fast enough to constitute a live funds event on its own.

The operating manual already exists

Treasury's Mar. 12 sanctions release put numbers on the problem: DPRK IT-worker fraud schemes generated nearly $800 million in 2024, using fraudulent documents, stolen identities, and fabricated personas.

The Department of Justice separately said North Korean operatives obtained employment at more than 100 US companies using fake and stolen identities. In one Atlanta blockchain R&D case, workers stole more than $900,000 in virtual currency.

These were workforce infiltrations sustained across multiple firms over extended periods.

Flare and IBM X-Force published their operational breakdown on Mar. 18. The research describes a tiered structure of recruiters, facilitators, IT workers, and collaborators who assist with identity verification and onboarding.

Once embedded, operatives use remote access tools, VPN and proxy services, and internal communication channels, leaving detectable but often-missed traces in device logs.

Flare and IBM frame this as a shared problem owned jointly by security teams and HR, requiring coordination across hiring, onboarding, access controls, and offboarding disciplines.

StageWho is involvedWhat happensWhat the warning sign looks likeWhy crypto teams miss it
Recruitment / identity fabricationRecruiters, facilitators, fake applicants, collaboratorsOperatives build false personas using fraudulent documents, stolen identities, and fabricated employment histories to get through screeningInconsistent biographical details, thin digital footprint, identity mismatches, suspicious referencesTeams optimize for speed and technical talent, not adversarial hiring review
Hiring / onboardingHR, hiring managers, collaborators / brokers, IT workersCollaborators help candidates pass identity verification, background checks, and onboarding stepsUnusual help during onboarding, documentation anomalies, device / location inconsistenciesHiring and security often operate separately, so no single team sees the whole pattern
Embedding inside teamsIT workers, managers, coworkers, contractorsOnce hired, operatives establish legitimacy over time through routine work and trusted relationshipsHeavy use of VPNs / proxies, unusual remote-access patterns, odd device logs, limited willingness for direct interactionNormal remote-work behavior can mask the indicators, and smaller teams lack monitoring depth
Access accumulationDevelopers, admins, signers, governance operatorsTrusted insiders gain permissions, signer influence, admin access, or visibility into sensitive workflowsPermission creep, over-broad role access, weak separation of duties, dormant approvals sitting in placeCrypto security is often code-centric, so human access design gets less scrutiny than smart contracts
Exploitation / theft or extortionCompromised insiders, external handlers, laundering networksAttackers convert ordinary access into privileged withdrawals, governance actions, key compromise, or post-access theftSudden use of privileged functions, suspicious governance migrations, unusual withdrawal behavior, emergency pausesBy the time on-chain activity looks abnormal, the trust failure happened much earlier
Post-incident responseProtocol teams, users, auditors, investigatorsTeams pause operations, ask users to withdraw, rotate access, commission audits, and investigate exposurePrecautionary withdrawal warnings, audit resets, access reviews, attribution updatesMost protocols do not have mature playbooks for insider-risk containment and offboarding

Reuters reported on Mar. 31 that a North Korea-linked operation compromised the widely used Axios npm package in a supply chain attack that could have affected millions of environments.

The actor behind that compromise, UNC1069, is distinct from UNC4736, the cluster Drift tied to the Radiant hack. Yet both cases exploit a trusted relationship comprising a trusted person, a trusted signer, and a trusted package before touching funds or systems.

What to expect

The bear case runs through what Drift's staging timeline exposes about latent exposure across DeFi.

If attackers spent from Mar. 11 to Apr. 1 embedding pre-signed authorizations and engineering approvals before executing the drain, this adds to months of complex social engineering. Other protocols may already host compromised signers, contractors, or contributors they have yet to identify.

Stabble's situation, where a suspected link to a flagged identity surfaced in ZachXBT's public research before the team's own controls caught it, illustrates how often organizations learn about their own exposure from the outside.

CryptoSlate Daily Brief

Daily signals, zero noise.

Market-moving headlines and context delivered every morning in one tight read.

5-minute digest 100k+ readers

Free. No spam. Unsubscribe any time.

Whoops, looks like there was a problem. Please try again.

You’re subscribed. Welcome aboard.

Treasury's $800 million figure for a single year puts a floor on the threat's already cost. DOJ's 100-plus-company figure suggests the target distribution is broad.

In that environment, the next major loss may already be inside the perimeter, waiting on a governance window or an admin key rotation.

The bull case is grounded in the sector's capacity to adapt once the threat model becomes concrete. Drift is the concrete proof, and the countermeasures are well documented.

Protocols can add timelocks to governance migrations, reduce signer powers, segment permissions across functions, and treat onboarding as a security checkpoint with the rigor applied to code audits.

Flare and IBM supply the operational framework: verify identity aggressively, monitor device logs and remote-access indicators, segment contractor access, and build offboarding discipline that revokes credentials and signing authority on exit. The zero-timelock governance design identified by TRM as central to Drift's exploit is fixable.

Protocols that fix it and add organizational controls alongside it materially narrow the attack surface.

If Drift becomes a forcing event, as the 2016 DAO hack did, forcing a reckoning with smart contract risk, the sector could close the gap between known DPRK tactics and actual defenses within a reasonable window.

The harder constraint on the bull case is institutional habit. Crypto teams built their security culture around audits, bounty programs, and formal verification.

Adding identity verification, access minimization, device controls, signer separation, and HR security coordination demands a different operating posture, one that most small-to-medium protocols have yet to build.

The market will price this in, with protocols that demonstrate governance hygiene and operational controls attracting a trust premium.

ScenarioWhat drives itWhat happens inside protocolsMarket consequenceWhat stronger teams do differently
Bear case: latent exposure is already inside the perimeterDrift’s long staging timeline suggests other protocols may already host compromised signers, contractors, or contributorsTeams discover exposure late, often after external research, suspicious activity, or a live incidentMore precautionary pauses, user withdrawals, TVL fragmentation, and a trust discount on smaller protocolsTighten signer controls, add timelocks, rotate credentials faster, segment permissions, and audit org access as aggressively as code
Bull case: Drift becomes a forcing eventThe sector treats Drift as a structural wake-up call, not an isolated hackProtocols upgrade governance design, identity verification, onboarding checks, device monitoring, and offboarding disciplineConfidence gradually stabilizes, with better-defended protocols recovering trust fasterAdd timelocks to governance changes, minimize access, verify identities aggressively, and integrate HR with security operations
Trust-premium case: market rewards operational securityUsers and capital begin distinguishing between audited code and audited organizationsProtocols that can prove governance hygiene and access discipline attract stickier users and counterpartiesA premium emerges for teams with visible controls; weaker teams face higher skepticism and slower liquidity returnPublish clearer security processes, separate signer roles, document offboarding, monitor remote-access indicators, and show repeatable operational hygiene
Stagnation case: the threat is known but habits do not change fast enoughSmall and mid-sized teams keep relying mainly on audits, bounties, and formal verificationCode security improves, but hiring, access, and trusted-software gaps remain openRepeated “surprise” incidents keep resetting confidence and raising the cost of trustTreat non-code controls as part of core protocol security, not as an optional compliance layer

The gap above the code layer

Treasury, DOJ, Flare, IBM, TRM, and Elliptic are each, in different ways, pointing to the same structural gap: smart contract audits address only the code layer.

Who holds signing keys, who vouches for contractors, who reviews device logs, and who has the authority to push a governance migration without a timelock are steps that live above that layer. The current generation of security tooling barely reaches it.

The next exploit may begin with a hiring decision, contractor onboarding, a trusted npm package, or a signer who, over months, earned enough confidence to authorize the one transaction that mattered.

Protocols that close that gap before the next attribution update lands will still have their users' trust when it does.

Read Entire Article
Patroli | Crypto | | |