Tether Hires Big Four Firm for First Full USDT Reserve Audit
Tether announced today it has selected a Big Four accounting firm to conduct its first comprehensive financial statement audit — a significant step beyond the periodic attestations the company has published to date.
CFO Simon McWilliams confirmed the firm was chosen through a competitive process, stating that Tether already operates at Big Four audit standards and that the audit "will be delivered."
A full audit is materially different from an attestation. It involves a detailed review of assets, liabilities, internal controls, and reporting systems — the kind of deep scrutiny that critics have long demanded for the $184 billion stablecoin.
Tether's reserves consist primarily of U.S. Treasury bills, with smaller allocations in gold, bitcoin, and various loans. That composition has drawn persistent questions, particularly around liquidity and risk during periods of market stress. An S&P downgrade late last year citing bitcoin price exposure as a risk factor underscored those concerns.
The company did not name the specific firm. Big Four refers to Deloitte, EY, KPMG, and PwC.
If completed and published, this audit could reset the transparency conversation around stablecoins entirely — or, depending on findings, intensify it.
Invesco Takes Over Superstate's $900M Tokenized Treasury Fund
Invesco, managing $2.2 trillion in assets, is formally entering the tokenized Treasuries space by assuming management of Superstate's USTB fund — one of the largest onchain Treasury products, holding over $900 million in short-term U.S. government securities.
The transition is expected in Q2 2026. The fund will rebrand as the Invesco Short Duration US Government Securities Fund, keeping its existing ticker and token infrastructure intact. Superstate stays on as the technology layer — handling token issuance, onchain settlement, and the digital transfer agent system — while Invesco's global liquidity team (which oversees $200B+ in short-term assets) manages the investment side.
This puts Invesco alongside BlackRock, Franklin Templeton, and Fidelity in the $12 billion tokenized U.S. Treasuries market. The thesis is familiar: blockchain rails offer near-instant settlement, transparent reserves, and 24/7 access — advantages that traditional infrastructure struggles to match.
The tokenization race among major asset managers keeps accelerating. The fact that a $2.2T firm is now acquiring an existing onchain fund rather than building from scratch says a lot about where institutional conviction stands.
Running a DApp on Midnight Doesn’t Feel Like Running a Typical Crypto App
A while ago I started looking into how Midnight Network treats app operators, and it made me pause longer than expected. Not because of the usual architecture or token design, but because of what it actually means to run something in production. Most chains don’t really go deep into that part. They explain how the system works, but not how it feels to operate inside it. The first thing that stood out to me is how data is handled. On most public chains, every user interaction becomes public metadata. Not just for the operator, but for everyone watching. That’s fine for DeFi experiments, but it becomes uncomfortable pretty quickly in more serious use cases. Midnight seems to approach this differently. Instead of giving operators access to raw data, it gives them proofs. You can verify that a user meets certain conditions without actually seeing the underlying information. That changes the role of the operator quite a bit. Because if you never collect sensitive data in the first place, you’re not just improving privacy — you’re reducing risk. It starts to feel less like a feature, and more like a structural shift. Then there’s the cost side. Using $night to generate DUST instead of paying directly per transaction makes things more predictable. Not necessarily cheaper, but more stable. And stability is something most chains still struggle with. If you’re trying to run a real application, unpredictable fees are a bigger problem than people think. You can’t plan around chaos. Another part I found interesting is the compliance layer. Instead of forcing everything to be public or hidden, Midnight seems to let operators define what gets revealed, and to whom. That sounds simple, but it’s actually closer to how real systems work outside crypto. Not everything is meant for everyone. At the same time, this introduces a different kind of complexity. Now operators have to make decisions about data governance, not just deployment. And that’s not purely technical anymore — it involves legal and operational thinking too. There’s also the question of how all of this behaves under real load. Things like DUST sponsorship or capacity planning look clean in theory, but real usage tends to surface edge cases quickly. Still, what makes this interesting to me is the direction. It doesn’t feel like Midnight is just adding features to the existing model. It feels more like it’s trying to remove the reasons why businesses hesitate to use blockchain in the first place. Whether that actually works at scale is still an open question. But the design at least feels like it’s aiming at the right problems. #night $NIGHT @MidnightNetwork
I almost dismissed TokenTable… until I realized it’s not really about tokens
At first I thought this was just another vesting or distribution tool in the $SIGN stack. We’ve seen a lot of those already. Lockups, schedules, streaming… nothing that surprising. But reading deeper, I think I was looking at the wrong layer. TokenTable isn’t really trying to solve tokenization. It’s trying to solve what happens after the token exists. Who gets it, when they get it, under what conditions, and more importantly… how you prove all of that in a way that doesn’t need to be reconstructed later. That shift feels subtle at first, but it changes a lot. The idea of an allocation manifest is what made it click for me. Instead of executing distributions and then figuring out reporting afterward, everything is defined upfront in a structured way. Who receives what, under which rules, at what time. And once that’s set, execution just follows it deterministically. So the “truth” of the distribution already exists before anything is sent. And then the audit part isn’t something layered on top. It’s produced as part of the process itself. Every distribution event carries proof that eligibility was checked, that rules were followed, and that outcomes match the original plan. Not logs you piece together later, but evidence generated in real time. This is where it starts to connect back to Sign Protocol in a more meaningful way. Each of those actions becomes an attestation. And if multiple programs, even across different organizations, are all generating these standardized pieces of evidence, then suddenly things like auditing or reconciliation don’t have to be siloed anymore. At least in theory. What also stood out to me is how directly this maps to real-world problems. Not abstract inefficiencies, but very specific failure points. Wrong recipients, duplicate payments, distributions happening without proper verification. These aren’t edge cases, they’re common enough at scale. And TokenTable seems designed around preventing those structurally, not just detecting them after. But yeah, I’m still not fully convinced on the adoption side. This kind of system only really matters if institutions actually coordinate around it. And that means shared schemas, shared standards, multiple parties agreeing on how evidence is defined and used. That’s usually the hardest part, not the tech. There’s also the reality that simpler tools already have traction. Even if they don’t solve the audit problem fully, they’re easier to integrate and faster to deploy. Still, I can’t really ignore where Sign is positioning this. Not at the token layer, but at the distribution layer with verifiable evidence baked in from the start. That feels like a part of the stack most people haven’t taken seriously yet. If that layer becomes important, then TokenTable isn’t just a feature. It’s kind of a foundation. I’m still thinking this through. @SignOfficial #SignDigitalSovereignInfra $SIGN
Bitcoin just experienced a rare 2-block chain reorganization at height 941,881 — and it tells a bigger story about where the mining industry is heading.
Foundry USA and AntPool found valid blocks within 12 seconds of each other on Monday. The network briefly split into two competing chains. ViaBTC extended AntPool's side while Foundry extended its own, creating parallel chains two blocks deep.
Then Foundry pulled away. Blocks 941,883 through 941,886 all went to Foundry, making its chain the heaviest. The network reorganized accordingly, orphaning AntPool and ViaBTC's blocks. Those miners earned nothing for their work.
The mechanics here aren't alarming — Bitcoin handled the reorg exactly as designed, with the longer chain winning and consensus re-establishing within minutes. Transactions from orphaned blocks returned to the mempool for inclusion in future blocks.
What's worth watching is the underlying trend. Mining difficulty dropped 7.76% on Saturday, the second-largest negative adjustment of 2026. Hashrate has retreated to roughly 920 EH/s from the 1 zetahash record hit in 2025.
The math is brutal: BTC at $70,000 sits well below the estimated $88,000 average production cost. Smaller and mid-sized miners are shutting down, and every exit concentrates the remaining hashrate into fewer pools.
When fewer pools control more hashrate, the probability of a single pool finding multiple consecutive blocks increases — and with it, the probability of exactly this kind of competing-chain scenario.
No security threat. But a clear signal that mining centralization is no longer theoretical.
Balancer Labs Is Shutting Down — But the Protocol Lives On
Balancer Labs, the company behind one of DeFi's earliest automated market makers, is winding down operations. Founder Fernando Martinelli confirmed the decision on Monday, citing unsustainable costs, zero revenue at the corporate level, and lingering legal exposure from the $116 million exploit back in November.
The numbers tell the story. Balancer peaked at $3.3 billion in TVL during the 2021 bull run. By October 2025, that had already dropped to $800 million. The hack accelerated the decline — $500 million in TVL evaporated within two weeks. Today, Balancer sits at $158 million.
CEO Marcus Hardt acknowledged the core issue: the protocol was spending far more on liquidity incentives than it was earning, diluting BAL holders in the process. Martinelli framed it bluntly — Balancer Labs had become "a liability rather than an asset to the protocol."
The protocol itself isn't dead, though. Both founders are pushing for a transition to the Balancer Foundation and its DAO. The proposed restructuring includes cutting BAL emissions to zero, redirecting fees to the DAO treasury, reducing headcount, and slashing operating costs.
Martinelli pointed out that Balancer still generated over $1 million in revenue across the past three months. "The problem isn't that Balancer doesn't work," he said. "The problem is that the economics around Balancer aren't working. Those are fixable."
Two governance proposals are now live for DAO members to vote on — one covering operational restructuring, the other a full BAL tokenomics revamp.
A textbook case of why DeFi protocols need sustainable economics from the start, not just TVL growth.
Prediction Markets Get Their Own VC Fund — Backed by Polymarket and Kalshi CEOs
A new venture capital firm called 5c(c) Capital is launching with a singular focus: prediction markets. The fund carries the backing of Polymarket founder Shayne Coplan and Kalshi co-founder Tarek Mansour, and is targeting a $35 million raise to deploy across roughly 20 early-stage companies over two years.
The name itself is a nod to the section of the Commodity Exchange Act that governs prediction markets — a clear signal of where the founders see this space heading.
Rather than funding more exchanges, 5c(c) Capital is going after the infrastructure layer: data tools, liquidity services, and compliance systems that support the broader prediction market ecosystem. The thesis is straightforward — as platforms like Polymarket and Kalshi expand, the surrounding toolkit needs to keep up.
The timing is hard to argue with. Since the U.S. presidential election cycle, prediction market volumes have surged. New users are flowing in, and major players including Coinbase, Kraken, and Robinhood have started offering event-based contracts. Polymarket runs on-chain; Kalshi operates within a CFTC-regulated framework. Both models are attracting serious attention.
More than 20 early investors have already committed, including a Millennium Management portfolio manager and founders from rival platforms like PredictIt. The fund positions itself as potentially the first VC vehicle built specifically around the regulatory and market structure of prediction markets.
The broader takeaway: prediction markets are no longer a niche experiment. Capital is now flowing not just into the platforms themselves, but into the second- and third-order businesses they create.
I tried to visualize what 100,000 TPS actually feels like, not just how it sounds
When I first read about $SIGN targeting 100,000 TPS, it didn’t really hit me. Numbers in crypto are always big, so it’s easy to just scroll past them. But then I paused and tried to think about it in a more real-world way.
If this is meant for a national payment system, then it’s not about peak performance, it’s about consistency under pressure. Millions of people transacting at the same time, everyday payments, subsidies, transfers… all happening without delay. In that context, 100,000 TPS starts to feel less like an ambitious target and more like a baseline requirement.
What caught my attention more was the “immediate finality” part. Not waiting for confirmations, not probabilistic settlement. Once a transaction is committed, it’s done. That’s a very different expectation compared to most public chains, and probably much closer to how existing financial systems operate.
The rest of the design also feels very intentional. Identity tied to certificates, ISO standards built in from the start, and different privacy modes depending on whether it’s wholesale or retail. It doesn’t read like something trying to attract developers, it reads like something trying to pass institutional scrutiny.
I’m not sure how all of this performs outside of controlled environments, but the direction is clear. This isn’t about experimenting with blockchain anymore, it’s about making it fit into systems that already exist.
The Solana Foundation just dropped a major play for institutional adoption — and this time, the pitch is privacy.
In a report titled "Privacy on Solana: A Full-Spectrum Approach for the Modern Enterprise," the foundation argues that the next wave of crypto adoption hinges on giving companies control over what they disclose and to whom. Not transparency for its own sake, but selective transparency as a design choice.
The core idea: privacy is a spectrum, not a switch. The report lays out four modes — pseudonymity, confidentiality, anonymity, and fully private systems — each suited to different enterprise needs. Think encrypted order books, private credit risk calculations, or payroll processing that doesn't broadcast salaries on-chain.
Solana's speed is the enabler here. High throughput and low latency make advanced privacy techniques like zero-knowledge proofs and multiparty computation run at near-web speeds, rather than remaining theoretical.
Critically, the foundation positions privacy and regulation as compatible. Features like auditor keys let designated parties decrypt transactions when required. Wallets can prove compliance status without revealing identity. The framework is built so enterprises can mix and match privacy tools while staying within regulatory boundaries.
"Customers expect it and applications require it," the report states. "On Solana, you choose your privacy level — from encrypted balances to zero-knowledge anonymity to multiparty confidential computing."
This is a calculated shift. Solana is no longer just competing on speed and fees. It's positioning itself as the chain where institutions can actually operate under real-world compliance constraints — without sacrificing the benefits of public blockchain infrastructure.
Backpack Exchange Launches BP Token on Solana — 25% Airdrop, Zero Insider Allocation
Backpack Exchange just rolled out its native token, BP, with a total supply of 1 billion. The standout move: 250 million tokens (25%) go directly to users at launch — mostly through an airdrop to existing points program participants and Mad Lads NFT holders.
No tokens were allocated to founders, team members, or investors at inception. That is a deliberate break from the typical exchange token playbook, where insiders usually hold significant portions from day one.
The remaining 75% follows a structured unlock schedule: • 37.5% unlocks over time, tied to operational milestones like market expansion and product launches • 37.5% stays locked in a corporate treasury until after a potential IPO
Backpack also introduced a mechanism allowing long-term stakers to convert BP into company equity — effectively linking the token to real ownership in the firm, not just trading incentives or governance rights.
Worth noting: Backpack was founded by former FTX and Alameda Research employees. After early scrutiny following the FTX collapse, the company acquired FTX's European arm and relaunched it as Backpack EU, pushing into regulated markets.
A user-first token structure with equity conversion and no insider allocation at launch — rare in this space.
I thought TokenTable was just another vesting tool… but it’s not really about vesting
At first glance, I almost skipped over this part of $SIGN . TokenTable sounded like something I’d seen before. Vesting, allocations, distribution logic… nothing new on the surface. But the more I read, the more I realized I was looking at it from the wrong angle. It’s not really about creating tokens or even managing them. It’s about what happens after. Who gets what, when they get it, and more importantly… how you prove that it was done correctly. And that last part is where things started to click for me. Most systems today seem to treat distribution and auditing as two separate steps. First you execute, then later you try to reconstruct what happened for reporting. TokenTable feels like it flips that. The distribution is defined upfront through something like an allocation manifest, and everything that happens afterward just follows that structure. So instead of asking “what happened?” after the fact, you already have a verifiable record of what was supposed to happen. I might be simplifying it, but that shift feels bigger than it sounds. What stood out even more is how the audit layer isn’t added later. It’s built into the process itself. Every distribution event comes with proof that eligibility was checked, rules were followed, and outcomes match the original plan. Not reconstructed logs, but actual evidence generated as part of execution. And then this connects back into Sign Protocol. Which kind of makes the whole thing feel less like a standalone tool and more like one piece of a larger system. Every action becomes an attestation. Everything is queryable. Different programs, even across different organizations, can theoretically rely on the same structure of evidence. That’s the part I keep thinking about. If multiple systems are generating compatible “proofs” of what they’re doing, then things like cross-agency audits or reconciliation don’t have to be manual anymore. At least in theory. Also, the scale of the problem here is not small. Government transfers alone are massive, and a lot of inefficiency seems to come from exactly these gaps. Wrong recipients, duplicated payments, lack of verification before funds go out. TokenTable seems designed directly around those failure points, not just around token mechanics. But yeah, I still have some doubts. A lot of this only really matters if institutions actually use it. And that means coordination between different departments, agreeing on shared formats, changing existing processes… which is usually the hardest part. There’s also the reality that simpler tools already exist and have developer traction. Even if they don’t solve the audit side fully, they’re easier to adopt. Still, I can’t shake the feeling that Sign is building in a part of the stack most people ignore. Not the creation of assets, but the distribution layer with verifiable evidence baked in. If that piece becomes important, this starts to look very different. I’m still watching how this develops. @SignOfficial #SignDigitalSovereignInfra $SIGN
I tried to picture what 100,000 TPS actually means in practice
When I first saw the number in $SIGN ’s CBDC architecture, it didn’t really register. 100,000 TPS sounds impressive, but also kind of abstract. Crypto throws around big numbers all the time.
But then I thought about it differently. If you’re talking about a national payment system, not just a chain with traders and bots, that throughput isn’t just a flex. It’s basically the difference between something that “works in theory” and something that can actually handle millions of people using it at the same time.
What made me pause more wasn’t even the TPS itself, it was the “immediate finality” part. No waiting, no probabilistic settlement. Once it’s done, it’s done. For retail payments at a national level, that feels like a requirement, not an upgrade.
The rest of the stack also feels very… institutional. Identity tied through certificates, ISO standards baked in, different privacy modes depending on whether it’s wholesale or retail. It doesn’t read like typical crypto infra, more like something designed to fit into how central banks already operate.
I’m not deep enough technically to judge how realistic all of this is at scale, but the direction is clear. This isn’t built for experimentation, it’s built for deployment.
Strategy (MSTR) just rolled out a fresh $42 billion at-the-market equity program — $21B in common stock and $21B in its new STRC preferred series. On top of that, a separate $2.1B ATM for STRK preferred stock replaces the prior program.
The company also expanded its sales syndicate to 19 agents, adding Moelis & Company, A.G.P./Alliance Global Partners, and StoneX Financial. More intermediaries means smoother, more gradual capital deployment into the market.
As of March 22, Strategy still had roughly $30B in remaining capacity across existing programs: ~$6.24B in common stock, $1.98B in STRC, $20.33B in STRK, and $1.62B in STRF.
Last week, the company bought another 1,031 BTC, pushing total holdings to 762,099 coins. Shares are modestly green on Monday with BTC trading just above $71,300.
The playbook hasn't changed — Saylor keeps stacking. The scale of the capital raise, though, signals that the appetite for bitcoin exposure through equity markets is far from exhausted.