POKT Network – Rolling into the Modular Future of the Protocol (a technical deep dive)

Pokt Logo
POKT Network
Jan 2, 2024
POKT Network – Rolling into the Modular Future of the Protocol (a technical deep dive)

TL;DR POKT Network’s core protocol team has decided to implement the first version of the next iteration of the protocol (v1) as a “micro-rollup” using Rollkit with Celestia as a Data Availability (DA) layer.

You can read the announcement by @MatthewRossi, our Head of Product (Protocol), on behalf of the Pocket Network Foundation, here to skip the technical history and jump straight to the point.

🛑 Taking a Pause to Forge Ahead 🛑

In early August, the POKT Network (or Pocket Network) core protocol team made a crucial decision: we temporarily halted all development on the upcoming version of the POKT Protocol: the Shannon upgrade (formerly referred to as v1). We embarked on a 3-week intensive research sprint to evaluate rapid advancements of other projects in the industry and ensure an optimal path forward.

Our research period was a mix of open exploration, daily research syncs lasting 1-2 hours throughout the whole period, in-person discussions at a three-day offsite, extensive deep dives into at least nine projects that particularly caught our attention, calls with core teams from many of these projects, input from industry experts, and the development of several Proof-of-Concepts (POCs) along the way.

To get to the punch line, we’re going to implement the Shannon upgrade as a micro-rollup using Rollkit’s modular framework that features Celestia as the Data Availability (DA) layer.

This post will delve into some of POKT Network’s history, explain how we got here, re-state the core problems the protocol solves, and detail the requirements and considerations that eventually led us to this final decision.

We’ll also touch on some trending topics, including Data Availability and various Proof methodologies. Though these topics guided our direction, it’s essential to note that our decisions were primarily steered by our medium-term product requirements so we can hone in our utility in creating the most powerful decentralized RPC protocol out there.

What is POKT Network❓

POKT Network is an open protocol that offers developers Reliable, Performant, and Cost-effective RPC access to the open internet. It’s the decentralized counterpart of your favourite neighborhood centralized node provider, designed to address the RPC Trilemma.

Most centralized “nodes as a service” providers excel in offering Performant and Cost-effective solutions, but what they lack is Reliable multi-tenant decentralization that can only be achieved via token-based incentivization.

POKT Network’s native cryptocurrency, POKT, facilitates access to a wide network of providers supporting more than 42 blockchain RPC endpoints. POKT incentivizes full nodes (non-validators) to exist, pessimistically validate state transitions, replicate data, and make it accessible to developers. It’s analogous to a universal API token, granting access to diverse RPC services from multiple providers.

The subsequent diagram (Fig 1) captures the interaction and relation among the various entities within POKT Network and its functionality. Stick around as we later delve into how POKT Protocol has operated as a pseudo-rollup since its inception! 🥕

Fig 1. Protocol actor integration post Shannon upgrade

Common questions we usually get are related to POKT tokenomics or our strategies for ensuring high QoS. However, here’s a quick breakdown of our core utility in 4 bullet points:

  • Applications remunerate in POKT for RPC service access
  • Servicers receive POKT in return for providing RPC services
  • Servicers submit claims reflecting their service volume
  • Validators verify proofs corresponding to the claimed service volume

Morse (v0️⃣) – Navigating Mainnet for more than 3 years

The genesis of POKT Network began in 2017 with the founding team: Michael (currently CEO of Grove), Luis, and Andrew. Back when a new Initial Coin Offering (ICO) would break the previous days record; when plasma and state channels were all the rage, they sought ways of designing POKT Protocol as a smart-contract based protocol. Concepts like rollups, modularity, data availability, and even Ethereum as a Proof of Stake chain were mere visions on some distant horizon.


However, upon realizing the high costs of smart contracts and recognizing the constraints and nascent state of scaling solutions, the team shipped POKT Protocol as its own sovereign L1 “app-chain” utilizing a Tendermint fork.

In three years, POKT Network  has grown to all-time highs of more than 2B daily relays across dozens of blockchains. One of the coolest achievements is also the fact that none of the nodes currently on the network are run by the corporation that paid for developing the original protocol – which speaks to the substantial growth and strength of the POKT Network community, ecosystem, and 60+ person DAO.


From Morse to Shannon : Navigating Constraints, Legacy & Insights

There is a lot of historical context leading to the decision to fork Tendermint in early 2020 (after beginning to design POKT Protocol in 2017), leading to various downstream decisions:

POKT Network’s first release, Morse (formerly known as v0), encountered a handful of chain halts, has had various scalability limitations, and due to our extensive modifications of Tendermint Core – has not kept up to date with Main.

Some of the problems we encountered included:

  1. Operational resource constraints: The team was heavily constrained on time & capital, but managed to design & ship the first decentralized RPC protocol.
  2. Tendermint Security Enhancements: We added randomness into the proposer leader selection algorithm, shielding against potential DDoS attacks. This was not part of Tendermint Core at the time.
  3. Tendermint Scaling: Our network pushed Tendermint’s validator set to its limit. We had more than 20,000 individual validators at one point, double Tendermint’s 10,000 validator limit, which ultimately led to a chain halt. As a result, we set a hard cap of 1,000 validators based on POKT stake weight.
  4. Scaling and Performance: We hit a few performance snags as POKT Network grew. Amino serialization, IAVL operations, and the aforementioned validator limitation. We switched to protobufs (aligning with mainline Tendermint) and found workarounds for the IAVL issues.
  5. Pocket Proofs & Sizes: One of the biggest bottlenecks to making Pocket completely scalable and permissionless is the size of the relay proofs and time to compute/verify them.

In late 2021, almost eighteen months after the genesis launch in 2020, with a growing network, a committed node runner community, and an influx of new capital, it felt right to start mapping out the Shannon upgrade specifications. As most of what has come into existence in the last 12 months (DA layers, rollup frameworks, etc…) did not exist at the time, the initial roadmap embraced Hotstuff consensus, a scalable P2P layer, and new ideas we couldn’t implement in the existing implementation of the network as we eyed crafting a top-tier L1 from the ground up.

Our development was open source since day 1 as we embarked on this ambitious task. As the core team grew and changed, the community became really active, finding solutions to scale the existing protocol on our forums. External contributors came to help us BUIDL, and we all know what happened to the markets…

As we kept building everything from the ground up, a lot of things started changing in the industry. Vitalik’s Rollup-Centric Ethereum Roadmap in the EVM ecosystem started coming to fruition, Tendermint was rebranded to CometBFT, and development picked up again. What started as a note on Ethereum’s Research Github, became an industry revolving around Data Availability, and started a movement around Modularity.

Having tooling, DA layers, settlement layers and interoperability protocols today enables us to delegate the responsibilities of some parts of a blockchain and focus on our utility instead.

Shannon (v1️⃣) – The Pivot 📐

More than two years after designing the new specifications for a custom L1 blockchain, and dedicating 1.5 years to R&D in our OSS repository, we’ve seen our highs and lows. The biggest realization? The need to focus on our utility, embrace modularity, and iterate on new features; a lesson every engineering org re-learns sooner or later.

Our decisions were always guided by first principles and steadfastly committed to the KISS and DRY methodologies. We’re proud of our documentation and the breadth of external contributors who joined along the way. Yet, over time, our development focus shifted towards the blockchain-back-end components of consensus, p2p, and persistence and away from our primary utility of decentralized RPC. Our ambitious move to design a complex system from the outset conflicted with Gall’s law, one of the pivotal Hacker Laws:

‘A complex system that works is invariably found to have evolved from a simple system that worked. A complex system designed from scratch never works and cannot be patched up to make it work. You have to start over with a working simple system.’

As we edged closer to launching the Shannon TestNet, we recognized that things were getting way more complex than we originally intended – a major driver in the decision to pivot to Rollkit was to get back to basics. Despite our vibrant community and committed node runners, we felt isolated as a protocol, navigating through the Ether in a separate Cosmos from leading ecosystems.

Why now ⁉️

Our team regularly reassesses our trajectory after every major conference, event or significant industry innovation.

We also kept an eye out on other projects to understand what’s actually being done instead of what’s being advertised and marketed.

ProjectPreviousCurrentHeliumCustom L1SolanaRenderEthereum → PolygonSolanaLivepeerEthereumArbitrumThe GraphEthereumArbitrumNucypherEthereumPolygonCeloL1OptimismFarcasterEthereumOptimismWorldcoinEthereumOptimism

While we were making lots of progress towards the Shannon TestNet, we also realized that there’s never a right time to Pivot and explore other options – the classic “sunk-cost fallacy” at work, so we made the hard call to stop and test our base assumptions and approach. In the worst case, we’d lose three weeks of development time, learn a lot along the way, and pick up where we left off afterwards.

POKT Protocol Requirements & Evaluation Criteria ⚖️

Identifying exact requirements and clearly understanding our needs forms the foundation of any good research spike. However, articulating and defining these is more challenging than it sounds. We began with initial guidelines, and as we progressed, these evolved, raised new questions, became clearer, and re-raised further questions. One of our protocol engineers, Bryan, put it very well when he phrased it as needing to understand the “degrees of freedom” we can operate around. I think it’s fair to say that everyone on the team self-reflected on both POKT Network’s direction and the broader implications for blockchains sooner or later.

Ultimately, our functional requirements could be summarized as follows:

  1. 🚪Permissionless Actors: Any actor, whether an Application, Servicer, Gateway, or Watcher, must be able to freely join or exit by staking or un-staking POKT.
  2. 💁‍♂️ Servicer Selection: Servicers should be chosen pseudorandomly, ensuring a fair distribution of work. This selection should factor in POKT stake weight, the quality of service, and the RPC services provided.
  3. 📞 Relay Request: Applications must be equipped to utilize the network, enabling Servicers to relay both read and write requests to any stateless service or any accessible data source through a unique RPC endpoint.
  4. 🧾 ⛏️Mine Relays: Applications must be rate-limited by Servicers based on POKT stakes. Later, POKT must be minted and burned based on the quality and volume of relays completed.
  5. 📈 Scalable Relay Volume: It is crucial for the network to be scalable, accommodating the anticipated growth in actors, sessions, chains, and relays in the foreseeable future.
  6. 🎛 Customizable Rewards & Cost Systems: Both the cost and reward structures should be adjustable based on the service/chain and reflect real-world compute usage.
  7. 🫥 Transparent Data Management: All records, service quality metrics, and other aggregated metadata should be openly available, verifiable, and stored on a decentralized ledger.
  8. 🌐 Interoperable: Compatibility at protocol, community, and tooling levels is essential, especially with major ecosystems like IBC and EVM.

Mapping a requirement to a “blockchain layer” can be considered a pseudoscience, but I think even an attempt can help build context. In the following table, the Incentive Layer can be treated as the design of the actual application, and the Execution Layer is what manages the state transition functions.

Some additional non-functional requirements, applicable to any team (engineering or not), are to move fast, iterate, focus on building what provides value to our users and to quote Albert Einstein: Everything should be made as simple as possible, but not simpler.

Hearing & Comparing 👂

With our requirements at hand, we started evaluating what’s out there.

We looked into the Cosmos Ecosystem (CometBFT, Cosmos SDK, Celestia), the EVM Ecosystem (Arbitrum, Optimism, Polygon, Aztec), compared ourselves against similar Smart Contract projects (LivePeer, TheGraph, Streamr), evaluated some Rollup-as-a-Service options (Caldera and Conduit) and also reviewed a few other projects that didn’t fit any of these buckets like Polygon Edge (Side-Chain) and Storj (Decentralized Data Storage).

Some of our internal and external discussions often had the following phrases come up over and over:

However, seeing how supportive everyone we spoke to is of other projects was very refreshing. At the end of the day, we’re all working towards the same vision:

  • “Their team has brilliant & hard working individuals.”
  • “Their team is providing us with all the support we need asap.”
  • “They have great ideas and making a ton of progress.”
  • “That framework is very easy to use.”
  • “Their docs are very educational & informative.”

We realized that there wouldn’t be a clear and definitive answer here, and given the need to examine the ecosystem benefits of choosing one opportunity over another too, it quickly became apparent that the set of tradeoffs resemble art more so than science.

Defining Ambiguity in the Search for Data Availability

Partway through our research spike, we had the idea of doing a “literature review”, like a document evaluating all the solutions that either exist or are under active development: smart contracts, L1, L2, L3, various DA layers, various rollup frameworks, etc… However, it is hard to draw a straightforward comparison. Not only would this be a huge undertaking, but it is very nuanced. We need to compare what exists today vs what will be available tomorrow. What are our timelines? What level of security risk are we willing to take? How much of a tradeoff will we take in exchange for the true long-term web3 ethos of complete sovereignty, permissionlessness and decentralization? What does settlement really mean? What ecosystem benefits do we expect? And what is the opportunity cost for choosing one option over another?

Settlement is very similar to 🎂. It can be delicious and bring joy, but comes with costs and tradeoffs since it’s not the healthiest food group. It could be a piece of cake if you leverage an existing framework, but could be very burdensome if you bake it from scratch. Some even question whether Settlement is real at all, or if it’s just a promise from GLaDOS. Choose your 🍰

The team drew inspiration from the posts by @jon_charb discussing how Rollups Aren’t real, and how Rollup’s Actually Work to understand the caveats and ideologies around these topics. Understanding that nuance, reading through Optimism’s amazing documentation on EVM rollups, and comparing that to documentation from Sovereign Labs and Celestia on sovereign rollups can build a really good conceptual model that I won’t try to re-explain here.

Source: Kelvin Fichter, How Rollups actually work

There are endless discussions around building truly decentralized and permissionless systems to tackle the Blockchain Scalability Trilemma. If blockchain is really a science, then we are very early in the process of discovering what’s true based on Alin Tomescu’s definition:

Science is a process that we engage in to discover truths. And it often leads us astray 3,4, which is why the idea of “science being true” is at best misleading and at worst dangerous.

Threads like this on X do a good job at capturing the idea. L1s and L2s provide applications with consensus, data availability and have various tradeoffs when it comes to latency, cost and security, but those details are outside the scope of this post.


A few common themes kept coming up when we spoke to different teams and individuals:

  • Everyone is starting with Fraud (optimistic) proofs and is aiming to have Validity (zk) proofs in the future
  • Everyone is starting with a single centralized sequencer (i.e. a fancy term for a web2 server) and aiming to decentralize it over time
  • Lots of teams are working on making validity proofs cheaper & more efficient to generate and verify in the future
  • Every sovereign rollup framework aims to have a configurable DA layer at some point in the future
  • There are lots of ways (e.g. bridging, etc…) to achieve settlement (whatever that means) and interoperability with other chains
  • It’s too early to have an analytical comparison of data availability costs across DA layers, sequencer speeds and decentralization, or proof cost and sizes. These will have impact across various factors including the censorship resistance of the underlying DA, its ability to enable trustless bridging, etc…

We put together an overview table of our deep dive findings.

Click on the image below or on this link to access it.

Why not the EVM?

When the team started its research spike, we were heavily focused on the OP Stack due to the signal in the industry and our bullish view on the EVM ecosystem, its developer community, security and interoperability. We managed to rule out EVM-based solutions after a little bit of work and a lot of great ideas from h5law, one of our lead protocol engineers.

We quickly realized that we might need to be an OP Stack Hack, but it would result in our new chain not being part of the official OP stack and, therefore, losing many of the ecosystem benefits we sought to obtain by joining the Superchain. We also discovered that we don’t need to fork the OP Stack because Pocket is an application-specific chain, and we do not need to be a standalone smart contract EVM rollup. This got us excited by the idea that Pocket could become a Smart Contract that acts as a registry for our on-chain actors deployed on one of the major EVM rollups similar to The Graph and LivePeer.

The team quickly put together a smart contract that captures the bare-bones functionality (i.e. parity with the live network) of what we need.

And then found itself having discussion around how to optimize for gas.

We realized that we’ll be at the mercy of various gas optimization discussions for our relay verification mechanism, while also being constrained by L2 to Ethereum block compression optimizations, and potential gas spikes in turbulent markets. That’s when Vitalik’s Ethereum’s rollup centric roadmap from a few years back became much evident, and when this line from the Optimism docs really started ringing (pun intended) true:

We need EVM compatibility not for the protocol’s core utility but to improve the experience related to wallets, our DAO UX, and access to best-in-class DeFi infra and token liquidity. This was an incredibly important component of why we sought to explore rollups, particularly the OP stack, in the first place. But if we can’t deploy directly where the wallet, DAO or DeFi infra we want to integrate with currently exists – ie on ETH L1, the OP stack, Arbitrum or Polygon etc – without needing to use a bridge; building outside of the EVM ecosystem can get us to a pretty similar place in the end without needing to be “EVM Native” if sufficiently robust and well-supported bridges like Axelar and Hyperlane support our network.

POKT Protocol has always been a Rollup 🧻

One of our key findings during this exercise, first mentioned by Dylan, is the realization that POKT Protocol is, and always has been, a pseudo-rollup.

Needing to handle more than a billion requests a day, with the goal of having the network handle a magnitude more within the next few years, it’s unfathomable (and unnecessary) to have every request be its own independent on-chain transaction. Having an on-chain transaction called Relay can be ruled out immediately.

Off-chain, Applications make thousands or millions of RPC requests to a Servicer during a session. Once the session ends, the Servicer uses a basic commit-and-reveal scheme that requires it to reveal one random branch in the tree. In other words, servicers rollup all the RPC requests over some time (e.g. 1 session = 1 hour) and post a claim for it on-chain. Later, they are requested to provide a pseudorandom proof that cannot be predicted ahead of time. It can be thought of as a hybrid between an optimistic and a validity proof with different probabilistic error guarantees, but those details are outside the scope of this specific post. The Relay Mining Paper and Shannon specifications go into it in greater depth for those interested in learning more.

A few of the key things in realizing the above is understanding where Pocket stands:

  • Scalability – Most of Pocket’s business logic happens off chain so our scalability is less-so limited by the number of transactions that we need to have on chain.
  • Latency – Accounting for POKT rewards distributions, slashes, POKT transfers and staking updates do not necessarily require sub-minute commitment, so longer block times are okay.

In POKT Network’s utility, the blockchain is used as a distributed registry (i.e. ledger) to transparently track actor properties, optimistically rate limit applications, verify the work done and keep permissionless actors honest through (dis)incentives. We have the leisure of not needing to optimize for milliseconds.

Committing to a Decision 🧑‍⚖️

We’ve already started prototyping an alpha version of the Shannon upgrade as a rollup using Rollkit and Celestia as a DA layer, and are committed to bringing this to mainnet early next year.

We had a few “runners ups” and wanted to share the reasons that led to our final decision. After going through the research above, identifying our requirements, and ruling out the need to be “EVM Native”, our top 3 choices were

  1. Rollkit; leveraging the Cosmos SDK and Celestia DA (configurable)
  2. Cosmos; leveraging the Cosmos SDK and CometBFT
  3. Sovereign; Leveraging the Sovereign SDK and Celestia DA (configurable)

These three solutions would cover all of our functional requirements, but in terms of non-functional requirements, we also took the following into account:

  • Framework Maturity – Though Cosmos has a lot of encapsulated complexity, it is also a very mature framework that provides tons of tooling out of the box.
  • Support – All three of these options have very supportive teams that have been invaluable in answering all our questions.
  • Risk Mitigation – If we realize that, for some reason, we need to quickly go back to being an L1, Rollkit makes that much easier as it is compliant with the ABCI interface and lets us fallback to Cosmos.
  • Interoperability – We filtered for options that are compatible with IBC out of the box, which also enables us to leverage projects like Axelar to bridge into the EVM ecosystem.
  • Simplicity – While we’re not too focused on the programming language, we did take into consideration that the team has experience in Go, and Rust has a slightly higher learning curve, which adds some complexity.
  • Time to Market – We biased towards more mature networks and frameworks to shorten the timeline for us to reach Mainnet.
  • External Contributions – Using a well known framework (i.e. Cosmos) would welcome more external contributors to build submodules specific to Pocket’s utility.

I anticipate that over the next 3-9 months, as these frameworks and DA layers mature, there will be a lot more visibility into the costs, scalability, security and decentralization of each solution. For now, we’re making the best decision we can given the data we have.

So What’s Happening to the Validators?

POKT Network was always committed to being the most decentralized, trustless and censorship-resistant protocol out there. By moving to a rollup, we’re delegating the security and decentralization to the underlying Data Availability layer, which could raise risk with respect to censorship resistance. However, modularity means that POKT Protocol could move from one DA layer to another, use multiple DA layers in parallel for redundancy, and eventually leverage its network of nodes for its own DA layer. It also moves the cost of the validators (on and off chain) to the underlying Data Availability layer.

Some of the the leading projects in the space, with very talented and hardworking teams, are working on solving the problems related to censorship resistance, DDoS attacks, trustless bridging, decentralizing sequencers, and many others while also making sure that these solutions scale. Being able to delegate that responsibility to a project for whom it is their top priority, allows us to focus on our core value add – Portals, Applications & Servicers and Relay Mining, rather than Validators & Sequencers.

Instead of scaling validators, we can scale the number of relays that the network can handle. Instead of researching proof mechanisms for state transitions, we can focus on proving different types of relays (gRPC, websockets, etc…) and their quality of service, on-chain. Instead of designing permissionless sequencers, we can design trust and tokenomic mechanisms around permissionless portals (i.e. gateways). We have plenty of other ideas and are hoping to tightly collaborate with the teams behind the projects that we’ll be working with.

This shift in work lets us focus solely on POKT Network’s utility. We can nail down the core mechanics specific to the POKT Network, such as the claim & proof lifecycle to validate relays completed, and spend more time listening to the community on how to expand other features over time. For example, we have heard, ideated and experimented with dozens of different ideas related to Quality of Service, but bringing them on-chain takes time and requires finding concrete solutions to various nuances. Offloading core blockchain building blocks to other frameworks in a modular fashion lets us be more productive and collaborate more closely with our community on our specialization. Similar to how Portal (by Valve Corp.) was a great game because it nailed one key game mechanic, we can do the same with POKT Network’s utility. In fact, tokenizing gateways (now known as portals) was an idea first brought up by one of our community members, Shane.

Special Thanks 🙏

Shoutout to Matteo for the meticulous edits as well as the protocol team for related discussion and review: Bryan, Dylan, Harry, Dima, Rampey and Red0ne.

Special thanks to Valeriy at 1kx, Adi at Informal Systems, Cem at Sovereign Labs and Kautuk at Stackr who jumped on calls, provided a ton of support along the way, and helped review this post.

If you liked this content and want to read more from me, you’ll find my substack here.


I assume some people might ask us “What should our team choose?” This isn’t an easy question, but below is an opinionated decision tree that does not take lots of things (tooling, maturity, interoperability, security, etc…) into account, but can hopefully help as a rough starting point.

Thank you! Your submission has been received!
Oops something went wrong.....