Outcome Based Funding for Web3 Popups: GG24 sensemaking report

Written by @nidiyia with help from @JamesFarrell and Devansh Mehta

this is a deprecated version; latest version below

https://gov.gitcoin.co/t/outcome-based-funding-for-web3-popups-gg24-sensemaking-report/23054/17

TLDR;

Build a transparent & impact driven funding model for Web3 popups & residencies where each program is documented, evaluated & rewarded based on its contribution to Ethereum

Require popup cities & residencies to publish Hypercerts with clear costs & documented outputs. Calculate an impact score for each program to direct funds in proportion to impact generated.

Problem & Impact

Web3 popup residencies, ephemeral, IRL gatherings like Zuzalu, Edge City, and ZuBerlin, have rapidly become catalytic hubs for Ethereum’s ecosystem. They co-locate builders, researchers, and creatives, fostering innovation, cross-pollination, and collaboration across disciplines, essential for a remote first industry.

Yet despite their growing influence, these gatherings face systemic friction: lack of transparent budgeting, diffuse impact measurement, and opaque coordination, issues that directly hinder Ethereum’s core goals of efficient capital deployment, ecosystem resilience, and reputation.

The need to sustainably fund web popups & residencies is urgent because:

  1. Ethereum popup funding is under strain

Community building in Ethereum is critical. Web3 popup cities and builder residencies have emerged as valuable physical spaces bringing an internet first, remote industry face-to-face & seeding the next big ideas to move the space forward.

Despite being IRL gatherings, there is no standardized reporting and funders lack insight into how funds are used at these events, making resource optimization and shared learnings difficult. Reports note growing concern over ā€œinconsistent evaluation frameworksā€ and transparency issues in Web3 grants, leading to inefficiencies in matching funding to outcomes (Francis, 2024).

  1. Evidence of opacity

Take Zuzalu (March–May 2023): a two-month popup city in Montenegro conceptualized by Vitalik Buterin. It brought together ~200 residents and up to 800 visitors, yet budget disclosures remain scarce (Bankless Report, 2024). Similarly, residencies like Edge City in Chiang Mai operated as live testing grounds ahead of Devcon but detailed cost breakdowns have not been publicly shared (Francis, 2024). Globally, 2024–25 has seen at least 10 to 15 major Web3 popup residencies, including MU Accra (2025) and ETHiopia but uniform outcome reporting is virtually nonexistent (TechCabal, 2025).

  1. Evaluation of popups is needed

Without transparency, funders may redirect capital away from community-driven innovation, and opt for more measurable, traditional paths. (SSSG, 2025, Binance). This post discusses the use of Hypercerts, on-chain impact certificates with cost and outcome metadata, to enable accountability and verifiability in the popup ecosystem funding. (Li, 2024).

Sensemaking Analysis

My entry into the web3 world was through the popup city Edge Esmeralda, where I spent a month in Healdsburg (April-May 2024). Since then I have participated in the Funding The Commons Residency in Chiang Mai (Oct-Nov 2024) where there was an archipelago of many popups happening simultaneously, ZuBerlin (2025) & also been on the organizing team for the IERR Residency in Iceland that just concluded on August 10, 2025.

Overall, I see popups not just as cultural experiments but as informal infrastructures where Ethereum’s global community tests ideas face-to-face. Yet without transparent costs or clear outcome metrics, they risk misallocating capital and eroding legitimacy

My understanding of the popup landscape draws heavily on each of these lived experiences as both participant & organizer of these events. Drawing on participant interviews, comparative analysis between popups and ecosystem research, I have come to the following conclusions

  • Popups generate disproportionate intangible value (community cohesion, trust, cross-pollination of ideas) but lack mechanisms to prove or measure this.

  • Financial opacity is widespread, most residencies do not publish budgets, leaving funders in the dark.

  • Impact stories circulate informally through whisper networks, but without systematic capture they remain anecdotal, limiting their usefulness for allocation.

The remaining sections will tackle these issues, proposing a structure of hypercerts & benefit-cost scoring of popups that can provide a missing layer of impact legibility. This aligns with Gitcoin’s shift in GG24 toward radically transparent, data-informed funding.

The sensemaking shows that without structured evaluation, Ethereum risks undervaluing some of its most fertile cultural and intellectual spaces, or worse funding them blindly in ways that invite reputational risk

Gitcoin’s Unique Role & Fundraising

Gitcoin has a history of operating multiple Zuzalu-related rounds, demonstrating legitimacy to steward an IRL coordination domain at scale. This domain will build on this history by positioning Gitcoin as the leader converging standards for funding of these popup movement

Specifically, we propose keeping the following eligibility requirements

  1. Must have already hosted a popup or builder residency

  2. Must create a hypercert listing out the costs of their past residency with a transparent breakdown. Amount paid by participants must also be included so we can provide matching on top of what is contributed by attendees

  3. Must list outputs of the popup that can then be quantified to obtain its benefit-cost (BC) ratio. An example of how this might look can be seen here

  4. Must have confirmed location and dates for the next popup that funds will go towards supporting

  5. Must have cofunding from other funders for the next popup so that GG24 funding is less than 30% of total outlay

Success Measurement & Reflection

Six-month outcomes:

  1. Transparency baseline: Onboard 8 to 12 popups or residencies who mint a Hypercert of their past program with total cost and line-item breakdowns. Simply having the finances of past popups become transparent is a win in itself.

  2. Impact legibility: For every residency, ask for a submission of outputs that are attributable to them. Use these to compute a standardized Benefit–Cost (BC) ratio

  3. Funding routed by evidence: Channel funding to purchase hypercerts of these popups, with the support predicated upon receiving an impact evaluation score of their past residency and secured funding and dates for their next edition

At the end of 6 months, publish a Gitcoin-style round report card summarizing allocations, evidence, and learning.

Domain Information

Name : Outcome based Funding for Web3 Popups

Purpose : Residencies and popups that (1) publish transparent costs, (2) issue a Hypercert capturing outputs/outcomes, and (3) undergo impact evaluation (Benefit Cost ratio) so future funding routes toward the highest impact per dollar.

Mechanism : We propose a Dedicated Domain Allocation (DDA) track with appointed stewards that calculate impact scores of past popups, based on which funds get allocated.

Even past popups with a high impact score but unable to lock down dates or secured cofunding for their next edition will be rendered ineligible.

Structure :

  1. Transparency through Hypercerts: each residency or pop-up program issues a Hypercert for their past event. These must include a clear total cost & a breakdown of total costs. This ensures financial transparency, which is currently lacking, and creates a verifiable record of funding needs.

  2. Accountability through Impact Measurement: for each residency, we ask them to give deliverables that occurred from their residency. This is then used along with other information to calculate a standardized impact score using a benefit cost ratio calculator. We document and publish all:

  • Quantifiable outcomes (e.g., number of participants, projects launched, protocols developed).
  • Non-quantifiable outcomes (e.g., community cohesion, knowledge transfer, cultural impact).

This provides a balanced view of both tangible and intangible contributions & will feed into the calculation of benefits in the impact score. Anyone can also challenge these scores and give a better calculation of them.

  1. Funding Distribution: funds from the matching pool are distributed programmatically via impact score calculations. Additional cofunding is done by reachouts to potential donors to buy units of the hypercert or contributions by participants at the popup and local citizens who want more such events to be held in their city.

Timeline :

  • Domain vote: Aug 22–29 (Snapshot).

  • Domains announced: Sept 1.

  • GG24 execution: September to January (hypercert onboarding, schema finalization, review committee, secured cofunding and dates for another popup by the applicant).

10 Likes

have we reached out to popup organizers and confirmed they are willing to try this experiment? is their participation contingent on a certain funding amount $$? what is their feedback on the proposal, if any?

7 Likes

I really like this approach. Will share with some of the pop up builders I’m close with to gauge interest though I am pretty sure they will happy to participate.

3 Likes

Thank you for your inputs :pray: I am in the process of reaching out to popup organizers & will get back with collated feedback this week.

<is their participation contingent on a certain funding amount $$?>
No, the only criteria would be for the next popup dates to be locked in & for there to be some co-funding commitment. This is current direction we are thinking in, open to feedback.

1 Like

That’ll be very helpful @Donny_Jerri :pray: thanks! would be nice to stay connected as we piece together the feedback, what’s the best way to stay in touch?

2 Likes

Draft Scorecard

2025/08/18 - Version 0.1.1

By Owocki

Prepared for nidiyia re: ā€œOutcome Based Funding for Web3 Popupsā€

(vibe-researched-and-written by an LLM using this prompt, iterated on, + edited for accuracy quality and legibility by owocki himself.)

Proposal Comprehension

Title: Outcome Based Funding for Web3 Popups: GG24 sensemaking report
Author: nidiyia, with help from James Farrell and Devansh Mehta
Link: https://gov.gitcoin.co/t/outcome-based-funding-for-web3-popups-gg24-sensemaking-report/23054

TLDR

Create a transparent, impact-driven funding model for popup cities and residencies. Require applicants to publish hypercerts with clear cost breakdowns and documented outputs, compute a standardized impact score, and route funding in proportion to verified impact.

Proposers

Primary: nidiyia
Collaborators: James Farrell, Devansh Mehta
Context: author has participant and organizer experience across multiple popups and residencies.

Domain Experts

Not explicitly listed beyond the proposers’ lived experience in Edge Esmeralda, Funding the Commons residency, ZuBerlin, and IERR Residency organizing.

Proposal might benefit from having some popup organizers onboard.

Problem

IRL popup cities and residencies generate real value for Ethereum but suffer from budget opacity, inconsistent outcome reporting, and weak impact legibility. This causes allocation inefficiency and reputational risk.
Background context on popup cities like Zuzalu supports the importance and scale of these gatherings, while public budgets and systematic evaluations remain sparse.

Solution

Stand up a GG24 domain. Eligibility requires: prior popup held, hypercert with total and line-item costs plus attendee contributions, documented outputs to compute a benefit-cost style impact score, confirmed next-edition dates and location, and co-funding so GG24 is under 30 percent of total outlay. Allocate matching by impact scores and publish a round report card at 6 months.

Hypercerts provide the shared data layer for claims, verification, and funding.

Risks

  1. Measurement burden and gaming. Applicants may optimize for the metric rather than durable impact.
  2. Software - who is going to input all of the hypercerts metada? will there be a good enough ux/network effects to make it worthwihle?
  3. Attribution and counterfactuals. Intangible benefits like trust and network formation are hard to apportion across overlapping IRL events.
  4. Data quality variance. Cost line items, participant counts, and output claims may be incomplete or non-standard.
  5. Coordinator capacity. Appointed stewards must stand up schema, review process, and dispute resolution quickly.
  6. Ecosystem uptake risk. Some popup organizers may decline transparency or hypercerts, limiting sample size.
  7. Timeline risk. Achieving meaningful outcomes by October is tight given Sept-Jan execution window.
  8. Buy pressure for the hypercerts, where will it come from?

Outside Funding

Not yet specified. Proposal sets a requirement that applicants secure co-funding so GG24 contributes less than 30 percent of total outlay. We will need confirmation that leading popup organizers have expressed willingness to meet this bar.

Why Gitcoin?

Gitcoin has prior rounds tied to Zuzalu and report-card infrastructure, which gives legitimacy to steward standards for IRL coordination domains and to publish transparent post-round analyses. ([Gitcoin Governance][5])

Owockis scorecard

# Criterion Score(0-2) Notes
1 Problem Focus – Clearly frames a real problem, avoids solutionism 2 Sharp articulation of opacity and impact-legibility gaps in popups that matter to Ethereum.
2 Credible, High-leverage, Evidence-Based Approach 1 Hypercerts plus benefit-cost scoring is credible and standardizable, but evidence of reliability across popups is early. Add evaluator independence and audits.
3 Domain Expertise – Recognized experts involved 1 Strong lived experience from proposers. Would like to see named independent evaluators or advisors with measurement chops and real pop up experience.
4 Co-Funding – Backing beyond Gitcoin 1 Co-funding is a requirement, but no commitments listed yet.
5 Fit-for-Purpose Capital Allocation Method 2 Hypercerts with transparent eligibility, impact scoring, and report cards fits the domain’s epistemology and Gitcoin’s capabilities.
6 Execution Readiness – Results by October 1 Sept-Jan window is workable, but October impact will likely be early outputs: schema, onboarding, initial hypercerts. Delivery risk remains.
7 Other – Vibe and blind spots 1 Positive alignment and standards focus. Risks: metric gaming, inconsistent participation, and attribution noise. Plan risk-mitigation explicitly.

Score

Total Score: 9 / 14
Confidence in score: 70%

Feedback:

Major

  • Get explicit comittments from cofunders.
  • Secure written expressions of interest from 6 to 10 target popup organizers agreeing to publish budgets and mint hypercerts, subject to a clear privacy policy and data schema. Include named targets and status.
  • Specify the evaluation pipeline: who computes scores, independence of evaluators, challenge process, and how intangible benefits are translated into the benefit side of the ratio.

Minor

  • Publish a draft hypercert schema with example field values and two worked examples from past popups.
  • Pre-commit to a public dashboard listing applicants, data completeness, preliminary scores, and co-funding status.
  • Outline fraud and gaming safeguards: random audits, document checks, peer challenges, and consequences for misreporting.

Steel man case for/against:

For

This domain operationalizes what the ecosystem says it wants: transparency, comparability, and retroactive funding based on evidence. Hypercerts plus benefit-cost style scoring can converge fragmented popup experiments into a legible market of impact claims, while Gitcoin’s prior Zuzalu rounds and report-card muscle give the right foundation to set standards.

Against

Impact measurement for IRL popups is messy. Budgets are sensitive, outcomes are diffuse, and attributions overlap. If key organizers refuse transparency or if scores are noisy or gameable, the domain could under-allocate to genuinely catalytic culture and over-allocate to well-documented but lower-leverage events. External narratives about popup opacity persist, and may not be solvable in a single GG cycle.

Rose/ Bud/Thorn

Rose
Clear, standards-first approach with eligibility rules that push the ecosystem toward transparency. The plan to publish report cards and route funding by evidence aligns with Gitcoin’s strengths.

Thorn
Evaluator independence, gaming resistance, and data quality are underspecified. Without organizer pre-commitments to share budgets and outcomes, volume may be too thin to benchmark across popups.

Bud
This could become a new way to solve pop up city funding.

Feedback

Did I miss anything or get anything wrong? FF to let me know in the comments.

Research Notes

  • Open questions: which organizers are pre-committed, evaluator names, challenge procedure, how to score intangible outcomes, and data privacy approach for sensitive budget items.
  • Future diligence: obtain letters of intent from target popups, review two example hypercerts with full cost lines, run a dry-run scoring on a past popup, and pressure-test the scoring rubric with independent evaluators.
1 Like

Love to see the activity on this one…

Evaluated using my steward scorecard — reviewed and iterated manually for consistency, clarity, and alignment with GG24 criteria.

:white_check_mark: Submission Compliance

  • Structured and thorough proposal with problem, sensemaking, success metrics, domain structure
  • Team has direct experience with popups but lacks named reviewers or external anchors
  • No confirmed co-funders or popup organizer commitments listed
  • Mechanism is specific and promising (hypercerts + cost-benefit scoring)
  • Verdict: Compliant, but execution readiness and traction unclear

:bar_chart: Scorecard Evaluation

Total Score: 9 / 14

Criteria Score Notes
Problem Clarity 2 Clear articulation of opacity and coordination issues across popups
Evidence-Based Approach 1 Hypercert + cost-benefit model is strong, but reliability across popups is untested
Domain Expertise 1 Strong lived experience; needs external reviewers and evaluators for rigor
Co-Funding 1 Co-funding required, but not secured yet; no letters of intent shared
Capital Allocation Design 2 Excellent match of method to domain: eligibility, transparency, scoring, and reports
Execution Readiness 1 Sept-Jan plan is viable, but October outputs will be minimal unless organizers are pre-committed
Other (Vibe, Alignment, Blind Spots) 1 Strong ethos and alignment; metric gaming and data quality are key risks

:pushpin: Feedback for Improvement

Where I agree with Owocki:

  • Securing popup organizer participation is critical — get written pre-commitments.
  • Evaluation needs to be independent and audit-friendly — define challenge/appeal process.
  • Publish example hypercert schema and data completeness dashboard to build early trust.

What I’d add:

  • A dry-run hypercert scoring for 1–2 past popups would be invaluable for stress-testing impact metrics.
  • Clarify whether Gitcoin infra or a third party will host dashboards, dispute resolution, and fraud reviews.
  • Consider anonymized financials or tiered transparency if organizers are budget-shy.

:yellow_circle: Conditional Support

Would support this domain if:

  • 5+ popups pre-commit to the proposed schema (or a version of it)
  • At least one co-funder confirms in writing
  • Evaluation committee and scoring guardrails are made public
  • Gitcoin’s role is clearly bounded (infra + distribution, not sole arbiter)

This proposal moves the ecosystem toward impact accountability in a messy but important area. Let’s give it the scaffolding it needs to work.

2 Likes

Hey folks, thank you so much for starting this conversation; this is awesome. It’s really exciting to see people thinking about funding mechanisms at this level, so props to @nidiyia, @JamesFarrell, and Devansh.

I’m happy to share my perspective from building Edge City. I’ll share a couple of quick thoughts below to start.

Sustainability & transparency
For us at Edge City, sustainability is something that we think about a lot. Our events are run through a nonprofit, and every edition is a big lift. They are budgeted to be breakeven through tickets and sponsorships, which can be a lot of work to align. So if there was a mechanism that could help meet up to 30% of our outlay, that would be a huge help.

At our last few gatherings, we’ve run sessions where we’ve shared the open budget with attendees, and people are always surprised at how much it actually takes to make one of these things happen. That’s why we’d be thrilled to see tools that help make popups sustainable in the long run, not just for us but for everyone else experimenting in this space.

Sustainability for everyone
Edge City is increasingly evolving into a network of residencies which happen alongside each other during the months of our popups (but also often have their own communities and other activations through the year). Each one brings its own theme, community, and in some cases, its own funding model. At Edge City Patagonia this fall, we’ll have 10–12 different residencies happening concurrently during the month before Devconnect in a beautiful mountain town in Patagonia; some sponsored by organizations, others completely community-run.

A few are even crowdfunding ahead of time using tools like ante.xyz, which is a great first step towards sustainability and helps solve the collective action problem of everyone feeling ā€œyeah I would totally join that residency if it happens, but I’m not willing to be the first moverā€.

It would be amazing if outcome-based funding could support these residencies as well, because if they’re able to be sustainable and drive impact, that would be great for the whole ecosystem.

Impact & evaluation
We couldn’t agree more that measuring impact is key, but for us, it is also one of the trickiest parts. We can definitely track how many startups, projects, academic papers, etc. get created during the popups (at this point there have been many), but part of the reason why it’s hard to track impact for popups is that often the most meaningful impact shows up months later.

Just the other day someone mentioned to me they’d raised funding for their startup and casually added, ā€œOh, I actually met my cofounder at Edge Esmeralda 2024.ā€ We would have had no idea, and I’m certain there are many more anecdotal cases like this where so much value has been driven through the community work that people are doing in the space, but it’s hard to measure those ripple effects, especially in a lightweight way that doesn’t add more administrative burden to an already stretched small team.

We’re huge fans of mechanisms like Hypercerts, so if there’s a way to plug into a lightweight, standardized approach that helps prove impact without creating a ton of extra work, we’d love to be part of it.

Really grateful to you all for starting this conversation. We’re locked in and will continue to execute, but we see a lot of our peers burning out from the sheer amount of work it takes to do these things well, and it’s a pity because they do drive so much value to the broader ecosystem. I see these gatherings as necessary building blocks for any kind of future network-society, because people need surface area to meet and connect and develop affinity for each other.

Finding ways to make them sustainable feels like something the whole ecosystem will benefit from.

With love,
Timour

9 Likes

thank you @owocki & @deltajuliet for the detailed scorecards. These are all valid points & we are speaking with popup organizers, potential cofunders & working on the rest of your feedback. We will get back with an update soon.

These scorecards themselves are good examples of how we could assess a popup. The idea is to build an LLM-assisted IE system that could generate a similar assessment with a score (BC ratio). I really liked the addition of the ā€œConfidence in the scoreā€, something I would love to add to our own thinking of the impact evaluation process & I’m curious to know how it was calculated @owocki

Also, as an example schema, here is a hypercert that was created for the Impact Evaluator’s Research Retreat that concluded on Aug 10: Impact Evaluator Research Retreat 2025 | VoiceDeck.

2 Likes

Thanks for this proposal on Outcome-Based Funding for Web3 Popups — it’s exciting to see experimentation with temporary, agile structures that can deliver public goods in focused bursts.

In our proposal, we didn’t define a domain but are instead testing whether CollabBerry’s peer-based allocation and accountability tools could serve across domains as complementary mechanisms.

What feels particularly relevant here is the time-bound nature of popups. When a team is formed quickly around a specific goal, it’s often even harder to ensure fair allocation of resources among contributors — because roles are fluid, contributions are diverse, and there’s little time to set up heavy governance. CollabBerry experiments with lightweight peer-to-peer assessment mechanisms that allow contributors to continuously recognize each other’s efforts, generating a reputation layer that directly informs payouts.

We imagine this could complement popup structures by reducing overhead and giving teams a way to distribute funds transparently and fairly within a short lifecycle.

Curious to hear your view: do you see contributor-level peer assessments as a useful fit for outcome-based popup models?

1 Like

I’m excited about your interesting proposal, @thedevanshmehta , @nidiyia , and @JamesFarrell ! Nidhi explained this proposal for me in a more detail.

I’m working on public goods funding, impact evaluation, and evidence-based practice(EBP). And then, I’m a co-organizer of popup village, ā€œkuu villageā€. kuu village would like to join in the round!

In Oct, first kuu village will be held in Nara City(former Tsukigase Village), in Japan. Multiple kuu village will happen all over the world, but first kuu village is still proto-type. I’d like to provide a quick introduction before my feedback for the proposal.

Quick introduction for ā€œkuu villageā€

kuu village is a one-week co-living program where participants immerse themselves in the worldview of ā€œkuuā€ through daily life. It serves as a living laboratory that combines an off-grid lifestyle with Decentralized technologies to simultaneously explore autonomy, collaboration, and circularity. Participants let go of attachment to ownership and hierarchy, co-creating shelter, energy, water, food, and governance through their relationships with others. The insights and outcomes gained from this experience are shared as open source and passed on to the next community.

kuu village is inspired by previous popup village like Zuzalu. However, we more focus on the physical experience and physical practice. According to d/acc, d/acc is including the factor of ā€œatom(physical)ā€, not only ā€œbits(digital)ā€, so we huge focus on physical technologies for resilient societies as well as digital technologies like Ethereum. For example, physical technologies mean that self-building house, digging a well, and more. Of course, we are planning the experiment of onchain governance, onchain reputation, funding allocation for public goods, SALSA NFTs, broad-listening, and more. We aim at a practical and experimental-centric popup village, not discussion-centric one.

The detail is available here: [PUBLIC]kuuvillage_2025 - Google Slides

Feedback for the proposal

I’m interested in this approach in this proposal since I’m huge contributing in evaluation outcomes and impacts. Traditional impact / outcome evaluation need a logic model before the projects, but a logic model is sometimes solid and is not flexible, I think. We will also introduce the portfolio approach and we’re interested in a dynamic, flexible and complex model, so I agree with your idea that LLM generates outcomes and benefits based-on the actual outputs after the projects. However, there is a problem for kuu village, especially. kuu village has more physical experience, and it is more difficult to record the experience than digital outputs like source cord and research papers. There is oracle problem. Of course, we can publish the recap report after kuu village, but it is hard to verify whether the information is true or false. I won’t publish a fake info haha

Conclusion

We ā€œkuu villageā€ would like to join the round if this proposal is accepted. And then, I would like to contribute to improve the evaluation system since I work in impact space. This proposal is totally interesting!

2 Likes

At Funding the Commons we’ve been working on strategies to measure, evaluate, and report impact to our stakeholders across our previous residencies. We know first-hand how difficult it can be to both capture the right data and evaluate it in a consistent way.

That’s why we see Hypercerts as a promising foundation: they provide a way to standardize claims about impact, align multiple funders and evaluators around shared metrics, and build transparency into reporting. In our residencies we’ve experimented with combining qualitative data (builder stories, collaborations sparked, downstream projects) with more quantifiable outcomes (bounties completed, follow-on funding raised), and we’re actively exploring how Hypercerts can anchor these contributions into a common standard.

We’d love to compare notes with others tackling outcome-based funding. There’s an opportunity here to co-develop shared reporting frameworks that can work across pop-ups, hackathons, residencies, and other short-cycle funding mechanisms.

Funding the Commons would be happy to support and facilitate the development and test-running of such a system through our upcoming residency near Buenos Aires this October. We see it as a great opportunity to ground these ideas in practice and generate learnings that the broader ecosystem can build upon.

1 Like

Amazing to see this initiative to improve sustainable and fair popup funding! :heart_eyes:

From ZuBerlin side we strongly support this and would be open to participate if it’s voted in.

As already brought up directly measuring impact for popups is hard, often time delayed, anecdotal and challenging to directly quantify. We appreciate attempts to do so but would also recommend being cautious with driving things too much towards primarily quantitative metrics.

Simply having the costs and breakdowns for popups is already helpful for the ecosystem as it improves planning ability and increases transparency.

Going forward a combination of automated / quantitative approaches and governance-based ones (individuals who can assess popups well) will likely be a good balance.

We are generally quite open and supportive of stepwise iteration with the suggested LLM-based approach here.

Thanks a lot to @nidiyia and all other contributors for pushing this forward! :muscle:

I’ve thought of a few potential additional criteria::

  1. Sustainability & Longevity: Is the project likely to persist and continue delivering impact?
  2. Community Engagement & Diversity: How inclusive is the project, and how much does it grow networks or participation?
  3. Knowledge Dissemination & Replicability: Are outputs sharable and usable by others in the ecosystem?
  4. Cross-Ecosystem Contribution – Does the project inform or support other residencies/projects?
  5. Risk Awareness – Are potential downsides or unintended consequences identified, with mitigation strategies in place?

Adding these dimensions could make impact evaluation more robust, capturing not just activity but influence, sustainability, and ecosystem value.

This a very intetersting proposition. I think measuring impact and cost-effectiveness could be valuable but as many have noted here, the challenge is comparing one impact metric to the other. One size fits all might not work for all the pop-ups as each could have a different approach.

One potential solution:

  • Establish a few impact metrics and reporting standards.
  • Pop-ups by participating in this round agree to respond to n/k impact metrics following the reporting standards.
  • compare only matching impact metrics from one pop-up to the other.

It is hard to come up with metrics that a group of people would agree as impactful.
For example, for person A, attending pop-up in a new country never visited before. This could be very impactful for person A.
In another pop-up person B goes to hack on some problems, this could be impactful for person B.

Another approach could be,

  • Each pop-up designs a theory of change (TOC), publishes it with how they would evaluate the change.
  • This would be considered as an entry point to the round.
  • This gives freedom to the pop-up to design their impact.
  • The community would vote on which TOC is more impactful and then funds are distributed based on that.

Having said the above and being the organizer of India’s pop-up ZuGrama, I would be interested in participating and experimenting on this.

Revised Proposal: Outcome driven funding for Web3 Popups & Residencies by @nidiyia with help from Dipanshu Singh and Devansh Mehta

TLDR

This is an update to the proposal published on August 16, 2025.

It incorporates the feedback received from @owocki (scorecard) & @deltajuliet (scorecard); and reflects a revised plan of action after conversations with six Web3 popups & residencies.

Please consider this revised proposal for ā€˜Outcome-Driven Funding for Web3 Popups & Residencies’ when casting your votes for GG24 domains on September 11, 2025.

Goal: build a funding model for Web3 popups and residencies that encourages a culture of robust impact documentation and rewards each popup for verified benefits generated for the broader Ethereum community, while taking into consideration the cost to run a popup

Mechanism: participating popups submit detailed impact assessments of their past or recently concluded programs. Impact reports are fed to various LLM based Impact Quantifiers that assign a $ value to outcomes submitted by each popup. Each popup applies with a hypercert that represents their funding target in GG24. Funds are allotted algorithmically, in proportion to the benefits from each popup divided by the fundraising target listed in their hypercert, with an appeal period for unfair calculations and a maximum cap.

Revision Highlights:

1. Focus on benefits and fundraising targets

Being mindful of budget privacy concerns & learnings from the earlier Gitcoin Zuzalu round, the revised mechanism focuses on calculating the benefit (in $) accruing from each popup. Sharing cost data of past hypercerts is optional & not required The project applicants are popup city hypercerts with fundraising targets for GG24 and GG25. The focus is on funding popups based on the verified benefits it has generated in the past divided by the amount it wishes to raise for the future, rather than only a reimbursement of past costs incurred to host it.

2. Popups are responsible for their own impact documentation and hypercert creation

Conversations with six web3 popups/residencies revealed that a one size-fits all impact framework is unviable given the unique mission & nature of every popup. Each participating popup is responsible to document its own impact based on its own defined metrics to submit a true & detailed report to domain operators.

Along with this, an application form is to be filled out by all participating teams eliciting program related information common to all popups, along with minting a hypercert stating clearly they amount they want to raise in Gitcoin rounds.

3. LLM based Impact Quantifier

Existing impact models often output similar scores across projects, when they shouldn’t. We have overcome this issue, by feeding it a formal evaluation schema-the Relentless Monetization Technique (Weinstein & Bradburd, 2013). This follows a structured process that delivers differentiated and higher-quality results. We have tested this quantifier with preliminary documentation from four popups with their permission. Results are published anonymously in Section C.

The revised proposal begins with describing the mechanism process (Section A), goes on to highlight important domain details like popups onboarded, co-funding, evaluations process, public dashboard, privacy policy, risks & timeline (Section B) & wraps up with test results from the LLM Impact Quantifier, based on trial runs using impact reports shared confidentially by four Web3 popups (Section C).

Section A: How it Works

Step-1: Popups submit their own records of impact

  • To enter the round, each popup fills out an application form that elicits general program related information like location, no. of attendees, no. & formats of events held, specific outputs/outcomes arising out of each event if any & clear impact metrics they are tracking.

  • They also submit their own impact evaluation reports which have already been prepared for other purposes. Impact evaluations can be submitted in varied formats (grant applications, detailed reports, narrative, spreadsheets etc.).
    Domain operators provide a suggested framework but do not interfere in evaluations and documentation of impact per popup.

  • Any popup that has taken place and recorded its impact on or after February 15, 2024 will be eligible for submission and funding in the proposed GG24 domain round. The cutoff date of February 15, 2024, is chosen because it marks the conclusion of the last web3 popup funding round, i.e., the Zuzalu QF Grants Program on Grants Stack.

  • Submitted impact metrics & reports are transparent & made publicly accessible. If popups do not want to disclose specific information, they must indicate the same in the application form. However, private information will not be considered by the impact quantifier in its calculations.

  • Popups also mint a hypercert representing these past impacts, as part of which they need to list a fundraising cap (price per unit of hypercert). This will be considered for funding in 2 ways; popups cannot raise more funds than that possible via purchase of their hypercert; their overall score determining allocations between popups is quantified benefits of past residencies divided by cost listed in the hypercert.

Step-2: LLM based Impact Quantifier calculates a $ value for impact per popup

  • All submitted reports are processed through an LLM-assisted Impact Quantifier.
  • The system uses 3** independent LLMs to generate a net benefits (in $) value for each popup. The mean or median value is published with an end result reading for e.g., ā€œ$20k in benefits generated by Popup Xā€.
    **Note: For the test run results (Section C) the following models were used: Chat GPT 5, Gemini & Claude 4/Grok 4.
  • If additional data is needed for a more thorough calculation, popups will be asked to provide supplemental information. If they do not have this additional data, the calculation would be produced independent of it.
  • During GG24, their community can purchase units of their hypercert and leave a comment describing the impact of the popup or residency on them. Both the dollar value purchased by their community and the comments left behind will be considered by the quantifier, example here
  • The main calculation is done by the LLM Impact Quantifier (see Section B). A technical lead fixes any bugs & ensures the smooth functioning of the LLM. An evaluations lead reviews the LLM calculations for general validity of method & data sources submitted by popups for factual accuracy (see Section B).

Step-3: Publication of results & appeals

  • The impact reports submitted by each popup are made public, along with the quantification of these reports by the LLM.

  • If popups do not want information disclosed to be made public, they can indicate the same in the application form. But private information will NOT be considered by the quantifier.

  • If popups feel the quantification of their impact report is unfair, they can lodge an appeal within a specified time period. Their responses will be fed back into the LLM after review from the evaluations lead, with updates to the quantification process made accordingly. The appeals will also be transparently documented on the forum.

  • The final quantified benefit will be divided by the cost listed in their hypercert to obtain the score according to which allocations are made. This step provides a moderating feature designed such that popups do not request excess funds as it reduces their score; and they do not list too low an amount as it caps the aggregate funding they can receive from the past impact they have created.

Step-4: Allocation mechanism

  • The final output will be an excel/.csv file with the names of participating popups, linked self-conducted impact evaluations & the median net benefit value (in $) accruing from each of them as calculated by 3 LLM models. It will also link to the hypercert created by each popup to apply in the round, which will be divided by the median net benefit to obtain their final score.

  • This file is programmed into a wallet with a formula that distributes the common pool of funds among participating popups in proportion to their benefit:cost ratio.

  • The maximum amount of funding per popup in GG24 will be capped at 20% of the funding pool at most.

  • We anticipate continuation of the popup funding round for GG25 with the same hypercerts (and new applicants that didn’t participate or those created more impact in the next 6 months), so participants are encouraged to think long term

Section B: Domain Details

Popups signaled for participation

We have had conversations with six popup organizers since August 16, 2025. The following popups have provided in principle commitments to participating in the proposed outcome driven funding round for Web3 popups & residencies for the upcoming GG24 cycle;

Funding

Based on consultations with popups, an average of $20k-50k per popup applicant is considered viable to incentivize their participation in the round.

One level of cofunding is from popups asking their participants to purchase units of their hypercert in GG24 to leave a comment listing the impact it has had on them, so that it can be considered by the quantifier in aggregating total benefit.

We have yet to confirm cofunding for the matching pool, our time this month was spent on speaking with popups and refining this proposal.

Evaluation

A human evaluator/domain operator feeds the impact reports submitted by popups to an LLM assisted Impact Quantifier.

The Impact Quantifer runs the popup submissions on 3 different LLMs & takes a median value to generate a net benefit (in $) value per popup.

The Quantifier is built on the Relentless Monetization (RM) method (Weinstein & Bradburd, 2013) & follows five steps to monetize outcomes per popup. See Section C for these detailed steps that go into the system prompt. The system prompt containing the evaluation mechanism is open source & accessible to all.

All calculations are ultimately based on the reports & data submitted by popups. It is the responsibility of popups to create the hypercert defining their own metrics, record outcomes unique to their own mission and list the total amount they want to raise . No standardized, one-size fits all metrics are suggested by the LLM or human evaluators in the domain team across popups (albeit some general program details will be collected via the application form).

The LLM Impact Quantifier makes the final calculations on $value generated per popup, drawing on the RM method. Humans in the loop in the form of a technical & evaluations lead ensure that the calculations are free of technical bugs, based on factually correct data & follow the method correctly. They also manually divide the $ value generated per popup with the cost listed in their hypercert to obtain the final score upon which allocations get made.

Evaluation Team

Team Member Role
LLM Impact Quantifier Calculates a $ value for outcomes generated by each popup
Technical Lead Dipanshu Singh Built the Quantifier, will oversee all technical matters pertaining to the Quantifier to ensure its sound functioning.
1. Human evaluator Nidhi Harihar Well versed with Relentless Monetization & web3 popups, will 1. Check for validity of sources and references cited by the LLM in making its calculations to ensure that the data sources are relevant, sound & valid 2. Cross check data & ensure sound methodology 3. Liaise with popup POCs to procure relevant additional data that may be needed for fairer & sounder calculations.

Privacy & data sharing policy

Popups submit their own impact metrics & evaluations in a format tailored uniquely to their program. The data they choose to submit (or not) is at their own discretion.

By default, all impact metrics per popup will be made public & published openly. This is an important step to making popup funding transparent & outcomes driven.

Impact evaluation reports submitted by each popup will also be published openly on the public dashboard. The LLM Quantifier reports with calculations of $value generated & detailed methodology followed to arrive at the value will be published publicly by default.

If popups want to keep certain sections of this report anonymous/ locked; they must indicate the same in the application form or simply not share it in their submissions. Any private impact data will NOT be considered by the quantifier.

Risks & sensitivities

The proposed mechanism comes with the following risks:

  • Popups may misrepresent or overstate their outcomes, resulting in the calculator assigning dollar values to impacts that did not actually occur.
  • Given that both the prompt and calculator are public, popups may strategically adjust the way they present information, without explicitly falsifying data, in order to maximize the benefits calculated.
  • To preserve budget privacy, cost data is kept optional. But a fundraising goal has to be specified in the application form, based on which a benefit cost ratio is calculated for allocation.
  • The Relentless Monetization technique widely used to calculate social impact benefits is built to measure & responsibly monetize intangible benefits.

Public dashboard

The round will be hosted on VoiceDeck: app.voicedeck.org, a platform to fund impact. VoiceDeck was launched in November 2024 to fund concrete outcomes resulting from journalism investigations. Outcome driven funding for popups & residencies extends its mission to the web3 domain. The dashboard will be open source & publicly accessible. It will serve as a repository to track key program details & specific outcomes resulting from web3 popups & residencies participating in the round, with the following details:

Details of popups published

  • Popup Name
  • Year/s for which impact evaluation has been conducted
  • Recurrence/frequency of event
  • Impact evaluations submitted by popups
  • Impact metrics tracked (all tangible & intangible metrics)
  • Costs (optional) , fundraising goal for GG24 and GG25 (mandatory)
  • Current funding status

Information from domain operators

  • System prompt to generate a $ value of outcomes per popup
  • Detailed LLM generated report complete with all calculations, assumptions & inferred data points
  • Funding allocation mechanism (formula, codes etc.)
  • Benefit cost ratio

Section C: LLM based Impact Quantifier & Test Results

LLM assisted Impact Quantifier

Our LLM assisted Impact Quantifier is built to assign a $ value to outcomes listed out in impact reports.

The Quantifier uses Relentless Monetization (Robin Hood Rules for Smart Giving) pioneered by the Robin Hood Foundation (Weinstein & Bradburd, 2013- chapters 3, 4, 5 & 8) to derive a dollar value for each impact, with much less overhead.

It gives a tangible representation of impact by saying, for every $1 given to this project, $X in value was created.

The Impact Quantifier can be run using:

Relentless Monetization: 5 steps of arriving at $ value of benefits per popup

A key risk of using LLMs (without a structured method) to score projects is that they lack variance or differentiation. This means that they usually end up generating similar values or outputs across projects, even though the projects may be significantly different in impact produced

To address this issue, we embedded a full evaluation schema, centered on the Relentless Monetization technique, into the LLM.

This guides our model to follow five distinct steps (widely recognized in social impact evaluations) to calculate a $value in benefits unique to the outcomes listed in reports submitted by each popup.

Outputs generated by our LLM calculator thus account for higher-quality differentiation among projects with significant variance between popups based on the quality of documentation they provide.

Step Description Formula
Define outcomes Clear outcomes of residencies & popups are defined. This includes listing both tangible outcomes (e.g., follow-on funding secured per project, jobs landed by attendees) & intangible outcomes (say, knowledge transfer & skill development) The LLM reads through the documents shared by popups & lists outcome/s
Measurement of Causal Effect Quantify what % of these outcomes can be fairly attributed to be a result of the residency/popup exclusively, vis-a-vis other factors/network effects The LLM assigns a % value drawing on relevant data points shared by popups & benchmarking it vis-a-vis relevant & latest industry studies
Calculating Gross Benefit Against each outcome, the no. of beneficiaries reached & a benefit per beneficiary (in $ value) is calculated Gross Benefit = Ī£ Number of Beneficiaries Ɨ Benefit per Beneficiary (per outcome)
Counterfactual Analysis Calculates the net incremental benefit of the residency/popup by adjusting for the loss/gain in benefits if the residency has not taken place Net benefit = Gross benefit - Total Counterfactual benefit
Discounted Future Benefits Finally, the net benefit amount is adjusted for the decreasing $ exchange value in the years following the residency/popup Discounted Net Benefit = Net Benefit / (1+r)^t
Calculate Cost of the Popup While Relentless Monetization considers past costs, we are tweaking this method to include so that cost of the hypercert they create to apply in the round Cost = Price of the Hypercert they created to apply in the round
Obtain a benefit cost ratio Impact is always relative to cost incurred; this intuition is applied here by allocating funds based on how much funding popups solicit Benefit Cost Ratio = Step 5 divided by Step 6

Notes:

  1. The RM technique emphasizes referencing related studies, papers & international reports and citing them as proxy sources in quantifying attribution & benefits. Providing clear evidence of data/numbers at every step is the most critical aspect of this method, which is enabled by the LLMs being plugged into the Internet.

  2. There is a subtle but important difference in steps 2 & 4: Causal effect measurement is about attribution, i.e., isolating the program’s outcomes from other influences.
    Counterfactual analysis is about adjusting the gross benefit to avoid overstating the impact by considering whether the impact would have taken place even without the popup.

  3. The discounted net benefit/net present value of benefit thus calculated is taken as the outcomes (in $) generated per popup/residency.

  4. The benefits are NOT to be taken as impact created by a popup. It is to be interpreted as the quantifiable value of the documentation they have undertaken of their popup.

Test run results of our LLM Impact Quantifier

Case study 1: Quantifying the $ value of Ethereum open source repos

We used our LLM based Impact Quantifier to generate a gross & net benefit value for 45 open source repos in the deep funding competition.

We ran our system prompt on GPT-5 & Gemini 2.5 Pro, with the question ā€˜what is the $ value that each repo generates for Ethereum?’. By taking an average of the results from GPT-5 & Gemini 2.5 Pro with an optimal weight percentage of 80% & 20% respectively, we were able to get results with an error value of 6.8, finishing 9th place on the leaderboard.

Case study 2: Quantifying the $ value of Web3 popup cities/residencies

Impact Evaluations shared by popups

We tested our LLM assisted Impact Quantifier by feeding it impact reports shared by four popups. These impact reports were fed into 4 different LLMs viz., Chat Gpt-5, Claude 4, Grok 4 & Gemini.

The names of these popups and the specific reports shared by them are kept confidential at this stage.

Broadly, the impact reports shared by popups were in the formats of:

  • Draft preliminary reports with impact metrics
  • Substack articles recapping the program
  • Tweets capturing event highlights
  • Proceedings/research articles produced at the end of the programs

Indicative results based only on benefit quantified are below. Please note that actual weights would be calculated by dividing the benefit with the cost of the hypercert created by the popup in applying to the round.

Popup Chat GPT5 Gemini Pro 2.5 Claude 4/ Grok 4 Median Value Weights
A $651k, See report here $450k in Gross benefits, Net benefit calculation needs more data, See report here $1.13 M, See report here $651,000 0.16
B Gross benefits: $1M (conservative), $2M (upper estimate) Net benefit calculation needs more data, See report here Calculation needs more data, See report here $30M in Gross benefits, with inferred ticket price data. Net benefit calculation needs more data See report here $2million 0.49
C $248k, See report here $292k, See report here $292k, See report here $292,000,Listed cost of their hypercert: $150,000 Benefit Cost ratio = 1.946 0.07
D $7.07M, See report here Calculation needs more data, See report here $1.17M , See report here $1.17M 0.28

General Observations and Notes from the Sensemaking Period with Popups

  1. The $ value in benefits in the table above is NOT to be taken as a measure of the impact generated by popups but as a reflection of the documentation submitted by popups.

  2. Popups generating intangible value (Popup B) are being valued higher than popups (e.g., IERR 2025) focused on tangible metrics of impact (like projects launched, papers published, follow-on funding raised etc.).

  3. We spoke with many popups and were struck by how impact documentation has often been minimal, leaving valuable outcomes under-recorded. Recording of impact remains an exception rather than a rule, lacks rigor and a systematized process.

  4. We need to create an incentive for popups to document their outcomes & contributions post their event. The proposed system of funding is a step towards instilling this culture of impact documentation in the popup ecosystem.

  5. Besides funding, a crucial benefit of popups participating in this round would be in enabling transparency in impact metrics. The breadth of impact metrics, unique to each popup, will be openly available on a public dashboard and teach other popups how to measure and record their impact. This provides much needed accountability of outcomes, which is a noted demand from web3 popup funders currently.

Hey @thedevanshmehta & @nidiyia

We’re missing some foundational information to evaluate this Domain for GG24 inclusion and Match Funding.

  1. What is the target budget for this Domain?
  2. What are the current $$ commitments you’ve secured?
  3. What, if any, fees will be charged by the Operator and tools used?

We would need these answers in the next 2-3 hours for us to be able to include this as a Domain proposal in our evaluation.

Thanks for this @LuukDAO

  1. Our target budget would be $200K at $20k per popup as indicated in our proposal, assuming 10 popups participate (6 pre-committed popups + considering 4 new ones).

  2. We have not secured any co-funding commitments yet. That said, since our revised proposal is now up we will be sharing it with potential cofunders & hope to hear from them.

  3. A flat fees of 10% or applying the fair fees model to cover operator/tools costs

Thanks for quick reply.

Just to confirm, the current budget suggests:
$220K target with ~20k USD going to operations
$0K USD confirmed third-party commits

My personal initial reaction is that the request seems significant, given the uncertainty around co-matching.

I’m wondering - would this Domain also be able to run with a target budget of, for example $110k?

Yes, without co-funding locked in, I understand that a budget of 100K+ is unfeasible.

I’m hopeful that an initial commitment enables this round to take place, which can then attract past funders of the Zuzalu round in February 2024 to also participate. The first funder is always the hardest!

Some other considerations:

  1. 100K to popups will be enough to validate the Impact Quantifier allocation method we are proposing & also get robust documentation from popups around impact created since their last funding round on Gitcoin.

  2. The Hypercerts that popups create in this round will carry over to GG25. So if we can show the legitimacy of the method in GG24, we can attract co-funding in the next cycle or even apply the same Quantifier to other domains.

  3. Even if the funding amount is deemed too low for all popups to participate in this round, that will make available a greater amount for popups that do participate.

So I think the amount you propose is a good start!

1 Like