In most funding rounds, selection procedures require a committee or donors to have an opinion on all eligible projects.
In deep funding, the jury votes carefully on a few projects. These selections serve as test data to choose a winning model that decides allocations to all projects.
Yes, but it is not “an” AI model. There is a data science competition (like the ones on Kaggle or Pond) where anyone can submit models giving weights to projects in a round.
The jury data is used to rank the different model submissions in the competition. Models most closely aligning with juror preferences feature get to decide the funding amounts to all projects in a round.
Very simply, it is a table showing different repos & their relative importance to one another. You can see an example below, based on Github stars.
The weights assigned by the winning model are deployed into a Splits Contract through which anyone can send money and have it distributed between repos.
Protocol Guild directly rewards individuals working on a repo, almost like a Universal Basic Income (UBI) for core contributors. Deep Funding seeks to allocate at the organizational level, not to individuals.
The second key difference is Protocol Guild has equal weights to contributors based on time spent, whereas in deep funding weights are assigned according to relative importance between projects.
A third difference is simply scale: Protocol Guild rewards ~160 contributors and is intended to remain tight knit. Deep Funding rewards 5000 open source repos.
There are 2 ways to do so, via deep funding or predictive funding:
In predictive funding, submitted Machine Learning (ML) models need to guess the amount that will be received by each project in a funding round before it even begins. This will show us what models are most likely to give an allocation similar to your funding mechanism. See some example submissions here.
In deep funding, submitted ML models give weights to the edges of a dependency graph. A jury rates a subset of edges to indirectly select winning models that then allocate rewards across the entire graph. See the competition to get weights between Ethereum core repos here.
Long term, the cost of quantifying impact (as Deep Funding seeks to do) should ideally be 1-2% of the effort in making the impact itself (as the repos are doing). So Deep Funding will be included in the splits contract distributing funds to repos with a weight of 0.01 - 0.02
In the short term, benefactors keep the Deep Funding mission alive. Please support us in Gitcoin and Giveth QF rounds!
The list of seed nodes was taken from:
The 5000 child nodes comprise dependencies of the 34 seed nodes.
The jury is constructed by either invitation or nomination. Select jurors are invited based on their application in the forum or outreach by our team for their special expertise. Other jurors are nominated by the existing jury. You can see the entire juror list and apply or nominate someone to be a juror here
All evaluations follow Chatham House rules. Ratings, comments, juror names and who they nominated are public but who made a specific rating will not be public.
Jurors vote on random pairwise comparisons of one repo vs another. So they may not even be shown the project they have a vested interest in.
In case they do vote on a project they have a conflict in, the indirection of their votes translating to selecting a model mitigates the effect compared to their votes directly allocating weights.
The bigger bias to worry about in deep funding is when the jury is mostly composed of one category (like say execution clients) which lead them to undervalue other categories (like dev tooling).
Pond is the host for the deep funding competition giving weights to ethereum core repos. You can see the leaderboard and make submissions here. This is also where we share the training data from the jurors for you to train a model.
You can use any method you want! Whether that relies on LLMs or a metrics based approach such as number of Github stars, contributors, open issues, etc. There are 15,000 weights requested so we encourage the use of AI.
We encourage you to check out the Pond competition page where this is shared in greater detail. In general, the jury data is divided at the project (repo) level between training, public test (leaderboard) and private test (selection of winning model) to prevent “mutual information” that limits ability to evaluate generalizability of models.
In the final week of the contest, all the jury data used for the public leaderboard will be released to either the training dataset or the private leaderboard.
As the originator of the mechanism, Vitalik Buterin is testing the mechanism with $170k to Ethereum repos based on weights given by the winning model. If successful, we will create a splits contract so that you can also fund Ethereum core development!
Vitalik has also kept aside a prize pool of $50,000 (total) for winning models whose weights are chosen for allocation and for those making open source submissions, contributing datasets and generally providing infrastructure useful to all participants.
There are 2 leaderboards, the first of which is indicative while the second is used to award prizes
a. Raw cost function value: Ranks models by total cost (lower is better). Scores calculated according to a cost function based on how far away your model is from the jury samples, on those questions where jurors have provided samples.
"Under the hood", each sample is converted into a "logit" form, for example, if a juror says that A contributed 4x more than B, then the sample will be in the form (A, B, -1.386), where 1.386 = ln(4), negative because A is higher.
If your model assigns A = 0.2 and B = 0.02, then those values are themselves converted into logits (here, -1.609 and -3.912), and the cost function term is (your_B - your_A - model_diff)^2 = (-3.912 + 1.609 + 1.386) ** 2 = 0.840.
If your model assigns A = 0.2 and B = 0.05, then you get (-2.995 + 1.609 + 1.386)^2 = 0: a perfect score, reflecting that your model exactly mirrors the juror's opinion that A contributed 4x more than B.
b. Contribution to composite model: the second leaderboard reflects how much your model is influencing the ensemble of models that determine the end allocation to projects. Most submissions will likely score 0 as only select models will form the composite model. This approach rewards originality in submissions rather than simply copying the leader. To give an example, if the second-best model is a copy of the first-best model but slightly adjusted to make it worse, it would score #2 in raw cost function value but its weight here may end up being zero.
You can see an example of the scoring script developed by Vitalik Buterin here