Skip to content

Commit 687783c

Browse files
committed
πŸ“ Enhance Impact Evaluators
1 parent d40031b commit 687783c

File tree

2 files changed

+6
-2
lines changed

2 files changed

+6
-2
lines changed

β€ŽDeep Funding.mdβ€Ž

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -106,12 +106,13 @@ Once the competition ends, extra comparisons could be gathered for projects that
106106
- There are better and more modern methods to derive weights from [noisy pairwise comparisons](https://arxiv.org/abs/2510.09333) ([from multiple annotators](https://arxiv.org/abs/1612.04413))
107107
- [Detect and correct for evaluators' bias in the task of ranking items from pairwise comparisons](https://link.springer.com/article/10.1007/s10618-024-01024-z)
108108
- Use active ranking or dueling bandits to [speed up the data gathering process](https://projecteuclid.org/journals/annals-of-statistics/volume-47/issue-6/Active-ranking-from-pairwise-comparisons-and-when-parametric-assumptions-do/10.1214/18-AOS1772.pdf)
109+
- Stop with a "budget stability" rule (expected absolute dollar change from one more batch is less than a threshold)
109110
- Do some post processing to the weights:
110111
- Report accuracy/Brier and use paired bootstrap to see if gap is statistically meaningful
111112
- If gaps are not statistically meaningful, bucket rewards (using Zipf's law) so it feels fair
112113
- If anyone (or jury selection is more relaxed) can rate you can remove low quality raters with heuristics or pick only the best N raters (crowd BT)
113114
- To gather more comparisons, a top-k method could be used instead of pairwise. Show 6 projects. Ask for the top 3 (no need to order them).
114-
- How would things look like if they were Bayesian instead of [classic Bradley-Terry](https://gwern.net/resorter)? Since comparisons are noisy and we have unreliable jurors, can we [compute distributions instead of "skills"](https://github.com/max-niederman/fullrank)?
115+
- How would things look like if they were [Bayesian Bradley Terry](https://erichorvitz.com/crowd_pairwise.pdf) instead of [classic Bradley-Terry](https://gwern.net/resorter)? Since comparisons are noisy and we have unreliable jurors, can we [compute distributions instead of "skills"](https://github.com/max-niederman/fullrank)?
115116
- Let the dependent set their weight percentage if they're around
116117
- Instead of one canonical graph, allow different stakeholder groups (developers, funders, users) to maintain their own weight overlays on the same edge structure. Aggregate these views using quadratic or other mechanisms
117118
- If there is a plurality of these "dependency graphs" (or just different set of weights), the funding organization can choose which one to use! The curators gain a % of the money for their service. This creates a market-like mechanism that incentivizes useful curation.
@@ -143,3 +144,4 @@ Once the competition ends, extra comparisons could be gathered for projects that
143144
- Self declaration needs a "contest process" to resolve issues/abuse.
144145
- Harberger Tax on self declarations? Bayesian Truth Serum for Weight Elicitation?
145146
- Projects continuously auction off "maintenance contracts" where funders bid on keeping projects maintained. The auction mechanism reveals willingness-to-pay for continued operation. Dependencies naturally emerge as projects that lose maintenance see their dependents bid up their contracts
147+
- [Explore Rank Centrality](https://arxiv.org/pdf/1209.1688). Theoretical and empirical results show that with a graph that has a decent spectral gap `O(n log(𝑛))` pair samples suffice for accurate scores and ranking.

β€ŽImpact Evaluators.mdβ€Ž

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -65,9 +65,11 @@ It's hard to do [[Public Goods Funding]], open-source software, research, etc. t
6565
- Let funders choose which lenses align with their values.
6666
- When collecting data, [pairwise comparisons and rankings are more reliable than absolute scoring](https://anishathalye.com/designing-a-better-judging-system/).
6767
- Humans excel at relative judgments, but struggle with absolute judgments.
68-
- Many algorithms can be used to convert pairwise comparisons into absolute scores.
68+
- [Many algorithms can be used to convert pairwise comparisons into absolute scores](https://crowd-kit.readthedocs.io/en/latest/).
6969
- Pairwise shines when all the context is in the UX.
7070
- [Data is good at providing comprehensive coverage of things that are countable. Data is bad at dealing with nuances and qualitative concepts that experts intuitively understand.](https://gov.optimism.io/t/lessons-learned-from-two-years-of-retroactive-public-goods-funding/9239)
71+
- Crowds bring natural diversity and help capture human semantics. [Disagreement is signal, not just noise](https://github.com/CrowdTruth/CrowdTruth-core/blob/master/tutorial/Part%20I_%20CrowdTruth%20Tutorial.pdf). There are niches of experts in the crowds.
72+
- Collecting good pairwise data [is similar to collecting good ML/AI training data](https://github.com/cleanlab/cleanlab).
7173
- **Design for composability**. Define clear data structures (graphs, weight vectors) as APIs between layers.
7274
- Multiple communities could share measurement infrastructure.
7375
- Different evaluation methods can operate on the same data.

0 commit comments

Comments
Β (0)