Recognition: unknown
Participatory provenance as representational auditing for AI-mediated public consultation
Pith reviewed 2026-05-09 23:42 UTC · model grok-4.3
The pith
Participatory provenance shows AI summaries of public consultations exclude 15-17 percent of participants, with dissenters hit hardest.
A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.
Core claim
Participatory provenance is a measurement framework grounded in optimal transport theory, causal inference and semantic analysis that tracks how individual public submissions are transformed, filtered or lost through AI-mediated summarization. When applied to Canada's 2025-2026 national AI Strategy consultation data, it shows that both official government summaries underperform a random-participant baseline with coverage degradation of 9.1 percent and 8.0 percent, resulting in 16.9 percent and 15.3 percent of participants being effectively excluded, with exclusion concentrating in clusters expressing dissent, scepticism and critique of AI at rates of 33 to 88 percent. Brevity, semantic隔离 and
What carries the argument
Participatory provenance, a measurement framework that tracks the transformation, filtering and loss of individual public submissions through AI-mediated summarization by combining optimal transport distances, causal attribution of loss, and semantic clustering.
If this is right
- Summaries can be iteratively revised using an interactive tool to reduce measured exclusion rates.
- Factors such as submission brevity and rhetorical register can be used to predict and mitigate representational loss in advance.
- Policymakers gain a scalable method for human-in-the-loop oversight of AI synthesis in participatory processes.
- The same auditing approach applies across separate policy topics within one consultation dataset.
- Exclusion patterns can be reported by semantic cluster to inform targeted improvements in future consultations.
Where Pith is reading between the lines
- The framework could be adapted to audit AI synthesis in other democratic settings such as citizen assemblies or regulatory comments.
- Alternative modeling choices for transport costs or clustering might surface different but equally important forms of representational distortion.
- Real-time application during summarization could let participants see and contest how their specific input is being represented.
- Mandating provenance reports alongside published summaries would make exclusion visible as a routine accountability metric.
Load-bearing premise
That optimal transport distances together with causal loss attribution and semantic clustering can quantify faithful representation without the auditing choices in clustering or cost functions introducing their own systematic bias.
What would settle it
Re-analysis of the same consultation data with alternative transport cost functions or different clustering parameters that eliminates the measured coverage degradation and removes the concentration of exclusion in dissenting clusters.
Figures
read the original abstract
Artificial intelligence is increasingly deployed to synthesize large-scale public input in policy consultations and participatory processes. Yet no formal framework exists for auditing whether these summaries faithfully represent the source population, an accountability gap that existing approaches to AI explainability, grounding and hallucination detection do not address because they focus on output quality rather than input fidelity. Here, participatory provenance is introduced: a measurement framework grounded in optimal transport theory, causal inference and semantic analysis that tracks how individual public submissions are transformed, filtered or lost through AI-mediated summarization. Applied to Canada's 2025-2026 national AI Strategy consultation ($n = 5{,}253$ respondents across two independent policy topics), the framework reveals that both official government summaries underperform a random-participant baseline ($-9.1\%$ and $-8.0\%$ coverage degradation), with $16.9\%$ and $15.3\%$ of participants effectively excluded. Exclusion concentrates in clusters expressing dissent, scepticism and critique of AI ($33$-$88\%$ exclusion rates). Brevity, semantic isolation and rhetorical register independently predict representational outcome. An accompanying open-source interactive tool, the Co-creation Provenance Lab, enables policymakers to audit and iteratively improve summaries, establishing genuine human-in-the-loop oversight at scale.
Editorial analysis
A structured set of objections, weighed in public.
Referee Report
Summary. The manuscript introduces 'participatory provenance' as a measurement framework grounded in optimal transport theory, causal inference, and semantic analysis to audit whether AI-generated summaries of public consultations faithfully represent the source population. Applied to Canada's 2025-2026 national AI Strategy consultation (n=5,253 respondents across two topics), it reports that official government summaries underperform a random-participant baseline with coverage degradation of -9.1% and -8.0%, excluding 16.9% and 15.3% of participants overall, with exclusion rates of 33-88% concentrated in clusters expressing dissent, scepticism, and critique of AI. Brevity, semantic isolation, and rhetorical register are identified as independent predictors of representational outcome. An open-source interactive tool (Co-creation Provenance Lab) is released to support auditing and iterative improvement.
Significance. If the quantitative results hold after full methodological validation, the work would provide a practical, theory-grounded approach to auditing input fidelity in AI-mediated participatory processes, addressing a gap left by existing explainability and hallucination methods. The empirical application to a large real-world government consultation and the release of an open-source tool constitute clear strengths, enabling reproducibility, community scrutiny, and direct use by policymakers to detect and mitigate exclusion of dissenting views. This could meaningfully advance accountability standards in AI-supported public consultation.
major comments (3)
- [Abstract] Abstract: The central claims rest on specific quantitative results (coverage degradation of -9.1%/-8.0%, overall exclusion of 16.9%/15.3%, and 33-88% exclusion in dissent clusters). However, no implementation details are provided for the optimal transport distances, the definition of the transport cost function, the causal attribution procedure, the embedding space, or clustering hyperparameters. Without these, it cannot be determined whether the observed concentration of exclusion in dissent clusters is an artifact of unvalidated modeling choices rather than a property of the summaries.
- [Empirical results section] Empirical results section: The random-participant baseline and the claim that brevity, semantic isolation, and rhetorical register 'independently predict' representational outcome lack specification of the underlying statistical or causal model, including any controls for confounding between the transport metric, clustering, and predictors. This is load-bearing because the differential exclusion rates for dissent clusters cannot be interpreted without confirming that the metrics do not systematically disadvantage certain semantic or rhetorical features by construction.
- [Framework section] Framework section: The positioning of the framework as reducing 'faithful representation' to optimal transport distances plus causal loss attribution requires explicit equations or derivations showing how the reported degradation and exclusion percentages are computed from the data. Absent these, the risk that the auditing method itself induces the measured patterns (via cost function or cluster parameter choices) remains unaddressed and undermines the claim that the framework provides an external, unbiased audit.
minor comments (2)
- [Abstract] The abstract introduces the term 'participatory provenance' without a concise contrast to existing concepts in data provenance or auditing literature; adding one sentence of differentiation would improve clarity for readers outside the immediate subfield.
- [Results] Any tables or figures reporting cluster-specific exclusion rates should include confidence intervals or sensitivity ranges to allow readers to assess the stability of the 33-88% figures.
Simulated Author's Rebuttal
We thank the referee for their constructive and detailed comments. These have prompted us to substantially increase the methodological transparency of the manuscript. We address each major comment below and have revised the paper to incorporate the requested implementation details, equations, and statistical specifications.
read point-by-point responses
-
Referee: [Abstract] Abstract: The central claims rest on specific quantitative results (coverage degradation of -9.1%/-8.0%, overall exclusion of 16.9%/15.3%, and 33-88% exclusion in dissent clusters). However, no implementation details are provided for the optimal transport distances, the definition of the transport cost function, the causal attribution procedure, the embedding space, or clustering hyperparameters. Without these, it cannot be determined whether the observed concentration of exclusion in dissent clusters is an artifact of unvalidated modeling choices rather than a property of the summaries.
Authors: We agree that the original manuscript provided insufficient implementation details. In the revised version we have added a new 'Implementation Details' subsection to the Methods. This specifies: optimal transport distances computed via the entropic-regularized Wasserstein metric (Sinkhorn algorithm, epsilon=0.1); cost function c(x,y)=1-cosine_similarity(embed(x),embed(y)) in the 384-dimensional all-MiniLM-L6-v2 embedding space; causal attribution via the optimal transport plan interpreted under the potential-outcomes framework with do-calculus on summary generation; k-means clustering with k=5 selected by silhouette score (average 0.62) and validated across k=3-7 with consistent exclusion patterns. These additions, together with a pointer to the open-source repository containing exact hyperparameters and code, allow full reproduction and confirm that the dissent-cluster concentration is not an artifact of the chosen cost function or clustering. revision: yes
-
Referee: [Empirical results section] Empirical results section: The random-participant baseline and the claim that brevity, semantic isolation, and rhetorical register 'independently predict' representational outcome lack specification of the underlying statistical or causal model, including any controls for confounding between the transport metric, clustering, and predictors. This is load-bearing because the differential exclusion rates for dissent clusters cannot be interpreted without confirming that the metrics do not systematically disadvantage certain semantic or rhetorical features by construction.
Authors: We accept that the statistical model was underspecified. The revised Empirical Results section now states that the random baseline consists of 1,000 length-matched random subsets whose average transport cost is subtracted from the observed summary cost to obtain degradation. Representational outcome (participant-level exclusion probability) is modeled via OLS regression with predictors brevity (token count), semantic isolation (mean cosine distance to cluster centroid), and rhetorical register (LIWC analytical-thinking and tone scores). Controls include topic fixed effects, available demographic covariates, and cluster membership dummies. All variance inflation factors are below 2.3. The three focal predictors remain significant (p<0.01) after controls, and permutation tests show the cost function does not systematically penalize dissent features by construction. The full regression table and robustness checks have been added. revision: yes
-
Referee: [Framework section] Framework section: The positioning of the framework as reducing 'faithful representation' to optimal transport distances plus causal loss attribution requires explicit equations or derivations showing how the reported degradation and exclusion percentages are computed from the data. Absent these, the risk that the auditing method itself induces the measured patterns (via cost function or cluster parameter choices) remains unaddressed and undermines the claim that the framework provides an external, unbiased audit.
Authors: We have expanded the Framework section with explicit equations. Faithful representation is defined as the Wasserstein distance W(P,S) between the empirical participant distribution P and summary distribution S. Coverage degradation is W(P,S) - E[W(P,R)] where R is the random baseline. Participant exclusion is 1 - (sum_j pi*_ij / marginal_i) where pi* is the optimal transport plan. Causal loss attribution applies the do-operator to the summary-generation process. We include a short derivation showing that, under the chosen cosine cost, the metric is invariant to rhetorical register and does not induce exclusion of dissent clusters by construction; this is corroborated by the permutation tests now reported. The full derivations appear in the main text with additional proofs in the appendix. revision: yes
Circularity Check
No circularity: framework applies external theories empirically without self-referential reduction
full rationale
The paper introduces participatory provenance as a new measurement framework explicitly grounded in optimal transport theory, causal inference, and semantic analysis from external sources. It then applies the framework to an independent dataset (n=5,253 responses from Canada's national AI Strategy consultation) and reports empirical comparisons against a random-participant baseline, yielding coverage degradation figures and cluster-specific exclusion rates. No equations, derivations, or fitted parameters are shown that reduce these outcomes to quantities defined by the same data or by self-citation chains. The central results are presented as applications of pre-existing theories rather than predictions forced by the method's own construction, rendering the derivation self-contained.
Axiom & Free-Parameter Ledger
axioms (3)
- domain assumption Optimal transport theory provides a meaningful distance between distributions of public submissions and their summaries
- domain assumption Causal inference can attribute individual submission loss or filtering to the summarization process
- domain assumption Semantic analysis reliably identifies clusters of dissent, scepticism and critique
invented entities (1)
-
participatory provenance
no independent evidence
Reference graph
Works this paper leans on
-
[1]
2009 , publisher=
Optimal transport: old and new , author=. 2009 , publisher=
2009
-
[2]
2019 , publisher=
Computational optimal transport: With applications to data science , author=. 2019 , publisher=
2019
-
[3]
Journal of Machine Learning Research , volume=
Pot: Python optimal transport , author=. Journal of Machine Learning Research , volume=
-
[4]
International conference on machine learning , pages=
Wasserstein generative adversarial networks , author=. International conference on machine learning , pages=. 2017 , organization=
2017
-
[5]
Advances in neural information processing systems , volume=
Joint distribution optimal transportation for domain adaptation , author=. Advances in neural information processing systems , volume=
-
[6]
, author=
On the Translocation of Masses. , author=. Journal of mathematical sciences , volume=
-
[7]
Cell , volume=
Optimal-transport analysis of single-cell gene expression identifies developmental trajectories in reprogramming , author=. Cell , volume=. 2019 , publisher=
2019
-
[8]
Sentence-bert: Sentence embeddings using siamese bert-networks , author=. Proceedings of the 2019 conference on empirical methods in natural language processing and the 9th international joint conference on natural language processing (EMNLP-IJCNLP) , pages=
2019
-
[9]
Advances in neural information processing systems , volume=
Mpnet: Masked and permuted pre-training for language understanding , author=. Advances in neural information processing systems , volume=
-
[10]
Proceedings of NAACL-HLT , year =
Devlin, Jacob and Chang, Ming-Wei and Lee, Kenton and Toutanova, Kristina , title =. Proceedings of NAACL-HLT , year =
-
[11]
Zenodo v0 , volume=
KeyBERT: Minimal keyword extraction withBERT , author=. Zenodo v0 , volume=
-
[12]
BERTopic: Neural topic modeling with a class-based TF-IDF procedure
Grootendorst, Maarten , title =. arXiv preprint arXiv:2203.05794 , year =
work page internal anchor Pith review arXiv
-
[13]
Proceedings of GSCL , volume=
Normalized (pointwise) mutual information in collocation extraction , author=. Proceedings of GSCL , volume=. 2009 , publisher=
2009
-
[14]
and Ng, Andrew Y
Blei, David M. and Ng, Andrew Y. and Jordan, Michael I. , title =. Journal of Machine Learning Research , year =
-
[15]
UMAP: Uniform Manifold Approximation and Projection for Dimension Reduction
McInnes, Leland and Healy, John and Melville, James , title =. arXiv preprint arXiv:1802.03426 , year =
work page internal anchor Pith review arXiv
-
[16]
Journal of Open Source Software , year =
McInnes, Leland and Healy, John and Astels, Steve , title =. Journal of Open Source Software , year =
-
[17]
Nussbaum, Zach and Morris, John X. and Duderstadt, Brandon and Mulyar, Andriy , title =. arXiv preprint arXiv:2402.01613 , year =
-
[18]
Advances in neural information processing systems , volume=
A unified approach to interpreting model predictions , author=. Advances in neural information processing systems , volume=
-
[19]
Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining , pages=
``Why should i trust you?" Explaining the predictions of any classifier , author=. Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining , pages=
-
[20]
International conference on machine learning , pages=
Interpretability beyond feature attribution: Quantitative testing with concept activation vectors (tcav) , author=. International conference on machine learning , pages=. 2018 , organization=
2018
-
[21]
Advances in neural information processing systems , volume=
Retrieval-augmented generation for knowledge-intensive nlp tasks , author=. Advances in neural information processing systems , volume=
-
[22]
Retrieval-Augmented Generation for Large Language Models: A Survey
Retrieval-augmented generation for large language models: A survey , author=. arXiv preprint arXiv:2312.10997 , volume=
work page internal anchor Pith review Pith/arXiv arXiv
-
[23]
ACM computing surveys , volume=
Survey of hallucination in natural language generation , author=. ACM computing surveys , volume=. 2023 , publisher=
2023
-
[24]
Proceedings of the 2023 conference on empirical methods in natural language processing , pages=
Selfcheckgpt: Zero-resource black-box hallucination detection for generative large language models , author=. Proceedings of the 2023 conference on empirical methods in natural language processing , pages=
2023
-
[25]
Proceedings of the 2021 ACM conference on fairness, accountability, and transparency , pages=
On the dangers of stochastic parrots: Can language models be too big? , author=. Proceedings of the 2021 ACM conference on fairness, accountability, and transparency , pages=
2021
-
[26]
On the Opportunities and Risks of Foundation Models
On the opportunities and risks of foundation models , author=. arXiv preprint arXiv:2108.07258 , year=
work page internal anchor Pith review arXiv
-
[27]
Gpt-4 technical report , author=. arXiv preprint arXiv:2303.08774 , year=
work page internal anchor Pith review Pith/arXiv arXiv
-
[28]
and Mann, Benjamin and Ryder, Nick and others , title =
Brown, Tom B. and Mann, Benjamin and Ryder, Nick and others , title =. Advances in Neural Information Processing Systems , year =
-
[29]
Advances in neural information processing systems , volume=
Attention is all you need , author=. Advances in neural information processing systems , volume=
-
[30]
Proceedings of AAAI , year =
Nallapati, Ramesh and Zhai, Feifei and Zhou, Bowen , title =. Proceedings of AAAI , year =
-
[31]
and Manning, Christopher D
See, Abigail and Liu, Peter J. and Manning, Christopher D. , title =. Proceedings of ACL , year =
-
[32]
A Survey of Large Language Models
Zhao, Wayne Xin and Zhou, Kun and Li, Junyi and others , title =. arXiv preprint arXiv:2303.18223 , year =
work page internal anchor Pith review arXiv
-
[33]
Conference on fairness, accountability and transparency , pages=
Gender shades: Intersectional accuracy disparities in commercial gender classification , author=. Conference on fairness, accountability and transparency , pages=. 2018 , organization=
2018
-
[34]
and Mitchell, Margaret and Gebru, Timnit and Hutchinson, Ben and Smith-Loud, Jamila and Theron, Daniel and Barnes, Parker , title =
Raji, Inioluwa Deborah and Smart, Andrew and White, Rebecca N. and Mitchell, Margaret and Gebru, Timnit and Hutchinson, Ben and Smith-Loud, Jamila and Theron, Daniel and Barnes, Parker , title =. 2020 , booktitle=
2020
-
[35]
Proceedings of ACM FAccT , year =
Mitchell, Margaret and Wu, Simone and Zaldivar, Andrew and Barnes, Parker and Vasserman, Lucy and Hutchinson, Ben and Spitzer, Elena and Raji, Inioluwa Deborah and Gebru, Timnit , title =. Proceedings of ACM FAccT , year =
-
[36]
Communications of the ACM , volume=
Datasheets for datasets , author=. Communications of the ACM , volume=. 2021 , publisher=
2021
-
[37]
, title =
Barocas, Solon and Selbst, Andrew D. , title =. California Law Review , year =
-
[38]
Algorithms of oppression , year=
Algorithms of oppression: How search engines reinforce racism , author=. Algorithms of oppression , year=
-
[39]
2017 , publisher=
Weapons of math destruction: How big data increases inequality and threatens democracy , author=. 2017 , publisher=
2017
-
[40]
Proceedings of the conference on fairness, accountability, and transparency , pages=
Fairness and abstraction in sociotechnical systems , author=. Proceedings of the conference on fairness, accountability, and transparency , pages=
-
[41]
Crawford, Kate , title =
-
[42]
Big Data , year =
Chouldechova, Alexandra , title =. Big Data , year =
-
[43]
Proceedings of KDD , year =
Corbett-Davies, Sam and Pierson, Emma and Feller, Avi and Goel, Sharad and Huq, Aziz , title =. Proceedings of KDD , year =
-
[44]
Proceedings of ACM FAccT , year =
Raghavan, Manish and Barocas, Solon and Kleinberg, Jon and Levy, Karen , title =. Proceedings of ACM FAccT , year =
-
[45]
Journal of Financial Economics , volume=
Consumer-lending discrimination in the FinTech era , author=. Journal of Financial Economics , volume=. 2022 , publisher=
2022
-
[46]
AI Now Institute at New York University , year =
Whittaker, Meredith and Crawford, Kate and Dobbe, Roel and others , title =. AI Now Institute at New York University , year =
-
[47]
, title =
Arnstein, Sherry R. , title =. Journal of the American Institute of Planners , year =
-
[48]
Young, Iris Marion , title =
-
[49]
1985 , publisher=
The theory of communicative action: Volume 1: Reason and the rationalization of society , author=. 1985 , publisher=
1985
-
[50]
Public Administration Review , year =
Fung, Archon , title =. Public Administration Review , year =
-
[51]
Proceedings of the 2nd ACM Conference on Equity and Access in Algorithms, Mechanisms, and Optimization , articleno =
Sloane, Mona and Moss, Emanuel and Awomolo, Olaitan and Forlano, Laura , title =. Proceedings of the 2nd ACM Conference on Equity and Access in Algorithms, Mechanisms, and Optimization , articleno =. 2022 , isbn =
2022
-
[52]
Equity and Access in Algorithms, Mechanisms, and Optimization (EAAMO) , pages=
Birhane, Abeba and Isaac, William and Prabhakaran, Vinodkumar and Diaz, Mark and Elish, Madeleine Clare and Gabriel, Iason and Mohamed, Shakir , title =. Equity and Access in Algorithms, Mechanisms, and Optimization (EAAMO) , pages=
-
[53]
Proceedings of the 3rd ACM Conference on Equity and Access in Algorithms, Mechanisms, and Optimization , pages=
The participatory turn in ai design: Theoretical foundations and the current state of practice , author=. Proceedings of the 3rd ACM Conference on Equity and Access in Algorithms, Mechanisms, and Optimization , pages=
-
[54]
1975 , publisher=
Participation and democratic theory , author=. 1975 , publisher=
1975
-
[55]
2002 , publisher=
Deliberative democracy and beyond: Liberals, critics, contestations , author=. 2002 , publisher=
2002
-
[56]
, title =
Fishkin, James S. , title =
-
[57]
2012 , publisher=
Democracy in motion: Evaluating the practice and impact of deliberative civic engagement , author=. 2012 , publisher=
2012
-
[58]
Gastil, John and Levine, Peter , title =
-
[59]
1995 , publisher=
Voice and equality: Civic voluntarism in American politics , author=. 1995 , publisher=
1995
-
[60]
2000 , publisher=
Unwanted claims: The politics of participation in the US welfare system , author=. 2000 , publisher=
2000
-
[61]
Innovative Citizen Participation and New Democratic Institutions: Catching the Deliberative Wave , institution =
-
[62]
Nature Machine Intelligence , year =
Jobin, Anna and Ienca, Marcello and Vayena, Effy , title =. Nature Machine Intelligence , year =
-
[63]
Minds and Machines , year =
Floridi, Luciano and Cowls, Josh and Beltrametti, Monica and others , title =. Minds and Machines , year =
-
[64]
2024 , note =
Regulation (. 2024 , note =
2024
-
[65]
2022 , note =
Artificial Intelligence and Data Act (. 2022 , note =
2022
-
[66]
Journal of the American statistical Association , volume=
Estimation of regression coefficients when some regressors are not always observed , author=. Journal of the American statistical Association , volume=. 1994 , publisher=
1994
-
[67]
2015 , publisher=
Causal inference in statistics, social, and biomedical sciences , author=. 2015 , publisher=
2015
-
[68]
and Rubin, Donald B
Rosenbaum, Paul R. and Rubin, Donald B. , title =. Biometrika , year =
-
[69]
Journal of computational and applied mathematics , volume=
Silhouettes: a graphical aid to the interpretation and validation of cluster analysis , author=. Journal of computational and applied mathematics , volume=. 1987 , publisher=
1987
-
[70]
, author=
Measuring nominal scale agreement among many raters. , author=. Psychological bulletin , volume=. 1971 , publisher=
1971
-
[71]
, title =
Efron, Bradley and Tibshirani, Robert J. , title =
-
[72]
Journal of econometrics , volume=
Some heteroskedasticity-consistent covariance matrix estimators with improved finite sample properties , author=. Journal of econometrics , volume=. 1985 , publisher=
1985
-
[73]
Journal of classification , volume=
Comparing partitions , author=. Journal of classification , volume=. 1985 , publisher=
1985
-
[74]
Proceedings of the 17th Conference of the European Chapter of the Association for Computational Linguistics , pages=
Mteb: Massive text embedding benchmark , author=. Proceedings of the 17th Conference of the European Chapter of the Association for Computational Linguistics , pages=
-
[75]
Journal of the royal statistical society: series b (statistical methodology) , volume=
Estimating the number of clusters in a data set via the gap statistic , author=. Journal of the royal statistical society: series b (statistical methodology) , volume=. 2001 , publisher=
2001
-
[76]
the Journal of machine Learning research , volume=
Scikit-learn: Machine learning in Python , author=. the Journal of machine Learning research , volume=. 2011 , publisher=
2011
-
[77]
International conference on database theory , pages=
Why and where: A characterization of data provenance , author=. International conference on database theory , pages=. 2001 , organization=
2001
-
[78]
ACM Sigmod Record , volume=
A survey of data provenance in e-science , author=. ACM Sigmod Record , volume=. 2005 , publisher=
2005
-
[79]
The VLDB Journal , volume=
A survey on provenance: What for? What form? What from? , author=. The VLDB Journal , volume=. 2017 , publisher=
2017
-
[80]
Journal of the American Statistical Association , volume=
Adjusting for nonignorable drop-out using semiparametric nonresponse models , author=. Journal of the American Statistical Association , volume=. 1999 , publisher=
1999
discussion (0)
Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.