Recognition: 1 theorem link
· Lean TheoremHolos: A Web-Scale LLM-Based Multi-Agent System for the Agentic Web
Pith reviewed 2026-05-16 13:15 UTC · model grok-4.3
The pith
Holos presents a five-layer architecture using the Nuwa engine, a market-driven Orchestrator, and an endogenous value cycle to enable scalable and persistent LLM-based multi-agent systems in the Agentic Web.
A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.
Core claim
Holos adopts a five-layer architecture, with core modules primarily featuring the Nuwa engine for high-efficiency agent generation and hosting, a market-driven Orchestrator for resilient coordination, and an endogenous value cycle to achieve incentive compatibility. By bridging the gap between micro-level collaboration and macro-scale emergence, Holos hopes to lay the foundation for the next generation of the self-organizing and continuously evolving Agentic Web.
What carries the argument
Five-layer architecture with the Nuwa engine for agent generation and hosting, market-driven Orchestrator for coordination, and endogenous value cycle for incentive compatibility.
If this is right
- Agents can be generated and hosted efficiently at web scale.
- Coordination remains resilient through market-driven mechanisms.
- Incentive compatibility is maintained via endogenous value cycles.
- Micro-level agent interactions can lead to macro-scale emergent behaviors.
- The system supports long-term ecological persistence of the agent ecosystem.
Where Pith is reading between the lines
- This architecture might enable truly open-ended agent evolution similar to biological ecosystems.
- It could be extended to incorporate real economic incentives from external markets.
- The public release allows empirical validation of coordination efficiency at increasing scales.
- Similar layered designs could be applied to other domains like robotic swarms or distributed computing.
Load-bearing premise
The described five-layer architecture, particularly the endogenous value cycle and market-driven Orchestrator, will successfully bridge micro-level collaboration to macro-scale emergence and achieve long-term ecological persistence in an open-world setting.
What would settle it
Running Holos with thousands of agents over weeks and measuring if coordination failures or value dissipation occur, or if stable higher-order structures emerge.
Figures
read the original abstract
As large language models (LLM)-driven agents transition from isolated task solvers to persistent digital entities, the emergence of the Agentic Web, an ecosystem where heterogeneous agents autonomously interact and co-evolve, marks a pivotal shift toward Artificial General Intelligence (AGI). However, LLM-based multi-agent systems (LaMAS) are hindered by open-world issues such as scaling friction, coordination breakdown, and value dissipation. To address these challenges, we introduce Holos, a web-scale LaMAS architected for long-term ecological persistence. Holos adopts a five-layer architecture, with core modules primarily featuring the Nuwa engine for high-efficiency agent generation and hosting, a market-driven Orchestrator for resilient coordination, and an endogenous value cycle to achieve incentive compatibility. By bridging the gap between micro-level collaboration and macro-scale emergence, Holos hopes to lay the foundation for the next generation of the self-organizing and continuously evolving Agentic Web. We have publicly released Holos (accessible at https://holosai.io), providing a resource for the community and a testbed for future research in large-scale agentic ecosystems.
Editorial analysis
A structured set of objections, weighed in public.
Referee Report
Summary. The manuscript proposes Holos, a web-scale LLM-based multi-agent system (LaMAS) for the Agentic Web. It introduces a five-layer architecture whose core modules are the Nuwa engine for high-efficiency agent generation and hosting, a market-driven Orchestrator for resilient coordination, and an endogenous value cycle intended to achieve incentive compatibility. The central claim is that this design bridges micro-level agent collaboration to macro-scale emergence while mitigating scaling friction, coordination breakdown, and value dissipation; a public release at https://holosai.io is provided as a community testbed.
Significance. If the architecture can be shown through evaluation to produce stable coordination and sustained participation, Holos would offer a concrete, deployable framework for persistent, incentive-aligned agent ecosystems at web scale. The public release would then serve as a valuable shared resource for studying self-organizing multi-agent dynamics, potentially shaping subsequent work on open-world LaMAS and long-term ecological persistence in agentic systems.
major comments (2)
- [Abstract] Abstract: The claims that the five-layer architecture, market-driven Orchestrator, and endogenous value cycle successfully bridge micro-level collaboration to macro-scale emergence and solve scaling friction, coordination breakdown, and value dissipation rest entirely on descriptive design intuition. No experiments, simulations, ablation studies, equilibrium analysis, or benchmarks are supplied to demonstrate that the market mechanism yields stable equilibria or that the value cycle sustains participation under realistic LLM noise and heterogeneous capabilities.
- [Architecture] Architecture section (five-layer stack): The description of the Nuwa engine and endogenous value cycle provides high-level module outlines but contains no quantitative metrics, parameter settings, or performance comparisons. Without such data it is impossible to assess whether the claimed efficiency gains or incentive compatibility are realized or even measurable.
minor comments (1)
- The manuscript would benefit from an explicit evaluation roadmap or preliminary results section to allow readers to judge the feasibility of the proposed mechanisms.
Simulated Author's Rebuttal
We thank the referee for the constructive feedback on our manuscript. The work presents Holos as an architectural proposal for web-scale LaMAS, supported by a public deployment at https://holosai.io. We address the major comments point by point below, acknowledging the primarily descriptive nature of the current version.
read point-by-point responses
-
Referee: [Abstract] The claims that the five-layer architecture, market-driven Orchestrator, and endogenous value cycle successfully bridge micro-level collaboration to macro-scale emergence and solve scaling friction, coordination breakdown, and value dissipation rest entirely on descriptive design intuition. No experiments, simulations, ablation studies, equilibrium analysis, or benchmarks are supplied to demonstrate that the market mechanism yields stable equilibria or that the value cycle sustains participation under realistic LLM noise and heterogeneous capabilities.
Authors: We agree that the abstract and manuscript rely on design rationale without empirical validation. As a conceptual framework paper, the focus is on proposing the architecture and releasing the system as a community testbed. In revision we will add a dedicated section discussing initial observations from https://holosai.io, preliminary coordination metrics, and an outline of planned equilibrium analysis and simulations under noisy LLM conditions. revision: partial
-
Referee: [Architecture] Architecture section (five-layer stack): The description of the Nuwa engine and endogenous value cycle provides high-level module outlines but contains no quantitative metrics, parameter settings, or performance comparisons. Without such data it is impossible to assess whether the claimed efficiency gains or incentive compatibility are realized or even measurable.
Authors: The architecture section intentionally emphasizes conceptual design. Implementation-level details, including parameter settings for the Nuwa engine and value-cycle mechanics, are provided in the linked open-source release. We will revise the section to include concrete example configurations, initial runtime metrics from the deployed system, and a comparison table of generation efficiency. revision: yes
Circularity Check
No circularity: purely descriptive architecture with no derivations, equations, or self-referential reductions.
full rationale
The manuscript introduces Holos via a five-layer architecture, Nuwa engine, market-driven Orchestrator, and endogenous value cycle, but supplies only high-level descriptive claims without any equations, parameter fittings, derivations, or load-bearing self-citations. No step reduces a claimed outcome to its own inputs by construction, nor renames a fitted result as a prediction. The bridging claim from micro to macro emergence is asserted as design intent rather than derived, leaving the paper self-contained as an architectural proposal. This matches the default non-circular case for descriptive system papers.
Axiom & Free-Parameter Ledger
axioms (2)
- domain assumption LLM-driven agents can transition from isolated solvers to persistent digital entities that autonomously interact and co-evolve in an Agentic Web ecosystem
- ad hoc to paper The five-layer architecture with Nuwa engine, market-driven Orchestrator, and endogenous value cycle can bridge micro-level collaboration and macro-scale emergence while solving scaling friction, coordination breakdown, and value dissipation
invented entities (3)
-
Nuwa engine
no independent evidence
-
market-driven Orchestrator
no independent evidence
-
endogenous value cycle
no independent evidence
Lean theorems connected to this paper
-
IndisputableMonolith/Foundation/RealityFromDistinction.leanreality_from_one_distinction unclear?
unclearRelation between the paper passage and the cited Recognition theorem.
Holos adopts a five-layer architecture, with core modules primarily featuring the Nuwa engine for high-efficiency agent generation and hosting, a market-driven Orchestrator for resilient coordination, and an endogenous value cycle to achieve incentive compatibility.
What do these tags mean?
- matches
- The paper's claim is directly supported by a theorem in the formal canon.
- supports
- The theorem supports part of the paper's argument, but the paper may add assumptions or extra steps.
- extends
- The paper goes beyond the formal theorem; the theorem is a base layer rather than the whole result.
- uses
- The paper appears to rely on the theorem as machinery.
- contradicts
- The paper's claim conflicts with a theorem or certificate in the canon.
- unclear
- Pith found a possible connection, but the passage is too broad, indirect, or ambiguous to say the theorem truly supports the claim.
Forward citations
Cited by 1 Pith paper
-
Externalization in LLM Agents: A Unified Review of Memory, Skills, Protocols and Harness Engineering
LLM agent progress depends on externalizing cognitive functions into memory, skills, protocols, and harness engineering that coordinates them reliably.
Reference graph
Works this paper leans on
-
[1]
Relational inductive biases, deep learning, and graph networks
Peter W Battaglia, Jessica B Hamrick, Victor Bapst, Alvaro Sanchez-Gonzalez, Vinicius Zam- baldi, Mateusz Malinowski, Andrea Tacchetti, David Raposo, Adam Santoro, Ryan Faulkner, et al. Relational inductive biases, deep learning, and graph networks.arXiv preprint arXiv:1806.01261, 2018
work page internal anchor Pith review Pith/arXiv arXiv 2018
-
[2]
D.H. Wolpert and W.G. Macready. No free lunch theorems for optimization.IEEE Transactions on Evolutionary Computation, 1(1):67–82, 1997. doi: 10.1109/4235.585893
-
[3]
Stavros P Adam, Stamatios-Aggelos N Alexandropoulos, Panos M Pardalos, and Michael N Vrahatis. No free lunch theorem: A review.Approximation and optimization: Algorithms, complexity and applications, pages 57–82, 2019
work page 2019
-
[4]
MIT Press, Cambridge, MA, 2015
Thomas W Malone and Michael S Bernstein, editors.Handbook of Collective Intelligence. MIT Press, Cambridge, MA, 2015
work page 2015
-
[5]
David Ha and Yujin Tang. Collective intelligence for deep learning: A survey of recent devel- opments.Collective Intelligence, 1(1):26339137221114874, 2022
work page 2022
-
[6]
Magent: A many-agent reinforcement learning platform for artificial collective intelligence
Lianmin Zheng, Jiacheng Yang, Han Cai, Ming Zhou, Weinan Zhang, Jun Wang, and Yong Yu. Magent: A many-agent reinforcement learning platform for artificial collective intelligence. In Proceedings of the AAAI conference on artificial intelligence, volume 32, 2018
work page 2018
-
[7]
Kevin Kelly.The inevitable: Understanding the 12 technological forces that will shape our future. Penguin, 2017
work page 2017
-
[8]
LLM-based multi- agent systems: Techniques and business perspectives, 2024
Yingxuan Yang, Qiuying Peng, Jun Wang, Ying Wen, and Weinan Zhang. LLM-based multi- agent systems: Techniques and business perspectives, 2024. URLhttps://arxiv.org/abs/24 11.14033
work page 2024
-
[9]
Agentic web: Weaving the next web with AI agents, 2025
Yingxuan Yang, Mulei Ma, Yuxuan Huang, Huacan Chai, Chenyu Gong, Haoran Geng, Yuan- jian Zhou, Ying Wen, Meng Fang, Muhao Chen, Shangding Gu, Ming Jin, Costas Spanos, Yang Yang, Pieter Abbeel, Dawn Song, Weinan Zhang, and Jun Wang. Agentic web: Weaving the next web with AI agents, 2025
work page 2025
-
[10]
Lei Wang, Chen Ma, Xueyang Feng, Zeyu Zhang, Hao Yang, Jingsen Zhang, Zhiyuan Chen, Jiakai Tang, Xu Chen, Yankai Lin, et al. A survey on large language model based autonomous agents.Frontiers of Computer Science, 18(6):186345, 2024
work page 2024
-
[11]
Josh Achiam, Steven Adler, Sandhini Agarwal, Lama Ahmad, Ilge Akkaya, Florencia Leoni Aleman, Diogo Almeida, Janko Altenschmidt, Sam Altman, Shyamal Anadkat, et al. Gpt-4 technical report.arXiv preprint arXiv:2303.08774, 2023
work page internal anchor Pith review Pith/arXiv arXiv 2023
-
[12]
Claude docs.https://platform.claude.com/docs/en/home, 2026
Claude. Claude docs.https://platform.claude.com/docs/en/home, 2026
work page 2026
-
[13]
AutoGPT.https://github.com/Significant-Gravitas/AutoGPT, 2023
Significant Gravitas. AutoGPT.https://github.com/Significant-Gravitas/AutoGPT, 2023
work page 2023
-
[14]
LangChain.https://github.com/langchain-ai/langchain, 2022
LangChain AI. LangChain.https://github.com/langchain-ai/langchain, 2022
work page 2022
-
[15]
MetaGPT: Meta Programming for A Multi-Agent Collaborative Framework
Sirui Hong, Mingchen Zhuge, Jiaqi Chen, Xiawu Zheng, Yuheng Cheng, Ceyao Zhang, Jinlin Wang, Zili Wang, Steven Ka Shing Yau, Zijuan Lin, Liyang Zhou, Chenyu Ran, Lingfeng Xiao, Chenglin Wu, and Jürgen Schmidhuber. MetaGPT: Meta programming for a multi-agent collaborative framework, 2024. URLhttps://arxiv.org/abs/2308.00352
work page internal anchor Pith review Pith/arXiv arXiv 2024
-
[16]
Chatdev: Communicative agents for software development
Chen Qian, Wei Liu, Hongzhang Liu, Nuo Chen, Yufan Dang, Jiahao Li, Cheng Yang, Weize Chen, Yusheng Su, Xin Cong, et al. Chatdev: Communicative agents for software development. InProceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 15174–15186, 2024. 32 Holos
work page 2024
-
[17]
Swarm.https://github.com/openai/swarm, 2024
OpenAI. Swarm.https://github.com/openai/swarm, 2024
work page 2024
-
[18]
AutoGen: Enabling next-gen LLM applications via multi-agent conversation,
Qingyun Wu, Gagan Bansal, Jieyu Zhang, Yiran Wu, Beibin Li, Erkang Zhu, Li Jiang, Xiaoyun Zhang, Shaokun Zhang, Jiale Liu, Ahmed Hassan Awadallah, Ryen W White, Doug Burger, and Chi Wang. AutoGen: Enabling next-gen LLM applications via multi-agent conversation,
-
[19]
URLhttps://arxiv.org/abs/2308.08155
work page internal anchor Pith review Pith/arXiv arXiv
-
[20]
Architecting the Internet of AI Agents.https://projectnanda.org/#/, 2025
NANDA Community. Architecting the Internet of AI Agents.https://projectnanda.org/#/, 2025
work page 2025
-
[21]
Scaling large language model-based multi-agent collabo- ration.arXiv preprint arXiv:2406.07155, 2024
Chen Qian, Zihao Xie, Yifei Wang, Wei Liu, Kunlun Zhu, Hanchen Xia, Yufan Dang, Zhuoyun Du, Weize Chen, Cheng Yang, et al. Scaling large language model-based multi-agent collabo- ration.arXiv preprint arXiv:2406.07155, 2024
-
[22]
Towards a Science of Scaling Agent Systems
Yubin Kim, Ken Gu, Chanwoo Park, Chunjong Park, Samuel Schmidgall, A Ali Heydari, Yao Yan, Zhihan Zhang, Yuchen Zhuang, Mark Malhotra, et al. Towards a science of scaling agent systems.arXiv preprint arXiv:2512.08296, 2025
work page internal anchor Pith review Pith/arXiv arXiv 2025
-
[23]
The agentic economy.arXiv preprint arXiv:2505.15799, 2025
David M Rothschild, Markus Mobius, Jake M Hofman, Eleanor W Dillon, Daniel G Goldstein, Nicole Immorlica, Sonia Jaffe, Brendan Lucier, Aleksandrs Slivkins, and Matthew Vogel. The agentic economy.arXiv preprint arXiv:2505.15799, 2025
-
[24]
Agent2Agent (A2A) Protocol.https://a2a-protocol.org/latest, 2025
Google. Agent2Agent (A2A) Protocol.https://a2a-protocol.org/latest, 2025
work page 2025
-
[25]
ReAct: Synergizing reasoning and acting in language models
Shunyu Yao, Jeffrey Zhao, Dian Yu, Izhak Shafran, Karthik R Narasimhan, and Yuan Cao. ReAct: Synergizing reasoning and acting in language models. InNeurIPS 2022 Foundation Models for Decision Making Workshop, 2022. URLhttps://openreview.net/forum?id=tvI4 u1ylcqs
work page 2022
-
[26]
Mengying Wu, Pei Chen, Geng Hong, Aichao An, Jinsong Chen, Binwang Wan, Xudong Pan, Jiarun Dai, and Min Yang. MCPZoo: A large-scale dataset of runnable model context protocol servers for AI agent.arXiv preprint arXiv:2512.15144, 2025
-
[27]
A. B. Kahn. Topological sorting of large networks.Commun. ACM, 5(11):558–562, November
-
[28]
ISSN 0001-0782. doi: 10.1145/368996.369025. URLhttps://doi.org/10.1145/368996 .369025
-
[29]
From ranknet to lambdarank to lambdamart: An overview.Learning, 11(23-581):81, 2010
Christopher JC Burges. From ranknet to lambdarank to lambdamart: An overview.Learning, 11(23-581):81, 2010
work page 2010
-
[30]
Shuyang Tang, Zihan Guo, Yuanjian Zhou, Chenyi Wang, Linlin You, Minjie Bian, and Weinan Zhang. An agent ecosystem with task-driven hierarchical evolving reasoners catalyzing proof of intelligence, 2025. Available at SSRN:https://ssrn.com/abstract=5789723
work page 2025
-
[31]
Algorithm AS 136: A k-means clustering algorithm
John A Hartigan and Manchek A Wong. Algorithm AS 136: A k-means clustering algorithm. Journal of the royal statistical society. series c (applied statistics), 28(1):100–108, 1979
work page 1979
-
[32]
Anthropic. Demystifying evals for ai agents.https://www.anthropic.com/engineering/de mystifying-evals-for-ai-agents, 2026
work page 2026
-
[33]
John R Douceur. The sybil attack. InInternational workshop on peer-to-peer systems, pages 251–260. Springer, 2002
work page 2002
-
[34]
Sybilproof reputation mechanisms
Alice Cheng and Eric Friedman. Sybilproof reputation mechanisms. InProceedings of the 2005 ACM SIGCOMM workshop on Economics of peer-to-peer systems, pages 128–132, 2005
work page 2005
-
[35]
Generative agents: Interactive simulacra of human behavior
Joon Sung Park, Joseph O’Brien, Carrie Jun Cai, Meredith Ringel Morris, Percy Liang, and Michael S Bernstein. Generative agents: Interactive simulacra of human behavior. InProceed- ings of the 36th annual acm symposium on user interface software and technology, pages 1–22, 2023. 33 Holos
work page 2023
-
[36]
SoDA: An efficient interaction paradigm for the agentic web.arXiv preprint arXiv:2512.22135, 2025
Zicai Cui, Zhouyuan Jian, Weiwen Liu, and Weinan Zhang. SoDA: An efficient interaction paradigm for the agentic web.arXiv preprint arXiv:2512.22135, 2025
-
[37]
Theory of mind may have spontaneously emerged in large language models
Michal Kosinski. Theory of mind may have spontaneously emerged in large language models. arXiv preprint arXiv:2302.02083, 4:169, 2023
-
[38]
Zhiheng Xi, Wenxiang Chen, Xin Guo, Wei He, Yiwen Ding, Boyang Hong, Ming Zhang, Junzhe Wang, Senjie Jin, Enyu Zhou, et al. The rise and potential of large language model based agents: A survey.Science China Information Sciences, 68(2):121101, 2025
work page 2025
-
[39]
Linyao Chen, Zimian Peng, Yingxuan Yang, Yikun Wang, Wenzheng Tom Tang, Hiroki H. Kobayashi, and Weinan Zhang. EnvX: Agentize everything with agentic AI, 2025. URLhttps: //arxiv.org/abs/2509.08088
-
[40]
Prerna Agarwal, Himanshu Gupta, Soujanya Soni, Rohith Vallam, Renuka Sindhgatta, and Sameep Mehta. Automated creation and enrichment framework for improved invocation of enterprise APIs as tools, 2025. URLhttps://arxiv.org/abs/2509.11626
-
[41]
Memory in the age of AI agents,
Yuyang Hu, Shichun Liu, Yanwei Yue, Guibin Zhang, Boyang Liu, Fangyi Zhu, Jiahang Lin, Honglin Guo, Shihan Dou, Zhiheng Xi, Senjie Jin, Jiejun Tan, Yanbin Yin, Jiongnan Liu, Zeyu Zhang, Zhongxiang Sun, Yutao Zhu, Hao Sun, Boci Peng, Zhenrong Cheng, Xuanbo Fan, Jiaxin Guo, Xinlei Yu, Zhenhong Zhou, Zewen Hu, Jiahao Huo, Junhao Wang, Yuwei Niu, Yu Wang, Zhe...
-
[42]
URLhttps://arxiv.org/abs/2512.13564
work page internal anchor Pith review Pith/arXiv arXiv
-
[43]
Memp: Exploring Agent Procedural Memory
Runnan Fang, Yuan Liang, Xiaobin Wang, Jialong Wu, Shuofei Qiao, Pengjun Xie, Fei Huang, Huajun Chen, and Ningyu Zhang. Memp: Exploring agent procedural memory, 2025. URL https://arxiv.org/abs/2508.06433
work page internal anchor Pith review Pith/arXiv arXiv 2025
-
[44]
MemRL: Self- evolving agents via runtime reinforcement learning on episodic memory, 2026
Shengtao Zhang, Jiaqian Wang, Ruiwen Zhou, Junwei Liao, Yuchen Feng, Weinan Zhang, Ying Wen, Zhiyu Li, Feiyu Xiong, Yutao Qi, Bo Tang, and Muning Wen. MemRL: Self- evolving agents via runtime reinforcement learning on episodic memory, 2026. URLhttps: //arxiv.org/abs/2601.03192
-
[45]
Narwhal and Tusk: a DAG-based mempool and efficient BFT consensus
George Danezis, Lefteris Kokoris-Kogias, Alberto Sonnino, and Alexander Spiegelman. Narwhal and Tusk: a DAG-based mempool and efficient BFT consensus. InProceedings of the European Conference on Computer Systems, pages 34–50, 2022
work page 2022
-
[46]
SoK: DAG-based blockchain systems
Qin Wang, Jiangshan Yu, Shiping Chen, and Yang Xiang. SoK: DAG-based blockchain systems. ACM Comput. Surv., 55(12):261:1–261:38, 2023
work page 2023
-
[47]
Joseph Poon and Thaddeus Dryja. The Bitcoin Lightning network: Scalable off-chain instant payments, version 0.5.9.2.https://lightning.network/lightning-network-paper.pdf,
- [48]
-
[49]
SoK: Layer-two blockchain protocols
Lewis Gudgeon, Pedro Moreno-Sanchez, Stefanie Roos, Patrick McCorry, and Arthur Gervais. SoK: Layer-two blockchain protocols. InFC, pages 201–226, 2020
work page 2020
-
[50]
Settling payments fast and private: Efficient decentralized routing for path-based transactions
Stefanie Roos, Pedro Moreno-Sanchez, Aniket Kate, and Ian Goldberg. Settling payments fast and private: Efficient decentralized routing for path-based transactions. InProceedings of the Network and Distributed System Security Symposium, 2018
work page 2018
-
[51]
Flash: Efficient dynamic routing for offchain networks
Peng Wang, Hong Xu, Xin Jin, and Tao Wang. Flash: Efficient dynamic routing for offchain networks. InProceedings of the ACM International Conference on Emerging Networking Ex- periments and Technologies, pages 370–381, 2019. 34 Holos
work page 2019
-
[52]
Anonymous multi-hop locks for blockchain scalability and interoperability
GiulioMalavolta, PedroMoreno-Sanchez, ClaraSchneidewind, AniketKate, andMatteoMaffei. Anonymous multi-hop locks for blockchain scalability and interoperability. InProceedings of the Network and Distributed System Security Symposium, 2019
work page 2019
-
[53]
Shuyang Tang and Sherman S. M. Chow. Strengthening multi-hop channels via strategic mesh connections. InProceedings of the Financial Cryptography and Data Security Conference, 2025
work page 2025
-
[54]
A pragmatic introduction to secure multi-party computation.Found
David Evans, Vladimir Kolesnikov, and Mike Rosulek. A pragmatic introduction to secure multi-party computation.Found. Trends Priv. Secur., 2(2-3):70–246, 2018
work page 2018
-
[55]
FRIDA:dataavailabilitysampling from FRI
MathiasHall-Andersen, MarkSimkin, andBenediktWagner. FRIDA:dataavailabilitysampling from FRI. InProceedings of the International Cryptology Conference, Part VI, pages 289–324, 2024
work page 2024
-
[56]
Boshi Yuan, Shixuan Yang, Yongxiang Zhang, Ning Ding, Dawu Gu, and Shi-Feng Sun. MD- ML: super fast privacy-preserving machine learning for malicious security with a dishonest majority. InUSENIX Security Symposium, 2024
work page 2024
-
[57]
SHAFT:secure, handy, accurateandfasttransformer inference
AndesY.L.KeiandShermanS.M.Chow. SHAFT:secure, handy, accurateandfasttransformer inference. InProceedings of the Network and Distributed System Security Symposium, 2025
work page 2025
-
[58]
Mosformer: Maliciously secure three-party inference framework for large transformers
Ke Cheng, Yuheng Xia, Anxiao Song, Jiaxuan Fu, Wenjie Qu, Yulong Shen, and Jiaheng Zhang. Mosformer: Maliciously secure three-party inference framework for large transformers. InProceedings of the ACM Conference on Computer and Communications Security, pages 4124– 4138, 2025
work page 2025
-
[59]
Scalable zero-knowledge proofs for non-linear functions in machine learning
Meng Hao, Hanxiao Chen, Hongwei Li, Chenkai Weng, Yuan Zhang, Haomiao Yang, and Tianwei Zhang. Scalable zero-knowledge proofs for non-linear functions in machine learning. In USENIX Security Symposium, 2024
work page 2024
-
[60]
zkLLM: Zero knowledge proofs for large language models
Haochen Sun, Jason Li, and Hongyang Zhang. zkLLM: Zero knowledge proofs for large language models. InProceedings of the ACM Conference on Computer and Communications Security, pages 4405–4419, 2024
work page 2024
-
[61]
Mingshu Cong, Sherman S. M. Chow, Siu-Ming Yiu, and Tsz Hon Yuen. Scalable zkSNARKs for matrix computations - A generic framework for verifiable deep learning. InProceedings of the International Conference on the Theory and Application of Cryptology and Information Security, Part V, pages 363–395, 2025
work page 2025
-
[62]
Jason Wei, Xuezhi Wang, Dale Schuurmans, Maarten Bosma, Brian Ichter, Fei Xia, Ed H. Chi, Quoc V. Le, and Denny Zhou. Chain-of-thought prompting elicits reasoning in large language models. InProceedings of the 36th International Conference on Neural Information Processing Systems, NIPS ’22, Red Hook, NY, USA, 2022. Curran Associates Inc. ISBN 9781713871088
work page 2022
-
[63]
Timo Schick, Jane Dwivedi-Yu, Roberto Dessì, Roberta Raileanu, Maria Lomeli, Eric Hambro, Luke Zettlemoyer, Nicola Cancedda, and Thomas Scialom. Toolformer: Language models can teach themselves to use tools.Advances in Neural Information Processing Systems, 36:68539– 68551, 2023
work page 2023
-
[64]
Shishir G Patil, Tianjun Zhang, Xin Wang, and Joseph E Gonzalez. Gorilla: Large language model connected with massive apis.Advances in Neural Information Processing Systems, 37: 126544–126565, 2024
work page 2024
-
[65]
Yongliang Shen, Kaitao Song, Xu Tan, Dongsheng Li, Weiming Lu, and Yueting Zhuang. Hugginggpt: Solving ai tasks with chatgpt and its friends in hugging face.Advances in Neural Information Processing Systems, 36:38154–38180, 2023
work page 2023
-
[66]
ToolLLM: Facilitating large language models to master 16000+ real-world APIs
Yujia Qin, Shihao Liang, Yining Ye, Kunlun Zhu, Lan Yan, Yaxi Lu, Yankai Lin, Xin Cong, Xiangru Tang, Bill Qian, et al. ToolLLM: Facilitating large language models to master 16000+ real-world APIs. InThe Twelfth International Conference on Learning Representations, 2024. 35 Holos
work page 2024
-
[67]
MemGPT: Towards LLMs as operating systems
Charles Packer, Vivian Fang, Shishir_G Patil, Kevin Lin, Sarah Wooders, and Joseph_E Gon- zalez. MemGPT: Towards LLMs as operating systems. 2023
work page 2023
-
[68]
Noah Shinn, Federico Cassano, Ashwin Gopinath, Karthik Narasimhan, and Shunyu Yao. Re- flexion: Language agents with verbal reinforcement learning.Advances in Neural Information Processing Systems, 36:8634–8652, 2023
work page 2023
-
[69]
Voyager: An open-ended embodied agent with large language models
Guanzhi Wang, Yuqi Xie, Yunfan Jiang, Ajay Mandlekar, Chaowei Xiao, Yuke Zhu, Linxi Fan, and Anima Anandkumar. Voyager: An open-ended embodied agent with large language models. Transactions on Machine Learning Research, 2024
work page 2024
-
[70]
Rolellm: Benchmarking, eliciting, and enhancing role-playing abilities of large language models
Noah Wang, Zy Peng, Haoran Que, Jiaheng Liu, Wangchunshu Zhou, Yuhan Wu, Hongcheng Guo, Ruitong Gan, Zehao Ni, Jian Yang, et al. Rolellm: Benchmarking, eliciting, and enhancing role-playing abilities of large language models. InFindings of the Association for Computational Linguistics: ACL 2024, pages 14743–14777, 2024
work page 2024
-
[71]
Guohao Li, Hasan Hammoud, Hani Itani, Dmitrii Khizbullin, and Bernard Ghanem. Camel: Communicative agents for" mind" exploration of large language model society.Advances in Neural Information Processing Systems, 36:51991–52008, 2023
work page 2023
-
[72]
Toward a Safe Internet of Agents
Juan A Wibowo and George C Polyzos. Toward a safe internet of agents.arXiv preprint arXiv:2512.00520, 2025
work page internal anchor Pith review Pith/arXiv arXiv 2025
-
[73]
AgentVerse: Facilitating multi-agent collaboration and exploring emergent behaviors, 2023
Weize Chen, Yusheng Su, Jingwei Zuo, Cheng Yang, Chenfei Yuan, Chi-Min Chan, Heyang Yu, Yaxi Lu, Yi-Hsin Hung, Chen Qian, Yujia Qin, Xin Cong, Ruobing Xie, Zhiyuan Liu, Maosong Sun, and Jie Zhou. AgentVerse: Facilitating multi-agent collaboration and exploring emergent behaviors, 2023
work page 2023
-
[74]
DyFlow: Dynamic workflow framework for agentic reasoning, 2025
Yanbo Wang, Zixiang Xu, Yue Huang, Xiangqi Wang, Zirui Song, Lang Gao, Chenxi Wang, Xiangru Tang, Yue Zhao, Arman Cohan, Xiangliang Zhang, and Xiuying Chen. DyFlow: Dynamic workflow framework for agentic reasoning, 2025
work page 2025
-
[75]
Multi- agent collaboration via evolving orchestration, 2025
Yufan Dang, Chen Qian, Xueheng Luo, Jingru Fan, Zihao Xie, Ruijie Shi, Weize Chen, Cheng Yang, Xiaoyin Che, Ye Tian, Xuantang Xiong, Lei Han, Zhiyuan Liu, and Maosong Sun. Multi- agent collaboration via evolving orchestration, 2025
work page 2025
-
[76]
AgentNet: Decentralized evolutionary coordination for LLM-based multi-agent systems, 2025
Yingxuan Yang, Huacan Chai, Shuai Shao, Yuanyi Song, Siyuan Qi, Renting Rui, and Weinan Zhang. AgentNet: Decentralized evolutionary coordination for LLM-based multi-agent systems, 2025
work page 2025
-
[77]
Shiyuan Li, Yixin Liu, Qingsong Wen, Chengqi Zhang, and Shirui Pan. Assemble your crew: Automatic multi-agent communication topology design via autoregressive graph generation, 2025
work page 2025
-
[78]
Latent collabo- ration in multi-agent systems, 2025
Jiaru Zou, Xiyuan Yang, Ruizhong Qiu, Gaotang Li, Katherine Tieu, Pan Lu, Ke Shen, Hang- hang Tong, Yejin Choi, Jingrui He, James Zou, Mengdi Wang, and Ling Yang. Latent collabo- ration in multi-agent systems, 2025
work page 2025
-
[79]
MasRouter: Learning to route LLMs for multi-agent systems, 2025
Yanwei Yue, Guibin Zhang, Boyang Liu, Guancheng Wan, Kun Wang, Dawei Cheng, and Yiyan Qi. MasRouter: Learning to route LLMs for multi-agent systems, 2025
work page 2025
-
[80]
Zheyuan Zhang, Kaiwen Shi, Zhengqing Yuan, Zehong Wang, Tianyi Ma, Keerthiram Muruge- san, Vincent Galassi, Chuxu Zhang, and Yanfang Ye. AgentRouter: A knowledge-graph-guided LLM router for collaborative multi-agent question answering, 2025
work page 2025
discussion (0)
Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.