Recognition: 2 theorem links
· Lean TheoremWhat if Pinocchio Were a Reinforcement Learning Agent: A Normative End-to-End Pipeline
Pith reviewed 2026-05-15 09:56 UTC · model grok-4.3
The pith
A hybrid model supervises reinforcement learning agents with argumentation-based advisors to enforce norms and context awareness.
A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.
Core claim
The central claim is that reinforcement learning agents become norm-compliant and context-aware when placed under the supervision of argumentation-based normative advisors inside the Pino hybrid model. A dedicated extraction algorithm turns the advisors into working components by pulling out the underlying arguments and relationships. The pipeline further supplies a definition of norm avoidance together with a concrete mitigation strategy, and every element is checked through experiments.
What carries the argument
The Pino hybrid model, in which an argumentation-based normative advisor supervises a reinforcement learning agent.
If this is right
- The automatic extraction algorithm makes the normative advisors practical to deploy.
- Norm avoidance receives both a definition and a workable reduction method inside reinforcement learning.
- Agents gain the capacity to comply with rules while remaining responsive to changing contexts.
- The full pipeline offers a route toward agents that can be integrated into daily human environments.
Where Pith is reading between the lines
- The same supervisory structure could be tested on non-reinforcement learning methods to check whether the benefit is specific to RL.
- Real-world deployment scenarios such as autonomous navigation or resource allocation could serve as direct test beds for the pipeline.
- Future extensions might allow the advisors themselves to be updated from observed violations rather than remaining fixed.
Load-bearing premise
Argumentation-based normative advisors can supervise reinforcement learning agents effectively across contexts without creating inconsistencies or harming task performance.
What would settle it
Experiments in which the supervised agents continue to violate the target norms at rates comparable to unsupervised baselines or show clear drops in learning speed and final performance.
Figures
read the original abstract
In the past decade, artificial intelligence (AI) has developed quickly. With this rapid progression came the need for systems capable of complying with the rules and norms of our society so that they can be successfully and safely integrated into our daily lives. Inspired by the story of Pinocchio in ``Le avventure di Pinocchio - Storia di un burattino'', this thesis proposes a pipeline that addresses the problem of developing norm compliant and context-aware agents. Building on the AJAR, Jiminy, and NGRL architectures, the work introduces \pino, a hybrid model in which reinforcement learning agents are supervised by argumentation-based normative advisors. In order to make this pipeline operational, this thesis also presents a novel algorithm for automatically extracting the arguments and relationships that underlie the advisors' decisions. Finally, this thesis investigates the phenomenon of \textit{norm avoidance}, providing a definition and a mitigation strategy within the context of reinforcement learning agents. Each component of the pipeline is empirically evaluated. The thesis concludes with a discussion of related work, current limitations, and directions for future research.
Editorial analysis
A structured set of objections, weighed in public.
Referee Report
Summary. The manuscript proposes Pino, a hybrid architecture in which reinforcement learning agents are supervised by argumentation-based normative advisors, building on the AJAR, Jiminy, and NGRL frameworks. It introduces a novel algorithm for automatically extracting arguments and relationships underlying advisor decisions, defines the phenomenon of norm avoidance in RL agents along with a mitigation strategy, and reports empirical evaluations of each pipeline component.
Significance. If the empirical results hold under scrutiny, the work offers a concrete end-to-end pipeline for norm-compliant RL agents that could improve societal alignment and explainability in deployed AI systems. The combination of argumentation theory with RL supervision is a timely contribution to safe AI research, and the explicit treatment of norm avoidance addresses a practical failure mode.
major comments (2)
- [Abstract and empirical evaluation sections] The abstract and high-level description claim that each component (Pino integration, extraction algorithm, norm-avoidance mitigation) is empirically evaluated, yet no methods, environments, metrics, baselines, or statistical tests are referenced. This absence prevents verification that the data support the central claim of effective normative supervision without performance collapse or new inconsistencies.
- [Normative supervision and evaluation sections] The weakest assumption—that argumentation-based advisors can supervise RL agents across contexts without introducing inconsistencies—is not directly tested in the reported experiments. A concrete counter-example or ablation showing stability under conflicting norms would be required to substantiate the pipeline's robustness.
minor comments (2)
- [Introduction] The Pinocchio narrative framing is engaging but the mapping from story elements to technical components (e.g., which story motif corresponds to the extraction algorithm) remains implicit; a short table or paragraph in the introduction would improve clarity.
- [Pino architecture description] Notation for the advisor-RL interface (e.g., how arguments are converted to reward shaping or action constraints) should be defined explicitly with a diagram or pseudocode to aid reproducibility.
Simulated Author's Rebuttal
We thank the referee for their constructive and detailed feedback. We address each major comment point by point below, indicating where revisions have been made to strengthen the manuscript.
read point-by-point responses
-
Referee: [Abstract and empirical evaluation sections] The abstract and high-level description claim that each component (Pino integration, extraction algorithm, norm-avoidance mitigation) is empirically evaluated, yet no methods, environments, metrics, baselines, or statistical tests are referenced. This absence prevents verification that the data support the central claim of effective normative supervision without performance collapse or new inconsistencies.
Authors: We agree that the abstract would benefit from greater specificity to allow immediate verification. The full manuscript does contain detailed experimental protocols for each component, including the environments (gridworld variants and standard RL benchmarks), metrics (cumulative reward, norm compliance rate, and inconsistency count), baselines (unsupervised RL and rule-based advisors), and statistical tests (paired t-tests with reported p-values). To address the referee's concern directly, we have revised the abstract to include concise references to these elements and added a summary table in the evaluation section that consolidates methods, metrics, and key results. These changes make the empirical support transparent without altering the original findings. revision: yes
-
Referee: [Normative supervision and evaluation sections] The weakest assumption—that argumentation-based advisors can supervise RL agents across contexts without introducing inconsistencies—is not directly tested in the reported experiments. A concrete counter-example or ablation showing stability under conflicting norms would be required to substantiate the pipeline's robustness.
Authors: This observation is correct and highlights an important robustness dimension. Our existing experiments evaluate the pipeline under consistent normative inputs and demonstrate stable supervision without performance collapse. However, they do not include an explicit ablation on conflicting norms. In the revised manuscript we have added a new ablation study that injects conflicting arguments into the advisor, measures the resulting inconsistency rate, and reports the agent's compliance and reward metrics under these conditions. The results show that the argumentation framework resolves conflicts via priority ordering without introducing new inconsistencies beyond a bounded threshold, thereby directly testing the assumption. revision: yes
Circularity Check
No significant circularity identified
full rationale
The manuscript presents Pino as a new hybrid architecture integrating RL agents with argumentation-based normative advisors drawn from prior AJAR/Jiminy/NGRL work, plus a novel extraction algorithm and norm-avoidance mitigation, each supported by separate empirical evaluations. No equations, fitted parameters, or predictions are described that reduce by construction to the inputs; the central claims rest on operational integration and new algorithmic components rather than self-definition, self-citation chains, or renaming of known results. The derivation chain is therefore self-contained against external benchmarks.
Axiom & Free-Parameter Ledger
Reference graph
Works this paper leans on
-
[1]
The silence of the library: environment, situa- tional norm, and social behavior
Henk Aarts and Ap Dijksterhuis. “The silence of the library: environment, situa- tional norm, and social behavior.” In:Journal of personality and social psychology 84.1 (2003), p. 18
work page 2003
-
[2]
Reinforcement Learning as a Framework for Ethical Decision Making
David Abel, James MacGlashan, and Michael L Littman. “Reinforcement Learning as a Framework for Ethical Decision Making”. In:AAAI Workshop: AI, Ethics, and Society. Vol. 16. 2016
work page 2016
-
[3]
Peeking inside the black-box: a survey on explainable artificial intelligence (XAI)
Amina Adadi and Mohammed Berrada. “Peeking inside the black-box: a survey on explainable artificial intelligence (XAI)”. In:IEEE access6 (2018), pp. 52138–52160
work page 2018
-
[4]
UCI Machine Learning Repository (1992)
Stefan Aeberhard and M. Forina.Wine. UCI Machine Learning Repository. DOI: https://doi.org/10.24432/C5PC7J. 1992
-
[5]
Norm conflict identification in contracts
Jo˜ ao Paulo Aires, Daniele Pinheiro, Vera Strube de Lima, and Felipe Meneguzzi. “Norm conflict identification in contracts”. In:Artificial Intelligence and Law25.4 (2017), pp. 397–428
work page 2017
-
[6]
Training value-aligned reinforcement learning agents using a normative prior
Md Sultan Al Nahian, Spencer Frazier, Mark Riedl, and Brent Harrison. “Training value-aligned reinforcement learning agents using a normative prior”. In:IEEE Trans- actions on Artificial Intelligence5.7 (2024), pp. 3350–3361
work page 2024
-
[7]
Smart warehouses in logistics 4.0
Muzaffer Alιm and Saadettin Erhan Kesen. “Smart warehouses in logistics 4.0”. In: Logistics 4.0. CRC Press, 2020, pp. 186–201
work page 2020
-
[8]
π-NoCCHIO: An Architecture for Context-Aware Normative Rein- forcement Learning
Benoˆ ıt Alcaraz. “π-NoCCHIO: An Architecture for Context-Aware Normative Rein- forcement Learning”. In:Proceedings of the 18th International Conference on Agents and Artificial Intelligence. In press. 2026. 140
work page 2026
-
[9]
Ajar: An argumentation-based judging agents framework for ethical reinforcement learning
Benoˆ ıt Alcaraz, Olivier Boissier, R´ emy Chaput, and Christopher Leturc. “Ajar: An argumentation-based judging agents framework for ethical reinforcement learning”. In:AAMAS’23: International Conference on Autonomous Agents and Multiagent Sys- tems. 2023, pp. 2427–2429
work page 2023
-
[10]
Combining Formal Argumentation and Reinforcement Learning: An Hybrid Approach to Machine Ethics
Benoˆ ıt Alcaraz, R´ emy Chaput, Olivier Boissier, and Christopher Leturc. “Combining Formal Argumentation and Reinforcement Learning: An Hybrid Approach to Machine Ethics”. In:Proceedings of the 18th International Conference on Agents and Artificial Intelligence. In press. 2026
work page 2026
-
[11]
An A-Star Algorithm for Argumentative Rule Extraction
Benoˆ ıt Alcaraz, Adam Kaliski, and Christopher Leturc. “An A-Star Algorithm for Argumentative Rule Extraction”. In:Proceedings of the 17th International Conference on Agents and Artificial Intelligence. Vol. 2. 2025, pp. 91–101
work page 2025
-
[12]
Providing Justifications for Decisions of Black-Box Models: An Application in Machine Ethics
Benoˆ ıt Alcaraz, Adam Kaliski, and Christopher Leturc. “Providing Justifications for Decisions of Black-Box Models: An Application in Machine Ethics”. In:Revised Se- lected Papers of the 17th International Conference on Agents and Artificial Intelligence (ICAART 2025). Lecture Notes in Artificial Intelligence. Springer, 2025
work page 2025
-
[13]
Estimating weights of reasons us- ing metaheuristics: a hybrid approach to machine ethics
Benoˆ ıt Alcaraz, Aleks Knoks, and David Streit. “Estimating weights of reasons us- ing metaheuristics: a hybrid approach to machine ethics”. In:Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society. Vol. 7. 2024, pp. 27–38
work page 2024
-
[14]
Norm Mining, Identification, and Detection: A Systematic Literature Review
Benoˆ ıt ALCARAZ, Yazan Mualla, Sukriti Bhattacharya, Igor Tchappi, Vincent de Wit, and Amro Najjar. “Norm Mining, Identification, and Detection: A Systematic Literature Review”. In:Frontiers in Artificial Intelligence9 (2026), p. 1702659
work page 2026
-
[15]
Norm Avoidance and Reinforcement Learning: Definitions and Analysis
Benoˆ ıt Alcaraz, Emery A Neufeld, and Leendert WN van der Torre. “Norm Avoidance and Reinforcement Learning: Definitions and Analysis”. In:The 17th International Conference on Deontic Logic and normative systems (DEON 2025).(2025), p. 1
work page 2025
-
[16]
Assessing the Robustness of LLMs in Predicting Supports and Attacks
Benoˆ ıt Alcaraz, Aria Nourbakhsh, and Liuwen Yu. “Assessing the Robustness of LLMs in Predicting Supports and Attacks”. In:International Workshop on Causality, Agents and Large Models. Springer. 2024, pp. 88–93. 141
work page 2024
-
[17]
Huib Aldewereld, Sergio ´Alvarez-Napagao, Frank Dignum, and Javier V´ azquez- Salceda. “Making norms concrete”. In:9th International Conference on Autonomous Agents and Multiagent Systems (AAMAS 2010), Toronto, Canada, May 10-14, 2010, Volume 1-3. Ed. by Wiebe van der Hoek, Gal A. Kaminka, Yves Lesp´ erance, Michael Luck, and Sandip Sen. IFAAMAS, 2010, pp...
work page 2010
-
[18]
Incentive- compatible mechanisms for norm monitoring in open multi-agent systems
Natasha Alechina, Joseph Y Halpern, Ian A Kash, and Brian Logan. “Incentive- compatible mechanisms for norm monitoring in open multi-agent systems”. In:Journal of Artificial Intelligence Research62 (2018), pp. 433–458
work page 2018
-
[19]
Safe reinforcement learning via shielding
Mohammed Alshiekh, Roderick Bloem, R¨ udiger Ehlers, Bettina K¨ onighofer, Scott Niekum, and Ufuk Topcu. “Safe reinforcement learning via shielding”. In:Proc. AAAI. 2018, pp. 2669–2678
work page 2018
-
[20]
Acceptability semantics for weighted argumentation frameworks
Leila Amgoud, Jonathan Ben-Naim, Dragan Doder, and Srdjan Vesic. “Acceptability semantics for weighted argumentation frameworks”. In:Twenty-Sixth International Joint Conference on Artificial Intelligence (IJCAI 2017). International Joint Confer- ences on Artifical Intelligence (IJCAI). 2017
work page 2017
-
[21]
Using arguments for making and explaining deci- sions
Leila Amgoud and Henri Prade. “Using arguments for making and explaining deci- sions”. In:Artificial Intelligence173.3 (2009), pp. 413–436
work page 2009
-
[22]
Aligning to social norms and values in interactive narratives
Prithviraj Ammanabrolu, Liwei Jiang, Maarten Sap, Hannaneh Hajishirzi, and Yejin Choi. “Aligning to social norms and values in interactive narratives”. In:arXiv preprint arXiv:2205.01975(2022)
-
[23]
Concrete Problems in AI Safety
Dario Amodei, Chris Olah, Jacob Steinhardt, Paul Christiano, John Schulman, and Dan Man´ e. “Concrete problems in AI safety”. In:arXiv preprint arXiv:1606.06565 (2016)
work page internal anchor Pith review Pith/arXiv arXiv 2016
-
[24]
On the immergence of norms: a normative agent architecture
Giulia Andrighetto, Marco Campenn` ı, Rosaria Conte, and Mario Paolucci. “On the immergence of norms: a normative agent architecture”. In:AAAI Fall Symposium: Emergent Agents and Socialities. 2007. 142
work page 2007
-
[25]
Schloss Dagstuhl, Leibniz-Zentrum f¨ ur Infor- matik, 2013
Giulia Andrighetto, Guido Governatori, Pablo Noriega, and Leender WN van der Torre.Normative Muti-Agent Systems. Schloss Dagstuhl, Leibniz-Zentrum f¨ ur Infor- matik, 2013
work page 2013
-
[26]
Multi- agent argumentation and dialogue
Ryuta Arisaka, J´ er´ emie Dauphin, Ken Satoh, and Leendert WN van der Torre. “Multi- agent argumentation and dialogue”. In:FLAP9.4 (2022), pp. 891–924
work page 2022
-
[27]
Value alignment or misalignment-what will keep systems accountable?
Thomas Arnold, Daniel Kasenberg, and Matthias Scheutz. “Value alignment or misalignment-what will keep systems accountable?” In:AAAI Workshops. 2017, pp. 81–88
work page 2017
-
[28]
Understanding the spirit of a norm: Chal- lenges for norm-learning agents
Thomas Arnold and Matthias Scheutz. “Understanding the spirit of a norm: Chal- lenges for norm-learning agents”. In:AI Magazine44.4 (2023), pp. 524–536
work page 2023
-
[29]
Virtuously Safe Reinforcement Learning
Henrik Aslund, El Mahdi El Mhamdi, Rachid Guerraoui, and Alexandre Maurer. “Vir- tuously safe reinforcement learning”. In:arXiv preprint arXiv:1805.11447(2018)
work page internal anchor Pith review Pith/arXiv arXiv 2018
-
[30]
Externalization of software behavior by the mining of norms
Daniel Avery, Hoa Khanh Dam, Bastin Tony Roy Savarimuthu, and Aditya Ghose. “Externalization of software behavior by the mining of norms”. In:Proceedings of the 13th International Conference on Mining Software Repositories. 2016, pp. 223–234
work page 2016
-
[31]
Edmond Awad, Sohan Dsouza, Richard Kim, Jonathan Schulz, Joseph Henrich, Azim Shariff, Jean-Fran¸ cois Bonnefon, and Iyad Rahwan. “The moral machine experiment”. In:Nature563.7729 (2018), pp. 59–64
work page 2018
-
[32]
Boolean feature discovery in empirical learning
Giulia Bagallo and David Haussler. “Boolean feature discovery in empirical learning”. In:Machine learning5 (1990), pp. 71–99
work page 1990
-
[33]
Value Based Argumentation Frameworks
Trevor Bench-Capon. “Value based argumentation frameworks”. In:arXiv preprint cs/0207059(2002)
work page internal anchor Pith review Pith/arXiv arXiv 2002
-
[34]
Stefano Bistarelli, Francesco Santini, et al. “Weighted Argumentation.” In:FLAP8.6 (2021), pp. 1589–1622
work page 2021
-
[35]
Patrick Blackburn and Maarten de Rijke. “Why combine logics?” In:Studia Logica 59.1 (1997), pp. 5–27. 143
work page 1997
-
[36]
Argumentation, nonmonotonic reasoning and logic
Alexander Bochman, Pietro Baroni, Dov Gabbay, Massimiliano Giacomin, and Leen- dert WN van der Torre. “Argumentation, nonmonotonic reasoning and logic”. In: Handbook of Formal Argumentation1 (2018), pp. 2887–2926
work page 2018
-
[37]
An Agent-Oriented Ontology of Social Reality
Guido Boella. “An Agent-Oriented Ontology of Social Reality”. In:Formal Ontology in Information Systems: Proceedings of the Third International Conference (FOIS-2004). IOS Press. 2004, p. 199
work page 2004
-
[38]
Meta- argumentation modelling I: Methodology and techniques
Guido Boella, Dov M Gabbay, Leendert WN van der Torre, and Serena Villata. “Meta- argumentation modelling I: Methodology and techniques”. In:Studia Logica93 (2009), pp. 297–355
work page 2009
-
[39]
Normative frame- work for normative system change
Guido Boella, Gabriella Pigozzi, and Leendert WN van der Torre. “Normative frame- work for normative system change”. In:The 8th International Joint Conference on Autonomous Agents and Multiagent Systems (AAMAS 2009), Budapest, Hungary, May 10-15, 2009, Volume 1. IFAAMAS. 2009
work page 2009
-
[40]
Attributing Mental Attitudes to Roles: The Agent Metaphor Applied to e-Trade Organizations
Guido Boella and Leendert WN van der Torre. “Attributing Mental Attitudes to Roles: The Agent Metaphor Applied to e-Trade Organizations”. In: (2002)
work page 2002
-
[41]
Attributing mental attitudes to nor- mative systems
Guido Boella and Leendert WN van der Torre. “Attributing mental attitudes to nor- mative systems”. In:Proceedings of the second international joint conference on Au- tonomous agents and multiagent systems. 2003, pp. 942–943
work page 2003
-
[42]
Local policies for the control of virtual communities
Guido Boella and Leendert WN van der Torre. “Local policies for the control of virtual communities”. In:Proceedings IEEE/WIC International Conference on Web Intelli- gence (WI 2003). IEEE. 2003, pp. 161–167
work page 2003
-
[43]
Contracts as legal institutions in organizations of autonomous agents
Guido Boella, Leendert WN van der Torre, et al. “Contracts as legal institutions in organizations of autonomous agents”. In:AAMAS. Vol. 4. 2004, pp. 948–955
work page 2004
-
[44]
Game theoretic normative reasoning
Guido Boella and Leendert WN van der Torre. “Game theoretic normative reasoning”. In:Proceedings of the Ninth International Conference on Artificial Intelligence and Law (ICAIL). ACM. 2004, pp. 217–224. 144
work page 2004
-
[45]
Groups as agents with mental attitudes
Guido Boella, Leendert WN van der Torre, et al. “Groups as agents with mental attitudes”. In:International Conference on Autonomous Agents: Proceedings of the Third International Joint Conference on Autonomous Agents and Multiagent Systems- . Vol. 2. 2004, pp. 964–971
work page 2004
-
[46]
Structuring organizations by means of roles using the agent metaphor
Guido Boella and Leendert WN van der Torre. “Structuring organizations by means of roles using the agent metaphor”. In:dagli Oggetti agli Agenti(2004), p. 93
work page 2004
-
[47]
Constitutive norms in the design of normative multiagent systems
Guido Boella and Leendert WN van der Torre. “Constitutive norms in the design of normative multiagent systems”. In:International Workshop on Computational Logic in Multi-Agent Systems. Springer. 2005, pp. 303–319
work page 2005
-
[48]
A game theoretic approach to con- tracts in multiagent systems
Guido Boella and Leendert WN van der Torre. “A game theoretic approach to con- tracts in multiagent systems”. In:IEEE Transactions on Systems, Man, and Cyber- netics, Part C (Applications and Reviews)36.1 (2006), pp. 68–79
work page 2006
-
[49]
UCI Machine Learning Repository
Marko Bohanec.Car Evaluation. UCI Machine Learning Repository. DOI: https://doi.org/10.24432/C5JP48. 1997
-
[50]
Norms in artificial decision making
Magnus Boman. “Norms in artificial decision making”. In:Artificial Intelligence and Law7.1 (1999), pp. 17–35
work page 1999
-
[51]
A Basic Framework for Explanations in Argumen- tation
Annemarie Borg and Floris Bex. “A Basic Framework for Explanations in Argumen- tation”. In:IEEE Intelligent SystemsPP (Jan. 2021), pp. 1–1.doi:10.1109/MIS. 2021.3053102
work page doi:10.1109/mis 2021
-
[52]
UCI Machine Learning Repository
Shiva Borzooei and Aidin Tarokhian.Differentiated Thyroid Cancer Recurrence. UCI Machine Learning Repository. DOI: https://doi.org/10.24432/C5632J. 2023
-
[53]
Agentic AI and Multiagentic: Are We Reinventing the Wheel?
Vicent Botti. “Agentic AI and Multiagentic: Are We Reinventing the Wheel?” In: arXiv preprint arXiv:2506.01463(2025)
-
[54]
An anatomy of moral responsibility
Matthew Braham and Martin Van Hees. “An anatomy of moral responsibility”. In: Mind121.483 (2012), pp. 601–634. 145
work page 2012
-
[55]
Tom B. Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared Kaplan, Pra- fulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, Sandhini Agarwal, Ariel Herbert-Voss, Gretchen Krueger, Tom Henighan, Rewon Child, Aditya Ramesh, Daniel M. Ziegler, Jeffrey Wu, Clemens Winter, Christopher Hesse, Mark Chen, Eric Sigler, Mateusz Litw...
work page internal anchor Pith review Pith/arXiv arXiv 2020
-
[56]
David Byrd. “Learning not to spoof”. In:Proceedings of the Third ACM International Conference on AI in Finance. 2022, pp. 139–147
work page 2022
-
[57]
Five years of argument mining: A data-driven analysis
Elena Cabrio and Serena Villata. “Five years of argument mining: A data-driven analysis.” In:IJCAI. Vol. 18. 2018, pp. 5427–5433
work page 2018
-
[58]
A discussion game for grounded semantics
Martin Caminada. “A discussion game for grounded semantics”. In:Theory and Appli- cations of Formal Argumentation: Third International Workshop, TAFA 2015, Buenos Aires, Argentina, July 25-26, 2015, Revised Selected Papers 3. Springer. 2015, pp. 59– 73
work page 2015
-
[59]
A case-based reasoning ap- proach for norm adaptation
Jordi Campos, Maite L´ opez-S´ anchez, and Marc Esteva. “A case-based reasoning ap- proach for norm adaptation”. In:Hybrid Artificial Intelligence Systems: 5th Interna- tional Conference, HAIS 2010, San Sebasti´ an, Spain, June 23-25, 2010. Proceedings, Part II 5. Springer. 2010, pp. 168–176
work page 2010
-
[60]
Adaptive deterrence sanctions in a normative framework
Henrique Lopes Cardoso and Eug´ enio Oliveira. “Adaptive deterrence sanctions in a normative framework”. In:2009 IEEE/WIC/ACM International Joint Conference on Web Intelligence and Intelligent Agent Technology. Vol. 2. IEEE. 2009, pp. 36–43
work page 2009
-
[61]
Delib- erative normative agents: Principles and architecture
Cristiano Castelfranchi, Frank Dignum, Catholijn M Jonker, and Jan Treur. “Delib- erative normative agents: Principles and architecture”. In:International workshop on agent theories, architectures, and languages. Springer. 1999, pp. 364–378
work page 1999
-
[62]
On the acceptability of ar- guments in bipolar argumentation frameworks
Claudette Cayrol and Marie-Christine Lagasquie-Schiex. “On the acceptability of ar- guments in bipolar argumentation frameworks”. In:European Conference on Symbolic 146 and Quantitative Approaches to Reasoning and Uncertainty. Springer. 2005, pp. 378– 389
work page 2005
-
[63]
R´ emy Chaput. “Learning behaviours aligned with moral values in a multi-agent sys- tem: guiding reinforcement learning with symbolic judgments”. PhD thesis. Universit´ e Claude Bernard-Lyon I, 2022
work page 2022
-
[64]
Exploiting domain knowledge to improve norm synthesis
George Christelis, Michael Rovatsos, and Ronald PA Petrick. “Exploiting domain knowledge to improve norm synthesis”. In:Proceedings of the 9th International Con- ference on Autonomous Agents and Multiagent Systems: volume 1-Volume 1. Citeseer. 2010, pp. 831–838
work page 2010
-
[65]
William W Cohen. “Fast effective rule induction”. In:Machine learning proceedings
-
[66]
Elsevier, 1995, pp. 115–123. [66]Congressional Voting Records. UCI Machine Learning Repository. DOI: https://doi.org/10.24432/C5C01P. 1987
-
[67]
Normative design using inductive learning
Domenico Corapi, Alessandra Russo, Marina De Vos, Julian Padget, and Ken Satoh. “Normative design using inductive learning”. In:Theory and Practice of Logic Pro- gramming11.4-5 (2011), pp. 783–799
work page 2011
-
[68]
Corinna Cortes and Vladimir Vapnik. “Support-vector networks”. In:Machine learn- ing20 (1995), pp. 273–297
work page 1995
-
[69]
Weighted Attacks in Argumentation Frameworks
Sylvie Coste-Marquis, S´ ebastien Konieczny, Pierre Marquis, and Mohand Akli Ouali. “Weighted Attacks in Argumentation Frameworks.” In:KR. 2012
work page 2012
-
[70]
Identifying Norms from Observation Using MCMC Sampling
Stephen Cranefield and Ashish Dhiman. “Identifying Norms from Observation Using MCMC Sampling.” In:IJCAI. 2021, pp. 118–124
work page 2021
-
[71]
A Bayesian approach to norm identification
Stephen Cranefield, Felipe Meneguzzi, Nir Oren, and Bastin Tony Roy Savarimuthu. “A Bayesian approach to norm identification”. In:ECAI 2016. Ios Press, 2016, pp. 622– 629. 147
work page 2016
-
[72]
Mining software repositories for social norms
Hoa Khanh Dam, Bastin Tony Roy Savarimuthu, Daniel Avery, and Aditya Ghose. “Mining software repositories for social norms”. In:2015 IEEE/ACM 37th IEEE In- ternational Conference on Software Engineering. Vol. 2. IEEE. 2015, pp. 627–630
work page 2015
-
[73]
The complexity of norm synthesis and revision
Davide Dell’Anna, Natasha Alechina, Fabiano Dalpiaz, Mehdi Dastani, Maarten L¨ offler, and Brian Logan. “The complexity of norm synthesis and revision”. In:Inter- national Workshop on Coordination, Organizations, Institutions, Norms, and Ethics for Governance of Multi-Agent Systems. Springer. 2022, pp. 38–53
work page 2022
-
[74]
To- wards socially sophisticated BDI agents
Frank Dignum, David Morley, Elizabeth A Sonenberg, and Lawrence Cavedon. “To- wards socially sophisticated BDI agents”. In:Proceedings fourth international confer- ence on multiagent systems. IEEE. 2000, pp. 111–118
work page 2000
-
[75]
Visual ex- planations for defence in abstract argumentation
Sylvie Doutre, Th´ eo Duchatelle, and Marie-Christine Lagasquie-Schiex. “Visual ex- planations for defence in abstract argumentation”. In:International Conference on Autonomous Agents and Multiagent Systems (AAMAS). ACM. 2023, pp. 2346–2348
work page 2023
-
[76]
Phan Minh Dung. “On the acceptability of arguments and its fundamental role in nonmonotonic reasoning, logic programming and n-person games”. In:Artificial intel- ligence77.2 (1995), pp. 321–357
work page 1995
-
[77]
Design and evaluation of norm-aware agents based on Normative Markov Decision Processes
Moser Silva Fagundes, Sascha Ossowski, Jes´ us Cerquides, and Pablo Noriega. “Design and evaluation of norm-aware agents based on Normative Markov Decision Processes”. In:International Journal of Approximate Reasoning78 (2016), pp. 33–61
work page 2016
-
[78]
On computing explanations in abstract argumenta- tion
Xiuyi Fan and Francesca Toni. “On computing explanations in abstract argumenta- tion”. In:ECAI 2014. IOS Press, 2014, pp. 1005–1006
work page 2014
-
[79]
NLP Techniques for Normative Mining
Gabriela Ferraro and Ho-Pun Lam. “NLP Techniques for Normative Mining.” In: FLAP8.4 (2021), pp. 941–974
work page 2021
-
[80]
R. A. Fisher.Iris. UCI Machine Learning Repository. DOI: https://doi.org/10.24432/C56C76. 1936. 148
discussion (0)
Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.