0
views
0
recommends
+1 Recommend
0 collections
    0
    shares
      • Record: found
      • Abstract: found
      • Article: found
      Is Open Access

      Can AI Rely on the Systematicity of Truth? The Challenge of Modelling Normative Domains

      research-article

      Read this article at

      Bookmark
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Abstract

          A key assumption fuelling optimism about the progress of Large Language Models (LLMs) in accurately and comprehensively modelling the world is that the truth is systematic: true statements about the world form a whole that is not just consistent, in that it contains no contradictions, but coherent, in that the truths are inferentially interlinked. This holds out the prospect that LLMs might in principle rely on that systematicity to fill in gaps and correct inaccuracies in the training data: consistency and coherence promise to facilitate progress towards comprehensiveness in an LLM’s representation of the world. However, philosophers have identified compelling reasons to doubt that the truth is systematic across all domains of thought, arguing that in normative domains, in particular, the truth is largely asystematic. I argue that insofar as the truth in normative domains is asystematic, this renders it correspondingly harder for LLMs to make progress, because they cannot then leverage the systematicity of truth. And the less LLMs can rely on the systematicity of truth, the less we can rely on them to do our practical deliberation for us, because the very asystematicity of normative domains requires human agency to play a greater role in practical thought.

          Related collections

          Most cited references67

          • Record: found
          • Abstract: not found
          • Book: not found

          How the Laws of Physics Lie

            Bookmark
            • Record: found
            • Abstract: found
            • Article: not found

            Training language models to follow instructions with human feedback

            Making language models bigger does not inherently make them better at following a user's intent. For example, large language models can generate outputs that are untruthful, toxic, or simply not helpful to the user. In other words, these models are not aligned with their users. In this paper, we show an avenue for aligning language models with user intent on a wide range of tasks by fine-tuning with human feedback. Starting with a set of labeler-written prompts and prompts submitted through the OpenAI API, we collect a dataset of labeler demonstrations of the desired model behavior, which we use to fine-tune GPT-3 using supervised learning. We then collect a dataset of rankings of model outputs, which we use to further fine-tune this supervised model using reinforcement learning from human feedback. We call the resulting models InstructGPT. In human evaluations on our prompt distribution, outputs from the 1.3B parameter InstructGPT model are preferred to outputs from the 175B GPT-3, despite having 100x fewer parameters. Moreover, InstructGPT models show improvements in truthfulness and reductions in toxic output generation while having minimal performance regressions on public NLP datasets. Even though InstructGPT still makes simple mistakes, our results show that fine-tuning with human feedback is a promising direction for aligning language models with human intent.
              Bookmark
              • Record: found
              • Abstract: not found
              • Book: not found

              Ethics Without Principles

                Bookmark

                Author and article information

                Contributors
                matthieu.queloz@unibe.ch
                Journal
                Philos Technol
                Philos Technol
                Philosophy & Technology
                Springer Netherlands (Dordrecht )
                2210-5433
                2210-5441
                13 March 2025
                13 March 2025
                2025
                : 38
                : 1
                : 34
                Affiliations
                Institute of Philosophy, University of Bern, ( https://ror.org/02k7v4d05) Laenggassstrasse 49a, 3012 Bern, Switzerland
                Author information
                http://orcid.org/0000-0001-6644-9992
                Article
                864
                10.1007/s13347-025-00864-x
                11906541
                40093820
                595ede9b-668d-4290-becb-0f4e9dfb434e
                © The Author(s) 2025

                Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.

                History
                : 15 November 2024
                : 20 February 2025
                Funding
                Funded by: FundRef http://dx.doi.org/10.13039/501100001711, Schweizerischer Nationalfonds zur Förderung der Wissenschaftlichen Forschung;
                Funded by: University of Bern
                Categories
                Research Article
                Custom metadata
                © Springer Nature B.V. 2025

                Philosophy of science
                artificial intelligence,language models,normativity,value pluralism,conflicts of values,agency,ai ethics,hard choices,consistency,coherence,authenticity

                Comments

                Comment on this article