Order:
See also
  1.  53
    Probing the quantitative–qualitative divide in probabilistic reasoning.Duligur Ibeling, Thomas Icard, Krzysztof Mierzewski & Milan Mossé - forthcoming - Annals of Pure and Applied Logic.
  2. Is Causal Reasoning Harder Than Probabilistic Reasoning?Milan Mossé, Duligur Ibeling & Thomas Icard - 2024 - Review of Symbolic Logic 17 (1):106-131.
    Many tasks in statistical and causal inference can be construed as problems of entailment in a suitable formal language. We ask whether those problems are more difficult, from a computational perspective, for causal probabilistic languages than for pure probabilistic (or “associational”) languages. Despite several senses in which causal reasoning is indeed more complex—both expressively and inferentially—we show that causal entailment (or satisfiability) problems can be systematically and robustly reduced to purely probabilistic problems. Thus there is no jump in computational complexity. (...)
    Direct download (3 more)  
     
    Export citation  
     
    Bookmark   1 citation  
  3. Social Choice for AI Alignment: Dealing with Diverse Human Feedback.Vincent Conitzer, Rachel Freedman, Jobst Heitzig, Wesley H. Holliday, Bob M. Jacobs, Nathan Lambert, Milan Mosse, Eric Pacuit, Stuart Russell, Hailey Schoelkopf, Emanuel Tewolde & William S. Zwicker - manuscript
    Foundation models such as GPT-4 are fine-tuned to avoid unsafe or otherwise problematic behavior, so that, for example, they refuse to comply with requests for help with committing crimes or with producing racist text. One approach to fine-tuning, called reinforcement learning from human feedback, learns from humans' expressed preferences over multiple outputs. Another approach is constitutional AI, in which the input from humans is a list of high-level principles. But how do we deal with potentially diverging input from humans? How (...)
    Direct download  
     
    Export citation  
     
    Bookmark