Figure 7 from "I'd Like to Have an Argument, Please": Argumentative Reasoning in Large Language Models | Semantic Scholar (2024)

Skip to search formSkip to main contentSkip to account menu

Semantic ScholarSemantic Scholar's Logo
  • Corpus ID: 263310788
@inproceedings{Wynter2023IdLT, title={"I'd Like to Have an Argument, Please": Argumentative Reasoning in Large Language Models}, author={Adrian de Wynter and Tommy Yuan}, year={2023}, url={https://api.semanticscholar.org/CorpusID:263310788}}
  • Adrian de Wynter, Tommy Yuan
  • Published 29 September 2023
  • Computer Science, Linguistics

This work evaluates two large language models (LLMs) ability to perform argumentative reasoning, and finds that scoring-wise the LLMs match or surpass the SOTA in AM and APE, and under certain I/O abstractions LLMs perform well, even beating chain-of-thought--the authors call this symbolic prompting.

3 Citations

Background Citations

2

Results Citations

1

Figures and Tables from this paper

  • figure 1
  • table 1
  • figure 2
  • table 2
  • figure 3
  • figure 4
  • figure 5
  • figure 6
  • figure 7
  • figure 8
  • figure 9
  • figure 10

3 Citations

ArgMed-Agents: Explainable Clinical Decision Reasoning with Large Language Models via Argumentation Schemes
    Shengxin HongLiang XiaoXin ZhangJian-Xing Chen

    Computer Science, Medicine

    ArXiv

  • 2024
  • 1
Will GPT-4 Run DOOM?
    Adrian de Wynter

    Computer Science

    ArXiv

  • 2024

It is found that GPT-4 can play the game to a passable degree: it is able to manipulate doors, combat enemies, and perform pathing, but more complex prompting strategies involving multiple model calls provide better results.

Computational Argumentation-based Chatbots: a Survey
    Federico CastagnaNadin KökciyanI. SassoonSimon ParsonsE. Sklar

    Computer Science

    ArXiv

  • 2024

The present survey sifts through the literature to review papers concerning this kind of argumentation-based bot, drawing conclusions about the benefits and drawbacks that this approach entails in comparison with standard chatbots, while also envisaging possible future development and integration with the Transformer-based architecture and state-of-the-art Large Language models.

35 References

Language Models Are Greedy Reasoners: A Systematic Formal Analysis of Chain-of-Thought
    Abulhair SaparovHe He

    Computer Science

    ICLR

  • 2023

This work presents a new synthetic question-answering dataset called PrOntoQA, where each example is generated from a synthetic world model represented in first-order logic, and shows that LLMs are quite capable of making correct individual deduction steps, and so are generally capable of reasoning, even in fictional contexts.

Have my arguments been replied to? Argument Pair Extraction as Machine Reading Comprehension
    Jianzhu BaoJingyi SunQinglin ZhuRuifeng Xu

    Computer Science, Linguistics

    ACL

  • 2022

This framework enables these two phases to be jointly trained in a single MRC model, thereby maximizing the mutual benefits of them and outperforming the state-of-the-art method.

  • 9
  • PDF
Explainable Unsupervised Argument Similarity Rating with Abstract Meaning Representation and Conclusion Generation
    J. OpitzP. HeinischPhilipp WiesenbachP. CimianoA. Frank

    Computer Science

    ARGMINING

  • 2021

It is shown that Abstract Meaning Representation (AMR) graphs can be useful for representing arguments, and that novel AMR graph metrics can offer explanations for argument similarity ratings and make argument similarity judgements more interpretable and may even support argument quality judgements.

  • 14
  • PDF
Do Prompt-Based Models Really Understand the Meaning of Their Prompts?
    Albert WebsonEllie Pavlick

    Computer Science

    NAACL

  • 2022

It is found that models can learn just as fast with many prompts that are intentionally irrelevant or even pathologically misleading as they do with instructively “good” prompts, and instruction-tuned models often produce good predictions with irrelevant and misleading prompts even at zero shots.

ReviewerGPT? An Exploratory Study on Using Large Language Models for Paper Reviewing
    Ryan LiuNihar B. Shah

    Computer Science, Linguistics

    ArXiv

  • 2023

It is thought that LLMs have a promising use as reviewing assistants for specific reviewing tasks, but not for complete evaluations of papers or proposals.

  • 28
  • PDF
Large Language Models are Zero-Shot Reasoners
    Takeshi KojimaS. GuMachel ReidYutaka MatsuoYusuke Iwasawa

    Computer Science

    NeurIPS

  • 2022

Experimental results demonstrate that the Zero-shot-CoT, using the same single prompt template, significantly outperforms zero-shot LLM performances on diverse benchmark reasoning tasks including arithmetics, symbolic reasoning, and other logical reasoning tasks, without any hand-crafted few-shot examples.

Chain of Thought Prompting Elicits Reasoning in Large Language Models
    Jason WeiXuezhi Wang Denny Zhou

    Computer Science

    NeurIPS

  • 2022

Experiments on three large language models show that chain of thought prompting improves performance on a range of arithmetic, commonsense, and symbolic reasoning tasks.

An Evaluation on Large Language Model Outputs: Discourse and Memorization
    Adrian de WynterXun WangAlex SokolovQilong GuSi-Qing Chen

    Computer Science, Linguistics

    Nat. Lang. Process. J.

  • 2023
Is GPT-4 a Good Data Analyst?
    Liying ChengXingxuan LiLidong Bing

    Computer Science

    EMNLP

  • 2023

This work regards GPT-4 as a data analyst to perform end-to-end data analysis with databases from a wide range of domains and designs several task-specific evaluation metrics to systematically compare the performance between several professional human data analysts and G PT-4.

Do Language Models Plagiarize?
    Jooyoung LeeThai LeJinghui ChenDongwon Lee

    Computer Science, Linguistics

    WWW

  • 2023

Study of three types of plagiarism among GPT-2 generated texts, in comparison to its training data, and the plagiarism patterns of fine-tuned LMs with domain-specific corpora which are extensively used in practice suggest that the practicality of current LMs in mission-critical writing tasks is questioned.

...

...

Related Papers

Showing 1 through 3 of 0 Related Papers

    Figure 7 from "I'd Like to Have an Argument, Please": Argumentative Reasoning in Large Language Models | Semantic Scholar (13)

    Figure 7. AM symbolic (indices) prompt with one exemplar and CoT. Refer to Figure 6 for a longer version of the exemplar. This prompt performs step-by-step reasoning on AM by following a templatized generation and…

    Published in 2023

    "I'd Like to Have an Argument, Please": Argumentative Reasoning in Large Language Models

    Adrian de WynterTommy Yuan

    Figure 9 of 12

    Figure 7 from "I'd Like to Have an Argument, Please": Argumentative Reasoning in Large Language Models | Semantic Scholar (2024)

    References

    Top Articles
    Latest Posts
    Article information

    Author: Kareem Mueller DO

    Last Updated:

    Views: 5884

    Rating: 4.6 / 5 (66 voted)

    Reviews: 81% of readers found this page helpful

    Author information

    Name: Kareem Mueller DO

    Birthday: 1997-01-04

    Address: Apt. 156 12935 Runolfsdottir Mission, Greenfort, MN 74384-6749

    Phone: +16704982844747

    Job: Corporate Administration Planner

    Hobby: Mountain biking, Jewelry making, Stone skipping, Lacemaking, Knife making, Scrapbooking, Letterboxing

    Introduction: My name is Kareem Mueller DO, I am a vivacious, super, thoughtful, excited, handsome, beautiful, combative person who loves writing and wants to share my knowledge and understanding with you.