<?xml version="1.0" encoding="UTF-8" standalone="no"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" article-type="brief-report">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Big Data</journal-id>
<journal-title>Frontiers in Big Data</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Big Data</abbrev-journal-title>
<issn pub-type="epub">2624-909X</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fdata.2019.00052</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Big Data</subject>
<subj-group>
<subject>Perspective</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Neural-Symbolic Argumentation Mining: An Argument in Favor of Deep Learning and Reasoning</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes">
<name><surname>Galassi</surname> <given-names>Andrea</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="corresp" rid="c001"><sup>&#x0002A;</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/819314/overview"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Kersting</surname> <given-names>Kristian</given-names></name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/321053/overview"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Lippi</surname> <given-names>Marco</given-names></name>
<xref ref-type="aff" rid="aff3"><sup>3</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/398480/overview"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Shao</surname> <given-names>Xiaoting</given-names></name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/882731/overview"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Torroni</surname> <given-names>Paolo</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/134807/overview"/>
</contrib>
</contrib-group>
<aff id="aff1"><sup>1</sup><institution>Department of Computer Science and Engineering, University of Bologna</institution>, <addr-line>Bologna</addr-line>, <country>Italy</country></aff>
<aff id="aff2"><sup>2</sup><institution>Computer Science Department and Centre for Cognitive Science, TU Darmstadt</institution>, <addr-line>Darmstadt</addr-line>, <country>Germany</country></aff>
<aff id="aff3"><sup>3</sup><institution>Department of Sciences and Methods for Engineering, University of Modena and Reggio Emilia</institution>, <addr-line>Reggio Emilia</addr-line>, <country>Italy</country></aff>
<author-notes>
<fn fn-type="edited-by"><p>Edited by: Balaraman Ravindran, Indian Institute of Technology Madras, India</p></fn>
<fn fn-type="edited-by"><p>Reviewed by: Elena Bellodi, University of Ferrara, Italy; Frank Zenker, Bo&#x0011F;azi&#x000E7;i University, Turkey</p></fn>
<corresp id="c001">&#x0002A;Correspondence: Andrea Galassi <email>a.galassi&#x00040;unibo.it</email></corresp>
<fn fn-type="other" id="fn001"><p>This article was submitted to Machine Learning and Artificial Intelligence, a section of the journal Frontiers in Big Data</p></fn></author-notes>
<pub-date pub-type="epub">
<day>22</day>
<month>01</month>
<year>2020</year>
</pub-date>
<pub-date pub-type="collection">
<year>2019</year>
</pub-date>
<volume>2</volume>
<elocation-id>52</elocation-id>
<history>
<date date-type="received">
<day>21</day>
<month>10</month>
<year>2019</year>
</date>
<date date-type="accepted">
<day>31</day>
<month>12</month>
<year>2019</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x000A9; 2020 Galassi, Kersting, Lippi, Shao and Torroni.</copyright-statement>
<copyright-year>2020</copyright-year>
<copyright-holder>Galassi, Kersting, Lippi, Shao and Torroni</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p></license>
</permissions>
<abstract><p>Deep learning is bringing remarkable contributions to the field of argumentation mining, but the existing approaches still need to fill the gap toward performing advanced reasoning tasks. In this position paper, we posit that neural-symbolic and statistical relational learning could play a crucial role in the integration of symbolic and sub-symbolic methods to achieve this goal.</p></abstract> <kwd-group>
<kwd>neural symbolic learning</kwd>
<kwd>argumentation mining</kwd>
<kwd>probabilistic logic programming</kwd>
<kwd>integrative AI</kwd>
<kwd>DeepProbLog</kwd>
<kwd>Ground-Specific Markov Logic Networks</kwd>
</kwd-group>
<contract-num rid="cn001">825619</contract-num>
<contract-num rid="cn002">KE 1686/3-1, SPP 1999</contract-num>
<contract-sponsor id="cn001">Horizon 2020<named-content content-type="fundref-id">10.13039/501100007601</named-content></contract-sponsor>
<contract-sponsor id="cn002">Deutsche Forschungsgemeinschaft<named-content content-type="fundref-id">10.13039/501100001659</named-content></contract-sponsor>
<counts>
<fig-count count="2"/>
<table-count count="0"/>
<equation-count count="7"/>
<ref-count count="46"/>
<page-count count="6"/>
<word-count count="4556"/>
</counts>
</article-meta>
</front>
<body>
<sec sec-type="intro" id="s1">
<title>1. Introduction</title>
<p>The goal of argumentation mining (AM) is to automatically extract arguments and their relations from a given document (Lippi and Torroni, <xref ref-type="bibr" rid="B30">2016</xref>). The majority of AM systems follows a pipeline scheme, starting with simpler tasks, such as argument component detection, down to more complex tasks, such as argumentation structure prediction. Recent years have seen the development of a large number of techniques in this area, on the wake of the advancements produced by deep learning on the whole research field of natural language processing (NLP). Yet, it is widely recognized that the existing AM systems still have a large margin of improvement, as good results have been obtained with some genres where prior knowledge on the structure of the text eases some AM tasks, but other genres, such as legal cases and social media documents still require more work (Cabrio and Villata, <xref ref-type="bibr" rid="B6">2018</xref>). Performing and understanding argumentation requires advanced reasoning capabilities, which are natural human skills, but are difficult to learn for a machine. Understanding whether a given piece of evidence supports a given claim, or whether two claims attack each other, are complex problems that humans can address thanks to their ability to exploit commonsense knowledge, and to perform reasoning and inference. Despite the remarkable impact of deep neural networks in NLP, we argue that these techniques alone will not suffice to address such complex issues.</p>
<p>We envisage that a significant advancement in AM could come from the combination of symbolic and sub-symbolic approaches, such as those developed in the Neural Symbolic (NeSy) (Garcez et al., <xref ref-type="bibr" rid="B19">2015</xref>) or Statistical Relational Learning (SRL) (Getoor and Taskar, <xref ref-type="bibr" rid="B21">2007</xref>; De Raedt et al., <xref ref-type="bibr" rid="B12">2016</xref>; Kordjamshidi et al., <xref ref-type="bibr" rid="B25">2018</xref>) communities. This issue is also widely recognized as one of the major challenges for the whole field of artificial intelligence in the coming years (LeCun et al., <xref ref-type="bibr" rid="B27">2015</xref>).</p>
<p>In computational argumentation, structured arguments have been studied and formalized for decades using models that can be expressed in a logic framework (Bench-Capon and Dunne, <xref ref-type="bibr" rid="B3">2007</xref>). At the same time, AM has rapidly evolved by exploiting state-of-the-art neural architectures coming from deep learning. So far, these two worlds have progressed largely independently of each other. Only recently, a few works have taken some steps toward the integration of such methods, by applying techniques combining sub-symbolic classifiers with knowledge expressed in the form of rules and constraints to AM. For instance, Niculae et al. (<xref ref-type="bibr" rid="B33">2017</xref>) adopted structured support vector machines and recurrent neural networks to collectively classify argument components and their relations in short documents, by hard-coding contextual dependencies and constraints of the argument model in a factor graph. A joint inference approach for argument component classification and relation identification was instead proposed by Persing and Ng (<xref ref-type="bibr" rid="B34">2016</xref>), following a pipeline scheme where integer linear programming is used to enforce mathematical constraints on the outcomes of a first-stage set of classifiers. More recently, Cocarascu and Toni (<xref ref-type="bibr" rid="B9">2018</xref>) combined a deep network for relation extraction with an argumentative reasoning system that computes the dialectical strength of arguments, for the task of determining whether a review is truthful or deceptive.</p>
<p>We propose to exploit the potential of both symbolic and sub-symbolic approaches for AM, by combining both results in systems that are capable of modeling knowledge and constraints with a logic formalism, while maintaining the computational power of deep networks. Differently from existing approaches, we advocate the use of a logic-based language for the definition of contextual dependencies and constraints, independently of the structure of the underlying classifiers. Most importantly, the approaches we outline do not exploit a pipeline scheme, but rather perform joint detection of argument components and relations through a single learning process.</p>
</sec>
<sec id="s2">
<title>2. Modeling Argumentation With Probabilistic Logic</title>
<p>Computational argumentation is concerned with modeling and analyzing argumentation in the computational settings of artificial intelligence (Bench-Capon and Dunne, <xref ref-type="bibr" rid="B3">2007</xref>; Simari and Rahwan, <xref ref-type="bibr" rid="B41">2009</xref>). The formalization of arguments is usually addressed at two levels. At the argument level, the definition of formal languages for representing knowledge and specifying how arguments and counterarguments can be constructed from that knowledge is the domain of <italic>structured argumentation</italic> (Besnard et al., <xref ref-type="bibr" rid="B4">2014</xref>). In structured argumentation, the premises and claim of the argument are made explicit, and their relationships are formally defined. However, when the discourse consists of multiple arguments, such arguments may conflict with one another and result in logical inconsistencies. A typical way of dealing with such inconsistencies is to identify sets of arguments that are mutually consistent and that collectively defeat their &#x0201C;attackers.&#x0201D; One way to do that is to abstract away from the internal structure of arguments and focus on the higher-level relations among arguments: a conceptual framework known as <italic>abstract argumentation</italic> (Dung, <xref ref-type="bibr" rid="B13">1995</xref>).</p>
<p>Similarly to structured argumentation, AM too builds on the definition of an argument model, and aims to identify parts of the input text that can be interpreted as argument components (Lippi and Torroni, <xref ref-type="bibr" rid="B30">2016</xref>). For example, if we take a basic claim/evidence argument model, possible tasks could be claim detection (Aharoni et al., <xref ref-type="bibr" rid="B1">2014</xref>; Lippi and Torroni, <xref ref-type="bibr" rid="B29">2015</xref>), evidence detection (Rinott et al., <xref ref-type="bibr" rid="B37">2015</xref>), and the prediction of links between claim and evidence (Niculae et al., <xref ref-type="bibr" rid="B33">2017</xref>; Galassi et al., <xref ref-type="bibr" rid="B17">2018</xref>). However, in structured argumentation the formalization of the model is the basis for an inferential process, whereby conclusions can be obtained starting from premises. In AM, instead, an argument model is usually defined in order to identify the target classes, and in some isolated cases to express relations, for instance among argument components (Niculae et al., <xref ref-type="bibr" rid="B33">2017</xref>), but not for producing inferences that could help the AM tasks.</p>
<p>The languages of structured argumentation are logic-based. An influential structured argumentation system is <italic>deductive argumentation</italic> (Besnard and Hunter, <xref ref-type="bibr" rid="B5">2001</xref>), where premises are logical formulae, which entail a claim, and entailment may be specified from a range of base logics, such as classical logic or modal logic. In <italic>assumption-based argumentation</italic> (Dung et al., <xref ref-type="bibr" rid="B14">2009</xref>) instead arguments correspond to assumptions, which like in deductive systems prove a claim, and attacks are obtained via a notion of contrary assumptions. Another powerful framework is <italic>defeasible logic programming</italic> (DeLP) (Garc&#x000ED;a and Simari, <xref ref-type="bibr" rid="B20">2004</xref>), where claims can be supported using strict or defeasible rules, and an argument supporting a claim is warranted if it defeats all its counterarguments. For example, that a cephalopod is a mollusc could be expressed by a strict rule, such as:</p>
<disp-formula id="E1"><mml:math id="M1"><mml:mtable columnalign="left"><mml:mtr><mml:mtd><mml:mi>m</mml:mi><mml:mi>o</mml:mi><mml:mi>l</mml:mi><mml:mi>l</mml:mi><mml:mi>u</mml:mi><mml:mi>s</mml:mi><mml:mi>c</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>&#x02190;</mml:mo><mml:mi>c</mml:mi><mml:mi>e</mml:mi><mml:mi>p</mml:mi><mml:mi>h</mml:mi><mml:mi>a</mml:mi><mml:mi>l</mml:mi><mml:mi>o</mml:mi><mml:mi>p</mml:mi><mml:mi>o</mml:mi><mml:mi>d</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>as these notions belong to an artificially defined, incontrovertible taxonomy. However, since in nature not all molluscs have a shell, and actually cephalopods are molluscs without a shell, rules used to conclude that a given specimen has or does not have a shell are best defined as defeasible. For instance, one could say:</p>
<graphic xlink:href="fdata-02-00052-e0001.tif"/>
<p>where <inline-graphic xlink:href="fdata-02-00052-i0001.tif"/> denotes defeasible inference.</p>
<p>The choice of logic notwithstanding, rules offer a convenient way to describe argumentative inference. Moreover, depending on the application domain, the document genre, and the employed argument model, different constraints and rules can be enforced on the structure of the underlying network of arguments. For example, if we adopt a DeLP-like approach, strict rules can be used to define the relations among argument components, and defeasible rules to define context knowledge. For example, in a hypothetical claim-premise model, support relations may be defined exclusively between a premise and a claim. Such structural properties could be expressed by the following strict rules:</p>
<disp-formula id="E2"><mml:math id="M2"><mml:mtable columnalign="left"><mml:mtr><mml:mtd><mml:mtext>&#x000A0;&#x000A0;&#x000A0;&#x000A0;</mml:mtext><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>a</mml:mi><mml:mi>i</mml:mi><mml:mi>m</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>Y</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>&#x02190;</mml:mo><mml:mi>s</mml:mi><mml:mi>u</mml:mi><mml:mi>p</mml:mi><mml:mi>p</mml:mi><mml:mi>o</mml:mi><mml:mi>r</mml:mi><mml:mi>t</mml:mi><mml:mi>s</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>X</mml:mi><mml:mo>,</mml:mo><mml:mi>Y</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mi>p</mml:mi><mml:mi>r</mml:mi><mml:mi>e</mml:mi><mml:mi>m</mml:mi><mml:mi>i</mml:mi><mml:mi>s</mml:mi><mml:mi>e</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>&#x02190;</mml:mo><mml:mi>s</mml:mi><mml:mi>u</mml:mi><mml:mi>p</mml:mi><mml:mi>p</mml:mi><mml:mi>o</mml:mi><mml:mi>r</mml:mi><mml:mi>t</mml:mi><mml:mi>s</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>X</mml:mi><mml:mo>,</mml:mo><mml:mi>Y</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>whereby if <italic>X</italic> supports <italic>Y</italic>, then <italic>X</italic> is a claim and <italic>Y</italic> is a premise. As another abstract example, two claims based on the same premise may not attack each other:</p>
<disp-formula id="E3"><mml:math id="M3"><mml:mtable columnalign="left"><mml:mtr><mml:mtd><mml:mo>&#x0007E;</mml:mo><mml:mi>a</mml:mi><mml:mi>t</mml:mi><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>c</mml:mi><mml:mi>k</mml:mi><mml:mi>s</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>Y</mml:mi><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mi>Y</mml:mi><mml:mn>2</mml:mn></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>&#x02190;</mml:mo><mml:mi>s</mml:mi><mml:mi>u</mml:mi><mml:mi>p</mml:mi><mml:mi>p</mml:mi><mml:mi>o</mml:mi><mml:mi>r</mml:mi><mml:mi>t</mml:mi><mml:mi>s</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>X</mml:mi><mml:mo>,</mml:mo><mml:mi>Y</mml:mi><mml:mn>1</mml:mn></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>&#x02227;</mml:mo><mml:mi>s</mml:mi><mml:mi>u</mml:mi><mml:mi>p</mml:mi><mml:mi>p</mml:mi><mml:mi>o</mml:mi><mml:mi>r</mml:mi><mml:mi>t</mml:mi><mml:mi>s</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>X</mml:mi><mml:mo>,</mml:mo><mml:mi>Y</mml:mi><mml:mn>2</mml:mn></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>As an example of defeasible rules, consider instead the context information about a political debate, where a republican candidate, <italic>R</italic>, faces a democrat candidate, <italic>D</italic>. Then one may want to use the knowledge that <italic>R</italic>&#x00027;s claims and <italic>D</italic>&#x00027;s claims are likely to attack each other:</p>
<graphic xlink:href="fdata-02-00052-e0002.tif"/>
<p>where predicate <italic>auth</italic>(<italic>A, B</italic>) denotes that claim <italic>A</italic> was made by <italic>B</italic>. There exist many non-monotonic reasoning systems that integrate defeasible and strict inference. However, an alternative approach that may successfully reconcile the computational argumentation view and the AM view is offered by probabilistic logic programming (PLP). PLP combines the capability of logic to represent complex relations among entities with the capability of probability to model uncertainty over attributes and relations (Riguzzi, <xref ref-type="bibr" rid="B36">2018</xref>). In a PLP framework, such as PRISM (Sato and Kameya, <xref ref-type="bibr" rid="B38">1997</xref>), LPAD (Vennekens et al., <xref ref-type="bibr" rid="B42">2004</xref>), or ProbLog (Raedt et al., <xref ref-type="bibr" rid="B35">2007</xref>), defeasible rules may be expressed by rules with a probability label. For instance, in LPAD syntax, one could write:</p>
<disp-formula id="E4"><mml:math id="M4"><mml:mtable columnalign="left"><mml:mtr><mml:mtd><mml:mi>a</mml:mi><mml:mi>t</mml:mi><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>c</mml:mi><mml:mi>k</mml:mi><mml:mi>s</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>Y</mml:mi><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mi>Y</mml:mi><mml:mn>2</mml:mn></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>:</mml:mo><mml:mn>0</mml:mn><mml:mo>.</mml:mo><mml:mn>8</mml:mn><mml:mo>&#x02190;</mml:mo><mml:mi>a</mml:mi><mml:mi>u</mml:mi><mml:mi>t</mml:mi><mml:mi>h</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>Y</mml:mi><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mi>R</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>&#x02227;</mml:mo><mml:mi>r</mml:mi><mml:mi>e</mml:mi><mml:mi>p</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>R</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>&#x02227;</mml:mo><mml:mi>a</mml:mi><mml:mi>u</mml:mi><mml:mi>t</mml:mi><mml:mi>h</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>Y</mml:mi><mml:mn>2</mml:mn><mml:mo>,</mml:mo><mml:mi>D</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mtext>&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;</mml:mtext><mml:mo>&#x02227;</mml:mo><mml:mi>d</mml:mi><mml:mi>e</mml:mi><mml:mi>m</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>D</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>to express that the above rule holds in 80% of cases. In this example, 0.8 could be interpreted as a weight or score suggesting how likely the given inference rule is to hold. In more recent approaches, such weights could be learned directly from a collection of examples, for example by exploiting likelihood maximization in a <italic>learning from interpretations</italic> setting (Gutmann et al., <xref ref-type="bibr" rid="B23">2011</xref>) or by using a generalization of expectation maximization applied to logic programs (Bellodi and Riguzzi, <xref ref-type="bibr" rid="B2">2013</xref>).</p>
<p>From a higher-level perspective, rules could be exploited also to model more complex relations between arguments or even to encode argument schemes (Walton et al., <xref ref-type="bibr" rid="B43">2008</xref>), for example to assess whether an argument is defeated by another, on the basis of the strength of its premises and claims. This is an even more sophisticated reasoning task, which yet could be addressed within the same kind of framework described so far.</p>
</sec>
<sec id="s3">
<title>3. Combining Symbolic and Sub-symbolic Approaches</title>
<p>The usefulness of deep networks has been tested and proven in many NLP tasks, such as machine translation (Young et al., <xref ref-type="bibr" rid="B44">2018</xref>), sentiment analysis (Zhang et al., <xref ref-type="bibr" rid="B45">2018a</xref>), text classification (Conneau et al., <xref ref-type="bibr" rid="B10">2017</xref>; Zhang et al., <xref ref-type="bibr" rid="B46">2018b</xref>), relations extraction (Huang and Wang, <xref ref-type="bibr" rid="B24">2017</xref>), as well as in AM (Cocarascu and Toni, <xref ref-type="bibr" rid="B8">2017</xref>, <xref ref-type="bibr" rid="B9">2018</xref>; Daxenberger et al., <xref ref-type="bibr" rid="B11">2017</xref>; Galassi et al., <xref ref-type="bibr" rid="B17">2018</xref>; Lauscher et al., <xref ref-type="bibr" rid="B26">2018</xref>; Lugini and Litman, <xref ref-type="bibr" rid="B31">2018</xref>; Schulz et al., <xref ref-type="bibr" rid="B39">2018</xref>). While a straightforward approach to exploit domain knowledge in AM is to apply a set of hand-crafted rules on the output of some first stage classifier (such as a neural network), NeSy or SRL approaches can directly enforce (hard or soft) constraints <italic>during training</italic>, so that a solution that does not satisfy them is penalized, or even ruled out. Therefore, if a neural network is trained to classify argument components, and another one<xref ref-type="fn" rid="fn0001"><sup>1</sup></xref> is trained to detect links between them, additional global constraints can be enforced to adjust the weights of the networks toward admissible solutions, as the learning process advances. Systems like DeepProbLog (Manhaeve et al., <xref ref-type="bibr" rid="B32">2018</xref>), Logic Tensor Networks (Serafini and Garcez, <xref ref-type="bibr" rid="B40">2016</xref>), or Grounding-Specific Markov Logic Networks (GS-MLN) (Lippi and Frasconi, <xref ref-type="bibr" rid="B28">2009</xref>), to mention a few, enable such a scheme.</p>
<p>By way of illustration, we report how to implement one of the cases mentioned in section 2 with DeepProbLog and with GS-MLNs. By extending the ProbLog framework, DeepProbLog allows to introduce the following kind of construct:</p>
<disp-formula id="E5"><mml:math id="M5"><mml:mtable columnalign="left"><mml:mtr><mml:mtd><mml:mi>n</mml:mi><mml:mi>n</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>m</mml:mi><mml:mo>,</mml:mo><mml:mover accent="true"><mml:mrow><mml:mi>t</mml:mi></mml:mrow><mml:mo>&#x02192;</mml:mo></mml:mover><mml:mo>,</mml:mo><mml:mover accent="true"><mml:mrow><mml:mi>u</mml:mi></mml:mrow><mml:mo>&#x02192;</mml:mo></mml:mover></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>:</mml:mo><mml:mo>:</mml:mo><mml:mi>q</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>t</mml:mi></mml:mrow><mml:mo>&#x02192;</mml:mo></mml:mover><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi>u</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>;</mml:mo><mml:mo>&#x02026;</mml:mo><mml:mo>;</mml:mo><mml:mi>q</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>t</mml:mi></mml:mrow><mml:mo>&#x02192;</mml:mo></mml:mover><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi>u</mml:mi></mml:mrow><mml:mrow><mml:mi>n</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>.</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>The effect of the construct is the creation of a set of ground probabilistic facts, whose probability is assigned by a neural network. This mechanism allows us to delegate to a neural network <italic>m</italic> the classification of a set of predicates <italic>q</italic> defined by some input features <inline-formula><mml:math id="M6"><mml:mover accent="true"><mml:mrow><mml:mi>t</mml:mi></mml:mrow><mml:mo>&#x02192;</mml:mo></mml:mover></mml:math></inline-formula>. The possible classes are given by <inline-formula><mml:math id="M7"><mml:mover accent="true"><mml:mrow><mml:mi>u</mml:mi></mml:mrow><mml:mo>&#x02192;</mml:mo></mml:mover></mml:math></inline-formula>. Therefore, in the AM scenario, it would be possible, for example, to exploit two networks <italic>m</italic>_<italic>t</italic> and <italic>m</italic>_<italic>r</italic> to classify, respectively, the type of a potential argumentative component and the potential relation between two components. <xref ref-type="fig" rid="F1">Figure 1A</xref> shows the corresponding DeepProbLog code. These predicates could be easily integrated within a probabilistic logic program designed for argumentation, so as to model (possibly weighted) constraints, rules, and preferences, such as those described in section 2. <xref ref-type="fig" rid="F1">Figure 1B</xref> illustrates one such possibility.</p>
<fig id="F1" position="float">
<label>Figure 1</label>
<caption><p>An excerpt of a DeepProbLog program for AM. <bold>(A)</bold> Definition of neural predicates. <bold>(B)</bold> Definition of (probabilistic) rules.</p></caption>
<graphic xlink:href="fdata-02-00052-g0001.tif"/>
</fig>
<p>The same scenario can be modeled using GS-MLNs. In the Markov logic framework, first-order logic rules are associated with a real-valued weight. The higher the weight, the higher the probability that the clause is satisfied, other things being equal. The weight could possibly be infinite, to model hard constraints. In the GS-MLN extension, different weights can be associated to different groundings of the same formula, and such weights can be computed by neural networks. Joint training and inference can be performed, as a straightforward extension of the classic Markov logic framework (Lippi and Frasconi, <xref ref-type="bibr" rid="B28">2009</xref>). <xref ref-type="fig" rid="F2">Figure 2</xref> shows an example of a GS-MLN used to model the AM scenario that we consider. Here, the first three rules model grounding-specific clauses (the dollar symbol indicating a reference to a generic vector of features describing the example) whose weights depend on the specific groundings (variables <monospace>x</monospace> and <monospace>y</monospace>); the three subsequent rules are hard constraints (ending with a period); the final rule is a classic Markov logic rule, with a weight attached to the first-order clause.</p>
<fig id="F2" position="float">
<label>Figure 2</label>
<caption><p>An excerpt of a GS-MLN for the definition of neural, hard, and weighted rules for AM.</p></caption>
<graphic xlink:href="fdata-02-00052-g0002.tif"/>
</fig>
<p>The kind of approach hereby described strongly differs from the existing approaches in AM. Whereas, Persing and Ng (<xref ref-type="bibr" rid="B34">2016</xref>) exploit a pipeline scheme to apply the constraints to the predictions made by deep networks at a first stage of computation, the framework we propose can perform a joint training, which includes the constraints within the learning phase. This can be viewed as an instance of <italic>Constraint Driven Learning</italic> (Chang et al., <xref ref-type="bibr" rid="B7">2012</xref>) and its continuous counterpart, posterior regularization (Ganchev et al., <xref ref-type="bibr" rid="B18">2010</xref>), where multiple signals contribute to a global decision, by being pushed to satisfy expectations on the global decision. Differently from the work by Niculae et al. (<xref ref-type="bibr" rid="B33">2017</xref>), who use factor graphs to encode inter-dependencies between random variables, our approach enables to exploit the interpretable formalism of logic to represent rules. Moreover, the models of NeSy and SRL are typically able to learn the weights or the probabilities of the rules, or even to learn the rules themselves, thus addressing a <italic>structure learning</italic> task.</p>
</sec>
<sec sec-type="discussion" id="s4">
<title>4. Discussion</title>
<p>After many years of growing interest and remarkable results, time is ripe for AM to move forward in its ability to support complex arguments. To this end, we argue that research in this area should aim at combining sub-symbolic and symbolic approaches, and that several state-of-the-art ML frameworks already provide the necessary ground for such a leap forward.</p>
<p>The combination of such approaches will leverage different forms of abstractions that we consider essential for AM. On the one hand, (probabilistic) logical representations enable to specify AM systems in terms of data, world knowledge and other constraints, and to express uncertainties at a logical and conceptual level rather than at the level of individual random variables. This would make AM systems easier to interpret&#x02014;a feature that is now becoming a need for AI in general (Guidotti et al., <xref ref-type="bibr" rid="B22">2018</xref>)&#x02014;since they could help explaining the logic and the reasons that lead them to produce their arguments, while still dealing with the uncertainties stemming from the data and the (incomplete) background knowledge. On the other hand, AM is too complex to fully specify the distributions of random variables and their global (in)dependency structure a priori. Sub-symbolic models can harness such a complexity by finding the right, general outline, in the form of computational graphs, and processing data.</p>
<p>In order to fully exploit the potential of this joint approach, clearly many challenges have to be faced. First of all, several languages and frameworks for NeSy and SRL exist, each with its own characteristics in terms of both expressive power and efficiency. In this sense, AM would represent an ideal test-bed for such frameworks, by presenting a challenging, large-scale application domain where the exploitation of background knowledge could play a crucial role to boost performance. Inference in this kind of models is clearly an issue, thus AM would provide additional benchmarks for the development of efficient algorithms, both in terms of memory consumption and running time. Finally, although there are already several NeSy and SRL frameworks available, being these research areas still relatively young and in rapid development, their tools are not yet mainstream. Here, an effort is needed in integrating such tools with state-of-the-art neural architectures for NLP.</p>
</sec>
<sec id="s5">
<title>Author Contributions</title>
<p>All authors listed have made a substantial, direct and intellectual contribution to the work, and approved it for publication.</p>
<sec>
<title>Conflict of Interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
</sec>
</body>
<back>
<ack><p>We thank the reviewers for their comments and contributions, which have increased the quality of this work. A preliminary version of this manuscript has been released as a Pre-Print at Galassi et al. (<xref ref-type="bibr" rid="B16">2019</xref>).</p>
</ack>
<ref-list>
<title>References</title>
<ref id="B1">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Aharoni</surname> <given-names>E.</given-names></name> <name><surname>Polnarov</surname> <given-names>A.</given-names></name> <name><surname>Lavee</surname> <given-names>T.</given-names></name> <name><surname>Hershcovich</surname> <given-names>D.</given-names></name> <name><surname>Levy</surname> <given-names>R.</given-names></name> <name><surname>Rinott</surname> <given-names>R.</given-names></name> <etal/></person-group>. (<year>2014</year>). <article-title>A benchmark dataset for automatic detection of claims and evidence in the context of controversial topics</article-title>, in <source>Proceedings of the First Workshop on Argumentation Mining</source> (<publisher-loc>Baltimore, MD</publisher-loc>: <publisher-name>Association for Computational Linguistics</publisher-name>), <fpage>64</fpage>&#x02013;<lpage>68</lpage>.</citation></ref>
<ref id="B2">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Bellodi</surname> <given-names>E.</given-names></name> <name><surname>Riguzzi</surname> <given-names>F.</given-names></name></person-group> (<year>2013</year>). <article-title>Expectation maximization over binary decision diagrams for probabilistic logic programs</article-title>. <source>Intell. Data Anal.</source> <volume>17</volume>, <fpage>343</fpage>&#x02013;<lpage>363</lpage>. <pub-id pub-id-type="doi">10.3233/IDA-130582</pub-id></citation></ref>
<ref id="B3">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Bench-Capon</surname> <given-names>T.</given-names></name> <name><surname>Dunne</surname> <given-names>P. E.</given-names></name></person-group> (<year>2007</year>). <article-title>Argumentation in artificial intelligence</article-title>. <source>Artif. Intell.</source> <volume>171</volume>, <fpage>619</fpage>&#x02013;<lpage>641</lpage>. <pub-id pub-id-type="doi">10.1016/j.artint.2007.05.001</pub-id></citation></ref>
<ref id="B4">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Besnard</surname> <given-names>P.</given-names></name> <name><surname>Garc&#x000ED;a</surname> <given-names>A. J.</given-names></name> <name><surname>Hunter</surname> <given-names>A.</given-names></name> <name><surname>Modgil</surname> <given-names>S.</given-names></name> <name><surname>Prakken</surname> <given-names>H.</given-names></name> <name><surname>Simari</surname> <given-names>G. R.</given-names></name> <etal/></person-group>. (<year>2014</year>). <article-title>Introduction to structured argumentation</article-title>. <source>Argum. Comput.</source> <volume>5</volume>, <fpage>1</fpage>&#x02013;<lpage>4</lpage>. <pub-id pub-id-type="doi">10.1080/19462166.2013.869764</pub-id></citation></ref>
<ref id="B5">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Besnard</surname> <given-names>P.</given-names></name> <name><surname>Hunter</surname> <given-names>A.</given-names></name></person-group> (<year>2001</year>). <article-title>A logic-based theory of deductive arguments</article-title>. <source>Artif. Intell.</source> <volume>128</volume>, <fpage>203</fpage>&#x02013;<lpage>235</lpage>. <pub-id pub-id-type="doi">10.1016/S0004-3702(01)00071-6</pub-id></citation></ref>
<ref id="B6">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Cabrio</surname> <given-names>E.</given-names></name> <name><surname>Villata</surname> <given-names>S.</given-names></name></person-group> (<year>2018</year>). <article-title>Five years of argument mining: a data-driven analysis</article-title>, in <source>Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence, IJCAI-18</source> (<publisher-loc>Stockholm</publisher-loc>: <publisher-name>International Joint Conferences on Artificial Intelligence Organization</publisher-name>), <fpage>5427</fpage>&#x02013;<lpage>5433</lpage>.</citation></ref>
<ref id="B7">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Chang</surname> <given-names>M.-W.</given-names></name> <name><surname>Ratinov</surname> <given-names>L.</given-names></name> <name><surname>Roth</surname> <given-names>D.</given-names></name></person-group> (<year>2012</year>). <article-title>Structured learning with constrained conditional models</article-title>. <source>Mach. Learn.</source> <volume>88</volume>, <fpage>399</fpage>&#x02013;<lpage>431</lpage>. <pub-id pub-id-type="doi">10.1007/s10994-012-5296-5</pub-id></citation></ref>
<ref id="B8">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Cocarascu</surname> <given-names>O.</given-names></name> <name><surname>Toni</surname> <given-names>F.</given-names></name></person-group> (<year>2017</year>). <article-title>Identifying attack and support argumentative relations using deep learning</article-title>, in <source>Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing</source> (<publisher-loc>Copenhagen</publisher-loc>: <publisher-name>Association for Computational Linguistics</publisher-name>), <fpage>1374</fpage>&#x02013;<lpage>1379</lpage>.</citation></ref>
<ref id="B9">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Cocarascu</surname> <given-names>O.</given-names></name> <name><surname>Toni</surname> <given-names>F.</given-names></name></person-group> (<year>2018</year>). <article-title>Combining deep learning and argumentative reasoning for the analysis of social media textual content using small data sets</article-title>. <source>Comput. Linguist.</source> <volume>44</volume>, <fpage>833</fpage>&#x02013;<lpage>858</lpage>. <pub-id pub-id-type="doi">10.1162/coli_a_00338</pub-id></citation></ref>
<ref id="B10">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Conneau</surname> <given-names>A.</given-names></name> <name><surname>Schwenk</surname> <given-names>H.</given-names></name> <name><surname>Barrault</surname> <given-names>L.</given-names></name> <name><surname>Lecun</surname> <given-names>Y.</given-names></name></person-group> (<year>2017</year>). <article-title>Very deep convolutional networks for text classification</article-title>, in <source>Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics: Volume 1, Long Papers</source> (<publisher-loc>Valencia</publisher-loc>: <publisher-name>Association for Computational Linguistics</publisher-name>), <fpage>1107</fpage>&#x02013;<lpage>1116</lpage>.</citation></ref>
<ref id="B11">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Daxenberger</surname> <given-names>J.</given-names></name> <name><surname>Eger</surname> <given-names>S.</given-names></name> <name><surname>Habernal</surname> <given-names>I.</given-names></name> <name><surname>Stab</surname> <given-names>C.</given-names></name> <name><surname>Gurevych</surname> <given-names>I.</given-names></name></person-group> (<year>2017</year>). <article-title>What is the essence of a claim? cross-domain claim identification</article-title>, in <source>Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing</source> (<publisher-loc>Copenhagen</publisher-loc>: <publisher-name>Association for Computational Linguistics</publisher-name>), <fpage>2055</fpage>&#x02013;<lpage>2066</lpage>.</citation></ref>
<ref id="B12">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>De Raedt</surname> <given-names>L.</given-names></name> <name><surname>Kersting</surname> <given-names>K.</given-names></name> <name><surname>Natarajan</surname> <given-names>S.</given-names></name> <name><surname>Poole</surname> <given-names>D.</given-names></name></person-group> (<year>2016</year>). <source>Statistical Relational Artificial Intelligence: Logic, Probability, and Computation</source>. Synthesis Lectures on Artificial Intelligence and Machine Learning. Morgan &#x00026; Claypool Publishers.</citation></ref>
<ref id="B13">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Dung</surname> <given-names>P. M.</given-names></name></person-group> (<year>1995</year>). <article-title>On the acceptability of arguments and its fundamental role in nonmonotonic reasoning, logic programming and n-person games</article-title>. <source>Artif. Intell.</source> <volume>77</volume>, <fpage>321</fpage>&#x02013;<lpage>358</lpage>. <pub-id pub-id-type="doi">10.1016/0004-3702(94)00041-X</pub-id></citation></ref>
<ref id="B14">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Dung</surname> <given-names>P. M.</given-names></name> <name><surname>Kowalski</surname> <given-names>R. A.</given-names></name> <name><surname>Toni</surname> <given-names>F.</given-names></name></person-group> (<year>2009</year>). <article-title>Assumption-based argumentation</article-title>, in <source>Argumentation in Artificial Intelligence</source>, eds <person-group person-group-type="editor"><name><surname>Simari</surname> <given-names>G.</given-names></name> <name><surname>Rahwan</surname> <given-names>I.</given-names></name></person-group> (<publisher-loc>Boston, MA</publisher-loc>: <publisher-name>Springer US</publisher-name>), <fpage>199</fpage>&#x02013;<lpage>218</lpage>.</citation></ref>
<ref id="B15">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Eger</surname> <given-names>S.</given-names></name> <name><surname>Daxenberger</surname> <given-names>J.</given-names></name> <name><surname>Gurevych</surname> <given-names>I.</given-names></name></person-group> (<year>2017</year>). <article-title>Neural end-to-end learning for computational argumentation mining</article-title>, in <source>Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)</source> (<publisher-loc>Vancouver, BC</publisher-loc>: <publisher-name>Association for Computational Linguistics</publisher-name>), <fpage>11</fpage>&#x02013;<lpage>22</lpage>.</citation></ref>
<ref id="B16">
<citation citation-type="web"><person-group person-group-type="author"><name><surname>Galassi</surname> <given-names>A.</given-names></name> <name><surname>Kersting</surname> <given-names>K.</given-names></name> <name><surname>Lippi</surname> <given-names>M.</given-names></name> <name><surname>Shao</surname> <given-names>X.</given-names></name> <name><surname>Torroni</surname> <given-names>P.</given-names></name></person-group> (<year>2019</year>). <article-title>Neural-symbolic argumentation mining: an argument in favour of deep learning and reasoning</article-title>. arXiv:1905.09103 [Preprint] Available online at: <ext-link ext-link-type="uri" xlink:href="https://arxiv.org/abs/1905.09103">https://arxiv.org/abs/1905.09103</ext-link></citation></ref>
<ref id="B17">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Galassi</surname> <given-names>A.</given-names></name> <name><surname>Lippi</surname> <given-names>M.</given-names></name> <name><surname>Torroni</surname> <given-names>P.</given-names></name></person-group> (<year>2018</year>). <article-title>Argumentative link prediction using residual networks and multi-objective learning</article-title>, in <source>Proceedings of the 5th Workshop on Argument Mining</source> (<publisher-loc>Brussels</publisher-loc>: <publisher-name>Association for Computational Linguistics</publisher-name>), <fpage>1</fpage>&#x02013;<lpage>10</lpage>.</citation></ref>
<ref id="B18">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ganchev</surname> <given-names>K.</given-names></name> <name><surname>Gra&#x000E7;a</surname> <given-names>J.</given-names></name> <name><surname>Gillenwater</surname> <given-names>J.</given-names></name> <name><surname>Taskar</surname> <given-names>B.</given-names></name></person-group> (<year>2010</year>). <article-title>Posterior regularization for structured latent variable models</article-title>. <source>J. Mach. Learn. Res.</source> <volume>11</volume>, <fpage>2001</fpage>&#x02013;<lpage>2049</lpage>. <pub-id pub-id-type="doi">10.5555/1756006.1859918</pub-id></citation></ref>
<ref id="B19">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Garcez</surname> <given-names>A.</given-names></name> <name><surname>Besold</surname> <given-names>T. R.</given-names></name> <name><surname>De Raedt</surname> <given-names>L.</given-names></name> <name><surname>F&#x000F6;ldiak</surname> <given-names>P.</given-names></name> <name><surname>Hitzler</surname> <given-names>P.</given-names></name> <name><surname>Icard</surname> <given-names>T.</given-names></name> <etal/></person-group>. (<year>2015</year>). <article-title>Neural-symbolic learning and reasoning: contributions and challenges</article-title>, in <source>Proceedings of the AAAI Spring Symposium on Knowledge Representation and Reasoning: Integrating Symbolic and Neural Approaches, Stanford</source> (<publisher-loc>Palo Alto, CA</publisher-loc>: <publisher-name>Stanford University</publisher-name>), <fpage>18</fpage>&#x02013;<lpage>21</lpage>.</citation></ref>
<ref id="B20">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Garc&#x000ED;a</surname> <given-names>A. J.</given-names></name> <name><surname>Simari</surname> <given-names>G. R.</given-names></name></person-group> (<year>2004</year>). <article-title>Defeasible logic programming: an argumentative approach</article-title>. <source>Theory Pract. Log. Program</source> <volume>4</volume>, <fpage>95</fpage>&#x02013;<lpage>138</lpage>. <pub-id pub-id-type="doi">10.1017/S1471068403001674</pub-id></citation></ref>
<ref id="B21">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Getoor</surname> <given-names>L.</given-names></name> <name><surname>Taskar</surname> <given-names>B.</given-names></name></person-group> (<year>2007</year>). <source>Introduction to Statistical Relational Learning</source>, <volume>Vol. 1</volume>. <publisher-loc>Cambridge, MA</publisher-loc>: <publisher-name>MIT Press</publisher-name>.</citation></ref>
<ref id="B22">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Guidotti</surname> <given-names>R.</given-names></name> <name><surname>Monreale</surname> <given-names>A.</given-names></name> <name><surname>Ruggieri</surname> <given-names>S.</given-names></name> <name><surname>Turini</surname> <given-names>F.</given-names></name> <name><surname>Giannotti</surname> <given-names>F.</given-names></name> <name><surname>Pedreschi</surname> <given-names>D.</given-names></name></person-group> (<year>2018</year>). <article-title>A survey of methods for explaining black box models</article-title>. <source>ACM Comput. Surv.</source> <volume>51</volume>, <fpage>93:1</fpage>&#x02013;<lpage>93:42</lpage>. <pub-id pub-id-type="doi">10.1145/3236009</pub-id></citation></ref>
<ref id="B23">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Gutmann</surname> <given-names>B.</given-names></name> <name><surname>Thon</surname> <given-names>I.</given-names></name> <name><surname>De Raedt</surname> <given-names>L.</given-names></name></person-group> (<year>2011</year>). <article-title>Learning the parameters of probabilistic logic programs from interpretations</article-title>, in <source>Joint European Conference on Machine Learning and Knowledge Discovery in Databases</source> (<publisher-loc>Athens</publisher-loc>: <publisher-name>Springer</publisher-name>), <fpage>581</fpage>&#x02013;<lpage>596</lpage>.</citation></ref>
<ref id="B24">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Huang</surname> <given-names>Y. Y.</given-names></name> <name><surname>Wang</surname> <given-names>W. Y.</given-names></name></person-group> (<year>2017</year>). <article-title>Deep residual learning for weakly-supervised relation extraction</article-title>, in <source>Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing</source> (<publisher-loc>Copenhagen</publisher-loc>: <publisher-name>Association for Computational Linguistics</publisher-name>), <fpage>1803</fpage>&#x02013;<lpage>1807</lpage>.</citation></ref>
<ref id="B25">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Kordjamshidi</surname> <given-names>P.</given-names></name> <name><surname>Roth</surname> <given-names>D.</given-names></name> <name><surname>Kersting</surname> <given-names>K.</given-names></name></person-group> (<year>2018</year>). <article-title>Systems AI: a declarative learning based programming perspective</article-title>, in <source>Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence, IJCAI-18</source> (<publisher-loc>Stockholm</publisher-loc>: <publisher-name>International Joint Conferences on Artificial Intelligence Organization</publisher-name>), <fpage>5464</fpage>&#x02013;<lpage>5471</lpage>.</citation></ref>
<ref id="B26">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Lauscher</surname> <given-names>A.</given-names></name> <name><surname>Glava&#x00161;</surname> <given-names>G.</given-names></name> <name><surname>Ponzetto</surname> <given-names>S. P.</given-names></name> <name><surname>Eckert</surname> <given-names>K.</given-names></name></person-group> (<year>2018</year>). <article-title>Investigating the role of argumentation in the rhetorical analysis of scientific publications with neural multi-task learning models</article-title>, in <source>Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing</source> (<publisher-loc>Brussels</publisher-loc>: <publisher-name>Association for Computational Linguistics</publisher-name>), <fpage>3326</fpage>&#x02013;<lpage>3338</lpage>.</citation></ref>
<ref id="B27">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>LeCun</surname> <given-names>Y.</given-names></name> <name><surname>Bengio</surname> <given-names>Y.</given-names></name> <name><surname>Hinton</surname> <given-names>G.</given-names></name></person-group> (<year>2015</year>). <article-title>Deep learning</article-title>. <source>Nature</source> <volume>531</volume>, <fpage>436</fpage>&#x02013;<lpage>444</lpage>. <pub-id pub-id-type="doi">10.1038/nature14539</pub-id></citation></ref>
<ref id="B28">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lippi</surname> <given-names>M.</given-names></name> <name><surname>Frasconi</surname> <given-names>P.</given-names></name></person-group> (<year>2009</year>). <article-title>Prediction of protein &#x003B2;-residue contacts by Markov logic networks with grounding-specific weights</article-title>. <source>Bioinformatics</source> <volume>25</volume>, <fpage>2326</fpage>&#x02013;<lpage>2333</lpage>. <pub-id pub-id-type="doi">10.1093/bioinformatics/btp421</pub-id><pub-id pub-id-type="pmid">19592394</pub-id></citation></ref>
<ref id="B29">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Lippi</surname> <given-names>M.</given-names></name> <name><surname>Torroni</surname> <given-names>P.</given-names></name></person-group> (<year>2015</year>). <article-title>Context-independent claim detection for argument mining</article-title>, in <source>Proceedings of the Twenty-Fourth International Joint Conference on Artificial Intelligence, IJCAI 2015</source>, eds <person-group person-group-type="editor"><name><surname>Yang</surname> <given-names>Q.</given-names></name> <name><surname>Wooldridge</surname> <given-names>M.</given-names></name></person-group> (<publisher-loc>Buenos Aires</publisher-loc>: <publisher-name>AAAI Press</publisher-name>), <fpage>185</fpage>&#x02013;<lpage>191</lpage>.</citation></ref>
<ref id="B30">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lippi</surname> <given-names>M.</given-names></name> <name><surname>Torroni</surname> <given-names>P.</given-names></name></person-group> (<year>2016</year>). <article-title>Argumentation mining: state of the art and emerging trends</article-title>. <source>ACM Trans. Internet Technol.</source> <volume>16</volume>, <fpage>10:1</fpage>&#x02013;<lpage>10:25</lpage>. <pub-id pub-id-type="doi">10.1145/2850417</pub-id></citation></ref>
<ref id="B31">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Lugini</surname> <given-names>L.</given-names></name> <name><surname>Litman</surname> <given-names>D.</given-names></name></person-group> (<year>2018</year>). <article-title>Argument component classification for classroom discussions</article-title>, in <source>Proceedings of the 5th Workshop on Argument Mining</source> (<publisher-loc>Brussels</publisher-loc>: <publisher-name>Association for Computational Linguistics</publisher-name>), <fpage>57</fpage>&#x02013;<lpage>67</lpage>.</citation></ref>
<ref id="B32">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Manhaeve</surname> <given-names>R.</given-names></name> <name><surname>Dumancic</surname> <given-names>S.</given-names></name> <name><surname>Kimmig</surname> <given-names>A.</given-names></name> <name><surname>Demeester</surname> <given-names>T.</given-names></name> <name><surname>De Raedt</surname> <given-names>L.</given-names></name></person-group> (<year>2018</year>). <article-title>DeepProbLog: neural probabilistic logic programming</article-title>, in <source>Advances in Neural Information Processing Systems</source>, eds <person-group person-group-type="editor"><name><surname>Bengio</surname> <given-names>S.</given-names></name> <name><surname>Wallach</surname> <given-names>H.</given-names></name> <name><surname>Larochelle</surname> <given-names>H.</given-names></name> <name><surname>Grauman</surname> <given-names>K.</given-names></name> <name><surname>Cesa-Bianchi</surname> <given-names>N.</given-names></name> <name><surname>Garnett</surname> <given-names>R.</given-names></name></person-group> (<publisher-loc>Montreal, QC</publisher-loc>: <publisher-name>Curran Associates, Inc.</publisher-name>), <fpage>3753</fpage>&#x02013;<lpage>3763</lpage>.</citation></ref>
<ref id="B33">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Niculae</surname> <given-names>V.</given-names></name> <name><surname>Park</surname> <given-names>J.</given-names></name> <name><surname>Cardie</surname> <given-names>C.</given-names></name></person-group> (<year>2017</year>). <article-title>Argument mining with structured SVMs and RNNs</article-title>, in <source>Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)</source> (<publisher-loc>Vancouver, BC</publisher-loc>: <publisher-name>Association for Computational Linguistics</publisher-name>), <fpage>985</fpage>&#x02013;<lpage>995</lpage>.</citation></ref>
<ref id="B34">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Persing</surname> <given-names>I.</given-names></name> <name><surname>Ng</surname> <given-names>V.</given-names></name></person-group> (<year>2016</year>). <article-title>End-to-end argumentation mining in student essays</article-title>, in <source>Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</source> (<publisher-loc>San Diego, CA</publisher-loc>: <publisher-name>Association for Computational Linguistics</publisher-name>), <fpage>1384</fpage>&#x02013;<lpage>1394</lpage>.</citation></ref>
<ref id="B35">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Raedt</surname> <given-names>L. D.</given-names></name> <name><surname>Kimmig</surname> <given-names>A.</given-names></name> <name><surname>Toivonen</surname> <given-names>H.</given-names></name></person-group> (<year>2007</year>). <article-title>ProbLog: a probabilistic prolog and its application in link discovery</article-title>, in <source>IJCAI</source> (<publisher-loc>Hyderabad</publisher-loc>), <fpage>2462</fpage>&#x02013;<lpage>2467</lpage>.</citation></ref>
<ref id="B36">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Riguzzi</surname> <given-names>F.</given-names></name></person-group> (<year>2018</year>). <source>Foundations of Probabilistic Logic Programming</source>. <publisher-loc>Gistrup</publisher-loc>: <publisher-name>River Publishers</publisher-name>.</citation></ref>
<ref id="B37">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Rinott</surname> <given-names>R.</given-names></name> <name><surname>Dankin</surname> <given-names>L.</given-names></name> <name><surname>Alzate Perez</surname> <given-names>C.</given-names></name> <name><surname>Khapra</surname> <given-names>M. M.</given-names></name> <name><surname>Aharoni</surname> <given-names>E.</given-names></name> <name><surname>Slonim</surname> <given-names>N.</given-names></name></person-group> (<year>2015</year>). <article-title>Show me your evidence&#x02013;an automatic method for context dependent evidence detection</article-title>, in <source>Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing</source> (<publisher-loc>Lisbon</publisher-loc>: <publisher-name>Association for Computational Linguistics</publisher-name>), <fpage>440</fpage>&#x02013;<lpage>450</lpage>.</citation></ref>
<ref id="B38">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Sato</surname> <given-names>T.</given-names></name> <name><surname>Kameya</surname> <given-names>Y.</given-names></name></person-group> (<year>1997</year>). <article-title>PRISM: a language for symbolic-statistical modeling</article-title>, in <source>Proceedings of the Fifteenth International Joint Conference on Artificial Intelligence, IJCAI 97</source> (<publisher-loc>Nagoya</publisher-loc>: <publisher-name>Morgan Kaufmann</publisher-name>), <fpage>1330</fpage>&#x02013;<lpage>1339</lpage>.</citation></ref>
<ref id="B39">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Schulz</surname> <given-names>C.</given-names></name> <name><surname>Eger</surname> <given-names>S.</given-names></name> <name><surname>Daxenberger</surname> <given-names>J.</given-names></name> <name><surname>Kahse</surname> <given-names>T.</given-names></name> <name><surname>Gurevych</surname> <given-names>I.</given-names></name></person-group> (<year>2018</year>). <article-title>Multi-task learning for argumentation mining in low-resource settings</article-title>, in <source>Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 2 (Short Papers)</source> (<publisher-loc>New Orleans, LA</publisher-loc>: <publisher-name>Association for Computational Linguistics</publisher-name>), <fpage>35</fpage>&#x02013;<lpage>41</lpage>.</citation></ref>
<ref id="B40">
<citation citation-type="web"><person-group person-group-type="author"><name><surname>Serafini</surname> <given-names>L.</given-names></name> <name><surname>Garcez</surname> <given-names>A. D.</given-names></name></person-group> (<year>2016</year>). <article-title>Logic tensor networks: deep learning and logical reasoning from data and knowledge</article-title>. arXiv 1606.04422 [Preprint] Available online at: <ext-link ext-link-type="uri" xlink:href="https://arxiv.org/abs/1606.04422">https://arxiv.org/abs/1606.04422</ext-link></citation></ref>
<ref id="B41">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Simari</surname> <given-names>G.</given-names></name> <name><surname>Rahwan</surname> <given-names>I.</given-names></name></person-group> (<year>2009</year>). <source>Argumentation in Artificial Intelligence</source>. Springer US.</citation></ref>
<ref id="B42">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Vennekens</surname> <given-names>J.</given-names></name> <name><surname>Verbaeten</surname> <given-names>S.</given-names></name> <name><surname>Bruynooghe</surname> <given-names>M.</given-names></name></person-group> (<year>2004</year>). <article-title>Logic programs with annotated disjunctions</article-title>, in <source>Logic Programming</source>, eds <person-group person-group-type="editor"><name><surname>Demoen</surname> <given-names>B.</given-names></name> <name><surname>Lifschitz</surname> <given-names>V.</given-names></name></person-group> (<publisher-loc>Berlin; Heidelberg</publisher-loc>: <publisher-name>Springer Berlin Heidelberg</publisher-name>), <fpage>431</fpage>&#x02013;<lpage>445</lpage>.</citation></ref>
<ref id="B43">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Walton</surname> <given-names>D.</given-names></name> <name><surname>Reed</surname> <given-names>C.</given-names></name> <name><surname>Macagno</surname> <given-names>F.</given-names></name></person-group> (<year>2008</year>). <source>Argumentation Schemes</source>. Cambridge University Press.</citation></ref>
<ref id="B44">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Young</surname> <given-names>T.</given-names></name> <name><surname>Hazarika</surname> <given-names>D.</given-names></name> <name><surname>Poria</surname> <given-names>S.</given-names></name> <name><surname>Cambria</surname> <given-names>E.</given-names></name></person-group> (<year>2018</year>). <article-title>Recent trends in deep learning based natural language processing</article-title>. <source>IEEE Comput. Intell. Mag.</source> <volume>13</volume>, <fpage>55</fpage>&#x02013;<lpage>75</lpage>. <pub-id pub-id-type="doi">10.1109/MCI.2018.2840738</pub-id></citation></ref>
<ref id="B45">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zhang</surname> <given-names>L.</given-names></name> <name><surname>Wang</surname> <given-names>S.</given-names></name> <name><surname>Liu</surname> <given-names>B.</given-names></name></person-group> (<year>2018a</year>). <article-title>Deep learning for sentiment analysis: a survey</article-title>. <source>Wiley Interdiscipl. Rev. Data Mining Knowl. Discov.</source> <volume>8</volume>:<fpage>e1253</fpage>. <pub-id pub-id-type="doi">10.1002/widm.1253</pub-id></citation></ref>
<ref id="B46">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Zhang</surname> <given-names>Y.</given-names></name> <name><surname>Henao</surname> <given-names>R.</given-names></name> <name><surname>Gan</surname> <given-names>Z.</given-names></name> <name><surname>Li</surname> <given-names>Y.</given-names></name> <name><surname>Carin</surname> <given-names>L.</given-names></name></person-group> (<year>2018b</year>). <article-title>Multi-label learning from medical plain text with convolutional residual models</article-title>, in <source>Proceedings of the 3rd Machine Learning for Healthcare Conference, Volume 85 of Proceedings of Machine Learning Research</source>, eds <person-group person-group-type="editor"><name><surname>Doshi-Velez</surname> <given-names>F.</given-names></name> <name><surname>Fackler</surname> <given-names>J.</given-names></name> <name><surname>Jung</surname> <given-names>K.</given-names></name> <name><surname>Kale</surname> <given-names>D.</given-names></name> <name><surname>Ranganath</surname> <given-names>R.</given-names></name> <name><surname>Wallace</surname> <given-names>B.</given-names></name> <etal/></person-group>. (<publisher-loc>Palo Alto, CA</publisher-loc>: <publisher-name>PMLR</publisher-name>), <fpage>280</fpage>&#x02013;<lpage>294</lpage>.</citation></ref>
</ref-list>
<fn-group>
<fn id="fn0001"><p><sup>1</sup>In a multi-task setting, instead of two networks, the same network could be used to perform both component classification and link detection at the same time. In AM, this approach has already been successfully explored (Eger et al., <xref ref-type="bibr" rid="B15">2017</xref>; Galassi et al., <xref ref-type="bibr" rid="B17">2018</xref>).</p></fn>
<fn fn-type="financial-disclosure"><p><bold>Funding.</bold> This work has been partially supported by the H2020 Project AI4EU (G.A. 825619), by the DFG project CAML (KE 1686/3-1, SPP 1999), and by the RMU project DeCoDeML.</p>
</fn>
</fn-group>
</back>
</article>
