Sunteți pe pagina 1din 50

Tarek R.

Besold, Kai-Uwe Kühnberger,


Marco Schorlemmer & Alan Smaill (eds.)

Proceedings of the Workshop


“Computational Creativity, Concept Invention,
and General Intelligence”

PICS
Publications of the Institute of Cognitive Science

Volume 2-2013
ISSN: 1610-5389

Series title: PICS


Publications of the Institute of Cognitive Science

Volume: 2-2013

Place of publication: Osnabrück, Germany

Date: August 2013

Editors: Kai-Uwe Kühnberger


Peter König
Sven Walter

Cover design: Thorsten Hinrichs

! Institute of Cognitive Science


Tarek R. Besold
Kai-Uwe Kühnberger
Marco Schorlemmer
Alan Smaill (Eds.)

Computational Creativity,
Concept Invention,
and General Intelligence

2nd International Workshop, C3GI @ IJCAI 2013,


Beijing, China, August 04, 2013

Proceedings
Volume Editors

Tarek R. Besold
Institute of Cognitive Science
University of Osnabrueck

Kai-Uwe Kühnberger
Institute of Cognitive Science
University of Osnabrueck

Marco Schorlemmer
IIIA-CSIC, Barcelona

Alan Smaill
School of Informatics
University of Edinburgh

This volume contains the proceedings of the workshop “Computational Creativity, Concept
Invention, and General Intelligence (C3GI)” at IJCAI-2013.
Preface

This volume contains the proceedings of the 2nd International Workshop for Computational
Creativity, Concept Invention, and General Intelligence held in conjunction with IJCAI 2013
in Beijing, China.

Continuing the mission started by the 1st edition of the workshop series in 2012 at
Montpellier, the aim of this workshop is to bring together researchers from different fields
of AI working on computational models for creativity, concept formation, concept
discovery, idea generation, and their overall relation and role to general intelligence as well
as researchers focusing on application areas, like computer-aided innovation.

Although the different approaches to questions concerning the aforementioned aspects do


share significant overlap in underlying ideas, the cooperation between the respective
communities is still in an early stage, and can greatly profit from interaction and discussion
between people from the respective fields, forming trans- and interdisciplinary alliances in
research and application.

Together with a growing community of researchers in both, academia and industry, we are
convinced that it is time to revitalize the old AI dream of “Thinking Machines” which now
has been almost completely abandoned for decades. Fortunately, more and more
researchers presently recognize the necessity and feasibility of returning to the original
goal of creating systems with human-like intelligence. By continuing the discussions
started one year ago at the 2012 edition of the workshop, we are now working towards
establishing the workshop series as an annual event gathering leading researchers in
computational creativity, general intelligence, and concept invention whose work is in
direct connection to our overall goal, providing and generating additional valuable support
in further turning the current spirit of optimism into a lasting research endeavor.

August, 2013 Tarek R. Besold


Kai-Uwe Kühnberger
Marco Schorlemmer
Alan Smaill
Program Committee

Committee Co-Chairs

! Tarek R. Besold, University of Osnabrueck


! Kai-Uwe Kühnberger, University of Osnabrueck
! Marco Schorlemmer, IIIA-CSIC, Barcelona
! Alan Smaill, University of Edinburgh

Committee Members

! John Barnden, University of Birmingham


! Mehul Bhatt, University of Bremen
! Mark Bishop, Goldsmiths, University of London
! Simon Colton, Goldsmiths, University of London
! Gaetano Cascini, Politecnico di Milano
! Joerg Cassens, University of Luebeck
! Hamid Ekbia, Indiana University Bloomington
! Markus Guhe, University of Edinburgh
! Helmar Gust, University of Osnabrueck
! Ivan M. Havel, Charles University Prague
! Bipin Indurkhya, IIIT Hyderabad/AGH University of Science and Technology, Krakow
! Anders Kofod-Petersen, NTNU Trondheim
! Ulf Krumnack, University of Osnabrueck
! Oliver Kutz, University of Bremen
! Maricarmen Martínez Baldares, University of the Andes, Bogota
! Alison Pease, University of Edinburgh
! Francisco Pereira, University of Coimbra
! Enric Plaza, IIIA-CSIC, Barcelona
! Thomas Roth-Berghofer, University of West London
! Ute Schmid, University of Bamberg
! Petros Stefaneas, National Technical University Athens
! Tony Veale, University College Dublin
! Pei Wang, Temple University Philadelphia
Table of Contents

Full presentations:

Creativity in Artificial Intelligence as a Hybrid of Logic and Spontaneity


S. Ellis, N. Sundar G., J. Valerio, S. Bringsjord & J. Braasch

How Models of Creativity and Analogy Need to Answer the Tailorability Concern
J. Licato, S. Bringsjord & N. Sundar G.

Creative Concept Invention and Representation in an Interactive Child-Robot Healthy


Snack Design Game
A. Williams, A. Stroud, J. Panka & J. Williams

Compact presentations:

Back-and-Forth Inception: Towards a Cognitively-Inspired Model of Concept Construction


via Conceptual Intellection
A. Abdel-Fattah & S. Schneider

Square, zero, kitchen: Start – Insights on cross-linguistic conceptual encoding


F. Quattri

Semantic Clues for Novel Metaphor Generator


R. Rzepka, P. Dybala, K. Sayama & K. Araki
Creativity in Artificial Intelligence as a Hybrid of Logic and Spontaneity

Simon Ellis • Naveen Sundar G. • Joseph Valerio • Selmer Bringsjord • Jonas Braasch
elliss5@rpi.edu • govinn@rpi.edu • valerj3@rpi.edu • selmer@rpi.edu • braasj@rpi.edu
Department of Cognitive Science • Department of Computer Science
The Rensselaer AI & Reasoning (RAIR) Lab • Rensselaer Polytechnic Institute (RPI) • Troy NY 12180 USA

Abstract 2 Setting the Context


2.1 Psychometric View of Creativity Presupposed
This paper explores machine creativity from the
perspective that genuine human-level creativity is a In earlier work, we discussed several ways an artificial agent
hybrid function of formal, computational logic on may via performance on certain tests at least seem creative,
the one hand, and intuition and spontaneity on the even though it may not be genuinely so [Ellis et al., 2012].
other. After setting the context and explaining the This test-based orientation reflects what is known as Psy-
resources upon which we draw, we provide an ex- chometric AI [Bringsjord and Schimanski, 2003; Bringsjord,
ample of this essentially synergistic relationship in 2011]. The toughest relevant test is the “Lovelace Test,”
the form of the improvisational functionality of the which is passed by a computing machine when the machine’s
CAIRA system. output cannot be anticipated by its designer [Bringsjord et al.,
2001]; a machine able to pass this test is said to be creativeB .
We concede that this standard exceeds what our current en-
gineering can reach. We have a more humble aim at present:
1 Introduction engineer a computing machine able to convince non-experts,
Turing-test style, that the output they observe is produced by
After setting the context, this paper reports on new develop- what they would call a “creative” computing machine; this
ments to our intelligent music agent, Handle, which was first is creativityT . In the literature, there are even more latitudi-
introduced at C3GI 2012 in Montpellier, France. Handle is narian tests for creativity: Cope, for instance, in his Com-
intended to operate as an artificial conductor, and to that end puter Models of Musical Creativity 2005, tells us that for him
was developed to interpret and provide feedback on a human creativity is “[t]he initialization of connections between two
performance of a score-based, classical solo piece of music. or more multifaceted things, ideas, or phenomena hitherto
The music calculus, a first-order sorted modal logic, was in- not otherwise considered actively connected” [Cope, 2005],
troduced as a means of evaluating certain aspects of musical where the evidence for such connections being made is pro-
performance, and cognitive attitudes directed thereto. The vided by performance on certain simple puzzles.
current version of Handle extends the work done in the pre-
vious version and explores similar questions in the area of 2.2 Two Assumptions that Imply Use of Logic
improvised “free music.” We demonstrate Handle together Despite the fact that we anchor our r&d to tests, we do make
with its sister agent FILTER as part of a larger musical sys- one assumption about the internals of a creativeT machine,
tem, CAIRA, and explore the intersections of logic, intuition, and one general assumption about the epistemic context of
and spontaneity that together enable the creative process. any computational artifact that we produce.
The plan of the paper is as follows. We begin by setting the Regarding internals, we assume that the computing ma-
context (§2): We note that in the present paper our AI work chine that is a candidate for creativityT have at least the
on computational creativity reflects a Bringsjordian perspec- minimum representation-and-reasoning power of quantified
tive (§2.1); explain two assumptions that serve to place logic epistemic logic. This minimality condition, and the method-
at the heart of our engineering (§2.2); give an overview of ology that accompanies it, have been defended elsewhere
the “music calculus,” the logic-based framework used for our [Bringsjord, 2008]. The condition reflects Bringsjord’s af-
modeling (§2.3); and adumbrate our general approach to the firmation of Piaget’s seminal position that mature, general-
nexus of logic and intuition (§2.4). Next (§3), we give a bar- purpose human cognition (including, most prominently,
baric summary of our CAIRA and its two main components, problem-solving) consists of processes operating largely on
and then, in section 4, we provide a demonstration, one that formulas having at least the expressivity of formulae in full
is broad but, on the logic side, thin (we end this section with a first-order logic [Inhelder and Piaget, 1958]. Given the af-
look at a deeper use of the music calculus). The paper wraps firmation of the condition in question, the present work re-
up with a brief discussion of next steps (§6). flects a desire to engineer machines that are creativeT+ : they
are both creativeT and their internal processing conforms
to Piagetian concepts of general intelligence and creative Figure 1: Syntax of DC EC ∗
problem-solving. For more on this machines that embody the Syntax
first condition, see [Bringsjord and Licato, 2012].
Object | Agent | Self � Agent | ActionType | Action � Event |
What is our second assumption? We assume that compu- S ::=
Moment | Boolean | Fluent | Numeric
tational creativity cannot be formalized and engineered with-
out yet another nod in the direction of logic, and specifically action : Agent × ActionType → Action
epistemic logic. To see our second assumption, consider a initially : Fluent → Boolean
person/agent A who has produced an artifact a through some holds : Fluent × Moment → Boolean
creative process p. Any formal model of A’s cognition before, happens : Event × Moment → Boolean
during, and after the creative process should by our lights clipped : Moment × Fluent × Moment → Boolean
have sufficient representational capacity to let us conclude, f ::= initiates : Event × Fluent × Moment → Boolean
or at least consider, whether: terminates : Event × Fluent × Moment → Boolean
1. A knows that A itself has produced a in a creative manner. prior : Moment × Moment → Boolean
2. A believes that a has not been produced by any other agents. interval : Moment × Boolean
3. A knew what existed and what did not exist before it started the ∗ : Agent → Self
process that produced a. payoff : Agent × ActionType × Moment → Numeric
4. A desired that it needed to create an artifact to fulfill some t ::= x : S | c : S | f (t1 , . . . ,tn )
need.
p : Boolean | ¬φ | φ ∧ ψ | φ ∨ ψ | φ → ψ | φ ↔ ψ | ∀x : S. φ | ∃x : S. φ
5. A intended to engage in a series of actions making up the pro- P(a,t, φ) | K(a,t, φ) | C(t, φ) | S(a, b,t, φ) | S(a,t, φ)
cess p. φ ::=
B(a,t, φ) | D(a,t, holds( f ,t � )) | I(a,t, happens(action(a∗ , α),t � ))
The above (informal) micro-theory seems to us is admit-
tedly incomplete, but does seem unavoidable, and suggests O(a,t, φ, happens(action(a∗ , α),t � ))

that higher cognition surrounding creativity that can be suc-


cinctly captured with an expressive formal logic. We do not Rules of Inference
claim here that we have invented and specified this logic, but which when performed will [Rlead to the agent simply[Rbeliev-
1] 2]
we have taken appreciable steps toward doing so, by way of ing that C(t,
theP(a,t,
value φ) →isK(a,t,
something.
φ)) There
C(t, K(a,t,are
φ) → costs associated
B(a,t, φ))
our modeling framework, DC EC ∗ , to which we now turn. with both C(t,
theφ)actions,
t ≤ t1 . . .t ≤andtn the agent K(a,t,would
φ) prefer knowing
[R3 ]
the value.K(aMore generally,
K(a φ) if the agent
φ believes
[R4 ]
that the cost
2.3 Our Logic-Based Framework (music calculus) ,t
1 1 , . . .
of a full non-intuitive process
n n
,t , . . .)
t1 ≤ t3 ,t2 is ≤ t3greater than the cost of an in-
Our core logical framework is DC EC ∗ (deontic cognitive tuitive process by a threshold γ (which factors in[Rthe 5 ] benefits
C(t, K(a,t1 , φ1 → φ2 ) → (K(a,t2 , φ1 ) → K(a,t3 , φ2 )))
event calculus) a multi-sorted quantified modal logic that has of a deductive search), the agent would engage in an intuite
t1 ≤ t3 ,t2 ≤ t3
a well-defined syntax and a proof calculus. The syntax of the process. Assuming that the agent knows some arithmetic, [R6 ] we
language of DC EC ∗ and the rules of inference for its proof can represent this model
C(t, B(a,t 1 , φ 1 → φ 2 ) →of 2 , φ1 ) →with
intuition
(B(a,t φ2 )))micro-theory:
B(a,t3 ,this
calculus are shown in Figure 2. DC EC ∗ syntax includes a t1 ≤ t3 ,t2 ≤ t3
[R7 ]
system of sorts S, a signature f , a grammar for terms t, and a  C(t, C(t1 , φ1 → φ2 ) → (C(t ∗ 2 , φ1 ) → C(t3 , φ2 ))) 
 K(a,t, happens(action(a , αdeduce ),t) ⇒ 
grammar for sentences φ; these are shown on the left half of 
 
[R9 ] 
the figure. The proof calculus is based on natural deduction


 C(t, ∀x. φ → φ[x →
� t])
[R8 ]
∃vK(a,t,valued(S(t
C(t, φ ↔ φ → ¬φ → ¬φ+ 1),
) v))),



 1 2 2 1 

[Jaśkowski, 1934], and includes all the standard introduction 
 K(a,t, happens(action(a ∗
, α ),t) ⇒ 


 intuit [R 10 ] 

 C(t, [φ1 ∧ . . . ∧ φn → φ] → [φ1 → . . . → φn → ψ]) 
and elimination rules for first-order logic, as well as addi- 



∃vB(a,t,valued(S(t + 1), v))) , 



tional rules for the modal operators; the rules are listed in the 


B(a,t,
K(a,t,
φ) B(a,t,
happens(action(a
φ → ψ) ∗
α
[R,11a
B(a,t,
] ),t) ⇒
φ) B(a,t, ψ) 


 intuit [R11b ] 
right half of the figure. 
 B(a,t, ψ) B(a,t, ψ ∧ φ) 


 ¬∃vK(a,t,valued(S(t + 1), v))),  
The music calculus is a specialization of DC EC ∗ with log- 

 S(s, h,t, φ)



 � �
happens(action(a ∗
, αintuit ),t)∧
�� 
ical symbols for reasoning over music, and was introduced in [R12 ]

earlier work [Ellis et al., 2012], where it is discussed in some


K a,t, ¬
B(h,t, B(s,t, φ)) ,
 happens(action(a∗ , α ),t) 
 � I(a,t, happens(action(a∗ , α),t � )) deduce �





detail. Figure 3 shows a part of the signature of the music cal- 
 happens(action(a ∗
, α [R13 ]
),t)∨ 

 intuit 
culus used in the present paper. (We omit parts of the calculus 
 K a,t,

P(a,t, happens(action(a ∗ , α),t))
∗ , 



 happens(action(a , α deduce ),t) 

not necessary for this paper.) The logic-based part of Handle 

 � B(a,t,
� ∗
φ) B(a,t, O(a ,t, φ, happens(action(a , α),t ))) �� ∗ � 


is based on the music calculus. 
 cost(α deduce ,t) − cost(α intuit ,t) ≤ γ ⇒ 



 B a,t,

O(a,t, φ, happens(action(a , α),t )) �
, 


More generally, the logical framework of DC EC ∗ lets us 


I(a,t, happens(action(a
K(a,t, I(a

, α deduce
∗ ,t, happens(action(a∗ , α),t � ))) ),t)) [R 14 ] 


model and compare different theories and implementations 
 � � �� 



 cost(α
φ ↔ ψdeduce ,t) − cost(α intuit ,t) > γ ⇒ 


of creativity, intuition, and cognition in a precise manner. As  B a,t, [R15 ] ∗ 
O(a,t,I(a,t,
φ, γ) ↔happens(action(a
O(a,t, ψ, γ) , αintuit ),t))
explained below, the model of intuition used in this paper is
based on the cost of computation. The agent needs to com-
pute a particular value S(t), the state of the performance at the
current time. The agent knows that there is a certain action 2.4 On Logic and Intuition
αdeduce which when performed will lead to the agent know- It is not unfair to say that logic and intuition appear to be
ing this value, and knows that there is another action αintuit contra-positional: ‘intuitive’ seems to be the opposite of ’rig-
t ::= x : S | c : S | f (t1 , . . . ,tn )

p : Boolean | ¬φ | φ ∧ ψ | φ ∨ ψ | φ → ψ | φ ↔ ψ | ∀x : S. φ | ∃x : S. φ
P(a,t, φ) | K(a,t, φ) | C(t, φ) | S(a, b,t, φ) | S(a,t, φ)
φ ::=
B(a,t, φ) | D(a,t, holds( f ,t � )) | I(a,t, happens(action(a∗ , α),t � ))

O(a,t, φ, happens(action(a∗ , α),t � )) and operating typically when a person seeks to understand a
Figure 2: Rules of Inference of DC EC ∗ situation or another person.
Rules of Inference 2. “Applied intuition,” directed towards the solution of a problem
[R1 ] [R2 ] or accomplishment of a task.
C(t, P(a,t, φ) → K(a,t, φ)) C(t, K(a,t, φ) → B(a,t, φ))
3. “Free intuition,” which arises in reaction neither to socioaffec-
C(t, φ) t ≤ t1 . . .t ≤ tn K(a,t, φ)
[R3 ] [R4 ] tive stimulus nor to a specific search process, but rather in-
K(a1 ,t1 , . . . K(an ,tn , φ) . . .) φ volves “a feeling of foreboding concerning the future.”
t1 ≤ t3 ,t2 ≤ t3
[R5 ] It is doubtless the third type of intuition which has caused
C(t, K(a,t1 , φ1 → φ2 ) → (K(a,t2 , φ1 ) → K(a,t3 , φ2 )))
the diminution of intuition’s perceived value: how can some-
t1 ≤ t3 ,t2 ≤ t3
[R6 ] thing so intangible as a ‘premonition’ have any scientific
C(t, B(a,t1 , φ1 → φ2 ) → (B(a,t2 , φ1 ) → B(a,t3 , φ2 ))) value? Yet all three forms of intuition, including the second
t1 ≤ t3 ,t2 ≤ t3
[R7 ]
type, which is most immediately relevant to our work, have
C(t, C(t1 , φ1 → φ2 ) → (C(t2 , φ1 ) → C(t3 , φ2 ))) their bases in a structured, if unconscious, analysis and inter-
[R8 ] [R9 ] pretation of perceived information. In our case, the perceived
C(t, ∀x. φ → φ[x �→ t]) C(t, φ1 ↔ φ2 → ¬φ2 → ¬φ1 ) information is structured.
C(t, [φ1 ∧ . . . ∧ φn → φ] → [φ1 → . . . → φn → ψ])
[R10 ] Because of its time-based nature, music is a good example
where timely decisions have to be made, which immediately
B(a,t, φ) B(a,t, φ → ψ) B(a,t, φ) B(a,t, ψ)
[R11a ] [R11b ] implies, on the Simon-Pollock view here affirmed, that in-
B(a,t, ψ) B(a,t, ψ ∧ φ)
tuition is needed. If time does not allow the deduction of a
S(s, h,t, φ) I(a,t, happens(action(a∗ , α),t � )) logical answer in the realm of music, humans must exploit
[R12 ] [R13 ]
B(h,t, B(s,t, φ)) P(a,t, happens(action(a∗ , α),t)) mechanisms to make a quick “best guess” using a sub-logical
B(a,t, φ) B(a,t, O(a∗ ,t, φ, happens(action(a∗ , α),t � )))
mechanism. Of course, we seek a system that does more than
O(a,t, φ, happens(action(a∗ , α),t � ))
make a random guess by using an informed approach that at
[R14 ] least approximates the correct answer.
K(a,t, I(a∗ ,t, happens(action(a∗ , α),t � )))
Our agent, CAIRA, includes a logic- and music-calculus-
φ↔ψ based reasoning system, Handle, so it can “understand” basic
[R15 ]
O(a,t, φ, γ) ↔ O(a,t, ψ, γ) concepts of music and use a hypothesis-driven approach to
perform with other musicians. The agent uses logic to de-
termine the current ensemble state from observing acousti-
Figure 3: Signature of the Music Calculus cal features in real time. CAIRA assesses the current ensem-
ble state of an improvisatory music group, thereby addressing
S : =Recommendation ⊆ Action | Person ⊆ Agent | Level | Label questions of who is playing a solo or whether it is likely that
MaxLevel : Person × Level → Boolean the improvisation will come to an end soon (see Table1). For
Level : Person × Level → Boolean this purpose, the agent analyzes the relationships between the
WithinOne : Level × Level → Boolean tension arcs of each musician, including the tension arc mea-
f : = LessThan : Level × Level → Boolean sured from CAIRAs own acoustic performance. The tension
a, b, c, d, e, f , g : Level arc describes the current perceived tension of an improvisa-
Tutti : Label → Boolean tion, instrumentally estimated used acoustical features such
low, high : Label and loudness and roughness [Braasch et al., 2011].
However, it often takes CAIRA too long to determine the
current ensemble state using logic, so we conceived a process
orous,’ or ’logical,’ or ’formal.’ Even merely trying to char- to simulate intuition, one that builds on a database of all cur-
acterize these two concepts brings an immediate contrast: rently known ensemble-state configurations. If an ensemble
logic seems well-defined, well-understood; intuition elusive state is unknown for an occurring tension arc configuration
and illusory. But for us, intuition is simply a rapid form of (a vector with individual tension arcs for each musician), the
non-intellectual processing containing no explicit reasoning process guesses the correct answer by taking into account the
(whether that reasoning is deductive, inductive, analogical,
1 [1995] has called “quick
abductive), essentially what Pollock
known states and making an informed guess using the eight
nearest neighbors and distance weighting.
and dirty” cognition. The same basic idea, as we understand
it, can be traced back to a basic idea presented by Herb Simon
1972, whose notion of “satisficing” corresponds to what we 3 CAIRA Overview
have in mind for intuition: rapid heuristic-driven decision- Put briefly, CAIRA, the Creative Artificially-Intuitive and
making and action employed when logic-based cognition is Reasoning Agent, the full architecture of which was pre-
too slow to get the job done. (See also [Bringsjord and Fer- sented at C3GI 2012, is a system able to provide guidance
rucci, 2000].) to ensembles, compose music, and perform in an improvi-
As to the nature of intuition, Raidl and Lubart [2000] enu- sational fashion. A major goal of the CAIRA project is to
merate three essential types: develop an understanding of the relationship between logic
1. “Socioaffective intuition,” concerning interpersonal relations and intuition within the scope of musical performances.
Instead of using a simple audio-to-MIDI converter, the
agent uses standard techniques of Computational Auditory Figure 4: Handle Architecture
Scene Analysis (CASA), including pitch perception, tracking
of rhythmical structures and timbre and texture recognition.
This approach allows CAIRA to extract further parameters re-
lated to sonic textures and gestures in addition to traditional
music parameters such as duration, pitch, and volume. This
multi-level architecture enables CAIRA to process sound us-
ing bottom-up processes simulating intuitive listening and
music performance skills as well as top-down processes in
the form of logic-based reasoning. The low-level stages are
characterized by a Hidden Markov Model (HMM) to recog-
nize musical gestures and an evolutionary algorithm to cre-
ate new material from memorized sound events. The evolu-
tionary algorithm presents audio material processed from the
input sound, which the agent trains itself on during a given
session, or from audio material that has been learned by the system as formerly. The MATLAB client interfaces with Fil-
agent in a prior live session. The material is analyzed using ter and MaxMSP via a simple set of OpenAudio commands
the HMM machine listening tools and CASA modules, re- sent periodically via the network, which provide information
structured through the evolutionary algorithms and then pre- on the current tension level of the three players as integers
sented in the context of what is being played live by the other in the range 1-7. These values are used to look up the cur-
musicians. rent ‘perceived state’ of the performance for Filter and the
CAIRA consists of two standalone modules, FILTER and appropriate information returned. If, however, the state has
Handle. not been encountered, a best guess is made, currently using a
weighted nearest-neighbor heuristic, and the state is marked
3.1 FILTER for formal calculation using an implementation of the mu-
The artificially-intuitive listening and music performance sic calculus in the Lisp-based SNARK automated theorem
processes of CAIRA are simulated using the Freely Improvis- prover. Due to the length of time it takes to perform the op-
ing, Learning and Transforming Evolutionary Recombination erations, all logic calculations are done off-line between per-
system (FILTER) [van Nort et al., 2009; van Nort et al., 2010; formances. An overview of the instantiation of the Handle
van Nort et al., 2012], which uses a Hidden Markov Model architecture corresponding to the work reported on herein is
(HMM) for sonic gesture recognition, and it utilizes Genetic given in Figure 5.
Algorithms (GA) for the creation of sonic material. In the
first step, the system extracts spectral and temporal sound
features on a continuous basis and tracks onsets and offsets Figure 5: Instantiation of Handle Architecture
from a filtered version of the signal. The analyzed cues are
processed through a set of parallel Hidden Markov Model HB+,+I'
(HMM)-based gesture recognizers. The recognizer deter-
mines a vector of probabilities in relation to a dictionary of $/01'
%23456738"9:3/;<43'
reference gestures. The vector analysis is used to determine
parameters related to maximum likelihood and confidence, !"#$%&'
*3</3A:'

,*+&='>/59909?'@5A04'B<C45C5AD'
and the data is then used to set the crossover, fitness, muta- 930?FG15/'
+9<CEA0A'
F35/0A64'
tion, and evolution rate of the genetic algorithm, which acts @5C67</0<:3'
<9<CEA0A'
on the parameter output space [van Nort et al., 2009]. -+.#"*%'

3.2 Handle =91JC3K?3'


#1?04L'M'
N3/;1/O<943'
G<A3'
Handle is a representative version of the logic- and intuition- <9<CEA0A'

based segments of the larger entity which is CAIRA, but ()*+,'

nonetheless functions as a standalone creativeT + machine


conductor, as explored in prior work by Ellis and Bringsjord
(see Figure 4). Whereas the prior version of Handle was de-
signed to work with score-based, structured music with min-
imal deviation from a known form, its current incarnation is
intended to explore the field of unstructured music in the form 4 Demonstration
of improvised, free “fusion jazz.”
Now as formerly, Handle’s architecture consists of two ma- 4.1 Method
jor components, one running in MATLAB and the other in In the demonstration, Handle, as the logical reasoning por-
Common Lisp. The most notable change is that the MAT- tion of CAIRA, accepts inputs from three audio sources: two
LAB system is now the controlling client rather than the Lisp of these correspond to two human players and the third is
the output from FILTER. Computational Audio Scene Anal-
ysis (CASA) is used on all three audio streams at various Figure 7: Sample Output from SNARK
points during the piece to extract data on various features
of each player’s performance. These features — tensionB ,
tensionMIR , dynamic, tempo, valence and activity — are
mapped to an integer scale in the range 0–6 [Braasch, 2012];
using combinations of 3-tuples of these features, informa-
tion on the current state of the performance may be derived.
Currently defined states and their definitions may be found
in Table 1; possible states are solos for players A, B and C,
low- and high-tension tutti, ending and an “uncertain” state.
The latter state indicates there is insufficient information for
Handle to supply meaningful data. When CASA informa-
tion is supplied to Handle, it returns an integer value repre-
senting current state of the piece, using the definitions in Ta-
ble 1. Based on this recommendation, FILTER then adjusts
its own parameters to take the recommendation into account
in its performance. At present, only tensionB (as defined in
[Braasch, 2012]) is used for state calculations.
As with human players, Handle may well have gaps in its 4.2 Logic Subsystem
knowledge when it is asked for its opinion on a state it has not
encountered. In this case it will attempt an “intuitive leap,” Handle operates first and foremost by logical deduction of
using a predetermined heuristic to calculate a “best guess” re- state using axioms derived from predetermined rules. We use
sponse. If such a guess is made, it is stored in the knowledge- the rules given in Table 1 to construct premises in first-order
base together with a marker that it is a guess. At the end of logic by which one may prove each state. Numerical values
the song, all states where a guess exists are passed to SNARK received via the CASA system are mapped to a correspond-
for formal proof, and the resultant values are stored perma- ing player-level value (0 = A, 1 = B, etc.); these are used for
nently in the knowledge-base. Figures 6 and 7 show output comparisons during theorem-proving, replacing the ordinary
from Handle operating over a short piece of music and the arithmetic operators which SNARK cannot reason over.
beginning of the resulting SNARK proof resulting from an
unknown state, respectively.
Table 1: Ensemble states as determined by TensionB . A, B,
and C represent the calculated state of Musicians A, B, and
Figure 6: Sample Output from Handle CAIRA respectively. CAIRA may disregard the recommen-
dation and respond differently. CAIRA must decide which
state to prefer. States are ordered hierarchically such that
overlap goes to the higher ranking state [Braasch, 2012].
Ensemble States Musician A Musician B CAIRA C
Solo A A>B+1 B+1<A C+1<A *
Solo B A+1<B B>A+1 C+1<B *
Solo C 0<A<5 0<B<5 ??
Low Level Tutti 0<A<5 0<B<5 ??
High Level Tutti A>4 B>4 C>4 *
Ending A=0 B=0 C=0 *
Uncertain None of the above ??

The following are some key examples of premises in FOL


that we use to define the music states [Valerio, 2013], some
of which are visible as SNARK axioms in Figure 7:
1. A person has a solo iff they have the highest level and there are
no other players within one level.

∀p1 : Person Solo(p1 ) ↔ ∃l1 : Level MaxLevel(p1 , l1 )∧


� � ��
∃p2 : Person Level(p2 , l2 ) ∧ WithinOne(l1 , l2 )
¬
∃l2 : Level ∧ (p1 �= p2 )
2. There is a low-level tutti iff every player is has a level greater
than zero and less than five, and no player has a solo. Figure 8: Chromatic Representation of Complete 3D State
Cube for tensionB
Tutti(low) ↔∀p : Person ∃l : Level
� �
Level(p, l) ∧ LessThan(l, f )
∧ LessThan(a, l)) ∧ ¬Solo(p)
3. There is a high-level tutti iff every player has a level greater
than four.
Tutti(high) ↔ ∀p : Person ∃l : Level(Level(p, l) ∧ LessThan(e, l))

4. It is the end iff all players are at level zero.


End ↔ ∀p : Person (Level(p, a))

By determining states logically in this fashion, the agent


is guided to playing in a fashion that is expected on a so-
cial level. Given that the state determines certain weight-
ings within the FILTER machine-learning aspect, however,
the agent is still able to improvise freely within the expected
range as a response to the other players. In effect, this is the
same as a human player reasoning about what the other play-
ers are doing and what they intend, and reacting accordingly.
index1 = find(x[i] ≥ αx & y[i] ≥ αy & z[i] ≥ αz
Similarly, when the agent is first initialized, it has no knowl-
edge about what combinations of inputs relate to which state index2 = find(x[i] ≥ αx & y[i] ≤ αy & z[i] ≥ αz
and must also learn by reasoning about new inputs over time. index3 = find(x[i] ≥ αx & y[i] ≤ αy & z[i] ≤ αz
index4 = find(x[i] ≥ αx & y[i] ≤ αy & z[i] ≤ αz
4.3 Approaching Intuition index5 = find(x[i] ≤ αx & y[i] ≥ αy & z[i] ≥ αz
Handle is a logical system that uses axioms provided and index6 = find(x[i] ≤ αx & y[i] ≥ αy & z[i] ≤ αz
computed in SNARK to derive state values matching to com- index7 = find(x[i] ≤ αx & y[i] ≤ αy & z[i] ≥ αz
binations of 3-tuples of musical features. However, again, in index8 = find(x[i] ≤ αx & y[i] ≤ αy & z[i] ≤ αz
cases where precomputed information is unavailable, Handle
has been designed to make an “intuitive leap,” by using one Admittedly simplistic, this approach nonetheless conforms
of a number of possible heuristics to guide its judgment. to our definition of ‘intuition’ given in section 2.4. In order
We note at this point that for the given three-player scenario to determine the efficacy of this approach, a simulation was
we may represent each state as a color corresponding to the run of over 400 sessions to determine the percentage error of
most extreme coordinates for a feature 3-tuple. For example, the nearest-neighbor approach by picking 10 random player
a solo for player A corresponds to red (6, 0, 0 → 255, 0, 0), inputs each session. It can be seen that, as new states were
Solo B is 0, 6, 0, which encodes to green, and Solo C is 0, 0, 6, accumulated in the knowledge base, the accuracy of the near-
blue. Low-tension tutti is represented by grey (calculated by est neighbor method improves (Figure 10). A comparator is
finding the centroid of the state), high-tension tutti is white, also provided, based on an approach which guesses random
end is black, and uncertainty is determined by the coordinate on uncomputed states.
values themselves. After computing all state combinations, We acknowledge that we may stand accused of not prop-
such a representation appears as in Figure 8. erly addressing, or exploring, the concept of ‘intuition’ in this
The current approach for resolving an unknown state at iteration of Handle. We do not deny this. However, in our
runtime is to use a nearest-neighbor algorithm. Given the own defense, we would reply by stating that ‘intuition’, as it
(safe) assumption that the states for all 8 corners of the state is commonly understood, requires a much larger dataset than
cube are absolutely known ab initio, the value of any un- the one we are currently using, one that in all likelihood con-
known state may be derived by taking the average of the sum tains a certain amount of conflicting data, together with an
of the distance vector between it and each (fully known) cor- appropriate means of accessing and using those data. The
ner. As more and more states are derived and the knowledge- current incarnation of Handle functions using only one mu-
base becomes progressively more complete, however, we sical feature, tensionB , yielding a dataset containing a pal-
would prefer to use more accurate information; that is, state try 7×7×7=343 states; as a consequence, ‘intuitive leaps’
values which are more proximate to the unknown state. These will inevitably be barely distinguishable from mere arithmetic
8 cubes, nearest to the unknown state in diagonal distance, are computation. But when we do have the additional informa-
determined using a simple vector approach and the mean of tion, the problem inevitably arises of how we manage it.
their values returned. Thus, given an unknown state α with One of the most interesting effects to arise from this itera-
coordinates (αx , αy , αz ) and �x�, �y�, �z� vectors of the co- tion of Handle is the concept of the state cube, shown above in
ordinates of all known states, the 8 nearest neighbors of α Figure 8. Originally developed by Valerio based on work by
can be found as follows: Ellis, it has considerable implications for computation using
can model this reasoning — and build agents that can per-
Figure 9: Nearest-Neighbor Error vs. Random Guess form such complex reasoning in real time. We now present
100
one embryonic example of this, by examining a pattern of
neighbor weighted pick reasoning carried out by a conductor:1
90 random pick
The conductor could be unsure whether a performance
80
should have a certain attribute; for example, proportional
tempo between movements. If the conductor then comes
Performance State Error [%]

70

60
to believe, via some process, that the composer believed
that the performance’s having this attribute would result
50
in a pleasing experience for the listener, then the conduc-
40 tor may start believing likewise. Naturally, this “epis-
30 temic chain” would result in performances that have pro-
portional tempo.
20
The reasoning pattern above can be computationally real-
ized using the following proof in DC EC ∗ . We first present
10

0 the necessary premises (Performance and Emotion are sub-


−10 sorts of Action and Fluent respectively):
0 50 100 150 200 250 300 350 400
Number of music sessions
A1 ∀α, f ,t,t � .D(h,t, holds( f ,t)) ∧ B(h,t, initiates(action(h∗ , α), f ,t � ))

⇒ I(h,t, happens(action(h∗ , α),t + 1))


multiple musical features, and may even lead to a mapping A2 ∀ρ : Performance, σ : Emotion,t,t � .
for the use of visual stimuli in the composition of music. (For  
B(h,t, B(c,t, initiates(ρ, holds(feels(σ, listener),t � ))))
full information, see [Valerio, 2013].) Most importantly to  � �
 � 
Handle, however, it offers us a resource to manage and easily  ∧ ¬ B(c,t, initiates(ρ, holds(feels(σ, listener),t )))∨ 
explore the large data-space resulting from the many possible

B(c,t, ¬initiates(ρ, holds(feels(σ, listener),t )))
combinations of musical features that we already capture, as ⇒ B(h,t + 1, initiates(perform(p), holds(feels(s, listener),t � )))
� � � ���
well as any we might choose to capture in the future. action(h∗ , perform(tempoprop )),
A3 ∀t,t � .¬ B c,t, initiates �
feels(pleased, listener),t
� � � ���
action(h∗ , perform(tempoprop )),
Figure 10: Chromatic State Cube Representing Addition of ∧ ¬ B c,t, ¬initiates �
feels(pleased, listener),t
Two Single Variables
A4 ∀t,t � D(h,t, holds(feels(pleased, listener),t � ))

The informal English translations of the premises are given


below:
A1 If the conductor desires something and believes a certain ac-
tion will accomplish it, the conductor intends to perform that
action.
A2 For all performances p and emotional states e, if the conduc-
tor does not know whether p can lead to state e in the listener,
but believes that the composer believes p can lead to emotion
e in the listener, then the conductor will then come to believe
the same.
A3 The conductor himself is ambivalent about the above state-
ment that movements with proportional tempos will lead to the
listener being pleased.
A4 The conductor desires that the listener be pleased.
Assume we have constant symbols with t1 < t2 < t3 < . . .
and that if t ≡ ti then ti+1 ≡ ti + 1, and that we also have the
5 Deeper Use of Logic: Epistemic Reasoning following along with the above five premises (where the com-
To indicate here a deeper use of formal logic than for encod- poser communicates with the conductor in some fashion):
� � ��
ing relatively simple extensional rules as in the demonstra- action(a, perform(tempoprop )),
tion just shared, consider that, as we have noted in earlier ∀t � S c, h,t1 , ∀a : initiates
feels(pleased, listener),t �
work [Ellis et al., 2012], the process of conducting is inter-
twined with rich epistemic, doxastic, and temporal reason- 1 This is our “logification” of actual reasoning performed by the
ing. DC EC ∗ provides a formal framework with which we famous conductor Marriner; see [Anderson, 1985].
[Bringsjord and Licato, 2012] S. Bringsjord and J. Licato. Psycho-
Using universal elimination, R12 , and universal introduc- metric Artificial General Intelligence: The Piaget-MacGuyver
tion with the above, we obtain the following: Room. In P. Wang and B. Goertzel, editors, Foundations of Arti-
ficial General Intelligence, pages 25–47. Atlantis Press, Amster-
φch ≡ dam, The Netherlands, 2012. This url is to a preprint only.
� � � ��� [Bringsjord and Schimanski, 2003] S. Bringsjord and B. Schiman-
action(a, perform(tempoprop )),
∀t � .B h,t1 , B c,t1 , ∀a : initiates ski. What is Artificial Intelligence? Psychometric AI as an An-
feels(pleased, listener),t �
swer. In Proceedings of the 18th International Joint Conference
on Artificial Intelligence (IJCAI–03), pages 887–893, San Fran-
Let cisco, CA, 2003. Morgan Kaufmann.
φh
listener

[Bringsjord et al., 2001] S. Bringsjord, D. Ferrucci, and P. Bello.
∀t � .B(h,t2 , initiates(action(h∗ , perform(tempoprop )), feels(pleased, listener),t � )) Creativity, the Turing test, and the (better) Lovelace test. Minds
and Machines, 11:3–27, 2001.
Finally, we have the following proof that the conductor will [Bringsjord, 2008] Selmer Bringsjord. Declarative/Logic-Based
try to bring about a performance with proportional tempo. Cognitive Modeling. In Ron Sun, editor, The Handbook of Com-
putational Psychology, pages 127–169. Cambridge University
A2 φch A3 Press, Cambridge, UK, 2008.
�DC EC ∗
φh
listener
A1 A4 [Bringsjord, 2011] Selmer Bringsjord. Psychometric Artificial In-
�FOL
I(h,t2 , happens(action(h∗ , perform(tempoprop )),t3 ))
telligence. Journal of Experimental and Theoretical Artificial In-
telligence, 23(3):271–277, 2011.
6 Next Steps [Cope, 2005] David Cope. Computer Models of Musical Creativity.
Cambridge: MIT Press, December 2005.
While the current research has had some genuinely interest-
[Ellis et al., 2012] Simon Ellis, Naveen Sundar Govindarajulu,
ing outcomes, as always it creates at least as many questions Selmer Bringsjord, Alex Haig, Colin Kuebler, Joshua Taylor,
as it answers, for example: Jonas Braasch, Pauline Oliveros, and Doug Van Nort. Creativity
1. Further development of axioms and rules in the music calculus and Conducting: Handle in the CAIRA Project. In Proceedings
is required. of the Workshop ”Computational Creativity, Concept Invention,
and General Intelligence”, 2012.
2. We will explore the possibilities of bi- and multi-variate anal-
yses of musical performances, together with different ways of [Inhelder and Piaget, 1958] B. Inhelder and J. Piaget. The Growth
encoding the combined state data. Metrics we have considered of Logical Thinking from Childhood to Adolescence. Basic
include tempo or dynamic, at given moments or across a pe- Books, New York, NY, 1958.
riod, and another form of tension derived from the ‘emotional’ [Jaśkowski, 1934] Stanislaw Jaśkowski. On the Rules of Supposi-
strength of a piece of music at a given time. tions in Formal Logic. Studia Logica, 1:5–32, 1934.
3. It would be useful and interesting to explore different heuris- [Pollock, 1995] J. Pollock. Cognitive Carpentry: A Blueprint for
tics for the nearest-neighbor analysis generally. Are there cal- How to Build a Person. MIT Press, Cambridge, MA, 1995.
culations which work well in some cases but not others? What
happens if the heuristic is biased toward states nearer the query [Raidl and Lubart, 2000] Marie-Helene Raidl and Todd I. Lubart.
state and against those further away? How good is the estimate An Empirical Study of Intuition and Creativity. Imagination,
compared to the actual calculated state, and is there any form Cognition and Personality, 20(3):217–230, 2000.
of pattern to the wrong guesses? [Simon, 1972] Herbert Simon. Theories of Bounded Rationality.
Acknowledgements: This project is made possible by generous In C. McGuire and Roy Radner, editors, Decision and Organi-
zation, pages 361–176. North-Holland, Amsterdam, The Nether-
sponsorship from both NSF (grant no. 1002851) & the John Tem-
lands, 1972.
pleton Foundation.
[Valerio, 2013] Joseph Valerio. Heuristic for Multi-Variable Music
State Recommendations. Master’s thesis, Rensselaer Polytechnic
References Institute, Troy, NY, April 2013.
[Anderson, 1985] Ronald Anderson. Marriner on Mozart. The [van Nort et al., 2009] Doug van Nort, Jonas Braasch, and Pauline
Choral Journal, 26(1):23–26, 1985. Oliveros. A system for musical improvisation combining sonic
[Braasch et al., 2011] J. Braasch, S. Bringsjord, C. Kuebler, gesture recognition and genetic algorithms. In Proceedings of the
P. Oliveros, A. Parks, and D. Van Nort. CAIRA – a creative SMC 2009–6th Sound and Music Computing Conference, pages
artificially-intuitive and reasoning agent as conductor of telem- 131–136, Porto, Portugal, July 2009.
atic music improvisations. In Proc. 131th Conv. Audio Eng. Soc., [van Nort et al., 2010] Doug van Nort, Pauline Oliveros, and Jonas
2011. Paper Number 8546. Braasch. Developing Systems for Improvisation based on Listen-
[Braasch, 2012] Jonas Braasch. The µ-cosm project: an introspec- ing. In Proceedings of the 2010 International Computer Music
tive platform to study intelligent agents in the context of music Conference (ICMC 2010), New York, NY, June 2010.
ensemble improvisation. (to appear), September 2012. [van Nort et al., 2012] Doug van Nort, Jonas Braasch, and Pauline
[Bringsjord and Ferrucci, 2000] S. Bringsjord and D. Ferrucci. Ar- Oliveros. Mapping to musical actions in the FILTER system. In
tificial Intelligence and Literary Creativity: Inside the Mind of The 12th International Conference on New Interfaces for Musical
Brutus, a Storytelling Machine. Lawrence Erlbaum, Mahwah, Expression (NIME), Ann Arbor, Michigan, May 2012.
NJ, 2000.
How Models of Creativity and Analogy Need to Answer the Tailorability Concern∗
John Licato, Selmer Bringsjord, Naveen Sundar Govindarajulu
{licatj,selmer,govinn}@rpi.edu
Rensselaer AI and Reasoning (RAIR) Lab
Department of Computer Science
Rensselaer Polytechnic Institute (RPI)
Troy NY 12180 USA

Abstract nent of artificial general intelligence as well [Bringsjord and


Licato, 2012; Hofstadter, 2001].
Analogy is a major component of human creativity. With the above in mind, it makes sense to develop models
Tasks from the ability to generate new stories to the of analogy, both computational and theoretical. Much work
ability to create new and insightful mathematical has been done in this direction; a few implementations in-
theorems can be shown to at least partially be ex- clude SME [Falkenhainer et al., 1989], LISA [Hummel and
plainable in terms of analogical processes. Artifi- Holyoak, 2003], HDTP [Gust et al., 2006], and recently our
cial creativity and AGI systems, then, require pow- own META-R [Licato et al., 2013]. On the surface, it seems
erful analogical sub-systems—or so we will soon that the current generation of analogical systems sufficiently
briefly argue. It quickly becomes obvious that a capture and explain all of the phenomena commonly associ-
roadblock to such a use for analogical systems is a ated with analogical reasoning, and that they will eventually
common critique that currently applies to every one reach levels characteristic of human cognition. It may well
in existence: the so-called ‘Tailorability Concern’ be the case that the most important principles underlying the
(TC). Unfortunately, TC currently lacks a canoni- nature of analogy have been expressed. But a serious objec-
cal formalization, and as a result the precise con- tion has been raised recently which, as will be argued, should
ditions that must be satisfied by an analogical sys- be the primary focus of analogical researchers over the next
tem intended to answer TC are unclear. We rem- few years — at least if any significant further progress is to
edy this problem by developing a still-informal but be made in the direction of creativity and AGI.
clear formulation of what it means to successfully
The objection is raised by Gentner and Forbus (2011).
answer TC, and offer guidelines for analogical sys-
They call it the ‘tailorability concern’ (TC), and the objec-
tems that hope to progress further toward AGI.
tion echoes a common criticism of cognitive systems in gen-
eral: that they operate on toy examples manually constructed
1 Introduction in such a way as to guarantee the desired solution. How-
ever, though this concern may have been stated in many forms
Creativity and the ability to reason analogically have a strong throughout the years [Mitchell and Hofstadter, 1995], it lacks,
relationship [Finke, 1990; Finke et al., 1992; Holyoak and to our knowledge, a formulation clear enough to anchor pro-
Thagard, 1995; Gentner et al., 1997]. For example, many of ductive scientific discussion. And this ambiguity in turn neg-
the greatest lines in literature make use of familiar analogical atively impacts not only the relevant science, but AI engi-
processes: “A rose by any other name would smell as sweet” neering as well: absent a definition of TC, it is difficult to
only makes sense if that rose and its name are understood to understand precisely what an analogical system must do in
correspond to Romeo and his family name (analogical map- order to successfully answer TC. In the remainder, we take
ping). The subsequent analogical inference is that just as the steps toward addressing this problem as it applies to analogi-
nature of the rose is independent of its name, so is Romeo’s cal systems.
nature independent of his.
Even if one believes that overuse of analogy can be harm-
ful to creative thought [Riesbeck and Schank, 1989], many 2 The Tailorability Concern
researchers argue that the ability to determine analogical sim-
ilarity is important at least for combinatorial creativity, which A frequently appearing criticism of cognitive systems in gen-
seems to be very easy for humans but very difficult for AI sys- eral is that they are only applied to manually constructed ‘toy
tems [Boden, 1995; Boden, 2009]. Furthermore, the impor- examples’, a problem many researchers in the field them-
tance of analogy is of course not limited to creativity, as ana- selves acknowledge. Gentner and Forbus (2011) referred to
logical ability has been identified as an indispensable compo- it as the tailorability concern (TC): “that is, that (whether
knowingly or not) the researchers have encoded the items in

With profound gratitude, the support of both the John Temple- such a way as to give them the desired results” [Gentner and
ton Foundation and AFOSR is acknowledged. Forbus, 2011].
B4#(=%/748"1C/ pers.
The field of case-based reasoning (CBR), which overlaps
!"#$%&'(%")%(&*(%$$#(%&+%",%(-./*(%$$#(%&01"2--./3245&+%",%(./01"2./5")%(./616%--/
'(%")%(&71"8%)%(&+%",%(-./71"8%)%(&01"2--/ quite heavily with that of analogical reasoning (AR), also
91:#1;&5")%(-/
32")<)46&5")%(-/
deals with some of the issues raised by the TC. There are dif-
!2%"(&+%",%(-/ fering opinions on what features distinguish the CBR and AR
approaches (see [Aamodt and Plaza, 1994; Burstein, 1989;
A"(D%)/748"1C/ Mantaras et al., 2005]), but two common themes are that CBR
3245&1=%>=#+%./=4?%%./@%")./+"(-/ tends to deal with source and target cases that come from the
'(%")%(&A%86%(")#(%&=4?%%-./A%86%(")#(%&1=%>=#+%--/ same domain, and cases are selected and adapted with some
91:#1;&=4?%%-/
32")<)46&=4?%%-/ clear pragmatic goal in mind. AR approaches, on the other
hand, try to be more generally applicable across different do-
mains, and tend to focus more on the mapping process that
actually determines analogical similarity. CBR approaches,
Figure 1: One version of the popular “heat flow” example then, deal with the TC by trading generality for effectiveness,
from Falkenhainer et al. (1989) used to test many analogical so that a program designed to work well in one domain (med-
systems. ical diagnosis, for example, is a popular field) may not work
so well in another without a significant amount of human as-
sistance.
Of course, nothing is wrong with toy examples per se. Unfortunately, the CBR approach of restricting generality
They can be extremely useful in demonstrating key concepts, does not sufficiently answer the TC. Analogy research can be
helping to illustrate particular qualitative strengths or weak- seen as centering around fundamental questions, one of them
nesses of computational models, or helping to get a new being: How can we find good analogies? The TC is especially
model off the ground. Indeed, the present authors plead guilty problematic because it forces analogy researchers to prove
to using toy examples this way; properly done, carefully cho- that their theoretical process is the answer to this question,
sen cases can be very useful as demonstrations-of-concept. and although it can be promising to see that some particular
But we should be careful not to treat these as the final proof approach produces good analogies in some limited domain,
of a system’s worth, since in most of these examples it is not no approach can constitute a completely satisfying answer to
clear that the principles used to solve them generalize to other this question unless it is versatile enough to perform well in
problems, and can be used to mechanically find useful solu- many domains.
tions just as effectively without human assistance. Any system that can answer the challenge of the TC will in-
For example, the well-known water-flow/heat-flow anal- stantly distinguish itself from every other analogical system
ogy (Figure 1) has been used as a demonstration of many currently in existence, since to our knowledge the only one
models of analogy [Falkenhainer et al., 1989; Holyoak and that has been able to do this with some degree of success is
Thagard, 1989; Wilson et al., 2001; Krumnack et al., 2007]. the SME-based family of systems [Gentner and Forbus, 2011;
But little to nothing is written about how the structural rep- Lovett et al., 2010; Forbus et al., 2009]. For this reason, we
resentations used in examples such as these are acquired in should clarify what it means to answer this challenge and dis-
the first place. One approach is to model the acquisition cuss why it is such a non-trivial feat.
of structured representations through sensory data (e.g., see
[Doumas et al., 2008]), and another is to presume the exis- 2.1 Answering the TC
tence of a large database of already-structured data (such as Gentner and Forbus (2011) suggest that there are two possible
that to which a neurobiologically normal adult might be ex- ways to answer the TC. One applies to visual domains, and
pected to have access), and some sort of filtering process that involves using automatic encodings of visual representations.
ensures that from this database, proper representations are se- The other more generally applicable direction involves two
lected and any unnecessary data that would produce incorrect key features: First, the use of pre-existing databases; second,
matching results are excluded.1 Yet even when such filter- an automated or semi-automated parsing process that goes
ing processes are proposed, they are not put to the test and from input text to a sufficiently rich semantic representation.
proven to perform well with a large source of data whose size A first attempt at a precise statement of what it means to an-
approaches that of a child, much less an adult. The TC right- swer the TC is as follows:
fully attempts to refocus efforts on these filtering processes, TCA1 A computational system of analogy answers the
by requiring that they demonstrate the ability to produce clean TC if, given no more than a pre-existing database and
source and target analogs as required by the analogical map- an unparsed input text, it is able to consistently produce
good analogies across many domains.
1
It would be less appropriate to urge a careful treatment of the TC At least one general intent behind TCA1 is clear: it at-
and tie it so closely to large semantic databases if they weren’t avail-
tempts to place emphasis on the filtering process (whose job
able. But over the past few years, natural-language processing and
semantic-web technologies have been progressing to the point where is, as we said, to either select some subset of available source
we now have access to large collections of semantic databases con- analogs from a large database and recommend only some of
taining wide-ranging general knowledge. These include Cyc [Ma- them for the more computationally expensive step of analog-
tuszek et al., 2006], Freebase [Bollacker et al., 2008], and DBPedia ical matching, or to automatically construct structured repre-
[Auer et al., 2007]. Many of these have easy-to-use interfaces. sentations from sensory data). By removing the reliance on
human intervention, TC ensures that the filtering is not done may not be as simple as it might seem. In the end, a model
manually in such a way that guarantees desired results. How- of analogy is a good one only if it produces good analogies,
ever, in order to truly answer the TC in a satisfactory way, and an analogy is a good one only if it produces plausible
we must be precise about its purpose and motivations: What hypotheses.2
concerns are behind the TC in the first place? Furthermore, This complicates things further. In toy examples, the infor-
TCA1 is hopelessly vague: the words ‘unparsed’ and ‘good,’ mation available to the analogical matcher is very limited;
if left open to interpretation, make it too easy for anyone to a hypothesis in these cases is plausible if the information
prematurely claim victory over TC. Also: Why is it important available to the analogical matcher does not provide a rea-
that the database be pre-existing? What degree of separation son to reject that hypothesis. But if we are suggesting that
must there be between the creators of the database and the the information available to the analogical system (or more
designers of the analogical system? For example, does the specifically, the filtering process) is actually a large database
database’s underlying knowledge-representation philosophy conceivably encompassing, say, all of the knowledge on the
need to overlap with that of the analogical system? World Wide Web, then how are we to realistically evaluate
the system’s performance?
What is a ‘good’ analogy? In light of these complications, we propose the following
Though the widely influential Structure-Mapping Theory alterations of TCA1 . If the word ‘good’ is replaced with
[Gentner, 1983] offers the systematicity principle and the ‘useful,’ then we connote an evaluation method that is not
one-to-one constraint as indicative features of a good anal- based in the particular metrics as defined by the analogical
ogy, it does not provide a clear quantitative measure of anal- theory itself (which can be controversial), but rather based
ogy quality. SME evaluates match scores by combining the in an intuitive notion that can and should be evaluated inde-
scores of the evidence provided by its match rules; this al- pendently of the model. In other words, researchers might
lows for a comparison between different matches of the same disagree on how to measure an analogical match, but whether
problem. But the resulting match score is not normalized, the resulting analogically inferred hypothesis is useful can be
and as a result match quality between different problems can- evaluated without any knowledge of the analogical matcher
not be compared [Forbus and Gentner, 1989]. Other analog- used. Of course, we do not claim that the word ‘useful’ is
ical models do not help much in this regard. Holyoak and completely unambiguous, but rather open to interpretation,
Thagard’s (1989) Multiconstraint Theory, for example, intro- as it might be domain dependent. Later in this paper, a better
duces additional criteria to evaluate what makes an analogi- replacement for the word ‘useful’ will be suggested. For now,
cal match a good one, making cross-domain analogical match the aim behind this move is to divorce the metric which deter-
quality more difficult to assess. mines the quality of an analogical match’s results (which may
This is especially problematic when considering TC. Given be very domain-dependent) from the theory-specific metric
a set of predefined problems and desired answers for all of that the matcher is specifically designed to optimize. That
them, the good analogy is simply the one that performs as (deceptively) small change gives us TCA2 :
intended. But when applied to problem sets where the best
analogy, or even the existence of a good analogy, is not clear TCA2 A computational system of analogy answers the
even to the persons running the experiment, do the guidelines TC if, given no more than a pre-existing database and
an unparsed input text, it is able to consistently produce
currently available still apply? useful analogies across many domains.
Paul Bartha (2010) offers a comprehensive theory of anal-
ogy that can be useful here. His goal is to produce a norma- What are acceptable databases?
tive account of analogy, and in the process he sheds light on In TCA2 , it is clear that the information available in the
the proper role of analogy within a larger context of reason- database used is a limiting factor in how knowledgeable the
ing. His Articulation Model is based on the idea that there inferences produced by the analogical system can be. The
are two features common to all good analogical arguments: suggestion phrased by Gentner and Forbus (2011) as “pre-
a prior association (a relationship in the source analog that existing databases” requires more clarification. The implica-
is transferred to the target), and a potential for generaliza- tion (at least as we interpret it) is that the dataset and the struc-
tion. But perhaps most relevant to our present discussion is tures within cannot have been constructed for the purpose of
his claim that by virtue of the analogical argument, the re- solving the particular toy examples that are of interest. Other-
sulting hypothesis inferred through an analogical argument wise this introduces bias and tailorability concerns, in spite of
contains no more than prima facie plausibility, which can be the best intentions of the designers. Two issues immediately
understood as something like a suggestion that the inferred come to mind. First, what is the proper level of separation
hypothesis is worth exploring further. If there is an indepen- between the database and the analogical system? Secondly,
dent reason to reject the hypothesis, such as a deductive argu- how large does the database have to be?
ment showing it leads to a contradiction, or contrary empirical
evidence, then it can be abandoned. 2
We might leave room here to exclude models of analogy that
The idea that the proper role of an analogical argument is have psychological or neurological plausibility as their primary end
to do no more than provide prima facie plausible hypotheses goals. In these cases, it might be the goal of the model to replicate
[Owen, 1990; Norton, Forthcoming] suggests that the rela- poor analogical reasoning as well, if it matches human performance.
tionship between an analogy’s match quality and its tendency But it is our assumption (at least in the present inquiry) that the ulti-
to produce hypotheses which can be independently verified mate goal of AGI research is not to model poor human reasoning.
The first question is at least partially answered by con- chitecture’s claim to psychological plausibility: if the archi-
sidering the level of representational agreement between the tecture can’t handle a dataset large enough to produce non-
database and the analogical system. For example, if the trivial answers, how can it be an accurate model of a process
database is a purely distributed one with no localist concepts used by human-level reasoners?4
whatsoever (which is, we acknowledge, an unlikely possi- Secondly, we could require robust and consistent perfor-
bility), and the analogical system is one that uses only lo- mance on a variety of input forms. For example, in the heat-
calist, explicitly structured data, then a significant level of flow problem (Figure 1) Mitchell and Hofstadter (1995) note
work will be needed to first extract the information from the that there are many possible ways to structure the input: heat
database and put it into the form that the analogical reasoner could be described as an object, or as an attribute of cof-
requires (this can be considered a re-representation step [Chi fee, or heat flow could be a relation with three rather than
et al., 1981; Yan et al., 2003; Krumnack et al., 2008]). The four arguments [Mitchell and Hofstadter, 1995]. Consistent
choice of database becomes important for this reason, and performance across various input forms puts more pressure
if no database exists that does not require a high level of on the analogical matcher’s re-representation algorithm(s),
re-representation, then it suggests a problem: Consider that rather than relying on a separate NLP module. This also al-
although proponents of localist, distributed, and hybrid rep- lows for a leveling of the playing field across different sys-
resentation styles make claims all the time about the scala- tems: In order to show an example adheres to this require-
bility of their assumptions of knowledge representation, the ment, a localist, structured analogical system would have to
researchers who have to design and work with large semantic demonstrate two things with regard to that particular exam-
databases actually have to put these ideas to the test. If the ple:
state-of-the-art research in the database field has significant
difficulty with some style of representation as required by an • Across multiple, more or less equivalent structural representa-
analogical matcher, then perhaps that matcher needs to care- tions of the same input data, and a wide variety of domains, the
fully consider its assumptions about the nature of knowledge matcher still produces the desired results.
representation, or else be able to easily extract the necessary • The desired results are still produced when the input is min-
data.3 The choice of database is therefore a sensitive one. imal; meaning any redundant information or structural con-
It may help to go back to the criticisms which may have structs which might be identified by critics as being used only
motivated this requirement in the first place. Hofstadter’s to aid the matcher can be removed.
group criticized SME by saying that “the analogy is already
effectively given in the representations” [Mitchell and Hof- Of course, if a system did happen to require no more than
stadter, 1995]. The information provided to the matcher is unstructured, natural-language descriptions as input, or direct
selected in such a way that it does not include a significant sensory data from visual representations, it would satisfy both
amount of extra data that would lead to false analogies, and of these conditions. This allows our criteria to encompass the
the choice of predicates and objects is done in such a way alternate route to answering the TC mentioned by Gentner
that presupposes the desired matching. This is a criticism and Forbus (2011)—a route which seeks to answer TC by not
composed of two points: the amount of information provided necessarily having a large database, but by having one that at
as input (which is not too little, or too much, but just right), least attempts to directly construct structured representations
and the nature of the information (the corresponding predi- from low-level sensory or natural-language data. The gener-
cates on both sides just happen to have the right amount of ality of these conditions allows us to claim the converse of
arguments in the necessary order). The system is not robust TCA2 , leading us to our next iteration:
enough to produce good analogies when given the same input TCA3 A computational system of analogy answers the
data in a variety of formats. TC if and only if given no more than either
Two possible changes can begin to answer these critiques.
One is to require that the databases are large enough, and • unstructured textural and/or visual data, or
the input minimal enough, to introduce a significant amount • a large, pre-existing database,
of false matches that would confuse less robust analogical
matchers. Critics claiming that the size of the input is too and minimal input, it is able to consistently produce use-
ful analogies and demonstrate stability through a variety
carefully chosen could be pointed to the large search space of input forms and domains.
that an analogical matcher must plumb given the small input
problem size. The larger the search space, the more impres- One might be satisfied with this set of criteria, which draws
sive the ability of the analogical matcher to select only a few its strength from its lack of commitment to any particular the-
potentially relevant source analogs. Furthermore, an inability ory of analogy, and its emphasis on large sets of non-tailored
to demonstrate scalability to large datasets weakens any ar- input. But TCA3 is undeniably ambiguous, and may not be
3 focused enough to guide any research program. We encour-
We do not mean here to say that what works best for large arti-
ficial databases is the same as what is employed by the human brain.
age the reader to take TCA3 and develop it further, but first,
But if a researcher declares that the structure of human knowledge to close out this paper we will take some steps of our own to
has certain properties, and large datasets cannot be created that do reduce some of its weaknesses.
not have those properties for reasons of scalability, then it should be
4
at least a weak hint that perhaps the assumption of those properties This is a common criticism of Hummel and Holyoak’s LISA
is not practicable. system; see [Gentner and Forbus, 2011].
((S,Q),V) • unstructured textual and/or visual data, or
• a vast, pre-existing database not significantly pre-
engineered ahead of time by humans for any partic-
NLP ular tests of A,
META-R VIVID is — in keeping with co-called Psychometric AI — able
(analogical (representation /
component) re-representation) to consistently generate analogies that enable A to per-
form provably well on precisely defined tests of cognitive
ability and skill.
To comment briefly on TCA4 , first note that we remove
Theorem
ATP
Provers
VATP the explicit requirement that the ability to find useful analo-
gies be stable across a variety of input forms and domains.
This is subsumed by the requirement of good performance
NLG on precisely defined tests; it is assumed that a sufficiently
difficult psychometric test would provide questions that are
both varied in their form (e.g., word problems, puzzle solv-
(A, J) ing, story comprehension) and in their domains. The implicit
requirement of domain variety rules out the possibility of an
artificial reasoning agent that can only process, for example,
Figure 2: Architecture-Sketch of an ADR System That An- certain types of math problems, as an acceptable answer to
swers TC. Input includes a Story and a Question, along with the TC.
Visual content. Output includes an Answer and—in the form Some might also see TCA4 and the use of PAI as too re-
of a proof or at least a rigorous argument—a Justification. strictive in that it relies too heavily on problem solving and
not enough on either creative thinking or the kind of every-
Strengthening the TC with Psychometric AI day analogical thinking that may not be goal-oriented in na-
We will make two important moves to sharpen TCA3 . ture.5 PAI, however, provides a tool for measuring those abil-
One, we turn to Psychometric AI (PAI) [Bringsjord, 2011; ities that, at least at the surface, don’t rely on directed prob-
Bringsjord and Schimanski, 2003], according to which, in a lem solving, such as reading comprehension. Additionally,
nutshell, commendable AI systems are those that demonstrate it is difficult to imagine that any research program in AGI
prowess on tests of various mental abilities from psychomet- would be able to demonstrate clear progress without showing
rics. Our second move is to imbed analogy-generation sys- increased performance in an ability that can be measured ac-
tems within broader AI problem-solving systems that make cording to some psychometric test. Non-goal-oriented ana-
use of additional forms of fundamental reasoning in human- logical reasoning is a good example of this principle: If
level intelligence; e.g., deduction. In particular, we place the cognitive processes underlying normal analogical reason-
TC within the context of the integration of analogical rea- ing when it is non-goal-oriented (as in everyday reasoning)
soning with deduction, which we dub analogico-deductive and when it is goal-oriented (as during psychometric test-
reasoning (ADR), and which we have explained and demon- ing) are largely the same, then an artificial system capable
strated elsewhere [Bringsjord and Licato, 2012]. An ADR of performing the latter has a strong claim to performing the
system does use analogy generation, but analogies are used former. A system that only has sporadic performance on
to guide solutions that can be rigorously verified by proof. psychometrically-measurable tasks is difficult to defend as
The architecture-sketch of an ADR system that accords with generally intelligent.
our pair of moves is shown in Figure 2. While we don’t have One might ask: Can a system do well on psychometric
the space to provide details here, this system receives puz- tests and still be subject to claims of tailorability? The an-
zles that are part linguistic and part visual in nature (e.g., swer, if the requirements in TCA4 are not met, is yes. PAI is
so-called seating puzzles), and harnesses not only our own not meant to be a replacement for the input format and large
analogical system (META-R), and not only ATP technology, database requirements we have been developing in this paper;
but visual theorem-proving; for formal details see [Arkoudas rather, it is only one possible sharpening of the ambiguous
and Bringsjord, 2009]. This architecture-sketch is inspired concepts in TCA3 . Other possibilities may exist, but we do
by, but abstracts and extends beyond, AI systems able to solve not at present know of any satisfying alternatives.
analogy-relevant problems. One such inspiring system is by
Lovett, Forbus, and Usher [Lovett et al., 2010]; it can solve 3 Conclusion and Future Work
items on the Raven’s Progressive Matrices. But it cannot for
The concerns reflected by TC have been more or less echoed
example prove that its answers are correct, which is part of
in many criticisms of analogical systems in the past 30 years
what the architecture-sketch in Figure 2 demands.
or so, but little to no discussion has attempted to precisely
Given the context we have now set, we can articulate a new
state the conditions under which an analogical system would
biconditional:
no longer be subject to this common complaint. We humbly
TCA4 A computational system A for analogy genera-
5
tion answers the TC if and only if, given as input no more We thank our reviewers for their insightful comments on this
than either topic and others.
acknowledge the limitations of our formulation here, and re- ical variations, and system approaches. AI Communica-
mind the reader that even a system with a strong claim to tions, 7(1):39–59, 1994.
meeting the conditions of TCA4 (e.g., an implementation of
[Arkoudas and Bringsjord, 2009] Konstantine Arkoudas
Figure 2) may not be completely immune to TC-based criti-
cisms. But it seems that such criticisms would be somewhat and Selmer Bringsjord. Vivid: An AI Frame-
difficult to justify. We hope that this paper inspires further work for Heterogeneous Problem Solving. Artificial
discussion, and invite researchers to pick up from and sharpen Intelligence, 173(15):1367–1405, 2009. The url
TCA4 , or to steer away from Psychometric AI and therefore http://kryten.mm.rpi.edu/vivid/vivid.pdf provides a
develop TCA3 in a different direction. preprint of the penultimate draft only. If for some reason
Those who strive for something deeper than a test-based it is not working, please contact either author directly by
anchor for a response to TC will presumably be more san- email.
guine than Bringsjord about the ultimate reach of comput- [Auer et al., 2007] Sören Auer, Christian Bizer, Georgi Ko-
ing machines. Bringsjord holds that it is impossible for any bilarov, Jens Lehmann, Richard Cyganiak, and Zachary
mere computing machine to originate anything [Bringsjord Ives. DBpedia: A Nucleus for a Web of Open Data. In
et al., 2001], or even to genuinely understanding anything Proceedings of the 6th International Semantic Web Con-
[Bringsjord and Noel, 2002]. But he holds on the other hand ference (ISWC2007), 2007.
that for any determinate behavioral test, a computing machine
can be programmed to pass it—even if the test is one which, [Bartha, 2010] Paul F.A. Bartha. By Parallel Reasoning: The
if passed, suggests to most human observers that the passing Construction and Evaluation of Analogical Arguments.
machine is not simply the product of human intelligence, har- Oxford University Press, 2010.
nessed [Bringsjord, 1995]. In short, the turn to TCA4 as [Boden, 1995] Margaret A. Boden. Creativity and unpre-
a way of fleshing out TCA3 reflects deep pessimism about dictability. Stanford Humanities Review, 4(2):123–139,
computing machines, and specifically a deep affirmation of 1995.
the penetrating and profound nature of TC. Licato, in con-
trast, is more optimistic about the potential of analogical- [Boden, 2009] Margaret A. Boden. Computer models of cre-
hybrid systems, but is not prepared to take any philosophical ativity. AI Magazine, 30(3), 2009.
positions on whether a system that passes TC demonstrates [Bollacker et al., 2008] Kurt Bollacker, Colin Evans,
true understanding or creativity. Praveen Paritosh, Tim Sturge, and Jamie Taylor. Freebase:
What about AI work devoted to building creative systems A Collaboratively Created Graph Database for Structuring
on the strength of analogical reasoning? Our work contin- Human Knowledge. In Proceedings of the 2008 ACM
ues along this path. We are currently investigating the po- SIGMOD International conference on Management of
tential of using our analogical matching system META-R, data (SIGMOD ’08). ACM, 2008.
which is designed to allow for a flexible application of heuris-
tics across a wide variety of domains, with the community- [Bringsjord and Licato, 2012] Selmer Bringsjord and John
built dataset Freebase [Bollacker et al., 2008]. We are Licato. Psychometric Artificial General Intelligence: The
also investigating the use of META-R in modeling the au- Piaget-MacGyver Room. In Pei Wang and Ben Goertzel,
tomatic generation of logical theorems and high-level proofs editors, Theoretical Foundations of Artificial General In-
from semi-formal and formal domains [Licato et al., 2013; telligence. Atlantis Press, 2012.
Govindarajulu et al., 2013]. Future work in this direction in- [Bringsjord and Noel, 2002] S. Bringsjord and R. Noel. Real
cludes the investigation of Analogico-Deductive Moral Rea-
robots and the missing thought experiment in the chinese
soning (ADMR) using a hybrid of CLARION [Sun, 2002]
room dialectic. In J. Preston and M. Bishop, editors, Views
and LISA [Hummel and Holyoak, 2003]. And finally, we are
into the Chinese Room: New Essays on Searle and Artifi-
seeking to leverage analogical reasoning in order to engineer
cial Intelligence, pages 144–166. Oxford University Press,
systems capable of automatic programming.6
Oxford, UK, 2002.
References [Bringsjord and Schimanski, 2003] S. Bringsjord and
[Aamodt and Plaza, 1994] Agnar Aamodt and Enric Plaza. B. Schimanski. What is Artificial Intelligence? Psy-
Case-based reasoning: Foundational issues, methodolog- chometric AI as an Answer. In Proceedings of the 18th
International Joint Conference on Artificial Intelligence
6
In the case of automatic programming, the input shown in Fig- (IJCAI–03), pages 887–893, San Francisco, CA, 2003.
ure 2 would be instantiated as the informal definition of a number- Morgan Kaufmann.
theoretic function (where that definition can be partly linguistic and
partly visual), and the answer is code in some conventional program- [Bringsjord et al., 2001] S. Bringsjord, D. Ferrucci, and
ming language, accompanied by a proof of the correctness of this P. Bello. Creativity, the Turing test, and the (better)
code relative to the input. Automatic-programming systems seem- Lovelace test. Minds and Machines, 11:3–27, 2001.
ingly require the ability to judge two programs analogous. More
precisely, such systems seemingly would need to be able to answer [Bringsjord, 1995] S. Bringsjord. Could, how could we tell
this question: Given a set of programs P in some programming lan- if, and why should–androids have inner lives? In K. Ford,
guage, can the system produce a similarity metric ρ : P × P → R C. Glymour, and P. Hayes, editors, Android Epistemology,
capturing which pairs of programs are semantically analogous? pages 93–122. MIT Press, Cambridge, MA, 1995.
[Bringsjord, 2011] Selmer Bringsjord. Psychometric Artifi- Projection (HDTP). Theoretical Computer Science,
cial Intelligence. Journal of Experimental and Theoretical 354(1):98–117, 2006.
Artificial Intelligence, 23(3):271–277, 2011. [Hofstadter, 2001] Douglas R. Hofstadter. Epilogue: Anal-
[Burstein, 1989] Mark H. Burstein. Analogy vs cbr: The ogy as the core of cognition. In Dedre Gentner, Keith J.
purpose of mapping. In Proceedings from the Case-based Holyoak, and Boicho N. Kokinov, editors, The Analogical
Reasoning Workshop, pages 133–136, Pensacola Beach, Mind: Perspectives from Cognitive Science, chapter 15.
Florida, 1989. The MIT Press, 2001.
[Chi et al., 1981] Michelene T.H. Chi, Paul J. Feltovich, and [Holyoak and Thagard, 1989] Keith J. Holyoak and Paul
Robert Glaser. Categorization and Representation of Thagard. Analogical Mapping by Constraint Satisfaction.
Physics Problems by Experts and Novices. Cognitive Sci- Cognitive Science, 13(3), 1989.
ence, 5, 1981. [Holyoak and Thagard, 1995] Keith J. Holyoak and Paul
[Doumas et al., 2008] Leonidas A. Doumas, John E. Hum- Thagard. Mental Leaps: Analogy in Creative Thought.
mel, and C.M. Sandhofer. A Theory of the Discovery and MIT Press, Cambridge, MA, 1995.
Predication of Relational Concepts. Psychological Review,
[Hummel and Holyoak, 2003] John E. Hummel and Keith J.
115:1–43, 2008.
Holyoak. A Symbolic-Connectionist Theory of Rela-
[Falkenhainer et al., 1989] Brian Falkenhainer, Kenneth D. tional Inference and Generalization. Psychological Re-
Forbus, and Dedre Gentner. The Structure-Mapping En- view, 110:220–264, 2003.
gine: Algorithm and Examples. Artificial Intelligence,
[Krumnack et al., 2007] Ulf Krumnack, Angela Schwering,
41:1–63, 1989.
Helmar Gust, and Kai-uwe Kühnberger. Restricted
[Finke et al., 1992] R.A. Finke, T.B. Ward, and S.M. Smith. Higher-Order Anti-Unification for Analogy Making. In
Creative Cognition: Theory, Research, and Applications. Proceedings of Twenties Australian Joint Conference
MIT Press, Cambridge, MA, 1992. on Artificial Intelligence, LNAI 4830, pages 273–282.
[Finke, 1990] R.A. Finke. Creative Imagery: Discoveries Springer, 2007.
and Inventions in Visualization. Lawrence Erlbaum As- [Krumnack et al., 2008] Ulf Krumnack, Helmar Gust,
sociates, 1990. Kai-uwe Kühnberger, and Angela Schwering. Re-
[Forbus and Gentner, 1989] Kenneth D. Forbus and Dedre Representation in a Logic-Based Model for Analogy-
Gentner. Structural Evaluation of Analogies: What Making. In Proceedings of the 21st Australasian Joint
Counts. In Proceedings of the Eleventh Annual Confer- Conference on Artificial Intelligence, Aukland, New
ence of the Cognitive Science Society, volume 34, 1989. Zealand, 2008.
[Forbus et al., 2009] Kenneth Forbus, Matthew Klenk, and [Licato et al., 2013] John Licato, Naveen S. Govindarajulu,
Thomas Hinrichs. Compaion Cognitive Systems: Design Selmer Bringsjord, Michael Pomeranz, and Logan Gittel-
Goals and Lessons Learned So Far. IEEE Intelligent Sys- son. Analogico-Deductive Generation of Gödel’s First In-
tems, 24(4):36–46, 2009. completeness Theorem from the Liar Paradox. Proceed-
[Gentner and Forbus, 2011] Dedre Gentner and Kenneth ings of the 23rd Annual International Joint Conference on
Forbus. Computational Models of Analogy. Wiley In- Artificial Intelligence (IJCAI-13), 2013.
terdisciplinary Reviews: Cognitive Science, 2(3):266–276, [Lovett et al., 2010] Andrew Lovett, Kenneth Forbus, and
2011. Jeffrey Usher. A Structure Mapping Model of Raven’s
[Gentner et al., 1997] Dedre Gentner, S. Brem, R.W. Fergu- Progressive Matrices. In Proceedings of CogSci, vol-
son, P. Wolff, A.B. Markman, and Ken Forbus. Anal- ume 10, 2010.
ogy and Creativity in the Works of Johannes Kepler. In [Mantaras et al., 2005] Ramon Lopez de Mantaras, David
T.B. Ward, S.M. Smith, and J. Vaid, editors, Creative Mcsherry, Derek Bridge, David Leake, Barry Smyth, Su-
Thought: An Investigation of Conceptual Structures and san Craw, Boi Faltings, Mary Lou Maher, Michael T. Cox,
Processes, pages 403–459. American Psychological Asso- Kenneth Forbus, Mark Keane, Agnar Aamodt, and Ian
ciation, 1997. Watson. Retrieval, reuse, revision and retention in case-
[Gentner, 1983] D Gentner. Structure-Mapping: A Theoret- based reasoning. The Knowledge Engineering Review,
ical Framework for Analogy. Cognitive Science, 7:155– 20(3):215–240, 2005.
170, 1983. [Matuszek et al., 2006] Cynthia Matuszek, John Cabral,
[Govindarajulu et al., 2013] Naveen S. Govindarajulu, John Michael Witbrock, and John DeOliveira. An Introduc-
Licato, and Selmer Bringsjord. Small Steps Toward Hy- tion to the Syntax and Content of Cyc. In Proceedings
percomputation via Infinitary Machine Proof Verification of the 2006 AAAI Spring Symposium on Formalizing and
and Proof Generation. In Unconventional Computation Compiling Background Knowledge and its Applications to
and Natural Computation - 12th International Conference Knowledge Representation and Question Answering, vol-
(UCNC 2013), 2013. ume 3864, 2006.
[Gust et al., 2006] Helmar Gust, Kai-uwe Kühnberger, and [Mitchell and Hofstadter, 1995] Melanie Mitchell and Dou-
U. Schmid. Metaphors and Heuristic-Driven Theory glas R. Hofstadter. Perspectives on Copycat: Comparisons
with Recent Work. In Douglas R. Hofstadter and the Fluid
Analogies Research Group, editors, Fluid Concepts and
Creative Analogies: Computer Models of the Fundamen-
tal Mechanisms of Thought, chapter 6. Basic Books, Inc,
1995.
[Norton, Forthcoming] John D. Norton. Formal and Mate-
rial Approaches to Analogical Inference. Forthcoming,
Forthcoming.
[Owen, 1990] S. Owen. Analogy for Automated Reasoning.
Academic Press, 1990.
[Riesbeck and Schank, 1989] Christopher K. Riesbeck and
Roger C. Schank. Inside Case-based Reasoning.
Lawrence Erlbaum Associates, 1989.
[Sun, 2002] Ron Sun. Duality of the Mind: A Bottom Up Ap-
proach Toward Cognition. Lawrence Erlbaum Associates,
Mah- wah, NJ, 2002.
[Wilson et al., 2001] William H. Wilson, Graeme S. Halford,
Brett Gray, and Steven Phillips. The STAR-2 Model for
Mapping Hierarchically Structured Analogs. In Dedre
Gentner, Keith J. Holyoak, and K. Kokinov Boicho, ed-
itors, The Analogical Mind: Perspectives from Cognitive
Science, chapter 4. The MIT Press, Cambridge, MA, 2001.
[Yan et al., 2003] Jin Yan, Ken Forbus, and Dedre Gentner.
A Theory of Rerepresentation in Analogical Matching. In
Proceedings of the 25th Annual Conference of the Cogni-
tive Science Society, 2003.
Creative Concept Invention and Representation in an Interactive Child-Robot
Healthy Snack Design Game

Andrew B. Williams, Adam Stroud, Joshua Panka, John C. Williams


Humanoid Engineering and Intelligent Robotics Laboratory
Marquette University, Milwaukee, WI USA
{andrew.williams, adam.stroud, joshua.panka, john.c.williams}@marquette.edu

Abstract
1.1 Creativity Framework for Concept Invention
In this paper, we describe how newly created
metaphor concepts will be invented and Concept invention is the process of discovering new
represented by a humanoid robot. The humanoid concepts that may require new interpretations and mappings
robot will invent these new concepts during of in a given situation [Reffat, 2002]. Creativity is closely
interaction with a child in a pretense play scenario. related to concept invention and is the quality that is marked
We demonstrate the feasibility of this concept by the ability or power to produce through imaginative skill.
invention and representation approach by Creativity has been defined as the ability to generate novel,
evaluating a knowledge-based metaphor discovery and valuable, ideas (e.g. concepts, theories, interpretations,
program using an open source reasoning engine. or stories) [Boden, 2009]. We use a componential
framework for creativity that is based on three factors for
creativity: domain-relevant skills, creativity relevant skills,
1 Introduction and task motivation [Amabile, 1983]. Amabile [1983]
The ability for a human to use her creativity to invent new presents a schematic for the creative process based on these
concepts or artifacts is a hallmark for human intelligence. skills and motivation that includes:
In order for a humanoid robot to attain human-like 1. Problem or task representation
intelligence, it must first be able to represent and 2. Preparation
incorporate creative new concepts into its knowledge base 3. Response generation
and be able to make new inferences using this new 4. Response validation
knowledge. We propose using PyClips, a python language 5. Outcome
interface to the CLIPS knowledge-based programming We incorporate this framework with metaphor-guided
language [Giarratano and Riley, 1998], as a tool to represent pretense play approach for creative concept invention and
and reason about these newly discovered creative semantic discovery.
concepts by a robot.
In [Williams, 2013] we described a newly proposed 1.2 Concept Invention through Metaphor
approach towards investigating how social humanoid robots Learning and Pretense Play
can learn creative conceptualizations through interaction In this section we describe how metaphor is related to
with children in metaphor-guided pretense play using a creativity, or concept invention. According to Indurkhya
componential creativity framework. We describe this [1992], a “metaphor is an unconventional way of describing
metaphor-guided pretense play approach and then illustrate (or representing) an object, event, or situation (real or
it by describing a social robot interaction pretense play imagined) as another object, event or situation”. [Genter et
scenario between a child and a humanoid robot. The rest of al., 2001] describe the critical steps for using metaphors for
this paper is organized as follows. First we review the problem-solving. One must extract a variety of unfamiliar
creativity framework that we combine with our metaphor- concepts from remote domains where possible relationships
guided pretense play creativity approach [Williams, 2013]. with the current task may not be initially apparent. Then
We then describe how our humanoid robot will learn new there must be a mapping of high-level or deep relationships
creative concepts using metaphors learned in a human-robot between the metaphor concept and the problem. Performing
interaction scenario. Then we describe how CLIPS will be generalization and abstraction techniques may make this
used to represent and reason about new creatively learned matching possible. Next, secondary relationships may be
concepts. discarded leaving only structural correspondence between
the metaphor source and the problem. Finally, the structural
matches, or correspondences, associated with the metaphor
source are transferred and applied to the problem, which
leads to a novel solution. Using metaphors is an example of how the child manipulates the objects and composes them.
a creativity-relevant skill of generating hypotheses that can It will then be up to the child to use her available food
lead to set-breaking and novel ideas. These metaphors can objects to create the concept conveyed by the picture. We
be used as heuristics to organize problem solving or design anticipate and will test in future experiments, that the child
thinking to solve loosely defined design problems [Rowe, will create a new metaphor for the subject using the food
1987; Antoniades 1992]. items. The resulting metaphors may consist of a stick of
celery (, or “log”) with soy butter (, or “dirt”) to glue the
raisins on (, or “bugs”). The child may name the new food
2 Concept Invention in an Interactive item, “bugs on a log”. The robot can respond with
“Yummy! What do you call it?” and then begin the verbal
Healthy Snack Food Design Game exchange process to incorporate the novel concept along
We propose using the context of a healthy snack food with the operators to produce the concept for future re-use.
design game to demonstrate how humanoid robots can
invent new concepts through interaction with children. 3 Creative Concept Invention and
[Morris et al., 2012] have developed a system for generating
novel recipes and use their system, PIERRE, to explore Representation
theoretical concepts in computational creativity. They show We target the problem of metaphor learning in the context
that a single inspiring set can be separated into two sets used of a pretense play setting. In this setting, the robot presents
for generation and evaluation to result in greater novelty. a picture not involving food items to a child and asks the
Other artificially intelligent food recipe generators include child to create a dish based on the picture. The underlying
CHEF [Hammond, 1986] which used case-based reasoning idea is that the child will see the objects in the picture as
to plan a recipe and Julia [Hinrichs, 1992] which also used metaphors for the food items it has been given, and create a
case-based reasoning but to plan out a meal. dish based on the metaphor. The robot then uses a
Children can create meaning from their social interactions description of the dish given by the child to recreate those
collectively through pretense play [Janes, 2002]. They can metaphors itself.
creatively take on imaginary roles, act out stories, and A humanoid robot’s ontology consists of the concepts
change toys into tools or instructions. In metaphor-guided known by the robot, or agent, and the interrelationships
pretense play [Williams, 2013], the child is given the between those concepts. The concepts can include
pretense play role of head chef with a set of healthy food attributes, actions, metaphors, and analogies. In a system in
items such as grapes, bananas, celery, soy butter, and raisins which more than one agent exists, or a multi-agent system
(Figure 1). The humanoid robot “customer” presents a (MAS), each agent, Ai, has a set of concepts, !i that it
“seed” image to the child “chef”. However, the seed image knows about.
is not a picture of food but rather a picture of another
unrelated subject, such as a muddy tree log lying on the Eq. 1) A = { A1, A2 ,…Ai,}
ground with a line of ants walking over it. Eq. 2) For each Ai there exists ! = K(Ai , !i )

A concept can be related to another concept, can define a


class of related sub-concepts, describe a function on a set of
objects, or describe a particular instance of a concept
[Gruber, 1993]. We have chosen to use PyCLIPS, a python
interface to CLIPS, to represent concepts known by the
humanoid robot and to be able to reason about the concepts
in the healthy snack food design game. CLIPS is a rule-
based knowledge-based programming language used to
represent and reason about knowledge [Giarratano and
Riley, 1998]. It contains the three main components of an
knowledge-based reasoning system: a fact list, knowledge
base, and inference engine.
In the previous section we described the scenario in
which the humanoid robot seeds the interaction with a
Figure 1 Metaphor-Guided Pretense Play to Learn picture of ants on a log. It will represent the concepts it
Creative Conceptualizations through Human-Child Food knows as facts in its knowledge base. For a given relation
Game Interaction. Humanoid robot seeds the interaction name, our reasoning has to be told which are valid symbolic
by projecting a seed metaphor image to the child. fields, or slots [Giarratano and Riley, 1998]. Groups of
facts that have the same relation name and contain common
information can be depicted using the deftemplate construct.
The particular ed image we are using in our example can
The humanoid robot will say to the child, “Make me
be represented by the CLIPS facts:
some food that looks like this!” and then begin observing
(Eq. 3) Our approach will make use of the Nao humanoid robot’s
(deffacts picture-concepts speech recognition abilities to conduct dialogue with the
(robot-concept (name nothing)) child to discover the new facts related to the food item using
(robot-concept (name ants)) approach (b) [Gouaillier, 2008].
(robot-concept (name log)) In order for the humanoid robot to discover these new
(robot-concept (name dirt)) metaphor concepts (e.g. raisin is a metaphor for ant), the
(robot-concept (name clouds)) CLIPS inference engine will be used to “fire”, or match, the
(robot-concept (name sky))) appropriate production rules. For example, the rules to find
the a new metaphor will have the following forms:
The location of the robots concepts in relation to each other
were represented using the following relations: (Eq. 7)

(Eq. 4) (defrule find-metaphor "Find a metaphor"


(deffacts picture-position-info ?f1 <- (robot-concept (name ?X))
(picture-position-relation (top nothing) (bottom ?f2 <- (robot-concept (name ?Y))
ant)) ?f3 <- (human-concept (name ?A))
(picture-position-relation (top ant) (bottom dirt)) ?f4 <- (human-concept (name ?B))
(picture-position-relation (top dirt) (bottom log))) ?f5 <- (picture-position-relation (top ?X) (bottom ?Y))
?f6 <- (dish-position-relation (top ?A) (bottom ?B))
=>
The child will use the seed image and along with its set of (retract ?f1)
healthy snack food elements it will attempt to create the (retract ?f3)
image using only the list of available ingredients as (assert (metaphor (robot-concept ?X) (human-concept
described in the previous section. Although the child uses ?A)))
her own mental model and does not use CLIPS, the (assert (metaphor (robot-concept ?Y) (human-concept
representation of the healthy snack food item that she ?B))))
creates is asserted as follows:
The production rule in Eq. 7 was used with our reasoning
(Eq. 5) engine to discover a new metaphor relation through the
(assert (human-concept (name nothing))) interaction between the robot and the child. This example
(assert (human-concept (name raisins))) rule should infer that the ant is stacked on top of the log as
(assert (human-concept (name celery))) well as the dirt and that raisins are stacked on top of the
(assert (human-concept (name soy_butter))) celery as well as the soy_butter. This asserts these new
facts. This rule will pattern match the contents of two
These assertions come through a verbal interaction concepts provided by the humanoid robot and the
between the robot and the child in which the robot queries corresponding child concepts and discover the metaphors
the child about the dish she has created. The locations of that the child creates so that they can be incorporated in the
the objects used by the child in relation to each other were robot’s knowledge base. This production rule
represented by the following assertions into the robot’s metaphorically should the concepts in the ant log picture
knowledge base: provided by the humanoid with the concepts in the ant log
food object. This rule was combined with other facts and
(Eq. 6) relations to demonstrate the feasibility to represent and infer
(assert (dish-position-relation (top nothing) (bottom new metaphor concepts through the child-robot interaction
raisins))) in the health snack food design game. The results are
(assert (dish-position-relation (top raisins) (bottom discussed in the next section.
soy_butter)))
(assert (dish-position-relation (top soy_butter) (bottom 4 Evaluation
celery)))
We executed our knowledge-based program using the
These facts would be incorporated to the robot’s CLIPS inference engine on a MacBook Pro with a 2.6 GHz
knowledge base in one of three ways: Intel Core i7 processor and 16 GB of RAM. CLIPS is a
(a) by observing the child combine and stack the rule-based reasoning engine that utlizes forward chaining to
ingredients to make the food item infer new knowledge.
(b) by having the robot interactively inquire what the The initial facts when the program constructs were loaded
new food item has and how they are stacked, or and the CLIPS system was reset included the following:
placed, on each other f-3 (robot-concept (name log))
(c) using a combination of (a) and (b). f-4 (robot-concept (name dirt))
f-7 (picture-position-relation (top nothing) (bottom ant))
f-8 (picture-position-relation (top ant) (bottom dirt)) needed. We believe that since our current program is based
f-9 (picture-position-relation (top dirt) (bottom log)) only on location relations, our system’s ability to find new
f-10 (human-concept (name nothing)) concepts, or metaphors, can be improved by incorporating
f-11 (human-concept (name raisins)) other key attributes such as shape or maybe texture.
f-12 (human-concept (name celery))
f-13 (human-concept (name soy_butter))
f-14 (dish-position-relation (top nothing) (bottom raisins)) 6 Conclusions and Future Work
f-15 (dish-position-relation (top raisins) (bottom
soy_butter)) We have described a method for representing concepts in a
f-16 (dish-position-relation (top soy_butter) (bottom metaphor-guided pretense play interaction and to infer new
celery)) metaphor concepts using a reasoning engine and have
demonstrated the feasibility of our approach. This paper
When executing the program, the find-metaphor rule was describes a rule-based method to learning new concepts
fired with facts: f-4, f-3, f-13, f-12, f-9, and f-16. The from observation. In this scenario, the robot initiates the
resulting knowledge base contains the following concepts learning by asking a child to ‘prepare’ a healthy snack based
and relations: on a picture. The child has to make an interpretation of the
image and come up with a novel concept. By formally
f-5 (robot-concept (name clouds)) annotating the original object with the novel concept and
f-6 (robot-concept (name sky)) annotating the ‘prepared’ snack in the same way, the robot
f-7 (picture-position-relation (top nothing) (bottom ant)) is able to figure out the metaphor and infer a new concept.
f-8 (picture-position-relation (top ant) (bottom dirt)) This paper describes a simplified and limited scenario
f-10 (human-concept (name nothing)) which allows this knowledge-based system to
f-11 (human-concept (name raisins)) computationally deal with the task. We have updated an
f-14 (dish-position-relation (top nothing) (bottom raisins)) earlier description of our general approach and test setting
f-15 (dish-position-relation (top raisins) (bottom by adding implementation details for our concept
soy_butter)) representation and learning along with some preliminary
f-17 (metaphor (robot-concept dirt) (human-concept empirical results. Further research is required to determine
soy_butter)) how far this approach will scale up with regard to the
f-18 (metaphor (robot-concept log) (human-concept knowledge acquisition of new metaphor concepts. We plan
celery)) to incorporate this approach into our prototype humanoid
health coach as one of the activities to teach children about
The creative concept invention knowledge based program how to make and eat health food snacks. Our approach
discovered two new metaphors. The child’s concept of holds promise for humanoid robots that can interactively
“dirt” was discovered to be a metaphor of “soy_butter”. learn new metaphor concepts by interacting with children in
Also, the child’s concept of “celery” was discovered by the food design scenarios.
robot to be a metaphor of “log”. These results are discussed
in the following section. Acknowledgments
This paper extends an earlier version of this concept from
5 Discussion of Results the AAAI 2013 Spring Symposium with further
Although there were two metaphors discovered and implementation details and preliminary empirical results for
incorporated into the robot’s knowledge base, there was a knowledge representation and reasoning to arrive at its
key metaphor missing. We expected the robot to discover current form. The authors wish to acknowledge the helpful
that “raisins” are a metaphor for “ants”. However, our comments from the reviewers that were incorporated into
knowledge base after the execution of our program failed to this paper.
discover this key metaphor relation. In other runs of the
program, incorrect metaphors were found but the resulting References
key find-metaphor rule was able to accurately uncover two [Amabile, 1983] The Social Psychology of Creativity: A
out of the three possible correct metaphors. This error Componential Conceptualization, Journal of Personality
occurred due to the fact that the find-metaphor rule retracted and Social Psychology, 45, 2, 357-376, 1983.
fact assertions necessary for the inference engine to discover
the additional metaphor for ants. Additional rules for [Antoniades, 1992] A. Antoniades. Poetics of architecture:
finding metaphors will have to rely on more than location Theory of design. New York: Van Nostrand Reinhold,
attributes and also be able to determine when nested 1992.
relationships occur such as the one in our example. [Boden, 2009] M.A. Boden. Computer Models of
Although our results demonstrate successfully that it is Creativity, AI Magazine, 30, 3, 23-24, 2009.
feasible to use our approach to invent new concepts through [Gentner, 2001] D. Gentner, B. Bowdle, P. Wolff, and C.
our metaphor-guided child-robot interaction in a health Boronat. Metaphor is like analogy. In D. Gentner, K.J.
snack food design game, we see that improvement is
and Kokinov, B.N. (Eds.), The analogical mind:
perspectives from cognitive science, 199-253.
Cambridge, MA: MIT Press, 2001.
[Giarratano and Riley, 1998] Joseph C. Giarratano and Gary
Riley. Expert Systems (3rd ed.). PWS Pub. Co., Boston,
MA, USA, 1998.
[Gouaillier et al., 2008] David Gouaillier, V. Hugel, P.
Blazevic, C. Kilner, J. Monceaux, P. Lafourcade, B.
Marnier, J. Serre and B. Maisonnier. The nao humanoid:
a combination of performance and affordability. CoRR,
vol. abs/0807.3223, 2008.
[Gruber, 1993] Thomas R. Gruber. A translation approach
to portable ontology specifications. Knowledge
acquisition 5, no. 2: 199-220, 1993.
[Hammond, 1986] Kristian Hammond. Chef: A model of
case-based planning. In Proceedings of the Fifth
National Conference on Artificial Intelligence (AAAI-
86), vol. 1, 267-271.
[Hinrichs, 1992] Tom Hinrichs. Problem solving in open
worlds: A case study in design. Lawrence Erlbaum
Associates. 1992.
[Indurkhya, 1992] Bipin Indurkhya, 1992. Metaphor and
Cognition: An Interactionist Approach, Dordrecht,
Boston: Kluwer Academic.
[Janes, 2002] P. Janes. Theory of the mind and pretense
play: play behaviors in 4-year olds. M.S. Thesis,
University of Georgia.
[Morris et al., 2012] Richard G. Morris, Scott H. Burton,
Paul M. Bodily, and Dan Ventura. Soup Over Bean of
Pure Joy: Culinary Ruminations of an Artificial Chef. In
International Conference on Computational Creativity,
p. 119. 2012.
[PyCLIPS 2013] PyCLIPS Python Module
http://pyclips.sourceforge.net/web/, accessed April 15,
2013.
[Reffat, 2002] Rabee M. Reffat. Intelligent agents for
concept invention of design forms. Proceedings of the
International Workshop on Agents in Design. 2002.
[Rowe, 1987] P. Rowe. Design Thinking. Cambridge, MA:
MIT Press, 1987.
[Williams, 2013] Andrew B. Williams, Towards Creative
Humanoid Conceptualization Learning from Metaphor-
Guided Pretense Play, Creativity and (Early) Cognitive
Development: A Perspective from Artificial Creativity,
Developmental AI, and Robotics, AAAI Spring
Symposium, Stanford University, AAAI Press, 79-83,
2013.
Back-And-Forth Inception: Towards a Cognitively-Inspired Model of Concept
Construction via Conceptual Intellection
Ahmed M. H. Abdel-Fattah and Stefan Schneider
Institute of Cognitive Science, University of Osnabrück,
Albrechtstr. 28, Osnabrück, Germany,
{ahabdelfatta,stefschn}@uni-osnabrueck.de

Abstract The interest in studying and developing computational


models of human-like creativity renders it inevitable to take
Based on experimental evidence, we introduce our these observations into consideration when proposing similar
ideas of ‘inception’ and ‘intellection’, then we use theories. One of the main motives is to contribute to the latter
the ideas to propose a preliminary model of concept need. We focus on how experimental results may influence
construction via ‘conceptual intellection’. First, the the design of computational models, based on stimulating the
experimental evidence is introduced and analyzed. way human participants construct their own understandings
Based on this analysis, we conceptually discuss of conceptions. We specifically suggest a ‘life-cycle’ of such
how the construction of concepts can function in construction, which has the potential to be simulated in com-
cognitively-inspired models, in which knowledge putational models, using computationally-plausible cognitive
can be represented as concept entities. processes. Later, we propose a preliminary model that is
based on this suggestion, but an overview summarizing the
1 Overview and Motivation central ideas is given in the following.
It has always been a big challenge for the interested re- Although humans are widely considered the best known
searchers, in a spectrum that ranges from closely related do- exemplar of cognitive beings, they are born knowledgeless,
mains to orthogonal and interdisciplinary ones, to give sen- then they acquire their knowledge from various sources over
tentious, exhaustive definitions of ‘creativity’. During only time. In principle, their knowledge is built of disparate pieces
the past few years, a multitude of relatively new (interdisci- of information (e.g. beliefs or facts), which we call concept
plinary) theories and ideas on creativity are introduced, which entities. These disparate concept entities can have implicit,
add to the numerous psychological studies that have already rather interrelated, tying links, that enable humans to utilize
been given during the past few decades. We refer the inter- their cognitive abilities in arranging the entities into concep-
ested reader to the list of references given in [McCormack tions; a network of conceptions that forms the background
and d’Inverno, 2012; Boden, 1996, p. 267]. knowledge. Depending on different situations and contexts
Recent theories on creativity excessively ground theoreti- (i.e. experience), the existing tying links between concept en-
cal ideas on creativity through practical applications and prac- tities can either be weakened, or strengthened. New links can
tice, and describe many ways to understand creativity through also be built. As people contemplate these links, the concep-
computers (e.g. in the interdisciplinary volume [McCormack tions themselves can be updated, and others may be created.
and d’Inverno, 2012]). Although this theoretical grounding is Conceptions can be learned and updated either by such con-
multi- and interdisciplinary in nature (since it bridges notions templation and deeper understanding, or by acquiring more
from computer science, psychology, neuroscience, visual art, knowledge. Both ways affect the already-existing knowledge
music, and philosophy), such theories and studies still do not conceptions, which results in the network of knowledge con-
fully capture deep insights that actually help in building com- ceptions being re-organized. This goes on and on, which
putational models of creativity (with a cognitively-inspired means that conceptions (as well as their composing entities)
flavor). For one thing, the recent theories illustrate how com- are being ‘intellectually’ treated and dealt with all the time:
puters are fundamentally changing what one can imagine and by updating already-existing conceptions, by creating new
create. Unfortunately, the theories neither widely nor deeply ones, or by re-arranging the hierarchy that possibly catego-
focus on how human participants can affect the proposed rizes them and their interrelated entities. We will call this
computational models. This disregards that participants are mechanism, of continuous intellectual treatment of the con-
very likely to have different knowledge, which affect what ceptual entities, the “intellection” mechanism.
the participants imagine or create. For another, defining cre- We claim that the above view has the potential to be par-
ativity and aspects thereof are extremely difficult, if at all pos- alleled in artificial models. We will propose a model that
sible, to express in a formal system, but can be described or mainly operates by employing a process that corresponds to
explained in terms of the observed outcomes and related phe- the intellection mechanism mentioned above. In the model,
nomena. we call this main process “conceptual intellection”, since it
stimulates, arranges, constructs, and updates concept entities hierarchies scaffold the construction and change the partici-
and concepts. In other words, ‘conceptual intellection’ is the pants’ conceptualizations. (N.B. Readers who are interested
name we give to the main process (in the proposed model), in other aspects of the study, such as the number of partici-
which utilizes other processes in producing and updating con- pants and the typicality of the presented data, need to refer to
cept entities and concepts. The article presents our views on [Schneider, 2012] for more information about the study and
how cognitive mechanisms are involved in ‘intellection’, and the investigated group of participants.)
how the proposed model potentially simulates them. Given the sequence of letter combinations A, B, C, D, BA,
Note that some terms may have different uses according BB (see figure 1-(a)), all the participants continued the se-
to different contexts. For example, we usually prefer to use quence by appending BC, BD. The protocols in [Schneider,
terms like ‘mechanisms’, and ‘conceptions’, whereas, from 2012] show that the task was apprehended as to create a
the modeling point of view, ‘processes’, and ‘concepts‘ (or ‘linear continuation’ of ‘two-letter combinations’. The way
concept entities) are preferred. In general, the mapping be- in which combinations were accomplished is by repeating
tween the notions and the contexts is not one-to-one, though the ‘ABCD-series’ at the right position while after each turn
the meaning should be understood (e.g. a cognitive mech- ‘prepending’ another symbol at the left. Some participants
anism and its corresponding computationally-plausible cog- were irritated that the first prepended symbol was a B, not
nitive process may both be termed ‘cognitive mechanisms’). an A, and perceived this as an inconsistency, so their con-
‘Conception’ is treated in most of the cases as the ideation in tinuation showed four major types of coping with what they
the human mind, while ‘concept’ is used when we are more conceived as a challenge.
interested in the representation issue. As already mentioned, In figure 1-(b), three types are shown: (i) the continuation
we use ‘intellection’ for the human cognitive mechanism, and to the left regularly repeats the missing of a symbol; (ii) the
‘conceptual intellection’ for the corresponding process in the branch in the middle defers to cope with the problem, which
model. resulted in later complications (cf. figure 1-(d)); and (iii)
The rest of the article is organized as follows. We start in the continuation to the right avoids the structural problem by
section 2 by giving evidence from cognitive science experi- appending an analogue version using the next letters of the
ments that help us in backing up the above views and claims, alphabet. In all three cases, the hierarchical understanding is
by arguing how humans use their cognitive capacities in stim- the basis of devising and realizing further continuations. The
ulating, creating, and updating their conceptions. In section 3, fourth continuation type is shown in figure 1-(c), where the
we state and explain our proposal of how the given evidence aspect of expanding positions was put in focus. The main
may help cognitively-inspired models to achieve concept con- idea was that, after each repetition of the ABCD series, an-
struction: (1) we suggest general assumptions about this kind other symbol is being prepended. This leads to an expansion
of models (cf. section 3.1), then (2) explain a ‘life-cycle’ of that one participant calls a ‘triangle’, another a ‘pyramid’, and
concept construction in details (cf. section 3.2). In section 3.3 yet a third associates the related metaphor of ‘atom splitting’.
we give quick hints on some related literature. In section 4, Drawings and gestures (which are being depicted by arrows
some concluding remarks are given. in figure 1-(c)) also show this figurative system comprehen-
sion. The pyramid metaphor only gives a coarse plan of what
2 Constructing Conceptions in Humans to do when positions get expanded, but as such, it sets the
frame for ‘filling in’ detail and ‘extracting’ other aspects.
Two case studies are investigated in [Schneider et al., 2013a; We now need to summarize how construction guides the
2013b; Schneider, 2012] to show how participants construe participants’ coping with the task, since this will turn out to
conceptions, and possibly solve problems related to inductive be highly related to the proposal given in the next section
production. (cf. section 3).
Participants naturally construct an understanding consist-
2.1 Stimulating Structure Construction: Study 1 ing of structures of differing scopes, when they apprehend
(Algorithmic Structures) the given symbol sequence and the underlying generative
A study in [Schneider et al., 2013a; Schneider, 2012] investi- principle. In general terms, one may speak of the ability to
gates how participants make up the regularities of a positional construct and develop a conception by the continuous grasp-
number system by way of example. A task in the study ‘dis- ing and organization of both (i) the entities that constitute a
guises’ the surface features of the familiar number notation given stimulation about this conception, and (ii) the underly-
by using a quaternary system with the four symbols A, B, C, ing connections between such entities. Moreover, more com-
and D. This task is presented to human participants. The par- prehensive structures guide their instantiation. As a result,
ticipants were only given the initial symbol sequence A, B, the whole construction determines its further application and
C, D, BA, BB, written on a tablet screen, then were asked to development. Structures tend to subordinate and re-arrange
continue it (cf. figure 1). No further restrictions were put onto other structures, and the degree of comprehensiveness does
the participants, so that they can continue the sequence while not restrict which structure is being modified. If an instan-
creating their own conceptualization. tiation (inception) of one’s conception is not readily at hand,
Two interrelated aspects of the findings are highlighted in modification of the assembled components may either change
the following. First, participants assemble occurrent rep- the detail structure or accommodate the comprehensive struc-
resentations of the system as a ‘hierarchized’ network of ture. For example, the case in which a ‘pyramid’ of letters
metaphors of differing scopes, and, secondly, the temporary is expanded indicates that, when one focuses on a stimulating
(a) (b) (c) (d)

Figure 1: (a) A sequence of letters, A, B, . . . , was written on a tablet screen by the interviewer until BB. All participants
continued it until BD. (b) The screenshot shows three possible branches of the sequence as proposed and written by one partic-
ipant. (c) The metaphor of a pyramid guides the vertical and diagonal application of the ABCD series. (d) This continuation
temporarily defers coping with the A-problem and later rectifies this by looping the initially linear progression.

aspect (the inception of position expansion) one may bring up type of assembly. When such a construction fails (and noth-
a conception (e.g. the pyramid metaphor) and thereby modify ing is being recognized), one attempts a new construction.
an earlier one (e.g. the linear succession). The new concep-
tion then guides in constructing further detail (e.g. fill center
digits). Similarly, figure 1-(d) indicates that the deferral of 2.3 Aspects of Construction via Intellection
the A-problem leads to completing the linear sequence with
AA, AB, AC, AD, followed in turn by BA, BB, BC, BD. This The first study investigated the inductive learning of a con-
modifies the linear continuation conception to become circu- cept. In study 2, the participants’ reports indicate that mental
lar, or loop. images are situations that need thoughtful intellection. That
is, intellections that have to be construed, and continuously
2.2 Stimulating Structural Construction: Study 2 reconstructed, to discover relations. In both studies, partici-
(Mental Imagery Construction) pants construed and operated with an intellection of the given
stimuli representing different aspects of it. Thus, an ‘intel-
Imagery is widely considered to be essential to creative pro- lection’ here means a structurally hierarchized apprehension
cesses [Finke et al., 1989]. The evidence given in [Schnei- of a situation, which gives an immediate directive of what
der et al., 2013b] show that imagery does not operate with could be done. An intellection is being carried out until the
static images, but requires to frequently reconstruct them. ‘step’ where it fails to ‘proceed forward’ (in the model, this
Apparently, one can mentally assemble and recognize a fig- corresponds to a concept being produced). One then has to re-
ure; a process which is, according to Hofstadter, based on construct and test the new ideation, where reconstruction usu-
“higher-level perception” and thought of as “the core of cog- ally starts with a very similar approach like before. The pro-
nition” [Hofstadter, 2001]. The assembly suggested in fig- cess can be viewed as an attempt to ‘fulfill’ the specification
given by the initial stimuli. The construction does not search
and find, but stimulates and instantiates a certain specifica-
tion (inception) and thereby creates something new. Mutual
modification of structures of different coarseness is another,
essential property of intellection construction. Construction
is guided by comprehensive structures, but a comprehensive
structure can also be modified as a result of working on struc-
Figure 2: The verbal description (inception) easily affects the ture details.
synthesis and apprehension of this figure, but analyzing it in In the model, proposed below, we view ‘conceptual intel-
order to find all triangles is much harder. lection’ as an open-ended process that is intended to com-
putationally simulate the ‘intellection’ mechanism discussed
ure 2 is given by verbal description in [Schneider et al., above: the latter is our view of the way in which humans
2013b], where the task is to count all triangles in it. Par- construct and update their conceptions (based on the analysis
ticipants easily found some of the triangles, but struggled to of the given experiments and the motivational view in sec-
find all. They readily ‘remembered’ the whole figure; but ap- tion 1), and the former is our view of the way the proposed
parently had to ‘reassemble’ varying parts of it in order to model simulates the latter. The construction and update is
consider ‘relations’ between lines that would ‘form’ a trian- mainly performed by means of the existing knowledge con-
gle. In the presented task, assembling parts is guided in that cepts and new stimulating entities. These stimulating entities
a prototype – a triangle in this case – ‘consummates’ relevant cause the process of inception to take place, again, and start
material. In the case where a prototype dominates thought a conceptual intellection process again. This back-and-forth
development we want to speak of construction as a special processing is thoroughly discussed in the following section.
3 Cognition-Based Concept Construction This aims at imitating aspects of knowledge emergence in
From a cognitive science point of view, we claim that creative humans that are based on their ability of acquiring knowl-
production of conceptions, in the human mind, is a (poten- edge from the environment, and using their experience in re-
tially unbounded) mechanism of cognition that utilizes sim- fining their understanding of repeatedly-encountered concep-
pler sub-mechanisms in the construction and organization of tions (cf. [Abdel-Fattah and Krumnack, 2013; Abdel-Fattah,
conceptual entities. So the simulation of this mechanism (in 2012]). In any case, we assume that the storage and orga-
a computational model of creativity, for instance) would be a nization of knowledge in structured concepts is unavoidable
process that needs to reflect the functionalities of other inter- for showing human-like thinking or creative behavior by this
related (cognitive) sub-processes; each of which is model type.
1. intermediate: an ultimate (creative) production has to go 3.2 The Life-Cycle of Conceptual Intellections
through a non-empty set of the sub-processes, and
A proposed life-cycle for concept construction in artificial
2. interacting: one sub-process, that may be involved in systems is explained below and illustrated in figure 3.
a construction process, affects the functioning and out- According to our proposal, a conceptual intellection com-
come of another. putationally simulates a (creative) construction of concep-
In what follows, the previously discussed aspects are trans- tions; a (fulfillment) mechanism that is always incremen-
formed into a proposal of a model that can eventually accom- tal and potentially unbounded. The system should always
plish (creative) production of concepts. Based on the way build and modify its (created) results, be they concept en-
conceptions are constructed in humans (the experimental re- tities, concepts, or arrangements thereof. The act of grasp-
sults and our claims in section 2), the model views (creative) ing conceptions caused by the stimulating entities (in a lively
concept construction as a life-cycle that utilizes cognitive pro- active human being’s mind) may never stop. So, a point
cesses in carrying out conceptual intellections. where (creative) construction stops should be generally un-
desirable in a computationally running system. This does
3.1 General Model Assumptions not have to conflict with notions concerning boundedness
We do not want to restrict our presentation to one spe- (e.g. bounded rationality [Gingerenzer and Selten, 2002;
cific model or another, because the conceptual ideas here Simon, 1984], knowledge and resources [Wang, 2011], etc.)
are intended to apply to a range of models that satisfy because these notions are concerned with how the cognitive
some assumptions. We will not be designing a computa- (natural or artificial) agent thinks and behaves as the limited
tional model of creativity, rather proposing our view about resources only allow (e.g. the processing and storage capaci-
how cognitively-inspired, computational models may employ ties).
cognitive processes to achieve (creative) production of con- Based on the given studies, we propose that the life-cycle
cepts. Building this kind of models needs to reduce com- of a conceptual intellection (that is, of constructing and up-
puting (creative) construction of concepts, in particular, to dating the concepts) consists of four facets: (1) inception, (2)
computationally-plausible cognitive mechanisms. For exam- arrangement, (3) production, and (4) adaptation. Each facet
ple, an agent (of a model that satisfies the assumptions) can constitutes specific cognitive processes. The depiction in fig-
update an already-existing concept, say CT , by making an ure 3 metaphorically uses a staircase object2 with four sides;
analogy to a well-known concept, say CS . The update can each corresponding to a facet of the potentially unbounded,
happen because an analogy-making may involve an analogi- construction process. Each facet’s stair-steps metaphorically
cal transfer that enriches the ‘structure’ of CT (humans would correspond to interrelated cognitive processes (not all of them
also update their understanding of CT in a similar analogy- can be shown), which may be involved in this facet’s func-
making situation). Another example is the utilization of con- tioning.
cept blending in creating new concepts and recognizing un- The utilized metaphor in the given figure tries to pictorially
precedented concept combinations.1 General assumptions deliver our own message that:
about how we conceive the model are given in the following. The more a system uses knowledge during the pro-
Let us first assume that one’s ultimate goal is to design a gressive inception (from brainstorming to identifi-
cognitively-inspired model of computational creativity and cation of context-dependent pieces of knowledge),
general intelligence. We also assume that the considered the more the system organizes and arranges the col-
model provides the ability to represent knowledge in the lected knowledge (during the progressive arrange-
form of conceptual entities. The organization of knowledge ment) before the ultimate (creative) productions are
in some form of concept domains follows the language-of- constructed, then finally adapted to the overall cog-
thought hypothesis [Fodor, 1983], where concepts can be nitive state (before a new life-cycle starts).
provided in modular groups to serve simulating several pro-
cesses. A model of this kind helps the acquisition, organiza- Using stair-steps implicitly reflects that the ‘proceeding’
tion, and development of knowledge using concept entities. through the staircase may go both ways at some times, caus-
ing some processes to be retracted and others re-executed.
1
Although these examples and figure 3 exemplify some sug-
2
gested mechanisms, such as analogy-making and concept blending, In fact, the staircase object, depicted in figure 3, ought to be
we do not intend here to give a list of the cognitive mechanisms. seen as a spiral of repeating stair-steps going up indefinitely. This is
Related discussions can be found in [Abdel-Fattah and Krumnack, also illustrated by the spiral and the up-arrow, drawn to the left of
2013; Abdel-Fattah et al., 2012; Abdel-Fattah, 2012]. the staircase object in the same figure.
Figure 3: An illustration of the suggested life-cycle for (creative) construction of concepts. The visible number of stair-steps in
each of the four turns is irrelevant (some steps may even be skipped when ‘proceeding’ through the staircase). The ‘proceeding’
is, in turn, depicted by a spiral to the left, where the up-arrow indicates that the life-cycle generally moves forward (up the stairs).

At the beginning of a conceptual intellection life-cycle, transfer). It is worth noting that, in the current proposal,
stimulating entities are instantiated. The stimulating entities ‘what counts as a creative construction and what counts
may have already existed (then activated somehow), or re- as not’ is not specified and left to another discussion.
ceived as new input. They can have structural relations tying This is mainly because the knowledge concepts can al-
them together (or to other knowledge entities). The inception ways be updated, though not considered new or ‘really
facet initiates the construction of concepts, which will lead creative’ concepts.
to the creation of a new concept, or the modification and the
4. The fourth facet, ‘adaptation’ parallels the deeper, in-
hierarchical arrangement of already-existing concepts. We
sightful understanding of conceptions (by humans), and
elaborate more on the cycle’s facets in the following:
reflects the changes in an input or a current context.
1. The ‘inception’ facet involves activating the set of pro- There is always the possibility that what might have once
cesses that leads to implanting and updating conceptual been considered just a recently produced concept by the
entities (which form the agent’s knowledge concepts; model (e.g. as a result of conceptual blending), becomes
cf. section 3.1 and [Abdel-Fattah and Krumnack, 2013; a well-entrenched concept later (perhaps after repeated3
Abdel-Fattah, 2012]). The inception can eventually re- usage).
sult in the model having its set of concepts updated. This
is because either (i) the existing conceptual entities are Digression – The Never-Ending Staircase
modified, or (ii) new entities are acquired. In the former
case, the links tying some of the conceptual entities are The visualization in figure 3 makes use of an attention-
affected and, consequently, the concepts involving them exciting construction instead of, for instance, drawing a dull
are also affected. The concepts can also be affected in circle in two dimensions (which might have as well achieved
the latter case, and new concepts may also be created. the same representational goal). It may be interesting for
some readers to see classical ideas represented in unusual
2. The ‘arrangement’ facet follows the ‘inception’ facet, ways, but there is more to this issue than only drawing an
and is responsible for organizing (currently existing) unusual, seemingly-complex shape.
conceptual entities. (Refer to the discussions in sec- A shape of an impossible triangle, famously known as the
tion 2.2 about hierarchical assembly and the attempts to Penrose triangle, was originally created for fun by Roger Pen-
re-construct when the assembly fails.) rose and Oscar Reutersvärd. Later, Roger Penrose and his
3. The ‘production’ facet starts with the ‘construction’ pro- father, Lionel Penrose, created the impossible staircase, and
cess, which constructs a concept (after assembling and presented an article about the two impossible shapes in the
arranging its components) and ends with such concept British Journal of Philosophy in 1958 (see [Penrose and Pen-
having its components enriched (the fulfillment of the rose, 1958]). M. C. Escher artistically implemented the two
initial model specifications; cf. section 2.3). As a pro- shapes in creating two of his most popular (and very creative)
cess, ‘construction’ is not the crowning result of the drawings; namely: “Waterfall” (lithograph, 1961 [Hofstadter,
‘production’ facet, rather the starting step (in a given cy-
cle). Newly constructed concepts may be re-constructed 3
The idea of entrenchment has been elaborated on in [Abdel-
and enriched in several ways (e.g. through analogical Fattah and Krumnack, 2013; Abdel-Fattah, 2012].
1999, fig. 5] and [Seckel, 2004, pp. 92]) and “Ascending and in the latter case, and not necessarily in the former. In fact,
Descending” (lithograph, 1960 [Hofstadter, 1999, fig. 6] and it is not even desirable that the participants reflect a creative
[Seckel, 2004, pp. 91]). A version of the latter drawing has ability (according to the aims of the tests in the former case).
theatrically been directed by Christopher Nolan in his mas- Another difference is that our proposed model has the po-
terpiece, “Inception” [Nolan and Nolan, 2010], which use the tential to utilize computationally-plausible cognitive mecha-
structure to represent impossible constructions. In fact, the nisms in achieving artificial systems that realize the model.
credit for using the term “inception”, as the initial process We are neither concerned with only testing deductive reason-
within the conceptual intellection life-cycle, is also due to ing, nor do we need results that only obey rules of deductive
Christopher Nolan, whose work in [Nolan and Nolan, 2010] reasoning. Rather, our main concern is in testing (creative)
inspired us to use the impossible staircase shape in figure 3 to concept construction, which seems to involve cognitive ca-
reflect a similar meaning. pacities other than deduction. For instance, concept blending
This article is concerned with creative constructions and possibly allows for unfamiliar, resulting concepts by blend-
the inception process. According to our proposal, the former ing two (or more) familiar concepts, opening thus the door
may indefinitely be performed in successive steps, whereas to non-deductively achieve thinking outside the box. Also,
the latter is the most essential facet of the given life-cycle. given two concepts, people do not usually see one, and only
Hence this particular visualization is used. one, analogical mapping between the two concepts’ entities.
Counterfactual reasoning is yet a third example that empha-
3.3 Related Literature sizes the importance of utilizing mechanisms other than de-
In addition to the earlier discussions, the ideas of the pro- ductive reasoning in (creative) concept construction(actually,
posed model are further supported by work in the literature. deduction may not at all work when computationally analyz-
The suggested (four-facet) cycle agrees with the “Geneplore ing counterfactual conditionals [Abdel-Fattah et al., 2013]).
model” [Finke et al., 1992], which proposes that creative
ideas can only be produced in an “exploratory phase”, after an 4 Concluding Remarks
individual already constructs mental representations, called
“pre-inventive structures”, in the first “generative phase”. It is not surprising that many ideas, of what currently is re-
Such pre-inventing, in our view, initiates the inception facet garded standard in (good old-fashioned) artificial intelligence
and finishes before creative construction starts. Usage of (AI), have already been laid out in Newell and Simon’s article
the term “inception” in our proposal is thus akin to “pre- [Newell and Simon, 1976]. The article very early anticipated
inventing”. Additionally, evidence is given in [Smith et improvements that seem inevitable from today’s perspective.
al., 1995, p. 157–178] that the already-existing concepts According to Newell and Simon’s pragmatic definition, an
greatly affect the (re-)structuring of the newly developed intelligent organism is required to be ‘adaptive’ to real world
ideas. Moreover, Weisberg argues that (human) creativity ‘situations’. Newell and Simon inspired the scientific society
only involves “ordinary cognitive processes” that yield “ex- to begin studying human-like intelligence and their ability to
traordinary results” [Weisberg, 1993], which consolidates our adapt. However, since the beginning of AI, it was “search”
motivation to focus on cognitively-inspired models. what they proposed as a cornerstone of computing [Newell
On the one hand, there are similarities between our pro- and Simon, 1976]. Less importance has been given to other
posal and the mental models theory, as proposed by Johnson- computationally-plausible cognitive mechanisms (which may
Laird and others (cf. [Johnson-Laird, 2006; 1995; 1983; not need, or be based on, search).
Gentner and Stevens, 1983]). The reader may for example Creativity, on the other hand, is a complex general intel-
recognize that our view of what we call “intellections” seems ligence ability, that is closely connected with the utilization
analogous to Johnson-Laird’s view of what he calls “mod- of various cognitive mechanisms. Creativity may definitely
els”. “Models are the natural way in which the human mind need searching in performing specific sub-processes, such as
constructs reality, conceives alternatives to it, and searches finding analogies to newly available concepts before modify-
out the consequences of assumptions” [Johnson-Laird, 1995, ing (or adapting) them. But, after all, it certainly differs from
pp. 999]. In addition, the general results of the many psy- mere search in several aspects. Our trial in this article claims
chological experiments given in [Johnson-Laird and Byrne, that investigating, testing, and implementing cognitive mech-
1991], for example, may be considered yet a further support anisms need to attract more attention, in particular when one
to our proposal about how humans construct their understand- is interested in modeling computational creativity.
ings. On the other hand, our approach differs from the the- A model of computational creativity is expected to greatly
ory of Johnson-Laird in at least the following essential points. differ from a classical AI model. An essential difference
Whilst mental model experiments are mainly concerned with between creativity and AI is that AI programs have to ac-
testing one particular ability of participants (namely, their cess huge amounts of (remarkably meaningless or unsorted)
ability to reason ‘deductively’), the first study in section 2, data before getting (meaningful and interconnected) results,
for example, investigated the ‘inductive’ learning of a con- whereas creativity reflects the general intelligence character-
cept. In our view, asking the participants to construct argu- istic of humans, who are much more selective and adaptive
ments, to show that one conclusion or another follows from in this regard. For example, humans can not only induce ap-
what already lies within what is already given, is different propriate meanings of unknown words, both within and with-
from giving them inceptions, then allow them to freely con- out a given context (e.g. interpreting unprecedentedly-seen
tinue their conception constructions. Creativity is revealed noun combinations that appear in general phrases; cf. [Abdel-
Fattah and Krumnack, 2013]), but can also produce words Krumnack, 2013; Abdel-Fattah et al., 2012]), though poten-
that reflect particular meanings or descriptively summarize tially unbounded. Such systems treat the behavior ‘to create’
specific situations (e.g. giving the name “Facebook” to the as that of bringing into existence a concept that has not pre-
online social networking service, yet a “face book” is, phys- viously ‘experienced’, or that is significantly ‘dissimilar’ to
ically, a directory that helps people get to know each other). another analogue (or even its older version). But, in essence,
As a further example, consider the analysis of counterfactual creativity cannot be about creating something from nothing
conditionals. Humans can flexibly restrict the judgement of (computationally-speaking, at least), rather triggered by ‘in-
whether or not a given counterfactual conditional is verifiable, ceptions’ every now and then.
although the conditional might indicate many, or possibly-
infinite, possibilities (cf. [Abdel-Fattah et al., 2013]). In Acknowledgments
essence, the ‘selectivity’ and ‘adaptivity’ capacities of hu-
mans are richly adorned with insightful constructions, regard- The authors would like to thank the three anonymous review-
less of whether or not humans already have a vast amount of ers of this article’s first draft, both for their valuable opin-
knowledge. Knowledge and experience certainly help in in- ions and encouraging comments. The concerns they raised
creasing these capacities: the more their knowledge, the more remarkably helped in improving the presentation of the given
humans can appropriately select, adapt, and (creatively) con- ideas in some parts of the article.
struct. The notion of search is partially motivated by the sub-
jective observation that one initially explores different ideas References
while actively working on a problem.
[Abdel-Fattah and Krumnack, 2013] A. Abdel-Fattah and
A cognitively-inspired model of creativity and general in-
telligence should be designed to reflect, and benefit from, the U. Krumnack. Creating analogy-based interpretations
essential differences between how a cognitive being, and how of blended noun concepts. In AAAI 2013 Spring Sym-
a classical AI analogue, solve the same problems. By a classi- posium: Creativity and (Early) Cognitive Development.
cal AI analogue we mean a classical AI model that solves the AAAI Press, Palo Alto, California, 2013.
problems ingeniously, regardless of whether or not the way [Abdel-Fattah et al., 2012] A. Abdel-Fattah, T. Besold, and
of solving agrees with that of a generally intelligent, cogni- K.-U. Kühnberger. Creativity, Cognitive Mechanisms, and
tive being. To base the models on representably-modifiable Logic. In J. Bach, B. Goertzel, and M. Iklé, editors, Proc.
conceptual constructions would hopefully be a promising so- of the 5th Conference on Artificial General Intelligence,
lution to achieving this design goal. We claim that the design Oxford, volume 7716 of LNCS, pages 1–10. Springer,
of a model of creativity should be inspired by the behavior 2012.
of humans in situations where construction and adaptation of [Abdel-Fattah et al., 2013] Ahmed M. H. Abdel-Fattah, Ulf
conceptions are needed, and can be implemented in the model Krumnack, and Kai-Uwe Kühnberger. Utilizing Cognitive
by employing specific4 cognitive mechanisms. The model
Mechanisms in the Analysis of Counterfactual Condition-
should as well allow for several (bidirectional) paths, some-
als by AGI Systems. In Kühnberger et al. [2013], pages
times even ‘bridges’, to construct and modify conceptions in a 1–10.
potentially unbounded process of construction, without cross-
ing the exponential bounds that search may require. Such a [Abdel-Fattah, 2012] A. Abdel-Fattah. On the feasibility of
potentially unbounded construction process is proposed here concept blending in interpreting novel noun compounds.
as a ‘looping’ process. It is therefore worth quoting part of In Proceedings of the Workshop “Computational Creativ-
Hofstadter’s comments on what he calls “the central cogni- ity, Concept Invention, and General Intelligence”, vol-
tive loop” (described in [Hofstadter, 2001]): ume 01-2012, pages 27–32. Institute of Cognitive Science,
2012.
“The broad-stroke pathway meandering through
the limitless space of potential ideas [... goes [Boden, 1996] Margaret A. Boden. Artificial Intelligence.
around] and around in [...] a loop, alternating be- Handbook Of Perception And Cognition. Elsevier Science,
tween fishing in long-term memory and unpacking 1996.
and reperceiving in short-term memory, rolls the [Finke et al., 1989] Ronald A. Finke, Steven Pinker, and
process of cognition.” [Hofstadter, 2001, p. 521] Martha J. Farah. Reinterpreting Visual Patterns in Men-
From where the authors stand, creativity produces uncom- tal Imagery. Cognitive Science, 13:51–78, 1989.
mon outcomes that provide contemporary applications of pre- [Finke et al., 1992] R.A. Finke, T.B. Ward, and S.M. Smith.
viously known constructs to new contexts. We propose that Creative Cognition: Theory, Research and Applications.
(computational) creativity should be phrased in terms of the Bradford Books. Bradford Books, 1992.
ability of an (artificial) intelligent system to construct concep-
[Fodor, 1983] Jerry A. Fodor. The Modularity of Mind: An
tions or modify them, using the already-existing knowledge
about former concepts; a process that is both computation- Essay on Faculty Psychology. Bradford Books. MIT Press,
ally plausible and cognitively inspired (cf. [Abdel-Fattah and 1983.
[Gentner and Stevens, 1983] Dedre Gentner and Albert L.
4
Again, the precise list of cognitive mechanisms can not be ex- Stevens. Mental Models. Cognitive Science. L. Erlbaum
tensively discussed in this article. Associates, 1983.
[Gingerenzer and Selten, 2002] G. Gingerenzer and R. Sel- number system. Master’s thesis, University of Osnabrück,
ten. Bounded Rationality: The Adaptative Toolbox. 2012.
Dahlem workshop reports. Mit Press, 2002. [Seckel, 2004] A. Seckel. Masters of Deception: Escher,
[Hofstadter, 1999] Douglas R. Hofstadter. Gödel, Escher, Dalí & the Artists of Optical Illusion. Sterling Publishing
Bach: An Eternal Golden Braid. Basic Books. Basic Company Incorporated, 2004.
Books, 1999. [Simon, 1984] H. A. Simon. Models of Bounded Rationality.
[Hofstadter, 2001] Douglas R. Hofstadter. Analogy as the MIT., 1984.
Core of Cognition. In Keith J Holyoak, Dedre Gen- [Smith et al., 1995] S.M. Smith, T.B. Ward, and R.A. Finke.
tner, and Boicho Kokinov, editors, The analogical mind: The Creative Cognition Approach. A Bradford book. MIT
Perspectives from cognitive science, pages 499–538. MIT Press, 1995.
Press, Cambridge, MA, 2001.
[Wang, 2011] Pei Wang. The Assumption on Knowledge
[Johnson-Laird and Byrne, 1991] Philip N. Johnson-Laird and Resources in Models of Rationality. International
and Ruth M. J. Byrne. Deduction. Essays in Cognitive Journal of Machine Consciousness (IJMC), 3:193–218,
Psychology. Lawrence Erlbaum Associates, Incorporated, 2011.
1991.
[Weisberg, 1993] R.W. Weisberg. Creativity: beyond the
[Johnson-Laird, 1983] Philip N. Johnson-Laird. Mental myth of genius. Books in psychology. W H Freeman &
models: towards a cognitive science of language, infer- Company, 1993.
ence, and consciousness. Harvard University Press, Cam-
bridge, MA, USA, 1983.
[Johnson-Laird, 1995] Philip N. Johnson-Laird. Mental
Models, Deductive Reasoning, and the Brain, pages 999–
1008. The MIT Press, 3 edition, 1995.
[Johnson-Laird, 2006] Philip N. Johnson-Laird. How We
Reason. Oxford University Press, 2006.
[Kühnberger et al., 2013] K.-U. Kühnberger, S. Rodolph,
and P. Wang, editors. Artificial General Intelligence - 6th
International Conference, AGI 2013, Beijing, China, 31
Juli – 3 August, 2013. Proceedings, volume 7999 of Lec-
ture Notes in Computer Science. Springer-Verlag Berlin
Heidelberg, 2013.
[McCormack and d’Inverno, 2012] Jon McCormack and
Mark d’Inverno. Computers and Creativity. Springer-
Verlag GmbH, 2012.
[Newell and Simon, 1976] Allen Newell and Herbert A. Si-
mon. Computer science as empirical inquiry: symbols
and search. Communications of the ACM, 19(3):113–126,
March 1976.
[Nolan and Nolan, 2010] Christopher Nolan and Jonathan
Nolan. Inception: The Shooting Script. Insight Editions,
2010.
[Penrose and Penrose, 1958] L. S. Penrose and R. Penrose.
Impossible objects: A special type of visual illusion.
British J. of Psychology, 49:31–33, 1958.
[Schneider et al., 2013a] Stefan Schneider, Ahmed M. H.
Abdel-Fattah, Benjamin Angerer, and Felix Weber. Model
Construction in General Intelligence. In Kühnberger et al.
[2013], pages 109–118.
[Schneider et al., 2013b] Stefan Schneider, Ulf Krumnack,
and Kai-Uwe Kühnberger. Dynamic Assembly of Figures
in Visuospatial Reasoning. In Proceedings of the Work-
shop SHAPES 2.0: The Shape of Things. Co-located with
UNI-LOG 2013, Rio de Janereiro, Brazil, 2013.
[Schneider, 2012] Stefan Schneider. The cognitive structure
of the enumeration algorithm: Making up a positional
Square, zero, kitchen: Start
Insights on cross-linguistic conceptual encoding
Francesca Quattri
The Hong Kong Polytechnic University
Hong Kong
francesca.quattri@connect.polyu.hk

Abstract closeness in the Non-Dominant Hemisphere and visual cor-


tex ([Nussbaum]).1 [Cytowic, 2002][Cytowic, 1993] claims
In this study, I approach the study of metaphor on the contrary that synesthetic and metaphorical reasoning
and shape from a linguistic/cognitive perspective. are two different features of the human brain (with the former
A distinction is made between metaphors of shape occurring more frequently than the latter). Back in the 1980s,
and metaphors of shaping. Two major case stud- scientists (e. g. Francis Galton, Darwin’s cousine, Nature ar-
ies are presented (going back to square one [i. e. ticle) believed that synesthesia was experiential, in particular
to start something anew because of a failure] that it was the remembrance of childhood memories in adult-
and vicious/virtuous circle) with the shape in the hood. The approach was later discarded for stronger scientific
metaphor’s source domain. ‘Metaphor’ is intended evidence.
here as form of speech with collocational/idiomatic While synesthesia blurs together senses, analogy picks up
behavior; the source and target domain frame is the common conceptual denominator beneath them. Ana-
experimentally applied to an idiom and not just a logical thinking is the ability to match meanings and collect
metaphor. words in our brain according to the concept/s they share ([Ra-
The study is qualitative, in progress and limited machandran, 2008]). Lack of analogy is lack of conceptual-
to two case studies, but it offers insights on the ization, and, with reference to this study, lack to interpret and
problem of conceptual encoding as observed from perceive space, shapes and surrounding reality as constructs
a new perspective (metaphors with shape words in of different properties, from width to depth, to color, tex-
the source domain) and values the literal and not ture and size (a pathology called agnesia, [Farah, 2004][Ra-
just conceptual meaning of metaphors. machandran, 1990]).
The final goal of the project is to help AI experts Combining “sharp”, a tactile perception with “cheese” to
in their struggle to find robust models of represen- define a taste is for instance a synesthetic ability ([Cytowic,
tation of metaphorical thinking, conceptual encod- 1993]); memorizing the words in a list (‘cake’, ‘pie’, ‘choco-
ing, as well as in-depth text and context analysis. late’ and ‘sugar’), and eventually cite along with them ‘sweet-
ness’ is on the contrary a demonstration of analogical think-
ing.
1 Introduction Idiomatic metaphorical language finds its reason of be-
As [Barnden, 2008] points out, the interest of AI to metaphor ing in these bytes of human creativity. Virtue and morality
and similes dates back to several decades. Researchers in- are attributed to circles (vicious or virtuous circle), lines are
cluding Russell 1976, 1985, Carbonell 1980, 1982, Norvig dropped, centuries go around the corner. Patients missing the
1989 and Way 1991 have paved the way of investigation in corpus callosum (a bunch of nerves that keeps the two hemi-
the field. spheres bound) miss the ability to cope with these forms (e. g.
autistic patients) and take them literally.
The author of this paper aims at underscoring metaphors by
the study of source-domain reasoning and backs up with cita- The study of metaphors and the study of shapes as taken
tions from cognitive science, another aim of AI (in [Barnden, together finds no record in my literature review experience.
2008]’s words). There are nevertheless plenty of studies on metaphors (but not
Metaphorical thinking remains, despite outstanding re- with shape words in the source domain)(e. g. [Gibbs (ed.),
search, a big black hole for cognitive and neuro-scientists as 2008] [Forceville & Urios-Aparisi (eds.), 2009][Lakoff and
well as AI experts. [Ramachandran & Hubbard, 2003][Ra- 1 Patients affected by chronic synesthesia cannot avoid to i. a. de-
machandran & Hubbard, 2011] claim that the backbone of fine numbers with colors and colors with personalities [Ramachan-
all metaphors is to be found in synesthesia (from the Greek dran, 2011] without making it up (e. g. “mint tasted like a cylindrical
roots syn, meaning “together,” and aisthesis, or “perception”), marble column”; “I enjoy music that has very metallic lines”; [Cy-
i. e. the ability to combine senses together given their spatial towic, 1993]:32f.).
Johnson, 1980][Koevecses, 2010]) and plenty of studies on something by putting the own name on it) allows us to lit-
shapes as geometrical entities for automated detection (e. g. erally imagine a dotted line. These two examples are in my
[Fonseca & Jorge, 2000]), or as a topic for discussion on for- definition metaphors of shape.
mal description and ontology (e. g. [Borgo & Kutz, 2011]). Metaphors of shaping are, as readers might intuitively un-
I present hereby two case scenarios. In the first, I analyze derstand, figures of speech that give indication of the process
the metaphor (vicious/virtuous circle) and want to show that of a shape coming into being, or, more importantly, the devel-
the perception of the shape greatly depends upon the fact of opment of an event/object. The shape does not play a relevant
whether the metaphor is of shape or of shaping. In the second role in the metaphor. It might be justifiable by etymology,
case (going back to square one), I try to apply the source- but its origins (which explain why it was originally used for
target domain parallelism to the idiom. Aware of the fact that the form) are not always traceable.4 “Invest in lines, not in
this approach might be biased by some linguists, I define the dots” is a common suggestion among market dealers. Ob-
case “experimental” and foresee to lead attention instead to at viously, the shapes are not used here as forms (‘lines’ are ro-
least two fact that have implications in conceptual encoding bust companies with history to count on; ‘dots’ stand for fresh
(and thus in the development in AI of sound models for the start-ups with no experience on the market). Eventually, the
same). study of shapes in metaphors can be extended to “compound”
The first observation is the ability of the human brain to metaphors, like in the idiom kick the bucket. The metaphor is
process abstract information by means of some more abstract entailed in ‘bucket’ (which stands for coffin) and means to
(and not by means of a simple paraphrase). When asked to die, to pass out. Another example of compound metaphor
declare the meaning of an idiomatic conventional form, other is the expression “the circular triangle” or in the idiom “to
abstract forms are used.2 square peg in a round hole”.
The second is the collection of observations on the literal Despite having anticipated the concept of compound
meaning of the metaphors, regarded as relevant information metaphors, I focus here on shape words as used in the source
carrier (sense versus meaning, description follows). domain. It might be useful to notice that the conceptual pair-
ing between source and target domain is for some scholars
2 Metaphors of shape and metaphors of not always a sound way to define how metaphors work.
[Agerri et al., 2007; Agerri, 2008] argues for instance that
shaping it is not always possible to establish a link or parallelism
The perception of shape contributes to the understanding of between source and target domain to explain a metaphor,
metaphors. In particular, I claim that the visual perception of since the expression is a highly contextualized linguistic
shape is strongly related to the fact of whether a metaphor is phenomenon that derives much of its information from the
defined of shape or of shaping. surrounding discourse (“enrichment of mappings”, [Agerri,
Notice that by ‘metaphor’ I define a figure of speech 2008]). The computational process deriving from these map-
which might contain idiomatic/collocation meaning and pings can be endless and provides little help for a clear inves-
form. Without going into details, the claim here is that the tigation of the subject.
[Hobbs, 1992] talks about metaphors in terms of relativity
traditional linguistic division between idioms, collocations,
metaphors, metonymies seems to be ad hoc and gives little theory. The scholar experiments with a novel and a conven-
indication of the true entanglement of these forms together.3 tional metaphor. In the novel (“John is an elephant”), he tries
I define by metaphor of shape is a figure of speech where to come up with a model to unambiguously map source and
the entailed shape (in the source and/or in the target domain) target domain and reasons on the properties that both John
has to be understood literally (as precise form with strong and an elephant may share (e. g. clumsiness). He does so
visualization details), in order to fully grasp the meaning of by citing different approaches undertaken in the analysis of
the expression itself. metaphors as proposed by i. a. Orthony et al. (1978), Glucks-
berg and Keysar (1990), Carbonell (1982), but claims that
Take for instance the sentence: “at a certain point in her
they are all a-contextual. Rather than (just) picking up the
life, Mary decided to draw the line”. Listeners cannot but
size or other physical properties of an elephant and compare
imagine a rather clear line which marks a cut, a division be-
them to John, Hobbs suggests that an analyst should con-
tween what it was and what is near to come. “I have not
sider the metaphor within the clause and within the situation
signed on the dotted lines yet, but will do so soon”: Also in
of use “Mary is graceful, but John is an elephant” ([Hobbs,
this case, the expression (which means: formally agree on
1992]:15). In a way, Hobbs validates the Sapir-Whorf hy-
2 The statement is not supported by experiment but by simple ob- pothesis of relativism, which claims that that there exists lin-
servation derived from a lexicographic search of the forms in differ- guistic relativity in commonsense language due to cultural
ent repositories. The viewer is naturally redirected to synonyms of contextual influences on language itself (a statement also in-
the forms, which are also conventional idiomatic expressions. directly sustained by [Pinker, 2009a][Pinker, 2009b][Pinker,
3 Further planned studies aim to find out whether these linguistic 1994]). The same theory is defended by [Polley, 2010], who
forms are not just parented with each other, but whether they are has proven in his doctoral dissertation the existence of emo-
bound by a sort of hiearchy (from collocations derive metaphors tion metaphors in English and Mandarin with varying degrees
[“the sharp bite is sat across the table”], some metonymies are
metaphors [“he drank a glass of bitter memories]”, from metaphors 4 A good archive for etymology of idiomatic form is the website
derive idioms ( “she’s in a good shape”). The Phrase Finder, www.phrases.org.uk.
in the patterns of organization and use of these metaphors metaphor in contexts as possible, I have proceeded with a ran-
across languages and individuals. The study, focused on � dom selection of examples. (around 1,700 entries for vicious
� (xı́ngfù, happiness) in Mandarin and English, is conducted circle and virtuous circle respectively (last search: April 21st,
through corpus analysis and adopts elements of the MPA, 2013).
Metaphor Pattern Analysis by Stefanowitch (2006) and of the The KWIC size has been set to 500 characters, which gives
Metaphor Identification Procedure (MIPVU), as developed at around three sentences before and after the sentence in which
the Vrije Universitaet by Stern et al. in 2010. the collocation is entailed.
The study of the metaphor is run always with the intent to
Ahrens et al. [Ahrens et al., 2003] defends the oppo- acknowledge whether the shape circle plays a role in the final
site front and proves, starting from Lakoff’s Conceptual understanding of the expression (in other words, whether vi-
Metaphor [Lakoff and Johnson, 1980], that all source do- cious and virtuous circle are metaphors of shape or metaphors
main can be put in correspondence with a target domain au- of shaping).
tomatically and with little ambiguity. The authors integrate Given the collocation form of the metaphor, I expect that
the Conceptual Mapping Model with ontology representa- vicious and virtuous are adjectives referring to morality and
tion [SUMO, 2013] to defend their thesis. Several studies on virtue. I can suppose that since the circle is also a religious
‘economy’ (jı̄ngjı̀) in Mandarin Chinese [Academia Sinica, symbol, the choice of the adjective is theologically motivated.
2006] prove that an automatic matching is feasible [Ahrens et Crosslinguistically, if I compare the metaphor as it exists in
al., 2003; Chung and Ahrens, 2006]. other languages, I bump into the same conclusions: “Teufel-
To make the Model work, a Mapping Principle Constraint skreis” (Ge, lit. “circle of the devil”); “circolo vizioso” (It; lit.
is introduced, for which a target domain will select only “circle of the vice”; the same in the Portuguese and Spanish,
source domains that involve unique mapping principles. In “circulo vicioso”); “���� (èxı̀ngxúnhuán)” (Ch; lit. the
this way, ECONOMY in Mandarin Chinese is WAR, where evil circle).
WAR is subset of the source domain of COMPETITION. In
another example, IDEA, a target domain, builds upon the Corpus-analysis applied on the form seems to revealed that
source domain BUILDING, and not for the wiring, or plumb- the literal meaning of the shape has no inference on the final
ing system of a building, but for its being a stable block understanding of the metaphor. The expressions come often
with pillows and corner stones. “Buildings involve a (phys- in contexts where different events are listed not in a random,
ical) structure and ideas involve an (abstract) organization“ but mostly in a cause-effect relation, or associated to verbs of
[Ahrens et al., 2003]. rotation and motion. This suggests that the human brain does
The authors come to these conclusions via a three-step pro- not need the shape of a circle to understand the metaphor,
cess: (a) first define entities, properties, and qualities that ex- but it needs a close frame and a sense of motion to give it a
ist between source and target domain; (b) then go back to the meaning.
source domain and compare it to the real-world knowledge; The actions involved with the circle are often: one or some-
(c) eventually postulate a unique mapping, as the one pre- thing get stuck in a circle; the circle can break, arise or simply
sented above for building and idea. We decide to adopt the be. The sequel of events follows a dynamics. In many cases
same process in our study for its clarity. this could be called causality, but there needs more evidence
for this. As for the other question, whether the property of
3 First case study: Vicious/virtuous circle being vicious or virtuous affects the behavior of the metaphor
in context, in none of the cases I have found that there exists
Given that the research is currently in process, only the major an explicit reference between the shape with morality or im-
outcomes of the study on the metaphor are reported. The re- morality. Vicious seems to stand more for negative, virtuous
search has been conducted in different stages, with different for positive.
granularity towards the syntactic elements surrounding the It seems that the metaphors of shape vicious/virtuous circle
expression (verbs, adjectives, prepositions) and with regards acknowledge the existence of a form in the source domain for
to different domains of use. two major reasons: an ontological function (the form stands
The corpus-based analysis has been conducted on both for mappings, enclosure of facts, boundness, network), and
[COCA, 2013] and [SketchEngine, 2007]. For the latter, the an orientational function (the close frame suggests round di-
corpus enTenTen has been selected given the high number of rectionality or causality more than a casual trajectory).The
tokens provided.5 image of the circle is from absent to vague, definitely not sig-
For each expression, I have worked on the sentences auto- nificant for the metaphor.
matically filtered by [SketchEngine, 2007]. In previous anal- Eventually, as stated above, it seems that the literal under-
ysis in fact, it has been noticed that going through the corpora standing of the adjectival form (morality, religion) is not rel-
manually (having set a number of sentences to analyze) re- evant for its conceptual use in context.
sulted in having many examples to support a theory about
the behavior of the metaphor and less examples for other If someone would ask to be explained the meaning of vi-
cases, which could therefore be considered the exception. To cious circle, it is probable (at least by native language speak-
avoid this and having to deal with as many applications of the ers) that the metaphor is explained via other metaphor or ab-
stract forms.
5 The Ten Ten English web corpus, 3,268,798,627 tokens, Vicious circle has in fact the same meaning of “daily
2,759,340,513 words. grind”, “treadmill”, “hamster cage”.
‘Circle’, ‘treadmill’ and ‘cage ’have at least two aspects What needs to be found are the qualities and relations be-
in common: (1) they are closed forms, especially ‘circle’, tween the two domains. The analysis can be done at several
‘grind’ and ‘cage’; (2) they rotate (‘grind’ is the product and levels of granularity. The properties between source and tar-
the action of a treadmill). get are either derived from the domains as we have defined
them, or they are extracted from every single components that
4 Second case study: Going back to square one make up each domain.
Following the first option, little relation is found. If going
The second expression considered is going back to square back to square one means start something anew because of a
one. This form is an idiom. failure, there are apparently no other linkings between the two
Given the intertwined nature of linguistic forms (explained sequences of information rather than the motion verb going
above), it also entails the collocation square one and the col- back associated to the motion verb start/begin.
location back to or back at square one6 . The second level of analysis provides more information,
The attempt at this stage is to transfer the source-target do- but it is unbalanced. We can in fact finely decompose the
main logics of metaphors (which works fine for some schol- source domain, but not the target domain. Even in Levin’s
ars) to idioms, in the higher goal to automatize the process. verb classification [Levin, 1993] little information is given on
Ahrens [Ahrens et al., 2003] suggests a three-step ap- the target.8
proach for the identification of a metaphorical mapping. A The word square is polysemous. As a noun, it means inter
quick recap of the stages: (a) define entities, properties, and alia “any artifact having a shape similar to a plane geometric
qualities existing between source and target domain; (b) go figure with four equal sides and four right angles”, “a hand
back to the source domain and compare it to the real-world tool [...] used to construct or test right angles” or “something
knowledge; (c) postulate a unique mapping, as the one pre- approximating the shape of a square”, [WordNet Search 3.1.,
sented above for building and idea. 2010].
An answer to the question: Why decide to adopt this model The collocation square one acts as collocation in certain
on idioms? The analysis of vicious/virtuous circle has showed domains of use, while in others it becomes a proper noun
recurrent patterns that cannot be totally proven until the re- (a jargon term) with exact meaning (as in the case of football
search switches to statistical methods and more data. The and rugby games, board games and the playground game hop-
outcome of the proposed research and of further researches scotch)9 This implicitly means that, even if we have decided
on shapes already shows nevertheless that one single shape to take it as an idiom, the expression going back to square one
might happen in a metaphor of shape and in a metaphor of can be idiomatic or not according to the context. Evidently,
shaping, something being perceived as one or the other in one the corpus-based analysis to be conducted will not take into
same form. The dizziness is justified by the context, which account the form as a jargon.
eventually determines the meaning of a metaphor per situa- Suppositions on the form might lead to think that going
tion (without exception for conventional forms). back to square one involves a return to a precise shape, pre-
cisely a square. The numbering (“one”) suggests that there
As in the current shape, the research reiterates what may be other numbered forms apart from that. The action of
[Hobbs, 1992] and other authors already previously cited tell going back involves a previous departure, ergo it may refer
us about metaphors in context, namely the happening of the to a temporal frame. A spatial setting is also involved, with
Sapir-Whorf theory of relativity. a movement from space A to space B and back to A in the
Ahrens proves that an automatic source-domain detection simplest visualization.
for metaphors is feasible, under certain constrains. So my The next question is: How does this fine-grain information
question is: Does it apply to idioms as well? Since I have relate to the concept or meaning of ‘start’? The verb entails a
adopted a definition of metaphor which enables the extension temporal and spatial location, but it would be a speculation to
of the research on other linguistic forms than metaphors, I test argue that a U-turn is also involved. That justifies the state-
it on an idiom with shape. ment about an unbalanced conceptual mapping.
The approach has been adopted with further variations: A final question relates to the choice of ‘square’ for build-
The whole idiom stands for the source domain (going back ing this idiom. Why is the concept of ‘start/begin’ precisely
to square one); the target domain stands for the meaning/s as- related to this geometric form (as it is ‘square’ taken in one of
sociated to it. This enables to identify stage (a) of the method its many senses)? Does the concept need to relate to this form
proposed. in particular? The attempt to find reasonable answers justifies
The third passage of the experiment can be anticipated as another variation in Ahrens [Ahrens et al., 2003]’ study. I
well. The mapping between source and target domain is de-
rived from encyclopedic knowledge: going back to square 8 The verb ‘start’ belongs to the category of the begin verbs,
one means ‘start something anew because of a failure’7 . which describes “the initiation, termination, or continuation of an
activity” ([Levin, 1993]:274-275). Since Levin’s work is not con-
6 Back to square one: 1 bln 10 ml results, Google; back at square cerned with sentential complement-taking properties, no specifica-
one: 1 bln 20 mln results, Google. tion is provided about a possible distinction between for instance
7 “If you are back to square one, you have to start working on a ‘start’ and ‘begin’. ‘Failure’ as a verb is not mentioned.
plan from the beginning because your previous attempt failed com- 9 From The Phrase Finder, http://www.phrases.org.
pletely”, [Cambridge, 2013] uk/meanings/back\%20to\%20square\%20one.html
extend the research to a cross-linguistic comparison. ◦ voltar ao principio (return back to the origin/the start)
◦ voltar a como era antes (return to how it was)
4.1 Cross-linguistic comparison ◦ atè au nı̀vel zero (to level zero)
A random sample of 250 sentences is taken [SketchEngine,
Italian
2007] from the enTenTen corpus (Web corpus, 3,268,798,627
◦ ricominciare da capo/daccapo (start/begin anew)
tokens, 2,759,340,513 words). The reason why random and
◦ cominciare da zero (start from zero)
not manually picked examples have been taken is the same as
◦ ricominciare da zero (start from zero)
for vicious/virtuous circle.
◦ partire dall’inizio (start from the beginning)
In this first stage of the analysis I am interested to find
◦ tornare al punto di partenza (start from the starting point)
out: (a) whether the idiom exists in other languages and
(b) whether it is translated by means of the same form. Chinese
In brackets the English literal translation. The expressions ◦ ���� (lı̀ngqǔlúzào) (to set up a separate kitchen, start
have been taken from [Linguee, 2013; HanDeDict, 2011; from scratch) ( ��) kitchen range, kitchen stove)
Idiomdatenbank, 2007].
Swedish
German ◦ tillbaka till grunden (going back to the ground)
◦ wieder da, wo wir angefangen haben (still there, where we ◦ böria om frän början (start from the beginning)
started/begun) ◦ tillbaka till utgångspunkten (go back to the point of exit)
◦ noch einmal von vorne beginnen (to start again from ◦ tillbaka vid utgångspunkten (go back to the point of exit)
scratch) ◦ börjar från noll (start from zero)
◦ wieder bei Null anfangen (to start again from zero)
◦ zurück zum Start (back to the start) The first observation from the data teaches that it is pos-
◦ auf Null zurück (back to zero) sible to paraphrase going back to square one via other ab-
◦ wieder zum Anfang zurückgehen (to go back to the begin- stract forms or idioms. To go back to square one has the same
ning/start) meaning of and can thus be translated into, inter alia, to start
◦ alles auf Null (all to zero) from scratch, to make a fresh start, to go back to point zero,
◦ zurückgehen (to go back) to set up a separate kitchen. Interestingly, some of these ex-
◦ zurück auf Platz eins sein (to be back to/at place one) pressions could only be found in the process of translating
the metaphors from other languages into English. Within the
French
limits of the tools and reference consulted, I have found no
◦ revenir à la case “départ” (come back to the box “start”)
repository that comprehensively reports all possible idioms
◦ ramener à la case “départ” (come back to the box “start”)
for the same concept in one language or more than one.
◦ en crèer de nouveau (to create something anew)
◦ retourner à la case départ (come back to the box “start”) In six of the seven languages presented, there are regu-
◦ être au point mort (to be at a deadlock) lar semantic patterns (at least regular in comparison to the
◦ remettre le compteur à zèro (set the counter back to zero) semantic elements of the original English form considered).
◦ reprendre à la case départ (to continue from the box “start”) These patterns are: The expressions contain a shape and a
◦ revenir au point du départ (come back to the back) number. Going back to square one is mostly translated into
◦ repartir à zèro (to start again from zero) start from zero or be at point zero.The choice of picking up
◦ faire table rase (start afresh/start anew/start over/turn a new Romanic, Neogermanic and Oriental languages at the same
leaf/start all over again) time makes the comparison less vulnerable to remarks on lin-
guistic inheritance.
Spanish Only in Chinese there is, in comparison to the other forms
◦ estar en el punto cero (to be at point zero) in the other languages, a completely different word, ‘kitchen
◦ partir de cero (start/begin from zero) stove’(which relates to the English idiomatic expression to set
◦ empezar desde cero (start/begin from zero) up a separate kitchen).
◦ una vuelta al principio (a turn to the beginning)
So, to answer the questions posed before the cross-
◦ hasta el nivel cero (by level zero)
linguistic analysis, it seems that the presence of a shape in
◦ recomenzar de nuevo (start again anew)
the source domain is not binding for defining the concept of
◦ devolver al punto de partida (return to the point of origin/to
start/begin in the target. Relating this finding to the above
the starting point)
presented definition of metaphor of shaping, the shaping or
Portuguese process of the form coming into being is thus located in the
◦ de volta au ponto de partida (back to the point of origin/to source.
the starting point) We may not find a form in all idioms in all languages, but
◦ estar no ponto zero (to be at point zero) we find in almost all of them a numbering, which enables us
◦ regressar ao ponto de partida (to go back to point zero) to state that “Starting something involves going back or start
◦ estar de volta à estaca zero (be at point zero) either from zero or from one.”
◦ voltar à estaca zero (turn back to point zero) The comparison shows, by linguistic comparison alone, a
◦ recomeçar do principio (start again from scratch) recurrent use of shapes and numbers for the same expression
◦ criaçao de nova (create anew) among different languages. But the translation with shapes is
not the only translation possible. This gives indication that shaping, than ‘kitchen’ can stand for more than a single room,
the concept of shape, let it be square or point, might not be but also for factory, business, facility, activity, productivity,
so indicative for the meaning of the idiom. If so, we are con- event. The same can be claimed for all the other metaphors
fronted with another metaphor of shaping and not of shape. in the list. ‘Head’ can then stand for ‘idea’ or ‘behavior’ or
‘trend’. To reduce the fuzziness of these implied conceptual-
4.2 Sense and Meaning ization, [Hobbs, 1992]’s model might become handy as well
In the language of current research on formal conceptual (the same model that declares why John is an elephant for his
mapping, lexical entries sense the same lexical sense. I am clumsiness and not for his weight).
referring here to the OntoLex terminology as adopted by John The conceptual and contextual and crosslinguistic analysis
McCrae in his diagrammatic rendering of semiotics.owl (pub- together trigger discussion not just in the domain of concep-
lic.ontolex w3.org). tual encoding, but also in creativity and in the way language
According to the scheme, a LexicalEntry denotes an Ontol- users indulge or procrastinate new ideas.
ogyEntity, expresses a Concept or Synset and senses a Lexi-
calSense. The LexicalSense refers to the OntologyEntity and 5 Conclusions and Future Work
the Concept or Synset conceptualizes it. From this Concept, Metaphors are at the backbone of everyday communication.
we can derive the semiotic Meaning. We utter almost six metaphors a minute, once said Geary
The scheme is still object of investigation by the W3C James in a TED talk11 , and yet it is unclear where they sit
members, so that a final word on the definiteness of the model in the brain.
might be inappropriate but group members seem to agree that Different theories have been suggested, the research is on-
a distinction between ‘sense’ and ‘meaning’ in ontological going.
and lexico-semantic formalizations represents added value. From the linguistic point of view, metaphors are interest-
(last update of the model checked on: April 19th, 2013). ing constructs that have been synthetized in a source and a
The interlingual analysis of the idiom going back to square target domain. For some scholars, this is a rather imprecise
one shows that: (a) the lexeme10 can be paraphrased in many distinction which leaves little attention to cultural and con-
other lexemes in the same language; (b) the lexeme exists in textual differences; for others, it is a feasible way to extract
different languages with different semantic forms including concepts automatically.
shapes and that (c) it always conveys the same meaning. I pick up two case studies, vicious/virtuous circle and going
Following the suggested OntoLex model, I want at this back to square one, and analyze their crosslinguistic contex-
point claim that the target domain of an idiom entails the id- tual behavior (with an emphasis of the latter for the form with
iom’s concept, but for the idiom to have a meaning, it needs ‘square’ and an emphasis on the monolingual corpus-based
a sense, contained in the source domain. analysis for ‘circle’).
The idiom to go back to square one denotes the concept or In this research, I work with metaphors and shape forms
OntologyEntity beneath ‘start/begin’. This is an encyclopedic in the source domain. In the best of my literature knowl-
truth, something we can learn from thesauri and dictionaries. edge I have so far encountered no cases of the two studied to-
Cross-linguistic analysis alone tells us though that a process gether. I make a distinction between metaphors of shape and
of starting or beginning something in conjunction with this id- metaphors of shaping, because I suppose that not all shapes
iomatic expression can only happen when it entails the sense are equally important in the understanding of a metaphor that
of going back to something, which is also visualized with a entails them, if they are relevant at all. In metaphors of shape,
space number zero or number one. the precise visualization of the form is necessary to under-
‘Sense’ is here used not with the connotation of ‘meaning’ stand the meaning of the expression. In metaphors of shap-
(as in the case of a polysemous word), but for indicating an ing, the geometric forms, its visualization and representation
adjunct which reinforces and fulfills the lexical meaning. The are irrelevant to the final meaning of the form. The shape
interpretation of the term does not dissociate from its standard is structural to the linguistic form, not to its meaning, but it
definition. gives indication of the spatial/temporal settings in which the
Lexical senses, like lexical entries, are language-specific. event occurs.
In the comparison, I have found different lexical entries for The qualitative analysis so far enables to tell that both
one language among different languages. Nevertheless, these forms are metaphors of shaping and not of shape, that in none
languages also share the same lexical sense for the same lex- of them the form is relevant. But the research beside this
ical entry. study has also showed that the are complex forms that might
The Chinese idiom for going back to square one, ���� be of shape or of shaping according to the context, so that a
(lı̀ngqǔlúzào), has been the only case with the most divergent further investigation in this direction is needed.
semantic formation (its English correspondent: to set up a Another remark is on the literal meaning of the metaphors,
separate kitchen). which enables to detect deeper “suppliers of meaning” or
senses, that otherwise would go lost in the commonsense use
If we argue once again (contextual use can further prove the of the form.
validity of this claim) that the metaphor is not of shape, but of The way metaphors behave, the concepts they trigger, the
place in the brain they claim, the process of creativity they
10 The idiom is defined here as lexeme or lexical entity; OntoLex
reference 11 blog.ted.com/2009/12/18/metaphorically/
fuel: All these interest topics are shared by AI researchers, [Farah, 2004] Martha J. Farah. Visual Agnosia, A Bradford
cognitive and neuro scientists and linguists. This preliminary Book, The MIT Press, Cambridge, Massachusetts, 2005.
research wants to sit at this crossroads and foster interdisci- [Fonseca & Jorge, 2000] Manuel J. Fonseca & Joaquim A.
plinary discussion. Jorge. Using fuzzy logic to recognize geometric shapes in-
teractively in Proceedings of the Ninth IEEE International
Conference on Fuzzy System (FUZZ IEEE), vol.1, pages
References 291-296, San Antonio, TX USA, May 7-10, 2000.
[Academia Sinica, 2006] Academia Sinica Balanced [Forceville & Urios-Aparisi (eds.), 2009] Charles Forceville
Corpus of Modern Chinese. 2006. http: & Eduardo Urios-Aparisi (eds.). Multimodal Metaphor,
//rocling.iis.sinica.edu.tw/CKIP/ Mouton de Gruyter, Berlin, New York, 2009.
engversion/20corpus.htm [Gibbs (ed.), 2008] Gibbs, Raymond W. Jr. The Cambridge
[Agerri, 2008] Rodrigo Agerri. Metaphor in Textual Entail- Handbook of Metaphor and Thought. 2008. Cambridge
ment in Proceedings of the 22nd International Conference University Press, London and New York.
on Computational Linguistics (Coling 2008), pages 2-6, [HanDeDict, 2011] Chinesisch-Deutsches Wörterbuch. Last
Manchester, UK, 2008. update: 2011, www.handedict.de
[Agerri et al., 2007] Rodrigo Agerri, John Barnden, Mark [Hobbs, 1992] Hobbs, Jerry R. Metaphor and Abduc-
Lee and Alan Wallington. Metaphor, inference and do- tion, in Andrew Ortony, Jon Slack & Olivier Stock
main independent mappings in Proceedings of the Re- (eds.), Communication from an Artificial Intelli-
search Advances in Natural Language Processing Confer- gence Perspective: Theoretical and Applied Issues,
ence, RANLP07, Borovets, Bulgaria, 27-29 September Springer Verlag, Berlin, pages 35-58, 1992. Also pub-
2007. lished as SRI Technical Note 508, SRI International,
[Ahrens et al., 2003] Kathleen Ahrens, Chung Siaw-Fong, Menlo Park, California, August 1991, http://www.
Chu-Ren Huang. Conceptual Metaphors: Ontology-Based isi.edu/˜hobbs/discourse-references/
Representation and Corpora Driven Mapping Principles in discourse-references.html.
Proceedings of the ACL Workshop on the Lexicon and Fig- [Idiomdatenbank, 2007] Berlin-Branderburgische Akademie
urative Language. pages 35-41, Sapporo, Japan, 2003. der Wissenschaften (BBAW) Kollokationen im
[Barnden, 2008] John A. Barnden. Metaphor and Artificial Wörterbuch und Idiomdatenbank. Last update 2007-
Intelligence: Why Do They Matter to Each Other. School 05-04, http://kollokationen.bbaw.de/htm/
of Computer Science, University of Birmingham, UK. idb_de.html
[Barnden & Lee (eds), 2001] John A. Barnden & Mark G. [Koevecses, 2010] Zoltán Kövecses. Metaphor. A Practical
Lee (eds.). Metaphor and Symbol. A Quarterly Jour- Introduction, Oxford University Press, New York, 2010.
nal, Special Issue of Metaphor and Artificial Intelligence, [Lakoff and Johnson, 1980] George Lakoff, Mark Johnson.
Lawrence Erlbaum Associates, Publishers, London, 2001. Metaphors we live by. 1980. The University of Chicago
[Bipin, 1992] Bipin, Indurkhya. Metaphor and Cognition: Press, Chicago and London.
an interactionist approach. Kluwer Academic. Dortrecht, [Levin, 1993] Beth Levin. English Verb Classes and Alterna-
Boston, 1992. tions. A Preliminary Investigation. 1993. The University
[Borgo & Kutz, 2011] Borgo Stefano & Kutz Oliver. A Gen- of Chicago Press, Chicago and London.
eral Framework for Shape Similarity in Proceedings [Linguee, 2013] Dictionary for German, French, Spanish
of SHAPES 1.0 – The Shape of Things. Workshop at and more. www.linguee.com
CONTEXT-11, Karlsruhe, September 20, 2011, CEUR- [Nussbaum] Paul David Nussbaum, Get to Know Your Brain,
WS, vol. 812, 2011. http://www.paulnussbaum.com/gettoknow.
[Chung and Ahrens, 2006] Chung, Siaw-Fong and Kathleen html
Ahrens. Source Domain Determination: WordNet-SUMO [Pinker, 2009a] Steven Pinker. Steven Pinker on Language,
and Collocation in Proceedings of Theme Session of Big Think talk, February 5th, 2009, bigthink.com/
Cognitive-Linguistic Approaches: What Can We Gain by videos/steven-pinker-on-language
Computational Treatment of Data? Second International [Pinker, 2009b] Steven Pinker. What our Language Habits
Conference of the German Cognitive Linguistics Associa- Reveal, TED talk, February 10th, 2009.
tion, 5-7 October, Munich, 2006. [Pinker, 1994] Steven Pinker. The Language Instinct. The
[Cambridge, 2013] Cambridge Dictionaries Online. 2013 New Science of Language and Mind, Penguin Books, New
c
�dictionary.cambridge.org York, 1994.
[COCA, 2013] Corpus of Contemporary American English [Polley, 2010] Polley, Carl A. Metaphors for Happiness in
(COCA), 2013, www.americancorpus.org English and Mandarin Chinese. Doctoral dissertation the-
[Cytowic, 2002] Richard E. Cytowic. Synesthesia. A Union sis. University of Hawai, 2012.
of Senses, The MIT Press, Cambridge, Massachusetts, [Ramachandran, 2011] Ramachandran, Vilaynur S..
2002. The ‘House’ of Neuroscience, 2011, http:
[Cytowic, 1993] Richard E. Cytowic. The Man Who Tasted //www.time.com/time/video/player/0,
Shapes. A Bizarre Medical Mystery Offers Revolutionary 32068,912832862001_2066607,00.html
Insights into Emotions, Reasoning, and Consciousness, [Ramachandran & Hubbard, 2011] Vilayanur R. Ramachan-
Tarcher and Putnam, New York, 1993. dran & Edward M. Hubbard. The Emergence of Qualia,
Metaphor, Language and Abstract Thought: Some Clues
from Synesthesia, in: Sejnowski TS, Van Hemmen L.
(eds.) 23 Problems in Systems Neuroscience, Oxford, UK:
Oxford Univ. Press, 2005, pp. 432-473.
[Ramachandran & Hubbard, 2003] Vilayanur S. Ramachan-
dran & Edward M. Hubbard. Hearing Colors, Tasting
Shapes, Scientific American, Vol. 288(5), pp. 53-61, 2003.
[Ramachandran, 2008] Phantoms in the Brain, Part
1-5, BBC4 documentary on blindsight, phantom-
limbs, anosognosia and hemispatial neglect, hosted by
V.S. Ramachandran, uploaded on January 23, 2008,
https://www.youtube.com/watch?feature=
endscreen&NR=1&v=ny5qMKTcURE
[Ramachandran, 1990] Ramachandran, Vilayanur S. Per-
ceiving shape from shading. The perceptual world: Read-
ings from Scientific American magazine. I. Rock. New
York, NY, US, W. H. Freeman & Co, Publishers, pp. 127-
138, 1990.
[SketchEngine, 2007] The SketchEngine (SkE), Corpus
Query System. Adam Kilgariff et al. 2007.
[SUMO, 2013] Adam Pease, Ian Niles et al. The Suggested
Upper Merged Ontology (SUMO). Last update checked:
04/09/2013.
[Wallington et. al, 2011] Alan Wallington, Rodrigo Agerri,
John Barnden, Mark Lee and Tim Rumbell. Affect Trans-
fer by Metaphor for an Intelligent Conversational Agent.
in K. Ahmad (ed.) Affecting Computing and Sentiment
Analysis: Metaphor, Ontology, Affect and Terminology.
Text, Speech, Language and Technology Series, ISBN
978-94-007-1756-5, 61-74. Springer, Heidelberg, 2011.
[Wallington et. al, 2006] Alan Wallington, Rodrigo Agerri,
John Barnden, Sheila Glasbey and Mark Lee. Considera-
tions on the Nature of Metaphorical Meaning Arising from
a Computational Treatment of Metaphor Interpretation. in
Proceedings of the 5th Workshop on Inference in Compu-
tational Semantics (ICoS-5), pages 1-10. 20-21 April,
Buxton, UK, 2006.
[WordNet Search 3.1., 2010] WordNet. Princeton Univer-
sity. 2010. http://wordnet.princeton.edu
Semantic Clues for Novel Metaphor Generator
Rafal Rzepka†, Pawel Dybala‡, Koichi Sayama‡and Kenji Araki†
†Hokkaido University, Sapporo, Japan
{kabura,araki}@media.eng.hokudai.ac.jp
‡Otaru University of Commerce, Otaru, Japan
{paweldybala,sayama}@res.otaru-uc.ac.jp

Abstract simple similes generating algorithm and finally describe pre-


liminary experiments for setting understandability treshold.
In this position paper we introduce our early find- We conclude with discussion about possible usage of the new
ings drawn from a new study on an over 30,000 en- data set and introduce our plans for the fuller usage of the
tries collection of metaphorical expressions manu- data.
ally extracted from Japanese literature. Such thor-
oughly chosen data assure higher standards when
figurative speech is analysed from the generation 2 Data Analysis
perspective and allow to develop an algorithm for 2.1 Onai’s Dictionary
explaining difficult or new notions. Such function Metaphors used in this study were acquired from Onai’s Great
is going to be implemented in our artificial tutor Dictionary of Japanese Metaphorical and Synonymical Ex-
project and here we describe a method for gener- pressions [Onai, 2005]. The dictionary contains metaphors
ating similes. We also show some statistical dis- selected from Japanese modern literature and Japanese trans-
coveries in the semantic level that could be useful lations of foreign works. The dictionary contains approxime-
knowledge for a dialog system with extended ex- tely 30,000 metaphorical entries, each of which includes:
planation capabilities.
• headline, i.e. a word or phrase used to look up
metaphors.
1 Introduction
• sometimes – sub-headlines, i.e. words or phrases similar
Figurative speech is one of the most spectacular thought pro- to the headline
cesses used in human communication. If we need to explain
some difficult word, to delicately suggest or emphasize some- • index – a phrase that was actually used in that particular
thing, we often use metaphors. Good teachers and famous metaphor (or its semantic equivalent)
poets have been using imaginative examples to help us un- • metaphor – actual metaphor example
derstand things, abstract phenomena and complex or straight-
• source – reference to the literature work from which the
forward emotions. Although human beings usually have no
metaphor was selected.
problems with creating such examples, choosing an under-
standable metaphor that will trigger somebody’s imagination According to the author, the dictionary was compiled to
is a difficult cognitive process [Carbonell, 1982], [Mcglone, assist in finding interesting and somewhat sophisticated ex-
1996]. The most famous theories on how we understand pressions that can be used instead of common phrases. If, as
metaphors are the categorization view [Glucksberg, 2001], in the example entry in Table 2.1, one needs to find an unusual
the comparison view [Gentner, 1983] and three hybrid views expression for “a rough woman”, first he would have to query
– the conventionality view [Bowdle and Gentner, 2004], the the word “woman” (headline), then search for the particular
aptness view [Jones and Estes, 2005] and the interpretive di- expression in the index and finally check the actual metaphor
versity view [Utsumi and Kuwabara, 2005]. This paper’s goal example.
is not to describe computational models for these methods
with their pros and cons but to support researchers working 2.2 Semantic Charactersistcs
on figurative speech generation as an addition to the existing To our best knowledge, the data we used is the biggest dig-
theories and prepare ourselves to implement them as a part of italized collection of Japanese metaphorical expressions and
explanation module for our artificial tutor project [Mazur et can be analysed from various angles. For the first trials with
al., 2012]. We used more than 12,000 of 30,000 metaphorical generation we have chosen the simplest and the most popu-
expressions (similes) gathered by Onai [Onai, 2005] which lar metaphorical figure of speech – a simile. A simile differs
allows for a new approach for computing or analysing alle- from a metaphor in that the latter compares two unlike things
gorical utterances in natural language interfaces. First, we by saying that the one thing is the other thing, while simile di-
show statisitcal data about words usage, then we propose a rectly compares two things through some connective, usually
Headline Index Metaphor Source
Japanese Transcription onna; josei hageshii onna hi no you ni hageshii onna Jakuchou Setouchi
English Translation a woman; a lady rough woman woman rough like a fire (a book by Jakuchou Setouchi)

Table 1: An example entry in Onai’s dictionary

“like”, “as” or by specific verbs like “resembles”. In order to Nouns Verbs Adjectives
select similes from our data set, we used a manually created eye (524) to look like (236) white (265)
set of such words (marks) used in Japanese. This allowed voice (437) to flow (204) black (156)
us to retrieve 12,214 similes, on which we performed some water (338) to shine (197) beautiful (138)
statistical tests. By using JUMAN morphological parser1 we sound (330) to stand (182) cold (135)
have separated and ranked 3,752 unique part of speech pat- face (325) to look (169) heavy (107)
terns that can be helpful while generating figurative expres- heart (309) to fall down (161) dark (101)
sions. Dependency parser KNP’s dictionary2 and semantic breast (282) to put out (146) sharp (101)
role tagger ASA3 were then used in order to rank most popu- light (265) to go (145) big (96)
lar categories and words needed to set weights for deciding on sky (235) to move (139) small (93)
the best simile candidate in the generation process. Most fre- head (218) to feel (137) detailed (86)
quent semantic categories characteristic to figurative speech
were colors, shapes, and patterns. The words with highest Table 2: Top 10 nouns, verbs and adjectives out of 10,658
frequency are shown in Table 2 (grouped by part of speech). morphological tokens found in corpus (similes only).
For comparison, semantic categories characteristic to random
web text (3,500 sentences from a blog corpus) were mostly
places, currencies, names, dates, organizations, education and
In order to examine the applicability of a generation task, the
the words most characteristic to a random web text were as
experimenter must conduct a metaphor generation task with a
follows.
huge number of concepts, therefore Abe et al. used Japanese
Nouns: newspaper corpus as a base for their language model. Re-
learning, media, bad, work, information, method, searchers also use the Web as a source for their models [Veale
enterprise, understanding, company, strength, and Hao, 2007][Masui et al., 2010] and utilize the latest the-
area, necessity, relationship, usage, utilization, sauri and ontologies as WordNet [Miller, 1995] to build so-
direction, United States, system, administration, phisticated generation algorithms [Huang et al., 2013]. Nu-
thought, two, city, money, district, caution merous examples extracted from Onai’s dictionary could be
Verbs: helpful for all existing approaches. Therefore we are planning
to be visible, to know, to be divided to test most popular approaches in nearest future. For the ba-
Adjectives: sic preparations we have used the 12,214 similes mentioned
individual, many, old in the previous section. Currently we are working on Ortony’s
Further semantic analysis data can broaden the system’s salience imbalance theory [Ortony, 1979] which predicts pos-
knowledge and become also helpful for recognizing figura- sible source-target shared attributes and their positions in
tive speech because when understanding users’ utterances as each ranking. Together, these concepts imply that low-high
metaphorical or idiomatic expression they need to be pro- topic-source pairings should cause increases in salience of
cessed by using different comprehension strategies. topic attributes. On Figure 1 we show the idea of two lists
of attributes that describe a word in an order of occurrences.
3 Example Usage: Novel Similes Generation So “sweet honey” is more natural than “sweet voice” but
The topic of machine generating metaphors is not as quite the same adjective can describe both nouns. However, as
popular as the understanding task. Most first trials were lim- Ortony’s theory suggests, if two adjectives are from the top
ited to narrow categories of target (topic/tenor) and source or bottom of the lists (distance between them increases), it is
(vehicle) as in [Hisano, 1996]. Ten years later [Abe et al., less likely that they can form an apt simile. We propose a
2006] have tackled problem of metaphorical data insuffi- method for calculating this distance in the following section.
ciency by using statistical analysis of language data to rep-
resent large scale human language knowledge stochastically. 3.1 Toward Calculating The Salience Imbalance
1 To observe which attributes could be used for a ground com-
Juman System, a User-Extensible Morphological Analyzer for
Japanese. Version 7.0: http://nlp.ist.i.kyoto-u.ac. bining a source and a target (as “strong” in man strong as
jp/index.php?Juman a bear) we experimented with two attribute lists. Prelimi-
2
Japanese Dependency and Case Structure Analyzer KNP 4.0: nary tests with different data sets suggested that it is fairly
http://nlp.ist.i.kyoto-u.ac.jp/index.php?KNP probable to find common attributes between distant source
3
ASA 1.0 - Semantic Role Tagger for Japanese Language: and target nouns using the Japanese case frames database
http://cl.it.okayama-u.ac.jp/study/project/ [Kawahara and Kurohashi, 2001], but as it is mostly verb-
asa/ centered, we also created attributes lists for nouns used in our
Figure 1: An example of attributes lists for a source - target pair created by ranking of adjective-noun bigram occurrences in a
blog corpus. Salience imbalance should, according to the Ortony’s theory, occur between attributes placed higher and lower on
such lists.

trials (an example is shown in Figure 1). We concentrate on • Source: matchbox


web-based textual resources because we aim at agent’s di- • Target: train
alog capabilities for using figurative speech mainly among
Japanese teenagers studying English. Newspapers have not • Ground: small
many metaphorical expressions and freely available corpus • Total source attributes: 64
for Japanese literature consists mostly of old books with old • Ground position in source attributes ranking (SourcePo-
written language. We chose one simple pattern: Source sition): 21
- M ark - Ground - T arget, which in English would be
T arget - Ground - M ark - Source as in Train small like a • Total target attributes: 7444
matchbox. The algorithm we created uses original source and • Ground position in target attributes ranking (TargetPosi-
target pair and compares lists of semantically related phrases tion): 5088
retrieved from the web. For instance associations list for fire
contains phrases as “to set”, “to put out”, “weak”, “fast” or 21 5088
“extinguisher”. When we input 100 random phrases accord- distance = · 100 − · 100 = 35
64 7444
ing to a chosen pattern, only 16 source-target pairs had com-
The results are shown in Tables 3 and 4.
mon attributes and this is because there are less adjectives
than verbs in the data. We have calculated the positions on 3.2 Preliminary Evaluation
the attribute lists that are sorted according to the frequency in
the web corpus so to set fire is on the top to carry fire is closer As we plan to use the generator for a English tutoring dia-
to the bottom. We have calculated the distance value between log system, we need to be sure that the example chosen by
common attributes using the following formula. computer is not too metaphorical and difficult to understand.
To set the tresholds we performed a preliminary evaluation
SourceP osition T argetP osition experiment and asked 5 Japanese language speakers (males
distance = 100 − 100
T otalSourceAttr T otalT argetAttr only) to evaluate phrases used in the previous section. The
subjects were asked to rate understandibility and aptness on
For example, from the metaphor “train small as a matchbox”, the 5 points scale. The results (see Table 3) show that hu-
the system first extracts “matchbox” (source) “train” (target) man evaluation averages for both aspects have set a borde-
and “small” (ground). Next, the rankings of attributes of line for distinguishing good similes (“voice sweet as honey”
“train” and “matchbox” are extracted from the case frames or “hand cold as ice”) from lesser ones as “hand delicate as
database, and the average position of “small” in each ranking woman” which is a repetition abbreviation of hand delicate
is checked. Finally, the system (after multiplying each posi- as woman’s hand.
tion by 100 to avoid dealing with very small fractions) calcu- By using human made examples from Onai’s dictionary
lates the difference of grounds position in these two rankings. we were able to set an average distance (D) to 20. Because
• Metaphor: Train small as a matchbox setting only distance was generating too many candidates, we
source position attribute common attribute position target attributive
in A-list usualness ground usualness in A-list distance
daughter 2892/29614 10 YOUNG 86 8531/9872 wife 76
fire 2340/3330 70 ROUGH 9 4371/47106 woman 61
horse 6193/12416 50 LONG 0.5 494/92105 face 49.5
matchbox 21/64 33 SMALL 68 5088/7444 train 35
honey 390/1162 34 SWEET 0.7 1124/151360 voice 33.3
blade 729/1934 38 THIN 7 527/7388 lips 31
fish 7798/17082 46 COLD 21 1673/7882 hand 25
zori thongs 201/310 65 BIG 42 376/893 oysters 23
be on the watch 367/1122 33 SHARP 10 1925/19620 eye 23
ice 2604/6532 40 COLD 20 10702/53506 hand 20
bleached 6/258 2 WHITE 7 758/10686 skin 5
elephant 688/3438 20 SMALL 16 15209/93516 eyes 4
death 425/4821 9 DEEP 10 246/2436 sleep 1
blood 225/7318 3 RED 2 164/7388 lips 1
paper 66/6002 1 THIN 0.1 11/10686 skin 0.9
woman 90/47106 0.2 DELICATE 0.3 164/53506 hand 0.1

Table 3: Sixteen metaphors which had common Ground in a Kyoto Case Frames-based attributes lists. Sorted in order in
distance which is the difference between attribute usualness values calculated from Ground position in both lists.

simile average average difference between difference between


understandability aptness understandability and aptness attribute positions
train small as a matchbox 4,8 4,2 0,6 35
skin thin as paper 3,966 2,8 1,166 0,9
hand cold as ice 4,4 4,4 0 20
woman rough as fire 4 4 0 61
hand delicate as woman 3 3 0 0,1
skin white as bleached 5 4,6 0,4 5
hand cold as a fish 2,6 2,6 0 25
eyes small as elephant 2,2 2,8 -0,6 4
voice sweet as honey 4,2 3,8 0,4 33,3
eye sharp like being on the watch 3,6 3 0,6 23
lips thin as a blade 3,4 3,48 -0,08 31
wife young as a daughter 4,4 3,4 1 76
lips red as blood 4,8 4,4 0,4 1
oysters big as zori thongs 4 3,4 0,6 23
sleep deep as death 5 4,8 0,2 1
face long as a horse 4 3,4 0,6 49,5
averages for tresholds: 3,96 3,63

Table 4: Results for understandability and aptness evaluation experiment. Lines in gray show similes which had both under-
standability and aptness higher than the averages.
have performed multiple aditional experiments with different [Jones and Estes, 2005] Lara Jones and Zachary Estes.
parameters to see which conditions are helping and which are Metaphor comprehension as attributive categorization.
not. Results of one strict set of conditions is shown in Table 5. Journal of Memory and Language, 53:110—124, 2005.
In most cases, if semi-correct results are counted as positive, [Kawahara and Kurohashi, 2001] Daisuke Kawahara and
the newly generated similes were significantly better that a Sadao Kurohashi. Japanese case frame construction
random generation, but further filtering semantically strange by coupling the verb and its closest case component.
outupts is needed. In Proceedings of the first international conference on
Human language technology research, HLT ’01, pages
4 Conclusions and Future Work 1–7, Stroudsburg, PA, USA, 2001. Association for
Computational Linguistics.
In this short paper we have introduced new possibilities for
figurative speech generation by using a new vast collection [Liu and Singh, 2004] Hugo Liu and Push Singh. Concept-
of Japanese metaphorical expressions. Because this is an net: A practical commonsense reasoning toolkit. BT Tech-
early stage of our study, we have performed only prelimi- nology Journal, 22:211–226, 2004.
nary tests and experiments in order to get a grasp of which [Masui et al., 2010] Fumito Masui, Rafal Rzepka, Yasutomo
tools and other repositories can be combined before we start Kimura, Jun-Ichi Fukumoto, and Kenji Araki. Www-
implementing the data to known theories about human’s abil- based figurative descriptions for japanese word. Journal
ity to use examples while explaining physical and abstract of Japan Society for Fuzzy Theory and Intelligent Infor-
objects. We have already started to work with more simile matics, 22(6):707–719, 2010.
patterns, also including verbs to fully utilize Kyoto Frames [Mazur et al., 2012] Michal Mazur, Rafal Rzepka, and Kenji
database. We are experimenting with N-gram frequencies Araki. Proposal for a conversational english tutoring sys-
of target, ground and source triplets to create vectors which tem that encourages user engagement. In Proceedings of
should help us discover more statistical dependencies. We are the 19th International Conference on Computers in Edu-
also testing WordNet and ConceptNet [Liu and Singh, 2004] cation, pages 10–12. Asia-Pacific Society for Computers
as a source for further calculation of semantic dependencies in Education (ICCE2011), 2012.
and show the latest progress during the workshop.
[Mcglone, 1996] Matthew S. Mcglone. Conceptual
Metaphors and Figurative Language Interpretation:
References Food for Thought? Journal of Memory and Language,
[Abe et al., 2006] Keiga Abe, Kayo Sakamoto, and 35(4):544–565, August 1996.
Masanori Nakagawa. A computational model of the [Miller, 1995] George A. Miller. Wordnet: A lexical
metaphor generation process. In Proceedings of the 28th database for english. Communications of the ACM, 38:39–
Annual Conference of the Cognitive Science Society, 41, 1995.
pages 937–942, 2006. [Onai, 2005] Hajime Onai. Great Dictionary of 33800
[Bowdle and Gentner, 2004] Brian Bowdle and Dedre Gen- Japanese Metaphors and Synonyms (in Japanese). Kodan-
tner. The career of metaphor. Psychological Review, sha, 2005.
111(1):193—216, 2004. [Ortony, 1979] Andrew Ortony. Beyond literal similarity.
[Carbonell, 1982] Jaime Carbonell. Metaphor: An in- Psychological Review, 86:161–180, 1979.
escapable phenomenon in natural language comprehen- [Utsumi and Kuwabara, 2005] Akira Utsumi and Yuu
sion. In Wendy Lehnert and Martin Ringle, editors, Strate- Kuwabara. Interpretive diversity as a source of metaphor-
gies for Natural Language Processing, pages 415–434. simile distinction. In In Proceedings of the 27th Annual
Lawrence Erlbaum, 1982. Meeting of the Cognitive Science Society, pages 2230–
[Gentner, 1983] Dedre Gentner. Structure mapping: A theo- 2235, 2005.
retical frame- work for analogy. Cognitive Science, 7:155– [Veale and Hao, 2007] Tony Veale and Yanfen Hao. Com-
170, 1983. prehending and generating apt metaphors: a web-driven,
case-based approach to figurative language. In Proceed-
[Glucksberg, 2001] Sam Glucksberg. Understanding figura-
ings of the 22nd national conference on Artificial intel-
tive language: From metaphors to idioms. Oxford Univer- ligence - Volume 2, AAAI’07, pages 1471–1476. AAAI
sity Press, 2001. Press, 2007.
[Hisano, 1996] Masaki Hisano. Influence of topic character-
istics on metaphor generation. In Proceedings of 38th An-
nual Conference of Japanese Association of Educational
Psychology (in Japanese)., page 449, 1996.
[Huang et al., 2013] Xiaoxi Huang, Huaxin Huang, Beishui
Liao, and Chihua Xu. An ontology-based approach to
metaphor cognitive computation. Mind and Machines,
23:105–121, 2013.
English translation Japanese in Roman letters
Lips as red as they were licking blood Chi-o nameru you-na akai kuchibiru
Lip as thin as somebody press a blade against it Ha-o oshitsukeru you-na usui kuchibiru
Voice so sweet as somebody was sucking up honey Mitsu-wo suiageru you-na amai koe
Wife as young as somebody was talking to daughter Musume-ni hanasu you-na wakai tsuma
Wife so young that resembles daughter Musume-ni niru you-na wakai tsuma
Woman so rough like she was dropping fire Hi-o otosu you-na hageshii onna
Sleep as deep as somebody was avoiding death Shi-o kaihi-suru you-na fukai nemuri
Hand as cold as somebody was biting ice Koori-o kamu youna tsumetai te
Hand as cold as somebody was putting it into ice Koori-ni ireru you-na tsumetai te
Skin as thin as somebody was cutting off paper Kami-o kiritoru you-na usui hifu
Skin as thin as somebody was peeling off paper Kami-o hagu you-na usui hifu
Skin as thin as somebody was scratching off paper Kami-o hikkaku you-na usui hifu
Skin as thin as somebody could stick paper (to it) Kami-ni haritsukeru you-na usui hifu
Face as long as somebody was being put on a horse (back) Uma-ni noseru you-na nagai kao
Face as long as somebody was aiming at a horse Uma-ni ateru you-na nagai kao
Face as long as somebody was separated from a horse Uma-kara hanasu you-na nagai kao
Wife as young as somebody was passing to daughter Musume-ni watasu you-na wakai tsuma
Wife so young that you could get used to daughter Musume-ni nareru you-na wakai tsuma
Wife so young that she could take away daughter Musume-wo ubau you-na wakai tsuma
Wife so young that you could find in daughter Musume-ni mitsukeru you-na wakai tsuma
Wife so young that you could be worried about daughter Musume-ni kizukau you-na wakai tsuma
Wife so young that you could let her go to daughter Musume-ni hanasu you-na wakai tsuma

Table 5: Examples for an Kyoto frames-based generation experiment for candidates with the same particles and grounds, where
there were more than 30 verb-source and 30 grounds in Ameba corpus. Conditions for salience imbalance distance values
were D < 100 and D > 5. First group of examples were classified by authors as correct, second as semi-correct and third as
incorrect, although there was no agreement in case of few examples from the last group suggesting that imagination plays a big
role in evaluating novel metaphors.

S-ar putea să vă placă și