Based on Howard’s comment on my last blog post, I will today give an overview of how I try to stay up to date with current research in the AI and Conceptual Spaces area. What are conferences, workshops, mailing lists, etc. that I think are relevant?
A few days ago, I had the chance to attend the workshop “Concept Learning and Reasoning in Conceptual Spaces” in Bochum. Here’s a link to the workshop’s website: CLRCS 2017
It was a really great event with researchers working on conceptual spaces from a wide variety of perspectives, ranging from AI and linguistics over psychology and neuroscience to philosophy. Today, I would like to give a short summary of the workshop for those who were not able to participate but who are nevertheless interested in what kinds of topics have been discussed.
It’s nice to have a mathematical definition of concepts in a conceptual space. It’s also nice that we can create new concepts based on old ones, for instance by intersecting them. But being able to talk about the relation of two concepts is certainly also useful. Last time, we talked about the size of a concept. We can use the size of concept to figure out that the concept of “animal” is more general than the concept of “Granny Smith” – simply because it is larger.
But there are also other ways of describing the relation of two concepts. Two of them, namely subsethood and implication, will be presented in today’s blog post.
A few weeks ago, I got the notification that my paper “Measuring Relations between Concepts in Conceptual Spaces”  (preprint available here) was accepted at the British SGAI Conference on Artificial Intelligence.
One of the question that I discuss there is posed in the title of this blog post: What’s the size of a concept?
In general, one can say that the size of a concept in a conceptual space tells you something about its specificity: Small concepts (like Granny Smith) are more specific, whereas large concepts (like fruit) are more general.
But how exactly can we measure the size of such a concept within my proposed formalization? My paper  gives a mathematical response to that, and today I would like to sketch the basic idea behind it.
Recently, another one of my papers  (look at the preprint here) has been accepted at the German Conference on Artificial Intelligence. It is a quite technical paper with a lot of formulas, but I’ll try to illustrate the overall high-level idea in this and one or two future blog posts.
Today I would like to talk about the starting point of the research presented in this paper: The observation that convex regions in a conceptual space are highly problematic if we want to represent correlations between domains.