Hosted here with Ian’s consent: Spoken Language Iconicity: An articulatory-based Analysis of 66 Languages
Recently, myself and other members of the LEvInSoN group hosted a Summer School at the Max Planck Institute for Psycholinguistics as part of our Minds, Mechanisms, and Interaction in the Evolution of Language Workshop. The goal of this summer school was ambitious – a synthetic view of the process of designing, creating, running, and analysing the results of an experiment – all in two days. We wanted to show the students the process of scientific collaboration, warts and all. The contents of the summer school, including the data collected for our experiment can be found in this GitHub repository.
We were, overall, very happy with the results of the summer school- both in the learning outcomes for the students and the actual experimental results – below I outline the basics of the study conducted, the topics covered, and the results of our experiment.
The September Tutorial in Empiricism was a massive undertaking with many moving parts and many contributors. I’d especially like to thank my fellow organisers, without whom nothing would have been possible – while I stressed about the details of the experiment and wrangling the instructors, they did a great job of making sure all of the bureaucratic boxes were ticked and making sure things ran smoothly.
The Summer School was jointly taught by 6 Instructors:
Alan Nielsen– Introduction, Context and Grounding, Experimental Basics, Descriptive Statistics
Hannah Little– Data Science and the Open Science Framework, Data Preparation
Yasamin Motamedi– Trial Structures in Python, Linear Models and Basic Statistics in R
Bill Thompson– Simulating Data, Making Predictions in Python, Advanced Statistics, Bayesian Basics, Agent-Based Bayesian Modelling
Justin Sulik– Building Experiments in jsPsych, Online Experiments with Mechanical Turk, Advanced Statistics, Visualisation and Reporting
Ashley Micklos– Linear Models and Basic Statistics in R
Twenty-six students participated in the Summer School, from a wide range of research backgrounds and levels of experience. Of our 26 students the majority (19) were PhD students, but we also had 2 postdoctoral students, 3 Master’s students, and even 2 undergraduates.
The following students would like to be recognized for their attendance:
Federica Bartolozzi – PhD Candidate – Max Planck Institute for Psycholinguistics
Miguel Borges – PhD Candidate – Max Planck Institute for Psycholinguistics
Giusy Cirillo – MA Student – University of Tubingen
Lara Clauss* – PhD Candidate – Max Planck Institute for Psycholinguistics
Varun deCastro-Arrazola – PhD Candidate – Leiden University and Meertens Instituut
Ian Joo – MA Student – National Chiao Tung University
Greta Kaufeld – PhD Candidate – Max Planck Institute for Psycholinguistics
Fiona Kirton – PhD Student – Center for Language Evolution, University of Edinburgh
Ezequiel Koile – Visiting Postdoctoral Researcher – Max Planck Institute for the Science of Human History
Elly Koutamanis – PhD Student – Center for Language Studies, Radboud University
Hannah Lutzenberger – PhD Student – Center for Language Studies, Radboud University
Katie Mudd – PhD Student – Vrije Universiteit Brussels
Limor Raviv – PhD Student – Max Planck Institute for Psycholinguistics
Constanze Schon* – Visiting Intern – Max Planck Institute for Psycholinguistics
Kazuki Sekine – Postdoctoral Researcher – Radboud University/Max Planck Institute for Psycholinguistics
Chen Shen* – PhD Student – Center for Langauge Studies, Radboud University
Anita Slonimska – PhD Candidate – Center for Language Studies, Radboud University. ITSC Fellow
Katja Stark – PhD Student – Max Planck Institute for Psycholinguistics
Katarina Stekic – BSc Student – Laboratory for Neurocognition and Applied Cognition, University of Belgrade
Jeroen van Paridon – PhD Candidate – Max Planck Institute for Psycholinguistics
Marieke Woensdregt – PhD Student – Center for Language Evolution, University of Edinburgh
Nezihe Zeybek* – PhD Student – University of Burgundy
Eirini Zormpa – PhD Student – Max Planck Institute for Psycholinguistics
As mentioned, our students had a broad range of experience with the various parts of designing, hosting, and analysing large-scale online experiments – in advance of the summer school we solicited self-reported competence ratings for a number of topics, which can be seen below:
Thus, we had students work through the summer school in groups, which we attempted to make as balanced as possible – you can see the average competence score for each group below. This approach proved especially valuable, because it allowed the more experienced students opportunities to pass on their knowledge to younger or otherwise less-experienced students in a hands on fashion.
This also served as a great first example of the ability of R (and ggplot specifically) to easily output informative graphs from pretty minimal code.
Human beings are not unbiased perceivers of the world, taking in information from the environment and processing it in a vacuum. One way that humans are biased is in the types of associations that they make between sensory modalities – starting in at least the early part of the 20th century it was recognized that humans are biased to associate certain types of sounds with certain types of meanings.
In the examples above, it’s likely that when tasked with choosing appropriate labels for the given images you’ll have chosen that the jagged star-like image should be called takete (rather than maluma), that the large table fits better with the name mal than mil, and that the correct word to describe the pictured dog is fuwafuwa (a Japanese ideophone meaning fluffy) rather than korokoro (a Japanese ideophone meaning ‘a small object rolling repeatedly’).
These types of iconic biases between sounds and meanings are often referred to as sound-symbolism – an area that has become a growth area in psycholinguistics research:
Associations between what would seem like otherwise unrelated perceptual modalities are not, however, limited to those that can be explored linguistically – experimental participants have been found to have dozens if not hundreds of these types of associations – they suggest for example that small objects are happy, bright, fast moving, and high pitched. As researchers have tested for more and more of these crossmodal biases, they have found increasing evidence that humans make associations both within and between sensory domains, which raises a number of questions.
From our perspective, the most important question is How are crossmodal biases related to each other?
Unfortunately, the current state of knowledge in the crossmodal literature makes this question vexingly difficult to answer – even answering fundamental questions that we need answers to in order to approach the above question is challenging. There are a number of reasons for this: first, researchers do not typically share stimuli, so finding, for example, that participants in one study associate high pitched sounds with fast objects while in another study they associate high pitched sounds with small objects can sometimes be minimally informative – often the pitch differences, size differences, or speed differences will be entirely idiosyncratic to the individual study in question. Relatedly, the findings of various studies are often isolated – even ambitious projects typically look at associations between less than a handful of domains. These studies are important, generally well-designed and implemented, and informative, but they make enumerating the types of crossmodal associations that human participants make difficult. If one could make a general summary of crossmodal research, it would be that where we look for crossmodal biases, we find them.
The focus of the present study then is to exhaustively test associations between a number of domains, cataloging their relative strengths and creating a network of associations.
The domains tested were:
Amplitude (Loudness) – Loud vs. Quiet
Pitch – High vs. Low
Noise – Noisy vs. Tonal
Size – Large vs. Small
Shape – Jagged vs. Curvy
Speed – Fast vs. Slow
Brightness – Bright vs. Dull
Color – Yellow vs. Blue, Red vs. Yellow, Red vs. Green, Red vs. Blue
Affect- Stressed vs. Calm, Pleased vs. Disgusted, Happy vs. Sad, Excited vs. Bored
Stimuli can be found on the github repository here.
In the experiment conducted for the Summer School, experimental participants were recruited via Amazon Mechanical Turk and tasked with making associations between perceptual domains – you can see the experimental interface here. The experimental interface was created in jsPsych by Justin Sulik.
Participants were tasked with responding to trials like the following below:
Participants were each shown 96 of these trials, and we solicited responses from a total of 210 participants via mechanical turk (60 pilot, 150 main experiment). Collectively, these participants were tested on all comparisons possible for our 9 chosen perceptual domains – thus they were asked whether Small things were Loud or Quiet, Noisy or Tonal, High or Low pitched, Jagged or Curvy, Fast or Slow, Bright or Dull, etc.
We found many interesting results, and continue to further analyse the data from the Summer School in preparation for publication, but our general suggestion that many or most associations are responded to in a biased fashion was upheld. Below you can see a heatmap of effect sizes for comparisons from our main experiment:
The heatmap may seem difficult to read, but it’s not – the value shown in each cell is the effect size of the association (calculated from a nonparametric Wilcoxon signed-ranks test), thus high values reflect larger effect sizes (more consistency between participants). Gray cells show non-significant associations – all other associations are significant at p < 0.05. The difference in colors (blue vs. red) informs you about the direction of association made by participants – the bright blue square (effect size = 0.87) in the top left corner of the heatmap tells you that generally, participants suggested that fast-moving images were excited (rather than bored). In the Noisy/Tonal row you can see an example of a strong negative association (effect size = 0.77) – participants responded that Noisy sounds were Disgusted (and thus than tonal sounds were pleased).
Any cell can be read in this fashion – If you take a Domain on a row, e.g. Bright/Dark, and a Domain on a column, e.g. Happy/Sad, the cell will be blue if participants associated the token on the left of the row domain (e.g. Bright) with the token on the left of the column domain (e.g. Happy; in this case most participants suggested that Bright colors are happy). if a cell is red, on the other hand, it suggests that the token on the left of the row domain was regularly paired with the token on the right of the column domain (and vice-versa – e.g. participants generally agreed that Jagged images were sad (and curvy images happy)).
As part of the Summer School, we encouraged students to make predictions about what the results of our main experiment would be, having provided them with some insight about how participants responded during the experimental pilot.
The rules of our hypothesis challenge were fairly simple- participants would make a set of predictions about how participants would respond, and we would compare their predictions to the actual results of the experiment. You can see the deviation of each set of predictions from the actual observed results (brighter red = larger difference) below:
Our top 3 sets of predictions were made by Jeroen van Paridon (average deviation = 0.134), Ezequiel Koile (average deviation = 0.178), and Ian Joo (average deviation = 0.234). From these three students a winner was chosen based on the participant who had made the fewest overall predictions (and allowed the imputation procedure covered by Bill Thompson during the summer school to fill in unmade predictions). This gave us a clear winner – in addition to producing the best overall predictions, Jeroen van Paridon also made less individual predictions than either of his competitors.
This provided an additional teaching opportunity for the other students about the power of computation. Whereas Ian Joo studies sound symbolism and made his predictions based on his own knowledge of the literature, Jeroen is a computationalist and took a brute-force mathematical/computational approach to the problem – generating a sampling procedure to best explain the pilot data with a minimal number of predictions. So congratulations to Jeroen for his interesting approach to the problem (which actually mirrors some of the more advanced ways we are continuing to look at the data). For his troubles, he won a copy of Rethinking Statistics.
Closing Thoughts and Future Directions
We all, as young academics, found the process of organising and running this summer school immensely rewarding – as a collaborative effort it highlighted good scientific practice not only to the students, but to all of us as well, and showed how productive it can be to bring together researchers with different types of expertise all working on a single project. There were a few hiccups along the way – but those were expected and desirable, given our framing of the Summer School as being a synthetic and honest view of science as it is done, rather than science as it is written.
I’d really like to thank the students who participated, especially those who came to the endeavour with a positive attitude. With the Summer School taking place over only two days, participants were presented with an impossible learning task, so rather than focusing on participants leaving with the ability to put together a similar experiment immediately, we aimed to provide them with some knowledge about what types of practices are possible for modern empiricists, and plenty of additional reading and supplementary materials that would allow them to later work through at their own pace and turn the tools they were provided with towards their own projects. As a nice bit of validation, some of our students provided us with lovely testimonials about their experience.
To that end, the GitHub repository for the summer school will remain public, and in the coming months we will assemble a public-facing website with all of the worked materials available in handbook format for anyone who is interested. We hope to be able to format the summer school materials such that any of us who participated in the teaching of this inaugural version would be able to present our own version of the school at other locations and later dates – so keep your eyes open for future possibilities.
On a personal level, I was immensely lucky that we chose to pursue one of my projects for the purposes of the Summer School, and I’m very happy with the results, so look for a more complete description in the near future. The upside of this project is not only in cataloging a set of crossmodal biases in our typical WEIRD population of experimental participants, but also in the possibility of extending further refinements of this procedure to other cultures and languages, allowing us to compare the degree to which crossmodal biases are universal vs. language specific. I recognize that the stimuli and the design of this preliminary experiment is imperfect (in fact, a major part of our analysis looks at the influence of task demands on these types of experiments), so am looking forward to feedback from other researchers.
Today I had the opportunity to present some of the new work I’ve been doing on modeling the relationship between cognitive biases and typology with my most excellent collaborator Bill Thompson at MPI Proudly Presents.
I must admit that when Bill and I signed up to give a talk about our work I was very skeptical- our work is relatively new and very preliminary, and I think we would both have worries about how it would be received by a general audience, it being as abstract as it was. I ended up being very happy with how our talk went, and we put a lot of work into it- I am continually amazed by how great of a thinker Bill is – how you can throw any problem at him and he can discuss it and make you think about things in ways that you might not have, despite being the supposed “expert” on a subject- but this isn’t a post about that.
Today I was blown away by the quality of the talks, mostly given by junior academics (MSc and PhD students). The event is titled “MPI Proudly Presents”, and in his closing remarks Peter Hagoort stressed that we should indeed be proud of the work we do- the MPI for Psycholinguistics is a truly unique place, and the work being done there on so many parts of the puzzle of language is really inspiring. But again – this isn’t a post about that.
What struck me most of all today, beyond the quality of the talks, was having my expectations about the *kinds* of talks that would be given being shattered- this was not a group of elitist “we have it all figured out” students preaching about how the gene, the animal model, the language, the computational model, or the experimental approach that they are using solves all of the problems of research into the evolution of language – far from it. My experience of talking to many of the other speakers after the day of talks was that they were hungry, gifted, and intellectually honest young academics. They were not prideful, but humble and acknowledging that they do not have all the answers. Maybe I am jaded beyond my years, but this has not been my experience with science generally, so I really found that today recharged my batteries.
As an early career academic I have many worries about the future of science, and especially the future of science communication- it is easy to be tempted down the dark path of making unsubstantiated claims to grab eyeballs – today I saw almost none of that. I saw young researchers with a similar mindset to my own- idealistic thinkers who are interested in solving difficult problems without simplifying away everything interesting and making impossible claims. There were many talks from areas outside of my own specialisation and they were delivered so wonderfully that I rarely found myself lost – the spirit of good science communication was very much alive at the MPI today: results were framed plainly, but without stripping them of nuance or their connection to the real world.
It’s not very often that I feel like I managed to give one of the worst talks at a workshop- and I stand behind the work that Bill and I presented today and think it was important- but even if every person in attendance agreed that our talk was the worst of 15, I wouldn’t feel bad about it- there truly was a rarefied air of excellent science going on- massive amounts of credit are due not only to the young academics who gave talks but also to their directions, the MPI, and the MPG. I could write many blog posts about each of the individual talks that I saw today – fascinating stuff about genetics, the neurobiology of language, comparative cognition, and experimental linguistics, but know that the work presented will all make it to publication, and trust the authors to present their own work much better than I ever could. Based on my personal interests a few talks stood out- Amanda Tilot‘s talk about relatively modern evolution of brain structure, my good friend Limor Raviv’s talk about the linguistic niche hypothesis, and Matthias Sjerp‘s talk about cortical tuning of speech processing specifically, but to focus too much on any of those talks would be a disservice to the amazing amount of information I was bombarded with today.
I said about that this isn’t a post about how unique the MPI is – In the words of my favorite comedian, I am not for the mawkish- but if something is true, then it is not sentimental. Ultimately I am both incredibly proud and incredibly humbled to be part of the MPI. Steve Levinson’s generosity and support in bringing in our group is something that I know I will look back upon fondly no matter where my academic career takes me. Today I was reminded of that fact – something it is easy to forget while bound up in the day-to-day grind of being a researcher. The path forward is not always clear, and no one has all the answers – today seeing so many talks that made substantive scientific contributions while remaining humble reinforced that. To be able to speak in the same company as those who presented today, and to such an established and important audience was an amazing opportunity that I never thought I would have. I am immensely privileged to be at the MPI, to work with my incredible collaborators, and to be surrounded with so many outstanding academics.
Academia… maybe it ain’t so bad after all.
Tomorrow I will be giving at a talk to the Iconicity Focus Group here at the MPI.
Here are the pertinent details:
Tuesday, December 6th, 2016
12:00 – 13:00
Recasting the magic bullet: Testing the limits of the effectiveness of iconicity for language learning
In this talk I will focus on outlining reasons to be more conservative in our expectations for the explanatory power of iconicity in langauge evolution and acquisition. I begin by discussing a pair of studies which demonstrate that some learning benefits previously attributed to iconicity are actually largely explainable in terms of a benefit for systematicity. Following from this, I explore the general idea that in addition to learnability benefits for iconic elements themselves, the acquisition of iconic tokens can bootstrap the acquisition of the rest of language. To do this, I explore a simple (and admittedly straw-man) version of the sound-symbolic bootstrapping hypothesis, making the argument that just because one thing (learning iconic tokens) happens before another (learning arbitrary tokens) does not mean that anything we can meaningfully call bootstrapping has taken place.
I’m really looking forward to giving this talk. It’s data from my PhD., so a few folks will already be familiar with the work, but I haven’t presented it before in this way, especially to a group of experts about iconicity, so I’m very much looking forward to hopefully getting lots of good feedback. For what it’s worth, my choice to present my data with this type of argument comes from discussions following Marcus Perlman’s talk to the LEvInSON group last month.
Slides and comments on the talk forthcoming later this week!
The projects page is now online and active, and I’ve added a few of the projects that I’m currently working on. Over the next few weeks I will post a few more. The goal is to have all of my projects listed by the New Year.
Then I can put in another section of the website devoted to my various crazy ideas that are not currently active projects.
The 11th International Symposium on Iconicity in Language and Literature is going down in April 2017 in Brighton, UK, and that’s exciting stuff.
Myself, Marcus Perlman, Ashley Micklos, Hannah Little, and a few others saw the call for abstracts and have opted to propose a double panel session titled “Experimental approaches to iconicity in language evolution.” We’re all very excited about the possibilty of presenting our work together, especially as Mark Dingemanse and Pamela Perniss have put themselves forward as discussants for the panel.