When a person suffers from a stroke, the blood flow to certain areas is rapidly disturbed either because a blood vessel starts bleeding (haemorrhage) or because a blood vessel is blocked. If the stroke affects brain areas which are important for language, then the language functions can be partially or fully lost. This condition is called aphasia, but it is sometimes reversible at least to some degree with time, treatment and rehabilitation. Bilinguals, that is, people who speak more than one language, are known to recover from aphasia in a number of different ways. The most common case is when the bilingual patient similarly recovers both languages equally well as in parallel aphasia. In some cases, however, the recovery of bilingual is disproportionately favouring one of his/her languages and this type of recovery is called selective aphasia.
The case of selective aphasia made researchers initially believe that each language of a bilingual person must be located in a different brain area, given that one language recovers better than the other after the stroke. However, with the help of brain imaging scans we now know that this idea is not correct. On the contrary, when a person speaks many languages, they all activate a common network of brain areas. While the whole picture of how the brain generates multiple languages is still partly unclear, we do know a number of factors that seem to influence to what extent the languages of a bilingual patient will recover after a stroke. If a person is less proficient in one of the two languages, this language may not recover as good as the more proficient language. This means that the more automatic a skill is, the easier it is to recover it, whereas something that takes effort such as language that one only speaks rarely is harder to recover. Social factors and emotional involvement are also important if we want to understand which language will recover after a stroke, for instance how often a specific language is used, or what emotions are associated with a specific language. However, it is still unclear exactly how these factors interplay in predicting the recovery success.
One of the current theories on why bilingual aphasics may disproportionately recover one language better than the other suggests that this happens when the stroke damages specific control mechanisms in the brain. When a bilingual person knows two languages, he/she needs to suppress or ‘switch off’ one of the languages while using the other language. If the mechanisms that control this switch are damaged during the stroke, the aphasic patient may no longer be able to similarly recover both languages as the ability to control the language use has been lost. In this case the person may appear to have completely lost one of the languages, but the problem is actually one of control. Recently researchers found that the control mechanisms are more impaired in bilinguals with selective aphasia who recover only one language than in bilinguals with parallel aphasia who recover both languages. Interestingly, when languages recover after stroke, the connections between language and control areas in the brain are re-established. While this interesting finding supports the theory linking selective aphasia to impaired control mechanisms, it is only one of several theories and researchers are currently trying to better understand what other causes may also underlie the surprising recovery patterns that can be seen in bilingual aphasia.
Written by Diana Dimitrova and Annika Hulten
Fabbro, F. (2001). The bilingual brain: Bilingual aphasia. Brain and Language, 79(2), 201-210. pdf
Green, D. W., & Abutalebi, J. (2008). Understanding the link between bilingual aphasia and language control. Journal of Neurolinguistics, 21(6), 558-576.
Verreyt, N. (2013). The underlying mechanism of selective and differential recovery in bilingual aphasia. Department of Experimental psychology, Ghent, Belgium. pdf
Children differ in their learning strategies and developmental patterns. Therefore age “benchmarks (for example, "first words appear around 12 months") only represent a very general average. It is completely normal for children to develop a skill either a few months earlier or a few months later than the “benchmark” age. With that in mind, we can summarize the general pattern for how children learn to use simple sentences.
Simple sentences start appearing in children's speech when they are 30 to 36 months old. To some, this might seem quite late, considering that children's first words appear around 12 months of age. But before they can string words into sentences, children need to acquire some basic knowledge about their language's grammar. For example, we think of nouns and verbs as the building blocks of sentences ("cat"+"want"+"milk" = "The cat wants some milk"), but even very simple sentences often require the speaker to add other words ("the"/"some") and inflections ("want"->"wants"). Learning function words and inflections is not trivial because, unlike content words like "cat" and "milk", children cannot experience the meaning of "the", "some", and "-s". To make things even more complicated, the words need to be put in the right order and said aloud with a melody and emphasis that matches the intended meaning (compare "The CAT wants some milk" vs. "The cat wants some MILK"). Managing all of this, even for simple sentences, can be difficult for young children.
But even though children under 30 months do not yet produce full sentences, they do learn how to combine words in other ways. For example, soon after their first birthdays, many children begin to combine words with gestures such as pointing, showing, and nodding. The combination of a word and a gesture (for example, the word "milk" + a nod) says more than just the word or the gesture alone. In fact, some studies suggest that these early word-gesture combinations are linked to children’s development of two-word combinations just a few months later.
From around 18 months, children begin to use two-word sequences like "Bear. Trolley." to express sentence-like meanings (in this case: "The bear is in the trolley”). At first, the words in these sequences sound unconnected, like individual mini-sentences of their own. But as children get more practice, the pauses between words get shorter and the melody of each word seems more connected to the next one.
At 24 months, when children are at the heart of this "two-word stage", they often produce sequences of 2–3 words that sound like normal sentences, only with most of the function words and inflections left out ("There kitty!") At this stage, some children even consistently place the words in a specific order, for example, with "pivot" words ("more", "no", "again", "it", "here") only being used in the first position for some children ("more apple", "here apple") and the second position for others ("apple more", "apple here"). As children begin to use shorter pauses, more consistency in word order, and speech melodies that join the individual words together, they show evidence that their word sequences were planned as a single act: a sentence.
Even when children finally begin producing simple sentences around 30–36 months, they still have a lot to learn. Their ideas about how to combine words into longer sequences are not yet adult-like, and rely quite a bit on what they hear most often around them. Marking words with the correct inflections can be complicated, and it is not uncommon for children to make errors as they learn which inflections are regular (walk, walks, walked...) and which are irregular (am, is, was...). In the 2 years following the onset of simple sentences, children continue to make huge advances in their vocabulary size and use of grammatical marking (among other things). Their sentences become more elaborate with age and, by the time they are 4–5 years old, they have learned a lot of what they need to know to communicate fluently with others.
If you would like to learn more about language development please have a look at the website of Nijmegen's Baby & Child Research Center.
By Marisa Casillas & Elma Hilbrink
Clark, E. V. (2009). “Part II: Constructions and meanings”. In First language acquisition (pp. 149–278). Cambridge University Press.
Iverson, J. M., & Goldin-Meadow, S. (2005). Gesture paves the way for language development. Psychological science, 16(5), 367-371.
Siegmund et al (2014) were the first to empirically investigate the link between programming and other cognitive domains, such as language processing, at least using modern neuroimaging methods. They used functional magnetic resonance imaging (fMRI) which measures changes in local blood oxygenation, as a result of brain activity in different networks across the brain. Undergraduate students of computer science were scanned while reading code snippets for comprehension and while reading similar code, looking for syntax errors without comprehension. The results showed activation in the classical language networks, including activation in Broca's, Wernicke's and Geschwind's territories, more in the left hemisphere.
The code was made to enhance so called bottom-up comprehension, which means reading and understanding expression by expression and line by line, rather than browsing the overall structure of the code. This process can be compared to language processes such as clipping words together, according to syntax, to arrive at a coherent sentence meaning (as well as connection meaning across sentences in discourse). It is possible, and has been suggested in the computer science literature on the skill set needed to be a good programmer (Dijkstra, 1982), that people master their native language well are also more efficient software developers. A mechanistic explanation could be the strength of the connections between the mentioned brain regions, which changes from person to person. In summary, similar brain networks are found for programming and language comprehension.
Answer by: Julia Udden, Harald Hammarström and Rick Jansen
Siegmund, J., Kästner, C., Apel, S., Parnin, C., Bethmann, A., Leich, T., Saake, G. & Brechmann, A. (2014) Understanding Source Code with Functional Magnetic Resonance Imaging. In Proceedings of the ACM/IEEE International Conference on Software Engineering (ICSE).
Dijkstra. How Do We Tell Truths that Might Hurt? In Selected Writings on Computing: A Personal Perspective, 129–131. Springer, 1982.
Programming languages are usually taught to teenagers or adults, much like when learning a second language. This kind of learning is called explicit learning. In contrast, everyone’s first language was learned implicitly during childhood. Children do not receive explicit instructions on how to use language, but learn by observation and practice. Part of what allows children to do this is the interactive nature of language: people ask questions and answer them, tell others when they don’t understand and negotiate until they do understand (Levinson, 2014). Programming languages, on the other hand, are passive: they carry out instructions and give error messages, but they don’t find the code interesting or boring and don’t ask the programmer questions.
Because of this, it is sometimes difficult to think in a programming language, i.e., to formulate instructions comprehensively and unambiguously. The good news is that many programming languages use similar concepts and structures, since they all are based on the principles of computation. This means that it is often quite easy to learn a second programming language after learning the first. Learning a second natural language can take much more effort. One thing is clear - it is becoming increasingly important to learn both kinds of language.
Answer by: Julia Udden, Harald Hammarström and Rick Jansen
Levinson, S. C. (2014). Pragmatics as the origin of recursion. In F. Lowenthal, & L. Lefebvre (Eds.), Language and recursion (pp. 3-13). Berlin: Springer.
Some programming languages certainly look a lot like natural languages. For example, here’s some Python code that searches through a list of names and prints one out if it’s also in a list called invited_people.
for name in my_list: if name in invited_people: print name
However, some other programming languages are much less readable. Here is some code in Scheme that does the same as the code above.
(map (lambda (name) (if (cond ((member name invited_people) name)) (display name) name)) my_list)
So, how similar or different are natural languages and programming languages really? To answer these questions, we need to understand some central terms that linguists use to describe the structure of languages, rather than just looking at the surface of what looks similar or not.
Similar structures of semantics and syntax
Two of the most central concepts in linguistics are the concepts of semantics and syntax. In short, semantics is the linguistic term for meaning, but a more precise explanation is that semantics contains the information connected to concepts. For instance, a word form like “sleep” (spelled S-L-E-E-P, in either letters or pronounced as sounds) designates a certain action of a living organism and that’s the semantics of that word. The syntax on the other hand is the structure of how words of different kinds (e.g. nouns and verbs) can be combined and inflected. The sentence “My ideas sleep” is a well-formed English sentence from the point of view of syntax, but the semantics of this sentence is not well-formed since ideas are not alive and thus cannot sleep. So, semantics and syntax have rules, but semantics relates to meaning and syntax relates to how words can be combined.
Having explained the semantics and syntax of natural languages, let’s turn back to programming. In programming languages, the coder has an intention of what the code should do. That could be called the semantics or the meaning of the code. The syntax of the programming language links a certain snippet of code, including its “words” (that is variables, functions, indices, different kinds of parenthesis etc.) to the intended meaning. The examples of python and scheme above have the same semantics, while the syntax of the two programming languages differ.
We have described many parallels between the basic structure of natural languages and programming languages, but how far does the analogy go? After all, natural languages are shaped by communicative needs and the functional constraints of human brains. Programming languages on the other hand are designed to have the capacities of a Turing machine, i.e. to do every computation that humans can do with pen and paper, again and again.
It is necessary for programming languages to be fixed and closed, while natural languages are open-ended and allow blends. Code allows long lists of input data to be read in, stored and rapidly parsed by shuffling around data in many steps, to finally arrive at some output data. The point is that this is done in a rigorous way. Natural languages on the other hand must allow their speakers to greet each other, make promises, give vague answers and tell lies. New meanings and syntax constantly appear in natural languages and there is a gradual change of e.g. word meanings. A sentence from a spoken language can have several possible meanings. For example, the sentence “I saw the dog with the telescope”, has two possible meanings (seeing a dog through a telescope or seeing a dog that owns a telescope). People use context and their knowledge of the world to tell the difference between these meanings. Natural languages thus depend on an ever changing culture, creating nuances and blends of meanings, for different people in different cultures and contexts. Programming languages don’t exhibit this kind of flexibility in interpretation. In programming languages, a line of code has a single meaning, so that the output can be reproduced with high fidelity.
Answer by: Julia Udden, Harald Hammarström and Rick Jansen
Languages organize their nouns into classes in various different ways. Some don’t have any of these noun classes (e.g., English: every noun is just ‘it’), some have two (e.g., French: every noun is either masculine or feminine), some have as many as 16 (e.g., Swahili: there are different classes for animate objects, inanimate objects, tools, fruits...). Some languages with two noun classes differentiate between masculine and feminine (e.g., French), some between common and neuter (e.g., Dutch). Clearly, all these languages differ in terms of what might be called their ‘grammatical gender system’. While Western European languages might give the impression that grammatical gender (e.g., whether nouns are male or female) primarily affects the articles placed in front of nouns (e.g., de versus het in Dutch), these differences often affect the noun itself and other words connected to it as well. Polish, for example, doesn’t even have articles (as if the English word ‘the’ didn’t exit) but still uses an intricate gender system which requires adjectives to agree with nouns. The reasons for these differences between languages remain mysterious.
Given that the gender system of a language permeates all sentences, one might wonder whether it goes further and also influences how people think in general. On the face of it, this appears unlikely. A grammatical gender system is just a set of rules for how words change when combined. There is no ‘deeper’ meaning to them. Nonetheless, a series of experiments have come up with surprising results.
In the 1980’s Alexander Guiora and colleagues noticed that two to three year old Hebrew-speaking children (whose language system does differentiate between masculine and feminine nouns) are about half a year ahead in their gender identity development compared to English- speaking children. It is as if the gender distinction found in Hebrew nouns gave these children a hint about a similar gender distinction in the natural world.
Adults too seem to make use of grammatical gender even when it doesn’t make any sense to do so. Roberto Cubelli and colleagues asked people to judge whether two objects were of the same category (e.g., tools or furniture) or not. When the grammatical gender of the objects matched, people were faster in their judgements than when there was a mismatch. This task didn’t require people to name these objects, and yet still they appear to use the arbitrary grammatical classification system of their native language.
Edward Segel and Lera Boroditsky found the influence of grammatical gender even outside the laboratory - in an encyclopedia of classical paintings. They looked at all the gendered depictions of naturally asexual concepts like love, justice, and time. They noticed that these asexual entities (e.g., time) tended to be personified by masculine characters if the grammatical gender was masculine in the painter’s language (e.g., French: ‘le temps’), and vice versa for female characters (e.g., German: ‘die Zeit’). The depicted gender agreed with grammatical gender in 78% of the cases for painters whose mother tongue was ‘gendered’, like Italian, French and German. On top of that, this effect was consistent even when only looking at those concepts with different grammatical genders in the studied languages.
These and similar studies have powerfully shown how a grammatical classification system for nouns affects the view language speakers have on the world. By forcing people to think in certain categories, general thinking habits appear to be affected. This illustrates quite nicely that thought is influenced by what you must say – rather than by what you can say. The grammatical gender effect on cognition highlights the fact that language is not an isolated skill but instead a central part of how the mind works.
Written by Richard Kunert and Gwilym Lockwood
Segel, E., & Boroditsky, L. (2011). Grammar in art. Frontiers in Psychology, 1,1. doi: 10.3389/fpsyg.2010.00244
Most people who try to learn a second language, or who interact with non-native speakers notice that the way non-native speakers speak their second language is influenced by their native language. They are likely to have a foreign accent, and they might use inappropriate words or an incorrect grammatical structure, because those words or that structure are used that way in their native language. A lesser known yet common phenomenon is the influence of a foreign language we learn on our native language.
People who start using a foreign language regularly (for example, after moving to a different country) often find themselves struggling to recall words when using their native language. Other common influences are the borrowing of words or collocations (two or more words that often go together). For example, Dutch non-native English speakers might insert English words, for which there is no literal translation, such as native, into a Dutch conversation. Or they may find themselves using a literal translation of the collocation ‘taking a picture’ while speaking in their native language, even if their native language does not use the verb take to express this action. Studies from the past couple of decades show that people show such an influence at all linguistic levels - as described above, they may borrow words or expressions from their second language, but they might also borrow grammatical structures or develop a non-native accent in their own native language.
In general, research has shown that all the languages we speak are always co-activated. This means that when a Dutch person speaks German, not only his German but also his Dutch, as well as any other language that person speaks, are automatically activated at the same time. This co-activation likely promotes cross-linguistic influence.
So will learning a foreign language necessarily influence one's native language at all linguistic levels? To a degree, but there are large individual differences. The influence is larger the more dominant the use of the foreign language is, and in particular if it is regularly used with native speakers of that language (as when moving to a foreign country). The influence also increases with time, so immigrants, for example, are likely to show more influence after 20 years abroad than after 2, although there is also a burst of influence in the initial period of using a foreign language regularly. Some studies also suggest that differences among people in certain cognitive abilities, like the ability to suppress irrelevant information, affect the magnitude of the influence of the second language on the native language. It is important to note though that some of these influences are relatively minor, and might not even be detectable in ordinary communication.
By Shiri Lev-Ari and Hans Rutger Bosker
Cook, V. (Ed.). (2003). Effects of the second language on the first. Clevedon: Multilingual Matters.
No language has a spelling system (or orthography) which absolutely and completely represents the sounds of that language, but some are definitely better than others. Italian has a shallow orthography, which means that the spelling of the words represent the sounds of Italian quite well (although Sicilian, Sardinian, and Neapolitan Italian speakers may disagree), while English has a deep orthography, which means that spelling and pronunciation don't match so well.
Italian is consistent for two main reasons. Firstly, the Accademia della Crusca was established in 1583 and has spent several centuries since regulating the Italian language; the existence of such an academy has enabled wide-ranging and effective spelling consistency. Secondly, Standard Italian only has five vowels; a, i, u, e, and o, which makes it much easier to distinguish between them on paper. Other examples of languages with five vowel systems are Spanish and Japanese, both of which also have shallow orthographies. Japanese is an interesting case; some words are written using the Japanese characters, which accurately represent the sound of the words, but other words are written with adapted Chinese characters, which represent the meaning of the words and don't represent the sound at all.
French has a deep orthography, but in one direction; while one sound can be written several different ways, there tends to be one specific way of pronouncing a particular vowel or combination of vowels. For example, the sound [o] can be written au, eau, or o, as in haut, oiseau, and mot; however, the spelling eau can only be pronounced as [o].
English, meanwhile, has a very deep orthography, and has happily resisted spelling reform for centuries (interestingly enough, this is not the case in the USA; Noah Webster's American Dictionary of the English Language introduced a successful modern spelling reform programme… or program). One obvious reason is the lack of a formal academy for the English language - English speakers are rather laissez-faire (certainly laissez-faire enough to use French to describe English speakers' attitudes towards English) - but there are several other reasons too.
Formed out of a melting pot of European languages - a dab of Latin and Greek here, a pinch of Celtic and French there, a fair old chunk of German, and a few handfuls of Norse - English has a long and complicated history. Some spelling irregularities in English reflect the original etymology of the words. The unpronounced b in doubt and debt harks back to their Latin roots, dubitare and debitum, while the pronunciation of ce- as "se-" in centre, certain, and celebrity is due to the influence of French (and send and sell are not "cend" and "cell" because they are Germanic in origin).
All languages change over time, but English had a particularly dramatic set of changes to the sound of its vowels in the middle ages known as the Great Vowel Shift. The early and middle phases of the Great Vowel Shift coincided with the invention of the printing press, which helped to freeze the English spelling system at that point; then, the sounds changed but the spellings didn't, meaning that Modern English spells many words the way they were pronounced 500 years ago. This means that the Shakespeare's plays were originally pronounced very differently from modern English, but the spelling is almost exactly the same. Moreover, the challenge of making the sounds of English match the spelling of English is harder because of the sheer number of vowels. Depending on their dialect, English speakers can have as many as 22 separate vowel sounds, but only the letters a, i, u, e, o, and y to represent them; it's no wonder that so many competing combinations of letters were created.
Deep orthography makes learning to read more difficult, as a native speaker and as a second language learner. Despite this, many people are resistant to spelling reform because the benefits may not make up for the loss of linguistic history. The English may love regularity when it comes to queuing and tea, but not when it comes to orthography.
By Gwilym Lockwood & Flora Vanlangendonck
Original Pronunciation in Shakespeare: http://www.youtube.com/watch?v=gPlpphT7n9s
Children usually start babbling at an age of two or three months – first they babble vowels, later consonants and finally, between an age of seven and eleven months, they produce word-like sounds. Babbling is basically used by children to explore how their speech apparatus works, how they can produce different sounds. Along with the production of word-like sounds comes the ability to extract words from a speech input. These are important steps towards the infant’s first words, which are usually produced at an age of around 12 months.
Simple one-word utterances are followed by two-word utterances during the second half of the children’s second year, in which one can already observe grammar. Children growing up learning a language like Dutch or German (subject-object-verb order in subordinate clauses, which are considered to have a stable order) or English (subject-verb-object order) produce their two-sentences in a subject-verb order, such as “I eat”, while learners of languages such as Arabic or Irish (languages with a verb–subject–object order) produce sentences like “eat I”. From then on, there is a rapid acceleration in the infant’s vocabulary growth as sentences also contain more words and get more complex. Grammar is said to have developed by an age of four or five years and by then, children are basically considered linguistic adults. The age at which children acquire these skills may vary strongly from one infant to another and the order may also vary depending on the linguistic environment in which the children grow up. But by the age of four or five, all healthy children will have acquired language. The development of language correlates with different processes in the brain, such as the formation of connective pathways, the increase of metabolic activity in different areas of the brain and myelination (the production of myelin sheaths that form a layer around the axon of a neuron and are essential for proper functioning of the nervous system).
By Mariella Paul and Antje Meyer
Bates E, Thal D, Finlay BL, Clancy, B (1999) Early Language Development and its Neural Correlates, in I. Rapin & S. Segalowitz (Eds.), Handbook of Neuropsychology, Vol. 6, Child Neurology (2nd edition). Amsterdam: Elsevier. (link)
Homophones are words that sound the same but have two or more distinct meanings. This phenomenon occurs in all spoken languages. Take for instance the English words FLOWER and FLOUR. These words sound the same, even though they differ in several letters when written down (therefore called heterographic homophones). Other homophones sound the same and also look the same, such as the words BANK (bench/river bed) and BANK (financial institution) in both Dutch and English. Such words are sometimes called homographic homophones. Words with very similar sounds but different meanings also exist between languages. An example is the word WIE, meaning 'who' in Dutch, but 'how' in German.
One might think that homophones would create serious problems for the hearer or listener. How can one possibly know what a speaker means when she says a sentence like "I hate the mouse"? Indeed, many studies have shown that listeners are a little slower to understand ambiguous words than unambiguous ones. However, in most cases, it is evident from the context what the intended meaning is. The above sentence might for example appear in the contexts of "I don't mind most of my daughter's pets, but I hate the mouse" or "I love my new computer, but I hate the mouse". People normally figure out the intended meaning so quickly, that they don't even perceive the alternative. Preceding linguistic context and common world knowledge thus help us in understanding a speaker’s intended message.
Why do homophones exist? It seems much less confusing to have separate sounds for separate concepts. Linguists take sound change as an important factor that can lead to the existence of homophones. For instance, in the early 18th century the first letter of the English word KNIGHT was no longer pronounced, making it a homophone with the word NIGHT. Also language contact creates homophones. The English word DATE was relatively recently adopted into Dutch, becoming a homophone with the already existing word DEED. Some changes over time thus create new homophones, whereas other changes undo the homophonic status of a word. Now the Dutch verb form ZOUDT (would) is no longer commonly used, the similarly sounding noun ZOUT (salt) is losing its homophonic status.
Finally, a particularly nice characteristic of homophones is that they are often used in puns or as stylistic elements in literary texts. In Shakespeare’s Romeo and Juliet (Act I, Scene IV, line 13-16), Romeo for instances uses a homophone when he refrains from following up his friend Mercutio‘s advice to dance:
|Mercutio:||Nay, gentle Romeo, we must have you dance.|
|Romeo:||Not I, believe me: you have dancing shoes|
|With nimble soles: I have a soul of lead|
|So stakes me to the ground I cannot move.|
Among other things, it is such elegant use of homophones that has led to Shakespeare’s literary success.
Bloomfield, L. (1933). Language. New York: Henry Holt and Company.
Cutler, A., & Van Donselaar, W. (2001). Voornaam is not (really) a homophone: Lexical prosody and lexical access in Dutch. Language and speech, 44(2), 171-195. (link)
Rodd, J., Gaskell, G., & Marslen-Wilson, W. (2002). Making sense of semantic ambiguity: Semantic competition in lexical access. Journal of Memory and Language, 46(2), 245-266. (link)
Tabossi, P. (1988). Accessing lexical ambiguity in different types of sentential contexts. Journal of Memory and Language, 27(3), 324-340. (link)
When a child has significant difficulties in learning to read and/or spell despite normal general intelligence and overall sensory abilities, then he or she may be diagnosed with developmental dyslexia. This condition was first described in the 1890's and referred to as 'congenital word blindness', because it was thought to result from problems with processing of visual symbols. Over the years it has become clear that visual deficits are not the core feature for most people with dyslexia. In many cases, it seems that subtle underlying difficulties with aspects of language could be contributing. To learn to read, a child needs to understand the way that words are made up by their individual units (phonemes), and must become adept at matching those phonemes to arbitrary written symbols (graphemes). Although the overall language proficiency of people with dyslexia usually appears normal, they often perform poorly on tests that involve manipulations of phonemes and processing of phonology, even when this does not involve any reading or writing.
Since dyslexia is defined as a failure to read, without being explained by an obvious known cause, it is possible that this is not one single syndrome, but instead represents a cluster of different disorders, involving distinct mechanisms. However, it has proved hard to clearly separate dyslexia out into subtypes. Studies have uncovered quite a few convincing behavioural markers (not only phonological deficits) that tend to be associated with the reading problems, and there is a lot of debate about how these features fit together into a coherent account. To give just one example, many people with dyslexia are less accurate when asked to rapidly name a visually-presented series of objects or colours. Some researchers now believe that dyslexia results from the convergence of several different cognitive deficits, co-occurring in the same person.
It is well established that dyslexia clusters in families and that inherited factors must play a substantial role in susceptibility. Nevertheless, there is no doubt that the genetic basis is complex and heterogeneous, involving multiple different genes of small effect size, interacting with the environment. Genetic mapping efforts have already enabled researchers to pinpoint a number of interesting candidate genes, such as DYX1C1, KIAA0319, DCDC2, and ROBO1, and with dramatic advances in DNA sequencing technology there is much promise for discovering others in the coming years. The neurobiological mechanisms that go awry in dyslexia are largely unknown. A prominent theory posits disruptions of a process in early development, a process in which brain cells move towards their final locations, known as neuronal migration. Indirect supporting evidence for this hypothesis comes from studies of post-mortem brain material in humans and investigations of functions of some candidate genes in rats. But there are still many open questions that need to be answered before we can fully understand the causal mechanisms that lead to this elusive syndrome.
By Simon Fisher
Carrion-Castillo, A., Franke, B., & Fisher, S. E. (2013). Molecular genetics of dyslexia: an overview. Dyslexia, 19, 214–240. (link)
Demonet, J. F., Taylor, M. J., & Chaix, Y. (2004). Developmental dyslexia. Lancet, 63, 1451–1460 (link)
Fisher, S. E. & Francks, C. (2006). Genes, cognition and dyslexia: learning to read the genome. Trends in Cognitive Science, 10, 250-257.(link)
When we tell people we investigate the sign languages of deaf people, or when people see us signing, they often ask us whether sign language is universal. The answer is that nearly every country is home to at least one national sign language which does not follow the structure of the dominant spoken language used in that country. So British Sign Language and American Sign Language in fact look really different. Chinese Sign Language and Sign Language of the Netherlands, for example, also have distinct vocabularies, deploy different fingerspelling systems, and have their own set of grammatical rules. At the same time, a Chinese and Dutch deaf person, who do not have any shared language, manage to bridge this language gap with relative ease when meeting for the first time.
This kind of ad hoc communication is also known as cross-signing. In collaboration with the International Institute for Sign Languages and Deaf Studies - iSLanDS, we are conducting a study of how cross-signing emerges among signers of varying countries for the first time. The recordings include signers from countries such as South Korea, Uzbekistan, and Indonesia. Our initial findings are that the communicative success in these dialogues is linked to the interlocutors’ ability to accommodate to their communicative partner by using creative signs that are not found in their own sign languages. This linguistic creativity often capitalizes on the depictive properties of visual signs (e.g. gesturing a circle in the air to talk about a round object, or representing a man by indicating a moustache), but also on general principles of interaction, e.g. repeating a sign to request more information.
Cross-signing is distinct from International Sign, which is used at international deaf meetings such as the World Federation of the Deaf (WFD) congress or the Deaflympics. International Sign is strongly influenced by signs from American Sign Language and is usually used to present in front of international deaf audiences who are familiar with its lexicon. Cross-signing, on the other hand, emerges in interaction among signers without knowledge of each other's native sign languages.
By Connie de Vos, Kang-Suk Byun & Elizabeth Manrique
Suggestion for further reading:
Information on differences and commonalities between different sign languages, and between spoken and signed languages by the World Federation of the Deaf: (link)
Mesch, J. (2010). Perspectives on the Concept and Definition of International Sign. World Federation of the Deaf. (link)
Supalla, T., & Webb, R. (1995). The grammar of International Sign: A new look at pidgin languages. In K. Emory and J. Reilly (Eds.), Sign, Gesture and Space. (pp.333-352) Mahwah, NJ: Lawrence Erlbaum.
Our bodies constantly communicate in various ways. One form of bodily communication is what we traditionally consider as “body-language”. In the context of social interactions, our body expresses attitudes and emotions influenced by the dynamics of the interaction, interpersonal relations and personality (see also answer to the question "What is body language?"). These bodily messages are often considered to be transmitted unwittingly. Because of this, it would be difficult to teach a universal shorthand suitable for expressing the kind of things considered to be body language; however, at least within one culture, there seems to be a great deal of commonality in how individuals express attitudes and emotions through their body.
Another form of bodily communication is the use of co-speech gesture. Co-speech gestures are movements of the hands, arms, and occasionally other body parts that interlocutors produce while talking. Importantly, co-speech gestures express meaning that is closely tied to the meaning communicated by the speech that they accompany, thus contrasting crucially with the kind of signals that constitute ‘body language’. Because speech and gesture are so tightly intertwined, co-speech gestures are only very rarely fully interpretable in the absence of speech. As such, co-speech gestures do not help communication much if interlocutors do not speak the same language. Further, co-speech gestures are shaped by the culture in which they are embedded (and therefore differ between cultures, at least to some extent), and there is no standard of form that applies to these gestures even within one culture; they are each person’s idiosyncratic creation at the moment of speaking (although there is overlap in how different people use gesture to represent meaning, there are also substantial differences). As such, co-speech gestures are the opposite of a universal nonverbal ‘code’ that could be used when interlocutors do not share a spoken language.
What people often tend to resort to when trying to communicate without a shared language are pantomimic gestures, or pantomimes. These gestures are highly iconic in nature (like some iconic co-speech gestures are), meaning that they map onto structures in the world around us. Even when produced while speaking, these gestures are designed to be understandable in the absence of speech. Without a shared spoken language, they are therefore more informative than co-speech gestures. As long as we share knowledge about the world around us - for instance about actions, objects, and their spatial relations - these gestures tend to be communicative, even if we don’t share the same language.
An important distinction has to be made between these pantomimic gestures that can communicate information in the absence of speech and sign languages. In contrast to pantomimes, sign languages of deaf communities are fully-fledged languages consisting of conventionalised meanings of individual manual forms and movements which equate to the components that constitute spoken language. There is not one universal sign language: different communities have different sign languages (Dutch, German, British, French or Turkish sign languages being a small number of examples).
Kendon, A. (2004). Gesture: Visible action as utterance. Cambridge University Press.(link)
McNeill, D. (1992). Hand and mind: What gestures reveal about thought. Chicago University press. (link book review)
Language appears to be unique in the natural world, a defining feature of the human condition. Although other species have complex communication systems of their own, even our closest living primate relatives do not speak, in part because they lack sufficient voluntary control of their vocalizations. After years of intensive tuition, some chimpanzees and bonobos have been able to acquire a rudimentary sign language. But still the skills of these exceptional cases have not come close to those of a typical human toddler, who will spontaneously use the generative power of language to express thoughts and ideas about present, past and future.
It is certain that genes are important for explaining this enigma. But, there is actually no such thing as a "language gene" or "gene for language", as in a special gene with the designated job of providing us with the unique skills in question. Genes do not specify cognitive or behavioural outputs; they contain the information for building proteins which carry out functions inside cells of the body. Some of these proteins have significant effects on the properties of brain cells, for example by influencing how they divide, grow and make connections with other brain cells that in turn are responsible for how the brain operates, including producing and understanding language. So, it is feasible that evolutionary changes in certain genes had impacts on the wiring of human brain circuits, and thereby played roles in the emergence of spoken language. Crucially, this might have depended on alterations in multiple genes, not just a single magic bullet, and there is no reason to think that the genes themselves should have appeared "out of the blue" in our species.
There is strong biological evidence that human linguistic capacities rely on modifications of genetic pathways that have a much deeper evolutionary history. A compelling argument comes from studies of FOXP2 (a gene that has often been misrepresented in the media as the mythical "language gene"). It is true that FOXP2 is relevant for language – its role in human language was originally discovered because rare mutations that disrupt it cause a severe speech and language disorder. But FOXP2 is not unique to humans. Quite the opposite, versions of this gene are found in remarkably similar forms in a great many vertebrate species (including primates, rodents, birds, reptiles and fish) and it seems to be active in corresponding parts of the brain in these different animals. For example, songbirds have their own version of FOXP2 which helps them learn to sing. In-depth studies of versions of the gene in multiple species indicate it plays roles in the ways that brain cells wire together. Intriguingly, while it has been around for many millions of years in evolutionary history, without changing very much, there have been at least two small but interesting alterations of FOXP2 that occurred on the branch that led to humans, after we split off from chimpanzees and bonobos. Scientists are now studying those changes to find out how they might have impacted the development of human brain circuits, as one piece of the jigsaw of our language origins.
By Simon Fisher, Katerina Kucera & Katherine Cronin
Revisiting Fox and the Origins of Language (link)
Fisher S.E. & Marcus G.F. (2006) The eloquent ape: genes, brains and the evolution of language. Nature Reviews Genetics, 7, 9-20. (link)
Fisher, S.E. & Ridley, M. (2013). Culture, genes, and the human revolution. Science, 340, 929-930.(link)
Learning a new language is not easy, largely because of the heavy burden on memory. A universal “best practice” for everyone probably does not exist, but vocabulary memorization can become more efficient with the help of some good strategies.
The conventional method of building up a vocabulary from scratch is remembering the words in the target language by translating to one’s own language. This might be advantageous if the new language and the person’s mother tongue are related, such as Dutch and German, but proves too indirect for unrelated languages, as for English and Chinese for instance. The learning process becomes more efficient when the translation step is removed and the new words are directly linked to the actual objects and actions. Many highly skilled second language speakers frequently run into words whose exact translations do not even exist in their native language, demonstrating that those words were not learned by translation, but from context in the new language.
To skip the translation step early in the language learning process, it is helpful to visualize what it was like when these words were said in in the learner’s native language and link the response to the new words. The idea is to mimic how a child learns a new language. Another way to build a vocabulary quicker is by grouping things that are conceptually related and practicing them at the same time. For example, naming things and events related to transportation as one is getting home from work, or naming objects on the dinner table. The key is to make the new language making “direct sense” instead of trying to understand it through a familiar media such as the native language. In a bit more advanced stage of building a vocabulary, one can use a dictionary in the target language, such as Thesaurus in English, to find the meaning of new words, rather than a language-to-language dictionary.
Employing a method called “Spaced Learning” may also be beneficial. Spaced Learning is a timed routine, in which new material (such as a set of new words in a studied language) is introduced, reviewed, and practiced in three timed blocks with two 10 minute breaks. It is important that distractor activities that are completely unrelated to the studied material, such as physical exercises, are performed during those breaks. It has been demonstrated in laboratory experiments that such repeated stimuli, separated by timed breaks, can initiate long-term connections between neurons in the brain and result in long-term memory encoding. These processes occur in minutes, and have been observed not only in humans, but also in other species.
It is inevitable to forget when we are learning new things and so is making mistakes. The more you use the words that you are learning, the better you will remember them.
Written by Sylvia Chen & Katerina Kucera
Kelly P. & Whatson T. (2013). Making long-term memories in minutes: a spaced learning pattern from memory research in education. Frontiers of Human Neuroscience, 7, 589. (link)
A basic assumption of language change is that if two linguistic groups are isolated from each other, then their languages will become more different over time. If the groups come into contact again, then they may become more similar to each other, for instance by borrowing parts of each others’ languages. This is the case in many parts of the world where aspects of language have been borrowed into many different languages. A recent example is the word ‘internet’ which has been adopted by many languages. However, before global communication was possible, borrowing was restricted to languages in the same geographic area. This might cause languages in the same area to become more similar to each other. Linguists call these ‘areal effects’, and they are influenced by factors such as migration or a group being more powerful or prestigious.
Another possibility that would make languages more similar is if there were certain words or linguistic rules were easier to learn or somehow ‘fit’ the human brain better. This is similar to different biological species sharing similar traits, such as birds, bats and some dinosaurs evolving wings. In this case, languages that were even on opposite sides of the world might change to become more similar. Researchers have noticed some features in words across languages which seem to make a direct link between a word's sound and its meaning (this is known as sound-symbolism). For instance, words for ‘nose’ tend to involve sounds made with the nose such as [n]. Languages all over the world use a similar word to question what someone else has said – ‘huh?’ or ‘what?’ – possibly because it’s a short, question-like word that’s effective at getting people’s attention.
Telling the difference between these effects and the similarities caused by borrowing is difficult because their effects can be similar. One of the goals of evolutionary linguistics is to find ways of teasing these effects apart.
This question strikes at the heart of linguistic research, because it involves asking whether there are limits on what languages can be like. Some of the first modern linguistic theories suggested that there were limits on how different languages could be because of biological constraints. More recently, field linguists have been documenting many languages that show a huge diversity in their sounds, words and rules. It may be the case that for every way in which languages become more similar, there is another way in which they become more different at the same time.
By Seán Roberts & Gwilym Lockwood
Nettle, D. (1999). Linguistic Diversity. Oxford: Oxford University Press.
Dingemanse, M., Torreira, F., & Enfield, N. J. (in press). Is “Huh?” a universal word? Conversational infrastructure and the convergent evolution of linguistic items. PLoS One. (link)
Dunn, M., Greenhill, S. J., Levinson, S. C., & Gray, R. D. (2011). Evolved structure of language shows lineage-specific trends in word-order universals. Nature, 473, 79-82. (link)
|Image: Eoin Dubsky|
When researchers talk about "maternal language", they talk about child-directed language. Modern research on the way in which caregivers talk to children started in the late seventies. Scholars who study language acquisition were interested in understanding how language learning is influenced by the way caregivers talk to their children. Since the main caregivers were usually mothers, most first studies focused on maternal language (often called Motherese), which was usually described as having higher tones, a wider tonal range and a simpler vocabulary. However, we now understand that the ways in which mothers modify their speech for their children are extremely variable. For example, some mothers use a wider tone range than usual, but many mothers use nearly the same tone range they use with adults. Mothers’ speech also changes to adapt to their children’s language abilities, so maternal speech can be very different from one month to the next.
Also, in some cultures mothers use little or no “baby talk” when interacting with their children. Because of this, it is impossible to define a universal maternal language. It’s also important to remember that fathers, grandparents, nannies, older siblings, cousins, and even non-family members often modify their speech when they address a young child. For this reason, researchers today prefer to use the label “child-directed speech” instead of "maternal language" or “motherese”. Some recent studies also show that we do more than just modify our speech for children: we also modify our hand gestures and demonstrative actions, usually making them slower, bigger, and closer to the child. In sum, child-directed language seems to be only one aspect of a broader communicative phenomenon between caregivers and children.
By Emanuela Campisi, Marisa Casillas & Elma Hilbrink
Fernald, A., Taeschner, T., Dunn, J., Papousek, M., de Boysson-Bardies, B., & Fukui, I. (1989). A cross-language study of prosodic modifications in mothers' and fathers' speech to preverbal infants. Journal of Child Language, 16, 477–501.
Rowe, M. L. (2008). Child-directed speech: relation to socioeconomic status, knowledge of child development and child vocabulary skill. Journal of Child Language, 35, 185–205.
|Image: Ars Electronica|
Our current understanding of the human brain holds that it is an information processor, comparable to a computer that enables us to cope with an extremely challenging environment. As such, it would be, theoretically, possible to extract the information the brain contains, through what is known as a `brain-computer interface’ (BCI): By measuring neural activity and applying sophisticated, self-learning computer algorithms, researchers are now able to extract and reproduce, for instance, image data that is being processed by the visual cortex. Such methods allow researchers to take a snap-shot of what someone is actually seeing though his or her own eyes.
With regards to language, recent studies have shown promising results as well. Brain imaging data was used to establish what kind of stimuli (spoken, written, photographs, natural sounds) people were observing in a task where they had to discriminate between animals and tools; researchers were able to determine the kind of stimulus by merely looking at a person’s brain activity. More invasive recordings have also demonstrated that we can extract information about speech sounds people are hearing and producing. This body of research suggests that it may be possible in the not-too-distant future to develop a neural prosthetic that would allow people to generate speech in a computer by just thinking about speaking.
When discussing voiceless communication, however, the application of BCI is still particularly challenging for a few reasons. First, human communication is substantially more than simply speaking and listening, and there are a whole host of other signals that we use to communicate (e.g., gestures, facial expression and other forms of non-verbal communication). Secondly, while BCIs to date have focused on extracting information from the brain (e.g., to-be-articulated syllables), inserting information into the brain would require more direct manipulation of brain activity. Cochlear implants represent one approach where we have been able to directly present incoming auditory information to auditory nerve cells in the ear in order to help congenitally deaf individuals hear. Such stimulation, however, still does not amount to directly stimulating the brain in the full richness needed for everyday communication. So, while both `reading’ and `writing’ language data from and to the brain with the aim to enable voiceless communication is theoretically possible, both are -extremely- challenging and would require tremendous advances in fields such as linguistics, neuroscience, computer science and statistics.
By Dan Acheson & Rick Janssen
 Schoenmakers, S., Barth, M., Heskes, T., & van Gerven, M. A. J. (2013). Linear Reconstruction of Perceived Images from Human Brain Activity. NeuroImage, 83, 951-961.(link)
 Simanova, I., Hagoort, P., Oostenveld, R., van Gerven, M.A.J. (2012). Modality-Independent Decoding of Semantic Information from the Human Brain. Cerebral Cortex, doi:10.1093/cercor/bhs324.(link)
 Chang, E. F., Niziolek, C. A., Knight, R. T., Nagarajan, S. S., & Houde, J. F. (2013). Human cortical sensorimotor network underlying feedback control of vocal pitch.Proceedings of the National Academy of Sciences, 110, 2653-2658. (link)
 Chang, E. F., Rieger, J. W., Johnson, K., Berger, M. S., Barbaro, N. M., & Knight, R. T. (2010). Categorical speech representation in human superior temporal gyrus. Nature neuroscience, 13, 1428-1432.(link)
 Bouchard, K. E., Mesgarani, N., Johnson, K., & Chang, E. F. (2013). Functional organization of human sensorimotor cortex for speech articulation. Nature, 495, 327-332. (link)
|Image: Duncan Rawlinson|
The terms "surface layer" and "deep layer" refer to different levels that information goes through in the language production system. For example, imagine that you see a dog chasing a mailman. When you encode this information, you create a representation that includes three pieces of information: a dog, a mailman, and the action chasing. This information exists in the mind of the speaker as a "deep" structure. If you want to express this information linguistically, you can, for example, produce a sentence like "The dog is chasing the mailman." This is the "surface" layer: it consists of the words and sounds produced by a speaker (or writer) and perceived by a listener (or reader). You can also produce a sentence like "The mailman is being chased by a dog" to describe the same event -- here, the order in which you mention the two characters (the "surface" layer) is different from the first sentence, but both sentences are derived from the same "deep" representation. Linguists propose that you can perform movement operations to transform the information encoded in the "deep" layer into the "surface" layer, and refer to these movement operations as linguistic rules. Linguistic rules are part of the grammar of a language and must be learned by speakers in order to produce grammatically correct sentences.
Rules exist for different types of utterances. Other examples of rules, or movement operations between "deep" and "surface" layers, include declarative sentences (You have a dog) and their corresponding interrogative sentences (Do you have a dog?). Here, the movement operations include switching the order of the first two words of the sentence.
By Gwilym Lockwood & Agnieszka Konopka
Chomsky, N. (1957). Syntactic Structures. Mouton.
Chomsky, N. (1965). Aspects of the Theory of Syntax. MIT Press.
There are actually two questions hidden in this one. For a clear answer we can best treat them separately:
(1) Why do we cry when we are in sudden pain?
(2) Why do we say ‘ouch’ and not something else?
With regard to the first question, let’s start by noting that we’re not alone: a lot of other animals also cry when in pain. Why? Darwin, who wrote a book in 1872 about emotions in humans and animals, thought it had to do with the fact that most animals experience strong muscle contractions when they’re in pain — a ritualized version of quickly escaping a painful stimulus. Still, why should this be accompanied with an actual cry? Research has since shown that pain cries also have communicative functions in the animal kingdom: for example to alarm others that there is danger, to call for help, or to prompt caring behavior. That last function already starts in the first seconds of our life, when we cry and our mothers caringly take us in their arms. Human babies, and in fact the young of many mammals, are born with a small repertoire of instinctive cries. The pain cry in this repertoire is clearly recognizable: it has a sudden start, a high intensity and a relatively short duration. Here we already see the contours of our ‘ouch’. Which brings us to the second part of our question.
Why do we say ‘ouch’ and not something else? Let us first take a more critical look at the question. Do we never shout anything different? Do we utter a neat ‘ouch’ when we hit our thumb with a hammer or could it also be ‘aaaah!’? In reality there is a lot of variation. Still, the variation is not endless: no one shouts out ‘bibibibi’ or ‘vuuuuu’ when in pain. Put differently, pain cries are variations on a theme — a theme that starts with an ‘aa’ sound because of the shape of our trachea when our mouth is wide open, and then sounds like ‘ow’ or ‘ouch’ when our mouth quickly returns to being closed. The word ‘ouch’ is a perfectly fine summary of this theme — and here we touch on an important function of language. Language helps us share experiences which are never exactly the same and yet can be categorised as similar. That is useful, because if we want to talk about “someone crying ouch” we don’t always need or want to imitate the cry. In that sense ‘ouch’ is more than just a pain cry: it is a proper word.
Does this all imply that ‘ouch’ may be the same in every language? Almost, but not quite, since each language will use its own inventory of sounds to describe a cry of pain. In German and Dutch it is ‘au’, in Israeli it is ‘oi’, and in Spanish it’s ‘ay’ — at least that is how Byington described it in 1942 in one of the first comparisons of these words.
Each of us is born with a repertoire of instinctive cries and then learns a language in addition to it. This enables us to do more than scream and cry: we can also talk about it. Which is a good thing, for otherwise this answer could not have been written.
Imagine if someone told you to write a message to a friend using only number and punctuation symbols. Perhaps you could decide on a way to use these signs to express the sounds of your language. But if you wrote a message using your new ‘alphabet’, how would your friend be able to decode it? The two of you would need to have already agreed on a shared way to use numbers and punctuation signs to write words. That is, you would need to have designed a shared orthography.
Like a key to a code, an orthography is a standard way of linking the symbols of an alphabet (or other script) to the sounds of a language. Having an orthography means that speakers of the same language can communicate with each other using writing. The shapes of the symbols that we choose don’t matter - what is important is that we understand how those symbols (graphemes) represent the sounds (phonemes) that we use in our spoken language.
Even languages that use the same script have different sounds, and thus different orthographies. The Dutch and English orthographies use symbols of the Roman alphabet, and share many sound-grapheme correspondences - for example, the letters ‘l’ and the sounds ‘l’ in the English word light and the Dutch word licht are pretty much the same. However, the letter ‘g’ stands for a different sound in the English orthography (e.g., in good) than it does in the Dutch orthography (goed). Readers of Dutch or English need to know this in order to link the written word forms to the spoken forms.
There are different ideas about what makes a ‘good’ orthography. General principles are that it should consistently represent all and only the distinctive sound contrasts in the language, with the fewest possible symbols and conventions. However, few established orthographies stick to this ideal - you only have to look at English spelling to realise that! Compare, for instance, the pronunciations of pint and print. Quirks and inconsistencies in orthographies can also have their own advantages, such as preserving historical information, highlighting cultural affiliations, and supporting dialect variation.
Written by Lila San Roque & Antje Meyer
|Image: Jurriaan Persyn|
Unlike most of the other answers here, language itself doesn't really come into it; English is perceived by many people as the universal language because of the former influence of the British Empire and the current influence of American political and economic hegemony.
It is possible to try giving a strictly linguistic explanation; it could be that English is a simple language which is relatively easy to pick up. English has no noun genders, no complicated morphology, no tone system, it is written in the Roman alphabet which is pretty good at accurately mapping sounds to symbols, and the prevalence of English-language films, TV, and music makes it readily accessible and easy to practise. However, English also has an extensive vocabulary, a highly inconsistent spelling system, many irregular verbs, some problematic sounds such as "th", and a large inventory of vowels which can make it difficult for non-native speakers to tell words apart. Arguments about which languages are easy or difficult to learn are ultimately circular, as the perception of what is easy and what is difficult to learn depends on the person doing the learning.
The second explanation is historical. The UK was the first industrialised nation, and discovered that one of the advantages to this was that they could colonise the rest of the developing world far faster than other European countries could. The British Empire covered a quarter of the globe at its largest, including North America, the Caribbean, Australia, New Zealand, much of West and Southern Africa, South Asia, and parts of South-East Asia. The UK set up English-speaking systems of government, industry, and exploitation in these areas, which established English as the language of global power in the industrial era. The British Empire finally fell apart after the Second World War, but the 20th century saw the transfer of power from one English-speaking expansionist power to another. The USA's cultural, economic, political, and military domination of the 20th and 21st centuries has ensured that English remains the most important and influential global language. As the official language of business, science, diplomacy, communications, and IT (not to mention the main language of the most popular websites), this is unlikely to change any time soon.
Written by Gwilym Lockwood & Katrin Bangel
|Image: Alexandre Duret-Lutz|
Nobody knows! This is one of the biggest problems in the field of language evolution. Unlike stone tools or skeletons, the language that people use doesn't fossilise, so we can’t study it directly. We have examples of writing from over 6,000 years ago which can help us work out what languages were like relatively recently, but recent research suggests that people have been using languages very much like our own for maybe 500,000 years.
In order to get an idea of how languages might have evolved, researchers use model systems. These might include computational models, experiments with real people or the study of languages that have developed very recently such as sign languages or creoles. We can’t know for sure what the first languages would have sounded like, but we can work out how communication systems tend to evolve and develop.
For example, one series of experiments studied how communication systems develop by using a game like Pictionary: players had to communicate about meanings by drawing. They found that negotiation and feedback during communication was very important for a successful communication system to develop.
Another experiment used a chain of learners who had to learn an artificial language. The first learner had to learn a language with no rules, then teach the next person in the chain whatever they could remember. This second person tried to learn the language and then passed it on to the third person in the chain, just like a game of ‘broken telephone’. In this way, we can simulate the process of passing a language down through the generations, but now it only takes one afternoon rather than thousands of years. What researchers noticed is that the language changed from having a different word for every possible meaning to having smaller words that referred to parts of the meaning that could be put together. That is, a language with rules emerged.
So, early languages may have had a different word for every meaning and gradually broken them apart to create rules. However, other researchers think that languages started with small words for every meaning and gradually learned to stick them together. Some think that language evolved very suddenly, while others think it evolved slowly. Some think that parts of language evolved in many different places at the same time, and contact between groups brought the ideas together. Some think that early languages would have sounded more like music. There are many suggestions that the first languages were signed languages, and that spoken language developed later.
Researchers in this field have to find creative ways of studying the problem and integrate knowledge from many fields such as linguistics, psychology, biology, neuroscience and anthropology.
Written by Seán Roberts & Harald Hammarström
Experiments on language evolution:
Galantucci, B. (2005). An Experimental Study of the Emergence of Human Communication Systems. Cognitive Science, 29, 737-767. (link)
Hurford. J. (forthcoming). Origins of language: A slim guide. Oxford University Press.
Johansson. S. (2005), Origins of Language: Constraints on hypotheses. Amsterdam: John Benjamins.
Kirby, S., Cornish, H. & Smith, K. (2008). Cumulative cultural evolution in the laboratory: An experimental approach to the origins of structure in human language. Proceedings of the National Academy of Sciences of the United States of America, 105,10681–10686. (link)
There are many ways to manipulate someone else: focusing on positive words while hiding negative information (‘95% fat-free’ rather than ‘5% fat’), giving neutral information an emotional twist through language melody, hiding important information in sentence positions which people usually don’t pay much attention to, etc. Language offers many instruments for manipulation. In this answer I will focus on how metaphors can be used in this way.
One way to manipulate someone else is by talking about one topic openly while raising another topic below the radar. The effect of this strategy is that the listener’s judgment is directed in a suitable direction. Stanford Psychologists Paul Thibodeau and Lera Boroditsky have investigated how this can be done. They presented participants with crime statistics which were embedded in a text which treated crime either as a beast (‘preying on a town’, ‘lurking in the neighbourhood’) or instead as a virus (‘infecting a town’, ‘plaguing the neighbourhood’). When participants were asked what to do about the crime problem, those who were exposed to the idea of crime as a beast were more likely to suggest law enforcement actions such as to capture criminals, enforce the law or punish wrong doers. In contrast, participants who were presented with crime as a virus often opted for reform-measures such as to diagnose, treat or inoculate crime. Remarkably, the effect of embedding the same information in different contexts had a bigger effect than the usual factors associated with opinions on how to address crime. A conservative political affiliation or male gender biases people towards the enforcement side but not nearly as strongly as slipping in a beast-metaphor.
Why don’t people realize that metaphors should not be taken literally? Crime was never meant as an actual animal or an actual virus, of course. And people presumably understood that. However, in order to process the metaphors, first their primary meaning gets activated regardless of whether the context demands it or not. This can be shown with brain research. A team of Cambridge neuroscientists led by Veronique Boulenger presented participants with sentences such as ‘He grasped the idea’ and compared the activation pattern to sentences such as ‘He grasped the object.’ It turns out that hand-related brain areas were activated in both kinds of sentences even though grasping an idea has nothing to do with actual grasping.
Manipulating through language can thus exploit the fact that language users cannot help but interpret the primary meaning of words. This way, manipulating through language can work via the use of metaphors. This, in turn, can even be exploited to influence something as politically sensitive as how to fight crime.
Written by Richard Kunert & Diana Dimitrova
Boulenger, V., Hauk, O., & Pulvermüller, F. (2009). Grasping Ideas with the Motor System: Semantic Somatotopy in Idiom Comprehension. Cebrebral Cortex, 19, 1905-1914. (link)
Thibodeau, P.H., & Boroditsky, L. (2011). Metaphors We Think With: The Role of Metaphor in Reasoning. Plos One, 6, e16782. (link)
Speaking requires planning and executing rapid sequences of movements. Several muscle systems are involved in the production of speech sounds. Not only the tongue, lips and jaw, but also the larynx and respiration muscles work together in coordination when we speak. As for any other movement, motor planning and sensorimotor control are essential for speaking.
In children, a tight relation between fine motor skills and language proficiency has been demonstrated. That is why speech therapists encourage sensory rich activities like finger painting, water or sand play, and manipulations involving small objects (coloring, buttoning, etc.) in children with speech delays. Such activities help to form new neural connections that are necessary for planning movement sequences and controlling fine-grain movements. For the same reason hand exercises can be beneficial as a part of complex therapy for speech and language recovery after a stroke or brain damage, in cases when language problems are caused by impaired articulation or motor control.
Another connection between movement and language lies in the domain of co-speech gestures. People often gesture when they speak and understanding the gestures is important in order to grasp the speaker’s intended message. Gesture may become essential to communicate at all in situations where verbal language use is constrained (for example in a noisy environment, or when speakers of different languages communicate). Usually people are remarkably fluent in extracting intended meaning from one’s hands and body movement. Interestingly, recent research demonstrates that similar brain areas are involved in constructing the meaning from linguistic and gestural input. Finally, the sign languages that deaf individuals use to communicate show that the language itself can be manifested in body movements, e.g. hands, arms, and facial expressions.
Why a Long Island Speech Therapist Incorporates Movement and Sensory Activities into Speech Therapy Sessions (link)
McNeill, David (2012). How Language Began: Gesture and Speech in Human Evolution. New York, USA; United Kingdom: Cambridge University Press. (link)
The vast majority of speech sounds are produced by creating a stream of air which flows from the lungs through the mouth or nose. We use this stream of air to form specific sounds with our vocal folds and/or by changing the configuration of our mouths.
When we produce consonants, a constriction is made somewhere in the mouth, either by stopping the air stream entirely (for example with our lips when saying 'p' or with our tongues when saying 't') or by leaving a very narrow gap which makes the air hiss as it passes (for example with our lips and teeth when saying 'f' or with our tongues when saying 's').
We also use our vocal-folds to differentiate consonants. When we bring our vocal folds close together, the stream of air makes them vibrate, which sounds like a hum; when they are apart, they do not vibrate. You can feel this difference by putting your finger on your Adam's apple when you say 'ssss' and 'zzzz' - can you feel how 'zzzz' is voiced and 'ssss' is not voiced? When we produce vowels, we change the shape of our mouths by moving our tongues, lips and jaw.
The different shapes of the vocal tract act as different acoustic filters, altering the hum produced by the vocal cords in different ways. For example, we move our tongues right to the front of our mouths and make our lips wide to make an 'ie' sound, and we move our tongues to the back of our mouths and make our lips round to make an 'oe' sound. For an 'aaa' sound, we move our tongue to the bottom of our mouth, lower the jaw and open our lips wide.
Finally, there are other specific ways of creating speech sounds, such as moving the stream of air through the nose to create nasal sounds like 'm', or creating a small pressure vacuum with the tongue before releasing it with a sharp popping sound, which is how people produce click sounds in some African languages.
Written by Matthias Sjerps, Matthias Franken & Gwilym Lockwood
Ladefoged, P. (1996). Elements of acoustic phonetics (second ed.) (link)
|Image: Paul Sapiano|
People do all kinds of things while sleeping. They move around, mumble, laugh, and some also sometimes whisper or produce speech while asleep. Sleep-talking (or: somniloquy) happens at all ages and may occur during all phases of sleep. But what exactly is the difference between sleep-talking and normal everyday speech?
Sleep-talk contains more speech errors than everyday speech. For instance, sleep-talkers can have trouble retrieving a word (word finding problems) or switch individual sounds when producing a word (for example beatag instead of teabag). While this of course also occurs during normal speech, it happens more frequently during sleep. Sleep-talk sometimes resembles speech produced by aphasic patients. In addition, sleep-talk resembles the speech that is sometimes produced by patients suffering from schizophrenia, in that there is less of a connection between utterances, which may lead to relatively incoherent discourse. Finally, sleep-talk may be less well articulated (mumbling) than everyday speech and contain incomprehensible words or newly invented words (neologisms).
However, perhaps the most striking thing is the similarity between sleep-talk and speech produced when awake. People produce full sentences while sleeping and the grammatical structure of their utterance is often perfectly correct. There are even some anecdotal reports describing people that would be more eloquent and creative during sleep compared to being awake, for instance when speaking a second language.
Sleep-talking does not necessarily indicate a psychological disorder or psychopathology. However, it may co-occur with sleep-disorder syndromes such as somnambulism (walking around while sleeping). Also, people that have encountered a traumatic event (such as soldiers who have fought in a war) are found to talk more in their sleep than non-traumatized control subjects. Besides such environmental factors, it has been found that there is also a genetic component to sleep-talking. If your parents are regular sleep-talkers, there is a higher chance that you are a sleep-talker yourself as well.
In conclusion, in linguistic terms sleep-talk differs less from talking while being awake than one may suspect. The main difference boils down to the popular belief that we have less control over what we say during sleep than during the day. Or as The Romantics put it in their 1984 hit: "I hear the secrets that you keep; When you're talking in your sleep; and I know that I’m right, cause I hear it in the night". Whether this is really the case has not been researched scientifically.
Arkin, A. (1981). Sleep talking. Psychology and psychophysiology. Hillsdale, NJ: Lawrence Erlbaum Associates.
People who speak more than one language are generally referred to as being bilingual, but there are many ways in which a person can become bilingual. Some grow up learning two languages at the same time, others learn them sequentially so that the mother tongue is learned first and a second language is learned later. However, in both cases the person may be equally fluent in each of the two languages. As it turns out, it is more important for the brain how fluent the person is, than how or when the languages are learnt. Current brain research studies suggest that when a person is equally good in both languages the brain uses the same areas, with the same level of activation and for the same reasons in both languages.
In many cases bilinguals are not equally good at both languages. If one compares the brain activation when a person uses his or her native and second language and is not equally fluent in both languages, the second language usually activates the same general language areas as the native language but these areas are more active in the second language. The less fluent language may also recruit areas not related directly to language but to cognitive control and attention. This means that when something is more difficult more brain resources are needed to process it.
Many have asked us if languages with different structure are processed differently if they are spoken by the same person, as may be the case for a bilingual. Languages vary greatly in how they express the relations of words in a sentence (who did what to whom). Some languages like English, Dutch and Chinese change the position of words in a sentence (which we call ‘word order’) while other languages like Japanese and Korean include additional short words without meaning to express these relations (which we call ‘case particles’). If the two languages of the bilingual belong to different language families (like English and Japanese), it is conceivable that they are not processed similarly, even when the bilingual is equally fluent in both languages. Currently very few studies exist on this, but at least there is one study on native Chinese people and native Korean people who speak both English and Japanese as a second language. The study found that brain activations depended on how similar the grammars of the non-native language was to that of the native language. Using English as a second language activated the language system stronger for the Korean native group than for the Chinese native group, because English and Korean are more different. Using Japanese as a second language activated slightly more brain areas for Chinese native speakers than for Korean native speakers, also because the grammars of Chinese and Japanese differ more. An explanation may be that besides proficiency and age of acquisition, brain activations depend on the differences and similarities of the languages’ grammars. Sometimes reaching automaticity on the level of the brain may take longer than reaching a proficient behavioral performance, especially if the language has a very different structure from your native language. However, this remains an open question. It seems clear that whichever language is spoken, on the level of the brain the same regions underlie language processing.
Written by Annika Hulten & Diana Dimitrova
Abutalebi, J. (2008). Neural aspects of second language representation and language control. Acta Psychologica,128, 466-478.
Kotz, S. A. (2009). A critical review of ERP and fMRI evidence on L2 syntactic processing. Brain Language, 109, 68-74.
Jeong, H., Sugiura, M., Sassa, Y., Yokoyama, S., Horie, K., Sato, S., & Kawashima, R. (2007). Cross-linguistic influence on brain activation during second language processing: An fMRI study. Bilingualism Language and Cognition, 10(2), 175.
Children and adult language learning differ in a number of ways. Firstly, the brain of a child is still developing whereas the adult brain is fully mature. The child’s is therefore more flexible and the neurons in the language system can adapt in accordance to both a native and a foreign language. This is why children are often able to learn a language without an accent in the speech and are able to distinguish the phonology of their second language better. Adults on the other hand, have a mature brain and are able to use their experience when learning. Adults are better and faster at learning by analogy, and at learning abstract rules and applying them. In addition to these differences in their brain capacity, children and adults typically also learn a second language in very different environments. Adults and teens learn a new language in a formal school setting, children learn by immersion e.g. in language day-care.
In order to answer when one should learn a second language, one should first determine what one wants to optimize: the end result or time devoted to learning. As a rule of thumb, a language which is learned before the ages of 6-9 is typically learnt to a level where the speaker has no detectable accent and is very comfortable using it. However, this requires that a child spends a considerable amount of time hearing and using the novel language. On the other hand, if a language is learnt later, speakers may have a slight accent in their pronunciation but may otherwise reach a very proficient native-like level. Moreover, in proportion adults spend typically less time on the actual learning compared to small children learning a language. Of course adults can also learn a language through the environment in which case the time devoted to learning is similar to that of small children. Adults who learn this way typically nevertheless have an accent while otherwise reaching a good level. It should also be noted that while children are able to learn a language without an accent, both adults and children learn vocabulary equally well.
Especially for elderly learners the motivation to learn plays a crucial part in the quality of the outcome. In fact, even elderly people above 60 years can show good language learning results, which has been shown to also have a “protective” effect for memory diseases such as Alzheimer’s disease. Humans are naturally curious and people of all ages can be encouraged to learn languages and explore other cultures in doing so.
Written by Annika Hulten & Diana Dimitrova
Kuhl, P. K. (2010). Brain mechanisms in early language acquisition. Neuron, 67, 713-727. (link)
Rodríguez-Fornells, A., T. Cunillera, A. Mestres-Missé & R. de Diego-Balaguer (2009). Neurophysiological mechanisms involved in language learning in adults. Philosophical Transactions of the Royal Society of London. Series B, Biological Sciences, 364, 3711-3735. (link)
Yes. This has indeed been found. However, the answer is a bit different for children learning their mother tongue and individuals learning a foreign language.
What advantages have been found for children who attend music lessons? At age ten these children show a more mature brain response when processing grammar. At age six to nine they read better. At age four they remember words better and make better use of grammatical rules for word formation. On top of all that, infants aged around one year who were assigned to early active music experiences gesture more to communicate with others. So, the answer for children learning their native language seems to be yes: music lessons do indeed improve language use or its precursors.
However, when it comes to learning a foreign language, the music advantage seems to be more restricted to hearing and producing the basic sounds which make up a language. For example, it has been found that among adults or children without music training, those who have more musical talent are also better at perceiving and producing non-native linguistic sounds like Chinese tones. Furthermore, individuals with music training are better at discriminating and learning foreign speech sounds, while they are also better at detecting mistakes in the way foreign speech is pronounced. There is evidence showing that the musician’s brain encodes speech sounds more efficiently and that parts of the brain that are dedicated to sound processing have a different structure in musicians compared to non-musicians. To summarize, people who are good at music are better in learning to perceive and to produce foreign speech sounds.
Might it be that more intelligent children stay in music classes while the others drop out, thereby explaining these results? This is unlikely: studies where children were randomly placed in music or painting classes have found better language abilities after music training. However, a word of caution is in order. It is not clear what it is about music that improves language use. Is it the training of long concentration during music lessons? Is it the training of complex, structured sound patterns during music listening? Yes, music classes make people good at music and somewhat good at language. But it is not exactly clear why that is.
Asaridou, S. S., & McQueen, J. M. (2013). Speech and music shape the listening brain: Evidence for shared domain-general mechanisms. Frontiers in Psychology, 4, 321. (link)
Kraus, N., & Chandrasekaran, B. (2010). Music training for the development of auditory skills. Nature reviews. Neuroscience, 11, 599–605. (link)
Patel, A. D. (in press). Can nonlinguistic musical training change the way the brain processes speech? The expanded opera hypothesis. Hearing Research. http://dx.doi.org/10.1016/j.heares.2013.08.011 (link)
Unlike humans, apes lack the anatomical pre-requisites for verbal language production. The organs within the vocal tract, such as larynx muscles and vocal cords, cannot be moved as freely and coordinated as in humans, especially not at a comparable speed. For this reason, we cannot talk with apes in the first place, even though they certainly understand and imitate a lot! (Click here for an example.)
Still, the communicative behavior of apes shares many characteristics with human language. Like many other animals, apes have developed ways to communicate that seem to resemble verbal language. However, most researchers refuse to call animal sign systems 'language' in the human sense because they are fairly restricted on many levels. For example, communicative signs are limited in number and cannot be combined to create new meaning. Also, apes can only communicate about things in their immediate surroundings, unlike humans who can refer to the past, the future or places and objects which are not present in the here and now.
Another reason why apes cannot speak is because they lack the cognitive capacity necessary for complex communication processes. Humans for instance are able to combine a limited number of words in such a way that they can express an infinite number of messages. We can also sequence smaller units of information, put them in a meaningful order and the result is a story. The brains of apes do not provide the cognitive resources to process such amounts of information. Apes can learn hundreds of words, but fail to use them in a creative way to convey complex meaning and intentions.
Recent evidence however indicates that brain structures supporting language production in the human brain might not only have been present in the ancestors of today's humans but also in the ancestors of chimpanzees. It was found that an area of the brain involved in the planning and production of spoken and signed language in humans (Broca's area) plays a similar role in chimpanzee communication.
Written by Katrin Bangel and Franziska Hartung
More videos on animals and language (link)
Chimps May Have A 'Language-Ready' Brain (link)
Did Neandertals have language? (link)
Pinker, S. (1994). The language instinct: how the mind creates language. New York: William Morrow & Co.
Taglialatela J.P., Russell J.L., Schaeffer J.A., Hopkins W.D. (2008). Communicative Signaling Activates 'Broca's' Homolog in Chimpanzees. Current Biology, 18, 343-348. (link)
We learn our language, use it to refer to things that are in another time or place, combine familiar words to create entirely novel messages, and use it intentionally to inform others. To some extent, each of these characteristics can be found in the communication systems of other species as well:
Dolphins develop personalized whistles that indicate to others who is calling, and pygmy marmosets “babble” like human infants, slowly pruning their vocal repertoire to match those of the adults. These examples of vocal learning demonstrate that animals can minimally shape or modify their communication as they develop.
|Image: Chittka L: Dances as Windows into Insect Perception. PLoS Biol 2/7/2004: e216.|
Bees communicate the location of distant food to members of their colony through a dance they perform in their hives, and vervet monkeys call out alarms that differ based on the identity of the predator. Behaviour of bees and the vervet monkeys shows us that some limited referential communication is possible, at least about referents that are in a different location.
The evidence for the combinatorial nature of nonhuman language (i.e., the ability to combine familiar communicative elements to create novel messages) is limited, but some cases do exist, such as when Campbell’s monkeys combine call components to create meanings that differ from the original parts.
Whether communication in other species is intentional has primarily been studied in one of the species most closely related to humans, chimpanzees. Chimpanzees use alarm calls to warn ignorant others of danger more often than knowledgeable ones, and change their screams during aggressive encounters when someone nearby is likely to be able to defeat their enemy.
Despite decades of research, determining exactly which features of human language are unique and which are shared with other species is still a largely unanswered question. Given the diversity of efficient communication systems in the animal world, the more interesting and productive question moving forward may be how the communication systems of other animals work in ways that are entirely different from our own.
Written by Katherine Cronin & Judith Holler
More videos on animals and language. (link)
Fedurek, P., & Slocombe, K. E. (2011). Primate vocal communication: A useful tool for understanding human speech and language evolution? Human Biology, 83, 153-173. (link)
Janik, V. M. (2013). Cognitive skills in bottlenose dolphin communication. Trends in Cognitive Sciences, 17, 157-159. (link)
Human children have a unique, and highly mysterious, ability to acquire proficient speech and language skills without the need to be formally taught. After only a few years, a typical infant has assembled a huge vocabulary of words, can use grammatical rules to combine them into a potentially limitless number of meaningful utterances, communicates these with rapid and precise coordination of the speech muscles, and is similarly adept at decoding the utterances of others. It has long been suspected that the answers to this puzzle may lie somewhere in our genetic makeup. Importantly, the human genome does not contain the information (the sets of words and rules) for any specific languages themselves; we all need exposure to a language in order to learn it. A child growing up surrounded by Dutch speakers becomes fluent in Dutch, while the same child growing up in Japan would learn to speak Japanese. Rather, our genes help to build brain circuits that are well-tuned to soak up language from the social environment.
For many years, it was only possible to speculate about potential genetic contributions to language acquisition. With the rise of new molecular techniques, scientists have now started to identify and study individual genes that are important. So far, most research has focused on searching for causative genes in children and adults who have problems with speech and/or language development that cannot be explained by another cause (like deafness or intellectual disability). It is clear that there is not just one single factor at play, instead an orchestra of genes and their interactions are responsible. Some genes have greater effects than others. For example, one gene with a large effect is FOXP2, the first gene implicated in an inherited speech and language disorder. If a child carries a disruptive mutation in this gene, it is enough to cause serious problems with learning to make sequences of speech sounds, problems that persist throughout life. In contrast to the rare severe mutations of FOXP2, more commonly occurring variants seen in other genes, such as CNTNAP2, ATP2C2, or CMIP have more subtle effects, increasing risk of language difficulties in small but significant ways.
While quite a lot is known about genetic variants that lead to language problems, there is less insight into genetic effects that make certain people in the general population better at acquiring language(s). Some of the genes involved in language impairments, like CNTNAP2, are also known to have effects on language development and function in people without disorders, yet further studies are required to uncover the genes and variants that might give some people a linguistic advantage. An exciting new step for this research will be to exploit the latest genomic techniques to look at the other extreme of the spectrum, people who have unusual talents in acquiring or using language.
Written by Katerina Kucera & Simon Fisher
The Language Fossils Buried in Every Cell of Your Body (link)
Graham S.A., Fisher, S.E. (2013). Decoding the genetics of speech and language. Current Opinion in Neurobiology, 23, 43-51. (link)
It is not unreasonable to expect that English will become the global language, as in many respects it already is; English is the language of the internet, English is the language of academia, English is the language of lots of mass media. Moreover, many languages are threatened with extinction; it is estimated that about half of the 7000 or so languages currently spoken will be extinct by 2100.
However, while new technology means that people all over the world are exposed to English, it also means that there are new and exciting ways to help keep languages alive. Social media is perfect for showing language diversity; many people use English (or Mandarin, or Swahili, or Spanish, or other dominant languages) for professional and formal work while using their own native languages for conversations with friends and family. Until recently, many of these native languages were unwritten and unused outside the home, but the rise of text messaging and social media has revitalised these languages and made them relevant to younger speakers.
Another reason that English (or any other language) is unlikely to become the one global language is that people always use language as a means of emphasising their own identity. Even within English, there are various different types of English, and the usage of English can be a highly divisive issue. For example, native speakers of British English are very protective of their own style of English, and British hostility towards Americanisms (words or phrases which are particular to American English, like the American "oftentimes" as opposed to the British "often" and the American "I have gotten" as opposed to the British "I have got") can be surprisingly vitriolic. Language protection is always linked to identity; if you are living in the Netherlands you only have to look as far as Belgium to see how.
English might be a bit like a sandwich. Just because you can buy a sandwich almost anywhere in the world, people are not going to stop eating curry, lasagne, burritos, or stroopwafels...
Written by Gwilym Lockwood & Katrien Segaert
50 of the most noted examples of Americanisms (link)
Katrien Segaert, Katerina Kucera, Judith Holler, Sean Roberts, Agnieszka Konopka, Gwilym Lockwood, Elma Hilbrink, Joost Rommers, Mark Dingemanse, Connie de Vos.