Paralanguage, also known as vocalics, is a component of meta-communication that may modify meaning, give nuanced meaning, or convey emotion, by using techniques such as prosody, pitch, volume, intonation, etc. It is sometimes defined as relating to nonphonemic properties only. Paralanguage may be expressed consciously or unconsciously.
The study of paralanguage is known as paralinguistics and was invented by George L. Trager in the 1950s, while he was working at the Foreign Service Institute of the U.S. Department of State. His colleagues at the time included Henry Lee Smith, Charles F. Hockett (working with him on using descriptive linguistics as a model for paralanguage), Edward T. Hall developing proxemics, and Ray Birdwhistell developing kinesics. Trager published his conclusions in 1958, 1960 and 1961.
His work has served as a basis for all later research, especially those investigating the relationship between paralanguage and culture (since paralanguage is learned, it differs by language and culture). A good example is the work of John J. Gumperz on language and social identity, which specifically describes paralinguistic differences between participants in intercultural interactions. The film Gumperz made for BBC in 1982, Multiracial Britain: Cross talk, does a particularly good job of demonstrating cultural differences in paralanguage and their impact on relationships.
Paralinguistic information, because it is phenomenal, belongs to the external speech signal (Ferdinand de Saussure's parole) but not to the arbitrary conventional code of language (Saussure's langue).
The paralinguistic properties of speech play an important role in human communication. There are no utterances or speech signals that lack paralinguistic properties, since speech requires the presence of a voice that can be modulated. This voice must have some properties, and all the properties of a voice as such are paralinguistic. However, the distinction linguistic vs. paralinguistic applies not only to speech but to writing and sign language as well, and it is not bound to any sensory modality. Even vocal language has some paralinguistic as well as linguistic properties that can be seen (lip reading, McGurk effect), and even felt, e.g. by the Tadoma method.
Aspects of the speech signalEdit
Speech signals arrive at a listener’s ears with acoustic properties that may allow listeners to identify location of the speaker (sensing distance and direction, for example). Sound localization functions in a similar way also for non-speech sounds. The perspectival aspects of lip reading are more obvious and have more drastic effects when head turning is involved.
The speech organs of different speakers differ in size. As children grow up, their organs of speech become larger, and there are differences between male and female adults. The differences concern not only size, but also proportions. They affect the pitch of the voice and to a substantial extent also the formant frequencies, which characterize the different speech sounds. The organic quality of speech has a communicative function in a restricted sense, since it is merely informative about the speaker. It will be expressed independently of the speaker’s intention.
Paralinguistic cues such as loudness, rate, pitch, pitch contour, and to some extent formant frequencies of an utterance, contribute to the emotive or attitudinal quality of an utterance. Typically, attitudes are expressed intentionally and emotions without intention, but attempts to fake or to hide emotions are not unusual.
Consequently, paralinguistic cues relating to expression have a moderate effect of semantic marking. That is, a message may be made more or less coherent by adjusting its expressive presentation. For instance, upon hearing an utterance such as "I drink a glass of wine every night before I go to sleep" is coherent when made by a speaker identified as an adult, but registers a small semantic anomaly when made by a speaker identified as a child. This anomaly is significant enough to be measured through electroencephalography, as an N400. Autistic individuals have a reduced sensitivity to this and similar effects.[failed verification]
Emotional tone of voice, itself paralinguistic information, has been shown to affect the resolution of lexical ambiguity. Some words have homophonous partners; some of these homophones appear to have an implicit emotive quality, for instance, the sad "die" contrasted with the neutral "dye"; uttering the sound /dai/ in a sad tone of voice can result in a listener writing the former word significantly more often than if the word is uttered in a neutral tone.
Ordinary phonetic transcriptions of utterances reflect only the linguistically informative quality. The problem of how listeners factor out the linguistically informative quality from speech signals is a topic of current research.
Some of the linguistic features of speech, in particular of its prosody, are paralinguistic or pre-linguistic in origin. A most fundamental and widespread phenomenon of this kind is described by John Ohala as the "frequency code". This code works even in communication across species. It has its origin in the fact that the acoustic frequencies in the voice of small vocalizers are high, while they are low in the voice of large vocalizers. This gives rise to secondary meanings such as "harmless", "submissive", "unassertive", which are naturally associated with smallness, while meanings such as "dangerous", "dominant", and "assertive" are associated with largeness. In most languages, the frequency code also serves the purpose of distinguishing questions from statements. It is universally reflected in expressive variation, and it is reasonable to assume that it has phylogenetically given rise to the sexual dimorphism that lies behind the large difference in pitch between average female and male adults.
In text-only communication such as email, chatrooms and instant messaging, paralinguistic elements can be displayed by emoticons, font and color choices, capitalization and the use of non-alphabetic or abstract characters. Nonetheless, paralanguage in written communication is limited in comparison with face-to-face conversation, sometimes leading to misunderstandings.
Specific forms of paralinguistic respirationEdit
A gasp is a kind of paralinguistic respiration in the form of a sudden and sharp inhalation of air through the mouth. A gasp may indicate difficulty breathing and a panicked effort to draw air into the lungs. Gasps also occur from an emotion of surprise, shock or disgust. Like a sigh, a yawn, or a moan, a gasp is often an automatic and unintentional act. Gasping is closely related to sighing, and the inhalation characterizing a gasp induced by shock or surprise may be released as a sigh if the event causing the initial emotional reaction is determined to be less shocking or surprising than the observer first believed.
As a symptom of physiological problems, apneustic respirations (a.k.a. apneusis), are gasps related to the brain damage associated with a stroke or other trauma.
A sigh is a kind of paralinguistic respiration in the form of a deep and especially audible, single exhalation of air out of the mouth or nose, that humans use to communicate emotion. It is voiced pharyngeal fricative, sometimes associated with a guttural glottal breath exuded in a low tone. It often arises from a negative emotion, such as dismay, dissatisfaction, boredom, or futility. A sigh can also arise from positive emotions such as relief, particularly in response to some negative situation ending or being avoided. Like a gasp, a yawn, or a moan, a sigh is often an automatic and unintentional act. In literature, a sigh is often used to signify that the person producing it is lovelorn.
Scientific studies show that babies sigh after 50 to 100 breaths. This serves to improve the mechanical properties of lung tissue, and it also helps babies to develop a regular breathing rhythm. Behaviors equivalent to sighing have also been observed in animals such as dogs, monkeys, and horses.
In text messages and internet chat rooms, or in comic books, a sigh is usually represented with the word itself, 'sigh', possibly within asterisks, *sigh*.
Moans and groansEdit
Moaning and groaning both refer to an extended guttural sounds emanating from the throat, which typically indicate displeasure, and are made by humans beginning in infancy. Although moaning is associated with pain and suffering, moans may also accompany pleasurable physical experiences such as eating stimulating food, receiving a massage, or engaging in sexual activity. Moans and groans are also noises traditionally associated with ghosts, and their supposed experience of suffering in the afterlife.
Clearing one's throat is a metamessaging nonverbal form of communication used in announcing one's presence upon entering the room or approaching a group. It is done by individuals who perceive themselves to be of higher rank than the group they are approaching and utilize the throat-clear as a form of communicating this perception to others.
As a form of metacommunication, the throat-clear is acceptable only to signal that a formal business meeting is about to start. It is not acceptable business etiquette to clear one's throat when approaching a group on an informal basis; the basis of one's authority has already been established and requires no further reiteration by this ancillary nonverbal communication.
Mhm is between a literal language and movement, by making a noise "hmm" or "mhm", to make a pause for the conversation or as a chance to stop and think.
The "mhm" utterance is often used in narrative interviews, such as an interview with a disaster survivor or sexual violence victim. In this kind of interview, it's better for the interviewers or counselors not to intervene too much when an interviewee is talking. The "mhm" assures the interviewee that they are being heard and can continue their story. Observing emotional differences and taking care of an interviewee's mental status is an important way to find slight changes during conversation.
Physiology of paralinguistic comprehensionEdit
Several studies have used the fMRI paradigm to observe brain states brought about by adjustments of paralinguistic information. One such study investigated the effect of interjections that differed along the criteria of lexical index (more or less "wordy") as well as neutral or emotional pronunciation; a higher hemodynamic response in auditory cortical gyri was found when more robust paralinguistic data was available. Some activation was found in lower brain structures such as the pons, perhaps indicating an emotional response.
- Leeds-Hurwitz, W. (1990). Notes in the history of intercultural communication: The Foreign Service Institute and the mandate for intercultural training. Quarterly Journal of Speech, 76, 262–281.
- Trager, G. L. (1958). Paralanguage: A first approximation. Studies in Linguistics, 13, 1–12.
- Trager, G. L. (1960). Taos III: Paralanguage. Anthropological Linguistics, 2, 24–30.
- Trager, G. L. (1961). The typology of paralanguage. Anthropological Linguistics, 3 (1), 17–21.
- Gumperz, J. J. (1982). Discourse strategies. Cambridge: Cambridge University Press.
- Van Berkum, J. J., Van den Brink, D., Tesink, C. M., Kos, M., & Hagoort, P. (2008). The neural integration of speaker and message. Journal of Cognitive Neuroscience, 20, 580–591.
- Groen, W. B., Tesink, C., Petersson, K. M., Van Berkum, J., Van der Gaag, R. J., Hagoort, P. and Buitelaar, J. K. (2010). Semantic, factual, and social language comprehension in adolescents with autism: an fMRI study. Cerebral Cortex, 20(8), 1937–1945.
- Nygaard, L. C., Lunders, E. R. (2002). Resolution of lexical ambiguity by emotional tone of voice. Memory & Cognition, 30(4), 583–593.
- Ohala, J. J. (1984) An ethological perspective on common cross-language utilization of F0 of voice. Phonetica, 41, 1–16. doi:10.1159/000261706.
- Rachel Broncher, A labor of love: a complete guide to childbirth for the mind, body, and soul (2004), p. 145.
- Fernando Poyatos, Paralanguage: a linguistic and interdisciplinary approach to interactive speech and sounds (1993), page 330.
- Paul Ekman, Emotions revealed: recognizing faces and feelings to improve communication (2007), p. 193.
- Schmidt, Elaine. "UCLA and Stanford researchers pinpoint origin of sighing reflex in the brain". UCLA Newsroom. Retrieved 2019-02-01.
- Determinants of Infant Behaviour: Proceedings, Volume 2 (1961), p. 123.
- Altman, Mara (August 21, 2018). "What Our Sex Sounds Say About Us". Time.
- Ben-Zeév, Aaron (April 16, 2019). "Why We Moan and Scream During Sex". Psychology Today.
- "Throat-Clear". www.nonverbal-dictionary.org.
- Question Authority: Think for Yourself, Estren & Potter, 2012:122.
- "ahem – Onomatopoeia List". 10 August 2013.
- "the definition of verbal". Dictionary.com.
- Primatology, Delmar 1984:46.
- Book of Etiquette and Manners, Chapter 3, by Nimeran Sahukar & Prem P. Bhalla.
- The healing effect of storytelling, Gabriele Rosenthal 2003
- Dietrich, S., Hertrich, I., Kai, A., Ischebeck, A., Ackermann, H. (2008). Understanding the emotional expression of verbal interjections: a functional MRI study. Brain Imaging, 19(18), 1751–1755.
- Cook, Guy (2001) The Discourse of Advertising. (second edition) London: Routledge. (chapter 4 on paralanguage and semiotics).
- Robbins, S. and Langton, N. (2001) Organizational Behaviour: Concepts, Controversies, Applications (2nd Canadian ed.). Upper Saddle River, NJ: Prentice-Hall.
- Traunmüller, H. (2005) "Paralinguale Phänomene" (Paralinguistic phenomena), chapter 76 in: SOCIOLINGUISTICS An International Handbook of the Science of Language and Society, 2nd ed., U. Ammon, N. Dittmar, K. Mattheier, P. Trudgill (eds.), Vol. 1, pp. 653–665. Walter de Gruyter, Berlin/New York.
- Matthew McKay, Martha Davis, Patrick Fanning  (1995) Messages: The Communication Skills Book, Second Edition, New Harbinger Publications, ISBN 1-57224-592-1, ISBN 978-1-57224-592-1, pp. 63–67.