Some restless infants don’t wait for birth to let out their first cry. They cry in the womb, a rare but well-documented phenomenon called vagitus uterinus (from the Latin word vagire, meaning to wail). Legend has it that Bartholomew the Apostle cried out in utero. The Iranian prophet Zarathustra is said to have been ‘noisy’ before his birth. Accounts of vagitus uterinus appear in writings from Babylon, Assyria, ancient Greece, Rome and India. The obstetrician Malcolm McLane described an incident that occurred in a hospital in the United States in 1897. He was prepping a patient for a c-section, when her unborn baby began to wail, and kept going for several minutes – prompting an attending nurse to drop to her knees, hands clasped in prayer. Yet another child is said to have cried a full 14 days before birth. In 1973, doctors in Belgium recorded the vitals of three wailing fetuses and concluded that vagitus uterinus is not a sign of distress. An Icelandic saga indicates that the phenomenon has been observed in other animals – ‘the whelps barked within the wombs of the bitches’ – vagitus uterinus in dogs, a foretelling of great events to come in Icelandic lore.
Air is necessary for crying. The coordinated movements of muscles in the stomach and rib cage force air out of the lungs and up through the vocal cords – two elastic curtains pulled over the top of the windpipe – causing them to vibrate and produce a buzz-like sound. These sound waves then pass through the mouth, where precise motions of the jaws, lips and tongue shape them into the vocal signals that we recognise. In this case, the rhythmic sounds of a cry.
Vagitus uterinus occurs – always in the last trimester – when there’s a tear in the uterine membrane. The tear lets air into the uterine cavity, thus enabling the fetus to vocalise. Vagitus uterinus provided scientists with some of the earliest insights into the fetus’s vocal apparatus, showing that the body parts and neural systems involved in the act of crying are fully functional before birth.
Loud, shrill and penetrating – a baby’s cry is its first act of communication. A simple adaptation that makes it less likely that the baby’s needs will be overlooked. And babies aren’t just crying for attention. While crying, they are practising the melodies of speech. In fact, newborns cry in the accent of their mother tongue. They make vowel-like sounds, growl and squeal – these are protophones, sounds that eventually turn into speech.
Babies communicate as soon as they are born. Rigorous analyses of the developmental origins of these behaviours reveal that, contrary to popular belief – even among scientists – they are not hardwired into our brain structures or preordained by our genes. Instead, the latest research – including my own – shows that these behaviours self-organise in utero through the continuous dance between brain, body and environment.
In the beginning, there is nothing. The world is formless and empty, darkness lies upon the amniotic waters. The embryo, still too young to be called a fetus, floats inside the amniotic sac of the uterus. It sees nothing. It cannot hear, smell or taste. It feels nothing – no heat, cold, pressure or pain. It cannot sense its own position or orientation. It cannot move.
In the seventh week of pregnancy, the embryo – c-shaped with a big head and tiny tail, eyes, ears, mouth, trunk and webbed limbs – begins to move. It measures about 10 mm from crown to rump – the size of a grape – and its heart, lungs, brain and spinal cord are forming. In that same week, the embryo begins to feel touch – which is the first of its senses to gain function. The world is no longer formless, or empty.
Movements trigger sensations, and sensations trigger movements. Through this cycle of sensorimotor activity, the embryo begins to discover its body, and world. It begins to develop vocal behaviour.
He got a First World War motion picture camera, and modified it by replacing the hand crank with a foot pedal
The first film of a living human fetus was made in January 1933, by Davenport Hooker, an anatomist at the University of Pittsburgh. We now consider it natural to observe the fetus in its home environment, but ultrasound technologies are shockingly young. Used originally to detect welding flaws in pipes, ultrasonography was first used to image a fetus in the late 1950s, by Ian Donald, an obstetrician in Glasgow. Real-time ultrasound scans have been around for about 50 years only, since the 1970s. To give you some perspective: Galileo Galilei first pointed a telescope towards the skies, and observed the craters of Earth’s moon, the rings of Saturn, the phases of Venus, the four largest moons of Jupiter and sunspots, in 1609. William Herschel discovered Uranus in 1781. Charles Wyville Thomson peered into the abyss for the first time, on board the HMS Challenger, in 1872. We had observed the skies and the depths of the oceans long before we had looked inside our own wombs.
Hooker wanted to study the developmental sequence of human sensorimotor activity. Scientists had done similar studies on a range of other species: salamanders, toadfish, terrapin, loggerhead turtles, pigeons, chickens, rats, cats, sheep and guinea pigs. But in 1933, Hooker had no way to look inside a human womb, so he arranged with a local hospital to film surgically aborted fetuses, while they were still alive. At the time, both the medical community and the general public viewed aborted fetuses as acceptable subjects for nontherapeutic research. The anthropologist Lynn M Morgan observes in her book Icons of Life: A Cultural History of Human Embryos (2009): ‘It is not that Hooker, his colleagues, or his audience de-humanised the fetus … they had never humanised fetuses to begin with.’
Early on, Hooker realised that fetal movements are too rapid to be studied accurately, unless the same movement can be viewed over and over again. So he purchased a surplus First World War motion picture camera, and modified it by replacing the hand crank with a foot pedal – thus freeing his hands to work with the fetus while also operating the camera.
On the day of surgical abortion, Hooker arrived at the hospital and set up his equipment in a nearby observation room. Time was of the essence, as the aborted fetuses in Hooker’s studies were all in the process of dying. Young fetuses would live for only seven to 10 minutes, and the older ones for a maximum of 20 minutes. To set up the camera in advance, Hooker picked out a plastic baby doll from a collection he had assembled, selecting the one that approximated the size of the fetus, based on its estimated age.
As soon as he received the fetus, Hooker positioned it beneath the camera, in a shallow basin of warm saline, and detached the placenta. This took him a few minutes to do. Then, as the camera recorded, Hooker used strands of thin human or thicker horse hair to lightly stroke various parts of the fetus – including its mouth, palms, trunk, limbs and back – to elicit movement.
Over a period of 25 years, Hooker filmed 149 fetuses, some of whom were spontaneously aborted, and others electively aborted for ‘the health, sanity, or life of the mother’, as Hooker wrote. They ranged in age from six weeks to 45 weeks (five weeks past the usual length of a pregnancy). After the fetus expired, Hooker handed over the fetal tissue to his collaborator Tryphena Humphrey. She then studied the structural development of the fetal nervous system.
Hooker and Humphrey found that the very young fetuses could not move. Then, at 7.5 weeks, lightly stroking the upper lip, lower lip or wings of the nostrils – collectively called the perioral area – caused the fetus to sharply move its head and trunk away from the offending strand of hair. The rest of its body still lacked sensation. Between 8 and 9.5 weeks, the fetus had feeling in its lower jaw area, and the sides of its mouth and nose. By 10.5 to 11 weeks, the upper regions of its face, like the eyelids, could feel touch. Two or three days after that, at around 11.5 weeks, the entire face of the fetus was sensitive to touch. In that same week, the fetus moved its lower jaw for the first time. At 12 to 12.5 weeks, the fetus had mature reflexes – it closed its lips when Hooker stroked them (in contrast to moving its whole upper body). By 13.5 to 14.5 weeks of age, the fetus was opening and closing its mouth, scrunching its face into a scowl-like expression, moving its tongue, and even swallowing.
The work of Hooker and Humphrey uncovered the close evolutionary ties we have to the other animals who share our planet. Though timelines may vary, terrapin reptiles, carrier pigeon birds and fellow mammals like rats, cats and sheep share a similar development sequence. In all of these species – ours included – the tactile system is the first to come online, and sensation begins in the perioral area, which is innervated by the fifth and largest cranial nerve: the trigeminal nerve. This nerve carries sensory inputs into the brain, giving us the ability to experience touch in our face and in the insides of our mouth. It also brings motor outputs from the brain to the muscles responsible for jaw movements, giving us the ability to feed and vocalise.
Some fetuses did, in fact, vocalise. ‘Between 18.5 and 23.5 weeks [about midway through the second trimester],’ Hooker reported, ‘the chest and abdominal contractions and expansions increase in incidence and amplitude, until brief but effective respiration … occurs … Whenever respiration occurs,’ he observed, ‘phonation – only a high-pitched cry, of course – accompanies its initiation and may be repeated at intervals following the establishment of breathing.’
Recent studies of prematurely born infants – conducted just a few years ago by Kimbrough Oller at the University of Memphis – have brought to light that fetuses born as young as 32 weeks (eight weeks before the usual date) do more than cry. They produce protophones, the infant sounds that eventually turn into speech. Meaning that a fetus in the last trimester of pregnancy can make all the sounds of a newborn infant.
Birth brings about tremendous changes in our physical and social environments, but life after birth is a continuation of life before birth.
Hooker’s findings had political consequences. When his science entered the public domain – in magazine spreads and a popular book called The First Nine Months of Life (1962), published just three years before his death – Hooker’s fetuses were used to illustrate the ‘wonder and beauty of human development’, without mention of abortion or where these fetuses came from. This was the Space Age, and his fetuses were pictured as little astronauts travelling through space in a placental capsule. The portrayals helped propagate a myth of independent fetal life – a way of thinking that had not existed publicly when Hooker began his research. In the 1970s, anti-abortion activists began using his work to plead their cases, including in briefs to the US Supreme Court.
Most visual representations of pregnancy are misleading, says the developmental biologist Scott Gilbert, professor emeritus at Swarthmore College in Pennsylvania. The mother and fetus are not separated by a cavity. Although often depicted this way, the fetus is not a tenant inside a vessel. Instead, the mother and fetus share arteries, veins and membranes. The placenta is an organ composed of both maternal and fetal tissues. ‘There is no clear boundary marking the place where the fetus ends and the uterus begins,’ Gilbert explains. The mother and fetus share a united anatomy, as well as cardiovascular, pulmonary, immune and metabolic systems. Pregnancy changes the mother’s physiology in ways that help both her and the fetus survive. Environmental factors that jolt the mother’s physiology – such as exposure to poverty, war, pollution or extreme heat events – will change the fetus. In strict scientific terms, neither the fetus nor the mother is a traditional biological ‘individual’. They are a fused entity. Political manoeuvrings to grant fetuses personhood, separate from their mothers, are not grounded in science.
I hope to convince you to drive a stake through the heart of the nature versus nurture debate
In the 65 years since Hooker completed his research, advances in ultrasound technology have allowed developmental psychologists to observe fetal faces in considerable detail. They have documented cry-like motions in fetuses that later went on to live full and healthy lives.
How then does our ability to vocalise come to be? Is it in our nature – ‘in-built’, ‘instinctive’, ‘inherent’, ‘innate’, ‘programmed’, ‘hardwired’, ‘in our genes’? Or is this ability ‘acquired’, ‘learned’, ‘absorbed from our environment’ – is there some nurture to speak of, even in the womb? If so, how much? Is it 80 per cent nature and 20 per cent nurture? Is it 90/10? 30/70? Or some other split? What I hope to do is convince you to drive a stake through the heart of the nature versus nurture debate – often framed as a fight for causal supremacy between two opposing factions – it gets us nowhere.
I learned this lesson while working in the Developmental Neuromechanics and Communication Lab, led by Asif Ghazanfar at Princeton University, where I earned my doctorate. When a behaviour – such as vocalising – is present at birth, it is thought to be largely in our nature (often even by scientists). But looking inside the womb reveals that the behaviour is the outcome of a developmental process – and at no point can that process be neatly partitioned into silos of nature or nurture.
Nature shapes nurture, and nurture shapes nature. It is not a duality – it’s more like a Möbius strip.
On the afternoon of 7 July 2011, I was with my lab mate Daniel Takahashi, a Brazilian-Japanese medical doctor turned neuroscientist. We were hunched over a padded bench in a dimly lit room of a 1920s neo-gothic building. I had an ultrasound wand in one hand and a popsicle stick coated with marshmallow fluff in the other. Our pregnant patient lay on the bench between us, as she had done countless times over the past year.
From head to toe, Takahashi and I wore medical-grade protective gear – gloves, gown, surgical mask, face shield, booties and bonnet. The equipment was in part to meet the stringent standards of the university and federal regulatory boards, but also for protection. The transmission of germs between us and our patient could be perilous, particularly for her.
Our patient was not concerned about germs, or the ultrasound gel dripping off her stomach, or even Takahashi’s gently restraining hands. Her mind was on the marshmallow fluff. We were thankful for this, as our task was not an easy one. We had to find her tiny acrobatic fetus in the vastness of her womb, and observe its face for as long as she would let us, while recording the fetus through a DVD player hooked up to the ultrasound machine. On this day, skill and luck collided. Takahashi and I were able to watch the face of the fetus for almost 30 minutes and, at moments, we saw it moving its mouth.
Our pregnant patient was a marmoset monkey (Callithrix jacchus), one of five females and five males, housed in pairs in a large indoor enclosure a few doors away from where we were conducting the ultrasound examination. We called her Laurie Rose. Her mate was John Holmes. We named them after porn stars, because they went at it like porn stars. Laurie Rose was the first of our monkeys to get pregnant and carry her baby to term. We were able to observe, quantify and characterise the mouth movements of her fetus from day one of movement to the day before birth – an approach not feasible in humans.
These first actions looked different from the movements Jack made after birth
Marmosets are a New World monkey species found in Brazil. They had a common ancestor with our species about 35 million years ago. Despite our distant ties and their squirrel-like appearance, marmosets resemble humans in a number of ways. Physiologically, they develop 12 times faster than we do, but have roughly the same brain and body parts. They live in family units of three to 15 – mother, father and children (the monkey equivalent of a nuclear family). Occasionally, extended family members and unrelated adults will live with them. They engage in both monogamy and polygamy, and older family members tend to assist in child-rearing. Marmosets are chatty, and take turns to speak to one another – like we do. Newborn marmosets cry and make a number of other sounds. These sounds are comparable to human protophones, but are more adult-like in their form. Most remarkably, Takahashi and Ghazanfar discovered that marmoset babies learn to ‘talk’ using parental feedback, similar to how human babies learn. It is highly likely that the vocal development of a marmoset fetus is similar to our vocal development.
At first, Laurie Rose’s fetus – we called him Jack – made very few mouth movements (and they were often accompanied by motions of the head). These first actions looked different from the movements Jack made after birth.
In scientific works, descriptions of early fetal movements, from a diversity of species, strongly resemble one another. Chicken embryos exhibit ‘convulsive-type jerks and twitches and occasional head thrusts’; cat fetuses show ‘purposeless, squirming trunk, head, and leg movements’; rat fetuses display ‘unintegrated, aimless movements’; and human fetuses give us ‘spontaneous startles, general movements (GMs), isolated movements and twitches’. This is how Jack was moving.
Hooker and other scientists of that era believed that all fetal movements are primarily reflexive, evoked by external stimuli. We now know that the earliest fetal motions are spontaneous, first triggered in the muscles themselves and later by electrical pulses sent from the spinal cord or brainstem to the muscles. Fetal movements can be spontaneous or evoked by a range of stimuli, such as the mother rubbing her belly or the fetus touching its own body.
Actions of the fetus – ‘purposeless’ though they may appear – are important. They support the growth of muscles, tendons, ligaments, cartilages and bones. They steer the development of the sensorimotor circuits. General movements are ‘storms’ of activity that help the fetus develop a sense of its own form and body position. Isolated movements help connect the body to the brain, through tactile feedback. Twitches, which occur during sleep, are the brain’s way of testing its circuitry. The behavioural neuroscientist Mark Blumberg likens the act to a switchboard operator throwing on switches, one at a time, to test which light turns on with each switch.
As Jack got older, he grew more active. He moved his mouth more. I calculated the entropy of his movement patterns and found that it was decreasing. This means Jack’s actions were getting less chaotic. He was now moving his mouth independently from his head. If you will recall, Hooker found a similar trajectory in humans. When he stroked the mouth of the 7.5-week-old fetus, it moved away its entire upper body. With time, its body parts acquired greater autonomy, and at 12.5 weeks the fetus responded by closing its mouth alone.
How does the fetal body become more finely tuned? It is hard to look under the hood of a living being and know exactly what is happening, but a group of roboticists in Japan are attempting to reverse-engineer human development and build humanoid robots. They provide us with clues. At the University of Tokyo, Yasuo Kuniyoshi, Hiroki Mori and Yasunori Yamada simulated the motor development of a fetus. At the start line, the computer-generated fetus was placed inside an artificial womb, enclosed in an elastic membrane, and bathed in fluid. It had immature neural circuits, with a biologically accurate musculoskeletal body and tactile receptor distribution. Even something as seemingly ‘hardwired’ as our physiology is not shaped by genes alone. Biophysical events (like chemical reactions in the cells, mechanical pressures inside and on the cells, and gravity) switch genes on and off – influencing how cells divide, move, determine their cell fate, and group into tissues and body parts.
The neural circuits of the simulated fetus triggered movement. As the fetus moved, its tactile receptors detected the pressure of fluid against its moving body. The information was passed on to its nervous system. The action of each muscle affected the sensory input it received, as well as the sensory input received by adjacent muscles. Sometimes, an isolated movement brought one body part into contact with another – inducing active stimulation of the touching part and passive stimulation of the touched part.
The neural circuits of the simulated fetus began to learn. Initially, large undifferentiated circuits controlled multiple body parts (eg, a single motor unit controlling the head and trunk). As development progressed, the brain responded to feedback from its body, causing these large units to subdivide into smaller, more precisely controlled units (eg, separate control of the head and trunk, driven by the separate experiences they have while moving).
The fetal body, like ours, varies in sensitivity to touch – the mouth region is dense with tactile receptors
Neuroscientists have a mnemonic for the biological process that adjusts the strength of connections between neurons – ‘Fire together, wire together. Out of sync, lose your link.’ The catchy rhyme is credited to the neurobiologist Carla Shatz. It means that the brain rewires and fine-tunes itself by strengthening the connections between neurons that are repeatedly activated together, and weakening the connections between neurons that are not activated together. The weak connections are ultimately pruned.
In the simulation, as in real life, the musculoskeletal body of the fetus determined both the possibilities and constraints of movement. Human heads, for example, rotate around two socket pivots, giving us a decent range of motion, but not as much as an owl’s head, which rotates around a single socket pivot.
As the simulated fetus grew larger, it came into more physical contact with the uterine wall. When this happened, the weak and continuous tactile signals from the amniotic fluid were followed by the short, blunt force of the wall hitting its flesh. The brain took note, and the fetus modified its actions.
The fetal body, like ours, varies in sensitivity to touch – the mouth region is dense with tactile receptors, whereas the top of the head is not. As the simulated fetus matured, it frequently touched its more sensitive parts while barely touching the more boring ones. This behaviour mirrors observations of real-life fetuses, who often direct their arm movements to their mouth, and even open their mouth in anticipation of their hand arriving there.
As the simulated fetus continued to grow, its larger body parts (like the head) came to be more constrained by the uterine wall, and only small regions (like the mouth) could move freely.
The neural circuits responded to all these contingencies, and the body of the fetus became more precisely controlled (eg, separate functions of the head and mouth, more manoeuvrable orofacial muscles). The fetus was now practising infant-like movements. In real life, the most frequent actions in the third trimester are face-related – hands touching the face, facial expressions like scowling, and mouth movements. All the behaviours seen in a newborn have at some point been observed in the fetus.
Fetal movements self-organise through the continuous and reciprocal interactions between brain, body and environment.
Like the simulated fetus, Jack began moving his mouth independently while rarely moving his head. I used a dynamic time warping (DTW) algorithm – often used to detect forged signatures – to match the temporal profiles of his mouth movements to those he eventually made as an infant. DTW generates a ‘distance’ score to indicate the extent to which one profile must be ‘warped’ to match the other. The algorithm uncovered that, in parallel to his mouth and head decoupling, a subset of Jack’s mouth motions were becoming nearly identical to those he made while calling out to his family as an infant. Fetal Jack had developed the behavioural repertoire he would so often use in life.
My work with Ghazanfar and Takahashi established – for the first time in a primate – that the ability to vocalise at birth is not ‘innate’. Rather, it undergoes a lengthy period of prenatal development, even before sound can be produced.
Marmosets and other nonhuman primates have precursors to language, but language is unique to humans. Even as newborns, we show a preference for speech, we recognise our native languages, and remember the stories we have frequently heard in the womb. As babies, we don’t merely cry – as already mentioned, we cry in the accent of our mother tongue. To understand the development of these uniquely human behaviours, we need to turn to hearing.
The sensory systems of birds and mammals on our planet, ours included, develop in a fixed order – with hearing developing second to last (just before the visual system). The human auditory system comes online between 24 and 28 weeks, when the spiral ganglion neurons connect our auditory receptors – the hair cells of the inner ear – to our brainstem and temporal lobe. Hearing becomes fully functional in the third trimester.
Hearing is often studied in sheep. The size of the middle and inner ear of sheep is similar to ours, and hearing develops in utero in both species. Ken Gerhardt, Robert Abrams and their colleagues at the University of Florida implanted a tiny electronic device in the inner ear of fetal sheep, to study what speech signals the auditory receptors of the fetus pick up and transmit to its brain. They then played 48 English sentences through a speaker positioned near the mother sheep. The sounds captured by the device were replayed to adult humans who attempted to decipher the original sentences, from the speech signals recorded in the fetus’s inner ear. In total, they deciphered about 41 per cent of the sentences. They tended to miss or confuse high-frequency consonants like ‘s’, ‘h’ or ‘f’. A word like ‘ship’ could be easily confused with ‘slit’ or ‘sit’. The low-frequency vowels were heard more reliably.
Vowel sounds penetrate the fetus’s inner ear better than high-frequency consonants because the uterine environment acts as a low-pass filter, absorbing and muffling sound frequencies above 600 Hz. While individual speech sounds are suppressed in the womb, what remains prominent are the variations in pitch, intensity and duration – what linguists refer to as the prosody of speech.
Exposure to speech in the womb leads to lasting changes in the brain
Prosody is what gives speech its musical quality. When we listen to someone speak, prosody helps us interpret their emotions, intentions and the overall meaning of their message. Different languages have different prosodic patterns. Native speakers of English, French or Italian use contrasts in syllable duration to make words more salient. When native English speakers utter the phrase ‘to Ro:me’, we draw out the vowel ‘o’ in ‘Rome’ (but not in ‘to’). A native Japanese speaker would use a change in pitch instead. They would say ‘^Tokyo kara’ (‘to Tokyo’) to draw the listener’s attention to the more important word ‘Tokyo’. Nawal Abboub, Thierry Nazzi and Judit Gervain used near-infrared spectroscopy (NIRS) to image the brains of newborns in one of Europe’s largest pediatric hospitals. They found that the infants had learned the prosodic patterns – duration contrast versus pitch contrast – of their mother tongue.
Language learning begins in the womb, and it begins with prosody. Exposure to speech in the womb leads to lasting changes in the brain, increasing the newborns’ sensitivity to previously heard languages. The mother’s voice is the most dominant and consistent sound in the womb, so the person carrying the fetus gets first dibs on influencing the fetus. If the mother speaks two languages, her infant will show equal preference and discrimination for both languages.
The fetus’s knowledge of native prosody goes beyond perception. Kathleen Wermke, Anne Christophe and their colleagues in Würzburg and Paris collected and analysed the cries of 30 German and 30 French newborns from strictly monolingual families. They found that the French newborns tended to cry with a rising – low to high – pitch, whereas the German newborns cried more with a falling – high to low – pitch. Their patterns were consistent with the accents of adults speaking French and German.
The newborns had not just memorised the prosody of their native languages; they were actively moving air through their vocal cords and controlling the movements of their mouth to mimic this prosody in their own vocalisations. Babies are communicating as soon as they are born, and these abilities are developing in the nine months before birth.
There is no genetic blueprint, programme or watchmaker who knows how it must turn out in the end. The reality of how these behaviours come to be is far more sophisticated and elegant. They develop through continuous interactions across multiple levels of causation – from genes to culture. The processes that shape them unfold over many timescales – from the milliseconds of cellular processes to the millennia of evolution.
Now here’s my challenge to you: which of these factors are the primary drivers of vocal development – our genes or brain? – and which ones are merely supporting – the body? How much of their communication do babies owe to nature versus nurture? Is it more nature or nurture? I guarantee you, there are no scientifically defensible answers to these questions.
Development does not rely on a distinction between primary, essential causes and secondary, supporting ones. They are all essential – intimately connected – causal contributors. Let us disregard our artificial hierarchies. Mother Nature herself pays little attention to them.