Why Cuts Work: The Science and Psychology of Film Editing
Section 4 of 13

Early Film Editing History from Lumière to Griffith

From Lumière to Griffith: The Birth of Editing Language

There's a famous story — possibly apocryphal, almost certainly embellished — about audiences at the Grand Café in Paris on December 28, 1895, screaming and leaping from their seats as a train locomotive appeared to bear down on them from the screen. The Lumière Brothers had shown L'Arrivée d'un train en gare de La Ciotat, and supposedly the crowd panicked, convinced they were about to be crushed.

Film historians have largely debunked the panic as myth. But here's what's actually interesting about the myth: it reveals something true about how quickly the brain adapts to something entirely new. Early cinema was shocking not because audiences couldn't tell images from reality — they obviously could — but because moving photographic images were categorically unprecedented. Your brain had never encountered anything like them. It had no template yet for how to segment, predict, and construct meaning from sequenced moving images.

Which means we're about to watch something remarkable: filmmakers discovering, mostly by accident, how to exploit the machinery we just explored. They had no EEG data. No neuroscience papers. No research grants. Just intuition, constant experimentation, and audiences who would immediately tell them — through engagement or confusion — whether a sequence worked. What's stunning is how quickly their intuitions converged on rules that align perfectly with how brains actually process information. Within roughly two decades of the Lumières' first films, filmmakers had invented editing as a genuine act of creative construction.

That speed is no accident. It tells us that cutting between images, despite being entirely artificial, must be exploiting something the brain already knows how to do.

So history becomes a kind of real-time demonstration. We'll start with the Lumières — cinema before editing, pure documentation — and trace the moment when filmmakers realized that the gaps between images could be more powerful than the images themselves.

Remember what we learned about event boundaries and perceptual suppression? Your brain is built to chunk experience into discrete events, to infer causality across gaps, to construct narrative from fragmentary glimpses. The Lumière model — a continuous, uninterrupted shot — actually works against these mechanisms. There's nothing to segment, no ambiguity to resolve, no gap to bridge. The brain is, in a sense, underemployed.

The Lumières made beautiful films. They just hadn't yet discovered that they were wasting the most powerful part of how human perception works.

Méliès: The Accident That Changed Everything

Georges Méliès was not trying to invent editing. He was trying to document Paris street life when, according to his own account, his camera jammed somewhere near the Place de l'Opéra. By the time he unjammed it and kept filming, the scene had changed: a horse-drawn omnibus had moved on and been replaced by a hearse. When Méliès projected the developed footage, the omnibus appeared to magically transform into a hearse — an instantaneous substitution that the camera had produced entirely by accident.

Méliès, being a professional stage magician, recognized opportunity immediately.

What he'd stumbled into was the substitution splice — the idea that you could cut the film and create apparent transformations in time or space. He ran with it. In The Vanishing Lady (1896), a woman disappears. In A Trip to the Moon (Le Voyage dans la lune, 1902), which runs approximately 14 minutes (though some sources cite 13-16 minutes depending on print) and contains approximately 20-30 shots, with most sources indicating around 20 distinct shots rather than more than thirty, Méliès takes scientists to the moon through a series of theatrical tableaux that smash together in ways that are spatially incoherent but narratively legible.

Here's what Méliès understood: editing could do time tricks. A cut could make something disappear, reappear, or transport characters across space. The discontinuity wasn't a problem — it was the whole point.

But here's what he didn't pursue: editing could create narrative continuity. That cuts between different vantage points on an unfolding story could actually feel seamless. For Méliès, each shot was still fundamentally a theatrical tableau. He changed them one after another, but the camera never moved relative to the action. Same distance, same height, same implicit proscenium arch. He invented temporal editing but didn't discover spatial editing.

This distinction matters. Temporal editing — skipping time, reversing it, creating magical substitutions — is the more obvious trick. It mimics something the human mind already understands. We know experience isn't continuous. We sleep, we remember in fragments. But spatial editing — cutting between two different positions in space as if the viewer is the same continuous observer — that's the genuinely strange invention. It required someone to discover that audiences would simply accept it.

graph TD
    A[Lumière Brothers 1895<br/>Single Shot - Pure Record] --> B[Georges Méliès 1896-1904<br/>Temporal Cuts - Time Tricks]
    B --> C[Edwin S. Porter 1903<br/>Spatial Cuts - Parallel Spaces]
    C --> D[D.W. Griffith 1908-1916<br/>Systematic Continuity Grammar]
    D --> E[Hollywood Studio System 1920s<br/>Invisible Editing Codified]
    
    style A fill:#2d4a6e,color:#fff
    style B fill:#4a6e2d,color:#fff
    style C fill:#6e4a2d,color:#fff
    style D fill:#6e2d4a,color:#fff
    style E fill:#2d6e6e,color:#fff

Edwin S. Porter: Editing Across Space

By 1902, Edwin S. Porter was working as a director for the Edison Manufacturing Company in New York, thinking seriously about whether you could cut between locations. His 1902 film Life of an American Fireman cross-cuts between a woman and child trapped inside a burning building and firefighters rushing to their rescue — the same rescue shown from two different vantage points, two different spaces, edited together as a single narrative sequence.

The Great Train Robbery (1903) pushed harder. Porter's twelve-minute film contains fourteen shots that cut between robbers holding up a train, fleeing through the wilderness, and a posse chasing them down. Interior to exterior. Close action to wide landscape. One group of characters to another. The editing is rough by later standards — spatial continuity is approximate at best — and shots aren't connected by matches on action. But the principle is unmistakable: a story can live across multiple shots from multiple positions, and the audience will understand it as continuous.

There's one moment at the end of The Great Train Robbery that feels almost experimentally brazen: a medium close-up of the outlaw Barnes pointing his revolver directly at the camera — directly at the audience — and firing. Porter apparently told exhibitors they could place this shot at the beginning or end of the film, wherever it would have the most impact. It was a purely affective insert, existing outside the story's geography, designed to provoke an immediate visceral response.

What Porter had intuited, without neuroscience language to express it, is that cuts could do two related but distinct things: they could extend narrative space by connecting different locations as part of one story, and they could punctuate emotional experience by cutting to something that exists purely for how it makes you feel. These two uses of the cut — still fundamental today, a century later — emerged from Porter almost accidentally.

What's worth noting about Porter is how rough his editing is. Cuts don't always match. Actors appear in slightly different positions before and after a cut. The geography of the robbery is impressionistic rather than precise. And yet — audiences followed it. They understood that the robbers on the train and the posse forming in town were connected. That the story was moving forward. The brain bridged the gaps Porter left ragged.

This is the first clear historical evidence that audiences' cognitive machinery for narrative inference was ahead of filmmakers' techniques. The brain didn't need perfect continuity. It just needed enough.

D.W. Griffith: The First Systematic Grammarian

Between 1908 and 1913, D.W. Griffith directed approximately 450 short films for the Biograph Company. Four hundred and fifty. That number is, in retrospect, something like a crash research program in editing grammar — a systematic exploration of what cuts could and couldn't do, conducted in real time with paying audiences as involuntary research subjects.

Griffith didn't invent most of the techniques credited to him. Close-ups existed before him. Cross-cutting existed before him. Reaction shot cutaways appeared in earlier films. What was genuinely new was systematization. Griffith understood what cognitive work each tool was doing and deployed them as a coherent grammar rather than isolated tricks.

The Close-Up as Psychological Space

Before Griffith, close-ups were occasional novelties. Showing a face larger than life seemed to many filmmakers excessive, even grotesque. Audiences trained by theater watched from a fixed distance; you didn't suddenly rush to within two feet of an actor's face.

Griffith argued — first through practice, then explicitly — that the close-up wasn't a spatial violation but a psychological one. It didn't say the camera is closer. It said we are now inside this character's emotional reality. Film historian Tom Gunning's work on Griffith's Biograph period documents how Griffith used the close-up to shift from objective observation of events to intimate access to individual interiority.

This exploits something we now understand neurologically. Facial perception is one of the most specialized processes in the human brain — dedicated cortical regions for it, processed faster than almost any other visual information, extraordinary sensitivity to minute expression changes. The close-up doesn't overwhelm this system; it feeds it. A face filling the frame gives your social cognition machinery something to process at maximum resolution.

What Griffith grasped intuitively, and what neuroscience has since confirmed, is that faces are the primary way humans understand other humans' mental states. Cut to a close-up of a face, and you're not just showing a face. You're inviting the viewer into intense psychological inference about what that person is thinking, feeling, about to do.

Cross-Cutting: Time, Tension, and the Ticking Clock

Griffith's most celebrated innovation was systematic cross-cutting — alternating between two or more lines of action occurring simultaneously in different locations. A woman in danger while rescuers race toward her. A trial proceeding while new evidence surfaces elsewhere. A battle fought while its outcome ripples through people waiting at home.

The effect on audiences was, by contemporary accounts, intense. Cutting between simultaneous events created a kind of physiological urgency — what we call suspense — that theatrical experience had never quite managed. The cognitive mechanism is straightforward: once you've established that two lines of action are simultaneous, the brain starts tracking them both and automatically generating predictions about their intersection. Will the rescuers arrive in time? Will the evidence arrive before the verdict? The editing hijacks your predictive narrative machinery and creates a state of suspended resolution that feels viscerally uncomfortable.

This only works because editing has established a convention that simultaneous cross-cutting means simultaneous action. An audience that had never seen cross-cutting might interpret the shots as sequential events in the same location. The convention had to be learned.

And it was learned, by Griffith's audiences, remarkably fast. Which brings us back to the deeper question running underneath this entire history.

The Fade, the Iris, and Punctuation Marks

Griffith also systematized visual transitions that function as punctuation. The fade to black — gradual darkening to full darkness before a new scene — signals that time has passed, that we've moved beyond a narrative moment. The iris in (circular mask narrowing to highlight a detail) and iris out (expanding from a detail to the full frame) were ways of directing attention and marking beginnings and endings.

These aren't arbitrary stylistic flourishes. They encode temporal information in visual form, the same way punctuation encodes temporal and logical relationships in writing. A period says this thought is complete. A fade says this scene is complete, time has passed. The parallel isn't exact, but it's real: both are agreed-upon signals that train the reader or viewer to organize information into meaningful chunks.

Birth of a Nation and Intolerance: Grammar at Scale

By 1915, Griffith had refined his grammar enough to sustain a three-hour feature. The Birth of a Nation — a film whose technical achievements are real and whose foundational racism is inexcusable, facts that must exist in the same sentence — was the first film to demonstrate that Griffith's editorial techniques could hold an entire feature-length narrative together with genuine emotional power.

Griffith used Birth of a Nation as a showcase for techniques considered experimental at the time: battle scenes cut from dozens of camera angles, close-ups intercut with wide shots, flashback structures, cross-cutting across three simultaneous story lines. Contemporary audiences responded with what can only be described as shock at the emotional intensity — a response we can now understand as a perceptual system being fed a far richer and more precisely targeted diet of information than theatrical staging had ever provided.

Intolerance (1916), Griffith's even more ambitious follow-up, intercut four entirely separate historical narratives — a modern story, 1572 France, the fall of Babylon in 539 BC, and the life of Christ — editing between them with increasing speed toward climax. This is one of the most formally radical experiments in cinema history. The film failed commercially, partly because Griffith's grammar had outrun audiences' ability to process it rapidly.

That failure is instructive. It confirms that the learning process was real — that audiences were genuinely building cognitive templates for film language, templates that could be exceeded. Griffith had pushed faster than the templates could keep up.

How Audiences Learned: The Normalization of Cuts

Here's what should strike you as genuinely strange: between 1895 and 1920, a new perceptual language was invented, distributed globally, and learned by mass audiences with no instruction manual. People in rural Kazakhstan and London and rural Japan all, within roughly a decade of cinema exposure, acquired the ability to watch a film cut from dozens of shots in multiple locations and experience it as continuous, coherent story.

This happened without teaching. Across wildly different cultures and languages. Even among populations with no literary tradition and no prior experience with representational narrative art.

Cognitive scientist Tim Smith's research on eye-tracking and film comprehension helps explain why. Smith and colleagues found that trained and untrained viewers — people with extensive cinema experience and people with almost none — respond similarly to cuts in terms of basic narrative comprehension. What varies is not comprehension but attention: trained viewers know where to look after a cut. But even novice viewers understand what happened. A different character is on screen. Time may have passed. The story continues.

This pattern suggests something crucial: much of what makes editing work is not learned but pre-existing. Your brain's machinery for event segmentation, causal inference, and narrative construction — mechanisms that evolved long before cinema existed — turn out to be exactly the mechanisms that editing exploits. Learning film language is mostly learning to recognize the signals that tell you which pre-existing mechanism to apply. The hard cognitive work — constructing coherent narrative from fragmented inputs — is something your brain was already doing.

The speed of learning, then, is exactly what you'd predict from this model. Audiences didn't need to build new cognitive machinery. They needed to recognize new triggers for machinery that was already running.

graph LR
    A[Pre-existing Brain Mechanisms] --> D[Rapid Language Acquisition]
    B[Event Segmentation] --> A
    C[Causal Inference] --> A
    E[Narrative Construction] --> A
    F[Social Cognition / Face Reading] --> A
    
    G[Film Conventions as Triggers] --> D
    H[Cut = Event Boundary] --> G
    I[Close-up = Emotional Access] --> G
    J[Cross-cut = Simultaneity] --> G
    K[Fade = Time Skip] --> G
    
    style D fill:#6e2d4a,color:#fff
    style A fill:#2d4a6e,color:#fff
    style G fill:#4a6e2d,color:#fff

The Resistance That Existed

It's important not to pretend this was frictionless. There was genuine resistance from people who thought editing was either cheap or deceptive.

Early critics complained that close-ups were "unnatural," that showing a severed head — as Griffith did repeatedly via close-ups of faces that implied decapitation — was grotesque. Cuts between distant locations confused some viewers initially. There are accounts of early cinema audiences calling out explanations to confused neighbors.

Theater practitioners were especially alarmed. If cinema could cut to a close-up to convey emotion that an actor had to project across a distance in theater, what did that mean for the skills actors had spent careers developing? The answer turned out to be: irrelevant. Which is why early film acting looks so wildly theatrical to modern eyes — performers trained for stage distance were now being filmed at portrait distance, and the expressiveness that read as naturalistic from row twenty looked deranged at three feet.

The resistance largely evaporated not because critics were convinced by argument but because audiences kept going. Commercial reality established what persisted. By the early 1920s, the grammar Griffith had systematized was becoming the default — and other filmmakers, watching what worked, were adopting and extending it.

Hollywood Codifies: The 1920s and Invisible Editing

By the early 1920s, the Hollywood studio system was consolidating around editing practices that would come to be called continuity editing — a system of rules designed to make cuts invisible, to create the sensation of smooth, continuous story flow rather than discrete photographic fragments.

David Bordwell's research on Hollywood's classical style traces how studios in the 1920s developed written guidelines, hired continuity editors (initially called script clerks), and institutionalized practices that had previously been improvised. The 180-degree rule, the eyeline match, the match on action, the establishing shot/shot-reverse-shot pattern — these weren't discovered by one filmmaker. They were refined through enormous quantities of practice, viewer feedback tracked through exhibitor reports and eventually preview screenings, and competitive imitation.

What studios were unconsciously doing was something like empirical cognitive science: testing what audiences would accept, what confused them, what created unintended disorientation, then iterating toward a system that minimized confusion while maximizing engagement. The continuity system is, in this sense, a set of engineering solutions to the problem of building a reliable perceptual illusion.

We'll spend significant time with that system later, but it's worth noting here where it came from. It didn't descend from aesthetic philosophy. It wasn't invented by theorists. It was built by practitioners — many of them women, in the early years, because editing was considered clerical rather than directorial — working under commercial pressure to make films audiences would pay to see and understand.

The role of women in establishing Hollywood's editorial conventions is documented extensively by Maureen Turim and others: editors like Anne Bauchens (who edited Cecil B. DeMille's films for decades), Blanche Sewell, and Viola Lawrence were central to developing the practices we now take as fundamental film grammar. The fact that these contributions were later largely erased from film history's official record is itself a cautionary tale about how expertise gets attributed.

The Historical Argument for Cognitive Pre-Wiring

Let's return to the central question this history raises, because it shapes everything that follows.

If editing were genuinely alien to human cognition — if cuts were truly perceptually disruptive in a way that required extensive training to overcome — we would expect a slow learning curve. Persistent confusion. Regional variation, with populations that had more cinema exposure learning faster than those with less. Significant individual differences. The grammar becoming entrenched over decades, not years.

What we actually see is something much faster and more universal. Marshall Segall's cross-cultural perception research, documented in his work on visual perception and cultural differences, has been discussed by film scholars in relation to how audiences understand visual narratives. Tim Smith's recent work confirms that novice viewers understand edited films at a narrative level even without trained viewing habits. And the historical record shows worldwide audiences acquiring working comprehension of film language within years of exposure, not decades.

This pattern strongly suggests that editing taps into something architectural about human cognition. The specific signals — fade here, close-up there, cut on action to suggest continuous movement — those are cultural conventions that have to be learned. But the underlying operations those signals invoke — narrative event segmentation, causal inference, social cognition, spatial mental model construction — those are the product of hundreds of thousands of years of evolution, not twenty-five years of cinema.

What Griffith stumbled into, through prodigious output and genuine artistic intelligence, was a system for triggering human cognitive machinery at scale. He didn't know the neuroscience. But his systematic empiricism — making 450 films in five years and watching what audiences responded to — produced an editing grammar that works precisely because it's a good map of how brains process narrative experience.

The Lumières pointed a camera at the world and recorded it. Méliès discovered that cuts could do time tricks. Porter found that cuts could connect different spaces. Griffith organized all of this into a language — one that, once you understand its cognitive foundations, turns out to be less arbitrary than it looks and less invented than it seems.

It was waiting to be found, because the minds it speaks to were already wired to receive it.