Neil Cohn

Neil Cohn
Born (1980-01-11) January 11, 1980
Fields Cognitive science, Linguistics, Comics Studies
Institutions Tilburg University
Alma mater Tufts
University of Chicago
UC Berkeley
Doctoral advisor Ray Jackendoff, Gina Kuperberg, Phillip Holcomb
Other academic advisors Marta Kutas, Jeff Elman
Known for Visual language theory
Contributions to comics theory

Neil Cohn (/kn/; born 1980) is an American cognitive scientist and comic author. His research offers the first serious scientific study of the cognition of understanding comics, and uses an interdisciplinary approach combining aspects of theoretical and corpus linguistics with cognitive psychology and cognitive neuroscience.[1][2]

Cohn’s work argues that common cognitive capacities underlie the processing of various expressive domains, especially verbal and signed languages and what he calls “visual language”—the structure and cognition of drawings and visual narratives, particularly those found in comics. His theories have examined the linguistic status of emoji.[3][4][5]

Biography

Cohn began developing his theories as an undergraduate at UC Berkeley where he graduated in 2002. He then spent several years as an independent scholar before studying under linguist Ray Jackendoff and psychologists Gina Kuperberg and Phillip Holcomb at Tufts University where he received his PhD in psychology in 2012. He then did a postdoctoral fellowship at UC San Diego working with Marta Kutas and Jeff Elman. In 2016, he joined the faculty of the Tilburg center for Cognition and Communication at Tilburg University. He is the son of Leigh Cohn and Lindsey Hall.

Visual language theory

Cohn’s work challenges many of the existing conceptions of both language and drawing. He argues that language involves an interaction between an expressive modality, meaning, and a grammar. Just as sign languages differ from gestures in that they use a vocabulary and grammar, “visual languages” differ from individual drawings because they have a vocabulary of patterned graphic representations and a grammar constraining the coherence of sequential images. Full visual languages primarily appear alongside written languages in comics of the world, though they also appear outside of comics, such as in sand drawings used by Australian Aboriginals.[6] Just as spoken languages differ, so do visual languages: Japanese manga are written in “Japanese Visual Language” while American comics are written in “American Visual Language.” In addition, Cohn has argued that the development of visual languages may follow similar constraints as learning spoken and signed languages, and that most people do not learn how to draw proficiently because they do not acquire visual vocabularies within a critical period.[7]

Cohn’s primary research program with visual language theory emphasizes that a narrative structure operates as a “grammar” to sequential images analogously to syntactic structure in sentences. While narrative grammar uses a discourse level of information, its function and structure is similar to syntax in that it organizes categorical roles in hierarchic constituents in order to express meaning. Cohn’s work in cognitive neuroscience has suggested that manipulation of this narrative grammar elicits similar brain responses as manipulations of syntax in language (i.e. N400, P600, and Left Anterior Negativity effects).[8][9][10]

Comic authorship

Cohn began working in the comic industry at age 14 by helping to run convention booths for Image Comics and Todd McFarlane Productions throughout his teenage years.[11] Beyond illustrating his academic books, Cohn’s creative work appears in several graphic novels, like We the People: A Call to Take Back America (2004) with Thom Hartmann, and illustrations for academic works, including Ray Jackendoff’s A User’s Guide to Thought and Meaning (2012), and the comic strip “Chinese Room” with philosopher Daniel Dennett.

References

Selected works

Footnotes

  1. Zimmer, Carl. 2012. The Charlie Brown Effect: A comic book-artist turned-neuroscientist says the images in Peanuts tap the same brain processes as sentences. Discover Magazine. Pp. 68-70
  2. Robson, David. 2013. How the visual language of comics could have its roots in the ice age. The Guardian. November 23, 2013
  3. Cohn, Neil. 2015. Will emoji become a new language? BBC Futures. October 12, 2015
  4. Gilmore, Garrett. 2015. Help! I can't stop thinking in emoji! VICE. April 21, 2015
  5. Barrett, Brian. 2016. Facebook messenger finally bridges the great emoji divide Wired Magazine. June 16, 2016
  6. Cohn, Neil. 2013. The Visual Language of Comics: Introduction to the Structure and Cognition of Sequential Images. London: Bloomsbury.
  7. Cohn, Neil. 2012. Explaining “I can’t draw”: Parallels between the structure and development of language and drawing. Human Development. 55(4): 167-192
  8. Cohn, Neil. 2013. Visual narrative structure. Cognitive Science. 37(3): 413-452
  9. Cohn, Neil, Martin Paczynski, Ray Jackendoff, Phillip Holcomb, and Gina Kuperberg. 2012. (Pea)nuts and bolts of visual narratives: Structure and meaning in sequential image comprehension. Cognitive Psychology. 65(1): 1-38
  10. Cohn, Neil, Ray Jackendoff, Phillip Holcomb, and Gina Kuperberg. 2014. The grammar of visual narratives: Neural evidence for constituent structure in visual narrative comprehension. Neuropsychologia. 64: 63-70.
  11. Cohen, Georgiana. Drawing Conclusions. Tufts University website. Jan. 26 - Feb. 2, 2009.

External links

This article is issued from Wikipedia - version of the 9/6/2016. The text is available under the Creative Commons Attribution/Share Alike but additional terms may apply for the media files.