Abstract: This paper emphasizes the importance of images as research objects within media texts for the field of linguistics. It provides an overview of different perspectives on multicodal texts that lead to the analysis of multicodal media texts. Alongside a linguistic perspective, approaches from other scientific disciplines such as image science and visual culture are considered. Definitions for the scientific terms mode and code are elaborated. Furthermore, the article outlines differences between multimodal and multicodal, hence stressing the importance of clear definitions as they inevitably lead to different focuses of such linguistic analysis. The author gives a detailed scientific overview on pictures and images from a phenomenological and a semiotic perspective to underline the potential of images as a medium of information within multicodal texts – for which a definition is stated. A multicodal text analysis approach is presented using the example of a multicodal blog entry.
Keywords: code, image, media texts, mode, multicodality, multicodal analysis
Media texts have changed considerably in recent years. Passive information intake is no longer predominant; on the contrary, the media today are all about active participation, whether through interactive TV shows or social media platforms on the Internet. Similar to the behavior of the recipients moving towards participation or even production, media texts have developed further in how and where they appear. They may contain different components: (moving or static) images, verbal language (written or spoken), colors, symbols, emoticons. The list could be extended in many respects. The co-occurrence of images and language can be described as multicodal (cf. 2.4). This form of multicodality will be at the center of my further deliberations.
Let us consider, as examples, the online text types online forum and blog and audiovisual media texts in television and cinema. Just as in a blog, the contributions made to an online forum, serving the communication exchanges between users of a platform, consist as a rule of verbal signs (such as text) and nonverbal signs (images and emoticons). The latter can additionally be moving or static, for instance, hopping about or waving to the user. The inclusion of short video sequences is also possible in some forums. Not only nonverbal signs, but also verbal ones offer the user a whole range of design opportunities: the verbal text can be formatted in some forums and can thus be altered according to preferences or to emphasize the content (cf. Fröhlich 2015). In addition to italics, bold type and underlining, this may include the size, font and color of the verbal (graphic) signs. Therefore, blogs and forums are consequently both varied in their appearance and often actively variable in their design on the part of the user. Depending on the content of the forum or blog, the content may consist more of verbal text parts and additionally integrate nonverbal ones, although communication by image alone is also possible.
Images or pictures as a means of communication are not a novelty: they have long been used for communication, e. g., in cave painting. One characteristic of the modern “visual era” is the growing number of images and their frequent presence (cf. Lobinger 2012, 20), above all on the Internet. The information intake is carried out by way of the visual sensory channel. We cannot assume that a user has necessarily activated the loudspeakers of their Internet-enabled device (e. g., a laptop, a tablet; cf. Fröhlich 2015). Videos in forums are sometimes included with no sound and run without any user intervention on an endless loop. Thus, above all in online forums, images and videos often have an effect purely through their visual appearance.
When focusing on films or TV programs in this respect, we find a whole selection of appropriate material in these media texts. On the other hand, because the latter are usually invariable from the audience point of view, we cannot talk of TV users, but rather of passive viewers – insofar as it is not a question of an interactive TV show. What we see is, on the contrary, usually dominated by moving images, although verbal text can be present, fulfilling in most cases an additional function. The information in film and on television is passed on via the visual and auditive sensory channels.
To sum up, media texts in forums and blogs, compared to films and TV programs, are similar concerning their codes, but the dominance of the verbal and nonverbal parts varies, with the result that certain key aspects of research come to the forefront.
Images dominate in the cinema and on TV; any further investigation would therefore need an analysis of their visual content. In contrast, verbal language signs may dominate the images in online media texts and relegate them to the background – depending on the aims of the texts. The question is whether important information, although relevant for the aims, might be overlooked in a restrictive analysis of particular codes in media texts. Texts consisting of different signs transport their (whole) meaning through the totality of all the signs. This is the reason why this contribution illustrates the necessity of the holistic multicodal analysis of online media texts.
In the following, indications are given of how one may approach multicodal research into online media texts. Initially, I will go into the terms mode and code and explain the possible limitations of the concepts. Secondly, I will examine the image as a scientific concept. This will be followed by a concise presentation of the disciplines which deal with the image in partly differing ways. Subsequently, I will ask what we basically understand as image or picture. Further, I will describe my understanding of a multicodal text concept, in order to demonstrate, with the aid of an example of analysis sequence, why images as a component of a whole text should be considered as an important object of analysis.
When working on code in linguistic research, it is impossible to avoid the widespread term mode or modality. Modus (plural: modi) can describe a procedural method. In linguistics, the term refers to the morphological, syntactic verb forms: indicative, subjunctive or imperative. From an information science viewpoint (e. g., Weidenmann 2002), mode and code refer to different fields of knowledge. Taking selected authors, I would like to take a closer look at different definitions of mode and code.
The two terms mode and code become blurred in use according to the branch of study and the author, as pointed out by, among others, Weidenmann (2002, 46), Schneider/Stöckl (2011, 25), Klemm/Stöckl (2011, 14s.). That makes a conceptual delineation of mode and code all the more important.
Fraas/Meier/Pentzold (2012, 56) thus consider images and language to be differently coded. The authors see the aim of the forms in the paraverbal code, such as bold type or capital letters, as “creating order” (Fraas/Meier/Pentzold 2012, 57). Following the subdivision in Weidenmann (2002), Fraas/Meier/Pentzold (2012, 64s.) initially also go along with the division into code, the type and arrangement of the signs, and mode, with a view to the sensory modalities which are employed for the perception of the signs. In social semiotics, especially through the research work of Kress/van Leeuwen (e. g., 2001), another understanding of mode and code is put forward, in which mainly the term mode is used.
In the following, the different characteristics are examined. Approaches which use these concept traditions in a slightly altered form are presented.
In terms of social semiotics, Kress/van Leeuwen (inter al. 2001) are above all concerned with modality. As researchers into semiotics (e. g., Stöckl or Klemm), they see images and language as signs. By mode they understand:
“[…] a socially shaped and culturally given resource for making meaning. Image, writing, layout, music, gesture, speech, moving image, soundtrack are examples of modes used in representation and communication” (Kress 2009, 54).
The decisive factor is thus the socially and culturally determined conventionalization of a resource, which leads to meaning being ascribed within a communication or presentation (cf. also Fraas/Meier/Pentzold 2012, 65). Furthermore, the connection to the concept of code or to a system of signs becomes clear (cf. Schneider/Stöckl 2011, 26). Modes are linked to a sort of grammar, or a general body of rules (cf. Kress/van Leeuwen 2001). Using these, it is possible to combine the signs of a mode. They thus take on a meaning in certain situations (cf. Stöckl 2004, 11). Colors hold a meaning within culture groups. For example, red means STOP and green GO on the public highway. Language corresponds to the most strongly conventionalized mode (cf. Fraas/Meier/Pentzold 2012, 65).
As Norris (2011, xvi) emphasizes, modes cannot be considered separately, for they are often dependent on one another (cf. Schneider/Stöckl 2011, 26). Norris (2011) mentions the example of a conversation in a restaurant. Perceivable modes here are the communication round the table or at neighboring tables, the background music and also formalities such as the furniture and décor, etc. Communication will be also influenced by the atmosphere created by the music and the environment, as the music will, for instance, have a calming or stimulating influence on the communication partners. The soundtrack in films has a particular influence on the mood of the cinemagoers and is targeted specifically.
As a consequence, Kress/van Leeuwen (2001, 20) define multimodality as “the use of several semiotic modes in the design of a semiotic product or event” (cf. also van Leeuwen 2005). Kress/van Leeuwen (2001) understand by mode any kind of sign whatever in the communication process (cf. Klemm/Stöckl 2011, 15). Mode and code seem to be interchangeable (cf. Klemm/Stöckl 2011, 10s.), even if this is not directly addressed in Kress/van Leeuwen (2001). The term mode is preferentially used (cf. Fraas/Meier/Pentzold 2012, 65).
In the tradition of social semiotics, Stöckl (2006) designates codes as “sign systems […], which regulate the way signs must be put together in order to generate intended meaning”99 (Stöckl 2006, 17). Depending on the degree of conventionalization, he distinguishes furthermore between strong codes (language), because of the “relativ feste Form-Inhalts-Passungen” [‘relatively fixed form-content matches’] (Stöckl 2006, 17) and rather weaker codes (such as images and paraverbal elements), where questions have to be answered regarding the rule conformity of their composition or combination, but also of the segmentability of individual signs (cf. Stöckl 2006, 17).
The vagueness of Kress and van Leeuwen – that mode and code seem interchangeable – is met explicitly by Stöckl (2011), who points out that his concept of modality refers to the type of sign (cf. Stöckl 2011, 45). Given this, modality and codality are interchangeable.
Based on the work of Kress and van Leeuwen, Fricke (2012) distinguishes between multimodality in the broad and narrow sense in her monography “Grammatik multimodal. Wie Wörter und Gesten zusammenwirken” [‘Grammar seen multimodally. How words and gestures interact’]. By multimodality in the narrow sense, she understands a state “[…] when, on the one hand, language types or tokens belong to two different sensory modalities and two different codification media”100 (Fricke 2012, 47) and, on the other hand, are functionally and/or structurally integrated in one code (ibid.). As an example of multimodality in the narrow sense, she offers an emblem to replace the verbal predicate ‘o.k.’ in the very same syntactically placed slot (ibid.). Then again, multimodality in the narrow sense is present according to Fricke (2012), insofar as a code can be shown that is present in two different codification media and belongs to two different sensory modalities (cf. Fricke 2012, 47). Fricke (ibid.) gives the example of gestures underpinning phonetic utterances.
By way of contrast, she describes multimodality in the broad sense exclusively with a view to a code-related media concept. By this, she understands, on the one hand, the use of a sensory channel within two media. She exemplifies this with the dovetailing of script and image in one text. On the other hand, for her there exists language multimodality, i.e., the structural and/or functional integration of “mindestens zwei verschiedene codebezogene Medien” [‘at least two different code-related media’] (Fricke 2012, 48). According to Fricke (2012), this is given when, for example, the deictic word here is made clear with an arrow within a text (cf. Fricke 2012, 48). It is striking that, in her work, the concept of code is used to define mode and modality.
In his article “Multicodierung und Multimodalität im Lernprozess” [‘Multicoding and multimodality in the learning process’] (2002), Weidenmann undertakes an explicit distinction between mode and code. From an information science viewpoint, he argues that by multimodality we may understand “Angebote, die unterschiedliche Sinnesmodalitäten bei den Nutzern ansprechen” [‘offers which address the users’ distinct sensory modalities’] (Weidenmann 2002, 47). Thus, we can say that he uses a sense-related modality concept.
Accordingly, a film or TV program is multimodal, because not only the visual, but also the auditive sensory channel is involved in the information intake.
In comparison, codification may describe “[…] the marking, abbreviation or changing of frequently reoccurring information”101 (Colin 1992, 8; quoted in Weidenmann 2002, 45). Images can be distinguished with regard to the codification of language (cf. Fraas/Meier/Pentzold 2012, 56), because they transport information by means of nonverbal signs. Weidenmann (2002, 47) describes a code as a “Symbolsystem” [‘system of symbols’].102 Multicodal, according to Weidenmann (ibid.), refers to “Angebote, die unterschiedliche Symbolsysteme bzw. Codierungen aufweisen” [‘offers which display distinct symbol systems or codifications’]. He thus supports a clear distinction between multimodal and multicodal, as the former refers to sensory channels and, in consequence, to reception and the latter to the composition of signs.
In the research area of technical communication, there is a tendency to link the concept of multimodality according to Weidenmann (2002) to sensory modalities.
Allwood/Ahlsén (2012) point to the fact that the term multimodal may have different meanings and so postulate a broad multimodality concept with direct reference to communication:
“[…] communication involving more than two of the sensory modalities (sight, hearing, touch, smell and taste). For practical purposes, we will also, besides ‘sensory or perception modalities’, talk about ‘production modalities’, by which we mean human bodily means that normally produce information for the different sensory modalities that is, gestures, speech organs etc.” (Allwood/Ahlsén 2012, 435s.).
Based on the work of Gibbon/Mertens/Moore (2000), Martin/Schultz (2012) define multimodal as a “[…] channel of communication involving more than one human output-input combination, e. g., oral-auditory, gesture-visual […]” (Martin/Schultz 2012, 189).
It may be concluded, in the terms of Weidenmann (2002), that multicodality refers to the use of different sign systems. By way of contrast, the use of different sensory channels in multimodality (e. g., in a film) is the key feature. Multicodality is given when dealing with (moving) images and language by the use of different codes (image and language). In the latter case, only one modality – the visual sensory channel – is called upon, so that we cannot really talk of multimodality. Thus, in analyses of online forums or blogs, we are not carrying out a multimodal, but rather multicodal analysis, in which different visual codes are of central importance (cf., for instance, Fröhlich 2015).
To sum up, the use of mode and code, or multimodality and multicodality, varies according to the field of research. Taken from a social-semiotic viewpoint (inter al. as in Kress/van Leeuwen 2001), the two concepts are not separate. Their use is not always clear-cut, whereby a preference for mode, or the concept of modality, can be observed. In contrast, others like Weidenmann (2002) make an explicit distinction from an information science perspective between mode and code. A sensory channel-related concept of modality is also used in technical communication.
The limitation and separation of the two concepts mode and code can thus be considered as more “precise”. An exact definition demands transparency and sheds light on the terminology connected to the two terms. This is the reason why I am following the definitions of Weidenmann (2002) in my deliberations. I will consider images and language in the following sections, thereby focusing on multicodality. I will, however, be specifically referring to a concept of signs and not symbols.
From a linguistic perspective, it is not a matter of course that images, even if they represent an important part of a media text, should come under analysis as the object of research. Those disciplines working more intensively on images make a serious case for considering the image as the object of scientific research (cf. Lobinger 2012, 16). In the following section, I will briefly introduce those sciences that deal with images, before taking a closer look at concepts of the image.
Several research institutions work on images as the object of their investigations. Some pertinent disciplines are image science, visual communication research, visual culture and image linguistics.
Image science pursues the aim of “giving a general introduction to a basic understanding of pictorial phenomena”103 (Lobinger 2012, 33). It is not confined to particular pictorial media (ibid.). Researchers in this field in part come from the history of art or philosophy.
In contrast to image science, research in visual communication deals with specific pictorial media. The research focuses on, for instance, “spezifische(n) Differenzen zwischen einzelnen Bildmedien” [‘specific differences between individual pictorial media’] (Huber 2004, 16).
In the Anglo-American world, the reception of images is above all discussed within the research area of visual culture. The central issue is the relationship between the viewer of the image and the image itself.
Image linguistics arose in linguistics research that was carried out into the connections and references between language and image (cf. Klemm/Stöckl 2011, 9). For many years, linguistics did not concern itself with the image as an object of research (cf. Klemm/Stöckl 2011, 8). Multicodality can be adduced as an argument for the strengthening of this discipline. Images are used to communicate and/or often refer to language utterances, in the same way as language utterances can be supported or distorted by images (cf. Fröhlich 2015).
The two codes image and language are different in many respects (cf. Nöth 2004). The differences can be classified, for example, using categories developed by Stöckl (2011, 48s.). He distinguishes the two codes with regard to perception, semantics, pragmatics and semiotics. I will single out particular differences as examples to characterize the different codes.
Considering the perception of image and language, we may state that images tend to be holistically perceived, whereas language is perceived successively and in a linear way according to the direction of reading or speed of talking (cf. Nöth 2004, 11). The result is a rapid perception of images and a comparatively slow one for language. Images can also be remembered better than language as a whole (cf. Stöckl 2011, 48). While the message may remain in the memory, the individual words in a text will most probably fade away.
Seen semantically, images may be vague and poorly determined compared to language (cf. Nöth 2004, 20), whereas language offers more precise possibilities of expression, and meanings are rooted in convention (cf. Stöckl 2011, 48).
With regard to their communicative functionality or their pragmatics, images can be distinguished from language through their clarity and vividness, which make it possible to show spatial locations. Language is, on the other hand, more appropriate than images for making logical connections clear (cf. Stöckl 2011, 48). We can thus see that language and images are different from one another in their functionality.
Semiotically, language and images can be seen as different codes because of the different nature of their signs. In the case of images, a continuous flow of signs arises, but in verbal language we find distinct and discreet individual signs. What is more, images as texts are perception-friendly or iconic, because they represent what is to be presented according to its appearance (cf. Nöth 2004, 12). Language is often called arbitrary and is fixed in its meaning by convention (cf. Stöckl 2011, 48). While language in the form of texts – although in another way – can also describe what is to be presented, the arbitrary correlation is located at the level of the word (and not at the level of the text). Thus, the prevalent classification of images as iconic and language as arbitrary is superficially correct, but the comparison takes place at different levels: in the case of images, texts are used as a reference, while in the case of language, the level of the word is the point of reference.
There is no single definition of the concept of the image that is accepted over the boundaries of many disciplines (cf. Halawa 2008, 15ss.). While it is clear in the case of language where one sign begins and ends – where letters end and begin, or where one word ends and another begins – this is not so clear in the case of the image. In the case of the (printed) image, the concept of the sign can, for instance, be used for pixels, i.e., the image-constructing signs. Transitions from one to another of these “signs” are smooth and are usually barely recognizable, if at all. Here, too, the (common) reference to one level, in the case in point with reference to the concept of signs, proves to be problematic, because language and images are different kinds of signs. With reference to the whole image, it remains unclear as to what belongs to the image and what does not (or no longer does). One approach to determining this in the sciences of the image is iconic difference (inter al. Boehm 2006). This describes the ability of the viewer to recognize an image as such. This happens when contrasts to the images and their direct environment are perceived (cf. Boehm 2006, 30).
Mitchell (1990) answers the question “What is an image?” with five types of image: optic (projections, mirror images), graphic (drawings, paintings etc.), perceptual (visions, forms), mental (memories, mind’s eye images) and language-driven images (descriptions, metaphors) (cf. Mitchell 1990, 20). These fine distinctions are compacted by others (as in Müller 2003, 20) into a division between material and immaterial images. With reference to language and images, a distinction is made in comparison to Mitchell’s classification into material and language images (cf. Stöckl 2011, 61ss.; cf. also Nöth 2004, 8). It is mostly as images in the narrow sense that material images are described (cf. Nöth 2004, 8).
According to Sachs-Hombach (2003) and Halawa (2008), definitions of images can be divided into two “Denktraditionen” [‘traditions of thought’] (cf. Halawa 2008, 17) – on the one hand, that of semiotics as a field in linguistics and, on the other, that of phenomenology, a subdiscipline of philosophy:
“While one side defines the image as a sign, the other would like to determine the phenomenon of the image beyond all semiotic categories starting from its visibility” (Halawa 2008, 17, emphasis in the original).104
Both semiotics specialists and phenomenologists are concerned with objects which are perceived, but they are of different opinions, however, “about how perception takes place, whether we have to have signs at our disposal in the process of perception or not”105 (Halawa 2008, 17). Many supporters of phenomenological approaches reject the semiotics perspective on images.
One of these is Brandt (2004). He sees in the concept of the image a purely everyday concept and neither an aesthetic nor a philosophical one. He justifies this with the human process of learning “to identify things as things, people as people and images as images”106 (Brandt 2004, 45). His phenomenological viewpoint becomes clear when he opposes semiotic approaches, underlining visibility as the key feature and seeing it as the necessary quality of the image:
“If visibility is a necessary characteristic of images, then images as such cannot consist of signs or symbols. Characteristic of signs is that they can be transferred to other forms of communication. […] Images cannot be read out aloud nor felt. We can but see them or imagine that we see them. What I see as an image, can, but need not be a sign” (Brandt 2004, 45s.).107
Sachs-Hombach, a researcher in image science, favors a semiotic approach. He also believes that the perception of images is essentially important for their description and/or definition (cf. Sachs-Hombach 2003, 50). He takes the middle ground between semiotics and phenomenology, trying to find common characteristics and not, like Brandt (2004), separating the two. For Sachs-Hombach (2003), an object is an image the moment it appears as an “Element eines Zeichensystems” [‘element in a system of signs’] (Sachs-Hombach 2003, 50). For him,
“images [are] in the narrower sense […] artificial, two-dimensional and relatively lasting objects, which serve within a communicative act to illustrate reality, but also fiction” (Sachs-Hombach 2003, 77).108
The concept of perception itself makes the image a special (pictorial) sign for Sachs-Hombach (2003, 73). This distinguishes it from other signs. Sachs-Hombach (ibid.) also talks of an “eigenständige Form des Zeichenhandelns” [‘independent form of a sign act’]. This becomes clear in online media texts, where images etc. are used in addition to verbal signs as a communicative means of expression. The function of images appears to have almost no limits, as Lobinger (2012) explains:
“Visual elements can complement verbal text illustratively, entertain, serve as design elements or decoration, independently transmit information or bring to mind the unspeakable in an illustrative way. A list of possible image functions could be extended at will” (Lobinger 2012, 19).109
With the help of culturally shared rules and processes, we can attribute a social and communicative sense to acts. “Human communication of all kinds is symbolic, interactive and social action”110 (Klemm 2011, 187). In terms of linguistic pragmatics, Klemm (2011) comments:
“Communicative action can be interpretatively attributed to action patterns or (in the terminology of speech act theory) “illocutions”, e. g., CLAIMING or WARNING. Actions are not, however, objective data” (Klemm 2011, 187; emphasis in the original).111
Based on Lenk (1978), he speaks of actions as “Interpretationskonstrukte” [‘interpretation constructs’] (Klemm 2011, 187), i.e., an action may lead to several, different interpretations. Additionally, speech acts can be carried out by means of different codes or sign systems (e. g., language and images). If several sign systems are used for the communication, then multicodal communication takes place (cf. Weidenmann 2002). For Klemm (2011, 187), this is the normal case. Similarly, online media texts can also contain multicodal communication. In the following, I shall be examining the question of what characterizes the textuality of these texts.
In order to classify online texts, it could be helpful to distinguish them from print texts. The textuality of online texts is, for example, for Pentzold, Fraas and Meier
“not only indicated and made visible by their linguisticality, but in relation to the updated action and sense horizon and to the implemented, resp. perceived, textual, paratextual and paraverbal aspects they are treated as texts” (Pentzold/Fraas/Meier 2013, 83).112
The possibility of updating the content is thus decisive for online media texts. It also becomes clear that the text-ness of online contributions or platforms is not exclusively confined to the verbal text (e. g., a posting), but that also the peritext (cf. Maaß 2014, referring to Genette 1987), like, for instance, information on the user and their status or paraverbal aspects (like the font, color and size of the print) plays an important role (cf. Fröhlich 2015). There are, too, depending on the individual contribution, the nonverbal elements such as (moving) images, smileys or emoticons. Compared to print texts that have appeared offline, online media texts can thus provide much information stemming from their “environment” which can be updated. On the basis of the textuality of online media texts that has been described above, the latter may be subsumed in a broad text concept which comprises different codes.
I consider not only images and emoticons, on the one hand, but also (verbal) language, on the other, as distinct codes, because they are built on different sign systems (cf. Weidenmann 2002). All information transmitted by different codes may in its entirety be considered as online media text. In consequence, we may speak of multicodal texts consisting of verbal, paraverbal and nonverbal signs and characterize these as a “Komplex aus Zeichen verschiedener Art” [‘aggregate of different kinds of signs’] (Fix 2011, 76; cf. also Weidenmann 2002). Through these different and partly intrareferential codes, it is likewise not possible to explain text comprehension alone with the verbal content. Multicodality has to be ascribed an important role here.
Multicodal text comprehension can be described as “the mutual activation and shaping of the meaning potentials of language and images with an aim of reconstructing a relevant, logical and minimal comprehensive message in context”113 (Stöckl 2011, 55). Comprehension is thus the result of cognitive and perceptual processes “which are based on language-image harmonization and a context-sensitive attribution of meaning”114 (Stöckl 2011, 55). We see that the interaction of the codes in use is decisive for multicodal text comprehension.
The following example illustrates multicodal text elements.
Elodie has been posting on her blog for many years. She lives in Paris and loves the world of fashion. She writes about fashion, beauty and lifestyle in her blog “Elodie in Paris”. She sees fashion as a continually unfolding new beginning:
“Analyser les tendances, les couleurs, faire attention aux moindres détails, anticiper, même si, en soit, la mode est un éternel recommencement…”115 (Elodie in Paris).
As may be recognized through the example of this blog, this online media text contains not only verbal signs (script) but also paraverbal (bold type, italics, underlining) and nonverbal signs (image116, a small heart and a star) and is consequently a multicodal (media) text. In the next subsection, I will be taking a closer look at the individual codes and making clear, through the interrelationship of the codes to one another, where the strengths of multicodal analysis lie.
For the scientific investigation of the (different) codes, it remains to be clarified not only which image definition is to be used (insofar as images are part of the texts to be analyzed). Also, the assignment to different codes is by no means trivial, but can, on the contrary, depending on the focus of the analysis, turn out quite differently. There may be agreement on assigning oral language to verbal communication, and, accordingly, written language will normally be assigned to a verbal code. This takes into account the fact that many media texts are oriented not only towards written, but also towards oral language use (cf. Koch/Oesterreicher 22011). We might think of an online forum, in which oral colloquial language is to be found in written form (cf. Fröhlich 2015). Vice versa, television programs may often contain rather formal written language in an oral form, e. g., when the content is read out or reported. Images can be assigned to a nonverbal code, because they do not contain any letters or anything similar in their pure form and, in fact, communicate in another way. There are different opinions on emoticons, for example, which are assigned to the nonverbal (cf., e. g., Fröhlich 2015), but also the paraverbal code (cf., e. g., Aldana Rueda 2011). The latter assignment is eminently plausible the moment we see them as oral forms of expression accompanying language. How far this is really the case, with regard to online media texts, is another question, because there they may also be seen as CMC-specific signs and not purely as compensation for the gestures and mimicry of oral face-to-face communication (cf. Döring 2003, 151s.; Fröhlich 2015). Different opinions on paraverbal phenomena are to be found, too. While discourse markers, like “Oh”, can be discussed under the heading of verbal code – emphasizing solely the verbal phenomenon – it is also possible to describe such markers as paraverbal, insofar as their context-dependent semantic meaning is at the center of attention and is intended to be highlighted in contrast to verbal signs with semantic content (cf. Fröhlich 2015). This follows the basics of oral communication. Therefore, in my opinion, the attribution to a certain code cannot be called right or wrong in one sweeping statement. The decisive factor is the reasoning behind the respective attribution with regard to the object under investigation.
Using the concrete example of the blog posting on “Elodie in Paris” (Fig. 1), it is possible to explain and illustrate the interrelationship of different codes.
The verbal code already directly refers in the headline “Paris, mon amour – Elodie in Paris” to the nonverbal code: in the background, the Eiffel Tower can be clearly seen. Elodie is posing in front of the Paris landmark, thereby expressing her close ties to, or maybe her “love” of, the city. The Eiffel Tower is depicted from top to bottom in the background, while in the foreground only half of Elodie’s face is visible. The famous landmark gains its importance from the central position in the picture.
In her verbal utterances, too, references to the picture are clear. While there is a hint in the French version with her comment “[. . .] malgré son côté tape-à-l’oeil” that she is playing on the striking color of her skirt, this supposition is confirmed in the English translation: “[…] even if it’s pink, metallic and not discreet at all [sic]”. The pink skirt is emphasized by the otherwise muted color mix in the nonverbal code and is the eye-catcher in the picture. Elodie uses bold type to highlight the names of brands and designers and even towns and thus uses the paraverbal code to make her references. On the one hand, the brand names refer to the clothes she is wearing, while on the other, she links them in part with their websites.
The English version of the preceding French verbal utterance is also printed in italics. The highlighting is not carried over. This may mean that the French version must be considered as the more important, whereas the English version is to be seen as a mere translation of the French or as an “extra”, and not as a verbal component of equal value. The peritext at the end of the posting sets itself apart from the main utterance in the text field by means of the typeface, using capitals, and assumes a framing position at the end of the posting.
The example shows that different codes can be harmonized to mutually complement one another.
As explained in the previous sections, the separation of the definitions of mode and code has the advantage of emphasizing the different research aims and of being able to better profile the latter through careful use of the specialist terms. If the analysis refers to media texts, like films, for example, which are received through different sensory perception, then multimodality is at the center of interest, even if multicodality can play a certain role, for instance, in the analysis of images and spoken language. Therefore, the terms of denomination should be chosen consciously, according to the focus of the research.
Especially the investigation of multicodal interconnections, like those of language and images, is seen by Klemm/Stöckl (2011, 9) as an important task for linguistics, which can then contribute to the image sciences and, vice versa, image science findings can enrich linguistics. In the way that image science attaches little importance to language, on the one hand (cf. Klemm/Stöckl 2011, 11), the focus of linguistics, on the other, is too narrowly directed towards language and too little towards images. At the same time, it is images that simply cannot be ignored in present-day media texts.
Visual communication researchers like Lobinger (2012) plead for the recognition of the (private) image in public contexts, e. g., in social networks, as a relevant object of research within media and communication sciences. For their proponents, these images belong to the daily life of many people in the digital age (cf. Lobinger 2012, 16). In support of image linguistics (e. g., Klemm/Stöckl 2011), it is even possible to go a step further.
To go to the extreme, you could say that it is a distortion of reality to “filter apart” language and images, although they are commonly embedded in a whole text amidst other codes or modes. As is clearly shown in the multicodal research by Fröhlich (2015) with the example of forum communication, images represent an important component for the meaning of the whole text which should not be ignored. Images are used in part through their specific characteristics to communicate in other ways. Images can provide cause for (verbal) communication. They transport a share of the meaning in the whole text that is not to be spurned, and which ought to be investigated especially in the field of media linguistics. The analysis of the (whole) meaning of a multicodal (or multimodal) media text requires research into all the codes it contains and into their interreferentiality.
Aldana Rueda, Virginia (2011), Verbale Konstruktion von Expertenstatus in asynchroner computervermittelter Kommunikation. Eine Untersuchung am Beispiel von Newsboards und Social-Networking-Plattformen, <http://www.as.uni-heidelberg.de/ejournal/si2/as-ejournal-si-2011_1.pdf> (17.01.2017).
Allwood, Jens/Ahlsén, Elisabeth (2012), Multimodal Communication, in: Alexander Mehler/Laurent Romary/Dafydd Gibbon (edd.), Handbook of technical communication, Berlin et al., de Gruyter, 435–460.
Boehm, Gottfried (2006), Die Wiederkehr der Bilder, in: Gottfried Boehm (ed.), Was ist ein Bild?, 4th edition, München, Fink, 11–38.
Brandt, Reinhard (2004), Bilderfahrungen – von der Wahrnehmung zum Bild, in: Christa Maar/Hubert Burda (edd.), Iconic turn. Die neue Macht der Bilder, Köln, DuMont, 44–54.
Colin, I. (1992), Optische Kodierung, Universität Frankfurt am Main (unpublished habilitation treatise).
Döring, Nicola (2003), Sozialpsychologie des Internet: die Bedeutung des Internet für Kommunikationsprozesse, Identitäten, soziale Beziehungen und Gruppen, Göttingen, Hogrefe.
Fix, Ulla (2011), Fraktale Narration. Eine semiotisch-textstilistische Analyse, in: Jan Georg Schneider/Hartmut Stöckl (edd.) (2011), Medientheorien und Multimodalität. Ein TV-Werbespot – Sieben methodische Beschreibungsansätze, Köln, von Halem, 70–87.
Fraas, Claudia/Meier, Stefan/Pentzold, Christian (2012), Online-Kommunikation. Grundlagen, Praxisfelder und Methoden, München, Oldenbourg.
Fricke, Ellen (2012), Grammatik multimodal. Wie Wörter und Gesten zusammenwirken, Berlin/Boston, de Gruyter.
Fröhlich, Uta (2015), Facework in multicodaler spanischer Foren-Kommunikation, Berlin/Boston, de Gruyter.
Genette, Gérard (1987), Paratextes, Paris, Seuil.
Gibbon, Dafydd/Mertens, Inge/Moore, Roger K. (edd.) (2000), Handbook of multimodal and spoken dialogue systems: Resources, terminology and product evaluation, Dordrecht, Kluwer Academic Publishers.
Halawa, Mark Ashraf (2008), Wie sind Bilder möglich? Argumente für eine semiotische Fundierung des Bildbegriffs, Köln, von Halem.
Huber, Hans Dieter (2004), Bild Beobachter Milieu. Entwurf einer allgemeinen Bildwissenschaft, Ostfildern, Hatje Cantz.
Klemm, Michael (2011), Bilder der Macht. Wie sich Spitzenpolitiker visuell inszenieren (lassen) – eine bildpragmatische Analyse, in: Hajo Diekmannshenke/Michael Klemm/Hartmut Stöckl (edd.), Bildlinguistik. Theorien – Methoden – Fallbeispiele, Berlin, Schmidt, 187–209.
Klemm, Michael/Stöckl, Hartmut (2011), Bildlinguistik – Standortbestimmung, Überblick, Forschungsdesiderate, in: Hajo Diekmannshenke/Michael Klemm/Hartmut Stöckl (edd.), Bildlinguistik. Theorien – Methoden – Fallbeispiele, Berlin, Schmidt, 7–18.
Koch, Peter/Oesterreicher, Wulf (22011), Gesprochene Sprache in der Romania. Französisch, Italienisch, Spanisch, Berlin/New York, de Gruyter.
Kress, Gunther R. (2009), What is mode?, in: Carey Jewitt (ed.), The Routledge handbook of multimodal analysis, London/New York, Routledge, 54–67.
Kress, Gunther R./van Leeuwen, Theo (2001), Multimodal Discourse: The Modes and Media of Contemporary Communication, London, Arnold.
Lenk, Hans (1978), Handlung als Interpretationskonstrukt. Entwurf einer konstituenten- und beschreibungstheoretischen Handlungsphilosophie, in: Hans Lenk (ed.), Handlungstheorien interdisziplinär, vol. 2.1, München, Fink, 279–350.
Lobinger, Katharina (2012), Visuelle Kommunikationsforschung. Medienbilder als Herausforderung für die Kommunikations- und Medienwissenschaft, Wiesbaden, VS Verlag für Sozialwissenschaften.
Maaß, Christiane (2014), On the markedness of communication on online message boards as part of a perception-oriented politeness approach, in: Kristina Bedijs/Gudrun Held/Christiane Maaß (edd.), Face Work and Social Media, Zürich/Münster, LIT, 237–275.
Martin, Jean-Claude/Schultz, Tanja (2012), Multimodal and Speech Technology, in: Alexander Mehler/Laurent Romary/Dafydd Gibbon (edd.), Handbook of technical communication, Berlin et al., de Gruyter, 189–254.
Mitchell, W. J. Thomas (1990), Was ist ein Bild?, in: Volker Bohn (ed.), Bildlichkeit, Frankfurt am Main, Suhrkamp, 17–68.
Müller, Marion G. (2003), Grundlagen der visuellen Kommunikation. Theorieansätze und Analysemethoden, Konstanz, UVK.
Nöth, Winfried (2004), Zur Komplementarität von Sprache und Bild aus semiotischer Sicht, Mitteilungen des deutschen Germanistenverbandes 51:1, 8–22.
Norris, Sigrid (2011), Identity in (Inter)action. Introducing multimodal (inter)action analysis, Berlin/New York, de Gruyter.
Pentzold, Christian/Fraas, Claudia/Meier, Stefan (2013), Online-mediale Texte: Kommunikationsformen, Affordanzen, Interfaces, Zeitschrift für Germanistische Linguistik 41:1, 81–101.
Sachs-Hombach, Klaus (2003), Das Bild als kommunikatives Medium. Elemente einer allgemeinen Bildwissenschaft, Köln, von Halem.
Schneider, Jan Georg/Stöckl, Hartmut (edd.) (2011), Medientheorien und Multimodalität. Ein TV-Werbespot – Sieben methodische Beschreibungsansätze, Köln, von Halem.
Stöckl, Hartmut (2004), In between modes. Language and image in printed media, in: Eija Ventola/Cassily Charles/Martin Kaltenbacher (edd.), Perspectives on multimodality, Amsterdam/Philadelphia, Benjamins, 9–30.
Stöckl, Hartmut (2006), Zeichen, Text und Sinn – Theorie und Praxis der multimodalen Textanalyse, in: Eva Martha Eckkrammer/Gudrun Held (edd.), Textsemiotik. Studien zu multimodalen Texten, Frankfurt am Main/New York, Lang, 11–36.
Stöckl, Hartmut (2011), Sprache-Bild-Texte lesen. Bausteine zur Methodik einer Grundkompetenz, in: Hajo Diekmannshenke/Michael Klemm/Hartmut Stöckl (edd.), Bildlinguistik. Theorien – Methoden – Fallbeispiele, Berlin, Schmidt, 45–70.
van Leeuwen, Theo (2005), Introducing social semiotics, London/New York, Routledge.
Weidenmann, Bernd (2002), Multicodierung und Multimodalität im Lernprozess, in: Ludwig J. Issing/Paul Klimsa (edd.), Information und Lernen mit Multimedia und Internet. Lehrbuch für Studium und Praxis, 3rd edition, Weinheim, Beltz, 45–62.
Elodie in Paris (2015), ABOUT, <http://www.elodieinparis.com/about-elodie-in-paris> (17.01.2017).
Elodie in Paris (2014), PARIS, MON AMOUR – ELODIE IN PARIS, <http://www.elodieinparis.com/2014/12/23/paris-mon-amour> (17.01.2016).
3.144.232.71