Visual inputs decrease brain activity in frontal areas during silent lipreading
Authors:
Julio Plata Bello aff001; Carlos García-Peña aff001; Cristián Modroño aff001; Estefanía Hernández-Martín aff001; Yaiza Pérez-Martín aff003; Francisco Marcano aff001; José Luis GonzálezMora aff001
Authors place of work:
Department of Physiology, Faculty of Medicine, University of La Laguna, S/C de Tenerife, Spain
aff001; Hospital Universitario de Canarias (Department of Neurosurgery), S/C de Tenerife, Spain
aff002; Hospital Universitario de Canarias (Department of Neurology), S/C de Tenerife, Spain
aff003
Published in the journal:
PLoS ONE 14(10)
Category:
Research Article
doi:
https://doi.org/10.1371/journal.pone.0223782
Summary
Aim
The aim of the present work is to analyze the modulation of the brain activity within the areas involved in lipreading when an additional visual stimulus is included.
Methods
The experiment consisted of two fMRI runs (lipreading_only and lipreading+picture) where two conditions were considered in each one (oral speech sentences condition [OSS] and oral speech syllables condition [OSSY]).
Results
During lipreading-only, higher activity in the left middle temporal gyrus (MTG) was identified for OSS than OSSY; during lipreading+picture, apart from the left MTG, higher activity was also present in the supplementary motor area (SMA), the left precentral gyrus (PreCG) and the left inferior frontal gyrus (IFG). The comparison between these two runs revealed higher activity for lipreading-only in the SMA and the left IFG.
Conclusion
The presence of a visual reference during a lipreading task leads to a decrease in activity in frontal areas.
Keywords:
functional magnetic resonance imaging – Neuroimaging – Sensory perception – Vision – Language – Tongue – Syllables – Speech
Introduction
Lipreading is the ability to visually perceive the speech of others by viewing the movements of their lips, teeth, tongue and jaw [1]. This ability may be considered a complex brain process because it engages sensory, perceptual, memory, cognitive and linguistic elements. Lipreading improves auditory speech discrimination [2] and it is particularly important for the hearing impaired population [2,3]. There is a larger inter-individual variability in lipreading ability (within and across normal hearing and deaf populations) than in auditory speech perception [4]. In this sense, the more skilled lip-readers are able to perceive all levels of speech patterns (from syllables to connected speech) [5]. Understanding how the lipreading ability can improve is essential given that lipreading can substitute for hearing in the education of deaf children [6].
Many brain regions involved in speech production and comprehension are also involved during lipreading. A robust activation in the superior and middle temporal gyrus (STG and MTG), the inferior frontal gyrus (IFG), the inferior parietal lobule (IPL) and the occipito-temporal junction is normally found [1,5,7,8]. Furthermore, there is usually greater activation in the left than the right hemisphere [1,9]. The implication of parietal and frontal regions during lipreading is also supported by the theoretical framework of the mirror neuron system (MNS) [10] and it is also compatible with the longstanding motor theory of speech perception [11]. On one hand, the MNS is formed by a group of regions located in the IPL and IFG bilaterally and they are activated during the execution and the observation of a motor action [12]. Bearing this in mind, considering the speaking language as a sequence of motor actions performed with the mouth and the tongue, IPL and IFG regions are going to be activated during speaking and the observation of speaking [10]. On the other hand, the motor theory of speech perception considered that phonetic information is perceived by detecting the gestures of the speaker [11]. Therefore, both theories are complementary, with the MNS giving an anatomical and functional support to the theory proposed by Liberman (1985).
Furthermore, the Dual Stream model proposed by Hickock and Poeppel (2007) also supports the presence of a lipreading network. Their model suggests a dorsal processing stream including frontal areas that process articulatory movements for the purpose of speech production and perception (IFG and dorsal premotor cortex), as well as a ventral stream formed by posterior areas which parse auditory speech (STG and MTG) and integrate phonology with articulatory movements (at the IPL) [13]. These regions are also known to be engaged even when syllables are presented only-visually [9] though it seems that modality-specific representations exist only to the level of the whole words [5]. Thus, the ventral stream is involved in processing speech signals for comprehension and the dorsal stream is involved in translating acoustic signals into articulatory representations [14].
The brain network involved in speech comprehension is known to be modulated by visual inputs. The recognition of perfectly audible speech is impaired when it is combined with visible facial movements that are incongruent with the acoustic signal [15]. A widely-known example of this is the McGurk effect, in which the auditory perception is modified by a visual input (e.g. an auditory /ba/ dubbed onto a visual /ga/ is heard as /da/) [16]. Using neuroimaging methods, Kislyuk et al (2008) also reported a modulation of the auditory cortex by visual information [17]. The visual modulation of speech perception may generate comprehension problems, mostly in the deaf population. This is especially important in educational and social contexts, where an inadequate speech perception can lead to significant misunderstandings. Therefore, the visual system modulates speech comprehension and this has a social relevance. However, this effect has not been completely studied. For example, what happens when only lipreading is considered (i.e. without auditory information) has not been investigated. This may be important for understanding the modulation of the visual system of lipreading perception in deaf people. In this regard, the presence of other visual stimuli (different from the mouth and tongue movements, e.g. objects) while lipreading may modulate the activity in brain regions involved in lipreading and, consequently, in speech comprehension.
Bearing all the above in mind, the aim of the present fMRI work is to study the modulation of the brain activity within the areas involved in lipreading when an additional visual stimulus is included. The experiment consisted of two fMRI runs: lipreading_only and lipreading+picture (i.e. the additional visual stimulus) with two studied conditions each: oral speech sentences (OSS) and oral speech syllables (OSSY) (i.e. the control condition).
Methods
Subjects
Thirty-six healthy, right handed subjects participated in the study (20 women), with an average age of 26.6 (SD = 6.64). All of participants were university students and they have no history of speech, writing and/or reading disability. Written informed consent was explained and signed. The study was approved by the University of La Laguna Ethics Committee, according to the Declaration of Helsinki.
Data acquisition and processing
Data for the experiment were collected at the Magnetic Resonance for Biomedical Research Service of the University of La Laguna. Functional images were obtained on a 3T General Electric (Milwaukee, WI, USA) scanner using an echo-planar imaging gradient-echo sequence and an 8 channel head coil (TR = 3000ms, TE = 21ms, flip angle = 90°, matrix size = 64 × 64 pixels, 57 slices/volume, spacing between slices = 1 mm, slice thickness = 3 mm). The slices were aligned to the anterior commissure–posterior commissure line and covered the whole cranium. Functional scanning was preceded by 18 s of dummy scans to ensure tissue steady-state magnetization.
A whole-brain three-dimensional structural image was acquired for anatomical reference. A 3D fast spoiled gradient–recalled pulse sequence was obtained with the following acquisition parameters: TR = 10.4 ms, TE = 4.2 ms, flip angle = 20, matrix size = 512 × 512 pixels, .5 × .5 mm in plane resolution, spacing between slices = 1 mm, slice thickness = 2 mm.
After checking the images for artefacts, data were preprocessed and analyzed using Statistical Parametric Mapping software SPM8 (Wellcome Trust Centre for Neuroimaging; http://www.fil.ion.ucl.ac.uk/spm/) and displayed using xjView 9 (http://www.alivelearn.net/xjview/). The images were spatially realigned, unwarped, and normalized to the Montreal Neurological Institute (MNI) space using standard SPM8 procedures. The normalized images of 2 × 2 × 2 mm were smoothed by a full width at half maximum (FWHM) 8 × 8 × 8 Gaussian kernel.
According to the PLOS Data policy the data used in the present manuscript have uploaded in a public repository (https://figshare.com/articles/Lipreading_fMRI_study/9904661).
Study design
Two fMRI runs were performed: lipreading_only and lipreading+picture (Fig 1). The order of the runs was counterbalanced among participants. During each run, participants had to watch a series of four video clips. All video clips showed a male or female speaker whose face was visible only below the nose. It has been shown that facial images showing mouth shape, lips, tongue and teeth position offer potential information about the filter state of the vocal tract [18].
Two conditions were considered in each run: oral speech sentences (OSS) and oral speech syllables (OSSY). The OSS clips consisted of facial movements saying ten short sentences (five sentences in each clip, each sentence lasted 3 seconds) where an object was denominated in Spanish. For example: “This is a helicopter”. The OSSY clips were composed of facial movements repetitively saying ten nonsense syllables during 3 seconds (/la/, /le/, /lo/, /lu/, /pa/, /pe/, /pi/, /ta/, /te/, /to/) (five syllables in each clip).
The videos were projected 12 times for 15 seconds each. OSS and OSSY conditions were presented in a randomized order and there was a 15 second cross fixation (a break with participants watching a black screen with a white cross in the centre) between each condition. In addition, the fMRI run lipreading+picture showed a picture of an object accompanying the speaker video. These pictures were selected from the Boston naming test [19] and they changed every 3 seconds, thus 5 pictures were shown in each video-clip. In the OSS condition, the picture shown coincided with the object named in the video (i.e. when the speaker was saying “this is a helicopter”, a picture of a helicopter was shown).
Simple T contrasts
A block design in the context of a general linear model was used, for individual subject analyses (first level), to look for differences in brain activity between OSS and OSSY within and between lipreading_only and lipreading+picture runs. The considered contrasts in the analysis were as follows: OSS > OSSY and OSS < OSSY in each run. The first-level contrast images were then used in a random-effects group analysis (second level). Group analysis was performed using the random effect approach, using an ANOVA design and including the age, gender and Edinburgh Handedness Inventory score [20] as covariates. The statistical significance was stablished using the Family Wise Error rate (FWE) = 0.05, with a minimum cluster size of ten voxels.
Region of interest (ROI) analysis
In addition to the voxel-wise analysis, we also conducted a region of interest (ROI) analysis in regions that have been implicated in lipreading[5]. ROIs of the IFG (Brodman areas [BA] 44 and 45); STG and MTG (BA 21, 22, 41 and 42); and IPL (BA 39 and 40) were created using the WFU Pickatlas software[21]. These images were used for a ROI analysis using MarsBaR 0.44 toolbox (http://marsbar.sourceforge.net/) in the contrasts previously described. The significance within this analysis was a corrected P-value = 0.05.
Results
Oral speech sentences (OSS) vs. oral speech syllables (OSSY)
During the lipreading+picture run, higher activity in the left MTG (Brodman areas [BA] 21) was identified for OSS than OSSY (Table 1, Fig 2A). The lipreading_only run also showed higher activity in the left MTG in the contrast OSS > OSSY but, additionally, this contrast showed higher activity in the supplementary motor area (SMA), the left precentral gyrus (PreCG), left superior frontal gyrus (SFG) and left IFG (Table 1, Fig 2A). The opposite contrast (OSSY > OSS) did not lead to significant activations in any of the runs with the selected level of significance (FWE = 0.05).
Lipreading+picture vs. lipreading_only
The comparison of the contrast OSS > OSSY between the two different runs revealed the presence of higher activity in the SMA and the left IFG during lipreading_only, whereas higher activity in the left medial frontal gyrus and the left posterior cingulate gyrus was identified for lipreading+picture (Table 1, Fig 2B).
ROI analysis
A region of interest (ROI) analysis in regions that have been implicated in lipreading was performed[5]. ROIs of the IFG (BA 44 and 45); STG and MTG (BA 21, 22, 41 and 42); and IPL (BA 39 and 40) in the left hemisphere were created. The ROI analysis showed a significant increase of activity in the primary associative auditory cortex (STG and MTG, BA 21) when the contrast OSS > OSSY was considered in both runs. Furthermore, the left IFG (BA 44 and 45, Broca’s area) showed higher activity during OSS than OSSY when there was no picture and when compared with the other fMRI run (S1 Table).
Discussion
The present work has studied the differences of brain activity in the lipreading network when the lipreading was accompanied with or without a visual stimulus (i.e. a picture of the object that was being named). The presence of an additional visual stimulus during a lipreading task led to a different pattern of brain activity in frontal areas (left IFG and SMA). This finding will be discussed below.
As described above, the activation of language related areas is highly consistent during lipreading tasks, using either words or syllables [9]. In fact, it is widely accepted that syllable-production related movements elicited a similar pattern of brain activity compared to spoken-words production related movements [5,22]. Here, the syllable lip-reading (OSSY condition) has been used as the control condition and it has been compared with a word/sentence lipreading (OSS condition). Thus, the pattern of brain activity that has been obtained should not be considered as the pattern of activity during a lipreading task, but the pattern of a lipreading task with a specific semantic component (i.e. a sentence naming an object). Bearing this in mind, the presence of the object that was named while lipreading the same object modulated the activity of regions that are normally activated during speech production and comprehension. In this sense, when the picture was absent, higher activity in frontal areas (IFG and SMA) was detected than when the picture was present.
Using a similar paradigm to the present work (i.e. a comparison between lexical [bisyllabic high-frequency words] lipreading vs. non lexical [bisyllabic legal nonwords] lipreading), Paulesu et al (2003), in a positron emission tomography (PET) study, described a similar pattern of brain activation as the present work, with higher activation for the lexical lipreading task (bisyllabic high-frequency words) in the L-IFG, L-MFG, L-PreCG and the R-IFG [22]. The higher implication of frontal regions during the OSS condition than OSSY in the lipreading_only task is assumed by the requirement of further activation in language related areas implicated in access to lexical semantic knowledge [23,24]. Nowadays, it is widely accepted that the left IFG (where Broca’s area is located) is involved in motor speech production as well as speech perception. Phonological processing is a necessary means of accessing lexical and semantic information during visual speech [1]. Thus, the left IFG is not only involved in articulatory-based mechanisms of speech perception, but also has a key role in language comprehension, supporting the information of syntactic and semantic structure [25,26].
However, different activity in frontal areas did not appear when the lipreading task is accompanied with a picture (i.e. the lipreading+picture task). In this respect, the presence of a visual stimulus (i.e. a picture) seems to modulate the activity during a lipreading task, leading to a decrease in frontal activity or an increase in frontal activity during the control condition (i.e. OSSY). This finding may be associated with a different use of the dorsal and ventral pathways of the speech processing system between the studied conditions. As stated above, the dual-loop model makes a distinction between a dorsal and a ventral pathway that anatomically consists of white matter tracts running either above or below the Sylvian fissure [27,28]. During lipreading comprehension, both streams are also involved, but there is a dominance of the dorsal stream, activated in this case, not by acoustic signals, but visual ones (i.e. mouth, lips and tongue movements). The ventral stream seems to be less activated during lipreading tasks when there is no semantic content (i.e. OSSY condition). Nevertheless, when an additional visual stimulus is incorporated (i.e. a picture of an object), the ventral stream is activated by the meaning of this picture and it is strong enough to eliminate the differences that exist between OSS and OSSY conditions when there are no additional visual stimuli. Changing the visual stimuli used here (i.e. a picture) for another visual cue (e.g. picture with parts of the object, a sentence, a word or even letters) would also lead to an activation of the ventral stream, although this activation would be variable according to the cue type. Furthermore, bearing in mind that the pictures that appear during lipreading+picture task were the same either during OSS or OSSY, the information that reached frontal areas via the ventral stream was the same in both conditions. This information activated per se frontal speech areas that are closely tied to the motor plan that has to be executed to produce the same speech movement [29]. The presentation of an incongruent visual stimuli would lead to different results [30].
Although the dual stream model may explain the existence of similar IFG activity in both conditions during lipreading+picture run, it would not explain the absence of SMA differences, which seem to be more associated with the information that flows through the dorsal stream. In any case, connectivity between IFG and SMA has already been described via the corticocortical frontal aslant tract, which allows language production, but not only its elaboration but also its monitoring [31]. Thus, the absence of differences in the activation of the IFG between OSS and OSSY conditions may be related to an activation of the ventral stream that led to the same activation because the stimulus was the same (i.e. object picture). Connections between IFG and SMA may also explain the absence of differences at this level.
Finally, higher activation in the left temporal lobe was identified in the OSS than in the OSSY on both fMRI runs. More specifically the left MTG, consistent with the primary associative auditory region (BA 21), showed this higher activity (Fig 2A). The presence of high level semantic elements during OSS could be the explanation for this higher activity in the left MTG because this area is more strongly recruited when the speech has more linguistic complexity [1]. Therefore, lipreading tasks recruit temporal areas involved in auditory processing (through thalamic connections or direct inputs from the visual cortex [32]) and they can be modulated by the semantic component of the task.
Bearing in mind the findings of the present work, one can make a parallelism between the modulation of lipreading processing by the presence of an additional visual stimulus reported here and the McGurk effect, which consists of a modification of the auditory perception by a visual input [16]. In any case, the identification of a modulatory effect of language perception (in any of their forms) is of great interest, because it allows a better understanding of language processing and, moreover, controlling those factors, one can improve communication. This may be particularly important in patients with difficulties in language processing as well as patients with any auditory impairment. In any case, more studies (preferably those focusing on patients with awake neurosurgical procedures and direct cortical stimulation procedures) are necessary to replicate the findings of the present study and to improve the knowledge about the factors influencing language processing.
Conclusion
Lipreading perception can be modulated by the presence of a different visual stimulus. During a lipreading task without an additional visual stimulus, higher activation of frontal areas (including IFG and SMA) was identified when observation of speech sentences is compared with observation of speech syllables. However, when an additional stimulus was added, that difference disappeared, and a similar pattern of activity between both conditions was identified.
Supporting information
S1 Table [doc]
ROI analysis (statistical significance was considered when corrected p-value was below 0.05).
Zdroje
1. Hall DA, Fussell C, Summerfield AQ. Reading fluent speech from talking faces: typical brain networks and individual differences. J Cogn Neurosci. 2005;17: 939–53. doi: 10.1162/0898929054021175 15969911
2. Grant KW, Seitz PF. The use of visible speech cues for improving auditory detection of spoken sentences. J Acoust Soc Am. 2000;108: 1197–208. doi: 10.1121/1.1288668 11008820
3. Grant KW, Seitz PF. Measures of auditory-visual integration in nonsense syllables and sentences. J Acoust Soc Am. 1998;104: 2438–50. doi: 10.1121/1.423751 10491705
4. MacLeod A, Summerfield Q. Quantifying the contribution of vision to speech perception in noise. Br J Audiol. 1987;21: 131–41. 3594015
5. Bernstein LE, Liebenthal E. Neural pathways for visual speech perception. Front Neurosci. 2014;8: 386. doi: 10.3389/fnins.2014.00386 25520611
6. Jeffers J, Barley M. Speechreading: (Lipreading). Charles C Thomas Pub Ltd (1971);
7. Campbell R, MacSweeney M, Surguladze S, Calvert G, McGuire P, Suckling J, et al. Cortical substrates for the perception of face actions: an fMRI study of the specificity of activation for seen speech and for meaningless lower-face acts (gurning). Brain Res Cogn Brain Res. 2001;12: 233–43. 11587893
8. Calvert GA, Campbell R, Brammer MJ. Evidence from functional magnetic resonance imaging of crossmodal binding in the human heteromodal cortex. Curr Biol. 2000;10: 649–57. doi: 10.1016/s0960-9822(00)00513-3 10837246
9. Calvert G a, Campbell R. Reading speech from still and moving faces: the neural substrates of visible speech. J Cogn Neurosci. 2003;15: 57–70. doi: 10.1162/089892903321107828 12590843
10. Rizzolatti G, Arbib MA. Language within our grasp. Trends Neurosci. 1998;21: 188–94. doi: 10.1016/s0166-2236(98)01260-0 9610880
11. Liberman AM, Mattingly IG. The motor theory of speech perception revised. Cognition. 1985;21: 1–36. 4075760
12. Cattaneo L, Rizzolatti G. The mirror neuron system. Arch Neurol. 2009;66: 557–60. 19433654
13. Hickok G, Poeppel D. The cortical organization of speech processing. Nat Rev Neurosci. 2007;8: 393–402. doi: 10.1038/nrn2113 17431404
14. Hickok G. The cortical organization of speech processing: feedback control and predictive coding the context of a dual-stream model. J Commun Disord. 45: 393–402. doi: 10.1016/j.jcomdis.2012.06.004 22766458
15. Dodd B. The role of vision in the perception of speech. Perception. 1977;6: 31–40. doi: 10.1068/p060031 840618
16. McGurk H, MacDonald J. Hearing lips and seeing voices. Nature. 264: 746–8. doi: 10.1038/264746a0 1012311
17. Kislyuk DS, Möttönen R, Sams M. Visual processing affects the neural basis of auditory discrimination. J Cogn Neurosci. 2008;20: 2175–84. doi: 10.1162/jocn.2008.20152 18457500
18. Summerfield Q. Lipreading and audio-visual speech perception. Philos Trans R Soc Lond B Biol Sci. 1992;335: 71–8. doi: 10.1098/rstb.1992.0009 1348140
19. Kaplan E, Goodglass H, Weintraub S. Boston Naming Test. Lea, Febiger, editors. Philadelphia; 1983.
20. Oldfield RC. The assessment and analysis of handedness: the Edinburgh inventory. Neuropsychologia. 1971;9: 97–113. doi: 10.1016/0028-3932(71)90067-4 5146491
21. Maldjian JA, Laurienti PJ, Kraft RA, Burdette JH. An automated method for neuroanatomic and cytoarchitectonic atlas-based interrogation of fMRI data sets. Neuroimage. 2003;19: 1233–9. 12880848
22. Paulesu E, Perani D, Blasi V, Silani G, Borghese NA, De Giovanni U, et al. A functional-anatomical model for lipreading. J Neurophysiol. 2003;90: 2005–13. doi: 10.1152/jn.00926.2002 12750414
23. Paulesu E, Goldacre B, Scifo P, Cappa SF, Gilardi MC, Castiglioni I, et al. Functional heterogeneity of left inferior frontal cortex as revealed by fMRI. Neuroreport. 1997;8: 2011–7. doi: 10.1097/00001756-199705260-00042 9223094
24. Poldrack RA, Wagner AD, Prull MW, Desmond JE, Glover GH, Gabrieli JD. Functional specialization for semantic and phonological processing in the left inferior prefrontal cortex. Neuroimage. 1999;10: 15–35. doi: 10.1006/nimg.1999.0441 10385578
25. Friederici AD. Towards a neural basis of auditory sentence processing. Trends Cogn Sci. 2002;6: 78–84. 15866191
26. Bookheimer S. Functional MRI of language: new approaches to understanding the cortical organization of semantic processing. Annu Rev Neurosci. 2002;25: 151–88. doi: 10.1146/annurev.neuro.25.112701.142946 12052907
27. Umarova RM, Saur D, Schnell S, Kaller CP, Vry M-S, Glauche V, et al. Structural connectivity for visuospatial attention: significance of ventral pathways. Cereb Cortex. 2010;20: 121–9. doi: 10.1093/cercor/bhp086 19406904
28. Rijntjes M, Weiller C, Bormann T, Musso M. The dual loop model: its relation to language and other modalities. Front Evol Neurosci. 2012;4: 9. doi: 10.3389/fnevo.2012.00009 22783188
29. Fridriksson J, Moss J, Davis B, Baylis GC, Bonilha L, Rorden C. Motor speech perception modulates the cortical language areas. Neuroimage. 2008;41: 605–13. doi: 10.1016/j.neuroimage.2008.02.046 18396063
30. Plata Bello J, Modroño C, Marcano F, González-Mora JL. Modulation in the mirror neuron system when action prediction is not satisfied. Eur J Neurosci. 2015;41. doi: 10.1111/ejn.12850 25722245
31. Duffau H. Stimulation mapping of white matter tracts to study brain functional connectivity. Nat Rev Neurol. 2015;11: 255–65. doi: 10.1038/nrneurol.2015.51 25848923
32. Schroeder CE, Lakatos P, Kajikawa Y, Partan S, Puce A. Neuronal oscillations and visual amplification of speech. Trends Cogn Sci. 2008;12: 106–13. doi: 10.1016/j.tics.2008.01.002 18280772
Článek vyšel v časopise
PLOS One
2019 Číslo 10
- S diagnostikou Parkinsonovy nemoci může nově pomoci AI nástroj pro hodnocení mrkacího reflexu
- Je libo čepici místo mozkového implantátu?
- Metamizol jako analgetikum první volby: kdy, pro koho, jak a proč?
- Pomůže v budoucnu s triáží na pohotovostech umělá inteligence?
- AI může chirurgům poskytnout cenná data i zpětnou vazbu v reálném čase
Nejčtenější v tomto čísle
- Correction: Low dose naltrexone: Effects on medication in rheumatoid and seropositive arthritis. A nationwide register-based controlled quasi-experimental before-after study
- Combining CDK4/6 inhibitors ribociclib and palbociclib with cytotoxic agents does not enhance cytotoxicity
- Experimentally validated simulation of coronary stents considering different dogboning ratios and asymmetric stent positioning
- Risk factors associated with IgA vasculitis with nephritis (Henoch–Schönlein purpura nephritis) progressing to unfavorable outcomes: A meta-analysis
Zvyšte si kvalifikaci online z pohodlí domova
Všechny kurzy