-
Notifications
You must be signed in to change notification settings - Fork 2
/
Copy path02-chap2.Rmd
executable file
·255 lines (149 loc) · 37.8 KB
/
02-chap2.Rmd
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
# Methodological framework {#chap2}
```{r setupCH2, echo = FALSE, message = FALSE, warning = FALSE, results = "hide"}
knitr::opts_chunk$set(
cache = TRUE, echo = FALSE, warning = FALSE, message = FALSE,
out.width = "100%",
fig.pos = "ht",
fig.align = "center"
)
```
\epigraph{"[...] we have been brainwashed by Fisherian statistics into thinking that refutation of $H_{0}$ is a powerful way of testing substantive theories"}{Meehl, P. (1986)}
\initial{I}n this chapter we briefly introduce some of the key concepts related to the methods we used in our work. More precisely, we cover the technical concepts related to speech production mechanisms, to electromyography and to our statistical approach. Finally, we give an overview of the following chapters.
## Biomechanical aspects of speech production
### Vocal apparatus
Speech production requires the involvement of more than 100 muscles in the face, the neck and the chest [@simonyan_laryngeal_2011]. The activity of these muscles is coordinated to produce an air flow moving from the lungs to the oral and nasal cavities, via the trachea, the larynx and the pharynx (see Figure \@ref(fig:vocal)). Broadly speaking, speech production can be said to consist essentially in i) *phonation*, which refers to the control of the air flow and to the vibration of the vocal folds and ii) *articulation*, which refers to movements of the *articulators*. The action of the *articulators* (e.g., lips, tongue) is to shape the oral and nasal cavities, resulting in modifications of the sound waves and in the production of different vowels and consonants.
```{r vocal, out.width = "50%", fig.cap = "Human respiratory and phonatory system. Figure from the OpenStax \\textit{Anatomy and Physiology} Textbook. Download for free at \\url{https://cnx.org/contents/14fb4ad7-39a1-4eee-ab6e-3ef2482e3e22@15.1}."}
knitr::include_graphics("assets/vocal.jpg")
```
The characteristics of the vocal folds (e.g., their length or thickness) influence what is known as the *fundamental frequency* (or F0) of the speech signal, which in turn determines the perceived pitch of the voice. The speech signal can be further decomposed in *resonant frequencies* or *formants*. We can relate changes in the state of the articulatory system with changes in the formant (and especially the F1-F2) space (see Figure \@ref(fig:ipa)). Indeed, the frequency of the first formant (F1) is mostly determined by the height of the tongue body (or rather by the aperture of the oral cavity) whereas the frequency of the second formant (F2) is mostly determined by the frontness/backness of the tongue body. For instance, when producing the /u/ vowel, the oral cavity is in a close position and the tongue lies in the back of the oral cavity (and the lips are rounded). However, when producing the /a/ vowel, the oral aperture is larger (and the lips are widely opened).
```{r ipa, out.width = "75%", fig.cap = "Illustration of the vocalic 'quadrilateral' and the relation between vowels and formants (F1 and F2). Figure adapted from the International Phonetic Association (2015) - IPA Chart, available under a Creative Commons Attribution-Sharealike 3.0 Unported License."}
knitr::include_graphics("assets/vowels.pdf")
```
In brief, modifications in the shape of the *vocal tract* result in the production of different vowels. Changes in the configuration of articulators such as the lips or the tongue may also produce consonants. Consonants are produced by applying some form of constriction to (or by closing) the vocal tract to constraint the air flow. We usually classify consonants according to *where* (the *place of articulation*) and *how* (the *manner of articulation*) this constriction takes place (see Figure \@ref(fig:consonants)). For instance, consonants such as /p/ or /b/ are produced by closing the lips together and are therefore known as *bilabial* plosives.
```{r consonants, out.width = "100%", fig.cap = "Table of consonants according to the manner (in rows) and place (in columns) of articulation. Figure from the International Phonetic Association (2015) - IPA Chart, available under a Creative Commons Attribution-Sharealike 3.0 Unported License."}
knitr::include_graphics("assets/consonants.png")
```
To sum up, besides being an essential communication tool for humans, speech production is also a complex motor action, involving the fine-grained coordination of numerous laryngeal and supralaryngeal structures. In the next section, we discuss in more details the specific facial muscles that were of interest in the present work.
### Orofacial speech muscles
In our work, we were especially interested in the activity of some of the orofacial muscles (i.e., the muscles situated around the mouth). More precisely, we studied the activity of the *orbicularis oris inferior* (OOI), the *orbicularis oris superior* (OOS), and the *zygomaticus major* (ZYG) muscles (cf. Figure \@ref(fig:facial)). Contrary to what was assumed until recently, the *orbicularis oris* muscle is not a sphincter muscle but is instead a complex of several distinct muscles that interlace in a way that gives the *orbicularis oris* complex its circular aspect. Among these muscles, the OOS and the OOI are placed above and below (respectively) the mouth and are responsible for rounding or protruding the lips. The ZYG muscle has its origin on the zygomatic bone and inserts at the labial commissure (the angle of the mouth) where it meets with fibers of the *levator anguli oris* and *orbicularis oris* muscles. Together with the *levator anguli oris*, it serves to move the labial commissure upwards and laterally, and is involved in laughing and more generally in pleasant reactions and positive emotions. It is also involved in speech production, especially during the production of *spread* sounds, that is, sounds that require a wide horizontal aperture of the mouth (e.g., /i/).
```{r facial, out.width = "75%", fig.cap = "Illustration of the main facial muscles of interest in the present work. Figure adapted from Patrick J. Lynch, medical illustrator, http://patricklynch.net."}
knitr::include_graphics("assets/facial_muscles.pdf")
```
For sensors placement, we followed guidelines and recommendations from @fridlund_guidelines_1986. In addition to speech-related orofacial muscles, we also routinely recorded the activity of other facial muscles such as the *frontalis* muscle (FRO) in Chapter \@ref(chap3), \@ref(chap4), \@ref(chap5), and the *corrugator supercilii* muscle (COR) in Chapter \@ref(chap5). The activity of these muscles was monitored to control for non speech-related facial muscular activity [as recommended by @garrity_electromyography_1977].
There are several ways to probe the involvement of specific articulators in a given speech production task. For instance, it is possible to selectively interfere with the activity of some articulators (or groups of articulators) to demonstrate their necessary involvement in this particular task. It is also possible to record the activity of facial muscles peripherally using surface electromyography, without interfering with (or with minimal interference to) the natural course of the speech production process. In the next section, we briefly introduce some core concepts of electromyography.
## A brief introduction to electromyography
Technically speaking, electromyography (EMG) is a technique concerned with the recording and analysis of myoelectric signals (i.e., signals resulting from physiological variations in the state of muscle fibers membranes). Broadly speaking, EMG is a measure of the electrical activity generated during muscle contraction. It is used both as a basic tool in (for instance) biomechanical and psychophysiology research and as an evaluation tool in applied research (e.g., physiotherapy, rehabilitation, human-computer interfaces). To facilitate the interpretation of the EMG signal, it is useful to briefly detail the meaning of its physiological components.
### Nature of the EMG signal
#### Muscular anatomy and physiology
A muscle is a collection of fibers that can vary in length, orientation, diameter, and architectural characteristics. For instance, deeper muscle fibers are usually composed of a greater proportion of slow-twitch fibers (type I muscle fibers) whereas more superficial muscle fibers comprise a greater proportion of larger and fast-twitch fibers [type II muscles fibers, @kamen_essentials_2010]. On the basis of their structure and contractile properties, we can identify three types of muscle tissues: i) the *skeletal muscles* are attached to bones, their function is to produce voluntary movements and to protect the organs, ii) the *cardiac muscles*, whose function is to pump blood and iii) the *smooth muscles*, involved in involuntary movements (e.g., respiration, moving food).
The contraction of the skeletal muscles is initiated by electrical impulses that propagate from the central nervous system to the muscle via the $\alpha$-motoneurons. Many (both larger and smaller) muscles are partitioned, with each portion having a specific role for the muscle function. Moreover, there is no one-to-one mapping between populations of motor neurons and muscle compartments. In other words, one population of motoneurons may innervate several compartments and reciprocally, several populations of motoneurons may innervate the same muscle compartment. Therefore, interpreting the EMG signal requires to be aware whether the recorded signal is characteristic of a whole muscle' activity or of a specific muscle compartment [@kamen_essentials_2010].
```{r muscle, out.width = "50%", fig.cap = "Structure of a skeletal muscle, muscle fascicle and muscle fiber. Figure from the OpenStax \\textit{Anatomy and Physiology} Textbook. Download for free at \\url{https://cnx.org/contents/14fb4ad7-39a1-4eee-ab6e-3ef2482e3e22@15.1}."}
knitr::include_graphics("assets/muscle.jpg")
```
The muscle fiber is surrounded by a membrane, the *sarcolemma* (see Figure \@ref(fig:muscle)). Under resting conditions, the voltage inside the membrane is around -90mV, relative to the outside. This voltage results from a particular combination of sodium ($\text{Na}^{+}$), potassium ($\text{K}^{+}$), chloride ($\text{Cl}^{-}$), and other anions. At rest, the concentration of $\text{Na}^{+}$ is relatively high outside the membrane and relatively low inside the membrane. The concentration of $\text{K}^{+}$ follows an opposite pattern, with a greater concentration inside the membrane, and a lower concentration outside the membrane.
#### The motor action potential
Because muscle membranes can change their electrical state, muscle fibers are excitable tissues. When a muscle fiber is depolarised, the *membrane potential* produces a response called the *muscle fiber ation potential* or more generally the *motor action potential* (MAP). The generated action potential proceeds along the muscle fiber in both directions from the neuromuscular junction [@kamen_essentials_2010].^[The *neuromuscular junction* is the site where the motoneuron meets the muscle fiber.] In the first phase of the MAP, the $\text{Na}^{+}$ permeability increases dramatically, inducing a massive income of $\text{Na}^{+}$ into the cell. This results in a temporary inversion of the cell polarity (see Figure \@ref(fig:mfap)).
```{r mfap, out.width = "75%", fig.cap = "Time course of a motor action potential (figure from Kamen \\& Gabriel, 2010)."}
knitr::include_graphics("assets/mfap.png")
```
The MAP is followed by a *refractory period*, characterised by a decrease in membrane excitability. This refractory period can be further decomposed in an *absolute* refractory period during which all $\text{Na}^{+}$ channels are closed, and a *relative* refractory period where some $\text{Na}^{+}$ channels are open (but to a lesser extent than before the MAP). Interestingly, this after-impulse hyperpolarisation limits the frequency of MAPs [@kamen_essentials_2010].
#### The motor unit
The *motor unit* is the smallest controllable muscular unit. It consists in a single $\alpha$-motoneuron, its neuromuscular junction, and all the muscle fibers it innervates. The number of motor units and their *innervation ratio* (i.e., the number of muscle fibers per motor unit) can vary by muscle. Because a single motoneuron can innervate multiple muscle fibers, the firing of a single motoneuron results in the simultaneous discharge of many muscle fibers. The *motor unit action potential* (MUAP) is the electric field resulting from the sum of the electric fields emitted by each fiber of the motor unit. In other words, it represents the spatiotemporal summation of individual MAPs originating from muscle fibers that are sufficiently close to a given electrode. This generates a *train* of MUAPs, called *motor unit action potential trains* (MUAPTs). The mixture of MUAPTs coming from different motor units constitute the raw EMG signal (cf. Figure \@ref(fig:muap)).
```{r muap, out.width = "75%", fig.cap = "Motor unit action potential representation (figure from De Luca et al., 2006)."}
knitr::include_graphics("assets/muap.jpg")
```
To sum up, the EMG signal results from a mixture of recruited motor units: it is composed of the sum of several to many MUAPTs. This signal can vary considerably because of factors such as the muscle length [@babault_effect_2003], the distance between the muscle fiber (of interest) and the electrode, the fiber length or the muscle temperature. In the next section, we discuss in more details how this signal can be acquired.
### EMG instrumentation and recording
Myoelectric measurements have a long history, starting in the XVII and XVIII centuries with the classical observations that muscle contraction can generate electricity and that electrical impulses can generate muscle contraction. The term of electromyography and the first EMG measures were realised at the end of the XIXth century, and the quality of EMG measurements did not cease to improve since [see @raez_techniques_2006, for a brief historical perspective].
Two main types of sensors have been used to record EMG signals, varying by their invasiveness. First, *indwelling* (intramuscular) recordings can be acquired via electrodes directly inserted into the muscle. This form of EMG is mostly used in rehabilitation, for diagnosis of muscle function and to examine nerve conduction [@fridlund_guidelines_1986]. Second, *surface electromyography* can be recorded at the surface of the skin. Each method is associated with its own type of sensors, its own advantages and disadvantages. Surface electrodes have the advantage of being easy to use and non-invasive. However, their use is limited to (large and) superficial muscles. Moreover, because of the phenomenon of *cross-talk*,^[The phenomenon of *cross-talk* can be defined as the mixing of the electrical activity of the muscle of interest with the electrical activity of adjacent or distant muscles, that are not of primary interest.] it is difficult to isolate the activity of specific muscles using surface EMG. On the opposite, intramuscular EMG (that can be recorded via a single needle or two wires implanted directly into the muscle) are highly selective and can sometimes record the activity of individual motor units. In addition, indwelling recordings are not submitted to *tissue filtering* (i.e., the fact that muscles tissues act as low-pass filters), in contrast to surface recordings.
Because of the important crossing and superposition of facial muscles, surface EMG recorded over facial muscles does not generally represent the activity of a single muscle, but rather a mixture of muscle activations [@de_luca_use_1997;@Rapin2011]. As a result, it is usually inappropriate, when using surface EMG, to attribute the recorded activity to a single muscle [@fridlund_guidelines_1986]. Whereas for the sake of simplicity, we designate sensors by the name of the underlying muscle of interest (e.g., "FRO" for the *frontalis* muscle), it should be kept in mind that these sensors reflect the activity of "sites", rather than the activity of single muscles.
Aside from *cross-talk*, many other factors can affect the EMG signals. These factors are usually described under three main categories [for more details, see @de_luca_use_1997]:
* The *causative factors*, that have a basic effect on EMG signals. These factors can be further subdivided into two classes: i) the *extrinsic* factors, including factors such as the type of electrode (e.g., size, shape, placement) or the inter-electrode distance and ii) the *intrinsic* factors such as physiological or anatomical factors (e.g., fiber type, fiber diameter, blood flow).
* The *intermediate factors*. These are the physiological phenomena that are influenced by one or more of the causative factors and that in turn influence the deterministic factors, such as the conduction velocity, spatial filtering or the signal *cross-talk*.
* Finally, the *deterministic factors* are influenced by the intermediate factors and have a direct effect on the EMG signal. These include the number of active motor units or the amplitude, duration and shape of the MUAPs.
All these factors contribute to modulating both the amplitude of the EMG signal and its spectral properties (e.g., its mean or median frequency). The importance of these perturbating factors should be acknowledged and controlled as far as possible. In our work, we use a state-of-the art surface EMG apparatus, specifically developed to tackle these issues, as well as standardised procedures (more details regarding the EMG apparatus are provided in Chapters \@ref(chap3) to \@ref(chap5)).
### EMG signal processing
The raw EMG signal is a stochastic train of MUAPs. As put by @fridlund_guidelines_1986, "when heard through a speaker, the raw EMG signal sounds like popcorn popping". Therefore, it is usually unsuitable for immediate quantification. In order to illustrate what the EMG signal looks like, we simulated EMG data based on a standard algorithm implemented in the `biosignalEMG` package [@R-biosignalEMG]. This simulated EMG signal is represented in Figure \@ref(fig:emgplot1).
```{r emgplot1, fig.width = 6, fig.height = 4, fig.cap = "Simulated EMG signal."}
library(biosignalEMG)
library(tidyverse)
emg <- syntheticemg(
off.sd = 1, on.sd = 2, on.mode.pos = 0.1, samplingrate = 1e3, units = "mV"
)
ma <- envelope(emg, method = "MA", wsize = 200)
rms <- envelope(emg, method = "RMS", wsize = 200, units = "samples")
ts.plot(
emg$values, xlab = "Time (samples)", ylab = "Raw EMG signal (mV)",
col = "steelblue"
)
```
We usually rectify the EMG signal by taking its absolute value and subtracting the mean in order to correct for any offset (bias) present in the raw data. The result of this operation is represented in Figure \@ref(fig:emgplot2).
```{r emgplot2, fig.width = 6, fig.height = 4, fig.cap = "Rectified EMG signal."}
rectified_emg <- abs(emg$values - mean(emg$values) )
ts.plot(
rectified_emg, xlab = "Time (samples)", ylab = "Rectified EMG signal",
col = "steelblue"
)
```
Then, the signal is usually low-pass filtered, with a cut-off frequency depending on the nature of the study. From there, two main measures can be used to represent the magnitude of the muscle activity.^[But see for instance @phinyomark_feature_2012, for a brief overview of other features that can be extracted from the surface EMG signal.] The first one is the *mean absolute value* (MAV), which is computed over a specific interval of $N$ samples. Its formula is given below, where $|x_{n}|$ is the absolute value of an EMG sample in the data window.
$$MAV = \frac{1}{N} \sum_{n=1}^{N} | x_{n} |$$
The unit of measurement is usually the $mV$ and the MAV calculation is generally similar to the numerical formula for integration [@kamen_essentials_2010]. The second one is the *root-mean-square* (RMS) amplitude:
$$RMS = \sqrt{ \frac{1}{N} \sum_{n=1}^{N} x^{2}_{n} }$$
where $x^{2}_{n}$ is the squared value of each EMG sample and has both physical and physiological meanings. Put broadly, the RMS it taken to reflect the level of the physiological activities in the motor unit during contraction. Both the MAV and the RMS are illustrated in Figure \@ref(fig:emgplot3).
```{r emgplot3, fig.width = 6, fig.height = 4, fig.cap = "Illustration of the MAV (in orange) and RMS (in green) values. These two features are usually highly correlated but differ in magnitude. More precisely, the RMS is proportional to the MAV when the signal has a Gaussian shape."}
ts.plot(
rectified_emg, xlab = "Time (samples)", ylab = "Rectified EMG signal",
col = "steelblue"
)
plot(ma, col = "orangered", lwd = 2, add = TRUE)
plot(rms, col = "darkgreen", lwd = 2, add = TRUE)
```
These features provide the *envelope* of the rectified EMG signal. This envelope can then be summarised (e.g., via its mean or median) over a period of interest (e.g., during the utterance of some phoneme).
## Statistical modelling and statistical inference
### Limitations of the standard statistical approach in Psychology
Numerous authors have highlighted the limitations inherent to the Null-Hypothesis Significance Testing (NHST) approach and the (exclusive) reliance on *p*-values and significance testing [e.g.,@bakan_test_1966;@Gigerenzer2004;@Kline2004;@Lambdin2012;@meehl_theory-testing_1967;@trafimow_manipulating_2018]. Considering these limitations, some authors have recommended to push away significance testing and to develop the use of effect size estimates and confidence intervals in order to favour accumulation of evidence and a meta-analytical mode of thinking [e.g.,@Cumming2012;@cumming_new_2014].
However, the apparent superiority of confidence intervals over *p*-values is an illusion. Indeed, as noted by many observers, confidence intervals are simply inverted significance tests. In other words, the confidence interval represents the range of values that are significant at some $\alpha$ level. Therefore, compared to a *p*-value, a confidence interval does not bring *any* new inferential value. Moreover, its interpretation might be as hard as the interpretation of *p*-values. For instance, contrary to a widely shared belief, confidence intervals do not contain the $(1 - \alpha) \cdot 100$% most probable values of the parameter [e.g.,@morey_fallacy_2015;@nalborczyk_pragmatism_2019].
That being said, it is fair to acknowledge that using confidence intervals (instead of or in addition to single *p*-values) does shift the emphasis from a mechanical (mindless) point-hypothesis testing procedure to a more careful consideration of the range of values that are *compatible* with some hypothesis. More importantly, it emphasises the uncertainty that accompanies every statistical procedure. Indeed, we think that most of the caveats that are attributed to a specific statistical procedure (e.g., to NHST) are really caveats of the way it is used. Namely, the fact that it is used in a categorical and absolute way. This tendency has been coined as *dichotomania* (i.e., the tendency to consider that results are either present --if significant-- or absent --if non-significant), or *trichotomania* (e.g., when considering evidence ratios thresholds).
These biases have been highlighted by many statisticians in the past [e.g.,@wasserstein_asas_2016]. Very recently, *The American Statistician* published a special issue on *Moving to a World Beyond "p<.05"*, with the intention to provide new recommendations for users of statistics (e.g., researchers, policy makers, journalists). This issue comprises 43 original papers aiming to provide new guidelines and practical alternatives to the "mindless" use of statistics. In the accompanying editorial, @wasserstein_moving_2019 summarise these recommendations in the form of the ATOM guidelines: "**A**ccept uncertainty. Be **t**houghtful, **o**pen, and **m**odest." We describe below how our statistical approach might be understood in light of these core principles.
- **Accept uncertainty**: we try to represent and to acknowledge uncertainty in our analyses and conclusions. For instance, we do not conclude and/or infer that an effect is either "present" or "absent", but we report the *estimated* magnitude of the effect and the uncertainty that comes with this estimation. Additionally, when relevant, we report probabilistic statements based on the posterior distribution.
- **Be thoughtful**: for each analysis opportunity (i.e., for each dataset to analyse), we consider what would be the most appropriate modelling strategy but we also acknowledge that there is no unique best way to analyse a given dataset. In most empirical chapters, we clearly distinguish between confirmatory (preregistered) and exploratory (non-preregistered) statistical analyses. We routinely evaluate the *validity* of the statistical model (and of its assumptions) and we are suspicious of statistical *defaults*. We try to consider the *practical* significance of the results, rather than their *statistical* significance. We use a variety of statistics (e.g., effect sizes, interval estimates, information criteria) to obtain a more diverse picture of the meaning of the results.
- **Be open**: the soundness of a statistical procedure (and more generally, of an inferential procedure) can only be evaluated if it is made transparent to peers and readers for critical examination. Therefore, we take some space in the next section (but also in each experimental chapter) to motivate our statistical modelling decisions. We also make all `R` scripts available to ensure the reproducibility of the analyses. We try to be exhaustive in the way we report our analyses and we beware of shortcuts than could hinder important information to the reader.
- **Be modest**: we recognise that there is no unique "true statistical model" and we discuss the limitations of our analyses and conclusions. We also recognise that scientific inference is much broader than statistical inference and we try not to conclude anything from a single study without the warranted uncertainty.
To sum up, we try to acknowledge the uncertainty that accompanies every (statistical) inference. In the next section, we present in more details what our approach does entail and we introduce some key technical concepts.
### Our statistical approach
In brief, we tried to move from the point-hypothesis mechanical testing to an approach that emphasises parameter estimation, model comparison, and continuous model expansion [e.g.,@Cumming2012;@cumming_new_2014;@gelman_data_2006;@gelman_bayesian_2013;@judd_data_2009;@R-rethinking;@kruschke_doing_2015;@kruschke_bayesian_2018;@kruschke_bayesian_2018-1]. In other words, our approach can be defined as a *statistical modelling* approach rather than a *statistical testing* approach. It means that we try to *model* the data (or rather the process that generated the data), rather than to "test" it. We carefully consider what could be the process that generated the data and we try to model it appropriately. For instance, we do not fit reaction time data, Likert data, or electromyographic data using the same model, as this practice would lead to high rates of erroneous inferences.
Throughout this work, we use Bayesian statistical modelling, not by dogmatism, but because we think the Bayesian approach provides richer inferences than the frequentist one. The main advantage of the Bayesian approach is the explicit use of probability to model the uncertainty (see Box \ref{logic}). By doing so, the Bayesian approach permits to evaluate the probability of a parameter (or a vector of parameters) $\theta$, given a set of data $y$:
$$p(\theta|y) = \frac{p(y|\theta)p(\theta)}{p(y)}$$
Using this equation (known as Bayes' theorem), a probability distribution $p(\theta|y)$ can be derived (called the *posterior distribution*), that reflects knowledge about the parameter, given the data and the prior information. This distribution is the goal of any Bayesian analysis and contains all the information needed for inference.
The term $p(\theta)$ corresponds to the *prior distribution*, which specifies the prior information about the parameters (i.e., what is known about $\theta$ before observing the data) as a probability distribution. The left hand of the numerator $p(y|\theta)$ represents the *likelihood*, also called the *sampling distribution* or *generative model*, and is the function through which the data affect the posterior distribution. The likelihood function indicates how likely the data are to appear, for each possible value of $\theta$. Finally, $p(y)$ is called the *marginal likelihood*. It is meant to normalise the posterior distribution, that is, to scale it in the "probability world". It gives the "probability of the data", summing over all values of $\theta$ and is described by $p(y) = \sum_{\theta} p(\theta) p(y|\theta)$ for discrete parameters, and by $p(y) = \int p(\theta) p(y|\theta) d\theta$ in the case of continuous parameters.
All this pieced together shows that the result of a Bayesian analysis, namely the posterior distribution $p(\theta|y)$, is given by the product of the information contained in the data (i.e., the likelihood) and the information available before observing the data (i.e., the prior). This constitutes the crucial principle of Bayesian inference, which can be seen as an updating mechanism. To sum up, Bayes' theorem allows a prior state of knowledge to be updated to a posterior state of knowledge, which represents a compromise between the prior knowledge and the empirical evidence.
\vspace{2mm}
\begin{mybox}[label = logic]{Probability theory as extended logic}
Probability theory has been argued to be nothing more than \textit{extended logic}. In other words, it generalises the rules of logic that apply to discrete events (e.g., TRUE or FALSE events) to continuous events. By doing this move, probability theory gives a way to represent and to quantify the \textit{uncertainty}. Importantly, the rules of probability have the same status as the rules of logic: these rules can be used to derive statements (conclusions) that are guaranteed to be correct given that some premises are correct.\\
Bayesian statistics can simply be presented as an application of probability theory to statistical analysis. Whereas the dependence of the inferential conclusions on prior assumptions is sometimes presented as a downside of this type of analysis, it is precisely what makes them \textit{optimal} or \textit{coherent} (in the sense of respecting the rules of probability). As put by Vandekerckhove (2018), the claim that Bayesian methods would be invalidated by their reliance on prior assumptions would be akin to conclude that logical deductions are somehow invalidated because they depend on premises.
\end{mybox}
We also use *multilevel models* (also known as *mixed-models*) to handle complex dependency structures and to obtain more precise estimates. A more accurate description of Bayesian multilevel models is outside the scope of this introductory section but the interested reader is redirected toward several existing tutorial papers [e.g.,@nalborczyk_introduction_2019;@nicenboim_statistical_2016;@sorensen_bayesian_2016] or Appendix \@ref(appendix-brms). Throughout this work, we also make use of several tools with very distinct properties and uses. For instance, we use Bayes factors (BFs) to quantify the relative evidence for a statistical hypothesis (see Box \ref{BF}), we use information criteria to assess the predictive abilities of our models (see Box \ref{IC}), we use posterior predictive checks as well as a diagnostics tools (e.g., convergence indexes, trace plots) to assess the validity of our models, and we use summary statistics when appropriate to convey the meaning of the main results.
\vspace{2mm}
\begin{mybox}[label = BF]{What is a Bayes factor ?}
It is a rule in statistics that every statistics has already been suggested as the \textit{new statistics}. Confidence intervals have been suggested as a replacement to \textit{p}-values, being purportedly more informative and less difficult to interpret. Credible intervals have been suggested as a replacement to confidence intervals, for roughly the same reasons. Bayesian hypothesis testing through Bayes factors (BFs) has also been suggested as a replacement for frequentist hypothesis testing. It has been argued that they permit a richer inference and that they come with a more straightforward interpretation. Whereas this might be true, they nonetheless come with their lot of misinterpretations.\\
To highlight what BFs are and what they are not, it might be useful to write down the formula used to compute them. To this end, it is useful to write the Bayes rule in its \textit{odds form}, making the BF explicitly visible:
$$\underbrace{\dfrac{p(H_{0}|D)}{p(H_{1}|D)}}_{posterior\ odds} = \underbrace{\dfrac{p(D|H_{0})}{p(D|H_{1})}}_{Bayes\ factor} \times \underbrace{\dfrac{p(H_{0})}{p(H_{1})}}_{prior\ odds}$$
This equation reveals that the \textit{posterior odds}, the ratio of the posterior probability (i.e., how much more probable is hypothesis 1 ($H_{1}$) as compared to hypothesis 2 ($H_{2}$), after seeing the data $D$), is equal to the ratio of the probability of the data given the first hypothesis and the probability of the data given the second hypothesis, multiplied by the \textit{prior odds} (i.e., how much more probable was hypothesis 1 ($H_{1}$) as compared to hypothesis 2 ($H_{2}$), before seeing the data $D$).\\
Importantly, what we consider as \textit{evidence} in the Bayesian framework is also known as a \textit{marginal likelihood} and represents the information contained in the data, weighted by the prior information. It is a sum when parameters are discrete or an integral when parameters are continuous.
$$\text{evidence}\ = p(D|H) = \int p(\theta|H) p(D|\theta,H) \text{d}\theta$$
Therefore, the BF does not indicate how much "probable" a hypothesis is, or how much more probable a hypothesis is, compared to another one (this would be to conflate the BF with the posterior odds). Instead, the BF can be (should be) interpreted either i) as a ratio of two \textit{marginal likelihoods} (i.e., a ratio of \textit{evidence}) or ii) as an updating factor, indicating how we should reallocate credibility from prior knowledge (what we knew before seeing the data) to posterior knowledge (what we know after seeing the data).
\end{mybox}
Bayes factors are often said to have desirable *asymptotic* (i.e., when the number of observations is very large) properties. Indeed, they are *consistent* for model identification. It means that if a "true" statistical model is in the set of models that are compared, using a BF will usually lead to selecting this "true" model with a probability approaching 1 with increasing sample size. Whether this seems an appealing property or not depends on the underlying statistical philosophy. Indeed, one could question whether it is sensible to assume a "true model" (an oxymoron) in real life, especially in the social sciences [e.g.,@burnham_model_2002;@burnham_multimodel_2004]. As @findley_unbiasedness_1985 notes: *"[...] consistency can be an undesirable property in the context of selecting a model"*. A more realistic question is then not to look for the "true" model, but rather for the *best* model for some practical purpose.
The usefulness of information criteria comes from them being approximations of the *out-of-sample deviance* (see Box \ref{IC}). In the present work, we used generalisations of the AIC (especially the WAIC and LOOIC) that also approximate the *out-of-sample deviance* and as such give an indication of how good/bad a model is to predict future (i.e., non-observed) data.
\vspace{2mm}
\begin{mybox}[label = IC]{Information criteria}
Hirotugu Akaike noticed that the negative log-likelihood of a model + 2 times its number of parameters was approximately equal to the \textbf{out-of-sample deviance} of a model, which led to what is nowadays famously known as the \textit{Akaike information criterion} (AIC):
$$
\text{AIC} = \underbrace{\underbrace{-2\log(\mathcal{L}(\hat{\theta}|\text{data}))}_{\text{in-sample deviance}} + 2K}_{\approx\ \text{out-of-sample deviance}}
$$
where $K$ is the number of parameters of the model and the \textit{deviance} is a measure of discrepancy between the data and the model.\\
Importantly, we make a distinction between two types of deviances. First, the \textbf{in-sample deviance} indicates how bad a model is to explain the current dataset (the dataset used to fit the model). Second, and more importantly, the \textbf{out-of-sample deviance} indicates how bad a model is to explain a \textbf{future} dataset issued from the same data generation process (the same population).
\end{mybox}
In brief, in the present work, we used various methods but coherently with a few (nuanced) guiding principles. Namely, we favoured a *model comparison* approach [e.g.,@burnham_model_2002;@burnham_multimodel_2004;@judd_data_2009], we used several statistics when they provide complementary information (e.g., using both posterior probabilities, information criteria or BFs), we assessed the validity of our models (e.g., via posterior predictive checks), we reported these analyses transparently, and we tried to convey uncertainty in our conclusions.
## Overview of the following chapters
The experiments carried out during this PhD will be presented as five empirical chapters that can be grouped under two main axes. In the first couple of experiments, we used surface electromyography and muscle-specific relaxation to investigate the involvement of the speech motor system during induced verbal and non-verbal rumination (Chapter \@ref(chap3) & \@ref(chap4)). In Chapter \@ref(chap5), we used surface electromyography and machine learning algorithms to decode the muscle-specific EMG correlates of inner speech production. In the last couple of experiments, we switched strategy from the "correlates strategy" to the "interference strategy", where the goal was to directly interfere with the activity of the speech motor system. More precisely, we used articulatory suppression to disrupt induced rumination in Chapter \@ref(chap6), and we used articulatory suppression to disrupt either induced rumination or problem-solving in Chapter \@ref(chap7). Finally, in Chapter \@ref(chap8), we summarise the main findings, discuss their implications and suggest ways forward from both a theoretical and an experimental perspective.