More political scientists doing bad linguistics

越来越多的政治学家演讲语言繁复

2021-01-18 05:50 Lingua Greca

本文共5227个字,阅读需53分钟

阅读模式 切换至中文

There’s a new article out that uses faulty methods to study the linguistic complexity of politicians’ speech. It makes many of the same mistakes that I criticized Schoonvelde et al. (2019) for – and even references that article. But it somehow comes to the right conclusion… for the wrong reasons. I know, it’s strange. Let’s check it out.In a recent study titled “The Language of Right-Wing Populist Leaders: Not So Simple” (2020), McDonnell & Ondelli analyze speeches by politicians in the US, UK, France and Italy. Their goal is to see whether populist politicians’ speech is simpler in comparison to their more mainstream political opponents. They are motivated by reports which say Trump speaks in a simpler way than Hillary Clinton. This is all fine and good, but the authors use faulty methods to study linguistic complexity. I’ll mention at the top that neither author is a linguist. I’m not saying that being a linguist or language scholar is necessary to do linguistic research, but in this case it probably would have helped prevent the authors from making simple mistakes. I’m going to go into detail below (a whole lotta boring detail), so I’ll summarize up here my main concerns with McDonnell & Ondelli’s study. Main criticisms The study needs more sources for the claims made, especially the ones supporting the methodology McDonnell & Ondelli claim multiple times that their methods are “well-established” but they’re really not – or at least it should be easy to slip a reference in when they make this claim. The authors cite sources which make the exact same mistakes. For example, they cite Schoonvelde et al. (2019), which is a study that I commented on and showed that it should not have been published. I spent a lot of time pointing out the problems with that study and I included many sources. Despite this, the McDonnell & Ondelli cite that study approvingly. Garbage in, garbage out McDonnell & Ondelli’s study uses the same faulty methodology as other studies (Schoonvelde et al. 2019, etc.), so it suffers from the same problems. These problems are: Confusing written language with spoken language Using an ineffectual test for written language on spoken language Not taking into account how transcriptions and punctuation affect the data Citing almost no linguistic sources in a study about language The tests used in this study are dependent on periods. Periods, like all other forms of punctuation, does not appear in speech. Punctuation is a feature of written language. But this study is a study of political speech. This is a basic error. Right answer, wrong reasoning The authors somehow come to the correct conclusion – that their methodology can’t show whether one politician’s speech is more complex – but they do so without recognizing that the issue is with their analysis. The tests that they use to study language could have shown anything. The fact that the tests showed that the politicians were quite similar in their complexity is a fluke, not a result of a sound analysis. The authors do end up making some good points in their study. I’ll show those too because I think they deserve credit for them. But first, the bad stuff. Lack of sources, lack of knowledge about the subject In the Intro, McDonnell & Ondelli lay out the tests that they use to study the linguistic complexity of politicians’ speeches. They are: The Flesch-Kincaid tests Type/token ratio & Lemma/token ratio Ratio of function words to content words. The Dale-Chall test (for measuring the amount of “difficult” words) The authors claim that these are the “main measures employed by linguistics scholars for evaluating simplicity” but they don’t offer any references for this. I don’t think linguists would use these tests, but what do I know? If these measures are so “main” it should be easy enough to provide some references. Later, in the Background, the authors state that their approach is “in line […] with a long tradition in linguistics research” but they don’t give any sources for this claim. I don’t know of any. Do you? Give me one of those studies in that long tradition in linguistics research. Instead they offer a list of recent studies which “have sought […] to analyze the simplicity of populist and other politicians’ language”, but only one of the seven sources that they cite appears in a linguistics journal (Wang and Liu 2018). This is worrying. They do it again in their Methods section, where they state: Saying the same thing over and over again doesn’t make it true. If these methods are so standard and well-established, then cite some of them. Who are they? When are they? Are linguists still doing this research? Or have they abandoned it because it’s based on shoddy tests (looking at you, Flesch-Kincaid)? It shouldn’t be hard to give us some studies throughout the last few decades which use these same tests. The lack or sources problem comes up again at the end of the Methods section: I really need some sources on these “well-established linguistic measures”. Just because some people have used some of these measures in the past doesn’t mean that we should keep using them. And one of the studies that these authors cites specifically says that the readability tests are a “crude” way to measure linguistic complexity (more on that later). By now you’ve probably figured out what’s up. Linguists don’t use these measures to study complexity in language. There have been studies which discuss how complex speech and writing are in comparison to each other, but they use other features to describe complexity. The studies by linguists discuss how many subordinate clauses are used, how many modifiers are in the noun phrases, or the various ways that cohesion is achieved. In speech, there are paralinguistic and non-verbal ways of establishing cohesion (eye gaze, gesture, intonation, etc.), while in writing, cohesion is achieved through lexical and syntactic structures (because writing can’t make use of the metalinguistic and non-verbal features, whereas practical memory constraints force speech to avoid lexical and syntactic cohesive measures). But linguists aren’t in agreement about whether speech or writing is more “complex” – there are various studies pointing different ways, and the methods used in each study are going to influence the results and conclusions of the studies. (See Biber 1989, Chapter 3, for an overview of these not-so-recent studies) Another way that the authors show a lack of knowledge and a lack of sources about the topic is when they discuss the average length of each word in syllables. They state: “we find that Trump uses slightly more words that are longer than three syllables (10.97% versus 10.75%, and an average word length in syllables of 1.44 versus 1.43)” But the authors need to prove that words with more syllables are more “complex” than words with fewer syllables. This isn’t something we can take at face value. Is December more complex than May? Is Saturday more complex than Monday? Is Afghanistan more complex than Russia? And besides, the numbers between Trump and Clinton are so close that the difference could come from quirks in the English language. There are two studies referenced in the Background that provide a foundation for this article: Oliver & Rahn (2016) and Kayam (2018). These are two studies on Trump’s language – neither of which was done by a linguist or published in a linguistics journal. Why do you think that is? Hmmmm…. Maybe this linguistic complexity stuff isn’t worth it. Because there are linguists studying Trump’s language, but not like this. McDonnell & Ondelli do manage to cite an article that was published in a linguistics journal (Language and Society, Wang and Liu 2018), but this article uses the F-K test. If you don’t know what’s wrong with this, stay tuned for a facepalm. And finally – saving the best for last – in the Background, McDonnell & Ondelli say that Schoonvelde et al. (2019) “provide more support for the claim that right-wing populists use simpler language than their opponents.” Friends, I lol’ed. I wonder if McDonnell & Ondelli read my comments on that article. The title of the comment thread was called “There are significant problems with this study”. I posted them almost immediately after the article was published and I showed (with resources) that the methods in that article (the same methods that McDonnell & Ondelli use) were critically flawed. Garbage in, Garbage out The fundamental problem with McDonnell & Ondelli’s study is their methodology. They test for linguistic complexity using measures which cannot show how simple or complex the language is. Let’s take these one by one. F-K test First, McDonnell & Ondelli use the Flesch-Kincaid (F-K) test. This test measures a text for average word length (in syllables) and average sentence length. It then applies a score to the text which says which education level is needed to understand the text (based on the US education system in the 1970s). The test is therefore dependent on sentence-final punctuation, aka “periods” (or full stops for our friends across the pond). Savvy readers will immediately notice that McDonnell & Ondelli are studying speech with a test that depends on periods… even though periods don’t exist in speech! This apparently flew right over the heads of the authors. They state: The suitability of the F-K test was also questioned in the comment section of Schoonvelde et al. (2019)… by me. ? And I showed exactly why the F-K test is garbage for studying speech (so I won’t bore you with the details here). Basically, the F-K test is overly simplistic and was designed to work on one style of WRITTEN English, so it can’t be applied to SPOKEN English. These are better grounds to question the suitability of the F-K test, and are COMPLETELY OBVIOUS to people with a basic knowledge of linguistics. This problem is sort of related to the problem of a lack of (knowledgeable) sources on the topic. In McDonnell & Ondelli’s defense, they didn’t apply the F-K test to the non-English languages that they studied. They could at least see that a test developed to study English might not be appropriate for languages that aren’t English. So good job! I’m not being sarcastic here. Schoonvelde et al. (2019) applied the F-K test to Spanish, French, German and Dutch because … fuck it, why not? So good on the McDonnell & Ondelli for not doing this. But… the tests that they do use on the other languages don’t seem to be much better. For French they use the Kandel-Moles index, which is just the F-K modified for French. For Italian, they used the Gulpease index, which is like an Italian version of the F-K test. And they supplemented these scores with something called LIX. I’m not here to discuss their analyses of French and Italian, but I would like to note that the authors say that LIX “has been proven to perform well on Western European languages” and to back this up they link to a company called SiteImprove (in endnote 9). What the heck is this? I’m supposed to believe SiteImprove when they say these tests are good stuff? Yeah, no thanks. Surprisingly, all of the tests listed by SiteImprove – tests which they will do for you if you pay them – are great tests. The best tests. A lotta people are saying they’ve never seen such tests. And, not for nothing, what the hell are “Western European languages”? Is Basque a Western European language? German? Catalan? Irish? Maltese? Welsh? Yiddish? Forget it. “Western European languages” is not a meaningful linguistic term. It’s marketing speak from a company trying to sell you linguistic snake oil. It’s Bruce Banner in Avengers saying that Wakandan is an “African dialect”. Type/token ration… and beyond! Moving on, the authors have a second language grinder to run their speeches through: Hoo-boy! We finally got a reference to linguists! Too bad that these linguists (Granger and Wynne) never say what they are cited as saying. Let’s go through the claims in this paragraph. “The first is lexical richness, which is based on the premise that the higher the repetition rate of word-types in a corpus, the easier the language being used, since lexical variation, in addition to increasing difficulty per se, may also imply a broader range of contents.” I’m not sure where they got this idea, but they certainly don’t cite anyone. It’s not in Granger & Wynne (1999). Whose premise is this? Nobody knows! “As an additional check of lexical richness, expressed as the Type/Token Ratio (TTR) and percentage of hapax legomena (i.e., words occurring only once in the corpus), we calculated the lemma/token ratio (LTR).” Granger & Wynne (1999) do talk about Type/Token Ratio. They call it “a rather crude measure”. *Sad trombone*. Granger & Wynne also discuss Lemma/Token Ratio. They say “Our study shows that it is not safe to use crude type/token or lemma/token ratios with learner corpora”. The “with learner corpora” part is important. Granger & Wynne investigated texts written by English learners – not professional speech writers – and they tested whether traditional measures of lexical richness should be used on them. And they found that these measures are too crude for investigating texts by English learners. They didn’t test them on speech, or written political texts, and they never ever talked about these tests showing how “complex” or “difficult” a text is. “This is because, when we refer in everyday speech to the “rich vocabulary of a speaker,” we tend to mean the unusually large number of different lemmas used (i.e., the basic forms for a paradigm, such as “love” as the basic form for “loves,” “loved,” “loving,” etc.) and not the great variety of inflected forms (Granger and Wynne 1999).” Granger & Wynne (1999) never refer to the “rich vocabulary of a speaker”. They only say that “A learner who uses five different forms of the verb go (go/goes/going/gone/went) in one and the same text has a less varied vocabulary than the one who uses five different lemmas (such as go/come/leave/enter/return).” That’s it. No rich vocabulary, no linguistic complexity. In the analysis, the authors do indeed report this crude analysis as if it’s supposed to tell us something: Again, there’s no research cited which indicates that greater TTR or LTR values mean a text is more complex. In fact, we’re not even offered a baseline of what TTR or LTR we’re supposed to expect. How high is high? Are both of the speakers high on the type-token ratio? Or low? This is a problem that will reoccur in the analysis. For the time being, let’s stay on the topic of type/token ratio. We’ve already seen that the source cited by McDonnell & Ondelli calls it a “crude” form of analysis (along with LTR). But they’re not the only ones. In the Longman Grammar of Spoken and Written English (which, to those who aren’t grammar G’s like me, is one of the grammars of English), we see some claims which the authors could’ve cited. For example, in section 2.2.1.2, Longman says “The high TTR in news reflects the extremely high density of nominal elements in that register, used to refer to a diverse range of people, places, objects, events, etc.” So perhaps the high TTR of Clinton really does mean that she refers to a greater variety of contents. But the next sentence in Longman says “At the other extreme, academic prose has the second lowest TTR, reflecting the fact that a great deal of academic writing has a restricted technical vocabulary and is therefore less variable than fiction and news reportage.” I think we can all agree that academic English is harder to follow than news reports – so a high TTR does not mean that a text is more complex. This is exactly the kind of thing I’m talking about. The authors make assumptions based on little or no evidence. And the very next sentence after that in Longman says that TTR is a “crude” measure. But that’s not all! Halliday also discusses TTR. But he cautions against using TTR to say that a text is more complex: Halliday notes that our conception of language and its complexity is therefore crucially dependent on which part of it we’re looking at. It’s not hard to see spoken language as more complex. If we want to use TTR to indicate complexity, we have to ground that analysis in something – more complex than what? Biber’s (1988) foundational work on the variation between different genres of language used multiple aspects of language to analyze how each genre differed. He (1988: 202) says about complexity in language: “the fact that [markers of discourse complexity] occur in a largely complementary pattern shows that there are different types of complexity, and that it is not adequate to simply characterize particular genres as complex or not – rather, different genres are complex in different ways to different extents. The discourse complexity of spoken informational genres takes the form of structural elaboration, while the complexity of planned, written genres takes the form of lexical elaboration and precision.” That raises the question of what McDonnell & Ondelli are studying here – written language or spoken language? Because complexity is likely to show up in different ways. But more on that question right now! Function words vs. Content words. Fight! The authors go on to describe another way that they are apparently complementing their use of the F-K tests and other readability measures: Says who? How do we know that a text which conveys more information is more difficult to understand? A source would be really nice here. Alas, we wait in vain. (And not for nothing, the term is “function word”, not “grammar word”. That’s the one used by the large majority of linguistic grammars.) There’s another very important question behind this assumption though. Basically, we don’t know – and the authors don’t tell us – what kinds of figures we’re supposed to expect. What do the numbers mean? As the authors show, Trump and Clinton have very similar percentages of content and grammar words. Maybe that’s just normal for English. No matter what kind of text you pick up, about 40-45% of the words are going to be function words and 50-55% of them are going to be content words (allowing for some percentage of “words” which don’t fit into either category – interjections, numbers, etc.). But we aren’t told what to expect and we aren’t told what an informative/complex text looks like. That means we have nothing to compare the figures to. We can only compare the figures for Trump and Clinton to each other. This doesn’t tell us much. The authors should have either cited a source or given us examples from other SPOKEN genres. Because that’s what you want to see, ain’t it? Don’t worry. I got you, bro/sis. For example, here’s a comparison of the percentages of function and content words in various corpora and novels: The percentages for other corpora of spoken language (from COCA and the BNC, two large representative corpora) look pretty close to Trump and Clinton. The corpus of TV and Movie dialogue is pretty darn close too. Even the novels aren’t that far off. Maybe a comparison of function and content words isn’t a good way to tell if one text is more complex than another. But Chall Knew That…? The authors use a test called the Dale-Chall score to see how many “difficult” words the English speeches contain. This is supposed to guard against problem I pointed out above: the fact that the F-K test is based on word length… but word length doesn’t really matter in terms of complexity. McDonnell & Ondelli show that ersatz is harder to comprehend than substitute, even though its shorter. Presumably this is because ersatz is less frequent, but the authors wrap it up in the idea that the average speaker can understand substitute better than ersatz. I don’t know why they do this. (We’ll talk about it more below) My first question is: if the tests are so bad that you need to guard against them – if they are dependent on something which doesn’t matter in terms of making the language more difficult, such as word length – then why use those tests at all? They don’t seem to be very good tests. But whatever. It’s done. The Dale-Chall score is pretty much the same as the F-K test, except it compares the text to a list of 3,000 words that are considered to be familiar to 80% of 4th graders in the US. If a word doesn’t appear on the list, then it’s considered “difficult”… to a ten-year-old American. The Dale-Chall test also suffers from the same problems as the F-K test. Dale-Chall includes sentence length in the calculation. So, again, a test made for WRITTEN language being used on SPOKEN language. When the authors state “Clinton uses longer sentences (on average 15.02 words per sentence compared to 12.55 for Trump)” they are measuring the punctuation styles of either the transcribers or the speechwriters. They are taking the “sentences” at face value – wherever a period was placed, that’s where the speaker ended a sentence. But of course that isn’t accurate. Speech doesn’t have periods. There is absolutely no punctuation in speech. None. Zip. Zilch. Periods are a feature of written language. And the decision of where to place them in a sentence is somewhat arbitrary. Tell me if Trump or Clinton used more coordinating conjunctions. And how long they paused when using these words. That would give us a better idea of the length of the speaker’s “sentences”. I swear to Thor, I’m going to keep repeating this until political scientists get it. The authors claim in Appendix A that the speeches were punctuated by either the transcribers (whoever they were) or university students. They also say that the students checked the transcriptions, and that the authors checked a portion of the transcriptions (20,000 words). I don’t know if the students checked the transcriptions against the actual speeches (by watching the videos, for example) or if they just checked the punctuation and spelling. The authors had each speech transcribed by multiple students. They were doing all this “to avoid individual transcription preferences/idiosyncrasies biasing results”. But I’m not so sure that they avoided this. First, let’s say that introducing punctuation (aka a feature of written language) into speech is methodologically fine when we want to study the complexity of language. It’s not, but go with me here. We have to put the punctuation symbols in because our complexity tests depend on them. So where do we put them? This isn’t an easy question to answer. English has these things called conjunctions. These are words like and and but. With these words, we can avoid using a period and connect two independent clauses. But we can also use a period and start the next sentence with one of these conjunctions. The decisions to use more periods is going to affect the scores of our totally reasonable readability tests. Yeah, but the authors checked against this, I hear you saying. Did they, though? Check this: The Clinton sub-corpus has 218 more sentences which start with a conjunction than the Trump sub-corpus. 157 of these sentences start with coordinating conjunction (and, or, but, etc.). Why is that? If we normalize these figures in order to make them comparable, we see that Clinton has 243 more sentences starting with a conjunction per 100,000 words. (This is some linguistics jargon. If you don’t get it, fine. I don’t have time to explain it here. But just know that the authors should know this stuff before doing their research and it appears that they don’t.) If we carry this forward a bit, we see that 24% of Clinton’s sentences start with a conjunction, while only 16% of Trump’s do. Why is that? Could it have something to do with where the periods were placed? Perhaps the preferences and idiosyncrasies of the transcribers weren’t so successfully avoided. But this whole matter could’ve been avoided. The authors could have placed the speeches into the tool on English-corpora.org and seen how frequent each word is. This would have given them an idea of which speaker uses more common vocabulary items and which speaker uses less common words. Yeah, linguists have these tools. And they’re better than some rinky-dink readability score. Why didn’t the authors do this? Oh yeah, because they’re not linguists and they don’t know what they’re doing. The authors performed some statistical tests to check their readability scores. But this doesn’t help the matter. The readability scores are flawed from the start. No amount of T-tests is going to make them better. Finally, remember Halliday? He was the linguist who wrote (waaay back in the days when Michael Keaton was Batman) that TTR was a bad way to analyze complexity in language. So right after Halliday advises against using TTR, he goes on to explain why we shouldn’t use “sentence” in our analyses of spoken language because the “sentence” is a feature of written language and is not applicable to spoken language. But, you know, you got these readability tests from the 70s and their just asking to be used. Who cares if they’re garbage? Here, have a Camel and chill, ok? Trust your T-Zone. Wrap it up, boys! Despite the critical methodological problems with their analysis, McDonnell and Ondelli believe that the tests have shown something significant. They write: As I said above: garbage in, garbage out. Flawed tests revealed flawed results. It doesn’t matter that the results don’t show support for the claim that populists use simpler language, or that Trump and Clinton are more or less linguistically similar in complexity, because the methodology was flawed from the start. No matter what results came, we would have to reject them (which is what I argued in my comments on the Schoonvelde et al. 2019 article, but whatevs) They attempt to address one of the major flaws in their research – that they are studying speech with tests designed for writing. They say “it is true that we cannot consider the effect of factors such as speech speed and intonation, as well as pauses, on simplicity. However, this applies also to other studies of the simplicity of political language.” But how is this an excuse? “The other studies were flawed and so is ours”? I mean, those aren’t the only aspects of spoken language that this study (and others) doesn’t consider. There’s also dialect and pitch. These things affect how speech is evaluated by listeners (as do other language-external factors). The other studies which used readability scores weren’t good. That doesn’t mean you can do the same thing they did because screw it who cares. Good points. Credit. To the authors’ credit, they do manage to make some good points about language. Some of these are self-evident – although they weren’t made by previous research (ahem, Schoonvelde et al. 2019). Some of the other points, however, the authors seem to arrive at through serendipity. The methods that they use in their research are not able to tell us anything (that’s what we saw in the last section), but the authors come to some astute conclusions. Quotes followed by my comments. We saw this one up above and it’s actually a good point. Credit where credit’s due. It’s not explained very well and it’s not backed up with any sources or evidence, but I think they’re on the right track here. Another good point, one that was completely lost on Schoonvelde et al. (2019) This is both a good and bad point. Yes, studying speeches is a way to test whether populists use simpler language. But McDonnell & Ondelli didn’t study them well. So this point is like saying “Taking tests is a way to finish fifth grade.” That’s true, you can’t fail all the tests and still pass fifth grade. Holy shit! I can’t believe it. They just came right out and said it. This is a great point. The F-K test can’t tell us whether language is complex or simple. SO WHY ARE YOU GUYS USING IT? I mean, good on them for straight up claiming that their results aren’t really useful, except for showing that people (and I’m looking at you, political scientists) should stop using the garbage readability scores. It’s almost like McDonnell & Ondelli are suggesting that someone should study language in a more rigorous way. I wonder who would do something like that. Maybe some sort of scientist of language? Yeah, no fucking shit. Sorry, but this good point makes me angry because it’s something a linguist could have just told you. Just fucking ask. You don’t have to go reinventing Linguistics. Just pick up a goddamn linguistics book and see that people already know this stuff – they’re called linguists and they study language for a living. Wild, huh? The right answer for the wrong reasons So, after this womp womp of an academic article, McDonnell & Ondelli somehow – surprisingly – come to the correct conclusion. They state: Did they come to the right solution by doing the wrong work? [Ron Howard voice: Yes. Yes, they did.] I mean, I guess I know how they got there – but since they didn’t see the problems with their methodology, I’m kind of shocked. I want to say one more thing though: Refuting this kind of research takes a lot of time – the work I’m doing here should have been done by the authors or reviewers. But their article is already out there and it’s been written up in the Conversation. Just like with Schoonvelde et al., lots of people are going to hear about it and believe it. This sucks and that crap is on the authors. I welcome scholars from other disciplines into the field of Linguistics. Scholars of other fields can bring new perspectives to the study of language. But you have to show up. If you’re going to study language, you have to do it right. You can’t do some half-baked slapdash study and call it good. If linguists published this kind of stuff in your field, would you take them seriously? Didn’t think so. References Biber, Doug. 1988. Variation in Speech and Writing. Cambridge: CUP. Longman Grammar of Written and Spoken English. 1999. Douglas Biber, Stig Johansson, Geoffrey Leech, Susan Conrad and Edward Finegan. London: Longman Halliday, M.A.K. 1989. Spoken and Written Language (2nd ed.). Oxford: OUP. Email Tweet
这有一篇新文献用错误的方法来研究政治家演讲语言的复杂性。它当中犯的许多错误与我2019年批评Schoonvelde等人时所犯的错误是一样,甚至还引用了我那篇文章。但不知道为什么他却得出了正确的结论…或许是因为研究方法错误的原因。我也知道,这是很奇怪的现象。最近,在一篇名为“右翼民粹主义领袖的演说语言:其实不简单”(2020年)的论文研究中,麦克唐纳和翁德利对美国,英国,法国和意大利的政治家的演讲进行了分析对比。他们是想验证是否民粹主义政客的言论与比他们更主流的政治对手的语言要更简单。他们的研究背景和原因是因为有报道称特朗普的演讲语言比希拉里·克林顿的要更简单。 其他的都不错,只是作者们用了错误的方法来研究政治家们语言的复杂性。我会先说明,该文两位作者都不是语言学家。我并不是说做语言学研究就必须是语言学家或语言学者,只是在类似这种情况下,这些身份可能会更有助于防止作者犯简单的错误。 下面我会一一进行详述(只是一大堆枯燥的细节),因此我将在这里总结一下我对麦克唐纳和翁德利(McDonnell&Ondelli) 俩人研究的主要关注。 最主要的问题 他们的论文研究可能还需要更多的资料支持,特别是支持其研究方法的资料与素材 麦克唐纳和翁德利(McDonnell&Ondelli) 曾多次表示说他们的方法是“成熟可信的”,不过事实并非如此--因为至少在他们提出这种说法时,应该很容易引用某一说法作为一个参考。而他们引用的文献就犯了一样的错误。比如说,他们引用斯库韦德(Schoonvelde)等人2019年的论文,这篇论文是我曾经评论过的一篇论文研究,我曾说明它本不该发表。我花了很多时间批评指出了那篇论文研究的问题,并运用了很多例证来支撑我的说法。尽管如此,麦克唐纳和翁德利(McDonnell&Ondelli) 还是肯定了这项我曾用心批评过论文研究。 材料不好,作品不佳 麦克唐纳和翁德利(McDonnell&Ondelli) 的研究与斯库韦德(Schoonvelde)等人的研究都用了同样错误的调查研究方法,所以他们 也难免会遇到同样的问题。这些问题包括: 把书面语和口语混为一谈 用测试书面语测试对口语进行测试 并不考虑转录和标点符号对研究数据的影响 还有是几乎没有引用关于语言的研究方面的参考文献 该项研究所使用的测试方法是以断句为依据的。而句号和其他所有的标点符号一样,都不会不出现在语音中。因为标点符号是书面语的特征之一。而本研究的研究对象是政治演讲语言。这就是犯了低级错误。 方法错了,结论正确 不知这两位作者为什么得出了正确的结论--尽管他们的研究方法并不能说明哪一个政客的演讲语言复杂与否--并且他们进行研究时也并没有发现问题就出在他们的研究方法上。而他们用来研究语言的测试结果可以得出任何结论。根据他们的测试结果,某两位政治家演讲语言在复杂性方面很相似,但这一结论只是一种碰巧,而非在合理分析之后得出的结果。 不过这两位作者在他们的研究中也确实提出了一些好想法。我在后文会给你说明,因为我认为他们这一点是值得称赞的。不过,先说说缺点与不足。 缺少例证,对研究主题还不够了解 在论文引言中,麦克唐纳和翁德利(McDonnell&Ondelli) 写下了他们用来研究政治家演讲语言复杂性的测试。如下: 弗莱什-金凯语言测试 字体/符号比率和字根/符号比率 虚词与实词之间的比率。 戴尔·查尔特(Dale-Chall)测试(用于测量“生僻”字的数量) 这两位作者在论文中认为他们的测试方法是“语言学学者用于评价语言简单性与否的主要衡量标准”,不过他们并没有引用任何文献加以论证。我不觉得语言学家会使用这种测试方法,据我所知,假如这测试方法是如此“重要”,那就应该很容易找到一论证材料。之后,在论文研究背景中,作者说明他们的研究测试方法“与符号语言学研究的悠久传统是相符的”,不过他们并没有对他们的说法加以论证。我不明白说明了什么。你明白了吗?就这样抛出一个关于语言学研究悠久传统的研究。不过,他们展示了他们最近的一份研究表,这项研究“想凭借符号语言来对民粹主义政客与其他政客演讲语言的复杂性程度进行分析”,不过他们所引用的七个参考文献里只有一个在语言学期刊上可以找到--王和刘,2018年的论文。因此这是很令人担心的。 而他们在论文的研究方法部分再次重申: 尽管你不停地重复说着同样的话也并不能表示你所说就是事实。假若他们的方法是如此成熟与标准完善,那就应该列举其中的一些研究方法。例如说明他们是谁?他们都是什么时候发表的?现在的语言学家还在这方面做相关的研究吗?或者说说他们放弃这一方法的原因,是因为它还是不够精致不够成熟的测试(就比如,Flesch-Kincaid语言测试法)?几十年的时间,我们用些相同的测试方法进行对比研究应该不是不难。 在论文研究方法的结尾,缺乏例证的问题再次出现: 我真的想看到一些关于“既定的语言衡量标准”的例证。仅仅因为有些人在过去使用过这些措施,并不意味着我们应该继续使用这些措施。这些作者引用的其中一项研究特别指出,可读性测试是衡量语言复杂性的一种“粗略”方法(稍后将详细介绍)。 到这,我想你已清楚到底是怎么回事。其实语言学家们根本并不使用这些测试方法来对语言的复杂性进行研究。不过确实已有研究分析了口语和书面语的复杂程度,不过他们使用其他方法来衡量复杂性。语言学家们通过研究分析讨论名词短语中从属分句的使用情况,修饰语的使用情况,以及各成分之间实现衔接而采用的各种衔接方式。在口语中,可通过元语言和非语言进行衔接(比如眼神,手势,语调等),而在书面语中,是以词汇和句法结构形式来实现衔接的(因为书面语不能借助元语言和非语言的方式,而记忆限制迫使口语尽量不用词汇和句法的衔接方式)。而语言学家们还没有一致的看法,对于口语与书面语那个更“复杂”--不同的研究表达了不同的看法,而每一项研究中使用的研究方法都会影响研究的结果。(参看比伯(Biber) 1989,第3章,“关于这些非最近的研究的概述”) 其他地方也表明这两位作者缺乏对研究主题知识的了解和缺乏例证。比如当他们讨论用词在音节的平均长度时说:“我们发现,特朗普使用三个音节以上的单词比较多(比率为10.97%对10.75%,平均音节长度为1.44对1.43)。” 只是这两位作者们需要进行论证的是,音节越多的单词就“复杂”。但是这并非从表面就能理解的。比如说英文单词中12月会比5月复杂吗?而星期六会比星期一复杂吗?还有阿富汗会比俄罗斯更复杂吗?还有就是,特朗普和希拉里之间的比率相差不大,而这种差距可能是由英语中的特例导致的。 论文研究背景中参引了两份文献,为他们的论文奠定基础:Oliver&Rahn(2016)和Kayam(2018)。这两份研究都是关于特朗普演讲语言的研究--并且都不是语言学家做的,也没有在语言学期刊上进行发表。您认为是什么原因呢?或许因为……。又也许对语言复杂性进行研究根本就是不合适。虽然有语言学家对特朗普演讲语言进行过研究,不过结论并非如此。 麦克唐纳和翁德利(McDonnell&Ondelli) 确实引用了一篇在语言学期刊上发表过的文章(《语言与社会》,王、刘,2018),不过这篇论文所用的是F-K测试法。要是您明白它存在的不足,请接着看前言。 结尾处,麦克唐纳和翁德利(McDonnell&Ondelli) 还引用了Schoonvelde等人2019年的论文《跟多例证--右翼民粹主义者的语言比对手的更简单》。各位,我想大声笑说,我不知麦克唐纳和翁德利(McDonnell&Ondelli) 读没读过我曾经对斯库韦德那篇论文的用心评论。评论文章的标题是“这篇论文有很多不足”。我差不多和斯库韦德的文章同时发表,并且(在参考文献处)展示了他的测试方法(与麦克唐纳和翁德利(McDonnell&Ondelli) 使用的方法相同)都存在很多缺陷。 摄入不好,产出不佳 麦克唐纳和翁德利(McDonnell&Ondelli) i研究的主要问题是他们使用了错误的方法。因为他们想分析语言的复杂性,但是使用的方法却不能语言的复杂性进行分析说明。 请听我一一道来。 首先介绍一下F-K测试方法 麦克唐纳和翁德利(McDonnell&Ondelli) 使用Flesch-Kincaid(F-K)检验。这个测试方法测量的是文本中单词音节的平均长度和句子的平均长度。然后,它给文本打出分值,并以分数来表示阅读所选文本所需的教育知识水平(以上世纪70年代美国的教育体系为衡量标准)。因此,这一测试方法依赖于句子最后的标点符号,也就是“句号”(或对我们大洋彼岸对面的朋友来说是句号)。睿智的读者会马上发现麦克唐纳和翁德利(McDonnell&Ondelli) 正在用一个依赖于句号的测试方法来对口语进行测试研究…就算口语中没有“句号”!这是在听者的脑海中闪过。他们声称: 我曾对斯库韦德Schoonvelde等人的论文分析部分也对F-K检验方法的适用性表示质疑。(2019年)。并且我说明了为什么F-K测试方法对于测试口语来说是不合适的(因此我就不赘述了)。从根本上说,F-K测试方法不够成熟,只适用于书面英语,所以它不适合用于测试口语。对于质疑F-K测试方法不适合测试口语来说,这都是不错的理由,而对于那些具有语言学基础知识的人来说,这个问题是毋庸置疑的。导致这一问题的根本在于这两位作者缺乏对这一主题的知识。 在麦克唐纳和翁德利(McDonnell&Ondelli) 的争辩说,他们没有将F-K测试应用于他们所学的其他语言。至少他们还明白,为英语开发的测试可能并不适用于其他非英语的语言。这一点做的不错!这不是在损你。2019年斯库韦德Schoonvelde等人曾把F-K测试方法运用到西班牙语,法语,德语和荷兰语的测试研究当中,是由于……管他呢,怎么不可以呢?麦克唐纳和翁德利没有跟着做,还算不错。 不过……麦克唐纳和翁德利(McDonnell&Ondelli) 在其他语言上使用的测试好像并没有很有效果。对法语的测试研究,他们使用康德尔模型Kandel-Moles指数,这相当于法语版的F-K测试法。对于意大利语,他们使用了格尔皮斯Gulpease指数,相当于意大利版的F-K测试方法。他们用一种叫做Lix的东西来补充这些分数。这里我不想说他们对法语和意大利语的分析测试,但我要说明的是,这两位作者说LIX“已经被证明在西欧语言测试上表现还不错”,为了支持这一点,他们参引了一家名为SiteImprove的公司(在尾注9中)。不过没人知道这到底是什么?难道他们说这些测试不错我就应该相信吗?哦,算了,很感谢。让人想不到的是,那个公司的所有测试方法--只要付钱就会帮你做的测试--都说是很不错的测试方法。甚至是最好的测试方法。好多人都说这是他们见过的最后的测试方法。 另外值得一说的是,所谓的“西欧语言”又是哪些呢?巴斯克语是西欧语言吗?德语呢?加泰罗尼亚语呢?爱尔兰语呢?马耳他语呢?威尔士语呢?意第绪语?我看就这样吧。他们所谓的“西欧语言”其实不存在于语言学术语中。这就好像一个想给你推销语言蛇油的公司的营销语言。布鲁斯·班纳在《复仇者联盟》中说过瓦坎丹语其实是一种“非洲方言”。 字体/引用量…甚至更多! 接着,这两位作者有一个第二语言分析机来分析他们的演讲语言: 嘿-兄弟!我们终于发现了来自语言学家的参考文献!不过,太差劲了,这两个语言学家(格兰杰和韦恩)从来没说过他们这里引用的话。让我们一起看看这段中所谓的“引文”吧。 “首先,了解词汇是否丰富,因为有这么一个前提:语料库中词汇类型的重复率越高,语言的使用频率就越高,因为词汇的变化除了本身增加了难度外,还可能意味着蕴含了更广泛的内容。”我不知道他们是从哪里得出这个结论的,但我敢肯定的是他们没有引用任何人的言论。这也不是格兰杰和韦恩的话Granger&Wynne(1999)。那么这是谁说的呢?无人知晓! “作为对词汇是否丰富的检查,我们计算了字体/引用比(LTR),用类型/令牌比(TTR)和hapax legomena(即在语料库中只出现一次的词)来表示。”格兰杰和韦恩Granger&Wynne(1999)确实谈到了类型/令牌比。他们称这是“非常简陋的方法”。*悲伤的长号*。格兰杰Granger和韦恩Wynne也讨论了引理/令牌比。他们说“我们的研究表明,在学习者语料库中使用粗略的类型/标记或引理/标记比率是不准确的”。“与学习者语料库”部分很重要。格兰杰Granger和韦恩Wynne调查了英语学习者--而非专业演讲稿文作者--所写的文章,他们测试了是否应该对他们使用传统的词汇丰富度衡量标准。并且他们发现这些方法对于英语学习者调查课文来说太简陋了。他们没有对他们进行演讲或政治文本写作的测试,他们也从来没有谈论过这些测试显示一篇文章有多么“复杂”或“艰涩难读”。 原因在于,当我们在平常提到“词汇表达丰富”时,我们倾向于指使用的很多的不同同根词(即,范式的基本形式,例如“love”作为“loves”,“loved”,“loved”等的基本形式),而非各种各样的屈折形式(Granger和Wynne1999)。Granger和Wynne(1999)从未提到“说话者的丰富词汇表达”。他们只是说:“在同一篇文章中使用五种不同动词形式的go(go/goes/going/gone/went)的学习者,其词汇量比使用五种不同引理(如go/come/leave/enter/return)的学习者少。”词汇表达不够丰富,语言的复杂程度也可见一般。 通过分析,作者确实检验了这种简陋的分析方法,只是她似乎应该给我们一些结论: 类似的,没有任何研究表明音节越多,句子越长越大就意味着文本越复杂。事实上,我们甚至没有提供我们期望的TTR或LTR的衡量标准。什么样的数值才叫高?是否两个政治演讲者的类型-标记比率都很高?或者说都很低?这是一个在测试研究中会再次出现的问题。 先略微让我们停留在参引比率这个话题上。我们已经看到McDonnell&Ondelli引用的消息源称它是一种“简易”的分析形式(还有LTR)。但他们不是唯一的。在朗文辞典的口语和书面语语法中(对于那些不像我这样的人来说,它是英语语法中的一种),我们看到了一些作者可以引用的观点。例如,在2.2.1.2节中,朗曼说:“新闻中的高TTR反映了该语域中名词性成分的使用频率极高,用来表示形形色色的人,地点,物体,事件等。”所以,也许希拉里的高词汇表达率确实意味着她说的内容更多样化更丰富。 不过朗文的接着说说:“在另一个极端,学术论文的TTR是第二低的,这凸显一个事实,即大量学术文章的技术词汇量有限,因此比小说和新闻报道的变化少。”我想大家都同意,学术英语比新闻报道英语更难学--所以证明高TTR并不表示一篇文章更复杂。这正是我说的那反证理由。这两位作者根据很少或根本没有证据而做出的不切实际的假设。 朗文中紧接的一句是,TTR是一个“大概”的度量。但这不是标准! 韩礼德Halliday也曾经讨论过TTR。但他也指出不要用TTR来验证文本更复杂: 韩礼德(Hanlliday)表示,我们对语言及其复杂性的界定与衡量非常依赖于我们所关注的语言的哪一部分。因此不难知道口语会比较复杂。但是要是我们想用TTR来表示语言的复杂性,必须应该建立在什么基础上进行比较呢? 比伯Biber(1988)在关于不同语言体裁之间的差异性的基础工作中通过语言的多个方面来分析每个体裁不同的原因。他谈到语言的复杂性时说:“[话语复杂性的标记]以一种很大程度上互补的模式出现,这一事实表明,语言存在不同类型的复杂性,简单地将特定体裁定性为复杂与否这还是不够的--但是,不同的体裁在不同程度上以不同的方式表示其复杂程度。口语信息体裁的语篇复杂性表现为结构的精炼与否,而计划书,书面体裁的语篇复杂性表现为词汇的精选。“ 这又出现了另一个问题,麦克唐奈和翁德利研究的到底是什么--究竟书面语还是口语的复杂性?因为它们的复杂性可能会以不同的表现方式显示出来。不过我现在更关心的是这个问题! 虚词和实词。比比看吧! 这两位作者又提供了另一种测试方法,很显然这是对F-K测试方法的补充,以作可读性度量的使用: 有谁说过这样的话?我们又如何断定一篇文章传达的信息越多,文章内容就会越难懂呢?要是这里有个懂的人就好了。可怜我们白白等了这许久。(不过还不不算白说,这个词是“虚词”,而非“语法词”。很多语言的语法都使用这个词。) 但是,在这个假设里面还有一个很重要的问题。我们基本都不知道--作者也并没有告诉我们--我们想要什么样的数据。而得到的数据又是什么意思?就像这两位作者所说的,特朗普与希拉里在词汇和语法方面的百分比都比较接近。可能这对英语来说并不奇怪。不管你选择哪种文本,其中40-45%的词会是虚词,而大约50-55%的词是实词(考虑到有些“词”不属于虚词也不是实词,比如感叹词,数字等等)。不过我们不知道我们的预期是什么,也不知道一个信息丰富/复杂的文本到底是什么样的。这意味着我们没有什么可用来比较的标准。我们只能把特朗普和希拉里的数据互相比较一下。但是这说明不了什么。这两位作者应该引用一个例证,或举例一些其他口语体裁的例子。 只因那是你想看到的,对吧?不急。我懂你们的意思,同胞们。 就比如,下文比较的是虚词和实词在各种语料库和小说中所占的比例: 其他口语语料库(来源于COCA和BNC这两大语料库)的比例与特朗普和希拉里之间的比率很接近。电视和电影对白的语料也很接近。就连小说的差距也不是很大。也许虚词和实词的比较并不是衡量一篇文章复杂与否的好方法。 不过查尔明白这个…? 这两位作者使用了一种名为戴尔-查尔分数的测试来计算英语演讲语言中包含了多少“生僻”字。这是为了避免出现我在上面提到的问题:实际上,F-K测试以是单词长度为基准的……不过单词长度在复杂度方面并不重要。马克唐纳McDonnell和Ondelli指出,虽然虚词的长度较短,但它比虚词更难理解。这大概是因为虚词的使用频率较低,但作者认为,一般说话者比替代物词更能理解代词。虽然我不明白他们因何如此。(下文将详细说到) 我想先问的一个问题:要是这些测试方法很差强人意,甚至你还需要对它们进行防范--并且假如它们以某种无关紧要的东西为基本来测试语言的复杂程度,比如单词的长度--那为何还继续使用这样的测试呢?它们好像算不上很好的测试。但无论如何。他们还是做了。 戴尔-查尔测试方法与F-K测验方法大致一样,只是它将文本与一份3000个单词的列表作了比较,而份单词表是美国80%的四年级学生所熟悉的。假若文本里一个单词没有出现在列表中,那么对于一个十岁的美国人来说,该选文被认为是“困难的”。戴尔·查尔Dale-Chall检验方法存在的问题与F-K检验方法的问题是一样的。戴尔·查尔Dale-Chall在分析处理中也包括句子长度。因此,同前,一个用来测试书面语的方法被用在测试口语上了。当作者指出“希拉里使用的句子长度更长的时候(平均每句15.02个单词,而特朗普为12.55个单词)”,他们衡量研究的是抄写者或演讲稿撰写者的标点符号风格。他们是从字面上看“句子”--他们的分句依据是看句号放在哪里,演讲者就在哪里分断一个句子。但这确实是不准确的。口语是没有句号的。口语中绝对没有标点符号。确实没有,断断续续会显得一文不值。句号是书面语的特征之一。而它们在句子中的位置多少有些无情。跟我说说特朗普或希拉里是否用了更多的连贯连词,还有他们在使用这些词时停顿了多久。这将让我们更好地知道演讲者“句子”的长度。 我向神明发誓,我会不断的说这句话直到政治学家明白为止。 这两位作者在附录A中说明,演讲稿中留有抄写员(不管是谁)或大学生的标点符号。他们还说,大学生们检查了抄本,而这两位作者也检查了抄本的一部分(20,000字)。我不知道学生们是否对照实际演讲情形检查了抄本(比如通过视频材料),或者他们只是检查了标点符号和拼写。作者让多名学生抄写同一篇演讲。他们这样做是为了“避免个体转录偏好/特质偏倚结果”。不过我不确定他们是否采取措施避免这件事。 先让我们说说,假如我们想要研究语言的复杂性时,在口语中引入标点符号(也就是书面语的特征)这在方法上可行。只是,请听我说。我们需要加入标点符号,因为我们选用的复杂性测试依赖于标点符号。那么我们应该把它们放在哪里呢?这确实是一个挺难回答的问题。英语中有种叫做连词的东西比如“and”和“but”。有了这些词,我们就可以减少使用句号,并且能把两个独立的分句连接起来。不过我们也可以继续使用句号,然后用这些连词中的一个来引导下一个句子。使用句号的多少将会影响我们所进行的可读性测试的分数。 对的,只是这两位作者已经查过了,这是听你说的。不过他们确实有吗?来看看这些数据:希拉里子语料库比特朗普子语料库多出了218个以连词开头的句子。这些句子中有157个以并列连词引导(比如and,or,but等)。这是为何?如若我们为了能让这些数字更有可比性而将其统一处理,我们可以知道,与特朗普相比,希拉里每10万个单词中多出243个以连词开头的句子。(这些都是语言学术语。要是你清楚也没关系。时间不多,在这里就不解释了。不过我们只要明白作者在做研究之前应该必须要了解这些东西,不过他们看起来好像并不十分了解。) 要是我们进一步推理,我们可以知道希拉里有24%的句子都是以连词开头,而特朗普只有16%。这个差异又是为何?是否与句号的位置有关?也可能抄录者的个人喜好和特质并没有被完全消除。 不过这件事本来是可以规避的。作者可以把这些演讲放到英语语料库域名上的工具里,检测一下每个单词的使用频率。这可以让人们了解哪个演讲者使用更多的常用词汇。不错,语言学家都有这些语料库测试工具。而且它们比一些毫无价值的工具可读性评分要好。只是这两位作者为何不这样做呢?嗯,对,只因他们不是语言学家,因而他们不清楚自己在做什么。 这两位作者做了一些统计测试来检测他们的可读性。只是这对他们所要做的事并没有什么帮助。可读性分数有先天不足。再多的测试也不会让他们变得更成熟。 最后,还记得韩礼德(Halliday)吗?他是语言学家,(在迈克尔·基顿还是蝙蝠侠的时候)他曾在书中写道,TTR并不是用来分析语言复杂性的好方法。之后,他还解释我们为何不该在口语分析中使用“句号”,因为“句号”是书面语的特征之一,它并不适用于测试分析口语。不过,你是否明白,70年代提出使用了这些可读性测试,他们只是被要求使用。谁会在意它们是否值得使用?这里,是否感到一丝凄凉,对吗?请相信你的感觉。 把它弄好做好总结,兄弟们! 虽然他两的研究方法有缺陷,马克唐纳(McDonnell)和韩德利(Ondelli)认为这些测试已经说明了一些重要的结论,他们这样表述: 正如我在前面说的那样:摄入不良,输出不佳。不成熟的测试得到的也是不完美的结论。研究测试结果并不能支持民粹主义者使用更简单语言这一说法,也不能支持特朗普和希拉里在复杂程度上语言有些相似的这一说法,不过这不是很重要,只因他们的研究方法有先天不足。不管结论怎么样,我们都必须拒绝它们(这一主张是我2019年在对斯库韦德(Schoonvelde)等人文章的评论中所说的,不管怎样) 他们希望可以消除他们研究中的存在的主要问题--用测试书面语的测试来研究口语。他们说:“对的,我们不能计算语速,语调以及停顿等因素对语言复杂性的影响。不过,这也适用于其他民族政治语言复杂性的研究。“但这怎么可以是理由呢?“其他的研究方法也有不足,我们的方法也有不足”我认为,这些皆非这项研究(以及其他研究)没有考虑到的口语的唯一理由。此外应该还有还有方言和音调的因素。这些东西影响着听者对口语的评价(其他语言也都是这样)。其他使用可读性评分的研究结论并不好。但这并不代表你可以跟着他们做他们曾经做过的测试方法,尽管没人在意。 好想法。讲究信用。 值得肯定的是,他们的确努力想出了一些关于语言研究的好想法。其中有一些是浅显易懂的--虽然它们并非前人的研究所得(斯库韦德等人,2019年)。不过,另外一些观点好像是这两位作者不经意之间发现的。毕竟他们的研究方法并不能得出任何想要的结论(这是我们在上一节所谈到的),不过这两位作者得出了一些较好的结论。引语后部分以及我所做的评论。 上面我们说到一个很好的想法。该欠的就欠。只是那没有很好的说明,也没有什么例证支持,不过我觉得他们在这方面走的是正确的道路。 另外还有一个很好的想法是斯库韦德等人完全没有的结论。 这是他的优势也是它的不足。不错,这是测试民粹主义者的语言是否使用更简单的语言的一种方法。但麦克唐纳和唐德利并没有好好研究这个方法。这就像“参加考试是完成五年级的一种方式。”确实是,你可能有些考试不及格,不过还是能完成五年级。 神啊!我真不敢相信。他们很直接地说。这是一个很好的想法。F-K测试方法并不能给我们说明语言的复杂与否。你们为何要用这一方法呢? 我想说的是,很开心他们直接地表明他们的研究结果并没有什么作用,除了表示人们(我认为是你们这些政治科学家)应该停止使用不好的可读性分数测试。这就像麦克唐纳和洪德里在建议人们应该用更严谨的方式来学习语言一样。我想知道谁会做那种事。可能是哪一位语言学家? 对的,啥也不是。不好意思 ,这个好想法却令我很生气,只因这个本来是语言学家就能告诉你的。你就只需要求教就行了。您也不必去重新发明自己的语言学。手上只需要拿一本语言学书,看到的人就会知道这些东西--我们称这些人为语言学家,他们以研究语言为生。很奇妙,对吗? 错误的研究方法却得出了的正确的结论 因此,这篇学术论文发表之后,麦克唐纳和欧德利却不知为何得出了正确的结论。它们表示: 是否因为做了错误的研究方法才促使他们得到正确的结论?罗恩·霍华德表示:不错,不错,他们得出正确的结论。] 我想表达的是,我认为我懂得了他们是如何实现的--不过他们没有发现他们的不足之处,我挺吃惊的。 我还要再说一点:批评反驳这类论文研究需要耗费不少时间--我在这里所做的解释按理来说应该由作者或审稿人来进行的。不过他们的论文业已发表,而且已经写在对话中了。就和斯库韦德等人的论文一样,很多人都有听说并会选择相信。真是太糟了,那些缺陷和不足都是作者的责任。 欢迎各学科领域的学者涉足语言学领域。因为其他领域的学者能给语言学研究开辟新的视角。不过你必须展示。如果你要学习语言,你必须把它做对。你不能做一些半生不熟的,草率的研究,并认为它是好的。假若语言学家也在你的专业领域发表类似漏洞百出的论文,你是否会严肃对待呢?我没认为你不会。 参考文献 比伯,道格。1988年. 《语言和书写的变化》剑桥:剑桥大学出版社 《朗文书面语和口语语法》.1999年。道格拉斯·比伯,斯蒂格·约翰逊,杰弗里·利奇,苏珊·康拉德和爱德华·费尼根. 伦敦:朗文出版社 韩礼德(Halliday,M.A.K).1989年。《口语与书面语》(第二版).牛津大学出版社 电子邮箱 推特

以上中文文本为机器翻译,存在不同程度偏差和错误,请理解并参考英文原文阅读。

阅读原文