-
Notifications
You must be signed in to change notification settings - Fork 18
/
014-writing.qmd
204 lines (110 loc) · 41.9 KB
/
014-writing.qmd
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
{{< include _setup.qmd >}}
# Writing {#sec-writing}
::: {.callout-note title="learning goals"}
* Write clearly by being concise, using structure, and adjusting to your audience
* Write reproducibly by interleaving writing and analysis code
* Write responsibly by acknowledging limitations, correcting errors, and calibrating your conclusions
:::
You've designed and run your experiment, and you have even analyzed your data. This final part of *Experimentology* discusses reporting your results. We begin by thinking through how to write clearly, reproducibility, and responsibly (this chapter); then we turn to the question of designing informative and pretty data visualizations (@sec-viz). Our final chapter in the section introduces meta-analysis as a tool for research synthesis, allowing us to contextualize research results. These chapters focus on themes of [transparency]{.smallcaps} as well as (especially for meta-analysis) [bias reduction]{.smallcaps} and [measurement precision]{.smallcaps}.
All of the effort you put into designing and running an effective experiment may be wasted if you cannot clearly communicate what you did. Writing is a powerful tool---though you contribute to the conversation only once, it enables you to speak to a potentially infinite number of readers. So it's important to get it right! Here we'll provide some guidance on how to write scientific papers---the primary method for reporting on experiments---clearly, reproducibly, and responsibly.^[Clarity of communication was a founding principle of modern science. Early proto-scientists conducting alchemical experiments often made their work deliberately obscure---even writing in cryptic codes---so that others could not discover the "powerful secrets of nature." Pioneers of scientific methodology, like Francis Bacon and Robert Boyle, pushed instead for transparency and clarity. Notoriously, Issac Newton (originally an alchemist and later a scientist), continued to write in a deliberately obscure fashion in order to "protect" his work [@heard2016].]
## Writing clearly
What is the purpose of writing? "Telepathy, of course," says Stephen King [@king2000]. The goal of writing is to transfer information from your mind to the reader's as effectively as possible. Unfortunately, for most of us, writing clearly does not come naturally; it is a craft we need to work at.
One of the most effective ways to learn to write clearly is to read and to imitate the writing you admire. Many scientific articles are not clearly written, so you will need to be selective in which models you imitate. Fortunately, as a reader, you will know good writing when you see it---you will feel like the writer is sending ideas directly from their mind to yours. When you come across writing like that, try to find more work by the same author. The more good scientific writing you are exposed to, the more you will develop a sense of what works and what does not. You may pick up bad habits as well as good ones (we sure have!), but over time, your writing will improve if you make a conscious effort to weed out the bad, and keep the good.
There are no strict rules of clear writing, but there are some generally accepted conventions that we will share with you here, drawing from both general style guides and those specific to scientific writing [@zinsser2006; @heard2016; @gernsbacher2018; @savage2019].
### The structure of a scientific paper
A scientific paper is not a novel. Rather than reading from beginning to end, readers typically jump between sections to extract information efficiently [@doumont2009]. This "random access" is possible because research articles typically follow the same conventional structure (see @fig-writing-imrad). The main body of the article includes four main sections: introduction, methods, results, and discussion (IMRaD).^[In the old old days, there were few conventions---scientists would share their latest findings by writing letters to each other. But as the number of scientists and studies increased, this approach became unsustainable. The IMRaD structure gained traction in the 1800s and became dominant in the mid-1900s as scientific productivity rapidly expanded in the post-war era. We think IMRaD style articles are a big improvement, even if it is nice to receive a letter every now and again.] This structure has a narrative logic: What's the knowledge gap? (introduction); how did you address it? (methods); what did you find? (results); what do the results mean? (discussion).
Structure helps writers as well as readers. Try starting the writing process with section headings as a structure, then flesh it out, layer by layer. In each section, start by making a list of the key points you want to convey, each representing the first sentence of a new paragraph. Then add the content of each paragraph, and you'll be well on your way to having a full first draft of your article.
Imagine that the breadth of focus in the body of your article has an "hourglass" structure (@fig-writing-imrad). The start of the introduction should have a broad focus, providing the reader with the general context of your study. From there, the focus of the introduction should get increasingly narrow until you are describing the specific knowledge gap or problem you will address and (briefly) how you are going to address it. The methods and results sections are at the center of the hourglass because they are tightly focused on your study alone. In the discussion section, the focus shifts in the opposite direction, from narrow to broad. Begin by summarizing the results of your study, discuss limitations, then integrate the findings with existing literature and describe practical and theoretical implications.
![Conventional structure of a research article. The main body of the article consists of introduction, methods, results, and discussion (IMRaD) sections.](images/writing/imrad.png){#fig-writing-imrad .column-margin width=70% fig-alt="A diagram of a funnel where breadth of writing is widest at introduction, narrower at methods/results, wider at discussion."}
Research articles are often packed with complex information; it is easy for readers to get lost. A "cross-reference" is a helpful signpost that tells readers where they can find relevant additional information without disrupting the flow of your writing. For example, you can refer the reader to data visualizations by cross-referencing to figures or tables (e.g., "see Figure 1"), or additional methodological information in the supplementary information (e.g., "see Supplementary Information A").
A useful trick for structuring complex arguments is to cross-reference your research aims or hypotheses with your results. For example, you could introduce numbered hypotheses in the introduction of an article and then refer to them directly when reporting the relevant analyses and results. These cross-references can serve to remind readers how different results or analyses relate back to your research goals.
### Paragraphs, sentences, and words
Writing an article is like drawing a human form. If you begin by sketching the clothes, you risk adding beautiful textures onto an impossible shape. Instead, you have to start by understanding the underlying skeleton, and then gradually adding layers until you can visualize how cloth hangs on the body. The structure of an article is the "skeleton" and the paragraphs and sentences are the "flesh." Only start thinking about paragraphs and sentences once you have a solid outline in place.
Ideally, each paragraph should correspond to a single point in the article's outline, with the specifics necessary to convince the reader embedded in it. "P-E-E-L" (point-explain-evidence-link) is a useful paragraph structure, particularly in the introduction and discussion sections. First, state the paragraph's message succinctly in the first sentence (P). The core of the paragraph is dedicated to further explaining the point and providing evidence (E-E; you can also include a third "E"---an example). At the end of the paragraph, take a couple of sentences to remind the reader of your point and set up a link to the next paragraph.
Since each sentence in a paragraph has a purpose, you can compose and edit the sentence by asking how its form serves that purpose. For example, short sentences are great for making strong initial points. On the other hand, if you only use short sentences, your writing may come across as monotonous and robotic. Try varying sentence lengths to give your writing a more natural rhythm. Just avoid cramming too much information into the same sentence; very long sentences can be confusing and difficult to process.
You can also use sentence structure as a scaffold to support the reader's thinking. Start sentences with something the reader already knows. For example, rather than writing "We performed a between-subjects $t$-test\index{t-test} comparing performance in the experimental and control groups to address the cognitive dissonance hypothesis," write "To address the cognitive dissonance hypothesis, we compared performance in the experimental group and control group using a between-subjects $t$-test."
Human readers are good at processing narratives about people. Yet, often scientists compromise the research narrative by removing themselves from the process, sometimes even using awkward grammatical constructions to do so. For example, scientists sometimes write "The data were analysed" or, worse, "An analysis of the data was carried out." Many of us were taught to write sentences like these, but it's much clearer to say "We analyzed the data."
Similarly, many of us tend to hide our views with frames and caveats: "[It is believed that/Research indicates that/Studies show that] money leads to increased happiness (Frog & Toad, 1963)." If you truly do believe that money causes happiness, simply assert it---with a citation if necessary. Save caveats for cases where *someone* believes that money causes happiness, but it's *not* you. Emphasize uncertainty where you in fact feel that uncertainty is warranted, and readers will take your doubts more seriously.
## Advice
Scientific writing has a reputation for being dry, dull, and soulless. While it's true that writing research articles is more constrained than writing fiction, there are still ways to surprise and entertain your reader with metaphor, alliteration, and humor. As long as your writing is clear and accurate, we see no reason why you can't also make it enjoyable. Enjoyable articles are easier to read and more fun to write.^[One of our favorite examples of an enjoyable article is @cutler1994, a delightful piece that uses the form of the article to make a point about human language processing. Read it: you'll see!] Here are a few pieces of advice about expressing yourself clearly:
**Be explicit**. Avoid vagueness and ambiguity. The more you leave the meaning of your writing to your reader's imagination, the greater the danger that different readers will imagine different things! So be direct and specific.
**Be concise**. Maximize the signal-to-noise ratio in your writing by omitting needless words and removing clutter [@zinsser2006]. For example, say *we investigated* rather than *we performed an investigation of* and say *if* rather than *in the event that*. Don't try to convey everything you know about a topic---a research report is not an essay. Include only what you need to achieve the purpose of the article and exclude everything else.
**Be concrete**. Concrete examples make abstract ideas easier to grasp. But some ideas are just hard to express in prose, and diagrams can be very helpful in these cases. For example, it may be clearer to illustrate a complex series of exclusion criteria using a flow chart rather than text. You can even use photos, videos, and screenshots to illustrate experimental tasks [@heycke2019].
**Be consistent**. Referring to the same concept using different words can be confusing because it may not be clear if you are referring to a different concept or just using a synonym. For example, in everyday conversation, "replication" and "reproducibility" may sound like two different ways to refer to the same thing, but in scientific writing, these two concepts have different technical definitions, so we should not use them interchangeably. Define each technical term once and then use the same term throughout the manuscript.
**Adjust to your audience**. Most of us adjust our conversation style depending on who we're talking to; the same principle applies to good writing. Knowing your audience is more difficult with writing, because we cannot see the reader's reactions and adjust accordingly. Nevertheless, we can make some educated guesses about who our readers might be. For example, if you are writing an introductory review article, you may need to pay more attention to explaining technical terms than if you are writing a research article for a specialty journal.
**Check your understanding**. Unclear writing can be a symptom of unclear thinking. If an idea doesn't make sense in your head, how will it ever make sense on the page? In fact, trying to communicate something in writing is an excellent way to probe your understanding and expose logical gaps in your arguments. So, if you are finding it difficult to write clearly, stop and ask yourself, *do I know what I want to say*? If the problem is unclear thinking, then it might be worth talking out the ideas with a colleague or advisor before you try to write them down.
<!-- Importantly, you must resist the temptation to mask unclear thinking in unclear writing. Sometimes, your thinking will be unclear simply because *nobody* understands the topic you are addressing. Indeed, we should expect that to happen; the whole point of research is to expand the frontiers of human knowledge, so naturally, you will find yourself working on the fault line of what is and what is not well understood. If you have unresolvable uncertainty about something, say so explicitly, don't hide it in bad writing. -->
**Use acronyms sparingly**. It's tempting to replace lengthy terminology with short acronyms--why say "cognitive dissonance theory" when you can say "CDT"? Unfortunately, acronyms can increase the reader's cognitive burden and cause misunderstandings.^[@barnett2020 found that acronyms are widely used in research articles and argued that they undermine clear communication. Here is one example of text Barnett and Doubleday extracted from a 2019 publication to illustrate the point: "Applying PROBAST showed that ADO, B-AE-D, B-AE-D-C, extended ADO, updated ADO, updated BODE, and a model developed by Bertens et al. were derived in studies assessed as being at low risk of bias."] For example, if you shorten "odds ratio"\index{odds ratio} to "OR," the reader has to take the extra step of translating "OR" back to "odds ratio" every time they encounter it. The problem multiplies as you introduce more acronyms into your article. Worse, for some readers, "OR" tends to mean "operating room," not "odds ratio." Acronyms can be useful, but usually only when they are widely used and understood.
<!-- For example, "magnetic resonance imaging" ("MRI") and "deoxyribonucleic acid" (DNA). At a minimum, be sure to define all acronyms, even when they are likely to be familiar to most readers. -->
### Drafting and revision
The clearest and most effortless-seeming scientific writing has probably gone through extensive revision to appear that way. It can surprise many students to know the amount of revision that has gone into many "breezy" articles. For example, Tversky and Kahneman repeatedly drafted and redrafted each word of their famous (and highly readable) articles on judgment and decision-making, hunched over the typewriter together [@lewis2016undoing].
Think of the article you are writing as a garden. Your first draft may be an unruly mess of intertwined fronds and branches. Several rounds of pruning and sculpting will be needed before your writing reaches its most effective form. You'll be amazed how often you find words you can omit or elaborate sentences you can simplify.
It can be difficult to judge if your own writing has achieved its telepathic goal, especially after several rounds of revision. Try to get feedback from somebody in your target audience. Their comments---even if not wholly positive---will give you a good sense of how much of your argument they understood (and agreed with).^[Seek out people who are willing to tell you that your writing is not good! They may not make you feel good, but they will help you improve.]
## Writing reproducibly
Many research results are not reproducible---that is, the numbers and graphs that they report can't be recreated by repeating the original analyses---even on the original data. As we discussed in @sec-replication, a lack of reproducibility is a big problem for the scientific literature; if you can't trust the numbers in the articles you read, it's much harder to build on the literature.
Fortunately, there are number of tools and techniques available that you can use to write fully reproducible research reports. The basic idea is to create an unbroken chain that links every part of the data analysis pipeline,\index{analysis pipeline} from the raw data through to the final numbers reported in your research article. This linkage enables you---and hopefully others as well---to trace the provenance\index{provenance} of every number and recreate (reproduce) it from scratch.
### Why write reproducible reports?
There are (at least) three reasons to write reproducible reports. First, data analysis is an error-prone activity. Without safeguards in place, it can be easy to accidentally overwrite data, mislabel experimental conditions,\index{experimental condition} or copy and paste the wrong statistics. As we discussed in @sec-replication, one study found that nearly half of a sample of psychology papers contained obvious statistical reporting errors [@nuijten2016]. You can reduce opportunities for error by adopting a reproducible analysis workflow that avoids error-prone manual actions, like copying and pasting.
Second, technical information about data analysis can be difficult to communicate in writing. Prose is often ambiguous, and authors can inadvertently leave out important details [@hardwicke2018b]. By contrast, a reproducible workflow documents the entire analysis pipeline\index{analysis pipeline} from raw data to research report exactly as it was implemented, describing the origin of any reported values and allowing readers to assess, verify, and repeat the analysis process.
Finally, reproducible workflows are typically more efficient workflows. For example, you may realize you forgot to perform data exclusions and need to rerun the analysis. You may produce a plot and then decide you'd prefer a different color scheme. Or perhaps you want to output the same results table in a PDF and in a PowerPoint slide. In a reproducible workflow, all of the analysis steps are scripted and can be easily rerun at the click of a button. You (and others) can also reuse parts of your code in other projects, rather than having to write from scratch.
### Principles of reproducible writing
Below we outline some general principles of reproducible writing. These can be put in practice in a number of different software ecosystems. We recommend R Markdown and its successor, Quarto, which are ways of writing data analysis code in R so that it compiles into spiffy documents or even websites. (This book was written in Quarto.) [Chapter @sec-rmarkdown] gives an introduction to the nuts and bolts of using these tools to create scientific papers.
* **Never break the chain**. Every part of the analysis pipeline\index{analysis pipeline}---from raw data^[Modulo the privacy concerns discussed in @sec-management, of course.] to final product---should be present in the project repository. By consulting the repository documentation, a reader should be able to follow the steps to go from the raw data to the final manuscript, including tables and figures.
* **Script everything**. Try to ensure that each step of the analysis pipeline\index{analysis pipeline} is executed by computer code rather than manual actions such as copying and pasting or directly editing spreadsheets. This practice ensures that every step is documented via executable code rather than ambiguous description, ensuring that it can be reproduced. Imagine, for example, that you decided to recode a variable in your dataset. You could use the "find and replace" function in Excel, but this action would not be documented---you might even forget that you did it! A better option would be to write a script. While a scripted pipeline can be a pain to set up the first time, by the third time you rerun it, it will save you time.
* **Use literate programming**. The meaning of a chunk of computer code is not always obvious to another user, especially if they're not an expert. Indeed, we frequently look at our own code and scratch our heads, wondering what on earth it's doing. To avoid this problem, try to structure your code around plain language comments that explain what it should be doing, a technique known as "literate programming" [@knuth1992].
* **Use defensive programming**. Errors can still occur in scripted analyses. Defensive programming is a series of strategies to help anticipate, detect, and avoid errors in advance. A typical defensive programming tool is the inclusion of tests in your code, snippets that check if the code or data meet some assumptions. For example, you might test if a variable storing reaction times has taken on values below zero (which should be impossible). If the test passes, the analysis pipeline\index{analysis pipeline} continues; if the test fails, the pipeline halts and an error message appears to alert you to the problem.
* **Use free/open-source software and programming languages**. If possible, avoid using commercial software, like SPSS or Matlab, and instead use free, open-source software and programming languages, like JASP, Jamovi, R, or Python. This practice will make it easier for others to access, reuse, and verify your work---including yourself!^[Several of us have libraries of old Matlab code. While discounted licenses are available to students, a full-price software license can be a major barrier to researchers with limited resources. if you move away from Matlab, it's also terrible to have to ask yourself whether it's worth the price of another year's license just to rerun one old analysis.]
* **Use version control**. In @sec-management, we introduced the benefits of version control---a great way to save your analysis pipeline\index{analysis pipeline} incrementally as you build it, allowing you to roll back to a previous version if you accidentally introduce errors.
* **Preserve the computational environment**. Even if your analysis pipeline\index{analysis pipeline} is entirely reproducible on your own computer, you still need to consider whether it will run on somebody else's computer, or even your own computer after software updates. You can address this issue by documenting and preserving the computational environment in which the analysis pipeline runs successfully. Various tools are available to help with this, including Docker, Code Ocean, renv (for R), and pip (for Python).^[If you are interested in going in this direction, we recommend @peikert2021, which gives an advanced tutorial for complete computational reproducibility using Docker and `make` as tools to supplement Git and R Markdown.]
### The reproducibility-collaboration trade-off
We would love to leave it there and watch you walk off into the sunset with a spring in your step and a reproducible report under your arm. Unfortunately, we have to admit that writing reproducibly can create a few practical difficulties when it comes to collaboration.
A major aspect of collaboration is exchanging comments and inline text edits with coauthors. You can do this exchange with R Markdown files and Git,\index{Git} but these tools are not as user-friendly as, say, Word or Google Docs, and some collaborators will be completely unfamiliar with them. Most journals also expect articles to be submitted as Word documents. Outputting R Markdown files to Word can often introduce formatting issues, especially for moderately complex tables. So, until more user-friendly tools are introduced, some compromise between reproducibility and collaboration may be necessary. Here are two workflow styles for you to consider.
First, the *maximal reproducibility* approach. If your collaborators are familiar with R Markdown and you don't mind exchanging comments and edits via Git\index{Git}---or if they don't mind giving you lists of comments and changes that you implement in the R Markdown document---then you can maintain a fully reproducible workflow for your project. The journal submission and publication process may still introduce some issues, such as incorporating changes made by the copy editor, but at least your submitted manuscript (and the preprint you have hopefully posted) will be fully reproducible.
Second, the *two worlds* approach. This workflow is a bit clunky, but it facilitates collaboration and maintains reproducibility. First, write your results section in R Markdown and generate a Word document (see @sec-rmarkdown). Then, write the remainder of the manuscript in Word, including incorporating comments and changes from collaborators. When you have a final version, copy and paste the abstract, introduction, methods, and discussion into the R Markdown document.^[You can also incorporate Google Docs into this workflow---we find that cloud platforms like Docs are especially useful when gathering comments from multiple collaborators on the same document. Unfortunately, you cannot generate a Google Doc from R Markdown, so you will need to import and convert or else copy and paste.] Integrating any changes made to the results section back into the R Markdown requires a bit more effort, either using manual checking or Word's "compare documents" feature.^[Packages such as `trackdown` [@kothe2021] could help as well (<https://claudiozandonella.github.io/trackdown>).] The advantage of this approach is that you have a reproducible document and your collaborators have not had to deviate from their preferred workflow. Unfortunately, it requires more effort from you and is slightly more error-prone than the maximal reproducibility approach.
## Writing responsibly
As a scientific writer, you have both professional and ethical responsibilities. You must communicate all relevant information about your research so as to enable proper evaluation and verification by other scientists. It is also important not to overstate your findings and calibrate your conclusions to the available evidence [@hoekstra2021]. If errors are found in your work, you must respond and correct them when possible [@bishop2018]. Finally, you must meet scholarly obligations with regards to authorship and citation practices.
### Responsible disclosure and interpretation
Back in school, we all learned that getting the right answer is not enough---you need to demonstrate how you arrived at that answer in order to get full marks. The same expectation applies to research reports. Don't just tell the reader what you found, tell them how you found it.^[It can be easy to overlook important details, especially when you reach the end of a project. Looking back at your study preregistration can be a helpful reminder. Reporting guidelines for different research designs can also provide useful checklists [@appelbaum2018].] That means describing the methods in full detail, as well as sharing data, materials, and analysis scripts.
In a journal article, you typically have some flexibility in terms of how much detail you provide in the main body of the article and how much you relegate to the supplementary information. Readers have different needs; some may just want to know the highlights, and some will need detailed methodological information in order to replicate your study. As a rule of thumb, try to make sure there is nothing relegated to the supplementary information that might surprise the reader. You certainty should not use the supplementary information to hide important details deliberately or use it as a disorganized dumping ground---the principles of clear writing still apply!
Here are a few more guidelines for responsible writing:
* **Don't overclaim**. Scientists often feel they are (and unfortunately, often are) evaluated based on the *results* of their research, rather than the *quality* of their research. Consequently, it can be tempting to make bigger and bolder claims than are really justified by the evidence. Think carefully about the limitations of your research and calibrate your conclusions to the evidence, rather than what you wish you were able to claim. Ensure that your conclusions are appropriately stated throughout the manuscript, especially in the title and abstract.
* **Acknowledge limitations on interpretation and generalizability**. Even if you calibrate your claims appropriately throughout, there are likely specific limitations that are worth discussing, either as you introduce the design of the study in the introduction or as you interpret it in the discussion section. For example, if your experiment used one particular manipulation to instantiate a construct of interest, you might discuss this limitation and how it might be addressed by future work. Think carefully about the limitations of your study, state them clearly, and consider how they impact your conclusions [@clarke2023].^[Should you just make your claims more modest and avoid writing about your study's limitations? The balance between claims and limitations is tricky. One way to navigate this issue is to ask yourself, "Is it OK to say X in the abstract of my article, if I later go on to say state a limitation relevant to that claim, or will the reader feel tricked?"] Discussions of limitations are a great point to make an explicit statement about the _generalizability_ of your findings [see @simons2017 for guidance about these kinds of statements].
* **Discuss, don't debate**. The purpose of the discussion section is to help the reader interpret your research. Importantly, a journal article is not a debate---don't feel the need to argue dogmatically for a particular position or interpretation. You should discuss the strengths and weaknesses of the evidence, and the relative merits of different interpretations. For example, perhaps there is a potential confounding variable that you were unable to eliminate with your research design. The reader might be able to spot this themselves, but regardless, its your responsibility to highlight it. Perhaps on balance you think the confound is unlikely to explain the results---that's fine, but you need to explain your reasoning to the reader.
* **Disclose conflicts of interest and funding**. Researchers are usually personally invested in the outcomes of their research, and this investment can lead to bias (for example, overclaiming or selective reporting\index{selective reporting}). But sometimes your potential personal gains from a piece of research rise above a threshold and are considered **conflicts of interest**\index{conflict of interest}. Where this threshold lies is not always completely clear. The most obvious conflicts of interest occur when you stand to benefit financially from the outcomes of your research (for example, a drug developer evaluating their own drug). If you are in doubt about whether you have a potential conflict of interest, then you should disclose it. You should also disclose any funding you received for the research, partly because this is often a requirement of the funder and partly because it may represent a conflict of interest, for example, if the funder has a particular stake in the outcome of the research. To avoid ambiguity, you should also disclose when you do *not* have a conflict of interest or funding to declare.
* **Report transparently**. In @sec-prereg, you learned about the problem of selective reporting\index{selective reporting} and how this practice can bias the research literature. There are several ways to avoid this issue in your own work. First, assuming you *have* reported everything, include a statement in the methods section that explicitly says so. A statement suggested by @simmons2012 is "We report how we determined our sample size, all data exclusions (if any), all manipulations, and all measures in the study." If you have preregistered your study, clearly link to the preregistration and state whether you deviated from your original plan. You can include a detailed preregistration disclosure table in the supplementary information and highlight any major deviations in the methods section. In the results section, clearly identify (e.g., with subheadings) which analyses were preplanned and included in the preregistration (confirmatory) and which were not planned (exploratory).
### Responsible handling of errors
It is not your responsibility to never make mistakes. But it *is* your responsibility to respond to errors in a timely, transparent, and professional manner [@bishop2018].^[As jazz musician Miles Davis once said, "If you hit a wrong note, it's the next note that you play that determines if it's good or bad."] Regardless of how the error was identified (e.g., by yourself or by a reader), we recommend contacting the journal and requesting that they publish a correction statement (sometimes called an *erratum*). Several of us have corrected papers in the past. If the error is serious and cannot be fixed, you should consider retracting the article.
A correction/retraction\index{retraction} statement should include the following:
1. **Acknowledge the error**. Be clear that an error has occurred.
2. **Describe the error**. Readers need to know the exact nature of the error.
3. **Describe the implications of the error**. Readers need to know how the error might affect their interpretation of the results.
4. **Describe how the error occurred**. Knowing how the error happened may help others avoid the same error.
5. **Describe what you have done to address the error**. Others may learn from solutions you've implemented.
6. **Acknowledge the person who identified the error**. Identifying errors can take a lot of work; if the person is willing to be identified, give credit where credit is due.
::: {.callout-note title="accident report"}
In 2018, at a crucial stage of her career, Julia Strand\index{Strand, Julia} published an important study in the prestigious journal *Psychonomic Bulletin & Review*. She presented the work at conferences and received additional funding to do follow-up studies. But several months later, her team found that they could not replicate the result.
Puzzled, she began searching for the cause of the discrepant results. Eventually, she found the culprit---a programming error. As she sat staring at her computer in horror, she realized that it was unlikely anyone else would ever find the bug. Hiding the error must have seemed like the easiest thing to do.
But she did the right thing. She spent the next day informing her students, her coauthors, the funding officer, the department chair overseeing her tenure review, and the journal---to initiate a retraction\index{retraction} of the article. And... it didn't ruin her career. Everybody was understanding and appreciated that she was doing the right thing. The journal corrected the article. She didn't lose her grant. She got tenure. And a lot of scientists, including us, admire her for what she did.
Honest mistakes happen---it's how you respond to them that matters [@strand2021]. In fact, survey research with both scientists and the general public suggests that scientists' reputations are built on the perception that they try to "get it right," not just to "be right" [@ebersole2016b].
:::
### Responsible citation
Citing prior work that your study builds upon ensures that researchers receive credit for their contributions and helps readers to verify the basis of your claims. You should certainly avoid copying the work of others and presenting it as your own (see @sec-ethics for more on plagiarism).\index{plagiarism} Try to be explicit about why you are citing a source. For example, does it provide evidence to support your point? Is it a review paper that gives the reader useful background? Or is it a description of a theory you are testing?
Make sure you read articles before you cite them. @stang2018 reports a cautionary tale in which a commentary criticizing a methodological tool was frequently cited as *supporting* the use of that tool! It seems that many authors had not read the paper they were citing, which is both misleading and embarrassing.
Try to avoid selective or uncritical citation. It is misleading to cite only research that supports your argument and ignoring research that doesn't. You should provide a balanced account of prior work, including contradictory evidence. Make sure to evaluate and integrate evidence from prior studies, rather than simply describing them. Remember---every study has limitations.
### Responsible authorship practices
It is an ethical responsibility to credit the individuals who worked on a research project---so that they can reap the benefits if the work is influential, but also so that they can take responsibility for errors.^[In 1975, physicist and mathematician Jack H. Hetherington wrote a paper he intended to submit to the journal *Physical Review Letters*. We're not sure why, but Hetherington wrote the paper in first person plural (i.e., referring to himself as "we" rather than "I"). He subsequently discovered that the journal would not accept the use of "we" for single-authored articles. Hetherington had painstakingly tapped out the article on his typewriter, an exercise he was not keen to repeat. Instead, he opted for a less taxing solution and named his cat---a feline by the name of F. D. C. Willard---as a coauthor. The paper was accepted and published [@hetherington1975].]
Currently in academia, the *authorship model* is dominant. Under this model, authorship and authorship order are important signals about researchers's contributions to a project. It is generally expected that to qualify for authorship, an individual should have made a substantial contribution to the research (e.g., design, data collection, analysis) and assisted with writing the research report, and that they take joint responsibility for the research along with the other coauthors. Individuals who worked on the project who do not reach this threshold are instead mentioned in a separate acknowledgements section and not considered authors.
**Authorship order** is often understood to signal the nature and extent of an author's contribution. In psychology (and neighboring disciplines), the first author and last author are typically the project leaders. Typically---though certainly not always!---the first author is a junior colleague who implements the project and the last author is a senior colleague who supervises the project.
It has been argued that the authorship model should be replaced with a more inclusive *contributorship* model in which all individuals who worked on the project are acknowledged as "contributors." Unlike the authorship model, there is no arbitrary threshold for contributorship. The actual contributions of each individual are explicitly described, rather than relying on the implicit conventions of authorship order. The contributorship model may facilitate collaboration and ensure student assistants are properly credited.
You will probably find that most journals still expect you to use the authorship model. Nevertheless, it is usually possible---and sometimes required---to include a contributorship statement in your article that describes what everybody did. For example, the CREDIT taxonomy provides a structured taxonomy of research tasks, making for uniform contributorship reporting.^[For larger projects, the tool Tenzing allows for CREDIT statements to be generated automatically from standardized forms [@holcombe2020].]
Because authorship is such an important signal in academia, it's important to agree on an authorship plan with your collaborators (particularly who will be the first and last authors) as early as possible.^[If you find yourself in a situation where all authors have contributed equally, you may have to draw inspiration from historical examples and determine authorship order based on a 25-game croquet series [@hassell1974]; rock, paper, scissors [@kupfer2004]; or a brownie bake-off [@young1992]. Alternatively, you can adopt the method of @lakens2018 and randomize the authorship order in R!]
## Chapter summary: Writing
Writing a scientific article can be a rewarding endpoint for the process of doing experimental research. But writing is a craft, and writing clearly---especially about complex and technical topics---can require substantial practice and many drafts. Further, writing about research comes with ethical and professional responsibilities that are different than the burdens of other kinds of writing. A scientific author must work to ensure the reproducibility of their findings and report on those findings responsibly, noting limitations and weaknesses as well as strengths.
::: {.callout-note title="discussion questions"}
1. Find a writing buddy and exchange feedback on a short piece of writing (the abstract of a paper in progress, a conference abstract, or even a class project proposal would be good examples). Think about how to improve each other's writing using the advice offered in this chapter.
2. Identify a published research article with openly available data and see if you can reproduce an analysis in their paper by recovering the exact numerical values they report. You can find support for this exercise at the Social Science Reproduction Platform (<https://www.socialsciencereproduction.org>) or ReproHack (<https://www.reprohack.org>). Discuss with a friend what challenges you faced in this exercise and how they might be avoided in your own work.
:::
::: {.callout-note title="readings"}
* Zinsser, William [-@zinsser2006]. *On Writing Well: The Classic Guide to Writing Nonfiction*. 7th ed. Harper Collins.
* Gernsbacher, Morton Ann [-@gernsbacher2018]. "Writing Empirical Articles: Transparency, Reproducibility, Clarity, and Memorability." *Advances in Methods and Practices in Psychological Science* 1 (3): 403–414. <https://doi.org/10.1177/2515245918754485>.
:::