5. SCIENCE AS CRAFT
Woodward and Goodstein question the general validity of the following principle: Scientists must report what they have done so fully that any other scientist can reproduce the experiment or calculation. They claim that science has a large "skill" or "craft" component, and that Conducting an experiment in a way that produces reliable results is not a matter of following algorithmic rules that specify exactly what is to be done at each step. This may be true of some areas in the biological sciences and other experimental sciences in which the behavior of living organisms or the functioning of complicated instrumentation may not be well understood, but this does not apply to computer simulation experiments. We can and must insist on exact reproducibility of simulation experiments; and this should, in fact, be a matter of following precisely stated, fully documented algorithms. There is of course a large "craft" component in building and using simulation models. Different individuals presented with the same system to be modeled neither build identical simulations nor apply those models in precisely the same way, just as different researchers in any other scientific discipline will neither build the same experimental apparatus nor carry out exactly the same experimental protocol to study a given effect. Nevertheless in these situations different simulationists should be able to reproduce each other's results in order to judge the significance and limitations of the conclusions based on the experiments in question. More generally, there is a large "craft" component in doing simulation research just as there is a large "craft" component in doing other types of scientific research--but this state of affairs does not mitigate the need for reproducibility of the main experiments associated with such research. 6. PEERS AND PUBLICATION 6.1 Is the Scientific Paper a Fraud? Woodward and Goodstein cite Peter Medawar's (1991) paper entitled "Is the Scientific Paper a Fraud?" to argue that because most archival papers in the scientific literature do not accurately portray the way scientific research is actually done, these papers fail to measure up to Feynman's ideal of "leaning over backwards." It is certainly true that primary journal articles in the scientific literature do not document all of the mistakes, dead ends, and backtracking that are an inevitable part of virtually every successful scientific investigation. Medawar (1982, p. 92) himself admitted that I reckon that for all the use it has been to science about four-fifths of my time has been wasted, and I believe this to be the common lot of people who are not merely playing follow-my-leader in research. In my view, the fundamental issue here is that there simply is not enough space in all the scientific journals to document the way that science is actually done; moreover no one has the time to absorb all the final results even in a relatively narrow area of specialization, much less to read the associated background material. Nowadays many high school students are sufficiently sophisticated to realize that primary journal articles are vehicles for efficiently communicating significant discoveries rather than for documenting the processes by which those discoveries were made. Moreover, this issue is rapidly becoming moot because of current trends toward complementing the printed version of a primary journal article with comprehensive supporting documentation (such as appendices containing lengthy proofs or detailed descriptions of experimental protocols) archived on a World Wide Web server that is maintained by the journal's sponsoring organization. 6.2 Problems with the Peer Review System Finally Woodward and Goodstein examine the peer review system for evaluation of research proposals and primary journal articles, concluding that the conflict of interest inherent in asking competitors to evaluate each other's work has inflicted genuine distress on the system. In my own experience, by far the most common form of misconduct by peer reviewers has nothing to do with conflicts of interest; instead the problem is simple dereliction of duty by reviewers who cannot be bothered to read and evaluate carefully the work of other researchers. Although this remark applies to evaluation of research proposals as well as refereeing of primary journal articles, I am most concerned with problems in refereeing. In my judgment, the problem of nonperformance by referees has reached epidemic proportions, and I believe it is urgently necessary for the scientific community to address this scandalous state of affairs. In preparing these remarks I solicited comments from numerous colleagues not only in the simulation community but also in the "hard" scientific disciplines, and I have been startled by the vehemence of their agreement with my evaluation of the current state of the refereeing system. Based on numerous conversations with colleagues in biology, electrical engineering, industrial engineering, mathematics, and statistics, I have a sense that problems with refereeing are much worse in these fields than in the simulation community. Perhaps the most egregious failure of the refereeing system in recent years was the publication of the initial paper on cold fusion by Fleischmann and Pons (1989a). This paper was published in the Journal of Electroanalytical Chemistry in just four weeks; and a long list of errata soon followed (Fleischmann and Pons 1989b)--including the name of M. Hawkins, a coauthor who was somehow omitted from the original paper. A detailed account of this infamous episode can be found on pp. 218-220 of Huizenga (1993). 6.3 Refereeing Remedies The two main reasons for breakdowns in the operation of the refereeing system are (a) misconceptions by referees about the job they are supposed to do, and (b) lack of incentives for doing a good job of refereeing. As Gleser (1986) points out, many referees think that a manuscript must be checked line by line for errors; and seeing that this will be extremely time-consuming, they continually put off the task. On the contrary, the referee's main responsibility is to serve the editor as an "expert witness" in answering certain key questions about the manuscript--and most of these questions can be answered under the assumption that the manuscript is error-free. These key questions are given in Table 2 and are elaborated in Forscher (1965), Gleser (1986), and Macrina (1995, pp. 84-89) along with general guidelines for refereeing that should be required reading for every research worker in the field of computer simulation. Table 2: Key Questions to be Answered in a Referee's Report _____ 1. Are the problems discussed in the paper of substantial interest? Would solutions of these problems materially advance knowledge of theory, methods, or applications? 2. Does the author either solve these problems or else make a contribution toward a solution that improves substantially upon previous work? 3. Are the methods of solution new? Can the proposed solution methods be used to solve other problems of interest? 4. Does the exposition of the paper help to clarify our understanding of this area of research or application? Does the paper hold our interest and make us want to give the paper the careful reading that we give to important papers in our area of specialization? 5. Are the topic and nature of this paper appropriate for this journal? Are the abstract and introduction accessible to a general reader of this journal? Is the rest of the paper accessible to a readily identified group of readers of this journal? 6. Are the clarity and readability of the manuscript acceptable? Is the writing grammatically correct? 7. Does the manuscript contain an adequate set of references? Is adequate credit given to prior work in the field upon which the present paper is built? 8. Is the material appropriately organized into an effective mix of text, figures and tables? Are data given in tables better presented in figures or in the text? 9. Is the work technically correct? Are the main conclusions justified by the experimental data and by logically valid arguments? Are the theorems stated and proved correctly given the assumptions? In practical applications of the theoretical results, do the authors check the validity of the underlying assumptions? 10. Are there gaps in the discussion of the experimental methods or results? If there are such gaps, can the closing of these gaps be considered (i) essential, (ii) desirable, or (iii) interesting? Are the experimental methods described in sufficient detail so that other investigators can reproduce the experiments? _____ If a paper passes the initial screening that consists of answering questions 1-8 in Table 2, then it is necessary to undertake the verification of technical correctness required to answer questions 9 and 10. If competent referees had scrutinized the initial paper on cold fusion by Fleischmann and Pons (1989a) with the objective of answering questions 9 and 10 in Table 2, then the fatal flaws in this work would have been uncovered immediately. In my view it is imperative that we protect the simulation literature against the long-lasting stigma that results from permitting the publication of technically incorrect work. If everyone in the simulation community followed the guidelines in Table 2 for preparing referee's reports, then I believe our problems with peer review would largely disappear. Additional tips on effective refereeing are given by Waser, Price, and Grosberg (1992). A set of questions similar to those given in Table 2 can be found on the home page of the ACM Transactions on Modeling and Computer Simulation by using the URL http://www.acm.org/pubs/tomacs/review/review.html. There remains the question of adequate incentives for good refereeing. In reviewing preliminary versions of these remarks, several individuals complained about general lack of editorial feedback on (a) the strengths and weaknesses of their reviews, and (b) the issues identified in other referees' reports on the same paper. As a routine professional courtesy, editors should include such feedback with their letters of appreciation to referees. Moreover, editors should strive to ensure that individuals who provide prompt and thorough refereeing will receive comparable service when those individuals submit their own papers for review. Ultimately refereeing is one of the professional responsibilities that each of us must fulfill to ensure the vitality of our chosen field, but doing this job well should be a source of pride and satisfaction commensurate with that of our other professional contributions to the field.

