__________________________________________________

Announcement

Theme: San Francisco Declaration on Research Assessment
Type: San Francisco Declaration on Research Assessment
Institution: American Society for Cell Biology (ASCB)
Web: http://am.ascb.org/dora/

__________________________________________________


San Francisco Declaration on Research Assessment
Putting science into the assessment of research

There is a pressing need to improve the ways in which the output of
scientific research is evaluated by funding agencies, academic
institutions, and other parties.

To address this issue, a group of editors and publishers of scholarly
journals met during the Annual Meeting of The American Society for
Cell Biology (ASCB) in San Francisco, CA, on December 16, 2012. The
group developed a set of recommendations, referred to as the San
Francisco Declaration on Research Assessment. We invite interested
parties across all scientific disciplines to indicate their support
by adding their names to this Declaration.

The outputs from scientific research are many and varied, including:
research articles reporting new knowledge, data, reagents, and
software; intellectual property; and highly trained young scientists.
Funding agencies, institutions that employ scientists, and scientists
themselves, all have a desire, and need, to assess the quality and
impact of scientific outputs. It is thus imperative that scientific
output is measured accurately and evaluated wisely.

The Journal Impact Factor is frequently used as the primary parameter
with which to compare the scientific output of individuals and
institutions. The Journal Impact Factor, as calculated by Thomson
Reuters, was originally created as a tool to help librarians identify
journals to purchase, not as a measure of the scientific quality of
research in an article. With that in mind, it is critical to
understand that the Journal Impact Factor has a number of
well-documented deficiencies as a tool for research assessment. These
limitations include: A) citation distributions within journals are
highly skewed [1–3]; B) the properties of the Journal Impact Factor
are field-specific: it is a composite of multiple, highly diverse
article types, including primary research papers and reviews [1, 4];
C) Journal Impact Factors can be manipulated (or "gamed") by
editorial policy [5]; and D) data used to calculate the Journal
Impact Factors are neither transparent nor openly available to the
public [4, 6, 7].

Below we make a number of recommendations for improving the way in
which the quality of research output is evaluated. Outputs other than
research articles will grow in importance in assessing research
effectiveness in the future, but the peer-reviewed research paper
will remain a central research output that informs research
assessment. Our recommendations therefore focus primarily on
practices relating to research articles published in peer-reviewed
journals but can and should be extended by recognizing additional
products, such as datasets, as important research outputs. These
recommendations are aimed at funding agencies, academic institutions,
journals, organizations that supply metrics, and individual
researchers.

A number of themes run through these recommendations:

the need to eliminate the use of journal-based metrics, such as
Journal Impact Factors, in funding, appointment, and promotion
considerations; the need to assess research on its own merits rather
than on the basis of the journal in which the research is published;
and the need to capitalize on the opportunities provided by online
publication (such as relaxing unnecessary limits on the number of
words, figures, and references in articles, and exploring new
indicators of significance and impact). We recognize that many
funding agencies, institutions, publishers, and researchers are
already encouraging improved practices in research assessment. Such
steps are beginning to increase the momentum toward more
sophisticated and meaningful approaches to research evaluation that
can now be built upon and adopted by all of the key constituencies
involved.

The signatories of the San Francisco Declaration on Research
Assessment support the adoption of the following practices in
research assessment.

General Recommendation

1. Do not use journal-based metrics, such as Journal Impact Factors,
as a surrogate measure of the quality of individual research
articles, to assess an individual scientist's contributions, or in
hiring, promotion, or funding decisions.

For funding agencies

2. Be explicit about the criteria used in evaluating the scientific
productivity of grant applicants and clearly highlight, especially
for early-stage investigators, that the scientific content of a paper
is much more important than publication metrics or the identity of
the journal in which it was published.

3. For the purposes of research assessment, consider the value and
impact of all research outputs (including datasets and software) in
addition to research publications, and consider a broad range of
impact measures including qualitative indicators of research impact,
such as influence on policy and practice.

For institutions

4. Be explicit about the criteria used to reach hiring, tenure, and
promotion decisions, clearly highlighting, especially for early-stage
investigators, that the scientific content of a paper is much more
important than publication metrics or the identity of the journal in
which it was published.

5. For the purposes of research assessment, consider the value and
impact of all research outputs (including datasets and software) in
addition to research publications, and consider a broad range of
impact measures including qualitative indicators of research impact,
such as influence on policy and practice.

For publishers

6. Greatly reduce emphasis on the journal impact factor as a
promotional tool, ideally by ceasing to promote the impact factor or
by presenting the metric in the context of a variety of journal-based
metrics (e.g., 5-year impact factor, EigenFactor [8], SCImago [9],
h-index, editorial and publication times, etc.) that provide a richer
view of journal performance.

7. Make available a range of article-level metrics to encourage a
shift toward assessment based on the scientific content of an article
rather than publication metrics of the journal in which it was
published.

8. Encourage responsible authorship practices and the provision of
information about the specific contributions of each author.

9. Whether a journal is open-access or subscription-based, remove all
reuse limitations on reference lists in research articles and make
them available under the Creative Commons Public Domain Dedication
[10].

10. Remove or reduce the constraints on the number of references in
research articles, and, where appropriate, mandate the citation of
primary literature in favor of reviews in order to give credit to the
group(s) who first reported a finding.

For organizations that supply metrics

11. Be open and transparent by providing data and methods used to
calculate all metrics.

12. Provide the data under a licence that allows unrestricted reuse,
and provide computational access to data, where possible.

13. Be clear that inappropriate manipulation of metrics will not be
tolerated; be explicit about what constitutes inappropriate
manipulation and what measures will be taken to combat this.

14. Account for the variation in article types (e.g., reviews versus
research articles), and in different subject areas when metrics are
used, aggregated, or compared.

For researchers

15. When involved in committees making decisions about funding,
hiring, tenure, or promotion, make assessments based on scientific
content rather than publication metrics.

16. Wherever appropriate, cite primary literature in which
observations are first reported rather than reviews in order to give
credit where credit is due.

17. Use a range of article metrics and indicators on
personal/supporting statements, as evidence of the impact of
individual published articles and other research outputs [11].

18. Challenge research assessment practices that rely inappropriately
on Journal Impact Factors and promote and teach best practice that
focuses on the value and influence of specific research outputs.

References

1.  Adler, R., Ewing, J., and Taylor, P. (2008) Citation statistics. A
    report from the International Mathematical Union.
    www.mathunion.org/publications/report/citationstatistics0
2.  Seglen, P.O. (1997) Why the impact factor of journals should not
    be used for evaluating research. BMJ 314, 498–502.
3.  Editorial (2005). Not so deep impact. Nature 435, 1003–1004.
4.  Vanclay, J.K. (2012) Impact Factor: Outdated artefact or
    stepping-stone to journal certification. Scientometrics 92,
    211–238.
5.  The PLoS Medicine Editors (2006). The impact factor game. PLoS Med
    3(6): e291 doi:10.1371/journal.pmed.0030291.
6.  Rossner, M., Van Epps, H., Hill, E. (2007). Show me the data. J.
    Cell Biol. 179, 1091–1092.
7.  Rossner M., Van Epps H., and Hill E. (2008). Irreproducible
    results: A response to Thomson Scientific. J. Cell Biol. 180,
    254–255. 8. http://www.eigenfactor.org/
9.  http://www.scimagojr.com/
10.
http://opencitations.wordpress.com/2013/01/03/open-letter-to-publishers
11. http://altmetrics.org/tools/


Website of the Declaration:
http://am.ascb.org/dora/
http://am.ascb.org/dora/index.php/sign-the-declaration




__________________________________________________


InterPhil List Administration:
http://interphil.polylog.org

Intercultural Philosophy Calendar:
http://cal.polylog.org

__________________________________________________

 

Reply via email to