Open Access
10 January 2019 Ensuring Scientific Publishing Credibility in Translational Biomedical Optics
Author Affiliations +
Abstract
Editor-in-Chief Brian Pogue writes about publishing credibility in the field of translational biomedical optics.

JBO_24_1_010101_f001.pngOptics has consistently been the largest singular technology sector used in medicine, and major advances in biomedical optics are documented daily in peer-reviewed publications. However, the academic stature of this field can be damaged by weaknesses in scientific publishing, where a “credibility crisis” has emerged as a popularized and increasingly studied dialogue. While there are still relatively few overt cases of fraud or erroneous research, more insidious aspects are seen in papers with results that have either low statistical power, selective reporting of observations, or data or computer codes that cannot be independently verified. Interestingly, the same solutions that improve scientific publishing quality and credibility can also be effective tools to foster growth of individual scientists. The solution for our biomedical optics community is to ensure that researchers allow and promote reproduction of results, effectively provide access to original data and computer codes, and stay actively involved in translating their results into practice. Along the way, researchers should benefit from transdisciplinary collaborations and mentoring networks of colleagues, involving both medical and commercial expertise. Publishing more impactful publications makes the entire field more impactful, through a sequence of quality measures and a focus on translation to improve industry and medicine.

1.

The Credibility Problems in Scientific Publishing

The nature of the scientific publishing credibility problem is in two distinct areas. The first is in the aspect of the highest impact journals, where career pressures and financial programs have created a situation where the intensity of desire to publish is driven by nonscientific goals.1 There have been high-profile retractions of fraudulent papers, and less acute but equally problematic lack of successes in translating therapeutics when going from basic animal studies to human clinical trials.2 The highest profile retractions have been documented by the journals themselves in recent years and publicized on websites,1,3 with some independent meta-analyses undertaken to understand the scope of the problem. The second distinct problem is at the bottom end of the journal impact range, where there has been enormous growth of journals with questionable credibility, willing to publish low-quality materials for a price with minimal or no real peer review.4 These two issues are separate, but both are symptoms of a system that is losing touch with core principles and appropriate driving influences of an open scientific community.

The high impact factor (IF) journal credibility crisis has been a growing and openly debated issue, documented by some of the same leading publishers. A poll of Nature readers indicated that 90% of respondents thought there was a credibility crisis in their own published papers, with 52% indicating it was a significant crisis.1 This self-reported survey has been backed by scientific meta-analysis studies verifying the magnitude of the problem, but others also argue that it remains in proportion to the number of articles published.5 There have been major discussion panels sponsored to try and solve these issues, leading to white papers and proposed methodological solutions.6 Ultimately it is still up to each publisher to self-police their credibility. But societal and governmental bodies can get involved when expensive drugs fail to successfully translate from the preclinical or early clinical phase into pivotal trials. These are the most visible failures, and the most costly for companies and funding agencies, where the cost is ultimately to all of society. However, while there are clearly high-profile investigations and retractions from fraudulent research activities, most argue that these are still rare events, and the more insidious concerns are those where authors are accused of selective reporting, low statistical power, lack of self-replication, and insufficient oversight or mentoring of junior scientists doing the work and writing the papers. These factors, combined with pressures to publish for career success, lead to some irreproducible papers that make it through the peer-review system. The problem becomes further amplified by some local institutional or governmental financial systems that provide financial incentives for individual papers being published in high impact factor journals,7 thereby prioritizing short-term successes over long-term productivity goals, where quality and longevity of success might be better assessed. The more measured reward system of grant funding and scientist promotions over years provides a balance over multiple or sustained successes, which rewards achievements with demonstrated longevity over quick publication successes, and can take into account a better measure of quality metrics such as citation rates and translation efficacy of discoveries.8,9

The growth of low-impact journals is also well documented through grassroots efforts to identify so-called “predatory journals,” which have the phenotype of low or absent standards for review.1012 There has been a robust grassroots effort to highlight these journals,13,14 which can be identified by a number of features, such as the lack of an editorial board, lack of transparency in their website and publisher identity, and lack of indexing in known services such as Medline. Still, the number of these journals has ballooned, and while there are many shades of grey in validity, estimates indicate there are nearly 8000 such journals. Most obviously, these journals may not appropriately peer review their published papers. Unfortunately, the scientists who are predisposed to submit to these journals are often the most marginalized, by being located at low-resource countries or institutions, or junior investigators whose guidance and motivations may not be well placed.11,15 It is critical for the development of these individuals, and for the field itself, to provide guidance on optimal routes for peer review and to help provide mentorship on the appropriate ways to complete and sustain impactful science.

The easiest way for scientists to know if the journal that they are submitting to is appropriate for them is to be aware of the other peer scientists in their field and choose journals that their peers use,14 and to focus on publishing with their community. This process contributes to the overall development of a network scientists who are working toward competitive success with each other. But beyond the single publishing event, the key linkage from scientist to scientist is the broader idea that scientific research requires open scholarly communication as well as scholarly competition. This is the missing ingredient that must be continually refreshed and renewed in our community, helping the field focus on long-term gains and successes instead of individual publishing events. The transition of the Journal of Biomedical Optics to fully open access and the ability to include supplementary materials with the original manuscript is the right direction for our field, so that all research can be fully understood and reproduced. We need to encourage our community to be open with their work, to ensure reproducibility and transparency in the process of promoting discoveries and developments.

2.

Accurate Measures of Success in Translational Research

The changes that are occurring in the publishing world need attention, but these shifts indicate that attention must also be paid to broader measures of success in scientific discovery, providing a reward system that has the right balance. The discussion is paralleled by a related issue, which is about how to evaluate scientists as they come up for tenure and promotion. Recently, a report by Ioannidis and Khoury16 proposed a new set of metrics for assessing success over the long term, proposing their PQRST approach, where this acronym represents the five metrics of productivity, quality, reproducibility, sharing, and translation, as outlined in Table 1. This set of metrics works quite well for a translational field such as biomedical optics.

Table 1

The five metrics in the PQRST approach.

PProductivitypublications, continuity of output, output per funded project, outreach
QQualitycitation metrics, longevity of value, documented quality standards
RReproducibilityrepeated measures, repeated by others
SSharingdata, computer code, materials, protocols
TTranslationin vitro → in vivo → human trial → startup → industry → level 1 evidence

These goals for measuring impact are all features of an academic pipeline for a translational researcher, who is seeking to go from publishing through to eventual translation. They form an outstanding way to think about the successes of an individual researcher, a research institute, or a scientific community publication. The steps in this sequence where the publishing world can help is to ensure that reproducibility and sharing are possible. Reproducibility must be part of the assessment of a published paper, with appropriate repeats of data, based upon the known or expected variance in it. Adjudicating this variance and repeatability is a key part of editing and reviewing a paper. Sharing of data, computer codes, materials, and protocols can be enforced by a publisher as a requirement to publish or can be stressed by reviewers and editors. Biomedical optical devices are often produced with unique algorithms, with competitive research intensity focused on making better algorithms than others in the field. In the end, these algorithms should either be freely available or commercialized, such that others can use them to validate the results. Without this level of transparency, it can lead to publications that lack merit or believability. Similarly, if there are unique data sets that would be required to reproduce the work, it seems logical that they need to be made available to others, so that the work can be fairly reviewed or reproduced. The field of biomedical optics should seriously consider these goals, as is happening in other fields.17,18

The focus on eventual translation in biomedical optics is a goal for most studies, where this can mean translation either to industry development or to human clinical trial testing.1921 Ultimately if both are successful, a new optical device can change clinical practice by improving outcomes of the patients. There are many examples of such optical devices throughout surgery, pathology, cardiology, and all aspects of medicine. By ensuring that quality, reproducibility, and sharing are possible, journals can improve the impact of the scientist’s discovery as well as the journal itself.

3.

Translational Science Networks

In the field of biomedical optics, there is extreme geographic and scientific diversity, together with aspects of the work in the disparate areas of academia including medicine and industry,22 and so it can be a challenge to promote collaboration and cooperation in this setting where successes are measured by individual events such as publishing, grant funding, and commercial triumphs. However, the features that benefit individual researchers can be the same ones that benefit the field as a whole, namely strong mentor networks of collaborative researchers, who have a healthy balance of competition and collaboration. Sharing data and algorithms is not necessarily in conflict with individual innovation and successes; rather, focusing on one without the other at all can lead to suspicion of publications and results. In translational science, the local environment is critical—where solid biomedical optics inventions require a three-way blend in academics-medicine-industry—incorporating not just communication but active participation. One of the most tangible features of more integrated collaboration is the ability to efficiently cut off projects that may not be beneficial or have a viable translation conduit. The biomedical optics field today has matured to the point where there is extremely high creativity and a wide diversity of topics; however, the number of discoveries that penetrate out to viable medical devices and procedures still can be a small fraction of this whole activity. Indeed, it is apparent that most clinical optical devices use 40-year-old technologies of lamps, lenses, and glass fibers, because they are reliable and inexpensive. As money and time is spent on translational devices, hard decisions need to be made about what works, incorporating all aspects of technical performance, medical need, and commercial viability.22 If these were all truly folded into every research program, we could likely realize improved money spent on those programs that match the needed goals and possibilities.19,20 To achieve this, it would be good for developing researchers to have stronger mentorship and advice on where focused efforts were needed. This does not need to displace technical innovation, but rather should complement it in an ongoing dialogue, which is often done in translational centers or in networks program grants.23 One of the indicators that this is happening to some extent is the documented growth in dual-PI grants at the U.S. National Institutes of Health, as well as industry-academic partnership program funding.24

4.

Initiatives at the Journal of Biomedical Optics

Some argue the problem is that the peer-review system is broken, leading to publications without true peer review and feedback. While this can be true in some fields and for some journals, the strength of the SPIE BiOS community is the enormous size and growth of the field, leading to a strong desire to be part of it. This journal has a large and recently expanded editorial board, which represents the diversity of our field, in gender, geographic submissions, and research topics.25 The long-term impact of papers published here are still the highest of all biomedical optics journals today, and the peer-review system metrics appear strong. Papers published routinely get lengthy feedback, and still most papers that have been accepted and pass quality-control checks to enter the external peer-review system eventually gain publication after productive revisions. This process of a community-focused biomedical optics journal ensures that the individuals in the field assist in the process of making our science stronger.

A recent addition to the journal is the Perspective review series of papers, which provides a conduit for guidance from expert medical professionals to describe their vision for what technologies are needed, to hopefully foster more of a two-way dialogue.26 Those who wish to participate should volunteer, participate in the forum, and seek ways to ensure that we provide strong friendly competition while not losing sight of the need for collaboration and mentoring. Additionally, the BiOS conference itself has become the world’s largest event for biomedical optics researchers to meet and present to each other. While this is just one annual opportunity for exchange, it is a critical piece of our overall community blending together engineering academics, medical professionals, and industry. I urge all involved in the field to work positively toward discoveries and innovations that are scientifically verifiable, translatable, and ultimately impactful to industry and medicine. This journal is one important part of the overall efforts of SPIE and the biomedical optics community in this direction.

References

1. 

M. Baker, “1, 500 scientists lift the lid on reproducibility: Survey sheds light on the ‘crisis’ rocking research,” Nature, 533 452 –454 (2016). https://doi.org/10.1038/533452a Google Scholar

2. 

D. J. Drucker, “Never waste a good crisis: confronting reproducibility in translational research,” Cell Metab., 24 (3), 348 –360 (2016). https://doi.org/10.1016/j.cmet.2016.08.006 1550-4131 Google Scholar

4. 

T. Hanscheid, D. W. Hardisty and S. O. Henriques, “The crisis in scientific publishing: a holistic perspective about background issues associated with predatory publishing,” Acta Med. Port., 31 (10), 524 –526 (2018). https://doi.org/10.20344/amp.10762 AMPOD2 Google Scholar

5. 

D. Fanelli, “Opinion: is science really facing a reproducibility crisis, and do we need it to?,” Proc. Natl. Acad. Sci. USA, 115 (11), 2628 –2631 (2018). https://doi.org/10.1073/pnas.1708272114 Google Scholar

6. 

M. Baker and E. Dolgin, “Cancer reproducibility project releases first results,” Nature, 541 (7637), 269 –270 (2017). https://doi.org/10.1038/541269a Google Scholar

7. 

W. Quan, B. Chen and F. Shu, “Publish or impoverish: An investigation of the monetary reward system of science in China (1999-2016),” Aslib J. Inform. Manag., 69 486 –502 (2017). https://doi.org/10.1108/AJIM-01-2017-0014 Google Scholar

8. 

D. B. Allison, R. M. Shiffrin and V. Stodden, “Reproducibility of research: Issues and proposed remedies,” Proc. Natl. Acad. Sci. USA, 115 (11), 2561 –2562 (2018). https://doi.org/10.1073/pnas.1802324115 Google Scholar

9. 

C. Fernandez-Patron and E. Hardy, “A new science publishing system for a budding science publishing crisis,” Sci. Eng. Ethics, 24 (2), 805 –808 (2018). Google Scholar

10. 

K. D. Cobey et al., “What is a predatory journal? A scoping review,” F1000Res, 7 1001 (2018). https://doi.org/10.12688/f1000research Google Scholar

11. 

S. N. Goodman, “A quality-control test for predatory journals,” Nature, 553 (7687), 155 –155 (2018). https://doi.org/10.1038/d41586-018-00403-z Google Scholar

12. 

W. Strielkowski, “Predatory publishing: what are the alternatives to Beall’s List?,” Am. J. Med., 131 (4), 333 –334 (2018). https://doi.org/10.1016/j.amjmed.2017.10.054 Google Scholar

15. 

J. Bayry, “Indian researchers must resist predatory open-access journals,” Nature, 563 (7729), 35 –35 (2018). https://doi.org/10.1038/d41586-018-07243-x Google Scholar

16. 

J. P. Ioannidis and M. J. Khoury, “Assessing value in biomedical research: the PQRST of appraisal and reward,” JAMA, 312 (5), 483 –484 (2014). https://doi.org/10.1001/jama.2014.6932 JAMAAP 0098-7484 Google Scholar

17. 

S. J. Eglen et al., “Toward standard practices for sharing computer code and programs in neuroscience,” Nat. Neurosci., 20 (6), 770 –773 (2017). https://doi.org/10.1038/nn.4550 NANEFN 1097-6256 Google Scholar

18. 

Regulatory and Ethics Working Group, Global Alliance for Genomics & Health and S. Sugano, “International code of conduct for genomic and health-related data sharing,” Hugo J., 8 (1), 1 (2014). https://doi.org/10.1186/1877-6566-8-1 Google Scholar

19. 

A. D. Aguirre et al., “Special section guest editorial: translational biophotonics,” J. Biomed. Opt., 22 (12), 121601 (2017). https://doi.org/10.1117/1.JBO.22.12.121601 JBOPFO 1083-3668 Google Scholar

20. 

G. Apiou-Sbirlea et al., “Special section guest editorial: translational biophotonics,” J. Biomed. Opt., 21 (12), 124002 (2016). https://doi.org/10.1117/1.JBO.21.12.124002 JBOPFO 1083-3668 Google Scholar

21. 

T. Tkaczyk, M. Pawlowski and M. Pierce, “Special section guest editorial: translational biophotonics,” J. Biomed. Opt., 22 (3), 034000 (2017). https://doi.org/10.1117/1.JBO.22.3.034000 JBOPFO 1083-3668 Google Scholar

22. 

B. C. Wilson, M. Jermyn and F. Leblond, “Challenges and opportunities in clinical translation of biomedical optical spectroscopy and imaging,” J. Biomed. Opt., 23 (3), 030901 (2018). https://doi.org/10.1117/1.JBO.23.3.030901 JBOPFO 1083-3668 Google Scholar

23. 

B. J. Tromberg et al., “Biomedical optics centers: forty years of multidisciplinary clinical translation for improving human health,” J. Biomed. Opt., 21 (12), 124001 (2016). https://doi.org/10.1117/1.JBO.21.12.124001 JBOPFO 1083-3668 Google Scholar

24. 

B. W. Pogue, “Grant funding needs parallel the start-up venture: an analogy for translational research success,” J. Biomed. Opt., 23 (8), 080101 (2018). https://doi.org/10.1117/1.JBO.23.8.080101 JBOPFO 1083-3668 Google Scholar

25. 

B. W. Pogue, “Biomedical optics scientific community,” J. Biomed. Opt., 23 (1), 010101 (2018). https://doi.org/10.1117/1.JBO.23.1.010101 JBOPFO 1083-3668 Google Scholar

26. 

B. W. Pogue, “Medical perspective articles to stimulate the field for needs-finding,” J. Biomed. Opt., 23 (6), 060101 (2018). https://doi.org/10.1117/1.JBO.23.6.060101 JBOPFO 1083-3668 Google Scholar
© 2019 Society of Photo-Optical Instrumentation Engineers (SPIE)
Brian W. Pogue "Ensuring Scientific Publishing Credibility in Translational Biomedical Optics," Journal of Biomedical Optics 24(1), 010101 (10 January 2019). https://doi.org/10.1117/1.JBO.24.1.010101
Published: 10 January 2019
Lens.org Logo
CITATIONS
Cited by 3 scholarly publications.
Advertisement
Advertisement
KEYWORDS
Biomedical optics

Medicine

Optical components

Medical research

Transparency

Clinical trials

Quality measurement

RELATED CONTENT

Mapping DICOM to OpenDocument format
Proceedings of SPIE (March 13 2009)
Health care technology assessment
Proceedings of SPIE (December 06 1994)

Back to Top