Submission guidelines

Instructions for Authors


Behavior Research Methods publishes articles concerned with the methods, techniques, and instrumentation of research in experimental psychology. The journal focuses particularly on the use of computer technology in psychological research. An annual special issue is devoted to this field.


Manuscripts are to be submitted electronically via ScholarOne:

If you have not submitted via the ScholarOne submission system before, then you will first be asked to create an account. Otherwise you can use your existing account.


To submit a manuscript, the corresponding author must affirm that:

(a) the work conforms to Standard 8 of the American Psychological Association’s Ethical Principles of Psychologist and Code of Conduct [click on “Standard 8” on ], which speaks to the ethics of conducting and publishing research and sharing data for the purpose of verification;

(b) if the manuscript includes any copyrighted material the author understands that if the manuscript is accepted for publication s/he will be responsible for obtaining written permission to use that material;

(c) if any of the authors has a potential conflict of interest pertaining to the manuscript that conflict has been disclosed in a message to the Editor;

(d) the author(s) understand(s) that before a manuscript can be published in Behavior Research Methods the copyright to that manuscript must be transferred to the Psychonomic Society (see for details);

(e) the manuscript includes appropriate measures of variability, effect size, and (when relevant) statistical power.

Open Practices

Since its inception, the core mission of the Psychonomic Society has been to foster the science of cognition through the advancement and communication of basic research in experimental psychology and allied sciences. To promote replicable research practices, the policy of the Psychonomic Society is to publish papers in which authors follow standards for disclosing all important aspects of the research design and data analysis. The Society does not enforce any single reporting standard, but authors are encouraged to review and adopt guidelines described, for example, by the American Psychological Association (APA).

In 2017, the Society signed on to the Open Science Initiative’s Level 1 Transparency and Openness Guidelines. All authors are required to respond to the questions below, and in addition, all submitted manuscripts must include an Open Practices Statement immediately prior to the References section of the paper. The statement must specify (1) whether data and/or materials are available, and if so, where (as per Level 1 TOP guidelines, URLs are required to have a persistent identifier); and (2) whether any experiments were preregistered, and if so, which.

The following are examples of appropriate Open Practices Statements:

The data and materials for all experiments are available at (url for the site hosting the data and materials) and Experiment 1 was preregistered (url for the preregistration).

None of the data or materials for the experiments reported here is available, and none of the experiments was preregistered.

Acknowledgments and Funding Information

Acknowledgments of people, grants, funds, etc. should be placed in a separate section before the reference list. The names of funding organizations should be written in full. In addition, please provide the funding information in a separate step of the submission process in the peer review system. Funder names should preferably be selected from the standardized list you will see during submission. If the funding institution you need is not listed, it can be entered as free text. Funding information will be published as searchable metadata for the accepted article, whereas acknowledgements are published within the paper.


The Psychonomic Society’s Publications Committee and Ethics Committee and the Editors-in-Chief of the Society’s seven journals worked together (with input from others) to create these guidelines on statistical issues. These guidelines focus on the analysis and reporting of quantitative data. Many of the issues described below pertain to vulnerabilities in null hypothesis significance testing (NHST), in which the central question is whether or not experimental measures differ from what would be expected due to chance. Below we emphasize some steps that researchers using NHST can take to avoid exacerbating those vulnerabilities. Many of the guidelines are long-standing norms about how to conduct experimental research in psychology. Nevertheless, researchers may benefit from being reminded of some of the ways that poor experimental procedure and analysis can compromise research conclusions. Authors are asked to consider the following issues for each manuscript submitted for publication in a Psychonomic Society journal. Some of these issues are specific to NHST, but many of them apply to other approaches as well.

1. It is important to address the issue of statistical power. Statistical power refers to the sensitivity of a test to detect hypothetical true effects. Studies with low statistical power often produce ambiguous results. Thus it is highly desirable to have ample statistical power for effects that would be of interest to others and to report a priori power at several effect sizes (not post hoc power) for tests of your main hypotheses. Best practice is to determine what effects would be interesting (e.g., those that one would consider non-negligible, useful, or theoretically meaningful) and then to test a sufficient number of participants to attain adequate power to detect an effect of that size. There is no hard-and-fast rule specifying “adequate” power, and Editors may judge that other considerations (e.g., novelty, difficulty) justify a low-powered design. If there is no smallest effect size of interest for an a priori power analysis, then authors can report the effect size at which the design and test have 50% power. Alternatively an analysis might focus on estimation of an effect size rather than on a hypothesis test. In any case, the Method section should make clear what criteria were used to determine the sample size. The main points here are to (a) do what you reasonably can to design an experiment that allows a sensitive test and (b) explain how the number of participants was determined.

2. Multiple NHST tests inflate null-hypothesis rejection rates. Tests of statistical significance (e.g., t-tests, analyses of variance) should not be used repeatedly on different subsets of the same data set (e.g., on varying numbers of participants in a study) without statistical correction, because the Type I error rate increases across multiple tests.

A. One concern is the practice of testing a small sample of participants and then analyzing the data and deciding what to do next depending on whether the predicted effect (a) is statistically significant (stop and publish!), (b) clearly is not being obtained (stop, tweak, and start a new experiment), or (c) looks like it might become significant if more participants are added to the sample (test more participants, then reanalyze; repeat as needed). If this “optional stopping rule” has been followed without appropriate corrections, then report that fact and acknowledge that the Type I error rate is inflated by the multiple tests. Depending on the views of the Editor and reviewers, having used this stopping rule may not preclude publication, but unless appropriate corrections to the Type I error rate are made it will lessen confidence in the reported results. Note that Bayesian data analysis methods are less sensitive to problems related to optional stopping than NHST methods; see Rouder (2014) for a discussion and pointers to other literature.

B. It is problematic to analyze data and then drop some participants or some observations, re-run the analyses, and then report only the last set of analyses. If participants or observations were eliminated, then explicitly indicate why, when, and how this was done and either (a) report or synopsize the results of analyses that include all of the observations or (b) explain why such analyses would not be appropriate.

C. Covariate analyses should either be planned in advance (and in that case would benefit from preregistration) or be described as exploratory. It is inappropriate to analyze data without a covariate, then re-analyze those same data with a covariate and report only the latter analysis as confirmation of an idea. It may be appropriate to conduct multiple analyses in exploratory research, but it is important to report those analyses as exploratory and to acknowledge possible inflations of the Type I error rate.

D. If multiple dependent variables (DVs) are individually analyzed with NHST, the probability that at least one of them will be “significant” by chance alone grows with the number of DVs. Therefore it is important to inform readers of all of the DVs collected that are relevant to the study. For example, if accuracy, latency, and confidence were measured, but the paper focuses on the accuracy data, then report the existence of the other measures and (if possible) adjust the analyses as appropriate. Similarly, if several different measures were used to tap a construct, then it is important to report the existence of all of those indices, not just the ones that yielded significant effects (although it may be reasonable to present a rationale for why discounting or not reporting detailed results for some of the measures is justified). There is no need to report measures that were available to you (e.g., via a participant pool data base) but that are irrelevant to the study. The selection of variables would again benefit from preregistration; see the Psychonomic Society’s recent digital event on this issue:

3. Rich descriptions of the data help reviewers, the Editor, and other readers understand your findings. Thus it is important to report appropriate measures of variability around means and around effects (e.g., confidence intervals or Baysesian credible intervals), and ideally plot the raw data or descriptions of the data (e.g., violin plots, box plots, scatterplots).

4. Cherry picking experiments, conditions, DVs, or observations can be misleading. Give readers the information they need to gain an accurate impression of the reliability and size of the effect in question.

A. Conducting multiple experiments with the same basic procedure and then reporting only the subset of those studies that yielded significant results (and putting the other experiments in an unpublished “file drawer”) can give a misleading impression of the size and replicability of an effect. If several experiments testing the same hypothesis with the same or very similar methods have been conducted and have varied in the pattern of significant and null effects obtained (as would be expected, if only due to chance), then you should report both the significant and the non-significant findings. Reporting the non-significant findings can actually strengthen evidence for the existence of an effect when meta-analytical techniques pool effect sizes across experiments. It is not generally necessary to report results from exploratory pilot experiments (although their existence might be mentioned), such as when pilot experiments were used to estimate effect size, provided the final experiment is sufficiently informative. In contrast, it is not appropriate to run multiple low-powered pilot experiments on a given topic and then report only the experiments that reject the null hypothesis.

B. Deciding whether or not to report data from experimental conditions post hoc, contingent on the outcome of NHST, inflates the Type I error rate. Therefore, please inform readers of all of the conditions tested in the study. If, for example, 2nd, 4th, and 6th graders were tested in a study of memory development then it is appropriate to report on all three of those groups, even if one of them yielded discrepant data. This holds even if there are reasons to believe that some data should be discounted (e.g., due to a confound, a ceiling or floor effect in one condition, etc.). Here again, anomalous results do not necessarily preclude publication (after all, even ideal procedures yield anomalous results sometimes by chance). Failing to report the existence of a condition that did not yield the expected data can be misleading.

C. Deciding to drop participants or observations post hoc contingent on the outcome of NHST inflates the Type I error rate. Best practice is to set inclusion/exclusion criteria in advance (and pre-register where appropriate) and stick to them, but if that is not done then whatever procedure was followed should be reported.

5. Be careful about using null results to infer “boundary conditions” for an effect. A single experiment that does not reject the null hypothesis may provide only weak evidence for the absence of an effect. Too much faith in the outcome of a single experiment can lead to hypothesizing after the results are known (HARKing), which can lead to theoretical ideas being defined by noise in experimental results. Unless the experimental evidence for a boundary condition is strong, it may be more appropriate to consider a non-significant experimental finding as a Type II error. Such errors occur at a rate that reflects experimental power (e.g., if power is .80, then 20% of exact replications would be expected to fail to reject the null). Bayesian tests may provide stronger evidence for the null hypothesis than is possible through frequentist statistics.

6. Authors should use statistical methods that best describe and convey the properties of their data. The Psychonomic Society does not require authors to use any particular data analysis method. The following sections highlight some important considerations.

A. Statistically significant findings are not a prerequisite for publication in Psychonomic Society journals. Indeed, too many significant findings relative to experimental power can indicate bias. Sometimes strong evidence for null or negligible effects can be deeply informative for theorizing and for identifying boundary conditions of an effect.

B. In many scientific investigations the goal of an experiment is to measure the magnitude of an effect with some degree of precision. In such a situation stronger scientific arguments may be available through confidence or credible intervals (of parameter values or of standardized effect sizes). Moreover, some of the bias issues described above can be avoided by designing experiments to measure effects to a desired degree of precision (range of confidence interval). Confidence intervals must be interpreted with care, see the Psychonomic Society’s digital event on this issue:

C. The Psychonomic Society encourages the use of data analysis methods other than NHST when appropriate. For example, Bayesian data analysis methods avoid some of the problems described above. They can be used instead of traditional NHST methods for both hypothesis testing and estimation. The Psychonomic Society’s recent digital event on Bayesian methods may be helpful here:

Last Word

Ultimately, journal Editors work with reviewers and authors to promote good scientific practice in publications in Psychonomic Society journals. A publication decision on any specific manuscript depends on much more than the above guidelines, and individual Editors and reviewers may stress some points more than others. Nonetheless, all else being equal submissions that comply with these guidelines will be better science and be more likely to be published than submissions that deviate from them.


There are many excellent sources for information on statistical issues. Listed below are some that the 2019 Publications Committee and Editors recommend.

• Confidence Intervals

o Masson, M. J., & Loftus, G. R. (2003). Using confidence intervals for graphically based data interpretation. Canadian Journal of Experimental Psychology/Revue Canadienne de Psychologie Expérimentale, 57, 203-220. doi:10.1037/h0087426

o Morey, R. D., Hoekstra, R., Rouder, J. N., Lee, M. D., & Wagenmakers, E. J. (2016). The fallacy of placing confidence in confidence intervals. Psychonomic bulletin & review, 23, 103-123.

• Effect Size Estimates

o Ellis, P. D. (2010). The essential guide to effect sizes: Statistical power, meta-analysis and the interpretation of research results. Cambridge University Press. ISBN 978-0-521-14246-5.

o Fritz, C. O., Morris, P. E., & Richler, J. J. (2011). Effect size estimates: Current use, calculations and interpretation. Journal of Experimental Psychology: General, 141, 2-18.

o Grissom, R. J., & Kim, J. J. (2012). Effect sizes for research: Univariate and multivariate applications (2nd ed.). New York, NY US: Routledge/Taylor & Francis Group

• Meta-analysis

o Cumming, G. (2012). Understanding the new statistics: Effect sizes, confidence intervals, and meta-analysis. New York, NY US: Routledge/Taylor & Francis Group.

o Littell, J. H., Corcoran, J., & Pillai, V. (2008). Systematic reviews and meta-analysis. New York: Oxford University Press.

• Bayesian Data Analysis

o Kruschke, J. K. (2011). Doing Bayesian data analysis: A tutorial with R and BUGS. San Diego, CA US: Elsevier Academic Press.

o Kruschke, J. K. (2013). Bayesian estimation supersedes the t test. Journal of Experimental Psychology: General, 144(2), 573-603.

o McElreath, R. (2015). Statistical rethinking: A Bayesian course with examples in R and Stan. Chapman and Hall/CRC.

o Rouder, J. N. (2014). Optional stopping: No problem for Bayesians. Psychonomic bulletin & review, 21, 301-308.

• Power Analysis

o Faul, F., Erdfelder, E., Lang, A., & Buchner, A. (2007). G*Power 3: A flexible statistical power analysis program for the social, behavioral, and biomedical sciences. Behavior Research Methods, 39(2), 175-191. (See


Manuscripts are to adhere to the conventions described in the Publication Manual of the American Psychological Association (6th ed.). See for information on APA style, or type “APA style” into a search engine to find numerous online sources of information about APA style. Here we highlight only the most fundamental aspects of that style.

  • Layout: All manuscripts are to be double spaced and have 1” margins with page numbers in the upper right corner of each page.

    Title Page: The title page must include the authors’ names and affiliations and the corresponding author’s address, telephone number, and e-mail address.

    Abstract: There must be an abstract of no more than 250 words.

    Sections: Manuscript should be divided into sections (and perhaps subsections) appropriate for their content (e.g., introduction/background, Method, Results, etc.), as per APA style.

    Acknowledgments: The Author Note should include sources of financial support and any possible conflicts of interest. If desirable, contributions of different authors may be briefly described here. Reviewers and the Editor should not be thanked in the Author Note.

    Figures and Tables: Figures and tables are to be designed as per APA style.

    Location of Figures, Tables, and Footnotes: In submitted manuscripts, figures and tables can be embedded in the body of the text and footnotes can be placed at the bottom of the page on which the footnoted material is referenced. Note that this is a departure from APA style; if you prefer you can submit the manuscript with the figures, tables, and footnotes at the end, but it is slightly easier for reviewers if these elements appear near the text that refers to them. When a paper is accepted, in the final version that the author submits for production each figure and table must be on a separate page near the end of the manuscript and all footnotes must be listed on a footnote page, as per the APA Publication Manual.

    Citations and References: These should conform to APA style.


Behavior Research Method is published in electronic format only. Graphs and figures can be published in color in the journal at no cost to the author

Research Data Policy

A submission to the journal implies that materials described in the manuscript, including all relevant raw data, will be freely available to any researcher wishing to use them for non-commercial purposes, without breaching participant confidentiality.

Data availability

All original research must include a data availability statement. Data availability statements should include information on where data supporting the results reported in the article can be found, if applicable. Statements should include, where applicable, hyperlinks to publicly archived datasets analysed or generated during the study. For the purposes of the data availability statement, “data” is defined as the minimal dataset that would be necessary to interpret, replicate and build upon the findings reported in the article. When it is not possible to share research data publicly, for instance when individual privacy could be compromised, data availability should still be stated in the manuscript along with any conditions for access. Data availability statements can take one of the following forms (or a combination of more than one if required for multiple datasets):

1. The datasets generated during and/or analysed during the current study are available in the [NAME] repository, [PERSISTENT WEB LINK TO DATASETS]

2. The datasets generated during and/or analysed during the current study are available from the corresponding author on reasonable request.

3. All data generated or analysed during this study are included in this published article [and its supplementary information files].

4. The datasets generated during and/or analysed during the current study are not publicly available due [REASON(S) WHY DATA ARE NOT PUBLIC] but are available from the corresponding author on reasonable request.].

5. Data sharing not applicable to this article as no datasets were generated or analysed during the current study.

6. The data that support the findings of this study are available from [THIRD PARTY NAME] but restrictions apply to the availability of these data, which were used under licence for the current study, and so are not publicly available. Data are however available from the authors upon reasonable request and with permission of [THIRD PARTY NAME].

More templates for data availability statements, including examples of openly available and restricted access datasets, are available here:

Data availability statements

Data repositories

This journal strongly encourages that all datasets on which the conclusions of the paper rely are available to readers. We encourage authors to ensure that their datasets are either deposited in publicly available repositories (where available and appropriate) or presented in the main manuscript or additional supporting files whenever possible. Please see Springer Nature’s information on recommended repositories.

List of Repositories

General repositories - for all types of research data - such as figshare and Dryad may be used where appropriate.

Data citation

The journal also requires that authors cite any publicly available data on which the conclusions of the paper rely. Data citations should include a persistent identifier (such as a DOI), should be included in the reference list using the minimum information recommended by DataCite, and follow journal style. Dataset identifiers including DOIs should be expressed as full URLs.

Research data and peer review

Peer reviewers are encouraged to check the manuscript’s Data availability statement, where applicable. They should consider if the authors have complied with the journal’s policy on the availability of research data, and whether reasonable effort has been made to make the data that support the findings of the study available for replication or reuse by other researchers. Peer reviewers are entitled to request access to underlying data (and code) when needed for them to perform their evaluation of a manuscript.

Research data helpdesk

Springer Nature provides a research data policy support service for authors and editors, which can be contacted at

This service provides advice on research data policy compliance and on finding research data repositories. It is independent of journal, book and conference proceedings editorial offices and does not advise on specific manuscripts.


For more information:

English Language Editing

For editors and reviewers to accurately assess the work presented in your manuscript you need to ensure the English language is of sufficient quality to be understood. If you need help with writing in English you should consider:

  • Asking a colleague who is a native English speaker to review your manuscript for clarity.
  • Visiting the English language tutorial which covers the common mistakes when writing in English.
  • Using a professional language editing service where editors will improve the English to ensure that your meaning is clear and identify problems that require your review. Two such services are provided by our affiliates Nature Research Editing Service and American Journal Experts. Springer authors are entitled to a 10% discount on their first submission to either of these services, simply follow the links below.

English language tutorial

Nature Research Editing Service

American Journal Experts

Please note that the use of a language editing service is not a requirement for publication in this journal and does not imply or guarantee that the article will be selected for peer review or accepted.

If your manuscript is accepted it will be checked by our copyeditors for spelling and formal style before publication.



● 请一位以英语为母语的同事审核您的稿件是否表意清晰。

● 查看一些有关英语写作中常见语言错误的教程。

● 使用专业语言编辑服务,编辑人员会对英语进行润色,以确保您的意思表达清晰,并识别需要您复核的问题。我们的附属机构 Nature Research Editing Service 和合作伙伴 American Journal Experts 即可提供此类服务。


Nature Research Editing Service

American Journal Experts







・プロの英文校正サービスを利用する。校正者が原稿の意味を明確にしたり、問題点を指摘し、英語の質を向上させます。Nature Research Editing Service とAmerican Journal Experts の2つは弊社と提携しているサービスです。Springer の著者は、いずれのサービスも初めて利用する際には10%の割引を受けることができます。以下のリンクを参照ください。


Nature Research Editing Service

American Journal Experts




영어 원고의 경우, 에디터 및 리뷰어들이 귀하의 원고에 실린 결과물을 정확하게 평가할 수 있도록, 그들이 충분히 이해할 수 있을 만한 수준으로 작성되어야 합니다. 만약 영작문과 관련하여 도움을 받기를 원하신다면 다음의 사항들을 고려하여 주십시오:

• 귀하의 원고의 표현을 명확히 해줄 영어 원어민 동료를 찾아서 리뷰를 의뢰합니다.

• 영어 튜토리얼 페이지에 방문하여 영어로 글을 쓸 때 자주하는 실수들을 확인합니다.

• 리뷰에 대비하여, 원고의 의미를 명확하게 해주고 리뷰에서 요구하는 문제점들을 식별해서 영문 수준을 향상시켜주는 전문 영문 교정 서비스를 이용합니다. Nature Research Editing Service와 American Journal Experts에서 저희와 협약을 통해 서비스를 제공하고 있습니다. Springer 저자들이 본 교정 서비스를 첫 논문 투고를 위해 사용하시는 경우 10%의 할인이 적용되며, 아래의 링크를 통하여 확인이 가능합니다.

영어 튜토리얼 페이지

Nature Research Editing Service

American Journal Experts

영문 교정 서비스는 게재를 위한 요구사항은 아니며, 해당 서비스의 이용이 피어 리뷰에 논문이 선택되거나 게재가 수락되는 것을 의미하거나 보장하지 않습니다.

원고가 수락될 경우, 출판 전 저희측 편집자에 의해 원고의 철자 및 문체를 검수하는 과정을 거치게 됩니다.


If you have questions not answered above, please direct them to the Editor of the journal in question:

Marc Brysbaert

Ghent University

Author tools to prepare your article

For your convenience, Springer has developed macros and templates to help you prepare your article. For information about the preferred format for this journal please refer to the Instructions for Authors.

Word template

The document template provides predefined style formats for all the types of content that are part of an article. It can be used with Word 2000 and higher.

Word template (Download zip, 158 kB)

LaTeX macro package

The LaTex2e macro package supports various journal layouts. For information about which layout option to use, please refer to the section "Text" on the right.

LaTeX macro package (Download zip, 188 kB)

Open access publishing

Behavior Research Methods publishes open access articles. Authors of open access articles published in this journal retain the copyright of their articles and are free to reproduce and disseminate their work.

Visit our Open access publishing page to learn more.