ArticlePDF Available

Validity practices in qualitative research in school psychology

Authors:

Abstract and Figures

Validity refers to the extent to which a given study reflects the dominant values of a research community regarding what constitutes “good” research. Methodological texts on qualitative research provide a wide range of criteria and strategies to help qualitative researchers validate their studies. Given the importance of these strategies to establish a study as trustworthy and legitimate, the objective of this study was to understand the strategies commonly used by school psychology researchers in qualitative research. We therefore reviewed qualitative research articles published in seven school psychology journals between 2006 and 2021. We found 15 strategies authors used to enhance the validity of their research. We also found that a strategy could be enacted in many different ways by different researchers depending on the context. We conclude by recommending four ways in which qualitative researchers in school psychology can improve their validation practices.
Content may be subject to copyright.
Validity practices in
qualitative research in
school psychology
Sujay V. Sabnis
Miami University, Oxford, OH, USA
Jennifer R. Wolgemuth
University of South Florida, Tampa, FL, USA
Abstract
Validity refers to the extent to which a given study reects the dominant values of a
research community regarding what constitutes goodresearch. Methodological
texts on qualitative research provide a wide range of criteria and strategies to help quali-
tative researchers validate their studies. Given the importance of these strategies to
establish a study as trustworthy and legitimate, the objective of this study was to under-
stand the strategies commonly used by school psychology researchers in qualitative
research. We therefore reviewed qualitative research articles published in seven school
psychology journals between 2006 and 2021. We found 15 strategies authors used to
enhance the validity of their research. We also found that a strategy could be enacted
in many different ways by different researchers depending on the context. We conclude
by recommending four ways in which qualitative researchers in school psychology can
improve their validation practices.
Keywords
research methods, methodology, qualitative research, coding, member checking
School psychology, as a branch of psychology, has traditionally been a quantitative eld
grounded in the postpositivist paradigm. Although methodological diversity character-
ized the early days of psychology, the discipline became more conservative over time,
Corresponding author:
Sujay V. Sabnis, Miami University, Oxford, OH, USA.
Email: sabnissv@miamioh.edu
Article
School Psychology International
128
© The Author(s) 2023
Article reuse guidelines:
sagepub.com/journals-permissions
DOI: 10.1177/01430343231194731
journals.sagepub.com/home/spi
prioritizing narrow hypothetico-deductive (quantitative) research in order to position
itself as a respectablescience similar to material sciences such as physics and chemistry
(Danzinger, 1979; Harre, 2004). Non-quantitative forms of research were thus margina-
lized within the applied branches of psychology, including school psychology.
Nevertheless, qualitative research seems to be growing in prevalence. In school psych-
ology, specically, two systematic studies (Powell et al., 2008; Sabnis et al., 2023) have
found a denite growth in the use of qualitative research in school psychology journals.
Between 2001 and 2005, Powell et al. (2008) found the number of qualitative publica-
tions per year increased from one to six. Sabnis et al. (2023) found the number to
grow from 7 qualitative articles in 2006 to 22 articles in 2021. Although these gures con-
tinue to be minuscule in comparison to the overall volume of articles published in these
journals, it is nevertheless a cause for cautious optimism, given that similar trends are
seen in other elds (e.g., Hays et al., 2016).
As the prevalence of qualitative research in the social sciences increases, numerous
organizations have published reporting standards in order to provide journal editors
and authors more guidance for evaluating the quality of manuscripts. For example, a
task force appointed by the American Psychological Association (APA) published report-
ing standards for qualitative studies (see Levitt et al., 2018). Similar efforts have been
seen in other elds such as health (Tong et al., 2007), business (Gephart, 2004), and edu-
cation (Reid & Gough, 2000). These reporting standards may vary from each other in
some ways, given that each discipline has its own particular topography which shapes
the research methods employed by its researchers. Nevertheless, the standards all seem
to emphasize an explicit discussion on the strategies used by the authors to establish
the validity of their study.
What is validity?
Validity refers to the extent to which a given study reects the dominant values of a
research community about what constitutes goodresearch. Validity has a specic
meaning in quantitative research that was shaped by the positivist belief in the existence
of an objective reality independent of the human mind as well as a desire to capture this
reality accurately while diminishing the noise of confounding variables, conceptual fuzzi-
ness, and probabilistic errors (Eketu, 2017). Social scientists thus use validity to refer to
the truth valueof a studysndings (i.e., in its ability to reveal truethings about the
world; Onwuegbuzie & Leech, 2007). How could this ability be determined? In other
words, what criteria could be used by the scientic community to evaluate a studys
truth value? And what strategies could be used by social scientists to ensure their
study met these criteria?
Quantitative researchers Campbell and Stanley (1963) proposed two main criteria for
determining the truth value of a studynamely by looking at the extent to which the nd-
ings were internally true and externally true. A study nding could be deemed internally
true when it was purely a result of the experimenters intentional manipulation of experi-
mental conditions. A nding was deemed externally true when it was collected under
conditions that resembled real-world conditions and populations of people. These two
2School Psychology International 0(0)
criteria have come to be known as internal and external validity among quantitative
researchers. Although no study could be 100% internally and externally valid,
Campbell and Stanley advocated for researchers to take steps to reduce threats to the
internal and external validity of their studies. Following the popularization of these
two criteria, the quantitative research community came to use many different strategies
to bolster the internal and external validity, such as closely controlling the conditions
of the data collection to minimize confounding variables, statistically controlling for
extraneous variables, random assignment, and so forth.
At one point in time, it was thought possible to import the Campbell and Stanleys
(1963) postpositivist criteria of validity to qualitative research (e.g., Miles &
Huberman, 1984 argue this position). However, their underlying onto-epistemological
assumptions about the existence of objective reality and objective knowledge were
incompatible with those of qualitative researchers who rejected the concept of a single
objective social reality and viewed all truth claimsto be co-constructed (e.g.,
Bochner, 2000; Wolcott, 1990). This incompatibility led qualitative researchers to seek
an alternative to validity in qualitative research (Hammersly, 1992; Leininger, 1994),
starting with substituting the term validityitself, using words such as trustworthiness
(Lincoln & Guba, 1985), legitimation (Onwuegbuzie & Johnson, 2006), or methodo-
logical integrity (Levitt et al., 2018) in its place because they believed these terms
better captured what was being evaluated in a manuscript. Nevertheless, a comprehensive
look at the contemporary qualitative literature suggests these terms are often used inter-
changeably. For this article, we will use the term validityas a signier of quality since
that is the term most familiar to school psychologists and still widely used in the quali-
tative research community.
Validity criteria. Just as quantitative researchers use Campbell and Stanleys (1963) two
criteria for judging the validity of a study, qualitative researchers have sought to
develop their own criteria of validity. Lincoln and Guba (1985), proposed four criteria
for judging a studys trustworthiness, namely credibility, transferability, dependability,
and conrmability. Smith and Glass (1987) proposed four criterianamely logical val-
idity, construct validity, internal validity, and external validity. Whereas these criteria
were informed by the interpretivist paradigm, Lather (1993) proposed criteria informed
by poststructural feminist theorynamely catalytic validity, rhizomatic validity, ironic
validity, and paralogic validity. Some methodologists have sought to establish validity
criteria for specic methodologies such as action research, discourse analysis, ethnog-
raphy, and poststructural research (Cho & Trent, 2006).
Whereas the quantitative research community shares a strong consensus about the two
criteria to evaluate a studys validity (internal validity and external validity), no such con-
sensus exists among qualitative researchers. As qualitative approaches to inquiry expand,
evolve, and hybridize, methodologists such as Tracy (2010) and Onwuegbuzie and
Leech (2007) have attempted to bring together the various criteria proposed by qualitative
researchers under a single big-tent framework. For example, Tracy (2010) synthesized val-
idity literature into eight criteria (worthy topic, rich rigor, sincerity, credibility, resonance,
signicant contribution, ethics, and meaningful coherence). Similarly, Onwuegbuzie and
Sabnis and Wolgemuth 3
Leech (2007) subsumed the various criteria found in literature under two broad categories
which they called internal credibility and external credibility, which are further divided into
sub-criteria.
Validity strategies. Qualitative researchers are expected to choose the criteria by which
they wish their work to be judgedcriteria well aligned with their methodology and
underlying philosophy of inquiryand then take steps to ensure their study meets
those criteria. Validity strategies thus refer to the specicmeans, practices, and
methods through which(Tracy, 2010, p. 840) qualitative researchers live up to their
chosen criteria. Here it is important to mention that no study, however well-designed,
can meet a validity criterion entirely. Validity criteria act as aspirational lodestars that
can be used by a researcher to navigate their inquiry. Once a qualitative researcher deter-
mines the criteria they want their study to be evaluated on, they can then take steps or
enact procedures to meet that criteria. For instance, a researcher who wants their ndings
to be judged on the realist criteria of credibility (i.e., the extent to which their account
seems true,accurate, or that the researcher got it right) may conduct member check-
ing where they take their ndings back to the participants for feedback to ensure their
account corresponds to the participantsreality (Tracy, 2010). Onwuegbuzie and Leech
(2007) surveyed qualitative literature and found 24 strategies that have been commonly
used by researchers to help bolster their studys validity, such as triangulation, audit trails,
member checks, and so forth.
Rationale for this study
By conducting this study, we hoped to help future school psychology researchers and
graduate students gain a better understanding of validity by showing them how past
researchers in school psychology have practiced it. Levitt et al. (2018) identify discussion
about validity (which they refer to as methodological integrity) as an important part of a
manuscript reporting on a qualitative study. Specically, they ask authors to clarify the
means by which they sought to achieve validity for their study. However, there are
various difculties associated with this mandate. First, methodological texts on qualita-
tive research present a dazzling array of validity criteria for judging the worth of a
study. Reporting standards thus refrain from specifying the validity criteria to be used
given the wide range of qualitative approaches, and the importance of choosing or tai-
loring validity criteria for ones project (Levitt et al., 2018, p. 29). Standardizing the cri-
teria for validity is also difcult given the need for qualitative researchers to balance rigor,
discipline, and systematicity on one hand with creativity and subjectivity on the other,
without going too far in one direction at the expense of the other, which Whittemore
et al. (2001, p. 522) described as the opposing dangers of methodological rigidity and
methodological anarchy.
Secondly, school psychology researchers typically do not receive much training in
qualitative research methods (Arora et al., 2022). Powell et al. (2008) found that out of
all the school psychology training programs in the US, only one program required its stu-
dents to learn qualitative research methods. Thus, school psychology researchers
4School Psychology International 0(0)
interested in using qualitative methods for a research project and publishing the study
may be unsure of how to establish the validity of their study to make it publishable.
Therefore, it may be helpful to understand how previous researchers publishing in
school psychology journals have gone about the task of validating their studies. To
this end, we asked the following research question: In what ways do school psychology
researchers establish the validity of their studies?
Process of inquiry
The data for this manuscript partially came from a larger project investigating the preva-
lence of qualitative research in school psychology journals. Sabnis et al. (2023) con-
ducted a bibliometric analysis of seven school psychology journals from 2006 through
2021. The journals were selected for having a primarily school psychology audience
and performance metrics used to evaluate journal quality such as an impact factor or
CiteScore listed on their websites. The seven journals were
Canadian Journal of School Psychology (CJSP)
Journal of Applied School Psychology (JASP)
Journal of School Psychology (JSP)
Psychology in the Schools (PITS)
School Psychology (SP)
School Psychology International (SPI)
School Psychology Review (SPR)
Sabnis et al. (2023) looked at all the articles in these journals between 2006 and 2021 to
identify qualitative articles, dened as any article meeting the following criteria:
(1) The data being analyzed were in the form of text or images, not numbers. (2) The text or
images were analyzed and reported in terms of themes, categories, domains, narratives, or
discourses. Quantication, if present, was limited to descriptive statistics (e.g., frequency
of themes, number of codes, and average age of participants) not inferential statistics. (3)
The data analysis procedure was clearly described in the article.(Sabnis et al., 2023, p. 2)
After practicing two journal issues together, the four researchers split into two teams, and
each team worked on their assigned journals, going through each journal issue published
between 2006 and 2021 to identify the qualitative articles, and extract relevant informa-
tion, such as the research topic and participant characteristics (which are discussed at
length in Sabnis et al. [2023]). The two teams met frequently to provide updates and
clarify criteria. Each team then reviewed each others work to identify false positives.
This search process lasted around ve months, and led them to 142 qualitative articles.
For this study, the rst author collaborated with a different researcher to conduct a sys-
tematic review of these 142 qualitative articles to understand various methodological
aspects such as commonly used study designs, forms of data collection and analysis,
methods of reexivity, and validity. Given space constraints and the timely importance
Sabnis and Wolgemuth 5
of the topic of validity for qualitative researchers, we decided to focus this manuscript
solely on the ndings related to validity.
Validity of the study
We had two goals for this article. We wanted to increase the understanding of validity
among quantitatively-trained school psychology researchers who may be considering a
qualitative study and among school psychology graduate students who may be consider-
ing a qualitative dissertation but without being sure how to proceed. We hoped to bring
them more clarity by highlighting the different ways in which validity has been practiced
in the past by school psychology researchers. We also wanted to highlight areas for
improvement in order to help raise the overall quality of qualitative research produced
in the eld. Given these motives, undertaking a rigorous rather than a haphazard
approach to collecting and analyzing the large quantity of published articles spanning
16 years would reduce our chances of missing important patterns in the data. Since we
wanted our research to have pedagogic value in teaching people about qualitative
research, we would also need to come across as credible narrators to the reader of this
article. For this reason, we selected two of the eight validity criteria suggested by
Tracy (2010) to help orient our study, namely rigor and credibility.
Rigor. We borrowed the rst validity criteria from Tracy (2010) who dened it as due
diligence, exercising appropriate time, effort, care, and thoroughnessin following a sys-
tematic process of inquiry (p. 841). We started by creating a protocol to extract informa-
tion from the articles (see Appendix A). We started by reading ve articles separately,
followed by a meeting to extract the information into the spreadsheet together. This
step helped us to iron out differences and rene the protocol to reect an improved under-
standing of what we were looking for. Specically, we decided to use a broad denition
of validity strategies: steps or measures researchers take to increase the intellectual
value, usefulness, and insightfulness of their study. These steps could be taken at any
stage from study conception, data collection, data analysis through how the study is pre-
sented to readers. Importantly the author had to explicitly disclose taking these steps.We
decided to make this denition broad to account for the multiple forms of validity strat-
egies we came across during the literature review phase. Although we approached the
coding process without any a priori codes, this stage was doubtlessly inuenced by
our prior knowledge of common validity strategies described in qualitative literature
such as interrater agreement, member checking, prolonged engagement, triangulation,
and so forth.
After the protocol was nalized, we met over the next few weeks to code articles, rst
separately and then together. We compared our respective coding, and used discussion to
arrive at a consensus which we noted in the nal spreadsheet. By the 25th article, we were
coding similarly with very few disagreements. At this point, we decided to split the
remaining articles into two, with each author coding the articles assigned to them.
During this phase, we met every other week for seven months to check on each
others progress, review each others work, and ask questions to clarify the other
6School Psychology International 0(0)
persons coding. This process generated information about the validity procedures used
in each of the articles. We reviewed this information to identify 15 different validity pro-
cedures. The rst author revisited all the 142 articles as well as the spreadsheets to verify
the calculations.
Credibility. Credibility refers to the extent to which readers can trust a researcher to study
a topic competently and arrive at reasonable results and conclusions based on the data.
One of the factors that affect the credibility of a study is the competence of the authors
with regard to the topic under study. We believe we are competent authors based on our
extensive training in qualitative research. The rst author completed a doctoral degree
in school psychology and a graduate certicate in qualitative research. He has con-
ducted and published various qualitative research studies and is also asked to
provide ad hoc reviews for qualitative manuscripts submitted to various school psych-
ology journals. The second author is a qualitative methodologist who has published
over 70 articles related to qualitative research in journals such as Qualitative Inquiry
and Qualitative Health Research. Overall, we believe this training provides us with
the expertise needed to conduct this systematic review and arrive at credible
conclusions.
Credibility of the ndings is also bolstered by providing thick descriptions of a phe-
nomenon (Tracy, 2010). To construct a thick description, a researcher must dive deeply
into a phenomenon to uncover complex patterns which are then communicated to the
reader through examples and anecdotes. In this report, we showcase the different ways
in which school psychology researchers went about establishing the validity of their
studies. We discuss each strategy not only in terms of its overall meaning but also in
terms of its heterogeneous execution. In other words, we show how two teams of
researchers may use the same strategy but in widely different ways to attain the goals
of their manuscript. We also provide many examples to illustrate the wide range of
topics that have been studied with these strategies.
In the following section, we discuss the different ways in which validity was estab-
lished by qualitative researchers in school psychology. Note that when referring to one
of the 142 articles, we italicize the citation to distinguish the article from general
citations.
Key ndings: validity strategies
Researchers used various strategies to establish validity of their studies. These strat-
egies were used in different phases of the inquiry. Figure 1 represents the different
validity strategies in order of their popularity in qualitative research in school psych-
ology. Popularity was judged by how many of the 142 articles featured a given val-
idity strategy. A number in the gure (e.g., 30) indicates that the validity strategy
appeared in 30 out of 142 articles published in school psychology journals between
2006 and 2021.
Sabnis and Wolgemuth 7
Intercoder agreement
The most commonly used strategy to establish the validity of a qualitative inquiry was
intercoder agreement, which was used in 44 out of the 142 articles (31%). This strategy
also went by other names such as interrater reliability. In this strategy, researchers
coded the transcripts separately, then came together to compare their codes. During
these meetings they monitored the rate of agreement in terms of percentage or the
kappa. Frequently they went through multiple iterations of this process to get the
rate of agreement above a preset threshold such as 75% (Harvey & Pearrow, 2010),
80% (Kelada et al., 2017; McMahon et al., 2020; Moy et al., 2014; Washington
et al., 2020) or 90% (Proctor et al., 2018). The value of preset thresholds was justied
by citing methodologists such as Miles and Huberman (1994) for 80% and Bakeman
and Gottman (1986) for 90%.
Ideally the process of intercoder agreement involves two or more coders coding the
entire dataset independently and then coming together to compare their codes with
each other. However, this process can be extremely laborious and require extensive
time commitment. For Newman and Ingrahams (2020) study of a classroom activity
for teaching multicultural school consultation, the process took approximately two
years, with over 40 collaborative meetings typically two to three hours long(p. 16).
Many research teams may not have secondary coders available to make such time
commitments.
Figure 1. Most common validity strategies in published qualitative literature in seven school
psychology journals (2006-2021).
8School Psychology International 0(0)
Given the clear logistical challenges presented by this approach, qualitative research-
ers in our review found many creative ways to balance the need for intercoder rigor with
limited time. One of the most common strategies we found was for the primary researcher
to use a secondary coder to co-code part of the data set, and then code the remaining
dataset alone. For example, Justin et al. (2021) and Frels et al. (2013) conducted interrater
agreement on 30% and 18% of the data respectively, and rened the coding scheme until
they reached a satisfactory level of agreement with the secondary coder. The primary
researcher in both the studies coded the rest of data alone using the rened coding
scheme.
In Grapin et al.s (2021) open-ended survey of graduate students in school psychology
programs, researchers randomly selected 20% of the survey items and used a predeve-
loped coding scheme to calculate the interrater reliability using Cohens kappa which
exceeded 0.8. Ni et al. (2013) had a graduate student code 20% of randomly selected
data and reported an interrater agreement of 89%. McIntosh et al. (2013) assigned a
number to each participant and used a random number generator to select 25% of the
data which was then sent to the third author to code. These strategies helped the
authors to uphold the rigor of their study while also respecting the time constraints of
fellow researchers.
One study also calculated intrarater reliability which refers to the process of ensuring
that a researcher applied the same rules consistently over time, in order to minimize
researcher drift. In Rush and Wheelers (2011) study of early career scholars attending
the School Psychology Research Collaboration Conference (SPRCC), both authors indi-
vidually coded an excerpt a second time and compared it to their rst individual coding of
the same excerpt, leading to 99.8% and 100% agreement, respectively, indicating strong
intrarater consistency of applying the codes and a low level of coder drift.
Member checking
At its core, member checking refers to researchers checkingwith their participants and
providing them an opportunity to give feedback about the research process. These oppor-
tunities can come in many different formsfrom modifying the data itself to modifying
the researchers interpretation of the data. Member checking helps to make the research
process more democratic and also helps researchers avoid unintentionally misrepresent-
ing or misunderstanding participantsviewpoints. Thirty percent (n=43) of the studies
involved member checking. Researchers took many different approaches to member
checking, and differed specically in what was sent to the participants.
Interview notes. For some like Davies (2020) exploration of the challenges faced by fam-
ilies of students with traumatic brain injuries (TBIs), member checking involved summar-
izing the main pointsof the interview to the interviewees immediately upon nishing
the interview, and asking them if the summary was accurate(p. 282).
Transcripts. For others like Cassidy et al. (2012) and Craggs and Kelly (2018), member
checking involved transcribing the interviews and then sending the raw transcripts to
Sabnis and Wolgemuth 9
the participants and allowing them to make any changes or add any information they had
missed.
Summary of results. Leath et al. (2021), who explored how Black girls navigate friend-
ships in high schools, sent a summary of the results of the data analysis to participants
to get their feedback and made changes to the results based on that feedback in order
to ensure that [their] ndings reected the young Black womens perspectives
(p. 39). Similarly, Newman and Ingraham (2020) sent their ndings (consisting of a
summary of themes and a conceptual model created by the authors based on the study
results) to participants and requested their feedback in specic areas such as (a) their per-
ceived accuracy of the theme; (b) what, if anything would enhance the accuracy; and (c)
what additional points they had for consideration regarding each theme(p. 17).
Draft of the manuscript. Finally, we found one study investigating a distance education
program for preparing rural school psychologists in Colorado, wherein Lahman et al.
(2006) sent a draft of their manuscript to their participants to respond with comments
regarding inaccuracies, clarication and reections(p. 442) prior to journal submission.
Non-quantied consensual coding
The third most common strategy we encountered in the qualitative articles was what we
called non-quantied consensual coding, appearing in 27% of the articles (n=38). This
procedure was similar to intercoder agreement except that it did not involve the research-
ers quantifying their agreement. Thus, the researchers coded the transcripts separately and
then came together to compare their coding without specically monitoring the rate of
agreement or disagreement. The researchers discussed areas of disagreement and came
to a consensus. The reason for choosing this strategy over intercoder agreement was
cogently explained by Suldo et al. (2008): As the research teams goal was to reach
full consensus in the coding scheme, percentage of initial agreement between raters
was not calculatedconsistent with the qualitative work initiated by Harry, Sturges
and Klingner (2005)(p. 965).
External auditing or peer feedback
This strategy was found in 26% of the articles (n=35) and involved researchers consult-
ing with people outside the research team to aid in the research process. A closer look into
the data revealed variations in terms of when the researchers sought external help.
Before the study. In some cases, external auditors were brought in at the beginning of the
study and helped to audit the entire research process. For example, Sansosti and Sansosti
(2012), who explored educatorsexperiences working with students with high-
functioning autism spectrum disorders(p. 917), invited a panel of colleagues with
expertise in qualitative research and special education to review all the study procedures
before the study began. Larsen and Samdal (2012) invited two teachers to be external
10 School Psychology International 0(0)
reviewers. Before beginning data collection, they pilot-tested the interview protocol with
the teachers and used their feedback to improve the protocol.
During data analysis. In other cases, external reviewers were brought in during data ana-
lysis to help the researchers code the transcripts. For instance, Canpolat and Atli (2021),
who explored how children experience changing schools due to parentsrelocation, inter-
viewed 25 such students, tasked external reviewers with coding their transcripts, and then
used this to calculate intercoder agreement. Phasha (2008) invited an independent analyst
to code 15 of the transcripts and then reviewed their coding in order to modify her own
coding.
After data analysis. More often than not, the external auditor was brought in after the data
analysis was completed. In Moy et al. (2014), for example, an external auditor reviewed
the codebook created by the researchers based on data analysis as well as the transcripts to
evaluate whether or not statements were accurately codedand indicating his agree-
ment or disagreement with the coding of each statement(p. 329). The researchers
also supplied him with all the documents from the research including the coding, and
the auditor went through this and pointed out discrepancies. Similarly, researchers
such as Pufpaff (2008), who studied the emergence of literacy skills in a 7-year-old
student with expressive and communication difculties, met with their academic collea-
gues (peers) to debrief them about their problems, concerns, and initial ndings with col-
leagues who had expertise and experience in a variety of related elds(p. 586). The
colleagues pointed out potential biases, suggested new meanings they may not have con-
sidered before, and identied places that needed clarication.
Most studies that used external reviewers or peer feedback did not specify the reasons
for inviting a particular individual to be the reviewer. It is possible that many researchers
simply turned to people in their personal networks (e.g., fellow researchers, graduate
assistants, etc.) to be the external reviewer. Among the ones that indicated a reason for
inviting a specic reviewer, we found four reasons.
Methodological proximity. Craggs and Kelly (2018) conducted interpretative phenomeno-
logical analysis (IPA) to understand student experiences of being removed from a regular
school and sent to an alternative school. For external auditing, they invited a colleague
who had used IPA in their own research projects and was thus aware of its conventions
and procedures.
Expertise. Two studies invited people based on their area of expertise. For example,
Diakow and Goforth (2021) invited people who were international experts in the eld
of migrant and refugee youth mental health, humanitarian assistance, Islamic studies,
and Middle Eastern culture(p. 244) to serve as external auditors for their study on
the wellbeing of Muslim refugee youth. Theron (2013) invited other resilience research-
ers for their study on the resilience of Black South African university students from dis-
advantaged backgrounds.
Sabnis and Wolgemuth 11
Linguistic background. Finally, reviewers played a role in studies carried out in languages
other than English. In Sluiter et al. (2019) study exploring teachersperspectives on
ADHD medications, the researchers consulted with a native speaker of Dutch and
English to review the quotes the authors wanted to include in the manuscript to ensure
their translation resemble(d) the language that was used by the respondents with a
focus on intended meaning rather than the proximity of words in literal translation
(p. 1264). Fulano et al. (2018) similarly invited a teacher uent in English and
Mozambican to translate the quotes from the interviews.
Lack of proximity. Finally, some researchers like Chen et al. (2016), who studied how stu-
dents intervene in bullying incidents, intentionally sought researchers from different
elds in order to get different perspectives on the studys provisional theoretical frame-
work and research results(p. 293).
Triangulation
The term triangulationemerges etymologically from navigation and refers to the act of
using multiple points to pinpoint the geographical location of an object (Smith, 1975). In
qualitative research, triangulation refers to the researcher eschewing reliance on any one
strategy and instead using multiple strategies to arrive at their conclusions. Twenty-three
percent of the studies (n=32) used this validity strategy. We found many different ways
in which researchers interpreted the term. Specically, these studies differed in what was
triangulated.
Triangulating the mode of data collection. The most common form of triangulation was
between multiple modes of data collection to gather data. For example, Frels et al.
(2013) used individual interviews, focus groups, and document review to arrive at
their ndings about the factors that seem to characterize the relationship between success-
ful mentors and their mentees.
Theory triangulation. Fontil and Petrakos (2015) triangulated by analyzing the data from
several theoretical perspectives. Specically, they reported triangulating their under-
standing of the participantsexperiences by looking at it through the lens of
Rimm-Kaufman and Pianta (2001) developmental model of transition and the theoretical
framework underlying the systems of care (SOC) programs.
Site triangulation. Nkoma and Hay (2018) talked about site triangulation wherein they
purposefully (sampled) three provincial ofces, ensuring a high level of variation of par-
ticipation characteristics in terms of gender, ethnicity, and time served in school psych-
ology services(p. 856). The use of multiple sites aided them in achieving the greatest
diversity(p. 856) in the cases thus enabling a well-rounded understanding of the
phenomenon.
12 School Psychology International 0(0)
Prolonged engagement
Lincoln and Guba (1985) identied prolonged engagement with the community as an
important strategy for enhancing the quality of the research. At the core of this principle
is the idea that developing an intimate understanding of participants and the context in
which they live can allow the researcher richer insights into their actions. Sixteen
percent of the articles (n=22) involved prolonged engagement as a way to establish
the validity of the study.
In comparison to other validity strategies we have discussed so far, prolonged engage-
ment may be the most difcult to operationalize. Given the variety of research topics,
there is no one standard way to engage with a community nor is there a concrete criterion
to understand when ones engagement has been deepor prolongedenough. It is thus
up to each researcher to explain to the reader their attempts at the engagement, and to
allow the readers to decide whether it qualies as prolonged engagement. Given the oper-
ational fuzziness of the criteria, we found a variety of scenarios described as constituting
prolonged engagement. McGeown et al. (2017) reported enacting deep engagement by
using one single site for all their data collection. They believed that this yielded rich
data because of the opportunities to create strong rapport and to gain peoples trust.
Balagna et al. (2013) enacted deep engagement with a Spanish-speaking community
through phone calls as well as making several home visits to various families prior to the
interviews. This was especially salient given that their research involved participants
from a Spanish-speaking community and that the interviews were to be in Spanish as
well. Following the data analysis, they visited the interviewed families at their homes
to present their data interpretation and seek feedback (member checking).
Some researchers leveraged their prior relationship with a school as a form of validity.
For example, Wood et al. (2017) discussed their working relationship with the two school
sites for over ten years.They explained how this intimate familiarity with the culture of
the schools aided them in the creation of the interview protocol as well as the interpret-
ation of the data in context(p. 6). It also helped them during the interviews because they
were able to express a shared sense of knowledgeabout the school to the students they
were interviewing (p. 6). Similarly, Pillay (2014) discussed how his career as an educa-
tional psychologist and academic involved in promoting childrens rights(p. 229)
allowed him rich insights and enhanced the credibility of his conclusions.
Not discussed
Fifteen percent (n=21) of the studies did not explicitly describe the steps they took to
enhance the validity of their study.
Thick rich descriptions
Eleven percent (n=16) of the articles mentioned thick rich descriptions as the validity-
bolstering mechanism used by their authors. Thick rich description as a form of validity
emerged from ethnographic research, which focuses on understanding the cultural
Sabnis and Wolgemuth 13
practices of various social groups. Geertz (1973) advocated for thick rich description as a
way for the ethnographers to describe various cultural practices. This could be done by
providing a rich account of a cultural milieu that revealed the ethnographers intricate
understanding of culture. An ethnographers ability to write thick rich descriptions testi-
es to their deep immersion in the culture. Although emerging from ethnography, thick
rich description spread to other disciplines as a means of judging the trustworthiness of an
account. In the studies we reviewed, authors enacted thick rich descriptions by providing
the reader with many different quotes from the participants to ensure that the partici-
pantsvoices were well-represented(Haegele et al., 2020). In addition to using direct
quotes, researchers such as Vilbas and King-Sears (2021) enacted this procedure by
describing their participants, setting, and emergent themesin rich detail (p. 877).
Saturation
Ten percent (n=15) of the studies cited saturation as a form of validity-boosting practice.
Saturation as a practice refers to the practice of collecting data and simultaneously coding
it until no new codes emerge. The researchers may thus not start with a rigid ceiling for
the number of participants. Rather they continue to interview participants until they reach
a point at which the existing codes are sufcient to analyze the new information. There is
no standard way to decide the ceiling. Castillo et al. (2016) determined the ceiling had
been reached when the last two interviews yielded no new information(p. 647).
Audit trail
Onwuegbuzie and Leech (2007) dened an audit trail as maintaining extensive docu-
mentation of records and data stemming from the study(p. 240). Nine percent (n=
13) of the articles cited the use of audit trails, which Vilbas and King-Sears (2021) con-
cisely described as detailing the process of data collection, analysis, and interpretation of
the data(p. 877). OBryon and Rogers (2016) described it as documenting the process
and product of the investigation(p. 228) including various methodological considera-
tions and research decisionsthey took at each step of the research process. They argue
the audit trail aided them in being reexive about the research. Parker et al. (2020) simi-
larly talked about recording theoretical memos and notes throughout each transcript to
ensure that all members of the research team had a clear understanding of data interpret-
ation(p. 119). Wood et al. (2017) also discussed documenting the various changes they
made to the interview protocol and codebook as the study progressed.
Providing a detailed account of the research process
Six percent (n=8) of the articles provided a detailed description of how the research was
conducted from design to analysis and interpretation, and claimed this as the mechanism
that validated their studies. For example, Wood et al. (2017) stated the current article
offers in depth details about the process, context, and results of the study to provide a
detailed description that is designed to allow readers to assess transferability of these
14 School Psychology International 0(0)
ndings to other settings(p. 7). Meyer and Cranmore (2020) stated to address depend-
ability, the research report contains detailed procedural descriptions, interview protocols,
codebook, excerpts, and other resources for replication(p. 1062).
Training in coding or interviewing
Six percent of the studies (n=8) reported using training to enhance the rigor of their
research process. In Castillo et al. (2016), one of the research members with expertise
in qualitative research trained the other team members in interviewing. They met with
the research team, and explained common tenets of interviewing, followed by partici-
pants conducting mock interviews using the study protocol. Finally, the expert provided
feedback to the team members. Once data were collected, the expert also trained the team
members in thematic analysis which included provid[ing] information on identifying
thought units, coding, and thematic analysis(p. 646). This was followed by the team
member coding a partial transcript together in a meeting. Following this meeting, the
team members coded the rest of that transcript separately and met again to discuss
their coding. Similarly, OBryon and Rogers (2016) talked about extensive training-
both didactic and applied(p. 228) for all the coders in the method of content analysis.
Cheng et al. (2011) reported all the researchers undergoing a three-hour training on
coding, although no details were provided of this training. Moy et al. (2014) discussed
how the research team generated a list of domains. Following the creation of this list,
their research team members were trained to identify the domains as consensually
dened and to code all transcripts according to these domains(p. 328).
Negative case analysis
Three percent of the articles (n=4) featured negative case analysis as a validation strat-
egy. This procedure emerged from grounded theory as a way to prevent overinterpreta-
tion of the data. Barrow and Hannah (2012) described it as scrutiniz[ing] the data for
deviant cases(p. 457). Giraldo-Garcia et al. (2021) described it as identify[ing] con-
cepts and incidents that did not t the resulting thematic framework, helping to safeguard
against a drift toward a priori assumption…” (p. 57). Balagna et al. (2013) used negative
case analysis by scanning the transcripts for information that contradicted the initial
interpretations. They then rened or broadened those interpretations to accommodate
the diversity in the participantsexperiences presented by the negative cases. For
Gross and Lo (2018), negative case analysis involved identifying the data that interrupted
or went counter to an emerging pattern; they looked deeper into this data to [analyze] the
possible source of the discrepancy(p. 384).
Threshold
Three percent of the articles (n=4) featured the threshold validity strategy. This strategy
occurs toward the end of the coding stage, where researchers are nalizing the thematic
schema. Although some researchers use an a priori approach to coding (wherein the
Sabnis and Wolgemuth 15
researcher predetermines a set of codes and then reads the data for instances of these
codes), many others take an inductive approach, wherein the codes are allowed to
emerge organically from the data. This liberal approach to coding can potentially generate
many more codes or subthemes than necessary to answer a research question. Threshold
provides a way to separate the signal from the noise by setting a criterion to prioritize
some codes while eliminating others from the nal thematic schema. For example,
McCabe and Rubinson (2008) eliminated codes that occurred in less than three focus
group interviews. Grapin et al. (2021) treated themes that occurred in at least 25% of
the interviews as major themes while themes that occurred lesser than that were desig-
nated as minor themes. Vega et al. (2019) only reported on themes that occurred in
two or more participant interviews.
Translational validity
In 3% of the studies (n=4), researchers conducted the interviews in a language other than
English, and transcribed them into English for the purpose of coding. For example, van
Schalkwyk and Sit (2013) interviewed participants in Chinese and then translated the
interviews into English for analysis. In doing so, they risked changing the meaning of
what the participants said given that the nuances of one language may not carry into
the other language. In such cases, the researchers would typically take one of several
routes. Van Schawlwyk and Sit (2013) coded the translated transcripts and then referred
to the original transcripts while writing the results (refer back to the original recordings
and transcripts in Chinese to verify credibility of interpretations[p. 158]).
Other researchers coded the transcripts in the original language and then translated the
quotes when writing the manuscripts. Sluiter et al. (2019) state that the focus of their
translation was on intended meaning rather than the proximity of words in literal trans-
lation(p. 1264). Similarly, Fulano et al. (2018) spoke about coding in the original lan-
guage and then translating the quotes at the manuscript-writing stage: Following
literature recommendations, data were used in the original language; namely, all
data analysis and direct quotations in the manuscript. In the nal version of the manu-
script, one of the coders, who is Mozambican, translated the direct quotations with the
support of an English teacher. Researchers discussed the nal version of quotations in
English compared to the original and reached total consensus(p. 202).
Relational ethics
Two percent of the studies (n=3) emphasized the researchersethical conduct with their
participants during data collection as a form of validity. Their actions went beyond the
standard ethical mandate about non-coercion, condentiality, and privacy, and included
taking steps to weaken the power imbalance between the researcher and the participant.
One of the ways to do this was by disclosing their positionality to participants. This was
based on the sensitivity to context, as a validity strategy presented by Yardley (2000).
Haegele et al. (2020) explained their positionality to participants to expose his biases
and included an abundant number of verbatim transcript extracts in the presentation of
16 School Psychology International 0(0)
the results to ensure that the participantsvoices were well-represented. Explaining ones
positionality can help weaken the power imbalance by reversing the ow of information
in a research study, wherein the participants are typically the ones revealing personal
information to the researcher.
Frels et al. (2013) were one of the few articles that discussed ethics as an important part
of validity. They used Patton (2002) argument that a research studys credibility was
deeply connected to the integrity of the researcher. To this end, they used Pattons
Ethical Issues Checklist to inform their behavior during the data collection process.
Specically they spoke about doing the following things to demonstrate their integrity
as researchers: purpose of inquiry and methods, (b) recognizing promises and reci-
procity, (c) assessing risks, (d) maintaining condentiality, (e) obtaining informed
consent and assent from all parties, (f) understanding data access and ownership, (g)
being cognizant of interviewer mental health, (h) consulting for advice, (i) recognizing
data collection boundaries, and (j) maintaining ethical or legal conduct issues(Frels
et al., 2013, pp. 621622).
Nkoma and Hay (2018) followed Elanglander (2012) advice to have a preliminary
meeting with the participants as a way to establish trust, comfort, and safety without col-
lecting any data. To this end, they met with the participants one week before the sched-
uled interviews and used this to introduce themselves to the participants, talk about
ethical considerations and informed consent, and to preview the questions they would
be asked in their interviews the following week in order to give them time to think
about their answers. The authors argued that this ethical conduct enabled them to
obtain a richer description during the interviews without making the researcher ask
too many questions(p. 855).
Strengthening validity practices in school psychology research:
discussion and a few recommendations
Validity is an important consideration for researchers, helping a research community
to evaluate research studies and elevate the ones that potentially advance the eld or
enrich the discourse. Whereas quantitative researchers are subjected to well-
established criteria for judging the validity of a study, qualitative researchers are
expected to decide the criteria against which they want a specic study to be evaluated.
Multiple reporting standards established to guide the writing of qualitative research
manuscripts therefore emphasize the importance of talking about what makes the
study valid, but largely leave it up to the authors to select the criteria and strategies
to argue their studys validity.
Our systematic review uncovered a variety of strategies qualitative researchers in
school psychology used to establish and communicate the validity of their studies. The
diversity of validation strategies shows the idiosyncratic nature of each qualitative
research project. While the variations in validity strategies may reect authors personal
preferences, they may also reect the editorsand reviewersrequests for authors to
clarify the process of their inquiries. Regardless of the validity strategy used, it is
Sabnis and Wolgemuth 17
important to remember that a researcher can work toward validity but never fully attain it
(Koro-Ljungberg, 2010). Validity is a lodestar that can guide a research team and help
them navigate the research process to produce goodndings. However, all ndings
coming out of a study are a product of the researchers limited sensemaking processes
as well as a certain degree of randomness that characterizes the world. Thus, qualitative
researchers should not expect their strategies to yield a dichotomous outcome (i.e., valid
vs. invalid)but understand that validity is a matter of degree (Onwuegbuzie & Leech,
2007, p. 239).
Having discussed the different ways to validate studies, we reected upon the ndings
to identify how future qualitative research in school psychology might be strengthened.
Below we discuss four recommendations for potentially improving the practice of valid-
ity in school psychology research.
Recognizing the distinction between criteria and strategies
We believe it is important to distinguish between the criteria for judging a manu-
scripts validity and the strategies used to meet the criteria. A criterion species a
researchers values about how they evaluate a research study or a research publica-
tion and what characteristics they expect to see in a goodstudy. A strategy is a pro-
cedure or the means that help them live up to these standards. Thus, a quantitative
researcher who judges experimental studies based on the criteria of external validity
may employ strategies to make the experimental conditions resemble real-world
conditions. A qualitative researcher who judges research studies based on the criteria
of resonance (i.e., the extent to which it meaningfully reverberate(s) and affect(s)
the reader; Tracy, 2010, p. 844) may use strategies such as evocative writing (i.e.,
writing in a way that is vivid, engaging, and structurally complex; Tracy, 2010,
p. 845).
Although we acknowledge the distinction between criteria and strategy may be blurred
in some cases, we nd it pragmatically useful to distinguish between the two when teach-
ing qualitative research. In the absence of this distinction, qualitative researchers often
tend to conate the two. In the current study, we found many cases where authors
named their validation strategy (e.g., interrater agreement) without clarifying the criteria
they were trying to meet with the strategy as if the criteria were self-evident. Kahn (1993)
argued that this way of approaching validity leads to a procedural charadewherein val-
idity procedures come to be used ritualistically simply because they are expected or have
been used before. Thus, it is important for researchers to clarify the criteria they use to
judge their studys validity before discussing the means or strategies they used to meet
those criteria.
In this article, we illustrated what we consider to be good practice in regard to discuss-
ing validity in an article or a dissertation. We started by describing the criteria we selected
to help orient our study and also explained why we used those two criteria over others.
We then discussed the means or strategies by which we believe we were able to honor
these criteria.
18 School Psychology International 0(0)
Aligning validity and the paradigmatic position
Paradigmatic position refers to the kind of knowledge researchers and research commu-
nities value. Conducting interpretivist research means valuing knowledge about the
meanings people assign to events and practices around them. Inquiring under a critical
paradigm means exposing the hidden workings of power in oppressive and unjust
systems. Postmodern inquiries value knowledge that afrms the uncertain and fragmen-
ted nature of reality against the totalizing power of grand narratives. And postpositivist
research seeks knowledge that appears to be objective and unbiased, regardless of
what the knowledge is about. As such, what constitutes valid research is contingent on
the paradigmatic position adopted for the study.
We recommend researchers carefully consider their reasons for selecting validity strat-
egies and criteria in light of their studys stated or implied paradigmatic position. This
careful consideration requires awareness and understanding of the various research para-
digms available to qualitative researchers and explicit adoption of a paradigmatic position
aligned with a studys aims, purposes, methodology, methods, and validation criteria and
strategies. For example, if a researcher seeks to understand how teachers experience and
make sense of a new policy implemented in their schools, then that researcher will likely
choose the interpretivist paradigm to undergird their research design because they value
teacherssubjective perspectives and meaning-making. Selecting a critical paradigm for
this study, for example, would be a mismatch unless the researcher makes clear their
interest in how the new policy maintains oppression and/or enacts social justice.
Having carefully considered and selected an interpretivist paradigm, the researcher
should therefore avoid stating that they used intercoder agreement to obtain a more
accurateor objectiveunderstanding of the data. This is because the interpretivist pos-
ition does not hold that researchers create objective knowledge. Rather, under interpreti-
vism, all research ndingsare interpretations. Therefore, a researchers decision to use
multiple coders in an interpretivist study would be better justied by stating that incorp-
orating the multiple perspectives during coding helped them achieve a well-rounded,
nuanced, and multifaceted understanding of the teachersexperiences and understandings
of the new policy.
Going beyond positivist and interpretivist traditions of validity
Reecting on Figure 1, we notice the pervasive inuence of two well-established research
traditionspositivism and interpretivismon school psychology researchersapproach
to validity in qualitative research. For example, the rst and third most popular validity
strategies in school psychology research are interrater agreement and consensual coding.
In our own experience as mentors, grantwork collaborators, and members of dissertation
committees we have come across the misconception that every qualitative study must
include either interrater agreement or consensual coding, which is untrue. Although
both are useful, they are not mandatory for every qualitative study (Morse, 1997,
2015), and believing so can disadvantage researchers who may be unable to nd a co-
researcher to help them code their transcripts due to the expertise or time commitment
Sabnis and Wolgemuth 19
required from the co-coder. These strategies may even be detrimental in studies that
involve interpretive coding (see Morse, 2015). Qualitative researchers who may be the
sole researcher (e.g., a graduate student collecting and analyzing data for their disserta-
tion) can nevertheless produce rigorous and high-quality work by employing other val-
idity strategies discussed above as well as those discussed in Morse (2015). The
remaining strategies in Figure 1 (barring relational ethics) are all traditionally associated
with the interpretivist paradigm which dominates qualitative research in the social
sciences.
The least commonly used validity strategyrelational ethicsis associated with the
critical paradigm of research, which seeks to uncover and challenge oppressive social
processes. Researchers working from this paradigm are acutely aware of researchershis-
toric role in legitimizing oppressive ideologies (e.g., eugenics) and exploiting margina-
lized communities for data (e.g., the Tuskegee Syphilis Study). Critical researchers
working toward equity and social justice thus start by challenging the unequal balance
of power in the research project itself, holding that the communities should have some
level of say in the inquiry that impacts them (Kemmis & McTaggert, 2000). This may
be done by engaging participants as co-researchers, taking active measures to increase
their agency and involvement in the project, and ensuring the project benets the parti-
cipants or the community where it was conducted. Although we found these practices
in some of the studies using participatory forms of research, the authors did not explicitly
discuss these practices as sources of validity for their studies.
It is possible that the researchers may not be aware that these strategies do serve as a
means of validation within the critical paradigm of research. Summarizing Denzin (2015)
account of critical qualitative research, Koro-Ljungberg and Canella (2017) note that crit-
ical qualitative inquiry places the voices of the oppressed at the center of inquiry, reveals
sites for change, promotes change in peoples lives, and has the potential to affect policy.
As such critical qualitative inquiry is always already ethical…” (p. 330). Validating crit-
ical qualitative research thus includes discussing the ethical aspects of the inquiry
throughout the study and including the authors ethical deliberation in the research
write-up.
Appreciating the role of creativity
Finally, we want to highlight the importance of creativity, which often gets oversha-
dowed by discussion of the techniques and strategies of qualitative research. Although
the postpositivist inuence can be credited for strengthening qualitative research by
emphasizing rigor, discipline, and systematicity in knowledge production, Tracy
(2010) warned that a rigorously designed study does not automatically result in high
quality work. Qualitative methodology is as much art as it is effort, piles of data, and
time in the eld. And just like following a recipe does not guarantee perfect presenta-
tionrigor does not guarantee a brilliant nal product(p. 841).
In fact, an excessive emphasis on procedural rigor can come at the expense of creativ-
ity that is vital to a strong qualitative study (Whittemore et al., 2001). That is because
procedural rigor can engender rigidity and compliance-orientation which conicts with
20 School Psychology International 0(0)
the open-mindedness and intuitiveness associated with creativity (Sandelowski, 1993). A
research team that is purely focused on adhering to a preplanned strategy for data analysis
may miss out on an important pattern in the data that would have interested the reader. A
good qualitative researcher is thus tasked with knowing when to follow rigorous proce-
dures and when to follow instincts. The former may help in the efcient completion of a
study and prevent a researcher from straying off course, whereas the latter can lead to
unexpected insights that push the eld forward. Qualitative researchers must therefore
understand that rigorous procedures alone do not make their studies valid or worthy of
readers engagement. Their manuscripts must also demonstrate a depth of thought, a
newness of insights, and richness of writing that stimulates the reader, increases interest
in the topic of inquiry, and helps to broaden qualitative researchs audience.
Conclusion
As interest in qualitative research grows among school psychology researchers, we will
likely continue to see an increase in the publication of qualitative studies in school psych-
ology journals and doctoral dissertations. It is therefore important to attend to issues of
validity in order to ensure high quality research that advances the eld. One of the
ways to improve validity practices in a eld is to look back at its use by past researchers
in the eld, studying how they adapted its strategies for different contexts and reecting
on how these practices could be improved for future research. We therefore conducted
this systematic review of qualitative literature in school psychology journals. Our nd-
ings highlighted the diversity of strategies used by school psychology researchers to
establish their studys validity, as well as the creativity that goes into adapting a validity
strategy for the specic context of each study. In addition to attending to these procedural
aspects of validity, we also recommend school psychology researchers to attend to the
deeper issues underlying validity practices, such as the paradigmatic connections and
the role of creativity in crafting a strong qualitative study.
Limitations of the study
As with all studies, ours was shaped as well as constrained by our own beliefs, world-
views, experiences, and social positionalities. Thus, we make no claims about the univer-
sality or objectivity of our ndings or claims. We believe it is ultimately up to the reader
to determine the usefulness of this manuscript based on the arguments, reasoning, and
evidence we provide. This is known as transferability which is analogous to what quan-
titative researchers call generalizability (Levitt et al., 2018). Nevertheless, we would like
to identify some limitations of our study. First, our ndings focused on the strategies the
authors used to validate their studies. Although we were also curious to understand the
criteria toward which the strategies were applied, it was difcult to code this information
in the data given that researchers often deployed terms like validity, trustworthiness,
dependability, credibility, and rigor interchangeably. For example, both Meyer and
Cranmore (2020) and Wood et al. (2017) use the validity strategy of providing a detailed
account of their research process to the reader. However, Meyer and Cranmore (2020)
Sabnis and Wolgemuth 21
claims to use this to increase their studysdependability whereas Wood et al. (2017) used
it to increase their studystransferability, indicating the two criteria are not adequately
differentiated from each other. Other authors simply excluded any mention of validity cri-
teria and only discussed validity strategies. Thus, we made a strategic choice to focus our
data analysis on validity strategies only, since they were more concrete, had less inter-
changeable use of terms, and were therefore more feasible to code. Future research
studies might focus on the criteria of validity used by school psychology researchers.
Another limitation of the study is that we only looked at the strategies that authors
explicitly identied as increasing their studys validity. Thus, although we found many
articles containing thick rich descriptions, we did not classify an article under the thick
rich description category unless the authors explicitly stated they used this strategy.
We did this because we were interested in understanding what researchers believed
made their studies valid. A future study interested in evaluating the quality or validity
of qualitative research published in school psychology journals might use a broadened
scope of coding that includes both stated and unstated ways in which authors strengthen
their studys validity.
Declaration of conicting interests
The authors declared no potential conicts of interest with respect to the research, authorship, and/
or publication of this article.
Funding
The authors received no nancial support for the research, authorship, and/or publication of this
article.
ORCID iD
Sujay V. Sabnis https://orcid.org/0000-0001-9424-2796
References
Arora, P. G., Sullivan, A. L., & Song, S. Y. (2022). On the imperative for reexivity in school
psychology scholarship. School Psychology Review. Advance online publication. https://
doi.org/10.1080/2372966X.2022.2105050
Bakeman, R., & Gottman, J. M. (1986). Observing interactions: An introduction to sequential ana-
lysis. Cambridge University Press.
Balagna, R. M., Young, E. L., & Smith, T. B. (2013). School experiences of early adolescent
latinos/as at risk for emotional and behavioral disorders. School Psychology Quarterly, 28(2),
101121. https://doi.org/10.1037/spq0000018
Barrow, W., & Hannah, E. F. (2012). Using computer-assisted interviewing to consult with children
with autism spectrum disorders: An exploratory study. School Psychology International,33(4),
450464. https://doi.org/10.1177/0143034311429167
Bochner, A. (2000). Criteria against ourselves. Qualitative Inquiry, 6(2), 266272. https://doi.org/
10.1177/107780040000600209
22 School Psychology International 0(0)
Campbell, D. T., & Stanley, J. C. (1963). Experimental and quasi-experimental designs for
research. Rand McNally.
Canpolat, M., & Atli, A. (2021). Children who changed schools due to their parentsrelocation for
employment: Where do we go now? Psychology in the Schools,58(11), 21892201. https://doi.
org/10.1002/pits.22584
Cassidy, W., Brown, K., & Jackson, M. (2012). Under the radar: Educators and cyberbullying
in schools. School Psychology International,33(5), 520532. https://doi.org/10.1177/
0143034312445245
Castillo, J. M., Wolgemuth, J. R., Barclay, C., Mattison, A., Tan, S. Y., Sabnis, S., Brundage, A., &
Marshall, L. (2016). A qualitative study of facilitators and barriers related to comprehensive and
integrated school psychological services. Psychology in the Schools,53(6), 641658. https://doi.
org/10.1002/pits.21932
Chen, L. M., Chang, L. Y. C., & Cheng, Y.-Y. (2016). Choosing to be a defender or an outsider in a
school bullying incident: Determining factors and the defending process. School Psychology
International,37(3), 289302. https://doi.org/10.1177/0143034316632282
Cheng, Y.-Y., Chen, L.-M., Ho, H.-C., & Cheng, C.-L. (2011). Denitions of school bullying in
Taiwan: A comparison of multiple perspectives. School Psychology International,32(3),
227243. https://doi.org/10.1177/0143034311404130
Cho, J., & Trent, A. (2006). Validity in qualitative research revisited. Qualitative Research,6(3),
319340. https://doi.org/10.1177/1468794106065006
Craggs, H., & Kelly, C. (2018). School belonging: Listening to the voices of secondary school stu-
dents who have undergone managed moves. School Psychology International,39(1), 5673.
https://doi.org/10.1177/0143034317741936.
Danziger, K. (1979). The positivist repudiation of Wundt. Journal of the History of the Behavioral Sciences,
15(3), 205230. https://doi.org/10.1002/1520-6696(197907)15:3<205::AID-JHBS2300150303>3.0.
CO;2-P
Davies, S. (2020). School-based support for families of students with traumatic brain injuries.
Journal of Applied School Psychology,36(3), 275292. https://doi.org/10.1080/15377903.
2020.1734708
Denzin, N. D. (2015). What is critical qualitative inquiry. In G. S. Cannella, M. S. Pérez, &
P. A. Pasque (Eds.), Critical qualitative inquiry: Foundations and futures (pp. 3150).
Walnut Creek.
Diaków, D. M., & Goforth, A. N. (2021). Supporting muslim refugee youth during displacement:
Implications for international school psychologists. School Psychology International,42(3),
238258. https://doi.org/10.1177/0143034320987280
Eketu, C. A. (2017). Management research: A thought onvalidity of positivism. International Journal
of Advanced Academic Research,3(11), 133139. https://www.ijaar.org/articles/Volume3-
Number11/Social-Management-Sciences/ijaar-sms-v3n10-oct17-p7.pdf
Elanglander, M. (2012). The interview: Data collection in descriptive phenomenological human
scientic research. Journal of Phenomenological Psychology,43(1), 1335. https://doi.org/
10.1163/156916212X632943
Fontil, L., & Petrakos, H. H. (2015). Transition to school: The experiences of Canadian and immi-
grant families of children with autism spectrum disorders. Psychology in the Schools,52(8),
773788. https://doi.org/10.1002/pits.21859
Frels, R. K., Onwuegbuzie, A. J., Bustamante, R. M., Garza, Y., Nelson, J. A., Nichter, M., & Soto
Leggett, E. (2013). Purposes and approaches of selected mentors in school-based mentoring: A
collective case study. Psychology in the Schools,50(6), 618633. https://doi.org/10.1002/pits.
21697
Sabnis and Wolgemuth 23
Fulano, C., Cunha, J., Carlos Nunez, J., Pereira, B., & Rosario, P. (2018). Mozambican adolescents
perspectives on the academic procrastination process. School Psychology International,39(2),
196213. https://doi.org/10.1177/0143034318760115
Geertz, C. (1973). The interpretation of cultures: Selected essays. Basic Books.
Gephart, R. P. Jr. (2004). Qualitative research and the academy of management journal. Academy of
Management Journal,47(4), 454462. https://doi.org/10.5465/amj.2004.14438580
Giraldo, G. R. J., Voight, A., & OMalley, M. (2021). Mandatory voice: Implementation of a
district-led student-voice program in urban high schools. Psychology in the Schools,58(1),
5168. https://doi.org/10.1002/pits.22436
Grapin, S. L., Bocanegra, J. O., Rossen, E., Woytas, T. M., & Rossel, J. (2021). Student perspec-
tives on recruiting and orienting prospective school psychology trainees. School Psychology,
36(4), 261270. https://doi.org/10.1037/spq0000435
Gross, N., & Lo, C. (2018). Relational teaching and learning after loss: Evidence from black ado-
lescent male students and their teachers. School Psychology Quarterly,33(3), 381389. https://
doi.org/10.1037/spq0000285
Haegele, J. A., Zhu, X., & Holland, S. K. (2020). School-based bullying experiences as reected by
adults with visual impairments. Psychology in the Schools,57(2), 296309. https://doi.org/10.
1002/pits.22314
Hammersley, M. (1992). Whats wrong with ethnography? Methodological exploration.
Routledge.
Harré, R. (2004). Staking our claim for qualitative psychology as science. Qualitative Research in
Psychology,1(1), 314. https://doi.org/10.1191/1478088704qp002oa
Harry, B., Sturges, K. M., & Klingner, J. K. (2005). Mapping the process: An exemplar of process
and challenge in grounded theory analysis. Educational Researcher,34(2), 313. https://doi.org/
10.3102/0013189X034002003
Harvey, V. S., & Pearrow, M. (2010). Identifying challenges in supervising school
psychologists. Psychology in the Schools,47(6), 567581. https://doi.org/10.1002/pits.
20491
Hays, D. G., Wood, C., Dahl, H., & Kirk-Jenkins, A. (2016). Methodological rigor in journal of
counseling and development qualitative research articles: A 15 year review. Journal of
Counseling & Development, 94(2), 172183. https://doi.org/10.1002/jcad.12074
Justin, P., Dorard, G., Vioulac, C., Leu, A., & Untas, A. (2021). What do French school staff know
about young carers? A qualitative study about their perceptions. Psychology in the Schools,
58(8), 15311544. https://doi.org/10.1002/pits.22510
Kahn, D. L. (1993). Ways of discussing validity in qualitative nursing research. Western Journal of
Nursing Research,15(1), 122126. https://doi.org/10.1177/019394599301500111
Kelada, L., Hasking, P., & Melvin, G. A. (2017). School response to self-injury: Concerns of
mental health staff and parents. School Psychology Quarterly,32(2), 173187. https://doi.org/
10.1037/spq0000194
Kemmis, S., & McTaggert, R. (2000). Participatory action research. In N. Denzin, & Y. Lincoln
(Eds.) Handbook of qualitative research. (pp. 567606). Sage Publications, Inc.
Koro-Ljungberg, M. (2010). Validity, responsibility, and aporia. Qualitative Inquiry,16(8), 603
610. https://doi.org/10.1177/1077800410374034
Koro-Ljungberg, M., & Cannella, G. S. (2017). Critical qualitative inquiry: Histories, methodolo-
gies, and possibilities. International Review of Qualitative Research,10(4), 327339. https://
doi.org/10.1525/irqr.2017.10.4.327
Lahman, M. K. E., Damato, R. C., Stecker, S., & Mcgrain, E. (2006). Addressing the shortage of
rural school psychologists via technology: Using candidate qualitative interviews to inform
24 School Psychology International 0(0)
practice. School Psychology International,27(4), 439461. https://doi.org/10.1177/
0143034306070429
Larsen, T., & Samdal, O. (2012). The importance of teachersfeelings of self efcacy in developing
their pupilssocial and emotional learning: A Norwegian study of teachersreactions to the
second step program. School Psychology International,33(6), 631645. https://doi.org/10.
1177/0143034311412848
Lather, P. (1993). Fertile obsession: Validity after poststructuralism. The Sociological Quarterly,
34(4), 673693. https://doi.org/10.1111/j.1533-8525.1993.tb00112.x
Leath, S., Pster, T., Ball, P., Butler-Barnes, S., & Evans, K. A. (2021). A qualitative exploration of
school racial diversity, friendship choices, and black womens identity-based experiences in
high school. Journal of School Psychology,89,3450. https://doi.org/10.1016/j.jsp.2021.09.
006
Leininger, M. (1994). Evaluation criteria and critique of qualitative research studies. In J. M. Morse
(Ed.), Critical issues in qualitative research methods (pp. 95115). Sage.
Levitt, H. M., Bamberg, M., Creswell, J. W., Frost, D. M., Josselson, R., & Suárez-Orozco, C.
(2018). Journal article reporting standards for qualitative primary, qualitative meta-analytic,
and mixed methods research in psychology: The APA publications and communications
board task force report. American Psychologist,73(1), 2646. https://doi.org/10.1037/
amp0000151
Lincoln, Y. S., & Guba, E. A. (1985). Naturalistic inquiry. Sage.
McCabe, P. C., & Rubinson, F. (2008). Committing to social justice: The behavioral intention of
school psychology and education trainees to advocate for lesbian, gay, bisexual, and transgen-
dered youth. School Psychology Review,37(4), 469486. https://doi.org/10.1080/02796015.
2008.12087861
McGeown, S., Putwain, D., Thompson St. Clair, H., & Clough, P. (2017). Understanding and sup-
porting adolescentsmental toughness in an education context. Psychology in the Schools,
54(2), 196209. https://doi.org/10.1002/pits.21986
McIntosh, K., Martinez, R. S., Ty, S. V., & McClain, M. B. (2013). Scientic research in school
psychology: Leading researchers weigh in on its past, present, and future. Journal of School
Psychology,51(3), 267318. https://doi.org/10.1016/j.jsp.2013.04.003
McMahon, S. D., Peist, E., Davis, J. O., McConnell, E., Reaves, S., Reddy, L. A., Anderman, E.
M., & Espelage, D. L. (2020). Addressing violence against teachers: A social-ecological ana-
lysis of teachersperspectives. Psychology in the Schools,57(7), 10401056. https://doi.org/
10.1002/pits.22382
Meyer, M. S., & Cranmore, J. (2020). Supporting the information gathering practices of academ-
ically advanced college-bound students. Psychology in the Schools,57(7), 10571076. https://
doi.org/10.1002/pits.22372
Miles, M. B., & Huberman, A. M. (1984). Drawing valid meaning from qualitative data: Toward a
shared craft. Educational Researcher,13(5),2030. https://doi.org/10.3102/0013189X013005020
Miles, M. B., & Huberman, A. M. (1994). Qualitative data analysis: An expanded source book.
Sage Publications.
Morse, J. M. (1997). Perfectly healthy, but dead: The myth of inter-rater reliability. Qualitative
Health Research,7(4), 445447. https://doi.org/10.1177/104973239700700401
Morse, J. M. (2015). Critical analysis of strategies for determining rigor in qualitative
inquiry. Qualitative Health Research,25(9), 12121222. https://doi.org/10.1177/
1049732315588501
Moy, G. E., Briggs, A., Shriberg, D., Furrey, K. J., Smith, P., & Tompkins, N. (2014). Developing
school psychologists as agents of social justice: A qualitative analysis of student understanding
Sabnis and Wolgemuth 25
across three years. Journal of School Psychology,52(3), 323341. https://doi.org/10.1016/j.jsp.
2014.03.001
Newman, D. S., & Ingraham, C. L. (2020). Cross-university dialogues to support multicultural school
consultation training. Journal of School Psychology,81,1127. https://doi.org/10.1016/j.jsp.2020.
06.001
Ni, H., Jones, C., & Bruning, R. (2013). Chinese Teachersevaluation criteria as reected in nar-
rative student evaluations: Implications for psychological services in schools. School
Psychology International,34(2), 223238. https://doi.org/10.1177/0143034312437079
Nkoma, E., & Hay, J. (2018). Educational psychologistssupport roles regarding the implementa-
tion of inclusive education in Zimbabwe. Psychology in the Schools,55(7), 850866. https://doi.
org/10.1002/pits.22147
OBryon, E. C., & Rogers, M. R. (2016). Using consultation to support English learners: The
experiences of bilingual school psychologists. Psychology in the Schools,53(3), 225239.
https://doi.org/10.1002/pits.21904
Onwuegbuzie, A. J., & Johnson, R. B. (2006). The validity issue in mixed research. Research in the
Schools,13(1), 4863.
Onwuegbuzie, A. J., & Leech, N. L. (2007). Validity and qualitative research: An oxymoron?
Quality & Quantity: International Journal of Methodology,41(2), 233249. https://doi.org/
10.1007/s11135-006-9000-3
Parker, J., Garnes, J., Oliver, E., Amabile, A., & Sarathy, A. (2020). It takes a village:
Understanding African American high school studentsself-determination in school. School
Psychology Review,49(2), 111129. https://doi.org/10.1080/2372966X.2020.1717371
Patton, M. Q. (2002). Qualitative research and evaluation methods (3rd ed). Sage.
Phasha, T. N. (2008). The role of the teacher in helping learners overcome the negative impact of
child sexual abuse: A South African perspective. School Psychology International,29(3), 303
327. https://doi.org/10.1177/0143034308093671
Pillay, J. (2014). Advancement of childrens rights in Africa: A social justice framework for school
psychologists. School Psychology International,35(3), 225240. https://doi.org/10.1177/
0143034313515990
Powell, H., Mihalas, S., Onwuegbuzie, A. J., Suldo, S., & Daley, C. E. (2008). Mixed methods
research in school psychology: A mixed methods investigation of trends in the literature.
Psychology in the Schools,45(4), 291309. https://doi.org/10.1002/pits.20296
Proctor, S. L., Nasir, A., Wilson, T., Li, K., & Castrillon, P. (2018). Retention and persistence of
African-American students in school psychology programs. Psychology in the Schools,55(5),
506522. https://doi.org/10.1002/pits.22124
Pufpaff, L. A. (2008). Barriers to participation in kindergarten literacy instruction for a student with
augmentative and alternative communication needs. Psychology in the Schools,45(7), 582599.
https://doi.org/10.1002/pits.20311
Reid, A., & Gough, S. (2000). Guidelines for reporting and evaluating qualitative research: What
are the alternatives? Environmental Education Research,6(1), 5991. https://doi.org/10.1080/
135046200110494
Rimm-Kaufman, S. E., & Pianta, R. C. (2000). An ecological perspective on transition to kinder-
garten: A theoretical framework to guide empirical research. Journal of Applied Developmental
Psychology,21(5), 491511. https://doi.org/10.1016/S0193-3973(00)00051-4
Rush, S. C., & Wheeler, J. (2011). Enhancing junior faculty research productivity through multi-
institution collaboration: Participantsimpressions of the school psychology research collabor-
ation conference. Canadian Journal of School Psychology,26(3), 220240. https://doi.org/10.
1177/0829573511413552
26 School Psychology International 0(0)
Sabnis, S. V., Newman, D., Whitford, D., & Mossing, K. (2023). Publication and characteristics of
qualitative research in school psychology journals between 2006 and 2021. School Psychology.
Advance online publications.
Sandelowski, M. (1993). Rigor or rigor mortis: The problem of rigor in qualitative research revis-
ited. Advances in Nursing Science,16(2), 18. https://doi.org/10.1097/00012272-199312000-
00002
Sansosti, J. M., & Sansosti, F. J. (2012). Inclusion for students with high-functioning autism spec-
trum disorders: Denitions and decision making. Psychology in the Schools,49(10), 917931.
https://doi.org/10.1002/pits.21652
Sluiter,M. N., Wienen, A. W., Thoutenhoofd, E.D., Doornenbal, J. M.,& Batstra, L. (2019). Teachers
role and attitudes concerning ADHD medication: A qualitative analysis. Psychologyin the Schools,
56(8), 12591270. https://doi.org/10.1002/pits.22270
Smith, H. W. (1975). Strategies of social research: The methodological imagination. Prentice Hall.
Smith, M. L., & Glass, G. V. (1987). Research and evaluation in education and the social sciences.
Prentice Hall.
Suldo, S. M., Shaunessy, E., Michalowski, J., & Shaffer, E. (2008). Coping strategies of high
school students in an international baccalaureate program. Psychology in the Schools,45(10),
960977. https://doi.org/10.1002/pits.20345
Theron, L. C. (2013). Black studentsrecollections of pathways to resilience: Lessons for school
psychologists. School Psychology International,34(5), 527539. https://doi.org/10.1177/
0143034312472762
Tong, A., Sainsbury, P., & Craig, J. (2007). Consolidated criteria for reporting qualitative
research (COREQ): A 32-item checklist for interviews and focus groups. International
Journal for Quality in Health Care,19(6), 349357. https://doi.org/10.1093/intqhc/
mzm042
Tracy, S. J. (2010). Qualitative quality: Eight big-tentcriteria for excellent qualitative research.
Qualitative Inquiry,16(10), 837851. https://doi.org/10.1177/1077800410383121
van Schalkwyk, G. J., & Sit, H. H. Y. (2013). Evaluating school-based psychological and counsel-
ling services in Macao using a qualitative approach. School Psychology International,34(2),
154165. https://doi.org/10.1177/0143034312453395
Vega, D., Wolf, J. N., Barton, D. O., Stathatos, M., Iurino, C., & Hammer, L. (2019). Identifying
the training experiences and needs of bilingual school psychologists. Psychology in the Schools,
56(10), 16871699. https://doi.org/10.1002/pits.22288
Vilbas, J. A., & King-Sears, M. E. (2021). An effective elementary school counselors support of
students with disabilities: A case study. Psychology in the Schools,58(5), 873892. https://doi.
org/10.1002/pits.22476
Washington, M. L., Truscott, S., & Franklin, A. (2020). Using grant funding to enhance school
psychology training. Psychology in the Schools,57(8), 13091326. https://doi.org/10.1002/
pits.22389
Whittemore, R., Chase, S. K., & Mandle, C. L. (2001). Pearls, pith, and provocation. Validity in
Qualitative Research. Qualitative Health Research,11(4), 522537. https://doi.org/10.1177/
104973201129119299
Wolcott, H. (1990). On seekingand rejectingvalidity in qualitative research. In E. Eisner, &
A. Peshkin (Eds.), Qualitative inquiry in education: The continuing debate (pp. 121152).
Teachers College Press.
Wood, L., Smith, J., Varjas, K., & Meyers, J. (2017). School personnel social support and nonsup-
port for bystanders of bullying: Exploring student perspectives. Journal of School Psychology,
61,117. https://doi.org/10.1016/j.jsp.2016.12.003
Sabnis and Wolgemuth 27
Yardley, L. (2000). Dilemmas in qualitative health research. Psychology & Health, 15(2), 215228.
https://doi.org/10.1080/08870440008400302
Author Biographies
Sujay Sabnis, PhD, is an assistant professor at Miami University, and teaches courses on
consultation, multiculturalism, school experiences, and school psychology practicum. He
is a Nationally Certied School Psychologist (NCSP). His research interests include crit-
ical race theory, critical qualitative methodologies, consultation, and policy analysis in
K-12 public education.
Jennifer Wolgemuth, PhD, is an associate professor of Measurement and Research at
the University of South Florida. His research focuses on evaluating and developing
equity-oriented approaches to qualitative and mixed-methods research, including system-
atic reviews.
Appendix A
Coding proforma
For each journal, we extracted the following information in a spreadsheet:
- Year
- Volume
- Issue
- Total Articles (does not include acknowledgement, editorials, test reviews, book
reviews, corrections, and announcement)
- Number of qualitative articles
- Additional comments (optional)
A different team of researchers downloaded the qualitative articles identied in the
previous stage, and extracted the following information in a spreadsheet:
- Article citation
- Journal
- Study design
- Data collection strategy (interviews, focus groups, etc.)
- Data analysis strategy
- Discussion of reexivity
- Discussion of validity
- Overall strengths and limitations of the articles methodology
28 School Psychology International 0(0)
ResearchGate has not been able to resolve any citations for this publication.
Article
Full-text available
To understand the evolution and current status of qualitative research in School Psychology, we reviewed 4,346 articles published across seven school psychology journals between 2006 and 2021. The bibliometric analysis indicates that publication of qualitative research has increased over the years, but remains small (3%) when seen against the total volume of journal publications. Less than 5% of articles in all but one journal were qualitative. The most commonly explored topic was diversity, equity, and social justice accounting for 23% of the qualitative articles. In total, 55% of the studies were conducted in the United States. Although many studies did not specify participants' race and gender, the most commonly reported research participants were K-12 students, female, White, and from the United States. We discuss these findings and provide recommendations. (PsycInfo Database Record (c) 2023 APA, all rights reserved).
Article
Full-text available
This study investigated current school psychology students' perspectives on recruiting and orienting prospective trainees to the profession. Two hundred sixty-two student members of the National Association of School Psychologists (NASP) responded to two open-ended survey items regarding: (a) information they wished they had known prior to beginning their training and (b) their beliefs about effective recruitment practices for increasing racial and ethnic (R/E) diversity in the profession. Qualitative analyses revealed that approximately one-third of participants from R/E minoritized backgrounds reported wishing they had known more about the roles and functions of school psychologists. Participants also generated a wide range of potentially valuable recruitment strategies. Implications for improving recruitment practices are discussed. (PsycInfo Database Record (c) 2021 APA, all rights reserved).
Article
Full-text available
Contestations on research paradigms in the social sciences are replete in the literature on the philosophy of knowledge. The management discipline which was recreated from the intellectual domain of the social sciences is also cut in the contestations. This paper identifies the validity of positivism, which is one of the prescribed preferences for social inquiry. The paper contends that management is an admixture of science and art, in which case, the science based phenomena with stable, predictable, controllable, and explainable characteristics in management tend to bear scientific elements. This aspect of management is skewed to ontological realism; epistemological positivism; human nature determinism; and nomothetic methodology, all corresponding to a choice in positivism based inquiry. This makes positivism a valid research paradigm in management inquiry, because of its strong validity and reliability in the progression of search for truth in management inquiry. Positivism has remained with the growing consciousness that the natural world of matter shares useful analytical precepts with social phenomenon.
Article
Reflexivity, defined as the critical analysis of how one’s identity and values influence their scholarship, has been underscored as a crucial element of antioppressive scholarship. Despite its importance, reflexivity, and particularly its documentation, remains relatively uncommon in school psychology scholarship. In the following commentary, we introduce the need for reflexivity as it relates to the field’s commitment to antiracism, followed by a brief review of the literature on reflexivity, including its historical foundations and its previous applications within the fields of psychological research. We conclude with recommendations for the school psychology scholarly community for conceptualizing, conducting, and disseminating scholarship. Impact Statement Despite its importance, reflexivity, or the critical analysis of how one’s identity and values influence their scholarship, in school psychology scholarship remains uncommon. In this commentary, we introduce the need for reflexivity within school psychology scholarship and provide recommendations for the school psychology scholarly community. This manuscript serves as a call to increase implementation of a reflexive approach in school psychology scholarship.
Article
Social integration is a critical component of adolescents' positive school adjustment. Although prior scholars have highlighted how Black women and girls' social identities (e.g., race, gender, social class) influence their academic and social experiences in school, very little work has focused on how school racial diversity shapes Black girls' peer networks throughout K–12 education. To address this gap in the literature, the present qualitative study explored the narratives of 44 Black undergraduate women (Mage = 20 years) who reflected on their friendship choices in high school. We used consensual qualitative research methods to examine how Black women navigated friendships during their time attending predominantly White (less than 20% Black), racially diverse (21%–60% Black), and predominantly Black (61%–100% Black) high schools. Coding analyses revealed five friendship themes: (a) Black female friends, (b) mostly Black friends, (c) mostly interracial friends, (d) mostly White friends, and (e) White friends in academic settings and Black friends in social settings. Our findings highlight how the young women's ongoing negotiation of racialized and gendered school norms influenced their sense of closeness with same-race and interracial peers. Black girls may have challenges with forming lasting and meaningful friendships when they cannot find peers who are affirming and supportive, particularly in predominantly White school contexts. This study underscores the need to look at how racial diversity in the student population offers school psychologists and educators insight into how to better support the social and emotional development of Black girls.
Article
Approximately 300,000 people change provinces in Turkey every year due to relocation for employment. Among them are school-age children who need to change schools due to the relocation of their parents. This qualitative research aims to understand what students, who have to change schools due to the relocation of their parents to another city, experience in the process of school adaptation. All participants had to change schools at least once due to relocation for the employment of their parents. As a result of the thematic analysis of the qualitative interviews with 25 participants between the ages of 11 and 18, the following five main themes were identified: “Receiving news of the employment relocation,” “Effects on academic success,” “Influences on friendships,” “Sources of support,” and “Suggestions to novices.” Within the framework of the findings obtained, practical recommendations are presented to professionals who provide psychological assistance services on this subject.
Article
Young carers (YCs) are children and adolescents who provide a significant help to family member(s) with a disease or a disability. They often remain hidden, especially in France where awareness levels and policy responses regarding YC are just emerging. Several studies show that YC often have difficulties at school because of their situation. However, school staff seems to have difficulties to identify and support YC. The aim of this study was to explore schools teachers and other professionals' levels of awareness and their perceptions of the potential impacts of caring on pupil attainment, achievement, and wellbeing. Thirty‐three professionals working with pupils from 6 to 18 years old (primary, secondary, and high schools) took part in a semi‐structured interview and in a short phone interview. The results showed that almost all had already met a YC. Despite knowledge and perceptions of potential impacts for their pupils, few undertook actions to support them. Healthcare professionals and school counsellors seemed more aware of these situations than teachers and more prone to offer support. These results underline the need to enhance all school staff's knowledge about YC and help them develop appropriate support.
Article
School psychologists around the world have an important role in advocating for refugee youth, including Muslim refugees who experience increased risk for mental health issues due to discrimination. There are few studies, however, that have examined the resiliency factors that impact Muslim refugee youth’s well-being. This qualitative research study investigated these factors through the voices of humanitarian workers who are the primary source of support for these youth during displacement. Humanitarian workers ( N = 28 ) were recruited using snowball sampling through organizations that serve refugee youth and families during displacement. Using a recursive thematic process, we identified four main themes associated with risk factors: (1) gender norm expectations, (2) challenges in help-seeking, (3) unsafe climate and limited resources, and (4) challenges in collaborating with family. We also identified two main themes of protective factors: (1) faith and culture, and (2) family, peer, and community connections. We discussed these themes in the context of school psychological practice and provided specific recommendations for international school psychologists who support Muslim refugee students.
Article
Using a qualitative single‐case study design, perceptions from multiple stakeholders about how an effective elementary school counselor provides supports for students with disabilities (SWD) were examined. Interviews occurred with a school counselor, a student with a disability, his parent, his general and special educators, and school administrators. Findings are described using an ecological perspective and highlight the roles of an effective school counselor as well as the importance of interactions among the school counselor, parents, teachers, and administrators to support SWD. In addition, unique perspectives were shared by an elementary student with a disability. Implications for school counseling practices and programs are provided.
Article
This qualitative case study explores how urban high schools implement a district‐initiated student‐voice program, referred to as Student Advisory Committees, intended to improve school climate. The study uses an implementation science framework to analyze interview data from principals and advisors in 22 urban high schools. An iterative, top-down thematic analytic procedure was used for analysis and interpretation of the data. The results suggest wide variation in how the program was implemented in terms of fidelity, dosage, quality, program reach, and adaptation, with contextual school factors—namely, principal and school characteristics—that help to explain this variation. In addition, the findings indicate that these contextual factors had a significant influence on how the school district implemented the student advisory committees. The authors discuss the implications for program implementation in the context of urban school districts and for school psychologists.