|
|
|
|
|
|
|
|
|
|
Cómo citar este artículo:
Farida, F., Alamsyah, Y. A., Anggoro, B. S., Andari, T., & Lusiana, R.
(2024). Validación de una Herramienta de Evaluación Basada en el Modelo Rasch para Medir la Resolución Creativa de Problemas en
Estudiantes Mediante el Uso de TIC [Rasch Measurement Validation of an Assessment
Tool for Measuring Students’ Creative Problem-Solving
through the Use of ICT]. Pixel-Bit.
Revista De Medios Y Educación, 71, 83–106. https://doi.org/10.12795/pixelbit.107973
ABSTRACT
Despite increasing recognition of the importance of
creative problem solving (CPS) through the use of ICT in
independent curriculum education, there is a lack of comprehensive psychometric
validation for CPS assessment instruments. This study aimed to develop and
evaluate an assessment instrument to measure CPS through the
use of ICT students using the Rasch model. A total of 137 higher
education students participated as respondents. For this purpose, 20 items were
created, covering different aspects of CPS. Data analysis was
performed using Winstep and SPSS software. The Rasch
model was employed to confirm the validity and reliability of the newly
developed measurement instrument. The findings of the analysis of the Rasch
model indicated a good fit between the assessment items and individual
students. The items demonstrated adequate fit with the Rasch model, allowing
for differentiation of difficulty levels among different items and exhibiting a
satisfactory level of reliability. The Wright map analysis revealed patterns of
interaction between the items and individuals, effectively discriminating
between varying levels of student abilities. In particular,
an item showed DIF based on gender, which favours male students in terms
of their response abilities. Furthermore, the study identified that female
students in the fourth semester exhibited higher average response abilities
compared to female students in the sixth and eighth semesters. Furthermore,
significant differences in response abilities were observed between male and
female students, as well as between students who resides in urban and rural
areas. These findings are crucial for educators, emphasising the need to
implement effective differentiation strategies.
RESUMEN
A pesar del creciente reconocimiento de la importancia
de la resolución creativa de problemas (CPS) a través del uso de las TIC en la
educación con un currículo independiente, existe una falta de validación
psicométrica integral para los instrumentos de evaluación de CPS. Este estudio
tuvo como objetivo desarrollar y evaluar un instrumento de evaluación para
medir la CPS a través del uso de las TIC en estudiantes, utilizando el modelo
Rasch. Participaron un total de 137 estudiantes de educación superior como
encuestados. Para este propósito, se crearon 20 ítems que cubrían diferentes
aspectos de la CPS. El análisis de datos se realizó utilizando el software
Winstep y SPSS. Se empleó el modelo Rasch para confirmar la validez y
fiabilidad del instrumento de medición recién desarrollado. Los hallazgos del
análisis del modelo Rasch indicaron un buen ajuste entre los ítems de
evaluación y los estudiantes individuales. Los ítems demostraron un ajuste
adecuado con el modelo Rasch, lo que permitió diferenciar los niveles de
dificultad entre diferentes ítems y mostró un nivel satisfactorio de
fiabilidad. El análisis del mapa de Wright reveló patrones de interacción entre
los ítems y los individuos, discriminando efectivamente entre los diversos
niveles de habilidades de los estudiantes. En particular, un ítem mostró DIF basado
en el género, lo que favorece a los estudiantes varones en términos de sus
habilidades de respuesta. Además, el estudio identificó que las estudiantes en
el cuarto semestre exhibieron habilidades de respuesta promedio más altas en
comparación con las estudiantes en el sexto y octavo semestre. Además, se
observaron diferencias significativas en las habilidades de respuesta entre
estudiantes varones y mujeres, así como entre estudiantes que residen en áreas
urbanas y rurales. Estos hallazgos son cruciales para los educadores,
enfatizando la necesidad de implementar estrategias de diferenciación
efectivas.
PALABRAS CLAVES· KEYWORDS
Creative problem solving, ICT, education, gender,
psychometric, Rasch measurement
Resolución creativa de
problemas, TIC, educación, género, psicometría, medición Rasch
1. Introductión
In recent years, there has been a growing interest in
evaluating creative problem solving (CPS) skills among students using
information and communication technologies (ICT), as it is recognised as a
crucial skill in the 21st century (Care & Kim,
2018; Hao et al., 2017). The ability to
think creatively and find innovative solutions to complex problems is crucial
in a rapidly changing world that demands adaptability and creative thinking (Suherman &
Vidákovich, 2022). Furthermore, ICT
plays a central role within the DigCompEdu framework,
where technologies are integrated into teaching practices in a pedagogically
meaningful way (Caena & Redecker, 2019). Understanding
and promoting CPS abilities among students is crucial for several reasons.
Firstly, fostering creativity equips people with the capacity to generate
innovative solutions (Lorusso et al., 2021), foster
entrepreneurship (Val et al., 2019), and drive
economic growth (Florida, 2014). Additionally,
CPS is vital in addressing complex societal challenges, such as sustainability
and social inequality (Mitchell & Walinga, 2017).
As an implication of the 21st century era,
numerous nations have acknowledged the necessity of incorporating abilities
such as creatice problem-solving (D. Lee & Lee,
2024), computational
thinking (Küçükaydın et al., 2024), ICT (Rahimi & Oh,
2024), and creativity (Suherman &
Vidákovich, 2024), which are
identified as essential skills for the 21st century. These competencies are
increasingly being integrated into educational curricula to prepare students
for the demands of modern society and the evolving job market (Abina et al.,
2024). As such, the
emphasis on developing these skills reflects a global recognition of their
importance for future success (Yu & Duchin,
2024). The impact of
the 21st century era extends beyond education into the workforce and daily
life. Research by Arredondo-Trapero
et al. (2024) emphasizes that
problem-solving, critical thinking, and ICT are crucial for innovation and
competitiveness in the global of education. Consequently, educational systems
and curricula are under pressure to reform and equip students with these skills
to ensure they are prepared for future challenges.
CPS in an independent curriculum also fosters
collaboration and teamwork. By incorporating CPS into the curriculum, students
are empowered to approach problems with an open mind and explore multiple
perspectives (Burns &
Norris, 2009). They are
encouraged to question assumptions, challenge conventional wisdom, and seek
alternative solutions. This process not only develops your analytical skills,
but also nurtures your creativity and divergent thinking skills (Suherman &
Vidákovich, 2022). An independent
curriculum provides students with the freedom to explore topics of interest and
engage in self-directed learning (Lestari et al.,
2023), collaborative
and teamwork (Zheng et al.,
2024). This approach
aligns with the needs of the 21st-century learner by emphasizing personalized
education paths that cater to individual strengths and preferences (Zhang et al.,
2024). Students are
encouraged to work together, leveraging their diverse perspectives and skills
to address complex challenges (Utami &
Suswanto, 2022). This
collaborative environment enhances their interpersonal and communication
skills, preparing them for future collaborative endeavours. Additionally,
incorporating CPS into the curriculum prepares students for the demands of the
rapidly evolving workforce of the 21st century (Stankovic et al.,
2017). As the world
becomes increasingly complex and interconnected (Brunner et al.,
2024), employers seek
individuals who can think critically (Carnevale &
Smith, 2013), adapt to change
and the work becomes easier (Sousa et al.,
2014), and generate
innovative solutions (Wolcott et al.,
2021).
CPS has gained recognition as a valuable skill set in
educational contexts, including independent curriculum. However, its
implementation can face several challenges that need to be addressed to ensure
its effectiveness and success. Previous research has highlighted the
significance of CPS abilities in various educational contexts (Greiff et al.,
2013; Wang et al., 2023; Wolcott et al., 2021). One challenge is
the lack of teacher training and familiarity with CPS techniques. Studies have
shown that educators can struggle to integrate CPS into their teaching
practices due to limited knowledge and experience in facilitating CPS
activities (van Hooijdonk et al., 2020). This can hinder
the effective implementation of CPS and limit its impact on student learning.
Furthermore, research has explored factors that contribute to the development
of CPS, including the influence of culture (Cho & Lin,
2010), instructional
approaches, and individual characteristics (Samson, 2015). However, there
is limited understanding of the specific factors that develop in CPS abilities
of students.
Regarding assessment, previous studies have explored
alternative approaches to assess CPS skills. Performance-based assessments,
portfolios, and rubrics that assess creativity, critical thinking, metaphorical
thinking, problem-solving related technology motivation, and problem solving
skills have been proposed as more comprehensive and authentic assessment
methods (Abosalem, 2016;
Farida et al., 2022; Liu et al., 2024; Montgomery,
2002; Suastra et al., 2019). These approaches
provide a more holistic view of the students' CPS skills and encourage the
development of higher-order thinking abilities is limited. However, research
also offers potential solutions, such as professional development of teachers,
integration into the curriculum, and alternative assessment methods. By
reviewing relevant literature, we aim to build upon existing knowledge and
identify gaps in understanding, providing a foundation for this study's
contribution to the field.
Despite increasing recognition of the importance of
CPS in independent curriculum education, there is a lack of comprehensive psychometric
validation of CPS assessment instruments. Validating the measurement instrument
is crucial, as CPS remains a poorly defined psychological construct from a
psychometric perspective (Tang et al., 2020). In the absence of valid and
reliable assessments, instructors face challenges in confidently measuring the
CPS learning of students in the classroom. Therefore, this study aims to
validate CPS using the Rasch model by investigating whether the data align with
the measurement of the Rasch model. The reseach
questions are followed:
1.
Does the developed instrument demonstrate reliability
and validity based on the Rasch measurement?
2.
What are the patterns of interaction between items and
persons in the developed instrument based on the Wright map?
3.
Are there any instrument biases based on gender
according to the Differential Item Functioning (DIF) analysis?
4.
How does collaborative problem solving
(CPS) development for students in terms of course grades?
1.1 CPS
CPS is a process that enables people to apply creative
and critical thinking to find solutions to everyday problems (T. Lee et al.,
2023; Van Hooijdonk et al., 2023). CPS helps to
eliminate the tendency to approach problems haphazardly and, as a result,
prevent surprises and/or disappointments with the solutions. Students learn to
work together or individually to find appropriate and unique solutions to
real-world problems they may encounter, using tried and tested methods. Most
importantly, they are challenged to think both creatively and critically as
they face each problem.
CPS can also be influenced by external factors such as
an individual's skill in achieving required goals through a creative process to
find new solutions. The importance of communication in the educational process
means that teachers must also possess various competencies such as personality,
communication, social, lifelong learning, methodology, planning, organisation,
leadership, and assessment, to discern the most significant problems for their
respondents (Suryanto, Degeng, Djatmika & Kuswandi, 2021). Research states
that creative problem solving can provide students with the skills to tackle
everyday problem solving (Abdulla Alabbasi et al., 2021). These skills
require extensive practice involving the creative process, and these activities
are crucial to developing social skills in the field of creativity. Evaluating
ideas and involving multiple people in decision-making with creative thinking
in everyday life - the process of generating new ideas and still discussing
different ways of thinking.
CPS and ICT integration represent crucial
intersections in contemporary education. In educational contexts, ICT serves as
a powerful toolset that not only enhances traditional learning methods, but
also fosters CPS skills among students (Guillén-Gámez et al., 2024; Mäkiö et al.,
2022). Furthermore, ICT
enables personalised learning experiences tailored to the needs of students (Gaeta, Miranda, Orciuoli, Paolozzi & Poce, 2013), empowering them
to develop CPS skills in diverse and engaging ways (Andrews-Todd et
al., 2023; Treffinger, 2007). As educational
paradigms evolve, the integration of CPS and ICT not only prepares students for
the challenges of the modern world but also equips them with essential skills
to thrive in a digitally driven society.
Recent research underscores the importance of
integrating ICT to improve CPS skills among students in educational settings.
According to Selfa-Sastre, Pifarre, Cujba, Cutillas
& Falguera (2022), ICT plays a
crucial role in promoting creativity through collaborative problem solving and
creative expression in language education. Their study highlights how digital
technologies enable diverse learning opportunities and facilitate three key
roles in enhancing collaborative creativity. These roles involve using
interactive technologies to engage students in co-creative language learning
experiences, equipping them with essential competencies to tackle complex
challenges in a globalised and interconnected world. Moreover, Wheeler, Waite,
& Bromfield (2002) emphasise that
ICT enables students to engage in complex problem solving tasks that require
creativity. Their findings suggest that integrating ICT tools into educational
classrooms not only enhances students' technical skills but also cultivates
their ability to think creatively and approach problems from different angles.
These studies collectively underscore the synergy between CPS and ICT in
education, highlighting ICT as a catalyst to nurture creative thinking and problem solving skills essential for 21st-century learners.
Integrating ICT effectively into pedagogical practices not only enriches
educational experiences but also prepares students to thrive in an increasingly
complex and digital world.
Several researchers have developed instruments to
assess CPS ability. For example, the study conducted by Hao et al. (2017) focused on
developing a standardised assessment of CPS skills. Researchers recognised the
importance of CPS in today's collaborative work environments and aimed to
address the practical challenges associated with assessing this complex
construct. The study also highlighted the importance of establishing clear
scoring rubrics and criteria for evaluating CPS performance. In another study, Harding et al.
(2017) focused on
measuring CPS using mathematics-based tasks. The study highlighted the
potential of mathematics-based tasks for assessing CPS skills. They employed
rigorous psychometric analyses to examine the reliability and validity of the
assessment instrument.
These instruments developed by different researchers
provide valuable resources for assessing the CPS capabilities. They offer a
comprehensive approach to measuring various aspects of CPS, including creative
thinking, problem-solving strategies, and collaboration. Using these
instruments, researchers and educators can gain insight into individuals' CPS
abilities and tailor instructional strategies to enhance students' creative
problem-solving skills.
1.2 Rasch measurement
Rasch measurement is a psychometric approach developed
by Georg Rasch in the 1960s (Panayides et al., 2010). It is used to
analyse and interpret data from educational and psychological assessments. The
Rasch model, also known as the Rasch measurement model or the Rasch model for
item response theory (IRT), is a mathematical model that relates the
probability of a response to an item to the ability or trait level of the
individual being assessed (Cappelleri et al., 2014; Rusch et al., 2017).
The Rasch model is based on the principle of
probabilistic measurement, which means that it assesses the probability of a
particular response pattern given the person's ability and the item's
difficulty (Kyngdon, 2008). Individuals with
higher abilities should have a higher probability of answering items correctly,
reflecting easier difficulty levels (Tesio et al.,
2023). In other words,
probabilities are closely related to differences between item difficulty and
individual ability (Boone et al.,
2014). The model
assumes that the probability of a correct response follows a logistic function and that the item's difficulty and the person's
ability can be placed on the same underlying continuum, often referred to as a
logit scale. In Rasch measurement, person abilities and item difficulties are
calibrated on an interval scale called logits, and the item and person
parameters are completely independent (Chan et al.,
2021). This means that
the measurement of student abilities remains the same regardless of the
difficulty level of the items, and item difficulties remain invariant
regardless of student abilities or test takers.
Rasch measurement provides several advantages. It
allows the development of linear measures that are independent of the specific
items used in the assessment (Caty et al.,
2008). This means that
the scores obtained from different sets of items can be compared and
aggregated. The Rasch measurement also provides information about the
reliability of the measurement and the fit of the data to the model, which
helps to assess the quality of the assessment instrument. In educational and
psychological research, the Rasch measurement is commonly used to evaluate the
quality of test items, calibrate item difficulty, estimate person ability, and
conduct item and person analysis. It has applications in various fields,
including educational evaluation, health outcomes research, and social sciences
(Planinic et al.,
2019). By employing the
Rasch model, researchers can gain valuable insight into the relationship
between individuals and items, refine measurement instruments, and make
meaningful inferences about the construct being measured.
2. Methodology
2.1.
Participants
In this cross-sectional study, a total of 137 higher
education students participated as respondents. These students were selected from
the Department of Mathematics Education in Indonesia using a stratified random
sampling technique. The ethical approval is being considered by the
Institutional Review Boards of the Universitas Islam Negeri Raden Intan
Lampung. This sampling technique was chosen to ensure that the sample
population accurately represents the entire population under investigation. The
average age of the participants was 20.84 years, with a standard deviation (SD)
of 1.34. In terms of gender distribution, 51.8% of the respondents were female,
while 48.2% were male. Regarding their residence, the
majority of students (50.4%) came from the city, while the remaining
49.6% came from other areas. The characteristics of the respondents are
presented in Table 1.
Table 1
Characteristics of
the Participants
2.2. Instrument
The instruments used in this study were developed by
researchers and specifically designed to assess the CPS abilities of students.
These instruments were aligned with the local curriculum in higher education to
ensure that they effectively measure the desired skills and competencies. A
total of 20 items were developed for this purpose, which encompass various
aspects of CPS in the use of ICT.
The items in the instruments were carefully designed
to assess the students' ability to apply higher-order thinking skills, critical
and creative thinking, problem solving strategies, and collaboration within the
context of real-world challenges. The items aimed to assess students' capacity
to generate innovative solutions, think critically about complex problems,
effectively communicate and collaborate with others, and demonstrate
adaptability and resilience in problem-solving situations.
Using these instruments, the researchers aimed to
obtain valuable insight into the CPS abilities of students and their ability to
apply these skills in different scenarios. Instruments were developed to
provide reliable and valid measurement of CPS, enabling researchers to gain a
comprehensive understanding of the strengths and areas for improvement in this
domain. The use of these instruments in this study allowed for a systematic and
standardised assessment of CPS, providing valuable data that can contribute to
improving educational practices and curriculum development.
2.3. Procedure
This study involved a one-week data collection period among
higher education students to assess their CPS skills. The CPS test was
administered using Google Forms during regular classroom sessions dedicated to
the respective courses. Students were given access to the test through their
laptop or mobile phone and were given 90 minutes to complete it. In the Google
Forms survey, students were required to provide their demographic information,
including sex, place of residence, ethnicity, and grade level. The CPS test
consisted of multiple-choice items, designed to assess various aspects of CPS
skills. Before starting the test, the researcher provided instructions to the
students and presented three example items to familiarise them with the format
of the question and the characters that would appear in the test. Upon
answering all the questions, the students submitted their responses by clicking
the 'Submit' button, which saved their answers for further analysis.
2.4. Data analysis
The data collected in this study were analysed using
Rasch measurement, a widely used psychometric approach to assess the fit
between the observed data and the underlying measurement model. In this study,
we used Winsteps version 4.7.0 (Linacre, 2020) to analyse the
data, the Rasch model was applied to analyse the responses to the CPS test
items. The model estimates the difficulty of each item and the ability of each
student on a common logit scale. Data obtained from the CPS assessment
instrument were analysed using various Rasch measurement parameters and
techniques. The Outfit mean square (MNSQ) and Outfit z-standardised (ZSTD) were
calculated to assess the fit of each item. The Outfit MNSQ provides an
indication of the extent to which the observed responses deviate from the
expected responses based on the Rasch model, while the Outfit ZSTD standardises
the MNSQ values to facilitate comparison across items. The point-measure
correlation (Pt-measure correlation) was computed to examine the relationship
between the item difficulty and the person's ability. This correlation
coefficient measures the strength of the association between item responses and
estimated person abilities on the logit scale.
The Wright map, a graphical representation, was used
to display the distribution of item difficulties and the corresponding
abilities of the students. This map provides a comprehensive overview of the
item difficulty hierarchy and the range of abilities
exhibited by the students. Additionally, a logit value person (LVP) analysis
was performed to identify the CPS abilities of the students. To explore
possible differences in item functioning based on gender and living of residence,
DIF analysis was performed. DIF analysis identifies items that may function
differently for different groups, indicating potential bias or differential
performance between groups. To analyse the differences in CPS abilities among
students, SPSS version 26 was used. Descriptive statistics such as mean and SD were calculated to provide an overview of the
data. Additionally, R package statistics were employed to see the map.
3. Results
3.1. RQ1: Does the developed
instrument demonstrate reliability and validity based on the Rasch measurement?
The results of the validation analysis using Rasch
analysis are presented in Table 2.
Table 2
The results of the
Rasch analysis conducted on CPS
Characteristics |
Item |
Person |
Number items |
20 |
137 |
infit MNSQ |
|
|
Mean |
1.00 |
1.00 |
SD |
0.18 |
0.14 |
outfit MNSQ |
|
|
Mean |
1.02 |
1.02 |
SD |
0.30 |
0.29 |
Separation |
4.01 |
1.40 |
Reliability |
0.66 |
0.94 |
Raw variance explained by measures |
76.6% |
|
3.2. RQ2: What are the
patterns of interaction between items and persons in the developed instrument
based on the Wright map?
The pattern of interaction between items and
individuals in the developed instrument, based on the Wright map, is presented
in Figure 1. It can be seen in Figure 1 that the instrument consists of 20
items and involves 137 students as respondents. The vertical line on the right
side represents the items, while the left side represents the number of
respondents. It can be noted that item number 13 (Q13) falls into the category
of easy items, whereas item number 12 (Q12) is classified as a difficult item. The
distribution or characteristics of difficult and easy items can be seen in
Figure 2. On the other hand, the distribution of the fit of the items are shown
in Figure 3.
Figure 1
Wright map
Figure 2
An item belongs to
difficult and easier item
Figure 3
The distribution
items are based on the Bubble Map
To determine the fit of the developed items based on
the Rasch model, three criteria were considered: Outfit MNSQ, Outfit ZSTD, and
Pt-Measure Corr. A range between 0.5 and 1.5 for Outfit MNSQ values for both
items and individuals indicates good fit between the data and the model . Outfit ZSTD values between -1.9 and 1.9 imply that
the items can be predicted. Additionally, Pt-Measure Corr is used to determine
if the items measure the intended construct. If the value is positive (+), it
indicates that the item measures the intended construct. Conversely, if the
value is negative (-), the item does not measure the intended construct.
Table 3
Distribution item-based 3
criteria of item fit
Items |
Measure |
Infit MNSQ |
Outfit MNSQ |
Outfit SZTD |
Pt-Measure Cor |
8 |
-0.08 |
1.60 |
1.95 |
8.51 |
-0.27 |
13 |
1.95 |
1.05 |
1.41 |
1.47 |
0.22 |
5 |
0.74 |
1.11 |
1.33 |
2.52 |
0.22 |
4 |
-1.43 |
1.15 |
1.28 |
1.47 |
0.16 |
9 |
0.86 |
1.09 |
1.16 |
1.17 |
0.27 |
16 |
0.90 |
1.16 |
1.08 |
0.63 |
0.24 |
17 |
0.33 |
1.04 |
1.06 |
0.67 |
0.34 |
19 |
-0.12 |
1.05 |
1.01 |
0.19 |
0.35 |
20 |
-0.19 |
1.02 |
0.96 |
-0.41 |
0.38 |
18 |
-0.12 |
1.00 |
0.96 |
-0.43 |
0.40 |
6 |
0.37 |
0.95 |
0.88 |
-1.26 |
0.45 |
14 |
-0.08 |
0.94 |
0.91 |
-1.05 |
0.45 |
1 |
0.19 |
0.91 |
0.91 |
-1.00 |
0.48 |
3 |
-0.64 |
0.91 |
0.86 |
-1.30 |
0.47 |
2 |
-0.53 |
0.89 |
0.90 |
-0.98 |
0.47 |
11 |
1.03 |
0.89 |
0.81 |
-1.35 |
0.49 |
10 |
-0.94 |
0.86 |
0.81 |
-1.48 |
0.50 |
15 |
-1.01 |
0.82 |
0.75 |
-1.90 |
0.54 |
12 |
-1.47 |
0.80 |
0.64 |
-2.13 |
0.54 |
7 |
0.23 |
0.69 |
0.63 |
-4.57 |
0.70 |
Based on the three criteria mentioned, it is evident
that item 8 (Q8) does not meet the above-mentioned criteria, indicating that
the item does not fit well. Therefore, it is recommended to remove or revise
item 8. As shown in Figure 5, Q8 appears to be approaching an underfit,
indicating that it does not align well with the Rasch model.
3.3. RQ3: Are there any
instrument biases based on gender according to the DIF analysis?
The DIF analysis was conducted to determine whether
there were items that favoured one gender (in the context of this study). An
item is considered to have DIF if the t-value is less than -2.0 or greater than
2.0, the DIF contrast value is less than -0.5 or greater than 0.5, and the
p-value is less than 0.05 or greater than -0.05 (Bond & Fox,
2015; Boone et al., 2014). Here are the
results of the analysis using the Rasch model.
Table 4
Potential DIF owing gender
Item |
DIF |
DIF Contrast |
t-value |
Prob. |
|
Female |
Male |
||||
Q12 |
-1.97 |
-1.06 |
-0.91 |
-2.04 |
0.0432 |
The analysis reveals that item Q12 is a difficult
item, indicating that it can differentiate the abilities between males and
females. This is further supported by the DIF analysis, which examines the
item's performance across gender groups. The DIF graph (Fig. 4) provides a
visual representation of the DIF values for each item.
In the graph, the DIF values for item Q12 are
noticeably higher compared to the other items. This suggests that there is a
significant difference in the performance of males and females on this particular item. The DIF analysis indicates that item Q12
may favour one gender over the other in terms of difficulty or discrimination.
These findings are important as they highlight potential gender-related biases
in the measurement of collaborative problem-solving abilities. Further
investigation and potential revision of the item may be necessary to ensure a
fair and unbiased assessment of all individuals, regardless of their gender.
Figure 4
Potential DIF owing gender
3.4. RQ 4: How does students'
CPS develop in terms of course grades?
The statistical description of the students' responses
to the given items is presented in Table 9. In Figure 5 (a), it can be observed
that among female students, those in the fourth semester have an average
response ability (M) of M = 10.19, SD = 3.73, followed by those in the sixth
semester with M = 10.00, SD = 3.32, and those in the eighth semester with M =
9.50, SD = 3.37. On the other hand, among male students, those in the 4th
semester have an average response ability of M = 9.67, SD = 3.11, followed by
those in the 6th semester with M = 5.67, SD = 1.51, and those in the 8th
semester with M = 9.86, SD = 4.02.
However, when comparing students' abilities based on
their place of residence (urban versus rural), there are differences as shown
in Figure 5 (b). Female students residing in urban areas have an average
response ability of M = 9.52, SD = 3.11, while those in rural areas have M =
10.10, SD = 3.59. Similarly, male students residing in urban areas have an
average response ability of M = 9.68, SD = 3.81, while those in rural areas
have M = 9.00, SD = 3.58. These findings indicate that the location of residence
may influence the collaborative problem-solving skills of students to some
extent.
Figure 5
The students'
ability to respond based on grade and gender (a), and grade and place of
residence (b)
4. Discussion
Overall, this analysis provides an understanding of
the measurement characteristics of the items and individuals in this study. The
findings indicate that the measurement used fits reasonably well with the Rasch
measurement model, the ability to differentiate difficulty levels among items,
and a sufficiently high level of reliability. The support of relevant research
in this field also confirms these findings and provides a strong foundation to
understand the measurement characteristics of this study. Previous studies,
such as the one conducted by Soeharto (2021), also found
similar results in terms of fit, separation, and reliability of the
measurement.
The analysis results that indicate a good fit between
items and individuals with the Rasch measurement model serve as an indicator
that the measurement accurately represents the measured characteristics. The
successful separation of difficulty levels among items also provides an
advantage in providing more detailed and accurate information about the
individual abilities measured. This is consistent with previous research
stating that adequate separation is crucial to ensure reliable and valid
measurement (Soeharto & Csapó, 2022).
Additionally, the reasonably good reliability for both
items and individuals provides confidence that the measurement results obtained
are reliable and consistent. In the context of this study, a reliability of
0.66 for items and 0.94 for persons indicates a satisfactory level of
reliability. Other relevant studies can also provide support for the analysis
of the measurement characteristics conducted in this study. For example, a
study conducted by Chan et al. (2021) found similar
results in terms of fit and reliability of the measurement. Furthermore,
research by Avinç & Doğan (2024) found that the
validity and relibality was confirmed by Rasch model.
However, they noted that it would be advantageous to test its validity and
reliability across various classes, age groups, and educational levels.
Additionally, this study similar to Welter et al.
(2024) that the
psychometric properties has valid and reliabel using
Rasch measurement. Overall, this analysis provides a deep understanding of the
measurement characteristics of the items and individuals in this study. The
support of relevant research and the analysis results showing good fit,
separation and reliability provide confidence that the measurement conducted in
this study is reliable and provides valid information on the measured
characteristics.
Based on the observed interaction pattern in Figure 3,
conclusions can be drawn about the difficulty level of the items. For example,
item 13 (Q13) is seen to be positioned lower on the vertical line, indicating
that it belongs to the category of easy items. On the contrary, item 12 (Q12)
is seen to be positioned higher, indicating that it belongs to the category of
difficult items.
This analysis provides important information about the
difficulty level of each element in the developed instrument. When the
difficulty level of the items is known, adjustments and further development can
be made to ensure that the items used cover appropriate levels of difficulty
aligned with the research objectives. However, it should be noted that the
assessment of the difficulty level of the items is not solely based on the
position of the items on the vertical line in Figure 3 but also takes into account other factors such as the characteristics
of the respondents and the deeper context of the measurement. The items were
effective in assessing the CPS abilities of students using ICT in the classroom
(Wheeler et al.,
2002) and impact on
students digital competencies (Guillén-Gámez et al., 2024)This suggests that
the difficulty level of the items was well-suited for the intended purpose of
the instrument (Hobani & Alharbi, 2024). The information
on item difficulty can guide further refinement and development of the
instrument. It allows researchers to identify areas where the difficulty level
may need to be adjusted, either by modifying existing items or adding new items
to cover different difficulty levels.
Moving on to the Differential Item Functioning (DIF)
analysis, DIF refers to differences in response characteristics to an item
between two or more groups of respondents who should have the same level of
ability. In this context, differences in ability between males and females are
explored using the concept of DIF. In this study, it is found that Q12 has the
potential to differentiate the ability between males and females, with a
category of DIF. This indicates that males and females have different probabilities
of answering Q12 despite having the same level of ability. In this context,
there is an indication that Q12 may be more difficult for a gender group.
However, it is important to note that the DIF analysis
only provides preliminary indications of potential differences in response
characteristics between respondent groups. It is crucial to view these DIF
findings as information that can help in instrument development and gain a
better understanding of how the behavioural items in the instrument perform in
specific groups. In other words, “DIF is not a synonym for bias,” as noted by Zieky (2012). Items identified
as DIF do not necessarily imply bias. According to Mollazehi & Abdel-Salam (2024), bias refers to
the differing performance among individuals of equal ability from different
subgroups due to irrelevant factors. DIF, introduced to distinguish the
statistical meaning of bias from its social implications, focusses on the
differing statistical properties that items exhibit among subgroups after
matching individual abilities (Angoff, 2012). Since DIF
interpretation is limited to differences in statistical properties, such as
item difficulty and discrimination, expert panel reviews are necessary to
determine if DIF items are biased (H. Lee &
Geisinger, 2014).Thus, items
showing DIF can be included in a test if no appropriate evidence of bias is
found through panel review (De Ayala et al.,
2002). In further
research, steps can be taken to assess the causes of DIF and ensure that the
instrument measures accurately without gender bias.
The statistical analysis presented in Figure 5
provides information on the ability of students to answer the given items based
on the categories of gender, semester and residential location. The data gives
an overview of the variation in the response abilities between different groups
based on gender, semester, and residential location. Differences in means and
standard deviations indicate variations in understanding of the material or
learning approaches among these groups.
The performance of female students in the fourth
semester, exhibiting an average ability score highest in creative
problem-solving through the use of ICT, appears to
surpass that of male students across different semesters. This observation
aligns with previous research indicating that female students often demonstrate
higher levels of proficiency in problem-solving tasks that require
collaborative and ICT-related skills (Andrews-Todd et
al., 2023; S. W.-Y. Lee et al., 2023). Studies have consistently shown that females
tend to excel in collaborative learning environments, leveraging ICT tools
effectively to enhance their problem-solving capabilities (Ma et al., 2023). This trend is
attributed to various factors, including greater attention to detail, enhanced
communication skills, and a preference for teamwork, which are critical in
creative problem-solving tasks (Thornhill-Miller
et al., 2023).
However, when comparing students living in urban and rural
areas, differences can be observed in the results. Female students living in
rural areas demonstrate the best performance in creative problem-solving through the use of ICT. Rural areas often face unique
challenges such as limited access to resources, including educational
infrastructure and technology (Alabdali et al., 2023). Despite these
challenges, female students in rural areas may exhibit higher problem-solving
abilities due to their adaptability and resilience in navigating these
constraints. Research suggests that females often demonstrate higher levels of
persistence and adaptability in learning environments (Dabas et al.,
2023), which could
contribute to their enhanced performance in creative problem-solving tasks
involving ICT. Furthermore, cultural and societal factors may play a role in
shaping educational outcomes (Min, 2023). In some
cultures, there may be a stronger emphasis on education for females,
particularly in rural settings where access to educational opportunities may be
seen as transformative for individuals and families (Robinson-Pant,
2023). This emphasis
could motivate female students to excel academically and in problem-solving
tasks, including those involving ICT.
These findings highlight the potential influence of
gender, semester, and residence location on the answering abilities of
students. Variations in means and standard deviations suggest differences in
learning experiences, exposure to educational resources, or other contextual
factors that may contribute to variations in response abilities.
Overall, these analyses provide valuable information
on the characteristics of the measurement, the functioning of differential
elements, and the relationship between answering abilities and various factors
such as gender, semester and residential location. They contribute to a better
understanding of the data and offer implications for future research and
instrument development in this field.
5. Limitation and
Future Research
This study provides important contributions to the
development of evaluation instruments to evaluate CPS students. However, there
are some limitations that need to be addressed. Firstly, the measurement
reliability for the items obtained a value of 0.66, indicating a moderate level
of reliability. While this reliability may be acceptable in some research
contexts, improving reliability is desirable for the development of more robust
evaluation instruments in the future.
Additionally, there is an item, item 8 (Q8), that does
not meet the criteria in the item fit analysis with the Rasch model. This item
could be removed or revised to ensure a better fit and validity of the
evaluation instrument. Revision and refinement of items that do not meet the
criteria is necessary to ensure that the evaluation instrument produces more
accurate and reliable results.
Furthermore, a potential gender-based differential
item functioning (DIF) is observed in item Q12. This indicates instrument bias
toward gender in terms of answering difficult questions. In the development of
future evaluation instruments, it is important to address this bias to ensure a
more neutral and fair instrument for all participants.
The study also provides information on the interaction
patterns between items and individuals in the developed instrument based on the
Wright map. By examining these patterns, the difficulty level of each item and
the distribution of respondents' abilities can be understood. However, no
further explanation of the implications of these interaction patterns in the
development of evaluation instruments is provided.
In addition to these limitations, this study lays a
strong foundation for future research in the development of evaluation
instruments based on higher-order thinking skills (HOTS). Future research can
focus on improving the reliability of the instrument, eliminating gender-based
instrument bias, and further exploring the patterns of patterns of interaction
between items and individuals.
In future studies, it is important to involve a more
representative sample and expand the scope of the analysis to obtain more
generalisable results. Additionally, instrument validation can also be
conducted using other methods that provide additional information on instrument
fit, validity, and reliability.
Overall, this study contributes to the development of
HOTS-based evaluation instruments using the Rasch model. Despite some
limitations that need to be addressed, this study serves as an important
foundation for further research in the development of more effective and robust
evaluation instruments.
6. Conclusions
On the basis of the analysis, the following
conclusions can be drawn: Measurement using the Rasch
model demonstrates a good fit between items and individuals in the evaluation
instrument. The items exhibit a good fit with the Rasch model, allowing for
differentiation of difficulty levels among different items, and they also have
a reasonably good level of reliability. There is an interaction pattern between
items and individuals in the evaluation instrument based on the Wright map. The
items in the instrument can effectively differentiate the abilities of
individuals, with some items being relatively easy and others more challenging.
There are items that do not meet the DIF criteria based on gender. Item Q12 in
the evaluation instrument tends to favour males over females in terms of the
ability to answer. Female students in the fourth semester have higher average
answering abilities compared to female students in the sixth and eighth
semesters. However, there are differences in answering abilities between male
and female students, as well as between students living in urban and rural
areas.
These conclusions indicate that the development of CPS
evaluation instruments can yield valid and reliable measurement results.
However, it should be noted that there are some items that need to be improved
for greater accuracy. Additionally, there is an indication of instrument bias
based on gender in terms of answering abilities. This should be considered when
developing instruments that are more gender-neutral and fair in measuring
participants' abilities.
In the development of CPS evaluation instruments and
the use of the Rasch model, there are positive impacts on curriculum and
instruction development. It is important for educators to adopt effective
differentiation approaches, ensure gender-neutral evaluation instruments, and
consider contextual factors in designing inclusive and learner-centred
instruction that aligns with the potential of students. Therefore, education
can become more relevant, responsive, and enable learners to face the
challenges of a complex world.
Authors’
Contribution
Farida Farida:
Conceptualization, Writing - Original Draft, Formal analysis; Yosep Aspat Alamsyah: Methodology,
Editing, and Visualization; Bambang Sri Anggoro:
Supervision, Funding acquisition, Writing – review & editing; Tri Andari:
Formal analysis and Visualization; Restu Lusiana:
Editing and Formal analysis.
Funding Agency
The Department of Research and Community Service
(LP2M) funded the study reported at Universitas Islam Negeri Raden Intan
Lampung, Indonesia (Grant number 99/2023).
References
Abdulla Alabbasi, A. M., Hafsyan, A. S., Runco, M. A., & AlSaleh, A.
(2021). Problem finding, divergent thinking, and evaluative
thinking among gifted and nongifted students. Journal for the Education of
the Gifted, 44(4), 398–413. https://doi.org/10.1515/ctra-2018-0019
Abina, A., Temeljotov Salaj,
A., Cestnik, B., Karalič,
A., Ogrinc, M., Kovačič Lukman, R., & Zidanšek, A. (2024).
Challenging 21st-Century Competencies for STEM Students: Companies’ Vision in
Slovenia and Norway in the Light of Global Initiatives for Competencies
Development. Sustainability, 16(3), 1295.
https://doi.org/10.3390/su16031295
Abosalem, Y. (2016).
Assessment techniques and students’ higher-order thinking skills. International
Journal of Secondary Education, 4(1), 1–11.
https://doi.org/10.11648/j.ijsedu.20160401.11
Alabdali, S. A., Pileggi,
S. F., & Cetindamar, D. (2023). Influential
factors, enablers, and barriers to adopting smart technology in rural regions:
A literature review. Sustainability, 15(10), 7908.
https://doi.org/10.3390/su15107908
Andrews-Todd, J., Jiang, Y., Steinberg, J., Pugh, S.
L., & D’Mello, S. K. (2023). Investigating collaborative problem
solving skills and outcomes across computer-based tasks. Computers
& Education, 207, 104928.
https://doi.org/10.1016/j.compedu.2023.104928
Angoff, W. H. (2012). Perspectives on differential
item functioning methodology. In Differential item functioning (pp.
3–23). Routledge.
https://www.taylorfrancis.com/chapters/edit/10.4324/9780203357811-2/perspectives-differential-item-functioning-methodology-william-angoff
Arredondo-Trapero, F. G.,
Guerra-Leal, E. M., Kim, J., & Vázquez-Parra, J. C. (2024). Competitiveness,
quality education and universities: The shift to the post-pandemic world. Journal
of Applied Research in Higher Education.
https://doi.org/10.1108/JARHE-08-2023-0376
Avinç, E., & Doğan, F. (2024). Digital
literacy scale: Validity and reliability study with the rasch
model. Education and Information Technologies.
https://doi.org/10.1007/s10639-024-12662-7
Bond, T., & Fox, C. M. (2015). Applying the
Rasch Model: Fundamental Measurement in the Human Sciences, Third Edition
(3rd ed.). Routledge. https://doi.org/10.4324/9781315814698
Boone, W. J., Staver, J. R., & Yale, M. S. (2014).
Rasch analysis in the human sciences. Springer.
https://doi.org/10.1007/978-94-007-6857-4
Brunner, L. G., Peer, R. A. M., Zorn, C., Paulik, R.,
& Logan, T. M. (2024). Understanding cascading risks through real-world
interdependent urban infrastructure. Reliability Engineering & System
Safety, 241, 109653. https://doi.org/10.1016/j.ress.2023.109653
Burns, D. P., & Norris, S. P. (2009). Open-minded
environmental education in the science classroom. Paideusis,
18(1), 36–43. https://doi.org/10.7202/1072337ar
Caena, F., & Redecker, C. (2019). Aligning
teacher competence frameworks to 21st century challenges: The case for the
European Digital Competence Framework for Educators (Digcompedu).
European Journal of Education, 54(3), 356–369.
https://doi.org/10.1111/ejed.12345
Cappelleri, J. C., Lundy, J.
J., & Hays, R. D. (2014). Overview of classical test theory and item
response theory for the quantitative assessment of items in developing
patient-reported outcomes measures. Clinical Therapeutics, 36(5),
648–662. https://doi.org/10.1016/j.clinthera.2014.04.006
Care, E., & Kim, H. (2018). Assessment of
twenty-first century skills: The issue of authenticity. Assessment and
Teaching of 21st Century Skills: Research and Applications, 21–39.
https://doi.org/10.1007/978-3-319-65368-6_2
Carnevale, A. P., & Smith, N. (2013). Workplace
basics: The skills employees need and employers want.
In Human Resource Development International (Vol. 16, Issue 5, pp.
491–501). Taylor & Francis. https://doi.org/10.1080/13678868.2013.821267
Caty, G. D., Arnould, C., Stoquart,
G. G., Thonnard, J.-L., & Lejeune, T. M. (2008). ABILOCO: a Rasch-built
13-item questionnaire to assess locomotion ability in stroke patients. Archives
of Physical Medicine and Rehabilitation, 89(2), 284–290.
https://doi.org/10.1016/j.apmr.2007.08.155
Chan, S. W., Looi, C. K., & Sumintono, B.
(2021). Assessing computational thinking abilities among
Singapore secondary students: A rasch model
measurement analysis. Journal of Computers in Education, 8(2),
213–236. https://doi.org/10.1007/s40692-020-00177-2
Cho, S., & Lin, C.-Y. (2010). Influence of family
processes, motivation, and beliefs about intelligence on creative problem
solving of scientifically talented individuals. Roeper Review, 33(1),
46–58. https://doi.org/10.1080/02783193.2011.530206
Dabas, C. S., Muljana, P. S., & Luo, T. (2023).
Female students in quantitative courses: An exploration of their motivational
sources, learning strategies, learning behaviors, and course achievement. Technology,
Knowledge and Learning, 28(3), 1033–1061.
https://doi.org/10.1007/s10758-021-09552-z
De Ayala, R. J., Kim, S.-H., Stapleton, L. M., &
Dayton, C. M. (2002). Differential Item Functioning: A Mixture Distribution
Conceptualization. International Journal of Testing, 2(3–4),
243–276. https://doi.org/10.1080/15305058.2002.9669495
Farida, F., Supriadi,
N., Andriani, S., Pratiwi,
D. D., Suherman, S., & Muhammad, R. R. (2022). STEM approach and
computer science impact the metaphorical thinking of Indonesian students’. Revista
de Educación a Distancia (RED), 22(69).
https://doi.org/10.6018/red.493721
Florida, R. (2014). The creative class and economic
development. Economic Development Quarterly, 28(3), 196–205.
https://doi.org/10.1177/0891242414541693
Gaeta, M., Miranda, S., Orciuoli, F., Paolozzi, S., &
Poce, A. (2013). An Approach To Personalized e-Learning. Journal of Education,
Informatics & Cybernetics, 11(1).
https://www.iiisci.org/Journal/pdv/sci/pdfs/HEB785ZU.pdf
Greiff, S., Wüstenberg, S.,
Molnár, G., Fischer, A., Funke, J., & Csapó, B. (2013). Complex problem
solving in educational contexts—Something beyond g: Concept, assessment,
measurement invariance, and construct validity. Journal of Educational
Psychology, 105(2), 364. https://doi.org/10.1037/a0031856
Guillén-Gámez, F. D.,
Gómez-García, M., & Ruiz-Palmero, J. (2024). Digital competence
in research work: Predictors that have an impact on it according to the type of
university and gender of the Higher Education teacher:[Digital
competence in research work: predictors that have an impact on it according to
the type of university and gender of the Higher Education teacher]. Pixel-Bit. Revista de Medios y Educación, 69, 7–34.
https://doi.org/10.12795/pixelbit.99992
Hao, J., Liu, L., von Davier, A. A., & Kyllonen, P. C. (2017). Initial steps
towards a standardized assessment for collaborative problem solving (CPS):
Practical challenges and strategies. Innovative Assessment of Collaboration,
135–156. https://doi.org/10.1007/978-3-319-33261-1_9
Harding, S.-M. E., Griffin, P. E., Awwal, N., Alom, B.
M., & Scoular, C. (2017). Measuring collaborative problem
solving using mathematics-based tasks. AERA Open, 3(3),
2332858417728046. https://doi.org/10.1177/2332858417728046
Hobani, F., & Alharbi, M. (2024). A Psychometric
Study of the Arabic Version of the “Searching for Hardships and Obstacles to
Shots (SHOT)” Instrument for Use in Saudi Arabia. Vaccines, 12(4), 391.
https://doi.org/10.3390/vaccines12040391
Küçükaydın, M. A., Çite,
H., & Ulum, H. (2024). Modelling the
relationships between STEM learning attitude, computational thinking, and 21st
century skills in primary school. Education and Information Technologies.
https://doi.org/10.1007/s10639-024-12492-7
Kyngdon, A. (2008). The
Rasch model from the perspective of the representational theory of measurement.
Theory & Psychology, 18(1), 89–109.
https://doi.org/10.1177/095935430708692
Lee, D., & Lee, Y. (2024). Productive failure-based
programming course to develop computational thinking and creative
Problem-Solving skills in a Korean elementary school. Informatics in
Education, 23(2), 385–408.
https://doi.org/10.1007/s10956-024-10130-y
Lee, H., & Geisinger, K. F. (2014). The Effect of
Propensity Scores on DIF Analysis: Inference on the Potential Cause of DIF. International
Journal of Testing, 14(4), 313–338.
https://doi.org/10.1080/15305058.2014.922567
Lee, S. W.-Y., Liang, J.-C., Hsu, C.-Y., Chien, F. P.,
& Tsai, M.-J. (2023). Exploring Potential Factors to Students’
Computational Thinking. Educational Technology & Society, 26(3),
176–189. https://doi.org/10.30191/ETS
Lee, T., O’Mahony, L., & Lebeck, P. (2023).
Creative Problem-Solving. In T. Lee, L. O’Mahony, & P. Lebeck, Creativity
and Innovation (pp. 117–147). Springer Nature Singapore.
https://doi.org/10.1007/978-981-19-8880-6_5
Lestari, W., Sari, M. M., Istyadji,
M., & Fahmi, F. (2023). Analysis of Implementation of the Independent
Curriculum in Science Learning at SMP Negeri 1 Tanah Grogot
Kalimantan Timur, Indonesia. Repository Universitas Lambung
Mangkurat.
https://doi.org/10.36348/jaep.2023.v07i06.001
Linacre, J. M. (2020). Winsteps®(Version 4.7. 0)[Computer Software].(4.7. 0). Winsteps. Com.
Liu, X., Gu, J., & Xu, J. (2024). The impact of
the design thinking model on pre-service teachers’ creativity self-efficacy,
inventive problem-solving skills, and technology-related motivation. International
Journal of Technology and Design Education, 34(1), 167–190.
https://doi.org/10.1007/s10798-023-09809-x
Lorusso, L., Lee, J. H.,
& Worden, E. A. (2021). Design thinking for healthcare: Transliterating the
creative problem-solving method into architectural practice. HERD: Health
Environments Research & Design Journal, 14(2), 16–29.
https://doi.org/10.1177/193758672199422
Ma, Y., Zhang, H., Ni, L., & Zhou, D. (2023).
Identifying collaborative problem-solver profiles based on collaborative
processing time, actions and skills on a computer-based task. International
Journal of Computer-Supported Collaborative Learning, 18(4),
465–488. https://doi.org/10.1007/s11412-023-09400-5
Mäkiö, E., Azmat, F., Ahmad, B., Harrison, R., &
Colombo, A. W. (2022). T-CHAT educational framework for teaching cyber-physical
system engineering. European Journal of Engineering Education, 47(4),
606–635. https://doi.org/10.1080/03043797.2021.2008879
Min, M. (2023). School culture, self-efficacy, outcome
expectation, and teacher agency toward reform with curricular autonomy in South
Korea: A social cognitive approach. Asia Pacific Journal of Education, 43(4),
951–967. https://doi.org/10.1080/02188791.2019.1626218
Mitchell, I. K., & Walinga,
J. (2017). The creative imperative: The role of creativity, creative problem
solving and insight as key drivers for sustainability. Journal of Cleaner
Production, 140, 1872–1884.
https://doi.org/10.1016/j.jclepro.2016.09.162
Mollazehi, M., &
Abdel-Salam, A.-S. G. (2024). Understanding the alternative Mantel-Haenszel
statistic: Factors affecting its robustness to detect non-uniform DIF. Communications
in Statistics - Theory and Methods, 1–25.
https://doi.org/10.1080/03610926.2024.2330668
Montgomery, K. (2002). Authentic tasks and rubrics:
Going beyond traditional assessments in college teaching. College Teaching,
50(1), 34–40. https://doi.org/10.1080/87567550209595870
Panayides, P., Robinson,
C., & Tymms, P. (2010). The assessment revolution
that has passed England by: Rasch measurement. British Educational Research
Journal, 36(4), 611–626. https://doi.org/10.1080/01411920903018182
Planinic, M., Boone, W. J., Susac,
A., & Ivanjek, L. (2019). Rasch analysis in physics education research: Why
measurement matters. Physical Review Physics Education Research, 15(2),
020111. https://doi.org/10.1103/PhysRevPhysEducRes.15.020111
Rahimi, R. A., & Oh, G. S. (2024). Rethinking the
role of educators in the 21st century: Navigating globalization, technology,
and pandemics. Journal of Marketing Analytics, 1–16.
https://doi.org/10.1057/s41270-024-00303-4
Robinson-Pant, A. (2023). Education for rural
development: Forty years on. International Journal of Educational
Development, 96, 102702.
https://doi.org/10.1016/j.ijedudev.2022.102702
Rusch, T., Lowry, P. B., Mair, P., & Treiblmaier, H. (2017). Breaking free from the limitations
of classical test theory: Developing and measuring information systems scales
using item response theory. Information & Management, 54(2),
189–203. https://doi.org/10.1016/j.im.2016.06.005
Samson, P. L. (2015). Fostering student engagement:
Creative problem-solving in small group facilitations. Collected Essays on
Learning and Teaching, 8, 153–164.
https://doi.org/10.22329/celt.v8i0.4227
Selfa-Sastre, M., Pifarre,
M., Cujba, A., Cutillas,
L., & Falguera, E. (2022). The role of
digital technologies to promote collaborative creativity in language education.
Frontiers in Psychology, 13, 828981.
https://doi.org/10.3389/fpsyg.2022.828981
Soeharto, S. (2021). Development
of a Diagnostic Assessment Test to Evaluate Science Misconceptions in Terms of
School Grades: A Rasch Measurement Approach. Journal of Turkish Science
Education, 18(3), 351–370. https://doi.org/10.36681/tused.2021.78
Soeharto, S., & Csapó,
B. (2022). Assessing Indonesian student inductive reasoning: Rasch analysis. Thinking
Skills and Creativity, 101132. https://doi.org/10.1016/j.tsc.2022.101132
Stankovic, J. A., Sturges, J. W., & Eisenberg, J.
(2017). A 21st century cyber-physical systems education. Computer, 50(12),
82–85. https://doi.org/10.1109/MC.2017.4451222
Suastra, I. W., Ristiati, N. P., Adnyana, P. P.
B., & Kanca, N. (2019). The effectiveness of
Problem Based Learning-Physics module with authentic assessment for enhancing
senior high school students’ physics problem solving ability and critical
thinking ability. Journal of Physics: Conference Series, 1171(1),
012027. https://doi.org/10.1088/1742-6596/1171/1/012027
Suherman, S., & Vidákovich, T. (2022). Assessment
of Mathematical Creative Thinking: A Systematic Review. Thinking Skills and
Creativity, 101019. https://doi.org/10.1016/j.tsc.2022.101019
Suherman, S., & Vidákovich, T. (2024).
Relationship between ethnic identity, attitude, and mathematical creative
thinking among secondary school students. Thinking Skills and Creativity,
51, 101448. https://doi.org/10.1016/j.tsc.2023.101448
Suryanto, H., Degeng, I.
N. S., Djatmika, E. T., & Kuswandi,
D. (2021). The effect of creative problem solving with the intervention social
skills on the performance of creative tasks. Creativity Studies, 14(2),
323–335. https://doi.org/10.3846/cs.2021.12364
Tesio, L., Caronni, A.,
Kumbhare, D., & Scarano, S. (2023). Interpreting results from Rasch
analysis 1. The “most likely” measures coming from the model. Disability and
Rehabilitation, 1–13. https://doi.org/10.1080/09638288.2023.2169771
Thornhill-Miller, B., Camarda, A., Mercier, M.,
Burkhardt, J.-M., Morisseau, T., Bourgeois-Bougrine,
S., Vinchon, F., El Hayek, S., Augereau-Landais,
M., & Mourey, F. (2023). Creativity, critical thinking, communication, and
collaboration: Assessment, certification, and promotion of 21st century skills
for the future of work and education. Journal of Intelligence, 11(3),
54. https://doi.org/10.3390/jintelligence11030054
Treffinger, D. J. (2007). Creative Problem Solving
(CPS): Powerful Tools for Managing Change and Developing Talent. Gifted and
Talented International, 22(2), 8–18.
https://doi.org/10.1080/15332276.2007.11673491
Utami, Y. P., & Suswanto, B. (2022). The
Educational Curriculum Reform in Indonesia: Supporting “Independent Learning
Independent Campus (MBKM)”. SHS Web of Conferences, 149.
https://doi.org/10.1051/shsconf/202214901041
Val, E., Gonzalez, I., Lauroba,
N., & Beitia, A. (2019). How can Design Thinking promote entrepreneurship
in young people? The Design Journal, 22(sup1), 111–121.
https://doi.org/10.1080/14606925.2019.1595853
van Hooijdonk, M., Mainhard, T., Kroesbergen, E. H.,
& van Tartwijk, J. (2020). Creative problem
solving in primary education: Exploring the role of fact finding, problem
finding, and solution finding across tasks. Thinking Skills and Creativity,
37, 100665. https://doi.org/10.1016/j.tsc.2020.100665
Van Hooijdonk, M., Mainhard, T., Kroesbergen, E. H.,
& Van Tartwijk, J. (2023). Creative problem
solving in primary school students. Learning and Instruction, 88,
101823. https://doi.org/10.1016/j.learninstruc.2023.101823
Wang, M., Yu, R., & Hu, J. (2023). The
relationship between social media-related factors and student collaborative
problem-solving achievement: An HLM analysis of 37 countries. Education and
Information Technologies, 1–19. https://doi.org/10.1007/s10639-023-11763-z
Welter, V. D. E., Dawborn-Gundlach,
M., Großmann, L., & Krell, M. (2024). Adapting a
self-efficacy scale to the task of teaching scientific reasoning: Collecting
evidence for its psychometric quality using Rasch measurement. Frontiers in
Psychology, 15, 1339615. https://doi.org/10.3389/fpsyg.2024.1339615
Wheeler, S., Waite, S. J., & Bromfield, C. (2002).
Promoting creative thinking through the use of ICT. Journal
of Computer Assisted Learning, 18(3), 367–378.
https://doi.org/10.1046/j.0266-4909.2002.00247.x
Wolcott, M. D., McLaughlin, J. E., Hubbard, D. K.,
Rider, T. R., & Umstead, K. (2021). Twelve tips to stimulate creative
problem-solving with design thinking. Medical Teacher, 43(5),
501–508. https://doi.org/10.1080/0142159X.2020.1807483
Yu, Y., & Duchin, F. (2024). Building a Curriculum
to Foster Global Competence and Promote the Public Interest: Social
Entrepreneurship and Digital Skills for American Community College Students. Community
College Journal of Research and Practice, 48(3), 164–174.
https://doi.org/10.1080/10668926.2022.2064374
Zhang, L., Carter, R. A., Bloom, L., Kennett, D. W.,
Hoekstra, N. J., Goldman, S. R., & Rujimora, J.
(2024). Are Pre-Service Educators Prepared to Implement Personalized Learning?: An Alignment Analysis of Educator Preparation
Standards. Journal of Teacher Education, 75(2), 219–235.
https://doi.org/10.1177/00224871231201367
Zheng, J., Cheung, K., & Sit, P. (2024). The
effects of perceptions toward Interpersonal relationships on collaborative
problem-solving competence: Comparing four ethnic Chinese communities assessed
in PISA 2015. The Asia-Pacific Education Researcher, 33(2),
481–493. https://doi.org/10.1007/s40299-023-00744-y
Zieky, M. (2012). Practical questions in the use of
DIF statistics in test development. In Differential item functioning
(pp. 337–347). Routledge.