Assessing The Effect Of The Multidimensionality Based Dif Analysis Paradigm On The Common Item Nonequivalent Group Equating Design Using Translated Tests

DOWNLOAD
Download Assessing The Effect Of The Multidimensionality Based Dif Analysis Paradigm On The Common Item Nonequivalent Group Equating Design Using Translated Tests PDF/ePub or read online books in Mobi eBooks. Click Download or Read Online button to get Assessing The Effect Of The Multidimensionality Based Dif Analysis Paradigm On The Common Item Nonequivalent Group Equating Design Using Translated Tests book now. This website allows unlimited access to, at the time of writing, more than 1.5 million titles, including hundreds of thousands of titles in various foreign languages. If the content not found or just blank you must refresh this page
Dissertation Abstracts International
DOWNLOAD
Author :
language : en
Publisher:
Release Date : 2006
Dissertation Abstracts International written by and has been published by this book supported file pdf, txt, epub, kindle and other format this book has been release on 2006 with Dissertations, Academic categories.
Assessing The Effect Of The Multidimensionality Based Dif Analysis Paradigm On The Common Item Nonequivalent Group Equating Design Using Translated Tests
DOWNLOAD
Author : Yuen Yee Li
language : en
Publisher:
Release Date : 2005
Assessing The Effect Of The Multidimensionality Based Dif Analysis Paradigm On The Common Item Nonequivalent Group Equating Design Using Translated Tests written by Yuen Yee Li and has been published by this book supported file pdf, txt, epub, kindle and other format this book has been release on 2005 with Achievement tests categories.
The Impact Of Multidimensionality On The Detection Of Differential Bundle Functioning Using Sibtest
DOWNLOAD
Author : Terris Raiford-Ross
language : en
Publisher:
Release Date : 2007
The Impact Of Multidimensionality On The Detection Of Differential Bundle Functioning Using Sibtest written by Terris Raiford-Ross and has been published by this book supported file pdf, txt, epub, kindle and other format this book has been release on 2007 with Educational tests and measurements categories.
In response to public concern over fairness in testing, conducting a differential item functioning (DIF) analysis is now standard practice for many large-scale testing programs (e.g., Scholastic Aptitude Test, intelligence tests, licensing exams). As highlighted by the Standards for Educational and Psychological Testing manual, the legal and ethical need to avoid bias when measuring examinee abilities is essential to fair testing practices (AERA-APA-NCME, 1999). Likewise, the development of statistical and substantive methods of investigating DIF is crucial to the goal of designing fair and valid educational and psychological tests. Douglas, Roussos and Stout (1996) introduced the concept of item bundle DIF and the implications of differential bundle functioning (DBF) for identifying the underlying causes of DIF. Since then, several studies have demonstrated DIF/DBF analyses within the framework of "unintended" multidimensionality (Oshima & Miller, 1992; Russell, 2005). Russell (2005), in particular, examined the effect of secondary traits on DBF/DTF detection. Like Russell, this study created item bundles by including multidimensional items on a simulated test designed in theory to be unidimensional. Simulating reference group members to have a higher mean ability than the focal group on the nuisance secondary dimension, resulted in DIF for each of the multidimensional items, that when examined together produced differential bundle functioning. The purpose of this Monte Carlo simulation study was to assess the Type I error and power performance of SIBTEST (Simultaneous Item Bias Test; Shealy & Stout, 1993a) for DBF analysis under various conditions with simulated data. The variables of interest included sample size and ratios of reference to focal group sample sizes, correlation between primary and secondary dimensions, magnitude of DIF/DBF, and angular item direction. Results showed SIBTEST to be quite powerful in detecting DBF and controlling Type I error for almost all of the simulated conditions. Specifically, power rates were .80 or above for 84% of all conditions and the average Type I error rate was approximately .05. Furthermore, the combined effect of the studied variables on SIBTEST power and Type I error rates provided much needed information to guide further use of SIBTEST for identifying potential sources of differential item/bundle functioning.
The Consequences Of Multidimensionality To Irt Equating Outcomes Using A Common Items Nonequivalent Groups Design
DOWNLOAD
Author : Kathryn Louise Ricker
language : en
Publisher:
Release Date : 2007
The Consequences Of Multidimensionality To Irt Equating Outcomes Using A Common Items Nonequivalent Groups Design written by Kathryn Louise Ricker and has been published by this book supported file pdf, txt, epub, kindle and other format this book has been release on 2007 with Educational tests and measurements categories.
A New Dimensionality Estimation Tool For Multiple Item Tests And A New Dif Analysis Paradigm Based On Multidimensionality And Construct Validity
DOWNLOAD
Author : Louis Ablen Roussos
language : en
Publisher:
Release Date : 1995
A New Dimensionality Estimation Tool For Multiple Item Tests And A New Dif Analysis Paradigm Based On Multidimensionality And Construct Validity written by Louis Ablen Roussos and has been published by this book supported file pdf, txt, epub, kindle and other format this book has been release on 1995 with Examinations categories.
This thesis is concerned with two critical issues facing the testing industry today: dimensionality analysis and DIF (Differential Item Functioning) analysis. Chapter 1 develops the use of new dimensionally-sensitive proximity measures with Hierarchical Cluster Analysis and DIMTEST to estimate the dimensionality structure of tests. The results of simulation studies and real data analyses indicate that the new tool represents a significant step forward in the ability of dimensionality assessment tools to identify reliably the latent dimensionality structure of a set of items. Chapter 2 of the thesis develops a new DIF analysis paradigm that unifies the substantive and statistical DIF research camps by linking both camps to a theoretically sound and mathematically rigorous multidimensional conceptualization of DIF. The new paradigm is shown to have the potential to improve the understanding of the root causes of DIF through the testing of substantive DIF hypotheses, to reduce Type 1 error through a better understanding of the dimensionality of the matching criterion, and to increase power through the testing of bundles of items with similar content. Two new paradigm-based DIF analysis methods, one of which employed the new dimensionality estimation tool of Chapter 1, were described and applied to real data. The analyses demonstrated that the new paradigm-based methods offered insights that cannot be obtained from the standard one-item-at-a-time DIF analysis.
Assessing Differential Bundle Functioning Using Meta Analysis
DOWNLOAD
Author : Lanrong Li
language : en
Publisher:
Release Date : 2021
Assessing Differential Bundle Functioning Using Meta Analysis written by Lanrong Li and has been published by this book supported file pdf, txt, epub, kindle and other format this book has been release on 2021 with Educational tests and measurements categories.
When developing a test, it is essential to ensure that the test is free of items with differential item functioning (DIF). DIF occurs when examinees of equal ability, but from different examinee subgroups, have different chances of getting the item correct. According to the multidimensional perspective, DIF occurs because the test measures more than one dimension, and examinees from different groups have unequal distributions on the secondary dimension(s) conditional on the primary dimension measured by the test. Often, more than one item is measured by the secondary dimension. The DIF of individual items may be hard to detect statistically but can accumulate to an unacceptable degree at the item cluster/bundle level. Differential bundle functioning (or DBF) occurs when examinees from different groups have unequal expected scores on an item bundle. Research on DBF has the potential to reveal the mechanisms underlying DIF and DBF. The simultaneous item bias test (SIBTEST, Shealy and Stout, 1993) has been developed to assess DIF and DBF. However, an unresolved issue is the lack of effect size for DBF, making it difficult to assess the amount of DBF within and between tests. Additionally, few procedures can be used to assess both DIF and DBF, which may be one of the reasons why DBF is less frequently examined than DIF among practitioners. In this study, I propose using meta-analysis techniques to synthesize differential item functioning (DIF) effect sizes and to assess differential bundle functioning (DBF). The test of nonzero average DIF can be used to test for the existence of DBF. Also, the weighted average DIF can be used as an average-based effect size, and the standard deviation of DIF in an item bundle can be used as a variance-based DBF effect size. A Monte Carlo simulation study was conducted to assess the performance of the proposed effect size for DBF and to compare the test of nonzero average DIF in an item bundle with that of a DBF test using SIBTEST. I used three DIF procedures (id est, the Mantel-Haenszel procedure, the logistic-regression procedure, and the SIBTEST procedure) to obtain DIF estimates and then applied the random-effects model to the DIF estimates. Seven factors, including sample size, between-group difference in the primary dimension, between-group difference in the secondary dimension, the correlation between the two dimensions, sample-size ratio between the focal and the reference groups, item-bundle length, and the presence of guessing in the item response, were manipulated in the simulation. When the two dimensions were moderately correlated or when there was no impact, the proposed DBF effect size based on the SIBTEST DIF procedure was essentially unbiased, and the proposed DBF tests based on the three DIF procedures had Type-I error and power rates comparable to those of the SIBTEST DBF test. When there was an impact and the two dimensions were not correlated or weakly correlated, the DBF effect sizes from the meta-analysis of DIF indices from the three procedures had upward biases, and the meta-analysis-based DBF tests tended to have inflated Type-I error. The variance-based DBF effect size for the whole test changed with the weighted average-based DBF effect size in the item bundle. The rejection rates of the DBF test, the weighted average-based DBF effect size, and the variance-based DBF effect size were largely determined by the potential for DIF. I discuss the findings and the implications for applied research and point out directions for future research.
A Comparison Of The Common Item And Random Groups Equating Designs Using Empirical Data
DOWNLOAD
Author : Dong-In Kim
language : en
Publisher:
Release Date : 2008
A Comparison Of The Common Item And Random Groups Equating Designs Using Empirical Data written by Dong-In Kim and has been published by this book supported file pdf, txt, epub, kindle and other format this book has been release on 2008 with categories.
We designed this study to evaluate several data collection and equating designs in the context of item response theory (IRT) equating. The random-groups design and the common-item design have been widely used for collecting data for IRT equating. In this study, we investigated four equating methods based upon these two data collection designs, using empirical data from a number of different testing programs. When the randomly equivalent group assumption was reasonably met, the four equating methods tended to produce highly comparable results. On the other hand, equating methods based upon either of the equating designs produced dissimilar results. Sample size can have differential effects on the equating results produced by the different equating methods. In practice, a common-item equivalent-groups design often produces unacceptably large differences in the group mean due to various anomalies such as context effects, poor quality of common items, or a very small number of common items. In such cases, a random-groups design would produce more stable equating results.
Equating Using Unidimensional Dichotomous And Polytomous Irt Models For Testlet Based Tests Under Common Item Nonequivalent Groups Design
DOWNLOAD
Author : Lidong Zhang
language : en
Publisher:
Release Date : 2013
Equating Using Unidimensional Dichotomous And Polytomous Irt Models For Testlet Based Tests Under Common Item Nonequivalent Groups Design written by Lidong Zhang and has been published by this book supported file pdf, txt, epub, kindle and other format this book has been release on 2013 with categories.
The relative equating performance of the Graded Response Model (GRM) and the Generalized Partial Credit (GPC) model was compared with that of the two parameter logistic (2PL) model using simulated testlet data under a common-item nonequivalent groups design. Impacts of various levels of testlet effects, calibration procedures, group differences, number of common items, sample size were investigated. Three traditional linear equating methods were used as criteria for the IRT true score equating and IRT observed score equating results from the three item response theory models. In general, the equating performance based on the two polytomous models yielded results that were more compatible with the results of the traditional equating methods with the presence of testlet effects. Even in some conditions without testlet effects, the equating performance of the two polytomous models was more similar to that of the traditional methods than the dichotomous 2PL model, particularly when the number of common items was larger. Of the two polytomous models, the GRM was found to render results in more agreement with those of traditional linear methods in conditions of separate calibration with linking. The characteristic curve linking methods outperformed the moment methods in a majority of conditions. The separate calibration procedures were better than the concurrent calibration procedure in most of the conditions, especially when the number of common items was small.
Analytic Smoothing For Equipercentile Equating Under The Common Item Nonequivalent Populations Design
DOWNLOAD
Author : Michael J. Kolen
language : en
Publisher:
Release Date : 1987
Analytic Smoothing For Equipercentile Equating Under The Common Item Nonequivalent Populations Design written by Michael J. Kolen and has been published by this book supported file pdf, txt, epub, kindle and other format this book has been release on 1987 with Educational tests and measurements categories.
Practical Issues In Linear Equating Using The Common Item Nonequivalent Populations Design
DOWNLOAD
Author : Robert L. Brennan
language : en
Publisher:
Release Date : 1986
Practical Issues In Linear Equating Using The Common Item Nonequivalent Populations Design written by Robert L. Brennan and has been published by this book supported file pdf, txt, epub, kindle and other format this book has been release on 1986 with Educational tests and measurements categories.