top of page
Sympo 1

Invited Symposium

Wooden Hut

Presentation 4

Language teacher development in relation to TOEFL family assessments 

시렁지렁.png

Veronika Timpe-Laughlin

Research scientist at - Educational Testing Service in Princeton, NJ.

Veronika Timpe-Laughlin is a research scientist at Educational Testing Service in Princeton, NJ. She received her Ph.D. in Applied Linguistics/TESOL in 2013 and her M.A. Ed. degree in teaching English and German in 2009—both from TU Dortmund University (Germany). She also holds an M.A. degree in Language Testing from Lancaster University (England). Since joining ETS in 2013, Veronika has conducted research in L2 pragmatics, young learner assessment, teacher education, and task-based language teaching (TBLT). Her research has recently been published in Intercultural Pragmatics, Language Teaching, Language Testing, and System. 

Renka Ohta

Research Project Manager in the Center for Language Education and Assessment Research at Educational Testing Service. 

Renka Ohta is a Research Project Manager in the Center for Language Education and Assessment Research at Educational Testing Service. She earned her Ph.D. in Foreign Language and ESL Education at the University of Iowa in 2018. Before working at ETS, she taught English as a second language at a community college in Cedar Rapids, Iowa, and was involved in professional development programs for K-12 teachers of English language learners in Iowa schools. Before pursuing graduate studies, she taught English at a high school in Japan. Her research interests include integrated writing assessment and teacher development.

지렁시롱.png

Invited Symposium 2

머 금고 사포

Invited Symposium 1

Language learning and assessment innovations in the TOEFL family

Abstract


The TOEFL iBT® test is an assessment of academic English language proficiency used for university admissions and other purposes by more than 11,000 institutions in more than 150 countries worldwide. Since its initial launch in 1964, the TOEFL® test has undergone major revisions motivated by advances in theories of language learning and improvements in English teaching practices. The most recent revision, the TOEFL iBT test, contains numerous innovative design features—including in particular integrated tasks that engage multiple skills—to simulate language use in academic settings and reflect the reading, listening, speaking, and writing demands of real-world academic environments. This alignment between the design of an assessment and best practices in language teaching and learning is a core value underlying all tests in the TOEFL Family of Assessments. The TOEFL® Young Students Series (the TOEFL Primary® and TOEFL Junior® tests) was designed to provide useful formative and summative information to teachers and learners of English in school settings. The TOEFL ITP® program offers institutions an affordable test for placement and progress monitoring within English training programs.  The newest addition to the family, the TOEFL Essentials® test, combines the convenience of at-home testing with reliable measurement of relevant language skills and abilities for academic and other purposes, in a student-friendly format.
At ETS, we understand that our English language assessments, and related products and services, should be designed to support language teaching and learning in a variety of educational settings. Accordingly, we engage in on-going research and innovation in response to the evolving needs of language teachers and learners worldwide. In this symposium, ETS researchers present on four recent projects that demonstrate this goal. Following an introduction to the teaching, learning, and assessment orientation of the TOEFL Family, presentations will address: (a) the application of  vertical scaling methodology across all tests in the TOEFL Family of Assessments to help teachers and learners understand English proficiency development in relation to test scores; (b) the development of a brand new English proficiency test in response to critical emerging needs of test takers and score users; (c) research that explores the impact of test use on young learner educational practices; and (d) the creation of two new English language teacher development courses to support improved teaching practices in specific educational sectors. The symposium concludes with ample opportunity for questions and discussion.

 

 

 

 

 

 

 

 

Introduction
A TOEFL perspective on language teaching,learning, and assessment

John Norris.png

John Norris

Senior Research Director of the Center for Language Education and Assessment Research at Educational Testing Service.

John Norris is Senior Research Director of the Center for Language Education and Assessment Research at Educational Testing Service. Prior to joining ETS, he worked at Georgetown University, University of Hawai'i at Mānoa, and Northern Arizona University. He began his career as an English teacher, and he completed his Ph.D. in Second Language Acquisition at the University of Hawai‘i. John’s research focuses on language education (task-based language teaching), assessment, program evaluation, and research methods. His most recent book (with Xiaoming Xi) is Assessing academic English for higher education admissions. John speaks German, Portuguese, and Spanish, and he is currently acquiring Japanese.

Presentation 1
The TOEFL learning journey: Facilitating test selection and score interpretation through application of vertical scaling methodology

.

시렁시렁.png

Spiros Papageorgiou

Managing Senior Research Scientist in the Center for Language Education and Assessment Research at Educational Testing Service (ETS).

Spiros Papageorgiou is a Managing Senior Research Scientist in the Center for Language Education and Assessment Research at Educational Testing Service (ETS). Prior to joining ETS, he worked at the University of Michigan, where he was responsible for leading the development of English assessment programs.  Spiros received his doctoral degree in linguistics specializing in language testing from Lancaster University, UK. His publications cover topics such as standard setting, score reporting and interpretation, and listening assessment. Spiros is the coeditor (with Kathleen M. Bailey) of the volume Global Perspectives on Language Assessment: Research, Theory, and Practice (2019).

Lixiong Gu

A Director of Psychometrics and Data analysis in Psychometric Analysis and Research at Educational Testing Service (ETS) in Princeton, New Jersey.

Lixiong Gu is a Director of Psychometrics and Data analysis in Psychometric Analysis and Research at Educational Testing Service (ETS) in Princeton, New Jersey. He holds an M.S. degree in Educational Psychology from Beijing Normal University and Ph.D in Measurement and Quantitative Methods from Michigan State University at East Lansing. Lixiong’s research interests include psychometric issues in language assessments, test equating and linking, and computerized adaptive testing. 

 

시렁시롱.png

Presentation 2

The TOEFL Essentials test: A new assessment to meet evolving test-taker and score-user needs 

시롱지롱.png

Larry Davis

Research scientist in the Center for Language Education and Assessment Research at Educational Testing Service (ETS) in Princeton, New Jersey.

Larry Davis is a research scientist in the Center for Language Education and Assessment Research at Educational Testing Service (ETS) in Princeton, New Jersey. He holds an M.A. degree in TESOL from the University of Minnesota and Ph.D in Second Language Studies from the University of Hawai’i at Manoa. Larry has taught English at the university level in China and the U.S. and taught university-level teacher training courses. His research interests include all aspects of the assessment of speaking, including development of technology-enhanced speaking tasks, assessment of spoken interaction, creation of rubrics, rater cognition, and automated evaluation of speaking ability. 

Shoko Sasayama

Associate Research Scientist at Educational Testing Service. 

Shoko Sasayama is an Associate Research Scientist at Educational Testing Service. She completed a Ph.D. in Linguistics at Georgetown University. Prior to joining ETS, Shoko worked as a faculty development specialist and as Assistant Managing Director of an English language program at the University of Tokyo. Her research focuses primarily on the role of cognition in communicative task design for language learning and assessment, including her award-winning publication, “Is a ‘complex’ task really complex? Validating the assumption of cognitive task complexity” (Modern Language Journal, 2016). At ETS, Shoko leads projects on assessment task design, language learning product innovation, and teacher training.

시렁지롱.png

Presentation 3

Investigating the impact of the TOEFL Young Students Series tests on teaching and learning 

시롱시롱.png

Mikyung Kim Wolf

managing principal research scientist at Educational Testing Service.

Mikyung Kim Wolf is a managing principal research scientist at Educational Testing Service. Her research areas span technology-enhanced language assessments, formative assessment, and validity issues in assessing K-12 English language learners in the U.S. and in global contexts. Mikyung has published widely including two edited books, English Language Proficiency Assessments for Young Learners and Assessing English Language Proficiency in U.S. K-12 Schools. She is a chair of the International Language Testing Association (ILTA)’s Language Assessments for Young Learners SIG. She also serves as an associate editor of Language Assessment Quarterly. Her recent co-authored article, Investigating the Benefits of Scaffolding in Assessments of Young English Learners: A Case for Scaffolded Retell Tasks won the 2019 ILTA Best Article Award. Mikyung received a B.A. in English language and literature and M.A. in psycholinguistics from Korea University, and a Ph.D. in applied linguistics with specialization in language assessment from UCLA. 

Presentation 4

Language teacher development in relation to TOEFL family assessments 

시렁지렁.png

Veronika Timpe-Laughlin

Research scientist at - Educational Testing Service in Princeton, NJ.

Veronika Timpe-Laughlin is a research scientist at Educational Testing Service in Princeton, NJ. She received her Ph.D. in Applied Linguistics/TESOL in 2013 and her M.A. Ed. degree in teaching English and German in 2009—both from TU Dortmund University (Germany). She also holds an M.A. degree in Language Testing from Lancaster University (England). Since joining ETS in 2013, Veronika has conducted research in L2 pragmatics, young learner assessment, teacher education, and task-based language teaching (TBLT). Her research has recently been published in Intercultural Pragmatics, Language Teaching, Language Testing, and System. 

Renka Ohta

Research Project Manager in the Center for Language Education and Assessment Research at Educational Testing Service. 

Renka Ohta is a Research Project Manager in the Center for Language Education and Assessment Research at Educational Testing Service. She earned her Ph.D. in Foreign Language and ESL Education at the University of Iowa in 2018. Before working at ETS, she taught English as a second language at a community college in Cedar Rapids, Iowa, and was involved in professional development programs for K-12 teachers of English language learners in Iowa schools. Before pursuing graduate studies, she taught English at a high school in Japan. Her research interests include integrated writing assessment and teacher development.

지렁시롱.png

Invited Symposium 2

National Policy, Consequences,

and English Learning: 

A Case of the TEPS-Family Tests

Abstract

Learning-oriented assessment (LOA) is primarily concerned with how to make assessment inform and guide learning and instruction. For this reason, LOA approaches require the development of not only appropriate tests designed to serve these purposes but also methodologies and procedures to extract specific information indicating where the learners currently stand and what they need to progress to the next stage of learning. The concept of effect-driven testing (Fulcher & Davidson, 2007) can be very useful in designing and developing an LOA system. In effect-driven testing, the intended effect of a test is clearly specified right from the beginning of a test design process and the test is structured and built with the very effect in mind. In fact, such test effects can also be conceptualized at many different levels of assessment evaluation and validity investigation.
Since the beneficial impact of assessment on learning is a crucial element in evaluating the results of LOA, one way to understand LOA is to investigate it from the perspective of consequential validity, particularly in terms of washback. Oftentimes, investigation of consequential validity and washback requires an expanded view of test effect, i.e., a macro-perspective on test effects, which can also be closely interconnected with the evaluation of language testing policies in a particular society. This is very true in the South Korean context where domestic and international English proficiency tests are used interchangeably for various admission and employment purposes. So it is very important to examine the impact of the domestic and international tests in tandem on English language learning and learners in South Korea along with various assessment variables/factors that can moderate or mediate such impact.
With this as a backdrop, the purpose of this symposium is to investigate and discuss the interrelationships among national testing policies, consequences of English proficiency testing, and English language learning in the South Korean context, with a focus on the fairness of score conversion between the TEPS family of tests and international tests and other related issues. The symposium consists of four related presentations followed by a discussion. The first presentation gives an overview of governmental, regulatory systems of English proficiency tests in Korea by reviewing clauses in relevant laws/decrees and discussing the limitations and problems of the systems. The second one reports a large-scale score comparability study undertaken to develop a new score conversion table for the TEPS and TOEIC. The third presents the results of a content comparison study and a test-taker survey that were done in support of the score conversion study. The fourth reports results of a research project conducted to develop a new intermediate-level English proficiency test targeting secondary school students and civil service exam candidates in South Korea, highlighting the difficulty of the new test compared to those of TEPS and TOEIC. The four presentations will be followed by the discussant’s critique and a Q&A session.

 

 

 

 

 

 

 

 

Presentation 1
National Certification and Score Conversion Between Domestic and International Tests in Korea: Legal and Policy Issues

Yong-Won Lee

YoungYu Yang

Abstract

   National policies on certification tests can have a huge impact on English language tests administered in a particular country, which may in turn influence English language learners’ attitude toward these tests. Where there are both domestic and international tests competing for market share, it is critically important to ensure that passing scores are determined in a fair and equitable manner across these tests, which require careful supervision/management of score conversion relationships among these tests by a credible public body or a relevant governmental organization. There are two major legal apparatuses currently in place to regulate English proficiency tests used in South Korea: (a) Framework Act on Qualifications (jagyeokgibonbeop) enforced by the Korea Research Institute for Vocational Education and Training under the auspice of the Ministries of Education and Labor and (b) Decree on Public Officials Appointment Examinations (gongmuwonimyongsiheomyeong) enforced by the Ministry of Personnel Management. The first one is primarily concerned with national certification/regulation of domestically-developed tests, whereas the second one has to do with determining English ability certification tests appropriate for civil service examinations, along with their passing score(s). One critical problem of the current systems, however, is that international tests, such TOEFL and TOEIC, are located outside the purview of such national certification/regulation systems while enjoying the privilege of becoming dominant English ability certification tests for civil service examinations in the South Korean context. Even worse, there is no regulatory procedure enforced to periodically review and evaluate the fairness of score conversion between domestic and international tests as well as the adequacy of cut scores established for each of these tests. With this as a backdrop, the main purposes of this presentation is: (a) to give an overview of these governmental, regulatory systems by reviewing clauses in relevant laws/decrees, (b) expose the limitations and problems of the systems, with a focus on test revision history, score trend analyses of these tests for the past two decades, and fairness of score conversion relationships among them, and finally (c) suggest possible ways to reform these certification/regulatory systems.

Presentation 2
Score Comparability of TEPS and TOEIC

Euijin Lim                  Guemin Lee                Heesung Jun 
   
Jiwon Choi                Youmin Hong              Nagap Park

Abstract

   Scores from the TEPS (Test of English Proficiency developed by Seoul National University) and the TOEIC (Test of English for International Communication) are widely used in Korea for various purposes, including employment, admission, and certification. However, there is more than one score conversion relationship between these two tests, causing confusion among test takers. Since being established in 2004 and 2011, moreover, the conversion relationships have not been updated for at least 10 years and up to 17 years. This is problematic, as both tests have undergone revision and the characteristics of the test-taker populations may have also shifted within this time. As a result, users of one test may have been disadvantaged over those of the other.
   With this as a background, a large-scale score comparability study was undertaken to develop a new score conversion table for the TEPS and TOEIC. Using a single group design, data were collected from a total of 1,415 test takers who had taken both tests. Log-linear pre-smoothing and equipercentile methods were used for linking, with cross validation across subgroups. 
   According to the results of this study, TEPS test takers need to obtain higher scores than TOEIC test takers when the old conversion tables are used. In fact, at a certain cut-score point widely used, the score disadvantage has been shown to be up to 102 points on the 600-point TEPS score scale. In our presentation, we will describe the methods and results of the score comparability study in greater detail and propose the social implications of the final score conversion table.

Presentation 3
Evaluating TEPS-TOEIC Score Conversion: Content Comparison and Test-Taker Survey

Heesung Jun                 Jiwon Choi                  Youngmi Lee

                  Yoo-Ree Chung              Yong-Won Lee

Abstract

   The TEPS (Test of English Proficiency developed by Seoul National University) and the TOEIC (Test of English for International Communication) are two standardized tests of English language proficiency whose scores are widely used for various purposes in Korea. To ensure that the uses of scores from these two tests are fair and valid, it is important to establish an accurate and unbiased score conversion relationship between the two tests that does not advantage the users of one test over those of the other.
   In our presentation, we question the accuracy of the TEPS-TOEIC score conversion relationship that was first established more than 15 years ago and is still being adopted by major decisions makers in Korea today. As a precursor to a large-scale score comparability study conducted in 2020 (Lee, Lim, Hong, & Park, 2021), we carried out a content comparison study and a test-taker survey to triangulate the findings.
   We conducted an expert survey of five language testing specialists regarding the comparative difficulty of the content of the two tests. The reading comprehension sections of the two tests were also analyzed and compared using measures of passage difficulty. Furthermore, 500 test takers who have taken both TEPS and TOEIC were surveyed regarding their perceptions of the score conversion table currently being used for the civil servant selection exam.
   In our presentation, we will focus on discussing the results of both the content comparison and test-taker survey along with their implications, particularly in terms of the need to regularly update score conversion relationships for the fairness and validity of test score use.

Presentation 4
Using the Delphi Method to Inform the Development of a Glocalized English Proficiency Test for Korean EFL Learners

  Euijin Lim              Heesung Jun               Jiwon Choi

Youngmi Lee           Yoo-Ree Chung           Yong-Won Lee

Abstract

   The TEPS (Test of English Proficiency developed by Seoul National University) is a general-purpose English test designed to assess the English proficiency of Korean EFL learners at a diverse range of proficiency levels. An important strength of this test lies in its excellent discriminating power for advanced-level English language learners, as is well evidenced by the wide use of its scores for undergraduate and graduate admissions at major universities and professional graduate schools in South Korea. While the usefulness of TEPS scores in such contexts has been proven, test-taker feedback over the years as well as needs analysis have consistently pointed to a growing need for a glocalized test with good discriminating power among intermediate-level test takers. With this as a background, a research project was undertaken to develop a new intermediate-level English proficiency test targeting secondary school students, civil service exam candidates, and job seekers in Korea.
   The project was conducted in three stages: (a) preliminary research, (b) creation of a test specification and pilot test form, and (c) pilot testing. First, preliminary research included domain and needs analyses, a Delphi study with 19 language education/assessment experts, and a review of existing language tests and their score reports. Second, based on the results of the preliminary research, a tentative test design framework was created for the new test, and then a pilot test form was assembled. Third, the pilot test form was administered to over 1,000 high school and adult test takers, who were also asked to complete a post-test survey questionnaire. In addition, an expert survey was conducted with 17 English education/assessment experts and 13 high school English teachers regarding the format, content, and difficulty level of the pilot test.
   In this presentation, we will focus on demonstrating how the results of our preliminary research, particularly the Delphi study, were utilized in the early stages of test development to make important decisions on various aspects of the new English proficiency test. We will also discuss how the results from the Delphi study can support the validity argument for the use of scores from the new test. We will conclude with implications and suggestions for test developers interested in finding out about what kinds of preliminary research can be done to inform and guide the test development process in its early stages.

Presenters

Photo_Yong-Won Lee.jpg

Yong-Won Lee

Professor of English linguistics in the Department of English Language and Literature at Seoul National University

Yong-Won Lee is a professor of English linguistics in the Department of English Language and Literature at Seoul National University. His research interests include multi-faceted analyses of writing and speaking assessments, diagnostic language assessment, and automated essay and speech evaluation.

Special professor in the School of Communications at Dankook University. 

Photo_YoungYu Yang.jpg

YoungYu Yang

YoungYu Yang is a special professor in the School of Communications at Dankook University. He worked for more than thirty years as a reporter for Joongang Ilbo Daily Newspaper.

 

Guemin Lee

Professor in the Department of Education in the College of Education at Yonsei University

Guemin Lee is a professor in the Department of Education in the College of Education at Yonsei University. His research interests include scaling, equating, college admission testing, generalizability, and item response theories.

Euijin Lim

Euijin Lim is an assistant professor in the Department of Education at Chinju National University of Education. Her research interests are equating, item response theory, and generalizability theory.

Photo_Euijin Lim.jpg

Assistant professor in the Department of Education at Chinju National University of Education

Principal researcher at the TEPS Center, Language Education Institute, Seoul National University

Youngmi Lee is a principal researcher at the TEPS Center, Language Education Institute, Seoul National University, where she works on the production and quality control of the TEPS and its family of tests.

 

Youngmi Lee

Photo_Youngmi Lee.jpg
Photo_Yoo-Ree Chung.jpg

Yoo-Ree Chung

Senior researcher at the TEPS Center, Seoul National Univeristy 

Senior researcher and Head of the Test Research and Development at the TEPS Center, Language Education Institute, Seoul National University

Heesung Jun is a senior researcher and Head of the Test Research and Development Department at the TEPS Center, Language Education Institute, Seoul National University, where she works on the development and validation of English proficiency tests, including the TEPS and its family of tests. Her research interests include integrated writing assessment and computer-assisted language testing.

 

Photo_Heesung Jun.jpg

Heesung Jun

Yoo-Ree Chung is a senior researcher at the TEPS Center, Seoul National University. She has earned a Ph.D. degree in applied linguistics at Iowa State University and recently taught courses in statistics for linguistics, Korean language assessment, and academic writing at Yonsei University. Her research interests include test validation and computer-assisted language testing.

Senior Researcher and Head of the Measurement and Statistical Research Department at the TEPS Center, Langauge Education Insitute, Seoul National University 

Jiwon Choi is a senior researcher and Head of the Measurement and Statistical Research Department at the TEPS Center, Language Education Institute, Seoul National University, where she works on scoring and quantitative research for the TEPS and its family of tests. Her research interests are test equating, scoring, and item response theory.

 

Photo_Jiwon Choi.jpg

Jiwon Choi

Photo_Nagap Park.JPG

Nagap Park

Graduate student in the Department of Education in the College of Education at Yonsei University

Nagap Park is a graduate student in the Department of Education in the College of Education at Yonsei University. His research interests include item response theories, statistical learning, and educational evaluation.

Graduate student in the Department of Education in the College of Education at Yonsei university

Youmin Hong is a graduate student in the Department of Education in the College of Education at Yonsei university. Her research interests are item response theory, computerized adaptive testing, and equating.

 

Photo_Youmin Hong.JPG

Youngmi Lee

sympo 2
bottom of page