Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                

Assessing English for Professional Purposes - Introduction

2020, Assessing English for Professional Purposes

Near final version of the ‘Introduction’ in Knoch, U. & Macqueen, S. (2020). Assessing English for professional purposes. Abingdon: Routledge. Introduction Modern life has become ever more dependent on ‘expert systems’, such as the airline industry, judicial and health systems (Giddens, 1990). At the same time, expertise itself is increasingly mobile. Doctors, engineers, programmers, nurses, accountants, pilots, teachers and many other professions, including workers with a plethora of hybrid professional capabilities, move across national borders in response to demands for the diverse array of expertises that sustain modern systems. A professional qualification now offers employment potential well beyond the boundaries of the state in which it was earned with the key proviso that the migrant professional has the ability to encode their expertise in the language of the destination nation-state. Hence, a need for language assessment arises at the governance boundaries of nations, professional certification bodies and workplaces. In this movement of expertise, standard varieties of the English language have accrued a high economic and cultural value. This means that it is readily used as a gate-keeping mechanism for participation in the professional workforce of migration destinations that have English as an official or national language. Commercial English language tests classify the language abilities of millions of test-takers annually, some of whom will undergo testing multiple times. As both a symbol and tool of globalization, English language skills are valued even in corporations which are predominantly non-English speaking. Therefore evidence of English language ability in the form of test scores converts to a competitive advantage for professional graduates seeking to work in both English-speaking and non-English-speaking countries. Some professions, especially those in industries serving global connectivity such as pilots, air traffic controllers and call centre staff, require people to work in lingua franca varieties of English such as Aviation English, that are not tied to national boundaries. These varieties too, are the objects of language tests. The regulatory role of English language tests is, in part, a response to the management of risk. Many professionals work in high-risk jobs, where communication breakdowns can lead to serious consequences. Even when the stakes are not so high, poor communication or miscommunication has negative effects on work practices such as relations with clients and the efficient achievment of work tasks. Language asessments for gate-keeping and screening purposes are efforts to avert the risk of miscommunication. Just what constitutes ‘risk’, however, is in the eye of the beholder. For skilled migration applicants, tests themselves, as potential barriers to employment, pose a risk. For skilled migration policy makers, the risk may have a political basis that is more concerned with immigration quotas and skill oversupply or shortages than effective communication. At the same time, ‘expert systems’ implicate laypeople in their expert procedures. The layperson converts to the role of ‘patient’ in the health professional world and is suddenly, and consequentially, surrounded by the need to understand the knowledge, activities, systems and relationships encoded in the language used there. A layperson who is a client signing off on a contract is legally commiting to an agreement about their comprehension of professional matters that are encoded in specialist register. The comprehension of laypeople as jurors in 1 legal proceedings is gravely consequential. Thus, even for laypeople, there is risk associated with not understanding the language used in professional worlds. This book is about various kinds of language assessment processes that are carried out in relation to participation in professionally-oriented settings. Our contribution comes some forty years after the first standardized English test for specific purposes, and with the benefit of the extensive development in research, theory and practice that has occurred since. Our aim is to recontextualize this body of work in the current times of rapid globalization, increasing transnational migration flows, massive technologization, English language spread and shift and increasingly complex infrastructures of professional regulation and statehood. We take account of these various dynamics and offer both a theoretical basis and practical guide for the development and validation of language assessment for professional purposes. What we hope to achieve is a renewed agenda for inquiry into Language Assessments for Professional Purposes (LAPP). It is one which may seem uncomfortably broad to some, but in the interests of greater awareness of the extent of language assessment activities and better assessment practices in society generally, it is an agenda worth pursuing we think. It has been nearly 20 years since Dan Douglas published his seminal text entitled ‘Assessing Languages for Specific Purposes’, and in the years since, new language assessment for professional purposes contexts have emerged or been formalized through various policies. Douglas’s work remains foundational and we do not attempt to repeat, revise or replace it here. We take up where he left off, by extending the remit of the field and theorizations about its various tools: those for understanding constructs, for carrying out needs analyses, for developing assessment instruments, for validation and for evaluating policy. Definition of LAPP and notes on terminology With the broader remit mentioned above in mind, we define language assessment for professional purposes (LAPP) as any assessment process, carried out by and for invested parties, which is used to determine a person's ability to understand and/or use the language of a professionally-oriented domain to a specified or necessary level. Although we use the term ‘test’ when appropriate throughout this book, and we draw upon the preponderance of test-based research and theory, it can be seen in the examples above, that there are non-test language evaluations that are highly consequential in terms of participation in a professional domain. Therefore, we use the term ‘assessment’ as the more general term, and ‘test’ to refer to instruments delivered under test conditions, such as time and resource-access constraints, often with multiple administrations. In the spirit of the broader remit, we refer to people who are assessed as ‘assessees’, and as ‘test-takers’ when we are discussing test instruments. This broadening allows us to consider assessment practices which are not explicitly named as such. Similarly, LAPP is intended to cover any language assessment in or related to workplaces and professional domains generally. We use the terms ‘domain insider’, ‘domain expert’ and ‘subject matter expert’ interchangeably throughout the book to refer to professionals with training and experience in a specific profession. We understand that this is a very simplistic use of the term, as professionals and professional specialisations differ greatly in and across workplaces, and the ‘experts’ therefore differ in their experiences. However, for consistency, we have decided to stick to this use of the term as it is currently used in the literature on language assessments for specific purposes. 2 Overview The chapters in this book are relatively self-contained but interrelated. Wherever text in a chapter relates to another chapter, we have noted this within the chapter. Readers may wish to read the six chapters in the book consecutively, or they may want to select individual chapters of interest from which they can explore the leads into other chapters. In Chapter 1, Scope, we set out the territory for the remainder of the book. The broadened conceptualisation of language assessment for professional purposes (LAPP) is introduced by drawing on two media stories we use as a basis for expanding LAPP into assessments for insiders and outsiders taking part, or seeking to take part in professional domains. We introduce the sociological concept of risk to conceptualize LAPP as a social exercise in risk management. We explore risk from the viewpoints of different stakeholder groups, including how the management of ‘language-associated risk’ implies responsibilities for different stakeholders in striving for assessment practices that are as beneficial and as fair as possible. We also consider the implications of rapid sociotechnological development for LAPP. In Chapter 2, Construct, we interrogate how constructs are represented in the language assessment literature, and we explore how they manifest in assessment practices and stakeholder activities. We propose a socially-oriented theory of construct for LAPP, starting with the dimensions of an assessment construct and then suggesting how these relate to the micro and macro contexts of an assessment. We then turn our attention to language varieties and professional registers, distinguished by Michael Halliday as ‘ways of speaking’ and ‘systems of meaning’ respectively (1985, p. 41). Because constructs are determined by what is sampled in an assessment procedure, we draw up a model for sampling from the gamut of linguistic repertoires that inhabit professional workplaces. In Chapter 3, Needs, we look closely at one key ingredient in the development of LAPPs, the needs analysis. We propose a broadened model of needs analysis, building on similar work in curriculum design, and discuss each component of our model in detail, drawing on methodological considerations in collecting this kind of information, where appropriate. The chapter concludes with a description of a needs analysis conducted for a large-scale LAPP, the Occupational English Test (OET), a language screening test for overseas-trained health professionals. Chapter 4, Development, then describes the various stages of LAPP development following a needs analysis. We start by describing the process of reconciling the findings of a needs analysis and developing a test blueprint and test specifications. We do this by taking test developers through a series of practical stages aimed at arriving at a representative list of tasks which could be included in a test blueprint. The remaining part of the chapter focusses on some key considerations specific to LAPP development, including the elicitation and development of scoring criteria based on indigenous criteria (that is, the criteria valued by domain insiders), setting standards on LAPP by drawing on the judgements of domain insiders and considerations associated with the operational use of a LAPP. In Chapter 5, Validation, we draw on the argument-based approach to validation, a framework proposed by researchers in educational assessment (e.g. Kane, 2006) to guide validation activities, to set out a comprehensive framework which researchers and test 3 developers can draw on when validating the score interpretations and uses of LAPPs. The framework we propose is designed to be flexible and adaptable to different contexts. This framework can be used to identify weaknesses in a chain of argumentation and to set research priorities. Where existing research on LAPPs is available, we have integrated that into our description of the framework. In Chapter 6, Policy, we discuss the fuzzy distinction between language assessments and language-associated policies. We take a broad view of policy, following work by Elana Shohamy and Tim McNamara, to include the covert work done by language assessments. We return to the notion of construct, particularly the public-facing dimension, in the question of the uses of language-based assessments as social sorting mechanisms. We consider the rise of national and international frameworks (e.g. the Common European Framework of Reference) which serve as supra-organisational mechanisms for linking different test instruments across different jurisdictions. We return to the principles of benefit and fairness, and the notion of language-associated risk in the making and enacting of policies, for both workplace insiders as well as outsiders. A hierarchy of affordances of assessment instruments proposed in this chapter is intended to guide policy decisions. We conclude the chapter by arguing that the type of validation work we describe in Chapter 5 may need to be complemented by policy evaluation in certain contexts. For this reason, we draw on the argumentative approach to policy evaluation and set out a framework that researchers may use as a basis for interrogating language-associated policies. In the concluding chapter, we return to the question of the use of linguistic evidence as a basis for decisions. We distinguish between two types of evidence: assessment evidence on which scores are based and validation evidence which may reveal something about how trustworthy a score is for a particular use. We revisit the tools, theories and concepts presented in this book under the overarching theme of ‘evidence’. We suggest some future directions for research and how a broadened agenda for assessment in professional worlds and workplaces might proceed. A broadened agenda Language Assessments for Professional Purposes (LAPPs) and Language Assessment for Specific Purposes (LSP) (the latter also focusses on language assessments for academic purposes) have been a key part of the language assessment literature for several decades. Because of the explicit connections made between the assessment instrument and the Target Language Use domain (TLU domain), these areas of inquiry have paid particular attention to the interaction between domain-related knowledge (or content) and language ability. We feel it is time to take stock of our directions in theory and practice as a basis for new developments and increased awareness of the risks and responsibilities associated with language assessment. In view of the various uses of these assessments, either as high-stakes sorting mechanisms or risk-reducing devices, our intention is to stretch the now-comfortable boundaries of the field. Doing so will offer some alternative directions to the centripetal force created by well-known commercial tests which attract attention away from the myriad of language assessing practices that are carried out in the maintenance of expert systems. In order to do this, we have explored within our field, and we have ventured beyond our discipline to offer a broader proposed agenda and some tools for working with it. At the same time, we pay close attention to how our methods are evolving in relation to our more traditional object of inquiry, the standardized test. While we consider our traditional objects 4 well worth the continued pursuit, we hope that we have offered some new lenses for familiar concepts, as well as some new avenues to pursue. References Giddens, A. (1990). The consequences of modernity. Cambridge: Polity. Halliday, M. A. K. (1985). Register variation. In M. A. K. Halliday & R. Hasan, Language, context and text: Aspects of language in a social-semiotic perspective (pp. 29-43). Victoria: Deakin University Press. 5