LEARNING@SCALE: INNOVATIONS IN SCALING AND ENHANCING LEARNING.
We are excited to announce that the Learning at Scale (L@S) conference will be held June 26-28, 2018 in London, UK.
L@S seeks contributions that address innovations in scaling and enhancing learning, empirical investigations of learning at scale, new technical systems for learning at scale, and novel syntheses of relevant research.
The conference will be part of a weeklong London Festival of Learning where L@S 2018 will be alongside the International Conference of the Learning Sciences and the Artificial Intelligence in Education Conference. The weeklong Festival will take place June 24-30. The L@S conference will be June 26-28, with a common festival day on June 27, and the single-track L@S events on June 26 and 28.
L@S investigates large-scale, technology-mediated learning environments with many learners and few experts to guide them. Large-scale learning environments are incredibly diverse: massive open online courses, intelligent tutoring systems, open learning courseware, learning games, citizen science communities, collaborative programming communities (such as Scratch), community tutorial systems (such as StackOverflow), shared critique communities (such as DeviantArt), and the countless informal communities of learners (such as the Explain It Like I’m Five sub-Reddit) are all examples of learning at scale. These systems either depend upon large numbers of learners, or they are enriched through use of data from previous use by many learners. They share a common purpose–to increase human potential–and a common infrastructure of data and computation to enable learning at scale.
Investigations of learning at scale naturally bring together two different research communities. Since the purpose of these environments is the advancement of human learning, learning scientists are drawn to study established and emerging forms of knowledge production, transfer, modeling, and co-creation. Since large-scale learning environments depend upon complex infrastructures of data storage, transmission, computation, and interface, computer scientists are drawn to the field as powerful site for the development and application of advanced computational techniques. At its very best, the L@S community supports the interdisciplinary investigation of these important sites of learning and human development.
The ultimate aim of the L@S community is the enhancement of human learning. In emerging education technology genres (such as intelligent tutors in the 1980s or MOOCs circa 2012), researchers often use a variety of proxy measures for learning, including measures of participation, persistence, completion, satisfaction, and activity. In the early stages of investigating a technological genre, it is entirely appropriate to begin lines of research by investigating these proxy outcomes. As lines of research mature, however, it is important for the community of researchers to hold each other to increasingly high standards and expectations for directly investigating thoughtfully-constructed measures of learning. In the early days of research on MOOCs, for instance, many researchers documented correlations between measures of activity (videos watched, forums posted, clicks) and other measures of activity, and between measures of activity and outcome proxies including participation, persistence, and completion. As MOOC research matures, additional studies that document these kinds of correlations should give way to more direct measures of student learning and of evidence of instructional techniques, technological infrastructures, learning habits, and experimental interventions that improve learning. As a community, we believe that the very best of our early papers define a foundation to build upon but anticipate that future papers will bring us well beyond.