Git is a widely used version control system that is powerful but complicated. Its complexity may not be an inevitable consequence of its power but rather evidence of flaws in its design. To explore this hypothesis, we analysed the design of Git using a theory that identifies concepts, purposes, and misfits. Some well-known difficulties with Git are described, and explained as misfits in which underlying concepts fail to meet their intended purpose. Based on this analysis, we designed a reworking of Git (called Gitless) that attempts to remedy these flaws.
To correlate misfits with issues reported by users, we conducted a study of Stack Overflow questions. And to determine whether users experienced fewer complications using Gitless in place of Git, we conducted a small user study. Results suggest our approach can be profitable in identifying, analysing, and fixing design problems.
Join us on Monday 7th September to discuss using theory in Computing Education Research at 11am. We’ll be talking about a paper  by Greg L. Nelson and Amy Ko at the University of Washington:
A primary goal of computing education research is to discover designs that produce better learning of computing. In this pursuit, we have increasingly drawn upon theories from learning science and education research, recognising the potential benefits of optimising our search for better designs by leveraging the predictions of general theories of learning. In this paper, we contribute an argument that theory can also inhibit our community’s search for better designs. We present three inhibitions: 1) our desire to both advance explanatory theory and advance design splits our attention, which prevents us from excelling at both; 2) our emphasis on applying and refining general theories of learning is done at the expense of domain-specific theories of computer science knowledge, and 3) our use of theory as a critical lens in peer review prevents the publication of designs that may accelerate design progress. We present several recommendations for how to improve our use of theory, viewing it as just one of many sources of design insight in pursuit of improving learning of computing.
Details of the zoom meeting will be posted on our slack workspace at uk-acm-sigsce.slack.com. If you don’t have access to the workspace, send me (Duncan Hull) an email to request an invite to join the workspace.
As Universities transition to online teaching during the global coronavirus pandemic, there’s increasing interest in the use of pre-recorded videos to replace traditional lectures in higher education. Join us to discuss how video production affects student engagement, based on a paper published by Philip Guo at the University of California, San Deigo (UCSD) from the Learning at Scale conference on How video production affects student engagement: an empirical study of MOOC videos. (MOOC stands for Massive Open Online Course).  Here is the abstract:
Videos are a widely-used kind of resource for online learning. This paper presents an empirical study of how video production decisions affect student engagement in online educational videos. To our knowledge, ours is the largest-scale study of video engagement to date, using data from 6.9 million video watching sessions across four courses on the edX MOOC platform. We measure engagement by how long students are watching each video, and whether they attempt to answer post-video assessment problems.
Our main findings are that shorter videos are much more engaging, that informal talking-head videos are more engaging, that Khan-style tablet drawings are more engaging, that even high-quality pre-recorded classroom lectures might not make for engaging online videos, and that students engage differently with lecture and tutorial videos.
Based upon these quantitative findings and qualitative insights from interviews with edX staff, we developed a set of recommendations to help instructors and video producers take better advantage of the online video format. Finally, to enable researchers to reproduce and build upon our findings, we have made our anonymized video watching data set and analysis scripts public. To our knowledge, ours is one of the first public data sets on MOOC resource usage.
Details of the zoom meeting will be posted on our slack workspace at uk-acm-sigsce.slack.com. If you don’t have access to the workspace, send me (Duncan Hull) an email to request an invite to join the workspace. The paper refers to several styles of video production, some examples below.
Join us for our next journal club meeting on Monday 6th July at 3pm, the papers we’ll be discussing below come from the #paper-suggestions channel of our slack workspace at uk-acm-sigsce.slack.com.
Show me the pedagogy!
The first paper is a short chapter by Katrina Falkner and Judy Sheard which gives an overview of pedagogic approaches including active learning, collaborative learning, cooperative learning, contributing student pedagogy (CSP), blended learning and MOOCs.  This was published last year as chapter 15 of the Cambridge Handbook on Computing Education Research edited by Sally Fincher and Anthony V. Robins. A lot of blended learning resources focus on technology, this chapter talks about where blended learning fits with a range of different pedagogic approaches.
Implementing blended learning
The second paper (suggested by Jane Waite) is Design and implementation factors in blended synchronous learning environments , here’s a summary from the abstract:
Increasingly, universities are using technology to provide students with more flexible modes of participation. This article presents a cross-case analysis of blended synchronous learning environments—contexts where remote students participated in face-to-face classes through the use of rich-media synchronous technologies such as video conferencing, web conferencing, and virtual worlds. The study examined how design and implementation factors influenced student learning activity and perceived learning outcomes, drawing on a synthesis of student, teacher, and researcher observations collected before, during, and after blended synchronous learning lessons. Key findings include the importance of designing for active learning, the need to select and utilise technologies appropriately to meet communicative requirements, varying degrees of co-presence depending on technological and human factors, and heightened cognitive load. Pedagogical, technological, and logistical implications are presented in the form of a Blended Synchronous Learning Design Framework that is grounded in the results of the study.
We look forward to seeing you there, zoom details are on the slack channel, email me if you’d like to request an invitation to the slack channel. Likewise, if you don’t have access to the papers let me know.
Short notes from the discussion
Some of the questions discussed on the day:
Inclusion raises a number of questions in terms of room management, gender balance – was this a consideration?
What effect do you think the absence of anyone F2F would have on the case studies and/or your outcomes?
How scalable is this approach? Can it be used with classes of 200 or 300 students?
Jaqueline Kenney, one of the co-authors of the paper we discussed joined us for the session (thanks again Jacqueline). Matt Bower also emailed some suggestions of work that follows on
See related work Collaborative learning across physical and virtual worlds: Factors supporting and constraining learners in a blended reality environment DOI:10.1111/bjet.12435 and blendsync.org
Bower, M. (2006). Virtual classroom pedagogy. Paper presented at the Proceedings of the 37th SIGCSE technical symposium on Computer science education, Houston, Texas, USA. DOI:10.1145/1121341.1121390
Bower, M. (2006). A learning system engineering approach to developing online courses. Paper presented at the Proceedings of the 8th Australasian Conference on Computing Education – Volume 52, Hobart, Australia.
Bower, M. (2007). Groupwork activities in synchronous online classroom spaces. Paper presented at the Proceedings of the 38th SIGCSE technical symposium on Computer science education, Covington, Kentucky, USA. DOI:10.1145/1227310.1227345
Bower, M. (2007). Independent, synchronous and asynchronous an analysis of approaches to online concept formation. Paper presented at the Proceedings of the 12th annual SIGCSE conference on Innovation and technology in computer science education, Dundee, Scotland. DOI:10.1145/1268784.1268827
Bower, M. (2008). The “instructed-teacher”: a computer science online learning pedagogical pattern. Paper presented at the Proceedings of the 13th annual conference on Innovation and technology in computer science education, Madrid, Spain. DOI:10.1145/1384271.1384323
Bower, M., & McIver, A. (2011). Continual and explicit comparison to promote proactive facilitation during second computer language learning. Paper presented at the Proceedings of the 16th annual joint conference on Innovation and technology in computer science education, Darmstadt, Germany. DOI:10.1145/1999747.1999809
Bower, M., & Richards, D. (2005). The impact of virtual classroom laboratories in CSE. Paper presented at the Proceedings of the 36th SIGCSE technical symposium on Computer science education, St. Louis, Missouri, USA. DOI:10.1145/1047344.1047447As well, this Computers & Education paper specifically relates to a study of teaching computing online:
Bower, M., & Hedberg, J. G. (2010). A quantitative multimodal discourse analysis of teaching and learning in a web-conferencing environment–the efficacy of student-centred learning designs. Computers & education, 54(2), 462-478.
Falkner, Katrina; Sheard, Judy (2019). “Pedagogic Approaches”: 445–480. doi:10.1017/9781108654555.016. Chapter 15 of the The Cambridge Handbook of Computing Education Research
Bower, Matt; Dalgarno, Barney; Kennedy, Gregor E.; Lee, Mark J.W.; Kenney, Jacqueline (2015). “Design and implementation factors in blended synchronous learning environments: Outcomes from a cross-case analysis”. Computers & Education. 86: 1–17. doi:10.1016/j.compedu.2015.03.006. ISSN0360-1315.
If you’ve not got access to the workspace yet, ping me or Alcywn Parker and we’ll add you to the group. Journal Club is part of the Association for Computing Machinery (ACM) Special Interest Group (SIG) on Computer Science Education (CSE) – all welcome!
ACM SIGCSE Journal Club returns Monday 4th May at 11am. The paper we’re discussing this month is “Relating Natural Language Aptitude to Individual Differences in Learning Programming Languages” by Chantel Prat et al published in Scientific Reports.  Here’s the abstract:
This experiment employed an individual differences approach to test the hypothesis that learning modern programming languages resembles second “natural” language learning in adulthood. Behavioral and neural (resting-state EEG) indices of language aptitude were used along with numeracy and fluid cognitive measures (e.g., fluid reasoning, working memory, inhibitory control) as predictors. Rate of learning, programming accuracy, and post-test declarative knowledge were used as outcome measures in 36 individuals who participated in ten 45-minute Python training sessions. The resulting models explained 50–72% of the variance in learning outcomes, with language aptitude measures explaining significant variance in each outcome even when the other factors competed for variance. Across outcome variables, fluid reasoning and working-memory capacity explained 34% of the variance, followed by language aptitude (17%), resting-state EEG power in beta and low-gamma bands (10%), and numeracy (2%). These results provide a novel framework for understanding programming aptitude, suggesting that the importance of numeracy may be overestimated in modern programming education environments
The paper describes an experiment which investigates the relationship between learning natural languages and programming languages and draws some interesting conclusions that provide some good discussion points. Does being good at learning natural languages like English make you good at learning programming language like Python? Do linguists make good coders?
Computing educators are often baffled by the misconceptions that their CS1 students hold. We need to understand these misconceptions more clearly in order to help students form correct conceptions. This paper describes one stage in the development of a concept inventory for Computing Fundamentals: investigation of student misconceptions in a series of core CS1 topics previously identified as both important and difficult. Formal interviews with students revealed four distinct themes, each containing many interesting misconceptions. Three of those misconceptions are detailed in this paper: two misconceptions about memory models, and data assignment when primitives are declared. Individual misconceptions are related, but vary widely, thus providing excellent material to use in the development of the CI. In addition, CS1 instructors are provided immediate usable material for helping their students understand some difficult introductory concepts.
In case you’re wondering, CS1 refers to the first course in the introductory sequence of a computer science major (in American parlance), roughly equivalent to first year undergraduate in the UK. CI refers to a Concept Inventory, a test designed to tell teachers exactly what students know and don’t know. According to Reinventing Nerds, the paper has been influential because it was the “first to apply rigorous research methods to investigating misconceptions”.