Mon 17 MayDisplayed time zone: Amsterdam, Berlin, Bern, Rome, Stockholm, Vienna change
02:00 - 02:50 | Opening/AwardsTechnical Papers / Tutorials / MIP Award / FOSS Award / content / Mining Challenge / Hackathon / MSR Awards / Registered Reports / Data Showcase / Shadow PC / Keynotes at MSR Room 1 | ||
02:50 - 03:10 | Break / Discussion Rooms Technical Papers / Tutorials / MIP Award / FOSS Award / content / Mining Challenge / Hackathon / MSR Awards / Registered Reports / Data Showcase / Shadow PC / Keynotes at MSR Room 1 | ||
03:10 - 04:00 | Welcome Event Technical Papers / Tutorials / MIP Award / FOSS Award / content / Mining Challenge / Hackathon / MSR Awards / Registered Reports / Data Showcase / Shadow PC / Keynotes at MSR Room 1 The MSR welcoming sessions will feature informal networking opportunities for newcomers to meet each other, learn about the MSR conference series, and interact with some established MSR veterans. All are welcome! | ||
10:00 - 10:50 | Resources for MSR ResearchTechnical Papers / Data Showcase at MSR Room 1 Chair(s): Felipe Ebert Eindhoven University of Technology | ||
10:01 3mTalk | PSIMiner: A Tool for Mining Rich Abstract Syntax Trees from Code Technical Papers Egor Spirin JetBrains Research; National Research University Higher School of Economics, Egor Bogomolov JetBrains Research, Vladimir Kovalenko JetBrains Research, Timofey Bryksin JetBrains Research, Saint Petersburg State University Pre-print | ||
10:04 3mTalk | Mining DEV for social and technical insights about software development Technical Papers Maria Papoutsoglou Aristotle University of Thessaloniki, Johannes Wachs Vienna University of Economics and Business & Complexity Science Hub Vienna, Georgia Kapitsaki University of Cyprus Pre-print | ||
10:07 3mTalk | TNM: A Tool for Mining of Socio-Technical Data from Git Repositories Technical Papers Nikolai Sviridov ITMO University, Mikhail Evtikhiev JetBrains Research, Vladimir Kovalenko JetBrains Research Pre-print | ||
10:10 3mTalk | Identifying Versions of Libraries used in Stack Overflow Code Snippets Technical Papers Ahmed Zerouali Vrije Universiteit Brussel, Camilo Velázquez-Rodríguez Vrije Universiteit Brussel, Coen De Roover Vrije Universiteit Brussel Pre-print Media Attached | ||
10:13 3mTalk | Sampling Projects in GitHub for MSR Studies Data Showcase Ozren Dabic Software Institute, Università della Svizzera italiana (USI), Switzerland, Emad Aghajani Software Institute, USI Università della Svizzera italiana, Gabriele Bavota Software Institute, USI Università della Svizzera italiana Pre-print | ||
10:16 3mTalk | gambit – An Open Source Name Disambiguation Tool for Version Control Systems Technical Papers Christoph Gote Chair of Systems Design, ETH Zurich, Christian Zingg Chair of Systems Design, ETH Zurich Pre-print Media Attached | ||
10:19 31mLive Q&A | Discussions and Q&A Technical Papers |
10:00 - 10:50 | Testing and code reviewTechnical Papers / Data Showcase / Registered Reports at MSR Room 2 Chair(s): Jürgen Cito TU Wien and Facebook | ||
10:01 3mTalk | A Traceability Dataset for Open Source Systems Data Showcase Mouna Hammoudi JOHANNES KEPLER UNIVERSITY LINZ, Christoph Mayr-Dorn Johannes Kepler University, Linz, Atif Mashkoor Johannes Kepler University Linz, Alexander Egyed Johannes Kepler University Media Attached | ||
10:04 4mTalk | How Java Programmers Test Exceptional Behavior Technical Papers Diego Marcilio USI Università della Svizzera italiana, Carlo A. Furia Università della Svizzera italiana (USI) Pre-print | ||
10:08 4mTalk | An Exploratory Study of Log Placement Recommendation in an Enterprise System Technical Papers Jeanderson Cândido Delft University of Technology, Jan Haesen Adyen N.V., Maurício Aniche Delft University of Technology, Arie van Deursen Delft University of Technology, Netherlands Pre-print Media Attached | ||
10:12 3mTalk | Does Code Review Promote Conformance? A Study of OpenStack Patches Technical Papers Panyawut Sri-iesaranusorn Nara Institute of Science and Technology, Raula Gaikovina Kula NAIST, Takashi Ishio Nara Institute of Science and Technology Pre-print | ||
10:15 4mTalk | A Replication Study on the Usability of Code Vocabulary in Predicting Flaky Tests Technical Papers Guillaume Haben University of Luxembourg, Sarra Habchi University of Luxembourg, Luxembourg, Mike Papadakis University of Luxembourg, Luxembourg, Maxime Cordy University of Luxembourg, Luxembourg, Yves Le Traon University of Luxembourg, Luxembourg Pre-print Media Attached | ||
10:19 3mTalk | On the Use of Mutation in Injecting Test Order-Dependency Registered Reports Sarra Habchi University of Luxembourg, Luxembourg, Maxime Cordy University of Luxembourg, Luxembourg, Mike Papadakis University of Luxembourg, Luxembourg, Yves Le Traon University of Luxembourg, Luxembourg Pre-print Media Attached | ||
10:22 28mLive Q&A | Discussions and Q&A Technical Papers |
10:50 - 11:10 | Break / Discussion Rooms Technical Papers / Tutorials / MIP Award / FOSS Award / content / Mining Challenge / Hackathon / MSR Awards / Registered Reports / Data Showcase / Shadow PC / Keynotes at MSR Room 1 | ||
11:10 - 12:00 | Welcome Event Technical Papers / Tutorials / MIP Award / FOSS Award / content / Mining Challenge / Hackathon / MSR Awards / Registered Reports / Data Showcase / Shadow PC / Keynotes at MSR Room 1 The MSR welcoming sessions will feature informal networking opportunities for newcomers to meet each other, learn about the MSR conference series, and interact with some established MSR veterans. All are welcome! | ||
17:50 - 18:10 | Break / Discussion Rooms Technical Papers / Tutorials / MIP Award / FOSS Award / content / Mining Challenge / Hackathon / MSR Awards / Registered Reports / Data Showcase / Shadow PC / Keynotes at MSR Room 1 | ||
18:10 - 19:00 | |||
Tue 18 MayDisplayed time zone: Amsterdam, Berlin, Bern, Rome, Stockholm, Vienna change
02:00 - 02:50 | |||
02:50 - 03:10 | Break / Discussion Rooms Technical Papers / Tutorials / MIP Award / FOSS Award / content / Mining Challenge / Hackathon / MSR Awards / Registered Reports / Data Showcase / Shadow PC / Keynotes at MSR Room 1 | ||
03:10 - 04:00 | Time series dataData Showcase / Technical Papers at MSR Room 2 Chair(s): Shane McIntosh University of Waterloo | ||
03:11 3mTalk | AndroCT: Ten Years of App Call Traces in Android Data Showcase Pre-print Media Attached | ||
03:14 4mTalk | Mining Workflows for Anomalous Data Transfers Technical Papers Huy Tu North Carolina State University, USA, George Papadimitriou University of Southern California, Mariam Kiran ESnet, LBNL, Cong Wang Renaissance Computing Institute, Anirban Mandal Renaissance Computing Institute, Ewa Deelman University of Southern California, Tim Menzies North Carolina State University, USA Pre-print | ||
03:18 4mTalk | Escaping the Time Pit: Pitfalls and Guidelines for Using Time-Based Git Data Technical Papers Samuel W. Flint University of Nebraska-Lincoln, Jigyasa Chauhan University of Nebraska-Lincoln, Robert Dyer University of Nebraska-Lincoln Pre-print Media Attached | ||
03:22 4mPaper | On the Naturalness and Localness of Software Logs Technical Papers Pre-print | ||
03:26 4mTalk | How Do Software Developers Use GitHub Actions to Automate Their Workflows? Technical Papers Timothy Kinsman University of Adelaide, Mairieli Wessel University of Sao Paulo, Marco Gerosa Northern Arizona University, USA, Christoph Treude University of Adelaide Pre-print | ||
03:30 30mLive Q&A | Discussions and Q&A Technical Papers |
10:00 - 10:50 | ML and Deep LearningTechnical Papers / Data Showcase / Registered Reports at MSR Room 2 Chair(s): Hongyu Zhang The University of Newcastle | ||
10:01 4mTalk | Fast and Memory-Efficient Neural Code Completion Technical Papers Alexey Svyatkovskiy Microsoft, Sebastian Lee University of Oxford, Anna Hadjitofi Alan Turing Institute, Maik Riechert Microsoft Research, Juliana Franco Microsoft Research, Miltiadis Allamanis Microsoft Research, UK Pre-print Media Attached | ||
10:05 4mResearch paper | Comparative Study of Feature Reduction Techniques in Software Change Prediction Technical Papers Ruchika Malhotra Delhi Technological University, Ritvik Kapoor Delhi Technological University, Deepti Aggarwal Delhi Technological University, Priya Garg Delhi Technological University Pre-print | ||
10:09 4mTalk | An Empirical Study on the Usage of BERT Models for Code Completion Technical Papers Matteo Ciniselli Università della Svizzera Italiana, Nathan Cooper William & Mary, Luca Pascarella Delft University of Technology, Denys Poshyvanyk College of William & Mary, Massimiliano Di Penta University of Sannio, Italy, Gabriele Bavota Software Institute, USI Università della Svizzera italiana Pre-print | ||
10:13 3mTalk | ManyTypes4Py: A benchmark Python dataset for machine learning-based type inference Data Showcase Amir Mir Delft University of Technology, Evaldas Latoskinas Delft University of Technology, Georgios Gousios Facebook & Delft University of Technology Pre-print | ||
10:16 3mTalk | KGTorrent: A Dataset of Python Jupyter Notebooks from Kaggle Data Showcase Luigi Quaranta University of Bari, Italy, Fabio Calefato University of Bari, Filippo Lanubile University of Bari | ||
10:19 3mTalk | Exploring the relationship between performance metrics and cost saving potential of defect prediction models Registered Reports Steffen Herbold University of Göttingen Pre-print | ||
10:22 28mLive Q&A | Discussions and Q&A Technical Papers |
10:50 - 11:10 | Break / Discussion Rooms Technical Papers / Tutorials / MIP Award / FOSS Award / content / Mining Challenge / Hackathon / MSR Awards / Registered Reports / Data Showcase / Shadow PC / Keynotes at MSR Room 1 | ||
11:10 - 12:00 | |||
11:10 50mTutorial | PyDriller 1.0 -- Ready to grow together Tutorials Pre-print |
17:00 - 17:50 | HackathonTechnical Papers / Hackathon at MSR Room 1 Chair(s): Jim Herbsleb Carnegie Mellon University, Audris Mockus The University of Tennessee, Alexander Nolte University of Tartu | ||
17:01 2m | Welcome by the MSR Hackathon Co-Chairs Hackathon Jim Herbsleb Carnegie Mellon University, Audris Mockus The University of Tennessee, Alexander Nolte University of Tartu | ||
17:03 3mTalk | An Exploratory Study of Project Activity Changepoints in Open Source Software Evolution Hackathon | ||
17:06 3mPaper | The Diversity-Innovation Paradox in Open-Source Software Hackathon Mengchen Sam Yong Carnegie Mellon University, Pittsburgh, Pennsylvania, United States, Lavinia Francesca Paganini Federal University of Pernambuco, Huilian Sophie Qiu Carnegie Mellon University, Pittsburgh, Pennsylvania, United States, José Bayoán Santiago Calderón University of Virginia, USA DOI Pre-print | ||
17:09 4mTalk | The Secret Life of Hackathon Code Technical Papers Ahmed Samir Imam Mahmoud University of Tartu, Tapajit Dey Lero - The Irish Software Research Centre and University of Limerick, Alexander Nolte University of Tartu, Audris Mockus The University of Tennessee, Jim Herbsleb Carnegie Mellon University Pre-print | ||
17:13 3mTalk | Tracing Vulnerable Code Lineage Hackathon David Reid University of Tennessee, Kalvin Eng University of Alberta, Chris Bogart Carnegie Mellon University, Adam Tutko University of Tennessee - Knoxville Pre-print | ||
17:16 3mTalk | Building the Collaboration Graph of Open-Source Software Ecosystem Hackathon Pre-print | ||
17:19 1mTalk | The Secret Life of Hackathon Code Hackathon Ahmed Samir Imam Mahmoud University of Tartu, Tapajit Dey Lero - The Irish Software Research Centre and University of Limerick Pre-print | ||
17:20 30mLive Q&A | Discussions and Q&A Technical Papers |
17:00 - 17:50 | |||
17:01 4mTalk | What Code Is Deliberately Excluded from Test Coverage and Why? Technical Papers Andre Hora UFMG Pre-print Media Attached | ||
17:05 3mTalk | AndroR2: A Dataset of Manually-Reproduced Bug Reports for Android apps Data Showcase Tyler Wendland University of Minnesota, Jingyang Sun University of Bristish Columbia, Junayed Mahmud George Mason University, S M Hasan Mansur George Mason University, Steven Huang University of Bristish Columbia, Kevin Moran George Mason University, Julia Rubin University of British Columbia, Canada, Mattia Fazzini University of Minnesota | ||
17:08 3mTalk | Apache Software Foundation Incubator Project Sustainability Dataset Data Showcase Likang Yin University of California, Davis, Zhiyuan Zhang University of California, Davis, Qi Xuan Institute of Cyberspace Security, Zhejiang University of Technology, Hangzhou 310023, China, Vladimir Filkov University of California at Davis, USA | ||
17:11 4mTalk | Leveraging Models to Reduce Test Cases in Software Repositories Technical Papers Pre-print Media Attached | ||
17:15 4mTalk | Which contributions count? Analysis of attribution in open source Technical Papers Jean-Gabriel Young University of Vermont, amanda casari Open Source Programs Office, Google, Katie McLaughlin Open Source Programs Office, Google, Milo Trujillo University of Vermont, Laurent Hébert-Dufresne University of Vermont, James P. Bagrow University of Vermont Pre-print Media Attached | ||
17:19 4mTalk | On Improving Deep Learning Trace Analysis with System Call Arguments Technical Papers Quentin Fournier Polytechnique Montréal, Daniel Aloise Polytechnique Montréal, Seyed Vahid Azhari Ciena, François Tetreault Ciena Pre-print | ||
17:23 27mLive Q&A | Discussions and Q&A Technical Papers |
17:50 - 18:10 | Break / Discussion RoomsTechnical Papers / Tutorials / MIP Award / FOSS Award / content / Mining Challenge / Hackathon / MSR Awards / Registered Reports / Data Showcase / Shadow PC / Keynotes at MSR Room 1 | ||
18:10 - 19:00 | |||
18:10 50mTutorial | Crafting your next MSR paper: suggestions from my (good and bad) experiences Tutorials Massimiliano Di Penta University of Sannio, Italy Pre-print |
Wed 19 MayDisplayed time zone: Amsterdam, Berlin, Bern, Rome, Stockholm, Vienna change
02:00 - 02:50 | |||
02:01 4mTalk | Automatic Part-of-Speech Tagging for Security Vulnerability Descriptions Technical Papers Sofonias Yitagesu Tianjin University, Xiaowang Zhang Tianjin University, Zhiyong Feng Tianjin University, Xiaohong Li TianJin University, Zhenchang Xing Australian National University Pre-print | ||
02:05 4mTalk | Attention-based model for predicting question relatedness on Stack Overflow Technical Papers Jiayan Pei South China University of Technology, Yimin Wu South China University of Technology, Research Institute of SCUT in Yangjiang, Zishan Qin South China University of Technology, Yao Cong South China University of Technology, Jingtao Guan Research Institute of SCUT in Yangjiang Pre-print | ||
02:09 4mTalk | Characterising the Knowledge about Primitive Variables in Java Code Comments Technical Papers Mahfouth Alghamdi The University of Adelaide, Shinpei Hayashi Tokyo Institute of Technology, Takashi Kobayashi Tokyo Institute of Technology, Christoph Treude University of Adelaide Pre-print | ||
02:13 4mTalk | Googling for Software Development: What Developers Search For and What They Find Technical Papers Andre Hora UFMG Pre-print Media Attached | ||
02:17 3mTalk | Evaluating Pre-Trained Models for User Feedback Analysis in Software Engineering: A Study on Classification of App-Reviews Registered Reports Mohammad Abdul Hadi University of British Columbia, Fatemeh Hendijani Fard University of British Columbia Pre-print | ||
02:20 3mTalk | Cross-status Communication and Project Outcomes in OSS Development–A Language Style Matching Perspective Registered Reports Yisi Han Nanjing University, Zhendong Wang University of California, Irvine, Yang Feng State Key Laboratory for Novel Software Technology, Nanjing University, Zhihong Zhao Nanjing Tech Unniversity, Yi Wang Beijing University of Posts and Telecommunications Pre-print | ||
02:23 27mLive Q&A | Discussions and Q&A Technical Papers |
02:50 - 03:10 | Break / Discussion Rooms Technical Papers / Tutorials / MIP Award / FOSS Award / content / Mining Challenge / Hackathon / MSR Awards / Registered Reports / Data Showcase / Shadow PC / Keynotes at MSR Room 1 | ||
03:10 - 04:00 | |||
03:10 50mTutorial | Elasticsearch Full-Text Search Internals Tutorials Philipp Krenn Elastic |
10:00 - 10:50 | DatasetsData Showcase / Technical Papers at MSR Room 1 Chair(s): Sridhar Chimalakonda Indian Institute of Technology Tirupati | ||
10:01 3mTalk | AndroidCompass: A Dataset of Android Compatibility Checks in Code Repositories Data Showcase Sebastian Nielebock Otto-von-Guericke University Magdeburg, Germany, Paul Blockhaus Otto-von-Guericke-University Magdeburg, Germany, Jacob Krüger Otto von Guericke University Magdeburg, Frank Ortmeier Otto-von-Guericke-University Magdeburg, Faculty of Computer Science, Chair of Software Engineering Pre-print Media Attached | ||
10:04 3mTalk | GE526: A Dataset of Open Source Game Engines Data Showcase Dheeraj Vagavolu Indian Institute of Technology Tirupati, Vartika Agrahari Indian Institute of Technology Tirupati, Sridhar Chimalakonda Indian Institute of Technology Tirupati, Akhila Sri Manasa Venigalla IIT Tirupati, India | ||
10:07 3mTalk | Andromeda: A Dataset of Ansible Galaxy Roles and Their Evolution Data Showcase Ruben Opdebeeck Vrije Universiteit Brussel, Ahmed Zerouali Vrije Universiteit Brussel, Coen De Roover Vrije Universiteit Brussel | ||
10:10 3mTalk | The Wonderless Dataset for Serverless Computing Data Showcase Pre-print | ||
10:13 3mTalk | DUETS: A Dataset of Reproducible Pairs of Java Library-Clients Data Showcase Thomas Durieux KTH Royal Institute of Technology, Sweden, César Soto-Valero KTH Royal Institute of Technology, Benoit Baudry KTH Royal Institute of Technology Pre-print | ||
10:16 3mTalk | EQBENCH: A Dataset of Equivalent and Non-equivalent Program Pairs Data Showcase Sahar Badihi University of British Columbia, Canada, Yi Li Nanyang Technological University, Julia Rubin University of British Columbia, Canada | ||
10:19 31mLive Q&A | Discussions and Q&A Technical Papers |
10:50 - 11:10 | Break / Discussion Rooms Technical Papers / Tutorials / MIP Award / FOSS Award / content / Mining Challenge / Hackathon / MSR Awards / Registered Reports / Data Showcase / Shadow PC / Keynotes at MSR Room 1 | ||
11:10 - 12:00 | |||
11:10 12mKeynote | Code review at speed: How can we use data to help developers do code review faster? Keynotes Patanamon Thongtanunam The University of Melbourne | ||
11:22 12mKeynote | To Sustain a Smart, Dependent and Social Software Ecosystem Keynotes Raula Gaikovina Kula NAIST | ||
11:34 8mKeynote | Measure what matters – but don’t be creepy: The ethics of using data about people Keynotes | ||
11:42 18m | Discussions and Q&A Keynotes |
17:00 - 17:50 | Energy, logging, and APIsTechnical Papers at MSR Room 1 Chair(s): Akond Rahman Tennessee Tech University | ||
17:01 3mTalk | S3M: Siamese Stack (Trace) Similarity Measure Technical Papers Aleksandr Khvorov JetBrains, ITMO University, Roman Vasiliev JetBrains, George Chernishev Saint-Petersburg State University, Irving Muller Rodrigues Polytechnique Montreal, Montreal, Canada, Dmitrij Koznov Saint-Petersburg State University, Nikita Povarov JetBrains Pre-print | ||
17:04 4mTalk | Mining the ROS ecosystem for Green Architectural Tactics in Robotics and an Empirical Evaluation Technical Papers Ivano Malavolta Vrije Universiteit Amsterdam, Katerina Chinnappan Vrije Universiteit Amsterdam, Stan Swanborn Vrije Universiteit Amsterdam, The Netherlands, Grace Lewis Carnegie Mellon Software Engineering Institute, Patricia Lago Vrije Universiteit Amsterdam Pre-print Media Attached | ||
17:08 4mTalk | Mining Energy-Related Practices in Robotics Software Technical Papers Michel Albonico UTFPR, Ivano Malavolta Vrije Universiteit Amsterdam, Gustavo Pinto Federal University of Pará, Emitzá Guzmán Vrije Universiteit Amsterdam, Katerina Chinnappan Vrije Universiteit Amsterdam, Patricia Lago Vrije Universiteit Amsterdam Pre-print Media Attached | ||
17:12 3mTalk | Mining API Interactions to Analyze Software Revisions for the Evolution of Energy Consumption Technical Papers Andreas Schuler University of Applied Sciences Upper Austria, Gabriele Anderst-Kotsis Johannes Kepler University, Linz, Austria Pre-print | ||
17:15 4mTalk | Can I Solve it? Identifying the APIs required to complete OSS tasks Technical Papers Fabio Marcos De Abreu Santos Northern Arizona University, USA, Igor Scaliante Wiese Federal University of Technology – Paraná - UTFPR, Bianca Trinkenreich Northern of Arizona Univeristy, Igor Steinmacher Northern Arizona University, USA, Anita Sarma Oregon State University, Marco Gerosa Northern Arizona University, USA Pre-print | ||
17:19 31mLive Q&A | Discussions and Q&A Technical Papers |
17:00 - 17:50 | Change Management and AnalysisTechnical Papers / Registered Reports at MSR Room 2 Chair(s): Sarah Nadi University of Alberta | ||
17:01 4mTalk | Studying the Change Histories of Stack Overflow and GitHub Snippets Technical Papers Pre-print Media Attached | ||
17:05 4mTalk | Learning Off-By-One Mistakes: An Empirical Study Technical Papers Hendrig Sellik Delft University of Technology, Onno van Paridon Adyen N.V., Georgios Gousios Facebook & Delft University of Technology, Maurício Aniche Delft University of Technology Pre-print | ||
17:09 4mTalk | Predicting Design Impactful Changes in Modern Code Review: A Large-Scale Empirical Study Technical Papers Anderson Uchôa Pontifical Catholic University of Rio de Janeiro (PUC-Rio), Caio Barbosa Pontifical Catholic University of Rio de Janeiro (PUC-Rio), Daniel Coutinho Pontifical Catholic University of Rio de Janeiro (PUC-Rio), Willian Oizumi Pontifical Catholic University of Rio de Janeiro (PUC-Rio), Wesley Assunção Pontifical Catholic University of Rio de Janeiro (PUC-Rio), Silvia Regina Vergilio Federal University of Paraná, Juliana Alves Pereira PUC-Rio, Anderson Oliveira PUC-Rio, Alessandro Garcia PUC-Rio Pre-print | ||
17:13 4mTalk | Rollback Edit Inconsistencies in Developer Forum Technical Papers Saikat Mondal University of Saskatchewan, Gias Uddin University of Calgary, Canada, Chanchal K. Roy University of Saskatchewan Pre-print | ||
17:17 3mTalk | Assessing the Exposure of Software Changes: The DiPiDi Approach Registered Reports Pre-print | ||
17:20 4mTalk | On the Use of Dependabot Security Pull Requests Technical Papers Mahmoud Alfadel Concordia Univerisity, Diego Costa Concordia University, Canada, Emad Shihab Concordia University, Mouafak Mkhallalati Concordia University Pre-print | ||
17:24 26mLive Q&A | Discussions and Q&A Technical Papers |
17:50 - 18:10 | Break / Discussion Rooms Technical Papers / Tutorials / MIP Award / FOSS Award / content / Mining Challenge / Hackathon / MSR Awards / Registered Reports / Data Showcase / Shadow PC / Keynotes at MSR Room 1 | ||
18:10 - 19:00 | ClosingTechnical Papers / Tutorials / MIP Award / FOSS Award / content / Mining Challenge / Hackathon / MSR Awards / Registered Reports / Data Showcase / Shadow PC / Keynotes at MSR Room 1 | ||
18:11 23mAwards | MIP Award 2011 MIP Award | ||
18:34 15mLive Q&A | Discussions and Q&A Technical Papers | ||
18:49 8m | MSR 2022 Technical Papers | ||
18:57 3m | Closing by the General Chair and the Program Co-Chair Technical Papers |
Accepted Papers
Scope
The technical track of MSR 2021 solicits high-quality submissions on a wide range of topics related to artificial intelligence (AI), machine learning (ML), and data science (DS) in one or more of the following three main themes.
1. AI/ML/DS and SE
The analysis should aim to improve understanding of development processes and practices or aid in the development of new techniques or models to support software developers. This includes (but is not limited to) analysis or models for:
- commits,
- execution traces and logs,
- interaction data,
- code review data,
- natural language artifacts,
- software licenses and copyrights,
- app store data,
- programming language features,
- release information,
- CI logs,
- deployment and delivery,
- test data,
- runtime information,
- software ecosystems,
- defect and software quality data,
- human and social aspects of development,
- development process,
- energy profile data.
2. New techniques, tools, and models.
The techniques, tools, and models should facilitate new ways to mine, analyze, or model software data. A submission could include (but is not limited to) techniques, tools, or models to:
- capture new forms of data,
- integrate data from multiple sources,
- visualize software data,
- model software data,
- solve SE problems,
- improve AI/ML/DS.
3. Considerations related to AI/ML/DS and SE.
These submissions should reflect on the current state-of-the-art research methods or current practices in mining, analyzing, or modeling software data. These submissions can also propose new research methods or guidelines. This theme includes topics such as (but not limited to)
- privacy of collected data,
- ethics of mining, analyzing, or modelling software data,
- biases in software data, analyses, and tools,
- fairness in software data, analyses, and tools,
- Replication studies.
Call for Papers
The Mining Software Repositories (MSR) conference is the premier conference for data science, machine learning, and artificial intelligence in software engineering. The goal of the conference is to improve software engineering practices by uncovering interesting and actionable information about software systems and projects using the vast amounts of software data such as source control systems, defect tracking systems, code review repositories, archived communications between project personnel, question-and-answer sites, CI build servers, and run-time telemetry. Mining this information can help to understand software development and evolution, software users, and runtime behavior; support the maintenance of software systems; improve software design/reuse; empirically validate novel ideas and techniques; support predictions about software development; and exploit this knowledge in planning future development. The goal of this two-day international conference is to advance the science and practice of software engineering with data-driven techniques. The 18th International Conference on Mining Software Repositories will be held virtually during the week of May 17, 2021.
Types of Technical Track Submissions
We accept both full (10 pages plus 2 additional pages of references) and short (4 pages plus 1 additional page of references) papers. Furthermore, in order to facilitate the reviewing process of your paper’s contribution, you should select one of the following paper categories:
1. Research Paper
Full research papers are expected to describe new methodologies and/or provide novel research results, and should be evaluated scientifically. While a high degree of technical rigor is expected for long papers, short research papers should discuss controversial issues in the field, or describe interesting or thought-provoking ideas that are not yet fully developed. Accepted short papers will be presented in a short lightning talk. Relevant review criteria:
- soundness of approach
- relevance to software engineering
- clarity of relation with related work
- quality of presentation
- quality of evaluation [for long papers]
- ability to replicate [for long papers]
- novelty
2. Practice Experience
MSR encourages the submission of papers that report on both positive and negative experiences of applying software analytics strategies in an industry/open source organization context. Adapting existing algorithms or proposing new algorithms or approaches for practical use are considered a plus. Relevant review criteria:
- quality of empirical evaluation
- explicit discussion on the usefulness/impact of the approach in practice
- explicit discussion of any adaptations required by the application of existing/new approach in practice
- quality of presentation • relevance to software engineering • clarity of relation with related work
3. Reusable Tool
MSR actively promotes and recognizes the creation and use of tools that are designed and built not only for a specific research project, but for the MSR community as a whole. Those tools enable other researchers to jumpstart their own research efforts, and also enable reproducibility of earlier work. Reusable Tool papers can be descriptions of tools built by the authors that can be used by other researchers, and/or descriptions of the use of tools built by others to obtain specific research results. Relevant review criteria:
- evaluation of usefulness/reusability of the tool [for long papers]
- novelty
- quality of presentation (details on tool’s internals, usage, etc.)
- relevance to software engineering
- clarity of relation with related work
- availability of the tool, clear installation instructions and example data set that allow the reviewers to run the tool
Submission Process
Submissions to the Technical Track can be made via the submission site by the submission deadline. We encourage authors to upload their paper info early (the PDF can be submitted later) to properly enter conflicts for double-blind reviewing. All submissions must adhere to the following requirements:
- Submissions must not exceed the page limit (10 pages plus 2 additional pages of references for full papers; 4 pages plus 1 additional page of references for short papers). The page limit is strict, and it will not be possible to purchase additional pages at any point in the process (including after acceptance).
- Submissions must conform to the IEEE formatting instructions IEEE Conference Proceedings Formatting Guidelines (title in 24pt font and full text in 10pt type, LaTeX users must use
\documentclass[10pt,conference]{IEEEtran}
without including thecompsoc
orcompsocconf
options). - Submissions must not reveal the authors’ identities. The authors must make every effort to honor the double-blind review process. In particular, the authors’ names must be omitted from the submission and references to their prior work should be in the third person. Further advice, guidance, and explanation about the double-blind review process can be found in the Q&A page for ICSE 2021.
Any submission that does not comply with these requirements is likely to be desk rejected by the Technical Track PC Chairs without further review. In addition, by submitting to the MSR Technical Track, the authors acknowledge that they are aware of and agree to be bound by the following policies:
- The ACM Policy and Procedures on Plagiarism and the IEEE Plagiarism FAQ. In particular, papers submitted to MSR 2021 must not have been published elsewhere and must not be under review or submitted for review elsewhere whilst under consideration for MSR 2021. Contravention of this concurrent submission policy will be deemed a serious breach of scientific ethics, and appropriate action will be taken in all such cases (including immediate rejection and reporting of the incident to ACM/IEEE). To check for double submission and plagiarism issues, the chairs reserve the right to (1) share the list of submissions with the PC Chairs of other conferences with overlapping review periods and (2) use external plagiarism detection software, under contract to the ACM or IEEE, to detect violations of these policies.
- The authorship policy of the ACM and the authorship policy of the IEEE.
Upon notification of acceptance, all authors of accepted papers will be asked to fill a copyright form and will receive further instructions for preparing the camera-ready version of their papers. At least one author of each paper is expected to register and present the paper at the MSR 2021 conference. All accepted contributions will be published in the electronic proceedings of the conference.
A selection of the best papers will be invited to an Empirical Software Engineering (EMSE) Special Issue. The authors of accepted papers that show outstanding contributions to the FOSS community will have a chance to self-nominate their paper for the MSR FOSS Impact Paper Award. Please note that providing a replication package is strongly recommended even in double-blind submissions, since not providing one effectively hinders the peer-review process. Since access to data and scripts is essential during peer review, we strongly recommend to archive data sets on online archival sites such as dropbox.com, zenodo.org or figshare.com (Instructions available in Open Science Policy below). The latter two even allow to receive a DOI and hence become citable.
Submission Link
Technical papers must be submitted through HotCRP: https://msr2021-technical.hotcrp.com/
Shadow PC
This year we are trialing a process called Shadow PC. In this process, we are going to recruit and train the next generation of MSR researchers on reviewing MSR papers. The primary audience would be PhD students and Post Docs who have not been in the MSR PC before. They will review papers submitted to MSR and have a parallel program committee. We will have experienced MSR researchers look at the reviews and comments and give feedback to the Shadow PC. This has been done in other venues like EuroSys and USENIX.
Since this is the first time, we are requesting authors to volunteer their paper to be reviewed by the Shadow PC. This is purely a learning experience, and the reviews from the Shadow PC will not be visible to the regular PC and will not impact the decision of the paper in MSR. The authors may request the reviews from the Shadow PC. The reviews and comments will be in a completely different HotCRP installation to prevent any cross-over. Therefore, we are requesting authors to please volunteer their papers. There will be an option in the submission page for this. When you volunteer you are not only helping future generations of MSR researchers become better reviewers, but you also will get more feedback on your work.
Important Dates
- Abstract Deadline: Tuesday, January 5, 2021, 23:59 AOE
- Papers Deadline: Tuesday, January 12, 2021, 23:59 AOE (No deadline extension or grace periods will be provided. Please plan accordingly)
- Author Response Period: February 10 - 12, 2021
- Author Notification: Monday, February 22, 2021
- Camera Ready: Monday, March 22, 2021, 23:59 AOE
Open Science Policy
Openness in science is key to fostering progress via transparency, reproducibility and replicability. Our steering principle is that all research output should be accessible to the public and that empirical studies should be reproducible. In particular, we actively support the adoption of open data and open source principles. The following guidelines are recommendations and not mandatory. Your choice to use open science or not will not affect the review process for your paper. However, to increase reproducibility and replicability, we encourage all contributing authors to disclose:
- the source code of relevant software used or proposed in the paper, including that used to retrieve and analyze data
- the data used in the paper (e.g., evaluation data, anonymized survey data, etc.)
- instructions for other researchers describing how to reproduce or replicate the results
Already upon submission, authors can privately share their anonymized data and software on preserved archives such as Zenodo or Figshare (tutorial available here – please make sure that any links shared during peer review are anonymized). Zenodo accepts up to 50GB per dataset (more upon request). There is no need to use Dropbox or Google Drive. Once accepted, an option can be toggled to publish the data and scripts with an official DOI. Zenodo and Figshare accounts can easily be linked with GitHub repositories to automatically archive software releases. In the unlikely case that authors need to upload terabytes of data, Archive.org may be used.
After acceptance, we encourage authors to self-archive pre-prints of their papers in open, preserved repositories such as arXiv.org. This is legal and allowed by all major publishers including ACM and IEEE, and it lets anybody in the world reach your paper. Note that you are usually not allowed to self-archive the PDF of the published article (that is, the publisher proof or the Digital Library version). Instead, use the manuscript with reviewer comments addressed, but before applying the camera-ready instructions and templates. Feel free to contact the MSR 2021 PC or proceedings chairs for more details. We recognise that anonymising artifacts such as source code is more difficult than preserving anonymity in a paper. We ask authors to take a best effort approach to not reveal their identities. We will also ask reviewers to avoid trying to identify authors by looking at commit histories and other such information that is not easily anonymised. Authors wanting to share GitHub repositories may want to look into using https://anonymous.4open.science/ which is an open source tool that helps you to quickly double-blind your repository. Please note that the success of the open science initiative depends on the willingness (and possibilities) of authors to disclose their data and that all submissions will undergo the same review process independent of whether or not they disclose their analysis code or data. We encourage authors who cannot disclose industrial or otherwise non-public data, for instance due to non-disclosure agreements, to provide an explicit (short) statement in the paper.