Top Banner
Five Days of Empirical Software Engineering: The PASED Experience Massimiliano Di Penta Giuliano Antoniol Daniel M. Germán Yann-Gaël Guéhéneuc Bram Adams
26
Welcome message from author
This document is posted to help you gain knowledge. Please leave a comment to let me know what you think about it! Share it to your friends and learn new things together.
Transcript
Page 1: ICSE12 SEE.ppt

Five Days of Empirical Software Engineering:

The PASED Experience

Massimiliano Di PentaGiuliano AntoniolDaniel M. Germán

Yann-Gaël GuéhéneucBram Adams

Page 2: ICSE12 SEE.ppt

MotivationEmpirical background important for graduate students

Courses on statistics insufficient to provide such a background

(Most) University curricula could not afford do have specific courses

Some exceptions (there are others for sure!):

• Easterbrook’s CSC2130: Empirical Research Methods in Software Engineering at the University of Toronto (2009)

• Herbsleb’s 08-803: Empirical Methods for Socio-Technical Research at CMU (2010)

• Dewayne E. Perry course, Univ. of Texas at Austin

Page 3: ICSE12 SEE.ppt

So students need that!

Page 4: ICSE12 SEE.ppt

General Info About the School

• Ecole Polytechnique de Montréal, June 2011

• Funded by MITACS

• low fee for students $250 all included

• 44 participants, 9 countries and 25 different institutions

• More on http://pased.soccerlab.polymtl.ca

Page 5: ICSE12 SEE.ppt

Learning Objectives

1. plan and conduct software engineering experiments with human subjects and collect related data

2. plan and conduct software engineering studies involving the mining of data from (un)structured software repositories

3. build prediction and classification models from the collected data, and to use these models

Page 6: ICSE12 SEE.ppt

Challenges

ChoosingTopics

Dealing withheterogeneous

participants

Combiningtheory and

practice

Page 7: ICSE12 SEE.ppt

What Topics?

Page 8: ICSE12 SEE.ppt

What Topics?

Planning thestudy

Page 9: ICSE12 SEE.ppt

What Topics?Getting the

data

Planning thestudy

Page 10: ICSE12 SEE.ppt

What Topics?Getting the

data

Planning thestudy

Analyzingresults

Page 11: ICSE12 SEE.ppt

...that’s too much!

Only 5 days....

Page 12: ICSE12 SEE.ppt

The Approach

• Learn by example and by doing format

• Experiment design principles and statistics introduced by presenting cases from studies in literature

• Practical application of theoretical concepts during labs

• Course material and laboratory packages available online, including course videos

Page 13: ICSE12 SEE.ppt

School Content

AMMining

SoftwareArchives

Exp.Design

Textmining

StatisticalAnalysis

PredictorModels

PM

Keynote Keynote Keynote Keynote Keynote

PM Handsonlab

Handsonlab

Handsonlab

Handsonlab

Handsonlab

Page 14: ICSE12 SEE.ppt

Learning by doing...

Page 15: ICSE12 SEE.ppt

Running example I

• Use of UML Stereotypes in comprehension and maintenance tasks

• Filippo Ricca, Massimiliano Di Penta, Marco Torchiano, Paolo Tonella, Mariano Ceccato: How Developers' Experience and Ability Influence Web Application Comprehension Tasks Supported by UML Stereotypes: A Series of Four Experiments. IEEE Trans. Software Eng. 36(1): 96-118 (2010)

• Filippo Ricca, Massimiliano Di Penta, Marco Torchiano, Paolo Tonella, Mariano Ceccato: The Role of Experience and Ability in Comprehension Tasks Supported by UML Stereotypes. ICSE 2007: 375-384

• In the following briefly referred as “Conallen”

from theschool

Page 16: ICSE12 SEE.ppt

Experiment Design

Group 1 Group 2 Group 3 Group 4

Lab 1Sys A Sys A Sys B Sys B

Lab 2Sys B Sys B Sys A Sys A

from theschool

Page 17: ICSE12 SEE.ppt

Data format: examplefrom theschool

Page 18: ICSE12 SEE.ppt

Boxplots: Conallenfrom theschool

Page 19: ICSE12 SEE.ppt

Paired analysis: example>wilcox.test(F.Conallen,F.UML, paired=TRUE,

alternative="greater")

Wilcoxon signed rank test with continuity correction

data: F.Conallen and F.UML

V = 138, p-value = 0.04354

alternative hypothesis: true location shift is greater than 0

ID F.Conallen F.UML

T20 0.74 0.74

T21 0.74 0.51

T22 0.7 0.29

T24 0.88 0.62

T25 0.75 0.8

T26 0.66 0.39

T27 0.35 0.51

T28 0.62 0.59

T29 0.57 0.68

T30 0.73 0.43

T32 0.74 0.56

• Must have data in a paired format• or you can use a proper R script

• Need to remove subjects that took part to one lab only• For parametric statistics just replace wilcox.test with t.test

from theschool

Page 20: ICSE12 SEE.ppt

Hands on Labs

• Mining software repository challenge: extract interesting facts from git

• Experiment design: groups working together on designing a study

• Data analysis: text mining, analyze working data sets of previous experiments and build bug predictors

• Working data sets from previous experiments, PROMISE data sets

• Tools: R and Weka

Page 21: ICSE12 SEE.ppt

Lab Script Example#UNPAIRED ANALYSIS

#Analysis of single experiment#Mann-Whitneyattach(tbn)wilcox.test(Correct[Fit=="yes"],Correct[Fit=="no"],paired=FALSE,alternative="greater")

attach(ttrento)wilcox.test(Correct[Fit=="yes"],Correct[Fit=="no"],paired=FALSE,alternative="greater")

attach(tphd)wilcox.test(Correct[Fit=="yes"],Correct[Fit=="no"],paired=FALSE,alternative="greater")

#All dataattach(t)wilcox.test(Correct[Fit=="yes"],Correct[Fit=="no"],paired=FALSE,alternative="greater")

#Exercises: # 1) perform a two-tailed test# 2) can t-test be applied instead of Wilcoxon test? test for data normality using the Wilk-Shapiro test# 3) repeat the analysis using the t-test?# 4) repeat the analysis for the time dependent variable

Page 22: ICSE12 SEE.ppt

Calibrating Courses to Participants’ Profiles

Excellent

Good

Basic

None

0 10 20 30

1

23

21

1

Statistical Analyses

Excellent

Good

Basic

None

0 10 20 30

2

18

20

4

Empirical Sw Engineering

Excellent

Good

Basic

None

0 10 20 30

1

24

16

4

Mining Sw repositories

Excellent

Good

Basic

None

0 10 20 30

7

25

12

1

Machine learning

Page 23: ICSE12 SEE.ppt

FeedbacksLonger Labs

Guidelines onwhat not to do

How to write empiricalpapers

Tutorialson tools

Page 24: ICSE12 SEE.ppt

Acknowledgments• Lecturers (other than the paper authors):

• Ahmed E. Hassan, Queen’s University, Canada

• Andrian Marcus, Wayne State University, USA

• Keynote Speakers

• Gail Murphy, University of British Columbia, USA

• Prem Devanbu, UC Davis, USA

• Alain Picard, Benchmark Consulting Services Inc., Canada

• Maria Codipiero, Peter Colligan, Kal Murtaia, SAP, Canada

• Marc-André Decoste, Google Montréal, Canada

• Student volunteers

• The attendees!

• MITACS (http://www.mitacs.ca/)

Page 25: ICSE12 SEE.ppt

Conclusions

Page 26: ICSE12 SEE.ppt