SlideShare a Scribd company logo
1 of 58
Download to read offline
Reflections on 40+ years of software
 engineering research and beyond
          an insider's view
           Carlo Ghezzi
           DeepSE Group
           Dipartimento di Elettronica e Informazione
           Politecnico di Milano, Italy
           carlo.ghezzi@polimi.it



                      ICSE 2009                         1
Outline of the talk
  A journey through 40+ years of software engineering
  research, mostly observed through ICSE
  What did we produce?
  How can their success (impact) be "defined"?
  How can it be measured?
  What can we learn from past?
  Can we do better?
  Will the paradigms followed in the past survive in the
  future?

                            ICSE 2009                      2
Part 1: Looking backwards


                         Where does ICSE come
                         from?
                         What can we learn by
                         mining ICSE data?




           ICSE 2009                        3
A bit of history
  The roots are in the NATO meetings in Garmisch-Partenkirchen (1968)
  and Roma (1969)
  ICSE started in 1975, this is the 31-st
   –  first conference called ICSE is actually ICSE2
         ICSE1 was NCSE---1st National Conf on Software Engineering, Sept 1975
   –  became annual in 1987
  TSE started in March 1975
  ESEC started in 1987, FSE started in 1993, they joined in 1997
  TOSEM started in January 1992



                                   ICSE 2009                                     4
Mining ICSE data
  International Conference on Software Engineering
   –  How much international?
   –  How are the different engineering research stakeholders
      (academia, industry) represented?
         scientific leadership, contributors
  Paper demographics
   –  Is ICSE an attractive venue?
  Research areas
   –  How did they evolve?

                                 ICSE 2009                  5
ICSE PCs
              avg
              1
              12
              32
              55




  ICSE 2009         6
ICSE PCs: industry/academia




            ICSE 2009         7
average 44




ICSE 2009           8
Submitted/accepted




       ICSE 2009     9
ICSE papers: geographical distribution



                                         average
                                         1%
                                         10%
                                         24%
                                         65%




                 ICSE 2009               10
Papers: industry vs. academia

                                               average
                                               AC 70%
                                               IND 30%




average 1976-1994          average 1995-2008
      AC 56%                     AC 83%
     IND 44%                    IND 17%
                    ICSE 2009                            11
Preliminary findings
ICSE is nurtured by an international community

                NorthAmerica   Europe      AsiaPacific   RestOfWorld
PC        all       55%             32%         12%           1%
       -5 yrs       46%             40%         11%           3%


                NorthAmerica   Europe      AsiaPacific   RestOfWorld
papers all          65%             24%         10%           1%
       -5 yrs       60%             27%         11%           2%

High number of submissions, high selection rate
There is a consensus that ICSE is a prestigious venue
                               ICSE 2009                               12
Preliminary findings
  Industry participation decreased over time
   –  research labs of large corporations disappeared
   –  ICSE became more research oriented
        initially: tool fair
        then: tutorials
        now: workshops
  –  now mostly co-authored papers (academia/industry)




                               ICSE 2009                 13
Topics
            Topic              %

Testing&Analysis               17

Design&Software Architecture   17

Specification&Verification     15

Tools&SDEs                     11

Management&Costs               9

Empirical Studies              7

Existing Systems               6

Others (13 topics)             18




                                    ICSE 2009   14
Topics
            Topic              %

Testing&Analysis               17

Design&Software Architecture   17

Specification&Verification     15

Tools&SDEs                     11

Management&Costs               9

Empirical Studies              7

Existing Systems               6

Others (13 topics)             18




                                    ICSE 2009   15
Topics
            Topic              %

Testing&Analysis               17

Design&Software Architecture   17

Specification&Verification     15

Tools&SDEs                     11

Management&Costs                9

Empirical Studies               7

Existing Systems                6

Others (13 topics)             18




                                    ICSE 2009   16
Topics
            Topic              %

Testing&Analysis               17

Design&Software Architecture   17

Specification&Verification     15

Tools&SDEs                     11

Management&Costs               9

Empirical Studies              7

Existing Systems               6

Others (13 topics)             18




                                    ICSE 2009   17
Topics

            Topic              %

Testing&Analysis               17

Design&Software Architecture   17

Specification&Verification     15

Tools&SDEs                     11

Management&Costs               9

Empirical Studies              7

Existing Systems               6

Others (13 topics)             18




                                    ICSE 2009   18
Topics

            Topic              %

Testing&Analysis               17

Design&Software Architecture   17

Specification&Verification     15

Tools&SDEs                     11

Management&Costs               9

Empirical Studies              7

Existing Systems               6

Others (13 topics)             18




                                    ICSE 2009   19
Topics

            Topic              %

Testing&Analysis               17

Design&Software Architecture   17

Specification&Verification     15

Tools&SDEs                     11

Management&Costs               9

Empirical Studies              7

Existing Systems               6

Others (13 topics)             18




                                    ICSE 2009   20
Programming languages




         ICSE 2009      21
Some findings
  Large variety of topics
  Trend from "methodology oriented", informal approaches to
  systematic and formally-based approaches
  Increased attention on empirical evaluations
  Still little on paradigmatic examples or domain-specific
  approaches




                           ICSE 2009                      22
Can we identify our best products?
  Which criteria to follow?
  –  Quantitative
        Citation count
        Downloads count
        Others?
  –  Qualitative
        Most influential -10 years?
        Others?

                            ICSE 2009     23
Quantitative criteria
  Citation count
   –  date of count Jan-March 2009
   –  source Google Scholar
   –  version (conference/journal) both (if j expands c)
   –  precision ???
  Number of downloads
   –  date, period Jan-March 2009, 12 months
   –  source ACM digital library
   –  version conference
   –  precision ???
                             ICSE 2009                     24
Outstanding products based on citations
1.  Program slicing (ICSE'81) 2120
2.  Statemate: a working environment for the development of complex
    reactive systems (ICSE'88) 1170
3.  N degrees of separation: multi-dimensional separation of concerns
    (ICSE'99) 997
4.  Bandera : extracting finite-state models from Java source code
    (ICSE'00) 791
5.  Software processes are software too (ICSE'87) 688
6.  Managing the development of large software systems: concepts and
    techniques (ICSE'87) 667
7.  Executable object modeling with statecharts (ICSE'96) 622
8.  Designing software for ease of extension and contraction (ICSE'78) 605
                                 ICSE 2009                           25
Citations vs expert judgment
               most influential papers
1.  Program slicing (ICSE'81) 2120
2.  Statemate: a working environment for the development of complex
    reactive systems (ICSE'88) 1170
3.  N degrees of separation: multi-dimensional separation of concerns
    (ICSE'99) 997
4.  Bandera : extracting finite-state models from Java source code
    (ICSE'00) 791
5.  Software processes are software too (ICSE'87) 688
6.  Managing the development of large software systems: concepts and
    techniques (ICSE'87) 667
7.  Executable object modeling with statecharts (ICSE'96) 622
8.  Designing software for ease of extension and contraction (ICSE'78) 605
                                          ICSE 2009                  26
MIP vs citations



                   380




      ICSE 2009          27
How do we compare with others?




   Citations of most influential papers

                      ICSE 2009           28
#Downloads (-12 months)?
From ACM DL
    1.  Automated Support for Development, Maintenance, and Testing in the
        Presence of Implicit Control Flow (ICSE'04): 781---13 cit
    2.  Managing the development of large software systems: concepts and
        techniques (ICSE'87): 705---667 cit
    3.  New directions on agile methods : a comparative analysis (ICSE'03):
        567---161 cit
    4.  Program slicing (ICSE'81): 520---2120 cit
    5.  Quantitative evaluation of software quality ICSE'76: 495---212 cit
    6.  Static detection of cross-site scripting vulnerabilities ICSE'08---396---8 cit
    7.  Analysis and testing of Web applications ICSE'01: 375---195 cit
    8.  Aspect-oriented programming and modular reasoning: ICSE'05: 300---32cit



                                        ICSE 2009                                    29
Findings: #citations vs. #downloads
  Time factor (both #citations an #downloads)
   –  my Jan-March numbers are different from the current ones
   –  how can you compare #cit of a 20 years and a 2 years old
      paper?
  No significant correlation between citations and downloads
  #citations and #downloads express different merits
        example: a highly downloaded paper may be useful in practice,
        but does not inspire further research
  #downloads demonstrates longevity of ICSE contributions
        3 out of the top 5 highest #downloads are papers over 20 years
        old, one is more than 30
                              ICSE 2009                            30
Findings: #citations vs. expert judgment
  Citations and expert judgment are better correlated,
  according to the ICSE experience
   –  most highly cited papers have been recognized
      as influential
   –  …but several influential papers have low citations




                         ICSE 2009                     31
What do the numbers tell us?
          Do they indicate impact?
  Does the number of paper
  citations indicate impact of
  a piece of work?
  Do citations of papers in a
  journal/proceedings
  indicate impact of a
  venue?
  Do citations of one's
  research measure impact
  of that research?
                             ICSE 2009   32
Part 2: A side road on numbers/citations
                        Joint Committee on Quantitative
                        Assessment of Research
                        Report CITATION STATISTICS
                        Intl Mathematical Union (IMU)
                        in cooperation with Intl Council of
                        Industrial and Applied Math (ICIAM)
                        and the Institute of Mathematical
                        Statistics (IMS), 6/12/2008

                        R. Adler, J. Ewing, P. Taylor (Eds.)

                  ICSE 2009                               33
A side road on numbers/citations
                    Informatics Europe
                    Report RESEARCH EVALUATION
                    FOR COMPUTER SCIENCE
                    Viewpoint article in CACM, April 2009
                    Eds. B. Meyer, C. Choppy, J.
                    Staunstrup, J. van Leewen (Eds.)

                    Also D. Parnas, CACM nov 2007
                    STOP THE NUMBERS GAME


              ICSE 2009                              34
Facts
  Numbers increasingly used to assess/rank
   –  papers
   –  journals and conferences---the (in)famous impact factor
   –  individuals
  Motivations
   –  different stakeholders need to evaluate research
   –  peer review can be subjective
   –  peer review is expensive
   –  numbers are simple and objective simplistic and misleading
                           ICSE 2009                       35
Findings
  Much of the modern bibliometrics is flawed (statistics
  improperly used)
  Objectivity and accuracy illusory
   –  the meaning of a citation can be even more subjective
      than peer review
  Sole reliance on citation data provides incomplete and
  shallow understanding of research
   –  only valid if reinforced by other judgments
     numbers are not inherently superior to and
       cannot substitute complex judgement
                            ICSE 2009                         36
Why are citations counted?
From Thomson Scientific
     "The value of information is determined by those who
     use it…the better way to measure the quality of work is
     by measuring the impact it makes on the community at
     large"

The statement makes an implicit equivalence assumption
                   #citations = impact



                            ICSE 2009                          37
Citations vs impact
  Is the equality
                     #citations = impact
   justified? (When) does it work?
  It presupposes a highly rational and uniform model of
  reference giving, but this is NOT true




                            ICSE 2009                     38
The complex sociology of citations
Average citations per article




citation practices differ substantially among disciplines
                           ICSE 2009                        39
The complex sociology of citations
  Citations are of two main kinds
   –  REWARD
         intellectual dependency on the cited
  –  RHETORICAL
         a way to carry out a scientific conversation
          –  reference included to show that topic is of interest to someone else or
             to prove that the author knows the literature
          –  the cited explains some result, not necessarily of the cited author
          –  the cited represents another approach, or is an example of…




                                  ICSE 2009                                      40
The complex sociology of citations
  Most citations are rhetorical
  Reward citations can be of many kinds
   –  currency, negative credit, operational information,
      persuasiveness, positive credit, reader alert, social
      consensus
  Obliteration effect
   –  incorporated into other work, which is cited



                             ICSE 2009                        41
An example
  The impact of MESA on system design, by HC Lauer, EH
  Satterthwaite, ICSE'79
   –  60 cit (highest of '79), 19 downloads … one of the few
      papers on MESA
                                BUT
  Its impact evidenced by indirection
   –  Implementing remote procedure calls, by Birrell&Nelson,
      ACM TOCS 1984, 1840 citations
        explicitly states its debt to MESA

                                ICSE 2009                   42
If impact is the question,
     are (citation) numbers the answer?
  Citation data and statistics provide some valuable
  information
   –  very high citation numbers tell more than smaller
  but only a limited and incomplete view of research quality
  Statistics derived from them often poorly understood and
  misused
  Expert judgment cannot be eliminated
  Research too important to measure only with a coarse
  tool
                            ICSE 2009                          43
Lost in the journey?
                                We started by looking back at
                                SE research through ICSE as
                                a magnifying lens
                                We tried to assess what we
                                accomplished by identifying
                                our impactful results
                                We got trapped by the
                                number game

Should we look for better assessment methods? Why?
                       ICSE 2009                          44
Should we care?
  Our social responsibility is to care about impact of research
   –  understanding
   –  measuring
   –  improving
  If we don't, others will do
   –  governments, funding bodies
         accountability, "return" to taxpayer or donor
  –  universities, schools, departments
         competition (CS versus other sciences, SE versus other
         CS areas)
                                   ICSE 2009                      45
Part 3: Where do we go from here?

                           Towards
                              understanding
                              measuring
                              improving
                           impact of SE research




               ICSE 2009                           46
Do we understand what impact is?
         How can we measure it?
  Not quite
  A notable counter-example
   –  IMPACT project
        aims at demonstrating impact of SE research on practice
        focuses on specific research areas in SE
  –  an initiative of ACM SIGSOFT
        the SIGSOFT Impact project.webarchive
  –  area leaders responsible for research
  –  backed by sound science history methodology
                             ICSE 2009                            47
Published results
    Osterweil, L.J., Ghezzi, C., Kramer, J., Wolf, A.L. Determining the Impact of Software
    Engineering Research on Practice . IEEE Computer, March 2008.
    Estublier, J., Leblang, D., Hoek, A., Conradi, R., Clemm, G., Tichy, W., and Wiborg-Weber,
    D.Impact of software engineering research on the practice of software configuration
    management. ACM TOSEM, Oct. 2005.
    Ryder, B. G., Soffa, M. L., and Burnett, M. The impact of software engineering research on
    modern progamming languages. ACM TOSEMl, Oct. 2005.
    Emmerich, W., Aoyama, M., Sventek, J. The Impact of Research on the Development of
    Middleware Technology. ACM TOSEM, August 2008.
    Clarke, L. A. and Rosenblum, D. S. Historical Perspective on Runtime Assertion Checking in
    Software Development. SIGSOFT SEN, March 2008.
    Rombach, D., Ciolkowski, M., Jeffery, R., Laitenberger, O., McGarry, F., Shull, S. Impact of
    research on practice in the field of inspections, reviews and walkthroughs: learning from
    successful industrial uses. In ACM SIGSOFT SEN, November 2008.


                                            ICSE 2009                                          48
Some key findings/confirmations
  SE research has had impact on SE practice
  Maturation needs time
   –  15-20 years between first publication of an idea and
      widespread availability in products
  Substantially different mechanisms have been successful at
  causing impact
   –  people are key to knowledge transmission
       people movement
       almost all impact traces lead back to PhD theses



                              ICSE 2009                    49
Problems with impact definition/measurement
   Scholarly assessment of impact of SE research on practice
   is a difficult and expensive
    –  it is research
   Substantially different methods are needed to assess
   impact of
    –  journals/conference
    –  individual researchers/papers



                            ICSE 2009                      50
Obstacles to impact
  The dominant reward/recognition system based on
  publication count does not favor impact
   –  from D. Parnas, CACM 2007
          "it encourages superficial research"
          "it encourages small, insignificant results"
          "it rewards publication of half-baked ideas"
          "it slow scientific progress: to increase their score,
          researchers must avoid tackling the tough problems and
          problems that will require years of dedicated work and instead
          work on easier ones"


                                 ICSE 2009                             51
Obstacles to impact
  Building on top of others' work too often too difficult or
  impossible
   –  TOSEM papers 2001-2006
         60% refer to a tool, only 20% installable
  ICSE'06 paper by Zannier, Melnik, Maurer evaluates ICSE
  empirical studies
         complete absence of replicated studies
         self-evaluations dominate empirical studies



                                ICSE 2009                      52
Conclusions: on impact
  Research on impact of SE research must continue and
  should become a community effort
  As a community we need to get to a common
  understanding and articulate methods for definition and
  measurement of impact for
   –  journals/conferences
   –  individual research
        define impactful research products beyond paper count
  We must be aware of risks
  –  be rigorous, avoid being incestuous
                              ICSE 2009                         53
Conclusions: on enhancing research flow
  Define community challenges and support repeated
  experiments
   –  catalogs of case-studies on which competing
      approaches can be applied and compared
  Favor research products backed by tools, but verify that the
  claimed results of their use can be replicated
  Use Internet facilities for community support
   –  research-as-service
  Reward successful reuse of one's research rather than
  paper citation
                            ICSE 2009                        54
The end of the journey



     In 40+ years we have gone far




          ICSE 2009                  55
Beginning of a new journey


        We have been successful, but we
        can do more and better
        Towards impact-aware research
        Can the ICSE community take the
        lead of this?




            ICSE 2009                     56
Acknowledgements
  L. Osterweil, J. Kramer, A. Wolf and IMPACT project
  members for their insights into impact of research
  Andrea Mocci, Mattia Monga, for helping me in mining data




                           ICSE 2009                      57
Thank you!!!



    ICSE 2009   58

More Related Content

Similar to ICSE 2009 keynote

Industry-Academia Communication In Empirical Software Engineering
Industry-Academia Communication In Empirical Software EngineeringIndustry-Academia Communication In Empirical Software Engineering
Industry-Academia Communication In Empirical Software EngineeringPer Runeson
 
Week 01-intro se
Week 01-intro seWeek 01-intro se
Week 01-intro seNguyen Tran
 
[2016/2017] RESEARCH in software engineering
[2016/2017] RESEARCH in software engineering[2016/2017] RESEARCH in software engineering
[2016/2017] RESEARCH in software engineeringIvano Malavolta
 
A proposal for a new common process scope for AutomotiveSPICE: Six reasons fo...
A proposal for a new common process scope for AutomotiveSPICE: Six reasons fo...A proposal for a new common process scope for AutomotiveSPICE: Six reasons fo...
A proposal for a new common process scope for AutomotiveSPICE: Six reasons fo...Luigi Buglione
 
ISO 29110 Software Quality Model For Software SMEs
ISO 29110 Software Quality Model For Software SMEsISO 29110 Software Quality Model For Software SMEs
ISO 29110 Software Quality Model For Software SMEsMoutasm Tamimi
 
Retrospective and Trends in Requirements Engineering for Embedded Systems: A ...
Retrospective and Trends in Requirements Engineering for Embedded Systems: A ...Retrospective and Trends in Requirements Engineering for Embedded Systems: A ...
Retrospective and Trends in Requirements Engineering for Embedded Systems: A ...Tarcísio Couto
 
OSAmI-Commons – an OSGi based platform supporting Open Ambient Intelligence f...
OSAmI-Commons – an OSGi based platform supporting Open Ambient Intelligence f...OSAmI-Commons – an OSGi based platform supporting Open Ambient Intelligence f...
OSAmI-Commons – an OSGi based platform supporting Open Ambient Intelligence f...mfrancis
 
Declarative Performance Testing Automation - Automating Performance Testing f...
Declarative Performance Testing Automation - Automating Performance Testing f...Declarative Performance Testing Automation - Automating Performance Testing f...
Declarative Performance Testing Automation - Automating Performance Testing f...Vincenzo Ferme
 
Research-Based Innovation with Industry: Project Experience and Lessons Learned
Research-Based Innovation with Industry: Project Experience and Lessons LearnedResearch-Based Innovation with Industry: Project Experience and Lessons Learned
Research-Based Innovation with Industry: Project Experience and Lessons LearnedLionel Briand
 
Software Engineering course
Software Engineering courseSoftware Engineering course
Software Engineering courseJeremy Rose
 
A World In Motion
A World In MotionA World In Motion
A World In Motionoose
 
status of INCOSE Systems Engineering Handbook 5th Edition - AISE Annual Event
status of INCOSE Systems Engineering Handbook 5th Edition - AISE Annual Eventstatus of INCOSE Systems Engineering Handbook 5th Edition - AISE Annual Event
status of INCOSE Systems Engineering Handbook 5th Edition - AISE Annual EventBernardo A. Delicado
 
Open Source Business Models
Open Source Business ModelsOpen Source Business Models
Open Source Business Modelsg2ix
 
Spatial data supply chains in Australia and New Zealand
Spatial data supply chains in Australia and New ZealandSpatial data supply chains in Australia and New Zealand
Spatial data supply chains in Australia and New ZealandMaurits van der Vlugt
 
Balanced Measurement Sets - Criteria for Improving Project Management Practices
Balanced Measurement Sets - Criteria for Improving Project Management PracticesBalanced Measurement Sets - Criteria for Improving Project Management Practices
Balanced Measurement Sets - Criteria for Improving Project Management PracticesSpagoWorld
 
Replication and Benchmarking in Software Analytics
Replication and Benchmarking in Software AnalyticsReplication and Benchmarking in Software Analytics
Replication and Benchmarking in Software AnalyticsUniversity of Zurich
 

Similar to ICSE 2009 keynote (20)

Industry-Academia Communication In Empirical Software Engineering
Industry-Academia Communication In Empirical Software EngineeringIndustry-Academia Communication In Empirical Software Engineering
Industry-Academia Communication In Empirical Software Engineering
 
Week 01-intro se
Week 01-intro seWeek 01-intro se
Week 01-intro se
 
[2016/2017] RESEARCH in software engineering
[2016/2017] RESEARCH in software engineering[2016/2017] RESEARCH in software engineering
[2016/2017] RESEARCH in software engineering
 
A proposal for a new common process scope for AutomotiveSPICE: Six reasons fo...
A proposal for a new common process scope for AutomotiveSPICE: Six reasons fo...A proposal for a new common process scope for AutomotiveSPICE: Six reasons fo...
A proposal for a new common process scope for AutomotiveSPICE: Six reasons fo...
 
ISO 29110 Software Quality Model For Software SMEs
ISO 29110 Software Quality Model For Software SMEsISO 29110 Software Quality Model For Software SMEs
ISO 29110 Software Quality Model For Software SMEs
 
Retrospective and Trends in Requirements Engineering for Embedded Systems: A ...
Retrospective and Trends in Requirements Engineering for Embedded Systems: A ...Retrospective and Trends in Requirements Engineering for Embedded Systems: A ...
Retrospective and Trends in Requirements Engineering for Embedded Systems: A ...
 
OSAmI-Commons – an OSGi based platform supporting Open Ambient Intelligence f...
OSAmI-Commons – an OSGi based platform supporting Open Ambient Intelligence f...OSAmI-Commons – an OSGi based platform supporting Open Ambient Intelligence f...
OSAmI-Commons – an OSGi based platform supporting Open Ambient Intelligence f...
 
Declarative Performance Testing Automation - Automating Performance Testing f...
Declarative Performance Testing Automation - Automating Performance Testing f...Declarative Performance Testing Automation - Automating Performance Testing f...
Declarative Performance Testing Automation - Automating Performance Testing f...
 
Research-Based Innovation with Industry: Project Experience and Lessons Learned
Research-Based Innovation with Industry: Project Experience and Lessons LearnedResearch-Based Innovation with Industry: Project Experience and Lessons Learned
Research-Based Innovation with Industry: Project Experience and Lessons Learned
 
Software Engineering course
Software Engineering courseSoftware Engineering course
Software Engineering course
 
A World In Motion
A World In MotionA World In Motion
A World In Motion
 
status of INCOSE Systems Engineering Handbook 5th Edition - AISE Annual Event
status of INCOSE Systems Engineering Handbook 5th Edition - AISE Annual Eventstatus of INCOSE Systems Engineering Handbook 5th Edition - AISE Annual Event
status of INCOSE Systems Engineering Handbook 5th Edition - AISE Annual Event
 
Presentation of the SEALS project
Presentation of the SEALS projectPresentation of the SEALS project
Presentation of the SEALS project
 
SDLC Smashup
SDLC SmashupSDLC Smashup
SDLC Smashup
 
Open Source Business Models
Open Source Business ModelsOpen Source Business Models
Open Source Business Models
 
Oss2015
Oss2015Oss2015
Oss2015
 
Spatial data supply chains in Australia and New Zealand
Spatial data supply chains in Australia and New ZealandSpatial data supply chains in Australia and New Zealand
Spatial data supply chains in Australia and New Zealand
 
Balanced Measurement Sets - Criteria for Improving Project Management Practices
Balanced Measurement Sets - Criteria for Improving Project Management PracticesBalanced Measurement Sets - Criteria for Improving Project Management Practices
Balanced Measurement Sets - Criteria for Improving Project Management Practices
 
Replication and Benchmarking in Software Analytics
Replication and Benchmarking in Software AnalyticsReplication and Benchmarking in Software Analytics
Replication and Benchmarking in Software Analytics
 
Smef2009
Smef2009Smef2009
Smef2009
 

More from Carlo Ghezzi

More from Carlo Ghezzi (6)

Laser 0-prologue
Laser 0-prologueLaser 0-prologue
Laser 0-prologue
 
Laser 4-dynamic update
Laser 4-dynamic updateLaser 4-dynamic update
Laser 4-dynamic update
 
Laser 3-incremental
Laser 3-incrementalLaser 3-incremental
Laser 3-incremental
 
Laser 2-change
Laser 2-changeLaser 2-change
Laser 2-change
 
Laser 1-background
Laser 1-backgroundLaser 1-background
Laser 1-background
 
Paderborn
PaderbornPaderborn
Paderborn
 

ICSE 2009 keynote

  • 1. Reflections on 40+ years of software engineering research and beyond an insider's view Carlo Ghezzi DeepSE Group Dipartimento di Elettronica e Informazione Politecnico di Milano, Italy carlo.ghezzi@polimi.it ICSE 2009 1
  • 2. Outline of the talk   A journey through 40+ years of software engineering research, mostly observed through ICSE   What did we produce?   How can their success (impact) be "defined"?   How can it be measured?   What can we learn from past?   Can we do better?   Will the paradigms followed in the past survive in the future? ICSE 2009 2
  • 3. Part 1: Looking backwards   Where does ICSE come from?   What can we learn by mining ICSE data? ICSE 2009 3
  • 4. A bit of history   The roots are in the NATO meetings in Garmisch-Partenkirchen (1968) and Roma (1969)   ICSE started in 1975, this is the 31-st –  first conference called ICSE is actually ICSE2   ICSE1 was NCSE---1st National Conf on Software Engineering, Sept 1975 –  became annual in 1987   TSE started in March 1975   ESEC started in 1987, FSE started in 1993, they joined in 1997   TOSEM started in January 1992 ICSE 2009 4
  • 5. Mining ICSE data   International Conference on Software Engineering –  How much international? –  How are the different engineering research stakeholders (academia, industry) represented?   scientific leadership, contributors   Paper demographics –  Is ICSE an attractive venue?   Research areas –  How did they evolve? ICSE 2009 5
  • 6. ICSE PCs avg 1 12 32 55 ICSE 2009 6
  • 9. Submitted/accepted ICSE 2009 9
  • 10. ICSE papers: geographical distribution average 1% 10% 24% 65% ICSE 2009 10
  • 11. Papers: industry vs. academia average AC 70% IND 30% average 1976-1994 average 1995-2008 AC 56% AC 83% IND 44% IND 17% ICSE 2009 11
  • 12. Preliminary findings ICSE is nurtured by an international community NorthAmerica Europe AsiaPacific RestOfWorld PC all 55% 32% 12% 1% -5 yrs 46% 40% 11% 3% NorthAmerica Europe AsiaPacific RestOfWorld papers all 65% 24% 10% 1% -5 yrs 60% 27% 11% 2% High number of submissions, high selection rate There is a consensus that ICSE is a prestigious venue ICSE 2009 12
  • 13. Preliminary findings   Industry participation decreased over time –  research labs of large corporations disappeared –  ICSE became more research oriented   initially: tool fair   then: tutorials   now: workshops –  now mostly co-authored papers (academia/industry) ICSE 2009 13
  • 14. Topics Topic % Testing&Analysis 17 Design&Software Architecture 17 Specification&Verification 15 Tools&SDEs 11 Management&Costs 9 Empirical Studies 7 Existing Systems 6 Others (13 topics) 18 ICSE 2009 14
  • 15. Topics Topic % Testing&Analysis 17 Design&Software Architecture 17 Specification&Verification 15 Tools&SDEs 11 Management&Costs 9 Empirical Studies 7 Existing Systems 6 Others (13 topics) 18 ICSE 2009 15
  • 16. Topics Topic % Testing&Analysis 17 Design&Software Architecture 17 Specification&Verification 15 Tools&SDEs 11 Management&Costs 9 Empirical Studies 7 Existing Systems 6 Others (13 topics) 18 ICSE 2009 16
  • 17. Topics Topic % Testing&Analysis 17 Design&Software Architecture 17 Specification&Verification 15 Tools&SDEs 11 Management&Costs 9 Empirical Studies 7 Existing Systems 6 Others (13 topics) 18 ICSE 2009 17
  • 18. Topics Topic % Testing&Analysis 17 Design&Software Architecture 17 Specification&Verification 15 Tools&SDEs 11 Management&Costs 9 Empirical Studies 7 Existing Systems 6 Others (13 topics) 18 ICSE 2009 18
  • 19. Topics Topic % Testing&Analysis 17 Design&Software Architecture 17 Specification&Verification 15 Tools&SDEs 11 Management&Costs 9 Empirical Studies 7 Existing Systems 6 Others (13 topics) 18 ICSE 2009 19
  • 20. Topics Topic % Testing&Analysis 17 Design&Software Architecture 17 Specification&Verification 15 Tools&SDEs 11 Management&Costs 9 Empirical Studies 7 Existing Systems 6 Others (13 topics) 18 ICSE 2009 20
  • 21. Programming languages ICSE 2009 21
  • 22. Some findings   Large variety of topics   Trend from "methodology oriented", informal approaches to systematic and formally-based approaches   Increased attention on empirical evaluations   Still little on paradigmatic examples or domain-specific approaches ICSE 2009 22
  • 23. Can we identify our best products?   Which criteria to follow? –  Quantitative   Citation count   Downloads count   Others? –  Qualitative   Most influential -10 years?   Others? ICSE 2009 23
  • 24. Quantitative criteria   Citation count –  date of count Jan-March 2009 –  source Google Scholar –  version (conference/journal) both (if j expands c) –  precision ???   Number of downloads –  date, period Jan-March 2009, 12 months –  source ACM digital library –  version conference –  precision ??? ICSE 2009 24
  • 25. Outstanding products based on citations 1.  Program slicing (ICSE'81) 2120 2.  Statemate: a working environment for the development of complex reactive systems (ICSE'88) 1170 3.  N degrees of separation: multi-dimensional separation of concerns (ICSE'99) 997 4.  Bandera : extracting finite-state models from Java source code (ICSE'00) 791 5.  Software processes are software too (ICSE'87) 688 6.  Managing the development of large software systems: concepts and techniques (ICSE'87) 667 7.  Executable object modeling with statecharts (ICSE'96) 622 8.  Designing software for ease of extension and contraction (ICSE'78) 605 ICSE 2009 25
  • 26. Citations vs expert judgment most influential papers 1.  Program slicing (ICSE'81) 2120 2.  Statemate: a working environment for the development of complex reactive systems (ICSE'88) 1170 3.  N degrees of separation: multi-dimensional separation of concerns (ICSE'99) 997 4.  Bandera : extracting finite-state models from Java source code (ICSE'00) 791 5.  Software processes are software too (ICSE'87) 688 6.  Managing the development of large software systems: concepts and techniques (ICSE'87) 667 7.  Executable object modeling with statecharts (ICSE'96) 622 8.  Designing software for ease of extension and contraction (ICSE'78) 605 ICSE 2009 26
  • 27. MIP vs citations 380 ICSE 2009 27
  • 28. How do we compare with others? Citations of most influential papers ICSE 2009 28
  • 29. #Downloads (-12 months)? From ACM DL 1.  Automated Support for Development, Maintenance, and Testing in the Presence of Implicit Control Flow (ICSE'04): 781---13 cit 2.  Managing the development of large software systems: concepts and techniques (ICSE'87): 705---667 cit 3.  New directions on agile methods : a comparative analysis (ICSE'03): 567---161 cit 4.  Program slicing (ICSE'81): 520---2120 cit 5.  Quantitative evaluation of software quality ICSE'76: 495---212 cit 6.  Static detection of cross-site scripting vulnerabilities ICSE'08---396---8 cit 7.  Analysis and testing of Web applications ICSE'01: 375---195 cit 8.  Aspect-oriented programming and modular reasoning: ICSE'05: 300---32cit ICSE 2009 29
  • 30. Findings: #citations vs. #downloads   Time factor (both #citations an #downloads) –  my Jan-March numbers are different from the current ones –  how can you compare #cit of a 20 years and a 2 years old paper?   No significant correlation between citations and downloads   #citations and #downloads express different merits   example: a highly downloaded paper may be useful in practice, but does not inspire further research   #downloads demonstrates longevity of ICSE contributions   3 out of the top 5 highest #downloads are papers over 20 years old, one is more than 30 ICSE 2009 30
  • 31. Findings: #citations vs. expert judgment   Citations and expert judgment are better correlated, according to the ICSE experience –  most highly cited papers have been recognized as influential –  …but several influential papers have low citations ICSE 2009 31
  • 32. What do the numbers tell us? Do they indicate impact?   Does the number of paper citations indicate impact of a piece of work?   Do citations of papers in a journal/proceedings indicate impact of a venue?   Do citations of one's research measure impact of that research? ICSE 2009 32
  • 33. Part 2: A side road on numbers/citations Joint Committee on Quantitative Assessment of Research Report CITATION STATISTICS Intl Mathematical Union (IMU) in cooperation with Intl Council of Industrial and Applied Math (ICIAM) and the Institute of Mathematical Statistics (IMS), 6/12/2008 R. Adler, J. Ewing, P. Taylor (Eds.) ICSE 2009 33
  • 34. A side road on numbers/citations Informatics Europe Report RESEARCH EVALUATION FOR COMPUTER SCIENCE Viewpoint article in CACM, April 2009 Eds. B. Meyer, C. Choppy, J. Staunstrup, J. van Leewen (Eds.) Also D. Parnas, CACM nov 2007 STOP THE NUMBERS GAME ICSE 2009 34
  • 35. Facts   Numbers increasingly used to assess/rank –  papers –  journals and conferences---the (in)famous impact factor –  individuals   Motivations –  different stakeholders need to evaluate research –  peer review can be subjective –  peer review is expensive –  numbers are simple and objective simplistic and misleading ICSE 2009 35
  • 36. Findings   Much of the modern bibliometrics is flawed (statistics improperly used)   Objectivity and accuracy illusory –  the meaning of a citation can be even more subjective than peer review   Sole reliance on citation data provides incomplete and shallow understanding of research –  only valid if reinforced by other judgments numbers are not inherently superior to and cannot substitute complex judgement ICSE 2009 36
  • 37. Why are citations counted? From Thomson Scientific "The value of information is determined by those who use it…the better way to measure the quality of work is by measuring the impact it makes on the community at large" The statement makes an implicit equivalence assumption #citations = impact ICSE 2009 37
  • 38. Citations vs impact   Is the equality #citations = impact justified? (When) does it work?   It presupposes a highly rational and uniform model of reference giving, but this is NOT true ICSE 2009 38
  • 39. The complex sociology of citations Average citations per article citation practices differ substantially among disciplines ICSE 2009 39
  • 40. The complex sociology of citations   Citations are of two main kinds –  REWARD   intellectual dependency on the cited –  RHETORICAL   a way to carry out a scientific conversation –  reference included to show that topic is of interest to someone else or to prove that the author knows the literature –  the cited explains some result, not necessarily of the cited author –  the cited represents another approach, or is an example of… ICSE 2009 40
  • 41. The complex sociology of citations   Most citations are rhetorical   Reward citations can be of many kinds –  currency, negative credit, operational information, persuasiveness, positive credit, reader alert, social consensus   Obliteration effect –  incorporated into other work, which is cited ICSE 2009 41
  • 42. An example   The impact of MESA on system design, by HC Lauer, EH Satterthwaite, ICSE'79 –  60 cit (highest of '79), 19 downloads … one of the few papers on MESA BUT   Its impact evidenced by indirection –  Implementing remote procedure calls, by Birrell&Nelson, ACM TOCS 1984, 1840 citations   explicitly states its debt to MESA ICSE 2009 42
  • 43. If impact is the question, are (citation) numbers the answer?   Citation data and statistics provide some valuable information –  very high citation numbers tell more than smaller but only a limited and incomplete view of research quality   Statistics derived from them often poorly understood and misused   Expert judgment cannot be eliminated   Research too important to measure only with a coarse tool ICSE 2009 43
  • 44. Lost in the journey?   We started by looking back at SE research through ICSE as a magnifying lens   We tried to assess what we accomplished by identifying our impactful results   We got trapped by the number game Should we look for better assessment methods? Why? ICSE 2009 44
  • 45. Should we care?   Our social responsibility is to care about impact of research –  understanding –  measuring –  improving   If we don't, others will do –  governments, funding bodies   accountability, "return" to taxpayer or donor –  universities, schools, departments   competition (CS versus other sciences, SE versus other CS areas) ICSE 2009 45
  • 46. Part 3: Where do we go from here? Towards   understanding   measuring   improving impact of SE research ICSE 2009 46
  • 47. Do we understand what impact is? How can we measure it?   Not quite   A notable counter-example –  IMPACT project   aims at demonstrating impact of SE research on practice   focuses on specific research areas in SE –  an initiative of ACM SIGSOFT   the SIGSOFT Impact project.webarchive –  area leaders responsible for research –  backed by sound science history methodology ICSE 2009 47
  • 48. Published results   Osterweil, L.J., Ghezzi, C., Kramer, J., Wolf, A.L. Determining the Impact of Software Engineering Research on Practice . IEEE Computer, March 2008.   Estublier, J., Leblang, D., Hoek, A., Conradi, R., Clemm, G., Tichy, W., and Wiborg-Weber, D.Impact of software engineering research on the practice of software configuration management. ACM TOSEM, Oct. 2005.   Ryder, B. G., Soffa, M. L., and Burnett, M. The impact of software engineering research on modern progamming languages. ACM TOSEMl, Oct. 2005.   Emmerich, W., Aoyama, M., Sventek, J. The Impact of Research on the Development of Middleware Technology. ACM TOSEM, August 2008.   Clarke, L. A. and Rosenblum, D. S. Historical Perspective on Runtime Assertion Checking in Software Development. SIGSOFT SEN, March 2008.   Rombach, D., Ciolkowski, M., Jeffery, R., Laitenberger, O., McGarry, F., Shull, S. Impact of research on practice in the field of inspections, reviews and walkthroughs: learning from successful industrial uses. In ACM SIGSOFT SEN, November 2008. ICSE 2009 48
  • 49. Some key findings/confirmations   SE research has had impact on SE practice   Maturation needs time –  15-20 years between first publication of an idea and widespread availability in products   Substantially different mechanisms have been successful at causing impact –  people are key to knowledge transmission   people movement   almost all impact traces lead back to PhD theses ICSE 2009 49
  • 50. Problems with impact definition/measurement   Scholarly assessment of impact of SE research on practice is a difficult and expensive –  it is research   Substantially different methods are needed to assess impact of –  journals/conference –  individual researchers/papers ICSE 2009 50
  • 51. Obstacles to impact   The dominant reward/recognition system based on publication count does not favor impact –  from D. Parnas, CACM 2007   "it encourages superficial research"   "it encourages small, insignificant results"   "it rewards publication of half-baked ideas"   "it slow scientific progress: to increase their score, researchers must avoid tackling the tough problems and problems that will require years of dedicated work and instead work on easier ones" ICSE 2009 51
  • 52. Obstacles to impact   Building on top of others' work too often too difficult or impossible –  TOSEM papers 2001-2006   60% refer to a tool, only 20% installable   ICSE'06 paper by Zannier, Melnik, Maurer evaluates ICSE empirical studies   complete absence of replicated studies   self-evaluations dominate empirical studies ICSE 2009 52
  • 53. Conclusions: on impact   Research on impact of SE research must continue and should become a community effort   As a community we need to get to a common understanding and articulate methods for definition and measurement of impact for –  journals/conferences –  individual research   define impactful research products beyond paper count   We must be aware of risks –  be rigorous, avoid being incestuous ICSE 2009 53
  • 54. Conclusions: on enhancing research flow   Define community challenges and support repeated experiments –  catalogs of case-studies on which competing approaches can be applied and compared   Favor research products backed by tools, but verify that the claimed results of their use can be replicated   Use Internet facilities for community support –  research-as-service   Reward successful reuse of one's research rather than paper citation ICSE 2009 54
  • 55. The end of the journey In 40+ years we have gone far ICSE 2009 55
  • 56. Beginning of a new journey We have been successful, but we can do more and better Towards impact-aware research Can the ICSE community take the lead of this? ICSE 2009 56
  • 57. Acknowledgements   L. Osterweil, J. Kramer, A. Wolf and IMPACT project members for their insights into impact of research   Andrea Mocci, Mattia Monga, for helping me in mining data ICSE 2009 57
  • 58. Thank you!!! ICSE 2009 58