Skip to content

Explainable AI for Software Engineering: A Hands-on Guide on How to Make Software Analytics More Practical, Explainable, and Actionable (https://xai4se.github.io)

License

Notifications You must be signed in to change notification settings

DiliSR/xai4se.github.io

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Website xai4se.github.io Made with Python Made with Jupyter License: MIT (Code), CC BY-NC-SA (Book) DOI

Explainable AI for Software Engineering:

A Hands-on Guide on How to Make Software Analytics More Practical, Explainable, and Actionable

Alt text

About this Book

The success of software engineering projects largely depends on complex decision-making. For example, which tasks should a developer do first, who should perform this task, is the software of high quality, is a software system reliable and resilient enough to deploy, etc. However, erroneous decision-making for these complex questions is costly in terms of money and reputation. Thus, Artificial Intelligence/Machine Learning (AI/ML) techniques have been widely used in software engineering for developing software analytics tools and techniques to improve decision-making, developer productivity, and software quality. However, the predictions of such AI/ML models for software engineering are still not practical (i.e., fine-grained), not explainable, and not actionable. These concerns often hinder the adoption of AI/ML models in software engineering practices. In addition, many recent studies still focus on improving the accuracy, while a few of them focus on improving explainability. Are we moving in the right direction? How can we better improve the SE community (both research and education)? In this book, we first provide a concise yet essential introduction to the most important aspects of Explainable AI and a hands-on tutorial of Explainable AI tools and techniques. Then, we introduce the fundamental knowledge of defect prediction (an example application of AI for Software Engineering). Finally, we demonstrate three successful case studies on how Explainable AI techniques can be used to address the aforementioned challenges by making the predictions of software defect prediction models more practical, explainable, and actionable.

What you will learn

This book consists of three parts:

  • Part 1-Explainable AI: We first provide a concise yet essential introduction to the most important aspects of Explainable AI and a hands-on tutorial of Explainable AI tools and techniques.

  • Part 2-Defect Prediction Models: We introduce the fundamental knowledge of defect prediction (an example application of AI for Software Engineering)

  • Part 3-Explainable AI for Software Engineering: We demonstrate three successful case studies on how Explainable AI techniques can be used to address the aforementioned challenges by making the predictions of software defect prediction models more practical, explainable, and actionable.

Who this book is for

Some of the potential readers of this book include:

  1. SE researchers and PhD students who want to learn more about the intersection of Explainable AI and Software Engineering.

  2. Software practitioners who already use Python for as data science, machine learning, research, and analysis and wish to apply their data science knowledge to software data.

  3. Software analysts and data scientists who want to understand and avoid pitfalls when desigining software analytics.

  4. Project managers who involve high-stakes decision-making and need software analytics to make smarter data-driven busineess decisions.

Caution: This book is not an introduction to data science, machine learning, or artifcial intelligence concepts. You must have some foundational knowledge and/or experience with machine learning libraries such as scikit-learn to make the most out of this book.

About the authors

Dr. Chakkrit (Kla) Tantithamthavorn is a Senior Lecturer in Software Engineering and a 2020 ARC DECRA Fellow in the Faculty of Information Technology, Monash University, Australia. He is leading a new research area of Explaianble AI for Software Engineering. His current fellowship is focusing on the development of Practical and Explainable Analytics to Prevent Future Software Defects. His work has been published at several top-tier software engineering venues, such as TSE, ICSE, EMSE, MSR, IST. Contact him at chakkrit@monash.edu.

Dr. Jirayus Jiarpakdee is graduated from Monash University, Australia. His research interests include empirical software engineering and mining software repositories (MSR). The goal of his Ph.D. is to apply the knowledge of statistical modelling, experimental design, and software engineering to improve the explainability of defect prediction models. Contact him at jirayus.jiarpakdee@monash.edu.

Cite this book:

@book{xai4sebook,
    author = {Chakkrit Tantithamthavorn and Jirayus Jiarpakdee},
    booktitle = {Explainable AI for Software Engineering},
    year = {2021},
    publisher = {Monash University},
    howpublished = {\url{http://xai4se.github.io/}},
    note = {Retrieved 2021-05-17},
    url = {http://xai4se.github.io/},
    urldate = {2021-05-17},
    doi = {10.5281/zenodo.4769127}
}

Contributing

  1. Create a conda environment conda env create -f environment.yml, then activate the environment conda activate xaitools

  2. Add kernel to jupyter python -m ipykernel install --name xaitools --display-name "xaitools"

  3. Build the book jupyter-book build docs

About

Explainable AI for Software Engineering: A Hands-on Guide on How to Make Software Analytics More Practical, Explainable, and Actionable (https://xai4se.github.io)

Resources

License

Stars

Watchers

Forks

Packages

No packages published

Languages

  • Shell 100.0%