Automatic bug-fixing is the automatic repair of software bugs without the intervention of a human programmer.[1][2] It is also commonly referred to as automatic patch generation, automatic bug repair, or automatic program repair.[3][4] The typical goal of such techniques is to automatically generate correct patches to eliminate bugs in software programs without causing software regression.[5]
Automatic bug fixing is made according to a specification of the expected behavior which can be for instance a formal specification or a test suite.[6]
A test-suite – the input/output pairs specify the functionality of the program, possibly captured in assertions can be used as a test oracle to drive the search. This oracle can in fact be divided between the bug oracle that exposes the faulty behavior, and the regression oracle, which encapsulates the functionality any program repair method must preserve. Note that a test suite is typically incomplete and does not cover all possible cases. Therefore, it is often possible for a validated patch to produce expected outputs for all inputs in the test suite but incorrect outputs for other inputs.[7] The existence of such validated but incorrect patches is a major challenge for generate-and-validate techniques.[7] Recent successful automatic bug-fixing techniques often rely on additional information other than the test suite, such as information learned from previous human patches, to further identify correct patches among validated patches.[8]
Another way to specify the expected behavior is to use formal specifications[9][10] Verification against full specifications that specify the whole program behavior including functionalities is less common because such specifications are typically not available in practice and the computation cost of such verification is prohibitive. For specific classes of errors, however, implicit partial specifications are often available. For example, there are targeted bug-fixing techniques validating that the patched program can no longer trigger overflow errors in the same execution path.[11]
Generate-and-validate approaches compile and test each candidate patch to collect all validated patches that produce expected outputs for all inputs in the test suite.[6][7] Such a technique typically starts with a test suite of the program, i.e., a set of test cases, at least one of which exposes the bug.[6][8][12][13] An early generate-and-validate bug-fixing systems is GenProg.[6] The effectiveness of generate-and-validate techniques remains controversial, because they typically do not provide patch correctness guarantees.[7][14] Nevertheless, the reported results of recent state-of-the-art techniques are generally promising. For example, on systematically collected 69 real world bugs in eight large C software programs, the state-of-the-art bug-fixing system Prophet generates correct patches for 18 out of the 69 bugs.[8]
One way to generate candidate patches is to apply mutation operators on the original program. Mutation operators manipulate the original program, potentially via its abstract syntax tree representation, or a more coarse-grained representation such as operating at the statement-level or block-level. Earlier genetic improvement approaches operate at the statement level and carry out simple delete/replace operations such as deleting an existing statement or replacing an existing statement with another statement in the same source file.[6][15] Recent approaches use more fine-grained operators at the abstract syntax tree level to generate more diverse set of candidate patches.[8][13]
Another way to generate candidate patches consists of using fix templates. Fix templates are typically predefined changes for fixing specific classes of bugs.[16] Examples of fix templates include inserting a conditional statement to check whether the value of a variable is null to fix null pointer exception, or changing an integer constant by one to fix off-by-one errors.[16] It is also possible to automatically mine fix templates for generate-and-validate approaches.[17][18]
Many generate-and-validate techniques rely on the redundancy insight: the code of the patch can be found elsewhere in the application. This idea was introduced in the Genprog system, where two operators, addition and replacement of AST nodes, were based on code taken from elsewhere (i.e. adding an existing AST node). This idea has been validated empirically, with two independent studies that have shown that a significant proportion of commits (3%-17%) are composed of existing code.[19][20] Beyond the fact that the code to reuse exists somewhere else, it has also been shown that the context of the potential repair ingredients is useful: often, the donor context is similar to the recipient context.[21][22]
Repair techniques exist that are based on symbolic execution. For example, Semfix[23] uses symbolic execution to extract a repair constraint. Angelix[24] introduced the concept of angelic forest in order to deal with multiline patches.
Under certain assumptions, it is possible to state the repair problem as a synthesis problem. SemFix[23] and Nopol[25] uses component-based synthesis.[26] Dynamoth[27] uses dynamic synthesis.[28] S3[29] is based on syntax-guided synthesis.[30] SearchRepair[31] converts potential patches into an SMT formula and queries candidate patches that allow the patched program to pass all supplied test cases.
Machine learning techniques can improve the effectiveness of automatic bug-fixing systems.[8] One example of such techniques learns from past successful patches from human developers collected from open source repositories in GitHub and SourceForge.[8] It then use the learned information to recognize and prioritize potentially correct patches among all generated candidate patches.[8] Alternatively, patches can be directly mined from existing sources. Example approaches include mining patches from donor applications[11] or from QA web sites.[32] Learning can done online, aka continual learning, with the known precedent of online learning of patches from the stream of open source build results from continuous integration.[33]
SequenceR uses sequence-to-sequence learning on source code in order to generate one-line patches.[34] It defines a neural network architecture that works well with source code, with the copy mechanism that allows to produce patches with tokens that are not in the learned vocabulary. Those tokens are taken from the code of the Java class under repair.
Getafix[35] is a language-agnostic approach developed and used in production at Facebook. Given a sample of code commits where engineers fixed a certain kind of bug, it learns human-like fix patterns that apply to future bugs of the same kind. Besides using Facebook's own code repositories as training data, Getafix learnt some fixes from open source Java repositories. When new bugs get detected, Getafix applies its previously learnt patterns to produce candidate fixes and ranks them within seconds. It presents only the top-ranked fix for final validation by tools or an engineer, in order to save resources and ideally be so fast that no human time was spent on fixing the same bug, yet.
Targeted automatic bug-fixing techniques generate repairs for specific classes of errors such as null pointer exception[36][37][16] integer overflow ,[11] buffer overflow ,[11] memory leak ,[38] etc.. Such techniques often use empirical fix templates to fix bugs in the targeted scope. For example, insert a conditional statement to check whether the value of a variable is null[16] or insert missing memory deallocation statements.[38] Comparing to generate-and-validate techniques, targeted techniques tend to have better bug-fixing accuracy but a much narrowed scope.[7][38]
There are multiple uses of automatic bug fixing:
In essence, automatic bug fixing is a search activity, whether deductive-based or heuristic-based. The search space of automatic bug fixing is composed of all edits that can be possibly made to a program. There have been studies to understand the structure of this search space. Qi et al.[47] showed that the original fitness function of Genprog is not better than random search to drive the search. Martinez et al.[48] explored the imbalance between possible repair actions, showing its significant impact on the search. Long et al.'s[49] study indicated that correct patches can be considered as sparse in the search space and that incorrect overfitting patches are vastly more abundant (see also discussion about overfitting below).
If one explicitly enumerates all possible variants in a repair algorithm, this defines a design space for program repair.[50] Each variant selects an algorithm involved at some point in the repair process (e.g. the fault localization algorithm), or selects a specific heuristic which yields different patches. For instance, in the design space of generate-and-validate program repair, there is one variation point about the granularity of the program elements to be modified: an expression, a statement, a block, etc.[50]
It is also possible to analyze whether known search spaces encompass existing commits.[51] Such a search space analysis means mining a software repository, this results in an approximation of the applicability and usefulness of a given repair algorithm.
Sometimes, in test-suite based program repair, tools generate patches that pass the test suite, yet are actually incorrect, this is known as the "overfitting" problem.[52] "Overfitting" in this context refers to the fact that the patch overfits to the test inputs. There are different kinds of overfitting:[53] incomplete fixing means that only some buggy inputs are fixed, regression introduction means some previously working features are broken after the patch (because they were poorly tested). Early prototypes for automatic repair suffered a lot from overfitting: on the Manybugs C benchmark, Qi et al.[7] reported that 104/110 of plausible GenProg patches were overfitting; on the Defects4J Java benchmark, Martinez et al.[54] reported that 73/84 plausible patches as overfitting. In the context of synthesis-based repair, Le et al.[55] obtained more than 80% of overfitting patches.
One way to avoid overfitting is to filter out the generated patches. This can be done based on dynamic analysis,[56] or static code analysis of the generated patches.[57] When a reference patch is available, a state of the art technique is to generate tests based on the patched version, such that the generated tests capture the expected behavior. While the sampling of the input domain by test generation is incomplete by construction, it has been shown to be effective at detecting overfitting patches, and even at finding human errors done during manual classification of patches.[58]
Automatic bug-fixing techniques that rely on a test suite do not provide patch correctness guarantees, because the test suite is incomplete and does not cover all cases.[7] A weak test suite may cause generate-and-validate techniques to produce validated but incorrect patches that have negative effects such as eliminating desirable functionalities, causing memory leaks, and introducing security vulnerabilities.[7] One possible approach is to amplify the failing test suite by automatically generating further test cases that are then labelled as passing or failing. To minimize the human labelling effort, an automatic test oracle can be trained that gradually learns to automatically classify test cases as passing or failing and only engages the bug-reporting user for uncertain cases.[59]
A limitation of generate-and-validate repair systems is the search space explosion.[49] For a program, there are a large number of statements to change and for each statement there are a large number of possible modifications. State-of-the-art systems address this problem by assuming that a small modification is enough for fixing a bug, resulting in a search space reduction.
The limitation of approaches based on symbolic analysis[23][24] is that real world programs are often converted to intractably large formulas especially for modifying statements with side effects.
Benchmarks of bugs typically focus on one specific programming language. In C, the Manybugs benchmark collected by GenProg authors contains 69 real world defects and it is widely used to evaluate many other bug-fixing tools for C.[15][8][13][24]
In Java, the main benchmark is Defects4J, initially explored by Martinez et al.,[54] and now extensively used in most research papers on program repair for Java.[22][60] Alternative benchmarks exist, such as the Quixbugs benchmark,[61] which contains original bugs for program repair.[62] Other benchmarks of Java bugs include Bugs.jar,[63] based on past commits, and BEARS[64] which is a benchmark of continuous integration build failures.
Automatic bug-fixing is an active research topic in computer science. There are many implementations of various bug-fixing techniques especially for C and Java programs. Note that most of these implementations are research prototypes for demonstrating their techniques, i.e., it is unclear whether their current implementations are ready for industrial usage or not.