Skip to content

chibinz/CSE231

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

58 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

CSE231

Notes and projects following UC San Diego's Advanced Compilers course

Running

# Build all the passes
mkdir Build
meson Build
ninja -C Build

# Run a pass
opt -load-pass-plugin ./Build/libCountStaticInstructions.so -passes=csi ./Tests/<input>.ll -disable-output

Collecting Static Instruction Counts

Sample Output

call    2
ret     1
load    5
store   5
alloca  5
add     2
ret     1
call    1
ret     1

Notes

This is just a warmup pass to get familiarized with LLVM passes. LLVM made it super simple for us to iterate over functions and basic blocks by implementing iterators for class Function and BasicBlock. A typical LLVM function pass looks something like this:

// Iterate over all basic blocks in function
for (auto &BB : F) {
    // Iterate over all instructions in basic block
    for (auto &Instr : BB) {
        // Filter and manipulate instructions
    }
}

Directions in project description are already very clear, the following are a few key points.

  • std::map <const char*, int> is a good candidate for storing instruction opcode names and their respective counts.
  • Output to stderr using llvm::errs() << name << "\t" << count << "\n"
  • use -disable-output instead of > /dev/null

Collecting Dynamic Instruction Counts

There are a few confusions that I met with through out section 2 of this project. First of all, since the input is compiled down to native code, we're actually executing x86_64 instructions instead of LLVM IR instructions. That is to say, the dynamic instruction count here is assumming that we're interpreting LLVM bitcode rather than running it. Second, the last direction states that functions instrumented will not make any calls to other functions. To have a runnable executable, there must be a main function and that main function cannot make any function calls. Why don't the direction just say that there is only one main function? This seems to conflict with "Each instrumented function should output the analysis results before function termination"...

This sections mainly aims to teach us how to insert function calls. In LLVM it usually works like this. First you need get a Context; Context, in my understanding, is an LLVM class encapsulating information about your machine. For example, type information about 32-bit integers. How these integers are represented on x86_64 may differ from, say, ARM64 machines. Once we've obtained the context, we can construct types ground up, from basic integer types to function types. Module::getOrInsertFunction gives us a handle to the function that we want to call, given the function name and type. Hints in the project description suggest that we use IRBuilder to insert the actual function call, but I find CallInst::Create easier to use. Just pass in the function handle, actual parameters, and the instruction that you want to insert the call before.

Tips

  • opt ... -S can directly output LLVM assembly, no need to use llvm-dis.
  • Using the provided runtime library lib231 requires you to construct arrays of key and value at compile time, which can be a bit complicated. My solution is to simply insert a function call for every opcode. Overhead is not an issue here.:)
  • You can .cpp .ll files together, and clang will still produce an executable happily.

Profiling Branch Bias

The main goal of section, I assume, is to teach you how to filter for a specific type of instruction, in this case the conditional branch instruction. There are probably a dozen ways to do that. Listed in the following are ways that I found comfortable using:

// Handle multiple cases
switch (Instr.getOpcode()) {
    case Instruction::Br: ...; break;
    ...
}

// Filter by downcasting
BranchInst* BrInstr = dyn_cast<BranchInst>(&Instr);
if (Instr != nullptr) {
    // `dyn_cast` is an LLVM helper function that attempts to cast a pointer
    // to an instance of a base class to its derived class, and returns
    // a `nullptr` if it fails to do so.
}

// A better name would be isinstance
if (isa<BranchInst>(&Instr)) {
    // Yes, `Instr` "is a" `BranchInst`
}

// Filter by instruction name
if (Instr.getOpcodeName() == "br") {
    // Manipulate `Instr`
}

The benefit of using Instruction::getOpcode is that its return type is unsigned. Scenario's in which you have to deal with various types of instructions can be handled efficiently with a single switch case. If you just want to check if an object is an instance of a certain class, use the isa function. isa is a bad name in my opinion because I often mistake it for Instruction Set Architecture... Downcasting with dyn_cast, on the other hand, gives us access to specific member functions of the derived class. In the conditional branch case, the isConditional function is only available in the BranchInst class, not implemented in the base Instruction class. A common gotcha in using dyn_cast is putting in the pointer type instead of the class type in the template parameter.

// This will not compile...
// The compiler will spit out screens of barely comprehensible errors.
BranchInst* BrInstr = dyn_cast<BranchInst*>(Instr);

The last filtering technique is quite straight forward, but is less used in practice because it offers similar functionality as the first technique, while incurring more overhead.

Implement Reaching Definition Analysis (Adhoc)

Unfortunately, as an outsider I do not have access to recitation and lectures. My only reference are slides from my own compiler course and the Compilers: Principles, Techniques, and Tools textbook, and thus jumping right into the given template in 231DFA.h is kind of overwhelming for me. So I decided to implement non-generic data flow analyses to get myself familiarized with the process. At the core of each data flow analyses is its transfer function, it determines the type of the analyses info, and how it is manipulated by each statement(instruction in implementation). In the reaching definition case, the type is set of defintions, and the transfer function states that the output is the definition generated by this statement plus the definitions from the input except those killed by this instruction. A definition, in my understanding, is just an assignment. In LLVM, most computational instructions have return values, with a few control flow instructions that don't have left handsides. Set of definitions can be represented by std::set<Instruction *>, containing instructions with a return value. The project description seems to handle phi instructions differently. I treat them just an any other instruction with a return value, not sure what is the problem here. LLVM IR adhering to the SSA requirement makes it super easy to compute the gen set and kill set. Since each instruction only assigns to one variable, the gen is simply the return value of the instruction(which turn out to be the instruction itself in LLVM, Instruction * is a subtype of Value *). And since no variable is assigned twice, the kill set is always an empty set. Note that in any cases, the gen and kill set only need to be computed once and stay fixed throughout the iterative procedure, what keeps changing is the in and out set of each statement. There are quite a few distinction between the project requirement and the textbook. First, the meet operator introduced by the text book operates on basic blocks, while the output printed by 231_solution.so is on instruction granularity. The problem with this is that llvm only supports getting successors / predecessors of basic blocks but not instructions. For terminating and leading instructions getting their respective successors and predecessors using getNext/PrevNode will return nullptr. I was able to get around this by wrapping edge cases in a function, but previously expected LLVM would offer readily available APIs... Second, transfer function in the textbook seems to correspond to flow function in the project description, which made it a little harder for me to comprehend at first sight.

Gotchas

  • When doing a union for 2 sets, be careful with std::set::merge. A.merge(B) does not have a return value (C = A.merge(b) will not compile), and the merge operation actually clears set B. I was trying to print out reaching definitions for each instruction at the end of the function pass, but found out that all but the terminating instructions' out set is empty. It took me quite a well to figure what the problem was(Rust's borrow checker won't allow such things to happen...)
  • Inconsistency in getting the last and first instruction of a basic block. To get the last instruction, I used the getTerminator method. So I was expecting getLeader method of some sort to get me the first instruction...The right method to use turned out to be front. Since BasicBlock is an iterable class, it shouldn't be surprising, but what really struck me was that the back method was deleted. What is the rationale behind such weirdness..."
  • To make a copy of custom structure in Rust, you have to explicitly invoke the clone method. In C++, this seems to be done by overloading operator =...

Generic Framework

C++ Virtual Functions stuff

//---- Inheriting constructors ----//
class B : public A {
    // Inherit all constructors of base class A
    using A::A;
}

//---- virtual function weirdness ----//
// Must be overriden
virtual auto func_name() = 0;

// Actually both a declaration and a definition / implementation
virtual auto func_name() {}

// What about the following?
virtual auto func_name();
virtual auto func_name() {};
  • Unnecessary polymorphic Info type? Delete base Info class, and remove inheritance declaration(:public Info), code compiles just fine...
  • Questionable use of static functions The point of Info being an abstract class is that its derived class must implement the equal and meet operator. Static function are specific to each class and cannot be virtual, i.e. derived class doesn't have to implement it. If this is the case, why not just use virtual operator overloading? The compiler does not check whether the derived class has function equal and meet function when Info is passed in as template parameter, but the linker will complain undefined reference...
  • I previously thought that derived class must implement pure virtual functions, or else the compiler would complain...But this doesn't seem to be case, it would compile just fine

Having implemented the generic Analysis framework in DFAFramework.h, I thought I only need to code much fewer lines to spin up a reaching definition pass. But by closer inspection, my initial expectations were wrong. Excluding comments and blank lines, the adhoc reaching definition pass amounts to 70 lines, while the generic reaching definition pass based on the framework totals 60 lines. I was really stunned by the verbosity despite the worklist algorithm refactored into the analysis framework. In essence, what really defines a dataflow analysis is its domain, direction, transfer function, and join operator. All of these information nice fits into a table of a slide's size. At a closer look, most of the code that I wrote are class / type definition, function signatures, and routine stuff that are non-functional, or has little correlation with the analysis itself. I am thinking of encoding the functionality of each analysis in its purest and distilled form will being comprenhendable by the compiler... The verbosity is totally unnecessary.

Currently how summary for each project is organized is a little bit messy. I assume it is ok for me to keep it as person development blog, but not so much for those who take it as a tutorial. Thus all summaries / notes must conform to a general structure. The main purpose of me doing this alien university project is to get my self familiarized with llvm, as this is the "goto tool / framework" for compiler instrumentation. Thus the first subsection will be a list of llvm apis that is newly introduced or is if of interesting use. Although I consider myself an advocate for Rust, C++ is and will still be an actively used language in the near future, so there's no way around getting myself comfortable with the use of C++. Plus, most LLVM APIs are exposed through C++, and most Rust bindings are either obsolete or far from usable. The second section will be explanation of interesting C++ features exercised. for (auto &it: iterable), for example, is something I wasn't taught of in my Introduction to Programming course. I avoided writing C++ as much as possible by writing C in my Algorithm class, and when it is inevitable, what I wrote was some sort of "C with classes" or "C with STL". Explanation must extend beyond superficial syntactic aspects, dive into the semantics, rationale and why is beneficial / a good practice in that style. There should also be a tip section to offer suggestions on how to automize building, testing and scripting. Since I am on my own doing this project, there are quite distinctions from the original project, regarding the implementation and output. Readers should be informed of this, too. Finally, a question section for my doubts and issues with the project.

Testing. New passes are added every now and then, but I never took the time to actually test my implementation against the reference implementation. Perhaps because 1. The reference pass uses the legacy pass scheduler which seems to schedule functions / basicblocks differently than the latest pass manager. 2. A lack of test corpus. While I've pulled some tests from other repositories on GitHub, it's not clear how strong do they stand up against edge cases. In order for my passes to be usable by other people, automated testing and a strong test suite must be available. Now that I have the generic framework up and working, implementing new dataflow analysis will be, in theory, at least as simple as filling out the lattice table. So implementing new analyses won't be much of a burden. My initial observation shows the output of my live variable analysis pass drastically differs from the reference passes'. I suspect it has something to do with phi instructions, which I treated just as any other instruction with a return value. Hope the passes I implemented will be of use to someone who come upon this repository.

I also have a plan of an extra wrap up part that make uses of all the passes that I've written and composes them into a compiler backend. But that will have to wait until I finish all the requiremenst of the project. There are also little Chinese literature on this topic, so writing a Chinese version when I'm not in the mood of progress seems to be alternative. So far so good, stepping into client analysis. The last part, Interprocedural analyses, from what learned in class, is hard. I still have little clue what I have to do in project 4. Anyways, keep up the pace, it's not about a day or a week, it's about months and semester's of prolonged effort.

Reaching Definition (Generic)

  • Direction
    • forward
    • AnalysisDirection::Forward
  • Domain of lattice values
    • set of reaching definitions
    • std::set<Instruction *>
  • Transfer function
    • out[i] = gen(i) U (in[i] - kill(i))
      • i is the current instruction
      • kill(i) are assignments (instructions) with the same lvalue as i
      • gen(i) is the lvalue of current instruction if i has a return value otherwise empty set
  • Join operator
    • union of 2 sets
    • set::union2

Live Variable Analysis

  • Direction
    • backward
    • AnalysisDirection::Backward
  • Domain of lattice values
    • set of live variables
    • std::set<Value *>
  • Transfer function
    • out[i] = use(i) U (in[i] - def(i))
      • use(i) is the operands of current instruction, or rvalues
      • def(i) is the return value of current instruction if any
  • Join operator
    • union of sets
    • set::union2

May Point to Analysis

  • Direction
    • Forward
    • AnalysisDirection::Forward
  • Domain of lattice values
    • map from pointers to set of pointees
    • std::map<Value *, std::set<Value *>>
  • Transfer function
    • As defined in project description
    • different for
      • alloca
      • bitcast
      • getelementptr
      • store
      • load
      • select
      • phi
  • Join operator
    • union of maps
    • if the 2 maps have the same key the keys' corresponding sets are merged

Gotchas

  • If you're having problem getting alloca instructions, try opt -reg2mem -S to convert phi instructions into memory operations
  • Mutating a container while iterating it may lead to segfaults... Painful debugging
  • Referencing an uninitialized map entry will also cause segfault
  • You must add curly braces around a switch case if you want to declare local variables inside that case. case A: {auto foo = bar; ...; break;}

Getting all uses of an llvm value

for (auto use: val->users()) { // No reference `&` before use
    // `User` is a subtype of `Value`
}

Inserting an llvm instruction

BB.getFirstInsertPt();
BB.getTerminator();
instr->insertBefore(Instruction *);

Never insert an instruction before a phi node or after the termination of a basic block, e.g. ret, br instructions. There is a reason why you could only insert before a given instruction, inserting after a terminator instruction breaks the definition of a basicblock and can be problematic. If you are that the instruction to insert after is not a terminator,use this instr->insertBefore(anchor->getNextNode());.

Removing an llvm instruction

instr->eraseFromParent() // Note that removeFromParent does not actually deletes the instruction, it is used for another purpose

Replacing an llvm instruction

instr->replaceAllUsesWith(Value *)
instr->eraseFromParent()

Manipulating data structures while iterating over it can be a dangerous practice. In the insert case, you might be adding instructions repeatedly, causing pass manager to loop indefinitely. In the removing case, the definitions that you remove may leave behind dangling uses (reference) to it. And the replace case, the definition of the replaces instruction might not post-dominated all uses of itself. The best practice is what I called a 2 phase commit. Since iterating and changing the data structure simultaneously is so bad, we split this 2 actions into 2 phase. In the first phase, we record the instruction that we want to insert before, remove, or replace in a set or map. And then in the second phase, we use the information previously stored to accomplish our goals.

Casting all types to i32 / Type casting

LLVM while named "Low Level Virtual Machine" actually contains aggregate types such as struct {i32, i32}. I assume that this sort of high level type information make certain optimization easier, but it means for pass writers extra edge cases to handle. LLVM is also explicitly and strongly typed, unlike c, making it very painful to deal with type casting. Pointers and integers have dedicated instructions for casting between them.

Declaring a function

Module &M;
auto funcType = FunctionType::get(...);
M.getOrInsertFunction("function_name", funcType);

FunctionPass and BasicBlock pass are not allowed to insert Functions into a module because they are invoked many times over a single module.

Using lit for pass transformation validation

Using functional programming idioms(map, filter, reduce) to reduce boilerplate code

LLVM pass scheduling

  • Pass pipeline
  • Pass dependency
  • Debugging pass pipeline
  • opt -load lib/LLVMHello.so -gvn -licm --debug-pass=Structure < hello.bc > /dev/null

How to compose LLVM passes, use result of analysis pass as input

Advantages of using the new pass manager over the legacy one

Windows

LLVM does not support running out-of-tree passes on windows, due to limitations of dll files. While it is possible to compile a pass in-tree using some hacky ways on windows, building and testing LLVM on linux or mac is much easier.

clang options

  • -mllvm # Pass one argument to opt

opt options

  • -debug-only=<key>
  • -debug
  • -stats
  • -print-after/before-all
  • -print-after/before=<pass>
  • -filter-print-funcs=<func>
  • -filter-passes=<passes>
  • -print-module-scope
  • experimental (Available LLVM 12.0.0)
    • -print-changed
    • -print-crashed
    • -print-changes
    • -dot-cfg-changes

Use llc to generate llvm cpp code

clang -emit-llvm -c input.c -o tmp.bc
llc -march=cpp tmp.bc -o output.cpp
llc -O0 -march=cpp tmp.bc -o output.cpp

Reference

  1. 2020 LLVM Developers’ Meeting: “Understanding Changes made by a Pass in the Opt Pipeline”
  2. Generate llvm cpp code

About

Notes and projects following UC San Diego's Advanced Compilers course

Topics

Resources

License

Stars

Watchers

Forks