LLVM and the future of optimization

后端 未结 7 1873
被撕碎了的回忆
被撕碎了的回忆 2021-02-19 07:11

I realize that LLVM has a long way to go, but theoretically, can the optimizations that are in GCC/ICC/etc. for individual languages be applied to LLVM byte code? If so, does t

相关标签:
7条回答
  • 2021-02-19 07:43

    LLVM has made a far way to be a promising alternative to GCC (it even succeeds in compiling the Linux kernel, with some patches). It is also in many cases faster than GCC (the compiler and the generated executables) and has a structure which makes it quite easy to write frontends for arbitrary languages.

    But to allow extensive optimizations also the frontend is equally important since it knows much more details about the program being compiled. Things the LLVM stack cannot easily find out. Therefore to generate efficient programs also the frontend has to optimize the code.

    0 讨论(0)
  • 2021-02-19 07:47

    It is an interesting question, but I am afraid that you are lacking some notions of what compilers do.

    There will always be several stages of optimizations within a compiler.

    • Some optimizations depend on languages rules, for example in C++ you may optimize away some copy/move constructors calls.
    • Some optimizations are broadly available (loop transforms, tail calls)
    • Some optimizations depend on the hardware (SSE, MMX)

    The LLVM compiler applies all 3... but starting from LLVM IR, not from C or Java or whatever.

    It is the job of the front-end to provide a suitable LLVM IR.

    For example, as noted by @Dietrich Epp, the IR is not really suitable for Functional Languages. A lot of Haskell optimizations must therefore be performed before the representation is lowered to the IR.

    Another source of non-optimization is that specific runtime may come with the languages. Haskell has a complicated runtime with sparks pool, lightweight-threads, preemption before system calls, work-stealing etc... The IR is not suitable to represent this rich environment, and no optimization is made on this organization.

    0 讨论(0)
  • 2021-02-19 07:51

    I don't know any details about the bytecode format used by LLVM but I think the answer to your question is no.

    Just consider the following: dynamic vs. static typing. A programming language which is dynamically typed probably will be slower than a statically typed language, because the majority of type checking is performed at run-time.

    There can also be some other differences between programming languages which may affect performance.

    0 讨论(0)
  • 2021-02-19 07:51

    Here I've found report, with easy for begginers overview of GCC vs LLVM differences : Clang/LLVM Maturity Evaluation Report by Dominic Fandrey

    0 讨论(0)
  • 2021-02-19 07:53

    Coming back to the your question "is it theoretically possible?" , let's imagine/assume - Just for a sake of discussion - following :

    • we have bytecode, or even machinecode
    • we know it's produced with given compiler X
    • source language was L
    • that input program had finite size.

    IMHO - from computer science point of view, it's mostly question of resources, to apply almost anything.

    Now let's try following

    function machinecode_to_sourcecode( given_machinecode ){
      it = all_posibble_strings_iterator()
      while (true) {
        possible_sourcecode = it->get_string()
        machine_code = compile possible_sourcecode with X
        if (compilation succeeded)
            if(machine_code == given_machinecode)
               return possible_sourcecode;
        else
           it = it->next_string()
      }
    }
    

    So we try all possible strings, as input for compiler X, in case of success compilation - ones results equal, we have source.

    Once we have source, we have as much information about program as we can, so we are able to apply all possible optimizations.

    All looks very costly, but as you've asked "theoretically", I will say

    • it will take finite amount of time

    because

    • input sourcecode have finite length

    so

    • iterating at all such strings takes finite amount of time
    • I assumed, that compiler process any sourcecode in finite time (That's what we usually assume when we think about compiler ;) ).

    whole operation is going to take finite amount of computation time.

    0 讨论(0)
  • 2021-02-19 07:57

    In general, no.

    For example, in Haskell a common optimization is strictness analysis, which allows the compiler to determine which variables are always in head-normal form and therefore can be forced + inlined without changing program semantics. This is not possible with LLVM.

    Explanation: In Haskell, a function (Int, Int) -> Int is more or less equivalent to the type in C:

    typedef int (*returns_int)();
    struct pair { returns_int first, second};
    typedef struct pair *(*returns_pair)();
    
    int function(returns_pair arg);
    

    The compiler can analyze function and determine that it always evaluates its argument and always extracts the contents, transforming the function into this:

    int function(int x, int y); // note that it takes *two* arguments now
    

    This is well beyond the capability of LLVM. Maybe, in the future, with some really heavy interprocedural optimization... but realistically speaking, this will not happen in the foreseeable future.

    Example 2: There are Java VMs which can transform virtual function calls into direct function calls. However, this isn't something that LLVM can do — because this transformation must be undone dynamically if another class is loaded which implements the same interface.

    In general, when you compile a program to LLVM you lose much of the semantic information about the original program. LLVM bytecode is capable of representing any code, but its type system is fairly limited — and your choice of type system affects what optimizations you can do.

    0 讨论(0)
提交回复
热议问题