This article includes a list of general references, but it remains largely unverified because it lacks sufficient corresponding inline citations. (February 2014) (Learn how and when to remove this template message)
In computer science, ahead-of-time compilation (AOT compilation) is the act of compiling a higher-level programming language such as C or C++, or an intermediate representation such as Java bytecode or .NET Framework Common Intermediate Language (CIL) code, into a native (system-dependent) machine code so that the resulting binary file can execute natively.
Reduced runtime overheadEdit
Some programming languages with a managed code runtime that can be compiled to an intermediate representation, use just-in-time (JIT) compiling. This, briefly, compiles intermediate code into machine code for a native run while the intermediate code is executing, which may slow an application's performance. Ahead-of-time compiling eliminates the need for this step by occurring before execution rather than during execution.
Ahead-of-time compiling for dynamically typed languages to native machine code or other static VM bytecode is possible in a limited number of cases only. For example, the High Performance Erlang Project (HiPE) AOT compiler for the language Erlang can do this because of advanced static type reconstruction techniques and type speculations.
In most situations with fully AOT compiled programs and libraries, it is possible to drop a useful fraction of a runtime environment, thus saving disk space, memory, battery life, and startup times (no JIT warmup phase), etc. Because of this, it can be useful in embedded or mobile devices.
AOT compilers can perform complex and advanced code optimizations, which in most cases of JITing will be considered much too costly. In contrast, AOT usually cannot perform some optimizations possible in JIT, like runtime profile-guided optimization (PGO), pseudo-constant propagation, or indirect-virtual function inlining. AOT must compile to a target architecture, while a JIT can compile the code to make the best use of the actual CPU it is running on, even years after the software has been released.
Further, JIT compilers can speculatively optimize hot code by making assumptions on the code. The generated code can be deoptimized if a speculative assumption later proves wrong. Such operation slows the performance of the running software until code is optimized again by adaptive optimization. An AOT compiler cannot make such assumptions and needs to infer as much information as possible at compile time. It needs to resort to less specialized code because it cannot know what types will go through a method. Such problems can be alleviated by profile-guided optimizations. But even in this case, the generated code cannot be adapted dynamically to the changing runtime profile as a JIT compiler would do.
The Android mobile operating system was delivered in 2008 with Dalvik, a virtual machine using a JIT compiler. In 2013, it was replaced by Android Runtime, a new virtual machine using AOT compilation, but in 2017, it received a JIT compiler.
- "Implementing ART Just-In-Time (JIT) Compiler". android.com. Retrieved 25 January 2018.
- Speed: NGen Revs Up Your Performance with Powerful New Features – MSDN Magazine, April 2005
- Mono AOT
- Excelsior JET – Certified Java SE Implementation with AOT compiler
- GNU Compiler for Java
- AOT compilation of asm.js
- Real-time Java, Part 2: Comparing compilation techniques – IBM developerWorks, April 2007
- Improving Swing Performance: JIT vs AOT Compilation – LinuxWorld Magazine, November 2004