changeset 2685:97bae9fd4fff

Merge
author henryjen
date Thu, 17 Aug 2017 12:17:39 -0700
parents d8af7c437748 a08cbfc0e4ec
children 5da62ff54752
files README-builds.html README-builds.md common/autoconf/generated-configure.sh common/autoconf/lib-freetype.m4 common/autoconf/spec.gmk.in common/bin/update-build-readme.sh make/CompileJavaModules.gmk make/Javadoc.gmk test/lib/jdk/test/lib/DynamicVMOption.java test/lib/jdk/test/lib/InMemoryJavaCompiler.java test/lib/jdk/test/lib/wrappers/InfiniteLoop.java test/lib/jdk/test/lib/wrappers/TimeLimitedRunner.java
diffstat 62 files changed, 7612 insertions(+), 5513 deletions(-) [+]
line wrap: on
line diff
--- a/.hgtags	Tue Feb 14 14:57:56 2017 -0800
+++ b/.hgtags	Thu Aug 17 12:17:39 2017 -0700
@@ -399,3 +399,28 @@
 8d26916eaa21b689835ffc1c0dbf12470aa9be61 jdk-9+154
 688a3863c00ebc089ab17ee1fc46272cbbd96815 jdk-9+155
 783ec7542cf7154e5d2b87f55bb97d28f81e9ada jdk-9+156
+4eb77fb98952dc477a4229575c81d2263a9ce711 jdk-9+157
+a4087bc10a88a43ea3ad0919b5b4af1c86977221 jdk-9+158
+fe8466adaef8178dba94be53c789a0aaa87d13bb jdk-9+159
+4d29ee32d926ebc960072d51a3bc558f95c1cbad jdk-9+160
+cda60babd152d889aba4d8f20a8f643ab151d3de jdk-9+161
+21b063d75b3edbffb9bebc8872d990920c4ae1e5 jdk-9+162
+c38c6b270ccc8e2b86d1631bcf42248241b54d2c jdk-9+163
+7810f75d016a52e32295c4233009de5ca90e31af jdk-9+164
+aff4f339acd40942d3dab499846b52acd87b3af1 jdk-9+165
+ba5b16c9c6d80632b61959a33d424b1c3398ce62 jdk-9+166
+35017c286513ddcbcc6b63b99679c604993fc639 jdk-9+167
+143d4c87bc1ef1ed6dadd613cd9dd4488fdefc29 jdk-9+168
+b25838a28195f4b6dab34668411eedd2d366a16c jdk-9+169
+4d163ec59d989a9261ed7f848bc6303f90869af5 jdk-9+170
+4c12464a907db4656c1033f56fa49cba643ac629 jdk-9+171
+2c25fc24103251f9711a1c280c31e1e41016d90f jdk-9+172
+88d7fd969e7df0e07a53b201cfd29393ca33ede9 jdk-9+173
+5466f409346e0446ee9a6daeb7f5d75c8fc76823 jdk-9+174
+8f7227c6012b0051ea4e0bcee040c627bf699b88 jdk-9+175
+84777531d994ef70163d35078ec9c4127f2eadb5 jdk-9+176
+a4371edb589c60db01142e45c317adb9ccbcb083 jdk-9+177
+ec4159ebe7050fcc5dcee8a2d150cf948ecc97db jdk-9+178
+252475ccfd84cc249f8d6faf4b7806b5e2c384ce jdk-9+179
+d2982a786f53814367698e63efe6349c9128e1db jdk-9+180
+b656dea9398ef601f7fc08d1a5157a560e0ccbe0 jdk-9+181
--- a/README	Tue Feb 14 14:57:56 2017 -0800
+++ b/README	Thu Aug 17 12:17:39 2017 -0700
@@ -1,40 +1,10 @@
-README:
-  This file should be located at the top of the OpenJDK Mercurial root
-  repository. A full OpenJDK repository set (forest) should also include
-  the following 7 nested repositories:
-    "jdk", "hotspot", "langtools", "nashorn", "corba", "jaxws"  and "jaxp".
+Welcome to OpenJDK!
+===================
 
-  The root repository can be obtained with something like:
-    hg clone http://hg.openjdk.java.net/jdk9/jdk9 openjdk9
+For information about building OpenJDK, including how to fully retrieve all
+source code, please see either of these:
 
-  You can run the get_source.sh script located in the root repository to get
-  the other needed repositories:
-    cd openjdk9 && sh ./get_source.sh
+  * common/doc/building.html   (html version)
+  * common/doc/building.md     (markdown version)
 
-  People unfamiliar with Mercurial should read the first few chapters of
-  the Mercurial book: http://hgbook.red-bean.com/read/
-
-  See http://openjdk.java.net/ for more information about OpenJDK.
-
-Simple Build Instructions:
-
-  0. Get the necessary system software/packages installed on your system, see
-     http://hg.openjdk.java.net/jdk9/jdk9/raw-file/tip/README-builds.html
-
-  1. If you don't have a jdk8 or newer jdk, download and install it from
-     http://java.sun.com/javase/downloads/index.jsp
-     Add the /bin directory of this installation to your PATH environment
-     variable.
-
-  2. Configure the build:
-       bash ./configure
-
-  3. Build the OpenJDK:
-       make all
-     The resulting JDK image should be found in build/*/images/jdk
-
-where make is GNU make 3.81 or newer, /usr/bin/make on Linux usually
-is 3.81 or newer. Note that on Solaris, GNU make is called "gmake".
-
-Complete details are available in the file:
-     http://hg.openjdk.java.net/jdk9/jdk9/raw-file/tip/README-builds.html
+See http://openjdk.java.net/ for more information about OpenJDK.
--- a/README-builds.html	Tue Feb 14 14:57:56 2017 -0800
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,1406 +0,0 @@
-<html>
-  <head>
-    <title>OpenJDK Build README</title>
-  </head>
-  <body>
-<p><img src="http://openjdk.java.net/images/openjdk.png" alt="OpenJDK" title="" /></p>
-
-<h1>OpenJDK Build README</h1>
-
-<hr />
-
-<p><a name="introduction"></a></p>
-
-<h2>Introduction</h2>
-
-<p>This README file contains build instructions for the
-<a href="http://openjdk.java.net">OpenJDK</a>. Building the source code for the OpenJDK
-requires a certain degree of technical expertise.</p>
-
-<h3>!!!!!!!!!!!!!!! THIS IS A MAJOR RE-WRITE of this document. !!!!!!!!!!!!!</h3>
-
-<p>Some Headlines:</p>
-
-<ul>
-<li>The build is now a "<code>configure &amp;&amp; make</code>" style build</li>
-<li>Any GNU make 3.81 or newer should work, except on Windows where 4.0 or newer
-is recommended.</li>
-<li>The build should scale, i.e. more processors should cause the build to be
-done in less wall-clock time</li>
-<li>Nested or recursive make invocations have been significantly reduced,
-as has the total fork/exec or spawning of sub processes during the build</li>
-<li>Windows MKS usage is no longer supported</li>
-<li>Windows Visual Studio <code>vsvars*.bat</code> and <code>vcvars*.bat</code> files are run
-automatically</li>
-<li>Ant is no longer used when building the OpenJDK</li>
-<li>Use of ALT_* environment variables for configuring the build is no longer
-supported</li>
-</ul>
-
-<hr />
-
-<h2>Contents</h2>
-
-<ul>
-<li><a href="#introduction">Introduction</a></li>
-<li><a href="#hg">Use of Mercurial</a>
-<ul>
-<li><a href="#get_source">Getting the Source</a></li>
-<li><a href="#repositories">Repositories</a></li>
-</ul></li>
-<li><a href="#building">Building</a>
-<ul>
-<li><a href="#setup">System Setup</a>
-<ul>
-<li><a href="#linux">Linux</a></li>
-<li><a href="#solaris">Solaris</a></li>
-<li><a href="#macosx">Mac OS X</a></li>
-<li><a href="#windows">Windows</a></li>
-</ul></li>
-<li><a href="#configure">Configure</a></li>
-<li><a href="#make">Make</a></li>
-</ul></li>
-<li><a href="#testing">Testing</a></li>
-</ul>
-
-<hr />
-
-<ul>
-<li><a href="#hints">Appendix A: Hints and Tips</a>
-<ul>
-<li><a href="#faq">FAQ</a></li>
-<li><a href="#performance">Build Performance Tips</a></li>
-<li><a href="#troubleshooting">Troubleshooting</a></li>
-</ul></li>
-<li><a href="#gmake">Appendix B: GNU Make Information</a></li>
-<li><a href="#buildenvironments">Appendix C: Build Environments</a></li>
-</ul>
-
-<hr />
-
-<p><a name="hg"></a></p>
-
-<h2>Use of Mercurial</h2>
-
-<p>The OpenJDK sources are maintained with the revision control system
-<a href="http://mercurial.selenic.com/wiki/Mercurial">Mercurial</a>. If you are new to
-Mercurial, please see the <a href="http://mercurial.selenic.com/wiki/
-BeginnersGuides">Beginner Guides</a> or refer to the <a href="http://hgbook.red-bean.com/">Mercurial Book</a>.
-The first few chapters of the book provide an excellent overview of Mercurial,
-what it is and how it works.</p>
-
-<p>For using Mercurial with the OpenJDK refer to the <a href="http://openjdk.java.net/guide/
-repositories.html#installConfig">Developer Guide: Installing
-and Configuring Mercurial</a> section for more information.</p>
-
-<p><a name="get_source"></a></p>
-
-<h3>Getting the Source</h3>
-
-<p>To get the entire set of OpenJDK Mercurial repositories use the script
-<code>get_source.sh</code> located in the root repository:</p>
-
-<pre><code>  hg clone http://hg.openjdk.java.net/jdk9/jdk9 YourOpenJDK
-  cd YourOpenJDK
-  bash ./get_source.sh
-</code></pre>
-
-<p>Once you have all the repositories, keep in mind that each repository is its
-own independent repository. You can also re-run <code>./get_source.sh</code> anytime to
-pull over all the latest changesets in all the repositories. This set of
-nested repositories has been given the term "forest" and there are various
-ways to apply the same <code>hg</code> command to each of the repositories. For
-example, the script <code>make/scripts/hgforest.sh</code> can be used to repeat the
-same <code>hg</code> command on every repository, e.g.</p>
-
-<pre><code>  cd YourOpenJDK
-  bash ./make/scripts/hgforest.sh status
-</code></pre>
-
-<p><a name="repositories"></a></p>
-
-<h3>Repositories</h3>
-
-<p>The set of repositories and what they contain:</p>
-
-<ul>
-<li><strong>. (root)</strong> contains common configure and makefile logic</li>
-<li><strong>hotspot</strong> contains source code and make files for building the OpenJDK
-Hotspot Virtual Machine</li>
-<li><strong>langtools</strong> contains source code for the OpenJDK javac and language tools</li>
-<li><strong>jdk</strong> contains source code and make files for building the OpenJDK runtime
-libraries and misc files</li>
-<li><strong>jaxp</strong> contains source code for the OpenJDK JAXP functionality</li>
-<li><strong>jaxws</strong> contains source code for the OpenJDK JAX-WS functionality</li>
-<li><strong>corba</strong> contains source code for the OpenJDK Corba functionality</li>
-<li><strong>nashorn</strong> contains source code for the OpenJDK JavaScript implementation</li>
-</ul>
-
-<h3>Repository Source Guidelines</h3>
-
-<p>There are some very basic guidelines:</p>
-
-<ul>
-<li>Use of whitespace in source files (.java, .c, .h, .cpp, and .hpp files) is
-restricted. No TABs, no trailing whitespace on lines, and files should not
-terminate in more than one blank line.</li>
-<li>Files with execute permissions should not be added to the source
-repositories.</li>
-<li>All generated files need to be kept isolated from the files maintained or
-managed by the source control system. The standard area for generated files
-is the top level <code>build/</code> directory.</li>
-<li>The default build process should be to build the product and nothing else,
-in one form, e.g. a product (optimized), debug (non-optimized, -g plus
-assert logic), or fastdebug (optimized, -g plus assert logic).</li>
-<li>The <code>.hgignore</code> file in each repository must exist and should include
-<code>^build/</code>, <code>^dist/</code> and optionally any <code>nbproject/private</code> directories. <strong>It
-should NEVER</strong> include anything in the <code>src/</code> or <code>test/</code> or any managed
-directory area of a repository.</li>
-<li>Directory names and file names should never contain blanks or non-printing
-characters.</li>
-<li>Generated source or binary files should NEVER be added to the repository
-(that includes <code>javah</code> output). There are some exceptions to this rule, in
-particular with some of the generated configure scripts.</li>
-<li>Files not needed for typical building or testing of the repository should
-not be added to the repository.</li>
-</ul>
-
-<hr />
-
-<p><a name="building"></a></p>
-
-<h2>Building</h2>
-
-<p>The very first step in building the OpenJDK is making sure the system itself
-has everything it needs to do OpenJDK builds. Once a system is setup, it
-generally doesn't need to be done again.</p>
-
-<p>Building the OpenJDK is now done with running a <code>configure</code> script which will
-try and find and verify you have everything you need, followed by running
-<code>make</code>, e.g.</p>
-
-<blockquote>
-  <p><strong><code>bash ./configure</code></strong> <br />
- <strong><code>make all</code></strong></p>
-</blockquote>
-
-<p>Where possible the <code>configure</code> script will attempt to located the various
-components in the default locations or via component specific variable
-settings. When the normal defaults fail or components cannot be found,
-additional <code>configure</code> options may be necessary to help <code>configure</code> find the
-necessary tools for the build, or you may need to re-visit the setup of your
-system due to missing software packages.</p>
-
-<p><strong>NOTE:</strong> The <code>configure</code> script file does not have execute permissions and
-will need to be explicitly run with <code>bash</code>, see the source guidelines.</p>
-
-<hr />
-
-<p><a name="setup"></a></p>
-
-<h3>System Setup</h3>
-
-<p>Before even attempting to use a system to build the OpenJDK there are some very
-basic system setups needed. For all systems:</p>
-
-<ul>
-<li><p>Be sure the GNU make utility is version 3.81 (4.0 on windows) or newer, e.g.
-run "<code>make -version</code>"</p>
-
-<p><a name="bootjdk"></a></p></li>
-<li><p>Install a Bootstrap JDK. All OpenJDK builds require access to a previously
-released JDK called the <em>bootstrap JDK</em> or <em>boot JDK.</em> The general rule is
-that the bootstrap JDK must be an instance of the previous major release of
-the JDK. In addition, there may be a requirement to use a release at or
-beyond a particular update level.</p>
-
-<p><strong><em>Building JDK 9 requires JDK 8. JDK 9 developers should not use JDK 9 as
-the boot JDK, to ensure that JDK 9 dependencies are not introduced into the
-parts of the system that are built with JDK 8.</em></strong></p>
-
-<p>The JDK 8 binaries can be downloaded from Oracle's <a href="http://www.oracle.com/technetwork/java/javase/downloads/index.html">JDK 8 download
-site</a>.
-For build performance reasons it is very important that this bootstrap JDK
-be made available on the local disk of the machine doing the build. You
-should add its <code>bin</code> directory to the <code>PATH</code> environment variable. If
-<code>configure</code> has any issues finding this JDK, you may need to use the
-<code>configure</code> option <code>--with-boot-jdk</code>.</p></li>
-<li><p>Ensure that GNU make, the Bootstrap JDK, and the compilers are all in your
-PATH environment variable.</p></li>
-</ul>
-
-<p>And for specific systems:</p>
-
-<ul>
-<li><p><strong>Linux</strong></p>
-
-<p>Install all the software development packages needed including
-<a href="#alsa">alsa</a>, <a href="#freetype">freetype</a>, <a href="#cups">cups</a>, and
-<a href="#xrender">xrender</a>. See <a href="#SDBE">specific system packages</a>.</p></li>
-<li><p><strong>Solaris</strong></p>
-
-<p>Install all the software development packages needed including <a href="#studio">Studio
-Compilers</a>, <a href="#freetype">freetype</a>, <a href="#cups">cups</a>, and
-<a href="#xrender">xrender</a>. See <a href="#SDBE">specific system packages</a>.</p></li>
-<li><p><strong>Windows</strong></p>
-
-<ul>
-<li>Install one of <a href="#cygwin">CYGWIN</a> or <a href="#msys">MinGW/MSYS</a></li>
-<li>Install <a href="#vs2013">Visual Studio 2013</a></li>
-</ul></li>
-<li><p><strong>Mac OS X</strong></p>
-
-<p>Install <a href="https://developer.apple.com/xcode/">XCode 6.3</a></p></li>
-</ul>
-
-<p><a name="linux"></a></p>
-
-<h4>Linux</h4>
-
-<p>With Linux, try and favor the system packages over building your own or getting
-packages from other areas. Most Linux builds should be possible with the
-system's available packages.</p>
-
-<p>Note that some Linux systems have a habit of pre-populating your environment
-variables for you, for example <code>JAVA_HOME</code> might get pre-defined for you to
-refer to the JDK installed on your Linux system. You will need to unset
-<code>JAVA_HOME</code>. It's a good idea to run <code>env</code> and verify the environment variables
-you are getting from the default system settings make sense for building the
-OpenJDK.</p>
-
-<p><a name="solaris"></a></p>
-
-<h4>Solaris</h4>
-
-<p><a name="studio"></a></p>
-
-<h5>Studio Compilers</h5>
-
-<p>At a minimum, the <a href="http://www.oracle.com/
-technetwork/server-storage/solarisstudio/downloads/index.htm">Studio 12 Update 4 Compilers</a> (containing
-version 5.13 of the C and C++ compilers) is required, including specific
-patches.</p>
-
-<p>The Solaris Studio installation should contain at least these packages:</p>
-
-<blockquote>
-  <p><table border="1">
-     <thead>
-       <tr>
-         <td><strong>Package</strong></td>
-         <td><strong>Version</strong></td>
-       </tr>
-     </thead>
-     <tbody>
-       <tr>
-         <td>developer/solarisstudio-124/backend</td>
-         <td>12.4-1.0.6.0</td>
-       </tr>
-       <tr>
-         <td>developer/solarisstudio-124/c++</td>
-         <td>12.4-1.0.10.0</td>
-       </tr>
-       <tr>
-         <td>developer/solarisstudio-124/cc</td>
-         <td>12.4-1.0.4.0</td>
-       </tr>
-       <tr>
-         <td>developer/solarisstudio-124/library/c++-libs</td>
-         <td>12.4-1.0.10.0</td>
-       </tr>
-       <tr>
-         <td>developer/solarisstudio-124/library/math-libs</td>
-         <td>12.4-1.0.0.1</td>
-       </tr>
-       <tr>
-         <td>developer/solarisstudio-124/library/studio-gccrt</td>
-         <td>12.4-1.0.0.1</td>
-       </tr>
-       <tr>
-         <td>developer/solarisstudio-124/studio-common</td>
-         <td>12.4-1.0.0.1</td>
-       </tr>
-       <tr>
-         <td>developer/solarisstudio-124/studio-ja</td>
-         <td>12.4-1.0.0.1</td>
-       </tr>
-       <tr>
-         <td>developer/solarisstudio-124/studio-legal</td>
-         <td>12.4-1.0.0.1</td>
-       </tr>
-       <tr>
-         <td>developer/solarisstudio-124/studio-zhCN</td>
-         <td>12.4-1.0.0.1</td>
-       </tr>
-     </tbody>
-   </table></p>
-</blockquote>
-
-<p>In particular backend 12.4-1.0.6.0 contains a critical patch for the sparc
-version.</p>
-
-<p>Place the <code>bin</code> directory in <code>PATH</code>.</p>
-
-<p>The Oracle Solaris Studio Express compilers at: <a href="http://www.oracle.com/technetwork/server-storage/solarisstudio/
-downloads/index-jsp-142582.html">Oracle Solaris Studio Express
-Download site</a> are also an option, although these compilers
-have not been extensively used yet.</p>
-
-<p><a name="windows"></a></p>
-
-<h4>Windows</h4>
-
-<h5>Windows Unix Toolkit</h5>
-
-<p>Building on Windows requires a Unix-like environment, notably a Unix-like
-shell. There are several such environments available of which
-<a href="http://www.cygwin.com/">Cygwin</a> and
-<a href="http://www.mingw.org/wiki/MSYS">MinGW/MSYS</a> are currently supported for the
-OpenJDK build. One of the differences of these systems from standard Windows
-tools is the way they handle Windows path names, particularly path names which
-contain spaces, backslashes as path separators and possibly drive letters.
-Depending on the use case and the specifics of each environment these path
-problems can be solved by a combination of quoting whole paths, translating
-backslashes to forward slashes, escaping backslashes with additional
-backslashes and translating the path names to their <a href="http://en.wikipedia.org/wiki/8.3_filename">"8.3"
-version</a>.</p>
-
-<p><a name="cygwin"></a></p>
-
-<h6>CYGWIN</h6>
-
-<p>CYGWIN is an open source, Linux-like environment which tries to emulate a
-complete POSIX layer on Windows. It tries to be smart about path names and can
-usually handle all kinds of paths if they are correctly quoted or escaped
-although internally it maps drive letters <code>&lt;drive&gt;:</code> to a virtual directory
-<code>/cygdrive/&lt;drive&gt;</code>.</p>
-
-<p>You can always use the <code>cygpath</code> utility to map pathnames with spaces or the
-backslash character into the <code>C:/</code> style of pathname (called 'mixed'), e.g.
-<code>cygpath -s -m "&lt;path&gt;"</code>.</p>
-
-<p>Note that the use of CYGWIN creates a unique problem with regards to setting
-<a href="#path"><code>PATH</code></a>. Normally on Windows the <code>PATH</code> variable contains directories
-separated with the ";" character (Solaris and Linux use ":"). With CYGWIN, it
-uses ":", but that means that paths like "C:/path" cannot be placed in the
-CYGWIN version of <code>PATH</code> and instead CYGWIN uses something like
-<code>/cygdrive/c/path</code> which CYGWIN understands, but only CYGWIN understands.</p>
-
-<p>The OpenJDK build requires CYGWIN version 1.7.16 or newer. Information about
-CYGWIN can be obtained from the CYGWIN website at
-<a href="http://www.cygwin.com">www.cygwin.com</a>.</p>
-
-<p>By default CYGWIN doesn't install all the tools required for building the
-OpenJDK. Along with the default installation, you need to install the following
-tools.</p>
-
-<blockquote>
-  <p><table border="1">
-     <thead>
-       <tr>
-         <td>Binary Name</td>
-         <td>Category</td>
-         <td>Package</td>
-         <td>Description</td>
-      </tr>
-     </thead>
-     <tbody>
-       <tr>
-         <td>ar.exe</td>
-         <td>Devel</td>
-         <td>binutils</td>
-         <td>The GNU assembler, linker and binary utilities</td>
-       </tr>
-       <tr>
-         <td>make.exe</td>
-         <td>Devel</td>
-         <td>make</td>
-         <td>The GNU version of the 'make' utility built for CYGWIN</td>
-       </tr>
-       <tr>
-         <td>m4.exe</td>
-         <td>Interpreters</td>
-         <td>m4</td>
-         <td>GNU implementation of the traditional Unix macro processor</td>
-       </tr>
-       <tr>
-         <td>cpio.exe</td>
-         <td>Utils</td>
-         <td>cpio</td>
-         <td>A program to manage archives of files</td>
-       </tr>
-       <tr>
-         <td>gawk.exe</td>
-         <td>Utils</td>
-         <td>awk</td>
-         <td>Pattern-directed scanning and processing language</td>
-       </tr>
-       <tr>
-         <td>file.exe</td>
-         <td>Utils</td>
-         <td>file</td>
-         <td>Determines file type using 'magic' numbers</td>
-       </tr>
-       <tr>
-         <td>zip.exe</td>
-         <td>Archive</td>
-         <td>zip</td>
-         <td>Package and compress (archive) files</td>
-       </tr>
-       <tr>
-         <td>unzip.exe</td>
-         <td>Archive</td>
-         <td>unzip</td>
-         <td>Extract compressed files in a ZIP archive</td>
-       </tr>
-       <tr>
-         <td>free.exe</td>
-         <td>System</td>
-         <td>procps</td>
-         <td>Display amount of free and used memory in the system</td>
-       </tr>
-     </tbody>
-   </table></p>
-</blockquote>
-
-<p>Note that the CYGWIN software can conflict with other non-CYGWIN software on
-your Windows system. CYGWIN provides a <a href="http://cygwin.com/faq/
-faq.using.html">FAQ</a> for known issues and problems, of particular interest is the
-section on <a href="http://cygwin.com/faq/faq.using.html#faq.using.bloda">BLODA (applications that interfere with
-CYGWIN)</a>.</p>
-
-<p><a name="msys"></a></p>
-
-<h6>MinGW/MSYS</h6>
-
-<p>MinGW ("Minimalist GNU for Windows") is a collection of free Windows specific
-header files and import libraries combined with GNU toolsets that allow one to
-produce native Windows programs that do not rely on any 3rd-party C runtime
-DLLs. MSYS is a supplement to MinGW which allows building applications and
-programs which rely on traditional UNIX tools to be present. Among others this
-includes tools like <code>bash</code> and <code>make</code>. See <a href="http://www.mingw.org/
-wiki/MSYS">MinGW/MSYS</a> for more information.</p>
-
-<p>Like Cygwin, MinGW/MSYS can handle different types of path formats. They are
-internally converted to paths with forward slashes and drive letters
-<code>&lt;drive&gt;:</code> replaced by a virtual directory <code>/&lt;drive&gt;</code>. Additionally, MSYS
-automatically detects binaries compiled for the MSYS environment and feeds them
-with the internal, Unix-style path names. If native Windows applications are
-called from within MSYS programs their path arguments are automatically
-converted back to Windows style path names with drive letters and backslashes
-as path separators. This may cause problems for Windows applications which use
-forward slashes as parameter separator (e.g. <code>cl /nologo /I</code>) because MSYS may
-wrongly <a href="http://mingw.org/wiki/
-Posix_path_conversion">replace such parameters by drive letters</a>.</p>
-
-<p>In addition to the tools which will be installed by default, you have to
-manually install the <code>msys-zip</code> and <code>msys-unzip</code> packages. This can be easily
-done with the MinGW command line installer:</p>
-
-<pre><code>  mingw-get.exe install msys-zip
-  mingw-get.exe install msys-unzip
-</code></pre>
-
-<p><a name="vs2013"></a></p>
-
-<h5>Visual Studio 2013 Compilers</h5>
-
-<p>The 32-bit and 64-bit OpenJDK Windows build requires Microsoft Visual Studio
-C++ 2013 (VS2013) Professional Edition or Express compiler. The compiler and
-other tools are expected to reside in the location defined by the variable
-<code>VS120COMNTOOLS</code> which is set by the Microsoft Visual Studio installer.</p>
-
-<p>Only the C++ part of VS2013 is needed. Try to let the installation go to the
-default install directory. Always reboot your system after installing VS2013.
-The system environment variable VS120COMNTOOLS should be set in your
-environment.</p>
-
-<p>Make sure that TMP and TEMP are also set in the environment and refer to
-Windows paths that exist, like <code>C:\temp</code>, not <code>/tmp</code>, not <code>/cygdrive/c/temp</code>,
-and not <code>C:/temp</code>. <code>C:\temp</code> is just an example, it is assumed that this area
-is private to the user, so by default after installs you should see a unique
-user path in these variables.</p>
-
-<p><a name="macosx"></a></p>
-
-<h4>Mac OS X</h4>
-
-<p>Make sure you get the right XCode version.</p>
-
-<hr />
-
-<p><a name="configure"></a></p>
-
-<h3>Configure</h3>
-
-<p>The basic invocation of the <code>configure</code> script looks like:</p>
-
-<blockquote>
-  <p><strong><code>bash ./configure [options]</code></strong></p>
-</blockquote>
-
-<p>This will create an output directory containing the "configuration" and setup
-an area for the build result. This directory typically looks like:</p>
-
-<blockquote>
-  <p><strong><code>build/linux-x64-normal-server-release</code></strong></p>
-</blockquote>
-
-<p><code>configure</code> will try to figure out what system you are running on and where all
-necessary build components are. If you have all prerequisites for building
-installed, it should find everything. If it fails to detect any component
-automatically, it will exit and inform you about the problem. When this
-happens, read more below in <a href="#configureoptions">the <code>configure</code> options</a>.</p>
-
-<p>Some examples:</p>
-
-<blockquote>
-  <p><strong>Windows 32bit build with freetype specified:</strong> <br />
- <code>bash ./configure --with-freetype=/cygdrive/c/freetype-i586 --with-target-
-bits=32</code></p>
-
-<p><strong>Debug 64bit Build:</strong> <br />
- <code>bash ./configure --enable-debug --with-target-bits=64</code></p>
-</blockquote>
-
-<p><a name="configureoptions"></a></p>
-
-<h4>Configure Options</h4>
-
-<p>Complete details on all the OpenJDK <code>configure</code> options can be seen with:</p>
-
-<blockquote>
-  <p><strong><code>bash ./configure --help=short</code></strong></p>
-</blockquote>
-
-<p>Use <code>-help</code> to see all the <code>configure</code> options available. You can generate any
-number of different configurations, e.g. debug, release, 32, 64, etc.</p>
-
-<p>Some of the more commonly used <code>configure</code> options are:</p>
-
-<blockquote>
-  <p><strong><code>--enable-debug</code></strong> <br />
- set the debug level to fastdebug (this is a shorthand for <code>--with-debug-
-   level=fastdebug</code>)</p>
-</blockquote>
-
-<p><a name="alsa"></a></p>
-
-<blockquote>
-  <p><strong><code>--with-alsa=</code></strong><em>path</em> <br />
- select the location of the Advanced Linux Sound Architecture (ALSA)</p>
-
-<p>Version 0.9.1 or newer of the ALSA files are required for building the
-   OpenJDK on Linux. These Linux files are usually available from an "alsa" of
-   "libasound" development package, and it's highly recommended that you try
-   and use the package provided by the particular version of Linux that you are
-   using.</p>
-
-<p><strong><code>--with-boot-jdk=</code></strong><em>path</em> <br />
- select the <a href="#bootjdk">Bootstrap JDK</a></p>
-
-<p><strong><code>--with-boot-jdk-jvmargs=</code></strong>"<em>args</em>" <br />
- provide the JVM options to be used to run the <a href="#bootjdk">Bootstrap JDK</a></p>
-
-<p><strong><code>--with-cacerts=</code></strong><em>path</em> <br />
- select the path to the cacerts file.</p>
-
-<p>See <a href="http://en.wikipedia.org/wiki/
-   Certificate_Authority">Certificate Authority on Wikipedia</a> for a better understanding of the Certificate
-   Authority (CA). A certificates file named "cacerts" represents a system-wide
-   keystore with CA certificates. In JDK and JRE binary bundles, the "cacerts"
-   file contains root CA certificates from several public CAs (e.g., VeriSign,
-   Thawte, and Baltimore). The source contain a cacerts file without CA root
-   certificates. Formal JDK builders will need to secure permission from each
-   public CA and include the certificates into their own custom cacerts file.
-   Failure to provide a populated cacerts file will result in verification
-   errors of a certificate chain during runtime. By default an empty cacerts
-   file is provided and that should be fine for most JDK developers.</p>
-</blockquote>
-
-<p><a name="cups"></a></p>
-
-<blockquote>
-  <p><strong><code>--with-cups=</code></strong><em>path</em> <br />
- select the CUPS install location</p>
-
-<p>The Common UNIX Printing System (CUPS) Headers are required for building the
-   OpenJDK on Solaris and Linux. The Solaris header files can be obtained by
-   installing the package <strong>print/cups</strong>.</p>
-
-<p>The CUPS header files can always be downloaded from
-   <a href="http://www.cups.org">www.cups.org</a>.</p>
-
-<p><strong><code>--with-cups-include=</code></strong><em>path</em> <br />
- select the CUPS include directory location</p>
-
-<p><strong><code>--with-debug-level=</code></strong><em>level</em> <br />
- select the debug information level of release, fastdebug, or slowdebug</p>
-
-<p><strong><code>--with-dev-kit=</code></strong><em>path</em> <br />
- select location of the compiler install or developer install location</p>
-</blockquote>
-
-<p><a name="freetype"></a></p>
-
-<blockquote>
-  <p><strong><code>--with-freetype=</code></strong><em>path</em> <br />
- select the freetype files to use.</p>
-
-<p>Expecting the freetype libraries under <code>lib/</code> and the headers under
-   <code>include/</code>.</p>
-
-<p>Version 2.3 or newer of FreeType is required. On Unix systems required files
-   can be available as part of your distribution (while you still may need to
-   upgrade them). Note that you need development version of package that
-   includes both the FreeType library and header files.</p>
-
-<p>You can always download latest FreeType version from the <a href="http://www.freetype.org">FreeType
-   website</a>. Building the freetype 2 libraries from
-   scratch is also possible, however on Windows refer to the <a href="http://freetype.freedesktop.org/wiki/FreeType_DLL">Windows FreeType
-   DLL build instructions</a>.</p>
-
-<p>Note that by default FreeType is built with byte code hinting support
-   disabled due to licensing restrictions. In this case, text appearance and
-   metrics are expected to differ from Sun's official JDK build. See the
-   <a href="http://freetype.sourceforge.net/freetype2">SourceForge FreeType2 Home Page</a>
-   for more information.</p>
-
-<p><strong><code>--with-import-hotspot=</code></strong><em>path</em> <br />
- select the location to find hotspot binaries from a previous build to avoid
-   building hotspot</p>
-
-<p><strong><code>--with-target-bits=</code></strong><em>arg</em> <br />
- select 32 or 64 bit build</p>
-
-<p><strong><code>--with-jvm-variants=</code></strong><em>variants</em> <br />
- select the JVM variants to build from, comma separated list that can
-   include: server, client, kernel, zero and zeroshark</p>
-
-<p><strong><code>--with-memory-size=</code></strong><em>size</em> <br />
- select the RAM size that GNU make will think this system has</p>
-
-<p><strong><code>--with-msvcr-dll=</code></strong><em>path</em> <br />
- select the <code>msvcr100.dll</code> file to include in the Windows builds (C/C++
-   runtime library for Visual Studio).</p>
-
-<p>This is usually picked up automatically from the redist directories of
-   Visual Studio 2013.</p>
-
-<p><strong><code>--with-num-cores=</code></strong><em>cores</em> <br />
- select the number of cores to use (processor count or CPU count)</p>
-</blockquote>
-
-<p><a name="xrender"></a></p>
-
-<blockquote>
-  <p><strong><code>--with-x=</code></strong><em>path</em> <br />
- select the location of the X11 and xrender files.</p>
-
-<p>The XRender Extension Headers are required for building the OpenJDK on
-   Solaris and Linux. The Linux header files are usually available from a
-   "Xrender" development package, it's recommended that you try and use the
-   package provided by the particular distribution of Linux that you are using.
-   The Solaris XRender header files is included with the other X11 header files
-   in the package <strong>SFWxwinc</strong> on new enough versions of Solaris and will be
-   installed in <code>/usr/X11/include/X11/extensions/Xrender.h</code> or
-   <code>/usr/openwin/share/include/X11/extensions/Xrender.h</code></p>
-</blockquote>
-
-<hr />
-
-<p><a name="make"></a></p>
-
-<h3>Make</h3>
-
-<p>The basic invocation of the <code>make</code> utility looks like:</p>
-
-<blockquote>
-  <p><strong><code>make all</code></strong></p>
-</blockquote>
-
-<p>This will start the build to the output directory containing the
-"configuration" that was created by the <code>configure</code> script. Run <code>make help</code> for
-more information on the available targets.</p>
-
-<p>There are some of the make targets that are of general interest:</p>
-
-<blockquote>
-  <p><em>empty</em> <br />
- build everything but no images</p>
-
-<p><strong><code>all</code></strong> <br />
- build everything including images</p>
-
-<p><strong><code>all-conf</code></strong> <br />
- build all configurations</p>
-
-<p><strong><code>images</code></strong> <br />
- create complete j2sdk and j2re images</p>
-
-<p><strong><code>install</code></strong> <br />
- install the generated images locally, typically in <code>/usr/local</code></p>
-
-<p><strong><code>clean</code></strong> <br />
- remove all files generated by make, but not those generated by <code>configure</code></p>
-
-<p><strong><code>dist-clean</code></strong> <br />
- remove all files generated by both and <code>configure</code> (basically killing the
-   configuration)</p>
-
-<p><strong><code>help</code></strong> <br />
- give some help on using <code>make</code>, including some interesting make targets</p>
-</blockquote>
-
-<hr />
-
-<p><a name="testing"></a></p>
-
-<h2>Testing</h2>
-
-<p>When the build is completed, you should see the generated binaries and
-associated files in the <code>j2sdk-image</code> directory in the output directory. In
-particular, the <code>build/*/images/j2sdk-image/bin</code> directory should contain
-executables for the OpenJDK tools and utilities for that configuration. The
-testing tool <code>jtreg</code> will be needed and can be found at: <a href="http://openjdk.java.net/jtreg/">the jtreg
-site</a>. The provided regression tests in the
-repositories can be run with the command:</p>
-
-<blockquote>
-  <p><strong><code>cd test &amp;&amp; make PRODUCT_HOME=`pwd`/../build/*/images/j2sdk-image all</code></strong></p>
-</blockquote>
-
-<hr />
-
-<p><a name="hints"></a></p>
-
-<h2>Appendix A: Hints and Tips</h2>
-
-<p><a name="faq"></a></p>
-
-<h3>FAQ</h3>
-
-<p><strong>Q:</strong> The <code>generated-configure.sh</code> file looks horrible! How are you going to
-edit it? <br />
-<strong>A:</strong> The <code>generated-configure.sh</code> file is generated (think "compiled") by the
-autoconf tools. The source code is in <code>configure.ac</code> and various .m4 files in
-common/autoconf, which are much more readable.</p>
-
-<p><strong>Q:</strong> Why is the <code>generated-configure.sh</code> file checked in, if it is 
-generated? <br />
-<strong>A:</strong> If it was not generated, every user would need to have the autoconf
-tools installed, and re-generate the <code>configure</code> file as the first step. Our
-goal is to minimize the work needed to be done by the user to start building
-OpenJDK, and to minimize the number of external dependencies required.</p>
-
-<p><strong>Q:</strong> Do you require a specific version of autoconf for regenerating
-<code>generated-configure.sh</code>? <br />
-<strong>A:</strong> Yes, version 2.69 is required and should be easy enough to aquire on all
-supported operating systems. The reason for this is to avoid large spurious
-changes in <code>generated-configure.sh</code>.</p>
-
-<p><strong>Q:</strong> How do you regenerate <code>generated-configure.sh</code> after making changes to
-the input files? <br />
-<strong>A:</strong> Regnerating <code>generated-configure.sh</code> should always be done using the
-script <code>common/autoconf/autogen.sh</code> to ensure that the correct files get
-updated. This script should also be run after mercurial tries to merge
-<code>generated-configure.sh</code> as a merge of the generated file is not guaranteed to
-be correct.</p>
-
-<p><strong>Q:</strong> What are the files in <code>common/makefiles/support/*</code> for? They look like
-gibberish. <br />
-<strong>A:</strong> They are a somewhat ugly hack to compensate for command line length
-limitations on certain platforms (Windows, Solaris). Due to a combination of
-limitations in make and the shell, command lines containing too many files will
-not work properly. These helper files are part of an elaborate hack that will
-compress the command line in the makefile and then uncompress it safely. We're
-not proud of it, but it does fix the problem. If you have any better
-suggestions, we're all ears! :-)</p>
-
-<p><strong>Q:</strong> I want to see the output of the commands that make runs, like in the old
-build. How do I do that? <br />
-<strong>A:</strong> You specify the <code>LOG</code> variable to make. There are several log levels:</p>
-
-<ul>
-<li><strong><code>warn</code></strong> -- Default and very quiet.</li>
-<li><strong><code>info</code></strong> -- Shows more progress information than warn.</li>
-<li><strong><code>debug</code></strong> -- Echos all command lines and prints all macro calls for
-compilation definitions.</li>
-<li><strong><code>trace</code></strong> -- Echos all $(shell) command lines as well.</li>
-</ul>
-
-<p><strong>Q:</strong> When do I have to re-run <code>configure</code>? <br />
-<strong>A:</strong> Normally you will run <code>configure</code> only once for creating a
-configuration. You need to re-run configuration only if you want to change any
-configuration options, or if you pull down changes to the <code>configure</code> script.</p>
-
-<p><strong>Q:</strong> I have added a new source file. Do I need to modify the makefiles? <br />
-<strong>A:</strong> Normally, no. If you want to create e.g. a new native library, you will
-need to modify the makefiles. But for normal file additions or removals, no
-changes are needed. There are certan exceptions for some native libraries where
-the source files are spread over many directories which also contain sources
-for other libraries. In these cases it was simply easier to create include
-lists rather than excludes.</p>
-
-<p><strong>Q:</strong> When I run <code>configure --help</code>, I see many strange options, like
-<code>--dvidir</code>. What is this? <br />
-<strong>A:</strong> Configure provides a slew of options by default, to all projects that
-use autoconf. Most of them are not used in OpenJDK, so you can safely ignore
-them. To list only OpenJDK specific features, use <code>configure --help=short</code>
-instead.</p>
-
-<p><strong>Q:</strong> <code>configure</code> provides OpenJDK-specific features such as <code>--with-
-builddeps-server</code> that are not described in this document. What about those? <br />
-<strong>A:</strong> Try them out if you like! But be aware that most of these are
-experimental features. Many of them don't do anything at all at the moment; the
-option is just a placeholder. Others depend on pieces of code or infrastructure
-that is currently not ready for prime time.</p>
-
-<p><strong>Q:</strong> How will you make sure you don't break anything? <br />
-<strong>A:</strong> We have a script that compares the result of the new build system with
-the result of the old. For most part, we aim for (and achieve) byte-by-byte
-identical output. There are however technical issues with e.g. native binaries,
-which might differ in a byte-by-byte comparison, even when building twice with
-the old build system. For these, we compare relevant aspects (e.g. the symbol
-table and file size). Note that we still don't have 100% equivalence, but we're
-close.</p>
-
-<p><strong>Q:</strong> I noticed this thing X in the build that looks very broken by design.
-Why don't you fix it? <br />
-<strong>A:</strong> Our goal is to produce a build output that is as close as technically
-possible to the old build output. If things were weird in the old build, they
-will be weird in the new build. Often, things were weird before due to
-obscurity, but in the new build system the weird stuff comes up to the surface.
-The plan is to attack these things at a later stage, after the new build system
-is established.</p>
-
-<p><strong>Q:</strong> The code in the new build system is not that well-structured. Will you
-fix this? <br />
-<strong>A:</strong> Yes! The new build system has grown bit by bit as we converted the old
-system. When all of the old build system is converted, we can take a step back
-and clean up the structure of the new build system. Some of this we plan to do
-before replacing the old build system and some will need to wait until after.</p>
-
-<p><strong>Q:</strong> Is anything able to use the results of the new build's default make
-target? <br />
-<strong>A:</strong> Yes, this is the minimal (or roughly minimal) set of compiled output
-needed for a developer to actually execute the newly built JDK. The idea is
-that in an incremental development fashion, when doing a normal make, you
-should only spend time recompiling what's changed (making it purely
-incremental) and only do the work that's needed to actually run and test your
-code. The packaging stuff that is part of the <code>images</code> target is not needed for
-a normal developer who wants to test his new code. Even if it's quite fast,
-it's still unnecessary. We're targeting sub-second incremental rebuilds! ;-)
-(Or, well, at least single-digit seconds...)</p>
-
-<p><strong>Q:</strong> I usually set a specific environment variable when building, but I can't
-find the equivalent in the new build. What should I do? <br />
-<strong>A:</strong> It might very well be that we have neglected to add support for an
-option that was actually used from outside the build system. Email us and we
-will add support for it!</p>
-
-<p><a name="performance"></a></p>
-
-<h3>Build Performance Tips</h3>
-
-<p>Building OpenJDK requires a lot of horsepower. Some of the build tools can be
-adjusted to utilize more or less of resources such as parallel threads and
-memory. The <code>configure</code> script analyzes your system and selects reasonable
-values for such options based on your hardware. If you encounter resource
-problems, such as out of memory conditions, you can modify the detected values
-with:</p>
-
-<ul>
-<li><strong><code>--with-num-cores</code></strong> -- number of cores in the build system, e.g.
-<code>--with-num-cores=8</code></li>
-<li><strong><code>--with-memory-size</code></strong> -- memory (in MB) available in the build system,
-e.g. <code>--with-memory-size=1024</code></li>
-</ul>
-
-<p>It might also be necessary to specify the JVM arguments passed to the Bootstrap
-JDK, using e.g. <code>--with-boot-jdk-jvmargs="-Xmx8G -enableassertions"</code>. Doing
-this will override the default JVM arguments passed to the Bootstrap JDK.</p>
-
-<p>One of the top goals of the new build system is to improve the build
-performance and decrease the time needed to build. This will soon also apply to
-the java compilation when the Smart Javac wrapper is fully supported.</p>
-
-<p>At the end of a successful execution of <code>configure</code>, you will get a performance
-summary, indicating how well the build will perform. Here you will also get
-performance hints. If you want to build fast, pay attention to those!</p>
-
-<h4>Building with ccache</h4>
-
-<p>The OpenJDK build supports building with ccache when using gcc or clang. Using
-ccache can radically speed up compilation of native code if you often rebuild
-the same sources. Your milage may vary however so we recommend evaluating it
-for yourself. To enable it, make sure it's on the path and configure with
-<code>--enable-ccache</code>.</p>
-
-<h4>Building on local disk</h4>
-
-<p>If you are using network shares, e.g. via NFS, for your source code, make sure
-the build directory is situated on local disk. The performance penalty is
-extremely high for building on a network share, close to unusable.</p>
-
-<h4>Building only one JVM</h4>
-
-<p>The old build builds multiple JVMs on 32-bit systems (client and server; and on
-Windows kernel as well). In the new build we have changed this default to only
-build server when it's available. This improves build times for those not
-interested in multiple JVMs. To mimic the old behavior on platforms that
-support it, use <code>--with-jvm-variants=client,server</code>.</p>
-
-<h4>Selecting the number of cores to build on</h4>
-
-<p>By default, <code>configure</code> will analyze your machine and run the make process in
-parallel with as many threads as you have cores. This behavior can be
-overridden, either "permanently" (on a <code>configure</code> basis) using
-<code>--with-num-cores=N</code> or for a single build only (on a make basis), using
-<code>make JOBS=N</code>.</p>
-
-<p>If you want to make a slower build just this time, to save some CPU power for
-other processes, you can run e.g. <code>make JOBS=2</code>. This will force the makefiles
-to only run 2 parallel processes, or even <code>make JOBS=1</code> which will disable
-parallelism.</p>
-
-<p>If you want to have it the other way round, namely having slow builds default
-and override with fast if you're impatient, you should call <code>configure</code> with
-<code>--with-num-cores=2</code>, making 2 the default. If you want to run with more cores,
-run <code>make JOBS=8</code></p>
-
-<p><a name="troubleshooting"></a></p>
-
-<h3>Troubleshooting</h3>
-
-<h4>Solving build problems</h4>
-
-<p>If the build fails (and it's not due to a compilation error in a source file
-you've changed), the first thing you should do is to re-run the build with more
-verbosity. Do this by adding <code>LOG=debug</code> to your make command line.</p>
-
-<p>The build log (with both stdout and stderr intermingled, basically the same as
-you see on your console) can be found as <code>build.log</code> in your build directory.</p>
-
-<p>You can ask for help on build problems with the new build system on either the
-<a href="http://mail.openjdk.java.net/mailman/listinfo/build-dev">build-dev</a> or the
-<a href="http://mail.openjdk.java.net/mailman/listinfo/build-infra-dev">build-infra-dev</a>
-mailing lists. Please include the relevant parts of the build log.</p>
-
-<p>A build can fail for any number of reasons. Most failures are a result of
-trying to build in an environment in which all the pre-build requirements have
-not been met. The first step in troubleshooting a build failure is to recheck
-that you have satisfied all the pre-build requirements for your platform.
-Scanning the <code>configure</code> log is a good first step, making sure that what it
-found makes sense for your system. Look for strange error messages or any
-difficulties that <code>configure</code> had in finding things.</p>
-
-<p>Some of the more common problems with builds are briefly described below, with
-suggestions for remedies.</p>
-
-<ul>
-<li><p><strong>Corrupted Bundles on Windows:</strong> <br />
-Some virus scanning software has been known to corrupt the downloading of
-zip bundles. It may be necessary to disable the 'on access' or 'real time'
-virus scanning features to prevent this corruption. This type of 'real time'
-virus scanning can also slow down the build process significantly.
-Temporarily disabling the feature, or excluding the build output directory
-may be necessary to get correct and faster builds.</p></li>
-<li><p><strong>Slow Builds:</strong> <br />
-If your build machine seems to be overloaded from too many simultaneous C++
-compiles, try setting the <code>JOBS=1</code> on the <code>make</code> command line. Then try
-increasing the count slowly to an acceptable level for your system. Also:</p>
-
-<p>Creating the javadocs can be very slow, if you are running javadoc, consider
-skipping that step.</p>
-
-<p>Faster CPUs, more RAM, and a faster DISK usually helps. The VM build tends
-to be CPU intensive (many C++ compiles), and the rest of the JDK will often
-be disk intensive.</p>
-
-<p>Faster compiles are possible using a tool called
-<a href="http://ccache.samba.org/">ccache</a>.</p></li>
-<li><p><strong>File time issues:</strong> <br />
-If you see warnings that refer to file time stamps, e.g.</p>
-
-<blockquote>
-  <p><em>Warning message:</em> <code>File 'xxx' has modification time in the future.</code> <br />
-<em>Warning message:</em> <code>Clock skew detected. Your build may be incomplete.</code></p>
-</blockquote>
-
-<p>These warnings can occur when the clock on the build machine is out of sync
-with the timestamps on the source files. Other errors, apparently unrelated
-but in fact caused by the clock skew, can occur along with the clock skew
-warnings. These secondary errors may tend to obscure the fact that the true
-root cause of the problem is an out-of-sync clock.</p>
-
-<p>If you see these warnings, reset the clock on the build machine, run
-"<code>gmake clobber</code>" or delete the directory containing the build output, and
-restart the build from the beginning.</p></li>
-<li><p><strong>Error message: <code>Trouble writing out table to disk</code></strong> <br />
-Increase the amount of swap space on your build machine. This could be
-caused by overloading the system and it may be necessary to use:</p>
-
-<blockquote>
-  <p><code>make JOBS=1</code></p>
-</blockquote>
-
-<p>to reduce the load on the system.</p></li>
-<li><p><strong>Error Message: <code>libstdc++ not found</code>:</strong> <br />
-This is caused by a missing libstdc++.a library. This is installed as part
-of a specific package (e.g. libstdc++.so.devel.386). By default some 64-bit
-Linux versions (e.g. Fedora) only install the 64-bit version of the
-libstdc++ package. Various parts of the JDK build require a static link of
-the C++ runtime libraries to allow for maximum portability of the built
-images.</p></li>
-<li><p><strong>Linux Error Message: <code>cannot restore segment prot after reloc</code></strong> <br />
-This is probably an issue with SELinux (See <a href="http://en.wikipedia.org/wiki/SELinux">SELinux on
-Wikipedia</a>). Parts of the VM is built
-without the <code>-fPIC</code> for performance reasons.</p>
-
-<p>To completely disable SELinux:</p>
-
-<ol>
-<li><code>$ su root</code></li>
-<li><code># system-config-securitylevel</code></li>
-<li><code>In the window that appears, select the SELinux tab</code></li>
-<li><code>Disable SELinux</code></li>
-</ol>
-
-<p>Alternatively, instead of completely disabling it you could disable just
-this one check.</p>
-
-<ol>
-<li>Select System->Administration->SELinux Management</li>
-<li>In the SELinux Management Tool which appears, select "Boolean" from the
-menu on the left</li>
-<li>Expand the "Memory Protection" group</li>
-<li>Check the first item, labeled "Allow all unconfined executables to use
-libraries requiring text relocation ..."</li>
-</ol></li>
-<li><p><strong>Windows Error Messages:</strong> <br />
-<code>*** fatal error - couldn't allocate heap, ...</code> <br />
-<code>rm fails with "Directory not empty"</code> <br />
-<code>unzip fails with "cannot create ... Permission denied"</code> <br />
-<code>unzip fails with "cannot create ... Error 50"</code></p>
-
-<p>The CYGWIN software can conflict with other non-CYGWIN software. See the
-CYGWIN FAQ section on <a href="http://cygwin.com/faq/faq.using.html#faq.using.bloda">BLODA (applications that interfere with
-CYGWIN)</a>.</p></li>
-<li><p><strong>Windows Error Message: <code>spawn failed</code></strong> <br />
-Try rebooting the system, or there could be some kind of issue with the disk
-or disk partition being used. Sometimes it comes with a "Permission Denied"
-message.</p></li>
-</ul>
-
-<hr />
-
-<p><a name="gmake"></a></p>
-
-<h2>Appendix B: GNU make</h2>
-
-<p>The Makefiles in the OpenJDK are only valid when used with the GNU version of
-the utility command <code>make</code> (usually called <code>gmake</code> on Solaris). A few notes
-about using GNU make:</p>
-
-<ul>
-<li>You need GNU make version 3.81 or newer. On Windows 4.0 or newer is
-recommended. If the GNU make utility on your systems is not of a suitable
-version, see "<a href="#buildgmake">Building GNU make</a>".</li>
-<li>Place the location of the GNU make binary in the <code>PATH</code>.</li>
-<li><strong>Solaris:</strong> Do NOT use <code>/usr/bin/make</code> on Solaris. If your Solaris system
-has the software from the Solaris Developer Companion CD installed, you
-should try and use <code>/usr/bin/gmake</code> or <code>/usr/gnu/bin/make</code>.</li>
-<li><strong>Windows:</strong> Make sure you start your build inside a bash shell.</li>
-<li><strong>Mac OS X:</strong> The XCode "command line tools" must be installed on your Mac.</li>
-</ul>
-
-<p>Information on GNU make, and access to ftp download sites, are available on the
-<a href="http://www.gnu.org/software/make/make.html">GNU make web site </a>. The latest
-source to GNU make is available at
-<a href="http://ftp.gnu.org/pub/gnu/make/">ftp.gnu.org/pub/gnu/make/</a>.</p>
-
-<p><a name="buildgmake"></a></p>
-
-<h3>Building GNU make</h3>
-
-<p>First step is to get the GNU make 3.81 or newer source from
-<a href="http://ftp.gnu.org/pub/gnu/make/">ftp.gnu.org/pub/gnu/make/</a>. Building is a
-little different depending on the OS but is basically done with:</p>
-
-<pre><code>  bash ./configure
-  make
-</code></pre>
-
-<hr />
-
-<p><a name="buildenvironments"></a></p>
-
-<h2>Appendix C: Build Environments</h2>
-
-<h3>Minimum Build Environments</h3>
-
-<p>This file often describes specific requirements for what we call the "minimum
-build environments" (MBE) for this specific release of the JDK. What is listed
-below is what the Oracle Release Engineering Team will use to build the Oracle
-JDK product. Building with the MBE will hopefully generate the most compatible
-bits that install on, and run correctly on, the most variations of the same
-base OS and hardware architecture. In some cases, these represent what is often
-called the least common denominator, but each Operating System has different
-aspects to it.</p>
-
-<p>In all cases, the Bootstrap JDK version minimum is critical, we cannot
-guarantee builds will work with older Bootstrap JDK's. Also in all cases, more
-RAM and more processors is better, the minimums listed below are simply
-recommendations.</p>
-
-<p>With Solaris and Mac OS X, the version listed below is the oldest release we
-can guarantee builds and works, and the specific version of the compilers used
-could be critical.</p>
-
-<p>With Windows the critical aspect is the Visual Studio compiler used, which due
-to it's runtime, generally dictates what Windows systems can do the builds and
-where the resulting bits can be used.</p>
-
-<p><strong>NOTE: We expect a change here off these older Windows OS releases and to a
-'less older' one, probably Windows 2008R2 X64.</strong></p>
-
-<p>With Linux, it was just a matter of picking a stable distribution that is a
-good representative for Linux in general.</p>
-
-<p>It is understood that most developers will NOT be using these specific
-versions, and in fact creating these specific versions may be difficult due to
-the age of some of this software. It is expected that developers are more often
-using the more recent releases and distributions of these operating systems.</p>
-
-<p>Compilation problems with newer or different C/C++ compilers is a common
-problem. Similarly, compilation problems related to changes to the
-<code>/usr/include</code> or system header files is also a common problem with older,
-newer, or unreleased OS versions. Please report these types of problems as bugs
-so that they can be dealt with accordingly.</p>
-
-<blockquote>
-  <p><table border="1">
-     <thead>
-       <tr>
-         <th>Base OS and Architecture</th>
-         <th>OS</th>
-         <th>C/C++ Compiler</th>
-         <th>Bootstrap JDK</th>
-         <th>Processors</th>
-         <th>RAM Minimum</th>
-         <th>DISK Needs</th>
-       </tr>
-     </thead>
-     <tbody>
-       <tr>
-         <td>Linux X86 (32-bit) and X64 (64-bit)</td>
-         <td>Oracle Enterprise Linux 6.4</td>
-         <td>gcc 4.9.2 </td>
-         <td>JDK 8</td>
-         <td>2 or more</td>
-         <td>1 GB</td>
-         <td>6 GB</td>
-       </tr>
-       <tr>
-         <td>Solaris SPARCV9 (64-bit)</td>
-         <td>Solaris 11 Update 1</td>
-         <td>Studio 12 Update 4 + patches</td>
-         <td>JDK 8</td>
-         <td>4 or more</td>
-         <td>4 GB</td>
-         <td>8 GB</td>
-       </tr>
-       <tr>
-         <td>Solaris X64 (64-bit)</td>
-         <td>Solaris 11 Update 1</td>
-         <td>Studio 12 Update 4 + patches</td>
-         <td>JDK 8</td>
-         <td>4 or more</td>
-         <td>4 GB</td>
-         <td>8 GB</td>
-       </tr>
-       <tr>
-         <td>Windows X86 (32-bit)</td>
-         <td>Windows Server 2012 R2 x64</td>
-         <td>Microsoft Visual Studio C++ 2013 Professional Edition</td>
-         <td>JDK 8</td>
-         <td>2 or more</td>
-         <td>2 GB</td>
-         <td>6 GB</td>
-       </tr>
-       <tr>
-         <td>Windows X64 (64-bit)</td>
-         <td>Windows Server 2012 R2 x64</td>
-         <td>Microsoft Visual Studio C++ 2013 Professional Edition</td>
-         <td>JDK 8</td>
-         <td>2 or more</td>
-         <td>2 GB</td>
-         <td>6 GB</td>
-       </tr>
-       <tr>
-         <td>Mac OS X X64 (64-bit)</td>
-         <td>Mac OS X 10.9 "Mavericks"</td>
-         <td>Xcode 6.3 or newer</td>
-         <td>JDK 8</td>
-         <td>2 or more</td>
-         <td>4 GB</td>
-         <td>6 GB</td>
-       </tr>
-     </tbody>
-   </table></p>
-</blockquote>
-
-<hr />
-
-<p><a name="SDBE"></a></p>
-
-<h3>Specific Developer Build Environments</h3>
-
-<p>We won't be listing all the possible environments, but we will try to provide
-what information we have available to us.</p>
-
-<p><strong>NOTE: The community can help out by updating this part of the document.</strong></p>
-
-<h4>Fedora</h4>
-
-<p>After installing the latest <a href="http://fedoraproject.org">Fedora</a> you need to
-install several build dependencies. The simplest way to do it is to execute the
-following commands as user <code>root</code>:</p>
-
-<pre><code>  yum-builddep java-1.7.0-openjdk
-  yum install gcc gcc-c++
-</code></pre>
-
-<p>In addition, it's necessary to set a few environment variables for the build:</p>
-
-<pre><code>  export LANG=C
-  export PATH="/usr/lib/jvm/java-openjdk/bin:${PATH}"
-</code></pre>
-
-<h4>CentOS 5.5</h4>
-
-<p>After installing <a href="http://www.centos.org/">CentOS 5.5</a> you need to make sure you
-have the following Development bundles installed:</p>
-
-<ul>
-<li>Development Libraries</li>
-<li>Development Tools</li>
-<li>Java Development</li>
-<li>X Software Development (Including XFree86-devel)</li>
-</ul>
-
-<p>Plus the following packages:</p>
-
-<ul>
-<li>cups devel: Cups Development Package</li>
-<li>alsa devel: Alsa Development Package</li>
-<li>Xi devel: libXi.so Development Package</li>
-</ul>
-
-<p>The freetype 2.3 packages don't seem to be available, but the freetype 2.3
-sources can be downloaded, built, and installed easily enough from <a href="http://downloads.sourceforge.net/freetype">the
-freetype site</a>. Build and install
-with something like:</p>
-
-<pre><code>  bash ./configure
-  make
-  sudo -u root make install
-</code></pre>
-
-<p>Mercurial packages could not be found easily, but a Google search should find
-ones, and they usually include Python if it's needed.</p>
-
-<h4>Debian 5.0 (Lenny)</h4>
-
-<p>After installing <a href="http://debian.org">Debian</a> 5 you need to install several
-build dependencies. The simplest way to install the build dependencies is to
-execute the following commands as user <code>root</code>:</p>
-
-<pre><code>  aptitude build-dep openjdk-7
-  aptitude install openjdk-7-jdk libmotif-dev
-</code></pre>
-
-<p>In addition, it's necessary to set a few environment variables for the build:</p>
-
-<pre><code>  export LANG=C
-  export PATH="/usr/lib/jvm/java-7-openjdk/bin:${PATH}"
-</code></pre>
-
-<h4>Ubuntu 12.04</h4>
-
-<p>After installing <a href="http://ubuntu.org">Ubuntu</a> 12.04 you need to install several
-build dependencies. The simplest way to do it is to execute the following
-commands:</p>
-
-<pre><code>  sudo aptitude build-dep openjdk-7
-  sudo aptitude install openjdk-7-jdk
-</code></pre>
-
-<p>In addition, it's necessary to set a few environment variables for the build:</p>
-
-<pre><code>  export LANG=C
-  export PATH="/usr/lib/jvm/java-7-openjdk/bin:${PATH}"
-</code></pre>
-
-<h4>OpenSUSE 11.1</h4>
-
-<p>After installing <a href="http://opensuse.org">OpenSUSE</a> 11.1 you need to install
-several build dependencies. The simplest way to install the build dependencies
-is to execute the following commands:</p>
-
-<pre><code>  sudo zypper source-install -d java-1_7_0-openjdk
-  sudo zypper install make
-</code></pre>
-
-<p>In addition, it is necessary to set a few environment variables for the build:</p>
-
-<pre><code>  export LANG=C
-  export PATH="/usr/lib/jvm/java-1.7.0-openjdk/bin:$[PATH}"
-</code></pre>
-
-<p>Finally, you need to unset the <code>JAVA_HOME</code> environment variable:</p>
-
-<pre><code>  export -n JAVA_HOME`
-</code></pre>
-
-<h4>Mandriva Linux One 2009 Spring</h4>
-
-<p>After installing <a href="http://mandriva.org">Mandriva</a> Linux One 2009 Spring you need
-to install several build dependencies. The simplest way to install the build
-dependencies is to execute the following commands as user <code>root</code>:</p>
-
-<pre><code>  urpmi java-1.7.0-openjdk-devel make gcc gcc-c++ freetype-devel zip unzip
-    libcups2-devel libxrender1-devel libalsa2-devel libstc++-static-devel
-    libxtst6-devel libxi-devel
-</code></pre>
-
-<p>In addition, it is necessary to set a few environment variables for the build:</p>
-
-<pre><code>  export LANG=C
-  export PATH="/usr/lib/jvm/java-1.7.0-openjdk/bin:${PATH}"
-</code></pre>
-
-<h4>OpenSolaris 2009.06</h4>
-
-<p>After installing <a href="http://opensolaris.org">OpenSolaris</a> 2009.06 you need to
-install several build dependencies. The simplest way to install the build
-dependencies is to execute the following commands:</p>
-
-<pre><code>  pfexec pkg install SUNWgmake SUNWj7dev sunstudioexpress SUNWcups SUNWzip
-    SUNWunzip SUNWxwhl SUNWxorg-headers SUNWaudh SUNWfreetype2
-</code></pre>
-
-<p>In addition, it is necessary to set a few environment variables for the build:</p>
-
-<pre><code>  export LANG=C
-  export PATH="/opt/SunStudioExpress/bin:${PATH}"
-</code></pre>
-
-<hr />
-
-<p>End of the OpenJDK build README document.</p>
-
-<p>Please come again!</p>
-  </body>
-</html>
--- a/README-builds.md	Tue Feb 14 14:57:56 2017 -0800
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,1285 +0,0 @@
-![OpenJDK](http://openjdk.java.net/images/openjdk.png)
-# OpenJDK Build README
-
-*****
-
-<a name="introduction"></a>
-## Introduction
-
-This README file contains build instructions for the
-[OpenJDK](http://openjdk.java.net). Building the source code for the OpenJDK
-requires a certain degree of technical expertise.
-
-### !!!!!!!!!!!!!!! THIS IS A MAJOR RE-WRITE of this document. !!!!!!!!!!!!!
-
-Some Headlines:
-
- * The build is now a "`configure && make`" style build
- * Any GNU make 3.81 or newer should work, except on Windows where 4.0 or newer
-   is recommended.
- * The build should scale, i.e. more processors should cause the build to be
-   done in less wall-clock time
- * Nested or recursive make invocations have been significantly reduced,
-   as has the total fork/exec or spawning of sub processes during the build
- * Windows MKS usage is no longer supported
- * Windows Visual Studio `vsvars*.bat` and `vcvars*.bat` files are run
-   automatically
- * Ant is no longer used when building the OpenJDK
- * Use of ALT_* environment variables for configuring the build is no longer
-   supported
-
-*****
-
-## Contents
-
-  * [Introduction](#introduction)
-  * [Use of Mercurial](#hg)
-    * [Getting the Source](#get_source)
-    * [Repositories](#repositories)
-  * [Building](#building)
-    * [System Setup](#setup)
-      * [Linux](#linux)
-      * [Solaris](#solaris)
-      * [Mac OS X](#macosx)
-      * [Windows](#windows)
-    * [Configure](#configure)
-    * [Make](#make)
-  * [Testing](#testing)
-
-*****
-
-  * [Appendix A: Hints and Tips](#hints)
-    * [FAQ](#faq)
-    * [Build Performance Tips](#performance)
-    * [Troubleshooting](#troubleshooting)
-  * [Appendix B: GNU Make Information](#gmake)
-  * [Appendix C: Build Environments](#buildenvironments)
-
-*****
-
-<a name="hg"></a>
-## Use of Mercurial
-
-The OpenJDK sources are maintained with the revision control system
-[Mercurial](http://mercurial.selenic.com/wiki/Mercurial). If you are new to
-Mercurial, please see the [Beginner Guides](http://mercurial.selenic.com/wiki/
-BeginnersGuides) or refer to the [Mercurial Book](http://hgbook.red-bean.com/).
-The first few chapters of the book provide an excellent overview of Mercurial,
-what it is and how it works.
-
-For using Mercurial with the OpenJDK refer to the [Developer Guide: Installing
-and Configuring Mercurial](http://openjdk.java.net/guide/
-repositories.html#installConfig) section for more information.
-
-<a name="get_source"></a>
-### Getting the Source
-
-To get the entire set of OpenJDK Mercurial repositories use the script
-`get_source.sh` located in the root repository:
-
-      hg clone http://hg.openjdk.java.net/jdk9/jdk9 YourOpenJDK
-      cd YourOpenJDK
-      bash ./get_source.sh
-
-Once you have all the repositories, keep in mind that each repository is its
-own independent repository. You can also re-run `./get_source.sh` anytime to
-pull over all the latest changesets in all the repositories. This set of
-nested repositories has been given the term "forest" and there are various
-ways to apply the same `hg` command to each of the repositories. For
-example, the script `make/scripts/hgforest.sh` can be used to repeat the
-same `hg` command on every repository, e.g.
-
-      cd YourOpenJDK
-      bash ./make/scripts/hgforest.sh status
-
-<a name="repositories"></a>
-### Repositories
-
-The set of repositories and what they contain:
-
- * **. (root)** contains common configure and makefile logic
- * **hotspot** contains source code and make files for building the OpenJDK
-   Hotspot Virtual Machine
- * **langtools** contains source code for the OpenJDK javac and language tools
- * **jdk** contains source code and make files for building the OpenJDK runtime
-   libraries and misc files
- * **jaxp** contains source code for the OpenJDK JAXP functionality
- * **jaxws** contains source code for the OpenJDK JAX-WS functionality
- * **corba** contains source code for the OpenJDK Corba functionality
- * **nashorn** contains source code for the OpenJDK JavaScript implementation
-
-### Repository Source Guidelines
-
-There are some very basic guidelines:
-
- * Use of whitespace in source files (.java, .c, .h, .cpp, and .hpp files) is
-   restricted. No TABs, no trailing whitespace on lines, and files should not
-   terminate in more than one blank line.
- * Files with execute permissions should not be added to the source
-   repositories.
- * All generated files need to be kept isolated from the files maintained or
-   managed by the source control system. The standard area for generated files
-   is the top level `build/` directory.
- * The default build process should be to build the product and nothing else,
-   in one form, e.g. a product (optimized), debug (non-optimized, -g plus
-   assert logic), or fastdebug (optimized, -g plus assert logic).
- * The `.hgignore` file in each repository must exist and should include
-   `^build/`, `^dist/` and optionally any `nbproject/private` directories. **It
-   should NEVER** include anything in the `src/` or `test/` or any managed
-   directory area of a repository.
- * Directory names and file names should never contain blanks or non-printing
-   characters.
- * Generated source or binary files should NEVER be added to the repository
-   (that includes `javah` output). There are some exceptions to this rule, in
-   particular with some of the generated configure scripts.
- * Files not needed for typical building or testing of the repository should
-   not be added to the repository.
-
-*****
-
-<a name="building"></a>
-## Building
-
-The very first step in building the OpenJDK is making sure the system itself
-has everything it needs to do OpenJDK builds. Once a system is setup, it
-generally doesn't need to be done again.
-
-Building the OpenJDK is now done with running a `configure` script which will
-try and find and verify you have everything you need, followed by running
-`make`, e.g.
-
->  **`bash ./configure`**  
->  **`make all`**
-
-Where possible the `configure` script will attempt to located the various
-components in the default locations or via component specific variable
-settings. When the normal defaults fail or components cannot be found,
-additional `configure` options may be necessary to help `configure` find the
-necessary tools for the build, or you may need to re-visit the setup of your
-system due to missing software packages.
-
-**NOTE:** The `configure` script file does not have execute permissions and
-will need to be explicitly run with `bash`, see the source guidelines.
-
-*****
-
-<a name="setup"></a>
-### System Setup
-
-Before even attempting to use a system to build the OpenJDK there are some very
-basic system setups needed. For all systems:
-
- * Be sure the GNU make utility is version 3.81 (4.0 on windows) or newer, e.g.
-   run "`make -version`"
-
-   <a name="bootjdk"></a>
- * Install a Bootstrap JDK. All OpenJDK builds require access to a previously
-   released JDK called the _bootstrap JDK_ or _boot JDK._ The general rule is
-   that the bootstrap JDK must be an instance of the previous major release of
-   the JDK. In addition, there may be a requirement to use a release at or
-   beyond a particular update level.
-
-   **_Building JDK 9 requires JDK 8. JDK 9 developers should not use JDK 9 as
-   the boot JDK, to ensure that JDK 9 dependencies are not introduced into the
-   parts of the system that are built with JDK 8._**
-
-   The JDK 8 binaries can be downloaded from Oracle's [JDK 8 download
-   site](http://www.oracle.com/technetwork/java/javase/downloads/index.html).
-   For build performance reasons it is very important that this bootstrap JDK
-   be made available on the local disk of the machine doing the build. You
-   should add its `bin` directory to the `PATH` environment variable. If
-   `configure` has any issues finding this JDK, you may need to use the
-   `configure` option `--with-boot-jdk`.
-
- * Ensure that GNU make, the Bootstrap JDK, and the compilers are all in your
-   PATH environment variable.
-
-And for specific systems:
-
- * **Linux**
-
-   Install all the software development packages needed including
-   [alsa](#alsa), [freetype](#freetype), [cups](#cups), and
-   [xrender](#xrender). See [specific system packages](#SDBE).
-
- * **Solaris**
-
-   Install all the software development packages needed including [Studio
-   Compilers](#studio), [freetype](#freetype), [cups](#cups), and
-   [xrender](#xrender). See [specific system packages](#SDBE).
-
- * **Windows**
-
-   * Install one of [CYGWIN](#cygwin) or [MinGW/MSYS](#msys)
-   * Install [Visual Studio 2013](#vs2013)
-
- * **Mac OS X**
-
-   Install [XCode 6.3](https://developer.apple.com/xcode/)
-
-<a name="linux"></a>
-#### Linux
-
-With Linux, try and favor the system packages over building your own or getting
-packages from other areas. Most Linux builds should be possible with the
-system's available packages.
-
-Note that some Linux systems have a habit of pre-populating your environment
-variables for you, for example `JAVA_HOME` might get pre-defined for you to
-refer to the JDK installed on your Linux system. You will need to unset
-`JAVA_HOME`. It's a good idea to run `env` and verify the environment variables
-you are getting from the default system settings make sense for building the
-OpenJDK.
-
-<a name="solaris"></a>
-#### Solaris
-
-<a name="studio"></a>
-##### Studio Compilers
-
-At a minimum, the [Studio 12 Update 4 Compilers](http://www.oracle.com/
-technetwork/server-storage/solarisstudio/downloads/index.htm) (containing
-version 5.13 of the C and C++ compilers) is required, including specific
-patches.
-
-The Solaris Studio installation should contain at least these packages:
-
->  <table border="1">
-     <thead>
-       <tr>
-         <td>**Package**</td>
-         <td>**Version**</td>
-       </tr>
-     </thead>
-     <tbody>
-       <tr>
-         <td>developer/solarisstudio-124/backend</td>
-         <td>12.4-1.0.6.0</td>
-       </tr>
-       <tr>
-         <td>developer/solarisstudio-124/c++</td>
-         <td>12.4-1.0.10.0</td>
-       </tr>
-       <tr>
-         <td>developer/solarisstudio-124/cc</td>
-         <td>12.4-1.0.4.0</td>
-       </tr>
-       <tr>
-         <td>developer/solarisstudio-124/library/c++-libs</td>
-         <td>12.4-1.0.10.0</td>
-       </tr>
-       <tr>
-         <td>developer/solarisstudio-124/library/math-libs</td>
-         <td>12.4-1.0.0.1</td>
-       </tr>
-       <tr>
-         <td>developer/solarisstudio-124/library/studio-gccrt</td>
-         <td>12.4-1.0.0.1</td>
-       </tr>
-       <tr>
-         <td>developer/solarisstudio-124/studio-common</td>
-         <td>12.4-1.0.0.1</td>
-       </tr>
-       <tr>
-         <td>developer/solarisstudio-124/studio-ja</td>
-         <td>12.4-1.0.0.1</td>
-       </tr>
-       <tr>
-         <td>developer/solarisstudio-124/studio-legal</td>
-         <td>12.4-1.0.0.1</td>
-       </tr>
-       <tr>
-         <td>developer/solarisstudio-124/studio-zhCN</td>
-         <td>12.4-1.0.0.1</td>
-       </tr>
-     </tbody>
-   </table>
-
-In particular backend 12.4-1.0.6.0 contains a critical patch for the sparc
-version.
-
-Place the `bin` directory in `PATH`.
-
-The Oracle Solaris Studio Express compilers at: [Oracle Solaris Studio Express
-Download site](http://www.oracle.com/technetwork/server-storage/solarisstudio/
-downloads/index-jsp-142582.html) are also an option, although these compilers
-have not been extensively used yet.
-
-<a name="windows"></a>
-#### Windows
-
-##### Windows Unix Toolkit
-
-Building on Windows requires a Unix-like environment, notably a Unix-like
-shell. There are several such environments available of which
-[Cygwin](http://www.cygwin.com/) and
-[MinGW/MSYS](http://www.mingw.org/wiki/MSYS) are currently supported for the
-OpenJDK build. One of the differences of these systems from standard Windows
-tools is the way they handle Windows path names, particularly path names which
-contain spaces, backslashes as path separators and possibly drive letters.
-Depending on the use case and the specifics of each environment these path
-problems can be solved by a combination of quoting whole paths, translating
-backslashes to forward slashes, escaping backslashes with additional
-backslashes and translating the path names to their ["8.3"
-version](http://en.wikipedia.org/wiki/8.3_filename).
-
-<a name="cygwin"></a>
-###### CYGWIN
-
-CYGWIN is an open source, Linux-like environment which tries to emulate a
-complete POSIX layer on Windows. It tries to be smart about path names and can
-usually handle all kinds of paths if they are correctly quoted or escaped
-although internally it maps drive letters `<drive>:` to a virtual directory
-`/cygdrive/<drive>`.
-
-You can always use the `cygpath` utility to map pathnames with spaces or the
-backslash character into the `C:/` style of pathname (called 'mixed'), e.g.
-`cygpath -s -m "<path>"`.
-
-Note that the use of CYGWIN creates a unique problem with regards to setting
-[`PATH`](#path). Normally on Windows the `PATH` variable contains directories
-separated with the ";" character (Solaris and Linux use ":"). With CYGWIN, it
-uses ":", but that means that paths like "C:/path" cannot be placed in the
-CYGWIN version of `PATH` and instead CYGWIN uses something like
-`/cygdrive/c/path` which CYGWIN understands, but only CYGWIN understands.
-
-The OpenJDK build requires CYGWIN version 1.7.16 or newer. Information about
-CYGWIN can be obtained from the CYGWIN website at
-[www.cygwin.com](http://www.cygwin.com).
-
-By default CYGWIN doesn't install all the tools required for building the
-OpenJDK. Along with the default installation, you need to install the following
-tools.
-
->  <table border="1">
-     <thead>
-       <tr>
-         <td>Binary Name</td>
-         <td>Category</td>
-         <td>Package</td>
-         <td>Description</td>
-      </tr>
-     </thead>
-     <tbody>
-       <tr>
-         <td>ar.exe</td>
-         <td>Devel</td>
-         <td>binutils</td>
-         <td>The GNU assembler, linker and binary utilities</td>
-       </tr>
-       <tr>
-         <td>make.exe</td>
-         <td>Devel</td>
-         <td>make</td>
-         <td>The GNU version of the 'make' utility built for CYGWIN</td>
-       </tr>
-       <tr>
-         <td>m4.exe</td>
-         <td>Interpreters</td>
-         <td>m4</td>
-         <td>GNU implementation of the traditional Unix macro processor</td>
-       </tr>
-       <tr>
-         <td>cpio.exe</td>
-         <td>Utils</td>
-         <td>cpio</td>
-         <td>A program to manage archives of files</td>
-       </tr>
-       <tr>
-         <td>gawk.exe</td>
-         <td>Utils</td>
-         <td>awk</td>
-         <td>Pattern-directed scanning and processing language</td>
-       </tr>
-       <tr>
-         <td>file.exe</td>
-         <td>Utils</td>
-         <td>file</td>
-         <td>Determines file type using 'magic' numbers</td>
-       </tr>
-       <tr>
-         <td>zip.exe</td>
-         <td>Archive</td>
-         <td>zip</td>
-         <td>Package and compress (archive) files</td>
-       </tr>
-       <tr>
-         <td>unzip.exe</td>
-         <td>Archive</td>
-         <td>unzip</td>
-         <td>Extract compressed files in a ZIP archive</td>
-       </tr>
-       <tr>
-         <td>free.exe</td>
-         <td>System</td>
-         <td>procps</td>
-         <td>Display amount of free and used memory in the system</td>
-       </tr>
-     </tbody>
-   </table>
-
-Note that the CYGWIN software can conflict with other non-CYGWIN software on
-your Windows system. CYGWIN provides a [FAQ](http://cygwin.com/faq/
-faq.using.html) for known issues and problems, of particular interest is the
-section on [BLODA (applications that interfere with
-CYGWIN)](http://cygwin.com/faq/faq.using.html#faq.using.bloda).
-
-<a name="msys"></a>
-###### MinGW/MSYS
-
-MinGW ("Minimalist GNU for Windows") is a collection of free Windows specific
-header files and import libraries combined with GNU toolsets that allow one to
-produce native Windows programs that do not rely on any 3rd-party C runtime
-DLLs. MSYS is a supplement to MinGW which allows building applications and
-programs which rely on traditional UNIX tools to be present. Among others this
-includes tools like `bash` and `make`. See [MinGW/MSYS](http://www.mingw.org/
-wiki/MSYS) for more information.
-
-Like Cygwin, MinGW/MSYS can handle different types of path formats. They are
-internally converted to paths with forward slashes and drive letters
-`<drive>:` replaced by a virtual directory `/<drive>`. Additionally, MSYS
-automatically detects binaries compiled for the MSYS environment and feeds them
-with the internal, Unix-style path names. If native Windows applications are
-called from within MSYS programs their path arguments are automatically
-converted back to Windows style path names with drive letters and backslashes
-as path separators. This may cause problems for Windows applications which use
-forward slashes as parameter separator (e.g. `cl /nologo /I`) because MSYS may
-wrongly [replace such parameters by drive letters](http://mingw.org/wiki/
-Posix_path_conversion).
-
-In addition to the tools which will be installed by default, you have to
-manually install the `msys-zip` and `msys-unzip` packages. This can be easily
-done with the MinGW command line installer:
-
-      mingw-get.exe install msys-zip
-      mingw-get.exe install msys-unzip
-
-<a name="vs2013"></a>
-##### Visual Studio 2013 Compilers
-
-The 32-bit and 64-bit OpenJDK Windows build requires Microsoft Visual Studio
-C++ 2013 (VS2013) Professional Edition or Express compiler. The compiler and
-other tools are expected to reside in the location defined by the variable
-`VS120COMNTOOLS` which is set by the Microsoft Visual Studio installer.
-
-Only the C++ part of VS2013 is needed. Try to let the installation go to the
-default install directory. Always reboot your system after installing VS2013.
-The system environment variable VS120COMNTOOLS should be set in your
-environment.
-
-Make sure that TMP and TEMP are also set in the environment and refer to
-Windows paths that exist, like `C:\temp`, not `/tmp`, not `/cygdrive/c/temp`,
-and not `C:/temp`. `C:\temp` is just an example, it is assumed that this area
-is private to the user, so by default after installs you should see a unique
-user path in these variables.
-
-<a name="macosx"></a>
-#### Mac OS X
-
-Make sure you get the right XCode version.
-
-*****
-
-<a name="configure"></a>
-### Configure
-
-The basic invocation of the `configure` script looks like:
-
->  **`bash ./configure [options]`**
-
-This will create an output directory containing the "configuration" and setup
-an area for the build result. This directory typically looks like:
-
->  **`build/linux-x64-normal-server-release`**
-
-`configure` will try to figure out what system you are running on and where all
-necessary build components are. If you have all prerequisites for building
-installed, it should find everything. If it fails to detect any component
-automatically, it will exit and inform you about the problem. When this
-happens, read more below in [the `configure` options](#configureoptions).
-
-Some examples:
-
->  **Windows 32bit build with freetype specified:**  
->  `bash ./configure --with-freetype=/cygdrive/c/freetype-i586 --with-target-
-bits=32`
-
->  **Debug 64bit Build:**  
->  `bash ./configure --enable-debug --with-target-bits=64`
-
-<a name="configureoptions"></a>
-#### Configure Options
-
-Complete details on all the OpenJDK `configure` options can be seen with:
-
->  **`bash ./configure --help=short`**
-
-Use `-help` to see all the `configure` options available. You can generate any
-number of different configurations, e.g. debug, release, 32, 64, etc.
-
-Some of the more commonly used `configure` options are:
-
->  **`--enable-debug`**  
->  set the debug level to fastdebug (this is a shorthand for `--with-debug-
-   level=fastdebug`)
-
-<a name="alsa"></a>
->  **`--with-alsa=`**_path_  
->  select the location of the Advanced Linux Sound Architecture (ALSA)
-
->  Version 0.9.1 or newer of the ALSA files are required for building the
-   OpenJDK on Linux. These Linux files are usually available from an "alsa" of
-   "libasound" development package, and it's highly recommended that you try
-   and use the package provided by the particular version of Linux that you are
-   using.
-
->  **`--with-boot-jdk=`**_path_  
->  select the [Bootstrap JDK](#bootjdk)
-
->  **`--with-boot-jdk-jvmargs=`**"_args_"  
->  provide the JVM options to be used to run the [Bootstrap JDK](#bootjdk)
-
->  **`--with-cacerts=`**_path_  
->  select the path to the cacerts file.
-
->  See [Certificate Authority on Wikipedia](http://en.wikipedia.org/wiki/
-   Certificate_Authority) for a better understanding of the Certificate
-   Authority (CA). A certificates file named "cacerts" represents a system-wide
-   keystore with CA certificates. In JDK and JRE binary bundles, the "cacerts"
-   file contains root CA certificates from several public CAs (e.g., VeriSign,
-   Thawte, and Baltimore). The source contain a cacerts file without CA root
-   certificates. Formal JDK builders will need to secure permission from each
-   public CA and include the certificates into their own custom cacerts file.
-   Failure to provide a populated cacerts file will result in verification
-   errors of a certificate chain during runtime. By default an empty cacerts
-   file is provided and that should be fine for most JDK developers.
-
-<a name="cups"></a>
->  **`--with-cups=`**_path_  
->  select the CUPS install location
-
->  The Common UNIX Printing System (CUPS) Headers are required for building the
-   OpenJDK on Solaris and Linux. The Solaris header files can be obtained by
-   installing the package **print/cups**.
-
->  The CUPS header files can always be downloaded from
-   [www.cups.org](http://www.cups.org).
-
->  **`--with-cups-include=`**_path_  
->  select the CUPS include directory location
-
->  **`--with-debug-level=`**_level_  
->  select the debug information level of release, fastdebug, or slowdebug
-
->  **`--with-dev-kit=`**_path_  
->  select location of the compiler install or developer install location
-
-<a name="freetype"></a>
->  **`--with-freetype=`**_path_  
->  select the freetype files to use.
-
->  Expecting the freetype libraries under `lib/` and the headers under
-   `include/`.
-
->  Version 2.3 or newer of FreeType is required. On Unix systems required files
-   can be available as part of your distribution (while you still may need to
-   upgrade them). Note that you need development version of package that
-   includes both the FreeType library and header files.
-
->  You can always download latest FreeType version from the [FreeType
-   website](http://www.freetype.org). Building the freetype 2 libraries from
-   scratch is also possible, however on Windows refer to the [Windows FreeType
-   DLL build instructions](http://freetype.freedesktop.org/wiki/FreeType_DLL).
-
->  Note that by default FreeType is built with byte code hinting support
-   disabled due to licensing restrictions. In this case, text appearance and
-   metrics are expected to differ from Sun's official JDK build. See the
-   [SourceForge FreeType2 Home Page](http://freetype.sourceforge.net/freetype2)
-   for more information.
-
->  **`--with-import-hotspot=`**_path_  
->  select the location to find hotspot binaries from a previous build to avoid
-   building hotspot
-
->  **`--with-target-bits=`**_arg_  
->  select 32 or 64 bit build
-
->  **`--with-jvm-variants=`**_variants_  
->  select the JVM variants to build from, comma separated list that can
-   include: server, client, kernel, zero and zeroshark
-
->  **`--with-memory-size=`**_size_  
->  select the RAM size that GNU make will think this system has
-
->  **`--with-msvcr-dll=`**_path_  
->  select the `msvcr100.dll` file to include in the Windows builds (C/C++
-   runtime library for Visual Studio).
-
->  This is usually picked up automatically from the redist directories of
-   Visual Studio 2013.
-
->  **`--with-num-cores=`**_cores_  
->  select the number of cores to use (processor count or CPU count)
-
-<a name="xrender"></a>
->  **`--with-x=`**_path_  
->  select the location of the X11 and xrender files.
-
->  The XRender Extension Headers are required for building the OpenJDK on
-   Solaris and Linux. The Linux header files are usually available from a
-   "Xrender" development package, it's recommended that you try and use the
-   package provided by the particular distribution of Linux that you are using.
-   The Solaris XRender header files is included with the other X11 header files
-   in the package **SFWxwinc** on new enough versions of Solaris and will be
-   installed in `/usr/X11/include/X11/extensions/Xrender.h` or
-   `/usr/openwin/share/include/X11/extensions/Xrender.h`
-
-*****
-
-<a name="make"></a>
-### Make
-
-The basic invocation of the `make` utility looks like:
-
->  **`make all`**
-
-This will start the build to the output directory containing the
-"configuration" that was created by the `configure` script. Run `make help` for
-more information on the available targets.
-
-There are some of the make targets that are of general interest:
-
->  _empty_  
->  build everything but no images
-
->  **`all`**  
->  build everything including images
-
->  **`all-conf`**  
->  build all configurations
-
->  **`images`**  
->  create complete j2sdk and j2re images
-
->  **`install`**  
->  install the generated images locally, typically in `/usr/local`
-
->  **`clean`**  
->  remove all files generated by make, but not those generated by `configure`
-
->  **`dist-clean`**  
->  remove all files generated by both and `configure` (basically killing the
-   configuration)
-
->  **`help`**  
->  give some help on using `make`, including some interesting make targets
-
-*****
-
-<a name="testing"></a>
-## Testing
-
-When the build is completed, you should see the generated binaries and
-associated files in the `j2sdk-image` directory in the output directory. In
-particular, the `build/*/images/j2sdk-image/bin` directory should contain
-executables for the OpenJDK tools and utilities for that configuration. The
-testing tool `jtreg` will be needed and can be found at: [the jtreg
-site](http://openjdk.java.net/jtreg/). The provided regression tests in the
-repositories can be run with the command:
-
->  **``cd test && make PRODUCT_HOME=`pwd`/../build/*/images/j2sdk-image all``**
-
-*****
-
-<a name="hints"></a>
-## Appendix A: Hints and Tips
-
-<a name="faq"></a>
-### FAQ
-
-**Q:** The `generated-configure.sh` file looks horrible! How are you going to
-edit it?  
-**A:** The `generated-configure.sh` file is generated (think "compiled") by the
-autoconf tools. The source code is in `configure.ac` and various .m4 files in
-common/autoconf, which are much more readable.
-
-**Q:** Why is the `generated-configure.sh` file checked in, if it is 
-generated?  
-**A:** If it was not generated, every user would need to have the autoconf
-tools installed, and re-generate the `configure` file as the first step. Our
-goal is to minimize the work needed to be done by the user to start building
-OpenJDK, and to minimize the number of external dependencies required.
-
-**Q:** Do you require a specific version of autoconf for regenerating
-`generated-configure.sh`?  
-**A:** Yes, version 2.69 is required and should be easy enough to aquire on all
-supported operating systems. The reason for this is to avoid large spurious
-changes in `generated-configure.sh`.
-
-**Q:** How do you regenerate `generated-configure.sh` after making changes to
-the input files?  
-**A:** Regnerating `generated-configure.sh` should always be done using the
-script `common/autoconf/autogen.sh` to ensure that the correct files get
-updated. This script should also be run after mercurial tries to merge
-`generated-configure.sh` as a merge of the generated file is not guaranteed to
-be correct.
-
-**Q:** What are the files in `common/makefiles/support/*` for? They look like
-gibberish.  
-**A:** They are a somewhat ugly hack to compensate for command line length
-limitations on certain platforms (Windows, Solaris). Due to a combination of
-limitations in make and the shell, command lines containing too many files will
-not work properly. These helper files are part of an elaborate hack that will
-compress the command line in the makefile and then uncompress it safely. We're
-not proud of it, but it does fix the problem. If you have any better
-suggestions, we're all ears! :-)
-
-**Q:** I want to see the output of the commands that make runs, like in the old
-build. How do I do that?  
-**A:** You specify the `LOG` variable to make. There are several log levels:
-
- * **`warn`** -- Default and very quiet.
- * **`info`** -- Shows more progress information than warn.
- * **`debug`** -- Echos all command lines and prints all macro calls for
-   compilation definitions.
- * **`trace`** -- Echos all $(shell) command lines as well.
-
-**Q:** When do I have to re-run `configure`?  
-**A:** Normally you will run `configure` only once for creating a
-configuration. You need to re-run configuration only if you want to change any
-configuration options, or if you pull down changes to the `configure` script.
-
-**Q:** I have added a new source file. Do I need to modify the makefiles?  
-**A:** Normally, no. If you want to create e.g. a new native library, you will
-need to modify the makefiles. But for normal file additions or removals, no
-changes are needed. There are certan exceptions for some native libraries where
-the source files are spread over many directories which also contain sources
-for other libraries. In these cases it was simply easier to create include
-lists rather than excludes.
-
-**Q:** When I run `configure --help`, I see many strange options, like
-`--dvidir`. What is this?  
-**A:** Configure provides a slew of options by default, to all projects that
-use autoconf. Most of them are not used in OpenJDK, so you can safely ignore
-them. To list only OpenJDK specific features, use `configure --help=short`
-instead.
-
-**Q:** `configure` provides OpenJDK-specific features such as `--with-
-builddeps-server` that are not described in this document. What about those?  
-**A:** Try them out if you like! But be aware that most of these are
-experimental features. Many of them don't do anything at all at the moment; the
-option is just a placeholder. Others depend on pieces of code or infrastructure
-that is currently not ready for prime time.
-
-**Q:** How will you make sure you don't break anything?  
-**A:** We have a script that compares the result of the new build system with
-the result of the old. For most part, we aim for (and achieve) byte-by-byte
-identical output. There are however technical issues with e.g. native binaries,
-which might differ in a byte-by-byte comparison, even when building twice with
-the old build system. For these, we compare relevant aspects (e.g. the symbol
-table and file size). Note that we still don't have 100% equivalence, but we're
-close.
-
-**Q:** I noticed this thing X in the build that looks very broken by design.
-Why don't you fix it?  
-**A:** Our goal is to produce a build output that is as close as technically
-possible to the old build output. If things were weird in the old build, they
-will be weird in the new build. Often, things were weird before due to
-obscurity, but in the new build system the weird stuff comes up to the surface.
-The plan is to attack these things at a later stage, after the new build system
-is established.
-
-**Q:** The code in the new build system is not that well-structured. Will you
-fix this?  
-**A:** Yes! The new build system has grown bit by bit as we converted the old
-system. When all of the old build system is converted, we can take a step back
-and clean up the structure of the new build system. Some of this we plan to do
-before replacing the old build system and some will need to wait until after.
-
-**Q:** Is anything able to use the results of the new build's default make
-target?  
-**A:** Yes, this is the minimal (or roughly minimal) set of compiled output
-needed for a developer to actually execute the newly built JDK. The idea is
-that in an incremental development fashion, when doing a normal make, you
-should only spend time recompiling what's changed (making it purely
-incremental) and only do the work that's needed to actually run and test your
-code. The packaging stuff that is part of the `images` target is not needed for
-a normal developer who wants to test his new code. Even if it's quite fast,
-it's still unnecessary. We're targeting sub-second incremental rebuilds! ;-)
-(Or, well, at least single-digit seconds...)
-
-**Q:** I usually set a specific environment variable when building, but I can't
-find the equivalent in the new build. What should I do?  
-**A:** It might very well be that we have neglected to add support for an
-option that was actually used from outside the build system. Email us and we
-will add support for it!
-
-<a name="performance"></a>
-### Build Performance Tips
-
-Building OpenJDK requires a lot of horsepower. Some of the build tools can be
-adjusted to utilize more or less of resources such as parallel threads and
-memory. The `configure` script analyzes your system and selects reasonable
-values for such options based on your hardware. If you encounter resource
-problems, such as out of memory conditions, you can modify the detected values
-with:
-
- * **`--with-num-cores`** -- number of cores in the build system, e.g.
-   `--with-num-cores=8`
- * **`--with-memory-size`** -- memory (in MB) available in the build system,
-    e.g. `--with-memory-size=1024`
-
-It might also be necessary to specify the JVM arguments passed to the Bootstrap
-JDK, using e.g. `--with-boot-jdk-jvmargs="-Xmx8G -enableassertions"`. Doing
-this will override the default JVM arguments passed to the Bootstrap JDK.
-
-One of the top goals of the new build system is to improve the build
-performance and decrease the time needed to build. This will soon also apply to
-the java compilation when the Smart Javac wrapper is fully supported.
-
-At the end of a successful execution of `configure`, you will get a performance
-summary, indicating how well the build will perform. Here you will also get
-performance hints. If you want to build fast, pay attention to those!
-
-#### Building with ccache
-
-The OpenJDK build supports building with ccache when using gcc or clang. Using
-ccache can radically speed up compilation of native code if you often rebuild
-the same sources. Your milage may vary however so we recommend evaluating it
-for yourself. To enable it, make sure it's on the path and configure with
-`--enable-ccache`.
-
-#### Building on local disk
-
-If you are using network shares, e.g. via NFS, for your source code, make sure
-the build directory is situated on local disk. The performance penalty is
-extremely high for building on a network share, close to unusable.
-
-#### Building only one JVM
-
-The old build builds multiple JVMs on 32-bit systems (client and server; and on
-Windows kernel as well). In the new build we have changed this default to only
-build server when it's available. This improves build times for those not
-interested in multiple JVMs. To mimic the old behavior on platforms that
-support it, use `--with-jvm-variants=client,server`.
-
-#### Selecting the number of cores to build on
-
-By default, `configure` will analyze your machine and run the make process in
-parallel with as many threads as you have cores. This behavior can be
-overridden, either "permanently" (on a `configure` basis) using
-`--with-num-cores=N` or for a single build only (on a make basis), using
-`make JOBS=N`.
-
-If you want to make a slower build just this time, to save some CPU power for
-other processes, you can run e.g. `make JOBS=2`. This will force the makefiles
-to only run 2 parallel processes, or even `make JOBS=1` which will disable
-parallelism.
-
-If you want to have it the other way round, namely having slow builds default
-and override with fast if you're impatient, you should call `configure` with
-`--with-num-cores=2`, making 2 the default. If you want to run with more cores,
-run `make JOBS=8`
-
-<a name="troubleshooting"></a>
-### Troubleshooting
-
-#### Solving build problems
-
-If the build fails (and it's not due to a compilation error in a source file
-you've changed), the first thing you should do is to re-run the build with more
-verbosity. Do this by adding `LOG=debug` to your make command line.
-
-The build log (with both stdout and stderr intermingled, basically the same as
-you see on your console) can be found as `build.log` in your build directory.
-
-You can ask for help on build problems with the new build system on either the
-[build-dev](http://mail.openjdk.java.net/mailman/listinfo/build-dev) or the
-[build-infra-dev](http://mail.openjdk.java.net/mailman/listinfo/build-infra-dev)
-mailing lists. Please include the relevant parts of the build log.
-
-A build can fail for any number of reasons. Most failures are a result of
-trying to build in an environment in which all the pre-build requirements have
-not been met. The first step in troubleshooting a build failure is to recheck
-that you have satisfied all the pre-build requirements for your platform.
-Scanning the `configure` log is a good first step, making sure that what it
-found makes sense for your system. Look for strange error messages or any
-difficulties that `configure` had in finding things.
-
-Some of the more common problems with builds are briefly described below, with
-suggestions for remedies.
-
- * **Corrupted Bundles on Windows:**  
-   Some virus scanning software has been known to corrupt the downloading of
-   zip bundles. It may be necessary to disable the 'on access' or 'real time'
-   virus scanning features to prevent this corruption. This type of 'real time'
-   virus scanning can also slow down the build process significantly.
-   Temporarily disabling the feature, or excluding the build output directory
-   may be necessary to get correct and faster builds.
-
- * **Slow Builds:**  
-   If your build machine seems to be overloaded from too many simultaneous C++
-   compiles, try setting the `JOBS=1` on the `make` command line. Then try
-   increasing the count slowly to an acceptable level for your system. Also:
-
-   Creating the javadocs can be very slow, if you are running javadoc, consider
-   skipping that step.
-
-   Faster CPUs, more RAM, and a faster DISK usually helps. The VM build tends
-   to be CPU intensive (many C++ compiles), and the rest of the JDK will often
-   be disk intensive.
-
-   Faster compiles are possible using a tool called
-   [ccache](http://ccache.samba.org/).
-
- * **File time issues:**  
-   If you see warnings that refer to file time stamps, e.g.
-
-   > _Warning message:_ ` File 'xxx' has modification time in the future.`  
-   > _Warning message:_ ` Clock skew detected. Your build may be incomplete.`
-
-   These warnings can occur when the clock on the build machine is out of sync
-   with the timestamps on the source files. Other errors, apparently unrelated
-   but in fact caused by the clock skew, can occur along with the clock skew
-   warnings. These secondary errors may tend to obscure the fact that the true
-   root cause of the problem is an out-of-sync clock.
-
-   If you see these warnings, reset the clock on the build machine, run
-   "`gmake clobber`" or delete the directory containing the build output, and
-   restart the build from the beginning.
-
- * **Error message: `Trouble writing out table to disk`**  
-   Increase the amount of swap space on your build machine. This could be
-   caused by overloading the system and it may be necessary to use:
-
-   > `make JOBS=1`
-
-   to reduce the load on the system.
-
- * **Error Message: `libstdc++ not found`:**  
-   This is caused by a missing libstdc++.a library. This is installed as part
-   of a specific package (e.g. libstdc++.so.devel.386). By default some 64-bit
-   Linux versions (e.g. Fedora) only install the 64-bit version of the
-   libstdc++ package. Various parts of the JDK build require a static link of
-   the C++ runtime libraries to allow for maximum portability of the built
-   images.
-
- * **Linux Error Message: `cannot restore segment prot after reloc`**  
-   This is probably an issue with SELinux (See [SELinux on
-   Wikipedia](http://en.wikipedia.org/wiki/SELinux)). Parts of the VM is built
-   without the `-fPIC` for performance reasons.
-
-   To completely disable SELinux:
-
-   1. `$ su root`
-   2. `# system-config-securitylevel`
-   3. `In the window that appears, select the SELinux tab`
-   4. `Disable SELinux`
-
-   Alternatively, instead of completely disabling it you could disable just
-   this one check.
-
-   1. Select System->Administration->SELinux Management
-   2. In the SELinux Management Tool which appears, select "Boolean" from the
-      menu on the left
-   3. Expand the "Memory Protection" group
-   4. Check the first item, labeled "Allow all unconfined executables to use
-      libraries requiring text relocation ..."
-
- * **Windows Error Messages:**  
-   `*** fatal error - couldn't allocate heap, ... `  
-   `rm fails with "Directory not empty"`  
-   `unzip fails with "cannot create ... Permission denied"`  
-   `unzip fails with "cannot create ... Error 50"`
-
-   The CYGWIN software can conflict with other non-CYGWIN software. See the
-   CYGWIN FAQ section on [BLODA (applications that interfere with
-   CYGWIN)](http://cygwin.com/faq/faq.using.html#faq.using.bloda).
-
- * **Windows Error Message: `spawn failed`**  
-   Try rebooting the system, or there could be some kind of issue with the disk
-   or disk partition being used. Sometimes it comes with a "Permission Denied"
-   message.
-
-*****
-
-<a name="gmake"></a>
-## Appendix B: GNU make
-
-The Makefiles in the OpenJDK are only valid when used with the GNU version of
-the utility command `make` (usually called `gmake` on Solaris). A few notes
-about using GNU make:
-
- * You need GNU make version 3.81 or newer. On Windows 4.0 or newer is
-   recommended. If the GNU make utility on your systems is not of a suitable
-   version, see "[Building GNU make](#buildgmake)".
- * Place the location of the GNU make binary in the `PATH`.
- * **Solaris:** Do NOT use `/usr/bin/make` on Solaris. If your Solaris system
-   has the software from the Solaris Developer Companion CD installed, you
-   should try and use `/usr/bin/gmake` or `/usr/gnu/bin/make`.
- * **Windows:** Make sure you start your build inside a bash shell.
- * **Mac OS X:** The XCode "command line tools" must be installed on your Mac.
-
-Information on GNU make, and access to ftp download sites, are available on the
-[GNU make web site ](http://www.gnu.org/software/make/make.html). The latest
-source to GNU make is available at
-[ftp.gnu.org/pub/gnu/make/](http://ftp.gnu.org/pub/gnu/make/).
-
-<a name="buildgmake"></a>
-### Building GNU make
-
-First step is to get the GNU make 3.81 or newer source from
-[ftp.gnu.org/pub/gnu/make/](http://ftp.gnu.org/pub/gnu/make/). Building is a
-little different depending on the OS but is basically done with:
-
-      bash ./configure
-      make
-
-*****
-
-<a name="buildenvironments"></a>
-## Appendix C: Build Environments
-
-### Minimum Build Environments
-
-This file often describes specific requirements for what we call the "minimum
-build environments" (MBE) for this specific release of the JDK. What is listed
-below is what the Oracle Release Engineering Team will use to build the Oracle
-JDK product. Building with the MBE will hopefully generate the most compatible
-bits that install on, and run correctly on, the most variations of the same
-base OS and hardware architecture. In some cases, these represent what is often
-called the least common denominator, but each Operating System has different
-aspects to it.
-
-In all cases, the Bootstrap JDK version minimum is critical, we cannot
-guarantee builds will work with older Bootstrap JDK's. Also in all cases, more
-RAM and more processors is better, the minimums listed below are simply
-recommendations.
-
-With Solaris and Mac OS X, the version listed below is the oldest release we
-can guarantee builds and works, and the specific version of the compilers used
-could be critical.
-
-With Windows the critical aspect is the Visual Studio compiler used, which due
-to it's runtime, generally dictates what Windows systems can do the builds and
-where the resulting bits can be used.
-
-**NOTE: We expect a change here off these older Windows OS releases and to a
-'less older' one, probably Windows 2008R2 X64.**
-
-With Linux, it was just a matter of picking a stable distribution that is a
-good representative for Linux in general.
-
-It is understood that most developers will NOT be using these specific
-versions, and in fact creating these specific versions may be difficult due to
-the age of some of this software. It is expected that developers are more often
-using the more recent releases and distributions of these operating systems.
-
-Compilation problems with newer or different C/C++ compilers is a common
-problem. Similarly, compilation problems related to changes to the
-`/usr/include` or system header files is also a common problem with older,
-newer, or unreleased OS versions. Please report these types of problems as bugs
-so that they can be dealt with accordingly.
-
->  <table border="1">
-     <thead>
-       <tr>
-         <th>Base OS and Architecture</th>
-         <th>OS</th>
-         <th>C/C++ Compiler</th>
-         <th>Bootstrap JDK</th>
-         <th>Processors</th>
-         <th>RAM Minimum</th>
-         <th>DISK Needs</th>
-       </tr>
-     </thead>
-     <tbody>
-       <tr>
-         <td>Linux X86 (32-bit) and X64 (64-bit)</td>
-         <td>Oracle Enterprise Linux 6.4</td>
-         <td>gcc 4.9.2 </td>
-         <td>JDK 8</td>
-         <td>2 or more</td>
-         <td>1 GB</td>
-         <td>6 GB</td>
-       </tr>
-       <tr>
-         <td>Solaris SPARCV9 (64-bit)</td>
-         <td>Solaris 11 Update 1</td>
-         <td>Studio 12 Update 4 + patches</td>
-         <td>JDK 8</td>
-         <td>4 or more</td>
-         <td>4 GB</td>
-         <td>8 GB</td>
-       </tr>
-       <tr>
-         <td>Solaris X64 (64-bit)</td>
-         <td>Solaris 11 Update 1</td>
-         <td>Studio 12 Update 4 + patches</td>
-         <td>JDK 8</td>
-         <td>4 or more</td>
-         <td>4 GB</td>
-         <td>8 GB</td>
-       </tr>
-       <tr>
-         <td>Windows X86 (32-bit)</td>
-         <td>Windows Server 2012 R2 x64</td>
-         <td>Microsoft Visual Studio C++ 2013 Professional Edition</td>
-         <td>JDK 8</td>
-         <td>2 or more</td>
-         <td>2 GB</td>
-         <td>6 GB</td>
-       </tr>
-       <tr>
-         <td>Windows X64 (64-bit)</td>
-         <td>Windows Server 2012 R2 x64</td>
-         <td>Microsoft Visual Studio C++ 2013 Professional Edition</td>
-         <td>JDK 8</td>
-         <td>2 or more</td>
-         <td>2 GB</td>
-         <td>6 GB</td>
-       </tr>
-       <tr>
-         <td>Mac OS X X64 (64-bit)</td>
-         <td>Mac OS X 10.9 "Mavericks"</td>
-         <td>Xcode 6.3 or newer</td>
-         <td>JDK 8</td>
-         <td>2 or more</td>
-         <td>4 GB</td>
-         <td>6 GB</td>
-       </tr>
-     </tbody>
-   </table>
-
-*****
-
-<a name="SDBE"></a>
-### Specific Developer Build Environments
-
-We won't be listing all the possible environments, but we will try to provide
-what information we have available to us.
-
-**NOTE: The community can help out by updating this part of the document.**
-
-#### Fedora
-
-After installing the latest [Fedora](http://fedoraproject.org) you need to
-install several build dependencies. The simplest way to do it is to execute the
-following commands as user `root`:
-
-      yum-builddep java-1.7.0-openjdk
-      yum install gcc gcc-c++
-
-In addition, it's necessary to set a few environment variables for the build:
-
-      export LANG=C
-      export PATH="/usr/lib/jvm/java-openjdk/bin:${PATH}"
-
-#### CentOS 5.5
-
-After installing [CentOS 5.5](http://www.centos.org/) you need to make sure you
-have the following Development bundles installed:
-
- * Development Libraries
- * Development Tools
- * Java Development
- * X Software Development (Including XFree86-devel)
-
-Plus the following packages:
-
- * cups devel: Cups Development Package
- * alsa devel: Alsa Development Package
- * Xi devel: libXi.so Development Package
-
-The freetype 2.3 packages don't seem to be available, but the freetype 2.3
-sources can be downloaded, built, and installed easily enough from [the
-freetype site](http://downloads.sourceforge.net/freetype). Build and install
-with something like:
-
-      bash ./configure
-      make
-      sudo -u root make install
-
-Mercurial packages could not be found easily, but a Google search should find
-ones, and they usually include Python if it's needed.
-
-#### Debian 5.0 (Lenny)
-
-After installing [Debian](http://debian.org) 5 you need to install several
-build dependencies. The simplest way to install the build dependencies is to
-execute the following commands as user `root`:
-
-      aptitude build-dep openjdk-7
-      aptitude install openjdk-7-jdk libmotif-dev
-
-In addition, it's necessary to set a few environment variables for the build:
-
-      export LANG=C
-      export PATH="/usr/lib/jvm/java-7-openjdk/bin:${PATH}"
-
-#### Ubuntu 12.04
-
-After installing [Ubuntu](http://ubuntu.org) 12.04 you need to install several
-build dependencies. The simplest way to do it is to execute the following
-commands:
-
-      sudo aptitude build-dep openjdk-7
-      sudo aptitude install openjdk-7-jdk
-
-In addition, it's necessary to set a few environment variables for the build:
-
-      export LANG=C
-      export PATH="/usr/lib/jvm/java-7-openjdk/bin:${PATH}"
-
-#### OpenSUSE 11.1
-
-After installing [OpenSUSE](http://opensuse.org) 11.1 you need to install
-several build dependencies. The simplest way to install the build dependencies
-is to execute the following commands:
-
-      sudo zypper source-install -d java-1_7_0-openjdk
-      sudo zypper install make
-
-In addition, it is necessary to set a few environment variables for the build:
-
-      export LANG=C
-      export PATH="/usr/lib/jvm/java-1.7.0-openjdk/bin:$[PATH}"
-
-Finally, you need to unset the `JAVA_HOME` environment variable:
-
-      export -n JAVA_HOME`
-
-#### Mandriva Linux One 2009 Spring
-
-After installing [Mandriva](http://mandriva.org) Linux One 2009 Spring you need
-to install several build dependencies. The simplest way to install the build
-dependencies is to execute the following commands as user `root`:
-
-      urpmi java-1.7.0-openjdk-devel make gcc gcc-c++ freetype-devel zip unzip
-        libcups2-devel libxrender1-devel libalsa2-devel libstc++-static-devel
-        libxtst6-devel libxi-devel
-
-In addition, it is necessary to set a few environment variables for the build:
-
-      export LANG=C
-      export PATH="/usr/lib/jvm/java-1.7.0-openjdk/bin:${PATH}"
-
-#### OpenSolaris 2009.06
-
-After installing [OpenSolaris](http://opensolaris.org) 2009.06 you need to
-install several build dependencies. The simplest way to install the build
-dependencies is to execute the following commands:
-
-      pfexec pkg install SUNWgmake SUNWj7dev sunstudioexpress SUNWcups SUNWzip
-        SUNWunzip SUNWxwhl SUNWxorg-headers SUNWaudh SUNWfreetype2
-
-In addition, it is necessary to set a few environment variables for the build:
-
-      export LANG=C
-      export PATH="/opt/SunStudioExpress/bin:${PATH}"
-
-*****
-
-End of the OpenJDK build README document.
-
-Please come again!
--- a/common/autoconf/basics.m4	Tue Feb 14 14:57:56 2017 -0800
+++ b/common/autoconf/basics.m4	Thu Aug 17 12:17:39 2017 -0700
@@ -530,6 +530,7 @@
   BASIC_PATH_PROGS(DF, df)
   BASIC_PATH_PROGS(CPIO, [cpio bsdcpio])
   BASIC_PATH_PROGS(NICE, nice)
+  BASIC_PATH_PROGS(PANDOC, pandoc)
 ])
 
 # Setup basic configuration paths, and platform-specific stuff related to PATHs.
@@ -1094,6 +1095,7 @@
     OTOOL="true"
   fi
   BASIC_PATH_PROGS(READELF, [greadelf readelf])
+  BASIC_PATH_PROGS(DOT, dot)
   BASIC_PATH_PROGS(HG, hg)
   BASIC_PATH_PROGS(STAT, stat)
   BASIC_PATH_PROGS(TIME, time)
--- a/common/autoconf/basics_windows.m4	Tue Feb 14 14:57:56 2017 -0800
+++ b/common/autoconf/basics_windows.m4	Thu Aug 17 12:17:39 2017 -0700
@@ -329,8 +329,8 @@
       AC_MSG_ERROR([Something is wrong with your cygwin installation since I cannot find cygpath.exe in your path])
     fi
     AC_MSG_CHECKING([cygwin root directory as unix-style path])
-    # The cmd output ends with Windows line endings (CR/LF), the grep command will strip that away
-    cygwin_winpath_root=`cd / ; cmd /c cd | $GREP ".*"`
+    # The cmd output ends with Windows line endings (CR/LF)
+    cygwin_winpath_root=`cd / ; cmd /c cd | $TR -d '\r\n'`
     # Force cygpath to report the proper root by including a trailing space, and then stripping it off again.
     CYGWIN_ROOT_PATH=`$CYGPATH -u "$cygwin_winpath_root " | $CUT -f 1 -d " "`
     AC_MSG_RESULT([$CYGWIN_ROOT_PATH])
--- a/common/autoconf/build-aux/install.sh	Tue Feb 14 14:57:56 2017 -0800
+++ b/common/autoconf/build-aux/install.sh	Thu Aug 17 12:17:39 2017 -0700
@@ -1,5 +1,28 @@
-#!/bin/sh
-echo >&2 "No suitable 'install' command found.'"
-echo >&2 "If automake is installed, running 'automake -fa'"
-echo >&2 "(and ignoring the errors) might produce one."
-exit 1
+#!/bin/bash
+#
+# Copyright (c) 2017, Oracle and/or its affiliates. All rights reserved.
+# DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
+#
+# This code is free software; you can redistribute it and/or modify it
+# under the terms of the GNU General Public License version 2 only, as
+# published by the Free Software Foundation.  Oracle designates this
+# particular file as subject to the "Classpath" exception as provided
+# by Oracle in the LICENSE file that accompanied this code.
+#
+# This code is distributed in the hope that it will be useful, but WITHOUT
+# ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
+# FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License
+# version 2 for more details (a copy is included in the LICENSE file that
+# accompanied this code).
+#
+# You should have received a copy of the GNU General Public License version
+# 2 along with this work; if not, write to the Free Software Foundation,
+# Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
+#
+# Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
+# or visit www.oracle.com if you need additional information or have any
+# questions.
+#
+
+# This file is empty on purpose. It's a placeholder which is required by
+# autoconf, but it serves no purpose for us.
--- a/common/autoconf/flags.m4	Tue Feb 14 14:57:56 2017 -0800
+++ b/common/autoconf/flags.m4	Thu Aug 17 12:17:39 2017 -0700
@@ -355,7 +355,7 @@
         SHARED_LIBRARY_FLAGS="-dynamiclib -compatibility_version 1.0.0 -current_version 1.0.0 $PICFLAG"
         JVM_CFLAGS="$JVM_CFLAGS $PICFLAG"
       fi
-      SET_EXECUTABLE_ORIGIN='-Wl,-rpath,@loader_path/.'
+      SET_EXECUTABLE_ORIGIN='-Wl,-rpath,@loader_path$(or [$]1,/.)'
       SET_SHARED_LIBRARY_ORIGIN="$SET_EXECUTABLE_ORIGIN"
       SET_SHARED_LIBRARY_NAME='-Wl,-install_name,@rpath/[$]1'
       SET_SHARED_LIBRARY_MAPFILE='-Wl,-exported_symbols_list,[$]1'
@@ -375,7 +375,7 @@
       # Linking is different on MacOSX
       PICFLAG=''
       SHARED_LIBRARY_FLAGS="-dynamiclib -compatibility_version 1.0.0 -current_version 1.0.0 $PICFLAG"
-      SET_EXECUTABLE_ORIGIN='-Wl,-rpath,@loader_path/.'
+      SET_EXECUTABLE_ORIGIN='-Wl,-rpath,@loader_path$(or [$]1,/.)'
       SET_SHARED_LIBRARY_ORIGIN="$SET_EXECUTABLE_ORIGIN"
       SET_SHARED_LIBRARY_NAME='-Wl,-install_name,@rpath/[$]1'
       SET_SHARED_LIBRARY_MAPFILE='-Wl,-exported_symbols_list,[$]1'
--- a/common/autoconf/generated-configure.sh	Tue Feb 14 14:57:56 2017 -0800
+++ b/common/autoconf/generated-configure.sh	Thu Aug 17 12:17:39 2017 -0700
@@ -692,6 +692,7 @@
 LIBFFI_CFLAGS
 ALSA_LIBS
 ALSA_CFLAGS
+FREETYPE_LICENSE
 FREETYPE_BUNDLE_LIB_PATH
 FREETYPE_LIBS
 FREETYPE_CFLAGS
@@ -922,6 +923,7 @@
 INCLUDE_SA
 UNLIMITED_CRYPTO
 CACERTS_FILE
+ENABLE_FULL_DOCS
 ENABLE_HEADLESS_ONLY
 DEFAULT_MAKE_TARGET
 OS_VERSION_MICRO
@@ -939,6 +941,7 @@
 TIME
 STAT
 HG
+DOT
 READELF
 OTOOL
 LDD
@@ -997,8 +1000,9 @@
 OPENJDK_TARGET_CPU_ISADIR
 OPENJDK_TARGET_CPU_LEGACY_LIB
 OPENJDK_TARGET_CPU_LEGACY
-REQUIRED_OS_VERSION
-REQUIRED_OS_NAME
+RELEASE_FILE_OS_ARCH
+RELEASE_FILE_OS_NAME
+OPENJDK_MODULE_TARGET_PLATFORM
 COMPILE_TYPE
 OPENJDK_TARGET_CPU_ENDIAN
 OPENJDK_TARGET_CPU_BITS
@@ -1028,6 +1032,7 @@
 build_vendor
 build_cpu
 build
+PANDOC
 NICE
 CPIO
 DF
@@ -1140,6 +1145,7 @@
 with_output_sync
 with_default_make_target
 enable_headless_only
+enable_full_docs
 with_cacerts_file
 enable_unlimited_crypto
 with_copyright_year
@@ -1198,6 +1204,7 @@
 with_freetype_lib
 with_freetype_src
 enable_freetype_bundling
+with_freetype_license
 with_alsa
 with_alsa_include
 with_alsa_lib
@@ -1288,12 +1295,14 @@
 DF
 CPIO
 NICE
+PANDOC
 MAKE
 UNZIP
 ZIPEXE
 LDD
 OTOOL
 READELF
+DOT
 HG
 STAT
 TIME
@@ -1974,6 +1983,8 @@
   --enable-debug          set the debug level to fastdebug (shorthand for
                           --with-debug-level=fastdebug) [disabled]
   --enable-headless-only  only build headless (no GUI) support [disabled]
+  --enable-full-docs      build complete documentation [enabled if all tools
+                          found]
   --disable-unlimited-crypto
                           Disable unlimited crypto policy [enabled]
   --disable-keep-packaged-modules
@@ -2149,6 +2160,7 @@
   --with-freetype-src     specify directory with freetype sources to
                           automatically build the library (experimental,
                           Windows-only)
+  --with-freetype-license if bundling freetype, also bundle this license file
   --with-alsa             specify prefix directory for the alsa package
                           (expecting the libraries under PATH/lib and the
                           headers under PATH/include)
@@ -2260,12 +2272,14 @@
   DF          Override default value for DF
   CPIO        Override default value for CPIO
   NICE        Override default value for NICE
+  PANDOC      Override default value for PANDOC
   MAKE        Override default value for MAKE
   UNZIP       Override default value for UNZIP
   ZIPEXE      Override default value for ZIPEXE
   LDD         Override default value for LDD
   OTOOL       Override default value for OTOOL
   READELF     Override default value for READELF
+  DOT         Override default value for DOT
   HG          Override default value for HG
   STAT        Override default value for STAT
   TIME        Override default value for TIME
@@ -4933,6 +4947,8 @@
 
 
 
+
+
 #%%% Build and target systems %%%
 
 
@@ -5089,7 +5105,7 @@
 #
 # $1 = compiler to test (CC or CXX)
 # $2 = human readable name of compiler (C or C++)
-# $3 = list of compiler names to search for
+# $3 = compiler name to search for
 
 
 # Detect the core components of the toolchain, i.e. the compilers (CC and CXX),
@@ -5216,7 +5232,7 @@
 #CUSTOM_AUTOCONF_INCLUDE
 
 # Do not change or remove the following line, it is needed for consistency checks:
-DATE_WHEN_GENERATED=1487097206
+DATE_WHEN_GENERATED=1502822879
 
 ###############################################################################
 #
@@ -15404,6 +15420,203 @@
 
 
 
+  # Publish this variable in the help.
+
+
+  if [ -z "${PANDOC+x}" ]; then
+    # The variable is not set by user, try to locate tool using the code snippet
+    for ac_prog in pandoc
+do
+  # Extract the first word of "$ac_prog", so it can be a program name with args.
+set dummy $ac_prog; ac_word=$2
+{ $as_echo "$as_me:${as_lineno-$LINENO}: checking for $ac_word" >&5
+$as_echo_n "checking for $ac_word... " >&6; }
+if ${ac_cv_path_PANDOC+:} false; then :
+  $as_echo_n "(cached) " >&6
+else
+  case $PANDOC in
+  [\\/]* | ?:[\\/]*)
+  ac_cv_path_PANDOC="$PANDOC" # Let the user override the test with a path.
+  ;;
+  *)
+  as_save_IFS=$IFS; IFS=$PATH_SEPARATOR
+for as_dir in $PATH
+do
+  IFS=$as_save_IFS
+  test -z "$as_dir" && as_dir=.
+    for ac_exec_ext in '' $ac_executable_extensions; do
+  if as_fn_executable_p "$as_dir/$ac_word$ac_exec_ext"; then
+    ac_cv_path_PANDOC="$as_dir/$ac_word$ac_exec_ext"
+    $as_echo "$as_me:${as_lineno-$LINENO}: found $as_dir/$ac_word$ac_exec_ext" >&5
+    break 2
+  fi
+done
+  done
+IFS=$as_save_IFS
+
+  ;;
+esac
+fi
+PANDOC=$ac_cv_path_PANDOC
+if test -n "$PANDOC"; then
+  { $as_echo "$as_me:${as_lineno-$LINENO}: result: $PANDOC" >&5
+$as_echo "$PANDOC" >&6; }
+else
+  { $as_echo "$as_me:${as_lineno-$LINENO}: result: no" >&5
+$as_echo "no" >&6; }
+fi
+
+
+  test -n "$PANDOC" && break
+done
+
+  else
+    # The variable is set, but is it from the command line or the environment?
+
+    # Try to remove the string !PANDOC! from our list.
+    try_remove_var=${CONFIGURE_OVERRIDDEN_VARIABLES//!PANDOC!/}
+    if test "x$try_remove_var" = "x$CONFIGURE_OVERRIDDEN_VARIABLES"; then
+      # If it failed, the variable was not from the command line. Ignore it,
+      # but warn the user (except for BASH, which is always set by the calling BASH).
+      if test "xPANDOC" != xBASH; then
+        { $as_echo "$as_me:${as_lineno-$LINENO}: WARNING: Ignoring value of PANDOC from the environment. Use command line variables instead." >&5
+$as_echo "$as_me: WARNING: Ignoring value of PANDOC from the environment. Use command line variables instead." >&2;}
+      fi
+      # Try to locate tool using the code snippet
+      for ac_prog in pandoc
+do
+  # Extract the first word of "$ac_prog", so it can be a program name with args.
+set dummy $ac_prog; ac_word=$2
+{ $as_echo "$as_me:${as_lineno-$LINENO}: checking for $ac_word" >&5
+$as_echo_n "checking for $ac_word... " >&6; }
+if ${ac_cv_path_PANDOC+:} false; then :
+  $as_echo_n "(cached) " >&6
+else
+  case $PANDOC in
+  [\\/]* | ?:[\\/]*)
+  ac_cv_path_PANDOC="$PANDOC" # Let the user override the test with a path.
+  ;;
+  *)
+  as_save_IFS=$IFS; IFS=$PATH_SEPARATOR
+for as_dir in $PATH
+do
+  IFS=$as_save_IFS
+  test -z "$as_dir" && as_dir=.
+    for ac_exec_ext in '' $ac_executable_extensions; do
+  if as_fn_executable_p "$as_dir/$ac_word$ac_exec_ext"; then
+    ac_cv_path_PANDOC="$as_dir/$ac_word$ac_exec_ext"
+    $as_echo "$as_me:${as_lineno-$LINENO}: found $as_dir/$ac_word$ac_exec_ext" >&5
+    break 2
+  fi
+done
+  done
+IFS=$as_save_IFS
+
+  ;;
+esac
+fi
+PANDOC=$ac_cv_path_PANDOC
+if test -n "$PANDOC"; then
+  { $as_echo "$as_me:${as_lineno-$LINENO}: result: $PANDOC" >&5
+$as_echo "$PANDOC" >&6; }
+else
+  { $as_echo "$as_me:${as_lineno-$LINENO}: result: no" >&5
+$as_echo "no" >&6; }
+fi
+
+
+  test -n "$PANDOC" && break
+done
+
+    else
+      # If it succeeded, then it was overridden by the user. We will use it
+      # for the tool.
+
+      # First remove it from the list of overridden variables, so we can test
+      # for unknown variables in the end.
+      CONFIGURE_OVERRIDDEN_VARIABLES="$try_remove_var"
+
+      # Check if we try to supply an empty value
+      if test "x$PANDOC" = x; then
+        { $as_echo "$as_me:${as_lineno-$LINENO}: Setting user supplied tool PANDOC= (no value)" >&5
+$as_echo "$as_me: Setting user supplied tool PANDOC= (no value)" >&6;}
+        { $as_echo "$as_me:${as_lineno-$LINENO}: checking for PANDOC" >&5
+$as_echo_n "checking for PANDOC... " >&6; }
+        { $as_echo "$as_me:${as_lineno-$LINENO}: result: disabled" >&5
+$as_echo "disabled" >&6; }
+      else
+        # Check if the provided tool contains a complete path.
+        tool_specified="$PANDOC"
+        tool_basename="${tool_specified##*/}"
+        if test "x$tool_basename" = "x$tool_specified"; then
+          # A command without a complete path is provided, search $PATH.
+          { $as_echo "$as_me:${as_lineno-$LINENO}: Will search for user supplied tool PANDOC=$tool_basename" >&5
+$as_echo "$as_me: Will search for user supplied tool PANDOC=$tool_basename" >&6;}
+          # Extract the first word of "$tool_basename", so it can be a program name with args.
+set dummy $tool_basename; ac_word=$2
+{ $as_echo "$as_me:${as_lineno-$LINENO}: checking for $ac_word" >&5
+$as_echo_n "checking for $ac_word... " >&6; }
+if ${ac_cv_path_PANDOC+:} false; then :
+  $as_echo_n "(cached) " >&6
+else
+  case $PANDOC in
+  [\\/]* | ?:[\\/]*)
+  ac_cv_path_PANDOC="$PANDOC" # Let the user override the test with a path.
+  ;;
+  *)
+  as_save_IFS=$IFS; IFS=$PATH_SEPARATOR
+for as_dir in $PATH
+do
+  IFS=$as_save_IFS
+  test -z "$as_dir" && as_dir=.
+    for ac_exec_ext in '' $ac_executable_extensions; do
+  if as_fn_executable_p "$as_dir/$ac_word$ac_exec_ext"; then
+    ac_cv_path_PANDOC="$as_dir/$ac_word$ac_exec_ext"
+    $as_echo "$as_me:${as_lineno-$LINENO}: found $as_dir/$ac_word$ac_exec_ext" >&5
+    break 2
+  fi
+done
+  done
+IFS=$as_save_IFS
+
+  ;;
+esac
+fi
+PANDOC=$ac_cv_path_PANDOC
+if test -n "$PANDOC"; then
+  { $as_echo "$as_me:${as_lineno-$LINENO}: result: $PANDOC" >&5
+$as_echo "$PANDOC" >&6; }
+else
+  { $as_echo "$as_me:${as_lineno-$LINENO}: result: no" >&5
+$as_echo "no" >&6; }
+fi
+
+
+          if test "x$PANDOC" = x; then
+            as_fn_error $? "User supplied tool $tool_basename could not be found" "$LINENO" 5
+          fi
+        else
+          # Otherwise we believe it is a complete path. Use it as it is.
+          { $as_echo "$as_me:${as_lineno-$LINENO}: Will use user supplied tool PANDOC=$tool_specified" >&5
+$as_echo "$as_me: Will use user supplied tool PANDOC=$tool_specified" >&6;}
+          { $as_echo "$as_me:${as_lineno-$LINENO}: checking for PANDOC" >&5
+$as_echo_n "checking for PANDOC... " >&6; }
+          if test ! -x "$tool_specified"; then
+            { $as_echo "$as_me:${as_lineno-$LINENO}: result: not found" >&5
+$as_echo "not found" >&6; }
+            as_fn_error $? "User supplied tool PANDOC=$tool_specified does not exist or is not executable" "$LINENO" 5
+          fi
+          { $as_echo "$as_me:${as_lineno-$LINENO}: result: $tool_specified" >&5
+$as_echo "$tool_specified" >&6; }
+        fi
+      fi
+    fi
+
+  fi
+
+
+
+
 # Now we can determine OpenJDK build and target platforms. This is required to
 # have early on.
 # Make sure we can run config.sub.
@@ -15867,30 +16080,38 @@
 $as_echo "$COMPILE_TYPE" >&6; }
 
 
+  if test "x$OPENJDK_TARGET_OS" = xmacosx; then
+    OPENJDK_MODULE_TARGET_OS_NAME="macos"
+  else
+    OPENJDK_MODULE_TARGET_OS_NAME="$OPENJDK_TARGET_OS"
+  fi
+
+  if test "x$OPENJDK_TARGET_CPU" = xx86_64; then
+    OPENJDK_MODULE_TARGET_OS_ARCH="amd64"
+  else
+    OPENJDK_MODULE_TARGET_OS_ARCH="$OPENJDK_TARGET_CPU"
+  fi
+
+  OPENJDK_MODULE_TARGET_PLATFORM="${OPENJDK_MODULE_TARGET_OS_NAME}-${OPENJDK_MODULE_TARGET_OS_ARCH}"
+
+
+
   if test "x$OPENJDK_TARGET_OS" = "xsolaris"; then
-    REQUIRED_OS_NAME=SunOS
-    REQUIRED_OS_VERSION=5.10
+    RELEASE_FILE_OS_NAME=SunOS
   fi
   if test "x$OPENJDK_TARGET_OS" = "xlinux"; then
-    REQUIRED_OS_NAME=Linux
-    REQUIRED_OS_VERSION=2.6
+    RELEASE_FILE_OS_NAME=Linux
   fi
   if test "x$OPENJDK_TARGET_OS" = "xwindows"; then
-    REQUIRED_OS_NAME=Windows
-    if test "x$OPENJDK_TARGET_CPU_BITS" = "x64"; then
-      REQUIRED_OS_VERSION=5.2
-    else
-      REQUIRED_OS_VERSION=5.1
-    fi
-  fi
-  if test "x$OPENJDK_TARGET_OS" = "xmacosx"; then
-    REQUIRED_OS_NAME=Darwin
-    REQUIRED_OS_VERSION=11.2
+    RELEASE_FILE_OS_NAME=Windows
+  fi
+  if test "x$OPENJDK_TARGET_OS" = xmacosx; then
+    RELEASE_FILE_OS_NAME="Darwin"
   fi
   if test "x$OPENJDK_TARGET_OS" = "xaix"; then
-    REQUIRED_OS_NAME=AIX
-    REQUIRED_OS_VERSION=7.1
-  fi
+    RELEASE_FILE_OS_NAME="AIX"
+  fi
+  RELEASE_FILE_OS_ARCH=${OPENJDK_TARGET_CPU}
 
 
 
@@ -16246,8 +16467,8 @@
     fi
     { $as_echo "$as_me:${as_lineno-$LINENO}: checking cygwin root directory as unix-style path" >&5
 $as_echo_n "checking cygwin root directory as unix-style path... " >&6; }
-    # The cmd output ends with Windows line endings (CR/LF), the grep command will strip that away
-    cygwin_winpath_root=`cd / ; cmd /c cd | $GREP ".*"`
+    # The cmd output ends with Windows line endings (CR/LF)
+    cygwin_winpath_root=`cd / ; cmd /c cd | $TR -d '\r\n'`
     # Force cygpath to report the proper root by including a trailing space, and then stripping it off again.
     CYGWIN_ROOT_PATH=`$CYGPATH -u "$cygwin_winpath_root " | $CUT -f 1 -d " "`
     { $as_echo "$as_me:${as_lineno-$LINENO}: result: $CYGWIN_ROOT_PATH" >&5
@@ -22316,6 +22537,203 @@
   # Publish this variable in the help.
 
 
+  if [ -z "${DOT+x}" ]; then
+    # The variable is not set by user, try to locate tool using the code snippet
+    for ac_prog in dot
+do
+  # Extract the first word of "$ac_prog", so it can be a program name with args.
+set dummy $ac_prog; ac_word=$2
+{ $as_echo "$as_me:${as_lineno-$LINENO}: checking for $ac_word" >&5
+$as_echo_n "checking for $ac_word... " >&6; }
+if ${ac_cv_path_DOT+:} false; then :
+  $as_echo_n "(cached) " >&6
+else
+  case $DOT in
+  [\\/]* | ?:[\\/]*)
+  ac_cv_path_DOT="$DOT" # Let the user override the test with a path.
+  ;;
+  *)
+  as_save_IFS=$IFS; IFS=$PATH_SEPARATOR
+for as_dir in $PATH
+do
+  IFS=$as_save_IFS
+  test -z "$as_dir" && as_dir=.
+    for ac_exec_ext in '' $ac_executable_extensions; do
+  if as_fn_executable_p "$as_dir/$ac_word$ac_exec_ext"; then
+    ac_cv_path_DOT="$as_dir/$ac_word$ac_exec_ext"
+    $as_echo "$as_me:${as_lineno-$LINENO}: found $as_dir/$ac_word$ac_exec_ext" >&5
+    break 2
+  fi
+done
+  done
+IFS=$as_save_IFS
+
+  ;;
+esac
+fi
+DOT=$ac_cv_path_DOT
+if test -n "$DOT"; then
+  { $as_echo "$as_me:${as_lineno-$LINENO}: result: $DOT" >&5
+$as_echo "$DOT" >&6; }
+else
+  { $as_echo "$as_me:${as_lineno-$LINENO}: result: no" >&5
+$as_echo "no" >&6; }
+fi
+
+
+  test -n "$DOT" && break
+done
+
+  else
+    # The variable is set, but is it from the command line or the environment?
+
+    # Try to remove the string !DOT! from our list.
+    try_remove_var=${CONFIGURE_OVERRIDDEN_VARIABLES//!DOT!/}
+    if test "x$try_remove_var" = "x$CONFIGURE_OVERRIDDEN_VARIABLES"; then
+      # If it failed, the variable was not from the command line. Ignore it,
+      # but warn the user (except for BASH, which is always set by the calling BASH).
+      if test "xDOT" != xBASH; then
+        { $as_echo "$as_me:${as_lineno-$LINENO}: WARNING: Ignoring value of DOT from the environment. Use command line variables instead." >&5
+$as_echo "$as_me: WARNING: Ignoring value of DOT from the environment. Use command line variables instead." >&2;}
+      fi
+      # Try to locate tool using the code snippet
+      for ac_prog in dot
+do
+  # Extract the first word of "$ac_prog", so it can be a program name with args.
+set dummy $ac_prog; ac_word=$2
+{ $as_echo "$as_me:${as_lineno-$LINENO}: checking for $ac_word" >&5
+$as_echo_n "checking for $ac_word... " >&6; }
+if ${ac_cv_path_DOT+:} false; then :
+  $as_echo_n "(cached) " >&6
+else
+  case $DOT in
+  [\\/]* | ?:[\\/]*)
+  ac_cv_path_DOT="$DOT" # Let the user override the test with a path.
+  ;;
+  *)
+  as_save_IFS=$IFS; IFS=$PATH_SEPARATOR
+for as_dir in $PATH
+do
+  IFS=$as_save_IFS
+  test -z "$as_dir" && as_dir=.
+    for ac_exec_ext in '' $ac_executable_extensions; do
+  if as_fn_executable_p "$as_dir/$ac_word$ac_exec_ext"; then
+    ac_cv_path_DOT="$as_dir/$ac_word$ac_exec_ext"
+    $as_echo "$as_me:${as_lineno-$LINENO}: found $as_dir/$ac_word$ac_exec_ext" >&5
+    break 2
+  fi
+done
+  done
+IFS=$as_save_IFS
+
+  ;;
+esac
+fi
+DOT=$ac_cv_path_DOT
+if test -n "$DOT"; then
+  { $as_echo "$as_me:${as_lineno-$LINENO}: result: $DOT" >&5
+$as_echo "$DOT" >&6; }
+else
+  { $as_echo "$as_me:${as_lineno-$LINENO}: result: no" >&5
+$as_echo "no" >&6; }
+fi
+
+
+  test -n "$DOT" && break
+done
+
+    else
+      # If it succeeded, then it was overridden by the user. We will use it
+      # for the tool.
+
+      # First remove it from the list of overridden variables, so we can test
+      # for unknown variables in the end.
+      CONFIGURE_OVERRIDDEN_VARIABLES="$try_remove_var"
+
+      # Check if we try to supply an empty value
+      if test "x$DOT" = x; then
+        { $as_echo "$as_me:${as_lineno-$LINENO}: Setting user supplied tool DOT= (no value)" >&5
+$as_echo "$as_me: Setting user supplied tool DOT= (no value)" >&6;}
+        { $as_echo "$as_me:${as_lineno-$LINENO}: checking for DOT" >&5
+$as_echo_n "checking for DOT... " >&6; }
+        { $as_echo "$as_me:${as_lineno-$LINENO}: result: disabled" >&5
+$as_echo "disabled" >&6; }
+      else
+        # Check if the provided tool contains a complete path.
+        tool_specified="$DOT"
+        tool_basename="${tool_specified##*/}"
+        if test "x$tool_basename" = "x$tool_specified"; then
+          # A command without a complete path is provided, search $PATH.
+          { $as_echo "$as_me:${as_lineno-$LINENO}: Will search for user supplied tool DOT=$tool_basename" >&5
+$as_echo "$as_me: Will search for user supplied tool DOT=$tool_basename" >&6;}
+          # Extract the first word of "$tool_basename", so it can be a program name with args.
+set dummy $tool_basename; ac_word=$2
+{ $as_echo "$as_me:${as_lineno-$LINENO}: checking for $ac_word" >&5
+$as_echo_n "checking for $ac_word... " >&6; }
+if ${ac_cv_path_DOT+:} false; then :
+  $as_echo_n "(cached) " >&6
+else
+  case $DOT in
+  [\\/]* | ?:[\\/]*)
+  ac_cv_path_DOT="$DOT" # Let the user override the test with a path.
+  ;;
+  *)
+  as_save_IFS=$IFS; IFS=$PATH_SEPARATOR
+for as_dir in $PATH
+do
+  IFS=$as_save_IFS
+  test -z "$as_dir" && as_dir=.
+    for ac_exec_ext in '' $ac_executable_extensions; do
+  if as_fn_executable_p "$as_dir/$ac_word$ac_exec_ext"; then
+    ac_cv_path_DOT="$as_dir/$ac_word$ac_exec_ext"
+    $as_echo "$as_me:${as_lineno-$LINENO}: found $as_dir/$ac_word$ac_exec_ext" >&5
+    break 2
+  fi
+done
+  done
+IFS=$as_save_IFS
+
+  ;;
+esac
+fi
+DOT=$ac_cv_path_DOT
+if test -n "$DOT"; then
+  { $as_echo "$as_me:${as_lineno-$LINENO}: result: $DOT" >&5
+$as_echo "$DOT" >&6; }
+else
+  { $as_echo "$as_me:${as_lineno-$LINENO}: result: no" >&5
+$as_echo "no" >&6; }
+fi
+
+
+          if test "x$DOT" = x; then
+            as_fn_error $? "User supplied tool $tool_basename could not be found" "$LINENO" 5
+          fi
+        else
+          # Otherwise we believe it is a complete path. Use it as it is.
+          { $as_echo "$as_me:${as_lineno-$LINENO}: Will use user supplied tool DOT=$tool_specified" >&5
+$as_echo "$as_me: Will use user supplied tool DOT=$tool_specified" >&6;}
+          { $as_echo "$as_me:${as_lineno-$LINENO}: checking for DOT" >&5
+$as_echo_n "checking for DOT... " >&6; }
+          if test ! -x "$tool_specified"; then
+            { $as_echo "$as_me:${as_lineno-$LINENO}: result: not found" >&5
+$as_echo "not found" >&6; }
+            as_fn_error $? "User supplied tool DOT=$tool_specified does not exist or is not executable" "$LINENO" 5
+          fi
+          { $as_echo "$as_me:${as_lineno-$LINENO}: result: $tool_specified" >&5
+$as_echo "$tool_specified" >&6; }
+        fi
+      fi
+    fi
+
+  fi
+
+
+
+
+  # Publish this variable in the help.
+
+
   if [ -z "${HG+x}" ]; then
     # The variable is not set by user, try to locate tool using the code snippet
     for ac_prog in hg
@@ -24366,6 +24784,101 @@
 
 
 
+  # Should we build the complete docs, or just a lightweight version?
+  # Check whether --enable-full-docs was given.
+if test "${enable_full_docs+set}" = set; then :
+  enableval=$enable_full_docs;
+fi
+
+
+  # Verify dependencies
+  { $as_echo "$as_me:${as_lineno-$LINENO}: checking for graphviz dot" >&5
+$as_echo_n "checking for graphviz dot... " >&6; }
+  if test "x$DOT" != "x"; then
+    { $as_echo "$as_me:${as_lineno-$LINENO}: result: yes" >&5
+$as_echo "yes" >&6; }
+  else
+    { $as_echo "$as_me:${as_lineno-$LINENO}: result: no, cannot generate full docs" >&5
+$as_echo "no, cannot generate full docs" >&6; }
+    FULL_DOCS_DEP_MISSING=true
+  fi
+
+  { $as_echo "$as_me:${as_lineno-$LINENO}: checking for pandoc" >&5
+$as_echo_n "checking for pandoc... " >&6; }
+  if test "x$PANDOC" != "x"; then
+    { $as_echo "$as_me:${as_lineno-$LINENO}: result: yes" >&5
+$as_echo "yes" >&6; }
+  else
+    { $as_echo "$as_me:${as_lineno-$LINENO}: result: no, cannot generate full docs" >&5
+$as_echo "no, cannot generate full docs" >&6; }
+    FULL_DOCS_DEP_MISSING=true
+  fi
+
+  { $as_echo "$as_me:${as_lineno-$LINENO}: checking full docs" >&5
+$as_echo_n "checking full docs... " >&6; }
+  if test "x$enable_full_docs" = xyes; then
+    if test "x$FULL_DOCS_DEP_MISSING" = "xtrue"; then
+      { $as_echo "$as_me:${as_lineno-$LINENO}: result: no, missing dependencies" >&5
+$as_echo "no, missing dependencies" >&6; }
+
+  # Print a helpful message on how to acquire the necessary build dependency.
+  # dot is the help tag: freetype, cups, alsa etc
+  MISSING_DEPENDENCY=dot
+
+  if test "x$OPENJDK_BUILD_OS_ENV" = "xwindows.cygwin"; then
+    cygwin_help $MISSING_DEPENDENCY
+  elif test "x$OPENJDK_BUILD_OS_ENV" = "xwindows.msys"; then
+    msys_help $MISSING_DEPENDENCY
+  else
+    PKGHANDLER_COMMAND=
+
+    case $PKGHANDLER in
+      apt-get)
+        apt_help     $MISSING_DEPENDENCY ;;
+      yum)
+        yum_help     $MISSING_DEPENDENCY ;;
+      brew)
+        brew_help    $MISSING_DEPENDENCY ;;
+      port)
+        port_help    $MISSING_DEPENDENCY ;;
+      pkgutil)
+        pkgutil_help $MISSING_DEPENDENCY ;;
+      pkgadd)
+        pkgadd_help  $MISSING_DEPENDENCY ;;
+    esac
+
+    if test "x$PKGHANDLER_COMMAND" != x; then
+      HELP_MSG="You might be able to fix this by running '$PKGHANDLER_COMMAND'."
+    fi
+  fi
+
+      as_fn_error $? "Cannot enable full docs with missing dependencies. See above. $HELP_MSG" "$LINENO" 5
+    else
+      ENABLE_FULL_DOCS=true
+      { $as_echo "$as_me:${as_lineno-$LINENO}: result: yes, forced" >&5
+$as_echo "yes, forced" >&6; }
+    fi
+  elif test "x$enable_full_docs" = xno; then
+    ENABLE_FULL_DOCS=false
+    { $as_echo "$as_me:${as_lineno-$LINENO}: result: no, forced" >&5
+$as_echo "no, forced" >&6; }
+  elif test "x$enable_full_docs" = x; then
+    # Check for prerequisites
+    if test "x$FULL_DOCS_DEP_MISSING" = xtrue; then
+      ENABLE_FULL_DOCS=false
+      { $as_echo "$as_me:${as_lineno-$LINENO}: result: no, missing dependencies" >&5
+$as_echo "no, missing dependencies" >&6; }
+    else
+      ENABLE_FULL_DOCS=true
+      { $as_echo "$as_me:${as_lineno-$LINENO}: result: yes, dependencies present" >&5
+$as_echo "yes, dependencies present" >&6; }
+    fi
+  else
+    as_fn_error $? "--enable-full-docs can only take yes or no" "$LINENO" 5
+  fi
+
+
+
   # Choose cacerts source file
 
 # Check whether --with-cacerts-file was given.
@@ -33184,10 +33697,9 @@
     if test -n "$TOOLCHAIN_PATH"; then
       PATH_save="$PATH"
       PATH="$TOOLCHAIN_PATH"
-      for ac_prog in $SEARCH_LIST
-do
-  # Extract the first word of "$ac_prog", so it can be a program name with args.
-set dummy $ac_prog; ac_word=$2
+      if test -n "$ac_tool_prefix"; then
+  # Extract the first word of "${ac_tool_prefix}$SEARCH_LIST", so it can be a program name with args.
+set dummy ${ac_tool_prefix}$SEARCH_LIST; ac_word=$2
 { $as_echo "$as_me:${as_lineno-$LINENO}: checking for $ac_word" >&5
 $as_echo_n "checking for $ac_word... " >&6; }
 if ${ac_cv_path_TOOLCHAIN_PATH_CC+:} false; then :
@@ -33226,20 +33738,73 @@
 fi
 
 
-  test -n "$TOOLCHAIN_PATH_CC" && break
-done
+fi
+if test -z "$ac_cv_path_TOOLCHAIN_PATH_CC"; then
+  ac_pt_TOOLCHAIN_PATH_CC=$TOOLCHAIN_PATH_CC
+  # Extract the first word of "$SEARCH_LIST", so it can be a program name with args.
+set dummy $SEARCH_LIST; ac_word=$2
+{ $as_echo "$as_me:${as_lineno-$LINENO}: checking for $ac_word" >&5
+$as_echo_n "checking for $ac_word... " >&6; }
+if ${ac_cv_path_ac_pt_TOOLCHAIN_PATH_CC+:} false; then :
+  $as_echo_n "(cached) " >&6
+else
+  case $ac_pt_TOOLCHAIN_PATH_CC in
+  [\\/]* | ?:[\\/]*)
+  ac_cv_path_ac_pt_TOOLCHAIN_PATH_CC="$ac_pt_TOOLCHAIN_PATH_CC" # Let the user override the test with a path.
+  ;;
+  *)
+  as_save_IFS=$IFS; IFS=$PATH_SEPARATOR
+for as_dir in $PATH
+do
+  IFS=$as_save_IFS
+  test -z "$as_dir" && as_dir=.
+    for ac_exec_ext in '' $ac_executable_extensions; do
+  if as_fn_executable_p "$as_dir/$ac_word$ac_exec_ext"; then
+    ac_cv_path_ac_pt_TOOLCHAIN_PATH_CC="$as_dir/$ac_word$ac_exec_ext"
+    $as_echo "$as_me:${as_lineno-$LINENO}: found $as_dir/$ac_word$ac_exec_ext" >&5
+    break 2
+  fi
+done
+  done
+IFS=$as_save_IFS
+
+  ;;
+esac
+fi
+ac_pt_TOOLCHAIN_PATH_CC=$ac_cv_path_ac_pt_TOOLCHAIN_PATH_CC
+if test -n "$ac_pt_TOOLCHAIN_PATH_CC"; then
+  { $as_echo "$as_me:${as_lineno-$LINENO}: result: $ac_pt_TOOLCHAIN_PATH_CC" >&5
+$as_echo "$ac_pt_TOOLCHAIN_PATH_CC" >&6; }
+else
+  { $as_echo "$as_me:${as_lineno-$LINENO}: result: no" >&5
+$as_echo "no" >&6; }
+fi
+
+  if test "x$ac_pt_TOOLCHAIN_PATH_CC" = x; then
+    TOOLCHAIN_PATH_CC=""
+  else
+    case $cross_compiling:$ac_tool_warned in
+yes:)
+{ $as_echo "$as_me:${as_lineno-$LINENO}: WARNING: using cross tools not prefixed with host triplet" >&5
+$as_echo "$as_me: WARNING: using cross tools not prefixed with host triplet" >&2;}
+ac_tool_warned=yes ;;
+esac
+    TOOLCHAIN_PATH_CC=$ac_pt_TOOLCHAIN_PATH_CC
+  fi
+else
+  TOOLCHAIN_PATH_CC="$ac_cv_path_TOOLCHAIN_PATH_CC"
+fi
 
       CC=$TOOLCHAIN_PATH_CC
       PATH="$PATH_save"
     fi
 
-    # AC_PATH_PROGS can't be run multiple times with the same variable,
+    # AC_PATH_TOOL can't be run multiple times with the same variable,
     # so create a new name for this run.
     if test "x$CC" = x; then
-      for ac_prog in $SEARCH_LIST
-do
-  # Extract the first word of "$ac_prog", so it can be a program name with args.
-set dummy $ac_prog; ac_word=$2
+      if test -n "$ac_tool_prefix"; then
+  # Extract the first word of "${ac_tool_prefix}$SEARCH_LIST", so it can be a program name with args.
+set dummy ${ac_tool_prefix}$SEARCH_LIST; ac_word=$2
 { $as_echo "$as_me:${as_lineno-$LINENO}: checking for $ac_word" >&5
 $as_echo_n "checking for $ac_word... " >&6; }
 if ${ac_cv_path_POTENTIAL_CC+:} false; then :
@@ -33278,8 +33843,62 @@
 fi
 
 
-  test -n "$POTENTIAL_CC" && break
-done
+fi
+if test -z "$ac_cv_path_POTENTIAL_CC"; then
+  ac_pt_POTENTIAL_CC=$POTENTIAL_CC
+  # Extract the first word of "$SEARCH_LIST", so it can be a program name with args.
+set dummy $SEARCH_LIST; ac_word=$2
+{ $as_echo "$as_me:${as_lineno-$LINENO}: checking for $ac_word" >&5
+$as_echo_n "checking for $ac_word... " >&6; }
+if ${ac_cv_path_ac_pt_POTENTIAL_CC+:} false; then :
+  $as_echo_n "(cached) " >&6
+else
+  case $ac_pt_POTENTIAL_CC in
+  [\\/]* | ?:[\\/]*)
+  ac_cv_path_ac_pt_POTENTIAL_CC="$ac_pt_POTENTIAL_CC" # Let the user override the test with a path.
+  ;;
+  *)
+  as_save_IFS=$IFS; IFS=$PATH_SEPARATOR
+for as_dir in $PATH
+do
+  IFS=$as_save_IFS
+  test -z "$as_dir" && as_dir=.
+    for ac_exec_ext in '' $ac_executable_extensions; do
+  if as_fn_executable_p "$as_dir/$ac_word$ac_exec_ext"; then
+    ac_cv_path_ac_pt_POTENTIAL_CC="$as_dir/$ac_word$ac_exec_ext"
+    $as_echo "$as_me:${as_lineno-$LINENO}: found $as_dir/$ac_word$ac_exec_ext" >&5
+    break 2
+  fi
+done
+  done
+IFS=$as_save_IFS
+
+  ;;
+esac
+fi
+ac_pt_POTENTIAL_CC=$ac_cv_path_ac_pt_POTENTIAL_CC
+if test -n "$ac_pt_POTENTIAL_CC"; then
+  { $as_echo "$as_me:${as_lineno-$LINENO}: result: $ac_pt_POTENTIAL_CC" >&5
+$as_echo "$ac_pt_POTENTIAL_CC" >&6; }
+else
+  { $as_echo "$as_me:${as_lineno-$LINENO}: result: no" >&5
+$as_echo "no" >&6; }
+fi
+
+  if test "x$ac_pt_POTENTIAL_CC" = x; then
+    POTENTIAL_CC=""
+  else
+    case $cross_compiling:$ac_tool_warned in
+yes:)
+{ $as_echo "$as_me:${as_lineno-$LINENO}: WARNING: using cross tools not prefixed with host triplet" >&5
+$as_echo "$as_me: WARNING: using cross tools not prefixed with host triplet" >&2;}
+ac_tool_warned=yes ;;
+esac
+    POTENTIAL_CC=$ac_pt_POTENTIAL_CC
+  fi
+else
+  POTENTIAL_CC="$ac_cv_path_POTENTIAL_CC"
+fi
 
       CC=$POTENTIAL_CC
     fi
@@ -34485,10 +35104,9 @@
     if test -n "$TOOLCHAIN_PATH"; then
       PATH_save="$PATH"
       PATH="$TOOLCHAIN_PATH"
-      for ac_prog in $SEARCH_LIST
-do
-  # Extract the first word of "$ac_prog", so it can be a program name with args.
-set dummy $ac_prog; ac_word=$2
+      if test -n "$ac_tool_prefix"; then
+  # Extract the first word of "${ac_tool_prefix}$SEARCH_LIST", so it can be a program name with args.
+set dummy ${ac_tool_prefix}$SEARCH_LIST; ac_word=$2
 { $as_echo "$as_me:${as_lineno-$LINENO}: checking for $ac_word" >&5
 $as_echo_n "checking for $ac_word... " >&6; }
 if ${ac_cv_path_TOOLCHAIN_PATH_CXX+:} false; then :
@@ -34527,20 +35145,73 @@
 fi
 
 
-  test -n "$TOOLCHAIN_PATH_CXX" && break
-done
+fi
+if test -z "$ac_cv_path_TOOLCHAIN_PATH_CXX"; then
+  ac_pt_TOOLCHAIN_PATH_CXX=$TOOLCHAIN_PATH_CXX
+  # Extract the first word of "$SEARCH_LIST", so it can be a program name with args.
+set dummy $SEARCH_LIST; ac_word=$2
+{ $as_echo "$as_me:${as_lineno-$LINENO}: checking for $ac_word" >&5
+$as_echo_n "checking for $ac_word... " >&6; }
+if ${ac_cv_path_ac_pt_TOOLCHAIN_PATH_CXX+:} false; then :
+  $as_echo_n "(cached) " >&6
+else
+  case $ac_pt_TOOLCHAIN_PATH_CXX in
+  [\\/]* | ?:[\\/]*)
+  ac_cv_path_ac_pt_TOOLCHAIN_PATH_CXX="$ac_pt_TOOLCHAIN_PATH_CXX" # Let the user override the test with a path.
+  ;;
+  *)
+  as_save_IFS=$IFS; IFS=$PATH_SEPARATOR
+for as_dir in $PATH
+do
+  IFS=$as_save_IFS
+  test -z "$as_dir" && as_dir=.
+    for ac_exec_ext in '' $ac_executable_extensions; do
+  if as_fn_executable_p "$as_dir/$ac_word$ac_exec_ext"; then
+    ac_cv_path_ac_pt_TOOLCHAIN_PATH_CXX="$as_dir/$ac_word$ac_exec_ext"
+    $as_echo "$as_me:${as_lineno-$LINENO}: found $as_dir/$ac_word$ac_exec_ext" >&5
+    break 2
+  fi
+done
+  done
+IFS=$as_save_IFS
+
+  ;;
+esac
+fi
+ac_pt_TOOLCHAIN_PATH_CXX=$ac_cv_path_ac_pt_TOOLCHAIN_PATH_CXX
+if test -n "$ac_pt_TOOLCHAIN_PATH_CXX"; then
+  { $as_echo "$as_me:${as_lineno-$LINENO}: result: $ac_pt_TOOLCHAIN_PATH_CXX" >&5
+$as_echo "$ac_pt_TOOLCHAIN_PATH_CXX" >&6; }
+else
+  { $as_echo "$as_me:${as_lineno-$LINENO}: result: no" >&5
+$as_echo "no" >&6; }
+fi
+
+  if test "x$ac_pt_TOOLCHAIN_PATH_CXX" = x; then
+    TOOLCHAIN_PATH_CXX=""
+  else
+    case $cross_compiling:$ac_tool_warned in
+yes:)
+{ $as_echo "$as_me:${as_lineno-$LINENO}: WARNING: using cross tools not prefixed with host triplet" >&5
+$as_echo "$as_me: WARNING: using cross tools not prefixed with host triplet" >&2;}
+ac_tool_warned=yes ;;
+esac
+    TOOLCHAIN_PATH_CXX=$ac_pt_TOOLCHAIN_PATH_CXX
+  fi
+else
+  TOOLCHAIN_PATH_CXX="$ac_cv_path_TOOLCHAIN_PATH_CXX"
+fi
 
       CXX=$TOOLCHAIN_PATH_CXX
       PATH="$PATH_save"
     fi
 
-    # AC_PATH_PROGS can't be run multiple times with the same variable,
+    # AC_PATH_TOOL can't be run multiple times with the same variable,
     # so create a new name for this run.
     if test "x$CXX" = x; then
-      for ac_prog in $SEARCH_LIST
-do
-  # Extract the first word of "$ac_prog", so it can be a program name with args.
-set dummy $ac_prog; ac_word=$2
+      if test -n "$ac_tool_prefix"; then
+  # Extract the first word of "${ac_tool_prefix}$SEARCH_LIST", so it can be a program name with args.
+set dummy ${ac_tool_prefix}$SEARCH_LIST; ac_word=$2
 { $as_echo "$as_me:${as_lineno-$LINENO}: checking for $ac_word" >&5
 $as_echo_n "checking for $ac_word... " >&6; }
 if ${ac_cv_path_POTENTIAL_CXX+:} false; then :
@@ -34579,8 +35250,62 @@
 fi
 
 
-  test -n "$POTENTIAL_CXX" && break
-done
+fi
+if test -z "$ac_cv_path_POTENTIAL_CXX"; then
+  ac_pt_POTENTIAL_CXX=$POTENTIAL_CXX
+  # Extract the first word of "$SEARCH_LIST", so it can be a program name with args.
+set dummy $SEARCH_LIST; ac_word=$2
+{ $as_echo "$as_me:${as_lineno-$LINENO}: checking for $ac_word" >&5
+$as_echo_n "checking for $ac_word... " >&6; }
+if ${ac_cv_path_ac_pt_POTENTIAL_CXX+:} false; then :
+  $as_echo_n "(cached) " >&6
+else
+  case $ac_pt_POTENTIAL_CXX in
+  [\\/]* | ?:[\\/]*)
+  ac_cv_path_ac_pt_POTENTIAL_CXX="$ac_pt_POTENTIAL_CXX" # Let the user override the test with a path.
+  ;;
+  *)
+  as_save_IFS=$IFS; IFS=$PATH_SEPARATOR
+for as_dir in $PATH
+do
+  IFS=$as_save_IFS
+  test -z "$as_dir" && as_dir=.
+    for ac_exec_ext in '' $ac_executable_extensions; do
+  if as_fn_executable_p "$as_dir/$ac_word$ac_exec_ext"; then
+    ac_cv_path_ac_pt_POTENTIAL_CXX="$as_dir/$ac_word$ac_exec_ext"
+    $as_echo "$as_me:${as_lineno-$LINENO}: found $as_dir/$ac_word$ac_exec_ext" >&5
+    break 2
+  fi
+done
+  done
+IFS=$as_save_IFS
+
+  ;;
+esac
+fi
+ac_pt_POTENTIAL_CXX=$ac_cv_path_ac_pt_POTENTIAL_CXX
+if test -n "$ac_pt_POTENTIAL_CXX"; then
+  { $as_echo "$as_me:${as_lineno-$LINENO}: result: $ac_pt_POTENTIAL_CXX" >&5
+$as_echo "$ac_pt_POTENTIAL_CXX" >&6; }
+else
+  { $as_echo "$as_me:${as_lineno-$LINENO}: result: no" >&5
+$as_echo "no" >&6; }
+fi
+
+  if test "x$ac_pt_POTENTIAL_CXX" = x; then
+    POTENTIAL_CXX=""
+  else
+    case $cross_compiling:$ac_tool_warned in
+yes:)
+{ $as_echo "$as_me:${as_lineno-$LINENO}: WARNING: using cross tools not prefixed with host triplet" >&5
+$as_echo "$as_me: WARNING: using cross tools not prefixed with host triplet" >&2;}
+ac_tool_warned=yes ;;
+esac
+    POTENTIAL_CXX=$ac_pt_POTENTIAL_CXX
+  fi
+else
+  POTENTIAL_CXX="$ac_cv_path_POTENTIAL_CXX"
+fi
 
       CXX=$POTENTIAL_CXX
     fi
@@ -47935,173 +48660,63 @@
 # Check whether --with-jtreg was given.
 if test "${with_jtreg+set}" = set; then :
   withval=$with_jtreg;
-else
-  with_jtreg=no
 fi
 
 
   if test "x$with_jtreg" = xno; then
     # jtreg disabled
-    { $as_echo "$as_me:${as_lineno-$LINENO}: checking for jtreg" >&5
-$as_echo_n "checking for jtreg... " >&6; }
-    { $as_echo "$as_me:${as_lineno-$LINENO}: result: no" >&5
-$as_echo "no" >&6; }
-  else
-    if test "x$with_jtreg" != xyes; then
-      # with path specified.
-      JT_HOME="$with_jtreg"
-    fi
-
+    { $as_echo "$as_me:${as_lineno-$LINENO}: checking for jtreg test harness" >&5
+$as_echo_n "checking for jtreg test harness... " >&6; }
+    { $as_echo "$as_me:${as_lineno-$LINENO}: result: no, disabled" >&5
+$as_echo "no, disabled" >&6; }
+  elif test "x$with_jtreg" != xyes && test "x$with_jtreg" != x; then
+    # An explicit path is specified, use it.
+    JT_HOME="$with_jtreg"
+    if test ! -d "$JT_HOME"; then
+      as_fn_error $? "jtreg home directory from --with-jtreg=$with_jtreg does not exist" "$LINENO" 5
+    fi
+
+    if test ! -e "$JT_HOME/lib/jtreg.jar"; then
+      as_fn_error $? "jtreg home directory from --with-jtreg=$with_jtreg is not a valid jtreg home" "$LINENO" 5
+    fi
+
+    JTREGEXE="$JT_HOME/bin/jtreg"
+    if test ! -x "$JTREGEXE"; then
+      as_fn_error $? "jtreg home directory from --with-jtreg=$with_jtreg does not contain valid jtreg executable" "$LINENO" 5
+    fi
+
+    { $as_echo "$as_me:${as_lineno-$LINENO}: checking for jtreg test harness" >&5
+$as_echo_n "checking for jtreg test harness... " >&6; }
+    { $as_echo "$as_me:${as_lineno-$LINENO}: result: $JT_HOME" >&5
+$as_echo "$JT_HOME" >&6; }
+  else
+    # Try to locate jtreg
     if test "x$JT_HOME" != x; then
-      { $as_echo "$as_me:${as_lineno-$LINENO}: checking for jtreg" >&5
-$as_echo_n "checking for jtreg... " >&6; }
-
-      # use JT_HOME enviroment var.
-
-  # Only process if variable expands to non-empty
-
-  if test "x$JT_HOME" != x; then
-    if test "x$OPENJDK_BUILD_OS_ENV" = "xwindows.cygwin"; then
-
-  # Input might be given as Windows format, start by converting to
-  # unix format.
-  path="$JT_HOME"
-  new_path=`$CYGPATH -u "$path"`
-
-  # Cygwin tries to hide some aspects of the Windows file system, such that binaries are
-  # named .exe but called without that suffix. Therefore, "foo" and "foo.exe" are considered
-  # the same file, most of the time (as in "test -f"). But not when running cygpath -s, then
-  # "foo.exe" is OK but "foo" is an error.
-  #
-  # This test is therefore slightly more accurate than "test -f" to check for file precense.
-  # It is also a way to make sure we got the proper file name for the real test later on.
-  test_shortpath=`$CYGPATH -s -m "$new_path" 2> /dev/null`
-  if test "x$test_shortpath" = x; then
-    { $as_echo "$as_me:${as_lineno-$LINENO}: The path of JT_HOME, which resolves as \"$path\", is invalid." >&5
-$as_echo "$as_me: The path of JT_HOME, which resolves as \"$path\", is invalid." >&6;}
-    as_fn_error $? "Cannot locate the the path of JT_HOME" "$LINENO" 5
-  fi
-
-  # Call helper function which possibly converts this using DOS-style short mode.
-  # If so, the updated path is stored in $new_path.
-
-  input_path="$new_path"
-  # Check if we need to convert this using DOS-style short mode. If the path
-  # contains just simple characters, use it. Otherwise (spaces, weird characters),
-  # take no chances and rewrite it.
-  # Note: m4 eats our [], so we need to use [ and ] instead.
-  has_forbidden_chars=`$ECHO "$input_path" | $GREP [^-._/a-zA-Z0-9]`
-  if test "x$has_forbidden_chars" != x; then
-    # Now convert it to mixed DOS-style, short mode (no spaces, and / instead of \)
-    shortmode_path=`$CYGPATH -s -m -a "$input_path"`
-    path_after_shortmode=`$CYGPATH -u "$shortmode_path"`
-    if test "x$path_after_shortmode" != "x$input_to_shortpath"; then
-      # Going to short mode and back again did indeed matter. Since short mode is
-      # case insensitive, let's make it lowercase to improve readability.
-      shortmode_path=`$ECHO "$shortmode_path" | $TR 'ABCDEFGHIJKLMNOPQRSTUVWXYZ' 'abcdefghijklmnopqrstuvwxyz'`
-      # Now convert it back to Unix-style (cygpath)
-      input_path=`$CYGPATH -u "$shortmode_path"`
-      new_path="$input_path"
-    fi
-  fi
-
-  test_cygdrive_prefix=`$ECHO $input_path | $GREP ^/cygdrive/`
-  if test "x$test_cygdrive_prefix" = x; then
-    # As a simple fix, exclude /usr/bin since it's not a real path.
-    if test "x`$ECHO $new_path | $GREP ^/usr/bin/`" = x; then
-      # The path is in a Cygwin special directory (e.g. /home). We need this converted to
-      # a path prefixed by /cygdrive for fixpath to work.
-      new_path="$CYGWIN_ROOT_PATH$input_path"
-    fi
-  fi
-
-
-  if test "x$path" != "x$new_path"; then
-    JT_HOME="$new_path"
-    { $as_echo "$as_me:${as_lineno-$LINENO}: Rewriting JT_HOME to \"$new_path\"" >&5
-$as_echo "$as_me: Rewriting JT_HOME to \"$new_path\"" >&6;}
-  fi
-
-    elif test "x$OPENJDK_BUILD_OS_ENV" = "xwindows.msys"; then
-
-  path="$JT_HOME"
-  has_colon=`$ECHO $path | $GREP ^.:`
-  new_path="$path"
-  if test "x$has_colon" = x; then
-    # Not in mixed or Windows style, start by that.
-    new_path=`cmd //c echo $path`
-  fi
-
-
-  input_path="$new_path"
-  # Check if we need to convert this using DOS-style short mode. If the path
-  # contains just simple characters, use it. Otherwise (spaces, weird characters),
-  # take no chances and rewrite it.
-  # Note: m4 eats our [], so we need to use [ and ] instead.
-  has_forbidden_chars=`$ECHO "$input_path" | $GREP [^-_/:a-zA-Z0-9]`
-  if test "x$has_forbidden_chars" != x; then
-    # Now convert it to mixed DOS-style, short mode (no spaces, and / instead of \)
-    new_path=`cmd /c "for %A in (\"$input_path\") do @echo %~sA"|$TR \\\\\\\\ / | $TR 'ABCDEFGHIJKLMNOPQRSTUVWXYZ' 'abcdefghijklmnopqrstuvwxyz'`
-  fi
-
-
-  windows_path="$new_path"
-  if test "x$OPENJDK_BUILD_OS_ENV" = "xwindows.cygwin"; then
-    unix_path=`$CYGPATH -u "$windows_path"`
-    new_path="$unix_path"
-  elif test "x$OPENJDK_BUILD_OS_ENV" = "xwindows.msys"; then
-    unix_path=`$ECHO "$windows_path" | $SED -e 's,^\\(.\\):,/\\1,g' -e 's,\\\\,/,g'`
-    new_path="$unix_path"
-  fi
-
-  if test "x$path" != "x$new_path"; then
-    JT_HOME="$new_path"
-    { $as_echo "$as_me:${as_lineno-$LINENO}: Rewriting JT_HOME to \"$new_path\"" >&5
-$as_echo "$as_me: Rewriting JT_HOME to \"$new_path\"" >&6;}
-  fi
-
-  # Save the first 10 bytes of this path to the storage, so fixpath can work.
-  all_fixpath_prefixes=("${all_fixpath_prefixes[@]}" "${new_path:0:10}")
-
-    else
-      # We're on a unix platform. Hooray! :)
-      path="$JT_HOME"
-      has_space=`$ECHO "$path" | $GREP " "`
-      if test "x$has_space" != x; then
-        { $as_echo "$as_me:${as_lineno-$LINENO}: The path of JT_HOME, which resolves as \"$path\", is invalid." >&5
-$as_echo "$as_me: The path of JT_HOME, which resolves as \"$path\", is invalid." >&6;}
-        as_fn_error $? "Spaces are not allowed in this path." "$LINENO" 5
-      fi
-
-      # Use eval to expand a potential ~
-      eval path="$path"
-      if test ! -f "$path" && test ! -d "$path"; then
-        as_fn_error $? "The path of JT_HOME, which resolves as \"$path\", is not found." "$LINENO" 5
-      fi
-
-      if test -d "$path"; then
-        JT_HOME="`cd "$path"; $THEPWDCMD -L`"
-      else
-        dir="`$DIRNAME "$path"`"
-        base="`$BASENAME "$path"`"
-        JT_HOME="`cd "$dir"; $THEPWDCMD -L`/$base"
-      fi
-    fi
-  fi
-
-
-      # jtreg win32 script works for everybody
-      JTREGEXE="$JT_HOME/bin/jtreg"
-
-      if test ! -f "$JTREGEXE"; then
-        as_fn_error $? "JTReg executable does not exist: $JTREGEXE" "$LINENO" 5
-      fi
-
-      { $as_echo "$as_me:${as_lineno-$LINENO}: result: $JTREGEXE" >&5
-$as_echo "$JTREGEXE" >&6; }
-    else
-      # try to find jtreg on path
-
+      # JT_HOME set in environment, use it
+      if test ! -d "$JT_HOME"; then
+        { $as_echo "$as_me:${as_lineno-$LINENO}: WARNING: Ignoring JT_HOME pointing to invalid directory: $JT_HOME" >&5
+$as_echo "$as_me: WARNING: Ignoring JT_HOME pointing to invalid directory: $JT_HOME" >&2;}
+        JT_HOME=
+      else
+        if test ! -e "$JT_HOME/lib/jtreg.jar"; then
+          { $as_echo "$as_me:${as_lineno-$LINENO}: WARNING: Ignoring JT_HOME which is not a valid jtreg home: $JT_HOME" >&5
+$as_echo "$as_me: WARNING: Ignoring JT_HOME which is not a valid jtreg home: $JT_HOME" >&2;}
+          JT_HOME=
+        elif test ! -x "$JT_HOME/bin/jtreg"; then
+          { $as_echo "$as_me:${as_lineno-$LINENO}: WARNING: Ignoring JT_HOME which does not contain valid jtreg executable: $JT_HOME" >&5
+$as_echo "$as_me: WARNING: Ignoring JT_HOME which does not contain valid jtreg executable: $JT_HOME" >&2;}
+          JT_HOME=
+        else
+          JTREGEXE="$JT_HOME/bin/jtreg"
+          { $as_echo "$as_me:${as_lineno-$LINENO}: Located jtreg using JT_HOME from environment" >&5
+$as_echo "$as_me: Located jtreg using JT_HOME from environment" >&6;}
+        fi
+      fi
+    fi
+
+    if test "x$JT_HOME" = x; then
+      # JT_HOME is not set in environment, or was deemed invalid.
+      # Try to find jtreg on path
 
 
   # Publish this variable in the help.
@@ -48299,13 +48914,451 @@
   fi
 
 
-
-  if test "x$JTREGEXE" = x; then
-    as_fn_error $? "Could not find required tool for JTREGEXE" "$LINENO" 5
-  fi
-
-
-      JT_HOME="`$DIRNAME $JTREGEXE`"
+      if test "x$JTREGEXE" != x; then
+        # That's good, now try to derive JT_HOME
+        JT_HOME=`(cd $($DIRNAME $JTREGEXE)/.. && pwd)`
+        if test ! -e "$JT_HOME/lib/jtreg.jar"; then
+          { $as_echo "$as_me:${as_lineno-$LINENO}: WARNING: Ignoring jtreg from path since a valid jtreg home cannot be found" >&5
+$as_echo "$as_me: WARNING: Ignoring jtreg from path since a valid jtreg home cannot be found" >&2;}
+          JT_HOME=
+          JTREGEXE=
+        else
+          { $as_echo "$as_me:${as_lineno-$LINENO}: Located jtreg using jtreg executable in path" >&5
+$as_echo "$as_me: Located jtreg using jtreg executable in path" >&6;}
+        fi
+      fi
+    fi
+
+    { $as_echo "$as_me:${as_lineno-$LINENO}: checking for jtreg test harness" >&5
+$as_echo_n "checking for jtreg test harness... " >&6; }
+    if test "x$JT_HOME" != x; then
+      { $as_echo "$as_me:${as_lineno-$LINENO}: result: $JT_HOME" >&5
+$as_echo "$JT_HOME" >&6; }
+    else
+      { $as_echo "$as_me:${as_lineno-$LINENO}: result: no, not found" >&5
+$as_echo "no, not found" >&6; }
+
+      if test "x$with_jtreg" = xyes; then
+        as_fn_error $? "--with-jtreg was specified, but no jtreg found." "$LINENO" 5
+      fi
+    fi
+  fi
+
+
+  # Only process if variable expands to non-empty
+
+  if test "x$JTREGEXE" != x; then
+    if test "x$OPENJDK_BUILD_OS_ENV" = "xwindows.cygwin"; then
+
+  # First separate the path from the arguments. This will split at the first
+  # space.
+  complete="$JTREGEXE"
+  path="${complete%% *}"
+  tmp="$complete EOL"
+  arguments="${tmp#* }"
+
+  # Input might be given as Windows format, start by converting to
+  # unix format.
+  new_path=`$CYGPATH -u "$path"`
+
+  # Now try to locate executable using which
+  new_path=`$WHICH "$new_path" 2> /dev/null`
+  # bat and cmd files are not always considered executable in cygwin causing which
+  # to not find them
+  if test "x$new_path" = x \
+      && test "x`$ECHO \"$path\" | $GREP -i -e \"\\.bat$\" -e \"\\.cmd$\"`" != x \
+      && test "x`$LS \"$path\" 2>/dev/null`" != x; then
+    new_path=`$CYGPATH -u "$path"`
+  fi
+  if test "x$new_path" = x; then
+    # Oops. Which didn't find the executable.
+    # The splitting of arguments from the executable at a space might have been incorrect,
+    # since paths with space are more likely in Windows. Give it another try with the whole
+    # argument.
+    path="$complete"
+    arguments="EOL"
+    new_path=`$CYGPATH -u "$path"`
+    new_path=`$WHICH "$new_path" 2> /dev/null`
+    # bat and cmd files are not always considered executable in cygwin causing which
+    # to not find them
+    if test "x$new_path" = x \
+        && test "x`$ECHO \"$path\" | $GREP -i -e \"\\.bat$\" -e \"\\.cmd$\"`" != x \
+        && test "x`$LS \"$path\" 2>/dev/null`" != x; then
+      new_path=`$CYGPATH -u "$path"`
+    fi
+    if test "x$new_path" = x; then
+      # It's still not found. Now this is an unrecoverable error.
+      { $as_echo "$as_me:${as_lineno-$LINENO}: The path of JTREGEXE, which resolves as \"$complete\", is not found." >&5
+$as_echo "$as_me: The path of JTREGEXE, which resolves as \"$complete\", is not found." >&6;}
+      has_space=`$ECHO "$complete" | $GREP " "`
+      if test "x$has_space" != x; then
+        { $as_echo "$as_me:${as_lineno-$LINENO}: You might be mixing spaces in the path and extra arguments, which is not allowed." >&5
+$as_echo "$as_me: You might be mixing spaces in the path and extra arguments, which is not allowed." >&6;}
+      fi
+      as_fn_error $? "Cannot locate the the path of JTREGEXE" "$LINENO" 5
+    fi
+  fi
+
+  # Cygwin tries to hide some aspects of the Windows file system, such that binaries are
+  # named .exe but called without that suffix. Therefore, "foo" and "foo.exe" are considered
+  # the same file, most of the time (as in "test -f"). But not when running cygpath -s, then
+  # "foo.exe" is OK but "foo" is an error.
+  #
+  # This test is therefore slightly more accurate than "test -f" to check for file presence.
+  # It is also a way to make sure we got the proper file name for the real test later on.
+  test_shortpath=`$CYGPATH -s -m "$new_path" 2> /dev/null`
+  if test "x$test_shortpath" = x; then
+    # Short path failed, file does not exist as specified.
+    # Try adding .exe or .cmd
+    if test -f "${new_path}.exe"; then
+      input_to_shortpath="${new_path}.exe"
+    elif test -f "${new_path}.cmd"; then
+      input_to_shortpath="${new_path}.cmd"
+    else
+      { $as_echo "$as_me:${as_lineno-$LINENO}: The path of JTREGEXE, which resolves as \"$new_path\", is invalid." >&5
+$as_echo "$as_me: The path of JTREGEXE, which resolves as \"$new_path\", is invalid." >&6;}
+      { $as_echo "$as_me:${as_lineno-$LINENO}: Neither \"$new_path\" nor \"$new_path.exe/cmd\" can be found" >&5
+$as_echo "$as_me: Neither \"$new_path\" nor \"$new_path.exe/cmd\" can be found" >&6;}
+      as_fn_error $? "Cannot locate the the path of JTREGEXE" "$LINENO" 5
+    fi
+  else
+    input_to_shortpath="$new_path"
+  fi
+
+  # Call helper function which possibly converts this using DOS-style short mode.
+  # If so, the updated path is stored in $new_path.
+  new_path="$input_to_shortpath"
+
+  input_path="$input_to_shortpath"
+  # Check if we need to convert this using DOS-style short mode. If the path
+  # contains just simple characters, use it. Otherwise (spaces, weird characters),
+  # take no chances and rewrite it.
+  # Note: m4 eats our [], so we need to use [ and ] instead.
+  has_forbidden_chars=`$ECHO "$input_path" | $GREP [^-._/a-zA-Z0-9]`
+  if test "x$has_forbidden_chars" != x; then
+    # Now convert it to mixed DOS-style, short mode (no spaces, and / instead of \)
+    shortmode_path=`$CYGPATH -s -m -a "$input_path"`
+    path_after_shortmode=`$CYGPATH -u "$shortmode_path"`
+    if test "x$path_after_shortmode" != "x$input_to_shortpath"; then
+      # Going to short mode and back again did indeed matter. Since short mode is
+      # case insensitive, let's make it lowercase to improve readability.
+      shortmode_path=`$ECHO "$shortmode_path" | $TR 'ABCDEFGHIJKLMNOPQRSTUVWXYZ' 'abcdefghijklmnopqrstuvwxyz'`
+      # Now convert it back to Unix-style (cygpath)
+      input_path=`$CYGPATH -u "$shortmode_path"`
+      new_path="$input_path"
+    fi
+  fi
+
+  test_cygdrive_prefix=`$ECHO $input_path | $GREP ^/cygdrive/`
+  if test "x$test_cygdrive_prefix" = x; then
+    # As a simple fix, exclude /usr/bin since it's not a real path.
+    if test "x`$ECHO $input_to_shortpath | $GREP ^/usr/bin/`" = x; then
+      # The path is in a Cygwin special directory (e.g. /home). We need this converted to
+      # a path prefixed by /cygdrive for fixpath to work.
+      new_path="$CYGWIN_ROOT_PATH$input_path"
+    fi
+  fi
+
+  # remove trailing .exe if any
+  new_path="${new_path/%.exe/}"
+
+    elif test "x$OPENJDK_BUILD_OS_ENV" = "xwindows.msys"; then
+
+  # First separate the path from the arguments. This will split at the first
+  # space.
+  complete="$JTREGEXE"
+  path="${complete%% *}"
+  tmp="$complete EOL"
+  arguments="${tmp#* }"
+
+  # Input might be given as Windows format, start by converting to
+  # unix format.
+  new_path="$path"
+
+  windows_path="$new_path"
+  if test "x$OPENJDK_BUILD_OS_ENV" = "xwindows.cygwin"; then
+    unix_path=`$CYGPATH -u "$windows_path"`
+    new_path="$unix_path"
+  elif test "x$OPENJDK_BUILD_OS_ENV" = "xwindows.msys"; then
+    unix_path=`$ECHO "$windows_path" | $SED -e 's,^\\(.\\):,/\\1,g' -e 's,\\\\,/,g'`
+    new_path="$unix_path"
+  fi
+
+
+  # Now try to locate executable using which
+  new_path=`$WHICH "$new_path" 2> /dev/null`
+
+  if test "x$new_path" = x; then
+    # Oops. Which didn't find the executable.
+    # The splitting of arguments from the executable at a space might have been incorrect,
+    # since paths with space are more likely in Windows. Give it another try with the whole
+    # argument.
+    path="$complete"
+    arguments="EOL"
+    new_path="$path"
+
+  windows_path="$new_path"
+  if test "x$OPENJDK_BUILD_OS_ENV" = "xwindows.cygwin"; then
+    unix_path=`$CYGPATH -u "$windows_path"`
+    new_path="$unix_path"
+  elif test "x$OPENJDK_BUILD_OS_ENV" = "xwindows.msys"; then
+    unix_path=`$ECHO "$windows_path" | $SED -e 's,^\\(.\\):,/\\1,g' -e 's,\\\\,/,g'`
+    new_path="$unix_path"
+  fi
+
+
+    new_path=`$WHICH "$new_path" 2> /dev/null`
+    # bat and cmd files are not always considered executable in MSYS causing which
+    # to not find them
+    if test "x$new_path" = x \
+        && test "x`$ECHO \"$path\" | $GREP -i -e \"\\.bat$\" -e \"\\.cmd$\"`" != x \
+        && test "x`$LS \"$path\" 2>/dev/null`" != x; then
+      new_path="$path"
+
+  windows_path="$new_path"
+  if test "x$OPENJDK_BUILD_OS_ENV" = "xwindows.cygwin"; then
+    unix_path=`$CYGPATH -u "$windows_path"`
+    new_path="$unix_path"
+  elif test "x$OPENJDK_BUILD_OS_ENV" = "xwindows.msys"; then
+    unix_path=`$ECHO "$windows_path" | $SED -e 's,^\\(.\\):,/\\1,g' -e 's,\\\\,/,g'`
+    new_path="$unix_path"
+  fi
+
+    fi
+
+    if test "x$new_path" = x; then
+      # It's still not found. Now this is an unrecoverable error.
+      { $as_echo "$as_me:${as_lineno-$LINENO}: The path of JTREGEXE, which resolves as \"$complete\", is not found." >&5
+$as_echo "$as_me: The path of JTREGEXE, which resolves as \"$complete\", is not found." >&6;}
+      has_space=`$ECHO "$complete" | $GREP " "`
+      if test "x$has_space" != x; then
+        { $as_echo "$as_me:${as_lineno-$LINENO}: You might be mixing spaces in the path and extra arguments, which is not allowed." >&5
+$as_echo "$as_me: You might be mixing spaces in the path and extra arguments, which is not allowed." >&6;}
+      fi
+      as_fn_error $? "Cannot locate the the path of JTREGEXE" "$LINENO" 5
+    fi
+  fi
+
+  # Now new_path has a complete unix path to the binary
+  if test "x`$ECHO $new_path | $GREP ^/bin/`" != x; then
+    # Keep paths in /bin as-is, but remove trailing .exe if any
+    new_path="${new_path/%.exe/}"
+    # Do not save /bin paths to all_fixpath_prefixes!
+  else
+    # Not in mixed or Windows style, start by that.
+    new_path=`cmd //c echo $new_path`
+
+  input_path="$new_path"
+  # Check if we need to convert this using DOS-style short mode. If the path
+  # contains just simple characters, use it. Otherwise (spaces, weird characters),
+  # take no chances and rewrite it.
+  # Note: m4 eats our [], so we need to use [ and ] instead.
+  has_forbidden_chars=`$ECHO "$input_path" | $GREP [^-_/:a-zA-Z0-9]`
+  if test "x$has_forbidden_chars" != x; then
+    # Now convert it to mixed DOS-style, short mode (no spaces, and / instead of \)
+    new_path=`cmd /c "for %A in (\"$input_path\") do @echo %~sA"|$TR \\\\\\\\ / | $TR 'ABCDEFGHIJKLMNOPQRSTUVWXYZ' 'abcdefghijklmnopqrstuvwxyz'`
+  fi
+
+    # Output is in $new_path
+
+  windows_path="$new_path"
+  if test "x$OPENJDK_BUILD_OS_ENV" = "xwindows.cygwin"; then
+    unix_path=`$CYGPATH -u "$windows_path"`
+    new_path="$unix_path"
+  elif test "x$OPENJDK_BUILD_OS_ENV" = "xwindows.msys"; then
+    unix_path=`$ECHO "$windows_path" | $SED -e 's,^\\(.\\):,/\\1,g' -e 's,\\\\,/,g'`
+    new_path="$unix_path"
+  fi
+
+    # remove trailing .exe if any
+    new_path="${new_path/%.exe/}"
+
+    # Save the first 10 bytes of this path to the storage, so fixpath can work.
+    all_fixpath_prefixes=("${all_fixpath_prefixes[@]}" "${new_path:0:10}")
+  fi
+
+    else
+      # We're on a unix platform. Hooray! :)
+      # First separate the path from the arguments. This will split at the first
+      # space.
+      complete="$JTREGEXE"
+      path="${complete%% *}"
+      tmp="$complete EOL"
+      arguments="${tmp#* }"
+
+      # Cannot rely on the command "which" here since it doesn't always work.
+      is_absolute_path=`$ECHO "$path" | $GREP ^/`
+      if test -z "$is_absolute_path"; then
+        # Path to executable is not absolute. Find it.
+        IFS_save="$IFS"
+        IFS=:
+        for p in $PATH; do
+          if test -f "$p/$path" && test -x "$p/$path"; then
+            new_path="$p/$path"
+            break
+          fi
+        done
+        IFS="$IFS_save"
+      else
+        # This is an absolute path, we can use it without further modifications.
+        new_path="$path"
+      fi
+
+      if test "x$new_path" = x; then
+        { $as_echo "$as_me:${as_lineno-$LINENO}: The path of JTREGEXE, which resolves as \"$complete\", is not found." >&5
+$as_echo "$as_me: The path of JTREGEXE, which resolves as \"$complete\", is not found." >&6;}
+        has_space=`$ECHO "$complete" | $GREP " "`
+        if test "x$has_space" != x; then
+          { $as_echo "$as_me:${as_lineno-$LINENO}: This might be caused by spaces in the path, which is not allowed." >&5
+$as_echo "$as_me: This might be caused by spaces in the path, which is not allowed." >&6;}
+        fi
+        as_fn_error $? "Cannot locate the the path of JTREGEXE" "$LINENO" 5
+      fi
+    fi
+
+    # Now join together the path and the arguments once again
+    if test "x$arguments" != xEOL; then
+      new_complete="$new_path ${arguments% *}"
+    else
+      new_complete="$new_path"
+    fi
+
+    if test "x$complete" != "x$new_complete"; then
+      JTREGEXE="$new_complete"
+      { $as_echo "$as_me:${as_lineno-$LINENO}: Rewriting JTREGEXE to \"$new_complete\"" >&5
+$as_echo "$as_me: Rewriting JTREGEXE to \"$new_complete\"" >&6;}
+    fi
+  fi
+
+
+  # Only process if variable expands to non-empty
+
+  if test "x$JT_HOME" != x; then
+    if test "x$OPENJDK_BUILD_OS_ENV" = "xwindows.cygwin"; then
+
+  # Input might be given as Windows format, start by converting to
+  # unix format.
+  path="$JT_HOME"
+  new_path=`$CYGPATH -u "$path"`
+
+  # Cygwin tries to hide some aspects of the Windows file system, such that binaries are
+  # named .exe but called without that suffix. Therefore, "foo" and "foo.exe" are considered
+  # the same file, most of the time (as in "test -f"). But not when running cygpath -s, then
+  # "foo.exe" is OK but "foo" is an error.
+  #
+  # This test is therefore slightly more accurate than "test -f" to check for file precense.
+  # It is also a way to make sure we got the proper file name for the real test later on.
+  test_shortpath=`$CYGPATH -s -m "$new_path" 2> /dev/null`
+  if test "x$test_shortpath" = x; then
+    { $as_echo "$as_me:${as_lineno-$LINENO}: The path of JT_HOME, which resolves as \"$path\", is invalid." >&5
+$as_echo "$as_me: The path of JT_HOME, which resolves as \"$path\", is invalid." >&6;}
+    as_fn_error $? "Cannot locate the the path of JT_HOME" "$LINENO" 5
+  fi
+
+  # Call helper function which possibly converts this using DOS-style short mode.
+  # If so, the updated path is stored in $new_path.
+
+  input_path="$new_path"
+  # Check if we need to convert this using DOS-style short mode. If the path
+  # contains just simple characters, use it. Otherwise (spaces, weird characters),
+  # take no chances and rewrite it.
+  # Note: m4 eats our [], so we need to use [ and ] instead.
+  has_forbidden_chars=`$ECHO "$input_path" | $GREP [^-._/a-zA-Z0-9]`
+  if test "x$has_forbidden_chars" != x; then
+    # Now convert it to mixed DOS-style, short mode (no spaces, and / instead of \)
+    shortmode_path=`$CYGPATH -s -m -a "$input_path"`
+    path_after_shortmode=`$CYGPATH -u "$shortmode_path"`
+    if test "x$path_after_shortmode" != "x$input_to_shortpath"; then
+      # Going to short mode and back again did indeed matter. Since short mode is
+      # case insensitive, let's make it lowercase to improve readability.
+      shortmode_path=`$ECHO "$shortmode_path" | $TR 'ABCDEFGHIJKLMNOPQRSTUVWXYZ' 'abcdefghijklmnopqrstuvwxyz'`
+      # Now convert it back to Unix-style (cygpath)
+      input_path=`$CYGPATH -u "$shortmode_path"`
+      new_path="$input_path"
+    fi
+  fi
+
+  test_cygdrive_prefix=`$ECHO $input_path | $GREP ^/cygdrive/`
+  if test "x$test_cygdrive_prefix" = x; then
+    # As a simple fix, exclude /usr/bin since it's not a real path.
+    if test "x`$ECHO $new_path | $GREP ^/usr/bin/`" = x; then
+      # The path is in a Cygwin special directory (e.g. /home). We need this converted to
+      # a path prefixed by /cygdrive for fixpath to work.
+      new_path="$CYGWIN_ROOT_PATH$input_path"
+    fi
+  fi
+
+
+  if test "x$path" != "x$new_path"; then
+    JT_HOME="$new_path"
+    { $as_echo "$as_me:${as_lineno-$LINENO}: Rewriting JT_HOME to \"$new_path\"" >&5
+$as_echo "$as_me: Rewriting JT_HOME to \"$new_path\"" >&6;}
+  fi
+
+    elif test "x$OPENJDK_BUILD_OS_ENV" = "xwindows.msys"; then
+
+  path="$JT_HOME"
+  has_colon=`$ECHO $path | $GREP ^.:`
+  new_path="$path"
+  if test "x$has_colon" = x; then
+    # Not in mixed or Windows style, start by that.
+    new_path=`cmd //c echo $path`
+  fi
+
+
+  input_path="$new_path"
+  # Check if we need to convert this using DOS-style short mode. If the path
+  # contains just simple characters, use it. Otherwise (spaces, weird characters),
+  # take no chances and rewrite it.
+  # Note: m4 eats our [], so we need to use [ and ] instead.
+  has_forbidden_chars=`$ECHO "$input_path" | $GREP [^-_/:a-zA-Z0-9]`
+  if test "x$has_forbidden_chars" != x; then
+    # Now convert it to mixed DOS-style, short mode (no spaces, and / instead of \)
+    new_path=`cmd /c "for %A in (\"$input_path\") do @echo %~sA"|$TR \\\\\\\\ / | $TR 'ABCDEFGHIJKLMNOPQRSTUVWXYZ' 'abcdefghijklmnopqrstuvwxyz'`
+  fi
+
+
+  windows_path="$new_path"
+  if test "x$OPENJDK_BUILD_OS_ENV" = "xwindows.cygwin"; then
+    unix_path=`$CYGPATH -u "$windows_path"`
+    new_path="$unix_path"
+  elif test "x$OPENJDK_BUILD_OS_ENV" = "xwindows.msys"; then
+    unix_path=`$ECHO "$windows_path" | $SED -e 's,^\\(.\\):,/\\1,g' -e 's,\\\\,/,g'`
+    new_path="$unix_path"
+  fi
+
+  if test "x$path" != "x$new_path"; then
+    JT_HOME="$new_path"
+    { $as_echo "$as_me:${as_lineno-$LINENO}: Rewriting JT_HOME to \"$new_path\"" >&5
+$as_echo "$as_me: Rewriting JT_HOME to \"$new_path\"" >&6;}
+  fi
+
+  # Save the first 10 bytes of this path to the storage, so fixpath can work.
+  all_fixpath_prefixes=("${all_fixpath_prefixes[@]}" "${new_path:0:10}")
+
+    else
+      # We're on a unix platform. Hooray! :)
+      path="$JT_HOME"
+      has_space=`$ECHO "$path" | $GREP " "`
+      if test "x$has_space" != x; then
+        { $as_echo "$as_me:${as_lineno-$LINENO}: The path of JT_HOME, which resolves as \"$path\", is invalid." >&5
+$as_echo "$as_me: The path of JT_HOME, which resolves as \"$path\", is invalid." >&6;}
+        as_fn_error $? "Spaces are not allowed in this path." "$LINENO" 5
+      fi
+
+      # Use eval to expand a potential ~
+      eval path="$path"
+      if test ! -f "$path" && test ! -d "$path"; then
+        as_fn_error $? "The path of JT_HOME, which resolves as \"$path\", is not found." "$LINENO" 5
+      fi
+
+      if test -d "$path"; then
+        JT_HOME="`cd "$path"; $THEPWDCMD -L`"
+      else
+        dir="`$DIRNAME "$path"`"
+        base="`$BASENAME "$path"`"
+        JT_HOME="`cd "$dir"; $THEPWDCMD -L`/$base"
+      fi
     fi
   fi
 
@@ -49120,7 +50173,7 @@
         SHARED_LIBRARY_FLAGS="-dynamiclib -compatibility_version 1.0.0 -current_version 1.0.0 $PICFLAG"
         JVM_CFLAGS="$JVM_CFLAGS $PICFLAG"
       fi
-      SET_EXECUTABLE_ORIGIN='-Wl,-rpath,@loader_path/.'
+      SET_EXECUTABLE_ORIGIN='-Wl,-rpath,@loader_path$(or $1,/.)'
       SET_SHARED_LIBRARY_ORIGIN="$SET_EXECUTABLE_ORIGIN"
       SET_SHARED_LIBRARY_NAME='-Wl,-install_name,@rpath/$1'
       SET_SHARED_LIBRARY_MAPFILE='-Wl,-exported_symbols_list,$1'
@@ -49140,7 +50193,7 @@
       # Linking is different on MacOSX
       PICFLAG=''
       SHARED_LIBRARY_FLAGS="-dynamiclib -compatibility_version 1.0.0 -current_version 1.0.0 $PICFLAG"
-      SET_EXECUTABLE_ORIGIN='-Wl,-rpath,@loader_path/.'
+      SET_EXECUTABLE_ORIGIN='-Wl,-rpath,@loader_path$(or $1,/.)'
       SET_SHARED_LIBRARY_ORIGIN="$SET_EXECUTABLE_ORIGIN"
       SET_SHARED_LIBRARY_NAME='-Wl,-install_name,@rpath/$1'
       SET_SHARED_LIBRARY_MAPFILE='-Wl,-exported_symbols_list,$1'
@@ -52689,12 +53742,12 @@
     # Only enable AOT on linux-X64.
     if test "x$OPENJDK_TARGET_OS-$OPENJDK_TARGET_CPU" = "xlinux-x86_64"; then
       if test -e "$HOTSPOT_TOPDIR/src/jdk.aot"; then
-        if test -e "$HOTSPOT_TOPDIR/src/jdk.vm.compiler"; then
+        if test -e "$HOTSPOT_TOPDIR/src/jdk.internal.vm.compiler"; then
           ENABLE_AOT="true"
         else
           ENABLE_AOT="false"
           if test "x$enable_aot" = "xyes"; then
-            as_fn_error $? "Cannot build AOT without hotspot/src/jdk.vm.compiler sources. Remove --enable-aot." "$LINENO" 5
+            as_fn_error $? "Cannot build AOT without hotspot/src/jdk.internal.vm.compiler sources. Remove --enable-aot." "$LINENO" 5
           fi
         fi
       else
@@ -56702,6 +57755,12 @@
 fi
 
 
+# Check whether --with-freetype-license was given.
+if test "${with_freetype_license+set}" = set; then :
+  withval=$with_freetype_license;
+fi
+
+
   # Need to specify explicitly since it needs to be overridden on some versions of macosx
   FREETYPE_BASE_NAME=freetype
   FREETYPE_CFLAGS=
@@ -62648,6 +63707,153 @@
 
   fi # end freetype needed
 
+  FREETYPE_LICENSE=""
+  if test "x$with_freetype_license" = "xyes"; then
+    as_fn_error $? "--with-freetype-license must have a value" "$LINENO" 5
+  elif test "x$with_freetype_license" != "x"; then
+    { $as_echo "$as_me:${as_lineno-$LINENO}: checking for freetype license" >&5
+$as_echo_n "checking for freetype license... " >&6; }
+    { $as_echo "$as_me:${as_lineno-$LINENO}: result: $with_freetype_license" >&5
+$as_echo "$with_freetype_license" >&6; }
+    FREETYPE_LICENSE="$with_freetype_license"
+
+  # Only process if variable expands to non-empty
+
+  if test "x$FREETYPE_LICENSE" != x; then
+    if test "x$OPENJDK_BUILD_OS_ENV" = "xwindows.cygwin"; then
+
+  # Input might be given as Windows format, start by converting to
+  # unix format.
+  path="$FREETYPE_LICENSE"
+  new_path=`$CYGPATH -u "$path"`
+
+  # Cygwin tries to hide some aspects of the Windows file system, such that binaries are
+  # named .exe but called without that suffix. Therefore, "foo" and "foo.exe" are considered
+  # the same file, most of the time (as in "test -f"). But not when running cygpath -s, then
+  # "foo.exe" is OK but "foo" is an error.
+  #
+  # This test is therefore slightly more accurate than "test -f" to check for file precense.
+  # It is also a way to make sure we got the proper file name for the real test later on.
+  test_shortpath=`$CYGPATH -s -m "$new_path" 2> /dev/null`
+  if test "x$test_shortpath" = x; then
+    { $as_echo "$as_me:${as_lineno-$LINENO}: The path of FREETYPE_LICENSE, which resolves as \"$path\", is invalid." >&5
+$as_echo "$as_me: The path of FREETYPE_LICENSE, which resolves as \"$path\", is invalid." >&6;}
+    as_fn_error $? "Cannot locate the the path of FREETYPE_LICENSE" "$LINENO" 5
+  fi
+
+  # Call helper function which possibly converts this using DOS-style short mode.
+  # If so, the updated path is stored in $new_path.
+
+  input_path="$new_path"
+  # Check if we need to convert this using DOS-style short mode. If the path
+  # contains just simple characters, use it. Otherwise (spaces, weird characters),
+  # take no chances and rewrite it.
+  # Note: m4 eats our [], so we need to use [ and ] instead.
+  has_forbidden_chars=`$ECHO "$input_path" | $GREP [^-._/a-zA-Z0-9]`
+  if test "x$has_forbidden_chars" != x; then
+    # Now convert it to mixed DOS-style, short mode (no spaces, and / instead of \)
+    shortmode_path=`$CYGPATH -s -m -a "$input_path"`
+    path_after_shortmode=`$CYGPATH -u "$shortmode_path"`
+    if test "x$path_after_shortmode" != "x$input_to_shortpath"; then
+      # Going to short mode and back again did indeed matter. Since short mode is
+      # case insensitive, let's make it lowercase to improve readability.
+      shortmode_path=`$ECHO "$shortmode_path" | $TR 'ABCDEFGHIJKLMNOPQRSTUVWXYZ' 'abcdefghijklmnopqrstuvwxyz'`
+      # Now convert it back to Unix-style (cygpath)
+      input_path=`$CYGPATH -u "$shortmode_path"`
+      new_path="$input_path"
+    fi
+  fi
+
+  test_cygdrive_prefix=`$ECHO $input_path | $GREP ^/cygdrive/`
+  if test "x$test_cygdrive_prefix" = x; then
+    # As a simple fix, exclude /usr/bin since it's not a real path.
+    if test "x`$ECHO $new_path | $GREP ^/usr/bin/`" = x; then
+      # The path is in a Cygwin special directory (e.g. /home). We need this converted to
+      # a path prefixed by /cygdrive for fixpath to work.
+      new_path="$CYGWIN_ROOT_PATH$input_path"
+    fi
+  fi
+
+
+  if test "x$path" != "x$new_path"; then
+    FREETYPE_LICENSE="$new_path"
+    { $as_echo "$as_me:${as_lineno-$LINENO}: Rewriting FREETYPE_LICENSE to \"$new_path\"" >&5
+$as_echo "$as_me: Rewriting FREETYPE_LICENSE to \"$new_path\"" >&6;}
+  fi
+
+    elif test "x$OPENJDK_BUILD_OS_ENV" = "xwindows.msys"; then
+
+  path="$FREETYPE_LICENSE"
+  has_colon=`$ECHO $path | $GREP ^.:`
+  new_path="$path"
+  if test "x$has_colon" = x; then
+    # Not in mixed or Windows style, start by that.
+    new_path=`cmd //c echo $path`
+  fi
+
+
+  input_path="$new_path"
+  # Check if we need to convert this using DOS-style short mode. If the path
+  # contains just simple characters, use it. Otherwise (spaces, weird characters),
+  # take no chances and rewrite it.
+  # Note: m4 eats our [], so we need to use [ and ] instead.
+  has_forbidden_chars=`$ECHO "$input_path" | $GREP [^-_/:a-zA-Z0-9]`
+  if test "x$has_forbidden_chars" != x; then
+    # Now convert it to mixed DOS-style, short mode (no spaces, and / instead of \)
+    new_path=`cmd /c "for %A in (\"$input_path\") do @echo %~sA"|$TR \\\\\\\\ / | $TR 'ABCDEFGHIJKLMNOPQRSTUVWXYZ' 'abcdefghijklmnopqrstuvwxyz'`
+  fi
+
+
+  windows_path="$new_path"
+  if test "x$OPENJDK_BUILD_OS_ENV" = "xwindows.cygwin"; then
+    unix_path=`$CYGPATH -u "$windows_path"`
+    new_path="$unix_path"
+  elif test "x$OPENJDK_BUILD_OS_ENV" = "xwindows.msys"; then
+    unix_path=`$ECHO "$windows_path" | $SED -e 's,^\\(.\\):,/\\1,g' -e 's,\\\\,/,g'`
+    new_path="$unix_path"
+  fi
+
+  if test "x$path" != "x$new_path"; then
+    FREETYPE_LICENSE="$new_path"
+    { $as_echo "$as_me:${as_lineno-$LINENO}: Rewriting FREETYPE_LICENSE to \"$new_path\"" >&5
+$as_echo "$as_me: Rewriting FREETYPE_LICENSE to \"$new_path\"" >&6;}
+  fi
+
+  # Save the first 10 bytes of this path to the storage, so fixpath can work.
+  all_fixpath_prefixes=("${all_fixpath_prefixes[@]}" "${new_path:0:10}")
+
+    else
+      # We're on a unix platform. Hooray! :)
+      path="$FREETYPE_LICENSE"
+      has_space=`$ECHO "$path" | $GREP " "`
+      if test "x$has_space" != x; then
+        { $as_echo "$as_me:${as_lineno-$LINENO}: The path of FREETYPE_LICENSE, which resolves as \"$path\", is invalid." >&5
+$as_echo "$as_me: The path of FREETYPE_LICENSE, which resolves as \"$path\", is invalid." >&6;}
+        as_fn_error $? "Spaces are not allowed in this path." "$LINENO" 5
+      fi
+
+      # Use eval to expand a potential ~
+      eval path="$path"
+      if test ! -f "$path" && test ! -d "$path"; then
+        as_fn_error $? "The path of FREETYPE_LICENSE, which resolves as \"$path\", is not found." "$LINENO" 5
+      fi
+
+      if test -d "$path"; then
+        FREETYPE_LICENSE="`cd "$path"; $THEPWDCMD -L`"
+      else
+        dir="`$DIRNAME "$path"`"
+        base="`$BASENAME "$path"`"
+        FREETYPE_LICENSE="`cd "$dir"; $THEPWDCMD -L`/$base"
+      fi
+    fi
+  fi
+
+    if test ! -f "$FREETYPE_LICENSE"; then
+      as_fn_error $? "$FREETYPE_LICENSE cannot be found" "$LINENO" 5
+    fi
+  fi
+
+
 
 
 
@@ -64558,8 +65764,8 @@
     JVM_FEATURES_jvmci=""
   fi
 
-  { $as_echo "$as_me:${as_lineno-$LINENO}: checking if jdk.vm.compiler should be built" >&5
-$as_echo_n "checking if jdk.vm.compiler should be built... " >&6; }
+  { $as_echo "$as_me:${as_lineno-$LINENO}: checking if jdk.internal.vm.compiler should be built" >&5
+$as_echo_n "checking if jdk.internal.vm.compiler should be built... " >&6; }
   if   [[ " $JVM_FEATURES " =~ " graal " ]]  ; then
     { $as_echo "$as_me:${as_lineno-$LINENO}: result: yes, forced" >&5
 $as_echo "yes, forced" >&6; }
--- a/common/autoconf/hotspot.m4	Tue Feb 14 14:57:56 2017 -0800
+++ b/common/autoconf/hotspot.m4	Thu Aug 17 12:17:39 2017 -0700
@@ -215,12 +215,12 @@
     # Only enable AOT on linux-X64.
     if test "x$OPENJDK_TARGET_OS-$OPENJDK_TARGET_CPU" = "xlinux-x86_64"; then
       if test -e "$HOTSPOT_TOPDIR/src/jdk.aot"; then
-        if test -e "$HOTSPOT_TOPDIR/src/jdk.vm.compiler"; then
+        if test -e "$HOTSPOT_TOPDIR/src/jdk.internal.vm.compiler"; then
           ENABLE_AOT="true"
         else
           ENABLE_AOT="false"
           if test "x$enable_aot" = "xyes"; then
-            AC_MSG_ERROR([Cannot build AOT without hotspot/src/jdk.vm.compiler sources. Remove --enable-aot.])
+            AC_MSG_ERROR([Cannot build AOT without hotspot/src/jdk.internal.vm.compiler sources. Remove --enable-aot.])
           fi
         fi
       else
@@ -327,7 +327,7 @@
     JVM_FEATURES_jvmci=""
   fi
 
-  AC_MSG_CHECKING([if jdk.vm.compiler should be built])
+  AC_MSG_CHECKING([if jdk.internal.vm.compiler should be built])
   if HOTSPOT_CHECK_JVM_FEATURE(graal); then
     AC_MSG_RESULT([yes, forced])
     if test "x$JVM_FEATURES_jvmci" != "xjvmci" ; then
--- a/common/autoconf/jdk-options.m4	Tue Feb 14 14:57:56 2017 -0800
+++ b/common/autoconf/jdk-options.m4	Thu Aug 17 12:17:39 2017 -0700
@@ -154,6 +154,55 @@
 
   AC_SUBST(ENABLE_HEADLESS_ONLY)
 
+  # Should we build the complete docs, or just a lightweight version?
+  AC_ARG_ENABLE([full-docs], [AS_HELP_STRING([--enable-full-docs],
+      [build complete documentation @<:@enabled if all tools found@:>@])])
+
+  # Verify dependencies
+  AC_MSG_CHECKING([for graphviz dot])
+  if test "x$DOT" != "x"; then
+    AC_MSG_RESULT([yes])
+  else
+    AC_MSG_RESULT([no, cannot generate full docs])
+    FULL_DOCS_DEP_MISSING=true
+  fi
+
+  AC_MSG_CHECKING([for pandoc])
+  if test "x$PANDOC" != "x"; then
+    AC_MSG_RESULT([yes])
+  else
+    AC_MSG_RESULT([no, cannot generate full docs])
+    FULL_DOCS_DEP_MISSING=true
+  fi
+
+  AC_MSG_CHECKING([full docs])
+  if test "x$enable_full_docs" = xyes; then
+    if test "x$FULL_DOCS_DEP_MISSING" = "xtrue"; then
+      AC_MSG_RESULT([no, missing dependencies])
+      HELP_MSG_MISSING_DEPENDENCY([dot])
+      AC_MSG_ERROR([Cannot enable full docs with missing dependencies. See above. $HELP_MSG])
+    else
+      ENABLE_FULL_DOCS=true
+      AC_MSG_RESULT([yes, forced])
+    fi
+  elif test "x$enable_full_docs" = xno; then
+    ENABLE_FULL_DOCS=false
+    AC_MSG_RESULT([no, forced])
+  elif test "x$enable_full_docs" = x; then
+    # Check for prerequisites
+    if test "x$FULL_DOCS_DEP_MISSING" = xtrue; then
+      ENABLE_FULL_DOCS=false
+      AC_MSG_RESULT([no, missing dependencies])
+    else
+      ENABLE_FULL_DOCS=true
+      AC_MSG_RESULT([yes, dependencies present])
+    fi
+  else
+    AC_MSG_ERROR([--enable-full-docs can only take yes or no])
+  fi
+
+  AC_SUBST(ENABLE_FULL_DOCS)
+
   # Choose cacerts source file
   AC_ARG_WITH(cacerts-file, [AS_HELP_STRING([--with-cacerts-file],
       [specify alternative cacerts file])])
--- a/common/autoconf/lib-freetype.m4	Tue Feb 14 14:57:56 2017 -0800
+++ b/common/autoconf/lib-freetype.m4	Thu Aug 17 12:17:39 2017 -0700
@@ -194,6 +194,8 @@
       [specify directory with freetype sources to automatically build the library (experimental, Windows-only)])])
   AC_ARG_ENABLE(freetype-bundling, [AS_HELP_STRING([--disable-freetype-bundling],
       [disable bundling of the freetype library with the build result @<:@enabled on Windows or when using --with-freetype, disabled otherwise@:>@])])
+  AC_ARG_WITH(freetype-license, [AS_HELP_STRING([--with-freetype-license],
+      [if bundling freetype, also bundle this license file])])
 
   # Need to specify explicitly since it needs to be overridden on some versions of macosx
   FREETYPE_BASE_NAME=freetype
@@ -443,7 +445,21 @@
 
   fi # end freetype needed
 
+  FREETYPE_LICENSE=""
+  if test "x$with_freetype_license" = "xyes"; then
+    AC_MSG_ERROR([--with-freetype-license must have a value])
+  elif test "x$with_freetype_license" != "x"; then
+    AC_MSG_CHECKING([for freetype license])
+    AC_MSG_RESULT([$with_freetype_license])
+    FREETYPE_LICENSE="$with_freetype_license"
+    BASIC_FIXUP_PATH(FREETYPE_LICENSE)
+    if test ! -f "$FREETYPE_LICENSE"; then
+      AC_MSG_ERROR([$FREETYPE_LICENSE cannot be found])
+    fi
+  fi
+
   AC_SUBST(FREETYPE_BUNDLE_LIB_PATH)
   AC_SUBST(FREETYPE_CFLAGS)
   AC_SUBST(FREETYPE_LIBS)
+  AC_SUBST(FREETYPE_LICENSE)
 ])
--- a/common/autoconf/platform.m4	Tue Feb 14 14:57:56 2017 -0800
+++ b/common/autoconf/platform.m4	Thu Aug 17 12:17:39 2017 -0700
@@ -436,32 +436,42 @@
 AC_DEFUN([PLATFORM_SET_RELEASE_FILE_OS_VALUES],
 [
   if test "x$OPENJDK_TARGET_OS" = "xsolaris"; then
-    REQUIRED_OS_NAME=SunOS
-    REQUIRED_OS_VERSION=5.10
+    RELEASE_FILE_OS_NAME=SunOS
   fi
   if test "x$OPENJDK_TARGET_OS" = "xlinux"; then
-    REQUIRED_OS_NAME=Linux
-    REQUIRED_OS_VERSION=2.6
+    RELEASE_FILE_OS_NAME=Linux
   fi
   if test "x$OPENJDK_TARGET_OS" = "xwindows"; then
-    REQUIRED_OS_NAME=Windows
-    if test "x$OPENJDK_TARGET_CPU_BITS" = "x64"; then
-      REQUIRED_OS_VERSION=5.2
-    else
-      REQUIRED_OS_VERSION=5.1
-    fi
+    RELEASE_FILE_OS_NAME=Windows
   fi
-  if test "x$OPENJDK_TARGET_OS" = "xmacosx"; then
-    REQUIRED_OS_NAME=Darwin
-    REQUIRED_OS_VERSION=11.2
+  if test "x$OPENJDK_TARGET_OS" = xmacosx; then
+    RELEASE_FILE_OS_NAME="Darwin"
   fi
   if test "x$OPENJDK_TARGET_OS" = "xaix"; then
-    REQUIRED_OS_NAME=AIX
-    REQUIRED_OS_VERSION=7.1
+    RELEASE_FILE_OS_NAME="AIX"
+  fi
+  RELEASE_FILE_OS_ARCH=${OPENJDK_TARGET_CPU}
+
+  AC_SUBST(RELEASE_FILE_OS_NAME)
+  AC_SUBST(RELEASE_FILE_OS_ARCH)
+])
+
+AC_DEFUN([PLATFORM_SET_MODULE_TARGET_OS_VALUES],
+[
+  if test "x$OPENJDK_TARGET_OS" = xmacosx; then
+    OPENJDK_MODULE_TARGET_OS_NAME="macos"
+  else
+    OPENJDK_MODULE_TARGET_OS_NAME="$OPENJDK_TARGET_OS"
   fi
 
-  AC_SUBST(REQUIRED_OS_NAME)
-  AC_SUBST(REQUIRED_OS_VERSION)
+  if test "x$OPENJDK_TARGET_CPU" = xx86_64; then
+    OPENJDK_MODULE_TARGET_OS_ARCH="amd64"
+  else
+    OPENJDK_MODULE_TARGET_OS_ARCH="$OPENJDK_TARGET_CPU"
+  fi
+
+  OPENJDK_MODULE_TARGET_PLATFORM="${OPENJDK_MODULE_TARGET_OS_NAME}-${OPENJDK_MODULE_TARGET_OS_ARCH}"
+  AC_SUBST(OPENJDK_MODULE_TARGET_PLATFORM)
 ])
 
 #%%% Build and target systems %%%
@@ -478,6 +488,7 @@
 
   PLATFORM_EXTRACT_TARGET_AND_BUILD
   PLATFORM_SETUP_TARGET_CPU_BITS
+  PLATFORM_SET_MODULE_TARGET_OS_VALUES
   PLATFORM_SET_RELEASE_FILE_OS_VALUES
   PLATFORM_SETUP_LEGACY_VARS
 ])
--- a/common/autoconf/spec.gmk.in	Tue Feb 14 14:57:56 2017 -0800
+++ b/common/autoconf/spec.gmk.in	Thu Aug 17 12:17:39 2017 -0700
@@ -101,9 +101,12 @@
 OPENJDK_BUILD_CPU_BITS:=@OPENJDK_BUILD_CPU_BITS@
 OPENJDK_BUILD_CPU_ENDIAN:=@OPENJDK_BUILD_CPU_ENDIAN@
 
-# Legacy OS values for use in release file.
-REQUIRED_OS_NAME:=@REQUIRED_OS_NAME@
-REQUIRED_OS_VERSION:=@REQUIRED_OS_VERSION@
+# Target platform value in ModuleTarget class file attribute.
+OPENJDK_MODULE_TARGET_PLATFORM:=@OPENJDK_MODULE_TARGET_PLATFORM@
+
+# OS_* properties in release file
+RELEASE_FILE_OS_NAME:=@RELEASE_FILE_OS_NAME@
+RELEASE_FILE_OS_ARCH:=@RELEASE_FILE_OS_ARCH@
 
 LIBM:=@LIBM@
 LIBDL:=@LIBDL@
@@ -249,6 +252,8 @@
 # Only build headless support or not
 ENABLE_HEADLESS_ONLY := @ENABLE_HEADLESS_ONLY@
 
+ENABLE_FULL_DOCS := @ENABLE_FULL_DOCS@
+
 # JDK_OUTPUTDIR specifies where a working jvm is built.
 # You can run $(JDK_OUTPUTDIR)/bin/java
 # Though the layout of the contents of $(JDK_OUTPUTDIR) is not
@@ -271,9 +276,6 @@
 TESTMAKE_OUTPUTDIR=$(BUILD_OUTPUT)/test-make
 MAKESUPPORT_OUTPUTDIR=$(BUILD_OUTPUT)/make-support
 
-# By default, output javadoc directly into image
-JAVADOC_OUTPUTDIR = $(DOCS_IMAGE_DIR)
-
 # This does not get overridden in a bootcycle build
 CONFIGURESUPPORT_OUTPUTDIR:=@CONFIGURESUPPORT_OUTPUTDIR@
 BUILDJDK_OUTPUTDIR=$(BUILD_OUTPUT)/buildjdk
@@ -315,6 +317,7 @@
 FREETYPE_LIBS:=@FREETYPE_LIBS@
 FREETYPE_CFLAGS:=@FREETYPE_CFLAGS@
 FREETYPE_BUNDLE_LIB_PATH=@FREETYPE_BUNDLE_LIB_PATH@
+FREETYPE_LICENSE=@FREETYPE_LICENSE@
 CUPS_CFLAGS:=@CUPS_CFLAGS@
 ALSA_LIBS:=@ALSA_LIBS@
 ALSA_CFLAGS:=@ALSA_CFLAGS@
@@ -579,20 +582,31 @@
 # Use ?= as this can be overridden from bootcycle-spec.gmk
 BOOT_JDK_MODULAR ?= @BOOT_JDK_MODULAR@
 
-INTERIM_OVERRIDE_MODULES := java.compiler jdk.compiler \
-    jdk.jdeps jdk.javadoc jdk.rmic
+INTERIM_LANGTOOLS_OVERRIDE_MODULES := java.compiler jdk.compiler \
+    jdk.jdeps jdk.javadoc
+INTERIM_RMIC_OVERRIDE_MODULES := jdk.rmic
 ifeq ($(BOOT_JDK_MODULAR), true)
-  INTERIM_OVERRIDE_MODULES_ARGS = $(foreach m, $(INTERIM_OVERRIDE_MODULES), \
+  INTERIM_LANGTOOLS_OVERRIDE_MODULES_ARGS = $(foreach m, \
+      $(INTERIM_LANGTOOLS_OVERRIDE_MODULES), \
       --patch-module $m=$(BUILDTOOLS_OUTPUTDIR)/override_modules/$m)
-  INTERIM_LANGTOOLS_ARGS = $(INTERIM_OVERRIDE_MODULES_ARGS)
+  INTERIM_RMIC_OVERRIDE_MODULES_ARGS = $(foreach m, \
+      $(INTERIM_LANGTOOLS_OVERRIDE_MODULES) \
+      $(INTERIM_RMIC_OVERRIDE_MODULES), \
+      --patch-module $m=$(BUILDTOOLS_OUTPUTDIR)/override_modules/$m)
+  INTERIM_LANGTOOLS_ARGS = $(INTERIM_LANGTOOLS_OVERRIDE_MODULES_ARGS)
   JAVAC_MAIN_CLASS = -m jdk.compiler/com.sun.tools.javac.Main
   JAVADOC_MAIN_CLASS = -m jdk.javadoc/jdk.javadoc.internal.tool.Main
 else
-  INTERIM_OVERRIDE_MODULES_ARGS = \
+  INTERIM_LANGTOOLS_OVERRIDE_MODULES_ARGS = \
       -Xbootclasspath/p:$(call PathList, \
           $(addprefix $(BUILDTOOLS_OUTPUTDIR)/override_modules/, \
-              $(INTERIM_OVERRIDE_MODULES)))
-  INTERIM_LANGTOOLS_ARGS = $(INTERIM_OVERRIDE_MODULES_ARGS) \
+              $(INTERIM_LANGTOOLS_OVERRIDE_MODULES)))
+  INTERIM_RMIC_OVERRIDE_MODULES_ARGS = \
+      -Xbootclasspath/p:$(call PathList, \
+          $(addprefix $(BUILDTOOLS_OUTPUTDIR)/override_modules/, \
+              $(INTERIM_LANGTOOLS_OVERRIDE_MODULES) \
+              $(INTERIM_RMIC_OVERRIDE_MODULES)))
+  INTERIM_LANGTOOLS_ARGS = $(INTERIM_LANGTOOLS_OVERRIDE_MODULES_ARGS) \
       -cp $(BUILDTOOLS_OUTPUTDIR)/override_modules/jdk.compiler
   JAVAC_MAIN_CLASS = com.sun.tools.javac.Main
   JAVADOC_MAIN_CLASS = jdk.javadoc.internal.tool.Main
@@ -642,6 +656,7 @@
 MV:=@MV@
 NAWK:=@NAWK@
 NICE:=@NICE@
+PANDOC:=@FIXPATH@ @PANDOC@
 PATCH:=@PATCH@
 PRINTF:=@PRINTF@
 RM:=@RM@
@@ -670,6 +685,7 @@
 READELF:=@READELF@
 EXPR:=@EXPR@
 FILE:=@FILE@
+DOT:=@DOT@
 HG:=@HG@
 OBJCOPY:=@OBJCOPY@
 SETFILE:=@SETFILE@
@@ -805,6 +821,8 @@
 # Docs image
 DOCS_IMAGE_SUBDIR := docs
 DOCS_IMAGE_DIR = $(IMAGES_OUTPUTDIR)/$(DOCS_IMAGE_SUBDIR)
+# Output docs directly into image
+DOCS_OUTPUTDIR := $(DOCS_IMAGE_DIR)
 
 # Macosx bundles directory definitions
 JDK_MACOSX_BUNDLE_SUBDIR=jdk-bundle
@@ -833,11 +851,6 @@
     jre-$(VERSION_SHORT)+$(VERSION_BUILD)-compact3_$(OPENJDK_TARGET_BUNDLE_PLATFORM)_bin$(DEBUG_PART).tar.gz
 JDK_SYMBOLS_BUNDLE_NAME := jdk-$(BASE_NAME)_bin$(DEBUG_PART)-symbols.tar.gz
 JRE_SYMBOLS_BUNDLE_NAME := jre-$(BASE_NAME)_bin$(DEBUG_PART)-symbols.tar.gz
-ifeq ($(OPENJDK_TARGET_OS), windows)
-  DEMOS_BUNDLE_NAME := jdk-$(BASE_NAME)_demo$(DEBUG_PART).zip
-else
-  DEMOS_BUNDLE_NAME := jdk-$(BASE_NAME)_demo$(DEBUG_PART).tar.gz
-endif
 TEST_BUNDLE_NAME := jdk-$(BASE_NAME)_bin-tests$(DEBUG_PART).tar.gz
 DOCS_BUNDLE_NAME := jdk-$(BASE_NAME)_doc-api-spec$(DEBUG_PART).tar.gz
 
@@ -845,7 +858,6 @@
 JRE_BUNDLE :=  $(BUNDLES_OUTPUTDIR)/$(JRE_BUNDLE_NAME)
 JDK_SYMBOLS_BUNDLE :=  $(BUNDLES_OUTPUTDIR)/$(JDK_SYMBOLS_BUNDLE_NAME)
 JRE_SYMBOLS_BUNDLE :=  $(BUNDLES_OUTPUTDIR)/$(JRE_SYMBOLS_BUNDLE_NAME)
-DEMOS_BUNDLE :=  $(BUNDLES_OUTPUTDIR)/$(DEMOS_BUNDLE_NAME)
 TEST_BUNDLE :=  $(BUNDLES_OUTPUTDIR)/$(TEST_BUNDLE_NAME)
 DOCS_BUNDLE :=  $(BUNDLES_OUTPUTDIR)/$(DOCS_BUNDLE_NAME)
 
--- a/common/autoconf/toolchain.m4	Tue Feb 14 14:57:56 2017 -0800
+++ b/common/autoconf/toolchain.m4	Thu Aug 17 12:17:39 2017 -0700
@@ -440,7 +440,7 @@
 #
 # $1 = compiler to test (CC or CXX)
 # $2 = human readable name of compiler (C or C++)
-# $3 = list of compiler names to search for
+# $3 = compiler name to search for
 AC_DEFUN([TOOLCHAIN_FIND_COMPILER],
 [
   COMPILER_NAME=$2
@@ -482,15 +482,15 @@
     if test -n "$TOOLCHAIN_PATH"; then
       PATH_save="$PATH"
       PATH="$TOOLCHAIN_PATH"
-      AC_PATH_PROGS(TOOLCHAIN_PATH_$1, $SEARCH_LIST)
+      AC_PATH_TOOL(TOOLCHAIN_PATH_$1, $SEARCH_LIST)
       $1=$TOOLCHAIN_PATH_$1
       PATH="$PATH_save"
     fi
 
-    # AC_PATH_PROGS can't be run multiple times with the same variable,
+    # AC_PATH_TOOL can't be run multiple times with the same variable,
     # so create a new name for this run.
     if test "x[$]$1" = x; then
-      AC_PATH_PROGS(POTENTIAL_$1, $SEARCH_LIST)
+      AC_PATH_TOOL(POTENTIAL_$1, $SEARCH_LIST)
       $1=$POTENTIAL_$1
     fi
 
@@ -926,41 +926,82 @@
 AC_DEFUN_ONCE([TOOLCHAIN_SETUP_JTREG],
 [
   AC_ARG_WITH(jtreg, [AS_HELP_STRING([--with-jtreg],
-      [Regression Test Harness @<:@probed@:>@])],
-      [],
-      [with_jtreg=no])
+      [Regression Test Harness @<:@probed@:>@])])
 
   if test "x$with_jtreg" = xno; then
     # jtreg disabled
-    AC_MSG_CHECKING([for jtreg])
-    AC_MSG_RESULT(no)
-  else
-    if test "x$with_jtreg" != xyes; then
-      # with path specified.
-      JT_HOME="$with_jtreg"
+    AC_MSG_CHECKING([for jtreg test harness])
+    AC_MSG_RESULT([no, disabled])
+  elif test "x$with_jtreg" != xyes && test "x$with_jtreg" != x; then
+    # An explicit path is specified, use it.
+    JT_HOME="$with_jtreg"
+    if test ! -d "$JT_HOME"; then
+      AC_MSG_ERROR([jtreg home directory from --with-jtreg=$with_jtreg does not exist])
     fi
 
+    if test ! -e "$JT_HOME/lib/jtreg.jar"; then
+      AC_MSG_ERROR([jtreg home directory from --with-jtreg=$with_jtreg is not a valid jtreg home])
+    fi
+
+    JTREGEXE="$JT_HOME/bin/jtreg"
+    if test ! -x "$JTREGEXE"; then
+      AC_MSG_ERROR([jtreg home directory from --with-jtreg=$with_jtreg does not contain valid jtreg executable])
+    fi
+
+    AC_MSG_CHECKING([for jtreg test harness])
+    AC_MSG_RESULT([$JT_HOME])
+  else
+    # Try to locate jtreg
     if test "x$JT_HOME" != x; then
-      AC_MSG_CHECKING([for jtreg])
+      # JT_HOME set in environment, use it
+      if test ! -d "$JT_HOME"; then
+        AC_MSG_WARN([Ignoring JT_HOME pointing to invalid directory: $JT_HOME])
+        JT_HOME=
+      else
+        if test ! -e "$JT_HOME/lib/jtreg.jar"; then
+          AC_MSG_WARN([Ignoring JT_HOME which is not a valid jtreg home: $JT_HOME])
+          JT_HOME=
+        elif test ! -x "$JT_HOME/bin/jtreg"; then
+          AC_MSG_WARN([Ignoring JT_HOME which does not contain valid jtreg executable: $JT_HOME])
+          JT_HOME=
+        else
+          JTREGEXE="$JT_HOME/bin/jtreg"
+          AC_MSG_NOTICE([Located jtreg using JT_HOME from environment])
+        fi
+      fi
+    fi
 
-      # use JT_HOME enviroment var.
-      BASIC_FIXUP_PATH([JT_HOME])
+    if test "x$JT_HOME" = x; then
+      # JT_HOME is not set in environment, or was deemed invalid.
+      # Try to find jtreg on path
+      BASIC_PATH_PROGS(JTREGEXE, jtreg)
+      if test "x$JTREGEXE" != x; then
+        # That's good, now try to derive JT_HOME
+        JT_HOME=`(cd $($DIRNAME $JTREGEXE)/.. && pwd)`
+        if test ! -e "$JT_HOME/lib/jtreg.jar"; then
+          AC_MSG_WARN([Ignoring jtreg from path since a valid jtreg home cannot be found])
+          JT_HOME=
+          JTREGEXE=
+        else
+          AC_MSG_NOTICE([Located jtreg using jtreg executable in path])
+        fi
+      fi
+    fi
 
-      # jtreg win32 script works for everybody
-      JTREGEXE="$JT_HOME/bin/jtreg"
+    AC_MSG_CHECKING([for jtreg test harness])
+    if test "x$JT_HOME" != x; then
+      AC_MSG_RESULT([$JT_HOME])
+    else
+      AC_MSG_RESULT([no, not found])
 
-      if test ! -f "$JTREGEXE"; then
-        AC_MSG_ERROR([JTReg executable does not exist: $JTREGEXE])
+      if test "x$with_jtreg" = xyes; then
+        AC_MSG_ERROR([--with-jtreg was specified, but no jtreg found.])
       fi
-
-      AC_MSG_RESULT($JTREGEXE)
-    else
-      # try to find jtreg on path
-      BASIC_REQUIRE_PROGS(JTREGEXE, jtreg)
-      JT_HOME="`$DIRNAME $JTREGEXE`"
     fi
   fi
 
+  BASIC_FIXUP_EXECUTABLE(JTREGEXE)
+  BASIC_FIXUP_PATH(JT_HOME)
   AC_SUBST(JT_HOME)
   AC_SUBST(JTREGEXE)
 ])
--- a/common/bin/compare_exceptions.sh.incl	Tue Feb 14 14:57:56 2017 -0800
+++ b/common/bin/compare_exceptions.sh.incl	Thu Aug 17 12:17:39 2017 -0700
@@ -38,25 +38,9 @@
 if [ "$OPENJDK_TARGET_OS" = "linux" ]; then
 
   STRIP_BEFORE_COMPARE="
-      ./demo/jvmti/compiledMethodLoad/lib/libcompiledMethodLoad.so
-      ./demo/jvmti/gctest/lib/libgctest.so
-      ./demo/jvmti/heapTracker/lib/libheapTracker.so
-      ./demo/jvmti/heapViewer/lib/libheapViewer.so
-      ./demo/jvmti/minst/lib/libminst.so
-      ./demo/jvmti/mtrace/lib/libmtrace.so
-      ./demo/jvmti/versionCheck/lib/libversionCheck.so
-      ./demo/jvmti/waiters/lib/libwaiters.so
       "
 
   ACCEPTED_BIN_DIFF="
-      ./demo/jvmti/compiledMethodLoad/lib/libcompiledMethodLoad.so
-      ./demo/jvmti/gctest/lib/libgctest.so
-      ./demo/jvmti/heapTracker/lib/libheapTracker.so
-      ./demo/jvmti/heapViewer/lib/libheapViewer.so
-      ./demo/jvmti/minst/lib/libminst.so
-      ./demo/jvmti/mtrace/lib/libmtrace.so
-      ./demo/jvmti/versionCheck/lib/libversionCheck.so
-      ./demo/jvmti/waiters/lib/libwaiters.so
       ./lib/client/libjsig.so
       ./lib/client/libjvm.so
       ./lib/libattach.so
@@ -151,15 +135,6 @@
 if [ "$OPENJDK_TARGET_OS" = "solaris" ] && [ "$OPENJDK_TARGET_CPU" = "x86_64" ]; then
 
   STRIP_BEFORE_COMPARE="
-      ./demo/jni/Poller/lib/libPoller.so
-      ./demo/jvmti/compiledMethodLoad/lib/libcompiledMethodLoad.so
-      ./demo/jvmti/gctest/lib/libgctest.so
-      ./demo/jvmti/heapTracker/lib/libheapTracker.so
-      ./demo/jvmti/heapViewer/lib/libheapViewer.so
-      ./demo/jvmti/minst/lib/libminst.so
-      ./demo/jvmti/mtrace/lib/libmtrace.so
-      ./demo/jvmti/versionCheck/lib/libversionCheck.so
-      ./demo/jvmti/waiters/lib/libwaiters.so
       "
 
   SORT_SYMBOLS="
@@ -174,15 +149,6 @@
   SKIP_BIN_DIFF="true"
 
   ACCEPTED_SMALL_SIZE_DIFF="
-      ./demo/jni/Poller/lib/libPoller.so
-      ./demo/jvmti/compiledMethodLoad/lib/libcompiledMethodLoad.so
-      ./demo/jvmti/gctest/lib/libgctest.so
-      ./demo/jvmti/heapTracker/lib/libheapTracker.so
-      ./demo/jvmti/heapViewer/lib/libheapViewer.so
-      ./demo/jvmti/minst/lib/libminst.so
-      ./demo/jvmti/mtrace/lib/libmtrace.so
-      ./demo/jvmti/versionCheck/lib/libversionCheck.so
-      ./demo/jvmti/waiters/lib/libwaiters.so
       ./lib/jli/libjli.so
       ./lib/jspawnhelper
       ./lib/libJdbcOdbc.so
@@ -279,19 +245,9 @@
 if [ "$OPENJDK_TARGET_OS" = "solaris" ] && [ "$OPENJDK_TARGET_CPU" = "sparcv9" ]; then
 
   STRIP_BEFORE_COMPARE="
-      ./demo/jni/Poller/lib/libPoller.so
-      ./demo/jvmti/compiledMethodLoad/lib/libcompiledMethodLoad.so
-      ./demo/jvmti/gctest/lib/libgctest.so
-      ./demo/jvmti/heapTracker/lib/libheapTracker.so
-      ./demo/jvmti/heapViewer/lib/libheapViewer.so
-      ./demo/jvmti/minst/lib/libminst.so
-      ./demo/jvmti/mtrace/lib/libmtrace.so
-      ./demo/jvmti/versionCheck/lib/libversionCheck.so
-      ./demo/jvmti/waiters/lib/libwaiters.so
       "
 
   SORT_SYMBOLS="
-      ./demo/jvmti/waiters/lib/libwaiters.so
       ./lib/libjsig.so
       ./lib/libfontmanager.so
       ./lib/libjimage.so
@@ -305,15 +261,6 @@
   SKIP_BIN_DIFF="true"
 
   ACCEPTED_SMALL_SIZE_DIFF="
-      ./demo/jni/Poller/lib/libPoller.so
-      ./demo/jvmti/compiledMethodLoad/lib/libcompiledMethodLoad.so
-      ./demo/jvmti/gctest/lib/libgctest.so
-      ./demo/jvmti/heapTracker/lib/libheapTracker.so
-      ./demo/jvmti/heapViewer/lib/libheapViewer.so
-      ./demo/jvmti/minst/lib/libminst.so
-      ./demo/jvmti/mtrace/lib/libmtrace.so
-      ./demo/jvmti/versionCheck/lib/libversionCheck.so
-      ./demo/jvmti/waiters/lib/libwaiters.so
       ./lib/client/libjvm.so
       ./lib/jli/libjli.so
       ./lib/jspawnhelper
@@ -438,9 +385,6 @@
 
   # Probably should add all libs here
   ACCEPTED_SMALL_SIZE_DIFF="
-      ./demo/jvmti/gctest/lib/gctest.dll
-      ./demo/jvmti/heapTracker/lib/heapTracker.dll
-      ./demo/jvmti/minst/lib/minst.dll
       ./bin/attach.dll
       ./bin/jsoundds.dll
       ./bin/client/jvm.dll
@@ -579,14 +523,6 @@
       ./bin/wsgen
       ./bin/wsimport
       ./bin/xjc
-      ./demo/jvmti/compiledMethodLoad/lib/libcompiledMethodLoad.dylib
-      ./demo/jvmti/gctest/lib/libgctest.dylib
-      ./demo/jvmti/heapTracker/lib/libheapTracker.dylib
-      ./demo/jvmti/heapViewer/lib/libheapViewer.dylib
-      ./demo/jvmti/minst/lib/libminst.dylib
-      ./demo/jvmti/mtrace/lib/libmtrace.dylib
-      ./demo/jvmti/versionCheck/lib/libversionCheck.dylib
-      ./demo/jvmti/waiters/lib/libwaiters.dylib
       ./Contents/Home/bin/_javaws
       ./Contents/Home/bin/javaws
       ./Contents/Home/bin/idlj
--- a/common/bin/hgforest.sh	Tue Feb 14 14:57:56 2017 -0800
+++ b/common/bin/hgforest.sh	Thu Aug 17 12:17:39 2017 -0700
@@ -1,6 +1,6 @@
 #!/bin/sh
 #
-# Copyright (c) 2009, 2016, Oracle and/or its affiliates. All rights reserved.
+# Copyright (c) 2009, 2017, Oracle and/or its affiliates. All rights reserved.
 # DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
 #
 # This code is free software; you can redistribute it and/or modify it
@@ -183,7 +183,7 @@
 
 subrepos="corba jaxp jaxws langtools jdk hotspot nashorn"
 jdk_subrepos_extra="closed jdk/src/closed jdk/make/closed jdk/test/closed hotspot/make/closed hotspot/src/closed hotspot/test/closed"
-subrepos_extra="$jdk_subrepos_extra deploy install sponsors pubs"
+subrepos_extra="$jdk_subrepos_extra deploy install sponsors"
 
 # Only look in specific locations for possible forests (avoids long searches)
 pull_default=""
--- a/common/bin/unshuffle_list.txt	Tue Feb 14 14:57:56 2017 -0800
+++ b/common/bin/unshuffle_list.txt	Thu Aug 17 12:17:39 2017 -0700
@@ -361,8 +361,8 @@
 jdk/src/java.base/share/native/libverify/check_format.c : jdk/src/share/native/common/check_format.c
 jdk/src/java.base/share/native/libverify/opcodes.in_out : jdk/src/share/native/common/opcodes.in_out
 jdk/src/java.base/share/native/libzip : jdk/src/share/native/java/util/zip
-jdk/src/java.base/share/native/libzip/zlib-1.2.8 : jdk/src/share/native/java/util/zip/zlib-1.2.8
-jdk/src/java.base/share/native/libzip/zlib-1.2.8/patches/ChangeLog_java : jdk/src/share/native/java/util/zip/zlib-1.2.8/patches/ChangeLog_java
+jdk/src/java.base/share/native/libzip/zlib : jdk/src/share/native/java/util/zip/zlib
+jdk/src/java.base/share/native/libzip/zlib/patches/ChangeLog_java : jdk/src/share/native/java/util/zip/zlib/patches/ChangeLog_java
 jdk/src/java.base/unix/classes/java/io : jdk/src/solaris/classes/java/io
 jdk/src/java.base/unix/classes/java/lang : jdk/src/solaris/classes/java/lang
 jdk/src/java.base/unix/classes/java/net : jdk/src/solaris/classes/java/net
--- a/common/bin/update-build-readme.sh	Tue Feb 14 14:57:56 2017 -0800
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,62 +0,0 @@
-#!/bin/bash
-
-# Get an absolute path to this script, since that determines the top-level
-# directory.
-this_script_dir=`dirname $0`
-TOPDIR=`cd $this_script_dir/../.. > /dev/null && pwd`
-
-GREP=grep
-MD_FILE=$TOPDIR/README-builds.md
-HTML_FILE=$TOPDIR/README-builds.html
-
-# Locate the markdown processor tool and check that it is the correct version.
-locate_markdown_processor() {
-  if [ -z "$MARKDOWN" ]; then
-    MARKDOWN=`which markdown 2> /dev/null`
-    if [ -z "$MARKDOWN" ]; then
-      echo "Error: Cannot locate markdown processor" 1>&2
-      exit 1
-    fi
-  fi
-
-  # Test version
-  MARKDOWN_VERSION=`$MARKDOWN -version | $GREP version`
-  if [ "x$MARKDOWN_VERSION" != "xThis is Markdown, version 1.0.1." ]; then
-    echo "Error: Expected markdown version 1.0.1." 1>&2
-    echo "Actual version found: $MARKDOWN_VERSION" 1>&2
-    echo "Download markdown here: https://daringfireball.net/projects/markdown/"  1>&2
-    exit 1
-  fi
-
-}
-
-# Verify that the source markdown file looks sound.
-verify_source_code() {
-  TOO_LONG_LINES=`$GREP -E -e '^.{80}.+$' $MD_FILE`
-  if [ "x$TOO_LONG_LINES" != x ]; then
-    echo "Warning: The following lines are longer than 80 characters:"
-    $GREP -E -e '^.{80}.+$' $MD_FILE
-  fi
-}
-
-# Convert the markdown file to html format.
-process_source() {
-  echo "Generating html file from markdown"
-  cat > $HTML_FILE << END
-<html>
-  <head>
-    <title>OpenJDK Build README</title>
-  </head>
-  <body>
-END
-  markdown $MD_FILE >> $HTML_FILE
-  cat >> $HTML_FILE <<END
-  </body>
-</html>
-END
-  echo "Done"
-}
-
-locate_markdown_processor
-verify_source_code
-process_source
--- a/common/conf/jib-profiles.js	Tue Feb 14 14:57:56 2017 -0800
+++ b/common/conf/jib-profiles.js	Thu Aug 17 12:17:39 2017 -0700
@@ -1,5 +1,5 @@
 /*
- * Copyright (c) 2015, 2016, Oracle and/or its affiliates. All rights reserved.
+ * Copyright (c) 2015, 2017, Oracle and/or its affiliates. All rights reserved.
  * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
  *
  * This code is free software; you can redistribute it and/or modify it
@@ -231,18 +231,16 @@
     // List of the main profile names used for iteration
     common.main_profile_names = [
         "linux-x64", "linux-x86", "macosx-x64", "solaris-x64",
-        "solaris-sparcv9", "windows-x64", "windows-x86"
+        "solaris-sparcv9", "windows-x64", "windows-x86",
+        "linux-arm64", "linux-arm-vfp-hflt", "linux-arm-vfp-hflt-dyn"
     ];
 
     // These are the base setttings for all the main build profiles.
     common.main_profile_base = {
         dependencies: ["boot_jdk", "gnumake", "jtreg"],
         default_make_targets: ["product-bundles", "test-bundles"],
-        configure_args: [
-            "--with-version-opt=" + common.build_id,
-            "--enable-jtreg-failure-handler",
-            "--with-version-build=" + common.build_number
-        ]
+        configure_args: concat(["--enable-jtreg-failure-handler"],
+                               versionArgs(input, common))
     };
     // Extra settings for debug profiles
     common.debug_suffix = "-debug";
@@ -268,10 +266,12 @@
 
     /**
      * Define common artifacts template for all main profiles
-     * @param pf - Name of platform in bundle names
-     * @param demo_ext - Type of extension for demo bundle
+     * @param o - Object containing data for artifacts
      */
-    common.main_profile_artifacts = function (pf, demo_ext) {
+    common.main_profile_artifacts = function (o) {
+        var jdk_subdir = (o.jdk_subdir != null ? o.jdk_subdir : "jdk-" + data.version);
+        var jre_subdir = (o.jre_subdir != null ? o.jre_subdir : "jre-" + data.version);
+        var pf = o.platform
         return {
             artifacts: {
                 jdk: {
@@ -280,7 +280,7 @@
                         "bundles/" + pf + "/jdk-" + data.version + "_" + pf + "_bin.tar.gz",
                         "bundles/" + pf + "/\\1"
                     ],
-                    subdir: "jdk-" + data.version,
+                    subdir: jdk_subdir,
                     exploded: "images/jdk"
                 },
                 jre: {
@@ -289,7 +289,7 @@
                         "bundles/" + pf + "/jre-" + data.version + "_" + pf + "_bin.tar.gz",
                         "bundles/" + pf + "/\\1"
                     ],
-                    subdir: "jre-" + data.version,
+                    subdir: jre_subdir,
                     exploded: "images/jre"
                 },
                 test: {
@@ -306,7 +306,7 @@
                         "bundles/" + pf + "/jdk-" + data.version + "_" + pf + "_bin-symbols.tar.gz",
                         "bundles/" + pf + "/\\1"
                     ],
-                    subdir: "jdk-" + data.version,
+                    subdir: jdk_subdir,
                     exploded: "images/jdk"
                 },
                 jre_symbols: {
@@ -315,15 +315,8 @@
                         "bundles/" + pf + "/jre-" + data.version + "_" + pf + "_bin-symbols.tar.gz",
                         "bundles/" + pf + "/\\1"
                     ],
-                    subdir: "jre-" + data.version,
+                    subdir: jre_subdir,
                     exploded: "images/jre"
-                },
-                demo: {
-                    local: "bundles/\\(jdk.*demo." + demo_ext + "\\)",
-                    remote: [
-                        "bundles/" + pf + "/jdk-" + data.version + "_" + pf + "_demo." + demo_ext,
-                        "bundles/" + pf + "/\\1"
-                    ],
                 }
             }
         };
@@ -332,9 +325,12 @@
 
     /**
      * Define common artifacts template for all debug profiles
-     * @param pf - Name of platform in bundle names
+     * @param o - Object containing data for artifacts
      */
-    common.debug_profile_artifacts = function (pf) {
+    common.debug_profile_artifacts = function (o) {
+        var jdk_subdir = "jdk-" + data.version + "/fastdebug";
+        var jre_subdir = "jre-" + data.version + "/fastdebug";
+        var pf = o.platform
         return {
             artifacts: {
                 jdk: {
@@ -343,7 +339,7 @@
                         "bundles/" + pf + "/jdk-" + data.version + "_" + pf + "_bin-debug.tar.gz",
                         "bundles/" + pf + "/\\1"
                     ],
-                    subdir: "jdk-" + data.version,
+                    subdir: jdk_subdir,
                     exploded: "images/jdk"
                 },
                 jre: {
@@ -352,7 +348,7 @@
                         "bundles/" + pf + "/jre-" + data.version + "_" + pf + "_bin-debug.tar.gz",
                         "bundles/" + pf + "/\\1"
                     ],
-                    subdir: "jre-" + data.version,
+                    subdir: jre_subdir,
                     exploded: "images/jre"
                 },
                 test: {
@@ -369,7 +365,7 @@
                         "bundles/" + pf + "/jdk-" + data.version + "_" + pf + "_bin-debug-symbols.tar.gz",
                         "bundles/" + pf + "/\\1"
                     ],
-                    subdir: "jdk-" + data.version,
+                    subdir: jdk_subdir,
                     exploded: "images/jdk"
                 },
                 jre_symbols: {
@@ -378,7 +374,7 @@
                         "bundles/" + pf + "/jre-" + data.version + "_" + pf + "_bin-debug-symbols.tar.gz",
                         "bundles/" + pf + "/\\1"
                     ],
-                    subdir: "jre-" + data.version,
+                    subdir: jre_subdir,
                     exploded: "images/jre"
                 }
             }
@@ -391,7 +387,7 @@
     // on such hardware.
     if (input.build_cpu == "sparcv9") {
        var cpu_brand = $EXEC("bash -c \"kstat -m cpu_info | grep brand | head -n1 | awk '{ print \$2 }'\"");
-       if (cpu_brand.trim() == 'SPARC-M7') {
+       if (cpu_brand.trim().match('SPARC-.7')) {
            boot_jdk_revision = "8u20";
            boot_jdk_subdirpart = "1.8.0_20";
        }
@@ -401,6 +397,8 @@
     common.boot_jdk_home = input.get("boot_jdk", "home_path") + "/jdk"
         + common.boot_jdk_subdirpart
         + (input.build_os == "macosx" ? ".jdk/Contents/Home" : "");
+    common.boot_jdk_platform = input.build_os + "-"
+        + (input.build_cpu == "x86" ? "i586" : input.build_cpu);
 
     return common;
 };
@@ -419,8 +417,9 @@
         "linux-x64": {
             target_os: "linux",
             target_cpu: "x64",
-            dependencies: ["devkit"],
-            configure_args: concat(common.configure_args_64bit, "--with-zlib=system"),
+            dependencies: ["devkit", "graphviz", "pandoc"],
+            configure_args: concat(common.configure_args_64bit,
+                "--enable-full-docs", "--with-zlib=system"),
             default_make_targets: ["docs-bundles"],
         },
 
@@ -469,8 +468,43 @@
             build_cpu: "x64",
             dependencies: ["devkit", "freetype"],
             configure_args: concat(common.configure_args_32bit),
+        },
+
+        "linux-arm64": {
+            target_os: "linux",
+            target_cpu: "aarch64",
+            build_cpu: "x64",
+            dependencies: ["devkit", "build_devkit", "cups", "headless_stubs"],
+            configure_args: [
+                "--with-cpu-port=arm64",
+                "--with-jvm-variants=server",
+                "--openjdk-target=aarch64-linux-gnu",
+                "--enable-headless-only"
+            ],
+        },
+
+        "linux-arm-vfp-hflt": {
+            target_os: "linux",
+            target_cpu: "arm",
+            build_cpu: "x64",
+            dependencies: ["devkit", "build_devkit", "cups"],
+            configure_args: [
+                "--with-jvm-variants=minimal1,client",
+                "--with-x=" + input.get("devkit", "install_path") + "/arm-linux-gnueabihf/libc/usr/X11R6-PI",
+                "--openjdk-target=arm-linux-gnueabihf",
+                "--with-abi-profile=arm-vfp-hflt"
+            ],
+        },
+
+        // Special version of the SE profile adjusted to be testable on arm64 hardware.
+        "linux-arm-vfp-hflt-dyn": {
+            configure_args: "--with-stdc++lib=dynamic"
         }
     };
+    // Let linux-arm-vfp-hflt-dyn inherit everything from linux-arm-vfp-hflt
+    profiles["linux-arm-vfp-hflt-dyn"] = concatObjects(
+        profiles["linux-arm-vfp-hflt-dyn"], profiles["linux-arm-vfp-hflt"]);
+
     // Add the base settings to all the main profiles
     common.main_profile_names.forEach(function (name) {
         profiles[name] = concatObjects(common.main_profile_base, profiles[name]);
@@ -582,7 +616,7 @@
     var testOnlyProfilesPrebuilt = {
         "run-test-prebuilt": {
             src: "src.conf",
-            dependencies: [ "jtreg", "gnumake", testedProfile + ".jdk",
+            dependencies: [ "jtreg", "gnumake", "boot_jdk", testedProfile + ".jdk",
                 testedProfile + ".test", "src.full"
             ],
             work_dir: input.get("src.full", "install_path") + "/test",
@@ -626,49 +660,53 @@
     //
     // Define artifacts for profiles
     //
-    // Macosx bundles are named osx and Windows demo bundles use zip instead of
+    // Macosx bundles are named osx
     // tar.gz.
     var artifactData = {
         "linux-x64": {
             platform: "linux-x64",
-            demo_ext: "tar.gz"
         },
         "linux-x86": {
             platform: "linux-x86",
-            demo_ext: "tar.gz"
         },
         "macosx-x64": {
             platform: "osx-x64",
-            demo_ext: "tar.gz"
+            jdk_subdir: "jdk-" + data.version +  ".jdk/Contents/Home",
+            jre_subdir: "jre-" + data.version +  ".jre/Contents/Home"
         },
         "solaris-x64": {
             platform: "solaris-x64",
-            demo_ext: "tar.gz"
         },
         "solaris-sparcv9": {
             platform: "solaris-sparcv9",
-            demo_ext: "tar.gz"
         },
         "windows-x64": {
             platform: "windows-x64",
-            demo_ext: "zip"
         },
         "windows-x86": {
             platform: "windows-x86",
-            demo_ext: "zip"
+        },
+       "linux-arm64": {
+            platform: "linux-arm64-vfp-hflt",
+        },
+        "linux-arm-vfp-hflt": {
+            platform: "linux-arm32-vfp-hflt",
+        },
+        "linux-arm-vfp-hflt-dyn": {
+            platform: "linux-arm32-vfp-hflt-dyn",
         }
     }
     // Generate common artifacts for all main profiles
-    common.main_profile_names.forEach(function (name) {
+    Object.keys(artifactData).forEach(function (name) {
         profiles[name] = concatObjects(profiles[name],
-            common.main_profile_artifacts(artifactData[name].platform, artifactData[name].demo_ext));
+            common.main_profile_artifacts(artifactData[name]));
     });
 
     // Generate common artifacts for all debug profiles
-    common.main_profile_names.forEach(function (name) {
+    Object.keys(artifactData).forEach(function (name) {
         var debugName = name + common.debug_suffix;
         profiles[debugName] = concatObjects(profiles[debugName],
-            common.debug_profile_artifacts(artifactData[name].platform));
+            common.debug_profile_artifacts(artifactData[name]));
     });
 
     // Extra profile specific artifacts
@@ -689,7 +727,12 @@
             artifacts: {
                 jdk: {
                     local: "bundles/\\(jdk.*bin.tar.gz\\)",
-                    remote: "bundles/openjdk/GPL/linux-x64/\\1",
+                    remote: [
+                        "bundles/openjdk/GPL/linux-x64/jdk-" + data.version
+                            + "_linux-x64_bin.tar.gz",
+                        "bundles/openjdk/GPL/linux-x64/\\1"
+                    ],
+                    subdir: "jdk-" + data.version
                 },
                 jre: {
                     local: "bundles/\\(jre.*bin.tar.gz\\)",
@@ -697,20 +740,25 @@
                 },
                 test: {
                     local: "bundles/\\(jdk.*bin-tests.tar.gz\\)",
-                    remote: "bundles/openjdk/GPL/linux-x64/\\1",
+                    remote: [
+                        "bundles/openjdk/GPL/linux-x64/jdk-" + data.version
+                            + "_linux-x64_bin-tests.tar.gz",
+                        "bundles/openjdk/GPL/linux-x64/\\1"
+                    ]
                 },
                 jdk_symbols: {
                     local: "bundles/\\(jdk.*bin-symbols.tar.gz\\)",
-                    remote: "bundles/openjdk/GPL/linux-x64/\\1",
+                    remote: [
+                        "bundles/openjdk/GPL/linux-x64/jdk-" + data.version
+                            + "_linux-x64_bin-symbols.tar.gz",
+                        "bundles/openjdk/GPL/linux-x64/\\1"
+                    ],
+                    subdir: "jdk-" + data.version
                 },
                 jre_symbols: {
                     local: "bundles/\\(jre.*bin-symbols.tar.gz\\)",
                     remote: "bundles/openjdk/GPL/linux-x64/\\1",
                 },
-                demo: {
-                    local: "bundles/\\(jdk.*demo.tar.gz\\)",
-                    remote: "bundles/openjdk/GPL/linux-x64/\\1",
-                },
                 doc_api_spec: {
                     local: "bundles/\\(jdk.*doc-api-spec.tar.gz\\)",
                     remote: "bundles/openjdk/GPL/linux-x64/\\1",
@@ -722,11 +770,29 @@
             artifacts: {
                 jdk: {
                     local: "bundles/\\(jdk.*bin.tar.gz\\)",
-                    remote: "bundles/openjdk/GPL/profile/linux-x86/\\1",
+                    remote: [
+                        "bundles/openjdk/GPL/linux-x86/jdk-" + data.version
+                            + "_linux-x86_bin.tar.gz",
+                        "bundles/openjdk/GPL/linux-x86/\\1"
+                    ],
+                    subdir: "jdk-" + data.version
                 },
                 jdk_symbols: {
                     local: "bundles/\\(jdk.*bin-symbols.tar.gz\\)",
-                    remote: "bundles/openjdk/GPL/profile/linux-x86/\\1",
+                    remote: [
+                        "bundles/openjdk/GPL/linux-x86/jdk-" + data.version
+                            + "_linux-x86_bin-symbols.tar.gz",
+                        "bundles/openjdk/GPL/linux-x86/\\1"
+                    ],
+                    subdir: "jdk-" + data.version
+                },
+                test: {
+                    local: "bundles/\\(jdk.*bin-tests.tar.gz\\)",
+                    remote: [
+                        "bundles/openjdk/GPL/linux-x86/jdk-" + data.version
+                            + "_linux-x86_bin-tests.tar.gz",
+                        "bundles/openjdk/GPL/linux-x86/\\1"
+                    ]
                 },
                 jre: {
                     // This regexp needs to not match the compact* files below
@@ -752,7 +818,12 @@
             artifacts: {
                 jdk: {
                     local: "bundles/\\(jdk.*bin.tar.gz\\)",
-                    remote: "bundles/openjdk/GPL/windows-x86/\\1",
+                    remote: [
+                        "bundles/openjdk/GPL/windows-x86/jdk-" + data.version
+                            + "_windows-x86_bin.tar.gz",
+                        "bundles/openjdk/GPL/windows-x86/\\1"
+                    ],
+                    subdir: "jdk-" + data.version
                 },
                 jre: {
                     local: "bundles/\\(jre.*bin.tar.gz\\)",
@@ -760,19 +831,24 @@
                 },
                 test: {
                     local: "bundles/\\(jdk.*bin-tests.tar.gz\\)",
-                    remote: "bundles/openjdk/GPL/windows-x86/\\1",
+                    remote: [
+                        "bundles/openjdk/GPL/windows-x86/jdk-" + data.version
+                            + "_windows-x86_bin-tests.tar.gz",
+                        "bundles/openjdk/GPL/windows-x86/\\1"
+                    ]
                 },
                 jdk_symbols: {
                     local: "bundles/\\(jdk.*bin-symbols.tar.gz\\)",
-                    remote: "bundles/openjdk/GPL/windows-x86/\\1"
+                    remote: [
+                        "bundles/openjdk/GPL/windows-x86/jdk-" + data.version
+                            + "_windows-x86_bin-symbols.tar.gz",
+                        "bundles/openjdk/GPL/windows-x86/\\1"
+                    ],
+                    subdir: "jdk-" + data.version
                 },
                 jre_symbols: {
                     local: "bundles/\\(jre.*bin-symbols.tar.gz\\)",
                     remote: "bundles/openjdk/GPL/windows-x86/\\1",
-                },
-                demo: {
-                    local: "bundles/\\(jdk.*demo.zip\\)",
-                    remote: "bundles/openjdk/GPL/windows-x86/\\1",
                 }
             }
         },
@@ -817,6 +893,16 @@
             }
         });
 
+    // The windows ri profile needs to add the freetype license file
+    profilesRiFreetype = {
+        "windows-x86-ri": {
+            configure_args: "--with-freetype-license="
+                + input.get("freetype", "install_path")
+                + "/freetype-2.7.1-v120-x86/freetype.md"
+        }
+    };
+    profiles = concatObjects(profiles, profilesRiFreetype);
+
     // Generate the missing platform attributes
     profiles = generatePlatformAttributes(profiles);
     profiles = generateDefaultMakeTargetsConfigureArg(common, profiles);
@@ -832,15 +918,16 @@
  */
 var getJibProfilesDependencies = function (input, common) {
 
-    var boot_jdk_platform = input.build_os + "-"
-        + (input.build_cpu == "x86" ? "i586" : input.build_cpu);
-
     var devkit_platform_revisions = {
         linux_x64: "gcc4.9.2-OEL6.4+1.1",
         macosx_x64: "Xcode6.3-MacOSX10.9+1.0",
         solaris_x64: "SS12u4-Solaris11u1+1.0",
         solaris_sparcv9: "SS12u4-Solaris11u1+1.0",
-        windows_x64: "VS2013SP4+1.0"
+        windows_x64: "VS2013SP4+1.0",
+        linux_aarch64: "gcc-linaro-aarch64-linux-gnu-4.8-2013.11_linux+1.0",
+        linux_arm: (input.profile != null && input.profile.indexOf("hflt") >= 0
+                    ? "gcc-linaro-arm-linux-gnueabihf-raspbian-2012.09-20120921_linux+1.0"
+                    : "arm-linaro-4.7+1.0")
     };
 
     var devkit_platform = (input.target_cpu == "x86"
@@ -853,9 +940,9 @@
             server: "javare",
             module: "jdk",
             revision: common.boot_jdk_revision,
-            checksum_file: boot_jdk_platform + "/MD5_VALUES",
-            file: boot_jdk_platform + "/jdk-" + common.boot_jdk_revision
-                + "-" + boot_jdk_platform + ".tar.gz",
+            checksum_file: common.boot_jdk_platform + "/MD5_VALUES",
+            file: common.boot_jdk_platform + "/jdk-" + common.boot_jdk_revision
+                + "-" + common.boot_jdk_platform + ".tar.gz",
             configure_args: "--with-boot-jdk=" + common.boot_jdk_home,
             environment_path: common.boot_jdk_home + "/bin"
         },
@@ -883,7 +970,7 @@
         jtreg: {
             server: "javare",
             revision: "4.2",
-            build_number: "b05",
+            build_number: "b07",
             checksum_file: "MD5_VALUES",
             file: "jtreg_bin-4.2.zip",
             environment_name: "JT_HOME",
@@ -911,9 +998,27 @@
         freetype: {
             organization: common.organization,
             ext: "tar.gz",
-            revision: "2.3.4+1.0",
+            revision: "2.7.1-v120+1.0",
             module: "freetype-" + input.target_platform
-        }
+        },
+
+        graphviz: {
+            organization: common.organization,
+            ext: "tar.gz",
+            revision: "2.38.0-1+1.1",
+            module: "graphviz-" + input.target_platform,
+            configure_args: "DOT=" + input.get("graphviz", "install_path") + "/dot",
+            environment_path: input.get("graphviz", "install_path")
+        },
+
+        pandoc: {
+            organization: common.organization,
+            ext: "tar.gz",
+            revision: "1.17.2+1.0",
+            module: "pandoc-" + input.target_platform,
+            configure_args: "PANDOC=" + input.get("pandoc", "install_path") + "/pandoc/pandoc",
+            environment_path: input.get("pandoc", "install_path") + "/pandoc"
+        },
     };
 
     return dependencies;
@@ -1080,12 +1185,29 @@
         + "." + (minor != null ? minor : version_numbers.get("DEFAULT_VERSION_MINOR"))
         + "." + (security != null ? security :  version_numbers.get("DEFAULT_VERSION_SECURITY"))
         + "." + (patch != null ? patch : version_numbers.get("DEFAULT_VERSION_PATCH"));
-    while (version.match(".*\.0$")) {
+    while (version.match(".*\\.0$")) {
         version = version.substring(0, version.length - 2);
     }
     return version;
 };
 
+/**
+ * Constructs the common version configure args based on build type and
+ * other version inputs
+ */
+var versionArgs = function(input, common) {
+    var args = ["--with-version-build=" + common.build_number];
+    if (input.build_type == "promoted") {
+        args = concat(args,
+                      // This needs to be changed when we start building release candidates
+                      "--with-version-pre=ea",
+                      "--without-version-opt");
+    } else {
+        args = concat(args, "--with-version-opt=" + common.build_id);
+    }
+    return args;
+}
+
 // Properties representation of the common/autoconf/version-numbers file. Lazily
 // initiated by the function below.
 var version_numbers;
--- /dev/null	Thu Jan 01 00:00:00 1970 +0000
+++ b/common/doc/building.html	Thu Aug 17 12:17:39 2017 -0700
@@ -0,0 +1,960 @@
+<!DOCTYPE html>
+<html>
+<head>
+  <meta charset="utf-8">
+  <meta name="generator" content="pandoc">
+  <meta name="viewport" content="width=device-width, initial-scale=1.0, user-scalable=yes">
+  <title>Building OpenJDK</title>
+  <style type="text/css">code{white-space: pre;}</style>
+  <link rel="stylesheet" href="../../jdk/make/data/docs-resources/resources/jdk-default.css">
+  <!--[if lt IE 9]>
+    <script src="//cdnjs.cloudflare.com/ajax/libs/html5shiv/3.7.3/html5shiv-printshiv.min.js"></script>
+  <![endif]-->
+  <style type="text/css">pre, code, tt { color: #1d6ae5; }</style>
+</head>
+<body>
+<header>
+<h1 class="title">Building OpenJDK</h1>
+</header>
+<nav id="TOC">
+<ul>
+<li><a href="#tldr-instructions-for-the-impatient">TL;DR (Instructions for the Impatient)</a></li>
+<li><a href="#introduction">Introduction</a></li>
+<li><a href="#getting-the-source-code">Getting the Source Code</a><ul>
+<li><a href="#special-considerations">Special Considerations</a></li>
+<li><a href="#using-get_source.sh">Using get_source.sh</a></li>
+<li><a href="#using-hgforest.sh">Using hgforest.sh</a></li>
+<li><a href="#using-the-trees-extension">Using the Trees Extension</a></li>
+</ul></li>
+<li><a href="#build-hardware-requirements">Build Hardware Requirements</a><ul>
+<li><a href="#building-on-x86">Building on x86</a></li>
+<li><a href="#building-on-sparc">Building on sparc</a></li>
+<li><a href="#building-on-armaarch64">Building on arm/aarch64</a></li>
+</ul></li>
+<li><a href="#operating-system-requirements">Operating System Requirements</a><ul>
+<li><a href="#windows">Windows</a></li>
+<li><a href="#solaris">Solaris</a></li>
+<li><a href="#macos">macOS</a></li>
+<li><a href="#linux">Linux</a></li>
+<li><a href="#aix">AIX</a></li>
+</ul></li>
+<li><a href="#native-compiler-toolchain-requirements">Native Compiler (Toolchain) Requirements</a><ul>
+<li><a href="#gcc">gcc</a></li>
+<li><a href="#clang">clang</a></li>
+<li><a href="#apple-xcode">Apple Xcode</a></li>
+<li><a href="#oracle-solaris-studio">Oracle Solaris Studio</a></li>
+<li><a href="#microsoft-visual-studio">Microsoft Visual Studio</a></li>
+<li><a href="#ibm-xl-cc">IBM XL C/C++</a></li>
+</ul></li>
+<li><a href="#boot-jdk-requirements">Boot JDK Requirements</a><ul>
+<li><a href="#jdk-8-on-linux">JDK 8 on Linux</a></li>
+<li><a href="#jdk-8-on-windows">JDK 8 on Windows</a></li>
+<li><a href="#jdk-8-on-macos">JDK 8 on macOS</a></li>
+<li><a href="#jdk-8-on-aix">JDK 8 on AIX</a></li>
+</ul></li>
+<li><a href="#external-library-requirements">External Library Requirements</a><ul>
+<li><a href="#freetype">FreeType</a></li>
+<li><a href="#cups">CUPS</a></li>
+<li><a href="#x11">X11</a></li>
+<li><a href="#alsa">ALSA</a></li>
+<li><a href="#libffi">libffi</a></li>
+<li><a href="#libelf">libelf</a></li>
+</ul></li>
+<li><a href="#other-tooling-requirements">Other Tooling Requirements</a><ul>
+<li><a href="#gnu-make">GNU Make</a></li>
+<li><a href="#gnu-bash">GNU Bash</a></li>
+<li><a href="#autoconf">Autoconf</a></li>
+</ul></li>
+<li><a href="#running-configure">Running Configure</a><ul>
+<li><a href="#common-configure-arguments">Common Configure Arguments</a></li>
+<li><a href="#configure-control-variables">Configure Control Variables</a></li>
+</ul></li>
+<li><a href="#running-make">Running Make</a><ul>
+<li><a href="#common-make-targets">Common Make Targets</a></li>
+<li><a href="#make-control-variables">Make Control Variables</a></li>
+</ul></li>
+<li><a href="#running-tests">Running Tests</a></li>
+<li><a href="#cross-compiling">Cross-compiling</a><ul>
+<li><a href="#boot-jdk-and-build-jdk">Boot JDK and Build JDK</a></li>
+<li><a href="#specifying-the-target-platform">Specifying the Target Platform</a></li>
+<li><a href="#toolchain-considerations">Toolchain Considerations</a></li>
+<li><a href="#native-libraries">Native Libraries</a></li>
+<li><a href="#building-for-armaarch64">Building for ARM/aarch64</a></li>
+<li><a href="#verifying-the-build">Verifying the Build</a></li>
+</ul></li>
+<li><a href="#build-performance">Build Performance</a><ul>
+<li><a href="#disk-speed">Disk Speed</a></li>
+<li><a href="#virus-checking">Virus Checking</a></li>
+<li><a href="#ccache">Ccache</a></li>
+<li><a href="#precompiled-headers">Precompiled Headers</a></li>
+<li><a href="#icecc-icecream">Icecc / icecream</a></li>
+<li><a href="#using-sjavac">Using sjavac</a></li>
+<li><a href="#building-the-right-target">Building the Right Target</a></li>
+</ul></li>
+<li><a href="#troubleshooting">Troubleshooting</a><ul>
+<li><a href="#locating-the-source-of-the-error">Locating the Source of the Error</a></li>
+<li><a href="#fixing-unexpected-build-failures">Fixing Unexpected Build Failures</a></li>
+<li><a href="#specific-build-issues">Specific Build Issues</a></li>
+<li><a href="#getting-help">Getting Help</a></li>
+</ul></li>
+<li><a href="#hints-and-suggestions-for-advanced-users">Hints and Suggestions for Advanced Users</a><ul>
+<li><a href="#setting-up-a-forest-for-pushing-changes-defpath">Setting Up a Forest for Pushing Changes (defpath)</a></li>
+<li><a href="#bash-completion">Bash Completion</a></li>
+<li><a href="#using-multiple-configurations">Using Multiple Configurations</a></li>
+<li><a href="#handling-reconfigurations">Handling Reconfigurations</a></li>
+<li><a href="#using-fine-grained-make-targets">Using Fine-Grained Make Targets</a></li>
+<li><a href="#learn-about-mercurial">Learn About Mercurial</a></li>
+</ul></li>
+<li><a href="#understanding-the-build-system">Understanding the Build System</a><ul>
+<li><a href="#configurations">Configurations</a></li>
+<li><a href="#build-output-structure">Build Output Structure</a></li>
+<li><a href="#fixpath">Fixpath</a></li>
+<li><a href="#native-debug-symbols">Native Debug Symbols</a></li>
+<li><a href="#autoconf-details">Autoconf Details</a></li>
+<li><a href="#developing-the-build-system-itself">Developing the Build System Itself</a></li>
+</ul></li>
+<li><a href="#contributing-to-openjdk">Contributing to OpenJDK</a></li>
+</ul>
+</nav>
+<h2 id="tldr-instructions-for-the-impatient">TL;DR (Instructions for the Impatient)</h2>
+<p>If you are eager to try out building OpenJDK, these simple steps works most of the time. They assume that you have installed Mercurial (and Cygwin if running on Windows) and cloned the top-level OpenJDK repository that you want to build.</p>
+<ol type="1">
+<li><p><a href="#getting-the-source-code">Get the complete source code</a>:<br />
+<code>bash get_source.sh</code></p></li>
+<li><p><a href="#running-configure">Run configure</a>:<br />
+<code>bash configure</code></p>
+<p>If <code>configure</code> fails due to missing dependencies (to either the <a href="#native-compiler-toolchain-requirements">toolchain</a>, <a href="#external-library-requirements">external libraries</a> or the <a href="#boot-jdk-requirements">boot JDK</a>), most of the time it prints a suggestion on how to resolve the situation on your platform. Follow the instructions, and try running <code>bash configure</code> again.</p></li>
+<li><p><a href="#running-make">Run make</a>:<br />
+<code>make images</code></p></li>
+<li><p>Verify your newly built JDK:<br />
+<code>./build/*/images/jdk/bin/java -version</code></p></li>
+<li><p><a href="##running-tests">Run basic tests</a>:<br />
+<code>make run-test-tier1</code></p></li>
+</ol>
+<p>If any of these steps failed, or if you want to know more about build requirements or build functionality, please continue reading this document.</p>
+<h2 id="introduction">Introduction</h2>
+<p>OpenJDK is a complex software project. Building it requires a certain amount of technical expertise, a fair number of dependencies on external software, and reasonably powerful hardware.</p>
+<p>If you just want to use OpenJDK and not build it yourself, this document is not for you. See for instance <a href="http://openjdk.java.net/install">OpenJDK installation</a> for some methods of installing a prebuilt OpenJDK.</p>
+<h2 id="getting-the-source-code">Getting the Source Code</h2>
+<p>OpenJDK uses <a href="http://www.mercurial-scm.org">Mercurial</a> for source control. The source code is contained not in a single Mercurial repository, but in a tree (&quot;forest&quot;) of interrelated repositories. You will need to check out all of the repositories to be able to build OpenJDK. To assist you in dealing with this somewhat unusual arrangement, there are multiple tools available, which are explained below.</p>
+<p>In any case, make sure you are getting the correct version. At the <a href="http://hg.openjdk.java.net/">OpenJDK Mercurial server</a> you can see a list of all available forests. If you want to build an older version, e.g. JDK 8, it is recommended that you get the <code>jdk8u</code> forest, which contains incremental updates, instead of the <code>jdk8</code> forest, which was frozen at JDK 8 GA.</p>
+<p>If you are new to Mercurial, a good place to start is the <a href="http://www.mercurial-scm.org/guide">Mercurial Beginner's Guide</a>. The rest of this document assumes a working knowledge of Mercurial.</p>
+<h3 id="special-considerations">Special Considerations</h3>
+<p>For a smooth building experience, it is recommended that you follow these rules on where and how to check out the source code.</p>
+<ul>
+<li><p>Do not check out the source code in a path which contains spaces. Chances are the build will not work. This is most likely to be an issue on Windows systems.</p></li>
+<li><p>Do not check out the source code in a path which has a very long name or is nested many levels deep. Chances are you will hit an OS limitation during the build.</p></li>
+<li><p>Put the source code on a local disk, not a network share. If possible, use an SSD. The build process is very disk intensive, and having slow disk access will significantly increase build times. If you need to use a network share for the source code, see below for suggestions on how to keep the build artifacts on a local disk.</p></li>
+<li><p>On Windows, extra care must be taken to make sure the <a href="#cygwin">Cygwin</a> environment is consistent. It is recommended that you follow this procedure:</p>
+<ul>
+<li><p>Create the directory that is going to contain the top directory of the OpenJDK clone by using the <code>mkdir</code> command in the Cygwin bash shell. That is, do <em>not</em> create it using Windows Explorer. This will ensure that it will have proper Cygwin attributes, and that it's children will inherit those attributes.</p></li>
+<li><p>Do not put the OpenJDK clone in a path under your Cygwin home directory. This is especially important if your user name contains spaces and/or mixed upper and lower case letters.</p></li>
+<li><p>Clone the OpenJDK repository using the Cygwin command line <code>hg</code> client as instructed in this document. That is, do <em>not</em> use another Mercurial client such as TortoiseHg.</p></li>
+</ul>
+<p>Failure to follow this procedure might result in hard-to-debug build problems.</p></li>
+</ul>
+<h3 id="using-get_source.sh">Using get_source.sh</h3>
+<p>The simplest way to get the entire forest is probably to clone the top-level repository and then run the <code>get_source.sh</code> script, like this:</p>
+<pre><code>hg clone http://hg.openjdk.java.net/jdk9/jdk9
+cd jdk9
+bash get_source.sh</code></pre>
+<p>The first time this is run, it will clone all the sub-repositories. Any subsequent execution of the script will update all sub-repositories to the latest revision.</p>
+<h3 id="using-hgforest.sh">Using hgforest.sh</h3>
+<p>The <code>hgforest.sh</code> script is more expressive than <code>get_source.sh</code>. It takes any number of arguments, and runs <code>hg</code> with those arguments on each sub-repository in the forest. The <code>get_source.sh</code> script is basically a simple wrapper that runs either <code>hgforest.sh clone</code> or <code>hgforest.sh pull -u</code>.</p>
+<ul>
+<li><p>Cloning the forest:</p>
+<pre><code>hg clone http://hg.openjdk.java.net/jdk9/jdk9
+cd jdk9
+bash common/bin/hgforest.sh clone</code></pre></li>
+<li><p>Pulling and updating the forest:</p>
+<pre><code>bash common/bin/hgforest.sh pull -u</code></pre></li>
+<li><p>Merging over the entire forest:</p>
+<pre><code>bash common/bin/hgforest.sh merge</code></pre></li>
+</ul>
+<h3 id="using-the-trees-extension">Using the Trees Extension</h3>
+<p>The trees extension is a Mercurial add-on that helps you deal with the forest. More information is available on the <a href="http://openjdk.java.net/projects/code-tools/trees">Code Tools trees page</a>.</p>
+<h4 id="installing-the-extension">Installing the Extension</h4>
+<p>Install the extension by cloning <code>http://hg.openjdk.java.net/code-tools/trees</code> and updating your <code>.hgrc</code> file. Here's one way to do this:</p>
+<pre><code>cd ~
+mkdir hg-ext
+cd hg-ext
+hg clone http://hg.openjdk.java.net/code-tools/trees
+cat &lt;&lt; EOT &gt;&gt; ~/.hgrc
+[extensions]
+trees=~/hg-ext/trees/trees.py
+EOT</code></pre>
+<h4 id="initializing-the-tree">Initializing the Tree</h4>
+<p>The trees extension needs to know the structure of the forest. If you have already cloned the entire forest using another method, you can initialize the forest like this:</p>
+<pre><code>hg tconf --set --walk --depth</code></pre>
+<p>Or you can clone the entire forest at once, if you substitute <code>clone</code> with <code>tclone</code> when cloning the top-level repository, e.g. like this:</p>
+<pre><code>hg tclone http://hg.openjdk.java.net/jdk9/jdk9</code></pre>
+<p>In this case, the forest will be properly initialized from the start.</p>
+<h4 id="other-operations">Other Operations</h4>
+<p>The trees extensions supplement many common operations with a trees version by prefixing a <code>t</code> to the normal Mercurial command, e.g. <code>tcommit</code>, <code>tstatus</code> or <code>tmerge</code>. For instance, to update the entire forest:</p>
+<pre><code>hg tpull -u</code></pre>
+<h2 id="build-hardware-requirements">Build Hardware Requirements</h2>
+<p>OpenJDK is a massive project, and require machines ranging from decent to powerful to be able to build in a reasonable amount of time, or to be able to complete a build at all.</p>
+<p>We <em>strongly</em> recommend usage of an SSD disk for the build, since disk speed is one of the limiting factors for build performance.</p>
+<h3 id="building-on-x86">Building on x86</h3>
+<p>At a minimum, a machine with 2-4 cores is advisable, as well as 2-4 GB of RAM. (The more cores to use, the more memory you need.) At least 6 GB of free disk space is required (8 GB minimum for building on Solaris).</p>
+<p>Even for 32-bit builds, it is recommended to use a 64-bit build machine, and instead create a 32-bit target using <code>--with-target-bits=32</code>.</p>
+<h3 id="building-on-sparc">Building on sparc</h3>
+<p>At a minimum, a machine with 4 cores is advisable, as well as 4 GB of RAM. (The more cores to use, the more memory you need.) At least 8 GB of free disk space is required.</p>
+<h3 id="building-on-armaarch64">Building on arm/aarch64</h3>
+<p>This is not recommended. Instead, see the section on <a href="#cross-compiling">Cross-compiling</a>.</p>
+<h2 id="operating-system-requirements">Operating System Requirements</h2>
+<p>The mainline OpenJDK project supports Linux, Solaris, macOS, AIX and Windows. Support for other operating system, e.g. BSD, exists in separate &quot;port&quot; projects.</p>
+<p>In general, OpenJDK can be built on a wide range of versions of these operating systems, but the further you deviate from what is tested on a daily basis, the more likely you are to run into problems.</p>
+<p>This table lists the OS versions used by Oracle when building JDK 9. Such information is always subject to change, but this table is up to date at the time of writing.</p>
+<table>
+<thead>
+<tr class="header">
+<th style="text-align: left;">Operating system</th>
+<th style="text-align: left;">Vendor/version used</th>
+</tr>
+</thead>
+<tbody>
+<tr class="odd">
+<td style="text-align: left;">Linux</td>
+<td style="text-align: left;">Oracle Enterprise Linux 6.4 / 7.1 (using kernel 3.8.13)</td>
+</tr>
+<tr class="even">
+<td style="text-align: left;">Solaris</td>
+<td style="text-align: left;">Solaris 11.1 SRU 21.4.1 / 11.2 SRU 5.5</td>
+</tr>
+<tr class="odd">
+<td style="text-align: left;">macOS</td>
+<td style="text-align: left;">Mac OS X 10.9 (Mavericks) / 10.10 (Yosemite)</td>
+</tr>
+<tr class="even">
+<td style="text-align: left;">Windows</td>
+<td style="text-align: left;">Windows Server 2012 R2</td>
+</tr>
+</tbody>
+</table>
+<p>The double version numbers for Linux, Solaris and macOS is due to the hybrid model used at Oracle, where header files and external libraries from an older version is used when building on a more modern version of the OS.</p>
+<p>The Build Group has a wiki page with <a href="https://wiki.openjdk.java.net/display/Build/Supported+Build+Platforms">Supported Build Platforms</a>. From time to time, this is updated by the community to list successes or failures of building on different platforms.</p>
+<h3 id="windows">Windows</h3>
+<p>Windows XP is not a supported platform, but all newer Windows should be able to build OpenJDK.</p>
+<p>On Windows, it is important that you pay attention to the instructions in the <a href="#special-considerations">Special Considerations</a>.</p>
+<p>Windows is the only non-POSIX OS supported by OpenJDK, and as such, requires some extra care. A POSIX support layer is required to build on Windows. For OpenJDK 9, the only supported such layer is Cygwin. (Msys is no longer supported due to a too old bash; msys2 and the new Windows Subsystem for Linux (WSL) would likely be possible to support in a future version but that would require a community effort to implement.)</p>
+<p>Internally in the build system, all paths are represented as Unix-style paths, e.g. <code>/cygdrive/c/hg/jdk9/Makefile</code> rather than <code>C:\hg\jdk9\Makefile</code>. This rule also applies to input to the build system, e.g. in arguments to <code>configure</code>. So, use <code>--with-freetype=/cygdrive/c/freetype</code> rather than <code>--with-freetype=c:\freetype</code>. For details on this conversion, see the section on <a href="#fixpath">Fixpath</a>.</p>
+<h4 id="cygwin">Cygwin</h4>
+<p>A functioning <a href="http://www.cygwin.com/">Cygwin</a> environment is thus required for building OpenJDK on Windows. If you have a 64-bit OS, we strongly recommend using the 64-bit version of Cygwin.</p>
+<p><strong>Note:</strong> Cygwin has a model of continuously updating all packages without any easy way to install or revert to a specific version of a package. This means that whenever you add or update a package in Cygwin, you might (inadvertently) update tools that are used by the OpenJDK build process, and that can cause unexpected build problems.</p>
+<p>OpenJDK requires GNU Make 4.0 or greater on Windows. This is usually not a problem, since Cygwin currently only distributes GNU Make at a version above 4.0.</p>
+<p>Apart from the basic Cygwin installation, the following packages must also be installed:</p>
+<ul>
+<li><code>make</code></li>
+<li><code>zip</code></li>
+<li><code>unzip</code></li>
+</ul>
+<p>Often, you can install these packages using the following command line:</p>
+<pre><code>&lt;path to Cygwin setup&gt;/setup-x86_64 -q -P make -P unzip -P zip</code></pre>
+<p>Unfortunately, Cygwin can be unreliable in certain circumstances. If you experience build tool crashes or strange issues when building on Windows, please check the Cygwin FAQ on the <a href="https://cygwin.com/faq/faq.html#faq.using.bloda">&quot;BLODA&quot; list</a> and the section on <a href="https://cygwin.com/faq/faq.html#faq.using.fixing-fork-failures">fork() failures</a>.</p>
+<h3 id="solaris">Solaris</h3>
+<p>See <code>make/devkit/solaris11.1-package-list.txt</code> for a list of recommended packages to install when building on Solaris. The versions specified in this list is the versions used by the daily builds at Oracle, and is likely to work properly.</p>
+<p>Older versions of Solaris shipped a broken version of <code>objcopy</code>. At least version 2.21.1 is needed, which is provided by Solaris 11 Update 1. Objcopy is needed if you want to have external debug symbols. Please make sure you are using at least version 2.21.1 of objcopy, or that you disable external debug symbols.</p>
+<h3 id="macos">macOS</h3>
+<p>Apple is using a quite aggressive scheme of pushing OS updates, and coupling these updates with required updates of Xcode. Unfortunately, this makes it difficult for a project like OpenJDK to keep pace with a continuously updated machine running macOS. See the section on <a href="#apple-xcode">Apple Xcode</a> on some strategies to deal with this.</p>
+<p>It is recommended that you use at least Mac OS X 10.9 (Mavericks). At the time of writing, OpenJDK has been successfully compiled on macOS versions up to 10.12.5 (Sierra), using XCode 8.3.2 and <code>--disable-warnings-as-errors</code>.</p>
+<p>The standard macOS environment contains the basic tooling needed to build, but for external libraries a package manager is recommended. OpenJDK uses <a href="https://brew.sh/">homebrew</a> in the examples, but feel free to use whatever manager you want (or none).</p>
+<h3 id="linux">Linux</h3>
+<p>It is often not much problem to build OpenJDK on Linux. The only general advice is to try to use the compilers, external libraries and header files as provided by your distribution.</p>
+<p>The basic tooling is provided as part of the core operating system, but you will most likely need to install developer packages.</p>
+<p>For apt-based distributions (Debian, Ubuntu, etc), try this:</p>
+<pre><code>sudo apt-get install build-essential</code></pre>
+<p>For rpm-based distributions (Fedora, Red Hat, etc), try this:</p>
+<pre><code>sudo yum groupinstall &quot;Development Tools&quot;</code></pre>
+<h3 id="aix">AIX</h3>
+<p>The regular builds by SAP is using AIX version 7.1, but AIX 5.3 is also supported. See the <a href="http://cr.openjdk.java.net/~simonis/ppc-aix-port">OpenJDK PowerPC Port Status Page</a> for details.</p>
+<h2 id="native-compiler-toolchain-requirements">Native Compiler (Toolchain) Requirements</h2>
+<p>Large portions of OpenJDK consists of native code, that needs to be compiled to be able to run on the target platform. In theory, toolchain and operating system should be independent factors, but in practice there's more or less a one-to-one correlation between target operating system and toolchain.</p>
+<table>
+<thead>
+<tr class="header">
+<th style="text-align: left;">Operating system</th>
+<th style="text-align: left;">Supported toolchain</th>
+</tr>
+</thead>
+<tbody>
+<tr class="odd">
+<td style="text-align: left;">Linux</td>
+<td style="text-align: left;">gcc, clang</td>
+</tr>
+<tr class="even">
+<td style="text-align: left;">macOS</td>
+<td style="text-align: left;">Apple Xcode (using clang)</td>
+</tr>
+<tr class="odd">
+<td style="text-align: left;">Solaris</td>
+<td style="text-align: left;">Oracle Solaris Studio</td>
+</tr>
+<tr class="even">
+<td style="text-align: left;">AIX</td>
+<td style="text-align: left;">IBM XL C/C++</td>
+</tr>
+<tr class="odd">
+<td style="text-align: left;">Windows</td>
+<td style="text-align: left;">Microsoft Visual Studio</td>
+</tr>
+</tbody>
+</table>
+<p>Please see the individual sections on the toolchains for version recommendations. As a reference, these versions of the toolchains are used, at the time of writing, by Oracle for the daily builds of OpenJDK. It should be possible to compile OpenJDK with both older and newer versions, but the closer you stay to this list, the more likely you are to compile successfully without issues.</p>
+<table>
+<thead>
+<tr class="header">
+<th style="text-align: left;">Operating system</th>
+<th style="text-align: left;">Toolchain version</th>
+</tr>
+</thead>
+<tbody>
+<tr class="odd">
+<td style="text-align: left;">Linux</td>
+<td style="text-align: left;">gcc 4.9.2</td>
+</tr>
+<tr class="even">
+<td style="text-align: left;">macOS</td>
+<td style="text-align: left;">Apple Xcode 6.3 (using clang 6.1.0)</td>
+</tr>
+<tr class="odd">
+<td style="text-align: left;">Solaris</td>
+<td style="text-align: left;">Oracle Solaris Studio 12.4 (with compiler version 5.13)</td>
+</tr>
+<tr class="even">
+<td style="text-align: left;">Windows</td>
+<td style="text-align: left;">Microsoft Visual Studio 2013 update 4</td>
+</tr>
+</tbody>
+</table>
+<h3 id="gcc">gcc</h3>
+<p>The minimum accepted version of gcc is 4.3. Older versions will not be accepted by <code>configure</code>.</p>
+<p>However, gcc 4.3 is quite old and OpenJDK is not regularly tested on this version, so it is recommended to use a more modern gcc.</p>
+<p>OpenJDK 9 includes patches that should allow gcc 6 to compile, but this should be considered experimental.</p>
+<p>In general, any version between these two should be usable.</p>
+<h3 id="clang">clang</h3>
+<p>The minimum accepted version of clang is 3.2. Older versions will not be accepted by <code>configure</code>.</p>
+<p>To use clang instead of gcc on Linux, use <code>--with-toolchain-type=clang</code>.</p>
+<h3 id="apple-xcode">Apple Xcode</h3>
+<p>The oldest supported version of Xcode is 5.</p>
+<p>You will need the Xcode command lines developers tools to be able to build OpenJDK. (Actually, <em>only</em> the command lines tools are needed, not the IDE.) The simplest way to install these is to run:</p>
+<pre><code>xcode-select --install</code></pre>
+<p>It is advisable to keep an older version of Xcode for building OpenJDK when updating Xcode. This <a href="http://iosdevelopertips.com/xcode/install-multiple-versions-of-xcode.html">blog page</a> has good suggestions on managing multiple Xcode versions. To use a specific version of Xcode, use <code>xcode-select -s</code> before running <code>configure</code>, or use <code>--with-toolchain-path</code> to point to the version of Xcode to use, e.g. <code>configure --with-toolchain-path=/Applications/Xcode5.app/Contents/Developer/usr/bin</code></p>
+<p>If you have recently (inadvertently) updated your OS and/or Xcode version, and OpenJDK can no longer be built, please see the section on <a href="#problems-with-the-build-environment">Problems with the Build Environment</a>, and <a href="#getting-help">Getting Help</a> to find out if there are any recent, non-merged patches available for this update.</p>
+<h3 id="oracle-solaris-studio">Oracle Solaris Studio</h3>
+<p>The minimum accepted version of the Solaris Studio compilers is 5.13 (corresponding to Solaris Studio 12.4). Older versions will not be accepted by configure.</p>
+<p>The Solaris Studio installation should contain at least these packages:</p>
+<table>
+<thead>
+<tr class="header">
+<th style="text-align: left;">Package</th>
+<th style="text-align: left;">Version</th>
+</tr>
+</thead>
+<tbody>
+<tr class="odd">
+<td style="text-align: left;">developer/solarisstudio-124/backend</td>
+<td style="text-align: left;">12.4-1.0.6.0</td>
+</tr>
+<tr class="even">
+<td style="text-align: left;">developer/solarisstudio-124/c++</td>
+<td style="text-align: left;">12.4-1.0.10.0</td>
+</tr>
+<tr class="odd">
+<td style="text-align: left;">developer/solarisstudio-124/cc</td>
+<td style="text-align: left;">12.4-1.0.4.0</td>
+</tr>
+<tr class="even">
+<td style="text-align: left;">developer/solarisstudio-124/library/c++-libs</td>
+<td style="text-align: left;">12.4-1.0.10.0</td>
+</tr>
+<tr class="odd">
+<td style="text-align: left;">developer/solarisstudio-124/library/math-libs</td>
+<td style="text-align: left;">12.4-1.0.0.1</td>
+</tr>
+<tr class="even">
+<td style="text-align: left;">developer/solarisstudio-124/library/studio-gccrt</td>
+<td style="text-align: left;">12.4-1.0.0.1</td>
+</tr>
+<tr class="odd">
+<td style="text-align: left;">developer/solarisstudio-124/studio-common</td>
+<td style="text-align: left;">12.4-1.0.0.1</td>
+</tr>
+<tr class="even">
+<td style="text-align: left;">developer/solarisstudio-124/studio-ja</td>
+<td style="text-align: left;">12.4-1.0.0.1</td>
+</tr>
+<tr class="odd">
+<td style="text-align: left;">developer/solarisstudio-124/studio-legal</td>
+<td style="text-align: left;">12.4-1.0.0.1</td>
+</tr>
+<tr class="even">
+<td style="text-align: left;">developer/solarisstudio-124/studio-zhCN</td>
+<td style="text-align: left;">12.4-1.0.0.1</td>
+</tr>
+</tbody>
+</table>
+<p>Compiling with Solaris Studio can sometimes be finicky. This is the exact version used by Oracle, which worked correctly at the time of writing:</p>
+<pre><code>$ cc -V
+cc: Sun C 5.13 SunOS_i386 2014/10/20
+$ CC -V
+CC: Sun C++ 5.13 SunOS_i386 151846-10 2015/10/30</code></pre>
+<h3 id="microsoft-visual-studio">Microsoft Visual Studio</h3>
+<p>The minimum accepted version of Visual Studio is 2010. Older versions will not be accepted by <code>configure</code>. The maximum accepted version of Visual Studio is 2013.</p>
+<p>If you have multiple versions of Visual Studio installed, <code>configure</code> will by default pick the latest. You can request a specific version to be used by setting <code>--with-toolchain-version</code>, e.g. <code>--with-toolchain-version=2010</code>.</p>
+<p>If you get <code>LINK: fatal error LNK1123: failure during conversion to COFF: file invalid</code> when building using Visual Studio 2010, you have encountered <a href="http://support.microsoft.com/kb/2757355">KB2757355</a>, a bug triggered by a specific installation order. However, the solution suggested by the KB article does not always resolve the problem. See <a href="https://stackoverflow.com/questions/10888391">this stackoverflow discussion</a> for other suggestions.</p>
+<h3 id="ibm-xl-cc">IBM XL C/C++</h3>
+<p>The regular builds by SAP is using version 12.1, described as <code>IBM XL C/C++ for AIX, V12.1 (5765-J02, 5725-C72) Version: 12.01.0000.0017</code>.</p>
+<p>See the <a href="http://cr.openjdk.java.net/~simonis/ppc-aix-port">OpenJDK PowerPC Port Status Page</a> for details.</p>
+<h2 id="boot-jdk-requirements">Boot JDK Requirements</h2>
+<p>Paradoxically, building OpenJDK requires a pre-existing JDK. This is called the &quot;boot JDK&quot;. The boot JDK does not have to be OpenJDK, though. If you are porting OpenJDK to a new platform, chances are that there already exists another JDK for that platform that is usable as boot JDK.</p>
+<p>The rule of thumb is that the boot JDK for building JDK major version <em>N</em> should be an JDK of major version <em>N-1</em>, so for building JDK 9 a JDK 8 would be suitable as boot JDK. However, OpenJDK should be able to &quot;build itself&quot;, so an up-to-date build of the current OpenJDK source is an acceptable alternative. If you are following the <em>N-1</em> rule, make sure you got the latest update version, since JDK 8 GA might not be able to build JDK 9 on all platforms.</p>
+<p>If the Boot JDK is not automatically detected, or the wrong JDK is picked, use <code>--with-boot-jdk</code> to point to the JDK to use.</p>
+<h3 id="jdk-8-on-linux">JDK 8 on Linux</h3>
+<p>On apt-based distros (like Debian and Ubuntu), <code>sudo apt-get install openjdk-8-jdk</code> is typically enough to install OpenJDK 8. On rpm-based distros (like Fedora and Red Hat), try <code>sudo yum install java-1.8.0-openjdk-devel</code>.</p>
+<h3 id="jdk-8-on-windows">JDK 8 on Windows</h3>
+<p>No pre-compiled binaries of OpenJDK 8 are readily available for Windows at the time of writing. An alternative is to download the <a href="http://www.oracle.com/technetwork/java/javase/downloads">Oracle JDK</a>. Another is the <a href="https://adoptopenjdk.net/">Adopt OpenJDK Project</a>, which publishes experimental prebuilt binaries for Windows.</p>
+<h3 id="jdk-8-on-macos">JDK 8 on macOS</h3>
+<p>No pre-compiled binaries of OpenJDK 8 are readily available for macOS at the time of writing. An alternative is to download the <a href="http://www.oracle.com/technetwork/java/javase/downloads">Oracle JDK</a>, or to install it using <code>brew cask install java</code>. Another option is the <a href="https://adoptopenjdk.net/">Adopt OpenJDK Project</a>, which publishes experimental prebuilt binaries for macOS.</p>
+<h3 id="jdk-8-on-aix">JDK 8 on AIX</h3>
+<p>No pre-compiled binaries of OpenJDK 8 are readily available for AIX at the time of writing. A starting point for working with OpenJDK on AIX is the <a href="http://openjdk.java.net/projects/ppc-aix-port/">PowerPC/AIX Port Project</a>.</p>
+<h2 id="external-library-requirements">External Library Requirements</h2>
+<p>Different platforms require different external libraries. In general, libraries are not optional - that is, they are either required or not used.</p>
+<p>If a required library is not detected by <code>configure</code>, you need to provide the path to it. There are two forms of the <code>configure</code> arguments to point to an external library: <code>--with-&lt;LIB&gt;=&lt;path&gt;</code> or <code>--with-&lt;LIB&gt;-include=&lt;path to include&gt; --with-&lt;LIB&gt;-lib=&lt;path to lib&gt;</code>. The first variant is more concise, but require the include files an library files to reside in a default hierarchy under this directory. In most cases, it works fine.</p>
+<p>As a fallback, the second version allows you to point to the include directory and the lib directory separately.</p>
+<h3 id="freetype">FreeType</h3>
+<p>FreeType2 from <a href="http://www.freetype.org/">The FreeType Project</a> is required on all platforms. At least version 2.3 is required.</p>
+<ul>
+<li>To install on an apt-based Linux, try running <code>sudo apt-get install libcups2-dev</code>.</li>
+<li>To install on an rpm-based Linux, try running <code>sudo yum install cups-devel</code>.</li>
+<li>To install on Solaris, try running <code>pkg install system/library/freetype-2</code>.</li>
+<li>To install on macOS, try running <code>brew install freetype</code>.</li>
+<li>To install on Windows, see <a href="#building-freetype-on-windows">below</a>.</li>
+</ul>
+<p>Use <code>--with-freetype=&lt;path&gt;</code> if <code>configure</code> does not properly locate your FreeType files.</p>
+<h4 id="building-freetype-on-windows">Building FreeType on Windows</h4>
+<p>On Windows, there is no readily available compiled version of FreeType. OpenJDK can help you compile FreeType from source. Download the FreeType sources and unpack them into an arbitrary directory:</p>
+<pre><code>wget http://download.savannah.gnu.org/releases/freetype/freetype-2.5.3.tar.gz
+tar -xzf freetype-2.5.3.tar.gz</code></pre>
+<p>Then run <code>configure</code> with <code>--with-freetype-src=&lt;freetype_src&gt;</code>. This will automatically build the freetype library into <code>&lt;freetype_src&gt;/lib64</code> for 64-bit builds or into <code>&lt;freetype_src&gt;/lib32</code> for 32-bit builds. Afterwards you can always use <code>--with-freetype-include=&lt;freetype_src&gt;/include</code> and <code>--with-freetype-lib=&lt;freetype_src&gt;/lib[32|64]</code> for other builds.</p>
+<p>Alternatively you can unpack the sources like this to use the default directory:</p>
+<pre><code>tar --one-top-level=$HOME/freetype --strip-components=1 -xzf freetype-2.5.3.tar.gz</code></pre>
+<h3 id="cups">CUPS</h3>
+<p>CUPS, <a href="http://www.cups.org">Common UNIX Printing System</a> header files are required on all platforms, except Windows. Often these files are provided by your operating system.</p>
+<ul>
+<li>To install on an apt-based Linux, try running <code>sudo apt-get install libcups2-dev</code>.</li>
+<li>To install on an rpm-based Linux, try running <code>sudo yum install cups-devel</code>.</li>
+<li>To install on Solaris, try running <code>pkg install print/cups</code>.</li>
+</ul>
+<p>Use <code>--with-cups=&lt;path&gt;</code> if <code>configure</code> does not properly locate your CUPS files.</p>
+<h3 id="x11">X11</h3>
+<p>Certain <a href="http://www.x.org/">X11</a> libraries and include files are required on Linux and Solaris.</p>
+<ul>
+<li>To install on an apt-based Linux, try running <code>sudo apt-get install libx11-dev libxext-dev libxrender-dev libxtst-dev libxt-dev</code>.</li>
+<li>To install on an rpm-based Linux, try running <code>sudo yum install libXtst-devel libXt-devel libXrender-devel libXi-devel</code>.</li>
+<li>To install on Solaris, try running <code>pkg install x11/header/x11-protocols x11/library/libice x11/library/libpthread-stubs x11/library/libsm x11/library/libx11 x11/library/libxau x11/library/libxcb x11/library/libxdmcp x11/library/libxevie x11/library/libxext x11/library/libxrender x11/library/libxscrnsaver x11/library/libxtst x11/library/toolkit/libxt</code>.</li>
+</ul>
+<p>Use <code>--with-x=&lt;path&gt;</code> if <code>configure</code> does not properly locate your X11 files.</p>
+<h3 id="alsa">ALSA</h3>
+<p>ALSA, <a href="https://www.alsa-project.org/">Advanced Linux Sound Architecture</a> is required on Linux. At least version 0.9.1 of ALSA is required.</p>
+<ul>
+<li>To install on an apt-based Linux, try running <code>sudo apt-get install libasound2-dev</code>.</li>
+<li>To install on an rpm-based Linux, try running <code>sudo yum install alsa-lib-devel</code>.</li>
+</ul>
+<p>Use <code>--with-alsa=&lt;path&gt;</code> if <code>configure</code> does not properly locate your ALSA files.</p>
+<h3 id="libffi">libffi</h3>
+<p>libffi, the <a href="http://sourceware.org/libffi">Portable Foreign Function Interface Library</a> is required when building the Zero version of Hotspot.</p>
+<ul>
+<li>To install on an apt-based Linux, try running <code>sudo apt-get install libffi-dev</code>.</li>
+<li>To install on an rpm-based Linux, try running <code>sudo yum install libffi-devel</code>.</li>
+</ul>
+<p>Use <code>--with-libffi=&lt;path&gt;</code> if <code>configure</code> does not properly locate your libffi files.</p>
+<h3 id="libelf">libelf</h3>
+<p>libelf from the <a href="http://sourceware.org/elfutils">elfutils project</a> is required when building the AOT feature of Hotspot.</p>
+<ul>
+<li>To install on an apt-based Linux, try running <code>sudo apt-get install libelf-dev</code>.</li>
+<li>To install on an rpm-based Linux, try running <code>sudo yum install elfutils-libelf-devel</code>.</li>
+</ul>
+<p>Use <code>--with-libelf=&lt;path&gt;</code> if <code>configure</code> does not properly locate your libelf files.</p>
+<h2 id="other-tooling-requirements">Other Tooling Requirements</h2>
+<h3 id="gnu-make">GNU Make</h3>
+<p>OpenJDK requires <a href="http://www.gnu.org/software/make">GNU Make</a>. No other flavors of make are supported.</p>
+<p>At least version 3.81 of GNU Make must be used. For distributions supporting GNU Make 4.0 or above, we strongly recommend it. GNU Make 4.0 contains useful functionality to handle parallel building (supported by <code>--with-output-sync</code>) and speed and stability improvements.</p>
+<p>Note that <code>configure</code> locates and verifies a properly functioning version of <code>make</code> and stores the path to this <code>make</code> binary in the configuration. If you start a build using <code>make</code> on the command line, you will be using the version of make found first in your <code>PATH</code>, and not necessarily the one stored in the configuration. This initial make will be used as &quot;bootstrap make&quot;, and in a second stage, the make located by <code>configure</code> will be called. Normally, this will present no issues, but if you have a very old <code>make</code>, or a non-GNU Make <code>make</code> in your path, this might cause issues.</p>
+<p>If you want to override the default make found by <code>configure</code>, use the <code>MAKE</code> configure variable, e.g. <code>configure MAKE=/opt/gnu/make</code>.</p>
+<p>On Solaris, it is common to call the GNU version of make by using <code>gmake</code>.</p>
+<h3 id="gnu-bash">GNU Bash</h3>
+<p>OpenJDK requires <a href="http://www.gnu.org/software/bash">GNU Bash</a>. No other shells are supported.</p>
+<p>At least version 3.2 of GNU Bash must be used.</p>
+<h3 id="autoconf">Autoconf</h3>
+<p>If you want to modify the build system itself, you need to install <a href="http://www.gnu.org/software/autoconf">Autoconf</a>.</p>
+<p>However, if you only need to build OpenJDK or if you only edit the actual OpenJDK source files, there is no dependency on autoconf, since the source distribution includes a pre-generated <code>configure</code> shell script.</p>
+<p>See the section on <a href="#autoconf-details">Autoconf Details</a> for details on how OpenJDK uses autoconf. This is especially important if you plan to contribute changes to OpenJDK that modifies the build system.</p>
+<h2 id="running-configure">Running Configure</h2>
+<p>To build OpenJDK, you need a &quot;configuration&quot;, which consists of a directory where to store the build output, coupled with information about the platform, the specific build machine, and choices that affect how OpenJDK is built.</p>
+<p>The configuration is created by the <code>configure</code> script. The basic invocation of the <code>configure</code> script looks like this:</p>
+<pre><code>bash configure [options]</code></pre>
+<p>This will create an output directory containing the configuration and setup an area for the build result. This directory typically looks like <code>build/linux-x64-normal-server-release</code>, but the actual name depends on your specific configuration. (It can also be set directly, see <a href="#using-multiple-configurations">Using Multiple Configurations</a>). This directory is referred to as <code>$BUILD</code> in this documentation.</p>
+<p><code>configure</code> will try to figure out what system you are running on and where all necessary build components are. If you have all prerequisites for building installed, it should find everything. If it fails to detect any component automatically, it will exit and inform you about the problem.</p>
+<p>Some command line examples:</p>
+<ul>
+<li><p>Create a 32-bit build for Windows with FreeType2 in <code>C:\freetype-i586</code>:</p>
+<pre><code>bash configure --with-freetype=/cygdrive/c/freetype-i586 --with-target-bits=32</code></pre></li>
+<li><p>Create a debug build with the <code>server</code> JVM and DTrace enabled:</p>
+<pre><code>bash configure --enable-debug --with-jvm-variants=server --enable-dtrace</code></pre></li>
+</ul>
+<h3 id="common-configure-arguments">Common Configure Arguments</h3>
+<p>Here follows some of the most common and important <code>configure</code> argument.</p>
+<p>To get up-to-date information on <em>all</em> available <code>configure</code> argument, please run:</p>
+<pre><code>bash configure --help</code></pre>
+<p>(Note that this help text also include general autoconf options, like <code>--dvidir</code>, that is not relevant to OpenJDK. To list only OpenJDK specific features, use <code>bash configure --help=short</code> instead.)</p>
+<h4 id="configure-arguments-for-tailoring-the-build">Configure Arguments for Tailoring the Build</h4>
+<ul>
+<li><code>--enable-debug</code> - Set the debug level to <code>fastdebug</code> (this is a shorthand for <code>--with-debug-level=fastdebug</code>)</li>
+<li><code>--with-debug-level=&lt;level&gt;</code> - Set the debug level, which can be <code>release</code>, <code>fastdebug</code>, <code>slowdebug</code> or <code>optimized</code>. Default is <code>release</code>. <code>optimized</code> is variant of <code>release</code> with additional Hotspot debug code.</li>
+<li><code>--with-native-debug-symbols=&lt;method&gt;</code> - Specify if and how native debug symbols should be built. Available methods are <code>none</code>, <code>internal</code>, <code>external</code>, <code>zipped</code>. Default behavior depends on platform. See <a href="#native-debug-symbols">Native Debug Symbols</a> for more details.</li>
+<li><code>--with-version-string=&lt;string&gt;</code> - Specify the version string this build will be identified with.</li>
+<li><code>--with-version-&lt;part&gt;=&lt;value&gt;</code> - A group of options, where <code>&lt;part&gt;</code> can be any of <code>pre</code>, <code>opt</code>, <code>build</code>, <code>major</code>, <code>minor</code>, <code>security</code> or <code>patch</code>. Use these options to modify just the corresponding part of the version string from the default, or the value provided by <code>--with-version-string</code>.</li>
+<li><code>--with-jvm-variants=&lt;variant&gt;[,&lt;variant&gt;...]</code> - Build the specified variant (or variants) of Hotspot. Valid variants are: <code>server</code>, <code>client</code>, <code>minimal</code>, <code>core</code>, <code>zero</code>, <code>zeroshark</code>, <code>custom</code>. Note that not all variants are possible to combine in a single build.</li>
+<li><code>--with-jvm-features=&lt;feature&gt;[,&lt;feature&gt;...]</code> - Use the specified JVM features when building Hotspot. The list of features will be enabled on top of the default list. For the <code>custom</code> JVM variant, this default list is empty. A complete list of available JVM features can be found using <code>bash configure --help</code>.</li>
+<li><code>--with-target-bits=&lt;bits&gt;</code> - Create a target binary suitable for running on a <code>&lt;bits&gt;</code> platform. Use this to create 32-bit output on a 64-bit build platform, instead of doing a full cross-compile. (This is known as a <em>reduced</em> build.)</li>
+</ul>
+<h4 id="configure-arguments-for-native-compilation">Configure Arguments for Native Compilation</h4>
+<ul>
+<li><code>--with-devkit=&lt;path&gt;</code> - Use this devkit for compilers, tools and resources</li>
+<li><code>--with-sysroot=&lt;path&gt;</code> - Use this directory as sysroot</li>
+<li><code>--with-extra-path=&lt;path&gt;[;&lt;path&gt;]</code> - Prepend these directories to the default path when searching for all kinds of binaries</li>
+<li><code>--with-toolchain-path=&lt;path&gt;[;&lt;path&gt;]</code> - Prepend these directories when searching for toolchain binaries (compilers etc)</li>
+<li><code>--with-extra-cflags=&lt;flags&gt;</code> - Append these flags when compiling JDK C files</li>
+<li><code>--with-extra-cxxflags=&lt;flags&gt;</code> - Append these flags when compiling JDK C++ files</li>
+<li><code>--with-extra-ldflags=&lt;flags&gt;</code> - Append these flags when linking JDK libraries</li>
+</ul>
+<h4 id="configure-arguments-for-external-dependencies">Configure Arguments for External Dependencies</h4>
+<ul>
+<li><code>--with-boot-jdk=&lt;path&gt;</code> - Set the path to the <a href="#boot-jdk-requirements">Boot JDK</a></li>
+<li><code>--with-freetype=&lt;path&gt;</code> - Set the path to <a href="#freetype">FreeType</a></li>
+<li><code>--with-cups=&lt;path&gt;</code> - Set the path to <a href="#cups">CUPS</a></li>
+<li><code>--with-x=&lt;path&gt;</code> - Set the path to <a href="#x11">X11</a></li>
+<li><code>--with-alsa=&lt;path&gt;</code> - Set the path to <a href="#alsa">ALSA</a></li>
+<li><code>--with-libffi=&lt;path&gt;</code> - Set the path to <a href="#libffi">libffi</a></li>
+<li><code>--with-libelf=&lt;path&gt;</code> - Set the path to <a href="#libelf">libelf</a></li>
+<li><code>--with-jtreg=&lt;path&gt;</code> - Set the path to JTReg. See <a href="#running-tests">Running Tests</a></li>
+</ul>
+<p>Certain third-party libraries used by OpenJDK (libjpeg, giflib, libpng, lcms and zlib) are included in the OpenJDK repository. The default behavior of the OpenJDK build is to use this version of these libraries, but they might be replaced by an external version. To do so, specify <code>system</code> as the <code>&lt;source&gt;</code> option in these arguments. (The default is <code>bundled</code>).</p>
+<ul>
+<li><code>--with-libjpeg=&lt;source&gt;</code> - Use the specified source for libjpeg</li>
+<li><code>--with-giflib=&lt;source&gt;</code> - Use the specified source for giflib</li>
+<li><code>--with-libpng=&lt;source&gt;</code> - Use the specified source for libpng</li>
+<li><code>--with-lcms=&lt;source&gt;</code> - Use the specified source for lcms</li>
+<li><code>--with-zlib=&lt;source&gt;</code> - Use the specified source for zlib</li>
+</ul>
+<p>On Linux, it is possible to select either static or dynamic linking of the C++ runtime. The default is static linking, with dynamic linking as fallback if the static library is not found.</p>
+<ul>
+<li><code>--with-stdc++lib=&lt;method&gt;</code> - Use the specified method (<code>static</code>, <code>dynamic</code> or <code>default</code>) for linking the C++ runtime.</li>
+</ul>
+<h3 id="configure-control-variables">Configure Control Variables</h3>
+<p>It is possible to control certain aspects of <code>configure</code> by overriding the value of <code>configure</code> variables, either on the command line or in the environment.</p>
+<p>Normally, this is <strong>not recommended</strong>. If used improperly, it can lead to a broken configuration. Unless you're well versed in the build system, this is hard to use properly. Therefore, <code>configure</code> will print a warning if this is detected.</p>
+<p>However, there are a few <code>configure</code> variables, known as <em>control variables</em> that are supposed to be overriden on the command line. These are variables that describe the location of tools needed by the build, like <code>MAKE</code> or <code>GREP</code>. If any such variable is specified, <code>configure</code> will use that value instead of trying to autodetect the tool. For instance, <code>bash configure MAKE=/opt/gnumake4.0/bin/make</code>.</p>
+<p>If a configure argument exists, use that instead, e.g. use <code>--with-jtreg</code> instead of setting <code>JTREGEXE</code>.</p>
+<p>Also note that, despite what autoconf claims, setting <code>CFLAGS</code> will not accomplish anything. Instead use <code>--with-extra-cflags</code> (and similar for <code>cxxflags</code> and <code>ldflags</code>).</p>
+<h2 id="running-make">Running Make</h2>
+<p>When you have a proper configuration, all you need to do to build OpenJDK is to run <code>make</code>. (But see the warning at <a href="#gnu-make">GNU Make</a> about running the correct version of make.)</p>
+<p>When running <code>make</code> without any arguments, the default target is used, which is the same as running <code>make default</code> or <code>make jdk</code>. This will build a minimal (or roughly minimal) set of compiled output (known as an &quot;exploded image&quot;) needed for a developer to actually execute the newly built JDK. The idea is that in an incremental development fashion, when doing a normal make, you should only spend time recompiling what's changed (making it purely incremental) and only do the work that's needed to actually run and test your code.</p>
+<p>The output of the exploded image resides in <code>$BUILD/jdk</code>. You can test the newly built JDK like this: <code>$BUILD/jdk/bin/java -version</code>.</p>
+<h3 id="common-make-targets">Common Make Targets</h3>
+<p>Apart from the default target, here are some common make targets:</p>
+<ul>
+<li><code>hotspot</code> - Build all of hotspot (but only hotspot)</li>
+<li><code>hotspot-&lt;variant&gt;</code> - Build just the specified jvm variant</li>
+<li><code>images</code> or <code>product-images</code> - Build the JRE and JDK images</li>
+<li><code>docs</code> or <code>docs-image</code> - Build the documentation image</li>
+<li><code>test-image</code> - Build the test image</li>
+<li><code>all</code> or <code>all-images</code> - Build all images (product, docs and test)</li>
+<li><code>bootcycle-images</code> - Build images twice, second time with newly built JDK (good for testing)</li>
+<li><code>clean</code> - Remove all files generated by make, but not those generated by configure</li>
+<li><code>dist-clean</code> - Remove all files, including configuration</li>
+</ul>
+<p>Run <code>make help</code> to get an up-to-date list of important make targets and make control variables.</p>
+<p>It is possible to build just a single module, a single phase, or a single phase of a single module, by creating make targets according to these followin patterns. A phase can be either of <code>gensrc</code>, <code>gendata</code>, <code>copy</code>, <code>java</code>, <code>launchers</code>, <code>libs</code> or <code>rmic</code>. See <a href="#using-fine-grained-make-targets">Using Fine-Grained Make Targets</a> for more details about this functionality.</p>
+<ul>
+<li><code>&lt;phase&gt;</code> - Build the specified phase and everything it depends on</li>
+<li><code>&lt;module&gt;</code> - Build the specified module and everything it depends on</li>
+<li><code>&lt;module&gt;-&lt;phase&gt;</code> - Compile the specified phase for the specified module and everything it depends on</li>
+</ul>
+<p>Similarly, it is possible to clean just a part of the build by creating make targets according to these patterns:</p>
+<ul>
+<li><code>clean-&lt;outputdir&gt;</code> - Remove the subdir in the output dir with the name</li>
+<li><code>clean-&lt;phase&gt;</code> - Remove all build results related to a certain build phase</li>
+<li><code>clean-&lt;module&gt;</code> - Remove all build results related to a certain module</li>
+<li><code>clean-&lt;module&gt;-&lt;phase&gt;</code> - Remove all build results related to a certain module and phase</li>
+</ul>
+<h3 id="make-control-variables">Make Control Variables</h3>
+<p>It is possible to control <code>make</code> behavior by overriding the value of <code>make</code> variables, either on the command line or in the environment.</p>
+<p>Normally, this is <strong>not recommended</strong>. If used improperly, it can lead to a broken build. Unless you're well versed in the build system, this is hard to use properly. Therefore, <code>make</code> will print a warning if this is detected.</p>
+<p>However, there are a few <code>make</code> variables, known as <em>control variables</em> that are supposed to be overriden on the command line. These make up the &quot;make time&quot; configuration, as opposed to the &quot;configure time&quot; configuration.</p>
+<h4 id="general-make-control-variables">General Make Control Variables</h4>
+<ul>
+<li><code>JOBS</code> - Specify the number of jobs to build with. See <a href="#build-performance">Build Performance</a>.</li>
+<li><code>LOG</code> - Specify the logging level and functionality. See <a href="#checking-the-build-log-file">Checking the Build Log File</a></li>
+<li><code>CONF</code> and <code>CONF_NAME</code> - Selecting the configuration(s) to use. See <a href="#using-multiple-configurations">Using Multiple Configurations</a></li>
+</ul>
+<h4 id="test-make-control-variables">Test Make Control Variables</h4>
+<p>These make control variables only make sense when running tests. Please see <a href="testing.html">Testing OpenJDK</a> for details.</p>
+<ul>
+<li><code>TEST</code></li>
+<li><code>TEST_JOBS</code></li>
+<li><code>JTREG</code></li>
+<li><code>GTEST</code></li>
+</ul>
+<h4 id="advanced-make-control-variables">Advanced Make Control Variables</h4>
+<p>These advanced make control variables can be potentially unsafe. See <a href="#hints-and-suggestions-for-advanced-users">Hints and Suggestions for Advanced Users</a> and <a href="#understanding-the-build-system">Understanding the Build System</a> for details.</p>
+<ul>
+<li><code>SPEC</code></li>
+<li><code>CONF_CHECK</code></li>
+<li><code>COMPARE_BUILD</code></li>
+<li><code>JDK_FILTER</code></li>
+</ul>
+<h2 id="running-tests">Running Tests</h2>
+<p>Most of the OpenJDK tests are using the <a href="http://openjdk.java.net/jtreg">JTReg</a> test framework. Make sure that your configuration knows where to find your installation of JTReg. If this is not picked up automatically, use the <code>--with-jtreg=&lt;path to jtreg home&gt;</code> option to point to the JTReg framework. Note that this option should point to the JTReg home, i.e. the top directory, containing <code>lib/jtreg.jar</code> etc.</p>
+<p>To execute the most basic tests (tier 1), use:</p>
+<pre><code>make run-test-tier1</code></pre>
+<p>For more details on how to run tests, please see the <a href="testing.html">Testing OpenJDK</a> document.</p>
+<h2 id="cross-compiling">Cross-compiling</h2>
+<p>Cross-compiling means using one platform (the <em>build</em> platform) to generate output that can ran on another platform (the <em>target</em> platform).</p>
+<p>The typical reason for cross-compiling is that the build is performed on a more powerful desktop computer, but the resulting binaries will be able to run on a different, typically low-performing system. Most of the complications that arise when building for embedded is due to this separation of <em>build</em> and <em>target</em> systems.</p>
+<p>This requires a more complex setup and build procedure. This section assumes you are familiar with cross-compiling in general, and will only deal with the particularities of cross-compiling OpenJDK. If you are new to cross-compiling, please see the <a href="https://en.wikipedia.org/wiki/Cross_compiler#External_links">external links at Wikipedia</a> for a good start on reading materials.</p>
+<p>Cross-compiling OpenJDK requires you to be able to build both for the build platform and for the target platform. The reason for the former is that we need to build and execute tools during the build process, both native tools and Java tools.</p>
+<p>If all you want to do is to compile a 32-bit version, for the same OS, on a 64-bit machine, consider using <code>--with-target-bits=32</code> instead of doing a full-blown cross-compilation. (While this surely is possible, it's a lot more work and will take much longer to build.)</p>
+<h3 id="boot-jdk-and-build-jdk">Boot JDK and Build JDK</h3>
+<p>When cross-compiling, make sure you use a boot JDK that runs on the <em>build</em> system, and not on the <em>target</em> system.</p>
+<p>To be able to build, we need a &quot;Build JDK&quot;, which is a JDK built from the current sources (that is, the same as the end result of the entire build process), but able to run on the <em>build</em> system, and not the <em>target</em> system. (In contrast, the Boot JDK should be from an older release, e.g. JDK 8 when building JDK 9.)</p>
+<p>The build process will create a minimal Build JDK for you, as part of building. To speed up the build, you can use <code>--with-build-jdk</code> to <code>configure</code> to point to a pre-built Build JDK. Please note that the build result is unpredictable, and can possibly break in subtle ways, if the Build JDK does not <strong>exactly</strong> match the current sources.</p>
+<h3 id="specifying-the-target-platform">Specifying the Target Platform</h3>
+<p>You <em>must</em> specify the target platform when cross-compiling. Doing so will also automatically turn the build into a cross-compiling mode. The simplest way to do this is to use the <code>--openjdk-target</code> argument, e.g. <code>--openjdk-target=arm-linux-gnueabihf</code>. or <code>--openjdk-target=aarch64-oe-linux</code>. This will automatically set the <code>--build</code>, <code>--host</code> and <code>--target</code> options for autoconf, which can otherwise be confusing. (In autoconf terminology, the &quot;target&quot; is known as &quot;host&quot;, and &quot;target&quot; is used for building a Canadian cross-compiler.)</p>
+<h3 id="toolchain-considerations">Toolchain Considerations</h3>
+<p>You will need two copies of your toolchain, one which generates output that can run on the target system (the normal, or <em>target</em>, toolchain), and one that generates output that can run on the build system (the <em>build</em> toolchain). Note that cross-compiling is only supported for gcc at the time being. The gcc standard is to prefix cross-compiling toolchains with the target denominator. If you follow this standard, <code>configure</code> is likely to pick up the toolchain correctly.</p>
+<p>The <em>build</em> toolchain will be autodetected just the same way the normal <em>build</em>/<em>target</em> toolchain will be autodetected when not cross-compiling. If this is not what you want, or if the autodetection fails, you can specify a devkit containing the <em>build</em> toolchain using <code>--with-build-devkit</code> to <code>configure</code>, or by giving <code>BUILD_CC</code> and <code>BUILD_CXX</code> arguments.</p>
+<p>It is often helpful to locate the cross-compilation tools, headers and libraries in a separate directory, outside the normal path, and point out that directory to <code>configure</code>. Do this by setting the sysroot (<code>--with-sysroot</code>) and appending the directory when searching for cross-compilations tools (<code>--with-toolchain-path</code>). As a compact form, you can also use <code>--with-devkit</code> to point to a single directory, if it is correctly setup. (See <code>basics.m4</code> for details.)</p>
+<p>If you are unsure what toolchain and versions to use, these have been proved working at the time of writing:</p>
+<ul>
+<li><a href="https://releases.linaro.org/archive/13.11/components/toolchain/binaries/gcc-linaro-aarch64-linux-gnu-4.8-2013.11_linux.tar.xz">aarch64</a></li>
+<li><a href="https://launchpad.net/linaro-toolchain-unsupported/trunk/2012.09/+download/gcc-linaro-arm-linux-gnueabihf-raspbian-2012.09-20120921_linux.tar.bz2">arm 32-bit hardware floating point</a></li>
+</ul>
+<h3 id="native-libraries">Native Libraries</h3>
+<p>You will need copies of external native libraries for the <em>target</em> system, present on the <em>build</em> machine while building.</p>
+<p>Take care not to replace the <em>build</em> system's version of these libraries by mistake, since that can render the <em>build</em> machine unusable.</p>
+<p>Make sure that the libraries you point to (ALSA, X11, etc) are for the <em>target</em>, not the <em>build</em>, platform.</p>
+<h4 id="alsa-1">ALSA</h4>
+<p>You will need alsa libraries suitable for your <em>target</em> system. For most cases, using Debian's pre-built libraries work fine.</p>
+<p>Note that alsa is needed even if you only want to build a headless JDK.</p>
+<ul>
+<li><p>Go to <a href="https://www.debian.org/distrib/packages">Debian Package Search</a> and search for the <code>libasound2</code> and <code>libasound2-dev</code> packages for your <em>target</em> system. Download them to /tmp.</p></li>
+<li><p>Install the libraries into the cross-compilation toolchain. For instance:</p>
+<pre><code>cd /tools/gcc-linaro-arm-linux-gnueabihf-raspbian-2012.09-20120921_linux/arm-linux-gnueabihf/libc
+dpkg-deb -x /tmp/libasound2_1.0.25-4_armhf.deb .
+dpkg-deb -x /tmp/libasound2-dev_1.0.25-4_armhf.deb .</code></pre></li>
+<li><p>If alsa is not properly detected by <code>configure</code>, you can point it out by <code>--with-alsa</code>.</p></li>
+</ul>
+<h4 id="x11-1">X11</h4>
+<p>You will need X11 libraries suitable for your <em>target</em> system. For most cases, using Debian's pre-built libraries work fine.</p>
+<p>Note that X11 is needed even if you only want to build a headless JDK.</p>
+<ul>
+<li>Go to <a href="https://www.debian.org/distrib/packages">Debian Package Search</a>, search for the following packages for your <em>target</em> system, and download them to /tmp/target-x11:
+<ul>
+<li>libxi</li>
+<li>libxi-dev</li>
+<li>x11proto-core-dev</li>
+<li>x11proto-input-dev</li>
+<li>x11proto-kb-dev</li>
+<li>x11proto-render-dev</li>
+<li>x11proto-xext-dev</li>
+<li>libice-dev</li>
+<li>libxrender</li>
+<li>libxrender-dev</li>
+<li>libsm-dev</li>
+<li>libxt-dev</li>
+<li>libx11</li>
+<li>libx11-dev</li>
+<li>libxtst</li>
+<li>libxtst-dev</li>
+<li>libxext</li>
+<li>libxext-dev</li>
+</ul></li>
+<li><p>Install the libraries into the cross-compilation toolchain. For instance:</p>
+<pre><code>cd /tools/gcc-linaro-arm-linux-gnueabihf-raspbian-2012.09-20120921_linux/arm-linux-gnueabihf/libc/usr
+mkdir X11R6
+cd X11R6
+for deb in /tmp/target-x11/*.deb ; do dpkg-deb -x $deb . ; done
+mv usr/* .
+cd lib
+cp arm-linux-gnueabihf/* .</code></pre>
+<p>You can ignore the following messages. These libraries are not needed to successfully complete a full JDK build.</p>
+<pre><code>cp: cannot stat `arm-linux-gnueabihf/libICE.so&#39;: No such file or directory
+cp: cannot stat `arm-linux-gnueabihf/libSM.so&#39;: No such file or directory
+cp: cannot stat `arm-linux-gnueabihf/libXt.so&#39;: No such file or directory</code></pre></li>
+<li><p>If the X11 libraries are not properly detected by <code>configure</code>, you can point them out by <code>--with-x</code>.</p></li>
+</ul>
+<h3 id="building-for-armaarch64">Building for ARM/aarch64</h3>
+<p>A common cross-compilation target is the ARM CPU. When building for ARM, it is useful to set the ABI profile. A number of pre-defined ABI profiles are available using <code>--with-abi-profile</code>: arm-vfp-sflt, arm-vfp-hflt, arm-sflt, armv5-vfp-sflt, armv6-vfp-hflt. Note that soft-float ABIs are no longer properly supported on OpenJDK.</p>
+<p>OpenJDK contains two different ports for the aarch64 platform, one is the original aarch64 port from the <a href="http://openjdk.java.net/projects/aarch64-port">AArch64 Port Project</a> and one is a 64-bit version of the Oracle contributed ARM port. When targeting aarch64, by the default the original aarch64 port is used. To select the Oracle ARM 64 port, use <code>--with-cpu-port=arm64</code>. Also set the corresponding value (<code>aarch64</code> or <code>arm64</code>) to --with-abi-profile, to ensure a consistent build.</p>
+<h3 id="verifying-the-build">Verifying the Build</h3>
+<p>The build will end up in a directory named like <code>build/linux-arm-normal-server-release</code>.</p>
+<p>Inside this build output directory, the <code>images/jdk</code> and <code>images/jre</code> will contain the newly built JDK and JRE, respectively, for your <em>target</em> system.</p>
+<p>Copy these folders to your <em>target</em> system. Then you can run e.g. <code>images/jdk/bin/java -version</code>.</p>
+<h2 id="build-performance">Build Performance</h2>
+<p>Building OpenJDK requires a lot of horsepower. Some of the build tools can be adjusted to utilize more or less of resources such as parallel threads and memory. The <code>configure</code> script analyzes your system and selects reasonable values for such options based on your hardware. If you encounter resource problems, such as out of memory conditions, you can modify the detected values with:</p>
+<ul>
+<li><p><code>--with-num-cores</code> -- number of cores in the build system, e.g. <code>--with-num-cores=8</code>.</p></li>
+<li><p><code>--with-memory-size</code> -- memory (in MB) available in the build system, e.g. <code>--with-memory-size=1024</code></p></li>
+</ul>
+<p>You can also specify directly the number of build jobs to use with <code>--with-jobs=N</code> to <code>configure</code>, or <code>JOBS=N</code> to <code>make</code>. Do not use the <code>-j</code> flag to <code>make</code>. In most cases it will be ignored by the makefiles, but it can cause problems for some make targets.</p>
+<p>It might also be necessary to specify the JVM arguments passed to the Boot JDK, using e.g. <code>--with-boot-jdk-jvmargs=&quot;-Xmx8G&quot;</code>. Doing so will override the default JVM arguments passed to the Boot JDK.</p>
+<p>At the end of a successful execution of <code>configure</code>, you will get a performance summary, indicating how well the build will perform. Here you will also get performance hints. If you want to build fast, pay attention to those!</p>
+<p>If you want to tweak build performance, run with <code>make LOG=info</code> to get a build time summary at the end of the build process.</p>
+<h3 id="disk-speed">Disk Speed</h3>
+<p>If you are using network shares, e.g. via NFS, for your source code, make sure the build directory is situated on local disk (e.g. by <code>ln -s /localdisk/jdk-build $JDK-SHARE/build</code>). The performance penalty is extremely high for building on a network share; close to unusable.</p>
+<p>Also, make sure that your build tools (including Boot JDK and toolchain) is located on a local disk and not a network share.</p>
+<p>As has been stressed elsewhere, do use SSD for source code and build directory, as well as (if possible) the build tools.</p>
+<h3 id="virus-checking">Virus Checking</h3>
+<p>The use of virus checking software, especially on Windows, can <em>significantly</em> slow down building of OpenJDK. If possible, turn off such software, or exclude the directory containing the OpenJDK source code from on-the-fly checking.</p>
+<h3 id="ccache">Ccache</h3>
+<p>The OpenJDK build supports building with ccache when using gcc or clang. Using ccache can radically speed up compilation of native code if you often rebuild the same sources. Your milage may vary however, so we recommend evaluating it for yourself. To enable it, make sure it's on the path and configure with <code>--enable-ccache</code>.</p>
+<h3 id="precompiled-headers">Precompiled Headers</h3>
+<p>By default, the Hotspot build uses preccompiled headers (PCH) on the toolchains were it is properly supported (clang, gcc, and Visual Studio). Normally, this speeds up the build process, but in some circumstances, it can actually slow things down.</p>
+<p>You can experiment by disabling precompiled headers using <code>--disable-precompiled-headers</code>.</p>
+<h3 id="icecc-icecream">Icecc / icecream</h3>
+<p><a href="http://github.com/icecc/icecream">icecc/icecream</a> is a simple way to setup a distributed compiler network. If you have multiple machines available for building OpenJDK, you can drastically cut individual build times by utilizing it.</p>
+<p>To use, setup an icecc network, and install icecc on the build machine. Then run <code>configure</code> using <code>--enable-icecc</code>.</p>
+<h3 id="using-sjavac">Using sjavac</h3>
+<p>To speed up Java compilation, especially incremental compilations, you can try the experimental sjavac compiler by using <code>--enable-sjavac</code>.</p>
+<h3 id="building-the-right-target">Building the Right Target</h3>
+<p>Selecting the proper target to build can have dramatic impact on build time. For normal usage, <code>jdk</code> or the default target is just fine. You only need to build <code>images</code> for shipping, or if your tests require it.</p>
+<p>See also <a href="#using-fine-grained-make-targets">Using Fine-Grained Make Targets</a> on how to build an even smaller subset of the product.</p>
+<h2 id="troubleshooting">Troubleshooting</h2>
+<p>If your build fails, it can sometimes be difficult to pinpoint the problem or find a proper solution.</p>
+<h3 id="locating-the-source-of-the-error">Locating the Source of the Error</h3>
+<p>When a build fails, it can be hard to pinpoint the actual cause of the error. In a typical build process, different parts of the product build in parallel, with the output interlaced.</p>
+<h4 id="build-failure-summary">Build Failure Summary</h4>
+<p>To help you, the build system will print a failure summary at the end. It looks like this:</p>
+<pre><code>ERROR: Build failed for target &#39;hotspot&#39; in configuration &#39;linux-x64&#39; (exit code 2)
+
+=== Output from failing command(s) repeated here ===
+* For target hotspot_variant-server_libjvm_objs_psMemoryPool.o:
+/localhome/hg/jdk9-sandbox/hotspot/src/share/vm/services/psMemoryPool.cpp:1:1: error: &#39;failhere&#39; does not name a type
+   ... (rest of output omitted)
+
+* All command lines available in /localhome/hg/jdk9-sandbox/build/linux-x64/make-support/failure-logs.
+=== End of repeated output ===
+
+=== Make failed targets repeated here ===
+lib/CompileJvm.gmk:207: recipe for target &#39;/localhome/hg/jdk9-sandbox/build/linux-x64/hotspot/variant-server/libjvm/objs/psMemoryPool.o&#39; failed
+make/Main.gmk:263: recipe for target &#39;hotspot-server-libs&#39; failed
+=== End of repeated output ===
+
+Hint: Try searching the build log for the name of the first failed target.
+Hint: If caused by a warning, try configure --disable-warnings-as-errors.</code></pre>
+<p>Let's break it down! First, the selected configuration, and the top-level target you entered on the command line that caused the failure is printed.</p>
+<p>Then, between the <code>Output from failing command(s) repeated here</code> and <code>End of repeated output</code> the first lines of output (stdout and stderr) from the actual failing command is repeated. In most cases, this is the error message that caused the build to fail. If multiple commands were failing (this can happen in a parallel build), output from all failed commands will be printed here.</p>
+<p>The path to the <code>failure-logs</code> directory is printed. In this file you will find a <code>&lt;target&gt;.log</code> file that contains the output from this command in its entirety, and also a <code>&lt;target&gt;.cmd</code>, which contain the complete command line used for running this command. You can re-run the failing command by executing <code>. &lt;path to failure-logs&gt;/&lt;target&gt;.cmd</code> in your shell.</p>
+<p>Another way to trace the failure is to follow the chain of make targets, from top-level targets to individual file targets. Between <code>Make failed targets repeated here</code> and <code>End of repeated output</code> the output from make showing this chain is repeated. The first failed recipe will typically contain the full path to the file in question that failed to compile. Following lines will show a trace of make targets why we ended up trying to compile that file.</p>
+<p>Finally, some hints are given on how to locate the error in the complete log. In this example, we would try searching the log file for &quot;<code>psMemoryPool.o</code>&quot;. Another way to quickly locate make errors in the log is to search for &quot;<code>] Error</code>&quot; or &quot;<code>***</code>&quot;.</p>
+<p>Note that the build failure summary will only help you if the issue was a compilation failure or similar. If the problem is more esoteric, or is due to errors in the build machinery, you will likely get empty output logs, and <code>No indication of failed target found</code> instead of the make target chain.</p>
+<h4 id="checking-the-build-log-file">Checking the Build Log File</h4>
+<p>The output (stdout and stderr) from the latest build is always stored in <code>$BUILD/build.log</code>. The previous build log is stored as <code>build.log.old</code>. This means that it is not necessary to redirect the build output yourself if you want to process it.</p>
+<p>You can increase the verbosity of the log file, by the <code>LOG</code> control variable to <code>make</code>. If you want to see the command lines used in compilations, use <code>LOG=cmdlines</code>. To increase the general verbosity, use <code>LOG=info</code>, <code>LOG=debug</code> or <code>LOG=trace</code>. Both of these can be combined with <code>cmdlines</code>, e.g. <code>LOG=info,cmdlines</code>. The <code>debug</code> log level will show most shell commands executed by make, and <code>trace</code> will show all. Beware that both these log levels will produce a massive build log!</p>
+<h3 id="fixing-unexpected-build-failures">Fixing Unexpected Build Failures</h3>
+<p>Most of the time, the build will fail due to incorrect changes in the source code.</p>
+<p>Sometimes the build can fail with no apparent changes that have caused the failure. If this is the first time you are building OpenJDK on this particular computer, and the build fails, the problem is likely with your build environment. But even if you have previously built OpenJDK with success, and it now fails, your build environment might have changed (perhaps due to OS upgrades or similar). But most likely, such failures are due to problems with the incremental rebuild.</p>
+<h4 id="problems-with-the-build-environment">Problems with the Build Environment</h4>
+<p>Make sure your configuration is correct. Re-run <code>configure</code>, and look for any warnings. Warnings that appear in the middle of the <code>configure</code> output is also repeated at the end, after the summary. The entire log is stored in <code>$BUILD/configure.log</code>.</p>
+<p>Verify that the summary at the end looks correct. Are you indeed using the Boot JDK and native toolchain that you expect?</p>
+<p>By default, OpenJDK has a strict approach where warnings from the compiler is considered errors which fail the build. For very new or very old compiler versions, this can trigger new classes of warnings, which thus fails the build. Run <code>configure</code> with <code>--disable-warnings-as-errors</code> to turn of this behavior. (The warnings will still show, but not make the build fail.)</p>
+<h4 id="problems-with-incremental-rebuilds">Problems with Incremental Rebuilds</h4>
+<p>Incremental rebuilds mean that when you modify part of the product, only the affected parts get rebuilt. While this works great in most cases, and significantly speed up the development process, from time to time complex interdependencies will result in an incorrect build result. This is the most common cause for unexpected build problems, together with inconsistencies between the different Mercurial repositories in the forest.</p>
+<p>Here are a suggested list of things to try if you are having unexpected build problems. Each step requires more time than the one before, so try them in order. Most issues will be solved at step 1 or 2.</p>
+<ol type="1">
+<li><p>Make sure your forest is up-to-date</p>
+<p>Run <code>bash get_source.sh</code> to make sure you have the latest version of all repositories.</p></li>
+<li><p>Clean build results</p>
+<p>The simplest way to fix incremental rebuild issues is to run <code>make clean</code>. This will remove all build results, but not the configuration or any build system support artifacts. In most cases, this will solve build errors resulting from incremental build mismatches.</p></li>
+<li><p>Completely clean the build directory.</p>
+<p>If this does not work, the next step is to run <code>make dist-clean</code>, or removing the build output directory (<code>$BUILD</code>). This will clean all generated output, including your configuration. You will need to re-run <code>configure</code> after this step. A good idea is to run <code>make print-configuration</code> before running <code>make dist-clean</code>, as this will print your current <code>configure</code> command line. Here's a way to do this:</p>
+<pre><code>make print-configuration &gt; current-configuration
+make dist-clean
+bash configure $(cat current-configuration)
+make</code></pre></li>
+<li><p>Re-clone the Mercurial forest</p>
+<p>Sometimes the Mercurial repositories themselves gets in a state that causes the product to be un-buildable. In such a case, the simplest solution is often the &quot;sledgehammer approach&quot;: delete the entire forest, and re-clone it. If you have local changes, save them first to a different location using <code>hg export</code>.</p></li>
+</ol>
+<h3 id="specific-build-issues">Specific Build Issues</h3>
+<h4 id="clock-skew">Clock Skew</h4>
+<p>If you get an error message like this:</p>
+<pre><code>File &#39;xxx&#39; has modification time in the future.
+Clock skew detected. Your build may be incomplete.</code></pre>
+<p>then the clock on your build machine is out of sync with the timestamps on the source files. Other errors, apparently unrelated but in fact caused by the clock skew, can occur along with the clock skew warnings. These secondary errors may tend to obscure the fact that the true root cause of the problem is an out-of-sync clock.</p>
+<p>If you see these warnings, reset the clock on the build machine, run <code>make clean</code> and restart the build.</p>
+<h4 id="out-of-memory-errors">Out of Memory Errors</h4>
+<p>On Solaris, you might get an error message like this:</p>
+<pre><code>Trouble writing out table to disk</code></pre>
+<p>To solve this, increase the amount of swap space on your build machine.</p>
+<p>On Windows, you might get error messages like this:</p>
+<pre><code>fatal error - couldn&#39;t allocate heap
+cannot create ... Permission denied
+spawn failed</code></pre>
+<p>This can be a sign of a Cygwin problem. See the information about solving problems in the <a href="#cygwin">Cygwin</a> section. Rebooting the computer might help temporarily.</p>
+<h3 id="getting-help">Getting Help</h3>
+<p>If none of the suggestions in this document helps you, or if you find what you believe is a bug in the build system, please contact the Build Group by sending a mail to <a href="mailto:build-dev@openjdk.java.net">build-dev@openjdk.java.net</a>. Please include the relevant parts of the configure and/or build log.</p>
+<p>If you need general help or advice about developing for OpenJDK, you can also contact the Adoption Group. See the section on <a href="#contributing-to-openjdk">Contributing to OpenJDK</a> for more information.</p>
+<h2 id="hints-and-suggestions-for-advanced-users">Hints and Suggestions for Advanced Users</h2>
+<h3 id="setting-up-a-forest-for-pushing-changes-defpath">Setting Up a Forest for Pushing Changes (defpath)</h3>
+<p>To help you prepare a proper push path for a Mercurial repository, there exists a useful tool known as <a href="http://openjdk.java.net/projects/code-tools/defpath">defpath</a>. It will help you setup a proper push path for pushing changes to OpenJDK.</p>
+<p>Install the extension by cloning <code>http://hg.openjdk.java.net/code-tools/defpath</code> and updating your <code>.hgrc</code> file. Here's one way to do this:</p>
+<pre><code>cd ~
+mkdir hg-ext
+cd hg-ext
+hg clone http://hg.openjdk.java.net/code-tools/defpath
+cat &lt;&lt; EOT &gt;&gt; ~/.hgrc
+[extensions]
+defpath=~/hg-ext/defpath/defpath.py
+EOT</code></pre>
+<p>You can now setup a proper push path using:</p>
+<pre><code>hg defpath -d -u &lt;your OpenJDK username&gt;</code></pre>
+<p>If you also have the <code>trees</code> extension installed in Mercurial, you will automatically get a <code>tdefpath</code> command, which is even more useful. By running <code>hg tdefpath -du &lt;username&gt;</code> in the top repository of your forest, all repos will get setup automatically. This is the recommended usage.</p>
+<h3 id="bash-completion">Bash Completion</h3>
+<p>The <code>configure</code> and <code>make</code> commands tries to play nice with bash command-line completion (using <code>&lt;tab&gt;</code> or <code>&lt;tab&gt;&lt;tab&gt;</code>). To use this functionality, make sure you enable completion in your <code>~/.bashrc</code> (see instructions for bash in your operating system).</p>
+<p>Make completion will work out of the box, and will complete valid make targets. For instance, typing <code>make jdk-i&lt;tab&gt;</code> will complete to <code>make jdk-image</code>.</p>
+<p>The <code>configure</code> script can get completion for options, but for this to work you need to help <code>bash</code> on the way. The standard way of running the script, <code>bash configure</code>, will not be understood by bash completion. You need <code>configure</code> to be the command to run. One way to achieve this is to add a simple helper script to your path:</p>
+<pre><code>cat &lt;&lt; EOT &gt; /tmp/configure
+#!/bin/bash
+if [ \$(pwd) = \$(cd \$(dirname \$0); pwd) ] ; then
+  echo &gt;&amp;2 &quot;Abort: Trying to call configure helper recursively&quot;
+  exit 1
+fi
+
+bash \$PWD/configure &quot;\$@&quot;
+EOT
+chmod +x /tmp/configure
+sudo mv /tmp/configure /usr/local/bin</code></pre>
+<p>Now <code>configure --en&lt;tab&gt;-dt&lt;tab&gt;</code> will result in <code>configure --enable-dtrace</code>.</p>
+<h3 id="using-multiple-configurations">Using Multiple Configurations</h3>
+<p>You can have multiple configurations for a single source forest. When you create a new configuration, run <code>configure --with-conf-name=&lt;name&gt;</code> to create a configuration with the name <code>&lt;name&gt;</code>. Alternatively, you can create a directory under <code>build</code> and run <code>configure</code> from there, e.g. <code>mkdir build/&lt;name&gt; &amp;&amp; cd build/&lt;name&gt; &amp;&amp; bash ../../configure</code>.</p>
+<p>Then you can build that configuration using <code>make CONF_NAME=&lt;name&gt;</code> or <code>make CONF=&lt;pattern&gt;</code>, where <code>&lt;pattern&gt;</code> is a substring matching one or several configurations, e.g. <code>CONF=debug</code>. The special empty pattern (<code>CONF=</code>) will match <em>all</em> available configuration, so <code>make CONF= hotspot</code> will build the <code>hotspot</code> target for all configurations. Alternatively, you can execute <code>make</code> in the configuration directory, e.g. <code>cd build/&lt;name&gt; &amp;&amp; make</code>.</p>
+<h3 id="handling-reconfigurations">Handling Reconfigurations</h3>
+<p>If you update the forest and part of the configure script has changed, the build system will force you to re-run <code>configure</code>.</p>
+<p>Most of the time, you will be fine by running <code>configure</code> again with the same arguments as the last time, which can easily be performed by <code>make reconfigure</code>. To simplify this, you can use the <code>CONF_CHECK</code> make control variable, either as <code>make CONF_CHECK=auto</code>, or by setting an environment variable. For instance, if you add <code>export CONF_CHECK=auto</code> to your <code>.bashrc</code> file, <code>make</code> will always run <code>reconfigure</code> automatically whenever the configure script has changed.</p>
+<p>You can also use <code>CONF_CHECK=ignore</code> to skip the check for a needed configure update. This might speed up the build, but comes at the risk of an incorrect build result. This is only recommended if you know what you're doing.</p>
+<p>From time to time, you will also need to modify the command line to <code>configure</code> due to changes. Use <code>make print-configure</code> to show the command line used for your current configuration.</p>
+<h3 id="using-fine-grained-make-targets">Using Fine-Grained Make Targets</h3>
+<p>The default behavior for make is to create consistent and correct output, at the expense of build speed, if necessary.</p>
+<p>If you are prepared to take some risk of an incorrect build, and know enough of the system to understand how things build and interact, you can speed up the build process considerably by instructing make to only build a portion of the product.</p>
+<h4 id="building-individual-modules">Building Individual Modules</h4>
+<p>The safe way to use fine-grained make targets is to use the module specific make targets. All source code in JDK 9 is organized so it belongs to a module, e.g. <code>java.base</code> or <code>jdk.jdwp.agent</code>. You can build only a specific module, by giving it as make target: <code>make jdk.jdwp.agent</code>. If the specified module depends on other modules (e.g. <code>java.base</code>), those modules will be built first.</p>
+<p>You can also specify a set of modules, just as you can always specify a set of make targets: <code>make jdk.crypto.cryptoki jdk.crypto.ec jdk.crypto.mscapi jdk.crypto.ucrypto</code></p>
+<h4 id="building-individual-module-phases">Building Individual Module Phases</h4>
+<p>The build process for each module is divided into separate phases. Not all modules need all phases. Which are needed depends on what kind of source code and other artifact the module consists of. The phases are:</p>
+<ul>
+<li><code>gensrc</code> (Generate source code to compile)</li>
+<li><code>gendata</code> (Generate non-source code artifacts)</li>
+<li><code>copy</code> (Copy resource artifacts)</li>
+<li><code>java</code> (Compile Java code)</li>
+<li><code>launchers</code> (Compile native executables)</li>
+<li><code>libs</code> (Compile native libraries)</li>
+<li><code>rmic</code> (Run the <code>rmic</code> tool)</li>
+</ul>
+<p>You can build only a single phase for a module by using the notation <code>$MODULE-$PHASE</code>. For instance, to build the <code>gensrc</code> phase for <code>java.base</code>, use <code>make java.base-gensrc</code>.</p>
+<p>Note that some phases may depend on others, e.g. <code>java</code> depends on <code>gensrc</code> (if present). Make will build all needed prerequisites before building the requested phase.</p>
+<h4 id="skipping-the-dependency-check">Skipping the Dependency Check</h4>
+<p>When using an iterative development style with frequent quick rebuilds, the dependency check made by make can take up a significant portion of the time spent on the rebuild. In such cases, it can be useful to bypass the dependency check in make.</p>
+<blockquote>
+<p><strong>Note that if used incorrectly, this can lead to a broken build!</strong></p>
+</blockquote>
+<p>To achieve this, append <code>-only</code> to the build target. For instance, <code>make jdk.jdwp.agent-java-only</code> will <em>only</em> build the <code>java</code> phase of the <code>jdk.jdwp.agent</code> module. If the required dependencies are not present, the build can fail. On the other hand, the execution time measures in milliseconds.</p>
+<p>A useful pattern is to build the first time normally (e.g. <code>make jdk.jdwp.agent</code>) and then on subsequent builds, use the <code>-only</code> make target.</p>
+<h4 id="rebuilding-part-of-java.base-jdk_filter">Rebuilding Part of java.base (JDK_FILTER)</h4>
+<p>If you are modifying files in <code>java.base</code>, which is the by far largest module in OpenJDK, then you need to rebuild all those files whenever a single file has changed. (This inefficiency will hopefully be addressed in JDK 10.)</p>
+<p>As a hack, you can use the make control variable <code>JDK_FILTER</code> to specify a pattern that will be used to limit the set of files being recompiled. For instance, <code>make java.base JDK_FILTER=javax/crypto</code> (or, to combine methods, <code>make java.base-java-only JDK_FILTER=javax/crypto</code>) will limit the compilation to files in the <code>javax.crypto</code> package.</p>
+<h3 id="learn-about-mercurial">Learn About Mercurial</h3>
+<p>To become an efficient OpenJDK developer, it is recommended that you invest in learning Mercurial properly. Here are some links that can get you started:</p>
+<ul>
+<li><a href="http://www.mercurial-scm.org/wiki/GitConcepts">Mercurial for git users</a></li>
+<li><a href="http://www.mercurial-scm.org/wiki/Tutorial">The official Mercurial tutorial</a></li>
+<li><a href="http://hginit.com/">hg init</a></li>
+<li><a href="http://hgbook.red-bean.com/read/">Mercurial: The Definitive Guide</a></li>
+</ul>
+<h2 id="understanding-the-build-system">Understanding the Build System</h2>
+<p>This section will give you a more technical description on the details of the build system.</p>
+<h3 id="configurations">Configurations</h3>
+<p>The build system expects to find one or more configuration. These are technically defined by the <code>spec.gmk</code> in a subdirectory to the <code>build</code> subdirectory. The <code>spec.gmk</code> file is generated by <code>configure</code>, and contains in principle the configuration (directly or by files included by <code>spec.gmk</code>).</p>
+<p>You can, in fact, select a configuration to build by pointing to the <code>spec.gmk</code> file with the <code>SPEC</code> make control variable, e.g. <code>make SPEC=$BUILD/spec.gmk</code>. While this is not the recommended way to call <code>make</code> as a user, it is what is used under the hood by the build system.</p>
+<h3 id="build-output-structure">Build Output Structure</h3>
+<p>The build output for a configuration will end up in <code>build/&lt;configuration name&gt;</code>, which we refer to as <code>$BUILD</code> in this document. The <code>$BUILD</code> directory contains the following important directories:</p>
+<pre><code>buildtools/
+configure-support/
+hotspot/
+images/
+jdk/
+make-support/
+support/
+test-results/
+test-support/</code></pre>
+<p>This is what they are used for:</p>
+<ul>
+<li><p><code>images</code>: This is the directory were the output of the <code>*-image</code> make targets end up. For instance, <code>make jdk-image</code> ends up in <code>images/jdk</code>.</p></li>
+<li><p><code>jdk</code>: This is the &quot;exploded image&quot;. After <code>make jdk</code>, you will be able to launch the newly built JDK by running <code>$BUILD/jdk/bin/java</code>.</p></li>
+<li><p><code>test-results</code>: This directory contains the results from running tests.</p></li>
+<li><p><code>support</code>: This is an area for intermediate files needed during the build, e.g. generated source code, object files and class files. Some noteworthy directories in <code>support</code> is <code>gensrc</code>, which contains the generated source code, and the <code>modules_*</code> directories, which contains the files in a per-module hierarchy that will later be collapsed into the <code>jdk</code> directory of the exploded image.</p></li>
+<li><p><code>buildtools</code>: This is an area for tools compiled for the build platform that are used during the rest of the build.</p></li>
+<li><p><code>hotspot</code>: This is an area for intermediate files needed when building hotspot.</p></li>
+<li><p><code>configure-support</code>, <code>make-support</code> and <code>test-support</code>: These directories contain files that are needed by the build system for <code>configure</code>, <code>make</code> and for running tests.</p></li>
+</ul>
+<h3 id="fixpath">Fixpath</h3>
+<p>Windows path typically look like <code>C:\User\foo</code>, while Unix paths look like <code>/home/foo</code>. Tools with roots from Unix often experience issues related to this mismatch when running on Windows.</p>
+<p>In the OpenJDK build, we always use Unix paths internally, and only just before calling a tool that does not understand Unix paths do we convert them to Windows paths.</p>
+<p>This conversion is done by the <code>fixpath</code> tool, which is a small wrapper that modifies unix-style paths to Windows-style paths in command lines. Fixpath is compiled automatically by <code>configure</code>.</p>
+<h3 id="native-debug-symbols">Native Debug Symbols</h3>
+<p>Native libraries and executables can have debug symbol (and other debug information) associated with them. How this works is very much platform dependent, but a common problem is that debug symbol information takes a lot of disk space, but is rarely needed by the end user.</p>
+<p>The OpenJDK supports different methods on how to handle debug symbols. The method used is selected by <code>--with-native-debug-symbols</code>, and available methods are <code>none</code>, <code>internal</code>, <code>external</code>, <code>zipped</code>.</p>
+<ul>
+<li><p><code>none</code> means that no debug symbols will be generated during the build.</p></li>
+<li><p><code>internal</code> means that debug symbols will be generated during the build, and they will be stored in the generated binary.</p></li>
+<li><p><code>external</code> means that debug symbols will be generated during the build, and after the compilation, they will be moved into a separate <code>.debuginfo</code> file. (This was previously known as FDS, Full Debug Symbols).</p></li>
+<li><p><code>zipped</code> is like <code>external</code>, but the .debuginfo file will also be zipped into a <code>.diz</code> file.</p></li>
+</ul>
+<p>When building for distribution, <code>zipped</code> is a good solution. Binaries built with <code>internal</code> is suitable for use by developers, since they facilitate debugging, but should be stripped before distributed to end users.</p>
+<h3 id="autoconf-details">Autoconf Details</h3>
+<p>The <code>configure</code> script is based on the autoconf framework, but in some details deviate from a normal autoconf <code>configure</code> script.</p>
+<p>The <code>configure</code> script in the top level directory of OpenJDK is just a thin wrapper that calls <code>common/autoconf/configure</code>. This in turn provides functionality that is not easily expressed in the normal Autoconf framework, and then calls into the core of the <code>configure</code> script, which is the <code>common/autoconf/generated-configure.sh</code> file.</p>
+<p>As the name implies, this file is generated by Autoconf. It is checked in after regeneration, to alleviate the common user to have to install Autoconf.</p>
+<p>The build system will detect if the Autoconf source files have changed, and will trigger a regeneration of <code>common/autoconf/generated-configure.sh</code> if needed. You can also manually request such an update by <code>bash common/autoconf/autogen.sh</code>.</p>
+<p>If you make changes to the build system that requires a re-generation, note the following:</p>
+<ul>
+<li><p>You must use <em>exactly</em> version 2.69 of autoconf for your patch to be accepted. This is to avoid spurious changes in the generated file. Note that Ubuntu 16.04 ships a patched version of autoconf which claims to be 2.69, but is not.</p></li>
+<li><p>You do not need to include the generated file in reviews.</p></li>
+<li><p>If the generated file needs updating, the Oracle JDK closed counter-part will also need to be updated. It is very much appreciated if you ask for an Oracle engineer to sponsor your push so this can be made in tandem.</p></li>
+</ul>
+<h3 id="developing-the-build-system-itself">Developing the Build System Itself</h3>
+<p>This section contains a few remarks about how to develop for the build system itself. It is not relevant if you are only making changes in the product source code.</p>
+<p>While technically using <code>make</code>, the make source files of the OpenJDK does not resemble most other Makefiles. Instead of listing specific targets and actions (perhaps using patterns), the basic modus operandi is to call a high-level function (or properly, macro) from the API in <code>make/common</code>. For instance, to compile all classes in the <code>jdk.internal.foo</code> package in the <code>jdk.foo</code> module, a call like this would be made:</p>
+<pre><code>$(eval $(call SetupJavaCompilation, BUILD_FOO_CLASSES, \
+    SETUP := GENERATE_OLDBYTECODE, \
+    SRC := $(JDK_TOPDIR)/src/jkd.foo/share/classes, \
+    INCLUDES := jdk/internal/foo, \
+    BIN := $(SUPPORT_OUTPUTDIR)/foo_classes, \
+))</code></pre>
+<p>By encapsulating and expressing the high-level knowledge of <em>what</em> should be done, rather than <em>how</em> it should be done (as is normal in Makefiles), we can build a much more powerful and flexible build system.</p>
+<p>Correct dependency tracking is paramount. Sloppy dependency tracking will lead to improper parallelization, or worse, race conditions.</p>
+<p>To test for/debug race conditions, try running <code>make JOBS=1</code> and <code>make JOBS=100</code> and see if it makes any difference. (It shouldn't).</p>
+<p>To compare the output of two different builds and see if, and how, they differ, run <code>$BUILD1/compare.sh -o $BUILD2</code>, where <code>$BUILD1</code> and <code>$BUILD2</code> are the two builds you want to compare.</p>
+<p>To automatically build two consecutive versions and compare them, use <code>COMPARE_BUILD</code>. The value of <code>COMPARE_BUILD</code> is a set of variable=value assignments, like this:</p>
+<pre><code>make COMPARE_BUILD=CONF=--enable-new-hotspot-feature:MAKE=hotspot</code></pre>
+<p>See <code>make/InitSupport.gmk</code> for details on how to use <code>COMPARE_BUILD</code>.</p>
+<p>To analyze build performance, run with <code>LOG=trace</code> and check <code>$BUILD/build-trace-time.log</code>. Use <code>JOBS=1</code> to avoid parallelism.</p>
+<p>Please check that you adhere to the <a href="http://openjdk.java.net/groups/build/doc/code-conventions.html">Code Conventions for the Build System</a> before submitting patches. Also see the section in <a href="#autoconf-details">Autoconf Details</a> about the generated configure script.</p>
+<h2 id="contributing-to-openjdk">Contributing to OpenJDK</h2>
+<p>So, now you've build your OpenJDK, and made your first patch, and want to contribute it back to the OpenJDK community.</p>
+<p>First of all: Thank you! We gladly welcome your contribution to the OpenJDK. However, please bear in mind that OpenJDK is a massive project, and we must ask you to follow our rules and guidelines to be able to accept your contribution.</p>
+<p>The official place to start is the <a href="http://openjdk.java.net/contribute/">'How to contribute' page</a>. There is also an official (but somewhat outdated and skimpy on details) <a href="http://openjdk.java.net/guide/">Developer's Guide</a>.</p>
+<p>If this seems overwhelming to you, the Adoption Group is there to help you! A good place to start is their <a href="https://wiki.openjdk.java.net/display/Adoption/New+Contributor">'New Contributor' page</a>, or start reading the comprehensive <a href="https://adoptopenjdk.gitbooks.io/adoptopenjdk-getting-started-kit/en/">Getting Started Kit</a>. The Adoption Group will also happily answer any questions you have about contributing. Contact them by <a href="http://mail.openjdk.java.net/mailman/listinfo/adoption-discuss">mail</a> or <a href="http://openjdk.java.net/irc/">IRC</a>.</p>
+</body>
+</html>
--- /dev/null	Thu Jan 01 00:00:00 1970 +0000
+++ b/common/doc/building.md	Thu Aug 17 12:17:39 2017 -0700
@@ -0,0 +1,1883 @@
+% Building OpenJDK
+
+## TL;DR (Instructions for the Impatient)
+
+If you are eager to try out building OpenJDK, these simple steps works most of
+the time. They assume that you have installed Mercurial (and Cygwin if running
+on Windows) and cloned the top-level OpenJDK repository that you want to build.
+
+ 1. [Get the complete source code](#getting-the-source-code): \
+    `bash get_source.sh`
+
+ 2. [Run configure](#running-configure): \
+    `bash configure`
+
+    If `configure` fails due to missing dependencies (to either the
+    [toolchain](#native-compiler-toolchain-requirements), [external libraries](
+    #external-library-requirements) or the [boot JDK](#boot-jdk-requirements)),
+    most of the time it prints a suggestion on how to resolve the situation on
+    your platform. Follow the instructions, and try running `bash configure`
+    again.
+
+ 3. [Run make](#running-make): \
+    `make images`
+
+ 4. Verify your newly built JDK: \
+    `./build/*/images/jdk/bin/java -version`
+
+ 5. [Run basic tests](##running-tests): \
+    `make run-test-tier1`
+
+If any of these steps failed, or if you want to know more about build
+requirements or build functionality, please continue reading this document.
+
+## Introduction
+
+OpenJDK is a complex software project. Building it requires a certain amount of
+technical expertise, a fair number of dependencies on external software, and
+reasonably powerful hardware.
+
+If you just want to use OpenJDK and not build it yourself, this document is not
+for you. See for instance [OpenJDK installation](
+http://openjdk.java.net/install) for some methods of installing a prebuilt
+OpenJDK.
+
+## Getting the Source Code
+
+OpenJDK uses [Mercurial](http://www.mercurial-scm.org) for source control. The
+source code is contained not in a single Mercurial repository, but in a tree
+("forest") of interrelated repositories. You will need to check out all of the
+repositories to be able to build OpenJDK. To assist you in dealing with this
+somewhat unusual arrangement, there are multiple tools available, which are
+explained below.
+
+In any case, make sure you are getting the correct version. At the [OpenJDK
+Mercurial server](http://hg.openjdk.java.net/) you can see a list of all
+available forests. If you want to build an older version, e.g. JDK 8, it is
+recommended that you get the `jdk8u` forest, which contains incremental
+updates, instead of the `jdk8` forest, which was frozen at JDK 8 GA.
+
+If you are new to Mercurial, a good place to start is the [Mercurial Beginner's
+Guide](http://www.mercurial-scm.org/guide). The rest of this document assumes a
+working knowledge of Mercurial.
+
+### Special Considerations
+
+For a smooth building experience, it is recommended that you follow these rules
+on where and how to check out the source code.
+
+  * Do not check out the source code in a path which contains spaces. Chances
+    are the build will not work. This is most likely to be an issue on Windows
+    systems.
+
+  * Do not check out the source code in a path which has a very long name or is
+    nested many levels deep. Chances are you will hit an OS limitation during
+    the build.
+
+  * Put the source code on a local disk, not a network share. If possible, use
+    an SSD. The build process is very disk intensive, and having slow disk
+    access will significantly increase build times. If you need to use a
+    network share for the source code, see below for suggestions on how to keep
+    the build artifacts on a local disk.
+
+  * On Windows, extra care must be taken to make sure the [Cygwin](#cygwin)
+    environment is consistent. It is recommended that you follow this
+    procedure:
+
+      * Create the directory that is going to contain the top directory of the
+        OpenJDK clone by using the `mkdir` command in the Cygwin bash shell.
+        That is, do *not* create it using Windows Explorer. This will ensure
+        that it will have proper Cygwin attributes, and that it's children will
+        inherit those attributes.
+
+      * Do not put the OpenJDK clone in a path under your Cygwin home
+        directory. This is especially important if your user name contains
+        spaces and/or mixed upper and lower case letters.
+
+      * Clone the OpenJDK repository using the Cygwin command line `hg` client
+        as instructed in this document. That is, do *not* use another Mercurial
+        client such as TortoiseHg.
+
+    Failure to follow this procedure might result in hard-to-debug build
+    problems.
+
+### Using get\_source.sh
+
+The simplest way to get the entire forest is probably to clone the top-level
+repository and then run the `get_source.sh` script, like this:
+
+```
+hg clone http://hg.openjdk.java.net/jdk9/jdk9
+cd jdk9
+bash get_source.sh
+```
+
+The first time this is run, it will clone all the sub-repositories. Any
+subsequent execution of the script will update all sub-repositories to the
+latest revision.
+
+### Using hgforest.sh
+
+The `hgforest.sh` script is more expressive than `get_source.sh`. It takes any
+number of arguments, and runs `hg` with those arguments on each sub-repository
+in the forest. The `get_source.sh` script is basically a simple wrapper that
+runs either `hgforest.sh clone` or `hgforest.sh pull -u`.
+
+  * Cloning the forest:
+    ```
+    hg clone http://hg.openjdk.java.net/jdk9/jdk9
+    cd jdk9
+    bash common/bin/hgforest.sh clone
+    ```
+
+  * Pulling and updating the forest:
+    ```
+    bash common/bin/hgforest.sh pull -u
+    ```
+
+  * Merging over the entire forest:
+    ```
+    bash common/bin/hgforest.sh merge
+    ```
+
+### Using the Trees Extension
+
+The trees extension is a Mercurial add-on that helps you deal with the forest.
+More information is available on the [Code Tools trees page](
+http://openjdk.java.net/projects/code-tools/trees).
+
+#### Installing the Extension
+
+Install the extension by cloning `http://hg.openjdk.java.net/code-tools/trees`
+and updating your `.hgrc` file. Here's one way to do this:
+
+```
+cd ~
+mkdir hg-ext
+cd hg-ext
+hg clone http://hg.openjdk.java.net/code-tools/trees
+cat << EOT >> ~/.hgrc
+[extensions]
+trees=~/hg-ext/trees/trees.py
+EOT
+```
+
+#### Initializing the Tree
+
+The trees extension needs to know the structure of the forest. If you have
+already cloned the entire forest using another method, you can initialize the
+forest like this:
+
+```
+hg tconf --set --walk --depth
+```
+
+Or you can clone the entire forest at once, if you substitute `clone` with
+`tclone` when cloning the top-level repository, e.g. like this:
+
+```
+hg tclone http://hg.openjdk.java.net/jdk9/jdk9
+```
+
+In this case, the forest will be properly initialized from the start.
+
+#### Other Operations
+
+The trees extensions supplement many common operations with a trees version by
+prefixing a `t` to the normal Mercurial command, e.g. `tcommit`, `tstatus` or
+`tmerge`. For instance, to update the entire forest:
+
+```
+hg tpull -u
+```
+
+## Build Hardware Requirements
+
+OpenJDK is a massive project, and require machines ranging from decent to
+powerful to be able to build in a reasonable amount of time, or to be able to
+complete a build at all.
+
+We *strongly* recommend usage of an SSD disk for the build, since disk speed is
+one of the limiting factors for build performance.
+
+### Building on x86
+
+At a minimum, a machine with 2-4 cores is advisable, as well as 2-4 GB of RAM.
+(The more cores to use, the more memory you need.) At least 6 GB of free disk
+space is required (8 GB minimum for building on Solaris).
+
+Even for 32-bit builds, it is recommended to use a 64-bit build machine, and
+instead create a 32-bit target using `--with-target-bits=32`.
+
+### Building on sparc
+
+At a minimum, a machine with 4 cores is advisable, as well as 4 GB of RAM. (The
+more cores to use, the more memory you need.) At least 8 GB of free disk space
+is required.
+
+### Building on arm/aarch64
+
+This is not recommended. Instead, see the section on [Cross-compiling](
+#cross-compiling).
+
+## Operating System Requirements
+
+The mainline OpenJDK project supports Linux, Solaris, macOS, AIX and Windows.
+Support for other operating system, e.g. BSD, exists in separate "port"
+projects.
+
+In general, OpenJDK can be built on a wide range of versions of these operating
+systems, but the further you deviate from what is tested on a daily basis, the
+more likely you are to run into problems.
+
+This table lists the OS versions used by Oracle when building JDK 9. Such
+information is always subject to change, but this table is up to date at the
+time of writing.
+
+ Operating system   Vendor/version used
+ -----------------  -------------------------------------------------------
+ Linux              Oracle Enterprise Linux 6.4 / 7.1 (using kernel 3.8.13)
+ Solaris            Solaris 11.1 SRU 21.4.1 / 11.2 SRU 5.5
+ macOS              Mac OS X 10.9 (Mavericks) / 10.10 (Yosemite)
+ Windows            Windows Server 2012 R2
+
+The double version numbers for Linux, Solaris and macOS is due to the hybrid
+model used at Oracle, where header files and external libraries from an older
+version is used when building on a more modern version of the OS.
+
+The Build Group has a wiki page with [Supported Build Platforms](
+https://wiki.openjdk.java.net/display/Build/Supported+Build+Platforms). From
+time to time, this is updated by the community to list successes or failures of
+building on different platforms.
+
+### Windows
+
+Windows XP is not a supported platform, but all newer Windows should be able to
+build OpenJDK.
+
+On Windows, it is important that you pay attention to the instructions in the
+[Special Considerations](#special-considerations).
+
+Windows is the only non-POSIX OS supported by OpenJDK, and as such, requires
+some extra care. A POSIX support layer is required to build on Windows. For
+OpenJDK 9, the only supported such layer is Cygwin. (Msys is no longer
+supported due to a too old bash; msys2 and the new Windows Subsystem for Linux
+(WSL) would likely be possible to support in a future version but that would
+require a community effort to implement.)
+
+Internally in the build system, all paths are represented as Unix-style paths,
+e.g. `/cygdrive/c/hg/jdk9/Makefile` rather than `C:\hg\jdk9\Makefile`. This
+rule also applies to input to the build system, e.g. in arguments to
+`configure`. So, use `--with-freetype=/cygdrive/c/freetype` rather than
+`--with-freetype=c:\freetype`. For details on this conversion, see the section
+on [Fixpath](#fixpath).
+
+#### Cygwin
+
+A functioning [Cygwin](http://www.cygwin.com/) environment is thus required for
+building OpenJDK on Windows. If you have a 64-bit OS, we strongly recommend
+using the 64-bit version of Cygwin.
+
+**Note:** Cygwin has a model of continuously updating all packages without any
+easy way to install or revert to a specific version of a package. This means
+that whenever you add or update a package in Cygwin, you might (inadvertently)
+update tools that are used by the OpenJDK build process, and that can cause
+unexpected build problems.
+
+OpenJDK requires GNU Make 4.0 or greater on Windows. This is usually not a
+problem, since Cygwin currently only distributes GNU Make at a version above
+4.0.
+
+Apart from the basic Cygwin installation, the following packages must also be
+installed:
+
+  * `make`
+  * `zip`
+  * `unzip`
+
+Often, you can install these packages using the following command line:
+```
+<path to Cygwin setup>/setup-x86_64 -q -P make -P unzip -P zip
+```
+
+Unfortunately, Cygwin can be unreliable in certain circumstances. If you
+experience build tool crashes or strange issues when building on Windows,
+please check the Cygwin FAQ on the ["BLODA" list](
+https://cygwin.com/faq/faq.html#faq.using.bloda) and the section on [fork()
+failures](https://cygwin.com/faq/faq.html#faq.using.fixing-fork-failures).
+
+### Solaris
+
+See `make/devkit/solaris11.1-package-list.txt` for a list of recommended
+packages to install when building on Solaris. The versions specified in this
+list is the versions used by the daily builds at Oracle, and is likely to work
+properly.
+
+Older versions of Solaris shipped a broken version of `objcopy`. At least
+version 2.21.1 is needed, which is provided by Solaris 11 Update 1. Objcopy is
+needed if you want to have external debug symbols. Please make sure you are
+using at least version 2.21.1 of objcopy, or that you disable external debug
+symbols.
+
+### macOS
+
+Apple is using a quite aggressive scheme of pushing OS updates, and coupling
+these updates with required updates of Xcode. Unfortunately, this makes it
+difficult for a project like OpenJDK to keep pace with a continuously updated
+machine running macOS. See the section on [Apple Xcode](#apple-xcode) on some
+strategies to deal with this.
+
+It is recommended that you use at least Mac OS X 10.9 (Mavericks). At the time
+of writing, OpenJDK has been successfully compiled on macOS versions up to
+10.12.5 (Sierra), using XCode 8.3.2 and `--disable-warnings-as-errors`.
+
+The standard macOS environment contains the basic tooling needed to build, but
+for external libraries a package manager is recommended. OpenJDK uses
+[homebrew](https://brew.sh/) in the examples, but feel free to use whatever
+manager you want (or none).
+
+### Linux
+
+It is often not much problem to build OpenJDK on Linux. The only general advice
+is to try to use the compilers, external libraries and header files as provided
+by your distribution.
+
+The basic tooling is provided as part of the core operating system, but you
+will most likely need to install developer packages.
+
+For apt-based distributions (Debian, Ubuntu, etc), try this:
+```
+sudo apt-get install build-essential
+```
+
+For rpm-based distributions (Fedora, Red Hat, etc), try this:
+```
+sudo yum groupinstall "Development Tools"
+```
+
+### AIX
+
+The regular builds by SAP is using AIX version 7.1, but AIX 5.3 is also
+supported. See the [OpenJDK PowerPC Port Status Page](
+http://cr.openjdk.java.net/~simonis/ppc-aix-port) for details.
+
+## Native Compiler (Toolchain) Requirements
+
+Large portions of OpenJDK consists of native code, that needs to be compiled to
+be able to run on the target platform. In theory, toolchain and operating
+system should be independent factors, but in practice there's more or less a
+one-to-one correlation between target operating system and toolchain.
+
+ Operating system   Supported toolchain
+ ------------------ -------------------------
+ Linux              gcc, clang
+ macOS              Apple Xcode (using clang)
+ Solaris            Oracle Solaris Studio
+ AIX                IBM XL C/C++
+ Windows            Microsoft Visual Studio
+
+Please see the individual sections on the toolchains for version
+recommendations. As a reference, these versions of the toolchains are used, at
+the time of writing, by Oracle for the daily builds of OpenJDK. It should be
+possible to compile OpenJDK with both older and newer versions, but the closer
+you stay to this list, the more likely you are to compile successfully without
+issues.
+
+ Operating system   Toolchain version
+ ------------------ -------------------------------------------------------
+ Linux              gcc 4.9.2
+ macOS              Apple Xcode 6.3 (using clang 6.1.0)
+ Solaris            Oracle Solaris Studio 12.4 (with compiler version 5.13)
+ Windows            Microsoft Visual Studio 2013 update 4
+
+### gcc
+
+The minimum accepted version of gcc is 4.3. Older versions will not be accepted
+by `configure`.
+
+However, gcc 4.3 is quite old and OpenJDK is not regularly tested on this
+version, so it is recommended to use a more modern gcc.
+
+OpenJDK 9 includes patches that should allow gcc 6 to compile, but this should
+be considered experimental.
+
+In general, any version between these two should be usable.
+
+### clang
+
+The minimum accepted version of clang is 3.2. Older versions will not be
+accepted by `configure`.
+
+To use clang instead of gcc on Linux, use `--with-toolchain-type=clang`.
+
+### Apple Xcode
+
+The oldest supported version of Xcode is 5.
+
+You will need the Xcode command lines developers tools to be able to build
+OpenJDK. (Actually, *only* the command lines tools are needed, not the IDE.)
+The simplest way to install these is to run:
+```
+xcode-select --install
+```
+
+It is advisable to keep an older version of Xcode for building OpenJDK when
+updating Xcode. This [blog page](
+http://iosdevelopertips.com/xcode/install-multiple-versions-of-xcode.html) has
+good suggestions on managing multiple Xcode versions. To use a specific version
+of Xcode, use `xcode-select -s` before running `configure`, or use
+`--with-toolchain-path` to point to the version of Xcode to use, e.g.
+`configure --with-toolchain-path=/Applications/Xcode5.app/Contents/Developer/usr/bin`
+
+If you have recently (inadvertently) updated your OS and/or Xcode version, and
+OpenJDK can no longer be built, please see the section on [Problems with the
+Build Environment](#problems-with-the-build-environment), and [Getting
+Help](#getting-help) to find out if there are any recent, non-merged patches
+available for this update.
+
+### Oracle Solaris Studio
+
+The minimum accepted version of the Solaris Studio compilers is 5.13
+(corresponding to Solaris Studio 12.4). Older versions will not be accepted by
+configure.
+
+The Solaris Studio installation should contain at least these packages:
+
+ Package                                            Version
+ -------------------------------------------------- -------------
+ developer/solarisstudio-124/backend                12.4-1.0.6.0
+ developer/solarisstudio-124/c++                    12.4-1.0.10.0
+ developer/solarisstudio-124/cc                     12.4-1.0.4.0
+ developer/solarisstudio-124/library/c++-libs       12.4-1.0.10.0
+ developer/solarisstudio-124/library/math-libs      12.4-1.0.0.1
+ developer/solarisstudio-124/library/studio-gccrt   12.4-1.0.0.1
+ developer/solarisstudio-124/studio-common          12.4-1.0.0.1
+ developer/solarisstudio-124/studio-ja              12.4-1.0.0.1
+ developer/solarisstudio-124/studio-legal           12.4-1.0.0.1
+ developer/solarisstudio-124/studio-zhCN            12.4-1.0.0.1
+
+Compiling with Solaris Studio can sometimes be finicky. This is the exact
+version used by Oracle, which worked correctly at the time of writing:
+```
+$ cc -V
+cc: Sun C 5.13 SunOS_i386 2014/10/20
+$ CC -V
+CC: Sun C++ 5.13 SunOS_i386 151846-10 2015/10/30
+```
+
+### Microsoft Visual Studio
+
+The minimum accepted version of Visual Studio is 2010. Older versions will not
+be accepted by `configure`. The maximum accepted version of Visual Studio is
+2013.
+
+If you have multiple versions of Visual Studio installed, `configure` will by
+default pick the latest. You can request a specific version to be used by
+setting `--with-toolchain-version`, e.g. `--with-toolchain-version=2010`.
+
+If you get `LINK: fatal error LNK1123: failure during conversion to COFF: file
+invalid` when building using Visual Studio 2010, you have encountered
+[KB2757355](http://support.microsoft.com/kb/2757355), a bug triggered by a
+specific installation order. However, the solution suggested by the KB article
+does not always resolve the problem. See [this stackoverflow discussion](
+https://stackoverflow.com/questions/10888391) for other suggestions.
+
+### IBM XL C/C++
+
+The regular builds by SAP is using version 12.1, described as `IBM XL C/C++ for
+AIX, V12.1 (5765-J02, 5725-C72) Version: 12.01.0000.0017`.
+
+See the [OpenJDK PowerPC Port Status Page](
+http://cr.openjdk.java.net/~simonis/ppc-aix-port) for details.
+
+## Boot JDK Requirements
+
+Paradoxically, building OpenJDK requires a pre-existing JDK. This is called the
+"boot JDK". The boot JDK does not have to be OpenJDK, though. If you are
+porting OpenJDK to a new platform, chances are that there already exists
+another JDK for that platform that is usable as boot JDK.
+
+The rule of thumb is that the boot JDK for building JDK major version *N*
+should be an JDK of major version *N-1*, so for building JDK 9 a JDK 8 would be
+suitable as boot JDK. However, OpenJDK should be able to "build itself", so an
+up-to-date build of the current OpenJDK source is an acceptable alternative. If
+you are following the *N-1* rule, make sure you got the latest update version,
+since JDK 8 GA might not be able to build JDK 9 on all platforms.
+
+If the Boot JDK is not automatically detected, or the wrong JDK is picked, use
+`--with-boot-jdk` to point to the JDK to use.
+
+### JDK 8 on Linux
+
+On apt-based distros (like Debian and Ubuntu), `sudo apt-get install
+openjdk-8-jdk` is typically enough to install OpenJDK 8. On rpm-based distros
+(like Fedora and Red Hat), try `sudo yum install java-1.8.0-openjdk-devel`.
+
+### JDK 8 on Windows
+
+No pre-compiled binaries of OpenJDK 8 are readily available for Windows at the
+time of writing. An alternative is to download the [Oracle JDK](
+http://www.oracle.com/technetwork/java/javase/downloads). Another is the [Adopt
+OpenJDK Project](https://adoptopenjdk.net/), which publishes experimental
+prebuilt binaries for Windows.
+
+### JDK 8 on macOS
+
+No pre-compiled binaries of OpenJDK 8 are readily available for macOS at the
+time of writing. An alternative is to download the [Oracle JDK](
+http://www.oracle.com/technetwork/java/javase/downloads), or to install it
+using `brew cask install java`. Another option is the [Adopt OpenJDK Project](
+https://adoptopenjdk.net/), which publishes experimental prebuilt binaries for
+macOS.
+
+### JDK 8 on AIX
+
+No pre-compiled binaries of OpenJDK 8 are readily available for AIX at the
+time of writing. A starting point for working with OpenJDK on AIX is
+the [PowerPC/AIX Port Project](http://openjdk.java.net/projects/ppc-aix-port/).
+
+## External Library Requirements
+
+Different platforms require different external libraries. In general, libraries
+are not optional - that is, they are either required or not used.
+
+If a required library is not detected by `configure`, you need to provide the
+path to it. There are two forms of the `configure` arguments to point to an
+external library: `--with-<LIB>=<path>` or `--with-<LIB>-include=<path to
+include> --with-<LIB>-lib=<path to lib>`. The first variant is more concise,
+but require the include files an library files to reside in a default hierarchy
+under this directory. In most cases, it works fine.
+
+As a fallback, the second version allows you to point to the include directory
+and the lib directory separately.
+
+### FreeType
+
+FreeType2 from [The FreeType Project](http://www.freetype.org/) is required on
+all platforms. At least version 2.3 is required.
+
+  * To install on an apt-based Linux, try running `sudo apt-get install
+    libcups2-dev`.
+  * To install on an rpm-based Linux, try running `sudo yum install
+    cups-devel`.
+  * To install on Solaris, try running `pkg install system/library/freetype-2`.
+  * To install on macOS, try running `brew install freetype`.
+  * To install on Windows, see [below](#building-freetype-on-windows).
+
+Use `--with-freetype=<path>` if `configure` does not properly locate your
+FreeType files.
+
+#### Building FreeType on Windows
+
+On Windows, there is no readily available compiled version of FreeType. OpenJDK
+can help you compile FreeType from source. Download the FreeType sources and
+unpack them into an arbitrary directory:
+
+```
+wget http://download.savannah.gnu.org/releases/freetype/freetype-2.5.3.tar.gz
+tar -xzf freetype-2.5.3.tar.gz
+```
+
+Then run `configure` with `--with-freetype-src=<freetype_src>`. This will
+automatically build the freetype library into `<freetype_src>/lib64` for 64-bit
+builds or into `<freetype_src>/lib32` for 32-bit builds. Afterwards you can
+always use `--with-freetype-include=<freetype_src>/include` and
+`--with-freetype-lib=<freetype_src>/lib[32|64]` for other builds.
+
+Alternatively you can unpack the sources like this to use the default
+directory:
+
+```
+tar --one-top-level=$HOME/freetype --strip-components=1 -xzf freetype-2.5.3.tar.gz
+```
+
+### CUPS
+
+CUPS, [Common UNIX Printing System](http://www.cups.org) header files are
+required on all platforms, except Windows. Often these files are provided by
+your operating system.
+
+  * To install on an apt-based Linux, try running `sudo apt-get install
+    libcups2-dev`.
+  * To install on an rpm-based Linux, try running `sudo yum install
+    cups-devel`.
+  * To install on Solaris, try running `pkg install print/cups`.
+
+Use `--with-cups=<path>` if `configure` does not properly locate your CUPS
+files.
+
+### X11
+
+Certain [X11](http://www.x.org/) libraries and include files are required on
+Linux and Solaris.
+
+  * To install on an apt-based Linux, try running `sudo apt-get install
+    libx11-dev libxext-dev libxrender-dev libxtst-dev libxt-dev`.
+  * To install on an rpm-based Linux, try running `sudo yum install
+    libXtst-devel libXt-devel libXrender-devel libXi-devel`.
+  * To install on Solaris, try running `pkg install x11/header/x11-protocols
+    x11/library/libice x11/library/libpthread-stubs x11/library/libsm
+    x11/library/libx11 x11/library/libxau x11/library/libxcb
+    x11/library/libxdmcp x11/library/libxevie x11/library/libxext
+    x11/library/libxrender x11/library/libxscrnsaver x11/library/libxtst
+    x11/library/toolkit/libxt`.
+
+Use `--with-x=<path>` if `configure` does not properly locate your X11 files.
+
+### ALSA
+
+ALSA, [Advanced Linux Sound Architecture](https://www.alsa-project.org/) is
+required on Linux. At least version 0.9.1 of ALSA is required.
+
+  * To install on an apt-based Linux, try running `sudo apt-get install
+    libasound2-dev`.
+  * To install on an rpm-based Linux, try running `sudo yum install
+    alsa-lib-devel`.
+
+Use `--with-alsa=<path>` if `configure` does not properly locate your ALSA
+files.
+
+### libffi
+
+libffi, the [Portable Foreign Function Interface Library](
+http://sourceware.org/libffi) is required when building the Zero version of
+Hotspot.
+
+  * To install on an apt-based Linux, try running `sudo apt-get install
+    libffi-dev`.
+  * To install on an rpm-based Linux, try running `sudo yum install
+    libffi-devel`.
+
+Use `--with-libffi=<path>` if `configure` does not properly locate your libffi
+files.
+
+### libelf
+
+libelf from the [elfutils project](http://sourceware.org/elfutils) is required
+when building the AOT feature of Hotspot.
+
+  * To install on an apt-based Linux, try running `sudo apt-get install
+    libelf-dev`.
+  * To install on an rpm-based Linux, try running `sudo yum install
+    elfutils-libelf-devel`.
+
+Use `--with-libelf=<path>` if `configure` does not properly locate your libelf
+files.
+
+## Other Tooling Requirements
+
+### GNU Make
+
+OpenJDK requires [GNU Make](http://www.gnu.org/software/make). No other flavors
+of make are supported.
+
+At least version 3.81 of GNU Make must be used. For distributions supporting
+GNU Make 4.0 or above, we strongly recommend it. GNU Make 4.0 contains useful
+functionality to handle parallel building (supported by `--with-output-sync`)
+and speed and stability improvements.
+
+Note that `configure` locates and verifies a properly functioning version of
+`make` and stores the path to this `make` binary in the configuration. If you
+start a build using `make` on the command line, you will be using the version
+of make found first in your `PATH`, and not necessarily the one stored in the
+configuration. This initial make will be used as "bootstrap make", and in a
+second stage, the make located by `configure` will be called. Normally, this
+will present no issues, but if you have a very old `make`, or a non-GNU Make
+`make` in your path, this might cause issues.
+
+If you want to override the default make found by `configure`, use the `MAKE`
+configure variable, e.g. `configure MAKE=/opt/gnu/make`.
+
+On Solaris, it is common to call the GNU version of make by using `gmake`.
+
+### GNU Bash
+
+OpenJDK requires [GNU Bash](http://www.gnu.org/software/bash). No other shells
+are supported.
+
+At least version 3.2 of GNU Bash must be used.
+
+### Autoconf
+
+If you want to modify the build system itself, you need to install [Autoconf](
+http://www.gnu.org/software/autoconf).
+
+However, if you only need to build OpenJDK or if you only edit the actual
+OpenJDK source files, there is no dependency on autoconf, since the source
+distribution includes a pre-generated `configure` shell script.
+
+See the section on [Autoconf Details](#autoconf-details) for details on how
+OpenJDK uses autoconf. This is especially important if you plan to contribute
+changes to OpenJDK that modifies the build system.
+
+## Running Configure
+
+To build OpenJDK, you need a "configuration", which consists of a directory
+where to store the build output, coupled with information about the platform,
+the specific build machine, and choices that affect how OpenJDK is built.
+
+The configuration is created by the `configure` script. The basic invocation of
+the `configure` script looks like this:
+
+```
+bash configure [options]
+```
+
+This will create an output directory containing the configuration and setup an
+area for the build result. This directory typically looks like
+`build/linux-x64-normal-server-release`, but the actual name depends on your
+specific configuration. (It can also be set directly, see [Using Multiple
+Configurations](#using-multiple-configurations)). This directory is referred to
+as `$BUILD` in this documentation.
+
+`configure` will try to figure out what system you are running on and where all
+necessary build components are. If you have all prerequisites for building
+installed, it should find everything. If it fails to detect any component
+automatically, it will exit and inform you about the problem.
+
+Some command line examples:
+
+  * Create a 32-bit build for Windows with FreeType2 in `C:\freetype-i586`:
+    ```
+    bash configure --with-freetype=/cygdrive/c/freetype-i586 --with-target-bits=32
+    ```
+
+  * Create a debug build with the `server` JVM and DTrace enabled:
+    ```
+    bash configure --enable-debug --with-jvm-variants=server --enable-dtrace
+    ```
+
+### Common Configure Arguments
+
+Here follows some of the most common and important `configure` argument.
+
+To get up-to-date information on *all* available `configure` argument, please
+run:
+```
+bash configure --help
+```
+
+(Note that this help text also include general autoconf options, like
+`--dvidir`, that is not relevant to OpenJDK. To list only OpenJDK specific
+features, use `bash configure --help=short` instead.)
+
+#### Configure Arguments for Tailoring the Build
+
+  * `--enable-debug` - Set the debug level to `fastdebug` (this is a shorthand
+    for `--with-debug-level=fastdebug`)
+  * `--with-debug-level=<level>` - Set the debug level, which can be `release`,
+    `fastdebug`, `slowdebug` or `optimized`. Default is `release`. `optimized`
+    is variant of `release` with additional Hotspot debug code.
+  * `--with-native-debug-symbols=<method>` - Specify if and how native debug
+    symbols should be built. Available methods are `none`, `internal`,
+    `external`, `zipped`. Default behavior depends on platform. See [Native
+    Debug Symbols](#native-debug-symbols) for more details.
+  * `--with-version-string=<string>` - Specify the version string this build
+    will be identified with.
+  * `--with-version-<part>=<value>` - A group of options, where `<part>` can be
+    any of `pre`, `opt`, `build`, `major`, `minor`, `security` or `patch`. Use
+    these options to modify just the corresponding part of the version string
+    from the default, or the value provided by `--with-version-string`.
+  * `--with-jvm-variants=<variant>[,<variant>...]` - Build the specified variant
+    (or variants) of Hotspot. Valid variants are: `server`, `client`,
+    `minimal`, `core`, `zero`, `zeroshark`, `custom`. Note that not all
+    variants are possible to combine in a single build.
+  * `--with-jvm-features=<feature>[,<feature>...]` - Use the specified JVM
+    features when building Hotspot. The list of features will be enabled on top
+    of the default list. For the `custom` JVM variant, this default list is
+    empty. A complete list of available JVM features can be found using `bash
+    configure --help`.
+  * `--with-target-bits=<bits>` - Create a target binary suitable for running
+    on a `<bits>` platform. Use this to create 32-bit output on a 64-bit build
+    platform, instead of doing a full cross-compile. (This is known as a
+    *reduced* build.)
+
+#### Configure Arguments for Native Compilation
+
+  * `--with-devkit=<path>` - Use this devkit for compilers, tools and resources
+  * `--with-sysroot=<path>` - Use this directory as sysroot
+  * `--with-extra-path=<path>[;<path>]` - Prepend these directories to the
+    default path when searching for all kinds of binaries
+  * `--with-toolchain-path=<path>[;<path>]` - Prepend these directories when
+    searching for toolchain binaries (compilers etc)
+  * `--with-extra-cflags=<flags>` - Append these flags when compiling JDK C
+    files
+  * `--with-extra-cxxflags=<flags>` - Append these flags when compiling JDK C++
+    files
+  * `--with-extra-ldflags=<flags>` - Append these flags when linking JDK
+    libraries
+
+#### Configure Arguments for External Dependencies
+
+  * `--with-boot-jdk=<path>` - Set the path to the [Boot JDK](
+    #boot-jdk-requirements)
+  * `--with-freetype=<path>` - Set the path to [FreeType](#freetype)
+  * `--with-cups=<path>` - Set the path to [CUPS](#cups)
+  * `--with-x=<path>` - Set the path to [X11](#x11)
+  * `--with-alsa=<path>` - Set the path to [ALSA](#alsa)
+  * `--with-libffi=<path>` - Set the path to [libffi](#libffi)
+  * `--with-libelf=<path>` - Set the path to [libelf](#libelf)
+  * `--with-jtreg=<path>` - Set the path to JTReg. See [Running Tests](
+    #running-tests)
+
+Certain third-party libraries used by OpenJDK (libjpeg, giflib, libpng, lcms
+and zlib) are included in the OpenJDK repository. The default behavior of the
+OpenJDK build is to use this version of these libraries, but they might be
+replaced by an external version. To do so, specify `system` as the `<source>`
+option in these arguments. (The default is `bundled`).
+
+  * `--with-libjpeg=<source>` - Use the specified source for libjpeg
+  * `--with-giflib=<source>` - Use the specified source for giflib
+  * `--with-libpng=<source>` - Use the specified source for libpng
+  * `--with-lcms=<source>` - Use the specified source for lcms
+  * `--with-zlib=<source>` - Use the specified source for zlib
+
+On Linux, it is possible to select either static or dynamic linking of the C++
+runtime. The default is static linking, with dynamic linking as fallback if the
+static library is not found.
+
+  * `--with-stdc++lib=<method>` - Use the specified method (`static`, `dynamic`
+    or `default`) for linking the C++ runtime.
+
+### Configure Control Variables
+
+It is possible to control certain aspects of `configure` by overriding the
+value of `configure` variables, either on the command line or in the
+environment.
+
+Normally, this is **not recommended**. If used improperly, it can lead to a
+broken configuration. Unless you're well versed in the build system, this is
+hard to use properly. Therefore, `configure` will print a warning if this is
+detected.
+
+However, there are a few `configure` variables, known as *control variables*
+that are supposed to be overriden on the command line. These are variables that
+describe the location of tools needed by the build, like `MAKE` or `GREP`. If
+any such variable is specified, `configure` will use that value instead of
+trying to autodetect the tool. For instance, `bash configure
+MAKE=/opt/gnumake4.0/bin/make`.
+
+If a configure argument exists, use that instead, e.g. use `--with-jtreg`
+instead of setting `JTREGEXE`.
+
+Also note that, despite what autoconf claims, setting `CFLAGS` will not
+accomplish anything. Instead use `--with-extra-cflags` (and similar for
+`cxxflags` and `ldflags`).
+
+## Running Make
+
+When you have a proper configuration, all you need to do to build OpenJDK is to
+run `make`. (But see the warning at [GNU Make](#gnu-make) about running the
+correct version of make.)
+
+When running `make` without any arguments, the default target is used, which is
+the same as running `make default` or `make jdk`. This will build a minimal (or
+roughly minimal) set of compiled output (known as an "exploded image") needed
+for a developer to actually execute the newly built JDK. The idea is that in an
+incremental development fashion, when doing a normal make, you should only
+spend time recompiling what's changed (making it purely incremental) and only
+do the work that's needed to actually run and test your code.
+
+The output of the exploded image resides in `$BUILD/jdk`. You can test the
+newly built JDK like this: `$BUILD/jdk/bin/java -version`.
+
+### Common Make Targets
+
+Apart from the default target, here are some common make targets:
+
+  * `hotspot` - Build all of hotspot (but only hotspot)
+  * `hotspot-<variant>` - Build just the specified jvm variant
+  * `images` or `product-images` - Build the JRE and JDK images
+  * `docs` or `docs-image` - Build the documentation image
+  * `test-image` - Build the test image
+  * `all` or `all-images` - Build all images (product, docs and test)
+  * `bootcycle-images` - Build images twice, second time with newly built JDK
+    (good for testing)
+  * `clean` - Remove all files generated by make, but not those generated by
+    configure
+  * `dist-clean` - Remove all files, including configuration
+
+Run `make help` to get an up-to-date list of important make targets and make
+control variables.
+
+It is possible to build just a single module, a single phase, or a single phase
+of a single module, by creating make targets according to these followin
+patterns. A phase can be either of `gensrc`, `gendata`, `copy`, `java`,
+`launchers`, `libs` or `rmic`. See [Using Fine-Grained Make Targets](
+#using-fine-grained-make-targets) for more details about this functionality.
+
+  * `<phase>` - Build the specified phase and everything it depends on
+  * `<module>` - Build the specified module and everything it depends on
+  * `<module>-<phase>` - Compile the specified phase for the specified module
+    and everything it depends on
+
+Similarly, it is possible to clean just a part of the build by creating make
+targets according to these patterns:
+
+  * `clean-<outputdir>` - Remove the subdir in the output dir with the name
+  * `clean-<phase>` - Remove all build results related to a certain build
+    phase
+  * `clean-<module>` - Remove all build results related to a certain module
+  * `clean-<module>-<phase>` - Remove all build results related to a certain
+    module and phase
+
+### Make Control Variables
+
+It is possible to control `make` behavior by overriding the value of `make`
+variables, either on the command line or in the environment.
+
+Normally, this is **not recommended**. If used improperly, it can lead to a
+broken build. Unless you're well versed in the build system, this is hard to
+use properly. Therefore, `make` will print a warning if this is detected.
+
+However, there are a few `make` variables, known as *control variables* that
+are supposed to be overriden on the command line. These make up the "make time"
+configuration, as opposed to the "configure time" configuration.
+
+#### General Make Control Variables
+
+  * `JOBS` - Specify the number of jobs to build with. See [Build
+    Performance](#build-performance).
+  * `LOG` - Specify the logging level and functionality. See [Checking the
+    Build Log File](#checking-the-build-log-file)
+  * `CONF` and `CONF_NAME` - Selecting the configuration(s) to use. See [Using
+    Multiple Configurations](#using-multiple-configurations)
+
+#### Test Make Control Variables
+
+These make control variables only make sense when running tests. Please see
+[Testing OpenJDK](testing.html) for details.
+
+  * `TEST`
+  * `TEST_JOBS`
+  * `JTREG`
+  * `GTEST`
+
+#### Advanced Make Control Variables
+
+These advanced make control variables can be potentially unsafe. See [Hints and
+Suggestions for Advanced Users](#hints-and-suggestions-for-advanced-users) and
+[Understanding the Build System](#understanding-the-build-system) for details.
+
+  * `SPEC`
+  * `CONF_CHECK`
+  * `COMPARE_BUILD`
+  * `JDK_FILTER`
+
+## Running Tests
+
+Most of the OpenJDK tests are using the [JTReg](http://openjdk.java.net/jtreg)
+test framework. Make sure that your configuration knows where to find your
+installation of JTReg. If this is not picked up automatically, use the
+`--with-jtreg=<path to jtreg home>` option to point to the JTReg framework.
+Note that this option should point to the JTReg home, i.e. the top directory,
+containing `lib/jtreg.jar` etc.
+
+To execute the most basic tests (tier 1), use:
+```
+make run-test-tier1
+```
+
+For more details on how to run tests, please see the [Testing
+OpenJDK](testing.html) document.
+
+## Cross-compiling
+
+Cross-compiling means using one platform (the *build* platform) to generate
+output that can ran on another platform (the *target* platform).
+
+The typical reason for cross-compiling is that the build is performed on a more
+powerful desktop computer, but the resulting binaries will be able to run on a
+different, typically low-performing system. Most of the complications that
+arise when building for embedded is due to this separation of *build* and
+*target* systems.
+
+This requires a more complex setup and build procedure. This section assumes
+you are familiar with cross-compiling in general, and will only deal with the
+particularities of cross-compiling OpenJDK. If you are new to cross-compiling,
+please see the [external links at Wikipedia](
+https://en.wikipedia.org/wiki/Cross_compiler#External_links) for a good start
+on reading materials.
+
+Cross-compiling OpenJDK requires you to be able to build both for the build
+platform and for the target platform. The reason for the former is that we need
+to build and execute tools during the build process, both native tools and Java
+tools.
+
+If all you want to do is to compile a 32-bit version, for the same OS, on a
+64-bit machine, consider using `--with-target-bits=32` instead of doing a
+full-blown cross-compilation. (While this surely is possible, it's a lot more
+work and will take much longer to build.)
+
+### Boot JDK and Build JDK
+
+When cross-compiling, make sure you use a boot JDK that runs on the *build*
+system, and not on the *target* system.
+
+To be able to build, we need a "Build JDK", which is a JDK built from the
+current sources (that is, the same as the end result of the entire build
+process), but able to run on the *build* system, and not the *target* system.
+(In contrast, the Boot JDK should be from an older release, e.g. JDK 8 when
+building JDK 9.)
+
+The build process will create a minimal Build JDK for you, as part of building.
+To speed up the build, you can use `--with-build-jdk` to `configure` to point
+to a pre-built Build JDK. Please note that the build result is unpredictable,
+and can possibly break in subtle ways, if the Build JDK does not **exactly**
+match the current sources.
+
+### Specifying the Target Platform
+
+You *must* specify the target platform when cross-compiling. Doing so will also
+automatically turn the build into a cross-compiling mode. The simplest way to
+do this is to use the `--openjdk-target` argument, e.g.
+`--openjdk-target=arm-linux-gnueabihf`. or `--openjdk-target=aarch64-oe-linux`.
+This will automatically set the `--build`, `--host` and `--target` options for
+autoconf, which can otherwise be confusing. (In autoconf terminology, the
+"target" is known as "host", and "target" is used for building a Canadian
+cross-compiler.)
+
+### Toolchain Considerations
+
+You will need two copies of your toolchain, one which generates output that can
+run on the target system (the normal, or *target*, toolchain), and one that
+generates output that can run on the build system (the *build* toolchain). Note
+that cross-compiling is only supported for gcc at the time being. The gcc
+standard is to prefix cross-compiling toolchains with the target denominator.
+If you follow this standard, `configure` is likely to pick up the toolchain
+correctly.
+
+The *build* toolchain will be autodetected just the same way the normal
+*build*/*target* toolchain will be autodetected when not cross-compiling. If
+this is not what you want, or if the autodetection fails, you can specify a
+devkit containing the *build* toolchain using `--with-build-devkit` to
+`configure`, or by giving `BUILD_CC` and `BUILD_CXX` arguments.
+
+It is often helpful to locate the cross-compilation tools, headers and
+libraries in a separate directory, outside the normal path, and point out that
+directory to `configure`. Do this by setting the sysroot (`--with-sysroot`) and
+appending the directory when searching for cross-compilations tools
+(`--with-toolchain-path`). As a compact form, you can also use `--with-devkit`
+to point to a single directory, if it is correctly setup. (See `basics.m4` for
+details.)
+
+If you are unsure what toolchain and versions to use, these have been proved
+working at the time of writing:
+
+  * [aarch64](
+https://releases.linaro.org/archive/13.11/components/toolchain/binaries/gcc-linaro-aarch64-linux-gnu-4.8-2013.11_linux.tar.xz)
+  * [arm 32-bit hardware floating  point](
+https://launchpad.net/linaro-toolchain-unsupported/trunk/2012.09/+download/gcc-linaro-arm-linux-gnueabihf-raspbian-2012.09-20120921_linux.tar.bz2)
+
+### Native Libraries
+
+You will need copies of external native libraries for the *target* system,
+present on the *build* machine while building.
+
+Take care not to replace the *build* system's version of these libraries by
+mistake, since that can render the *build* machine unusable.
+
+Make sure that the libraries you point to (ALSA, X11, etc) are for the
+*target*, not the *build*, platform.
+
+#### ALSA
+
+You will need alsa libraries suitable for your *target* system. For most cases,
+using Debian's pre-built libraries work fine.
+
+Note that alsa is needed even if you only want to build a headless JDK.
+
+  * Go to [Debian Package Search](https://www.debian.org/distrib/packages) and
+    search for the `libasound2` and `libasound2-dev` packages for your *target*
+    system. Download them to /tmp.
+
+  * Install the libraries into the cross-compilation toolchain. For instance:
+```
+cd /tools/gcc-linaro-arm-linux-gnueabihf-raspbian-2012.09-20120921_linux/arm-linux-gnueabihf/libc
+dpkg-deb -x /tmp/libasound2_1.0.25-4_armhf.deb .
+dpkg-deb -x /tmp/libasound2-dev_1.0.25-4_armhf.deb .
+```
+
+  * If alsa is not properly detected by `configure`, you can point it out by
+    `--with-alsa`.
+
+#### X11
+
+You will need X11 libraries suitable for your *target* system. For most cases,
+using Debian's pre-built libraries work fine.
+
+Note that X11 is needed even if you only want to build a headless JDK.
+
+  * Go to [Debian Package Search](https://www.debian.org/distrib/packages),
+    search for the following packages for your *target* system, and download them
+    to /tmp/target-x11:
+      * libxi
+      * libxi-dev
+      * x11proto-core-dev
+      * x11proto-input-dev
+      * x11proto-kb-dev
+      * x11proto-render-dev
+      * x11proto-xext-dev
+      * libice-dev
+      * libxrender
+      * libxrender-dev
+      * libsm-dev
+      * libxt-dev
+      * libx11
+      * libx11-dev
+      * libxtst
+      * libxtst-dev
+      * libxext
+      * libxext-dev
+
+  * Install the libraries into the cross-compilation toolchain. For instance:
+    ```
+    cd /tools/gcc-linaro-arm-linux-gnueabihf-raspbian-2012.09-20120921_linux/arm-linux-gnueabihf/libc/usr
+    mkdir X11R6
+    cd X11R6
+    for deb in /tmp/target-x11/*.deb ; do dpkg-deb -x $deb . ; done
+    mv usr/* .
+    cd lib
+    cp arm-linux-gnueabihf/* .
+    ```
+
+    You can ignore the following messages. These libraries are not needed to
+    successfully complete a full JDK build.
+    ```
+    cp: cannot stat `arm-linux-gnueabihf/libICE.so': No such file or directory
+    cp: cannot stat `arm-linux-gnueabihf/libSM.so': No such file or directory
+    cp: cannot stat `arm-linux-gnueabihf/libXt.so': No such file or directory
+    ```
+
+  * If the X11 libraries are not properly detected by `configure`, you can
+    point them out by `--with-x`.
+
+### Building for ARM/aarch64
+
+A common cross-compilation target is the ARM CPU. When building for ARM, it is
+useful to set the ABI profile. A number of pre-defined ABI profiles are
+available using `--with-abi-profile`: arm-vfp-sflt, arm-vfp-hflt, arm-sflt,
+armv5-vfp-sflt, armv6-vfp-hflt. Note that soft-float ABIs are no longer
+properly supported on OpenJDK.
+
+OpenJDK contains two different ports for the aarch64 platform, one is the
+original aarch64 port from the [AArch64 Port Project](
+http://openjdk.java.net/projects/aarch64-port) and one is a 64-bit version of
+the Oracle contributed ARM port. When targeting aarch64, by the default the
+original aarch64 port is used. To select the Oracle ARM 64 port, use
+`--with-cpu-port=arm64`. Also set the corresponding value (`aarch64` or
+`arm64`) to --with-abi-profile, to ensure a consistent build.
+
+### Verifying the Build
+
+The build will end up in a directory named like
+`build/linux-arm-normal-server-release`.
+
+Inside this build output directory, the `images/jdk` and `images/jre` will
+contain the newly built JDK and JRE, respectively, for your *target* system.
+
+Copy these folders to your *target* system. Then you can run e.g.
+`images/jdk/bin/java -version`.
+
+## Build Performance
+
+Building OpenJDK requires a lot of horsepower. Some of the build tools can be
+adjusted to utilize more or less of resources such as parallel threads and
+memory. The `configure` script analyzes your system and selects reasonable
+values for such options based on your hardware. If you encounter resource
+problems, such as out of memory conditions, you can modify the detected values
+with:
+
+  * `--with-num-cores` -- number of cores in the build system, e.g.
+    `--with-num-cores=8`.
+
+  * `--with-memory-size` -- memory (in MB) available in the build system, e.g.
+    `--with-memory-size=1024`
+
+You can also specify directly the number of build jobs to use with
+`--with-jobs=N` to `configure`, or `JOBS=N` to `make`. Do not use the `-j` flag
+to `make`. In most cases it will be ignored by the makefiles, but it can cause
+problems for some make targets.
+
+It might also be necessary to specify the JVM arguments passed to the Boot JDK,
+using e.g. `--with-boot-jdk-jvmargs="-Xmx8G"`. Doing so will override the
+default JVM arguments passed to the Boot JDK.
+
+At the end of a successful execution of `configure`, you will get a performance
+summary, indicating how well the build will perform. Here you will also get
+performance hints. If you want to build fast, pay attention to those!
+
+If you want to tweak build performance, run with `make LOG=info` to get a build
+time summary at the end of the build process.
+
+### Disk Speed
+
+If you are using network shares, e.g. via NFS, for your source code, make sure
+the build directory is situated on local disk (e.g. by `ln -s
+/localdisk/jdk-build $JDK-SHARE/build`). The performance penalty is extremely
+high for building on a network share; close to unusable.
+
+Also, make sure that your build tools (including Boot JDK and toolchain) is
+located on a local disk and not a network share.
+
+As has been stressed elsewhere, do use SSD for source code and build directory,
+as well as (if possible) the build tools.
+
+### Virus Checking
+
+The use of virus checking software, especially on Windows, can *significantly*
+slow down building of OpenJDK. If possible, turn off such software, or exclude
+the directory containing the OpenJDK source code from on-the-fly checking.
+
+### Ccache
+
+The OpenJDK build supports building with ccache when using gcc or clang. Using
+ccache can radically speed up compilation of native code if you often rebuild
+the same sources. Your milage may vary however, so we recommend evaluating it
+for yourself. To enable it, make sure it's on the path and configure with
+`--enable-ccache`.
+
+### Precompiled Headers
+
+By default, the Hotspot build uses preccompiled headers (PCH) on the toolchains
+were it is properly supported (clang, gcc, and Visual Studio). Normally, this
+speeds up the build process, but in some circumstances, it can actually slow
+things down.
+
+You can experiment by disabling precompiled headers using
+`--disable-precompiled-headers`.
+
+### Icecc / icecream
+
+[icecc/icecream](http://github.com/icecc/icecream) is a simple way to setup a
+distributed compiler network. If you have multiple machines available for
+building OpenJDK, you can drastically cut individual build times by utilizing
+it.
+
+To use, setup an icecc network, and install icecc on the build machine. Then
+run `configure` using `--enable-icecc`.
+
+### Using sjavac
+
+To speed up Java compilation, especially incremental compilations, you can try
+the experimental sjavac compiler by using `--enable-sjavac`.
+
+### Building the Right Target
+
+Selecting the proper target to build can have dramatic impact on build time.
+For normal usage, `jdk` or the default target is just fine. You only need to
+build `images` for shipping, or if your tests require it.
+
+See also [Using Fine-Grained Make Targets](#using-fine-grained-make-targets) on
+how to build an even smaller subset of the product.
+
+## Troubleshooting
+
+If your build fails, it can sometimes be difficult to pinpoint the problem or
+find a proper solution.
+
+### Locating the Source of the Error
+
+When a build fails, it can be hard to pinpoint the actual cause of the error.
+In a typical build process, different parts of the product build in parallel,
+with the output interlaced.
+
+#### Build Failure Summary
+
+To help you, the build system will print a failure summary at the end. It looks
+like this:
+
+```
+ERROR: Build failed for target 'hotspot' in configuration 'linux-x64' (exit code 2)
+
+=== Output from failing command(s) repeated here ===
+* For target hotspot_variant-server_libjvm_objs_psMemoryPool.o:
+/localhome/hg/jdk9-sandbox/hotspot/src/share/vm/services/psMemoryPool.cpp:1:1: error: 'failhere' does not name a type
+   ... (rest of output omitted)
+
+* All command lines available in /localhome/hg/jdk9-sandbox/build/linux-x64/make-support/failure-logs.
+=== End of repeated output ===
+
+=== Make failed targets repeated here ===
+lib/CompileJvm.gmk:207: recipe for target '/localhome/hg/jdk9-sandbox/build/linux-x64/hotspot/variant-server/libjvm/objs/psMemoryPool.o' failed
+make/Main.gmk:263: recipe for target 'hotspot-server-libs' failed
+=== End of repeated output ===
+
+Hint: Try searching the build log for the name of the first failed target.
+Hint: If caused by a warning, try configure --disable-warnings-as-errors.
+```
+
+Let's break it down! First, the selected configuration, and the top-level
+target you entered on the command line that caused the failure is printed.
+
+Then, between the `Output from failing command(s) repeated here` and `End of
+repeated output` the first lines of output (stdout and stderr) from the actual
+failing command is repeated. In most cases, this is the error message that
+caused the build to fail. If multiple commands were failing (this can happen in
+a parallel build), output from all failed commands will be printed here.
+
+The path to the `failure-logs` directory is printed. In this file you will find
+a `<target>.log` file that contains the output from this command in its
+entirety, and also a `<target>.cmd`, which contain the complete command line
+used for running this command. You can re-run the failing command by executing
+`. <path to failure-logs>/<target>.cmd` in your shell.
+
+Another way to trace the failure is to follow the chain of make targets, from
+top-level targets to individual file targets. Between `Make failed targets
+repeated here` and `End of repeated output` the output from make showing this
+chain is repeated. The first failed recipe will typically contain the full path
+to the file in question that failed to compile. Following lines will show a
+trace of make targets why we ended up trying to compile that file.
+
+Finally, some hints are given on how to locate the error in the complete log.
+In this example, we would try searching the log file for "`psMemoryPool.o`".
+Another way to quickly locate make errors in the log is to search for "`]
+Error`" or "`***`".
+
+Note that the build failure summary will only help you if the issue was a
+compilation failure or similar. If the problem is more esoteric, or is due to
+errors in the build machinery, you will likely get empty output logs, and `No
+indication of failed target found` instead of the make target chain.
+
+#### Checking the Build Log File
+
+The output (stdout and stderr) from the latest build is always stored in
+`$BUILD/build.log`. The previous build log is stored as `build.log.old`. This
+means that it is not necessary to redirect the build output yourself if you
+want to process it.
+
+You can increase the verbosity of the log file, by the `LOG` control variable
+to `make`. If you want to see the command lines used in compilations, use
+`LOG=cmdlines`. To increase the general verbosity, use `LOG=info`, `LOG=debug`
+or `LOG=trace`. Both of these can be combined with `cmdlines`, e.g.
+`LOG=info,cmdlines`. The `debug` log level will show most shell commands
+executed by make, and `trace` will show all. Beware that both these log levels
+will produce a massive build log!
+
+### Fixing Unexpected Build Failures
+
+Most of the time, the build will fail due to incorrect changes in the source
+code.
+
+Sometimes the build can fail with no apparent changes that have caused the
+failure. If this is the first time you are building OpenJDK on this particular
+computer, and the build fails, the problem is likely with your build
+environment. But even if you have previously built OpenJDK with success, and it
+now fails, your build environment might have changed (perhaps due to OS
+upgrades or similar). But most likely, such failures are due to problems with
+the incremental rebuild.
+
+#### Problems with the Build Environment
+
+Make sure your configuration is correct. Re-run `configure`, and look for any
+warnings. Warnings that appear in the middle of the `configure` output is also
+repeated at the end, after the summary. The entire log is stored in
+`$BUILD/configure.log`.
+
+Verify that the summary at the end looks correct. Are you indeed using the Boot
+JDK and native toolchain that you expect?
+
+By default, OpenJDK has a strict approach where warnings from the compiler is
+considered errors which fail the build. For very new or very old compiler
+versions, this can trigger new classes of warnings, which thus fails the build.
+Run `configure` with `--disable-warnings-as-errors` to turn of this behavior.
+(The warnings will still show, but not make the build fail.)
+
+#### Problems with Incremental Rebuilds
+
+Incremental rebuilds mean that when you modify part of the product, only the
+affected parts get rebuilt. While this works great in most cases, and
+significantly speed up the development process, from time to time complex
+interdependencies will result in an incorrect build result. This is the most
+common cause for unexpected build problems, together with inconsistencies
+between the different Mercurial repositories in the forest.
+
+Here are a suggested list of things to try if you are having unexpected build
+problems. Each step requires more time than the one before, so try them in
+order. Most issues will be solved at step 1 or 2.
+
+ 1. Make sure your forest is up-to-date
+
+    Run `bash get_source.sh` to make sure you have the latest version of all
+    repositories.
+
+ 2. Clean build results
+
+    The simplest way to fix incremental rebuild issues is to run `make clean`.
+    This will remove all build results, but not the configuration or any build
+    system support artifacts. In most cases, this will solve build errors
+    resulting from incremental build mismatches.
+
+ 3. Completely clean the build directory.
+
+    If this does not work, the next step is to run `make dist-clean`, or
+    removing the build output directory (`$BUILD`). This will clean all
+    generated output, including your configuration. You will need to re-run
+    `configure` after this step. A good idea is to run `make
+    print-configuration` before running `make dist-clean`, as this will print
+    your current `configure` command line. Here's a way to do this:
+
+    ```
+    make print-configuration > current-configuration
+    make dist-clean
+    bash configure $(cat current-configuration)
+    make
+    ```
+
+ 4. Re-clone the Mercurial forest
+
+    Sometimes the Mercurial repositories themselves gets in a state that causes
+    the product to be un-buildable. In such a case, the simplest solution is
+    often the "sledgehammer approach": delete the entire forest, and re-clone
+    it. If you have local changes, save them first to a different location
+    using `hg export`.
+
+### Specific Build Issues
+
+#### Clock Skew
+
+If you get an error message like this:
+```
+File 'xxx' has modification time in the future.
+Clock skew detected. Your build may be incomplete.
+```
+then the clock on your build machine is out of sync with the timestamps on the
+source files. Other errors, apparently unrelated but in fact caused by the
+clock skew, can occur along with the clock skew warnings. These secondary
+errors may tend to obscure the fact that the true root cause of the problem is
+an out-of-sync clock.
+
+If you see these warnings, reset the clock on the build machine, run `make
+clean` and restart the build.
+
+#### Out of Memory Errors
+
+On Solaris, you might get an error message like this:
+```
+Trouble writing out table to disk
+```
+To solve this, increase the amount of swap space on your build machine.
+
+On Windows, you might get error messages like this:
+```
+fatal error - couldn't allocate heap
+cannot create ... Permission denied
+spawn failed
+```
+This can be a sign of a Cygwin problem. See the information about solving
+problems in the [Cygwin](#cygwin) section. Rebooting the computer might help
+temporarily.
+
+### Getting Help
+
+If none of the suggestions in this document helps you, or if you find what you
+believe is a bug in the build system, please contact the Build Group by sending
+a mail to [build-dev@openjdk.java.net](mailto:build-dev@openjdk.java.net).
+Please include the relevant parts of the configure and/or build log.
+
+If you need general help or advice about developing for OpenJDK, you can also
+contact the Adoption Group. See the section on [Contributing to OpenJDK](
+#contributing-to-openjdk) for more information.
+
+## Hints and Suggestions for Advanced Users
+
+### Setting Up a Forest for Pushing Changes (defpath)
+
+To help you prepare a proper push path for a Mercurial repository, there exists
+a useful tool known as [defpath](
+http://openjdk.java.net/projects/code-tools/defpath). It will help you setup a
+proper push path for pushing changes to OpenJDK.
+
+Install the extension by cloning
+`http://hg.openjdk.java.net/code-tools/defpath` and updating your `.hgrc` file.
+Here's one way to do this:
+
+```
+cd ~
+mkdir hg-ext
+cd hg-ext
+hg clone http://hg.openjdk.java.net/code-tools/defpath
+cat << EOT >> ~/.hgrc
+[extensions]
+defpath=~/hg-ext/defpath/defpath.py
+EOT
+```
+
+You can now setup a proper push path using:
+```
+hg defpath -d -u <your OpenJDK username>
+```
+
+If you also have the `trees` extension installed in Mercurial, you will
+automatically get a `tdefpath` command, which is even more useful. By running
+`hg tdefpath -du <username>` in the top repository of your forest, all repos
+will get setup automatically. This is the recommended usage.
+
+### Bash Completion
+
+The `configure` and `make` commands tries to play nice with bash command-line
+completion (using `<tab>` or `<tab><tab>`). To use this functionality, make
+sure you enable completion in your `~/.bashrc` (see instructions for bash in
+your operating system).
+
+Make completion will work out of the box, and will complete valid make targets.
+For instance, typing `make jdk-i<tab>` will complete to `make jdk-image`.
+
+The `configure` script can get completion for options, but for this to work you
+need to help `bash` on the way. The standard way of running the script, `bash
+configure`, will not be understood by bash completion. You need `configure` to
+be the command to run. One way to achieve this is to add a simple helper script
+to your path:
+
+```
+cat << EOT > /tmp/configure
+#!/bin/bash
+if [ \$(pwd) = \$(cd \$(dirname \$0); pwd) ] ; then
+  echo >&2 "Abort: Trying to call configure helper recursively"
+  exit 1
+fi
+
+bash \$PWD/configure "\$@"
+EOT
+chmod +x /tmp/configure
+sudo mv /tmp/configure /usr/local/bin
+```
+
+Now `configure --en<tab>-dt<tab>` will result in `configure --enable-dtrace`.
+
+### Using Multiple Configurations
+
+You can have multiple configurations for a single source forest. When you
+create a new configuration, run `configure --with-conf-name=<name>` to create a
+configuration with the name `<name>`. Alternatively, you can create a directory
+under `build` and run `configure` from there, e.g. `mkdir build/<name> && cd
+build/<name> && bash ../../configure`.
+
+Then you can build that configuration using `make CONF_NAME=<name>` or `make
+CONF=<pattern>`, where `<pattern>` is a substring matching one or several
+configurations, e.g. `CONF=debug`. The special empty pattern (`CONF=`) will
+match *all* available configuration, so `make CONF= hotspot` will build the
+`hotspot` target for all configurations. Alternatively, you can execute `make`
+in the configuration directory, e.g. `cd build/<name> && make`.
+
+### Handling Reconfigurations
+
+If you update the forest and part of the configure script has changed, the
+build system will force you to re-run `configure`.
+
+Most of the time, you will be fine by running `configure` again with the same
+arguments as the last time, which can easily be performed by `make
+reconfigure`. To simplify this, you can use the `CONF_CHECK` make control
+variable, either as `make CONF_CHECK=auto`, or by setting an environment
+variable. For instance, if you add `export CONF_CHECK=auto` to your `.bashrc`
+file, `make` will always run `reconfigure` automatically whenever the configure
+script has changed.
+
+You can also use `CONF_CHECK=ignore` to skip the check for a needed configure
+update. This might speed up the build, but comes at the risk of an incorrect
+build result. This is only recommended if you know what you're doing.
+
+From time to time, you will also need to modify the command line to `configure`
+due to changes. Use `make print-configure` to show the command line used for
+your current configuration.
+
+### Using Fine-Grained Make Targets
+
+The default behavior for make is to create consistent and correct output, at
+the expense of build speed, if necessary.
+
+If you are prepared to take some risk of an incorrect build, and know enough of
+the system to understand how things build and interact, you can speed up the
+build process considerably by instructing make to only build a portion of the
+product.
+
+#### Building Individual Modules
+
+The safe way to use fine-grained make targets is to use the module specific
+make targets. All source code in JDK 9 is organized so it belongs to a module,
+e.g. `java.base` or `jdk.jdwp.agent`. You can build only a specific module, by
+giving it as make target: `make jdk.jdwp.agent`. If the specified module
+depends on other modules (e.g. `java.base`), those modules will be built first.
+
+You can also specify a set of modules, just as you can always specify a set of
+make targets: `make jdk.crypto.cryptoki jdk.crypto.ec jdk.crypto.mscapi
+jdk.crypto.ucrypto`
+
+#### Building Individual Module Phases
+
+The build process for each module is divided into separate phases. Not all
+modules need all phases. Which are needed depends on what kind of source code
+and other artifact the module consists of. The phases are:
+
+  * `gensrc` (Generate source code to compile)
+  * `gendata` (Generate non-source code artifacts)
+  * `copy` (Copy resource artifacts)
+  * `java` (Compile Java code)
+  * `launchers` (Compile native executables)
+  * `libs` (Compile native libraries)
+  * `rmic` (Run the `rmic` tool)
+
+You can build only a single phase for a module by using the notation
+`$MODULE-$PHASE`. For instance, to build the `gensrc` phase for `java.base`,
+use `make java.base-gensrc`.
+
+Note that some phases may depend on others, e.g. `java` depends on `gensrc` (if
+present). Make will build all needed prerequisites before building the
+requested phase.
+
+#### Skipping the Dependency Check
+
+When using an iterative development style with frequent quick rebuilds, the
+dependency check made by make can take up a significant portion of the time
+spent on the rebuild. In such cases, it can be useful to bypass the dependency
+check in make.
+
+> **Note that if used incorrectly, this can lead to a broken build!**
+
+To achieve this, append `-only` to the build target. For instance, `make
+jdk.jdwp.agent-java-only` will *only* build the `java` phase of the
+`jdk.jdwp.agent` module. If the required dependencies are not present, the
+build can fail. On the other hand, the execution time measures in milliseconds.
+
+A useful pattern is to build the first time normally (e.g. `make
+jdk.jdwp.agent`) and then on subsequent builds, use the `-only` make target.
+
+#### Rebuilding Part of java.base (JDK\_FILTER)
+
+If you are modifying files in `java.base`, which is the by far largest module
+in OpenJDK, then you need to rebuild all those files whenever a single file has
+changed. (This inefficiency will hopefully be addressed in JDK 10.)
+
+As a hack, you can use the make control variable `JDK_FILTER` to specify a
+pattern that will be used to limit the set of files being recompiled. For
+instance, `make java.base JDK_FILTER=javax/crypto` (or, to combine methods,
+`make java.base-java-only JDK_FILTER=javax/cr