#!/bin/sh # Locate python2 interpreter and re-execute the script. Note that the # mix of single and double quotes is intentional, as is the fact that # the ] goes on a new line. _=[ 'exec' '/bin/sh' '-c' ''' which python2.7 >/dev/null && exec python2.7 "$0" "$@" which python2 >/dev/null && exec python2 "$0" "$@" exec python "$0" "$@" ''' "$0" "$@" ] del _ import sys from distutils.spawn import find_executable as which if sys.version_info[0] != 2 or sys.version_info[1] not in (6, 7): sys.stderr.write('Please use either Python 2.6 or 2.7') python2 = which('python2') or which('python2.6') or which('python2.7') if python2: sys.stderr.write(':\n\n') sys.stderr.write(' ' + python2 + ' ' + ' '.join(sys.argv)) sys.stderr.write('\n') sys.exit(1) import errno import optparse import os import pprint import re import shlex import subprocess import shutil import string # If not run from node/, cd to node/. os.chdir(os.path.dirname(__file__) or '.') # gcc and g++ as defaults matches what GYP's Makefile generator does, # except on OS X. CC = os.environ.get('CC', 'cc' if sys.platform == 'darwin' else 'gcc') CXX = os.environ.get('CXX', 'c++' if sys.platform == 'darwin' else 'g++') sys.path.insert(0, os.path.join('tools', 'gyp', 'pylib')) from gyp.common import GetFlavor # imports in tools/configure.d sys.path.insert(0, os.path.join('tools', 'configure.d')) import nodedownload # imports in tools/ sys.path.insert(0, 'tools') import getmoduleversion from gyp_node import run_gyp # parse our options parser = optparse.OptionParser() valid_os = ('win', 'mac', 'solaris', 'freebsd', 'openbsd', 'linux', 'android', 'aix', 'cloudabi') valid_arch = ('arm', 'arm64', 'ia32', 'mips', 'mipsel', 'mips64el', 'ppc', 'ppc64', 'x32','x64', 'x86', 's390', 's390x') valid_arm_float_abi = ('soft', 'softfp', 'hard') valid_arm_fpu = ('vfp', 'vfpv3', 'vfpv3-d16', 'neon') valid_mips_arch = ('loongson', 'r1', 'r2', 'r6', 'rx') valid_mips_fpu = ('fp32', 'fp64', 'fpxx') valid_mips_float_abi = ('soft', 'hard') valid_intl_modes = ('none', 'small-icu', 'full-icu', 'system-icu') # create option groups shared_optgroup = optparse.OptionGroup(parser, "Shared libraries", "Flags that allows you to control whether you want to build against " "built-in dependencies or its shared representations. If necessary, " "provide multiple libraries with comma.") intl_optgroup = optparse.OptionGroup(parser, "Internationalization", "Flags that lets you enable i18n features in Node.js as well as which " "library you want to build against.") http2_optgroup = optparse.OptionGroup(parser, "HTTP2", "Flags that allows you to control HTTP2 features in Node.js") # Options should be in alphabetical order but keep --prefix at the top, # that's arguably the one people will be looking for most. parser.add_option('--prefix', action='store', dest='prefix', default='/usr/local', help='select the install prefix [default: %default]') parser.add_option('--coverage', action='store_true', dest='coverage', help='Build node with code coverage enabled') parser.add_option('--debug', action='store_true', dest='debug', help='also build debug build') parser.add_option('--dest-cpu', action='store', dest='dest_cpu', choices=valid_arch, help='CPU architecture to build for ({0})'.format(', '.join(valid_arch))) parser.add_option('--cross-compiling', action='store_true', dest='cross_compiling', default=None, help='force build to be considered as cross compiled') parser.add_option('--no-cross-compiling', action='store_false', dest='cross_compiling', default=None, help='force build to be considered as NOT cross compiled') parser.add_option('--dest-os', action='store', dest='dest_os', choices=valid_os, help='operating system to build for ({0})'.format(', '.join(valid_os))) parser.add_option('--gdb', action='store_true', dest='gdb', help='add gdb support') parser.add_option('--no-ifaddrs', action='store_true', dest='no_ifaddrs', help='use on deprecated SunOS systems that do not support ifaddrs.h') parser.add_option("--fully-static", action="store_true", dest="fully_static", help="Generate an executable without external dynamic libraries. This " "will not work on OSX when using the default compilation environment") parser.add_option("--partly-static", action="store_true", dest="partly_static", help="Generate an executable with libgcc and libstdc++ libraries. This " "will not work on OSX when using the default compilation environment") parser.add_option("--enable-vtune-profiling", action="store_true", dest="enable_vtune_profiling", help="Enable profiling support for Intel VTune profiler to profile " "JavaScript code executed in nodejs. This feature is only available " "for x32, x86, and x64 architectures.") parser.add_option("--link-module", action="append", dest="linked_module", help="Path to a JS file to be bundled in the binary as a builtin. " "This module will be referenced by path without extension; " "e.g. /root/x/y.js will be referenced via require('root/x/y'). " "Can be used multiple times") parser.add_option("--openssl-no-asm", action="store_true", dest="openssl_no_asm", help="Do not build optimized assembly for OpenSSL") parser.add_option('--openssl-fips', action='store', dest='openssl_fips', help='Build OpenSSL using FIPS canister .o file in supplied folder') parser.add_option('--openssl-use-def-ca-store', action='store_true', dest='use_openssl_ca_store', help='Use OpenSSL supplied CA store instead of compiled-in Mozilla CA copy.') parser.add_option('--openssl-system-ca-path', action='store', dest='openssl_system_ca_path', help='Use the specified path to system CA (PEM format) in addition to ' 'the OpenSSL supplied CA store or compiled-in Mozilla CA copy.') shared_optgroup.add_option('--shared-http-parser', action='store_true', dest='shared_http_parser', help='link to a shared http_parser DLL instead of static linking') shared_optgroup.add_option('--shared-http-parser-includes', action='store', dest='shared_http_parser_includes', help='directory containing http_parser header files') shared_optgroup.add_option('--shared-http-parser-libname', action='store', dest='shared_http_parser_libname', default='http_parser', help='alternative lib name to link to [default: %default]') shared_optgroup.add_option('--shared-http-parser-libpath', action='store', dest='shared_http_parser_libpath', help='a directory to search for the shared http_parser DLL') shared_optgroup.add_option('--shared-libuv', action='store_true', dest='shared_libuv', help='link to a shared libuv DLL instead of static linking') shared_optgroup.add_option('--shared-libuv-includes', action='store', dest='shared_libuv_includes', help='directory containing libuv header files') shared_optgroup.add_option('--shared-libuv-libname', action='store', dest='shared_libuv_libname', default='uv', help='alternative lib name to link to [default: %default]') shared_optgroup.add_option('--shared-libuv-libpath', action='store', dest='shared_libuv_libpath', help='a directory to search for the shared libuv DLL') shared_optgroup.add_option('--shared-nghttp2', action='store_true', dest='shared_nghttp2', help='link to a shared nghttp2 DLL instead of static linking') shared_optgroup.add_option('--shared-nghttp2-includes', action='store', dest='shared_nghttp2_includes', help='directory containing nghttp2 header files') shared_optgroup.add_option('--shared-nghttp2-libname', action='store', dest='shared_nghttp2_libname', default='nghttp2', help='alternative lib name to link to [default: %default]') shared_optgroup.add_option('--shared-nghttp2-libpath', action='store', dest='shared_nghttp2_libpath', help='a directory to search for the shared nghttp2 DLLs') shared_optgroup.add_option('--shared-openssl', action='store_true', dest='shared_openssl', help='link to a shared OpenSSl DLL instead of static linking') shared_optgroup.add_option('--shared-openssl-includes', action='store', dest='shared_openssl_includes', help='directory containing OpenSSL header files') shared_optgroup.add_option('--shared-openssl-libname', action='store', dest='shared_openssl_libname', default='crypto,ssl', help='alternative lib name to link to [default: %default]') shared_optgroup.add_option('--shared-openssl-libpath', action='store', dest='shared_openssl_libpath', help='a directory to search for the shared OpenSSL DLLs') shared_optgroup.add_option('--shared-zlib', action='store_true', dest='shared_zlib', help='link to a shared zlib DLL instead of static linking') shared_optgroup.add_option('--shared-zlib-includes', action='store', dest='shared_zlib_includes', help='directory containing zlib header files') shared_optgroup.add_option('--shared-zlib-libname', action='store', dest='shared_zlib_libname', default='z', help='alternative lib name to link to [default: %default]') shared_optgroup.add_option('--shared-zlib-libpath', action='store', dest='shared_zlib_libpath', help='a directory to search for the shared zlib DLL') shared_optgroup.add_option('--shared-cares', action='store_true', dest='shared_libcares', help='link to a shared cares DLL instead of static linking') shared_optgroup.add_option('--shared-cares-includes', action='store', dest='shared_libcares_includes', help='directory containing cares header files') shared_optgroup.add_option('--shared-cares-libname', action='store', dest='shared_libcares_libname', default='cares', help='alternative lib name to link to [default: %default]') shared_optgroup.add_option('--shared-cares-libpath', action='store', dest='shared_libcares_libpath', help='a directory to search for the shared cares DLL') parser.add_option_group(shared_optgroup) parser.add_option('--systemtap-includes', action='store', dest='systemtap_includes', help='directory containing systemtap header files') parser.add_option('--tag', action='store', dest='tag', help='custom build tag') parser.add_option('--release-urlbase', action='store', dest='release_urlbase', help='Provide a custom URL prefix for the `process.release` properties ' '`sourceUrl` and `headersUrl`. When compiling a release build, this ' 'will default to https://nodejs.org/download/release/') parser.add_option('--enable-d8', action='store_true', dest='enable_d8', help=optparse.SUPPRESS_HELP) # Unsupported, undocumented. parser.add_option('--enable-trace-maps', action='store_true', dest='trace_maps', help='Enable the --trace-maps flag in V8 (use at your own risk)') parser.add_option('--v8-options', action='store', dest='v8_options', help='v8 options to pass, see `node --v8-options` for examples.') parser.add_option('--with-arm-float-abi', action='store', dest='arm_float_abi', choices=valid_arm_float_abi, help='specifies which floating-point ABI to use ({0}).'.format( ', '.join(valid_arm_float_abi))) parser.add_option('--with-arm-fpu', action='store', dest='arm_fpu', choices=valid_arm_fpu, help='ARM FPU mode ({0}) [default: %default]'.format( ', '.join(valid_arm_fpu))) parser.add_option('--with-mips-arch-variant', action='store', dest='mips_arch_variant', default='r2', choices=valid_mips_arch, help='MIPS arch variant ({0}) [default: %default]'.format( ', '.join(valid_mips_arch))) parser.add_option('--with-mips-fpu-mode', action='store', dest='mips_fpu_mode', default='fp32', choices=valid_mips_fpu, help='MIPS FPU mode ({0}) [default: %default]'.format( ', '.join(valid_mips_fpu))) parser.add_option('--with-mips-float-abi', action='store', dest='mips_float_abi', default='hard', choices=valid_mips_float_abi, help='MIPS floating-point ABI ({0}) [default: %default]'.format( ', '.join(valid_mips_float_abi))) parser.add_option('--with-dtrace', action='store_true', dest='with_dtrace', help='build with DTrace (default is true on sunos and darwin)') parser.add_option('--with-lttng', action='store_true', dest='with_lttng', help='build with Lttng (Only supported on Linux)') parser.add_option('--with-etw', action='store_true', dest='with_etw', help='build with ETW (default is true on Windows)') intl_optgroup.add_option('--with-intl', action='store', dest='with_intl', default='small-icu', choices=valid_intl_modes, help='Intl mode (valid choices: {0}) [default: %default]'.format( ', '.join(valid_intl_modes))) intl_optgroup.add_option('--without-intl', action='store_const', dest='with_intl', const='none', help='Disable Intl, same as --with-intl=none (disables inspector)') intl_optgroup.add_option('--with-icu-path', action='store', dest='with_icu_path', help='Path to icu.gyp (ICU i18n, Chromium version only.)') icu_default_locales='root,en' intl_optgroup.add_option('--with-icu-locales', action='store', dest='with_icu_locales', default=icu_default_locales, help='Comma-separated list of locales for "small-icu". "root" is assumed. ' '[default: %default]') intl_optgroup.add_option('--with-icu-source', action='store', dest='with_icu_source', help='Intl mode: optional local path to icu/ dir, or path/URL of icu source archive.') intl_optgroup.add_option('--download', action='store', dest='download_list', help=nodedownload.help()) intl_optgroup.add_option('--download-path', action='store', dest='download_path', default='deps', help='Download directory [default: %default]') parser.add_option_group(intl_optgroup) http2_optgroup.add_option('--debug-http2', action='store_true', dest='debug_http2', help='build with http2 debug statements on (default is false)') http2_optgroup.add_option('--debug-nghttp2', action='store_true', dest='debug_nghttp2', help='build nghttp2 with DEBUGBUILD (default is false)') parser.add_option_group(http2_optgroup) parser.add_option('--with-perfctr', action='store_true', dest='with_perfctr', help='build with performance counters (default is true on Windows)') parser.add_option('--without-dtrace', action='store_true', dest='without_dtrace', help='build without DTrace') parser.add_option('--without-etw', action='store_true', dest='without_etw', help='build without ETW') parser.add_option('--without-npm', action='store_true', dest='without_npm', help='do not install the bundled npm (package manager)') parser.add_option('--without-perfctr', action='store_true', dest='without_perfctr', help='build without performance counters') # Dummy option for backwards compatibility parser.add_option('--with-snapshot', action='store_true', dest='unused_with_snapshot', help=optparse.SUPPRESS_HELP) parser.add_option('--without-snapshot', action='store_true', dest='without_snapshot', help=optparse.SUPPRESS_HELP) parser.add_option('--without-ssl', action='store_true', dest='without_ssl', help='build without SSL (disables crypto, https, inspector, etc.)') parser.add_option('--without-node-options', action='store_true', dest='without_node_options', help='build without NODE_OPTIONS support') parser.add_option('--xcode', action='store_true', dest='use_xcode', help='generate build files for use with xcode') parser.add_option('--ninja', action='store_true', dest='use_ninja', help='generate build files for use with Ninja') parser.add_option('--enable-asan', action='store_true', dest='enable_asan', help='build with asan') parser.add_option('--enable-static', action='store_true', dest='enable_static', help='build as static library') parser.add_option('--no-browser-globals', action='store_true', dest='no_browser_globals', help='do not export browser globals like setTimeout, console, etc. ' + '(This mode is not officially supported for regular applications)') parser.add_option('--without-inspector', action='store_true', dest='without_inspector', help='disable the V8 inspector protocol') parser.add_option('--shared', action='store_true', dest='shared', help='compile shared library for embedding node in another project. ' + '(This mode is not officially supported for regular applications)') parser.add_option('--without-v8-platform', action='store_true', dest='without_v8_platform', default=False, help='do not initialize v8 platform during node.js startup. ' + '(This mode is not officially supported for regular applications)') parser.add_option('--without-bundled-v8', action='store_true', dest='without_bundled_v8', default=False, help='do not use V8 includes from the bundled deps folder. ' + '(This mode is not officially supported for regular applications)') # Create compile_commands.json in out/Debug and out/Release. parser.add_option('-C', action='store_true', dest='compile_commands_json', help=optparse.SUPPRESS_HELP) (options, args) = parser.parse_args() # Expand ~ in the install prefix now, it gets written to multiple files. options.prefix = os.path.expanduser(options.prefix or '') # set up auto-download list auto_downloads = nodedownload.parse(options.download_list) def warn(msg): warn.warned = True prefix = '\033[1m\033[93mWARNING\033[0m' if os.isatty(1) else 'WARNING' print('%s: %s' % (prefix, msg)) # track if warnings occurred warn.warned = False def b(value): """Returns the string 'true' if value is truthy, 'false' otherwise.""" if value: return 'true' else: return 'false' def pkg_config(pkg): pkg_config = os.environ.get('PKG_CONFIG', 'pkg-config') retval = () for flag in ['--libs-only-l', '--cflags-only-I', '--libs-only-L']: try: proc = subprocess.Popen( shlex.split(pkg_config) + ['--silence-errors', flag, pkg], stdout=subprocess.PIPE) val = proc.communicate()[0].strip() except OSError as e: if e.errno != errno.ENOENT: raise e # Unexpected error. return (None, None, None) # No pkg-config/pkgconf installed. retval += (val,) return retval def try_check_compiler(cc, lang): try: proc = subprocess.Popen(shlex.split(cc) + ['-E', '-P', '-x', lang, '-'], stdin=subprocess.PIPE, stdout=subprocess.PIPE) except OSError: return (False, False, '', '') proc.stdin.write('__clang__ __GNUC__ __GNUC_MINOR__ __GNUC_PATCHLEVEL__ ' '__clang_major__ __clang_minor__ __clang_patchlevel__') values = (proc.communicate()[0].split() + ['0'] * 7)[0:7] is_clang = values[0] == '1' gcc_version = '%s.%s.%s' % tuple(values[1:1+3]) clang_version = '%s.%s.%s' % tuple(values[4:4+3]) return (True, is_clang, clang_version, gcc_version) # # The version of asm compiler is needed for building openssl asm files. # See deps/openssl/openssl.gypi for detail. # Commands and regular expressions to obtain its version number are taken from # https://github.com/openssl/openssl/blob/OpenSSL_1_0_2-stable/crypto/sha/asm/sha512-x86_64.pl#L112-L129 # def get_version_helper(cc, regexp): try: proc = subprocess.Popen(shlex.split(cc) + ['-v'], stdin=subprocess.PIPE, stderr=subprocess.PIPE, stdout=subprocess.PIPE) except OSError: print('''Node.js configure error: No acceptable C compiler found! Please make sure you have a C compiler installed on your system and/or consider adjusting the CC environment variable if you installed it in a non-standard prefix. ''') sys.exit() match = re.search(regexp, proc.communicate()[1]) if match: return match.group(2) else: return 0 def get_llvm_version(cc): return get_version_helper( cc, r"(^(?:FreeBSD )?clang version|based on LLVM) ([3-9]\.[0-9]+)") def get_xcode_version(cc): return get_version_helper( cc, r"(^Apple LLVM version) ([5-9]\.[0-9]+)") def get_gas_version(cc): try: proc = subprocess.Popen(shlex.split(cc) + ['-Wa,-v', '-c', '-o', '/dev/null', '-x', 'assembler', '/dev/null'], stdin=subprocess.PIPE, stderr=subprocess.PIPE, stdout=subprocess.PIPE) except OSError: print('''Node.js configure error: No acceptable C compiler found! Please make sure you have a C compiler installed on your system and/or consider adjusting the CC environment variable if you installed it in a non-standard prefix. ''') sys.exit() match = re.match(r"GNU assembler version ([2-9]\.[0-9]+)", proc.communicate()[1]) if match: return match.group(1) else: return 0 # Note: Apple clang self-reports as clang 4.2.0 and gcc 4.2.1. It passes # the version check more by accident than anything else but a more rigorous # check involves checking the build number against a whitelist. I'm not # quite prepared to go that far yet. def check_compiler(o): if sys.platform == 'win32': return ok, is_clang, clang_version, gcc_version = try_check_compiler(CXX, 'c++') if not ok: warn('failed to autodetect C++ compiler version (CXX=%s)' % CXX) elif clang_version < '3.4.2' if is_clang else gcc_version < '4.9.4': warn('C++ compiler too old, need g++ 4.9.4 or clang++ 3.4.2 (CXX=%s)' % CXX) ok, is_clang, clang_version, gcc_version = try_check_compiler(CC, 'c') if not ok: warn('failed to autodetect C compiler version (CC=%s)' % CC) elif not is_clang and gcc_version < '4.2.0': # clang 3.2 is a little white lie because any clang version will probably # do for the C bits. However, we might as well encourage people to upgrade # to a version that is not completely ancient. warn('C compiler too old, need gcc 4.2 or clang 3.2 (CC=%s)' % CC) o['variables']['llvm_version'] = get_llvm_version(CC) if is_clang else 0 # Need xcode_version or gas_version when openssl asm files are compiled. if options.without_ssl or options.openssl_no_asm or options.shared_openssl: return if is_clang: if sys.platform == 'darwin': o['variables']['xcode_version'] = get_xcode_version(CC) else: o['variables']['gas_version'] = get_gas_version(CC) def cc_macros(cc=None): """Checks predefined macros using the C compiler command.""" try: p = subprocess.Popen(shlex.split(cc or CC) + ['-dM', '-E', '-'], stdin=subprocess.PIPE, stdout=subprocess.PIPE, stderr=subprocess.PIPE) except OSError: print('''Node.js configure error: No acceptable C compiler found! Please make sure you have a C compiler installed on your system and/or consider adjusting the CC environment variable if you installed it in a non-standard prefix. ''') sys.exit() p.stdin.write('\n') out = p.communicate()[0] out = str(out).split('\n') k = {} for line in out: lst = shlex.split(line) if len(lst) > 2: key = lst[1] val = lst[2] k[key] = val return k def is_arch_armv7(): """Check for ARMv7 instructions""" cc_macros_cache = cc_macros() return cc_macros_cache.get('__ARM_ARCH') == '7' def is_arch_armv6(): """Check for ARMv6 instructions""" cc_macros_cache = cc_macros() return cc_macros_cache.get('__ARM_ARCH') == '6' def is_arm_hard_float_abi(): """Check for hardfloat or softfloat eabi on ARM""" # GCC versions 4.6 and above define __ARM_PCS or __ARM_PCS_VFP to specify # the Floating Point ABI used (PCS stands for Procedure Call Standard). # We use these as well as a couple of other defines to statically determine # what FP ABI used. return '__ARM_PCS_VFP' in cc_macros() def host_arch_cc(): """Host architecture check using the CC command.""" if sys.platform.startswith('aix'): # we only support gcc at this point and the default on AIX # would be xlc so hard code gcc k = cc_macros('gcc') else: k = cc_macros(os.environ.get('CC_host')) matchup = { '__aarch64__' : 'arm64', '__arm__' : 'arm', '__i386__' : 'ia32', '__MIPSEL__' : 'mipsel', '__mips__' : 'mips', '__PPC64__' : 'ppc64', '__PPC__' : 'ppc64', '__x86_64__' : 'x64', '__s390__' : 's390', '__s390x__' : 's390x', } rtn = 'ia32' # default for i in matchup: if i in k and k[i] != '0': rtn = matchup[i] if rtn != 's390': break if rtn == 'mipsel' and '_LP64' in k: rtn = 'mips64el' return rtn def host_arch_win(): """Host architecture check using environ vars (better way to do this?)""" observed_arch = os.environ.get('PROCESSOR_ARCHITECTURE', 'x86') arch = os.environ.get('PROCESSOR_ARCHITEW6432', observed_arch) matchup = { 'AMD64' : 'x64', 'x86' : 'ia32', 'arm' : 'arm', 'mips' : 'mips', } return matchup.get(arch, 'ia32') def configure_arm(o): if options.arm_float_abi: arm_float_abi = options.arm_float_abi elif is_arm_hard_float_abi(): arm_float_abi = 'hard' else: arm_float_abi = 'default' arm_fpu = 'vfp' if is_arch_armv7(): arm_fpu = 'vfpv3' o['variables']['arm_version'] = '7' else: o['variables']['arm_version'] = '6' if is_arch_armv6() else 'default' o['variables']['arm_thumb'] = 0 # -marm o['variables']['arm_float_abi'] = arm_float_abi if options.dest_os == 'android': arm_fpu = 'vfpv3' o['variables']['arm_version'] = '7' o['variables']['arm_fpu'] = options.arm_fpu or arm_fpu def configure_mips(o): can_use_fpu_instructions = (options.mips_float_abi != 'soft') o['variables']['v8_can_use_fpu_instructions'] = b(can_use_fpu_instructions) o['variables']['v8_use_mips_abi_hardfloat'] = b(can_use_fpu_instructions) o['variables']['mips_arch_variant'] = options.mips_arch_variant o['variables']['mips_fpu_mode'] = options.mips_fpu_mode def configure_node(o): if options.dest_os == 'android': o['variables']['OS'] = 'android' o['variables']['node_prefix'] = options.prefix o['variables']['node_install_npm'] = b(not options.without_npm) o['default_configuration'] = 'Debug' if options.debug else 'Release' host_arch = host_arch_win() if os.name == 'nt' else host_arch_cc() target_arch = options.dest_cpu or host_arch # ia32 is preferred by the build tools (GYP) over x86 even if we prefer the latter # the Makefile resets this to x86 afterward if target_arch == 'x86': target_arch = 'ia32' o['variables']['host_arch'] = host_arch o['variables']['target_arch'] = target_arch o['variables']['node_byteorder'] = sys.byteorder cross_compiling = (options.cross_compiling if options.cross_compiling is not None else target_arch != host_arch) want_snapshots = not options.without_snapshot o['variables']['want_separate_host_toolset'] = int( cross_compiling and want_snapshots) if target_arch == 'arm': configure_arm(o) elif target_arch in ('mips', 'mipsel', 'mips64el'): configure_mips(o) if flavor == 'aix': o['variables']['node_core_target_name'] = 'node_base' o['variables']['node_target_type'] = 'static_library' if target_arch in ('x86', 'x64', 'ia32', 'x32'): o['variables']['node_enable_v8_vtunejit'] = b(options.enable_vtune_profiling) elif options.enable_vtune_profiling: raise Exception( 'The VTune profiler for JavaScript is only supported on x32, x86, and x64 ' 'architectures.') else: o['variables']['node_enable_v8_vtunejit'] = 'false' if flavor in ('solaris', 'mac', 'linux', 'freebsd'): use_dtrace = not options.without_dtrace # Don't enable by default on linux and freebsd if flavor in ('linux', 'freebsd'): use_dtrace = options.with_dtrace if flavor == 'linux': if options.systemtap_includes: o['include_dirs'] += [options.systemtap_includes] o['variables']['node_use_dtrace'] = b(use_dtrace) o['variables']['uv_use_dtrace'] = b(use_dtrace) o['variables']['uv_parent_path'] = '/deps/uv/' elif options.with_dtrace: raise Exception( 'DTrace is currently only supported on SunOS, MacOS or Linux systems.') else: o['variables']['node_use_dtrace'] = 'false' # Enable Lttng if --with-lttng was defined. Use logic similar to # ETW for windows. Lttng is only available on the Linux platform. if flavor == 'linux': o['variables']['node_use_lttng'] = b(options.with_lttng) elif options.with_lttng: raise Exception('lttng is only supported on Linux.') else: o['variables']['node_use_lttng'] = 'false' if options.no_ifaddrs: o['defines'] += ['SUNOS_NO_IFADDRS'] # By default, enable ETW on Windows. if flavor == 'win': o['variables']['node_use_etw'] = b(not options.without_etw) elif options.with_etw: raise Exception('ETW is only supported on Windows.') else: o['variables']['node_use_etw'] = 'false' # By default, enable Performance counters on Windows. if flavor == 'win': o['variables']['node_use_perfctr'] = b(not options.without_perfctr) elif options.with_perfctr: raise Exception('Performance counter is only supported on Windows.') else: o['variables']['node_use_perfctr'] = 'false' if options.tag: o['variables']['node_tag'] = '-' + options.tag else: o['variables']['node_tag'] = '' o['variables']['node_release_urlbase'] = options.release_urlbase or '' if options.v8_options: o['variables']['node_v8_options'] = options.v8_options.replace('"', '\\"') if options.enable_static: o['variables']['node_target_type'] = 'static_library' if options.debug_http2: o['variables']['debug_http2'] = 1 else: o['variables']['debug_http2'] = 'false' if options.debug_nghttp2: o['variables']['debug_nghttp2'] = 1 else: o['variables']['debug_nghttp2'] = 'false' o['variables']['node_no_browser_globals'] = b(options.no_browser_globals) o['variables']['node_shared'] = b(options.shared) node_module_version = getmoduleversion.get_version() if sys.platform == 'darwin': shlib_suffix = '%s.dylib' elif sys.platform.startswith('aix'): shlib_suffix = '%s.a' else: shlib_suffix = 'so.%s' shlib_suffix %= node_module_version o['variables']['node_module_version'] = int(node_module_version) o['variables']['shlib_suffix'] = shlib_suffix if options.linked_module: o['variables']['library_files'] = options.linked_module o['variables']['asan'] = int(options.enable_asan or 0) o['variables']['debug_devtools'] = 'node' if options.use_xcode and options.use_ninja: raise Exception('--xcode and --ninja cannot be used together.') if options.coverage: o['variables']['coverage'] = 'true' else: o['variables']['coverage'] = 'false' def configure_library(lib, output): shared_lib = 'shared_' + lib output['variables']['node_' + shared_lib] = b(getattr(options, shared_lib)) if getattr(options, shared_lib): (pkg_libs, pkg_cflags, pkg_libpath) = pkg_config(lib) if options.__dict__[shared_lib + '_includes']: output['include_dirs'] += [options.__dict__[shared_lib + '_includes']] elif pkg_cflags: output['include_dirs'] += ( filter(None, map(str.strip, pkg_cflags.split('-I')))) # libpath needs to be provided ahead libraries if options.__dict__[shared_lib + '_libpath']: output['libraries'] += [ '-L%s' % options.__dict__[shared_lib + '_libpath']] elif pkg_libpath: output['libraries'] += [pkg_libpath] default_libs = getattr(options, shared_lib + '_libname') default_libs = map('-l{0}'.format, default_libs.split(',')) if default_libs: output['libraries'] += default_libs elif pkg_libs: output['libraries'] += pkg_libs.split() def configure_v8(o): o['variables']['v8_enable_gdbjit'] = 1 if options.gdb else 0 o['variables']['v8_no_strict_aliasing'] = 1 # Work around compiler bugs. o['variables']['v8_optimized_debug'] = 0 # Compile with -O0 in debug builds. o['variables']['v8_random_seed'] = 0 # Use a random seed for hash tables. o['variables']['v8_promise_internal_field_count'] = 1 # Add internal field to promises for async hooks. o['variables']['v8_use_snapshot'] = 'false' if options.without_snapshot else 'true' o['variables']['v8_trace_maps'] = 1 if options.trace_maps else 0 o['variables']['node_use_v8_platform'] = b(not options.without_v8_platform) o['variables']['node_use_bundled_v8'] = b(not options.without_bundled_v8) o['variables']['force_dynamic_crt'] = 1 if options.shared else 0 o['variables']['node_enable_d8'] = b(options.enable_d8) # Unconditionally force typed arrays to allocate outside the v8 heap. This # is to prevent memory pointers from being moved around that are returned by # Buffer::Data(). o['variables']['v8_typed_array_max_size_in_heap'] = 0 if options.enable_d8: o['variables']['test_isolation_mode'] = 'noop' # Needed by d8.gyp. if options.without_bundled_v8 and options.enable_d8: raise Exception('--enable-d8 is incompatible with --without-bundled-v8.') def configure_openssl(o): o['variables']['node_use_openssl'] = b(not options.without_ssl) o['variables']['node_shared_openssl'] = b(options.shared_openssl) o['variables']['openssl_no_asm'] = 1 if options.openssl_no_asm else 0 if options.use_openssl_ca_store: o['defines'] += ['NODE_OPENSSL_CERT_STORE'] if options.openssl_system_ca_path: o['variables']['openssl_system_ca_path'] = options.openssl_system_ca_path o['variables']['node_without_node_options'] = b(options.without_node_options) if options.without_node_options: o['defines'] += ['NODE_WITHOUT_NODE_OPTIONS'] if options.openssl_fips: o['variables']['openssl_fips'] = options.openssl_fips fips_dir = os.path.join('deps', 'openssl', 'fips') fips_ld = os.path.abspath(os.path.join(fips_dir, 'fipsld')) # LINK is for Makefiles, LD/LDXX is for ninja o['make_fips_settings'] = [ ['LINK', fips_ld + ' <(openssl_fips)/bin/fipsld'], ['LD', fips_ld + ' <(openssl_fips)/bin/fipsld'], ['LDXX', fips_ld + ' <(openssl_fips)/bin/fipsld'], ] else: o['variables']['openssl_fips'] = '' try: os.remove('config_fips.gypi') except OSError: pass if options.without_ssl: def without_ssl_error(option): print('Error: --without-ssl is incompatible with %s' % option) exit(1) if options.shared_openssl: without_ssl_error('--shared-openssl') if options.openssl_no_asm: without_ssl_error('--openssl-no-asm') if options.openssl_fips: without_ssl_error('--openssl-fips') return configure_library('openssl', o) def configure_static(o): if options.fully_static or options.partly_static: if flavor == 'mac': print("Generation of static executable will not work on OSX " "when using the default compilation environment") return if options.fully_static: o['libraries'] += ['-static'] elif options.partly_static: o['libraries'] += ['-static-libgcc', '-static-libstdc++'] if options.enable_asan: o['libraries'] += ['-static-libasan'] def write(filename, data): filename = filename print('creating %s' % filename) f = open(filename, 'w+') f.write(data) do_not_edit = '# Do not edit. Generated by the configure script.\n' def glob_to_var(dir_base, dir_sub, patch_dir): list = [] dir_all = '%s/%s' % (dir_base, dir_sub) files = os.walk(dir_all) for ent in files: (path, dirs, files) = ent for file in files: if file.endswith('.cpp') or file.endswith('.c') or file.endswith('.h'): # srcfile uses "slash" as dir separator as its output is consumed by gyp srcfile = '%s/%s' % (dir_sub, file) if patch_dir: patchfile = '%s/%s/%s' % (dir_base, patch_dir, file) if os.path.isfile(patchfile): srcfile = '%s/%s' % (patch_dir, file) print('Using version-specific floating patch %s' % patchfile) list.append(srcfile) break return list def configure_intl(o): icus = [ { 'url': 'https://ssl.icu-project.org/files/icu4c/60.1/icu4c-60_1-src.zip', 'md5': 'e6cb990ac2a3161d31a3def8435f80cb', }, ] def icu_download(path): # download ICU, if needed if not os.access(options.download_path, os.W_OK): print('Error: cannot write to desired download path. ' \ 'Either create it or verify permissions.') sys.exit(1) for icu in icus: url = icu['url'] md5 = icu['md5'] local = url.split('/')[-1] targetfile = os.path.join(options.download_path, local) if not os.path.isfile(targetfile): if nodedownload.candownload(auto_downloads, "icu"): nodedownload.retrievefile(url, targetfile) else: print('Re-using existing %s' % targetfile) if os.path.isfile(targetfile): print('Checking file integrity with MD5:\r') gotmd5 = nodedownload.md5sum(targetfile) print('MD5: %s %s' % (gotmd5, targetfile)) if (md5 == gotmd5): return targetfile else: print('Expected: %s *MISMATCH*' % md5) print('\n ** Corrupted ZIP? Delete %s to retry download.\n' % targetfile) return None icu_config = { 'variables': {} } icu_config_name = 'icu_config.gypi' def write_config(data, name): return # write an empty file to start with write(icu_config_name, do_not_edit + pprint.pformat(icu_config, indent=2) + '\n') # always set icu_small, node.gyp depends on it being defined. o['variables']['icu_small'] = b(False) with_intl = options.with_intl with_icu_source = options.with_icu_source have_icu_path = bool(options.with_icu_path) if have_icu_path and with_intl != 'none': print('Error: Cannot specify both --with-icu-path and --with-intl') sys.exit(1) elif have_icu_path: # Chromium .gyp mode: --with-icu-path o['variables']['v8_enable_i18n_support'] = 1 # use the .gyp given o['variables']['icu_gyp_path'] = options.with_icu_path return # --with-intl= # set the default if with_intl in (None, 'none'): o['variables']['v8_enable_i18n_support'] = 0 return # no Intl elif with_intl == 'small-icu': # small ICU (English only) o['variables']['v8_enable_i18n_support'] = 1 o['variables']['icu_small'] = b(True) locs = set(options.with_icu_locales.split(',')) locs.add('root') # must have root o['variables']['icu_locales'] = string.join(locs,',') # We will check a bit later if we can use the canned deps/icu-small elif with_intl == 'full-icu': # full ICU o['variables']['v8_enable_i18n_support'] = 1 elif with_intl == 'system-icu': # ICU from pkg-config. o['variables']['v8_enable_i18n_support'] = 1 pkgicu = pkg_config('icu-i18n') if pkgicu[0] is None: print('Error: could not load pkg-config data for "icu-i18n".') print('See above errors or the README.md.') sys.exit(1) (libs, cflags, libpath) = pkgicu # libpath provides linker path which may contain spaces if libpath: o['libraries'] += [libpath] # safe to split, cannot contain spaces o['libraries'] += libs.split() if cflags: o['include_dirs'] += filter(None, map(str.strip, cflags.split('-I'))) # use the "system" .gyp o['variables']['icu_gyp_path'] = 'tools/icu/icu-system.gyp' return # this is just the 'deps' dir. Used for unpacking. icu_parent_path = 'deps' # The full path to the ICU source directory. Should not include './'. icu_full_path = 'deps/icu' # icu-tmp is used to download and unpack the ICU tarball. icu_tmp_path = os.path.join(icu_parent_path, 'icu-tmp') # canned ICU. see tools/icu/README.md to update. canned_icu_dir = 'deps/icu-small' # We can use 'deps/icu-small' - pre-canned ICU *iff* # - with_intl == small-icu (the default!) # - with_icu_locales == 'root,en' (the default!) # - deps/icu-small exists! # - with_icu_source is unset (i.e. no other ICU was specified) # (Note that this is the *DEFAULT CASE*.) # # This is *roughly* equivalent to # $ configure --with-intl=small-icu --with-icu-source=deps/icu-small # .. Except that we avoid copying icu-small over to deps/icu. # In this default case, deps/icu is ignored, although make clean will # still harmlessly remove deps/icu. # are we using default locales? using_default_locales = ( options.with_icu_locales == icu_default_locales ) # make sure the canned ICU really exists canned_icu_available = os.path.isdir(canned_icu_dir) if (o['variables']['icu_small'] == b(True)) and using_default_locales and (not with_icu_source) and canned_icu_available: # OK- we can use the canned ICU. icu_config['variables']['icu_small_canned'] = 1 icu_full_path = canned_icu_dir # --with-icu-source processing # now, check that they didn't pass --with-icu-source=deps/icu elif with_icu_source and os.path.abspath(icu_full_path) == os.path.abspath(with_icu_source): print('Ignoring redundant --with-icu-source=%s' % with_icu_source) with_icu_source = None # if with_icu_source is still set, try to use it. if with_icu_source: if os.path.isdir(icu_full_path): print('Deleting old ICU source: %s' % icu_full_path) shutil.rmtree(icu_full_path) # now, what path was given? if os.path.isdir(with_icu_source): # it's a path. Copy it. print('%s -> %s' % (with_icu_source, icu_full_path)) shutil.copytree(with_icu_source, icu_full_path) else: # could be file or URL. # Set up temporary area if os.path.isdir(icu_tmp_path): shutil.rmtree(icu_tmp_path) os.mkdir(icu_tmp_path) icu_tarball = None if os.path.isfile(with_icu_source): # it's a file. Try to unpack it. icu_tarball = with_icu_source else: # Can we download it? local = os.path.join(icu_tmp_path, with_icu_source.split('/')[-1]) # local part icu_tarball = nodedownload.retrievefile(with_icu_source, local) # continue with "icu_tarball" nodedownload.unpack(icu_tarball, icu_tmp_path) # Did it unpack correctly? Should contain 'icu' tmp_icu = os.path.join(icu_tmp_path, 'icu') if os.path.isdir(tmp_icu): os.rename(tmp_icu, icu_full_path) shutil.rmtree(icu_tmp_path) else: print('Error: --with-icu-source=%s did not result in an "icu" dir.' % \ with_icu_source) shutil.rmtree(icu_tmp_path) sys.exit(1) # ICU mode. (icu-generic.gyp) o['variables']['icu_gyp_path'] = 'tools/icu/icu-generic.gyp' # ICU source dir relative to tools/icu (for .gyp file) o['variables']['icu_path'] = icu_full_path if not os.path.isdir(icu_full_path): print('* ECMA-402 (Intl) support didn\'t find ICU in %s..' % icu_full_path) # can we download (or find) a zipfile? localzip = icu_download(icu_full_path) if localzip: nodedownload.unpack(localzip, icu_parent_path) if not os.path.isdir(icu_full_path): print('Cannot build Intl without ICU in %s.' % icu_full_path) print('(Fix, or disable with "--with-intl=none" )') sys.exit(1) else: print('* Using ICU in %s' % icu_full_path) # Now, what version of ICU is it? We just need the "major", such as 54. # uvernum.h contains it as a #define. uvernum_h = os.path.join(icu_full_path, 'source/common/unicode/uvernum.h') if not os.path.isfile(uvernum_h): print('Error: could not load %s - is ICU installed?' % uvernum_h) sys.exit(1) icu_ver_major = None matchVerExp = r'^\s*#define\s+U_ICU_VERSION_SHORT\s+"([^"]*)".*' match_version = re.compile(matchVerExp) for line in open(uvernum_h).readlines(): m = match_version.match(line) if m: icu_ver_major = m.group(1) if not icu_ver_major: print('Could not read U_ICU_VERSION_SHORT version from %s' % uvernum_h) sys.exit(1) icu_endianness = sys.byteorder[0]; o['variables']['icu_ver_major'] = icu_ver_major o['variables']['icu_endianness'] = icu_endianness icu_data_file_l = 'icudt%s%s.dat' % (icu_ver_major, 'l') icu_data_file = 'icudt%s%s.dat' % (icu_ver_major, icu_endianness) # relative to configure icu_data_path = os.path.join(icu_full_path, 'source/data/in', icu_data_file_l) # relative to dep.. icu_data_in = os.path.join('..','..', icu_full_path, 'source/data/in', icu_data_file_l) if not os.path.isfile(icu_data_path) and icu_endianness != 'l': # use host endianness icu_data_path = os.path.join(icu_full_path, 'source/data/in', icu_data_file) # relative to dep.. icu_data_in = os.path.join('..', icu_full_path, 'source/data/in', icu_data_file) # this is the input '.dat' file to use .. icudt*.dat # may be little-endian if from a icu-project.org tarball o['variables']['icu_data_in'] = icu_data_in # this is the icudt*.dat file which node will be using (platform endianness) o['variables']['icu_data_file'] = icu_data_file if not os.path.isfile(icu_data_path): print('Error: ICU prebuilt data file %s does not exist.' % icu_data_path) print('See the README.md.') # .. and we're not about to build it from .gyp! sys.exit(1) # map from variable name to subdirs icu_src = { 'stubdata': 'stubdata', 'common': 'common', 'i18n': 'i18n', 'tools': 'tools/toolutil', 'genccode': 'tools/genccode', 'genrb': 'tools/genrb', 'icupkg': 'tools/icupkg', } # this creates a variable icu_src_XXX for each of the subdirs # with a list of the src files to use for i in icu_src: var = 'icu_src_%s' % i path = '../../%s/source/%s' % (icu_full_path, icu_src[i]) icu_config['variables'][var] = glob_to_var('tools/icu', path, 'patches/%s/source/%s' % (icu_ver_major, icu_src[i]) ) # write updated icu_config.gypi with a bunch of paths write(icu_config_name, do_not_edit + pprint.pformat(icu_config, indent=2) + '\n') return # end of configure_intl def configure_inspector(o): disable_inspector = (options.without_inspector or options.with_intl in (None, 'none') or options.without_ssl) o['variables']['v8_enable_inspector'] = 0 if disable_inspector else 1 def make_bin_override(): if sys.platform == 'win32': raise Exception('make_bin_override should not be called on win32.') # If the system python is not the python we are running (which should be # python 2), then create a directory with a symlink called `python` to our # sys.executable. This directory will be prefixed to the PATH, so that # other tools that shell out to `python` will use the appropriate python which_python = which('python') if (which_python and os.path.realpath(which_python) == os.path.realpath(sys.executable)): return bin_override = os.path.abspath('out/tools/bin') try: os.makedirs(bin_override) except OSError as e: if e.errno != errno.EEXIST: raise e python_link = os.path.join(bin_override, 'python') try: os.unlink(python_link) except OSError as e: if e.errno != errno.ENOENT: raise e os.symlink(sys.executable, python_link) # We need to set the environment right now so that when gyp (in run_gyp) # shells out, it finds the right python (specifically at # https://github.com/nodejs/node/blob/d82e107/deps/v8/gypfiles/toolchain.gypi#L43) os.environ['PATH'] = bin_override + ':' + os.environ['PATH'] return bin_override output = { 'variables': {}, 'include_dirs': [], 'libraries': [], 'defines': [], 'cflags': [], } # Print a warning when the compiler is too old. check_compiler(output) # determine the "flavor" (operating system) we're building for, # leveraging gyp's GetFlavor function flavor_params = {} if (options.dest_os): flavor_params['flavor'] = options.dest_os flavor = GetFlavor(flavor_params) configure_node(output) configure_library('zlib', output) configure_library('http_parser', output) configure_library('libuv', output) configure_library('libcares', output) configure_library('nghttp2', output) # stay backwards compatible with shared cares builds output['variables']['node_shared_cares'] = \ output['variables'].pop('node_shared_libcares') configure_v8(output) configure_openssl(output) configure_intl(output) configure_static(output) configure_inspector(output) # variables should be a root level element, # move everything else to target_defaults variables = output['variables'] del output['variables'] # make_global_settings for special FIPS linking # should not be used to compile modules in node-gyp config_fips = { 'make_global_settings' : [] } if 'make_fips_settings' in output: config_fips['make_global_settings'] = output['make_fips_settings'] del output['make_fips_settings'] write('config_fips.gypi', do_not_edit + pprint.pformat(config_fips, indent=2) + '\n') # make_global_settings should be a root level element too if 'make_global_settings' in output: make_global_settings = output['make_global_settings'] del output['make_global_settings'] else: make_global_settings = False output = { 'variables': variables, 'target_defaults': output, } if make_global_settings: output['make_global_settings'] = make_global_settings pprint.pprint(output, indent=2) write('config.gypi', do_not_edit + pprint.pformat(output, indent=2) + '\n') config = { 'BUILDTYPE': 'Debug' if options.debug else 'Release', 'USE_XCODE': str(int(options.use_xcode or 0)), 'PYTHON': sys.executable, } if options.prefix: config['PREFIX'] = options.prefix config = '\n'.join(map('='.join, config.iteritems())) + '\n' # On Windows there's no reason to search for a different python binary. bin_override = None if sys.platform == 'win32' else make_bin_override() if bin_override: config = 'export PATH:=' + bin_override + ':$(PATH)\n' + config write('config.mk', do_not_edit + config) gyp_args = ['--no-parallel'] if options.use_xcode: gyp_args += ['-f', 'xcode'] elif options.use_ninja: gyp_args += ['-f', 'ninja'] elif flavor == 'win' and sys.platform != 'msys': gyp_args += ['-f', 'msvs', '-G', 'msvs_version=auto'] else: gyp_args += ['-f', 'make-' + flavor] if options.compile_commands_json: gyp_args += ['-f', 'compile_commands_json'] gyp_args += args if warn.warned: warn('warnings were emitted in the configure phase') run_gyp(gyp_args)