diff options
Diffstat (limited to 'user')
-rw-r--r-- | user/node/APKBUILD | 66 | ||||
-rw-r--r-- | user/node/fix-all-caps-bool.patch | 84 | ||||
-rw-r--r-- | user/node/gyp-python3.patch | 3862 | ||||
-rw-r--r-- | user/node/libatomic.patch | 14 | ||||
-rw-r--r-- | user/node/ppc32.patch | 29 | ||||
-rw-r--r-- | user/node/ppc64.patch | 40 | ||||
-rw-r--r-- | user/node/python3.patch | 163 | ||||
-rw-r--r-- | user/node/stack-silliness.patch | 14 | ||||
-rw-r--r-- | user/node/v8-python3.patch | 169 |
9 files changed, 15 insertions, 4426 deletions
diff --git a/user/node/APKBUILD b/user/node/APKBUILD index 578040d07..954eee721 100644 --- a/user/node/APKBUILD +++ b/user/node/APKBUILD @@ -1,8 +1,8 @@ # Contributor: A. Wilcox <awilfox@adelielinux.org> # Maintainer: A. Wilcox <awilfox@adelielinux.org> pkgname=node -pkgver=10.21.0 -pkgrel=1 +pkgver=16.15.0 +pkgrel=0 pkgdesc="JavaScript runtime" url="https://nodejs.org/" arch="all" @@ -10,16 +10,9 @@ options="net" # Required in check() license="MIT AND ICU AND BSD-3-Clause AND BSD-2-Clause AND ISC AND Public-Domain AND Zlib AND Artistic-2.0 AND Apache-2.0 AND CC0-1.0" depends="" makedepends="c-ares-dev http-parser-dev icu-dev libexecinfo-dev libuv-dev - nghttp2-dev>=1.41 openssl-dev python3 zlib-dev" + nghttp2-dev openssl-dev python3 zlib-dev samurai" subpackages="$pkgname-dev $pkgname-doc" -source="https://nodejs.org/download/release/v$pkgver/node-v$pkgver.tar.xz - https://www.python.org/ftp/python/2.7.15/Python-2.7.15.tar.xz - libatomic.patch - ppc32.patch - ppc64.patch - stack-silliness.patch - fix-all-caps-bool.patch - " +source="https://nodejs.org/download/release/v$pkgver/node-v$pkgver.tar.xz" builddir="$srcdir/$pkgname-v$pkgver" # secfixes: @@ -36,59 +29,30 @@ builddir="$srcdir/$pkgname-v$pkgver" # - CVE-2020-7598 # - CVE-2020-8174 -unpack() { - default_unpack - [ -z $SKIP_PYTHON ] || return 0 - - # TODO: when bumping to 12.x, python3 should be usable - msg "Killing all remaining hope for humanity and building Python 2..." - cd "$srcdir/Python-2.7.15" - [ -d ../python ] && rm -r ../python - # 19:39 <+solar> just make the firefox build process build its own py2 copy - # 20:03 <calvin> TheWilfox: there's always violence - ./configure --prefix="$srcdir/python" - make -j $JOBS - make -j $JOBS install -} - build() { - export PATH="$srcdir/python/bin:$PATH" - python ./configure.py \ + python3 configure.py \ --prefix=/usr \ - --with-intl=system-icu \ + --shared-zlib \ + --shared-openssl \ --shared-cares \ - --shared-http-parser \ - --shared-libuv \ --shared-nghttp2 \ - --shared-openssl \ + --ninja \ --openssl-use-def-ca-store \ - --shared-zlib - # keep DESTDIR set, to avoid a full rebuild in package() - make DESTDIR="$pkgdir" + --with-icu-default-data-dir=$(icu-config --icudatadir) \ + --with-intl=small-icu \ + --without-corepack \ + --without-npm + + make BUILDTYPE=Release DESTDIR="$pkgdir" } check() { - case "$CARCH" in - pmmx) - # https://bts.adelielinux.org/show_bug.cgi?id=306 - _skip="parallel/test-http-invalid-te,parallel/test-worker-stdio" - ;; - esac - - export PATH="$srcdir/python/bin:$PATH" make DESTDIR="$pkgdir" test-only \ ${_skip:+CI_SKIP_TESTS="$_skip"} } package() { - export PATH="$srcdir/python/bin:$PATH" make DESTDIR="$pkgdir" install } -sha512sums="613d3c1bca79ea5f127dc6793de2b5cfdfa056c01ec092e3b7ee79205894b21ca5ec4a367265122641dd1d360c675cfb36a4f7892894194ddd18abd1b2206544 node-v10.21.0.tar.xz -27ea43eb45fc68f3d2469d5f07636e10801dee11635a430ec8ec922ed790bb426b072da94df885e4dfa1ea8b7a24f2f56dd92f9b0f51e162330f161216bd6de6 Python-2.7.15.tar.xz -8f64922d586bce9d82c83042a989739cc55ecc5e015778cdfbda21c257aa50527ddb18740985bcb2068e4a749b71eb8a135d9a8152b374d361589df7f33c9b60 libatomic.patch -6d37794c7c78ef92ebb845852af780e22dc8c14653b63a8609c21ab6860877b9dffc5cf856a8516b7978ec704f312c0627075c6440ace55d039f95bdc4c85add ppc32.patch -583326353de5b0ac14a6c42321f6b031bd943a80550624794e15bd7526470f67bfa14a66558db3c94b4ee2db3053d2e4efed2117f4e7b6dca3c59c171048c094 ppc64.patch -3ea09e36ed0cc31e0475ebc9c92b7609b70e9c1637c5db6c92cf1d6363fb8c6f884ffa20dd81054ca390b721695185327d80c9eeff0688a959e9d46947602471 stack-silliness.patch -a29d21c7336631aa099ef2fdf8ed63f1f75d10127f26103b9638cbeab5837f00603996f05bbce4c19694b19cd08f7f0fd516f011827a7fb912177334076f1bf0 fix-all-caps-bool.patch" +sha512sums="c85cf4a94a0dccdcf358a3e0383499fe1fd74ac0e7b6aa888e9524c070ae8be68b9f718c8c0940d51115bbc908202cd7819e370507b7191facd375a5be231c69 node-v16.15.0.tar.xz" diff --git a/user/node/fix-all-caps-bool.patch b/user/node/fix-all-caps-bool.patch deleted file mode 100644 index fa1007a3b..000000000 --- a/user/node/fix-all-caps-bool.patch +++ /dev/null @@ -1,84 +0,0 @@ -diff -ur a/deps/v8/src/objects/intl-objects.cc b/deps/v8/src/objects/intl-objects.cc ---- a/deps/v8/src/objects/intl-objects.cc 2022-04-22 08:12:51.454294680 +0000 -+++ b/deps/v8/src/objects/intl-objects.cc 2022-04-22 08:19:29.655077696 +0000 -@@ -243,7 +243,7 @@ - char result[ULOC_FULLNAME_CAPACITY]; - status = U_ZERO_ERROR; - uloc_toLanguageTag(icu_locale.getName(), result, ULOC_FULLNAME_CAPACITY, -- FALSE, &status); -+ false, &status); - if (U_SUCCESS(status)) { - JSObject::SetProperty(resolved, factory->NewStringFromStaticChars("locale"), - factory->NewStringFromAsciiChecked(result), -@@ -422,7 +422,7 @@ - char result[ULOC_FULLNAME_CAPACITY]; - UErrorCode status = U_ZERO_ERROR; - uloc_toLanguageTag(icu_locale.getName(), result, ULOC_FULLNAME_CAPACITY, -- FALSE, &status); -+ false, &status); - if (U_SUCCESS(status)) { - JSObject::SetProperty(resolved, factory->NewStringFromStaticChars("locale"), - factory->NewStringFromAsciiChecked(result), -@@ -657,7 +657,7 @@ - char result[ULOC_FULLNAME_CAPACITY]; - status = U_ZERO_ERROR; - uloc_toLanguageTag(icu_locale.getName(), result, ULOC_FULLNAME_CAPACITY, -- FALSE, &status); -+ false, &status); - if (U_SUCCESS(status)) { - JSObject::SetProperty(resolved, factory->NewStringFromStaticChars("locale"), - factory->NewStringFromAsciiChecked(result), -@@ -796,7 +796,7 @@ - char result[ULOC_FULLNAME_CAPACITY]; - status = U_ZERO_ERROR; - uloc_toLanguageTag(icu_locale.getName(), result, ULOC_FULLNAME_CAPACITY, -- FALSE, &status); -+ false, &status); - if (U_SUCCESS(status)) { - JSObject::SetProperty(resolved, factory->NewStringFromStaticChars("locale"), - factory->NewStringFromAsciiChecked(result), -diff -ur a/deps/v8/src/runtime/runtime-intl.cc b/deps/v8/src/runtime/runtime-intl.cc ---- a/deps/v8/src/runtime/runtime-intl.cc 2022-04-22 08:27:09.577527378 +0000 -+++ b/deps/v8/src/runtime/runtime-intl.cc 2022-04-22 08:27:24.397928433 +0000 -@@ -85,7 +85,7 @@ - char result[ULOC_FULLNAME_CAPACITY]; - - // Force strict BCP47 rules. -- uloc_toLanguageTag(icu_result, result, ULOC_FULLNAME_CAPACITY, TRUE, &error); -+ uloc_toLanguageTag(icu_result, result, ULOC_FULLNAME_CAPACITY, true, &error); - - if (U_FAILURE(error) || error == U_STRING_NOT_TERMINATED_WARNING) { - return *factory->NewStringFromAsciiChecked(kInvalidTag); -diff -ur a/deps/v8/src/runtime/runtime-intl.cc b/deps/v8/src/runtime/runtime-intl.cc ---- a/deps/v8/src/runtime/runtime-intl.cc 2022-04-22 08:29:09.750779385 +0000 -+++ b/deps/v8/src/runtime/runtime-intl.cc 2022-04-22 08:29:41.901649421 +0000 -@@ -132,7 +132,7 @@ - - error = U_ZERO_ERROR; - // No need to force strict BCP47 rules. -- uloc_toLanguageTag(icu_name, result, ULOC_FULLNAME_CAPACITY, FALSE, &error); -+ uloc_toLanguageTag(icu_name, result, ULOC_FULLNAME_CAPACITY, false, &error); - if (U_FAILURE(error) || error == U_STRING_NOT_TERMINATED_WARNING) { - // This shouldn't happen, but lets not break the user. - continue; -@@ -164,7 +164,7 @@ - char result[ULOC_FULLNAME_CAPACITY]; - UErrorCode status = U_ZERO_ERROR; - uloc_toLanguageTag(default_locale.getName(), result, ULOC_FULLNAME_CAPACITY, -- FALSE, &status); -+ false, &status); - if (U_SUCCESS(status)) { - return *factory->NewStringFromAsciiChecked(result); - } -@@ -399,9 +399,9 @@ - std::unique_ptr<uc16[]> sap1; - std::unique_ptr<uc16[]> sap2; - icu::UnicodeString string_val1( -- FALSE, GetUCharBufferFromFlat(flat1, &sap1, length1), length1); -+ false, GetUCharBufferFromFlat(flat1, &sap1, length1), length1); - icu::UnicodeString string_val2( -- FALSE, GetUCharBufferFromFlat(flat2, &sap2, length2), length2); -+ false, GetUCharBufferFromFlat(flat2, &sap2, length2), length2); - result = collator->compare(string_val1, string_val2, status); - } - if (U_FAILURE(status)) return isolate->ThrowIllegalOperation(); diff --git a/user/node/gyp-python3.patch b/user/node/gyp-python3.patch deleted file mode 100644 index e52ef19ec..000000000 --- a/user/node/gyp-python3.patch +++ /dev/null @@ -1,3862 +0,0 @@ -diff --git a/tools/gyp/pylib/gyp/input.py b/pylib/gyp/input.py -index a046a15..21b4606 100644 ---- a/tools/gyp/pylib/gyp/input.py -+++ b/tools/gyp/pylib/gyp/input.py -@@ -2,14 +2,8 @@ - # Use of this source code is governed by a BSD-style license that can be - # found in the LICENSE file. - --from compiler.ast import Const --from compiler.ast import Dict --from compiler.ast import Discard --from compiler.ast import List --from compiler.ast import Module --from compiler.ast import Node --from compiler.ast import Stmt --import compiler -+import ast -+ - import gyp.common - import gyp.simple_copy - import multiprocessing -@@ -183,43 +177,38 @@ - Note that this is slower than eval() is. - """ - -- ast = compiler.parse(file_contents) -- assert isinstance(ast, Module) -- c1 = ast.getChildren() -- assert c1[0] is None -- assert isinstance(c1[1], Stmt) -- c2 = c1[1].getChildren() -- assert isinstance(c2[0], Discard) -- c3 = c2[0].getChildren() -- assert len(c3) == 1 -- return CheckNode(c3[0], []) -+ syntax_tree = ast.parse(file_contents) -+ assert isinstance(syntax_tree, ast.Module) -+ c1 = syntax_tree.body -+ assert len(c1) == 1 -+ c2 = c1[0] -+ assert isinstance(c2, ast.Expr) -+ return CheckNode(c2.value, []) - - - def CheckNode(node, keypath): -- if isinstance(node, Dict): -- c = node.getChildren() -+ if isinstance(node, ast.Dict): - dict = {} -- for n in range(0, len(c), 2): -- assert isinstance(c[n], Const) -- key = c[n].getChildren()[0] -+ for key, value in zip(node.keys, node.values): -+ assert isinstance(key, ast.Str) -+ key = key.s - if key in dict: - raise GypError("Key '" + key + "' repeated at level " + - repr(len(keypath) + 1) + " with key path '" + - '.'.join(keypath) + "'") - kp = list(keypath) # Make a copy of the list for descending this node. - kp.append(key) -- dict[key] = CheckNode(c[n + 1], kp) -+ dict[key] = CheckNode(value, kp) - return dict -- elif isinstance(node, List): -- c = node.getChildren() -+ elif isinstance(node, ast.List): - children = [] -- for index, child in enumerate(c): -+ for index, child in enumerate(node.elts): - kp = list(keypath) # Copy list. - kp.append(repr(index)) - children.append(CheckNode(child, kp)) - return children -- elif isinstance(node, Const): -- return node.getChildren()[0] -+ elif isinstance(node, ast.Str): -+ return node.s - else: - raise TypeError("Unknown AST node at key path '" + '.'.join(keypath) + - "': " + repr(node)) -diff --git a/tools/gyp/pylib/gyp/simple_copy.py b/pylib/gyp/simple_copy.py -index 74c98c5..eaf5f8b 100644 ---- a/tools/gyp/pylib/gyp/simple_copy.py -+++ b/tools/gyp/pylib/gyp/simple_copy.py -@@ -28,8 +28,19 @@ - def _deepcopy_atomic(x): - return x - --for x in (type(None), int, long, float, -- bool, str, unicode, type): -+try: -+ _string_types = (str, unicode) -+# There's no unicode in python3 -+except NameError: -+ _string_types = (str, ) -+ -+try: -+ _integer_types = (int, long) -+# There's no long in python3 -+except NameError: -+ _integer_types = (int, ) -+ -+for x in (type(None), float, bool, type) + _integer_types + _string_types: - d[x] = _deepcopy_atomic - - def _deepcopy_list(x): -diff --git a/tools/gyp/PRESUBMIT.py b/PRESUBMIT.py -index 4bc1b8c..5ee669b 100644 ---- a/tools/gyp/PRESUBMIT.py -+++ b/tools/gyp/PRESUBMIT.py -@@ -76,8 +76,7 @@ - def _LicenseHeader(input_api): - # Accept any year number from 2009 to the current year. - current_year = int(input_api.time.strftime('%Y')) -- allowed_years = (str(s) for s in reversed(xrange(2009, current_year + 1))) -- -+ allowed_years = (str(s) for s in reversed(range(2009, current_year + 1))) - years_re = '(' + '|'.join(allowed_years) + ')' - - # The (c) is deprecated, but tolerate it until it's removed from all files. -diff --git a/tools/gyp/README.md b/README.md -index c0d73ac..b4766c9 100644 ---- a/tools/gyp/README.md -+++ b/tools/gyp/README.md -@@ -1,4 +1,5 @@ - GYP can Generate Your Projects. - =================================== - --Documents are available at [gyp.gsrc.io](https://gyp.gsrc.io), or you can check out ```md-pages``` branch to read those documents offline. -+Documents are available at [gyp.gsrc.io](https://gyp.gsrc.io), or you can -+check out ```md-pages``` branch to read those documents offline. -diff --git a/tools/gyp/buildbot/buildbot_run.py b/buildbot/buildbot_run.py -index 9a2b71f..8941652 100755 ---- a/tools/gyp/buildbot/buildbot_run.py -+++ b/tools/gyp/buildbot/buildbot_run.py -@@ -5,6 +5,8 @@ - - """Argument-less script to select what to run on the buildbots.""" - -+from __future__ import print_function -+ - import os - import shutil - import subprocess -@@ -24,14 +26,14 @@ - with open(os.devnull) as devnull_fd: - retcode = subprocess.call(stdin=devnull_fd, *args, **kwargs) - if retcode != 0: -- print '@@@STEP_EXCEPTION@@@' -+ print('@@@STEP_EXCEPTION@@@') - sys.exit(1) - - - def PrepareCmake(): - """Build CMake 2.8.8 since the version in Precise is 2.8.7.""" - if os.environ['BUILDBOT_CLOBBER'] == '1': -- print '@@@BUILD_STEP Clobber CMake checkout@@@' -+ print('@@@BUILD_STEP Clobber CMake checkout@@@') - shutil.rmtree(CMAKE_DIR) - - # We always build CMake 2.8.8, so no need to do anything -@@ -39,10 +41,10 @@ - if os.path.isdir(CMAKE_DIR): - return - -- print '@@@BUILD_STEP Initialize CMake checkout@@@' -+ print('@@@BUILD_STEP Initialize CMake checkout@@@') - os.mkdir(CMAKE_DIR) - -- print '@@@BUILD_STEP Sync CMake@@@' -+ print('@@@BUILD_STEP Sync CMake@@@') - CallSubProcess( - ['git', 'clone', - '--depth', '1', -@@ -53,7 +55,7 @@ - CMAKE_DIR], - cwd=CMAKE_DIR) - -- print '@@@BUILD_STEP Build CMake@@@' -+ print('@@@BUILD_STEP Build CMake@@@') - CallSubProcess( - ['/bin/bash', 'bootstrap', '--prefix=%s' % CMAKE_DIR], - cwd=CMAKE_DIR) -@@ -74,7 +76,7 @@ - if not format: - format = title - -- print '@@@BUILD_STEP ' + title + '@@@' -+ print('@@@BUILD_STEP ' + title + '@@@') - sys.stdout.flush() - env = os.environ.copy() - if msvs_version: -@@ -89,17 +91,17 @@ - retcode = subprocess.call(command, cwd=ROOT_DIR, env=env, shell=True) - if retcode: - # Emit failure tag, and keep going. -- print '@@@STEP_FAILURE@@@' -+ print('@@@STEP_FAILURE@@@') - return 1 - return 0 - - - def GypBuild(): - # Dump out/ directory. -- print '@@@BUILD_STEP cleanup@@@' -- print 'Removing %s...' % OUT_DIR -+ print('@@@BUILD_STEP cleanup@@@') -+ print('Removing %s...' % OUT_DIR) - shutil.rmtree(OUT_DIR, ignore_errors=True) -- print 'Done.' -+ print('Done.') - - retcode = 0 - if sys.platform.startswith('linux'): -@@ -128,7 +130,7 @@ - # after the build proper that could be used for cumulative failures), - # use that instead of this. This isolates the final return value so - # that it isn't misattributed to the last stage. -- print '@@@BUILD_STEP failures@@@' -+ print('@@@BUILD_STEP failures@@@') - sys.exit(retcode) - - -diff --git a/tools/gyp/pylib/gyp/MSVSSettings.py b/pylib/gyp/MSVSSettings.py -index 8ae1918..1d2e25a 100644 ---- a/tools/gyp/pylib/gyp/MSVSSettings.py -+++ b/tools/gyp/pylib/gyp/MSVSSettings.py -@@ -14,9 +14,17 @@ - MSBuild install directory, e.g. c:\Program Files (x86)\MSBuild - """ - -+from __future__ import print_function -+ - import sys - import re - -+try: -+ # basestring was removed in python3. -+ basestring -+except NameError: -+ basestring = str -+ - # Dictionaries of settings validators. The key is the tool name, the value is - # a dictionary mapping setting names to validation functions. - _msvs_validators = {} -@@ -400,7 +408,7 @@ - - if unrecognized: - # We don't know this setting. Give a warning. -- print >> stderr, error_msg -+ print(error_msg, file=stderr) - - - def FixVCMacroSlashes(s): -@@ -433,7 +441,7 @@ - '$(PlatformName)': '$(Platform)', - '$(SafeInputName)': '%(Filename)', - } -- for old, new in replace_map.iteritems(): -+ for old, new in replace_map.items(): - s = s.replace(old, new) - s = FixVCMacroSlashes(s) - return s -@@ -453,17 +461,18 @@ - dictionaries of settings and their values. - """ - msbuild_settings = {} -- for msvs_tool_name, msvs_tool_settings in msvs_settings.iteritems(): -+ for msvs_tool_name, msvs_tool_settings in msvs_settings.items(): - if msvs_tool_name in _msvs_to_msbuild_converters: - msvs_tool = _msvs_to_msbuild_converters[msvs_tool_name] -- for msvs_setting, msvs_value in msvs_tool_settings.iteritems(): -+ for msvs_setting, msvs_value in msvs_tool_settings.items(): - if msvs_setting in msvs_tool: - # Invoke the translation function. - try: - msvs_tool[msvs_setting](msvs_value, msbuild_settings) -- except ValueError, e: -- print >> stderr, ('Warning: while converting %s/%s to MSBuild, ' -- '%s' % (msvs_tool_name, msvs_setting, e)) -+ except ValueError as e: -+ print(('Warning: while converting %s/%s to MSBuild, ' -+ '%s' % (msvs_tool_name, msvs_setting, e)), -+ file=stderr) - else: - _ValidateExclusionSetting(msvs_setting, - msvs_tool, -@@ -472,8 +481,8 @@ - (msvs_tool_name, msvs_setting)), - stderr) - else: -- print >> stderr, ('Warning: unrecognized tool %s while converting to ' -- 'MSBuild.' % msvs_tool_name) -+ print(('Warning: unrecognized tool %s while converting to ' -+ 'MSBuild.' % msvs_tool_name), file=stderr) - return msbuild_settings - - -@@ -513,13 +522,13 @@ - for tool_name in settings: - if tool_name in validators: - tool_validators = validators[tool_name] -- for setting, value in settings[tool_name].iteritems(): -+ for setting, value in settings[tool_name].items(): - if setting in tool_validators: - try: - tool_validators[setting](value) -- except ValueError, e: -- print >> stderr, ('Warning: for %s/%s, %s' % -- (tool_name, setting, e)) -+ except ValueError as e: -+ print(('Warning: for %s/%s, %s' % -+ (tool_name, setting, e)), file=stderr) - else: - _ValidateExclusionSetting(setting, - tool_validators, -@@ -528,7 +537,7 @@ - stderr) - - else: -- print >> stderr, ('Warning: unrecognized tool %s' % tool_name) -+ print(('Warning: unrecognized tool %s' % tool_name), file=stderr) - - - # MSVS and MBuild names of the tools. -diff --git a/tools/gyp/pylib/gyp/MSVSSettings_test.py b/pylib/gyp/MSVSSettings_test.py -index bf6ea6b..73ed25e 100755 ---- a/tools/gyp/pylib/gyp/MSVSSettings_test.py -+++ b/tools/gyp/pylib/gyp/MSVSSettings_test.py -@@ -6,7 +6,10 @@ - - """Unit tests for the MSVSSettings.py file.""" - --import StringIO -+try: -+ from StringIO import StringIO -+except ImportError: -+ from io import StringIO - import unittest - import gyp.MSVSSettings as MSVSSettings - -@@ -14,7 +17,7 @@ - class TestSequenceFunctions(unittest.TestCase): - - def setUp(self): -- self.stderr = StringIO.StringIO() -+ self.stderr = StringIO() - - def _ExpectedWarnings(self, expected): - """Compares recorded lines to expected warnings.""" -diff --git a/tools/gyp/pylib/gyp/MSVSUserFile.py b/pylib/gyp/MSVSUserFile.py -index 6c07e9a..2264d64 100644 ---- a/tools/gyp/pylib/gyp/MSVSUserFile.py -+++ b/tools/gyp/pylib/gyp/MSVSUserFile.py -@@ -91,7 +91,7 @@ - - if environment and isinstance(environment, dict): - env_list = ['%s="%s"' % (key, val) -- for (key,val) in environment.iteritems()] -+ for (key,val) in environment.items()] - environment = ' '.join(env_list) - else: - environment = '' -@@ -135,7 +135,7 @@ - def WriteIfChanged(self): - """Writes the user file.""" - configs = ['Configurations'] -- for config, spec in sorted(self.configurations.iteritems()): -+ for config, spec in sorted(self.configurations.items()): - configs.append(spec) - - content = ['VisualStudioUserFile', -diff --git a/tools/gyp/pylib/gyp/MSVSUtil.py b/pylib/gyp/MSVSUtil.py -index 96dea6c..f24530b 100644 ---- a/tools/gyp/pylib/gyp/MSVSUtil.py -+++ b/tools/gyp/pylib/gyp/MSVSUtil.py -@@ -236,7 +236,7 @@ - - # Set up the shim to output its PDB to the same location as the final linker - # target. -- for config_name, config in shim_dict.get('configurations').iteritems(): -+ for config_name, config in shim_dict.get('configurations').items(): - pdb_path = _GetPdbPath(target_dict, config_name, vars) - - # A few keys that we don't want to propagate. -diff --git a/tools/gyp/pylib/gyp/MSVSVersion.py b/pylib/gyp/MSVSVersion.py -index 44b958d..5f316b6 100644 ---- a/tools/gyp/pylib/gyp/MSVSVersion.py -+++ b/tools/gyp/pylib/gyp/MSVSVersion.py -@@ -189,7 +189,7 @@ - text = None - try: - text = _RegistryQueryBase('Sysnative', key, value) -- except OSError, e: -+ except OSError as e: - if e.errno == errno.ENOENT: - text = _RegistryQueryBase('System32', key, value) - else: -@@ -207,12 +207,15 @@ - contents of the registry key's value, or None on failure. Throws - ImportError if _winreg is unavailable. - """ -- import _winreg -+ try: -+ import _winreg as winreg -+ except ImportError: -+ import winreg - try: - root, subkey = key.split('\\', 1) - assert root == 'HKLM' # Only need HKLM for now. -- with _winreg.OpenKey(_winreg.HKEY_LOCAL_MACHINE, subkey) as hkey: -- return _winreg.QueryValueEx(hkey, value)[0] -+ with winreg.OpenKey(winreg.HKEY_LOCAL_MACHINE, subkey) as hkey: -+ return winreg.QueryValueEx(hkey, value)[0] - except WindowsError: - return None - -diff --git a/tools/gyp/pylib/gyp/__init__.py b/pylib/gyp/__init__.py -index 668f38b..e038151 100755 ---- a/tools/gyp/pylib/gyp/__init__.py -+++ b/tools/gyp/pylib/gyp/__init__.py -@@ -4,6 +4,8 @@ - # Use of this source code is governed by a BSD-style license that can be - # found in the LICENSE file. - -+from __future__ import print_function -+ - import copy - import gyp.input - import optparse -@@ -14,6 +16,12 @@ - import traceback - from gyp.common import GypError - -+try: -+ # basestring was removed in python3. -+ basestring -+except NameError: -+ basestring = str -+ - # Default debug modes for GYP - debug = {} - -@@ -22,7 +30,6 @@ - DEBUG_VARIABLES = 'variables' - DEBUG_INCLUDES = 'includes' - -- - def DebugOutput(mode, message, *args): - if 'all' in gyp.debug or mode in gyp.debug: - ctx = ('unknown', 0, 'unknown') -@@ -34,8 +41,8 @@ - pass - if args: - message %= args -- print '%s:%s:%d:%s %s' % (mode.upper(), os.path.basename(ctx[0]), -- ctx[1], ctx[2], message) -+ print('%s:%s:%d:%s %s' % (mode.upper(), os.path.basename(ctx[0]), -+ ctx[1], ctx[2], message)) - - def FindBuildFiles(): - extension = '.gyp' -@@ -207,7 +214,7 @@ - # We always want to ignore the environment when regenerating, to avoid - # duplicate or changed flags in the environment at the time of regeneration. - flags = ['--ignore-environment'] -- for name, metadata in options._regeneration_metadata.iteritems(): -+ for name, metadata in options._regeneration_metadata.items(): - opt = metadata['opt'] - value = getattr(options, name) - value_predicate = metadata['type'] == 'path' and FixPath or Noop -@@ -226,12 +233,13 @@ - (action == 'store_false' and not value)): - flags.append(opt) - elif options.use_environment and env_name: -- print >>sys.stderr, ('Warning: environment regeneration unimplemented ' -+ print(('Warning: environment regeneration unimplemented ' - 'for %s flag %r env_name %r' % (action, opt, -- env_name)) -+ env_name)), -+ file=sys.stderr) - else: -- print >>sys.stderr, ('Warning: regeneration unimplemented for action %r ' -- 'flag %r' % (action, opt)) -+ print(('Warning: regeneration unimplemented for action %r ' -+ 'flag %r' % (action, opt)), file=sys.stderr) - - return flags - -@@ -431,12 +439,11 @@ - for build_file in build_files: - build_file_dir = os.path.abspath(os.path.dirname(build_file)) - build_file_dir_components = build_file_dir.split(os.path.sep) -- components_len = len(build_file_dir_components) -- for index in xrange(components_len - 1, -1, -1): -- if build_file_dir_components[index] == 'src': -+ for component in reversed(build_file_dir_components): -+ if component == 'src': - options.depth = os.path.sep.join(build_file_dir_components) - break -- del build_file_dir_components[index] -+ del build_file_dir_components[-1] - - # If the inner loop found something, break without advancing to another - # build file. -@@ -475,7 +482,7 @@ - if home_dot_gyp != None: - default_include = os.path.join(home_dot_gyp, 'include.gypi') - if os.path.exists(default_include): -- print 'Using overrides found in ' + default_include -+ print('Using overrides found in ' + default_include) - includes.append(default_include) - - # Command-line --include files come after the default include. -@@ -490,7 +497,7 @@ - if options.generator_flags: - gen_flags += options.generator_flags - generator_flags = NameValueListToDict(gen_flags) -- if DEBUG_GENERAL in gyp.debug.keys(): -+ if DEBUG_GENERAL in gyp.debug: - DebugOutput(DEBUG_GENERAL, "generator_flags: %s", generator_flags) - - # Generate all requested formats (use a set in case we got one format request -@@ -523,7 +530,7 @@ - generator.GenerateOutput(flat_list, targets, data, params) - - if options.configs: -- valid_configs = targets[flat_list[0]]['configurations'].keys() -+ valid_configs = targets[flat_list[0]]['configurations'] - for conf in options.configs: - if conf not in valid_configs: - raise GypError('Invalid config specified via --build: %s' % conf) -@@ -536,7 +543,7 @@ - def main(args): - try: - return gyp_main(args) -- except GypError, e: -+ except GypError as e: - sys.stderr.write("gyp: %s\n" % e) - return 1 - -diff --git a/tools/gyp/pylib/gyp/common.py b/pylib/gyp/common.py -index 1b245ec..1823de8 100644 ---- a/tools/gyp/pylib/gyp/common.py -+++ b/tools/gyp/pylib/gyp/common.py -@@ -345,7 +345,7 @@ - prefix=os.path.split(filename)[1] + '.gyp.', - dir=os.path.split(filename)[0]) - try: -- self.tmp_file = os.fdopen(tmp_fd, 'wb') -+ self.tmp_file = os.fdopen(tmp_fd, 'w') - except Exception: - # Don't leave turds behind. - os.unlink(self.tmp_path) -@@ -363,7 +363,7 @@ - same = False - try: - same = filecmp.cmp(self.tmp_path, filename, False) -- except OSError, e: -+ except OSError as e: - if e.errno != errno.ENOENT: - raise - -@@ -382,9 +382,9 @@ - # - # No way to get the umask without setting a new one? Set a safe one - # and then set it back to the old value. -- umask = os.umask(077) -+ umask = os.umask(0o77) - os.umask(umask) -- os.chmod(self.tmp_path, 0666 & ~umask) -+ os.chmod(self.tmp_path, 0o666 & ~umask) - if sys.platform == 'win32' and os.path.exists(filename): - # NOTE: on windows (but not cygwin) rename will not replace an - # existing file, so it must be preceded with a remove. Sadly there -@@ -471,7 +471,7 @@ - ''.join([source[0], header] + source[1:])) - - # Make file executable. -- os.chmod(tool_path, 0755) -+ os.chmod(tool_path, 0o755) - - - # From Alex Martelli, -diff --git a/tools/gyp/pylib/gyp/common_test.py b/pylib/gyp/common_test.py -index ad6f9a1..0b8ada3 100755 ---- a/tools/gyp/pylib/gyp/common_test.py -+++ b/tools/gyp/pylib/gyp/common_test.py -@@ -63,6 +63,7 @@ - self.assertFlavor('solaris', 'sunos' , {}); - self.assertFlavor('linux' , 'linux2' , {}); - self.assertFlavor('linux' , 'linux3' , {}); -+ self.assertFlavor('linux' , 'linux' , {}); - - def test_param(self): - self.assertFlavor('foobar', 'linux2' , {'flavor': 'foobar'}) -diff --git a/tools/gyp/pylib/gyp/easy_xml.py b/pylib/gyp/easy_xml.py -index 2522efb..15c6651 100644 ---- a/tools/gyp/pylib/gyp/easy_xml.py -+++ b/tools/gyp/pylib/gyp/easy_xml.py -@@ -6,6 +6,11 @@ - import os - import locale - -+try: -+ # reduce moved to functools in python3. -+ reduce -+except NameError: -+ from functools import reduce - - def XmlToString(content, encoding='utf-8', pretty=False): - """ Writes the XML content to disk, touching the file only if it has changed. -@@ -80,7 +85,7 @@ - # Optionally in second position is a dictionary of the attributes. - rest = specification[1:] - if rest and isinstance(rest[0], dict): -- for at, val in sorted(rest[0].iteritems()): -+ for at, val in sorted(rest[0].items()): - xml_parts.append(' %s="%s"' % (at, _XmlEscape(val, attr=True))) - rest = rest[1:] - if rest: -diff --git a/tools/gyp/pylib/gyp/easy_xml_test.py b/pylib/gyp/easy_xml_test.py -index df64354..a1fdb18 100755 ---- a/tools/gyp/pylib/gyp/easy_xml_test.py -+++ b/tools/gyp/pylib/gyp/easy_xml_test.py -@@ -8,13 +8,16 @@ - - import gyp.easy_xml as easy_xml - import unittest --import StringIO -+try: -+ from StringIO import StringIO -+except ImportError: -+ from io import StringIO - - - class TestSequenceFunctions(unittest.TestCase): - - def setUp(self): -- self.stderr = StringIO.StringIO() -+ self.stderr = StringIO() - - def test_EasyXml_simple(self): - self.assertEqual( -diff --git a/tools/gyp/pylib/gyp/flock_tool.py b/pylib/gyp/flock_tool.py -index b38d866..81fb79d 100755 ---- a/tools/gyp/pylib/gyp/flock_tool.py -+++ b/tools/gyp/pylib/gyp/flock_tool.py -@@ -39,7 +39,7 @@ - # where fcntl.flock(fd, LOCK_EX) always fails - # with EBADF, that's why we use this F_SETLK - # hack instead. -- fd = os.open(lockfile, os.O_WRONLY|os.O_NOCTTY|os.O_CREAT, 0666) -+ fd = os.open(lockfile, os.O_WRONLY|os.O_NOCTTY|os.O_CREAT, 0o666) - if sys.platform.startswith('aix'): - # Python on AIX is compiled with LARGEFILE support, which changes the - # struct size. -diff --git a/tools/gyp/pylib/gyp/generator/analyzer.py b/pylib/gyp/generator/analyzer.py -index 921c1a6..b3484dc 100644 ---- a/tools/gyp/pylib/gyp/generator/analyzer.py -+++ b/tools/gyp/pylib/gyp/generator/analyzer.py -@@ -62,6 +62,8 @@ - then the "all" target includes "b1" and "b2". - """ - -+from __future__ import print_function -+ - import gyp.common - import gyp.ninja_syntax as ninja_syntax - import json -@@ -155,7 +157,7 @@ - continue - result.append(base_path + source) - if debug: -- print 'AddSource', org_source, result[len(result) - 1] -+ print('AddSource', org_source, result[len(result) - 1]) - - - def _ExtractSourcesFromAction(action, base_path, base_path_components, -@@ -185,7 +187,7 @@ - base_path += '/' - - if debug: -- print 'ExtractSources', target, base_path -+ print('ExtractSources', target, base_path) - - results = [] - if 'sources' in target_dict: -@@ -278,7 +280,7 @@ - the root of the source tree.""" - if _ToLocalPath(toplevel_dir, _ToGypPath(build_file)) in files: - if debug: -- print 'gyp file modified', build_file -+ print('gyp file modified', build_file) - return True - - # First element of included_files is the file itself. -@@ -291,8 +293,8 @@ - _ToGypPath(gyp.common.UnrelativePath(include_file, build_file)) - if _ToLocalPath(toplevel_dir, rel_include_file) in files: - if debug: -- print 'included gyp file modified, gyp_file=', build_file, \ -- 'included file=', rel_include_file -+ print('included gyp file modified, gyp_file=', build_file, \ -+ 'included file=', rel_include_file) - return True - return False - -@@ -373,7 +375,7 @@ - # If a build file (or any of its included files) is modified we assume all - # targets in the file are modified. - if build_file_in_files[build_file]: -- print 'matching target from modified build file', target_name -+ print('matching target from modified build file', target_name) - target.match_status = MATCH_STATUS_MATCHES - matching_targets.append(target) - else: -@@ -381,7 +383,7 @@ - toplevel_dir) - for source in sources: - if _ToGypPath(os.path.normpath(source)) in files: -- print 'target', target_name, 'matches', source -+ print('target', target_name, 'matches', source) - target.match_status = MATCH_STATUS_MATCHES - matching_targets.append(target) - break -@@ -433,7 +435,7 @@ - for dep in target.deps: - if _DoesTargetDependOnMatchingTargets(dep): - target.match_status = MATCH_STATUS_MATCHES_BY_DEPENDENCY -- print '\t', target.name, 'matches by dep', dep.name -+ print('\t', target.name, 'matches by dep', dep.name) - return True - target.match_status = MATCH_STATUS_DOESNT_MATCH - return False -@@ -445,7 +447,7 @@ - supplied as input to analyzer. - possible_targets: targets to search from.""" - found = [] -- print 'Targets that matched by dependency:' -+ print('Targets that matched by dependency:') - for target in possible_targets: - if _DoesTargetDependOnMatchingTargets(target): - found.append(target) -@@ -484,12 +486,13 @@ - (add_if_no_ancestor or target.requires_build)) or - (target.is_static_library and add_if_no_ancestor and - not target.is_or_has_linked_ancestor)): -- print '\t\tadding to compile targets', target.name, 'executable', \ -- target.is_executable, 'added_to_compile_targets', \ -- target.added_to_compile_targets, 'add_if_no_ancestor', \ -- add_if_no_ancestor, 'requires_build', target.requires_build, \ -- 'is_static_library', target.is_static_library, \ -+ print('\t\tadding to compile targets', target.name, 'executable', -+ target.is_executable, 'added_to_compile_targets', -+ target.added_to_compile_targets, 'add_if_no_ancestor', -+ add_if_no_ancestor, 'requires_build', target.requires_build, -+ 'is_static_library', target.is_static_library, - 'is_or_has_linked_ancestor', target.is_or_has_linked_ancestor -+ ) - result.add(target) - target.added_to_compile_targets = True - -@@ -500,7 +503,7 @@ - supplied_targets: set of targets supplied to analyzer to search from.""" - result = set() - for target in matching_targets: -- print 'finding compile targets for match', target.name -+ print('finding compile targets for match', target.name) - _AddCompileTargets(target, supplied_targets, True, result) - return result - -@@ -508,46 +511,46 @@ - def _WriteOutput(params, **values): - """Writes the output, either to stdout or a file is specified.""" - if 'error' in values: -- print 'Error:', values['error'] -+ print('Error:', values['error']) - if 'status' in values: -- print values['status'] -+ print(values['status']) - if 'targets' in values: - values['targets'].sort() -- print 'Supplied targets that depend on changed files:' -+ print('Supplied targets that depend on changed files:') - for target in values['targets']: -- print '\t', target -+ print('\t', target) - if 'invalid_targets' in values: - values['invalid_targets'].sort() -- print 'The following targets were not found:' -+ print('The following targets were not found:') - for target in values['invalid_targets']: -- print '\t', target -+ print('\t', target) - if 'build_targets' in values: - values['build_targets'].sort() -- print 'Targets that require a build:' -+ print('Targets that require a build:') - for target in values['build_targets']: -- print '\t', target -+ print('\t', target) - if 'compile_targets' in values: - values['compile_targets'].sort() -- print 'Targets that need to be built:' -+ print('Targets that need to be built:') - for target in values['compile_targets']: -- print '\t', target -+ print('\t', target) - if 'test_targets' in values: - values['test_targets'].sort() -- print 'Test targets:' -+ print('Test targets:') - for target in values['test_targets']: -- print '\t', target -+ print('\t', target) - - output_path = params.get('generator_flags', {}).get( - 'analyzer_output_path', None) - if not output_path: -- print json.dumps(values) -+ print(json.dumps(values)) - return - try: - f = open(output_path, 'w') - f.write(json.dumps(values) + '\n') - f.close() - except IOError as e: -- print 'Error writing to output file', output_path, str(e) -+ print('Error writing to output file', output_path, str(e)) - - - def _WasGypIncludeFileModified(params, files): -@@ -556,7 +559,7 @@ - if params['options'].includes: - for include in params['options'].includes: - if _ToGypPath(os.path.normpath(include)) in files: -- print 'Include file modified, assuming all changed', include -+ print('Include file modified, assuming all changed', include) - return True - return False - -@@ -638,13 +641,13 @@ - set(self._root_targets))] - else: - test_targets = [x for x in test_targets_no_all] -- print 'supplied test_targets' -+ print('supplied test_targets') - for target_name in self._test_target_names: -- print '\t', target_name -- print 'found test_targets' -+ print('\t', target_name) -+ print('found test_targets') - for target in test_targets: -- print '\t', target.name -- print 'searching for matching test targets' -+ print('\t', target.name) -+ print('searching for matching test targets') - matching_test_targets = _GetTargetsDependingOnMatchingTargets(test_targets) - matching_test_targets_contains_all = (test_target_names_contains_all and - set(matching_test_targets) & -@@ -654,14 +657,14 @@ - # 'all' is subsequentely added to the matching names below. - matching_test_targets = [x for x in (set(matching_test_targets) & - set(test_targets_no_all))] -- print 'matched test_targets' -+ print('matched test_targets') - for target in matching_test_targets: -- print '\t', target.name -+ print('\t', target.name) - matching_target_names = [gyp.common.ParseQualifiedTarget(target.name)[1] - for target in matching_test_targets] - if matching_test_targets_contains_all: - matching_target_names.append('all') -- print '\tall' -+ print('\tall') - return matching_target_names - - def find_matching_compile_target_names(self): -@@ -669,7 +672,7 @@ - assert self.is_build_impacted(); - # Compile targets are found by searching up from changed targets. - # Reset the visited status for _GetBuildTargets. -- for target in self._name_to_target.itervalues(): -+ for target in self._name_to_target.values(): - target.visited = False - - supplied_targets = _LookupTargets(self._supplied_target_names_no_all(), -@@ -677,10 +680,10 @@ - if 'all' in self._supplied_target_names(): - supplied_targets = [x for x in (set(supplied_targets) | - set(self._root_targets))] -- print 'Supplied test_targets & compile_targets' -+ print('Supplied test_targets & compile_targets') - for target in supplied_targets: -- print '\t', target.name -- print 'Finding compile targets' -+ print('\t', target.name) -+ print('Finding compile targets') - compile_targets = _GetCompileTargets(self._changed_targets, - supplied_targets) - return [gyp.common.ParseQualifiedTarget(target.name)[1] -@@ -699,7 +702,7 @@ - - toplevel_dir = _ToGypPath(os.path.abspath(params['options'].toplevel_dir)) - if debug: -- print 'toplevel_dir', toplevel_dir -+ print('toplevel_dir', toplevel_dir) - - if _WasGypIncludeFileModified(params, config.files): - result_dict = { 'status': all_changed_string, -diff --git a/tools/gyp/pylib/gyp/generator/cmake.py b/pylib/gyp/generator/cmake.py -index a2b9629..4a2041c 100644 ---- a/tools/gyp/pylib/gyp/generator/cmake.py -+++ b/tools/gyp/pylib/gyp/generator/cmake.py -@@ -28,6 +28,8 @@ - CMakeLists.txt file. - """ - -+from __future__ import print_function -+ - import multiprocessing - import os - import signal -@@ -36,6 +38,12 @@ - import gyp.common - import gyp.xcode_emulation - -+try: -+ # maketrans moved to str in python3. -+ _maketrans = string.maketrans -+except NameError: -+ _maketrans = str.maketrans -+ - generator_default_variables = { - 'EXECUTABLE_PREFIX': '', - 'EXECUTABLE_SUFFIX': '', -@@ -238,7 +246,7 @@ - Invalid for make: ':' - Invalid for unknown reasons but cause failures: '.' - """ -- return a.translate(string.maketrans(' /():."', '_______')) -+ return a.translate(_maketrans(' /():."', '_______')) - - - def WriteActions(target_name, actions, extra_sources, extra_deps, -@@ -644,8 +652,8 @@ - - cmake_target_type = cmake_target_type_from_gyp_target_type.get(target_type) - if cmake_target_type is None: -- print ('Target %s has unknown target type %s, skipping.' % -- ( target_name, target_type ) ) -+ print('Target %s has unknown target type %s, skipping.' % -+ ( target_name, target_type )) - return - - SetVariable(output, 'TARGET', target_name) -@@ -868,8 +876,8 @@ - default_product_ext = generator_default_variables['SHARED_LIB_SUFFIX'] - - elif target_type != 'executable': -- print ('ERROR: What output file should be generated?', -- 'type', target_type, 'target', target_name) -+ print(('ERROR: What output file should be generated?', -+ 'type', target_type, 'target', target_name)) - - product_prefix = spec.get('product_prefix', default_product_prefix) - product_name = spec.get('product_name', default_product_name) -@@ -1207,11 +1215,11 @@ - output_dir, - config_name)) - arguments = ['cmake', '-G', 'Ninja'] -- print 'Generating [%s]: %s' % (config_name, arguments) -+ print('Generating [%s]: %s' % (config_name, arguments)) - subprocess.check_call(arguments, cwd=build_dir) - - arguments = ['ninja', '-C', build_dir] -- print 'Building [%s]: %s' % (config_name, arguments) -+ print('Building [%s]: %s' % (config_name, arguments)) - subprocess.check_call(arguments) - - -@@ -1230,7 +1238,7 @@ - GenerateOutputForConfig(target_list, target_dicts, data, - params, user_config) - else: -- config_names = target_dicts[target_list[0]]['configurations'].keys() -+ config_names = target_dicts[target_list[0]]['configurations'] - if params['parallel']: - try: - pool = multiprocessing.Pool(len(config_names)) -@@ -1239,7 +1247,7 @@ - arglists.append((target_list, target_dicts, data, - params, config_name)) - pool.map(CallGenerateOutputForConfig, arglists) -- except KeyboardInterrupt, e: -+ except KeyboardInterrupt as e: - pool.terminate() - raise e - else: -diff --git a/tools/gyp/pylib/gyp/generator/dump_dependency_json.py b/pylib/gyp/generator/dump_dependency_json.py -index 160eafe..2bf3f39 100644 ---- a/tools/gyp/pylib/gyp/generator/dump_dependency_json.py -+++ b/tools/gyp/pylib/gyp/generator/dump_dependency_json.py -@@ -2,6 +2,8 @@ - # Use of this source code is governed by a BSD-style license that can be - # found in the LICENSE file. - -+from __future__ import print_function -+ - import collections - import os - import gyp -@@ -96,4 +98,4 @@ - f = open(filename, 'w') - json.dump(edges, f) - f.close() -- print 'Wrote json to %s.' % filename -+ print('Wrote json to %s.' % filename) -diff --git a/tools/gyp/pylib/gyp/generator/eclipse.py b/pylib/gyp/generator/eclipse.py -index 3544347..d039f03 100644 ---- a/tools/gyp/pylib/gyp/generator/eclipse.py -+++ b/tools/gyp/pylib/gyp/generator/eclipse.py -@@ -141,7 +141,7 @@ - compiler_includes_list.append(include_dir) - - # Find standard gyp include dirs. -- if config.has_key('include_dirs'): -+ if 'include_dirs' in config: - include_dirs = config['include_dirs'] - for shared_intermediate_dir in shared_intermediate_dirs: - for include_dir in include_dirs: -@@ -272,7 +272,7 @@ - out.write(' <language name="holder for library settings"></language>\n') - for lang in eclipse_langs: - out.write(' <language name="%s">\n' % lang) -- for key in sorted(defines.iterkeys()): -+ for key in sorted(defines.keys()): - out.write(' <macro><name>%s</name><value>%s</value></macro>\n' % - (escape(key), escape(defines[key]))) - out.write(' </language>\n') -@@ -418,7 +418,7 @@ - GenerateOutputForConfig(target_list, target_dicts, data, params, - user_config) - else: -- config_names = target_dicts[target_list[0]]['configurations'].keys() -+ config_names = target_dicts[target_list[0]]['configurations'] - for config_name in config_names: - GenerateOutputForConfig(target_list, target_dicts, data, params, - config_name) -diff --git a/tools/gyp/pylib/gyp/generator/gypd.py b/pylib/gyp/generator/gypd.py -index 3efdb99..78eeaa6 100644 ---- a/tools/gyp/pylib/gyp/generator/gypd.py -+++ b/tools/gyp/pylib/gyp/generator/gypd.py -@@ -88,7 +88,7 @@ - if not output_file in output_files: - output_files[output_file] = input_file - -- for output_file, input_file in output_files.iteritems(): -+ for output_file, input_file in output_files.items(): - output = open(output_file, 'w') - pprint.pprint(data[input_file], output) - output.close() -diff --git a/tools/gyp/pylib/gyp/generator/make.py b/pylib/gyp/generator/make.py -index fb4f918..2057e3a 100644 ---- a/tools/gyp/pylib/gyp/generator/make.py -+++ b/tools/gyp/pylib/gyp/generator/make.py -@@ -21,6 +21,8 @@ - # toplevel Makefile. It may make sense to generate some .mk files on - # the side to keep the the files readable. - -+from __future__ import print_function -+ - import os - import re - import sys -@@ -668,7 +670,7 @@ - basenames.setdefault(basename, []).append(source) - - error = '' -- for basename, files in basenames.iteritems(): -+ for basename, files in basenames.items(): - if len(files) > 1: - error += ' %s: %s\n' % (basename, ' '.join(files)) - -@@ -816,7 +818,7 @@ - gyp.xcode_emulation.MacPrefixHeader( - self.xcode_settings, lambda p: Sourceify(self.Absolutify(p)), - self.Pchify)) -- sources = filter(Compilable, all_sources) -+ sources = [x for x in all_sources if Compilable(x)] - if sources: - self.WriteLn(SHARED_HEADER_SUFFIX_RULES_COMMENT1) - extensions = set([os.path.splitext(s)[1] for s in sources]) -@@ -945,7 +947,7 @@ - '%s%s' - % (name, cd_action, command)) - self.WriteLn() -- outputs = map(self.Absolutify, outputs) -+ outputs = [self.Absolutify(o) for o in outputs] - # The makefile rules are all relative to the top dir, but the gyp actions - # are defined relative to their containing dir. This replaces the obj - # variable for the action rule with an absolute version so that the output -@@ -1035,7 +1037,7 @@ - outputs = [gyp.xcode_emulation.ExpandEnvVars(o, env) for o in outputs] - inputs = [gyp.xcode_emulation.ExpandEnvVars(i, env) for i in inputs] - -- outputs = map(self.Absolutify, outputs) -+ outputs = [self.Absolutify(o) for o in outputs] - all_outputs += outputs - # Only write the 'obj' and 'builddir' rules for the "primary" output - # (:1); it's superfluous for the "extra outputs", and this avoids -@@ -1233,11 +1235,11 @@ - self.WriteList(cflags_objcc, 'CFLAGS_OBJCC_%s' % configname) - includes = config.get('include_dirs') - if includes: -- includes = map(Sourceify, map(self.Absolutify, includes)) -+ includes = [Sourceify(self.Absolutify(include)) for include in includes] - self.WriteList(includes, 'INCS_%s' % configname, prefix='-I') - - compilable = filter(Compilable, sources) -- objs = map(self.Objectify, map(self.Absolutify, map(Target, compilable))) -+ objs = [self.Objectify(self.Absolutify(Target(x))) for x in compilable] - self.WriteList(objs, 'OBJS') - - for obj in objs: -@@ -1309,7 +1311,7 @@ - - # If there are any object files in our input file list, link them into our - # output. -- extra_link_deps += filter(Linkable, sources) -+ extra_link_deps += [source for source in sources if Linkable(source)] - - self.WriteLn() - -@@ -1377,8 +1379,8 @@ - elif self.type == 'none': - target = '%s.stamp' % target - elif self.type != 'executable': -- print ("ERROR: What output file should be generated?", -- "type", self.type, "target", target) -+ print(("ERROR: What output file should be generated?", -+ "type", self.type, "target", target)) - - target_prefix = spec.get('product_prefix', target_prefix) - target = spec.get('product_name', target) -@@ -1542,9 +1544,9 @@ - # Postbuilds expect to be run in the gyp file's directory, so insert an - # implicit postbuild to cd to there. - postbuilds.insert(0, gyp.common.EncodePOSIXShellList(['cd', self.path])) -- for i in xrange(len(postbuilds)): -- if not postbuilds[i].startswith('$'): -- postbuilds[i] = EscapeShellArgument(postbuilds[i]) -+ for i, postbuild in enumerate(postbuilds): -+ if not postbuild.startswith('$'): -+ postbuilds[i] = EscapeShellArgument(postbuild) - self.WriteLn('%s: builddir := $(abs_builddir)' % QuoteSpaces(self.output)) - self.WriteLn('%s: POSTBUILDS := %s' % ( - QuoteSpaces(self.output), ' '.join(postbuilds))) -@@ -1634,7 +1636,7 @@ - self.WriteDoCmd([self.output_binary], deps, 'touch', part_of_all, - postbuilds=postbuilds) - else: -- print "WARNING: no output for", self.type, target -+ print("WARNING: no output for", self.type, target) - - # Add an alias for each target (if there are any outputs). - # Installable target aliases are created below. -@@ -1741,7 +1743,7 @@ - output is just a name to run the rule - command: (optional) command name to generate unambiguous labels - """ -- outputs = map(QuoteSpaces, outputs) -+ outputs = [QuoteSpaces(o) for o in outputs] - inputs = map(QuoteSpaces, inputs) - - if comment: -@@ -1986,7 +1988,7 @@ - if options.toplevel_dir and options.toplevel_dir != '.': - arguments += '-C', options.toplevel_dir - arguments.append('BUILDTYPE=' + config) -- print 'Building [%s]: %s' % (config, arguments) -+ print('Building [%s]: %s' % (config, arguments)) - subprocess.check_call(arguments) - - -diff --git a/tools/gyp/pylib/gyp/generator/msvs.py b/pylib/gyp/generator/msvs.py -index 8fe9e5a..e8a2b36 100644 ---- a/tools/gyp/pylib/gyp/generator/msvs.py -+++ b/tools/gyp/pylib/gyp/generator/msvs.py -@@ -2,6 +2,9 @@ - # Use of this source code is governed by a BSD-style license that can be - # found in the LICENSE file. - -+from __future__ import print_function -+ -+import collections - import copy - import ntpath - import os -@@ -23,16 +26,6 @@ - from gyp.common import GypError - from gyp.common import OrderedSet - --# TODO: Remove once bots are on 2.7, http://crbug.com/241769 --def _import_OrderedDict(): -- import collections -- try: -- return collections.OrderedDict -- except AttributeError: -- import gyp.ordered_dict -- return gyp.ordered_dict.OrderedDict --OrderedDict = _import_OrderedDict() -- - - # Regular expression for validating Visual Studio GUIDs. If the GUID - # contains lowercase hex letters, MSVS will be fine. However, -@@ -202,7 +195,7 @@ - if not prefix: prefix = [] - result = [] - excluded_result = [] -- folders = OrderedDict() -+ folders = collections.OrderedDict() - # Gather files into the final result, excluded, or folders. - for s in sources: - if len(s) == 1: -@@ -469,7 +462,7 @@ - 'CommandLine': cmd, - }) - # Add to the properties of primary input for each config. -- for config_name, c_data in spec['configurations'].iteritems(): -+ for config_name, c_data in spec['configurations'].items(): - p.AddFileConfig(_FixPath(primary_input), - _ConfigFullName(config_name, c_data), tools=[tool]) - -@@ -775,8 +768,8 @@ - # the VCProj but cause the same problem on the final command-line. Moving - # the item to the end of the list does works, but that's only possible if - # there's only one such item. Let's just warn the user. -- print >> sys.stderr, ('Warning: MSVS may misinterpret the odd number of ' + -- 'quotes in ' + s) -+ print(('Warning: MSVS may misinterpret the odd number of ' + -+ 'quotes in ' + s), file=sys.stderr) - return s - - -@@ -991,7 +984,7 @@ - basenames.setdefault(basename, []).append(source) - - error = '' -- for basename, files in basenames.iteritems(): -+ for basename, files in basenames.items(): - if len(files) > 1: - error += ' %s: %s\n' % (basename, ' '.join(files)) - -@@ -1023,7 +1016,7 @@ - relative_path_of_gyp_file = gyp.common.RelativePath(gyp_path, project_dir) - - config_type = _GetMSVSConfigurationType(spec, project.build_file) -- for config_name, config in spec['configurations'].iteritems(): -+ for config_name, config in spec['configurations'].items(): - _AddConfigurationToMSVSProject(p, spec, config_type, config_name, config) - - # MSVC08 and prior version cannot handle duplicate basenames in the same -@@ -1392,10 +1385,10 @@ - A list of Tool objects. - """ - tool_list = [] -- for tool, settings in tools.iteritems(): -+ for tool, settings in tools.items(): - # Collapse settings with lists. - settings_fixed = {} -- for setting, value in settings.iteritems(): -+ for setting, value in settings.items(): - if type(value) == list: - if ((tool == 'VCLinkerTool' and - setting == 'AdditionalDependencies') or -@@ -1570,7 +1563,7 @@ - def _GetPrecompileRelatedFiles(spec): - # Gather a list of precompiled header related sources. - precompiled_related = [] -- for _, config in spec['configurations'].iteritems(): -+ for _, config in spec['configurations'].items(): - for k in precomp_keys: - f = config.get(k) - if f: -@@ -1581,7 +1574,7 @@ - def _ExcludeFilesFromBeingBuilt(p, spec, excluded_sources, excluded_idl, - list_excluded): - exclusions = _GetExcludedFilesFromBuild(spec, excluded_sources, excluded_idl) -- for file_name, excluded_configs in exclusions.iteritems(): -+ for file_name, excluded_configs in exclusions.items(): - if (not list_excluded and - len(excluded_configs) == len(spec['configurations'])): - # If we're not listing excluded files, then they won't appear in the -@@ -1598,7 +1591,7 @@ - # Exclude excluded sources from being built. - for f in excluded_sources: - excluded_configs = [] -- for config_name, config in spec['configurations'].iteritems(): -+ for config_name, config in spec['configurations'].items(): - precomped = [_FixPath(config.get(i, '')) for i in precomp_keys] - # Don't do this for ones that are precompiled header related. - if f not in precomped: -@@ -1608,7 +1601,7 @@ - # Exclude them now. - for f in excluded_idl: - excluded_configs = [] -- for config_name, config in spec['configurations'].iteritems(): -+ for config_name, config in spec['configurations'].items(): - excluded_configs.append((config_name, config)) - exclusions[f] = excluded_configs - return exclusions -@@ -1617,7 +1610,7 @@ - def _AddToolFilesToMSVS(p, spec): - # Add in tool files (rules). - tool_files = OrderedSet() -- for _, config in spec['configurations'].iteritems(): -+ for _, config in spec['configurations'].items(): - for f in config.get('msvs_tool_files', []): - tool_files.add(f) - for f in tool_files: -@@ -1630,7 +1623,7 @@ - # kind (i.e. C vs. C++) as the precompiled header source stub needs - # to have use of precompiled headers disabled. - extensions_excluded_from_precompile = [] -- for config_name, config in spec['configurations'].iteritems(): -+ for config_name, config in spec['configurations'].items(): - source = config.get('msvs_precompiled_source') - if source: - source = _FixPath(source) -@@ -1651,7 +1644,7 @@ - else: - basename, extension = os.path.splitext(source) - if extension in extensions_excluded_from_precompile: -- for config_name, config in spec['configurations'].iteritems(): -+ for config_name, config in spec['configurations'].items(): - tool = MSVSProject.Tool('VCCLCompilerTool', - {'UsePrecompiledHeader': '0', - 'ForcedIncludeFiles': '$(NOINHERIT)'}) -@@ -1702,7 +1695,7 @@ - return # Nothing to add - # Write out the user file. - user_file = _CreateMSVSUserFile(project_path, version, spec) -- for config_name, c_data in spec['configurations'].iteritems(): -+ for config_name, c_data in spec['configurations'].items(): - user_file.AddDebugSettings(_ConfigFullName(config_name, c_data), - action, environment, working_directory) - user_file.WriteIfChanged() -@@ -1756,7 +1749,7 @@ - def _DictsToFolders(base_path, bucket, flat): - # Convert to folders recursively. - children = [] -- for folder, contents in bucket.iteritems(): -+ for folder, contents in bucket.items(): - if type(contents) == dict: - folder_children = _DictsToFolders(os.path.join(base_path, folder), - contents, flat) -@@ -1778,8 +1771,8 @@ - # such projects up one level. - if (type(node) == dict and - len(node) == 1 and -- node.keys()[0] == parent + '.vcproj'): -- return node[node.keys()[0]] -+ next(iter(node)) == parent + '.vcproj'): -+ return node[next(iter(node))] - if type(node) != dict: - return node - for child in node: -@@ -1798,8 +1791,8 @@ - # Walk down from the top until we hit a folder that has more than one entry. - # In practice, this strips the top-level "src/" dir from the hierarchy in - # the solution. -- while len(root) == 1 and type(root[root.keys()[0]]) == dict: -- root = root[root.keys()[0]] -+ while len(root) == 1 and type(root[next(iter(root))]) == dict: -+ root = root[next(iter(root))] - # Collapse singles. - root = _CollapseSingles('', root) - # Merge buckets until everything is a root entry. -@@ -1828,7 +1821,7 @@ - # Prepare a dict indicating which project configurations are used for which - # solution configurations for this target. - config_platform_overrides = {} -- for config_name, c in spec['configurations'].iteritems(): -+ for config_name, c in spec['configurations'].items(): - config_fullname = _ConfigFullName(config_name, c) - platform = c.get('msvs_target_platform', _ConfigPlatform(c)) - fixed_config_fullname = '%s|%s' % ( -@@ -1967,7 +1960,7 @@ - msvs_version = params['msvs_version'] - devenv = os.path.join(msvs_version.path, 'Common7', 'IDE', 'devenv.com') - -- for build_file, build_file_dict in data.iteritems(): -+ for build_file, build_file_dict in data.items(): - (build_file_root, build_file_ext) = os.path.splitext(build_file) - if build_file_ext != '.gyp': - continue -@@ -1977,7 +1970,7 @@ - - for config in configurations: - arguments = [devenv, sln_path, '/Build', config] -- print 'Building [%s]: %s' % (config, arguments) -+ print('Building [%s]: %s' % (config, arguments)) - rtn = subprocess.check_call(arguments) - - -@@ -2029,7 +2022,7 @@ - configs = set() - for qualified_target in target_list: - spec = target_dicts[qualified_target] -- for config_name, config in spec['configurations'].iteritems(): -+ for config_name, config in spec['configurations'].items(): - configs.add(_ConfigFullName(config_name, config)) - configs = list(configs) - -@@ -2072,7 +2065,7 @@ - if generator_flags.get('msvs_error_on_missing_sources', False): - raise GypError(error_message) - else: -- print >> sys.stdout, "Warning: " + error_message -+ print("Warning: " + error_message, file=sys.stdout) - - - def _GenerateMSBuildFiltersFile(filters_path, source_files, -@@ -2669,7 +2662,7 @@ - - def _GetMSBuildProjectConfigurations(configurations): - group = ['ItemGroup', {'Label': 'ProjectConfigurations'}] -- for (name, settings) in sorted(configurations.iteritems()): -+ for (name, settings) in sorted(configurations.items()): - configuration, platform = _GetConfigurationAndPlatform(name, settings) - designation = '%s|%s' % (configuration, platform) - group.append( -@@ -2742,7 +2735,7 @@ - - def _GetMSBuildConfigurationDetails(spec, build_file): - properties = {} -- for name, settings in spec['configurations'].iteritems(): -+ for name, settings in spec['configurations'].items(): - msbuild_attributes = _GetMSBuildAttributes(spec, settings, build_file) - condition = _GetConfigurationCondition(name, settings) - character_set = msbuild_attributes.get('CharacterSet') -@@ -2776,9 +2769,9 @@ - user_props = r'$(UserRootDir)\Microsoft.Cpp.$(Platform).user.props' - additional_props = {} - props_specified = False -- for name, settings in sorted(configurations.iteritems()): -+ for name, settings in sorted(configurations.items()): - configuration = _GetConfigurationCondition(name, settings) -- if settings.has_key('msbuild_props'): -+ if 'msbuild_props' in settings: - additional_props[configuration] = _FixPaths(settings['msbuild_props']) - props_specified = True - else: -@@ -2798,7 +2791,7 @@ - ] - else: - sheets = [] -- for condition, props in additional_props.iteritems(): -+ for condition, props in additional_props.items(): - import_group = [ - 'ImportGroup', - {'Label': 'PropertySheets', -@@ -2831,7 +2824,7 @@ - elif a == 'ConfigurationType': - msbuild_attributes[a] = _ConvertMSVSConfigurationType(msvs_attributes[a]) - else: -- print 'Warning: Do not know how to convert MSVS attribute ' + a -+ print('Warning: Do not know how to convert MSVS attribute ' + a) - return msbuild_attributes - - -@@ -2927,7 +2920,7 @@ - new_paths = '$(ExecutablePath);' + ';'.join(new_paths) - - properties = {} -- for (name, configuration) in sorted(configurations.iteritems()): -+ for (name, configuration) in sorted(configurations.items()): - condition = _GetConfigurationCondition(name, configuration) - attributes = _GetMSBuildAttributes(spec, configuration, build_file) - msbuild_settings = configuration['finalized_msbuild_settings'] -@@ -2952,7 +2945,7 @@ - _AddConditionalProperty(properties, condition, 'ExecutablePath', - new_paths) - tool_settings = msbuild_settings.get('', {}) -- for name, value in sorted(tool_settings.iteritems()): -+ for name, value in sorted(tool_settings.items()): - formatted_value = _GetValueFormattedForMSBuild('', name, value) - _AddConditionalProperty(properties, condition, name, formatted_value) - return _GetMSBuildPropertyGroup(spec, None, properties) -@@ -3021,7 +3014,7 @@ - # NOTE: reverse(topsort(DAG)) = topsort(reverse_edges(DAG)) - for name in reversed(properties_ordered): - values = properties[name] -- for value, conditions in sorted(values.iteritems()): -+ for value, conditions in sorted(values.items()): - if len(conditions) == num_configurations: - # If the value is the same all configurations, - # just add one unconditional entry. -@@ -3034,18 +3027,18 @@ - - def _GetMSBuildToolSettingsSections(spec, configurations): - groups = [] -- for (name, configuration) in sorted(configurations.iteritems()): -+ for (name, configuration) in sorted(configurations.items()): - msbuild_settings = configuration['finalized_msbuild_settings'] - group = ['ItemDefinitionGroup', - {'Condition': _GetConfigurationCondition(name, configuration)} - ] -- for tool_name, tool_settings in sorted(msbuild_settings.iteritems()): -+ for tool_name, tool_settings in sorted(msbuild_settings.items()): - # Skip the tool named '' which is a holder of global settings handled - # by _GetMSBuildConfigurationGlobalProperties. - if tool_name: - if tool_settings: - tool = [tool_name] -- for name, value in sorted(tool_settings.iteritems()): -+ for name, value in sorted(tool_settings.items()): - formatted_value = _GetValueFormattedForMSBuild(tool_name, name, - value) - tool.append([name, formatted_value]) -@@ -3078,8 +3071,8 @@ - for ignored_setting in ignored_settings: - value = configuration.get(ignored_setting) - if value: -- print ('Warning: The automatic conversion to MSBuild does not handle ' -- '%s. Ignoring setting of %s' % (ignored_setting, str(value))) -+ print('Warning: The automatic conversion to MSBuild does not handle ' -+ '%s. Ignoring setting of %s' % (ignored_setting, str(value))) - - defines = [_EscapeCppDefineForMSBuild(d) for d in defines] - disabled_warnings = _GetDisabledWarnings(configuration) -@@ -3245,7 +3238,7 @@ - {'Condition': condition}, - 'true']) - # Add precompile if needed -- for config_name, configuration in spec['configurations'].iteritems(): -+ for config_name, configuration in spec['configurations'].items(): - precompiled_source = configuration.get('msvs_precompiled_source', '') - if precompiled_source != '': - precompiled_source = _FixPath(precompiled_source) -@@ -3291,7 +3284,7 @@ - ['Project', guid], - ['ReferenceOutputAssembly', 'false'] - ] -- for config in dependency.spec.get('configurations', {}).itervalues(): -+ for config in dependency.spec.get('configurations', {}).values(): - if config.get('msvs_use_library_dependency_inputs', 0): - project_ref.append(['UseLibraryDependencyInputs', 'true']) - break -@@ -3360,7 +3353,7 @@ - extension_to_rule_name) - missing_sources = _VerifySourcesExist(sources, project_dir) - -- for configuration in configurations.itervalues(): -+ for configuration in configurations.values(): - _FinalizeMSBuildSettings(spec, configuration) - - # Add attributes to root element -@@ -3486,7 +3479,7 @@ - """ - sources_handled_by_action = OrderedSet() - actions_spec = [] -- for primary_input, actions in actions_to_add.iteritems(): -+ for primary_input, actions in actions_to_add.items(): - inputs = OrderedSet() - outputs = OrderedSet() - descriptions = [] -diff --git a/tools/gyp/pylib/gyp/generator/msvs_test.py b/pylib/gyp/generator/msvs_test.py -index c0b021d..838d236 100755 ---- a/tools/gyp/pylib/gyp/generator/msvs_test.py -+++ b/tools/gyp/pylib/gyp/generator/msvs_test.py -@@ -7,13 +7,16 @@ - - import gyp.generator.msvs as msvs - import unittest --import StringIO -+try: -+ from StringIO import StringIO -+except ImportError: -+ from io import StringIO - - - class TestSequenceFunctions(unittest.TestCase): - - def setUp(self): -- self.stderr = StringIO.StringIO() -+ self.stderr = StringIO() - - def test_GetLibraries(self): - self.assertEqual( -diff --git a/tools/gyp/pylib/gyp/generator/ninja.py b/pylib/gyp/generator/ninja.py -index 6de87b7..66faabc 100644 ---- a/tools/gyp/pylib/gyp/generator/ninja.py -+++ b/tools/gyp/pylib/gyp/generator/ninja.py -@@ -2,6 +2,8 @@ - # Use of this source code is governed by a BSD-style license that can be - # found in the LICENSE file. - -+from __future__ import print_function -+ - import collections - import copy - import hashlib -@@ -18,7 +20,10 @@ - import gyp.msvs_emulation - import gyp.MSVSUtil as MSVSUtil - import gyp.xcode_emulation --from cStringIO import StringIO -+try: -+ from cStringIO import StringIO -+except ImportError: -+ from io import StringIO - - from gyp.common import GetEnvironFallback - import gyp.ninja_syntax as ninja_syntax -@@ -350,7 +355,7 @@ - - Uses a stamp file if necessary.""" - -- assert targets == filter(None, targets), targets -+ assert targets == [t for t in targets if t], targets - if len(targets) == 0: - assert not order_only - return None -@@ -427,8 +432,8 @@ - compile_depends.append(target.PreCompileInput()) - if target.uses_cpp: - self.target.uses_cpp = True -- actions_depends = filter(None, actions_depends) -- compile_depends = filter(None, compile_depends) -+ actions_depends = [d for d in actions_depends if d] -+ compile_depends = [d for d in compile_depends if d] - actions_depends = self.WriteCollapsedDependencies('actions_depends', - actions_depends) - compile_depends = self.WriteCollapsedDependencies('compile_depends', -@@ -455,8 +460,8 @@ - try: - sources = extra_sources + spec.get('sources', []) - except TypeError: -- print 'extra_sources: ', str(extra_sources) -- print 'spec.get("sources"): ', str(spec.get('sources')) -+ print('extra_sources: ', str(extra_sources)) -+ print('spec.get("sources"): ', str(spec.get('sources'))) - raise - if sources: - if self.flavor == 'mac' and len(self.archs) > 1: -@@ -485,8 +490,9 @@ - if self.flavor != 'mac' or len(self.archs) == 1: - link_deps += [self.GypPathToNinja(o) for o in obj_outputs] - else: -- print "Warning: Actions/rules writing object files don't work with " \ -- "multiarch targets, dropping. (target %s)" % spec['target_name'] -+ print("Warning: Actions/rules writing object files don't work with " \ -+ "multiarch targets, dropping. (target %s)" % -+ spec['target_name']) - elif self.flavor == 'mac' and len(self.archs) > 1: - link_deps = collections.defaultdict(list) - -@@ -838,7 +844,7 @@ - 'XCASSETS_LAUNCH_IMAGE': 'launch-image', - } - settings = self.xcode_settings.xcode_settings[self.config_name] -- for settings_key, arg_name in settings_to_arg.iteritems(): -+ for settings_key, arg_name in settings_to_arg.items(): - value = settings.get(settings_key) - if value: - extra_arguments[arg_name] = value -@@ -1772,7 +1778,7 @@ - - # VS 2015 uses 20% more working set than VS 2013 and can consume all RAM - # on a 64 GB machine. -- mem_limit = max(1, stat.ullTotalPhys / (5 * (2 ** 30))) # total / 5GB -+ mem_limit = max(1, stat.ullTotalPhys // (5 * (2 ** 30))) # total / 5GB - hard_cap = max(1, int(os.environ.get('GYP_LINK_CONCURRENCY_MAX', 2**32))) - return min(mem_limit, hard_cap) - elif sys.platform.startswith('linux'): -@@ -1784,14 +1790,14 @@ - if not match: - continue - # Allow 8Gb per link on Linux because Gold is quite memory hungry -- return max(1, int(match.group(1)) / (8 * (2 ** 20))) -+ return max(1, int(match.group(1)) // (8 * (2 ** 20))) - return 1 - elif sys.platform == 'darwin': - try: - avail_bytes = int(subprocess.check_output(['sysctl', '-n', 'hw.memsize'])) - # A static library debug build of Chromium's unit_tests takes ~2.7GB, so - # 4GB per ld process allows for some more bloat. -- return max(1, avail_bytes / (4 * (2 ** 30))) # total / 4GB -+ return max(1, avail_bytes // (4 * (2 ** 30))) # total / 4GB - except: - return 1 - else: -@@ -1946,7 +1952,7 @@ - wrappers[key[:-len('_wrapper')]] = os.path.join(build_to_root, value) - - # Support wrappers from environment variables too. -- for key, value in os.environ.iteritems(): -+ for key, value in os.environ.items(): - if key.lower().endswith('_wrapper'): - key_prefix = key[:-len('_wrapper')] - key_prefix = re.sub(r'\.HOST$', '.host', key_prefix) -@@ -1966,7 +1972,7 @@ - configs, generator_flags) - cl_paths = gyp.msvs_emulation.GenerateEnvironmentFiles( - toplevel_build, generator_flags, shared_system_includes, OpenOutput) -- for arch, path in sorted(cl_paths.iteritems()): -+ for arch, path in sorted(cl_paths.items()): - if clang_cl: - # If we have selected clang-cl, use that instead. - path = clang_cl -@@ -2381,6 +2387,7 @@ - - qualified_target_for_hash = gyp.common.QualifiedTarget(build_file, name, - toolset) -+ qualified_target_for_hash = qualified_target_for_hash.encode('utf-8') - hash_for_rules = hashlib.md5(qualified_target_for_hash).hexdigest() - - base_path = os.path.dirname(build_file) -@@ -2447,7 +2454,7 @@ - for config in configurations: - builddir = os.path.join(options.toplevel_dir, 'out', config) - arguments = ['ninja', '-C', builddir] -- print 'Building [%s]: %s' % (config, arguments) -+ print('Building [%s]: %s' % (config, arguments)) - subprocess.check_call(arguments) - - -@@ -2475,7 +2482,7 @@ - GenerateOutputForConfig(target_list, target_dicts, data, params, - user_config) - else: -- config_names = target_dicts[target_list[0]]['configurations'].keys() -+ config_names = target_dicts[target_list[0]]['configurations'] - if params['parallel']: - try: - pool = multiprocessing.Pool(len(config_names)) -@@ -2484,7 +2491,7 @@ - arglists.append( - (target_list, target_dicts, data, params, config_name)) - pool.map(CallGenerateOutputForConfig, arglists) -- except KeyboardInterrupt, e: -+ except KeyboardInterrupt as e: - pool.terminate() - raise e - else: -diff --git a/tools/gyp/pylib/gyp/generator/ninja_test.py b/pylib/gyp/generator/ninja_test.py -index 1767b2f..1ad68e4 100644 ---- a/tools/gyp/pylib/gyp/generator/ninja_test.py -+++ b/tools/gyp/pylib/gyp/generator/ninja_test.py -@@ -8,7 +8,6 @@ - - import gyp.generator.ninja as ninja - import unittest --import StringIO - import sys - import TestCommon - -diff --git a/tools/gyp/pylib/gyp/generator/xcode.py b/pylib/gyp/generator/xcode.py -index b35372a..8bc22be 100644 ---- a/tools/gyp/pylib/gyp/generator/xcode.py -+++ b/tools/gyp/pylib/gyp/generator/xcode.py -@@ -2,6 +2,8 @@ - # Use of this source code is governed by a BSD-style license that can be - # found in the LICENSE file. - -+from __future__ import print_function -+ - import filecmp - import gyp.common - import gyp.xcodeproj_file -@@ -129,7 +131,7 @@ - try: - os.makedirs(self.path) - self.created_dir = True -- except OSError, e: -+ except OSError as e: - if e.errno != errno.EEXIST: - raise - -@@ -183,7 +185,7 @@ - # the tree tree view for UI display. - # Any values set globally are applied to all configurations, then any - # per-configuration values are applied. -- for xck, xcv in self.build_file_dict.get('xcode_settings', {}).iteritems(): -+ for xck, xcv in self.build_file_dict.get('xcode_settings', {}).items(): - xccl.SetBuildSetting(xck, xcv) - if 'xcode_config_file' in self.build_file_dict: - config_ref = self.project.AddOrGetFileInRootGroup( -@@ -197,7 +199,7 @@ - if build_file_configuration_named: - xcc = xccl.ConfigurationNamed(config_name) - for xck, xcv in build_file_configuration_named.get('xcode_settings', -- {}).iteritems(): -+ {}).items(): - xcc.SetBuildSetting(xck, xcv) - if 'xcode_config_file' in build_file_configuration_named: - config_ref = self.project.AddOrGetFileInRootGroup( -@@ -273,7 +275,7 @@ - script = script + "\n".join( - ['export %s="%s"' % - (key, gyp.xcodeproj_file.ConvertVariablesToShellSyntax(val)) -- for (key, val) in command.get('environment').iteritems()]) + "\n" -+ for (key, val) in command.get('environment').items()]) + "\n" - - # Some test end up using sockets, files on disk, etc. and can get - # confused if more then one test runs at a time. The generator -@@ -444,7 +446,7 @@ - dir=self.path) - - try: -- output_file = os.fdopen(output_fd, 'wb') -+ output_file = os.fdopen(output_fd, 'w') - - self.project_file.Print(output_file) - output_file.close() -@@ -454,7 +456,7 @@ - same = False - try: - same = filecmp.cmp(pbxproj_path, new_pbxproj_path, False) -- except OSError, e: -+ except OSError as e: - if e.errno != errno.ENOENT: - raise - -@@ -473,10 +475,10 @@ - # - # No way to get the umask without setting a new one? Set a safe one - # and then set it back to the old value. -- umask = os.umask(077) -+ umask = os.umask(0o77) - os.umask(umask) - -- os.chmod(new_pbxproj_path, 0666 & ~umask) -+ os.chmod(new_pbxproj_path, 0o666 & ~umask) - os.rename(new_pbxproj_path, pbxproj_path) - - except Exception: -@@ -566,7 +568,7 @@ - def PerformBuild(data, configurations, params): - options = params['options'] - -- for build_file, build_file_dict in data.iteritems(): -+ for build_file, build_file_dict in data.items(): - (build_file_root, build_file_ext) = os.path.splitext(build_file) - if build_file_ext != '.gyp': - continue -@@ -577,7 +579,7 @@ - for config in configurations: - arguments = ['xcodebuild', '-project', xcodeproj_path] - arguments += ['-configuration', config] -- print "Building [%s]: %s" % (config, arguments) -+ print("Building [%s]: %s" % (config, arguments)) - subprocess.check_call(arguments) - - -@@ -625,7 +627,7 @@ - skip_excluded_files = \ - not generator_flags.get('xcode_list_excluded_files', True) - xcode_projects = {} -- for build_file, build_file_dict in data.iteritems(): -+ for build_file, build_file_dict in data.items(): - (build_file_root, build_file_ext) = os.path.splitext(build_file) - if build_file_ext != '.gyp': - continue -@@ -744,7 +746,7 @@ - xctarget_type = gyp.xcodeproj_file.PBXNativeTarget - try: - target_properties['productType'] = _types[type_bundle_key] -- except KeyError, e: -+ except KeyError as e: - gyp.common.ExceptionAppend(e, "-- unknown product type while " - "writing target %s" % target_name) - raise -@@ -1016,22 +1018,21 @@ - makefile_name) - # TODO(mark): try/close? Write to a temporary file and swap it only - # if it's got changes? -- makefile = open(makefile_path, 'wb') -+ makefile = open(makefile_path, 'w') - - # make will build the first target in the makefile by default. By - # convention, it's called "all". List all (or at least one) - # concrete output for each rule source as a prerequisite of the "all" - # target. - makefile.write('all: \\\n') -- for concrete_output_index in \ -- xrange(0, len(concrete_outputs_by_rule_source)): -+ for concrete_output_index, concrete_output_by_rule_source in \ -+ enumerate(concrete_outputs_by_rule_source): - # Only list the first (index [0]) concrete output of each input - # in the "all" target. Otherwise, a parallel make (-j > 1) would - # attempt to process each input multiple times simultaneously. - # Otherwise, "all" could just contain the entire list of - # concrete_outputs_all. -- concrete_output = \ -- concrete_outputs_by_rule_source[concrete_output_index][0] -+ concrete_output = concrete_output_by_rule_source[0] - if concrete_output_index == len(concrete_outputs_by_rule_source) - 1: - eol = '' - else: -@@ -1047,8 +1048,8 @@ - # rule source. Collect the names of the directories that are - # required. - concrete_output_dirs = [] -- for concrete_output_index in xrange(0, len(concrete_outputs)): -- concrete_output = concrete_outputs[concrete_output_index] -+ for concrete_output_index, concrete_output in \ -+ enumerate(concrete_outputs): - if concrete_output_index == 0: - bol = '' - else: -@@ -1066,8 +1067,7 @@ - # the set of additional rule inputs, if any. - prerequisites = [rule_source] - prerequisites.extend(rule.get('inputs', [])) -- for prerequisite_index in xrange(0, len(prerequisites)): -- prerequisite = prerequisites[prerequisite_index] -+ for prerequisite_index, prerequisite in enumerate(prerequisites): - if prerequisite_index == len(prerequisites) - 1: - eol = '' - else: -@@ -1279,7 +1279,7 @@ - set_define = EscapeXcodeDefine(define) - xcbc.AppendBuildSetting('GCC_PREPROCESSOR_DEFINITIONS', set_define) - if 'xcode_settings' in configuration: -- for xck, xcv in configuration['xcode_settings'].iteritems(): -+ for xck, xcv in configuration['xcode_settings'].items(): - xcbc.SetBuildSetting(xck, xcv) - if 'xcode_config_file' in configuration: - config_ref = pbxp.AddOrGetFileInRootGroup( -@@ -1287,7 +1287,7 @@ - xcbc.SetBaseConfiguration(config_ref) - - build_files = [] -- for build_file, build_file_dict in data.iteritems(): -+ for build_file, build_file_dict in data.items(): - if build_file.endswith('.gyp'): - build_files.append(build_file) - -diff --git a/tools/gyp/pylib/gyp/input.py b/pylib/gyp/input.py -index 21b4606..8ac47cb 100644 ---- a/tools/gyp/pylib/gyp/input.py -+++ b/tools/gyp/pylib/gyp/input.py -@@ -2,8 +2,9 @@ - # Use of this source code is governed by a BSD-style license that can be - # found in the LICENSE file. - --import ast -+from __future__ import print_function - -+import ast - import gyp.common - import gyp.simple_copy - import multiprocessing -@@ -231,10 +232,10 @@ - else: - build_file_data = eval(build_file_contents, {'__builtins__': None}, - None) -- except SyntaxError, e: -+ except SyntaxError as e: - e.filename = build_file_path - raise -- except Exception, e: -+ except Exception as e: - gyp.common.ExceptionAppend(e, 'while reading ' + build_file_path) - raise - -@@ -254,7 +255,7 @@ - else: - LoadBuildFileIncludesIntoDict(build_file_data, build_file_path, data, - aux_data, None, check) -- except Exception, e: -+ except Exception as e: - gyp.common.ExceptionAppend(e, - 'while reading includes of ' + build_file_path) - raise -@@ -291,7 +292,7 @@ - subdict_path, include) - - # Recurse into subdictionaries. -- for k, v in subdict.iteritems(): -+ for k, v in subdict.items(): - if type(v) is dict: - LoadBuildFileIncludesIntoDict(v, subdict_path, data, aux_data, - None, check) -@@ -456,7 +457,7 @@ - try: - LoadTargetBuildFile(dependency, data, aux_data, variables, - includes, depth, check, load_dependencies) -- except Exception, e: -+ except Exception as e: - gyp.common.ExceptionAppend( - e, 'while loading dependencies of %s' % build_file_path) - raise -@@ -477,7 +478,7 @@ - signal.signal(signal.SIGINT, signal.SIG_IGN) - - # Apply globals so that the worker process behaves the same. -- for key, value in global_flags.iteritems(): -+ for key, value in global_flags.items(): - globals()[key] = value - - SetGeneratorGlobals(generator_input_info) -@@ -499,12 +500,12 @@ - return (build_file_path, - build_file_data, - dependencies) -- except GypError, e: -+ except GypError as e: - sys.stderr.write("gyp: %s\n" % e) - return None -- except Exception, e: -- print >>sys.stderr, 'Exception:', e -- print >>sys.stderr, traceback.format_exc() -+ except Exception as e: -+ print('Exception:', e, file=sys.stderr) -+ print(traceback.format_exc(), file=sys.stderr) - return None - - -@@ -594,7 +595,7 @@ - args = (global_flags, dependency, - variables, includes, depth, check, generator_input_info), - callback = parallel_state.LoadTargetBuildFileCallback) -- except KeyboardInterrupt, e: -+ except KeyboardInterrupt as e: - parallel_state.pool.terminate() - raise e - -@@ -894,7 +895,7 @@ - stderr=subprocess.PIPE, - stdin=subprocess.PIPE, - cwd=build_file_dir) -- except Exception, e: -+ except Exception as e: - raise GypError("%s while executing command '%s' in %s" % - (e, contents, build_file)) - -@@ -1008,9 +1009,9 @@ - - # Convert all strings that are canonically-represented integers into integers. - if type(output) is list: -- for index in xrange(0, len(output)): -- if IsStrCanonicalInt(output[index]): -- output[index] = int(output[index]) -+ for index, outstr in enumerate(output): -+ if IsStrCanonicalInt(outstr): -+ output[index] = int(outstr) - elif IsStrCanonicalInt(output): - output = int(output) - -@@ -1079,13 +1080,13 @@ - if eval(ast_code, {'__builtins__': None}, variables): - return true_dict - return false_dict -- except SyntaxError, e: -+ except SyntaxError as e: - syntax_error = SyntaxError('%s while evaluating condition \'%s\' in %s ' - 'at character %d.' % - (str(e.args[0]), e.text, build_file, e.offset), - e.filename, e.lineno, e.offset, e.text) - raise syntax_error -- except NameError, e: -+ except NameError as e: - gyp.common.ExceptionAppend(e, 'while evaluating condition \'%s\' in %s' % - (cond_expr_expanded, build_file)) - raise GypError(e) -@@ -1140,7 +1141,7 @@ - def LoadAutomaticVariablesFromDict(variables, the_dict): - # Any keys with plain string values in the_dict become automatic variables. - # The variable name is the key name with a "_" character prepended. -- for key, value in the_dict.iteritems(): -+ for key, value in the_dict.items(): - if type(value) in (str, int, list): - variables['_' + key] = value - -@@ -1153,7 +1154,7 @@ - # the_dict in the_dict's parent dict. If the_dict's parent is not a dict - # (it could be a list or it could be parentless because it is a root dict), - # the_dict_key will be None. -- for key, value in the_dict.get('variables', {}).iteritems(): -+ for key, value in the_dict.get('variables', {}).items(): - if type(value) not in (str, int, list): - continue - -@@ -1192,7 +1193,7 @@ - # list before we process them so that you can reference one - # variable from another. They will be fully expanded by recursion - # in ExpandVariables. -- for key, value in the_dict['variables'].iteritems(): -+ for key, value in the_dict['variables'].items(): - variables[key] = value - - # Handle the associated variables dict first, so that any variable -@@ -1205,7 +1206,7 @@ - - LoadVariablesFromVariablesDict(variables, the_dict, the_dict_key) - -- for key, value in the_dict.iteritems(): -+ for key, value in the_dict.items(): - # Skip "variables", which was already processed if present. - if key != 'variables' and type(value) is str: - expanded = ExpandVariables(value, phase, variables, build_file) -@@ -1263,7 +1264,7 @@ - - # Recurse into child dicts, or process child lists which may result in - # further recursion into descendant dicts. -- for key, value in the_dict.iteritems(): -+ for key, value in the_dict.items(): - # Skip "variables" and string values, which were already processed if - # present. - if key == 'variables' or type(value) is str: -@@ -1360,14 +1361,14 @@ - for dep in dependency_sections - for op in ('', '!', '/')] - -- for target, target_dict in targets.iteritems(): -+ for target, target_dict in targets.items(): - target_build_file = gyp.common.BuildFile(target) - toolset = target_dict['toolset'] - for dependency_key in all_dependency_sections: - dependencies = target_dict.get(dependency_key, []) -- for index in xrange(0, len(dependencies)): -+ for index, dep in enumerate(dependencies): - dep_file, dep_target, dep_toolset = gyp.common.ResolveTarget( -- target_build_file, dependencies[index], toolset) -+ target_build_file, dep, toolset) - if not multiple_toolsets: - # Ignore toolset specification in the dependency if it is specified. - dep_toolset = toolset -@@ -1400,7 +1401,7 @@ - dependency list, must be qualified when this function is called. - """ - -- for target, target_dict in targets.iteritems(): -+ for target, target_dict in targets.items(): - toolset = target_dict['toolset'] - target_build_file = gyp.common.BuildFile(target) - for dependency_key in dependency_sections: -@@ -1462,7 +1463,7 @@ - def RemoveDuplicateDependencies(targets): - """Makes sure every dependency appears only once in all targets's dependency - lists.""" -- for target_name, target_dict in targets.iteritems(): -+ for target_name, target_dict in targets.items(): - for dependency_key in dependency_sections: - dependencies = target_dict.get(dependency_key, []) - if dependencies: -@@ -1478,7 +1479,7 @@ - def RemoveSelfDependencies(targets): - """Remove self dependencies from targets that have the prune_self_dependency - variable set.""" -- for target_name, target_dict in targets.iteritems(): -+ for target_name, target_dict in targets.items(): - for dependency_key in dependency_sections: - dependencies = target_dict.get(dependency_key, []) - if dependencies: -@@ -1491,7 +1492,7 @@ - def RemoveLinkDependenciesFromNoneTargets(targets): - """Remove dependencies having the 'link_dependency' attribute from the 'none' - targets.""" -- for target_name, target_dict in targets.iteritems(): -+ for target_name, target_dict in targets.items(): - for dependency_key in dependency_sections: - dependencies = target_dict.get(dependency_key, []) - if dependencies: -@@ -1783,14 +1784,14 @@ - # Create a DependencyGraphNode for each target. Put it into a dict for easy - # access. - dependency_nodes = {} -- for target, spec in targets.iteritems(): -+ for target, spec in targets.items(): - if target not in dependency_nodes: - dependency_nodes[target] = DependencyGraphNode(target) - - # Set up the dependency links. Targets that have no dependencies are treated - # as dependent on root_node. - root_node = DependencyGraphNode(None) -- for target, spec in targets.iteritems(): -+ for target, spec in targets.items(): - target_node = dependency_nodes[target] - target_build_file = gyp.common.BuildFile(target) - dependencies = spec.get('dependencies') -@@ -1814,7 +1815,7 @@ - if not root_node.dependents: - # If all targets have dependencies, add the first target as a dependent - # of root_node so that the cycle can be discovered from root_node. -- target = targets.keys()[0] -+ target = next(iter(targets)) - target_node = dependency_nodes[target] - target_node.dependencies.append(root_node) - root_node.dependents.append(target_node) -@@ -1833,20 +1834,20 @@ - # Create a DependencyGraphNode for each gyp file containing a target. Put - # it into a dict for easy access. - dependency_nodes = {} -- for target in targets.iterkeys(): -+ for target in targets.keys(): - build_file = gyp.common.BuildFile(target) - if not build_file in dependency_nodes: - dependency_nodes[build_file] = DependencyGraphNode(build_file) - - # Set up the dependency links. -- for target, spec in targets.iteritems(): -+ for target, spec in targets.items(): - build_file = gyp.common.BuildFile(target) - build_file_node = dependency_nodes[build_file] - target_dependencies = spec.get('dependencies', []) - for dependency in target_dependencies: - try: - dependency_build_file = gyp.common.BuildFile(dependency) -- except GypError, e: -+ except GypError as e: - gyp.common.ExceptionAppend( - e, 'while computing dependencies of .gyp file %s' % build_file) - raise -@@ -1864,7 +1865,7 @@ - - # Files that have no dependencies are treated as dependent on root_node. - root_node = DependencyGraphNode(None) -- for build_file_node in dependency_nodes.itervalues(): -+ for build_file_node in dependency_nodes.values(): - if len(build_file_node.dependencies) == 0: - build_file_node.dependencies.append(root_node) - root_node.dependents.append(build_file_node) -@@ -1877,7 +1878,7 @@ - if not root_node.dependents: - # If all files have dependencies, add the first file as a dependent - # of root_node so that the cycle can be discovered from root_node. -- file_node = dependency_nodes.values()[0] -+ file_node = next(iter(dependency_nodes.values())) - file_node.dependencies.append(root_node) - root_node.dependents.append(file_node) - cycles = [] -@@ -2104,7 +2105,7 @@ - - def MergeDicts(to, fro, to_file, fro_file): - # I wanted to name the parameter "from" but it's a Python keyword... -- for k, v in fro.iteritems(): -+ for k, v in fro.items(): - # It would be nice to do "if not k in to: to[k] = v" but that wouldn't give - # copy semantics. Something else may want to merge from the |fro| dict - # later, and having the same dict ref pointed to twice in the tree isn't -@@ -2239,13 +2240,13 @@ - if not 'configurations' in target_dict: - target_dict['configurations'] = {'Default': {}} - if not 'default_configuration' in target_dict: -- concrete = [i for (i, config) in target_dict['configurations'].iteritems() -+ concrete = [i for (i, config) in target_dict['configurations'].items() - if not config.get('abstract')] - target_dict['default_configuration'] = sorted(concrete)[0] - - merged_configurations = {} - configs = target_dict['configurations'] -- for (configuration, old_configuration_dict) in configs.iteritems(): -+ for (configuration, old_configuration_dict) in configs.items(): - # Skip abstract configurations (saves work only). - if old_configuration_dict.get('abstract'): - continue -@@ -2253,7 +2254,7 @@ - # Get the inheritance relationship right by making a copy of the target - # dict. - new_configuration_dict = {} -- for (key, target_val) in target_dict.iteritems(): -+ for (key, target_val) in target_dict.items(): - key_ext = key[-1:] - if key_ext in key_suffixes: - key_base = key[:-1] -@@ -2274,10 +2275,9 @@ - merged_configurations[configuration]) - - # Now drop all the abstract ones. -- for configuration in target_dict['configurations'].keys(): -- old_configuration_dict = target_dict['configurations'][configuration] -- if old_configuration_dict.get('abstract'): -- del target_dict['configurations'][configuration] -+ configs = target_dict['configurations'] -+ target_dict['configurations'] = \ -+ {k: v for k, v in configs.items() if not v.get('abstract')} - - # Now that all of the target's configurations have been built, go through - # the target dict's keys and remove everything that's been moved into a -@@ -2337,7 +2337,7 @@ - - lists = [] - del_lists = [] -- for key, value in the_dict.iteritems(): -+ for key, value in the_dict.items(): - operation = key[-1] - if operation != '!' and operation != '/': - continue -@@ -2385,8 +2385,8 @@ - exclude_key = list_key + '!' - if exclude_key in the_dict: - for exclude_item in the_dict[exclude_key]: -- for index in xrange(0, len(the_list)): -- if exclude_item == the_list[index]: -+ for index, list_item in enumerate(the_list): -+ if exclude_item == list_item: - # This item matches the exclude_item, so set its action to 0 - # (exclude). - list_actions[index] = 0 -@@ -2411,8 +2411,7 @@ - raise ValueError('Unrecognized action ' + action + ' in ' + name + \ - ' key ' + regex_key) - -- for index in xrange(0, len(the_list)): -- list_item = the_list[index] -+ for index, list_item in enumerate(the_list): - if list_actions[index] == action_value: - # Even if the regex matches, nothing will change so continue (regex - # searches are expensive). -@@ -2442,7 +2441,7 @@ - # the indices of items that haven't been seen yet don't shift. That means - # that things need to be prepended to excluded_list to maintain them in the - # same order that they existed in the_list. -- for index in xrange(len(list_actions) - 1, -1, -1): -+ for index in range(len(list_actions) - 1, -1, -1): - if list_actions[index] == 0: - # Dump anything with action 0 (exclude). Keep anything with action 1 - # (include) or -1 (no include or exclude seen for the item). -@@ -2455,7 +2454,7 @@ - the_dict[excluded_key] = excluded_list - - # Now recurse into subdicts and lists that may contain dicts. -- for key, value in the_dict.iteritems(): -+ for key, value in the_dict.items(): - if type(value) is dict: - ProcessListFiltersInDict(key, value) - elif type(value) is list: -@@ -2512,7 +2511,7 @@ - basenames.setdefault(basename, []).append(source) - - error = '' -- for basename, files in basenames.iteritems(): -+ for basename, files in basenames.items(): - if len(files) > 1: - error += ' %s: %s\n' % (basename, ' '.join(files)) - -@@ -2651,8 +2650,7 @@ - def TurnIntIntoStrInList(the_list): - """Given list the_list, recursively converts all integers into strings. - """ -- for index in xrange(0, len(the_list)): -- item = the_list[index] -+ for index, item in enumerate(the_list): - if type(item) is int: - the_list[index] = str(item) - elif type(item) is dict: -@@ -2769,7 +2767,7 @@ - try: - LoadTargetBuildFile(build_file, data, aux_data, - variables, includes, depth, check, True) -- except Exception, e: -+ except Exception as e: - gyp.common.ExceptionAppend(e, 'while trying to load %s' % build_file) - raise - -@@ -2791,7 +2789,7 @@ - RemoveLinkDependenciesFromNoneTargets(targets) - - # Apply exclude (!) and regex (/) list filters only for dependency_sections. -- for target_name, target_dict in targets.iteritems(): -+ for target_name, target_dict in targets.items(): - tmp_dict = {} - for key_base in dependency_sections: - for op in ('', '!', '/'): -diff --git a/tools/gyp/pylib/gyp/input_test.py b/pylib/gyp/input_test.py -index 4234fbb..6c4b1cc 100755 ---- a/tools/gyp/pylib/gyp/input_test.py -+++ b/tools/gyp/pylib/gyp/input_test.py -@@ -22,7 +22,7 @@ - dependency.dependents.append(dependent) - - def test_no_cycle_empty_graph(self): -- for label, node in self.nodes.iteritems(): -+ for label, node in self.nodes.items(): - self.assertEquals([], node.FindCycles()) - - def test_no_cycle_line(self): -@@ -30,7 +30,7 @@ - self._create_dependency(self.nodes['b'], self.nodes['c']) - self._create_dependency(self.nodes['c'], self.nodes['d']) - -- for label, node in self.nodes.iteritems(): -+ for label, node in self.nodes.items(): - self.assertEquals([], node.FindCycles()) - - def test_no_cycle_dag(self): -@@ -38,7 +38,7 @@ - self._create_dependency(self.nodes['a'], self.nodes['c']) - self._create_dependency(self.nodes['b'], self.nodes['c']) - -- for label, node in self.nodes.iteritems(): -+ for label, node in self.nodes.items(): - self.assertEquals([], node.FindCycles()) - - def test_cycle_self_reference(self): -diff --git a/tools/gyp/pylib/gyp/mac_tool.py b/pylib/gyp/mac_tool.py -index 0ad7e7a..7d3a8c2 100755 ---- a/tools/gyp/pylib/gyp/mac_tool.py -+++ b/tools/gyp/pylib/gyp/mac_tool.py -@@ -8,6 +8,8 @@ - These functions are executed via gyp-mac-tool when using the Makefile generator. - """ - -+from __future__ import print_function -+ - import fcntl - import fnmatch - import glob -@@ -16,7 +18,6 @@ - import plistlib - import re - import shutil --import string - import struct - import subprocess - import sys -@@ -155,11 +156,11 @@ - fp.close() - return None - fp.close() -- if header.startswith("\xFE\xFF"): -+ if header.startswith(b"\xFE\xFF"): - return "UTF-16" -- elif header.startswith("\xFF\xFE"): -+ elif header.startswith(b"\xFF\xFE"): - return "UTF-16" -- elif header.startswith("\xEF\xBB\xBF"): -+ elif header.startswith(b"\xEF\xBB\xBF"): - return "UTF-8" - else: - return None -@@ -174,7 +175,7 @@ - # Insert synthesized key/value pairs (e.g. BuildMachineOSBuild). - plist = plistlib.readPlistFromString(lines) - if keys: -- plist = dict(plist.items() + json.loads(keys[0]).items()) -+ plist.update(json.loads(keys[0])) - lines = plistlib.writePlistToString(plist) - - # Go through all the environment variables and replace them as variables in -@@ -185,7 +186,7 @@ - continue - evar = '${%s}' % key - evalue = os.environ[key] -- lines = string.replace(lines, evar, evalue) -+ lines = lines.replace(evar, evalue) - - # Xcode supports various suffices on environment variables, which are - # all undocumented. :rfc1034identifier is used in the standard project -@@ -195,11 +196,11 @@ - # in a URL either -- oops, hence :rfc1034identifier was born. - evar = '${%s:identifier}' % key - evalue = IDENT_RE.sub('_', os.environ[key]) -- lines = string.replace(lines, evar, evalue) -+ lines = lines.replace(evar, evalue) - - evar = '${%s:rfc1034identifier}' % key - evalue = IDENT_RE.sub('-', os.environ[key]) -- lines = string.replace(lines, evar, evalue) -+ lines = lines.replace(evar, evalue) - - # Remove any keys with values that haven't been replaced. - lines = lines.split('\n') -@@ -270,7 +271,7 @@ - _, err = libtoolout.communicate() - for line in err.splitlines(): - if not libtool_re.match(line) and not libtool_re5.match(line): -- print >>sys.stderr, line -+ print(line, file=sys.stderr) - # Unconditionally touch the output .a file on the command line if present - # and the command succeeded. A bit hacky. - if not libtoolout.returncode: -@@ -385,7 +386,7 @@ - ]) - if keys: - keys = json.loads(keys) -- for key, value in keys.iteritems(): -+ for key, value in keys.items(): - arg_name = '--' + key - if isinstance(value, bool): - if value: -@@ -480,8 +481,9 @@ - profiles_dir = os.path.join( - os.environ['HOME'], 'Library', 'MobileDevice', 'Provisioning Profiles') - if not os.path.isdir(profiles_dir): -- print >>sys.stderr, ( -- 'cannot find mobile provisioning for %s' % bundle_identifier) -+ print(( -+ 'cannot find mobile provisioning for %s' % bundle_identifier), -+ file=sys.stderr) - sys.exit(1) - provisioning_profiles = None - if profile: -@@ -502,8 +504,9 @@ - valid_provisioning_profiles[app_id_pattern] = ( - profile_path, profile_data, team_identifier) - if not valid_provisioning_profiles: -- print >>sys.stderr, ( -- 'cannot find mobile provisioning for %s' % bundle_identifier) -+ print(( -+ 'cannot find mobile provisioning for %s' % bundle_identifier), -+ file=sys.stderr) - sys.exit(1) - # If the user has multiple provisioning profiles installed that can be - # used for ${bundle_identifier}, pick the most specific one (ie. the -@@ -527,7 +530,7 @@ - - def _MergePlist(self, merged_plist, plist): - """Merge |plist| into |merged_plist|.""" -- for key, value in plist.iteritems(): -+ for key, value in plist.items(): - if isinstance(value, dict): - merged_value = merged_plist.get(key, {}) - if isinstance(merged_value, dict): -@@ -637,7 +640,7 @@ - the key was not found. - """ - if isinstance(data, str): -- for key, value in substitutions.iteritems(): -+ for key, value in substitutions.items(): - data = data.replace('$(%s)' % key, value) - return data - if isinstance(data, list): -diff --git a/tools/gyp/pylib/gyp/msvs_emulation.py b/pylib/gyp/msvs_emulation.py -index 6d5b5bd..63d40e6 100644 ---- a/tools/gyp/pylib/gyp/msvs_emulation.py -+++ b/tools/gyp/pylib/gyp/msvs_emulation.py -@@ -7,6 +7,7 @@ - build systems, primarily ninja. - """ - -+import collections - import os - import re - import subprocess -@@ -16,6 +17,12 @@ - import gyp.MSVSUtil - import gyp.MSVSVersion - -+try: -+ # basestring was removed in python3. -+ basestring -+except NameError: -+ basestring = str -+ - - windows_quoter_regex = re.compile(r'(\\*)"') - -@@ -84,8 +91,8 @@ - """Add |prefix| to |element| or each subelement if element is iterable.""" - if element is None: - return element -- # Note, not Iterable because we don't want to handle strings like that. -- if isinstance(element, list) or isinstance(element, tuple): -+ if (isinstance(element, collections.Iterable) and -+ not isinstance(element, basestring)): - return [prefix + e for e in element] - else: - return prefix + element -@@ -97,7 +104,8 @@ - if map is not None and element is not None: - if not callable(map): - map = map.get # Assume it's a dict, otherwise a callable to do the remap. -- if isinstance(element, list) or isinstance(element, tuple): -+ if (isinstance(element, collections.Iterable) and -+ not isinstance(element, basestring)): - element = filter(None, [map(elem) for elem in element]) - else: - element = map(element) -@@ -109,7 +117,8 @@ - then add |element| to it, adding each item in |element| if it's a list or - tuple.""" - if append is not None and element is not None: -- if isinstance(element, list) or isinstance(element, tuple): -+ if (isinstance(element, collections.Iterable) and -+ not isinstance(element, basestring)): - append.extend(element) - else: - append.append(element) -@@ -209,7 +218,7 @@ - configs = spec['configurations'] - for field, default in supported_fields: - setattr(self, field, {}) -- for configname, config in configs.iteritems(): -+ for configname, config in configs.items(): - getattr(self, field)[configname] = config.get(field, default()) - - self.msvs_cygwin_dirs = spec.get('msvs_cygwin_dirs', ['.']) -@@ -482,7 +491,7 @@ - # https://msdn.microsoft.com/en-us/library/dn502518.aspx - cflags.append('/FS') - # ninja handles parallelism by itself, don't have the compiler do it too. -- cflags = filter(lambda x: not x.startswith('/MP'), cflags) -+ cflags = [x for x in cflags if not x.startswith('/MP')] - return cflags - - def _GetPchFlags(self, config, extension): -@@ -649,19 +658,17 @@ - - # If the base address is not specifically controlled, DYNAMICBASE should - # be on by default. -- base_flags = filter(lambda x: 'DYNAMICBASE' in x or x == '/FIXED', -- ldflags) -- if not base_flags: -+ if not any('DYNAMICBASE' in flag or flag == '/FIXED' for flag in ldflags): - ldflags.append('/DYNAMICBASE') - - # If the NXCOMPAT flag has not been specified, default to on. Despite the - # documentation that says this only defaults to on when the subsystem is - # Vista or greater (which applies to the linker), the IDE defaults it on - # unless it's explicitly off. -- if not filter(lambda x: 'NXCOMPAT' in x, ldflags): -+ if not any('NXCOMPAT' in flag for flag in ldflags): - ldflags.append('/NXCOMPAT') - -- have_def_file = filter(lambda x: x.startswith('/DEF:'), ldflags) -+ have_def_file = any(flag.startswith('/DEF:') for flag in ldflags) - manifest_flags, intermediate_manifest, manifest_files = \ - self._GetLdManifestFlags(config, manifest_base_name, gyp_to_build_path, - is_executable and not have_def_file, build_dir) -@@ -953,7 +960,7 @@ - """Expand $(Variable) per expansions dict. See MsvsSettings.GetVSMacroEnv - for the canonical way to retrieve a suitable dict.""" - if '$' in string: -- for old, new in expansions.iteritems(): -+ for old, new in expansions.items(): - assert '$(' not in new, new - string = string.replace(old, new) - return string -@@ -1001,7 +1008,7 @@ - CreateProcess documentation for more details.""" - block = '' - nul = '\0' -- for key, value in envvar_dict.iteritems(): -+ for key, value in envvar_dict.items(): - block += key + '=' + value + nul - block += nul - return block -@@ -1056,7 +1063,7 @@ - env['INCLUDE'] = ';'.join(system_includes) - - env_block = _FormatAsEnvironmentBlock(env) -- f = open_out(os.path.join(toplevel_build_dir, 'environment.' + arch), 'wb') -+ f = open_out(os.path.join(toplevel_build_dir, 'environment.' + arch), 'w') - f.write(env_block) - f.close() - -@@ -1078,7 +1085,7 @@ - if int(generator_flags.get('msvs_error_on_missing_sources', 0)): - no_specials = filter(lambda x: '$' not in x, sources) - relative = [os.path.join(build_dir, gyp_to_ninja(s)) for s in no_specials] -- missing = filter(lambda x: not os.path.exists(x), relative) -+ missing = [x for x in relative if not os.path.exists(x)] - if missing: - # They'll look like out\Release\..\..\stuff\things.cc, so normalize the - # path for a slightly less crazy looking output. -diff --git a/tools/gyp/pylib/gyp/ordered_dict.py b/pylib/gyp/ordered_dict.py -deleted file mode 100644 -index a1e89f9..0000000 ---- a/tools/gyp/pylib/gyp/ordered_dict.py -+++ /dev/null -@@ -1,289 +0,0 @@ --# Unmodified from http://code.activestate.com/recipes/576693/ --# other than to add MIT license header (as specified on page, but not in code). --# Linked from Python documentation here: --# http://docs.python.org/2/library/collections.html#collections.OrderedDict --# --# This should be deleted once Py2.7 is available on all bots, see --# http://crbug.com/241769. --# --# Copyright (c) 2009 Raymond Hettinger. --# --# Permission is hereby granted, free of charge, to any person obtaining a copy --# of this software and associated documentation files (the "Software"), to deal --# in the Software without restriction, including without limitation the rights --# to use, copy, modify, merge, publish, distribute, sublicense, and/or sell --# copies of the Software, and to permit persons to whom the Software is --# furnished to do so, subject to the following conditions: --# --# The above copyright notice and this permission notice shall be included in --# all copies or substantial portions of the Software. --# --# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR --# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, --# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE --# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER --# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, --# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN --# THE SOFTWARE. -- --# Backport of OrderedDict() class that runs on Python 2.4, 2.5, 2.6, 2.7 and pypy. --# Passes Python2.7's test suite and incorporates all the latest updates. -- --try: -- from thread import get_ident as _get_ident --except ImportError: -- from dummy_thread import get_ident as _get_ident -- --try: -- from _abcoll import KeysView, ValuesView, ItemsView --except ImportError: -- pass -- -- --class OrderedDict(dict): -- 'Dictionary that remembers insertion order' -- # An inherited dict maps keys to values. -- # The inherited dict provides __getitem__, __len__, __contains__, and get. -- # The remaining methods are order-aware. -- # Big-O running times for all methods are the same as for regular dictionaries. -- -- # The internal self.__map dictionary maps keys to links in a doubly linked list. -- # The circular doubly linked list starts and ends with a sentinel element. -- # The sentinel element never gets deleted (this simplifies the algorithm). -- # Each link is stored as a list of length three: [PREV, NEXT, KEY]. -- -- def __init__(self, *args, **kwds): -- '''Initialize an ordered dictionary. Signature is the same as for -- regular dictionaries, but keyword arguments are not recommended -- because their insertion order is arbitrary. -- -- ''' -- if len(args) > 1: -- raise TypeError('expected at most 1 arguments, got %d' % len(args)) -- try: -- self.__root -- except AttributeError: -- self.__root = root = [] # sentinel node -- root[:] = [root, root, None] -- self.__map = {} -- self.__update(*args, **kwds) -- -- def __setitem__(self, key, value, dict_setitem=dict.__setitem__): -- 'od.__setitem__(i, y) <==> od[i]=y' -- # Setting a new item creates a new link which goes at the end of the linked -- # list, and the inherited dictionary is updated with the new key/value pair. -- if key not in self: -- root = self.__root -- last = root[0] -- last[1] = root[0] = self.__map[key] = [last, root, key] -- dict_setitem(self, key, value) -- -- def __delitem__(self, key, dict_delitem=dict.__delitem__): -- 'od.__delitem__(y) <==> del od[y]' -- # Deleting an existing item uses self.__map to find the link which is -- # then removed by updating the links in the predecessor and successor nodes. -- dict_delitem(self, key) -- link_prev, link_next, key = self.__map.pop(key) -- link_prev[1] = link_next -- link_next[0] = link_prev -- -- def __iter__(self): -- 'od.__iter__() <==> iter(od)' -- root = self.__root -- curr = root[1] -- while curr is not root: -- yield curr[2] -- curr = curr[1] -- -- def __reversed__(self): -- 'od.__reversed__() <==> reversed(od)' -- root = self.__root -- curr = root[0] -- while curr is not root: -- yield curr[2] -- curr = curr[0] -- -- def clear(self): -- 'od.clear() -> None. Remove all items from od.' -- try: -- for node in self.__map.itervalues(): -- del node[:] -- root = self.__root -- root[:] = [root, root, None] -- self.__map.clear() -- except AttributeError: -- pass -- dict.clear(self) -- -- def popitem(self, last=True): -- '''od.popitem() -> (k, v), return and remove a (key, value) pair. -- Pairs are returned in LIFO order if last is true or FIFO order if false. -- -- ''' -- if not self: -- raise KeyError('dictionary is empty') -- root = self.__root -- if last: -- link = root[0] -- link_prev = link[0] -- link_prev[1] = root -- root[0] = link_prev -- else: -- link = root[1] -- link_next = link[1] -- root[1] = link_next -- link_next[0] = root -- key = link[2] -- del self.__map[key] -- value = dict.pop(self, key) -- return key, value -- -- # -- the following methods do not depend on the internal structure -- -- -- def keys(self): -- 'od.keys() -> list of keys in od' -- return list(self) -- -- def values(self): -- 'od.values() -> list of values in od' -- return [self[key] for key in self] -- -- def items(self): -- 'od.items() -> list of (key, value) pairs in od' -- return [(key, self[key]) for key in self] -- -- def iterkeys(self): -- 'od.iterkeys() -> an iterator over the keys in od' -- return iter(self) -- -- def itervalues(self): -- 'od.itervalues -> an iterator over the values in od' -- for k in self: -- yield self[k] -- -- def iteritems(self): -- 'od.iteritems -> an iterator over the (key, value) items in od' -- for k in self: -- yield (k, self[k]) -- -- # Suppress 'OrderedDict.update: Method has no argument': -- # pylint: disable=E0211 -- def update(*args, **kwds): -- '''od.update(E, **F) -> None. Update od from dict/iterable E and F. -- -- If E is a dict instance, does: for k in E: od[k] = E[k] -- If E has a .keys() method, does: for k in E.keys(): od[k] = E[k] -- Or if E is an iterable of items, does: for k, v in E: od[k] = v -- In either case, this is followed by: for k, v in F.items(): od[k] = v -- -- ''' -- if len(args) > 2: -- raise TypeError('update() takes at most 2 positional ' -- 'arguments (%d given)' % (len(args),)) -- elif not args: -- raise TypeError('update() takes at least 1 argument (0 given)') -- self = args[0] -- # Make progressively weaker assumptions about "other" -- other = () -- if len(args) == 2: -- other = args[1] -- if isinstance(other, dict): -- for key in other: -- self[key] = other[key] -- elif hasattr(other, 'keys'): -- for key in other.keys(): -- self[key] = other[key] -- else: -- for key, value in other: -- self[key] = value -- for key, value in kwds.items(): -- self[key] = value -- -- __update = update # let subclasses override update without breaking __init__ -- -- __marker = object() -- -- def pop(self, key, default=__marker): -- '''od.pop(k[,d]) -> v, remove specified key and return the corresponding value. -- If key is not found, d is returned if given, otherwise KeyError is raised. -- -- ''' -- if key in self: -- result = self[key] -- del self[key] -- return result -- if default is self.__marker: -- raise KeyError(key) -- return default -- -- def setdefault(self, key, default=None): -- 'od.setdefault(k[,d]) -> od.get(k,d), also set od[k]=d if k not in od' -- if key in self: -- return self[key] -- self[key] = default -- return default -- -- def __repr__(self, _repr_running={}): -- 'od.__repr__() <==> repr(od)' -- call_key = id(self), _get_ident() -- if call_key in _repr_running: -- return '...' -- _repr_running[call_key] = 1 -- try: -- if not self: -- return '%s()' % (self.__class__.__name__,) -- return '%s(%r)' % (self.__class__.__name__, self.items()) -- finally: -- del _repr_running[call_key] -- -- def __reduce__(self): -- 'Return state information for pickling' -- items = [[k, self[k]] for k in self] -- inst_dict = vars(self).copy() -- for k in vars(OrderedDict()): -- inst_dict.pop(k, None) -- if inst_dict: -- return (self.__class__, (items,), inst_dict) -- return self.__class__, (items,) -- -- def copy(self): -- 'od.copy() -> a shallow copy of od' -- return self.__class__(self) -- -- @classmethod -- def fromkeys(cls, iterable, value=None): -- '''OD.fromkeys(S[, v]) -> New ordered dictionary with keys from S -- and values equal to v (which defaults to None). -- -- ''' -- d = cls() -- for key in iterable: -- d[key] = value -- return d -- -- def __eq__(self, other): -- '''od.__eq__(y) <==> od==y. Comparison to another OD is order-sensitive -- while comparison to a regular mapping is order-insensitive. -- -- ''' -- if isinstance(other, OrderedDict): -- return len(self)==len(other) and self.items() == other.items() -- return dict.__eq__(self, other) -- -- def __ne__(self, other): -- return not self == other -- -- # -- the following methods are only used in Python 2.7 -- -- -- def viewkeys(self): -- "od.viewkeys() -> a set-like object providing a view on od's keys" -- return KeysView(self) -- -- def viewvalues(self): -- "od.viewvalues() -> an object providing a view on od's values" -- return ValuesView(self) -- -- def viewitems(self): -- "od.viewitems() -> a set-like object providing a view on od's items" -- return ItemsView(self) -- -diff --git a/tools/gyp/pylib/gyp/simple_copy.py b/pylib/gyp/simple_copy.py -index eaf5f8b..58a61c3 100644 ---- a/tools/gyp/pylib/gyp/simple_copy.py -+++ b/tools/gyp/pylib/gyp/simple_copy.py -@@ -49,7 +49,7 @@ - - def _deepcopy_dict(x): - y = {} -- for key, value in x.iteritems(): -+ for key, value in x.items(): - y[deepcopy(key)] = deepcopy(value) - return y - d[dict] = _deepcopy_dict -diff --git a/tools/gyp/pylib/gyp/win_tool.py b/pylib/gyp/win_tool.py -index 1c843a0..8973484 100755 ---- a/tools/gyp/pylib/gyp/win_tool.py -+++ b/tools/gyp/pylib/gyp/win_tool.py -@@ -9,6 +9,8 @@ - These functions are executed via gyp-win-tool when using the ninja generator. - """ - -+from __future__ import print_function -+ - import os - import re - import shutil -@@ -134,7 +136,7 @@ - if (not line.startswith(' Creating library ') and - not line.startswith('Generating code') and - not line.startswith('Finished generating code')): -- print line -+ print(line) - return link.returncode - - def ExecLinkWithManifests(self, arch, embed_manifest, out, ldcmd, resname, -@@ -193,16 +195,18 @@ - our_manifest = '%(out)s.manifest' % variables - # Load and normalize the manifests. mt.exe sometimes removes whitespace, - # and sometimes doesn't unfortunately. -- with open(our_manifest, 'rb') as our_f: -- with open(assert_manifest, 'rb') as assert_f: -+ with open(our_manifest, 'r') as our_f: -+ with open(assert_manifest, 'r') as assert_f: - our_data = our_f.read().translate(None, string.whitespace) - assert_data = assert_f.read().translate(None, string.whitespace) - if our_data != assert_data: - os.unlink(out) - def dump(filename): -- sys.stderr.write('%s\n-----\n' % filename) -- with open(filename, 'rb') as f: -- sys.stderr.write(f.read() + '\n-----\n') -+ print(filename, file=sys.stderr) -+ print('-----', file=sys.stderr) -+ with open(filename, 'r') as f: -+ print(f.read(), file=sys.stderr) -+ print('-----', file=sys.stderr) - dump(intermediate_manifest) - dump(our_manifest) - dump(assert_manifest) -@@ -223,7 +227,7 @@ - out, _ = popen.communicate() - for line in out.splitlines(): - if line and 'manifest authoring warning 81010002' not in line: -- print line -+ print(line) - return popen.returncode - - def ExecManifestToRc(self, arch, *args): -@@ -231,7 +235,7 @@ - |args| is tuple containing path to resource file, path to manifest file - and resource name which can be "1" (for executables) or "2" (for DLLs).""" - manifest_path, resource_path, resource_name = args -- with open(resource_path, 'wb') as output: -+ with open(resource_path, 'w') as output: - output.write('#include <windows.h>\n%s RT_MANIFEST "%s"' % ( - resource_name, - os.path.abspath(manifest_path).replace('\\', '/'))) -@@ -263,7 +267,7 @@ - for x in lines if x.startswith(prefixes)) - for line in lines: - if not line.startswith(prefixes) and line not in processing: -- print line -+ print(line) - return popen.returncode - - def ExecAsmWrapper(self, arch, *args): -@@ -277,7 +281,7 @@ - not line.startswith('Microsoft (R) Macro Assembler') and - not line.startswith(' Assembling: ') and - line): -- print line -+ print(line) - return popen.returncode - - def ExecRcWrapper(self, arch, *args): -@@ -291,7 +295,7 @@ - if (not line.startswith('Microsoft (R) Windows (R) Resource Compiler') and - not line.startswith('Copyright (C) Microsoft Corporation') and - line): -- print line -+ print(line) - return popen.returncode - - def ExecActionWrapper(self, arch, rspfile, *dir): -@@ -300,7 +304,7 @@ - env = self._GetEnv(arch) - # TODO(scottmg): This is a temporary hack to get some specific variables - # through to actions that are set after gyp-time. http://crbug.com/333738. -- for k, v in os.environ.iteritems(): -+ for k, v in os.environ.items(): - if k not in env: - env[k] = v - args = open(rspfile).read() -diff --git a/tools/gyp/pylib/gyp/xcode_emulation.py b/pylib/gyp/xcode_emulation.py -index dba8e76..4c875de 100644 ---- a/tools/gyp/pylib/gyp/xcode_emulation.py -+++ b/tools/gyp/pylib/gyp/xcode_emulation.py -@@ -7,6 +7,8 @@ - other build systems, such as make and ninja. - """ - -+from __future__ import print_function -+ - import copy - import gyp.common - import os -@@ -73,7 +75,7 @@ - if arch not in expanded_archs: - expanded_archs.append(arch) - except KeyError as e: -- print 'Warning: Ignoring unsupported variable "%s".' % variable -+ print('Warning: Ignoring unsupported variable "%s".' % variable) - elif arch not in expanded_archs: - expanded_archs.append(arch) - return expanded_archs -@@ -171,7 +173,7 @@ - # the same for all configs are implicitly per-target settings. - self.xcode_settings = {} - configs = spec['configurations'] -- for configname, config in configs.iteritems(): -+ for configname, config in configs.items(): - self.xcode_settings[configname] = config.get('xcode_settings', {}) - self._ConvertConditionalKeys(configname) - if self.xcode_settings[configname].get('IPHONEOS_DEPLOYMENT_TARGET', -@@ -197,8 +199,8 @@ - new_key = key.split("[")[0] - settings[new_key] = settings[key] - else: -- print 'Warning: Conditional keys not implemented, ignoring:', \ -- ' '.join(conditional_keys) -+ print('Warning: Conditional keys not implemented, ignoring:', \ -+ ' '.join(conditional_keys)) - del settings[key] - - def _Settings(self): -@@ -216,7 +218,7 @@ - - def _WarnUnimplemented(self, test_key): - if test_key in self._Settings(): -- print 'Warning: Ignoring not yet implemented key "%s".' % test_key -+ print('Warning: Ignoring not yet implemented key "%s".' % test_key) - - def IsBinaryOutputFormat(self, configname): - default = "binary" if self.isIOS else "xml" -@@ -963,7 +965,7 @@ - result = dict(self.xcode_settings[configname]) - first_pass = False - else: -- for key, value in self.xcode_settings[configname].iteritems(): -+ for key, value in self.xcode_settings[configname].items(): - if key not in result: - continue - elif result[key] != value: -@@ -1084,8 +1086,8 @@ - unimpl = ['OTHER_CODE_SIGN_FLAGS'] - unimpl = set(unimpl) & set(self.xcode_settings[configname].keys()) - if unimpl: -- print 'Warning: Some codesign keys not implemented, ignoring: %s' % ( -- ', '.join(sorted(unimpl))) -+ print('Warning: Some codesign keys not implemented, ignoring: %s' % ( -+ ', '.join(sorted(unimpl)))) - - if self._IsXCTest(): - # For device xctests, Xcode copies two extra frameworks into $TEST_HOST. -@@ -1737,7 +1739,7 @@ - order = gyp.common.TopologicallySorted(env.keys(), GetEdges) - order.reverse() - return order -- except gyp.common.CycleError, e: -+ except gyp.common.CycleError as e: - raise GypError( - 'Xcode environment variables are cyclically dependent: ' + str(e.nodes)) - -@@ -1774,10 +1776,11 @@ - def _AddIOSDeviceConfigurations(targets): - """Clone all targets and append -iphoneos to the name. Configure these targets - to build for iOS devices and use correct architectures for those builds.""" -- for target_dict in targets.itervalues(): -+ for target_dict in targets.values(): - toolset = target_dict['toolset'] - configs = target_dict['configurations'] -- for config_name, simulator_config_dict in dict(configs).iteritems(): -+ -+ for config_name, simulator_config_dict in dict(configs).items(): - iphoneos_config_dict = copy.deepcopy(simulator_config_dict) - configs[config_name + '-iphoneos'] = iphoneos_config_dict - configs[config_name + '-iphonesimulator'] = simulator_config_dict -diff --git a/tools/gyp/pylib/gyp/xcode_ninja.py b/pylib/gyp/xcode_ninja.py -index bc76fff..1d71b8c 100644 ---- a/tools/gyp/pylib/gyp/xcode_ninja.py -+++ b/tools/gyp/pylib/gyp/xcode_ninja.py -@@ -28,7 +28,7 @@ - workspace_path = os.path.join(options.generator_output, workspace_path) - try: - os.makedirs(workspace_path) -- except OSError, e: -+ except OSError as e: - if e.errno != errno.EEXIST: - raise - output_string = '<?xml version="1.0" encoding="UTF-8"?>\n' + \ -@@ -85,7 +85,7 @@ - "%s/$(CONFIGURATION)$(EFFECTIVE_PLATFORM_NAME)" % ninja_toplevel - - if 'configurations' in old_spec: -- for config in old_spec['configurations'].iterkeys(): -+ for config in old_spec['configurations'].keys(): - old_xcode_settings = \ - old_spec['configurations'][config].get('xcode_settings', {}) - if 'IPHONEOS_DEPLOYMENT_TARGET' in old_xcode_settings: -@@ -167,7 +167,7 @@ - params: Dict of global options for gyp. - """ - orig_gyp = params['build_files'][0] -- for gyp_name, gyp_dict in data.iteritems(): -+ for gyp_name, gyp_dict in data.items(): - if gyp_name == orig_gyp: - depth = gyp_dict['_DEPTH'] - -@@ -238,7 +238,7 @@ - not generator_flags.get('xcode_ninja_list_excluded_files', True) - - sources = [] -- for target, target_dict in target_dicts.iteritems(): -+ for target, target_dict in target_dicts.items(): - base = os.path.dirname(target) - files = target_dict.get('sources', []) + \ - target_dict.get('mac_bundle_resources', []) -diff --git a/tools/gyp/pylib/gyp/xcodeproj_file.py b/pylib/gyp/xcodeproj_file.py -index e69235f..bd238f6 100644 ---- a/tools/gyp/pylib/gyp/xcodeproj_file.py -+++ b/tools/gyp/pylib/gyp/xcodeproj_file.py -@@ -154,6 +154,11 @@ - import sha - _new_sha1 = sha.new - -+try: -+ # basestring was removed in python3. -+ basestring -+except NameError: -+ basestring = str - - # See XCObject._EncodeString. This pattern is used to determine when a string - # can be printed unquoted. Strings that match this pattern may be printed -@@ -314,7 +319,7 @@ - """ - - that = self.__class__(id=self.id, parent=self.parent) -- for key, value in self._properties.iteritems(): -+ for key, value in self._properties.items(): - is_strong = self._schema[key][2] - - if isinstance(value, XCObject): -@@ -324,8 +329,7 @@ - that._properties[key] = new_value - else: - that._properties[key] = value -- elif isinstance(value, str) or isinstance(value, unicode) or \ -- isinstance(value, int): -+ elif isinstance(value, basestring) or isinstance(value, int): - that._properties[key] = value - elif isinstance(value, list): - if is_strong: -@@ -449,10 +453,10 @@ - # is 160 bits. Instead of throwing out 64 bits of the digest, xor them - # into the portion that gets used. - assert hash.digest_size % 4 == 0 -- digest_int_count = hash.digest_size / 4 -+ digest_int_count = hash.digest_size // 4 - digest_ints = struct.unpack('>' + 'I' * digest_int_count, hash.digest()) - id_ints = [0, 0, 0] -- for index in xrange(0, digest_int_count): -+ for index in range(0, digest_int_count): - id_ints[index % 3] ^= digest_ints[index] - self.id = '%08X%08X%08X' % tuple(id_ints) - -@@ -475,7 +479,7 @@ - """Returns a list of all of this object's owned (strong) children.""" - - children = [] -- for property, attributes in self._schema.iteritems(): -+ for property, attributes in self._schema.items(): - (is_list, property_type, is_strong) = attributes[0:3] - if is_strong and property in self._properties: - if not is_list: -@@ -603,7 +607,12 @@ - comment = value.Comment() - elif isinstance(value, str): - printable += self._EncodeString(value) -- elif isinstance(value, unicode): -+ # A python3 compatible way of saying isinstance(value, unicode). -+ # basestring is str in python3 so this is equivalent to the above -+ # isinstance. Thus if it failed above it will fail here. -+ # In python2 we test against str and unicode at this point. str has already -+ # failed in the above isinstance so we test against unicode. -+ elif isinstance(value, basestring): - printable += self._EncodeString(value.encode('utf-8')) - elif isinstance(value, int): - printable += str(value) -@@ -622,7 +631,7 @@ - printable += end_tabs + ')' - elif isinstance(value, dict): - printable = '{' + sep -- for item_key, item_value in sorted(value.iteritems()): -+ for item_key, item_value in sorted(value.items()): - printable += element_tabs + \ - self._XCPrintableValue(tabs + 1, item_key, flatten_list) + ' = ' + \ - self._XCPrintableValue(tabs + 1, item_value, flatten_list) + ';' + \ -@@ -691,7 +700,7 @@ - printable_value[0] == '"' and printable_value[-1] == '"': - printable_value = printable_value[1:-1] - printable += printable_key + ' = ' + printable_value + ';' + after_kv -- except TypeError, e: -+ except TypeError as e: - gyp.common.ExceptionAppend(e, - 'while printing key "%s"' % key) - raise -@@ -730,7 +739,7 @@ - self._XCKVPrint(file, 3, 'isa', self.__class__.__name__) - - # The remaining elements of an object dictionary are sorted alphabetically. -- for property, value in sorted(self._properties.iteritems()): -+ for property, value in sorted(self._properties.items()): - self._XCKVPrint(file, 3, property, value) - - # End the object. -@@ -752,7 +761,7 @@ - if properties is None: - return - -- for property, value in properties.iteritems(): -+ for property, value in properties.items(): - # Make sure the property is in the schema. - if not property in self._schema: - raise KeyError(property + ' not in ' + self.__class__.__name__) -@@ -766,7 +775,7 @@ - ' must be list, not ' + value.__class__.__name__) - for item in value: - if not isinstance(item, property_type) and \ -- not (item.__class__ == unicode and property_type == str): -+ not (isinstance(item, basestring) and property_type == str): - # Accept unicode where str is specified. str is treated as - # UTF-8-encoded. - raise TypeError( -@@ -774,7 +783,7 @@ - ' must be ' + property_type.__name__ + ', not ' + \ - item.__class__.__name__) - elif not isinstance(value, property_type) and \ -- not (value.__class__ == unicode and property_type == str): -+ not (isinstance(value, basestring) and property_type == str): - # Accept unicode where str is specified. str is treated as - # UTF-8-encoded. - raise TypeError( -@@ -788,8 +797,7 @@ - self._properties[property] = value.Copy() - else: - self._properties[property] = value -- elif isinstance(value, str) or isinstance(value, unicode) or \ -- isinstance(value, int): -+ elif isinstance(value, basestring) or isinstance(value, int): - self._properties[property] = value - elif isinstance(value, list): - if is_strong: -@@ -865,7 +873,7 @@ - - # TODO(mark): A stronger verification mechanism is needed. Some - # subclasses need to perform validation beyond what the schema can enforce. -- for property, attributes in self._schema.iteritems(): -+ for property, attributes in self._schema.items(): - (is_list, property_type, is_strong, is_required) = attributes[0:4] - if is_required and not property in self._properties: - raise KeyError(self.__class__.__name__ + ' requires ' + property) -@@ -875,7 +883,7 @@ - overwrite properties that have already been set.""" - - defaults = {} -- for property, attributes in self._schema.iteritems(): -+ for property, attributes in self._schema.items(): - (is_list, property_type, is_strong, is_required) = attributes[0:4] - if is_required and len(attributes) >= 5 and \ - not property in self._properties: -@@ -1426,8 +1434,8 @@ - xche = self - while xche != None and isinstance(xche, XCHierarchicalElement): - xche_hashables = xche.Hashables() -- for index in xrange(0, len(xche_hashables)): -- hashables.insert(index, xche_hashables[index]) -+ for index, xche_hashable in enumerate(xche_hashables): -+ hashables.insert(index, xche_hashable) - xche = xche.parent - return hashables - -@@ -2468,8 +2476,7 @@ - # The headers phase should come before the resources, sources, and - # frameworks phases, if any. - insert_at = len(self._properties['buildPhases']) -- for index in xrange(0, len(self._properties['buildPhases'])): -- phase = self._properties['buildPhases'][index] -+ for index, phase in enumerate(self._properties['buildPhases']): - if isinstance(phase, PBXResourcesBuildPhase) or \ - isinstance(phase, PBXSourcesBuildPhase) or \ - isinstance(phase, PBXFrameworksBuildPhase): -@@ -2489,8 +2496,7 @@ - # The resources phase should come before the sources and frameworks - # phases, if any. - insert_at = len(self._properties['buildPhases']) -- for index in xrange(0, len(self._properties['buildPhases'])): -- phase = self._properties['buildPhases'][index] -+ for index, phase in enumerate(self._properties['buildPhases']): - if isinstance(phase, PBXSourcesBuildPhase) or \ - isinstance(phase, PBXFrameworksBuildPhase): - insert_at = index -@@ -2911,7 +2917,7 @@ - # determine the sort order. - return cmp(x_index, y_index) - -- for other_pbxproject, ref_dict in self._other_pbxprojects.iteritems(): -+ for other_pbxproject, ref_dict in self._other_pbxprojects.items(): - # Build up a list of products in the remote project file, ordered the - # same as the targets that produce them. - remote_products = [] -diff --git a/tools/gyp/pylib/gyp/xml_fix.py b/pylib/gyp/xml_fix.py -index 5de8481..4308d99 100644 ---- a/tools/gyp/pylib/gyp/xml_fix.py -+++ b/tools/gyp/pylib/gyp/xml_fix.py -@@ -32,8 +32,7 @@ - writer.write(indent+"<" + self.tagName) - - attrs = self._get_attributes() -- a_names = attrs.keys() -- a_names.sort() -+ a_names = sorted(attrs.keys()) - - for a_name in a_names: - writer.write(" %s=\"" % a_name) -diff --git a/tools/gyp/tools/graphviz.py b/tools/graphviz.py -index 326ae22..538b059 100755 ---- a/tools/gyp/tools/graphviz.py -+++ b/tools/gyp/tools/graphviz.py -@@ -8,6 +8,8 @@ - generate input suitable for graphviz to render a dependency graph of - targets.""" - -+from __future__ import print_function -+ - import collections - import json - import sys -@@ -50,9 +52,9 @@ - build_file, target_name, toolset = ParseTarget(src) - files[build_file].append(src) - -- print 'digraph D {' -- print ' fontsize=8' # Used by subgraphs. -- print ' node [fontsize=8]' -+ print('digraph D {') -+ print(' fontsize=8') # Used by subgraphs. -+ print(' node [fontsize=8]') - - # Output nodes by file. We must first write out each node within - # its file grouping before writing out any edges that may refer -@@ -63,31 +65,31 @@ - # the display by making it a box without an internal node. - target = targets[0] - build_file, target_name, toolset = ParseTarget(target) -- print ' "%s" [shape=box, label="%s\\n%s"]' % (target, filename, -- target_name) -+ print(' "%s" [shape=box, label="%s\\n%s"]' % (target, filename, -+ target_name)) - else: - # Group multiple nodes together in a subgraph. -- print ' subgraph "cluster_%s" {' % filename -- print ' label = "%s"' % filename -+ print(' subgraph "cluster_%s" {' % filename) -+ print(' label = "%s"' % filename) - for target in targets: - build_file, target_name, toolset = ParseTarget(target) -- print ' "%s" [label="%s"]' % (target, target_name) -- print ' }' -+ print(' "%s" [label="%s"]' % (target, target_name)) -+ print(' }') - - # Now that we've placed all the nodes within subgraphs, output all - # the edges between nodes. - for src, dsts in edges.items(): - for dst in dsts: -- print ' "%s" -> "%s"' % (src, dst) -+ print(' "%s" -> "%s"' % (src, dst)) - -- print '}' -+ print('}') - - - def main(): - if len(sys.argv) < 2: -- print >>sys.stderr, __doc__ -- print >>sys.stderr -- print >>sys.stderr, 'usage: %s target1 target2...' % (sys.argv[0]) -+ print(__doc__, file=sys.stderr) -+ print(file=sys.stderr) -+ print('usage: %s target1 target2...' % (sys.argv[0]), file=sys.stderr) - return 1 - - edges = LoadEdges('dump.json', sys.argv[1:]) -diff --git a/tools/gyp/tools/pretty_gyp.py b/tools/pretty_gyp.py -index d5736bb..5060d1d 100755 ---- a/tools/gyp/tools/pretty_gyp.py -+++ b/tools/gyp/tools/pretty_gyp.py -@@ -6,6 +6,8 @@ - - """Pretty-prints the contents of a GYP file.""" - -+from __future__ import print_function -+ - import sys - import re - -@@ -125,15 +127,15 @@ - (brace_diff, after) = count_braces(line) - if brace_diff != 0: - if after: -- print " " * (basic_offset * indent) + line -+ print(" " * (basic_offset * indent) + line) - indent += brace_diff - else: - indent += brace_diff -- print " " * (basic_offset * indent) + line -+ print(" " * (basic_offset * indent) + line) - else: -- print " " * (basic_offset * indent) + line -+ print(" " * (basic_offset * indent) + line) - else: -- print "" -+ print("") - last_line = line - - -diff --git a/tools/gyp/tools/pretty_sln.py b/tools/pretty_sln.py -index ca8cf4a..12a6dad 100755 ---- a/tools/gyp/tools/pretty_sln.py -+++ b/tools/gyp/tools/pretty_sln.py -@@ -12,6 +12,8 @@ - Then it outputs a possible build order. - """ - -+from __future__ import print_function -+ - __author__ = 'nsylvain (Nicolas Sylvain)' - - import os -@@ -26,7 +28,7 @@ - for dep in deps[project]: - if dep not in built: - BuildProject(dep, built, projects, deps) -- print project -+ print(project) - built.append(project) - - def ParseSolution(solution_file): -@@ -100,44 +102,44 @@ - return (projects, dependencies) - - def PrintDependencies(projects, deps): -- print "---------------------------------------" -- print "Dependencies for all projects" -- print "---------------------------------------" -- print "-- --" -+ print("---------------------------------------") -+ print("Dependencies for all projects") -+ print("---------------------------------------") -+ print("-- --") - - for (project, dep_list) in sorted(deps.items()): -- print "Project : %s" % project -- print "Path : %s" % projects[project][0] -+ print("Project : %s" % project) -+ print("Path : %s" % projects[project][0]) - if dep_list: - for dep in dep_list: -- print " - %s" % dep -- print "" -+ print(" - %s" % dep) -+ print("") - -- print "-- --" -+ print("-- --") - - def PrintBuildOrder(projects, deps): -- print "---------------------------------------" -- print "Build order " -- print "---------------------------------------" -- print "-- --" -+ print("---------------------------------------") -+ print("Build order ") -+ print("---------------------------------------") -+ print("-- --") - - built = [] - for (project, _) in sorted(deps.items()): - if project not in built: - BuildProject(project, built, projects, deps) - -- print "-- --" -+ print("-- --") - - def PrintVCProj(projects): - - for project in projects: -- print "-------------------------------------" -- print "-------------------------------------" -- print project -- print project -- print project -- print "-------------------------------------" -- print "-------------------------------------" -+ print("-------------------------------------") -+ print("-------------------------------------") -+ print(project) -+ print(project) -+ print(project) -+ print("-------------------------------------") -+ print("-------------------------------------") - - project_path = os.path.abspath(os.path.join(os.path.dirname(sys.argv[1]), - projects[project][2])) -@@ -153,7 +155,7 @@ - def main(): - # check if we have exactly 1 parameter. - if len(sys.argv) < 2: -- print 'Usage: %s "c:\\path\\to\\project.sln"' % sys.argv[0] -+ print('Usage: %s "c:\\path\\to\\project.sln"' % sys.argv[0]) - return 1 - - (projects, deps) = ParseSolution(sys.argv[1]) -diff --git a/tools/gyp/tools/pretty_vcproj.py b/tools/pretty_vcproj.py -index 6099bd7..f02e59e 100755 ---- a/tools/gyp/tools/pretty_vcproj.py -+++ b/tools/gyp/tools/pretty_vcproj.py -@@ -12,6 +12,8 @@ - It outputs the resulting xml to stdout. - """ - -+from __future__ import print_function -+ - __author__ = 'nsylvain (Nicolas Sylvain)' - - import os -@@ -73,23 +75,23 @@ - - # Print the main tag - if attr_count == 0: -- print '%s<%s>' % (' '*indent, node.nodeName) -+ print('%s<%s>' % (' '*indent, node.nodeName)) - else: -- print '%s<%s' % (' '*indent, node.nodeName) -+ print('%s<%s' % (' '*indent, node.nodeName)) - - all_attributes = [] - for (name, value) in node.attributes.items(): - all_attributes.append((name, value)) -- all_attributes.sort(CmpTuple()) -+ all_attributes.sort(key=(lambda attr: attr[0])) - for (name, value) in all_attributes: -- print '%s %s="%s"' % (' '*indent, name, value) -- print '%s>' % (' '*indent) -+ print('%s %s="%s"' % (' '*indent, name, value)) -+ print('%s>' % (' '*indent)) - if node.nodeValue: -- print '%s %s' % (' '*indent, node.nodeValue) -+ print('%s %s' % (' '*indent, node.nodeValue)) - - for sub_node in node.childNodes: - PrettyPrintNode(sub_node, indent=indent+2) -- print '%s</%s>' % (' '*indent, node.nodeName) -+ print('%s</%s>' % (' '*indent, node.nodeName)) - - - def FlattenFilter(node): -@@ -283,8 +285,8 @@ - - # check if we have exactly 1 parameter. - if len(argv) < 2: -- print ('Usage: %s "c:\\path\\to\\vcproj.vcproj" [key1=value1] ' -- '[key2=value2]' % argv[0]) -+ print('Usage: %s "c:\\path\\to\\vcproj.vcproj" [key1=value1] ' -+ '[key2=value2]' % argv[0]) - return 1 - - # Parse the keys -diff --git a/tools/gyp/gyptest.py b/gyptest.py -index 9930e78..1a9ffca 100755 ---- a/tools/gyp/gyptest.py -+++ b/tools/gyp/gyptest.py -@@ -58,7 +58,7 @@ - os.chdir(args.chdir) - - if args.path: -- extra_path = [os.path.abspath(p) for p in opts.path] -+ extra_path = [os.path.abspath(p) for p in args.path] - extra_path = os.pathsep.join(extra_path) - os.environ['PATH'] = extra_path + os.pathsep + os.environ['PATH'] - -diff --git a/tools/gyp/pylib/gyp/MSVSNew.py b/pylib/gyp/MSVSNew.py -index 593f0e5..0445931 100644 ---- a/tools/gyp/pylib/gyp/MSVSNew.py -+++ b/tools/gyp/pylib/gyp/MSVSNew.py -@@ -21,6 +21,13 @@ - _new_md5 = md5.new - - -+try: -+ # cmp was removed in python3. -+ cmp -+except NameError: -+ def cmp(a, b): -+ return (a > b) - (a < b) -+ - # Initialize random number generator - random.seed() - -diff --git a/tools/gyp/pylib/gyp/common.py b/pylib/gyp/common.py -index 1823de8..b268d22 100644 ---- a/tools/gyp/pylib/gyp/common.py -+++ b/tools/gyp/pylib/gyp/common.py -@@ -584,7 +584,7 @@ - graph = {'a': '$(b) $(c)', 'b': 'hi', 'c': '$(b)'} - def GetEdges(node): - return re.findall(r'\$\(([^))]\)', graph[node]) -- print TopologicallySorted(graph.keys(), GetEdges) -+ print(TopologicallySorted(graph.keys(), GetEdges)) - ==> - ['a', 'c', b'] - """ -diff --git a/tools/gyp/pylib/gyp/generator/make.py b/pylib/gyp/generator/make.py -index 2057e3a..8c2827e 100644 ---- a/tools/gyp/pylib/gyp/generator/make.py -+++ b/tools/gyp/pylib/gyp/generator/make.py -@@ -1636,7 +1636,7 @@ - self.WriteDoCmd([self.output_binary], deps, 'touch', part_of_all, - postbuilds=postbuilds) - else: -- print("WARNING: no output for", self.type, target) -+ print("WARNING: no output for", self.type, self.target) - - # Add an alias for each target (if there are any outputs). - # Installable target aliases are created below. -diff --git a/tools/gyp/pylib/gyp/generator/msvs.py b/pylib/gyp/generator/msvs.py -index e8a2b36..9eac028 100644 ---- a/tools/gyp/pylib/gyp/generator/msvs.py -+++ b/tools/gyp/pylib/gyp/generator/msvs.py -@@ -308,10 +308,8 @@ - if names: - return names[0] - else: -- print >> sys.stdout, ( -- 'Warning: No include files found for ' -- 'detected Windows SDK version %s' % (version) -- ) -+ print('Warning: No include files found for ' -+ 'detected Windows SDK version %s' % (version)) - - - def _BuildCommandLineForRuleRaw(spec, cmd, cygwin_shell, has_input_path, -@@ -2065,7 +2063,7 @@ - if generator_flags.get('msvs_error_on_missing_sources', False): - raise GypError(error_message) - else: -- print("Warning: " + error_message, file=sys.stdout) -+ print("Warning: " + error_message) - - - def _GenerateMSBuildFiltersFile(filters_path, source_files, -diff --git a/tools/gyp/pylib/gyp/mac_tool.py b/pylib/gyp/mac_tool.py -index 7d3a8c2..84f8863 100755 ---- a/tools/gyp/pylib/gyp/mac_tool.py -+++ b/tools/gyp/pylib/gyp/mac_tool.py -@@ -670,7 +670,7 @@ - count = len(filelist) - capacity = NextGreaterPowerOf2(count) - strings_offset = 24 + (12 * capacity) -- max_value_length = len(max(filelist.items(), key=lambda (k,v):len(v))[1]) -+ max_value_length = len(max(filelist.items(), key=lambda t: len(t[1]))[1]) - - out = open(output_name, "wb") - out.write(struct.pack('<LHHLLLL', magic, version, _reserved, strings_offset, -diff --git a/tools/gyp/pylib/gyp/xcodeproj_file.py b/pylib/gyp/xcodeproj_file.py -index bd238f6..bc9814d 100644 ---- a/tools/gyp/pylib/gyp/xcodeproj_file.py -+++ b/tools/gyp/pylib/gyp/xcodeproj_file.py -@@ -160,6 +160,13 @@ - except NameError: - basestring = str - -+try: -+ # cmp was removed in python3. -+ cmp -+except NameError: -+ def cmp(a, b): -+ return (a > b) - (a < b) -+ - # See XCObject._EncodeString. This pattern is used to determine when a string - # can be printed unquoted. Strings that match this pattern may be printed - # unquoted. Strings that do not match must be quoted and may be further -diff --git a/tools/gyp/samples/samples b/samples/samples -index 804b618..ff26de3 100755 ---- a/tools/gyp/samples/samples -+++ b/tools/gyp/samples/samples -@@ -4,6 +4,8 @@ - # Use of this source code is governed by a BSD-style license that can be - # found in the LICENSE file. - -+from __future__ import print_function -+ - import os.path - import shutil - import sys -@@ -57,7 +59,7 @@ - - def Main(argv): - if len(argv) != 3 or argv[1] not in ['push', 'pull']: -- print 'Usage: %s push/pull PATH_TO_CHROME' % argv[0] -+ print('Usage: %s push/pull PATH_TO_CHROME' % argv[0]) - return 1 - - path_to_chrome = argv[2] -@@ -66,10 +68,10 @@ - chrome_file = os.path.join(path_to_chrome, g) - local_file = os.path.join(os.path.dirname(argv[0]), os.path.split(g)[1]) - if argv[1] == 'push': -- print 'Copying %s to %s' % (local_file, chrome_file) -+ print('Copying %s to %s' % (local_file, chrome_file)) - shutil.copyfile(local_file, chrome_file) - elif argv[1] == 'pull': -- print 'Copying %s to %s' % (chrome_file, local_file) -+ print('Copying %s to %s' % (chrome_file, local_file)) - shutil.copyfile(chrome_file, local_file) - else: - assert False -diff --git a/tools/gyp/tools/pretty_vcproj.py b/tools/pretty_vcproj.py -index f02e59e..4454d9b 100755 ---- a/tools/gyp/tools/pretty_vcproj.py -+++ b/tools/gyp/tools/pretty_vcproj.py -@@ -22,6 +22,13 @@ - from xml.dom.minidom import parse - from xml.dom.minidom import Node - -+try: -+ # cmp was removed in python3. -+ cmp -+except NameError: -+ def cmp(a, b): -+ return (a > b) - (a < b) -+ - REPLACEMENTS = dict() - ARGUMENTS = None - -@@ -63,7 +70,7 @@ - def PrettyPrintNode(node, indent=0): - if node.nodeType == Node.TEXT_NODE: - if node.data.strip(): -- print '%s%s' % (' '*indent, node.data.strip()) -+ print('%s%s' % (' '*indent, node.data.strip())) - return - - if node.childNodes: -@@ -322,7 +329,6 @@ - - # Finally, we use the prett xml function to print the vcproj back to the - # user. -- #print dom.toprettyxml(newl="\n") - PrettyPrintNode(dom.documentElement) - return 0 - ---- node-v10.15.3/tools/gyp/pylib/gyp/input.py.old 2019-04-02 06:44:13.086310973 +0000 -+++ node-v10.15.3/tools/gyp/pylib/gyp/input.py 2019-04-02 06:45:35.987250735 +0000 -@@ -900,6 +900,9 @@ - (e, contents, build_file)) - - p_stdout, p_stderr = p.communicate('') -+ if getattr(p_stdout, 'decode'): -+ p_stdout = p_stdout.decode('utf-8') -+ p_stderr = p_stderr.decode('utf-8') - - if p.wait() != 0 or p_stderr: - sys.stderr.write(p_stderr) diff --git a/user/node/libatomic.patch b/user/node/libatomic.patch deleted file mode 100644 index fc2e4ffd2..000000000 --- a/user/node/libatomic.patch +++ /dev/null @@ -1,14 +0,0 @@ ---- node-v10.15.3/node.gyp.old 2019-03-05 15:16:32.000000000 +0000 -+++ node-v10.15.3/node.gyp 2019-04-06 13:54:51.204939193 +0000 -@@ -479,6 +479,11 @@ - 'msvs_disabled_warnings!': [4244], - - 'conditions': [ -+ [ 'host_arch=="mips" or host_arch=="mipsel" or host_arch=="ppc"', { -+ 'link_settings': { -+ 'libraries': [ '-latomic' ], -+ }, -+ }], - [ 'node_code_cache_path!=""', { - 'sources': [ '<(node_code_cache_path)' ] - }, { diff --git a/user/node/ppc32.patch b/user/node/ppc32.patch deleted file mode 100644 index 45b051534..000000000 --- a/user/node/ppc32.patch +++ /dev/null @@ -1,29 +0,0 @@ ---- node-v10.15.3/deps/v8/src/libsampler/sampler.cc.old 2019-03-05 15:16:28.000000000 +0000 -+++ node-v10.15.3/deps/v8/src/libsampler/sampler.cc 2019-04-06 13:44:07.224653262 +0000 -@@ -502,9 +502,15 @@ - reinterpret_cast<void*>(ucontext->uc_mcontext.regs->gpr[PT_R31]); - #else - // Some C libraries, notably Musl, define the regs member as a void pointer -+ #if !V8_TARGET_ARCH_32_BIT - state->pc = reinterpret_cast<void*>(ucontext->uc_mcontext.gp_regs[32]); - state->sp = reinterpret_cast<void*>(ucontext->uc_mcontext.gp_regs[1]); - state->fp = reinterpret_cast<void*>(ucontext->uc_mcontext.gp_regs[31]); -+ #else -+ state->pc = reinterpret_cast<void*>(ucontext->uc_mcontext.gregs[32]); -+ state->sp = reinterpret_cast<void*>(ucontext->uc_mcontext.gregs[1]); -+ state->fp = reinterpret_cast<void*>(ucontext->uc_mcontext.gregs[31]); -+ #endif - #endif - #elif V8_HOST_ARCH_S390 - #if V8_TARGET_ARCH_32_BIT ---- node-v10.16.3/configure.py.old 2019-08-15 19:20:03.000000000 +0000 -+++ node-v10.16.3/configure.py 2019-08-16 14:11:23.086489117 +0000 -@@ -848,7 +848,7 @@ - '__MIPSEL__' : 'mipsel', - '__mips__' : 'mips', - '__PPC64__' : 'ppc64', -- '__PPC__' : 'ppc64', -+ '__PPC__' : 'ppc', - '__x86_64__' : 'x64', - '__s390__' : 's390', - '__s390x__' : 's390x', diff --git a/user/node/ppc64.patch b/user/node/ppc64.patch deleted file mode 100644 index f76618562..000000000 --- a/user/node/ppc64.patch +++ /dev/null @@ -1,40 +0,0 @@ ---- a/deps/v8/src/ppc/assembler-ppc.h.old 2019-03-05 15:16:29.000000000 +0000 -+++ b/deps/v8/src/ppc/assembler-ppc.h 2019-04-02 07:05:25.977213735 +0000 -@@ -48,7 +48,8 @@ - #include "src/ppc/constants-ppc.h" - - #if V8_HOST_ARCH_PPC && \ -- (V8_OS_AIX || (V8_TARGET_ARCH_PPC64 && V8_TARGET_BIG_ENDIAN)) -+ (V8_OS_AIX || (V8_TARGET_ARCH_PPC64 && \ -+ (V8_TARGET_BIG_ENDIAN && (!defined(_CALL_ELF) || _CALL_ELF == 1)))) - #define ABI_USES_FUNCTION_DESCRIPTORS 1 - #else - #define ABI_USES_FUNCTION_DESCRIPTORS 0 -@@ -60,13 +61,15 @@ - #define ABI_PASSES_HANDLES_IN_REGS 0 - #endif - --#if !V8_HOST_ARCH_PPC || !V8_TARGET_ARCH_PPC64 || V8_TARGET_LITTLE_ENDIAN -+#if !V8_HOST_ARCH_PPC || !V8_TARGET_ARCH_PPC64 || \ -+ (V8_TARGET_LITTLE_ENDIAN || (defined(_CALL_ELF) && _CALL_ELF == 2)) - #define ABI_RETURNS_OBJECT_PAIRS_IN_REGS 1 - #else - #define ABI_RETURNS_OBJECT_PAIRS_IN_REGS 0 - #endif - --#if !V8_HOST_ARCH_PPC || (V8_TARGET_ARCH_PPC64 && V8_TARGET_LITTLE_ENDIAN) -+#if !V8_HOST_ARCH_PPC || (V8_TARGET_ARCH_PPC64 && \ -+ (V8_TARGET_LITTLE_ENDIAN || (defined(_CALL_ELF) && _CALL_ELF == 2))) - #define ABI_CALL_VIA_IP 1 - #else - #define ABI_CALL_VIA_IP 0 -@@ -220,7 +220,8 @@ - // The following constants describe the stack frame linkage area as - // defined by the ABI. Note that kNumRequiredStackFrameSlots must - // satisfy alignment requirements (rounding up if required). --#if V8_TARGET_ARCH_PPC64 && V8_TARGET_LITTLE_ENDIAN -+#if V8_TARGET_ARCH_PPC64 && (V8_TARGET_LITTLE_ENDIAN || \ -+ defined(_CALL_ELF) && _CALL_ELF == 2) - // [0] back chain - // [1] condition register save area - // [2] link register save area diff --git a/user/node/python3.patch b/user/node/python3.patch deleted file mode 100644 index dca32c7e4..000000000 --- a/user/node/python3.patch +++ /dev/null @@ -1,163 +0,0 @@ ---- node-v10.15.3/tools/configure.d/nodedownload.py (original) -+++ node-v10.15.3/tools/configure.d/nodedownload.py (refactored) -@@ -1,7 +1,10 @@ - #!/usr/bin/env python - # Moved some utilities here from ../../configure - -+try: --import urllib -+ import urllib -+except ImportError: -+ import urllib.request, urllib.parse, urllib.error - import hashlib - import sys - import zipfile -@@ -18,10 +18,10 @@ - spin = ".:|'" - return (spin[c % len(spin)]) - --class ConfigOpener(urllib.FancyURLopener): -+class ConfigOpener(urllib.request.FancyURLopener): - """fancy opener used by retrievefile. Set a UA""" - # append to existing version (UA) -- version = '%s node.js/configure' % urllib.URLopener.version -+ version = '%s node.js/configure' % urllib.request.URLopener.version - - def reporthook(count, size, total): - """internal hook used by retrievefile""" -@@ -36,10 +36,10 @@ - sys.stdout.write(' <%s>\nConnecting...\r' % url) - sys.stdout.flush() - ConfigOpener().retrieve(url, targetfile, reporthook=reporthook) -- print '' # clear the line -+ print('') # clear the line - return targetfile - except: -- print ' ** Error occurred while downloading\n <%s>' % url -+ print(' ** Error occurred while downloading\n <%s>' % url) - raise - - def md5sum(targetfile): -@@ -56,12 +56,12 @@ - """Unpacks packedfile into parent_path. Assumes .zip. Returns parent_path""" - if zipfile.is_zipfile(packedfile): - with contextlib.closing(zipfile.ZipFile(packedfile, 'r')) as icuzip: -- print ' Extracting zipfile: %s' % packedfile -+ print(' Extracting zipfile: %s' % packedfile) - icuzip.extractall(parent_path) - return parent_path - elif tarfile.is_tarfile(packedfile): - with contextlib.closing(tarfile.TarFile.open(packedfile, 'r')) as icuzip: -- print ' Extracting tarfile: %s' % packedfile -+ print(' Extracting tarfile: %s' % packedfile) - icuzip.extractall(parent_path) - return parent_path - else: -@@ -112,16 +112,16 @@ - theRet[anOpt] = True - else: - # future proof: ignore unknown types -- print 'Warning: ignoring unknown --download= type "%s"' % anOpt -+ print('Warning: ignoring unknown --download= type "%s"' % anOpt) - # all done - return theRet - - def candownload(auto_downloads, package): -- if not (package in auto_downloads.keys()): -+ if not (package in list(auto_downloads.keys())): - raise Exception('Internal error: "%s" is not in the --downloads list. Check nodedownload.py' % package) - if auto_downloads[package]: - return True - else: -- print """Warning: Not downloading package "%s". You could pass "--download=all" -- (Windows: "download-all") to try auto-downloading it.""" % package -+ print("""Warning: Not downloading package "%s". You could pass "--download=all" -+ (Windows: "download-all") to try auto-downloading it.""" % package) - return False ---- node-v10.15.3/tools/configure.d/nodedownload.py.old 2019-04-02 00:56:07.533200475 +0000 -+++ node-v10.15.3/tools/configure.d/nodedownload.py 2019-04-02 00:58:09.019947842 +0000 -@@ -6,12 +6,11 @@ - import sys - import zipfile - import tarfile --import fpformat - import contextlib - - def formatSize(amt): - """Format a size as a string in MB""" -- return fpformat.fix(amt / 1024000., 1) -+ return "%{size}.1f" % (amt / 1024000.) - - def spin(c): - """print out an ASCII 'spinner' based on the value of counter 'c'""" ---- node-v10.15.3/configure.py.old 2019-03-05 15:16:24.000000000 +0000 -+++ node-v10.15.3/configure.py 2019-04-02 01:09:04.948394534 +0000 -@@ -649,8 +649,8 @@ - except OSError: - return (False, False, '', '') - -- proc.stdin.write('__clang__ __GNUC__ __GNUC_MINOR__ __GNUC_PATCHLEVEL__ ' -- '__clang_major__ __clang_minor__ __clang_patchlevel__') -+ proc.stdin.write(b'__clang__ __GNUC__ __GNUC_MINOR__ __GNUC_PATCHLEVEL__ ' -+ b'__clang_major__ __clang_minor__ __clang_patchlevel__') - - values = (proc.communicate()[0].split() + ['0'] * 7)[0:7] - is_clang = values[0] == '1' -@@ -727,7 +727,7 @@ - consider adjusting the CC environment variable if you installed - it in a non-standard prefix.''') - -- gas_ret = proc.communicate()[1] -+ gas_ret = proc.communicate()[1].decode('utf-8') - match = re.match(r"GNU assembler version ([2-9]\.[0-9]+)", gas_ret) - - if match: -@@ -794,7 +794,7 @@ - consider adjusting the CC environment variable if you installed - it in a non-standard prefix.''') - -- p.stdin.write('\n') -+ p.stdin.write(b'\n') - out = p.communicate()[0] - - out = str(out).split('\n') -@@ -1351,7 +1351,7 @@ - o['variables']['icu_small'] = b(True) - locs = set(options.with_icu_locales.split(',')) - locs.add('root') # must have root -- o['variables']['icu_locales'] = string.join(locs,',') -+ o['variables']['icu_locales'] = ','.join(locs) - # We will check a bit later if we can use the canned deps/icu-small - elif with_intl == 'full-icu': - # full ICU -@@ -1655,7 +1655,7 @@ - if options.prefix: - config['PREFIX'] = options.prefix - --config = '\n'.join(map('='.join, config.iteritems())) + '\n' -+config = '\n'.join(map('='.join, config.items())) + '\n' - - # On Windows there's no reason to search for a different python binary. - bin_override = None if sys.platform == 'win32' else make_bin_override() ---- node-v10.15.3/configure.py.old 2019-04-02 01:12:29.786049396 +0000 -+++ node-v10.15.3/configure.py 2019-04-02 01:21:08.499637208 +0000 -@@ -634,7 +634,7 @@ - proc = subprocess.Popen( - shlex.split(pkg_config) + ['--silence-errors', flag, pkg], - stdout=subprocess.PIPE) -- val = proc.communicate()[0].strip() -+ val = proc.communicate()[0].strip().decode('utf-8') - except OSError as e: - if e.errno != errno.ENOENT: raise e # Unexpected error. - return (None, None, None, None) # No pkg-config/pkgconf installed. ---- node-v10.15.3/configure.py.old 2019-04-02 01:27:36.437454388 +0000 -+++ node-v10.15.3/configure.py 2019-04-02 01:28:06.954136125 +0000 -@@ -795,7 +795,7 @@ - it in a non-standard prefix.''') - - p.stdin.write(b'\n') -- out = p.communicate()[0] -+ out = p.communicate()[0].decode('utf-8') - - out = str(out).split('\n') - diff --git a/user/node/stack-silliness.patch b/user/node/stack-silliness.patch deleted file mode 100644 index e8d775b73..000000000 --- a/user/node/stack-silliness.patch +++ /dev/null @@ -1,14 +0,0 @@ ---- node-v10.15.3/src/inspector_agent.cc.old 2019-03-05 15:16:32.000000000 +0000 -+++ node-v10.15.3/src/inspector_agent.cc 2019-04-06 02:03:11.082016553 +0000 -@@ -104,9 +104,9 @@ - // Don't shrink the thread's stack on FreeBSD. Said platform decided to - // follow the pthreads specification to the letter rather than in spirit: - // https://lists.freebsd.org/pipermail/freebsd-current/2014-March/048885.html --#ifndef __FreeBSD__ -+#ifdef __GLIBC__ - CHECK_EQ(0, pthread_attr_setstacksize(&attr, PTHREAD_STACK_MIN)); --#endif // __FreeBSD__ -+#endif // __GLIBC__ - CHECK_EQ(0, pthread_attr_setdetachstate(&attr, PTHREAD_CREATE_DETACHED)); - sigset_t sigmask; - // Mask all signals. diff --git a/user/node/v8-python3.patch b/user/node/v8-python3.patch deleted file mode 100644 index 64151be4f..000000000 --- a/user/node/v8-python3.patch +++ /dev/null @@ -1,169 +0,0 @@ ---- node-v10.15.3/deps/v8/tools/node/fetch_deps.py.old 2019-03-05 15:16:31.000000000 +0000 -+++ node-v10.15.3/deps/v8/tools/node/fetch_deps.py 2019-04-02 01:01:00.421250178 +0000 -@@ -9,6 +9,9 @@ - Usage: fetch_deps.py <v8-path> - """ - -+# for py2/py3 compatibility -+from __future__ import print_function -+ - import os - import subprocess - import sys -@@ -51,9 +55,9 @@ - expected_git_dir = os.path.join(v8_path, ".git") - actual_git_dir = git("rev-parse --absolute-git-dir") - if expected_git_dir == actual_git_dir: -- print "V8 is tracked stand-alone by git." -+ print("V8 is tracked stand-alone by git.") - return False -- print "Initializing temporary git repository in v8." -+ print("Initializing temporary git repository in v8.") - git("init") - git("config user.name \"Ada Lovelace\"") - git("config user.email ada@lovela.ce") -@@ -70,7 +74,7 @@ - - temporary_git = EnsureGit(v8_path) - try: -- print "Fetching dependencies." -+ print("Fetching dependencies.") - env = os.environ.copy() - # gclient needs to have depot_tools in the PATH. - env["PATH"] = depot_tools + os.pathsep + env["PATH"] ---- node-v10.15.3/deps/v8/tools/node/node_common.py.old 2019-03-05 15:16:31.000000000 +0000 -+++ node-v10.15.3/deps/v8/tools/node/node_common.py 2019-04-02 01:00:45.522875398 +0000 -@@ -3,11 +3,15 @@ - # Use of this source code is governed by a BSD-style license that can be - # found in the LICENSE file. - -+# for py2/py3 compatibility -+from __future__ import print_function -+ - import os - import pipes - import shutil - import stat - import subprocess -+import sys - - DEPOT_TOOLS_URL = \ - "https://chromium.googlesource.com/chromium/tools/depot_tools.git" -@@ -22,23 +26,27 @@ - except: - pass - if fetch_if_not_exist: -- print "Checking out depot_tools." -+ print("Checking out depot_tools.") - # shell=True needed on Windows to resolve git.bat. - subprocess.check_call("git clone {} {}".format( - pipes.quote(DEPOT_TOOLS_URL), - pipes.quote(depot_tools)), shell=True) -+ # Using check_output to hide warning messages. -+ subprocess.check_output( -+ [sys.executable, gclient_path, "metrics", "--opt-out"], -+ cwd=depot_tools) - return depot_tools - return None - depot_tools = _Get(v8_path) - assert depot_tools is not None -- print "Using depot tools in %s" % depot_tools -+ print("Using depot tools in %s" % depot_tools) - return depot_tools - - def UninitGit(v8_path): -- print "Uninitializing temporary git repository" -+ print("Uninitializing temporary git repository") - target = os.path.join(v8_path, ".git") - if os.path.isdir(target): -- print ">> Cleaning up %s" % target -+ print(">> Cleaning up %s" % target) - def OnRmError(func, path, exec_info): - # This might happen on Windows - os.chmod(path, stat.S_IWRITE) ---- node-v10.15.3/deps/v8/tools/node/update_node.py.old 2019-03-05 15:16:31.000000000 +0000 -+++ node-v10.15.3/deps/v8/tools/node/update_node.py 2019-04-02 01:00:27.184875836 +0000 -@@ -23,6 +23,9 @@ - --with-patch Also include currently staged files in the V8 checkout. - """ - -+# for py2/py3 compatibility -+from __future__ import print_function -+ - import argparse - import os - import shutil -@@ -61,9 +64,9 @@ - # Node.js owns deps/v8/gypfiles in their downstream repository. - FILES_TO_KEEP = [ "gypfiles" ] - - def RunGclient(path): - assert os.path.isdir(path) -- print ">> Running gclient sync" -+ print(">> Running gclient sync") - subprocess.check_call(["gclient", "sync", "--nohooks"], cwd=path) - - def CommitPatch(options): -@@ -74,7 +77,7 @@ - the fake git clone fetch it into node.js. We can leave the commit, as - bot_update will ensure a clean state on each run. - """ -- print ">> Committing patch" -+ print(">> Committing patch") - subprocess.check_call( - ["git", "-c", "user.name=fake", "-c", "user.email=fake@chromium.org", - "commit", "--allow-empty", "-m", "placeholder-commit"], -@@ -84,8 +87,8 @@ - def UpdateTarget(repository, options, files_to_keep): - source = os.path.join(options.v8_path, *repository) - target = os.path.join(options.node_path, TARGET_SUBDIR, *repository) -- print ">> Updating target directory %s" % target -- print ">> from active branch at %s" % source -+ print(">> Updating target directory %s" % target) -+ print(">> from active branch at %s" % source) - if not os.path.exists(target): - os.makedirs(target) - # Remove possible remnants of previous incomplete runs. -@@ -98,10 +101,11 @@ - git_args.append(["add"] + files_to_keep) # add and commit - git_args.append(["commit", "-m", "keep files"]) # files we want to keep - -+ git_args.append(["clean", "-fxd"]) # nuke everything else - git_args.append(["remote", "add", "source", source]) # point to source repo - git_args.append(["fetch", "source", "HEAD"]) # sync to current branch - git_args.append(["checkout", "-f", "FETCH_HEAD"]) # switch to that branch -- git_args.append(["clean", "-fd"]) # delete removed files -+ git_args.append(["clean", "-fxd"]) # delete removed files - - if files_to_keep: - git_args.append(["cherry-pick", "master"]) # restore kept files -@@ -117,17 +120,17 @@ - def UpdateGitIgnore(options): - file_name = os.path.join(options.node_path, TARGET_SUBDIR, ".gitignore") - assert os.path.isfile(file_name) -- print ">> Updating .gitignore with lines" -+ print(">> Updating .gitignore with lines") - with open(file_name) as gitignore: - content = gitignore.readlines() - content = [x.strip() for x in content] - for x in DELETE_FROM_GITIGNORE: - if x in content: -- print "- %s" % x -+ print("- %s" % x) - content.remove(x) - for x in ADD_TO_GITIGNORE: - if x not in content: -- print "+ %s" % x -+ print("+ %s" % x) - content.append(x) - content.sort(key=lambda x: x[1:] if x.startswith("!") else x) - with open(file_name, "w") as gitignore: -@@ -135,7 +138,7 @@ - gitignore.write("%s\n" % x) - - def CreateCommit(options): -- print ">> Creating commit." -+ print(">> Creating commit.") - # Find git hash from source. - githash = subprocess.check_output(["git", "rev-parse", "--short", "HEAD"], - cwd=options.v8_path).strip() |