summaryrefslogtreecommitdiff
path: root/user
diff options
context:
space:
mode:
authorZach van Rijn <me@zv.io>2022-04-28 15:38:29 +0000
committerZach van Rijn <me@zv.io>2022-10-21 18:33:58 -0500
commitcb92b4063486a5ebb52b6b5622f361a35f75714d (patch)
tree1c4c0e641879bde7a84347baec0ebf92b8562105 /user
parent20ff69194828da4a1f125d5e4e9962989060591b (diff)
downloadpackages-cb92b4063486a5ebb52b6b5622f361a35f75714d.tar.gz
packages-cb92b4063486a5ebb52b6b5622f361a35f75714d.tar.bz2
packages-cb92b4063486a5ebb52b6b5622f361a35f75714d.tar.xz
packages-cb92b4063486a5ebb52b6b5622f361a35f75714d.zip
user/node: bump { 10.21.0 --> 16.15.0 }. fixes #604.
Diffstat (limited to 'user')
-rw-r--r--user/node/APKBUILD66
-rw-r--r--user/node/fix-all-caps-bool.patch84
-rw-r--r--user/node/gyp-python3.patch3862
-rw-r--r--user/node/libatomic.patch14
-rw-r--r--user/node/ppc32.patch29
-rw-r--r--user/node/ppc64.patch40
-rw-r--r--user/node/python3.patch163
-rw-r--r--user/node/stack-silliness.patch14
-rw-r--r--user/node/v8-python3.patch169
9 files changed, 15 insertions, 4426 deletions
diff --git a/user/node/APKBUILD b/user/node/APKBUILD
index 578040d07..954eee721 100644
--- a/user/node/APKBUILD
+++ b/user/node/APKBUILD
@@ -1,8 +1,8 @@
# Contributor: A. Wilcox <awilfox@adelielinux.org>
# Maintainer: A. Wilcox <awilfox@adelielinux.org>
pkgname=node
-pkgver=10.21.0
-pkgrel=1
+pkgver=16.15.0
+pkgrel=0
pkgdesc="JavaScript runtime"
url="https://nodejs.org/"
arch="all"
@@ -10,16 +10,9 @@ options="net" # Required in check()
license="MIT AND ICU AND BSD-3-Clause AND BSD-2-Clause AND ISC AND Public-Domain AND Zlib AND Artistic-2.0 AND Apache-2.0 AND CC0-1.0"
depends=""
makedepends="c-ares-dev http-parser-dev icu-dev libexecinfo-dev libuv-dev
- nghttp2-dev>=1.41 openssl-dev python3 zlib-dev"
+ nghttp2-dev openssl-dev python3 zlib-dev samurai"
subpackages="$pkgname-dev $pkgname-doc"
-source="https://nodejs.org/download/release/v$pkgver/node-v$pkgver.tar.xz
- https://www.python.org/ftp/python/2.7.15/Python-2.7.15.tar.xz
- libatomic.patch
- ppc32.patch
- ppc64.patch
- stack-silliness.patch
- fix-all-caps-bool.patch
- "
+source="https://nodejs.org/download/release/v$pkgver/node-v$pkgver.tar.xz"
builddir="$srcdir/$pkgname-v$pkgver"
# secfixes:
@@ -36,59 +29,30 @@ builddir="$srcdir/$pkgname-v$pkgver"
# - CVE-2020-7598
# - CVE-2020-8174
-unpack() {
- default_unpack
- [ -z $SKIP_PYTHON ] || return 0
-
- # TODO: when bumping to 12.x, python3 should be usable
- msg "Killing all remaining hope for humanity and building Python 2..."
- cd "$srcdir/Python-2.7.15"
- [ -d ../python ] && rm -r ../python
- # 19:39 <+solar> just make the firefox build process build its own py2 copy
- # 20:03 <calvin> TheWilfox: there's always violence
- ./configure --prefix="$srcdir/python"
- make -j $JOBS
- make -j $JOBS install
-}
-
build() {
- export PATH="$srcdir/python/bin:$PATH"
- python ./configure.py \
+ python3 configure.py \
--prefix=/usr \
- --with-intl=system-icu \
+ --shared-zlib \
+ --shared-openssl \
--shared-cares \
- --shared-http-parser \
- --shared-libuv \
--shared-nghttp2 \
- --shared-openssl \
+ --ninja \
--openssl-use-def-ca-store \
- --shared-zlib
- # keep DESTDIR set, to avoid a full rebuild in package()
- make DESTDIR="$pkgdir"
+ --with-icu-default-data-dir=$(icu-config --icudatadir) \
+ --with-intl=small-icu \
+ --without-corepack \
+ --without-npm
+
+ make BUILDTYPE=Release DESTDIR="$pkgdir"
}
check() {
- case "$CARCH" in
- pmmx)
- # https://bts.adelielinux.org/show_bug.cgi?id=306
- _skip="parallel/test-http-invalid-te,parallel/test-worker-stdio"
- ;;
- esac
-
- export PATH="$srcdir/python/bin:$PATH"
make DESTDIR="$pkgdir" test-only \
${_skip:+CI_SKIP_TESTS="$_skip"}
}
package() {
- export PATH="$srcdir/python/bin:$PATH"
make DESTDIR="$pkgdir" install
}
-sha512sums="613d3c1bca79ea5f127dc6793de2b5cfdfa056c01ec092e3b7ee79205894b21ca5ec4a367265122641dd1d360c675cfb36a4f7892894194ddd18abd1b2206544 node-v10.21.0.tar.xz
-27ea43eb45fc68f3d2469d5f07636e10801dee11635a430ec8ec922ed790bb426b072da94df885e4dfa1ea8b7a24f2f56dd92f9b0f51e162330f161216bd6de6 Python-2.7.15.tar.xz
-8f64922d586bce9d82c83042a989739cc55ecc5e015778cdfbda21c257aa50527ddb18740985bcb2068e4a749b71eb8a135d9a8152b374d361589df7f33c9b60 libatomic.patch
-6d37794c7c78ef92ebb845852af780e22dc8c14653b63a8609c21ab6860877b9dffc5cf856a8516b7978ec704f312c0627075c6440ace55d039f95bdc4c85add ppc32.patch
-583326353de5b0ac14a6c42321f6b031bd943a80550624794e15bd7526470f67bfa14a66558db3c94b4ee2db3053d2e4efed2117f4e7b6dca3c59c171048c094 ppc64.patch
-3ea09e36ed0cc31e0475ebc9c92b7609b70e9c1637c5db6c92cf1d6363fb8c6f884ffa20dd81054ca390b721695185327d80c9eeff0688a959e9d46947602471 stack-silliness.patch
-a29d21c7336631aa099ef2fdf8ed63f1f75d10127f26103b9638cbeab5837f00603996f05bbce4c19694b19cd08f7f0fd516f011827a7fb912177334076f1bf0 fix-all-caps-bool.patch"
+sha512sums="c85cf4a94a0dccdcf358a3e0383499fe1fd74ac0e7b6aa888e9524c070ae8be68b9f718c8c0940d51115bbc908202cd7819e370507b7191facd375a5be231c69 node-v16.15.0.tar.xz"
diff --git a/user/node/fix-all-caps-bool.patch b/user/node/fix-all-caps-bool.patch
deleted file mode 100644
index fa1007a3b..000000000
--- a/user/node/fix-all-caps-bool.patch
+++ /dev/null
@@ -1,84 +0,0 @@
-diff -ur a/deps/v8/src/objects/intl-objects.cc b/deps/v8/src/objects/intl-objects.cc
---- a/deps/v8/src/objects/intl-objects.cc 2022-04-22 08:12:51.454294680 +0000
-+++ b/deps/v8/src/objects/intl-objects.cc 2022-04-22 08:19:29.655077696 +0000
-@@ -243,7 +243,7 @@
- char result[ULOC_FULLNAME_CAPACITY];
- status = U_ZERO_ERROR;
- uloc_toLanguageTag(icu_locale.getName(), result, ULOC_FULLNAME_CAPACITY,
-- FALSE, &status);
-+ false, &status);
- if (U_SUCCESS(status)) {
- JSObject::SetProperty(resolved, factory->NewStringFromStaticChars("locale"),
- factory->NewStringFromAsciiChecked(result),
-@@ -422,7 +422,7 @@
- char result[ULOC_FULLNAME_CAPACITY];
- UErrorCode status = U_ZERO_ERROR;
- uloc_toLanguageTag(icu_locale.getName(), result, ULOC_FULLNAME_CAPACITY,
-- FALSE, &status);
-+ false, &status);
- if (U_SUCCESS(status)) {
- JSObject::SetProperty(resolved, factory->NewStringFromStaticChars("locale"),
- factory->NewStringFromAsciiChecked(result),
-@@ -657,7 +657,7 @@
- char result[ULOC_FULLNAME_CAPACITY];
- status = U_ZERO_ERROR;
- uloc_toLanguageTag(icu_locale.getName(), result, ULOC_FULLNAME_CAPACITY,
-- FALSE, &status);
-+ false, &status);
- if (U_SUCCESS(status)) {
- JSObject::SetProperty(resolved, factory->NewStringFromStaticChars("locale"),
- factory->NewStringFromAsciiChecked(result),
-@@ -796,7 +796,7 @@
- char result[ULOC_FULLNAME_CAPACITY];
- status = U_ZERO_ERROR;
- uloc_toLanguageTag(icu_locale.getName(), result, ULOC_FULLNAME_CAPACITY,
-- FALSE, &status);
-+ false, &status);
- if (U_SUCCESS(status)) {
- JSObject::SetProperty(resolved, factory->NewStringFromStaticChars("locale"),
- factory->NewStringFromAsciiChecked(result),
-diff -ur a/deps/v8/src/runtime/runtime-intl.cc b/deps/v8/src/runtime/runtime-intl.cc
---- a/deps/v8/src/runtime/runtime-intl.cc 2022-04-22 08:27:09.577527378 +0000
-+++ b/deps/v8/src/runtime/runtime-intl.cc 2022-04-22 08:27:24.397928433 +0000
-@@ -85,7 +85,7 @@
- char result[ULOC_FULLNAME_CAPACITY];
-
- // Force strict BCP47 rules.
-- uloc_toLanguageTag(icu_result, result, ULOC_FULLNAME_CAPACITY, TRUE, &error);
-+ uloc_toLanguageTag(icu_result, result, ULOC_FULLNAME_CAPACITY, true, &error);
-
- if (U_FAILURE(error) || error == U_STRING_NOT_TERMINATED_WARNING) {
- return *factory->NewStringFromAsciiChecked(kInvalidTag);
-diff -ur a/deps/v8/src/runtime/runtime-intl.cc b/deps/v8/src/runtime/runtime-intl.cc
---- a/deps/v8/src/runtime/runtime-intl.cc 2022-04-22 08:29:09.750779385 +0000
-+++ b/deps/v8/src/runtime/runtime-intl.cc 2022-04-22 08:29:41.901649421 +0000
-@@ -132,7 +132,7 @@
-
- error = U_ZERO_ERROR;
- // No need to force strict BCP47 rules.
-- uloc_toLanguageTag(icu_name, result, ULOC_FULLNAME_CAPACITY, FALSE, &error);
-+ uloc_toLanguageTag(icu_name, result, ULOC_FULLNAME_CAPACITY, false, &error);
- if (U_FAILURE(error) || error == U_STRING_NOT_TERMINATED_WARNING) {
- // This shouldn't happen, but lets not break the user.
- continue;
-@@ -164,7 +164,7 @@
- char result[ULOC_FULLNAME_CAPACITY];
- UErrorCode status = U_ZERO_ERROR;
- uloc_toLanguageTag(default_locale.getName(), result, ULOC_FULLNAME_CAPACITY,
-- FALSE, &status);
-+ false, &status);
- if (U_SUCCESS(status)) {
- return *factory->NewStringFromAsciiChecked(result);
- }
-@@ -399,9 +399,9 @@
- std::unique_ptr<uc16[]> sap1;
- std::unique_ptr<uc16[]> sap2;
- icu::UnicodeString string_val1(
-- FALSE, GetUCharBufferFromFlat(flat1, &sap1, length1), length1);
-+ false, GetUCharBufferFromFlat(flat1, &sap1, length1), length1);
- icu::UnicodeString string_val2(
-- FALSE, GetUCharBufferFromFlat(flat2, &sap2, length2), length2);
-+ false, GetUCharBufferFromFlat(flat2, &sap2, length2), length2);
- result = collator->compare(string_val1, string_val2, status);
- }
- if (U_FAILURE(status)) return isolate->ThrowIllegalOperation();
diff --git a/user/node/gyp-python3.patch b/user/node/gyp-python3.patch
deleted file mode 100644
index e52ef19ec..000000000
--- a/user/node/gyp-python3.patch
+++ /dev/null
@@ -1,3862 +0,0 @@
-diff --git a/tools/gyp/pylib/gyp/input.py b/pylib/gyp/input.py
-index a046a15..21b4606 100644
---- a/tools/gyp/pylib/gyp/input.py
-+++ b/tools/gyp/pylib/gyp/input.py
-@@ -2,14 +2,8 @@
- # Use of this source code is governed by a BSD-style license that can be
- # found in the LICENSE file.
-
--from compiler.ast import Const
--from compiler.ast import Dict
--from compiler.ast import Discard
--from compiler.ast import List
--from compiler.ast import Module
--from compiler.ast import Node
--from compiler.ast import Stmt
--import compiler
-+import ast
-+
- import gyp.common
- import gyp.simple_copy
- import multiprocessing
-@@ -183,43 +177,38 @@
- Note that this is slower than eval() is.
- """
-
-- ast = compiler.parse(file_contents)
-- assert isinstance(ast, Module)
-- c1 = ast.getChildren()
-- assert c1[0] is None
-- assert isinstance(c1[1], Stmt)
-- c2 = c1[1].getChildren()
-- assert isinstance(c2[0], Discard)
-- c3 = c2[0].getChildren()
-- assert len(c3) == 1
-- return CheckNode(c3[0], [])
-+ syntax_tree = ast.parse(file_contents)
-+ assert isinstance(syntax_tree, ast.Module)
-+ c1 = syntax_tree.body
-+ assert len(c1) == 1
-+ c2 = c1[0]
-+ assert isinstance(c2, ast.Expr)
-+ return CheckNode(c2.value, [])
-
-
- def CheckNode(node, keypath):
-- if isinstance(node, Dict):
-- c = node.getChildren()
-+ if isinstance(node, ast.Dict):
- dict = {}
-- for n in range(0, len(c), 2):
-- assert isinstance(c[n], Const)
-- key = c[n].getChildren()[0]
-+ for key, value in zip(node.keys, node.values):
-+ assert isinstance(key, ast.Str)
-+ key = key.s
- if key in dict:
- raise GypError("Key '" + key + "' repeated at level " +
- repr(len(keypath) + 1) + " with key path '" +
- '.'.join(keypath) + "'")
- kp = list(keypath) # Make a copy of the list for descending this node.
- kp.append(key)
-- dict[key] = CheckNode(c[n + 1], kp)
-+ dict[key] = CheckNode(value, kp)
- return dict
-- elif isinstance(node, List):
-- c = node.getChildren()
-+ elif isinstance(node, ast.List):
- children = []
-- for index, child in enumerate(c):
-+ for index, child in enumerate(node.elts):
- kp = list(keypath) # Copy list.
- kp.append(repr(index))
- children.append(CheckNode(child, kp))
- return children
-- elif isinstance(node, Const):
-- return node.getChildren()[0]
-+ elif isinstance(node, ast.Str):
-+ return node.s
- else:
- raise TypeError("Unknown AST node at key path '" + '.'.join(keypath) +
- "': " + repr(node))
-diff --git a/tools/gyp/pylib/gyp/simple_copy.py b/pylib/gyp/simple_copy.py
-index 74c98c5..eaf5f8b 100644
---- a/tools/gyp/pylib/gyp/simple_copy.py
-+++ b/tools/gyp/pylib/gyp/simple_copy.py
-@@ -28,8 +28,19 @@
- def _deepcopy_atomic(x):
- return x
-
--for x in (type(None), int, long, float,
-- bool, str, unicode, type):
-+try:
-+ _string_types = (str, unicode)
-+# There's no unicode in python3
-+except NameError:
-+ _string_types = (str, )
-+
-+try:
-+ _integer_types = (int, long)
-+# There's no long in python3
-+except NameError:
-+ _integer_types = (int, )
-+
-+for x in (type(None), float, bool, type) + _integer_types + _string_types:
- d[x] = _deepcopy_atomic
-
- def _deepcopy_list(x):
-diff --git a/tools/gyp/PRESUBMIT.py b/PRESUBMIT.py
-index 4bc1b8c..5ee669b 100644
---- a/tools/gyp/PRESUBMIT.py
-+++ b/tools/gyp/PRESUBMIT.py
-@@ -76,8 +76,7 @@
- def _LicenseHeader(input_api):
- # Accept any year number from 2009 to the current year.
- current_year = int(input_api.time.strftime('%Y'))
-- allowed_years = (str(s) for s in reversed(xrange(2009, current_year + 1)))
--
-+ allowed_years = (str(s) for s in reversed(range(2009, current_year + 1)))
- years_re = '(' + '|'.join(allowed_years) + ')'
-
- # The (c) is deprecated, but tolerate it until it's removed from all files.
-diff --git a/tools/gyp/README.md b/README.md
-index c0d73ac..b4766c9 100644
---- a/tools/gyp/README.md
-+++ b/tools/gyp/README.md
-@@ -1,4 +1,5 @@
- GYP can Generate Your Projects.
- ===================================
-
--Documents are available at [gyp.gsrc.io](https://gyp.gsrc.io), or you can check out ```md-pages``` branch to read those documents offline.
-+Documents are available at [gyp.gsrc.io](https://gyp.gsrc.io), or you can
-+check out ```md-pages``` branch to read those documents offline.
-diff --git a/tools/gyp/buildbot/buildbot_run.py b/buildbot/buildbot_run.py
-index 9a2b71f..8941652 100755
---- a/tools/gyp/buildbot/buildbot_run.py
-+++ b/tools/gyp/buildbot/buildbot_run.py
-@@ -5,6 +5,8 @@
-
- """Argument-less script to select what to run on the buildbots."""
-
-+from __future__ import print_function
-+
- import os
- import shutil
- import subprocess
-@@ -24,14 +26,14 @@
- with open(os.devnull) as devnull_fd:
- retcode = subprocess.call(stdin=devnull_fd, *args, **kwargs)
- if retcode != 0:
-- print '@@@STEP_EXCEPTION@@@'
-+ print('@@@STEP_EXCEPTION@@@')
- sys.exit(1)
-
-
- def PrepareCmake():
- """Build CMake 2.8.8 since the version in Precise is 2.8.7."""
- if os.environ['BUILDBOT_CLOBBER'] == '1':
-- print '@@@BUILD_STEP Clobber CMake checkout@@@'
-+ print('@@@BUILD_STEP Clobber CMake checkout@@@')
- shutil.rmtree(CMAKE_DIR)
-
- # We always build CMake 2.8.8, so no need to do anything
-@@ -39,10 +41,10 @@
- if os.path.isdir(CMAKE_DIR):
- return
-
-- print '@@@BUILD_STEP Initialize CMake checkout@@@'
-+ print('@@@BUILD_STEP Initialize CMake checkout@@@')
- os.mkdir(CMAKE_DIR)
-
-- print '@@@BUILD_STEP Sync CMake@@@'
-+ print('@@@BUILD_STEP Sync CMake@@@')
- CallSubProcess(
- ['git', 'clone',
- '--depth', '1',
-@@ -53,7 +55,7 @@
- CMAKE_DIR],
- cwd=CMAKE_DIR)
-
-- print '@@@BUILD_STEP Build CMake@@@'
-+ print('@@@BUILD_STEP Build CMake@@@')
- CallSubProcess(
- ['/bin/bash', 'bootstrap', '--prefix=%s' % CMAKE_DIR],
- cwd=CMAKE_DIR)
-@@ -74,7 +76,7 @@
- if not format:
- format = title
-
-- print '@@@BUILD_STEP ' + title + '@@@'
-+ print('@@@BUILD_STEP ' + title + '@@@')
- sys.stdout.flush()
- env = os.environ.copy()
- if msvs_version:
-@@ -89,17 +91,17 @@
- retcode = subprocess.call(command, cwd=ROOT_DIR, env=env, shell=True)
- if retcode:
- # Emit failure tag, and keep going.
-- print '@@@STEP_FAILURE@@@'
-+ print('@@@STEP_FAILURE@@@')
- return 1
- return 0
-
-
- def GypBuild():
- # Dump out/ directory.
-- print '@@@BUILD_STEP cleanup@@@'
-- print 'Removing %s...' % OUT_DIR
-+ print('@@@BUILD_STEP cleanup@@@')
-+ print('Removing %s...' % OUT_DIR)
- shutil.rmtree(OUT_DIR, ignore_errors=True)
-- print 'Done.'
-+ print('Done.')
-
- retcode = 0
- if sys.platform.startswith('linux'):
-@@ -128,7 +130,7 @@
- # after the build proper that could be used for cumulative failures),
- # use that instead of this. This isolates the final return value so
- # that it isn't misattributed to the last stage.
-- print '@@@BUILD_STEP failures@@@'
-+ print('@@@BUILD_STEP failures@@@')
- sys.exit(retcode)
-
-
-diff --git a/tools/gyp/pylib/gyp/MSVSSettings.py b/pylib/gyp/MSVSSettings.py
-index 8ae1918..1d2e25a 100644
---- a/tools/gyp/pylib/gyp/MSVSSettings.py
-+++ b/tools/gyp/pylib/gyp/MSVSSettings.py
-@@ -14,9 +14,17 @@
- MSBuild install directory, e.g. c:\Program Files (x86)\MSBuild
- """
-
-+from __future__ import print_function
-+
- import sys
- import re
-
-+try:
-+ # basestring was removed in python3.
-+ basestring
-+except NameError:
-+ basestring = str
-+
- # Dictionaries of settings validators. The key is the tool name, the value is
- # a dictionary mapping setting names to validation functions.
- _msvs_validators = {}
-@@ -400,7 +408,7 @@
-
- if unrecognized:
- # We don't know this setting. Give a warning.
-- print >> stderr, error_msg
-+ print(error_msg, file=stderr)
-
-
- def FixVCMacroSlashes(s):
-@@ -433,7 +441,7 @@
- '$(PlatformName)': '$(Platform)',
- '$(SafeInputName)': '%(Filename)',
- }
-- for old, new in replace_map.iteritems():
-+ for old, new in replace_map.items():
- s = s.replace(old, new)
- s = FixVCMacroSlashes(s)
- return s
-@@ -453,17 +461,18 @@
- dictionaries of settings and their values.
- """
- msbuild_settings = {}
-- for msvs_tool_name, msvs_tool_settings in msvs_settings.iteritems():
-+ for msvs_tool_name, msvs_tool_settings in msvs_settings.items():
- if msvs_tool_name in _msvs_to_msbuild_converters:
- msvs_tool = _msvs_to_msbuild_converters[msvs_tool_name]
-- for msvs_setting, msvs_value in msvs_tool_settings.iteritems():
-+ for msvs_setting, msvs_value in msvs_tool_settings.items():
- if msvs_setting in msvs_tool:
- # Invoke the translation function.
- try:
- msvs_tool[msvs_setting](msvs_value, msbuild_settings)
-- except ValueError, e:
-- print >> stderr, ('Warning: while converting %s/%s to MSBuild, '
-- '%s' % (msvs_tool_name, msvs_setting, e))
-+ except ValueError as e:
-+ print(('Warning: while converting %s/%s to MSBuild, '
-+ '%s' % (msvs_tool_name, msvs_setting, e)),
-+ file=stderr)
- else:
- _ValidateExclusionSetting(msvs_setting,
- msvs_tool,
-@@ -472,8 +481,8 @@
- (msvs_tool_name, msvs_setting)),
- stderr)
- else:
-- print >> stderr, ('Warning: unrecognized tool %s while converting to '
-- 'MSBuild.' % msvs_tool_name)
-+ print(('Warning: unrecognized tool %s while converting to '
-+ 'MSBuild.' % msvs_tool_name), file=stderr)
- return msbuild_settings
-
-
-@@ -513,13 +522,13 @@
- for tool_name in settings:
- if tool_name in validators:
- tool_validators = validators[tool_name]
-- for setting, value in settings[tool_name].iteritems():
-+ for setting, value in settings[tool_name].items():
- if setting in tool_validators:
- try:
- tool_validators[setting](value)
-- except ValueError, e:
-- print >> stderr, ('Warning: for %s/%s, %s' %
-- (tool_name, setting, e))
-+ except ValueError as e:
-+ print(('Warning: for %s/%s, %s' %
-+ (tool_name, setting, e)), file=stderr)
- else:
- _ValidateExclusionSetting(setting,
- tool_validators,
-@@ -528,7 +537,7 @@
- stderr)
-
- else:
-- print >> stderr, ('Warning: unrecognized tool %s' % tool_name)
-+ print(('Warning: unrecognized tool %s' % tool_name), file=stderr)
-
-
- # MSVS and MBuild names of the tools.
-diff --git a/tools/gyp/pylib/gyp/MSVSSettings_test.py b/pylib/gyp/MSVSSettings_test.py
-index bf6ea6b..73ed25e 100755
---- a/tools/gyp/pylib/gyp/MSVSSettings_test.py
-+++ b/tools/gyp/pylib/gyp/MSVSSettings_test.py
-@@ -6,7 +6,10 @@
-
- """Unit tests for the MSVSSettings.py file."""
-
--import StringIO
-+try:
-+ from StringIO import StringIO
-+except ImportError:
-+ from io import StringIO
- import unittest
- import gyp.MSVSSettings as MSVSSettings
-
-@@ -14,7 +17,7 @@
- class TestSequenceFunctions(unittest.TestCase):
-
- def setUp(self):
-- self.stderr = StringIO.StringIO()
-+ self.stderr = StringIO()
-
- def _ExpectedWarnings(self, expected):
- """Compares recorded lines to expected warnings."""
-diff --git a/tools/gyp/pylib/gyp/MSVSUserFile.py b/pylib/gyp/MSVSUserFile.py
-index 6c07e9a..2264d64 100644
---- a/tools/gyp/pylib/gyp/MSVSUserFile.py
-+++ b/tools/gyp/pylib/gyp/MSVSUserFile.py
-@@ -91,7 +91,7 @@
-
- if environment and isinstance(environment, dict):
- env_list = ['%s="%s"' % (key, val)
-- for (key,val) in environment.iteritems()]
-+ for (key,val) in environment.items()]
- environment = ' '.join(env_list)
- else:
- environment = ''
-@@ -135,7 +135,7 @@
- def WriteIfChanged(self):
- """Writes the user file."""
- configs = ['Configurations']
-- for config, spec in sorted(self.configurations.iteritems()):
-+ for config, spec in sorted(self.configurations.items()):
- configs.append(spec)
-
- content = ['VisualStudioUserFile',
-diff --git a/tools/gyp/pylib/gyp/MSVSUtil.py b/pylib/gyp/MSVSUtil.py
-index 96dea6c..f24530b 100644
---- a/tools/gyp/pylib/gyp/MSVSUtil.py
-+++ b/tools/gyp/pylib/gyp/MSVSUtil.py
-@@ -236,7 +236,7 @@
-
- # Set up the shim to output its PDB to the same location as the final linker
- # target.
-- for config_name, config in shim_dict.get('configurations').iteritems():
-+ for config_name, config in shim_dict.get('configurations').items():
- pdb_path = _GetPdbPath(target_dict, config_name, vars)
-
- # A few keys that we don't want to propagate.
-diff --git a/tools/gyp/pylib/gyp/MSVSVersion.py b/pylib/gyp/MSVSVersion.py
-index 44b958d..5f316b6 100644
---- a/tools/gyp/pylib/gyp/MSVSVersion.py
-+++ b/tools/gyp/pylib/gyp/MSVSVersion.py
-@@ -189,7 +189,7 @@
- text = None
- try:
- text = _RegistryQueryBase('Sysnative', key, value)
-- except OSError, e:
-+ except OSError as e:
- if e.errno == errno.ENOENT:
- text = _RegistryQueryBase('System32', key, value)
- else:
-@@ -207,12 +207,15 @@
- contents of the registry key's value, or None on failure. Throws
- ImportError if _winreg is unavailable.
- """
-- import _winreg
-+ try:
-+ import _winreg as winreg
-+ except ImportError:
-+ import winreg
- try:
- root, subkey = key.split('\\', 1)
- assert root == 'HKLM' # Only need HKLM for now.
-- with _winreg.OpenKey(_winreg.HKEY_LOCAL_MACHINE, subkey) as hkey:
-- return _winreg.QueryValueEx(hkey, value)[0]
-+ with winreg.OpenKey(winreg.HKEY_LOCAL_MACHINE, subkey) as hkey:
-+ return winreg.QueryValueEx(hkey, value)[0]
- except WindowsError:
- return None
-
-diff --git a/tools/gyp/pylib/gyp/__init__.py b/pylib/gyp/__init__.py
-index 668f38b..e038151 100755
---- a/tools/gyp/pylib/gyp/__init__.py
-+++ b/tools/gyp/pylib/gyp/__init__.py
-@@ -4,6 +4,8 @@
- # Use of this source code is governed by a BSD-style license that can be
- # found in the LICENSE file.
-
-+from __future__ import print_function
-+
- import copy
- import gyp.input
- import optparse
-@@ -14,6 +16,12 @@
- import traceback
- from gyp.common import GypError
-
-+try:
-+ # basestring was removed in python3.
-+ basestring
-+except NameError:
-+ basestring = str
-+
- # Default debug modes for GYP
- debug = {}
-
-@@ -22,7 +30,6 @@
- DEBUG_VARIABLES = 'variables'
- DEBUG_INCLUDES = 'includes'
-
--
- def DebugOutput(mode, message, *args):
- if 'all' in gyp.debug or mode in gyp.debug:
- ctx = ('unknown', 0, 'unknown')
-@@ -34,8 +41,8 @@
- pass
- if args:
- message %= args
-- print '%s:%s:%d:%s %s' % (mode.upper(), os.path.basename(ctx[0]),
-- ctx[1], ctx[2], message)
-+ print('%s:%s:%d:%s %s' % (mode.upper(), os.path.basename(ctx[0]),
-+ ctx[1], ctx[2], message))
-
- def FindBuildFiles():
- extension = '.gyp'
-@@ -207,7 +214,7 @@
- # We always want to ignore the environment when regenerating, to avoid
- # duplicate or changed flags in the environment at the time of regeneration.
- flags = ['--ignore-environment']
-- for name, metadata in options._regeneration_metadata.iteritems():
-+ for name, metadata in options._regeneration_metadata.items():
- opt = metadata['opt']
- value = getattr(options, name)
- value_predicate = metadata['type'] == 'path' and FixPath or Noop
-@@ -226,12 +233,13 @@
- (action == 'store_false' and not value)):
- flags.append(opt)
- elif options.use_environment and env_name:
-- print >>sys.stderr, ('Warning: environment regeneration unimplemented '
-+ print(('Warning: environment regeneration unimplemented '
- 'for %s flag %r env_name %r' % (action, opt,
-- env_name))
-+ env_name)),
-+ file=sys.stderr)
- else:
-- print >>sys.stderr, ('Warning: regeneration unimplemented for action %r '
-- 'flag %r' % (action, opt))
-+ print(('Warning: regeneration unimplemented for action %r '
-+ 'flag %r' % (action, opt)), file=sys.stderr)
-
- return flags
-
-@@ -431,12 +439,11 @@
- for build_file in build_files:
- build_file_dir = os.path.abspath(os.path.dirname(build_file))
- build_file_dir_components = build_file_dir.split(os.path.sep)
-- components_len = len(build_file_dir_components)
-- for index in xrange(components_len - 1, -1, -1):
-- if build_file_dir_components[index] == 'src':
-+ for component in reversed(build_file_dir_components):
-+ if component == 'src':
- options.depth = os.path.sep.join(build_file_dir_components)
- break
-- del build_file_dir_components[index]
-+ del build_file_dir_components[-1]
-
- # If the inner loop found something, break without advancing to another
- # build file.
-@@ -475,7 +482,7 @@
- if home_dot_gyp != None:
- default_include = os.path.join(home_dot_gyp, 'include.gypi')
- if os.path.exists(default_include):
-- print 'Using overrides found in ' + default_include
-+ print('Using overrides found in ' + default_include)
- includes.append(default_include)
-
- # Command-line --include files come after the default include.
-@@ -490,7 +497,7 @@
- if options.generator_flags:
- gen_flags += options.generator_flags
- generator_flags = NameValueListToDict(gen_flags)
-- if DEBUG_GENERAL in gyp.debug.keys():
-+ if DEBUG_GENERAL in gyp.debug:
- DebugOutput(DEBUG_GENERAL, "generator_flags: %s", generator_flags)
-
- # Generate all requested formats (use a set in case we got one format request
-@@ -523,7 +530,7 @@
- generator.GenerateOutput(flat_list, targets, data, params)
-
- if options.configs:
-- valid_configs = targets[flat_list[0]]['configurations'].keys()
-+ valid_configs = targets[flat_list[0]]['configurations']
- for conf in options.configs:
- if conf not in valid_configs:
- raise GypError('Invalid config specified via --build: %s' % conf)
-@@ -536,7 +543,7 @@
- def main(args):
- try:
- return gyp_main(args)
-- except GypError, e:
-+ except GypError as e:
- sys.stderr.write("gyp: %s\n" % e)
- return 1
-
-diff --git a/tools/gyp/pylib/gyp/common.py b/pylib/gyp/common.py
-index 1b245ec..1823de8 100644
---- a/tools/gyp/pylib/gyp/common.py
-+++ b/tools/gyp/pylib/gyp/common.py
-@@ -345,7 +345,7 @@
- prefix=os.path.split(filename)[1] + '.gyp.',
- dir=os.path.split(filename)[0])
- try:
-- self.tmp_file = os.fdopen(tmp_fd, 'wb')
-+ self.tmp_file = os.fdopen(tmp_fd, 'w')
- except Exception:
- # Don't leave turds behind.
- os.unlink(self.tmp_path)
-@@ -363,7 +363,7 @@
- same = False
- try:
- same = filecmp.cmp(self.tmp_path, filename, False)
-- except OSError, e:
-+ except OSError as e:
- if e.errno != errno.ENOENT:
- raise
-
-@@ -382,9 +382,9 @@
- #
- # No way to get the umask without setting a new one? Set a safe one
- # and then set it back to the old value.
-- umask = os.umask(077)
-+ umask = os.umask(0o77)
- os.umask(umask)
-- os.chmod(self.tmp_path, 0666 & ~umask)
-+ os.chmod(self.tmp_path, 0o666 & ~umask)
- if sys.platform == 'win32' and os.path.exists(filename):
- # NOTE: on windows (but not cygwin) rename will not replace an
- # existing file, so it must be preceded with a remove. Sadly there
-@@ -471,7 +471,7 @@
- ''.join([source[0], header] + source[1:]))
-
- # Make file executable.
-- os.chmod(tool_path, 0755)
-+ os.chmod(tool_path, 0o755)
-
-
- # From Alex Martelli,
-diff --git a/tools/gyp/pylib/gyp/common_test.py b/pylib/gyp/common_test.py
-index ad6f9a1..0b8ada3 100755
---- a/tools/gyp/pylib/gyp/common_test.py
-+++ b/tools/gyp/pylib/gyp/common_test.py
-@@ -63,6 +63,7 @@
- self.assertFlavor('solaris', 'sunos' , {});
- self.assertFlavor('linux' , 'linux2' , {});
- self.assertFlavor('linux' , 'linux3' , {});
-+ self.assertFlavor('linux' , 'linux' , {});
-
- def test_param(self):
- self.assertFlavor('foobar', 'linux2' , {'flavor': 'foobar'})
-diff --git a/tools/gyp/pylib/gyp/easy_xml.py b/pylib/gyp/easy_xml.py
-index 2522efb..15c6651 100644
---- a/tools/gyp/pylib/gyp/easy_xml.py
-+++ b/tools/gyp/pylib/gyp/easy_xml.py
-@@ -6,6 +6,11 @@
- import os
- import locale
-
-+try:
-+ # reduce moved to functools in python3.
-+ reduce
-+except NameError:
-+ from functools import reduce
-
- def XmlToString(content, encoding='utf-8', pretty=False):
- """ Writes the XML content to disk, touching the file only if it has changed.
-@@ -80,7 +85,7 @@
- # Optionally in second position is a dictionary of the attributes.
- rest = specification[1:]
- if rest and isinstance(rest[0], dict):
-- for at, val in sorted(rest[0].iteritems()):
-+ for at, val in sorted(rest[0].items()):
- xml_parts.append(' %s="%s"' % (at, _XmlEscape(val, attr=True)))
- rest = rest[1:]
- if rest:
-diff --git a/tools/gyp/pylib/gyp/easy_xml_test.py b/pylib/gyp/easy_xml_test.py
-index df64354..a1fdb18 100755
---- a/tools/gyp/pylib/gyp/easy_xml_test.py
-+++ b/tools/gyp/pylib/gyp/easy_xml_test.py
-@@ -8,13 +8,16 @@
-
- import gyp.easy_xml as easy_xml
- import unittest
--import StringIO
-+try:
-+ from StringIO import StringIO
-+except ImportError:
-+ from io import StringIO
-
-
- class TestSequenceFunctions(unittest.TestCase):
-
- def setUp(self):
-- self.stderr = StringIO.StringIO()
-+ self.stderr = StringIO()
-
- def test_EasyXml_simple(self):
- self.assertEqual(
-diff --git a/tools/gyp/pylib/gyp/flock_tool.py b/pylib/gyp/flock_tool.py
-index b38d866..81fb79d 100755
---- a/tools/gyp/pylib/gyp/flock_tool.py
-+++ b/tools/gyp/pylib/gyp/flock_tool.py
-@@ -39,7 +39,7 @@
- # where fcntl.flock(fd, LOCK_EX) always fails
- # with EBADF, that's why we use this F_SETLK
- # hack instead.
-- fd = os.open(lockfile, os.O_WRONLY|os.O_NOCTTY|os.O_CREAT, 0666)
-+ fd = os.open(lockfile, os.O_WRONLY|os.O_NOCTTY|os.O_CREAT, 0o666)
- if sys.platform.startswith('aix'):
- # Python on AIX is compiled with LARGEFILE support, which changes the
- # struct size.
-diff --git a/tools/gyp/pylib/gyp/generator/analyzer.py b/pylib/gyp/generator/analyzer.py
-index 921c1a6..b3484dc 100644
---- a/tools/gyp/pylib/gyp/generator/analyzer.py
-+++ b/tools/gyp/pylib/gyp/generator/analyzer.py
-@@ -62,6 +62,8 @@
- then the "all" target includes "b1" and "b2".
- """
-
-+from __future__ import print_function
-+
- import gyp.common
- import gyp.ninja_syntax as ninja_syntax
- import json
-@@ -155,7 +157,7 @@
- continue
- result.append(base_path + source)
- if debug:
-- print 'AddSource', org_source, result[len(result) - 1]
-+ print('AddSource', org_source, result[len(result) - 1])
-
-
- def _ExtractSourcesFromAction(action, base_path, base_path_components,
-@@ -185,7 +187,7 @@
- base_path += '/'
-
- if debug:
-- print 'ExtractSources', target, base_path
-+ print('ExtractSources', target, base_path)
-
- results = []
- if 'sources' in target_dict:
-@@ -278,7 +280,7 @@
- the root of the source tree."""
- if _ToLocalPath(toplevel_dir, _ToGypPath(build_file)) in files:
- if debug:
-- print 'gyp file modified', build_file
-+ print('gyp file modified', build_file)
- return True
-
- # First element of included_files is the file itself.
-@@ -291,8 +293,8 @@
- _ToGypPath(gyp.common.UnrelativePath(include_file, build_file))
- if _ToLocalPath(toplevel_dir, rel_include_file) in files:
- if debug:
-- print 'included gyp file modified, gyp_file=', build_file, \
-- 'included file=', rel_include_file
-+ print('included gyp file modified, gyp_file=', build_file, \
-+ 'included file=', rel_include_file)
- return True
- return False
-
-@@ -373,7 +375,7 @@
- # If a build file (or any of its included files) is modified we assume all
- # targets in the file are modified.
- if build_file_in_files[build_file]:
-- print 'matching target from modified build file', target_name
-+ print('matching target from modified build file', target_name)
- target.match_status = MATCH_STATUS_MATCHES
- matching_targets.append(target)
- else:
-@@ -381,7 +383,7 @@
- toplevel_dir)
- for source in sources:
- if _ToGypPath(os.path.normpath(source)) in files:
-- print 'target', target_name, 'matches', source
-+ print('target', target_name, 'matches', source)
- target.match_status = MATCH_STATUS_MATCHES
- matching_targets.append(target)
- break
-@@ -433,7 +435,7 @@
- for dep in target.deps:
- if _DoesTargetDependOnMatchingTargets(dep):
- target.match_status = MATCH_STATUS_MATCHES_BY_DEPENDENCY
-- print '\t', target.name, 'matches by dep', dep.name
-+ print('\t', target.name, 'matches by dep', dep.name)
- return True
- target.match_status = MATCH_STATUS_DOESNT_MATCH
- return False
-@@ -445,7 +447,7 @@
- supplied as input to analyzer.
- possible_targets: targets to search from."""
- found = []
-- print 'Targets that matched by dependency:'
-+ print('Targets that matched by dependency:')
- for target in possible_targets:
- if _DoesTargetDependOnMatchingTargets(target):
- found.append(target)
-@@ -484,12 +486,13 @@
- (add_if_no_ancestor or target.requires_build)) or
- (target.is_static_library and add_if_no_ancestor and
- not target.is_or_has_linked_ancestor)):
-- print '\t\tadding to compile targets', target.name, 'executable', \
-- target.is_executable, 'added_to_compile_targets', \
-- target.added_to_compile_targets, 'add_if_no_ancestor', \
-- add_if_no_ancestor, 'requires_build', target.requires_build, \
-- 'is_static_library', target.is_static_library, \
-+ print('\t\tadding to compile targets', target.name, 'executable',
-+ target.is_executable, 'added_to_compile_targets',
-+ target.added_to_compile_targets, 'add_if_no_ancestor',
-+ add_if_no_ancestor, 'requires_build', target.requires_build,
-+ 'is_static_library', target.is_static_library,
- 'is_or_has_linked_ancestor', target.is_or_has_linked_ancestor
-+ )
- result.add(target)
- target.added_to_compile_targets = True
-
-@@ -500,7 +503,7 @@
- supplied_targets: set of targets supplied to analyzer to search from."""
- result = set()
- for target in matching_targets:
-- print 'finding compile targets for match', target.name
-+ print('finding compile targets for match', target.name)
- _AddCompileTargets(target, supplied_targets, True, result)
- return result
-
-@@ -508,46 +511,46 @@
- def _WriteOutput(params, **values):
- """Writes the output, either to stdout or a file is specified."""
- if 'error' in values:
-- print 'Error:', values['error']
-+ print('Error:', values['error'])
- if 'status' in values:
-- print values['status']
-+ print(values['status'])
- if 'targets' in values:
- values['targets'].sort()
-- print 'Supplied targets that depend on changed files:'
-+ print('Supplied targets that depend on changed files:')
- for target in values['targets']:
-- print '\t', target
-+ print('\t', target)
- if 'invalid_targets' in values:
- values['invalid_targets'].sort()
-- print 'The following targets were not found:'
-+ print('The following targets were not found:')
- for target in values['invalid_targets']:
-- print '\t', target
-+ print('\t', target)
- if 'build_targets' in values:
- values['build_targets'].sort()
-- print 'Targets that require a build:'
-+ print('Targets that require a build:')
- for target in values['build_targets']:
-- print '\t', target
-+ print('\t', target)
- if 'compile_targets' in values:
- values['compile_targets'].sort()
-- print 'Targets that need to be built:'
-+ print('Targets that need to be built:')
- for target in values['compile_targets']:
-- print '\t', target
-+ print('\t', target)
- if 'test_targets' in values:
- values['test_targets'].sort()
-- print 'Test targets:'
-+ print('Test targets:')
- for target in values['test_targets']:
-- print '\t', target
-+ print('\t', target)
-
- output_path = params.get('generator_flags', {}).get(
- 'analyzer_output_path', None)
- if not output_path:
-- print json.dumps(values)
-+ print(json.dumps(values))
- return
- try:
- f = open(output_path, 'w')
- f.write(json.dumps(values) + '\n')
- f.close()
- except IOError as e:
-- print 'Error writing to output file', output_path, str(e)
-+ print('Error writing to output file', output_path, str(e))
-
-
- def _WasGypIncludeFileModified(params, files):
-@@ -556,7 +559,7 @@
- if params['options'].includes:
- for include in params['options'].includes:
- if _ToGypPath(os.path.normpath(include)) in files:
-- print 'Include file modified, assuming all changed', include
-+ print('Include file modified, assuming all changed', include)
- return True
- return False
-
-@@ -638,13 +641,13 @@
- set(self._root_targets))]
- else:
- test_targets = [x for x in test_targets_no_all]
-- print 'supplied test_targets'
-+ print('supplied test_targets')
- for target_name in self._test_target_names:
-- print '\t', target_name
-- print 'found test_targets'
-+ print('\t', target_name)
-+ print('found test_targets')
- for target in test_targets:
-- print '\t', target.name
-- print 'searching for matching test targets'
-+ print('\t', target.name)
-+ print('searching for matching test targets')
- matching_test_targets = _GetTargetsDependingOnMatchingTargets(test_targets)
- matching_test_targets_contains_all = (test_target_names_contains_all and
- set(matching_test_targets) &
-@@ -654,14 +657,14 @@
- # 'all' is subsequentely added to the matching names below.
- matching_test_targets = [x for x in (set(matching_test_targets) &
- set(test_targets_no_all))]
-- print 'matched test_targets'
-+ print('matched test_targets')
- for target in matching_test_targets:
-- print '\t', target.name
-+ print('\t', target.name)
- matching_target_names = [gyp.common.ParseQualifiedTarget(target.name)[1]
- for target in matching_test_targets]
- if matching_test_targets_contains_all:
- matching_target_names.append('all')
-- print '\tall'
-+ print('\tall')
- return matching_target_names
-
- def find_matching_compile_target_names(self):
-@@ -669,7 +672,7 @@
- assert self.is_build_impacted();
- # Compile targets are found by searching up from changed targets.
- # Reset the visited status for _GetBuildTargets.
-- for target in self._name_to_target.itervalues():
-+ for target in self._name_to_target.values():
- target.visited = False
-
- supplied_targets = _LookupTargets(self._supplied_target_names_no_all(),
-@@ -677,10 +680,10 @@
- if 'all' in self._supplied_target_names():
- supplied_targets = [x for x in (set(supplied_targets) |
- set(self._root_targets))]
-- print 'Supplied test_targets & compile_targets'
-+ print('Supplied test_targets & compile_targets')
- for target in supplied_targets:
-- print '\t', target.name
-- print 'Finding compile targets'
-+ print('\t', target.name)
-+ print('Finding compile targets')
- compile_targets = _GetCompileTargets(self._changed_targets,
- supplied_targets)
- return [gyp.common.ParseQualifiedTarget(target.name)[1]
-@@ -699,7 +702,7 @@
-
- toplevel_dir = _ToGypPath(os.path.abspath(params['options'].toplevel_dir))
- if debug:
-- print 'toplevel_dir', toplevel_dir
-+ print('toplevel_dir', toplevel_dir)
-
- if _WasGypIncludeFileModified(params, config.files):
- result_dict = { 'status': all_changed_string,
-diff --git a/tools/gyp/pylib/gyp/generator/cmake.py b/pylib/gyp/generator/cmake.py
-index a2b9629..4a2041c 100644
---- a/tools/gyp/pylib/gyp/generator/cmake.py
-+++ b/tools/gyp/pylib/gyp/generator/cmake.py
-@@ -28,6 +28,8 @@
- CMakeLists.txt file.
- """
-
-+from __future__ import print_function
-+
- import multiprocessing
- import os
- import signal
-@@ -36,6 +38,12 @@
- import gyp.common
- import gyp.xcode_emulation
-
-+try:
-+ # maketrans moved to str in python3.
-+ _maketrans = string.maketrans
-+except NameError:
-+ _maketrans = str.maketrans
-+
- generator_default_variables = {
- 'EXECUTABLE_PREFIX': '',
- 'EXECUTABLE_SUFFIX': '',
-@@ -238,7 +246,7 @@
- Invalid for make: ':'
- Invalid for unknown reasons but cause failures: '.'
- """
-- return a.translate(string.maketrans(' /():."', '_______'))
-+ return a.translate(_maketrans(' /():."', '_______'))
-
-
- def WriteActions(target_name, actions, extra_sources, extra_deps,
-@@ -644,8 +652,8 @@
-
- cmake_target_type = cmake_target_type_from_gyp_target_type.get(target_type)
- if cmake_target_type is None:
-- print ('Target %s has unknown target type %s, skipping.' %
-- ( target_name, target_type ) )
-+ print('Target %s has unknown target type %s, skipping.' %
-+ ( target_name, target_type ))
- return
-
- SetVariable(output, 'TARGET', target_name)
-@@ -868,8 +876,8 @@
- default_product_ext = generator_default_variables['SHARED_LIB_SUFFIX']
-
- elif target_type != 'executable':
-- print ('ERROR: What output file should be generated?',
-- 'type', target_type, 'target', target_name)
-+ print(('ERROR: What output file should be generated?',
-+ 'type', target_type, 'target', target_name))
-
- product_prefix = spec.get('product_prefix', default_product_prefix)
- product_name = spec.get('product_name', default_product_name)
-@@ -1207,11 +1215,11 @@
- output_dir,
- config_name))
- arguments = ['cmake', '-G', 'Ninja']
-- print 'Generating [%s]: %s' % (config_name, arguments)
-+ print('Generating [%s]: %s' % (config_name, arguments))
- subprocess.check_call(arguments, cwd=build_dir)
-
- arguments = ['ninja', '-C', build_dir]
-- print 'Building [%s]: %s' % (config_name, arguments)
-+ print('Building [%s]: %s' % (config_name, arguments))
- subprocess.check_call(arguments)
-
-
-@@ -1230,7 +1238,7 @@
- GenerateOutputForConfig(target_list, target_dicts, data,
- params, user_config)
- else:
-- config_names = target_dicts[target_list[0]]['configurations'].keys()
-+ config_names = target_dicts[target_list[0]]['configurations']
- if params['parallel']:
- try:
- pool = multiprocessing.Pool(len(config_names))
-@@ -1239,7 +1247,7 @@
- arglists.append((target_list, target_dicts, data,
- params, config_name))
- pool.map(CallGenerateOutputForConfig, arglists)
-- except KeyboardInterrupt, e:
-+ except KeyboardInterrupt as e:
- pool.terminate()
- raise e
- else:
-diff --git a/tools/gyp/pylib/gyp/generator/dump_dependency_json.py b/pylib/gyp/generator/dump_dependency_json.py
-index 160eafe..2bf3f39 100644
---- a/tools/gyp/pylib/gyp/generator/dump_dependency_json.py
-+++ b/tools/gyp/pylib/gyp/generator/dump_dependency_json.py
-@@ -2,6 +2,8 @@
- # Use of this source code is governed by a BSD-style license that can be
- # found in the LICENSE file.
-
-+from __future__ import print_function
-+
- import collections
- import os
- import gyp
-@@ -96,4 +98,4 @@
- f = open(filename, 'w')
- json.dump(edges, f)
- f.close()
-- print 'Wrote json to %s.' % filename
-+ print('Wrote json to %s.' % filename)
-diff --git a/tools/gyp/pylib/gyp/generator/eclipse.py b/pylib/gyp/generator/eclipse.py
-index 3544347..d039f03 100644
---- a/tools/gyp/pylib/gyp/generator/eclipse.py
-+++ b/tools/gyp/pylib/gyp/generator/eclipse.py
-@@ -141,7 +141,7 @@
- compiler_includes_list.append(include_dir)
-
- # Find standard gyp include dirs.
-- if config.has_key('include_dirs'):
-+ if 'include_dirs' in config:
- include_dirs = config['include_dirs']
- for shared_intermediate_dir in shared_intermediate_dirs:
- for include_dir in include_dirs:
-@@ -272,7 +272,7 @@
- out.write(' <language name="holder for library settings"></language>\n')
- for lang in eclipse_langs:
- out.write(' <language name="%s">\n' % lang)
-- for key in sorted(defines.iterkeys()):
-+ for key in sorted(defines.keys()):
- out.write(' <macro><name>%s</name><value>%s</value></macro>\n' %
- (escape(key), escape(defines[key])))
- out.write(' </language>\n')
-@@ -418,7 +418,7 @@
- GenerateOutputForConfig(target_list, target_dicts, data, params,
- user_config)
- else:
-- config_names = target_dicts[target_list[0]]['configurations'].keys()
-+ config_names = target_dicts[target_list[0]]['configurations']
- for config_name in config_names:
- GenerateOutputForConfig(target_list, target_dicts, data, params,
- config_name)
-diff --git a/tools/gyp/pylib/gyp/generator/gypd.py b/pylib/gyp/generator/gypd.py
-index 3efdb99..78eeaa6 100644
---- a/tools/gyp/pylib/gyp/generator/gypd.py
-+++ b/tools/gyp/pylib/gyp/generator/gypd.py
-@@ -88,7 +88,7 @@
- if not output_file in output_files:
- output_files[output_file] = input_file
-
-- for output_file, input_file in output_files.iteritems():
-+ for output_file, input_file in output_files.items():
- output = open(output_file, 'w')
- pprint.pprint(data[input_file], output)
- output.close()
-diff --git a/tools/gyp/pylib/gyp/generator/make.py b/pylib/gyp/generator/make.py
-index fb4f918..2057e3a 100644
---- a/tools/gyp/pylib/gyp/generator/make.py
-+++ b/tools/gyp/pylib/gyp/generator/make.py
-@@ -21,6 +21,8 @@
- # toplevel Makefile. It may make sense to generate some .mk files on
- # the side to keep the the files readable.
-
-+from __future__ import print_function
-+
- import os
- import re
- import sys
-@@ -668,7 +670,7 @@
- basenames.setdefault(basename, []).append(source)
-
- error = ''
-- for basename, files in basenames.iteritems():
-+ for basename, files in basenames.items():
- if len(files) > 1:
- error += ' %s: %s\n' % (basename, ' '.join(files))
-
-@@ -816,7 +818,7 @@
- gyp.xcode_emulation.MacPrefixHeader(
- self.xcode_settings, lambda p: Sourceify(self.Absolutify(p)),
- self.Pchify))
-- sources = filter(Compilable, all_sources)
-+ sources = [x for x in all_sources if Compilable(x)]
- if sources:
- self.WriteLn(SHARED_HEADER_SUFFIX_RULES_COMMENT1)
- extensions = set([os.path.splitext(s)[1] for s in sources])
-@@ -945,7 +947,7 @@
- '%s%s'
- % (name, cd_action, command))
- self.WriteLn()
-- outputs = map(self.Absolutify, outputs)
-+ outputs = [self.Absolutify(o) for o in outputs]
- # The makefile rules are all relative to the top dir, but the gyp actions
- # are defined relative to their containing dir. This replaces the obj
- # variable for the action rule with an absolute version so that the output
-@@ -1035,7 +1037,7 @@
- outputs = [gyp.xcode_emulation.ExpandEnvVars(o, env) for o in outputs]
- inputs = [gyp.xcode_emulation.ExpandEnvVars(i, env) for i in inputs]
-
-- outputs = map(self.Absolutify, outputs)
-+ outputs = [self.Absolutify(o) for o in outputs]
- all_outputs += outputs
- # Only write the 'obj' and 'builddir' rules for the "primary" output
- # (:1); it's superfluous for the "extra outputs", and this avoids
-@@ -1233,11 +1235,11 @@
- self.WriteList(cflags_objcc, 'CFLAGS_OBJCC_%s' % configname)
- includes = config.get('include_dirs')
- if includes:
-- includes = map(Sourceify, map(self.Absolutify, includes))
-+ includes = [Sourceify(self.Absolutify(include)) for include in includes]
- self.WriteList(includes, 'INCS_%s' % configname, prefix='-I')
-
- compilable = filter(Compilable, sources)
-- objs = map(self.Objectify, map(self.Absolutify, map(Target, compilable)))
-+ objs = [self.Objectify(self.Absolutify(Target(x))) for x in compilable]
- self.WriteList(objs, 'OBJS')
-
- for obj in objs:
-@@ -1309,7 +1311,7 @@
-
- # If there are any object files in our input file list, link them into our
- # output.
-- extra_link_deps += filter(Linkable, sources)
-+ extra_link_deps += [source for source in sources if Linkable(source)]
-
- self.WriteLn()
-
-@@ -1377,8 +1379,8 @@
- elif self.type == 'none':
- target = '%s.stamp' % target
- elif self.type != 'executable':
-- print ("ERROR: What output file should be generated?",
-- "type", self.type, "target", target)
-+ print(("ERROR: What output file should be generated?",
-+ "type", self.type, "target", target))
-
- target_prefix = spec.get('product_prefix', target_prefix)
- target = spec.get('product_name', target)
-@@ -1542,9 +1544,9 @@
- # Postbuilds expect to be run in the gyp file's directory, so insert an
- # implicit postbuild to cd to there.
- postbuilds.insert(0, gyp.common.EncodePOSIXShellList(['cd', self.path]))
-- for i in xrange(len(postbuilds)):
-- if not postbuilds[i].startswith('$'):
-- postbuilds[i] = EscapeShellArgument(postbuilds[i])
-+ for i, postbuild in enumerate(postbuilds):
-+ if not postbuild.startswith('$'):
-+ postbuilds[i] = EscapeShellArgument(postbuild)
- self.WriteLn('%s: builddir := $(abs_builddir)' % QuoteSpaces(self.output))
- self.WriteLn('%s: POSTBUILDS := %s' % (
- QuoteSpaces(self.output), ' '.join(postbuilds)))
-@@ -1634,7 +1636,7 @@
- self.WriteDoCmd([self.output_binary], deps, 'touch', part_of_all,
- postbuilds=postbuilds)
- else:
-- print "WARNING: no output for", self.type, target
-+ print("WARNING: no output for", self.type, target)
-
- # Add an alias for each target (if there are any outputs).
- # Installable target aliases are created below.
-@@ -1741,7 +1743,7 @@
- output is just a name to run the rule
- command: (optional) command name to generate unambiguous labels
- """
-- outputs = map(QuoteSpaces, outputs)
-+ outputs = [QuoteSpaces(o) for o in outputs]
- inputs = map(QuoteSpaces, inputs)
-
- if comment:
-@@ -1986,7 +1988,7 @@
- if options.toplevel_dir and options.toplevel_dir != '.':
- arguments += '-C', options.toplevel_dir
- arguments.append('BUILDTYPE=' + config)
-- print 'Building [%s]: %s' % (config, arguments)
-+ print('Building [%s]: %s' % (config, arguments))
- subprocess.check_call(arguments)
-
-
-diff --git a/tools/gyp/pylib/gyp/generator/msvs.py b/pylib/gyp/generator/msvs.py
-index 8fe9e5a..e8a2b36 100644
---- a/tools/gyp/pylib/gyp/generator/msvs.py
-+++ b/tools/gyp/pylib/gyp/generator/msvs.py
-@@ -2,6 +2,9 @@
- # Use of this source code is governed by a BSD-style license that can be
- # found in the LICENSE file.
-
-+from __future__ import print_function
-+
-+import collections
- import copy
- import ntpath
- import os
-@@ -23,16 +26,6 @@
- from gyp.common import GypError
- from gyp.common import OrderedSet
-
--# TODO: Remove once bots are on 2.7, http://crbug.com/241769
--def _import_OrderedDict():
-- import collections
-- try:
-- return collections.OrderedDict
-- except AttributeError:
-- import gyp.ordered_dict
-- return gyp.ordered_dict.OrderedDict
--OrderedDict = _import_OrderedDict()
--
-
- # Regular expression for validating Visual Studio GUIDs. If the GUID
- # contains lowercase hex letters, MSVS will be fine. However,
-@@ -202,7 +195,7 @@
- if not prefix: prefix = []
- result = []
- excluded_result = []
-- folders = OrderedDict()
-+ folders = collections.OrderedDict()
- # Gather files into the final result, excluded, or folders.
- for s in sources:
- if len(s) == 1:
-@@ -469,7 +462,7 @@
- 'CommandLine': cmd,
- })
- # Add to the properties of primary input for each config.
-- for config_name, c_data in spec['configurations'].iteritems():
-+ for config_name, c_data in spec['configurations'].items():
- p.AddFileConfig(_FixPath(primary_input),
- _ConfigFullName(config_name, c_data), tools=[tool])
-
-@@ -775,8 +768,8 @@
- # the VCProj but cause the same problem on the final command-line. Moving
- # the item to the end of the list does works, but that's only possible if
- # there's only one such item. Let's just warn the user.
-- print >> sys.stderr, ('Warning: MSVS may misinterpret the odd number of ' +
-- 'quotes in ' + s)
-+ print(('Warning: MSVS may misinterpret the odd number of ' +
-+ 'quotes in ' + s), file=sys.stderr)
- return s
-
-
-@@ -991,7 +984,7 @@
- basenames.setdefault(basename, []).append(source)
-
- error = ''
-- for basename, files in basenames.iteritems():
-+ for basename, files in basenames.items():
- if len(files) > 1:
- error += ' %s: %s\n' % (basename, ' '.join(files))
-
-@@ -1023,7 +1016,7 @@
- relative_path_of_gyp_file = gyp.common.RelativePath(gyp_path, project_dir)
-
- config_type = _GetMSVSConfigurationType(spec, project.build_file)
-- for config_name, config in spec['configurations'].iteritems():
-+ for config_name, config in spec['configurations'].items():
- _AddConfigurationToMSVSProject(p, spec, config_type, config_name, config)
-
- # MSVC08 and prior version cannot handle duplicate basenames in the same
-@@ -1392,10 +1385,10 @@
- A list of Tool objects.
- """
- tool_list = []
-- for tool, settings in tools.iteritems():
-+ for tool, settings in tools.items():
- # Collapse settings with lists.
- settings_fixed = {}
-- for setting, value in settings.iteritems():
-+ for setting, value in settings.items():
- if type(value) == list:
- if ((tool == 'VCLinkerTool' and
- setting == 'AdditionalDependencies') or
-@@ -1570,7 +1563,7 @@
- def _GetPrecompileRelatedFiles(spec):
- # Gather a list of precompiled header related sources.
- precompiled_related = []
-- for _, config in spec['configurations'].iteritems():
-+ for _, config in spec['configurations'].items():
- for k in precomp_keys:
- f = config.get(k)
- if f:
-@@ -1581,7 +1574,7 @@
- def _ExcludeFilesFromBeingBuilt(p, spec, excluded_sources, excluded_idl,
- list_excluded):
- exclusions = _GetExcludedFilesFromBuild(spec, excluded_sources, excluded_idl)
-- for file_name, excluded_configs in exclusions.iteritems():
-+ for file_name, excluded_configs in exclusions.items():
- if (not list_excluded and
- len(excluded_configs) == len(spec['configurations'])):
- # If we're not listing excluded files, then they won't appear in the
-@@ -1598,7 +1591,7 @@
- # Exclude excluded sources from being built.
- for f in excluded_sources:
- excluded_configs = []
-- for config_name, config in spec['configurations'].iteritems():
-+ for config_name, config in spec['configurations'].items():
- precomped = [_FixPath(config.get(i, '')) for i in precomp_keys]
- # Don't do this for ones that are precompiled header related.
- if f not in precomped:
-@@ -1608,7 +1601,7 @@
- # Exclude them now.
- for f in excluded_idl:
- excluded_configs = []
-- for config_name, config in spec['configurations'].iteritems():
-+ for config_name, config in spec['configurations'].items():
- excluded_configs.append((config_name, config))
- exclusions[f] = excluded_configs
- return exclusions
-@@ -1617,7 +1610,7 @@
- def _AddToolFilesToMSVS(p, spec):
- # Add in tool files (rules).
- tool_files = OrderedSet()
-- for _, config in spec['configurations'].iteritems():
-+ for _, config in spec['configurations'].items():
- for f in config.get('msvs_tool_files', []):
- tool_files.add(f)
- for f in tool_files:
-@@ -1630,7 +1623,7 @@
- # kind (i.e. C vs. C++) as the precompiled header source stub needs
- # to have use of precompiled headers disabled.
- extensions_excluded_from_precompile = []
-- for config_name, config in spec['configurations'].iteritems():
-+ for config_name, config in spec['configurations'].items():
- source = config.get('msvs_precompiled_source')
- if source:
- source = _FixPath(source)
-@@ -1651,7 +1644,7 @@
- else:
- basename, extension = os.path.splitext(source)
- if extension in extensions_excluded_from_precompile:
-- for config_name, config in spec['configurations'].iteritems():
-+ for config_name, config in spec['configurations'].items():
- tool = MSVSProject.Tool('VCCLCompilerTool',
- {'UsePrecompiledHeader': '0',
- 'ForcedIncludeFiles': '$(NOINHERIT)'})
-@@ -1702,7 +1695,7 @@
- return # Nothing to add
- # Write out the user file.
- user_file = _CreateMSVSUserFile(project_path, version, spec)
-- for config_name, c_data in spec['configurations'].iteritems():
-+ for config_name, c_data in spec['configurations'].items():
- user_file.AddDebugSettings(_ConfigFullName(config_name, c_data),
- action, environment, working_directory)
- user_file.WriteIfChanged()
-@@ -1756,7 +1749,7 @@
- def _DictsToFolders(base_path, bucket, flat):
- # Convert to folders recursively.
- children = []
-- for folder, contents in bucket.iteritems():
-+ for folder, contents in bucket.items():
- if type(contents) == dict:
- folder_children = _DictsToFolders(os.path.join(base_path, folder),
- contents, flat)
-@@ -1778,8 +1771,8 @@
- # such projects up one level.
- if (type(node) == dict and
- len(node) == 1 and
-- node.keys()[0] == parent + '.vcproj'):
-- return node[node.keys()[0]]
-+ next(iter(node)) == parent + '.vcproj'):
-+ return node[next(iter(node))]
- if type(node) != dict:
- return node
- for child in node:
-@@ -1798,8 +1791,8 @@
- # Walk down from the top until we hit a folder that has more than one entry.
- # In practice, this strips the top-level "src/" dir from the hierarchy in
- # the solution.
-- while len(root) == 1 and type(root[root.keys()[0]]) == dict:
-- root = root[root.keys()[0]]
-+ while len(root) == 1 and type(root[next(iter(root))]) == dict:
-+ root = root[next(iter(root))]
- # Collapse singles.
- root = _CollapseSingles('', root)
- # Merge buckets until everything is a root entry.
-@@ -1828,7 +1821,7 @@
- # Prepare a dict indicating which project configurations are used for which
- # solution configurations for this target.
- config_platform_overrides = {}
-- for config_name, c in spec['configurations'].iteritems():
-+ for config_name, c in spec['configurations'].items():
- config_fullname = _ConfigFullName(config_name, c)
- platform = c.get('msvs_target_platform', _ConfigPlatform(c))
- fixed_config_fullname = '%s|%s' % (
-@@ -1967,7 +1960,7 @@
- msvs_version = params['msvs_version']
- devenv = os.path.join(msvs_version.path, 'Common7', 'IDE', 'devenv.com')
-
-- for build_file, build_file_dict in data.iteritems():
-+ for build_file, build_file_dict in data.items():
- (build_file_root, build_file_ext) = os.path.splitext(build_file)
- if build_file_ext != '.gyp':
- continue
-@@ -1977,7 +1970,7 @@
-
- for config in configurations:
- arguments = [devenv, sln_path, '/Build', config]
-- print 'Building [%s]: %s' % (config, arguments)
-+ print('Building [%s]: %s' % (config, arguments))
- rtn = subprocess.check_call(arguments)
-
-
-@@ -2029,7 +2022,7 @@
- configs = set()
- for qualified_target in target_list:
- spec = target_dicts[qualified_target]
-- for config_name, config in spec['configurations'].iteritems():
-+ for config_name, config in spec['configurations'].items():
- configs.add(_ConfigFullName(config_name, config))
- configs = list(configs)
-
-@@ -2072,7 +2065,7 @@
- if generator_flags.get('msvs_error_on_missing_sources', False):
- raise GypError(error_message)
- else:
-- print >> sys.stdout, "Warning: " + error_message
-+ print("Warning: " + error_message, file=sys.stdout)
-
-
- def _GenerateMSBuildFiltersFile(filters_path, source_files,
-@@ -2669,7 +2662,7 @@
-
- def _GetMSBuildProjectConfigurations(configurations):
- group = ['ItemGroup', {'Label': 'ProjectConfigurations'}]
-- for (name, settings) in sorted(configurations.iteritems()):
-+ for (name, settings) in sorted(configurations.items()):
- configuration, platform = _GetConfigurationAndPlatform(name, settings)
- designation = '%s|%s' % (configuration, platform)
- group.append(
-@@ -2742,7 +2735,7 @@
-
- def _GetMSBuildConfigurationDetails(spec, build_file):
- properties = {}
-- for name, settings in spec['configurations'].iteritems():
-+ for name, settings in spec['configurations'].items():
- msbuild_attributes = _GetMSBuildAttributes(spec, settings, build_file)
- condition = _GetConfigurationCondition(name, settings)
- character_set = msbuild_attributes.get('CharacterSet')
-@@ -2776,9 +2769,9 @@
- user_props = r'$(UserRootDir)\Microsoft.Cpp.$(Platform).user.props'
- additional_props = {}
- props_specified = False
-- for name, settings in sorted(configurations.iteritems()):
-+ for name, settings in sorted(configurations.items()):
- configuration = _GetConfigurationCondition(name, settings)
-- if settings.has_key('msbuild_props'):
-+ if 'msbuild_props' in settings:
- additional_props[configuration] = _FixPaths(settings['msbuild_props'])
- props_specified = True
- else:
-@@ -2798,7 +2791,7 @@
- ]
- else:
- sheets = []
-- for condition, props in additional_props.iteritems():
-+ for condition, props in additional_props.items():
- import_group = [
- 'ImportGroup',
- {'Label': 'PropertySheets',
-@@ -2831,7 +2824,7 @@
- elif a == 'ConfigurationType':
- msbuild_attributes[a] = _ConvertMSVSConfigurationType(msvs_attributes[a])
- else:
-- print 'Warning: Do not know how to convert MSVS attribute ' + a
-+ print('Warning: Do not know how to convert MSVS attribute ' + a)
- return msbuild_attributes
-
-
-@@ -2927,7 +2920,7 @@
- new_paths = '$(ExecutablePath);' + ';'.join(new_paths)
-
- properties = {}
-- for (name, configuration) in sorted(configurations.iteritems()):
-+ for (name, configuration) in sorted(configurations.items()):
- condition = _GetConfigurationCondition(name, configuration)
- attributes = _GetMSBuildAttributes(spec, configuration, build_file)
- msbuild_settings = configuration['finalized_msbuild_settings']
-@@ -2952,7 +2945,7 @@
- _AddConditionalProperty(properties, condition, 'ExecutablePath',
- new_paths)
- tool_settings = msbuild_settings.get('', {})
-- for name, value in sorted(tool_settings.iteritems()):
-+ for name, value in sorted(tool_settings.items()):
- formatted_value = _GetValueFormattedForMSBuild('', name, value)
- _AddConditionalProperty(properties, condition, name, formatted_value)
- return _GetMSBuildPropertyGroup(spec, None, properties)
-@@ -3021,7 +3014,7 @@
- # NOTE: reverse(topsort(DAG)) = topsort(reverse_edges(DAG))
- for name in reversed(properties_ordered):
- values = properties[name]
-- for value, conditions in sorted(values.iteritems()):
-+ for value, conditions in sorted(values.items()):
- if len(conditions) == num_configurations:
- # If the value is the same all configurations,
- # just add one unconditional entry.
-@@ -3034,18 +3027,18 @@
-
- def _GetMSBuildToolSettingsSections(spec, configurations):
- groups = []
-- for (name, configuration) in sorted(configurations.iteritems()):
-+ for (name, configuration) in sorted(configurations.items()):
- msbuild_settings = configuration['finalized_msbuild_settings']
- group = ['ItemDefinitionGroup',
- {'Condition': _GetConfigurationCondition(name, configuration)}
- ]
-- for tool_name, tool_settings in sorted(msbuild_settings.iteritems()):
-+ for tool_name, tool_settings in sorted(msbuild_settings.items()):
- # Skip the tool named '' which is a holder of global settings handled
- # by _GetMSBuildConfigurationGlobalProperties.
- if tool_name:
- if tool_settings:
- tool = [tool_name]
-- for name, value in sorted(tool_settings.iteritems()):
-+ for name, value in sorted(tool_settings.items()):
- formatted_value = _GetValueFormattedForMSBuild(tool_name, name,
- value)
- tool.append([name, formatted_value])
-@@ -3078,8 +3071,8 @@
- for ignored_setting in ignored_settings:
- value = configuration.get(ignored_setting)
- if value:
-- print ('Warning: The automatic conversion to MSBuild does not handle '
-- '%s. Ignoring setting of %s' % (ignored_setting, str(value)))
-+ print('Warning: The automatic conversion to MSBuild does not handle '
-+ '%s. Ignoring setting of %s' % (ignored_setting, str(value)))
-
- defines = [_EscapeCppDefineForMSBuild(d) for d in defines]
- disabled_warnings = _GetDisabledWarnings(configuration)
-@@ -3245,7 +3238,7 @@
- {'Condition': condition},
- 'true'])
- # Add precompile if needed
-- for config_name, configuration in spec['configurations'].iteritems():
-+ for config_name, configuration in spec['configurations'].items():
- precompiled_source = configuration.get('msvs_precompiled_source', '')
- if precompiled_source != '':
- precompiled_source = _FixPath(precompiled_source)
-@@ -3291,7 +3284,7 @@
- ['Project', guid],
- ['ReferenceOutputAssembly', 'false']
- ]
-- for config in dependency.spec.get('configurations', {}).itervalues():
-+ for config in dependency.spec.get('configurations', {}).values():
- if config.get('msvs_use_library_dependency_inputs', 0):
- project_ref.append(['UseLibraryDependencyInputs', 'true'])
- break
-@@ -3360,7 +3353,7 @@
- extension_to_rule_name)
- missing_sources = _VerifySourcesExist(sources, project_dir)
-
-- for configuration in configurations.itervalues():
-+ for configuration in configurations.values():
- _FinalizeMSBuildSettings(spec, configuration)
-
- # Add attributes to root element
-@@ -3486,7 +3479,7 @@
- """
- sources_handled_by_action = OrderedSet()
- actions_spec = []
-- for primary_input, actions in actions_to_add.iteritems():
-+ for primary_input, actions in actions_to_add.items():
- inputs = OrderedSet()
- outputs = OrderedSet()
- descriptions = []
-diff --git a/tools/gyp/pylib/gyp/generator/msvs_test.py b/pylib/gyp/generator/msvs_test.py
-index c0b021d..838d236 100755
---- a/tools/gyp/pylib/gyp/generator/msvs_test.py
-+++ b/tools/gyp/pylib/gyp/generator/msvs_test.py
-@@ -7,13 +7,16 @@
-
- import gyp.generator.msvs as msvs
- import unittest
--import StringIO
-+try:
-+ from StringIO import StringIO
-+except ImportError:
-+ from io import StringIO
-
-
- class TestSequenceFunctions(unittest.TestCase):
-
- def setUp(self):
-- self.stderr = StringIO.StringIO()
-+ self.stderr = StringIO()
-
- def test_GetLibraries(self):
- self.assertEqual(
-diff --git a/tools/gyp/pylib/gyp/generator/ninja.py b/pylib/gyp/generator/ninja.py
-index 6de87b7..66faabc 100644
---- a/tools/gyp/pylib/gyp/generator/ninja.py
-+++ b/tools/gyp/pylib/gyp/generator/ninja.py
-@@ -2,6 +2,8 @@
- # Use of this source code is governed by a BSD-style license that can be
- # found in the LICENSE file.
-
-+from __future__ import print_function
-+
- import collections
- import copy
- import hashlib
-@@ -18,7 +20,10 @@
- import gyp.msvs_emulation
- import gyp.MSVSUtil as MSVSUtil
- import gyp.xcode_emulation
--from cStringIO import StringIO
-+try:
-+ from cStringIO import StringIO
-+except ImportError:
-+ from io import StringIO
-
- from gyp.common import GetEnvironFallback
- import gyp.ninja_syntax as ninja_syntax
-@@ -350,7 +355,7 @@
-
- Uses a stamp file if necessary."""
-
-- assert targets == filter(None, targets), targets
-+ assert targets == [t for t in targets if t], targets
- if len(targets) == 0:
- assert not order_only
- return None
-@@ -427,8 +432,8 @@
- compile_depends.append(target.PreCompileInput())
- if target.uses_cpp:
- self.target.uses_cpp = True
-- actions_depends = filter(None, actions_depends)
-- compile_depends = filter(None, compile_depends)
-+ actions_depends = [d for d in actions_depends if d]
-+ compile_depends = [d for d in compile_depends if d]
- actions_depends = self.WriteCollapsedDependencies('actions_depends',
- actions_depends)
- compile_depends = self.WriteCollapsedDependencies('compile_depends',
-@@ -455,8 +460,8 @@
- try:
- sources = extra_sources + spec.get('sources', [])
- except TypeError:
-- print 'extra_sources: ', str(extra_sources)
-- print 'spec.get("sources"): ', str(spec.get('sources'))
-+ print('extra_sources: ', str(extra_sources))
-+ print('spec.get("sources"): ', str(spec.get('sources')))
- raise
- if sources:
- if self.flavor == 'mac' and len(self.archs) > 1:
-@@ -485,8 +490,9 @@
- if self.flavor != 'mac' or len(self.archs) == 1:
- link_deps += [self.GypPathToNinja(o) for o in obj_outputs]
- else:
-- print "Warning: Actions/rules writing object files don't work with " \
-- "multiarch targets, dropping. (target %s)" % spec['target_name']
-+ print("Warning: Actions/rules writing object files don't work with " \
-+ "multiarch targets, dropping. (target %s)" %
-+ spec['target_name'])
- elif self.flavor == 'mac' and len(self.archs) > 1:
- link_deps = collections.defaultdict(list)
-
-@@ -838,7 +844,7 @@
- 'XCASSETS_LAUNCH_IMAGE': 'launch-image',
- }
- settings = self.xcode_settings.xcode_settings[self.config_name]
-- for settings_key, arg_name in settings_to_arg.iteritems():
-+ for settings_key, arg_name in settings_to_arg.items():
- value = settings.get(settings_key)
- if value:
- extra_arguments[arg_name] = value
-@@ -1772,7 +1778,7 @@
-
- # VS 2015 uses 20% more working set than VS 2013 and can consume all RAM
- # on a 64 GB machine.
-- mem_limit = max(1, stat.ullTotalPhys / (5 * (2 ** 30))) # total / 5GB
-+ mem_limit = max(1, stat.ullTotalPhys // (5 * (2 ** 30))) # total / 5GB
- hard_cap = max(1, int(os.environ.get('GYP_LINK_CONCURRENCY_MAX', 2**32)))
- return min(mem_limit, hard_cap)
- elif sys.platform.startswith('linux'):
-@@ -1784,14 +1790,14 @@
- if not match:
- continue
- # Allow 8Gb per link on Linux because Gold is quite memory hungry
-- return max(1, int(match.group(1)) / (8 * (2 ** 20)))
-+ return max(1, int(match.group(1)) // (8 * (2 ** 20)))
- return 1
- elif sys.platform == 'darwin':
- try:
- avail_bytes = int(subprocess.check_output(['sysctl', '-n', 'hw.memsize']))
- # A static library debug build of Chromium's unit_tests takes ~2.7GB, so
- # 4GB per ld process allows for some more bloat.
-- return max(1, avail_bytes / (4 * (2 ** 30))) # total / 4GB
-+ return max(1, avail_bytes // (4 * (2 ** 30))) # total / 4GB
- except:
- return 1
- else:
-@@ -1946,7 +1952,7 @@
- wrappers[key[:-len('_wrapper')]] = os.path.join(build_to_root, value)
-
- # Support wrappers from environment variables too.
-- for key, value in os.environ.iteritems():
-+ for key, value in os.environ.items():
- if key.lower().endswith('_wrapper'):
- key_prefix = key[:-len('_wrapper')]
- key_prefix = re.sub(r'\.HOST$', '.host', key_prefix)
-@@ -1966,7 +1972,7 @@
- configs, generator_flags)
- cl_paths = gyp.msvs_emulation.GenerateEnvironmentFiles(
- toplevel_build, generator_flags, shared_system_includes, OpenOutput)
-- for arch, path in sorted(cl_paths.iteritems()):
-+ for arch, path in sorted(cl_paths.items()):
- if clang_cl:
- # If we have selected clang-cl, use that instead.
- path = clang_cl
-@@ -2381,6 +2387,7 @@
-
- qualified_target_for_hash = gyp.common.QualifiedTarget(build_file, name,
- toolset)
-+ qualified_target_for_hash = qualified_target_for_hash.encode('utf-8')
- hash_for_rules = hashlib.md5(qualified_target_for_hash).hexdigest()
-
- base_path = os.path.dirname(build_file)
-@@ -2447,7 +2454,7 @@
- for config in configurations:
- builddir = os.path.join(options.toplevel_dir, 'out', config)
- arguments = ['ninja', '-C', builddir]
-- print 'Building [%s]: %s' % (config, arguments)
-+ print('Building [%s]: %s' % (config, arguments))
- subprocess.check_call(arguments)
-
-
-@@ -2475,7 +2482,7 @@
- GenerateOutputForConfig(target_list, target_dicts, data, params,
- user_config)
- else:
-- config_names = target_dicts[target_list[0]]['configurations'].keys()
-+ config_names = target_dicts[target_list[0]]['configurations']
- if params['parallel']:
- try:
- pool = multiprocessing.Pool(len(config_names))
-@@ -2484,7 +2491,7 @@
- arglists.append(
- (target_list, target_dicts, data, params, config_name))
- pool.map(CallGenerateOutputForConfig, arglists)
-- except KeyboardInterrupt, e:
-+ except KeyboardInterrupt as e:
- pool.terminate()
- raise e
- else:
-diff --git a/tools/gyp/pylib/gyp/generator/ninja_test.py b/pylib/gyp/generator/ninja_test.py
-index 1767b2f..1ad68e4 100644
---- a/tools/gyp/pylib/gyp/generator/ninja_test.py
-+++ b/tools/gyp/pylib/gyp/generator/ninja_test.py
-@@ -8,7 +8,6 @@
-
- import gyp.generator.ninja as ninja
- import unittest
--import StringIO
- import sys
- import TestCommon
-
-diff --git a/tools/gyp/pylib/gyp/generator/xcode.py b/pylib/gyp/generator/xcode.py
-index b35372a..8bc22be 100644
---- a/tools/gyp/pylib/gyp/generator/xcode.py
-+++ b/tools/gyp/pylib/gyp/generator/xcode.py
-@@ -2,6 +2,8 @@
- # Use of this source code is governed by a BSD-style license that can be
- # found in the LICENSE file.
-
-+from __future__ import print_function
-+
- import filecmp
- import gyp.common
- import gyp.xcodeproj_file
-@@ -129,7 +131,7 @@
- try:
- os.makedirs(self.path)
- self.created_dir = True
-- except OSError, e:
-+ except OSError as e:
- if e.errno != errno.EEXIST:
- raise
-
-@@ -183,7 +185,7 @@
- # the tree tree view for UI display.
- # Any values set globally are applied to all configurations, then any
- # per-configuration values are applied.
-- for xck, xcv in self.build_file_dict.get('xcode_settings', {}).iteritems():
-+ for xck, xcv in self.build_file_dict.get('xcode_settings', {}).items():
- xccl.SetBuildSetting(xck, xcv)
- if 'xcode_config_file' in self.build_file_dict:
- config_ref = self.project.AddOrGetFileInRootGroup(
-@@ -197,7 +199,7 @@
- if build_file_configuration_named:
- xcc = xccl.ConfigurationNamed(config_name)
- for xck, xcv in build_file_configuration_named.get('xcode_settings',
-- {}).iteritems():
-+ {}).items():
- xcc.SetBuildSetting(xck, xcv)
- if 'xcode_config_file' in build_file_configuration_named:
- config_ref = self.project.AddOrGetFileInRootGroup(
-@@ -273,7 +275,7 @@
- script = script + "\n".join(
- ['export %s="%s"' %
- (key, gyp.xcodeproj_file.ConvertVariablesToShellSyntax(val))
-- for (key, val) in command.get('environment').iteritems()]) + "\n"
-+ for (key, val) in command.get('environment').items()]) + "\n"
-
- # Some test end up using sockets, files on disk, etc. and can get
- # confused if more then one test runs at a time. The generator
-@@ -444,7 +446,7 @@
- dir=self.path)
-
- try:
-- output_file = os.fdopen(output_fd, 'wb')
-+ output_file = os.fdopen(output_fd, 'w')
-
- self.project_file.Print(output_file)
- output_file.close()
-@@ -454,7 +456,7 @@
- same = False
- try:
- same = filecmp.cmp(pbxproj_path, new_pbxproj_path, False)
-- except OSError, e:
-+ except OSError as e:
- if e.errno != errno.ENOENT:
- raise
-
-@@ -473,10 +475,10 @@
- #
- # No way to get the umask without setting a new one? Set a safe one
- # and then set it back to the old value.
-- umask = os.umask(077)
-+ umask = os.umask(0o77)
- os.umask(umask)
-
-- os.chmod(new_pbxproj_path, 0666 & ~umask)
-+ os.chmod(new_pbxproj_path, 0o666 & ~umask)
- os.rename(new_pbxproj_path, pbxproj_path)
-
- except Exception:
-@@ -566,7 +568,7 @@
- def PerformBuild(data, configurations, params):
- options = params['options']
-
-- for build_file, build_file_dict in data.iteritems():
-+ for build_file, build_file_dict in data.items():
- (build_file_root, build_file_ext) = os.path.splitext(build_file)
- if build_file_ext != '.gyp':
- continue
-@@ -577,7 +579,7 @@
- for config in configurations:
- arguments = ['xcodebuild', '-project', xcodeproj_path]
- arguments += ['-configuration', config]
-- print "Building [%s]: %s" % (config, arguments)
-+ print("Building [%s]: %s" % (config, arguments))
- subprocess.check_call(arguments)
-
-
-@@ -625,7 +627,7 @@
- skip_excluded_files = \
- not generator_flags.get('xcode_list_excluded_files', True)
- xcode_projects = {}
-- for build_file, build_file_dict in data.iteritems():
-+ for build_file, build_file_dict in data.items():
- (build_file_root, build_file_ext) = os.path.splitext(build_file)
- if build_file_ext != '.gyp':
- continue
-@@ -744,7 +746,7 @@
- xctarget_type = gyp.xcodeproj_file.PBXNativeTarget
- try:
- target_properties['productType'] = _types[type_bundle_key]
-- except KeyError, e:
-+ except KeyError as e:
- gyp.common.ExceptionAppend(e, "-- unknown product type while "
- "writing target %s" % target_name)
- raise
-@@ -1016,22 +1018,21 @@
- makefile_name)
- # TODO(mark): try/close? Write to a temporary file and swap it only
- # if it's got changes?
-- makefile = open(makefile_path, 'wb')
-+ makefile = open(makefile_path, 'w')
-
- # make will build the first target in the makefile by default. By
- # convention, it's called "all". List all (or at least one)
- # concrete output for each rule source as a prerequisite of the "all"
- # target.
- makefile.write('all: \\\n')
-- for concrete_output_index in \
-- xrange(0, len(concrete_outputs_by_rule_source)):
-+ for concrete_output_index, concrete_output_by_rule_source in \
-+ enumerate(concrete_outputs_by_rule_source):
- # Only list the first (index [0]) concrete output of each input
- # in the "all" target. Otherwise, a parallel make (-j > 1) would
- # attempt to process each input multiple times simultaneously.
- # Otherwise, "all" could just contain the entire list of
- # concrete_outputs_all.
-- concrete_output = \
-- concrete_outputs_by_rule_source[concrete_output_index][0]
-+ concrete_output = concrete_output_by_rule_source[0]
- if concrete_output_index == len(concrete_outputs_by_rule_source) - 1:
- eol = ''
- else:
-@@ -1047,8 +1048,8 @@
- # rule source. Collect the names of the directories that are
- # required.
- concrete_output_dirs = []
-- for concrete_output_index in xrange(0, len(concrete_outputs)):
-- concrete_output = concrete_outputs[concrete_output_index]
-+ for concrete_output_index, concrete_output in \
-+ enumerate(concrete_outputs):
- if concrete_output_index == 0:
- bol = ''
- else:
-@@ -1066,8 +1067,7 @@
- # the set of additional rule inputs, if any.
- prerequisites = [rule_source]
- prerequisites.extend(rule.get('inputs', []))
-- for prerequisite_index in xrange(0, len(prerequisites)):
-- prerequisite = prerequisites[prerequisite_index]
-+ for prerequisite_index, prerequisite in enumerate(prerequisites):
- if prerequisite_index == len(prerequisites) - 1:
- eol = ''
- else:
-@@ -1279,7 +1279,7 @@
- set_define = EscapeXcodeDefine(define)
- xcbc.AppendBuildSetting('GCC_PREPROCESSOR_DEFINITIONS', set_define)
- if 'xcode_settings' in configuration:
-- for xck, xcv in configuration['xcode_settings'].iteritems():
-+ for xck, xcv in configuration['xcode_settings'].items():
- xcbc.SetBuildSetting(xck, xcv)
- if 'xcode_config_file' in configuration:
- config_ref = pbxp.AddOrGetFileInRootGroup(
-@@ -1287,7 +1287,7 @@
- xcbc.SetBaseConfiguration(config_ref)
-
- build_files = []
-- for build_file, build_file_dict in data.iteritems():
-+ for build_file, build_file_dict in data.items():
- if build_file.endswith('.gyp'):
- build_files.append(build_file)
-
-diff --git a/tools/gyp/pylib/gyp/input.py b/pylib/gyp/input.py
-index 21b4606..8ac47cb 100644
---- a/tools/gyp/pylib/gyp/input.py
-+++ b/tools/gyp/pylib/gyp/input.py
-@@ -2,8 +2,9 @@
- # Use of this source code is governed by a BSD-style license that can be
- # found in the LICENSE file.
-
--import ast
-+from __future__ import print_function
-
-+import ast
- import gyp.common
- import gyp.simple_copy
- import multiprocessing
-@@ -231,10 +232,10 @@
- else:
- build_file_data = eval(build_file_contents, {'__builtins__': None},
- None)
-- except SyntaxError, e:
-+ except SyntaxError as e:
- e.filename = build_file_path
- raise
-- except Exception, e:
-+ except Exception as e:
- gyp.common.ExceptionAppend(e, 'while reading ' + build_file_path)
- raise
-
-@@ -254,7 +255,7 @@
- else:
- LoadBuildFileIncludesIntoDict(build_file_data, build_file_path, data,
- aux_data, None, check)
-- except Exception, e:
-+ except Exception as e:
- gyp.common.ExceptionAppend(e,
- 'while reading includes of ' + build_file_path)
- raise
-@@ -291,7 +292,7 @@
- subdict_path, include)
-
- # Recurse into subdictionaries.
-- for k, v in subdict.iteritems():
-+ for k, v in subdict.items():
- if type(v) is dict:
- LoadBuildFileIncludesIntoDict(v, subdict_path, data, aux_data,
- None, check)
-@@ -456,7 +457,7 @@
- try:
- LoadTargetBuildFile(dependency, data, aux_data, variables,
- includes, depth, check, load_dependencies)
-- except Exception, e:
-+ except Exception as e:
- gyp.common.ExceptionAppend(
- e, 'while loading dependencies of %s' % build_file_path)
- raise
-@@ -477,7 +478,7 @@
- signal.signal(signal.SIGINT, signal.SIG_IGN)
-
- # Apply globals so that the worker process behaves the same.
-- for key, value in global_flags.iteritems():
-+ for key, value in global_flags.items():
- globals()[key] = value
-
- SetGeneratorGlobals(generator_input_info)
-@@ -499,12 +500,12 @@
- return (build_file_path,
- build_file_data,
- dependencies)
-- except GypError, e:
-+ except GypError as e:
- sys.stderr.write("gyp: %s\n" % e)
- return None
-- except Exception, e:
-- print >>sys.stderr, 'Exception:', e
-- print >>sys.stderr, traceback.format_exc()
-+ except Exception as e:
-+ print('Exception:', e, file=sys.stderr)
-+ print(traceback.format_exc(), file=sys.stderr)
- return None
-
-
-@@ -594,7 +595,7 @@
- args = (global_flags, dependency,
- variables, includes, depth, check, generator_input_info),
- callback = parallel_state.LoadTargetBuildFileCallback)
-- except KeyboardInterrupt, e:
-+ except KeyboardInterrupt as e:
- parallel_state.pool.terminate()
- raise e
-
-@@ -894,7 +895,7 @@
- stderr=subprocess.PIPE,
- stdin=subprocess.PIPE,
- cwd=build_file_dir)
-- except Exception, e:
-+ except Exception as e:
- raise GypError("%s while executing command '%s' in %s" %
- (e, contents, build_file))
-
-@@ -1008,9 +1009,9 @@
-
- # Convert all strings that are canonically-represented integers into integers.
- if type(output) is list:
-- for index in xrange(0, len(output)):
-- if IsStrCanonicalInt(output[index]):
-- output[index] = int(output[index])
-+ for index, outstr in enumerate(output):
-+ if IsStrCanonicalInt(outstr):
-+ output[index] = int(outstr)
- elif IsStrCanonicalInt(output):
- output = int(output)
-
-@@ -1079,13 +1080,13 @@
- if eval(ast_code, {'__builtins__': None}, variables):
- return true_dict
- return false_dict
-- except SyntaxError, e:
-+ except SyntaxError as e:
- syntax_error = SyntaxError('%s while evaluating condition \'%s\' in %s '
- 'at character %d.' %
- (str(e.args[0]), e.text, build_file, e.offset),
- e.filename, e.lineno, e.offset, e.text)
- raise syntax_error
-- except NameError, e:
-+ except NameError as e:
- gyp.common.ExceptionAppend(e, 'while evaluating condition \'%s\' in %s' %
- (cond_expr_expanded, build_file))
- raise GypError(e)
-@@ -1140,7 +1141,7 @@
- def LoadAutomaticVariablesFromDict(variables, the_dict):
- # Any keys with plain string values in the_dict become automatic variables.
- # The variable name is the key name with a "_" character prepended.
-- for key, value in the_dict.iteritems():
-+ for key, value in the_dict.items():
- if type(value) in (str, int, list):
- variables['_' + key] = value
-
-@@ -1153,7 +1154,7 @@
- # the_dict in the_dict's parent dict. If the_dict's parent is not a dict
- # (it could be a list or it could be parentless because it is a root dict),
- # the_dict_key will be None.
-- for key, value in the_dict.get('variables', {}).iteritems():
-+ for key, value in the_dict.get('variables', {}).items():
- if type(value) not in (str, int, list):
- continue
-
-@@ -1192,7 +1193,7 @@
- # list before we process them so that you can reference one
- # variable from another. They will be fully expanded by recursion
- # in ExpandVariables.
-- for key, value in the_dict['variables'].iteritems():
-+ for key, value in the_dict['variables'].items():
- variables[key] = value
-
- # Handle the associated variables dict first, so that any variable
-@@ -1205,7 +1206,7 @@
-
- LoadVariablesFromVariablesDict(variables, the_dict, the_dict_key)
-
-- for key, value in the_dict.iteritems():
-+ for key, value in the_dict.items():
- # Skip "variables", which was already processed if present.
- if key != 'variables' and type(value) is str:
- expanded = ExpandVariables(value, phase, variables, build_file)
-@@ -1263,7 +1264,7 @@
-
- # Recurse into child dicts, or process child lists which may result in
- # further recursion into descendant dicts.
-- for key, value in the_dict.iteritems():
-+ for key, value in the_dict.items():
- # Skip "variables" and string values, which were already processed if
- # present.
- if key == 'variables' or type(value) is str:
-@@ -1360,14 +1361,14 @@
- for dep in dependency_sections
- for op in ('', '!', '/')]
-
-- for target, target_dict in targets.iteritems():
-+ for target, target_dict in targets.items():
- target_build_file = gyp.common.BuildFile(target)
- toolset = target_dict['toolset']
- for dependency_key in all_dependency_sections:
- dependencies = target_dict.get(dependency_key, [])
-- for index in xrange(0, len(dependencies)):
-+ for index, dep in enumerate(dependencies):
- dep_file, dep_target, dep_toolset = gyp.common.ResolveTarget(
-- target_build_file, dependencies[index], toolset)
-+ target_build_file, dep, toolset)
- if not multiple_toolsets:
- # Ignore toolset specification in the dependency if it is specified.
- dep_toolset = toolset
-@@ -1400,7 +1401,7 @@
- dependency list, must be qualified when this function is called.
- """
-
-- for target, target_dict in targets.iteritems():
-+ for target, target_dict in targets.items():
- toolset = target_dict['toolset']
- target_build_file = gyp.common.BuildFile(target)
- for dependency_key in dependency_sections:
-@@ -1462,7 +1463,7 @@
- def RemoveDuplicateDependencies(targets):
- """Makes sure every dependency appears only once in all targets's dependency
- lists."""
-- for target_name, target_dict in targets.iteritems():
-+ for target_name, target_dict in targets.items():
- for dependency_key in dependency_sections:
- dependencies = target_dict.get(dependency_key, [])
- if dependencies:
-@@ -1478,7 +1479,7 @@
- def RemoveSelfDependencies(targets):
- """Remove self dependencies from targets that have the prune_self_dependency
- variable set."""
-- for target_name, target_dict in targets.iteritems():
-+ for target_name, target_dict in targets.items():
- for dependency_key in dependency_sections:
- dependencies = target_dict.get(dependency_key, [])
- if dependencies:
-@@ -1491,7 +1492,7 @@
- def RemoveLinkDependenciesFromNoneTargets(targets):
- """Remove dependencies having the 'link_dependency' attribute from the 'none'
- targets."""
-- for target_name, target_dict in targets.iteritems():
-+ for target_name, target_dict in targets.items():
- for dependency_key in dependency_sections:
- dependencies = target_dict.get(dependency_key, [])
- if dependencies:
-@@ -1783,14 +1784,14 @@
- # Create a DependencyGraphNode for each target. Put it into a dict for easy
- # access.
- dependency_nodes = {}
-- for target, spec in targets.iteritems():
-+ for target, spec in targets.items():
- if target not in dependency_nodes:
- dependency_nodes[target] = DependencyGraphNode(target)
-
- # Set up the dependency links. Targets that have no dependencies are treated
- # as dependent on root_node.
- root_node = DependencyGraphNode(None)
-- for target, spec in targets.iteritems():
-+ for target, spec in targets.items():
- target_node = dependency_nodes[target]
- target_build_file = gyp.common.BuildFile(target)
- dependencies = spec.get('dependencies')
-@@ -1814,7 +1815,7 @@
- if not root_node.dependents:
- # If all targets have dependencies, add the first target as a dependent
- # of root_node so that the cycle can be discovered from root_node.
-- target = targets.keys()[0]
-+ target = next(iter(targets))
- target_node = dependency_nodes[target]
- target_node.dependencies.append(root_node)
- root_node.dependents.append(target_node)
-@@ -1833,20 +1834,20 @@
- # Create a DependencyGraphNode for each gyp file containing a target. Put
- # it into a dict for easy access.
- dependency_nodes = {}
-- for target in targets.iterkeys():
-+ for target in targets.keys():
- build_file = gyp.common.BuildFile(target)
- if not build_file in dependency_nodes:
- dependency_nodes[build_file] = DependencyGraphNode(build_file)
-
- # Set up the dependency links.
-- for target, spec in targets.iteritems():
-+ for target, spec in targets.items():
- build_file = gyp.common.BuildFile(target)
- build_file_node = dependency_nodes[build_file]
- target_dependencies = spec.get('dependencies', [])
- for dependency in target_dependencies:
- try:
- dependency_build_file = gyp.common.BuildFile(dependency)
-- except GypError, e:
-+ except GypError as e:
- gyp.common.ExceptionAppend(
- e, 'while computing dependencies of .gyp file %s' % build_file)
- raise
-@@ -1864,7 +1865,7 @@
-
- # Files that have no dependencies are treated as dependent on root_node.
- root_node = DependencyGraphNode(None)
-- for build_file_node in dependency_nodes.itervalues():
-+ for build_file_node in dependency_nodes.values():
- if len(build_file_node.dependencies) == 0:
- build_file_node.dependencies.append(root_node)
- root_node.dependents.append(build_file_node)
-@@ -1877,7 +1878,7 @@
- if not root_node.dependents:
- # If all files have dependencies, add the first file as a dependent
- # of root_node so that the cycle can be discovered from root_node.
-- file_node = dependency_nodes.values()[0]
-+ file_node = next(iter(dependency_nodes.values()))
- file_node.dependencies.append(root_node)
- root_node.dependents.append(file_node)
- cycles = []
-@@ -2104,7 +2105,7 @@
-
- def MergeDicts(to, fro, to_file, fro_file):
- # I wanted to name the parameter "from" but it's a Python keyword...
-- for k, v in fro.iteritems():
-+ for k, v in fro.items():
- # It would be nice to do "if not k in to: to[k] = v" but that wouldn't give
- # copy semantics. Something else may want to merge from the |fro| dict
- # later, and having the same dict ref pointed to twice in the tree isn't
-@@ -2239,13 +2240,13 @@
- if not 'configurations' in target_dict:
- target_dict['configurations'] = {'Default': {}}
- if not 'default_configuration' in target_dict:
-- concrete = [i for (i, config) in target_dict['configurations'].iteritems()
-+ concrete = [i for (i, config) in target_dict['configurations'].items()
- if not config.get('abstract')]
- target_dict['default_configuration'] = sorted(concrete)[0]
-
- merged_configurations = {}
- configs = target_dict['configurations']
-- for (configuration, old_configuration_dict) in configs.iteritems():
-+ for (configuration, old_configuration_dict) in configs.items():
- # Skip abstract configurations (saves work only).
- if old_configuration_dict.get('abstract'):
- continue
-@@ -2253,7 +2254,7 @@
- # Get the inheritance relationship right by making a copy of the target
- # dict.
- new_configuration_dict = {}
-- for (key, target_val) in target_dict.iteritems():
-+ for (key, target_val) in target_dict.items():
- key_ext = key[-1:]
- if key_ext in key_suffixes:
- key_base = key[:-1]
-@@ -2274,10 +2275,9 @@
- merged_configurations[configuration])
-
- # Now drop all the abstract ones.
-- for configuration in target_dict['configurations'].keys():
-- old_configuration_dict = target_dict['configurations'][configuration]
-- if old_configuration_dict.get('abstract'):
-- del target_dict['configurations'][configuration]
-+ configs = target_dict['configurations']
-+ target_dict['configurations'] = \
-+ {k: v for k, v in configs.items() if not v.get('abstract')}
-
- # Now that all of the target's configurations have been built, go through
- # the target dict's keys and remove everything that's been moved into a
-@@ -2337,7 +2337,7 @@
-
- lists = []
- del_lists = []
-- for key, value in the_dict.iteritems():
-+ for key, value in the_dict.items():
- operation = key[-1]
- if operation != '!' and operation != '/':
- continue
-@@ -2385,8 +2385,8 @@
- exclude_key = list_key + '!'
- if exclude_key in the_dict:
- for exclude_item in the_dict[exclude_key]:
-- for index in xrange(0, len(the_list)):
-- if exclude_item == the_list[index]:
-+ for index, list_item in enumerate(the_list):
-+ if exclude_item == list_item:
- # This item matches the exclude_item, so set its action to 0
- # (exclude).
- list_actions[index] = 0
-@@ -2411,8 +2411,7 @@
- raise ValueError('Unrecognized action ' + action + ' in ' + name + \
- ' key ' + regex_key)
-
-- for index in xrange(0, len(the_list)):
-- list_item = the_list[index]
-+ for index, list_item in enumerate(the_list):
- if list_actions[index] == action_value:
- # Even if the regex matches, nothing will change so continue (regex
- # searches are expensive).
-@@ -2442,7 +2441,7 @@
- # the indices of items that haven't been seen yet don't shift. That means
- # that things need to be prepended to excluded_list to maintain them in the
- # same order that they existed in the_list.
-- for index in xrange(len(list_actions) - 1, -1, -1):
-+ for index in range(len(list_actions) - 1, -1, -1):
- if list_actions[index] == 0:
- # Dump anything with action 0 (exclude). Keep anything with action 1
- # (include) or -1 (no include or exclude seen for the item).
-@@ -2455,7 +2454,7 @@
- the_dict[excluded_key] = excluded_list
-
- # Now recurse into subdicts and lists that may contain dicts.
-- for key, value in the_dict.iteritems():
-+ for key, value in the_dict.items():
- if type(value) is dict:
- ProcessListFiltersInDict(key, value)
- elif type(value) is list:
-@@ -2512,7 +2511,7 @@
- basenames.setdefault(basename, []).append(source)
-
- error = ''
-- for basename, files in basenames.iteritems():
-+ for basename, files in basenames.items():
- if len(files) > 1:
- error += ' %s: %s\n' % (basename, ' '.join(files))
-
-@@ -2651,8 +2650,7 @@
- def TurnIntIntoStrInList(the_list):
- """Given list the_list, recursively converts all integers into strings.
- """
-- for index in xrange(0, len(the_list)):
-- item = the_list[index]
-+ for index, item in enumerate(the_list):
- if type(item) is int:
- the_list[index] = str(item)
- elif type(item) is dict:
-@@ -2769,7 +2767,7 @@
- try:
- LoadTargetBuildFile(build_file, data, aux_data,
- variables, includes, depth, check, True)
-- except Exception, e:
-+ except Exception as e:
- gyp.common.ExceptionAppend(e, 'while trying to load %s' % build_file)
- raise
-
-@@ -2791,7 +2789,7 @@
- RemoveLinkDependenciesFromNoneTargets(targets)
-
- # Apply exclude (!) and regex (/) list filters only for dependency_sections.
-- for target_name, target_dict in targets.iteritems():
-+ for target_name, target_dict in targets.items():
- tmp_dict = {}
- for key_base in dependency_sections:
- for op in ('', '!', '/'):
-diff --git a/tools/gyp/pylib/gyp/input_test.py b/pylib/gyp/input_test.py
-index 4234fbb..6c4b1cc 100755
---- a/tools/gyp/pylib/gyp/input_test.py
-+++ b/tools/gyp/pylib/gyp/input_test.py
-@@ -22,7 +22,7 @@
- dependency.dependents.append(dependent)
-
- def test_no_cycle_empty_graph(self):
-- for label, node in self.nodes.iteritems():
-+ for label, node in self.nodes.items():
- self.assertEquals([], node.FindCycles())
-
- def test_no_cycle_line(self):
-@@ -30,7 +30,7 @@
- self._create_dependency(self.nodes['b'], self.nodes['c'])
- self._create_dependency(self.nodes['c'], self.nodes['d'])
-
-- for label, node in self.nodes.iteritems():
-+ for label, node in self.nodes.items():
- self.assertEquals([], node.FindCycles())
-
- def test_no_cycle_dag(self):
-@@ -38,7 +38,7 @@
- self._create_dependency(self.nodes['a'], self.nodes['c'])
- self._create_dependency(self.nodes['b'], self.nodes['c'])
-
-- for label, node in self.nodes.iteritems():
-+ for label, node in self.nodes.items():
- self.assertEquals([], node.FindCycles())
-
- def test_cycle_self_reference(self):
-diff --git a/tools/gyp/pylib/gyp/mac_tool.py b/pylib/gyp/mac_tool.py
-index 0ad7e7a..7d3a8c2 100755
---- a/tools/gyp/pylib/gyp/mac_tool.py
-+++ b/tools/gyp/pylib/gyp/mac_tool.py
-@@ -8,6 +8,8 @@
- These functions are executed via gyp-mac-tool when using the Makefile generator.
- """
-
-+from __future__ import print_function
-+
- import fcntl
- import fnmatch
- import glob
-@@ -16,7 +18,6 @@
- import plistlib
- import re
- import shutil
--import string
- import struct
- import subprocess
- import sys
-@@ -155,11 +156,11 @@
- fp.close()
- return None
- fp.close()
-- if header.startswith("\xFE\xFF"):
-+ if header.startswith(b"\xFE\xFF"):
- return "UTF-16"
-- elif header.startswith("\xFF\xFE"):
-+ elif header.startswith(b"\xFF\xFE"):
- return "UTF-16"
-- elif header.startswith("\xEF\xBB\xBF"):
-+ elif header.startswith(b"\xEF\xBB\xBF"):
- return "UTF-8"
- else:
- return None
-@@ -174,7 +175,7 @@
- # Insert synthesized key/value pairs (e.g. BuildMachineOSBuild).
- plist = plistlib.readPlistFromString(lines)
- if keys:
-- plist = dict(plist.items() + json.loads(keys[0]).items())
-+ plist.update(json.loads(keys[0]))
- lines = plistlib.writePlistToString(plist)
-
- # Go through all the environment variables and replace them as variables in
-@@ -185,7 +186,7 @@
- continue
- evar = '${%s}' % key
- evalue = os.environ[key]
-- lines = string.replace(lines, evar, evalue)
-+ lines = lines.replace(evar, evalue)
-
- # Xcode supports various suffices on environment variables, which are
- # all undocumented. :rfc1034identifier is used in the standard project
-@@ -195,11 +196,11 @@
- # in a URL either -- oops, hence :rfc1034identifier was born.
- evar = '${%s:identifier}' % key
- evalue = IDENT_RE.sub('_', os.environ[key])
-- lines = string.replace(lines, evar, evalue)
-+ lines = lines.replace(evar, evalue)
-
- evar = '${%s:rfc1034identifier}' % key
- evalue = IDENT_RE.sub('-', os.environ[key])
-- lines = string.replace(lines, evar, evalue)
-+ lines = lines.replace(evar, evalue)
-
- # Remove any keys with values that haven't been replaced.
- lines = lines.split('\n')
-@@ -270,7 +271,7 @@
- _, err = libtoolout.communicate()
- for line in err.splitlines():
- if not libtool_re.match(line) and not libtool_re5.match(line):
-- print >>sys.stderr, line
-+ print(line, file=sys.stderr)
- # Unconditionally touch the output .a file on the command line if present
- # and the command succeeded. A bit hacky.
- if not libtoolout.returncode:
-@@ -385,7 +386,7 @@
- ])
- if keys:
- keys = json.loads(keys)
-- for key, value in keys.iteritems():
-+ for key, value in keys.items():
- arg_name = '--' + key
- if isinstance(value, bool):
- if value:
-@@ -480,8 +481,9 @@
- profiles_dir = os.path.join(
- os.environ['HOME'], 'Library', 'MobileDevice', 'Provisioning Profiles')
- if not os.path.isdir(profiles_dir):
-- print >>sys.stderr, (
-- 'cannot find mobile provisioning for %s' % bundle_identifier)
-+ print((
-+ 'cannot find mobile provisioning for %s' % bundle_identifier),
-+ file=sys.stderr)
- sys.exit(1)
- provisioning_profiles = None
- if profile:
-@@ -502,8 +504,9 @@
- valid_provisioning_profiles[app_id_pattern] = (
- profile_path, profile_data, team_identifier)
- if not valid_provisioning_profiles:
-- print >>sys.stderr, (
-- 'cannot find mobile provisioning for %s' % bundle_identifier)
-+ print((
-+ 'cannot find mobile provisioning for %s' % bundle_identifier),
-+ file=sys.stderr)
- sys.exit(1)
- # If the user has multiple provisioning profiles installed that can be
- # used for ${bundle_identifier}, pick the most specific one (ie. the
-@@ -527,7 +530,7 @@
-
- def _MergePlist(self, merged_plist, plist):
- """Merge |plist| into |merged_plist|."""
-- for key, value in plist.iteritems():
-+ for key, value in plist.items():
- if isinstance(value, dict):
- merged_value = merged_plist.get(key, {})
- if isinstance(merged_value, dict):
-@@ -637,7 +640,7 @@
- the key was not found.
- """
- if isinstance(data, str):
-- for key, value in substitutions.iteritems():
-+ for key, value in substitutions.items():
- data = data.replace('$(%s)' % key, value)
- return data
- if isinstance(data, list):
-diff --git a/tools/gyp/pylib/gyp/msvs_emulation.py b/pylib/gyp/msvs_emulation.py
-index 6d5b5bd..63d40e6 100644
---- a/tools/gyp/pylib/gyp/msvs_emulation.py
-+++ b/tools/gyp/pylib/gyp/msvs_emulation.py
-@@ -7,6 +7,7 @@
- build systems, primarily ninja.
- """
-
-+import collections
- import os
- import re
- import subprocess
-@@ -16,6 +17,12 @@
- import gyp.MSVSUtil
- import gyp.MSVSVersion
-
-+try:
-+ # basestring was removed in python3.
-+ basestring
-+except NameError:
-+ basestring = str
-+
-
- windows_quoter_regex = re.compile(r'(\\*)"')
-
-@@ -84,8 +91,8 @@
- """Add |prefix| to |element| or each subelement if element is iterable."""
- if element is None:
- return element
-- # Note, not Iterable because we don't want to handle strings like that.
-- if isinstance(element, list) or isinstance(element, tuple):
-+ if (isinstance(element, collections.Iterable) and
-+ not isinstance(element, basestring)):
- return [prefix + e for e in element]
- else:
- return prefix + element
-@@ -97,7 +104,8 @@
- if map is not None and element is not None:
- if not callable(map):
- map = map.get # Assume it's a dict, otherwise a callable to do the remap.
-- if isinstance(element, list) or isinstance(element, tuple):
-+ if (isinstance(element, collections.Iterable) and
-+ not isinstance(element, basestring)):
- element = filter(None, [map(elem) for elem in element])
- else:
- element = map(element)
-@@ -109,7 +117,8 @@
- then add |element| to it, adding each item in |element| if it's a list or
- tuple."""
- if append is not None and element is not None:
-- if isinstance(element, list) or isinstance(element, tuple):
-+ if (isinstance(element, collections.Iterable) and
-+ not isinstance(element, basestring)):
- append.extend(element)
- else:
- append.append(element)
-@@ -209,7 +218,7 @@
- configs = spec['configurations']
- for field, default in supported_fields:
- setattr(self, field, {})
-- for configname, config in configs.iteritems():
-+ for configname, config in configs.items():
- getattr(self, field)[configname] = config.get(field, default())
-
- self.msvs_cygwin_dirs = spec.get('msvs_cygwin_dirs', ['.'])
-@@ -482,7 +491,7 @@
- # https://msdn.microsoft.com/en-us/library/dn502518.aspx
- cflags.append('/FS')
- # ninja handles parallelism by itself, don't have the compiler do it too.
-- cflags = filter(lambda x: not x.startswith('/MP'), cflags)
-+ cflags = [x for x in cflags if not x.startswith('/MP')]
- return cflags
-
- def _GetPchFlags(self, config, extension):
-@@ -649,19 +658,17 @@
-
- # If the base address is not specifically controlled, DYNAMICBASE should
- # be on by default.
-- base_flags = filter(lambda x: 'DYNAMICBASE' in x or x == '/FIXED',
-- ldflags)
-- if not base_flags:
-+ if not any('DYNAMICBASE' in flag or flag == '/FIXED' for flag in ldflags):
- ldflags.append('/DYNAMICBASE')
-
- # If the NXCOMPAT flag has not been specified, default to on. Despite the
- # documentation that says this only defaults to on when the subsystem is
- # Vista or greater (which applies to the linker), the IDE defaults it on
- # unless it's explicitly off.
-- if not filter(lambda x: 'NXCOMPAT' in x, ldflags):
-+ if not any('NXCOMPAT' in flag for flag in ldflags):
- ldflags.append('/NXCOMPAT')
-
-- have_def_file = filter(lambda x: x.startswith('/DEF:'), ldflags)
-+ have_def_file = any(flag.startswith('/DEF:') for flag in ldflags)
- manifest_flags, intermediate_manifest, manifest_files = \
- self._GetLdManifestFlags(config, manifest_base_name, gyp_to_build_path,
- is_executable and not have_def_file, build_dir)
-@@ -953,7 +960,7 @@
- """Expand $(Variable) per expansions dict. See MsvsSettings.GetVSMacroEnv
- for the canonical way to retrieve a suitable dict."""
- if '$' in string:
-- for old, new in expansions.iteritems():
-+ for old, new in expansions.items():
- assert '$(' not in new, new
- string = string.replace(old, new)
- return string
-@@ -1001,7 +1008,7 @@
- CreateProcess documentation for more details."""
- block = ''
- nul = '\0'
-- for key, value in envvar_dict.iteritems():
-+ for key, value in envvar_dict.items():
- block += key + '=' + value + nul
- block += nul
- return block
-@@ -1056,7 +1063,7 @@
- env['INCLUDE'] = ';'.join(system_includes)
-
- env_block = _FormatAsEnvironmentBlock(env)
-- f = open_out(os.path.join(toplevel_build_dir, 'environment.' + arch), 'wb')
-+ f = open_out(os.path.join(toplevel_build_dir, 'environment.' + arch), 'w')
- f.write(env_block)
- f.close()
-
-@@ -1078,7 +1085,7 @@
- if int(generator_flags.get('msvs_error_on_missing_sources', 0)):
- no_specials = filter(lambda x: '$' not in x, sources)
- relative = [os.path.join(build_dir, gyp_to_ninja(s)) for s in no_specials]
-- missing = filter(lambda x: not os.path.exists(x), relative)
-+ missing = [x for x in relative if not os.path.exists(x)]
- if missing:
- # They'll look like out\Release\..\..\stuff\things.cc, so normalize the
- # path for a slightly less crazy looking output.
-diff --git a/tools/gyp/pylib/gyp/ordered_dict.py b/pylib/gyp/ordered_dict.py
-deleted file mode 100644
-index a1e89f9..0000000
---- a/tools/gyp/pylib/gyp/ordered_dict.py
-+++ /dev/null
-@@ -1,289 +0,0 @@
--# Unmodified from http://code.activestate.com/recipes/576693/
--# other than to add MIT license header (as specified on page, but not in code).
--# Linked from Python documentation here:
--# http://docs.python.org/2/library/collections.html#collections.OrderedDict
--#
--# This should be deleted once Py2.7 is available on all bots, see
--# http://crbug.com/241769.
--#
--# Copyright (c) 2009 Raymond Hettinger.
--#
--# Permission is hereby granted, free of charge, to any person obtaining a copy
--# of this software and associated documentation files (the "Software"), to deal
--# in the Software without restriction, including without limitation the rights
--# to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
--# copies of the Software, and to permit persons to whom the Software is
--# furnished to do so, subject to the following conditions:
--#
--# The above copyright notice and this permission notice shall be included in
--# all copies or substantial portions of the Software.
--#
--# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
--# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
--# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
--# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
--# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
--# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
--# THE SOFTWARE.
--
--# Backport of OrderedDict() class that runs on Python 2.4, 2.5, 2.6, 2.7 and pypy.
--# Passes Python2.7's test suite and incorporates all the latest updates.
--
--try:
-- from thread import get_ident as _get_ident
--except ImportError:
-- from dummy_thread import get_ident as _get_ident
--
--try:
-- from _abcoll import KeysView, ValuesView, ItemsView
--except ImportError:
-- pass
--
--
--class OrderedDict(dict):
-- 'Dictionary that remembers insertion order'
-- # An inherited dict maps keys to values.
-- # The inherited dict provides __getitem__, __len__, __contains__, and get.
-- # The remaining methods are order-aware.
-- # Big-O running times for all methods are the same as for regular dictionaries.
--
-- # The internal self.__map dictionary maps keys to links in a doubly linked list.
-- # The circular doubly linked list starts and ends with a sentinel element.
-- # The sentinel element never gets deleted (this simplifies the algorithm).
-- # Each link is stored as a list of length three: [PREV, NEXT, KEY].
--
-- def __init__(self, *args, **kwds):
-- '''Initialize an ordered dictionary. Signature is the same as for
-- regular dictionaries, but keyword arguments are not recommended
-- because their insertion order is arbitrary.
--
-- '''
-- if len(args) > 1:
-- raise TypeError('expected at most 1 arguments, got %d' % len(args))
-- try:
-- self.__root
-- except AttributeError:
-- self.__root = root = [] # sentinel node
-- root[:] = [root, root, None]
-- self.__map = {}
-- self.__update(*args, **kwds)
--
-- def __setitem__(self, key, value, dict_setitem=dict.__setitem__):
-- 'od.__setitem__(i, y) <==> od[i]=y'
-- # Setting a new item creates a new link which goes at the end of the linked
-- # list, and the inherited dictionary is updated with the new key/value pair.
-- if key not in self:
-- root = self.__root
-- last = root[0]
-- last[1] = root[0] = self.__map[key] = [last, root, key]
-- dict_setitem(self, key, value)
--
-- def __delitem__(self, key, dict_delitem=dict.__delitem__):
-- 'od.__delitem__(y) <==> del od[y]'
-- # Deleting an existing item uses self.__map to find the link which is
-- # then removed by updating the links in the predecessor and successor nodes.
-- dict_delitem(self, key)
-- link_prev, link_next, key = self.__map.pop(key)
-- link_prev[1] = link_next
-- link_next[0] = link_prev
--
-- def __iter__(self):
-- 'od.__iter__() <==> iter(od)'
-- root = self.__root
-- curr = root[1]
-- while curr is not root:
-- yield curr[2]
-- curr = curr[1]
--
-- def __reversed__(self):
-- 'od.__reversed__() <==> reversed(od)'
-- root = self.__root
-- curr = root[0]
-- while curr is not root:
-- yield curr[2]
-- curr = curr[0]
--
-- def clear(self):
-- 'od.clear() -> None. Remove all items from od.'
-- try:
-- for node in self.__map.itervalues():
-- del node[:]
-- root = self.__root
-- root[:] = [root, root, None]
-- self.__map.clear()
-- except AttributeError:
-- pass
-- dict.clear(self)
--
-- def popitem(self, last=True):
-- '''od.popitem() -> (k, v), return and remove a (key, value) pair.
-- Pairs are returned in LIFO order if last is true or FIFO order if false.
--
-- '''
-- if not self:
-- raise KeyError('dictionary is empty')
-- root = self.__root
-- if last:
-- link = root[0]
-- link_prev = link[0]
-- link_prev[1] = root
-- root[0] = link_prev
-- else:
-- link = root[1]
-- link_next = link[1]
-- root[1] = link_next
-- link_next[0] = root
-- key = link[2]
-- del self.__map[key]
-- value = dict.pop(self, key)
-- return key, value
--
-- # -- the following methods do not depend on the internal structure --
--
-- def keys(self):
-- 'od.keys() -> list of keys in od'
-- return list(self)
--
-- def values(self):
-- 'od.values() -> list of values in od'
-- return [self[key] for key in self]
--
-- def items(self):
-- 'od.items() -> list of (key, value) pairs in od'
-- return [(key, self[key]) for key in self]
--
-- def iterkeys(self):
-- 'od.iterkeys() -> an iterator over the keys in od'
-- return iter(self)
--
-- def itervalues(self):
-- 'od.itervalues -> an iterator over the values in od'
-- for k in self:
-- yield self[k]
--
-- def iteritems(self):
-- 'od.iteritems -> an iterator over the (key, value) items in od'
-- for k in self:
-- yield (k, self[k])
--
-- # Suppress 'OrderedDict.update: Method has no argument':
-- # pylint: disable=E0211
-- def update(*args, **kwds):
-- '''od.update(E, **F) -> None. Update od from dict/iterable E and F.
--
-- If E is a dict instance, does: for k in E: od[k] = E[k]
-- If E has a .keys() method, does: for k in E.keys(): od[k] = E[k]
-- Or if E is an iterable of items, does: for k, v in E: od[k] = v
-- In either case, this is followed by: for k, v in F.items(): od[k] = v
--
-- '''
-- if len(args) > 2:
-- raise TypeError('update() takes at most 2 positional '
-- 'arguments (%d given)' % (len(args),))
-- elif not args:
-- raise TypeError('update() takes at least 1 argument (0 given)')
-- self = args[0]
-- # Make progressively weaker assumptions about "other"
-- other = ()
-- if len(args) == 2:
-- other = args[1]
-- if isinstance(other, dict):
-- for key in other:
-- self[key] = other[key]
-- elif hasattr(other, 'keys'):
-- for key in other.keys():
-- self[key] = other[key]
-- else:
-- for key, value in other:
-- self[key] = value
-- for key, value in kwds.items():
-- self[key] = value
--
-- __update = update # let subclasses override update without breaking __init__
--
-- __marker = object()
--
-- def pop(self, key, default=__marker):
-- '''od.pop(k[,d]) -> v, remove specified key and return the corresponding value.
-- If key is not found, d is returned if given, otherwise KeyError is raised.
--
-- '''
-- if key in self:
-- result = self[key]
-- del self[key]
-- return result
-- if default is self.__marker:
-- raise KeyError(key)
-- return default
--
-- def setdefault(self, key, default=None):
-- 'od.setdefault(k[,d]) -> od.get(k,d), also set od[k]=d if k not in od'
-- if key in self:
-- return self[key]
-- self[key] = default
-- return default
--
-- def __repr__(self, _repr_running={}):
-- 'od.__repr__() <==> repr(od)'
-- call_key = id(self), _get_ident()
-- if call_key in _repr_running:
-- return '...'
-- _repr_running[call_key] = 1
-- try:
-- if not self:
-- return '%s()' % (self.__class__.__name__,)
-- return '%s(%r)' % (self.__class__.__name__, self.items())
-- finally:
-- del _repr_running[call_key]
--
-- def __reduce__(self):
-- 'Return state information for pickling'
-- items = [[k, self[k]] for k in self]
-- inst_dict = vars(self).copy()
-- for k in vars(OrderedDict()):
-- inst_dict.pop(k, None)
-- if inst_dict:
-- return (self.__class__, (items,), inst_dict)
-- return self.__class__, (items,)
--
-- def copy(self):
-- 'od.copy() -> a shallow copy of od'
-- return self.__class__(self)
--
-- @classmethod
-- def fromkeys(cls, iterable, value=None):
-- '''OD.fromkeys(S[, v]) -> New ordered dictionary with keys from S
-- and values equal to v (which defaults to None).
--
-- '''
-- d = cls()
-- for key in iterable:
-- d[key] = value
-- return d
--
-- def __eq__(self, other):
-- '''od.__eq__(y) <==> od==y. Comparison to another OD is order-sensitive
-- while comparison to a regular mapping is order-insensitive.
--
-- '''
-- if isinstance(other, OrderedDict):
-- return len(self)==len(other) and self.items() == other.items()
-- return dict.__eq__(self, other)
--
-- def __ne__(self, other):
-- return not self == other
--
-- # -- the following methods are only used in Python 2.7 --
--
-- def viewkeys(self):
-- "od.viewkeys() -> a set-like object providing a view on od's keys"
-- return KeysView(self)
--
-- def viewvalues(self):
-- "od.viewvalues() -> an object providing a view on od's values"
-- return ValuesView(self)
--
-- def viewitems(self):
-- "od.viewitems() -> a set-like object providing a view on od's items"
-- return ItemsView(self)
--
-diff --git a/tools/gyp/pylib/gyp/simple_copy.py b/pylib/gyp/simple_copy.py
-index eaf5f8b..58a61c3 100644
---- a/tools/gyp/pylib/gyp/simple_copy.py
-+++ b/tools/gyp/pylib/gyp/simple_copy.py
-@@ -49,7 +49,7 @@
-
- def _deepcopy_dict(x):
- y = {}
-- for key, value in x.iteritems():
-+ for key, value in x.items():
- y[deepcopy(key)] = deepcopy(value)
- return y
- d[dict] = _deepcopy_dict
-diff --git a/tools/gyp/pylib/gyp/win_tool.py b/pylib/gyp/win_tool.py
-index 1c843a0..8973484 100755
---- a/tools/gyp/pylib/gyp/win_tool.py
-+++ b/tools/gyp/pylib/gyp/win_tool.py
-@@ -9,6 +9,8 @@
- These functions are executed via gyp-win-tool when using the ninja generator.
- """
-
-+from __future__ import print_function
-+
- import os
- import re
- import shutil
-@@ -134,7 +136,7 @@
- if (not line.startswith(' Creating library ') and
- not line.startswith('Generating code') and
- not line.startswith('Finished generating code')):
-- print line
-+ print(line)
- return link.returncode
-
- def ExecLinkWithManifests(self, arch, embed_manifest, out, ldcmd, resname,
-@@ -193,16 +195,18 @@
- our_manifest = '%(out)s.manifest' % variables
- # Load and normalize the manifests. mt.exe sometimes removes whitespace,
- # and sometimes doesn't unfortunately.
-- with open(our_manifest, 'rb') as our_f:
-- with open(assert_manifest, 'rb') as assert_f:
-+ with open(our_manifest, 'r') as our_f:
-+ with open(assert_manifest, 'r') as assert_f:
- our_data = our_f.read().translate(None, string.whitespace)
- assert_data = assert_f.read().translate(None, string.whitespace)
- if our_data != assert_data:
- os.unlink(out)
- def dump(filename):
-- sys.stderr.write('%s\n-----\n' % filename)
-- with open(filename, 'rb') as f:
-- sys.stderr.write(f.read() + '\n-----\n')
-+ print(filename, file=sys.stderr)
-+ print('-----', file=sys.stderr)
-+ with open(filename, 'r') as f:
-+ print(f.read(), file=sys.stderr)
-+ print('-----', file=sys.stderr)
- dump(intermediate_manifest)
- dump(our_manifest)
- dump(assert_manifest)
-@@ -223,7 +227,7 @@
- out, _ = popen.communicate()
- for line in out.splitlines():
- if line and 'manifest authoring warning 81010002' not in line:
-- print line
-+ print(line)
- return popen.returncode
-
- def ExecManifestToRc(self, arch, *args):
-@@ -231,7 +235,7 @@
- |args| is tuple containing path to resource file, path to manifest file
- and resource name which can be "1" (for executables) or "2" (for DLLs)."""
- manifest_path, resource_path, resource_name = args
-- with open(resource_path, 'wb') as output:
-+ with open(resource_path, 'w') as output:
- output.write('#include <windows.h>\n%s RT_MANIFEST "%s"' % (
- resource_name,
- os.path.abspath(manifest_path).replace('\\', '/')))
-@@ -263,7 +267,7 @@
- for x in lines if x.startswith(prefixes))
- for line in lines:
- if not line.startswith(prefixes) and line not in processing:
-- print line
-+ print(line)
- return popen.returncode
-
- def ExecAsmWrapper(self, arch, *args):
-@@ -277,7 +281,7 @@
- not line.startswith('Microsoft (R) Macro Assembler') and
- not line.startswith(' Assembling: ') and
- line):
-- print line
-+ print(line)
- return popen.returncode
-
- def ExecRcWrapper(self, arch, *args):
-@@ -291,7 +295,7 @@
- if (not line.startswith('Microsoft (R) Windows (R) Resource Compiler') and
- not line.startswith('Copyright (C) Microsoft Corporation') and
- line):
-- print line
-+ print(line)
- return popen.returncode
-
- def ExecActionWrapper(self, arch, rspfile, *dir):
-@@ -300,7 +304,7 @@
- env = self._GetEnv(arch)
- # TODO(scottmg): This is a temporary hack to get some specific variables
- # through to actions that are set after gyp-time. http://crbug.com/333738.
-- for k, v in os.environ.iteritems():
-+ for k, v in os.environ.items():
- if k not in env:
- env[k] = v
- args = open(rspfile).read()
-diff --git a/tools/gyp/pylib/gyp/xcode_emulation.py b/pylib/gyp/xcode_emulation.py
-index dba8e76..4c875de 100644
---- a/tools/gyp/pylib/gyp/xcode_emulation.py
-+++ b/tools/gyp/pylib/gyp/xcode_emulation.py
-@@ -7,6 +7,8 @@
- other build systems, such as make and ninja.
- """
-
-+from __future__ import print_function
-+
- import copy
- import gyp.common
- import os
-@@ -73,7 +75,7 @@
- if arch not in expanded_archs:
- expanded_archs.append(arch)
- except KeyError as e:
-- print 'Warning: Ignoring unsupported variable "%s".' % variable
-+ print('Warning: Ignoring unsupported variable "%s".' % variable)
- elif arch not in expanded_archs:
- expanded_archs.append(arch)
- return expanded_archs
-@@ -171,7 +173,7 @@
- # the same for all configs are implicitly per-target settings.
- self.xcode_settings = {}
- configs = spec['configurations']
-- for configname, config in configs.iteritems():
-+ for configname, config in configs.items():
- self.xcode_settings[configname] = config.get('xcode_settings', {})
- self._ConvertConditionalKeys(configname)
- if self.xcode_settings[configname].get('IPHONEOS_DEPLOYMENT_TARGET',
-@@ -197,8 +199,8 @@
- new_key = key.split("[")[0]
- settings[new_key] = settings[key]
- else:
-- print 'Warning: Conditional keys not implemented, ignoring:', \
-- ' '.join(conditional_keys)
-+ print('Warning: Conditional keys not implemented, ignoring:', \
-+ ' '.join(conditional_keys))
- del settings[key]
-
- def _Settings(self):
-@@ -216,7 +218,7 @@
-
- def _WarnUnimplemented(self, test_key):
- if test_key in self._Settings():
-- print 'Warning: Ignoring not yet implemented key "%s".' % test_key
-+ print('Warning: Ignoring not yet implemented key "%s".' % test_key)
-
- def IsBinaryOutputFormat(self, configname):
- default = "binary" if self.isIOS else "xml"
-@@ -963,7 +965,7 @@
- result = dict(self.xcode_settings[configname])
- first_pass = False
- else:
-- for key, value in self.xcode_settings[configname].iteritems():
-+ for key, value in self.xcode_settings[configname].items():
- if key not in result:
- continue
- elif result[key] != value:
-@@ -1084,8 +1086,8 @@
- unimpl = ['OTHER_CODE_SIGN_FLAGS']
- unimpl = set(unimpl) & set(self.xcode_settings[configname].keys())
- if unimpl:
-- print 'Warning: Some codesign keys not implemented, ignoring: %s' % (
-- ', '.join(sorted(unimpl)))
-+ print('Warning: Some codesign keys not implemented, ignoring: %s' % (
-+ ', '.join(sorted(unimpl))))
-
- if self._IsXCTest():
- # For device xctests, Xcode copies two extra frameworks into $TEST_HOST.
-@@ -1737,7 +1739,7 @@
- order = gyp.common.TopologicallySorted(env.keys(), GetEdges)
- order.reverse()
- return order
-- except gyp.common.CycleError, e:
-+ except gyp.common.CycleError as e:
- raise GypError(
- 'Xcode environment variables are cyclically dependent: ' + str(e.nodes))
-
-@@ -1774,10 +1776,11 @@
- def _AddIOSDeviceConfigurations(targets):
- """Clone all targets and append -iphoneos to the name. Configure these targets
- to build for iOS devices and use correct architectures for those builds."""
-- for target_dict in targets.itervalues():
-+ for target_dict in targets.values():
- toolset = target_dict['toolset']
- configs = target_dict['configurations']
-- for config_name, simulator_config_dict in dict(configs).iteritems():
-+
-+ for config_name, simulator_config_dict in dict(configs).items():
- iphoneos_config_dict = copy.deepcopy(simulator_config_dict)
- configs[config_name + '-iphoneos'] = iphoneos_config_dict
- configs[config_name + '-iphonesimulator'] = simulator_config_dict
-diff --git a/tools/gyp/pylib/gyp/xcode_ninja.py b/pylib/gyp/xcode_ninja.py
-index bc76fff..1d71b8c 100644
---- a/tools/gyp/pylib/gyp/xcode_ninja.py
-+++ b/tools/gyp/pylib/gyp/xcode_ninja.py
-@@ -28,7 +28,7 @@
- workspace_path = os.path.join(options.generator_output, workspace_path)
- try:
- os.makedirs(workspace_path)
-- except OSError, e:
-+ except OSError as e:
- if e.errno != errno.EEXIST:
- raise
- output_string = '<?xml version="1.0" encoding="UTF-8"?>\n' + \
-@@ -85,7 +85,7 @@
- "%s/$(CONFIGURATION)$(EFFECTIVE_PLATFORM_NAME)" % ninja_toplevel
-
- if 'configurations' in old_spec:
-- for config in old_spec['configurations'].iterkeys():
-+ for config in old_spec['configurations'].keys():
- old_xcode_settings = \
- old_spec['configurations'][config].get('xcode_settings', {})
- if 'IPHONEOS_DEPLOYMENT_TARGET' in old_xcode_settings:
-@@ -167,7 +167,7 @@
- params: Dict of global options for gyp.
- """
- orig_gyp = params['build_files'][0]
-- for gyp_name, gyp_dict in data.iteritems():
-+ for gyp_name, gyp_dict in data.items():
- if gyp_name == orig_gyp:
- depth = gyp_dict['_DEPTH']
-
-@@ -238,7 +238,7 @@
- not generator_flags.get('xcode_ninja_list_excluded_files', True)
-
- sources = []
-- for target, target_dict in target_dicts.iteritems():
-+ for target, target_dict in target_dicts.items():
- base = os.path.dirname(target)
- files = target_dict.get('sources', []) + \
- target_dict.get('mac_bundle_resources', [])
-diff --git a/tools/gyp/pylib/gyp/xcodeproj_file.py b/pylib/gyp/xcodeproj_file.py
-index e69235f..bd238f6 100644
---- a/tools/gyp/pylib/gyp/xcodeproj_file.py
-+++ b/tools/gyp/pylib/gyp/xcodeproj_file.py
-@@ -154,6 +154,11 @@
- import sha
- _new_sha1 = sha.new
-
-+try:
-+ # basestring was removed in python3.
-+ basestring
-+except NameError:
-+ basestring = str
-
- # See XCObject._EncodeString. This pattern is used to determine when a string
- # can be printed unquoted. Strings that match this pattern may be printed
-@@ -314,7 +319,7 @@
- """
-
- that = self.__class__(id=self.id, parent=self.parent)
-- for key, value in self._properties.iteritems():
-+ for key, value in self._properties.items():
- is_strong = self._schema[key][2]
-
- if isinstance(value, XCObject):
-@@ -324,8 +329,7 @@
- that._properties[key] = new_value
- else:
- that._properties[key] = value
-- elif isinstance(value, str) or isinstance(value, unicode) or \
-- isinstance(value, int):
-+ elif isinstance(value, basestring) or isinstance(value, int):
- that._properties[key] = value
- elif isinstance(value, list):
- if is_strong:
-@@ -449,10 +453,10 @@
- # is 160 bits. Instead of throwing out 64 bits of the digest, xor them
- # into the portion that gets used.
- assert hash.digest_size % 4 == 0
-- digest_int_count = hash.digest_size / 4
-+ digest_int_count = hash.digest_size // 4
- digest_ints = struct.unpack('>' + 'I' * digest_int_count, hash.digest())
- id_ints = [0, 0, 0]
-- for index in xrange(0, digest_int_count):
-+ for index in range(0, digest_int_count):
- id_ints[index % 3] ^= digest_ints[index]
- self.id = '%08X%08X%08X' % tuple(id_ints)
-
-@@ -475,7 +479,7 @@
- """Returns a list of all of this object's owned (strong) children."""
-
- children = []
-- for property, attributes in self._schema.iteritems():
-+ for property, attributes in self._schema.items():
- (is_list, property_type, is_strong) = attributes[0:3]
- if is_strong and property in self._properties:
- if not is_list:
-@@ -603,7 +607,12 @@
- comment = value.Comment()
- elif isinstance(value, str):
- printable += self._EncodeString(value)
-- elif isinstance(value, unicode):
-+ # A python3 compatible way of saying isinstance(value, unicode).
-+ # basestring is str in python3 so this is equivalent to the above
-+ # isinstance. Thus if it failed above it will fail here.
-+ # In python2 we test against str and unicode at this point. str has already
-+ # failed in the above isinstance so we test against unicode.
-+ elif isinstance(value, basestring):
- printable += self._EncodeString(value.encode('utf-8'))
- elif isinstance(value, int):
- printable += str(value)
-@@ -622,7 +631,7 @@
- printable += end_tabs + ')'
- elif isinstance(value, dict):
- printable = '{' + sep
-- for item_key, item_value in sorted(value.iteritems()):
-+ for item_key, item_value in sorted(value.items()):
- printable += element_tabs + \
- self._XCPrintableValue(tabs + 1, item_key, flatten_list) + ' = ' + \
- self._XCPrintableValue(tabs + 1, item_value, flatten_list) + ';' + \
-@@ -691,7 +700,7 @@
- printable_value[0] == '"' and printable_value[-1] == '"':
- printable_value = printable_value[1:-1]
- printable += printable_key + ' = ' + printable_value + ';' + after_kv
-- except TypeError, e:
-+ except TypeError as e:
- gyp.common.ExceptionAppend(e,
- 'while printing key "%s"' % key)
- raise
-@@ -730,7 +739,7 @@
- self._XCKVPrint(file, 3, 'isa', self.__class__.__name__)
-
- # The remaining elements of an object dictionary are sorted alphabetically.
-- for property, value in sorted(self._properties.iteritems()):
-+ for property, value in sorted(self._properties.items()):
- self._XCKVPrint(file, 3, property, value)
-
- # End the object.
-@@ -752,7 +761,7 @@
- if properties is None:
- return
-
-- for property, value in properties.iteritems():
-+ for property, value in properties.items():
- # Make sure the property is in the schema.
- if not property in self._schema:
- raise KeyError(property + ' not in ' + self.__class__.__name__)
-@@ -766,7 +775,7 @@
- ' must be list, not ' + value.__class__.__name__)
- for item in value:
- if not isinstance(item, property_type) and \
-- not (item.__class__ == unicode and property_type == str):
-+ not (isinstance(item, basestring) and property_type == str):
- # Accept unicode where str is specified. str is treated as
- # UTF-8-encoded.
- raise TypeError(
-@@ -774,7 +783,7 @@
- ' must be ' + property_type.__name__ + ', not ' + \
- item.__class__.__name__)
- elif not isinstance(value, property_type) and \
-- not (value.__class__ == unicode and property_type == str):
-+ not (isinstance(value, basestring) and property_type == str):
- # Accept unicode where str is specified. str is treated as
- # UTF-8-encoded.
- raise TypeError(
-@@ -788,8 +797,7 @@
- self._properties[property] = value.Copy()
- else:
- self._properties[property] = value
-- elif isinstance(value, str) or isinstance(value, unicode) or \
-- isinstance(value, int):
-+ elif isinstance(value, basestring) or isinstance(value, int):
- self._properties[property] = value
- elif isinstance(value, list):
- if is_strong:
-@@ -865,7 +873,7 @@
-
- # TODO(mark): A stronger verification mechanism is needed. Some
- # subclasses need to perform validation beyond what the schema can enforce.
-- for property, attributes in self._schema.iteritems():
-+ for property, attributes in self._schema.items():
- (is_list, property_type, is_strong, is_required) = attributes[0:4]
- if is_required and not property in self._properties:
- raise KeyError(self.__class__.__name__ + ' requires ' + property)
-@@ -875,7 +883,7 @@
- overwrite properties that have already been set."""
-
- defaults = {}
-- for property, attributes in self._schema.iteritems():
-+ for property, attributes in self._schema.items():
- (is_list, property_type, is_strong, is_required) = attributes[0:4]
- if is_required and len(attributes) >= 5 and \
- not property in self._properties:
-@@ -1426,8 +1434,8 @@
- xche = self
- while xche != None and isinstance(xche, XCHierarchicalElement):
- xche_hashables = xche.Hashables()
-- for index in xrange(0, len(xche_hashables)):
-- hashables.insert(index, xche_hashables[index])
-+ for index, xche_hashable in enumerate(xche_hashables):
-+ hashables.insert(index, xche_hashable)
- xche = xche.parent
- return hashables
-
-@@ -2468,8 +2476,7 @@
- # The headers phase should come before the resources, sources, and
- # frameworks phases, if any.
- insert_at = len(self._properties['buildPhases'])
-- for index in xrange(0, len(self._properties['buildPhases'])):
-- phase = self._properties['buildPhases'][index]
-+ for index, phase in enumerate(self._properties['buildPhases']):
- if isinstance(phase, PBXResourcesBuildPhase) or \
- isinstance(phase, PBXSourcesBuildPhase) or \
- isinstance(phase, PBXFrameworksBuildPhase):
-@@ -2489,8 +2496,7 @@
- # The resources phase should come before the sources and frameworks
- # phases, if any.
- insert_at = len(self._properties['buildPhases'])
-- for index in xrange(0, len(self._properties['buildPhases'])):
-- phase = self._properties['buildPhases'][index]
-+ for index, phase in enumerate(self._properties['buildPhases']):
- if isinstance(phase, PBXSourcesBuildPhase) or \
- isinstance(phase, PBXFrameworksBuildPhase):
- insert_at = index
-@@ -2911,7 +2917,7 @@
- # determine the sort order.
- return cmp(x_index, y_index)
-
-- for other_pbxproject, ref_dict in self._other_pbxprojects.iteritems():
-+ for other_pbxproject, ref_dict in self._other_pbxprojects.items():
- # Build up a list of products in the remote project file, ordered the
- # same as the targets that produce them.
- remote_products = []
-diff --git a/tools/gyp/pylib/gyp/xml_fix.py b/pylib/gyp/xml_fix.py
-index 5de8481..4308d99 100644
---- a/tools/gyp/pylib/gyp/xml_fix.py
-+++ b/tools/gyp/pylib/gyp/xml_fix.py
-@@ -32,8 +32,7 @@
- writer.write(indent+"<" + self.tagName)
-
- attrs = self._get_attributes()
-- a_names = attrs.keys()
-- a_names.sort()
-+ a_names = sorted(attrs.keys())
-
- for a_name in a_names:
- writer.write(" %s=\"" % a_name)
-diff --git a/tools/gyp/tools/graphviz.py b/tools/graphviz.py
-index 326ae22..538b059 100755
---- a/tools/gyp/tools/graphviz.py
-+++ b/tools/gyp/tools/graphviz.py
-@@ -8,6 +8,8 @@
- generate input suitable for graphviz to render a dependency graph of
- targets."""
-
-+from __future__ import print_function
-+
- import collections
- import json
- import sys
-@@ -50,9 +52,9 @@
- build_file, target_name, toolset = ParseTarget(src)
- files[build_file].append(src)
-
-- print 'digraph D {'
-- print ' fontsize=8' # Used by subgraphs.
-- print ' node [fontsize=8]'
-+ print('digraph D {')
-+ print(' fontsize=8') # Used by subgraphs.
-+ print(' node [fontsize=8]')
-
- # Output nodes by file. We must first write out each node within
- # its file grouping before writing out any edges that may refer
-@@ -63,31 +65,31 @@
- # the display by making it a box without an internal node.
- target = targets[0]
- build_file, target_name, toolset = ParseTarget(target)
-- print ' "%s" [shape=box, label="%s\\n%s"]' % (target, filename,
-- target_name)
-+ print(' "%s" [shape=box, label="%s\\n%s"]' % (target, filename,
-+ target_name))
- else:
- # Group multiple nodes together in a subgraph.
-- print ' subgraph "cluster_%s" {' % filename
-- print ' label = "%s"' % filename
-+ print(' subgraph "cluster_%s" {' % filename)
-+ print(' label = "%s"' % filename)
- for target in targets:
- build_file, target_name, toolset = ParseTarget(target)
-- print ' "%s" [label="%s"]' % (target, target_name)
-- print ' }'
-+ print(' "%s" [label="%s"]' % (target, target_name))
-+ print(' }')
-
- # Now that we've placed all the nodes within subgraphs, output all
- # the edges between nodes.
- for src, dsts in edges.items():
- for dst in dsts:
-- print ' "%s" -> "%s"' % (src, dst)
-+ print(' "%s" -> "%s"' % (src, dst))
-
-- print '}'
-+ print('}')
-
-
- def main():
- if len(sys.argv) < 2:
-- print >>sys.stderr, __doc__
-- print >>sys.stderr
-- print >>sys.stderr, 'usage: %s target1 target2...' % (sys.argv[0])
-+ print(__doc__, file=sys.stderr)
-+ print(file=sys.stderr)
-+ print('usage: %s target1 target2...' % (sys.argv[0]), file=sys.stderr)
- return 1
-
- edges = LoadEdges('dump.json', sys.argv[1:])
-diff --git a/tools/gyp/tools/pretty_gyp.py b/tools/pretty_gyp.py
-index d5736bb..5060d1d 100755
---- a/tools/gyp/tools/pretty_gyp.py
-+++ b/tools/gyp/tools/pretty_gyp.py
-@@ -6,6 +6,8 @@
-
- """Pretty-prints the contents of a GYP file."""
-
-+from __future__ import print_function
-+
- import sys
- import re
-
-@@ -125,15 +127,15 @@
- (brace_diff, after) = count_braces(line)
- if brace_diff != 0:
- if after:
-- print " " * (basic_offset * indent) + line
-+ print(" " * (basic_offset * indent) + line)
- indent += brace_diff
- else:
- indent += brace_diff
-- print " " * (basic_offset * indent) + line
-+ print(" " * (basic_offset * indent) + line)
- else:
-- print " " * (basic_offset * indent) + line
-+ print(" " * (basic_offset * indent) + line)
- else:
-- print ""
-+ print("")
- last_line = line
-
-
-diff --git a/tools/gyp/tools/pretty_sln.py b/tools/pretty_sln.py
-index ca8cf4a..12a6dad 100755
---- a/tools/gyp/tools/pretty_sln.py
-+++ b/tools/gyp/tools/pretty_sln.py
-@@ -12,6 +12,8 @@
- Then it outputs a possible build order.
- """
-
-+from __future__ import print_function
-+
- __author__ = 'nsylvain (Nicolas Sylvain)'
-
- import os
-@@ -26,7 +28,7 @@
- for dep in deps[project]:
- if dep not in built:
- BuildProject(dep, built, projects, deps)
-- print project
-+ print(project)
- built.append(project)
-
- def ParseSolution(solution_file):
-@@ -100,44 +102,44 @@
- return (projects, dependencies)
-
- def PrintDependencies(projects, deps):
-- print "---------------------------------------"
-- print "Dependencies for all projects"
-- print "---------------------------------------"
-- print "-- --"
-+ print("---------------------------------------")
-+ print("Dependencies for all projects")
-+ print("---------------------------------------")
-+ print("-- --")
-
- for (project, dep_list) in sorted(deps.items()):
-- print "Project : %s" % project
-- print "Path : %s" % projects[project][0]
-+ print("Project : %s" % project)
-+ print("Path : %s" % projects[project][0])
- if dep_list:
- for dep in dep_list:
-- print " - %s" % dep
-- print ""
-+ print(" - %s" % dep)
-+ print("")
-
-- print "-- --"
-+ print("-- --")
-
- def PrintBuildOrder(projects, deps):
-- print "---------------------------------------"
-- print "Build order "
-- print "---------------------------------------"
-- print "-- --"
-+ print("---------------------------------------")
-+ print("Build order ")
-+ print("---------------------------------------")
-+ print("-- --")
-
- built = []
- for (project, _) in sorted(deps.items()):
- if project not in built:
- BuildProject(project, built, projects, deps)
-
-- print "-- --"
-+ print("-- --")
-
- def PrintVCProj(projects):
-
- for project in projects:
-- print "-------------------------------------"
-- print "-------------------------------------"
-- print project
-- print project
-- print project
-- print "-------------------------------------"
-- print "-------------------------------------"
-+ print("-------------------------------------")
-+ print("-------------------------------------")
-+ print(project)
-+ print(project)
-+ print(project)
-+ print("-------------------------------------")
-+ print("-------------------------------------")
-
- project_path = os.path.abspath(os.path.join(os.path.dirname(sys.argv[1]),
- projects[project][2]))
-@@ -153,7 +155,7 @@
- def main():
- # check if we have exactly 1 parameter.
- if len(sys.argv) < 2:
-- print 'Usage: %s "c:\\path\\to\\project.sln"' % sys.argv[0]
-+ print('Usage: %s "c:\\path\\to\\project.sln"' % sys.argv[0])
- return 1
-
- (projects, deps) = ParseSolution(sys.argv[1])
-diff --git a/tools/gyp/tools/pretty_vcproj.py b/tools/pretty_vcproj.py
-index 6099bd7..f02e59e 100755
---- a/tools/gyp/tools/pretty_vcproj.py
-+++ b/tools/gyp/tools/pretty_vcproj.py
-@@ -12,6 +12,8 @@
- It outputs the resulting xml to stdout.
- """
-
-+from __future__ import print_function
-+
- __author__ = 'nsylvain (Nicolas Sylvain)'
-
- import os
-@@ -73,23 +75,23 @@
-
- # Print the main tag
- if attr_count == 0:
-- print '%s<%s>' % (' '*indent, node.nodeName)
-+ print('%s<%s>' % (' '*indent, node.nodeName))
- else:
-- print '%s<%s' % (' '*indent, node.nodeName)
-+ print('%s<%s' % (' '*indent, node.nodeName))
-
- all_attributes = []
- for (name, value) in node.attributes.items():
- all_attributes.append((name, value))
-- all_attributes.sort(CmpTuple())
-+ all_attributes.sort(key=(lambda attr: attr[0]))
- for (name, value) in all_attributes:
-- print '%s %s="%s"' % (' '*indent, name, value)
-- print '%s>' % (' '*indent)
-+ print('%s %s="%s"' % (' '*indent, name, value))
-+ print('%s>' % (' '*indent))
- if node.nodeValue:
-- print '%s %s' % (' '*indent, node.nodeValue)
-+ print('%s %s' % (' '*indent, node.nodeValue))
-
- for sub_node in node.childNodes:
- PrettyPrintNode(sub_node, indent=indent+2)
-- print '%s</%s>' % (' '*indent, node.nodeName)
-+ print('%s</%s>' % (' '*indent, node.nodeName))
-
-
- def FlattenFilter(node):
-@@ -283,8 +285,8 @@
-
- # check if we have exactly 1 parameter.
- if len(argv) < 2:
-- print ('Usage: %s "c:\\path\\to\\vcproj.vcproj" [key1=value1] '
-- '[key2=value2]' % argv[0])
-+ print('Usage: %s "c:\\path\\to\\vcproj.vcproj" [key1=value1] '
-+ '[key2=value2]' % argv[0])
- return 1
-
- # Parse the keys
-diff --git a/tools/gyp/gyptest.py b/gyptest.py
-index 9930e78..1a9ffca 100755
---- a/tools/gyp/gyptest.py
-+++ b/tools/gyp/gyptest.py
-@@ -58,7 +58,7 @@
- os.chdir(args.chdir)
-
- if args.path:
-- extra_path = [os.path.abspath(p) for p in opts.path]
-+ extra_path = [os.path.abspath(p) for p in args.path]
- extra_path = os.pathsep.join(extra_path)
- os.environ['PATH'] = extra_path + os.pathsep + os.environ['PATH']
-
-diff --git a/tools/gyp/pylib/gyp/MSVSNew.py b/pylib/gyp/MSVSNew.py
-index 593f0e5..0445931 100644
---- a/tools/gyp/pylib/gyp/MSVSNew.py
-+++ b/tools/gyp/pylib/gyp/MSVSNew.py
-@@ -21,6 +21,13 @@
- _new_md5 = md5.new
-
-
-+try:
-+ # cmp was removed in python3.
-+ cmp
-+except NameError:
-+ def cmp(a, b):
-+ return (a > b) - (a < b)
-+
- # Initialize random number generator
- random.seed()
-
-diff --git a/tools/gyp/pylib/gyp/common.py b/pylib/gyp/common.py
-index 1823de8..b268d22 100644
---- a/tools/gyp/pylib/gyp/common.py
-+++ b/tools/gyp/pylib/gyp/common.py
-@@ -584,7 +584,7 @@
- graph = {'a': '$(b) $(c)', 'b': 'hi', 'c': '$(b)'}
- def GetEdges(node):
- return re.findall(r'\$\(([^))]\)', graph[node])
-- print TopologicallySorted(graph.keys(), GetEdges)
-+ print(TopologicallySorted(graph.keys(), GetEdges))
- ==>
- ['a', 'c', b']
- """
-diff --git a/tools/gyp/pylib/gyp/generator/make.py b/pylib/gyp/generator/make.py
-index 2057e3a..8c2827e 100644
---- a/tools/gyp/pylib/gyp/generator/make.py
-+++ b/tools/gyp/pylib/gyp/generator/make.py
-@@ -1636,7 +1636,7 @@
- self.WriteDoCmd([self.output_binary], deps, 'touch', part_of_all,
- postbuilds=postbuilds)
- else:
-- print("WARNING: no output for", self.type, target)
-+ print("WARNING: no output for", self.type, self.target)
-
- # Add an alias for each target (if there are any outputs).
- # Installable target aliases are created below.
-diff --git a/tools/gyp/pylib/gyp/generator/msvs.py b/pylib/gyp/generator/msvs.py
-index e8a2b36..9eac028 100644
---- a/tools/gyp/pylib/gyp/generator/msvs.py
-+++ b/tools/gyp/pylib/gyp/generator/msvs.py
-@@ -308,10 +308,8 @@
- if names:
- return names[0]
- else:
-- print >> sys.stdout, (
-- 'Warning: No include files found for '
-- 'detected Windows SDK version %s' % (version)
-- )
-+ print('Warning: No include files found for '
-+ 'detected Windows SDK version %s' % (version))
-
-
- def _BuildCommandLineForRuleRaw(spec, cmd, cygwin_shell, has_input_path,
-@@ -2065,7 +2063,7 @@
- if generator_flags.get('msvs_error_on_missing_sources', False):
- raise GypError(error_message)
- else:
-- print("Warning: " + error_message, file=sys.stdout)
-+ print("Warning: " + error_message)
-
-
- def _GenerateMSBuildFiltersFile(filters_path, source_files,
-diff --git a/tools/gyp/pylib/gyp/mac_tool.py b/pylib/gyp/mac_tool.py
-index 7d3a8c2..84f8863 100755
---- a/tools/gyp/pylib/gyp/mac_tool.py
-+++ b/tools/gyp/pylib/gyp/mac_tool.py
-@@ -670,7 +670,7 @@
- count = len(filelist)
- capacity = NextGreaterPowerOf2(count)
- strings_offset = 24 + (12 * capacity)
-- max_value_length = len(max(filelist.items(), key=lambda (k,v):len(v))[1])
-+ max_value_length = len(max(filelist.items(), key=lambda t: len(t[1]))[1])
-
- out = open(output_name, "wb")
- out.write(struct.pack('<LHHLLLL', magic, version, _reserved, strings_offset,
-diff --git a/tools/gyp/pylib/gyp/xcodeproj_file.py b/pylib/gyp/xcodeproj_file.py
-index bd238f6..bc9814d 100644
---- a/tools/gyp/pylib/gyp/xcodeproj_file.py
-+++ b/tools/gyp/pylib/gyp/xcodeproj_file.py
-@@ -160,6 +160,13 @@
- except NameError:
- basestring = str
-
-+try:
-+ # cmp was removed in python3.
-+ cmp
-+except NameError:
-+ def cmp(a, b):
-+ return (a > b) - (a < b)
-+
- # See XCObject._EncodeString. This pattern is used to determine when a string
- # can be printed unquoted. Strings that match this pattern may be printed
- # unquoted. Strings that do not match must be quoted and may be further
-diff --git a/tools/gyp/samples/samples b/samples/samples
-index 804b618..ff26de3 100755
---- a/tools/gyp/samples/samples
-+++ b/tools/gyp/samples/samples
-@@ -4,6 +4,8 @@
- # Use of this source code is governed by a BSD-style license that can be
- # found in the LICENSE file.
-
-+from __future__ import print_function
-+
- import os.path
- import shutil
- import sys
-@@ -57,7 +59,7 @@
-
- def Main(argv):
- if len(argv) != 3 or argv[1] not in ['push', 'pull']:
-- print 'Usage: %s push/pull PATH_TO_CHROME' % argv[0]
-+ print('Usage: %s push/pull PATH_TO_CHROME' % argv[0])
- return 1
-
- path_to_chrome = argv[2]
-@@ -66,10 +68,10 @@
- chrome_file = os.path.join(path_to_chrome, g)
- local_file = os.path.join(os.path.dirname(argv[0]), os.path.split(g)[1])
- if argv[1] == 'push':
-- print 'Copying %s to %s' % (local_file, chrome_file)
-+ print('Copying %s to %s' % (local_file, chrome_file))
- shutil.copyfile(local_file, chrome_file)
- elif argv[1] == 'pull':
-- print 'Copying %s to %s' % (chrome_file, local_file)
-+ print('Copying %s to %s' % (chrome_file, local_file))
- shutil.copyfile(chrome_file, local_file)
- else:
- assert False
-diff --git a/tools/gyp/tools/pretty_vcproj.py b/tools/pretty_vcproj.py
-index f02e59e..4454d9b 100755
---- a/tools/gyp/tools/pretty_vcproj.py
-+++ b/tools/gyp/tools/pretty_vcproj.py
-@@ -22,6 +22,13 @@
- from xml.dom.minidom import parse
- from xml.dom.minidom import Node
-
-+try:
-+ # cmp was removed in python3.
-+ cmp
-+except NameError:
-+ def cmp(a, b):
-+ return (a > b) - (a < b)
-+
- REPLACEMENTS = dict()
- ARGUMENTS = None
-
-@@ -63,7 +70,7 @@
- def PrettyPrintNode(node, indent=0):
- if node.nodeType == Node.TEXT_NODE:
- if node.data.strip():
-- print '%s%s' % (' '*indent, node.data.strip())
-+ print('%s%s' % (' '*indent, node.data.strip()))
- return
-
- if node.childNodes:
-@@ -322,7 +329,6 @@
-
- # Finally, we use the prett xml function to print the vcproj back to the
- # user.
-- #print dom.toprettyxml(newl="\n")
- PrettyPrintNode(dom.documentElement)
- return 0
-
---- node-v10.15.3/tools/gyp/pylib/gyp/input.py.old 2019-04-02 06:44:13.086310973 +0000
-+++ node-v10.15.3/tools/gyp/pylib/gyp/input.py 2019-04-02 06:45:35.987250735 +0000
-@@ -900,6 +900,9 @@
- (e, contents, build_file))
-
- p_stdout, p_stderr = p.communicate('')
-+ if getattr(p_stdout, 'decode'):
-+ p_stdout = p_stdout.decode('utf-8')
-+ p_stderr = p_stderr.decode('utf-8')
-
- if p.wait() != 0 or p_stderr:
- sys.stderr.write(p_stderr)
diff --git a/user/node/libatomic.patch b/user/node/libatomic.patch
deleted file mode 100644
index fc2e4ffd2..000000000
--- a/user/node/libatomic.patch
+++ /dev/null
@@ -1,14 +0,0 @@
---- node-v10.15.3/node.gyp.old 2019-03-05 15:16:32.000000000 +0000
-+++ node-v10.15.3/node.gyp 2019-04-06 13:54:51.204939193 +0000
-@@ -479,6 +479,11 @@
- 'msvs_disabled_warnings!': [4244],
-
- 'conditions': [
-+ [ 'host_arch=="mips" or host_arch=="mipsel" or host_arch=="ppc"', {
-+ 'link_settings': {
-+ 'libraries': [ '-latomic' ],
-+ },
-+ }],
- [ 'node_code_cache_path!=""', {
- 'sources': [ '<(node_code_cache_path)' ]
- }, {
diff --git a/user/node/ppc32.patch b/user/node/ppc32.patch
deleted file mode 100644
index 45b051534..000000000
--- a/user/node/ppc32.patch
+++ /dev/null
@@ -1,29 +0,0 @@
---- node-v10.15.3/deps/v8/src/libsampler/sampler.cc.old 2019-03-05 15:16:28.000000000 +0000
-+++ node-v10.15.3/deps/v8/src/libsampler/sampler.cc 2019-04-06 13:44:07.224653262 +0000
-@@ -502,9 +502,15 @@
- reinterpret_cast<void*>(ucontext->uc_mcontext.regs->gpr[PT_R31]);
- #else
- // Some C libraries, notably Musl, define the regs member as a void pointer
-+ #if !V8_TARGET_ARCH_32_BIT
- state->pc = reinterpret_cast<void*>(ucontext->uc_mcontext.gp_regs[32]);
- state->sp = reinterpret_cast<void*>(ucontext->uc_mcontext.gp_regs[1]);
- state->fp = reinterpret_cast<void*>(ucontext->uc_mcontext.gp_regs[31]);
-+ #else
-+ state->pc = reinterpret_cast<void*>(ucontext->uc_mcontext.gregs[32]);
-+ state->sp = reinterpret_cast<void*>(ucontext->uc_mcontext.gregs[1]);
-+ state->fp = reinterpret_cast<void*>(ucontext->uc_mcontext.gregs[31]);
-+ #endif
- #endif
- #elif V8_HOST_ARCH_S390
- #if V8_TARGET_ARCH_32_BIT
---- node-v10.16.3/configure.py.old 2019-08-15 19:20:03.000000000 +0000
-+++ node-v10.16.3/configure.py 2019-08-16 14:11:23.086489117 +0000
-@@ -848,7 +848,7 @@
- '__MIPSEL__' : 'mipsel',
- '__mips__' : 'mips',
- '__PPC64__' : 'ppc64',
-- '__PPC__' : 'ppc64',
-+ '__PPC__' : 'ppc',
- '__x86_64__' : 'x64',
- '__s390__' : 's390',
- '__s390x__' : 's390x',
diff --git a/user/node/ppc64.patch b/user/node/ppc64.patch
deleted file mode 100644
index f76618562..000000000
--- a/user/node/ppc64.patch
+++ /dev/null
@@ -1,40 +0,0 @@
---- a/deps/v8/src/ppc/assembler-ppc.h.old 2019-03-05 15:16:29.000000000 +0000
-+++ b/deps/v8/src/ppc/assembler-ppc.h 2019-04-02 07:05:25.977213735 +0000
-@@ -48,7 +48,8 @@
- #include "src/ppc/constants-ppc.h"
-
- #if V8_HOST_ARCH_PPC && \
-- (V8_OS_AIX || (V8_TARGET_ARCH_PPC64 && V8_TARGET_BIG_ENDIAN))
-+ (V8_OS_AIX || (V8_TARGET_ARCH_PPC64 && \
-+ (V8_TARGET_BIG_ENDIAN && (!defined(_CALL_ELF) || _CALL_ELF == 1))))
- #define ABI_USES_FUNCTION_DESCRIPTORS 1
- #else
- #define ABI_USES_FUNCTION_DESCRIPTORS 0
-@@ -60,13 +61,15 @@
- #define ABI_PASSES_HANDLES_IN_REGS 0
- #endif
-
--#if !V8_HOST_ARCH_PPC || !V8_TARGET_ARCH_PPC64 || V8_TARGET_LITTLE_ENDIAN
-+#if !V8_HOST_ARCH_PPC || !V8_TARGET_ARCH_PPC64 || \
-+ (V8_TARGET_LITTLE_ENDIAN || (defined(_CALL_ELF) && _CALL_ELF == 2))
- #define ABI_RETURNS_OBJECT_PAIRS_IN_REGS 1
- #else
- #define ABI_RETURNS_OBJECT_PAIRS_IN_REGS 0
- #endif
-
--#if !V8_HOST_ARCH_PPC || (V8_TARGET_ARCH_PPC64 && V8_TARGET_LITTLE_ENDIAN)
-+#if !V8_HOST_ARCH_PPC || (V8_TARGET_ARCH_PPC64 && \
-+ (V8_TARGET_LITTLE_ENDIAN || (defined(_CALL_ELF) && _CALL_ELF == 2)))
- #define ABI_CALL_VIA_IP 1
- #else
- #define ABI_CALL_VIA_IP 0
-@@ -220,7 +220,8 @@
- // The following constants describe the stack frame linkage area as
- // defined by the ABI. Note that kNumRequiredStackFrameSlots must
- // satisfy alignment requirements (rounding up if required).
--#if V8_TARGET_ARCH_PPC64 && V8_TARGET_LITTLE_ENDIAN
-+#if V8_TARGET_ARCH_PPC64 && (V8_TARGET_LITTLE_ENDIAN || \
-+ defined(_CALL_ELF) && _CALL_ELF == 2)
- // [0] back chain
- // [1] condition register save area
- // [2] link register save area
diff --git a/user/node/python3.patch b/user/node/python3.patch
deleted file mode 100644
index dca32c7e4..000000000
--- a/user/node/python3.patch
+++ /dev/null
@@ -1,163 +0,0 @@
---- node-v10.15.3/tools/configure.d/nodedownload.py (original)
-+++ node-v10.15.3/tools/configure.d/nodedownload.py (refactored)
-@@ -1,7 +1,10 @@
- #!/usr/bin/env python
- # Moved some utilities here from ../../configure
-
-+try:
--import urllib
-+ import urllib
-+except ImportError:
-+ import urllib.request, urllib.parse, urllib.error
- import hashlib
- import sys
- import zipfile
-@@ -18,10 +18,10 @@
- spin = ".:|'"
- return (spin[c % len(spin)])
-
--class ConfigOpener(urllib.FancyURLopener):
-+class ConfigOpener(urllib.request.FancyURLopener):
- """fancy opener used by retrievefile. Set a UA"""
- # append to existing version (UA)
-- version = '%s node.js/configure' % urllib.URLopener.version
-+ version = '%s node.js/configure' % urllib.request.URLopener.version
-
- def reporthook(count, size, total):
- """internal hook used by retrievefile"""
-@@ -36,10 +36,10 @@
- sys.stdout.write(' <%s>\nConnecting...\r' % url)
- sys.stdout.flush()
- ConfigOpener().retrieve(url, targetfile, reporthook=reporthook)
-- print '' # clear the line
-+ print('') # clear the line
- return targetfile
- except:
-- print ' ** Error occurred while downloading\n <%s>' % url
-+ print(' ** Error occurred while downloading\n <%s>' % url)
- raise
-
- def md5sum(targetfile):
-@@ -56,12 +56,12 @@
- """Unpacks packedfile into parent_path. Assumes .zip. Returns parent_path"""
- if zipfile.is_zipfile(packedfile):
- with contextlib.closing(zipfile.ZipFile(packedfile, 'r')) as icuzip:
-- print ' Extracting zipfile: %s' % packedfile
-+ print(' Extracting zipfile: %s' % packedfile)
- icuzip.extractall(parent_path)
- return parent_path
- elif tarfile.is_tarfile(packedfile):
- with contextlib.closing(tarfile.TarFile.open(packedfile, 'r')) as icuzip:
-- print ' Extracting tarfile: %s' % packedfile
-+ print(' Extracting tarfile: %s' % packedfile)
- icuzip.extractall(parent_path)
- return parent_path
- else:
-@@ -112,16 +112,16 @@
- theRet[anOpt] = True
- else:
- # future proof: ignore unknown types
-- print 'Warning: ignoring unknown --download= type "%s"' % anOpt
-+ print('Warning: ignoring unknown --download= type "%s"' % anOpt)
- # all done
- return theRet
-
- def candownload(auto_downloads, package):
-- if not (package in auto_downloads.keys()):
-+ if not (package in list(auto_downloads.keys())):
- raise Exception('Internal error: "%s" is not in the --downloads list. Check nodedownload.py' % package)
- if auto_downloads[package]:
- return True
- else:
-- print """Warning: Not downloading package "%s". You could pass "--download=all"
-- (Windows: "download-all") to try auto-downloading it.""" % package
-+ print("""Warning: Not downloading package "%s". You could pass "--download=all"
-+ (Windows: "download-all") to try auto-downloading it.""" % package)
- return False
---- node-v10.15.3/tools/configure.d/nodedownload.py.old 2019-04-02 00:56:07.533200475 +0000
-+++ node-v10.15.3/tools/configure.d/nodedownload.py 2019-04-02 00:58:09.019947842 +0000
-@@ -6,12 +6,11 @@
- import sys
- import zipfile
- import tarfile
--import fpformat
- import contextlib
-
- def formatSize(amt):
- """Format a size as a string in MB"""
-- return fpformat.fix(amt / 1024000., 1)
-+ return "%{size}.1f" % (amt / 1024000.)
-
- def spin(c):
- """print out an ASCII 'spinner' based on the value of counter 'c'"""
---- node-v10.15.3/configure.py.old 2019-03-05 15:16:24.000000000 +0000
-+++ node-v10.15.3/configure.py 2019-04-02 01:09:04.948394534 +0000
-@@ -649,8 +649,8 @@
- except OSError:
- return (False, False, '', '')
-
-- proc.stdin.write('__clang__ __GNUC__ __GNUC_MINOR__ __GNUC_PATCHLEVEL__ '
-- '__clang_major__ __clang_minor__ __clang_patchlevel__')
-+ proc.stdin.write(b'__clang__ __GNUC__ __GNUC_MINOR__ __GNUC_PATCHLEVEL__ '
-+ b'__clang_major__ __clang_minor__ __clang_patchlevel__')
-
- values = (proc.communicate()[0].split() + ['0'] * 7)[0:7]
- is_clang = values[0] == '1'
-@@ -727,7 +727,7 @@
- consider adjusting the CC environment variable if you installed
- it in a non-standard prefix.''')
-
-- gas_ret = proc.communicate()[1]
-+ gas_ret = proc.communicate()[1].decode('utf-8')
- match = re.match(r"GNU assembler version ([2-9]\.[0-9]+)", gas_ret)
-
- if match:
-@@ -794,7 +794,7 @@
- consider adjusting the CC environment variable if you installed
- it in a non-standard prefix.''')
-
-- p.stdin.write('\n')
-+ p.stdin.write(b'\n')
- out = p.communicate()[0]
-
- out = str(out).split('\n')
-@@ -1351,7 +1351,7 @@
- o['variables']['icu_small'] = b(True)
- locs = set(options.with_icu_locales.split(','))
- locs.add('root') # must have root
-- o['variables']['icu_locales'] = string.join(locs,',')
-+ o['variables']['icu_locales'] = ','.join(locs)
- # We will check a bit later if we can use the canned deps/icu-small
- elif with_intl == 'full-icu':
- # full ICU
-@@ -1655,7 +1655,7 @@
- if options.prefix:
- config['PREFIX'] = options.prefix
-
--config = '\n'.join(map('='.join, config.iteritems())) + '\n'
-+config = '\n'.join(map('='.join, config.items())) + '\n'
-
- # On Windows there's no reason to search for a different python binary.
- bin_override = None if sys.platform == 'win32' else make_bin_override()
---- node-v10.15.3/configure.py.old 2019-04-02 01:12:29.786049396 +0000
-+++ node-v10.15.3/configure.py 2019-04-02 01:21:08.499637208 +0000
-@@ -634,7 +634,7 @@
- proc = subprocess.Popen(
- shlex.split(pkg_config) + ['--silence-errors', flag, pkg],
- stdout=subprocess.PIPE)
-- val = proc.communicate()[0].strip()
-+ val = proc.communicate()[0].strip().decode('utf-8')
- except OSError as e:
- if e.errno != errno.ENOENT: raise e # Unexpected error.
- return (None, None, None, None) # No pkg-config/pkgconf installed.
---- node-v10.15.3/configure.py.old 2019-04-02 01:27:36.437454388 +0000
-+++ node-v10.15.3/configure.py 2019-04-02 01:28:06.954136125 +0000
-@@ -795,7 +795,7 @@
- it in a non-standard prefix.''')
-
- p.stdin.write(b'\n')
-- out = p.communicate()[0]
-+ out = p.communicate()[0].decode('utf-8')
-
- out = str(out).split('\n')
-
diff --git a/user/node/stack-silliness.patch b/user/node/stack-silliness.patch
deleted file mode 100644
index e8d775b73..000000000
--- a/user/node/stack-silliness.patch
+++ /dev/null
@@ -1,14 +0,0 @@
---- node-v10.15.3/src/inspector_agent.cc.old 2019-03-05 15:16:32.000000000 +0000
-+++ node-v10.15.3/src/inspector_agent.cc 2019-04-06 02:03:11.082016553 +0000
-@@ -104,9 +104,9 @@
- // Don't shrink the thread's stack on FreeBSD. Said platform decided to
- // follow the pthreads specification to the letter rather than in spirit:
- // https://lists.freebsd.org/pipermail/freebsd-current/2014-March/048885.html
--#ifndef __FreeBSD__
-+#ifdef __GLIBC__
- CHECK_EQ(0, pthread_attr_setstacksize(&attr, PTHREAD_STACK_MIN));
--#endif // __FreeBSD__
-+#endif // __GLIBC__
- CHECK_EQ(0, pthread_attr_setdetachstate(&attr, PTHREAD_CREATE_DETACHED));
- sigset_t sigmask;
- // Mask all signals.
diff --git a/user/node/v8-python3.patch b/user/node/v8-python3.patch
deleted file mode 100644
index 64151be4f..000000000
--- a/user/node/v8-python3.patch
+++ /dev/null
@@ -1,169 +0,0 @@
---- node-v10.15.3/deps/v8/tools/node/fetch_deps.py.old 2019-03-05 15:16:31.000000000 +0000
-+++ node-v10.15.3/deps/v8/tools/node/fetch_deps.py 2019-04-02 01:01:00.421250178 +0000
-@@ -9,6 +9,9 @@
- Usage: fetch_deps.py <v8-path>
- """
-
-+# for py2/py3 compatibility
-+from __future__ import print_function
-+
- import os
- import subprocess
- import sys
-@@ -51,9 +55,9 @@
- expected_git_dir = os.path.join(v8_path, ".git")
- actual_git_dir = git("rev-parse --absolute-git-dir")
- if expected_git_dir == actual_git_dir:
-- print "V8 is tracked stand-alone by git."
-+ print("V8 is tracked stand-alone by git.")
- return False
-- print "Initializing temporary git repository in v8."
-+ print("Initializing temporary git repository in v8.")
- git("init")
- git("config user.name \"Ada Lovelace\"")
- git("config user.email ada@lovela.ce")
-@@ -70,7 +74,7 @@
-
- temporary_git = EnsureGit(v8_path)
- try:
-- print "Fetching dependencies."
-+ print("Fetching dependencies.")
- env = os.environ.copy()
- # gclient needs to have depot_tools in the PATH.
- env["PATH"] = depot_tools + os.pathsep + env["PATH"]
---- node-v10.15.3/deps/v8/tools/node/node_common.py.old 2019-03-05 15:16:31.000000000 +0000
-+++ node-v10.15.3/deps/v8/tools/node/node_common.py 2019-04-02 01:00:45.522875398 +0000
-@@ -3,11 +3,15 @@
- # Use of this source code is governed by a BSD-style license that can be
- # found in the LICENSE file.
-
-+# for py2/py3 compatibility
-+from __future__ import print_function
-+
- import os
- import pipes
- import shutil
- import stat
- import subprocess
-+import sys
-
- DEPOT_TOOLS_URL = \
- "https://chromium.googlesource.com/chromium/tools/depot_tools.git"
-@@ -22,23 +26,27 @@
- except:
- pass
- if fetch_if_not_exist:
-- print "Checking out depot_tools."
-+ print("Checking out depot_tools.")
- # shell=True needed on Windows to resolve git.bat.
- subprocess.check_call("git clone {} {}".format(
- pipes.quote(DEPOT_TOOLS_URL),
- pipes.quote(depot_tools)), shell=True)
-+ # Using check_output to hide warning messages.
-+ subprocess.check_output(
-+ [sys.executable, gclient_path, "metrics", "--opt-out"],
-+ cwd=depot_tools)
- return depot_tools
- return None
- depot_tools = _Get(v8_path)
- assert depot_tools is not None
-- print "Using depot tools in %s" % depot_tools
-+ print("Using depot tools in %s" % depot_tools)
- return depot_tools
-
- def UninitGit(v8_path):
-- print "Uninitializing temporary git repository"
-+ print("Uninitializing temporary git repository")
- target = os.path.join(v8_path, ".git")
- if os.path.isdir(target):
-- print ">> Cleaning up %s" % target
-+ print(">> Cleaning up %s" % target)
- def OnRmError(func, path, exec_info):
- # This might happen on Windows
- os.chmod(path, stat.S_IWRITE)
---- node-v10.15.3/deps/v8/tools/node/update_node.py.old 2019-03-05 15:16:31.000000000 +0000
-+++ node-v10.15.3/deps/v8/tools/node/update_node.py 2019-04-02 01:00:27.184875836 +0000
-@@ -23,6 +23,9 @@
- --with-patch Also include currently staged files in the V8 checkout.
- """
-
-+# for py2/py3 compatibility
-+from __future__ import print_function
-+
- import argparse
- import os
- import shutil
-@@ -61,9 +64,9 @@
- # Node.js owns deps/v8/gypfiles in their downstream repository.
- FILES_TO_KEEP = [ "gypfiles" ]
-
- def RunGclient(path):
- assert os.path.isdir(path)
-- print ">> Running gclient sync"
-+ print(">> Running gclient sync")
- subprocess.check_call(["gclient", "sync", "--nohooks"], cwd=path)
-
- def CommitPatch(options):
-@@ -74,7 +77,7 @@
- the fake git clone fetch it into node.js. We can leave the commit, as
- bot_update will ensure a clean state on each run.
- """
-- print ">> Committing patch"
-+ print(">> Committing patch")
- subprocess.check_call(
- ["git", "-c", "user.name=fake", "-c", "user.email=fake@chromium.org",
- "commit", "--allow-empty", "-m", "placeholder-commit"],
-@@ -84,8 +87,8 @@
- def UpdateTarget(repository, options, files_to_keep):
- source = os.path.join(options.v8_path, *repository)
- target = os.path.join(options.node_path, TARGET_SUBDIR, *repository)
-- print ">> Updating target directory %s" % target
-- print ">> from active branch at %s" % source
-+ print(">> Updating target directory %s" % target)
-+ print(">> from active branch at %s" % source)
- if not os.path.exists(target):
- os.makedirs(target)
- # Remove possible remnants of previous incomplete runs.
-@@ -98,10 +101,11 @@
- git_args.append(["add"] + files_to_keep) # add and commit
- git_args.append(["commit", "-m", "keep files"]) # files we want to keep
-
-+ git_args.append(["clean", "-fxd"]) # nuke everything else
- git_args.append(["remote", "add", "source", source]) # point to source repo
- git_args.append(["fetch", "source", "HEAD"]) # sync to current branch
- git_args.append(["checkout", "-f", "FETCH_HEAD"]) # switch to that branch
-- git_args.append(["clean", "-fd"]) # delete removed files
-+ git_args.append(["clean", "-fxd"]) # delete removed files
-
- if files_to_keep:
- git_args.append(["cherry-pick", "master"]) # restore kept files
-@@ -117,17 +120,17 @@
- def UpdateGitIgnore(options):
- file_name = os.path.join(options.node_path, TARGET_SUBDIR, ".gitignore")
- assert os.path.isfile(file_name)
-- print ">> Updating .gitignore with lines"
-+ print(">> Updating .gitignore with lines")
- with open(file_name) as gitignore:
- content = gitignore.readlines()
- content = [x.strip() for x in content]
- for x in DELETE_FROM_GITIGNORE:
- if x in content:
-- print "- %s" % x
-+ print("- %s" % x)
- content.remove(x)
- for x in ADD_TO_GITIGNORE:
- if x not in content:
-- print "+ %s" % x
-+ print("+ %s" % x)
- content.append(x)
- content.sort(key=lambda x: x[1:] if x.startswith("!") else x)
- with open(file_name, "w") as gitignore:
-@@ -135,7 +138,7 @@
- gitignore.write("%s\n" % x)
-
- def CreateCommit(options):
-- print ">> Creating commit."
-+ print(">> Creating commit.")
- # Find git hash from source.
- githash = subprocess.check_output(["git", "rev-parse", "--short", "HEAD"],
- cwd=options.v8_path).strip()