summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
-rw-r--r--.mailmap9
-rwxr-xr-xbin/spack8
-rwxr-xr-xlib/spack/env/cc4
l---------lib/spack/env/nag/nagfor1
-rw-r--r--lib/spack/spack/build_environment.py4
-rw-r--r--lib/spack/spack/cmd/bootstrap.py2
-rw-r--r--lib/spack/spack/cmd/create.py2
-rw-r--r--lib/spack/spack/cmd/deactivate.py2
-rw-r--r--lib/spack/spack/cmd/location.py2
-rw-r--r--lib/spack/spack/cmd/md5.py33
-rw-r--r--lib/spack/spack/cmd/mirror.py14
-rw-r--r--lib/spack/spack/cmd/pkg.py2
-rw-r--r--lib/spack/spack/cmd/test-install.py40
-rw-r--r--lib/spack/spack/compiler.py2
-rw-r--r--lib/spack/spack/compilers/__init__.py2
-rw-r--r--lib/spack/spack/compilers/clang.py32
-rw-r--r--lib/spack/spack/compilers/nag.py33
-rw-r--r--lib/spack/spack/database.py2
-rw-r--r--lib/spack/spack/fetch_strategy.py6
-rw-r--r--lib/spack/spack/mirror.py19
-rw-r--r--lib/spack/spack/modules.py13
-rw-r--r--lib/spack/spack/package.py14
-rw-r--r--lib/spack/spack/repository.py12
-rw-r--r--lib/spack/spack/stage.py18
-rw-r--r--lib/spack/spack/test/cc.py6
-rw-r--r--lib/spack/spack/test/git_fetch.py5
-rw-r--r--lib/spack/spack/test/hg_fetch.py5
-rw-r--r--lib/spack/spack/test/install.py4
-rw-r--r--lib/spack/spack/test/make_executable.py48
-rw-r--r--lib/spack/spack/test/mirror.py72
-rw-r--r--lib/spack/spack/test/mock_repo.py10
-rw-r--r--lib/spack/spack/test/svn_fetch.py7
-rw-r--r--lib/spack/spack/util/executable.py95
-rw-r--r--var/spack/repos/builtin/packages/SuiteSparse/package.py27
-rw-r--r--var/spack/repos/builtin/packages/arpack-ng/package.py57
-rw-r--r--var/spack/repos/builtin/packages/boost/boost_11856.patch34
-rw-r--r--var/spack/repos/builtin/packages/boost/package.py114
-rw-r--r--var/spack/repos/builtin/packages/caliper/package.py25
-rw-r--r--var/spack/repos/builtin/packages/cmake/package.py20
-rw-r--r--var/spack/repos/builtin/packages/cmocka/package.py16
-rw-r--r--var/spack/repos/builtin/packages/dakota/package.py55
-rw-r--r--var/spack/repos/builtin/packages/eigen/package.py69
-rw-r--r--var/spack/repos/builtin/packages/elpa/package.py55
-rw-r--r--var/spack/repos/builtin/packages/espresso/package.py65
-rw-r--r--var/spack/repos/builtin/packages/fish/package.py3
-rw-r--r--var/spack/repos/builtin/packages/gcc/package.py15
-rw-r--r--var/spack/repos/builtin/packages/hdf/package.py45
-rw-r--r--var/spack/repos/builtin/packages/hdf5/package.py89
-rw-r--r--var/spack/repos/builtin/packages/hpx5/package.py52
-rw-r--r--var/spack/repos/builtin/packages/hwloc/package.py13
-rw-r--r--var/spack/repos/builtin/packages/libedit/package.py2
-rw-r--r--var/spack/repos/builtin/packages/libgpg-error/package.py1
-rw-r--r--var/spack/repos/builtin/packages/m4/package.py13
-rw-r--r--var/spack/repos/builtin/packages/mpfr/package.py2
-rw-r--r--var/spack/repos/builtin/packages/mpich/package.py17
-rw-r--r--var/spack/repos/builtin/packages/mumps/Makefile.inc38
-rw-r--r--var/spack/repos/builtin/packages/mumps/package.py139
-rw-r--r--var/spack/repos/builtin/packages/mvapich2/package.py34
-rw-r--r--var/spack/repos/builtin/packages/netcdf/netcdf-4.3.3-mpi.patch25
-rw-r--r--var/spack/repos/builtin/packages/netcdf/package.py86
-rw-r--r--var/spack/repos/builtin/packages/netlib-scalapack/package.py50
-rw-r--r--var/spack/repos/builtin/packages/openblas/package.py2
-rw-r--r--var/spack/repos/builtin/packages/opencv/package.py50
-rw-r--r--var/spack/repos/builtin/packages/openmpi/package.py27
-rw-r--r--var/spack/repos/builtin/packages/openssl/package.py2
-rw-r--r--var/spack/repos/builtin/packages/parallel-netcdf/package.py20
-rw-r--r--var/spack/repos/builtin/packages/paraview/package.py1
-rw-r--r--var/spack/repos/builtin/packages/py-mpi4py/package.py2
-rw-r--r--var/spack/repos/builtin/packages/py-phonopy/package.py18
-rw-r--r--var/spack/repos/builtin/packages/py-pyyaml/package.py13
-rw-r--r--var/spack/repos/builtin/packages/py-wheel/package.py15
-rw-r--r--var/spack/repos/builtin/packages/qhull/package.py14
-rw-r--r--var/spack/repos/builtin/packages/qt/package.py3
-rw-r--r--var/spack/repos/builtin/packages/trilinos/package.py38
74 files changed, 1557 insertions, 337 deletions
diff --git a/.mailmap b/.mailmap
index 1b99da32b5..39ec183241 100644
--- a/.mailmap
+++ b/.mailmap
@@ -1,13 +1,20 @@
Todd Gamblin <tgamblin@llnl.gov> George Todd Gamblin <gamblin2@llnl.gov>
+Todd Gamblin <tgamblin@llnl.gov> Todd Gamblin <gamblin2@llnl.gov>
Adam Moody <moody20@llnl.gov> Adam T. Moody <moody20@llnl.gov>
Alfredo Gimenez <gimenez1@llnl.gov> Alfredo Gimenez <alfredo.gimenez@gmail.com>
David Boehme <boehme3@llnl.gov> David Boehme <boehme3@sierra324.llnl.gov>
David Boehme <boehme3@llnl.gov> David Boehme <boehme3@sierra648.llnl.gov>
Kevin Brandstatter <kjbrandstatter@gmail.com> Kevin Brandstatter <kbrandst@hawk.iit.edu>
Luc Jaulmes <luc.jaulmes@bsc.es> Luc Jaulmes <jaulmes1@llnl.gov>
-Saravan Pantham <saravan.pantham@gmail.com> Saravan Pantham <pantham1@surface86.llnl.gov
+Saravan Pantham <saravan.pantham@gmail.com> Saravan Pantham <pantham1@surface86.llnl.gov>
Tom Scogland <tscogland@llnl.gov> Tom Scogland <scogland1@llnl.gov>
Tom Scogland <tscogland@llnl.gov> Tom Scogland <tom.scogland@gmail.com>
Joachim Protze <protze@rz.rwth-aachen.de> jprotze <protze@rz.rwth-aachen.de>
Gregory L. Lee <lee218@llnl.gov> Gregory L. Lee <lee218@surface86.llnl.gov>
+Gregory L. Lee <lee218@llnl.gov> Gregory L. Lee <lee218@cab687.llnl.gov>
+Gregory L. Lee <lee218@llnl.gov> Gregory L. Lee <lee218@cab690.llnl.gov>
+Gregory L. Lee <lee218@llnl.gov> Gregory L. Lee <lee218@catalyst159.llnl.gov>
Gregory L. Lee <lee218@llnl.gov> Gregory Lee <lee218@llnl.gov>
+Massimiliano Culpo <massimiliano.culpo@epfl.ch> Massimiliano Culpo <massimiliano.culpo@googlemail.com>
+Massimiliano Culpo <massimiliano.culpo@epfl.ch> alalazo <massimiliano.culpo@googlemail.com>
+Mark Miller <miller86@llnl.gov> miller86 <miller86@llnl.gov>
diff --git a/bin/spack b/bin/spack
index 93263217ac..31165bba9d 100755
--- a/bin/spack
+++ b/bin/spack
@@ -42,9 +42,9 @@ SPACK_EXTERNAL_LIBS = os.path.join(SPACK_LIB_PATH, "external")
sys.path.insert(0, SPACK_EXTERNAL_LIBS)
import warnings
-# Avoid warnings when nose is installed with the python exe being used to run
-# spack. Note this must be done after Spack's external libs directory is added
-# to sys.path.
+# Avoid warnings when nose is installed with the python exe being used to run
+# spack. Note this must be done after Spack's external libs directory is added
+# to sys.path.
with warnings.catch_warnings():
warnings.filterwarnings("ignore", ".*nose was already imported")
import nose
@@ -62,7 +62,7 @@ for pyc_file in orphaned_pyc_files:
try:
os.remove(pyc_file)
except OSError as e:
- print "WARNING: Spack may fail mysteriously. Couldn't remove orphaned .pyc file: %s" % pyc
+ print "WARNING: Spack may fail mysteriously. Couldn't remove orphaned .pyc file: %s" % pyc_file
# If there is no working directory, use the spack prefix.
try:
diff --git a/lib/spack/env/cc b/lib/spack/env/cc
index 0966277a91..aacba996b3 100755
--- a/lib/spack/env/cc
+++ b/lib/spack/env/cc
@@ -94,11 +94,11 @@ case "$command" in
command="$SPACK_CXX"
language="C++"
;;
- f90|fc|f95|gfortran|ifort|pgf90|xlf90)
+ f90|fc|f95|gfortran|ifort|pgf90|xlf90|nagfor)
command="$SPACK_FC"
language="Fortran 90"
;;
- f77|gfortran|ifort|pgf77|xlf)
+ f77|gfortran|ifort|pgf77|xlf|nagfor)
command="$SPACK_F77"
language="Fortran 77"
;;
diff --git a/lib/spack/env/nag/nagfor b/lib/spack/env/nag/nagfor
new file mode 120000
index 0000000000..82c2b8e90a
--- /dev/null
+++ b/lib/spack/env/nag/nagfor
@@ -0,0 +1 @@
+../cc \ No newline at end of file
diff --git a/lib/spack/spack/build_environment.py b/lib/spack/spack/build_environment.py
index b2db83acb7..1b87778080 100644
--- a/lib/spack/spack/build_environment.py
+++ b/lib/spack/spack/build_environment.py
@@ -237,9 +237,9 @@ def set_module_variables_for_package(pkg, m):
def get_rpaths(pkg):
"""Get a list of all the rpaths for a package."""
rpaths = [pkg.prefix.lib, pkg.prefix.lib64]
- rpaths.extend(d.prefix.lib for d in pkg.spec.traverse(root=False)
+ rpaths.extend(d.prefix.lib for d in pkg.spec.dependencies.values()
if os.path.isdir(d.prefix.lib))
- rpaths.extend(d.prefix.lib64 for d in pkg.spec.traverse(root=False)
+ rpaths.extend(d.prefix.lib64 for d in pkg.spec.dependencies.values()
if os.path.isdir(d.prefix.lib64))
return rpaths
diff --git a/lib/spack/spack/cmd/bootstrap.py b/lib/spack/spack/cmd/bootstrap.py
index e4ec7da35d..bdbd623b39 100644
--- a/lib/spack/spack/cmd/bootstrap.py
+++ b/lib/spack/spack/cmd/bootstrap.py
@@ -42,7 +42,7 @@ def get_origin_url():
git = which('git', required=True)
origin_url = git(
'--git-dir=%s' % git_dir, 'config', '--get', 'remote.origin.url',
- return_output=True)
+ output=str)
return origin_url.strip()
diff --git a/lib/spack/spack/cmd/create.py b/lib/spack/spack/cmd/create.py
index 7cea39cb55..edcea0718c 100644
--- a/lib/spack/spack/cmd/create.py
+++ b/lib/spack/spack/cmd/create.py
@@ -132,7 +132,7 @@ class ConfigureGuesser(object):
# Peek inside the tarball.
tar = which('tar')
output = tar(
- "--exclude=*/*/*", "-tf", stage.archive_file, return_output=True)
+ "--exclude=*/*/*", "-tf", stage.archive_file, output=str)
lines = output.split("\n")
# Set the configure line to the one that matched.
diff --git a/lib/spack/spack/cmd/deactivate.py b/lib/spack/spack/cmd/deactivate.py
index a0c78bf755..d6b23d6a08 100644
--- a/lib/spack/spack/cmd/deactivate.py
+++ b/lib/spack/spack/cmd/deactivate.py
@@ -37,7 +37,7 @@ def setup_parser(subparser):
help="Run deactivation even if spec is NOT currently activated.")
subparser.add_argument(
'-a', '--all', action='store_true',
- help="Deactivate all extensions of an extendable pacakge, or "
+ help="Deactivate all extensions of an extendable package, or "
"deactivate an extension AND its dependencies.")
subparser.add_argument(
'spec', nargs=argparse.REMAINDER, help="spec of package extension to deactivate.")
diff --git a/lib/spack/spack/cmd/location.py b/lib/spack/spack/cmd/location.py
index 39c225e9b2..307ee8982d 100644
--- a/lib/spack/spack/cmd/location.py
+++ b/lib/spack/spack/cmd/location.py
@@ -32,7 +32,7 @@ from llnl.util.filesystem import join_path
import spack
import spack.cmd
-description="Print out locations of various diectories used by Spack"
+description="Print out locations of various directories used by Spack"
def setup_parser(subparser):
global directories
diff --git a/lib/spack/spack/cmd/md5.py b/lib/spack/spack/cmd/md5.py
index ef1e4f3475..879ef9f7b7 100644
--- a/lib/spack/spack/cmd/md5.py
+++ b/lib/spack/spack/cmd/md5.py
@@ -23,15 +23,31 @@
# Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
##############################################################################
import os
-import hashlib
import argparse
+import hashlib
+
+from contextlib import contextmanager
import llnl.util.tty as tty
from llnl.util.filesystem import *
import spack.util.crypto
+from spack.stage import Stage, FailedDownloadError
+
+description = "Calculate md5 checksums for files/urls."
-description = "Calculate md5 checksums for files."
+@contextmanager
+def stager(url):
+ _cwd = os.getcwd()
+ _stager = Stage(url)
+ try:
+ _stager.fetch()
+ yield _stager
+ except FailedDownloadError:
+ tty.msg("Failed to fetch %s" % url)
+ finally:
+ _stager.destroy()
+ os.chdir(_cwd) # the Stage class changes the current working dir so it has to be restored
def setup_parser(subparser):
setup_parser.parser = subparser
@@ -45,9 +61,12 @@ def md5(parser, args):
for f in args.files:
if not os.path.isfile(f):
- tty.die("Not a file: %s" % f)
- if not can_access(f):
- tty.die("Cannot read file: %s" % f)
+ with stager(f) as stage:
+ checksum = spack.util.crypto.checksum(hashlib.md5, stage.archive_file)
+ print "%s %s" % (checksum, f)
+ else:
+ if not can_access(f):
+ tty.die("Cannot read file: %s" % f)
- checksum = spack.util.crypto.checksum(hashlib.md5, f)
- print "%s %s" % (checksum, f)
+ checksum = spack.util.crypto.checksum(hashlib.md5, f)
+ print "%s %s" % (checksum, f)
diff --git a/lib/spack/spack/cmd/mirror.py b/lib/spack/spack/cmd/mirror.py
index 885483a840..8e9438c1a3 100644
--- a/lib/spack/spack/cmd/mirror.py
+++ b/lib/spack/spack/cmd/mirror.py
@@ -53,11 +53,13 @@ def setup_parser(subparser):
create_parser.add_argument('-d', '--directory', default=None,
help="Directory in which to create mirror.")
create_parser.add_argument(
- 'specs', nargs=argparse.REMAINDER, help="Specs of packages to put in mirror")
+ 'specs', nargs=argparse.REMAINDER,
+ help="Specs of packages to put in mirror")
create_parser.add_argument(
'-f', '--file', help="File with specs of packages to put in mirror.")
create_parser.add_argument(
- '-D', '--dependencies', action='store_true', help="Also fetch all dependencies")
+ '-D', '--dependencies', action='store_true',
+ help="Also fetch all dependencies")
create_parser.add_argument(
'-o', '--one-version-per-spec', action='store_const', const=1, default=0,
help="Only fetch one 'preferred' version per spec, not all known versions.")
@@ -74,7 +76,8 @@ def setup_parser(subparser):
help="Configuration scope to modify.")
# Remove
- remove_parser = sp.add_parser('remove', aliases=['rm'], help=mirror_remove.__doc__)
+ remove_parser = sp.add_parser('remove', aliases=['rm'],
+ help=mirror_remove.__doc__)
remove_parser.add_argument('name')
remove_parser.add_argument(
'--scope', choices=scopes, default=spack.cmd.default_modify_scope,
@@ -141,15 +144,17 @@ def mirror_list(args):
def _read_specs_from_file(filename):
+ specs = []
with open(filename, "r") as stream:
for i, string in enumerate(stream):
try:
s = Spec(string)
s.package
- args.specs.append(s)
+ specs.append(s)
except SpackError, e:
tty.die("Parse error in %s, line %d:" % (args.file, i+1),
">>> " + string, str(e))
+ return specs
def mirror_create(args):
@@ -169,6 +174,7 @@ def mirror_create(args):
specs = [Spec(n) for n in spack.repo.all_package_names()]
specs.sort(key=lambda s: s.format("$_$@").lower())
+ # If the user asked for dependencies, traverse spec DAG get them.
if args.dependencies:
new_specs = set()
for spec in specs:
diff --git a/lib/spack/spack/cmd/pkg.py b/lib/spack/spack/cmd/pkg.py
index 448f762841..cf478d3763 100644
--- a/lib/spack/spack/cmd/pkg.py
+++ b/lib/spack/spack/cmd/pkg.py
@@ -79,7 +79,7 @@ def list_packages(rev):
git = get_git()
relpath = spack.packages_path[len(spack.prefix + os.path.sep):] + os.path.sep
output = git('ls-tree', '--full-tree', '--name-only', rev, relpath,
- return_output=True)
+ output=str)
return sorted(line[len(relpath):] for line in output.split('\n') if line)
diff --git a/lib/spack/spack/cmd/test-install.py b/lib/spack/spack/cmd/test-install.py
index 74017f59fb..656873a2f0 100644
--- a/lib/spack/spack/cmd/test-install.py
+++ b/lib/spack/spack/cmd/test-install.py
@@ -37,20 +37,20 @@ from spack.build_environment import InstallError
from spack.fetch_strategy import FetchError
import spack.cmd
-description = "Treat package installations as unit tests and output formatted test results"
+description = "Run package installation as a unit test, output formatted results."
def setup_parser(subparser):
subparser.add_argument(
'-j', '--jobs', action='store', type=int,
help="Explicitly set number of make jobs. Default is #cpus.")
-
+
subparser.add_argument(
'-n', '--no-checksum', action='store_true', dest='no_checksum',
help="Do not check packages against checksum")
-
+
subparser.add_argument(
'-o', '--output', action='store', help="test output goes in this file")
-
+
subparser.add_argument(
'package', nargs=argparse.REMAINDER, help="spec of package to install")
@@ -59,10 +59,10 @@ class JunitResultFormat(object):
def __init__(self):
self.root = ET.Element('testsuite')
self.tests = []
-
+
def add_test(self, buildId, testResult, buildInfo=None):
self.tests.append((buildId, testResult, buildInfo))
-
+
def write_to(self, stream):
self.root.set('tests', '{0}'.format(len(self.tests)))
for buildId, testResult, buildInfo in self.tests:
@@ -84,25 +84,25 @@ class TestResult(object):
PASSED = 0
FAILED = 1
SKIPPED = 2
-
+
class BuildId(object):
def __init__(self, spec):
self.name = spec.name
self.version = spec.version
self.hashId = spec.dag_hash()
-
+
def stringId(self):
return "-".join(str(x) for x in (self.name, self.version, self.hashId))
def __hash__(self):
return hash((self.name, self.version, self.hashId))
-
+
def __eq__(self, other):
if not isinstance(other, BuildId):
return False
-
- return ((self.name, self.version, self.hashId) ==
+
+ return ((self.name, self.version, self.hashId) ==
(other.name, other.version, other.hashId))
@@ -114,12 +114,12 @@ def fetch_log(path):
def failed_dependencies(spec):
- return set(childSpec for childSpec in spec.dependencies.itervalues() if not
+ return set(childSpec for childSpec in spec.dependencies.itervalues() if not
spack.repo.get(childSpec).installed)
def create_test_output(topSpec, newInstalls, output, getLogFunc=fetch_log):
- # Post-order traversal is not strictly required but it makes sense to output
+ # Post-order traversal is not strictly required but it makes sense to output
# tests for dependencies first.
for spec in topSpec.traverse(order='post'):
if spec not in newInstalls:
@@ -143,12 +143,12 @@ def create_test_output(topSpec, newInstalls, output, getLogFunc=fetch_log):
re.search('error:', line, re.IGNORECASE))
errOutput = errMessages if errMessages else lines[-10:]
errOutput = '\n'.join(itertools.chain(
- [spec.to_yaml(), "Errors:"], errOutput,
+ [spec.to_yaml(), "Errors:"], errOutput,
["Build Log:", package.build_log_path]))
else:
result = TestResult.PASSED
errOutput = None
-
+
bId = BuildId(spec)
output.add_test(bId, result, errOutput)
@@ -163,18 +163,18 @@ def test_install(parser, args):
if args.no_checksum:
spack.do_checksum = False # TODO: remove this global.
-
+
specs = spack.cmd.parse_specs(args.package, concretize=True)
if len(specs) > 1:
tty.die("Only 1 top-level package can be specified")
topSpec = iter(specs).next()
-
+
newInstalls = set()
for spec in topSpec.traverse():
package = spack.repo.get(spec)
if not package.installed:
newInstalls.add(spec)
-
+
if not args.output:
bId = BuildId(topSpec)
outputDir = join_path(os.getcwd(), "test-output")
@@ -183,7 +183,7 @@ def test_install(parser, args):
outputFpath = join_path(outputDir, "test-{0}.xml".format(bId.stringId()))
else:
outputFpath = args.output
-
+
for spec in topSpec.traverse(order='post'):
# Calling do_install for the top-level package would be sufficient but
# this attempts to keep going if any package fails (other packages which
@@ -202,7 +202,7 @@ def test_install(parser, args):
pass
except FetchError:
pass
-
+
jrf = JunitResultFormat()
handled = {}
create_test_output(topSpec, newInstalls, jrf)
diff --git a/lib/spack/spack/compiler.py b/lib/spack/spack/compiler.py
index a665f6062d..12c02e0ea2 100644
--- a/lib/spack/spack/compiler.py
+++ b/lib/spack/spack/compiler.py
@@ -51,7 +51,7 @@ _version_cache = {}
def get_compiler_version(compiler_path, version_arg, regex='(.*)'):
if not compiler_path in _version_cache:
compiler = Executable(compiler_path)
- output = compiler(version_arg, return_output=True, error=os.devnull)
+ output = compiler(version_arg, output=str, error=str)
match = re.search(regex, output)
_version_cache[compiler_path] = match.group(1) if match else 'unknown'
diff --git a/lib/spack/spack/compilers/__init__.py b/lib/spack/spack/compilers/__init__.py
index facc9c338b..6159ef576c 100644
--- a/lib/spack/spack/compilers/__init__.py
+++ b/lib/spack/spack/compilers/__init__.py
@@ -51,7 +51,7 @@ _required_instance_vars = ['cc', 'cxx', 'f77', 'fc']
if platform.system() == 'Darwin':
_default_order = ['clang', 'gcc', 'intel']
else:
- _default_order = ['gcc', 'intel', 'pgi', 'clang', 'xlc']
+ _default_order = ['gcc', 'intel', 'pgi', 'clang', 'xlc', 'nag']
def _auto_compiler_spec(function):
diff --git a/lib/spack/spack/compilers/clang.py b/lib/spack/spack/compilers/clang.py
index 340051019c..e406d86a24 100644
--- a/lib/spack/spack/compilers/clang.py
+++ b/lib/spack/spack/compilers/clang.py
@@ -22,7 +22,10 @@
# along with this program; if not, write to the Free Software Foundation,
# Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
##############################################################################
+import re
+import spack.compiler as cpr
from spack.compiler import *
+from spack.util.executable import *
class Clang(Compiler):
# Subclasses use possible names of C compiler
@@ -47,11 +50,34 @@ class Clang(Compiler):
@classmethod
def default_version(self, comp):
"""The '--version' option works for clang compilers.
- Output looks like this::
+ On most platforms, output looks like this::
clang version 3.1 (trunk 149096)
Target: x86_64-unknown-linux-gnu
Thread model: posix
+
+ On Mac OS X, it looks like this:
+
+ Apple LLVM version 7.0.2 (clang-700.1.81)
+ Target: x86_64-apple-darwin15.2.0
+ Thread model: posix
+
"""
- return get_compiler_version(
- comp, '--version', r'(?:clang version|based on LLVM) ([^ )]+)')
+ if comp not in cpr._version_cache:
+ compiler = Executable(comp)
+ output = compiler('--version', output=str, error=str)
+
+ ver = 'unknown'
+ match = re.search(r'^Apple LLVM version ([^ )]+)', output)
+ if match:
+ # Apple's LLVM compiler has its own versions, so suffix them.
+ ver = match.group(1) + '-apple'
+ else:
+ # Normal clang compiler versions are left as-is
+ match = re.search(r'^clang version ([^ )]+)', output)
+ if match:
+ ver = match.group(1)
+
+ cpr._version_cache[comp] = ver
+
+ return cpr._version_cache[comp]
diff --git a/lib/spack/spack/compilers/nag.py b/lib/spack/spack/compilers/nag.py
new file mode 100644
index 0000000000..527a05a090
--- /dev/null
+++ b/lib/spack/spack/compilers/nag.py
@@ -0,0 +1,33 @@
+from spack.compiler import *
+
+class Nag(Compiler):
+ # Subclasses use possible names of C compiler
+ cc_names = []
+
+ # Subclasses use possible names of C++ compiler
+ cxx_names = []
+
+ # Subclasses use possible names of Fortran 77 compiler
+ f77_names = ['nagfor']
+
+ # Subclasses use possible names of Fortran 90 compiler
+ fc_names = ['nagfor']
+
+ # Named wrapper links within spack.build_env_path
+ link_paths = { # Use default wrappers for C and C++, in case provided in compilers.yaml
+ 'cc' : 'cc',
+ 'cxx' : 'c++',
+ 'f77' : 'nag/nagfor',
+ 'fc' : 'nag/nagfor' }
+
+ @classmethod
+ def default_version(self, comp):
+ """The '-V' option works for nag compilers.
+ Output looks like this::
+
+ NAG Fortran Compiler Release 6.0(Hibiya) Build 1037
+ Product NPL6A60NA for x86-64 Linux
+ Copyright 1990-2015 The Numerical Algorithms Group Ltd., Oxford, U.K.
+ """
+ return get_compiler_version(
+ comp, '-V', r'NAG Fortran Compiler Release ([0-9.]+)')
diff --git a/lib/spack/spack/database.py b/lib/spack/spack/database.py
index 0fa18db34b..9cbe7de44a 100644
--- a/lib/spack/spack/database.py
+++ b/lib/spack/spack/database.py
@@ -489,7 +489,7 @@ class Database(object):
1. Marks the spec as not installed.
2. Removes the spec if it has no more dependents.
3. If removed, recursively updates dependencies' ref counts
- and remvoes them if they are no longer needed.
+ and removes them if they are no longer needed.
"""
# Take a lock around the entire removal.
diff --git a/lib/spack/spack/fetch_strategy.py b/lib/spack/spack/fetch_strategy.py
index 0657146bf6..337dd1e198 100644
--- a/lib/spack/spack/fetch_strategy.py
+++ b/lib/spack/spack/fetch_strategy.py
@@ -154,7 +154,7 @@ class URLFetchStrategy(FetchStrategy):
# Run curl but grab the mime type from the http headers
headers = spack.curl(
- *curl_args, return_output=True, fail_on_error=False)
+ *curl_args, output=str, fail_on_error=False)
if spack.curl.returncode != 0:
# clean up archive on failure.
@@ -375,7 +375,7 @@ class GitFetchStrategy(VCSFetchStrategy):
@property
def git_version(self):
- vstring = self.git('--version', return_output=True).lstrip('git version ')
+ vstring = self.git('--version', output=str).lstrip('git version ')
return Version(vstring)
@@ -518,7 +518,7 @@ class SvnFetchStrategy(VCSFetchStrategy):
def _remove_untracked_files(self):
"""Removes untracked files in an svn repository."""
- status = self.svn('status', '--no-ignore', return_output=True)
+ status = self.svn('status', '--no-ignore', output=str)
self.svn('status', '--no-ignore')
for line in status.split('\n'):
if not re.match('^[I?]', line):
diff --git a/lib/spack/spack/mirror.py b/lib/spack/spack/mirror.py
index 1d9b0e7ef2..341cc4cb88 100644
--- a/lib/spack/spack/mirror.py
+++ b/lib/spack/spack/mirror.py
@@ -78,6 +78,7 @@ def get_matching_versions(specs, **kwargs):
continue
num_versions = kwargs.get('num_versions', 0)
+ matching_spec = []
for i, v in enumerate(reversed(sorted(pkg.versions))):
# Generate no more than num_versions versions for each spec.
if num_versions and i >= num_versions:
@@ -88,7 +89,11 @@ def get_matching_versions(specs, **kwargs):
s = Spec(pkg.name)
s.versions = VersionList([v])
s.variants = spec.variants.copy()
- matching.append(s)
+ matching_spec.append(s)
+
+ if not matching_spec:
+ tty.warn("No known version matches spec: %s" % spec)
+ matching.extend(matching_spec)
return matching
@@ -147,7 +152,11 @@ def create(path, specs, **kwargs):
# Get the absolute path of the root before we start jumping around.
mirror_root = os.path.abspath(path)
if not os.path.isdir(mirror_root):
- mkdirp(mirror_root)
+ try:
+ mkdirp(mirror_root)
+ except OSError as e:
+ raise MirrorError(
+ "Cannot create directory '%s':" % mirror_root, str(e))
# Things to keep track of while parsing specs.
present = []
@@ -164,7 +173,11 @@ def create(path, specs, **kwargs):
# create a subdirectory for the current package@version
archive_path = os.path.abspath(join_path(mirror_root, mirror_archive_path(spec)))
subdir = os.path.dirname(archive_path)
- mkdirp(subdir)
+ try:
+ mkdirp(subdir)
+ except OSError as e:
+ raise MirrorError(
+ "Cannot create directory '%s':" % subdir, str(e))
if os.path.exists(archive_path):
tty.msg("Already added %s" % spec.format("$_$@"))
diff --git a/lib/spack/spack/modules.py b/lib/spack/spack/modules.py
index 7036626e29..c834763564 100644
--- a/lib/spack/spack/modules.py
+++ b/lib/spack/spack/modules.py
@@ -194,12 +194,14 @@ class Dotkit(EnvModule):
@property
def file_name(self):
return join_path(Dotkit.path, self.spec.architecture,
- self.spec.format('$_$@$%@$+$#.dk'))
+ '%s.dk' % self.use_name)
@property
def use_name(self):
- return self.spec.format('$_$@$%@$+$#')
-
+ return "%s-%s-%s-%s-%s" % (self.spec.name, self.spec.version,
+ self.spec.compiler.name,
+ self.spec.compiler.version,
+ self.spec.dag_hash())
def _write(self, dk_file):
# Category
@@ -235,7 +237,10 @@ class TclModule(EnvModule):
@property
def use_name(self):
- return self.spec.format('$_$@$%@$+$#')
+ return "%s-%s-%s-%s-%s" % (self.spec.name, self.spec.version,
+ self.spec.compiler.name,
+ self.spec.compiler.version,
+ self.spec.dag_hash())
def _write(self, m_file):
diff --git a/lib/spack/spack/package.py b/lib/spack/spack/package.py
index fba2912b75..8cb947c276 100644
--- a/lib/spack/spack/package.py
+++ b/lib/spack/spack/package.py
@@ -631,7 +631,7 @@ class Package(object):
spack.install_layout.remove_install_directory(self.spec)
- def do_fetch(self):
+ def do_fetch(self, mirror_only=False):
"""Creates a stage directory and downloads the taball for this package.
Working directory will be set to the stage directory.
"""
@@ -656,7 +656,7 @@ class Package(object):
raise FetchError(
"Will not fetch %s." % self.spec.format('$_$@'), checksum_msg)
- self.stage.fetch()
+ self.stage.fetch(mirror_only)
##########
# Fetch resources
@@ -677,7 +677,8 @@ class Package(object):
if spack.do_checksum and self.version in self.versions:
self.stage.check()
- def do_stage(self):
+
+ def do_stage(self, mirror_only=False):
"""Unpacks the fetched tarball, then changes into the expanded tarball
directory."""
if not self.spec.concrete:
@@ -691,8 +692,7 @@ class Package(object):
else:
tty.msg("Already staged %s in %s." % (name, stage.path))
-
- self.do_fetch()
+ self.do_fetch(mirror_only)
_expand_archive(self.stage)
##########
@@ -835,10 +835,6 @@ class Package(object):
resource_stage_folder = '-'.join(pieces)
return resource_stage_folder
- def _build_logger(self, log_path):
- """Create a context manager to log build output."""
-
-
def do_install(self,
keep_prefix=False, keep_stage=False, ignore_deps=False,
diff --git a/lib/spack/spack/repository.py b/lib/spack/spack/repository.py
index 31596cee7a..f58cd52125 100644
--- a/lib/spack/spack/repository.py
+++ b/lib/spack/spack/repository.py
@@ -233,6 +233,11 @@ class RepoPath(object):
return providers
+ @_autospec
+ def extensions_for(self, extendee_spec):
+ return [p for p in self.all_packages() if p.extends(extendee_spec)]
+
+
def find_module(self, fullname, path=None):
"""Implements precedence for overlaid namespaces.
@@ -295,8 +300,11 @@ class RepoPath(object):
for repo in self.repos:
if spec.name in repo:
return repo
- else:
- raise UnknownPackageError(spec.name)
+
+ # If the package isn't in any repo, return the one with
+ # highest precedence. This is for commands like `spack edit`
+ # that can operate on packages that don't exist yet.
+ return self.first_repo()
@_autospec
diff --git a/lib/spack/spack/stage.py b/lib/spack/spack/stage.py
index 31635a854a..79c9030e20 100644
--- a/lib/spack/spack/stage.py
+++ b/lib/spack/spack/stage.py
@@ -97,7 +97,6 @@ class Stage(object):
self.name = kwargs.get('name')
self.mirror_path = kwargs.get('mirror_path')
-
self.tmp_root = find_tmp_root()
self.path = None
@@ -240,11 +239,13 @@ class Stage(object):
tty.die("Setup failed: no such directory: " + self.path)
- def fetch(self):
+ def fetch(self, mirror_only=False):
"""Downloads an archive or checks out code from a repository."""
self.chdir()
- fetchers = [self.default_fetcher]
+ fetchers = []
+ if not mirror_only:
+ fetchers.append(self.default_fetcher)
# TODO: move mirror logic out of here and clean it up!
# TODO: Or @alalazo may have some ideas about how to use a
@@ -252,7 +253,13 @@ class Stage(object):
self.skip_checksum_for_mirror = True
if self.mirror_path:
mirrors = spack.config.get_config('mirrors')
- urls = [urljoin(u, self.mirror_path) for name, u in mirrors.items()]
+
+ # Join URLs of mirror roots with mirror paths. Because
+ # urljoin() will strip everything past the final '/' in
+ # the root, so we add a '/' if it is not present.
+ mirror_roots = [root if root.endswith('/') else root + '/'
+ for root in mirrors.values()]
+ urls = [urljoin(root, self.mirror_path) for root in mirror_roots]
# If this archive is normally fetched from a tarball URL,
# then use the same digest. `spack mirror` ensures that
@@ -261,10 +268,11 @@ class Stage(object):
if isinstance(self.default_fetcher, fs.URLFetchStrategy):
digest = self.default_fetcher.digest
- # Have to skip the checkesum for things archived from
+ # Have to skip the checksum for things archived from
# repositories. How can this be made safer?
self.skip_checksum_for_mirror = not bool(digest)
+ # Add URL strategies for all the mirrors with the digest
for url in urls:
fetchers.insert(0, fs.URLFetchStrategy(url, digest))
diff --git a/lib/spack/spack/test/cc.py b/lib/spack/spack/test/cc.py
index 4188b8d550..905af28a06 100644
--- a/lib/spack/spack/test/cc.py
+++ b/lib/spack/spack/test/cc.py
@@ -65,17 +65,17 @@ class CompilerTest(unittest.TestCase):
def check_cc(self, command, args, expected):
os.environ['SPACK_TEST_COMMAND'] = command
- self.assertEqual(self.cc(*args, return_output=True).strip(), expected)
+ self.assertEqual(self.cc(*args, output=str).strip(), expected)
def check_ld(self, command, args, expected):
os.environ['SPACK_TEST_COMMAND'] = command
- self.assertEqual(self.ld(*args, return_output=True).strip(), expected)
+ self.assertEqual(self.ld(*args, output=str).strip(), expected)
def check_cpp(self, command, args, expected):
os.environ['SPACK_TEST_COMMAND'] = command
- self.assertEqual(self.cpp(*args, return_output=True).strip(), expected)
+ self.assertEqual(self.cpp(*args, output=str).strip(), expected)
def test_vcheck_mode(self):
diff --git a/lib/spack/spack/test/git_fetch.py b/lib/spack/spack/test/git_fetch.py
index 3813079065..d84433176a 100644
--- a/lib/spack/spack/test/git_fetch.py
+++ b/lib/spack/spack/test/git_fetch.py
@@ -56,10 +56,7 @@ class GitFetchTest(MockPackagesTest):
def tearDown(self):
"""Destroy the stage space used by this test."""
super(GitFetchTest, self).tearDown()
-
- if self.repo.stage is not None:
- self.repo.stage.destroy()
-
+ self.repo.destroy()
self.pkg.do_clean()
diff --git a/lib/spack/spack/test/hg_fetch.py b/lib/spack/spack/test/hg_fetch.py
index ee8327aec8..bbcb64e4c1 100644
--- a/lib/spack/spack/test/hg_fetch.py
+++ b/lib/spack/spack/test/hg_fetch.py
@@ -53,10 +53,7 @@ class HgFetchTest(MockPackagesTest):
def tearDown(self):
"""Destroy the stage space used by this test."""
super(HgFetchTest, self).tearDown()
-
- if self.repo.stage is not None:
- self.repo.stage.destroy()
-
+ self.repo.destroy()
self.pkg.do_clean()
diff --git a/lib/spack/spack/test/install.py b/lib/spack/spack/test/install.py
index 628329a423..c09bd24c8e 100644
--- a/lib/spack/spack/test/install.py
+++ b/lib/spack/spack/test/install.py
@@ -59,9 +59,7 @@ class InstallTest(MockPackagesTest):
def tearDown(self):
super(InstallTest, self).tearDown()
-
- if self.repo.stage is not None:
- self.repo.stage.destroy()
+ self.repo.destroy()
# Turn checksumming back on
spack.do_checksum = True
diff --git a/lib/spack/spack/test/make_executable.py b/lib/spack/spack/test/make_executable.py
index 09efec8580..d568a28d44 100644
--- a/lib/spack/spack/test/make_executable.py
+++ b/lib/spack/spack/test/make_executable.py
@@ -56,47 +56,47 @@ class MakeExecutableTest(unittest.TestCase):
def test_make_normal(self):
make = MakeExecutable('make', 8)
- self.assertEqual(make(return_output=True).strip(), '-j8')
- self.assertEqual(make('install', return_output=True).strip(), '-j8 install')
+ self.assertEqual(make(output=str).strip(), '-j8')
+ self.assertEqual(make('install', output=str).strip(), '-j8 install')
def test_make_explicit(self):
make = MakeExecutable('make', 8)
- self.assertEqual(make(parallel=True, return_output=True).strip(), '-j8')
- self.assertEqual(make('install', parallel=True, return_output=True).strip(), '-j8 install')
+ self.assertEqual(make(parallel=True, output=str).strip(), '-j8')
+ self.assertEqual(make('install', parallel=True, output=str).strip(), '-j8 install')
def test_make_one_job(self):
make = MakeExecutable('make', 1)
- self.assertEqual(make(return_output=True).strip(), '')
- self.assertEqual(make('install', return_output=True).strip(), 'install')
+ self.assertEqual(make(output=str).strip(), '')
+ self.assertEqual(make('install', output=str).strip(), 'install')
def test_make_parallel_false(self):
make = MakeExecutable('make', 8)
- self.assertEqual(make(parallel=False, return_output=True).strip(), '')
- self.assertEqual(make('install', parallel=False, return_output=True).strip(), 'install')
+ self.assertEqual(make(parallel=False, output=str).strip(), '')
+ self.assertEqual(make('install', parallel=False, output=str).strip(), 'install')
def test_make_parallel_disabled(self):
make = MakeExecutable('make', 8)
os.environ['SPACK_NO_PARALLEL_MAKE'] = 'true'
- self.assertEqual(make(return_output=True).strip(), '')
- self.assertEqual(make('install', return_output=True).strip(), 'install')
+ self.assertEqual(make(output=str).strip(), '')
+ self.assertEqual(make('install', output=str).strip(), 'install')
os.environ['SPACK_NO_PARALLEL_MAKE'] = '1'
- self.assertEqual(make(return_output=True).strip(), '')
- self.assertEqual(make('install', return_output=True).strip(), 'install')
+ self.assertEqual(make(output=str).strip(), '')
+ self.assertEqual(make('install', output=str).strip(), 'install')
# These don't disable (false and random string)
os.environ['SPACK_NO_PARALLEL_MAKE'] = 'false'
- self.assertEqual(make(return_output=True).strip(), '-j8')
- self.assertEqual(make('install', return_output=True).strip(), '-j8 install')
+ self.assertEqual(make(output=str).strip(), '-j8')
+ self.assertEqual(make('install', output=str).strip(), '-j8 install')
os.environ['SPACK_NO_PARALLEL_MAKE'] = 'foobar'
- self.assertEqual(make(return_output=True).strip(), '-j8')
- self.assertEqual(make('install', return_output=True).strip(), '-j8 install')
+ self.assertEqual(make(output=str).strip(), '-j8')
+ self.assertEqual(make('install', output=str).strip(), '-j8 install')
del os.environ['SPACK_NO_PARALLEL_MAKE']
@@ -106,20 +106,20 @@ class MakeExecutableTest(unittest.TestCase):
# These should work
os.environ['SPACK_NO_PARALLEL_MAKE'] = 'true'
- self.assertEqual(make(parallel=True, return_output=True).strip(), '')
- self.assertEqual(make('install', parallel=True, return_output=True).strip(), 'install')
+ self.assertEqual(make(parallel=True, output=str).strip(), '')
+ self.assertEqual(make('install', parallel=True, output=str).strip(), 'install')
os.environ['SPACK_NO_PARALLEL_MAKE'] = '1'
- self.assertEqual(make(parallel=True, return_output=True).strip(), '')
- self.assertEqual(make('install', parallel=True, return_output=True).strip(), 'install')
+ self.assertEqual(make(parallel=True, output=str).strip(), '')
+ self.assertEqual(make('install', parallel=True, output=str).strip(), 'install')
# These don't disable (false and random string)
os.environ['SPACK_NO_PARALLEL_MAKE'] = 'false'
- self.assertEqual(make(parallel=True, return_output=True).strip(), '-j8')
- self.assertEqual(make('install', parallel=True, return_output=True).strip(), '-j8 install')
+ self.assertEqual(make(parallel=True, output=str).strip(), '-j8')
+ self.assertEqual(make('install', parallel=True, output=str).strip(), '-j8 install')
os.environ['SPACK_NO_PARALLEL_MAKE'] = 'foobar'
- self.assertEqual(make(parallel=True, return_output=True).strip(), '-j8')
- self.assertEqual(make('install', parallel=True, return_output=True).strip(), '-j8 install')
+ self.assertEqual(make(parallel=True, output=str).strip(), '-j8')
+ self.assertEqual(make('install', parallel=True, output=str).strip(), '-j8 install')
del os.environ['SPACK_NO_PARALLEL_MAKE']
diff --git a/lib/spack/spack/test/mirror.py b/lib/spack/spack/test/mirror.py
index 04e9e3db2e..046ec56604 100644
--- a/lib/spack/spack/test/mirror.py
+++ b/lib/spack/spack/test/mirror.py
@@ -44,8 +44,16 @@ class MirrorTest(MockPackagesTest):
self.repos = {}
+ def tearDown(self):
+ """Destroy all the stages created by the repos in setup."""
+ super(MirrorTest, self).tearDown()
+ for repo in self.repos.values():
+ repo.destroy()
+ self.repos.clear()
+
+
def set_up_package(self, name, MockRepoClass, url_attr):
- """Use this to set up a mock package to be mirrored.
+ """Set up a mock package to be mirrored.
Each package needs us to:
1. Set up a mock repo/archive to fetch from.
2. Point the package's version args at that repo.
@@ -65,21 +73,14 @@ class MirrorTest(MockPackagesTest):
pkg.versions[v][url_attr] = repo.url
- def tearDown(self):
- """Destroy all the stages created by the repos in setup."""
- super(MirrorTest, self).tearDown()
-
- for name, repo in self.repos.items():
- if repo.stage:
- pass #repo.stage.destroy()
-
- self.repos.clear()
-
-
def check_mirror(self):
stage = Stage('spack-mirror-test')
mirror_root = join_path(stage.path, 'test-mirror')
+ # register mirror with spack config
+ mirrors = { 'spack-mirror-test' : 'file://' + mirror_root }
+ spack.config.update_config('mirrors', mirrors)
+
try:
os.chdir(stage.path)
spack.mirror.create(
@@ -88,7 +89,7 @@ class MirrorTest(MockPackagesTest):
# Stage directory exists
self.assertTrue(os.path.isdir(mirror_root))
- # subdirs for each package
+ # check that there are subdirs for each package
for name in self.repos:
subdir = join_path(mirror_root, name)
self.assertTrue(os.path.isdir(subdir))
@@ -96,40 +97,37 @@ class MirrorTest(MockPackagesTest):
files = os.listdir(subdir)
self.assertEqual(len(files), 1)
- # Decompress archive in the mirror
- archive = files[0]
- archive_path = join_path(subdir, archive)
- decomp = decompressor_for(archive_path)
-
- with working_dir(subdir):
- decomp(archive_path)
+ # Now try to fetch each package.
+ for name, mock_repo in self.repos.items():
+ spec = Spec(name).concretized()
+ pkg = spec.package
- # Find the untarred archive directory.
- files = os.listdir(subdir)
- self.assertEqual(len(files), 2)
- self.assertTrue(archive in files)
- files.remove(archive)
-
- expanded_archive = join_path(subdir, files[0])
- self.assertTrue(os.path.isdir(expanded_archive))
+ saved_checksum_setting = spack.do_checksum
+ try:
+ # Stage the archive from the mirror and cd to it.
+ spack.do_checksum = False
+ pkg.do_stage(mirror_only=True)
# Compare the original repo with the expanded archive
- repo = self.repos[name]
- if not 'svn' in name:
- original_path = repo.path
- else:
- co = 'checked_out'
- svn('checkout', repo.url, co)
- original_path = join_path(subdir, co)
+ original_path = mock_repo.path
+ if 'svn' in name:
+ # have to check out the svn repo to compare.
+ original_path = join_path(mock_repo.path, 'checked_out')
+ svn('checkout', mock_repo.url, original_path)
- dcmp = dircmp(original_path, expanded_archive)
+ dcmp = dircmp(original_path, pkg.stage.source_path)
# make sure there are no new files in the expanded tarball
self.assertFalse(dcmp.right_only)
+
+ # and that all original files are present.
self.assertTrue(all(l in exclude for l in dcmp.left_only))
+ finally:
+ spack.do_checksum = saved_checksum_setting
+ pkg.do_clean()
finally:
- pass #stage.destroy()
+ stage.destroy()
def test_git_mirror(self):
diff --git a/lib/spack/spack/test/mock_repo.py b/lib/spack/spack/test/mock_repo.py
index c454b1f106..ed94023b0e 100644
--- a/lib/spack/spack/test/mock_repo.py
+++ b/lib/spack/spack/test/mock_repo.py
@@ -55,6 +55,12 @@ class MockRepo(object):
mkdirp(self.path)
+ def destroy(self):
+ """Destroy resources associated with this mock repo."""
+ if self.stage:
+ self.stage.destroy()
+
+
class MockArchive(MockRepo):
"""Creates a very simple archive directory with a configure script and a
makefile that installs to a prefix. Tars it up into an archive."""
@@ -141,7 +147,7 @@ class MockGitRepo(MockVCSRepo):
self.url = self.path
def rev_hash(self, rev):
- return git('rev-parse', rev, return_output=True).strip()
+ return git('rev-parse', rev, output=str).strip()
class MockSvnRepo(MockVCSRepo):
@@ -193,4 +199,4 @@ class MockHgRepo(MockVCSRepo):
def get_rev(self):
"""Get current mercurial revision."""
- return hg('id', '-i', return_output=True).strip()
+ return hg('id', '-i', output=str).strip()
diff --git a/lib/spack/spack/test/svn_fetch.py b/lib/spack/spack/test/svn_fetch.py
index 2ee4748fdb..454a7f1d1f 100644
--- a/lib/spack/spack/test/svn_fetch.py
+++ b/lib/spack/spack/test/svn_fetch.py
@@ -55,17 +55,14 @@ class SvnFetchTest(MockPackagesTest):
def tearDown(self):
"""Destroy the stage space used by this test."""
super(SvnFetchTest, self).tearDown()
-
- if self.repo.stage is not None:
- self.repo.stage.destroy()
-
+ self.repo.destroy()
self.pkg.do_clean()
def assert_rev(self, rev):
"""Check that the current revision is equal to the supplied rev."""
def get_rev():
- output = svn('info', return_output=True)
+ output = svn('info', output=str)
self.assertTrue("Revision" in output)
for line in output.split('\n'):
match = re.match(r'Revision: (\d+)', line)
diff --git a/lib/spack/spack/util/executable.py b/lib/spack/spack/util/executable.py
index ba765eb662..fc27b789d0 100644
--- a/lib/spack/spack/util/executable.py
+++ b/lib/spack/spack/util/executable.py
@@ -55,24 +55,80 @@ class Executable(object):
def __call__(self, *args, **kwargs):
- """Run the executable with subprocess.check_output, return output."""
- return_output = kwargs.get("return_output", False)
- fail_on_error = kwargs.get("fail_on_error", True)
- ignore_errors = kwargs.get("ignore_errors", ())
+ """Run this executable in a subprocess.
+
+ Arguments
+ args
+ command line arguments to the executable to run.
+
+ Optional arguments
+
+ fail_on_error
+
+ Raise an exception if the subprocess returns an
+ error. Default is True. When not set, the return code is
+ avaiale as `exe.returncode`.
+
+ ignore_errors
+
+ An optional list/tuple of error codes that can be
+ *ignored*. i.e., if these codes are returned, this will
+ not raise an exception when `fail_on_error` is `True`.
+
+ output, error
+
+ These arguments allow you to specify new stdout and stderr
+ values. They default to `None`, which means the
+ subprocess will inherit the parent's file descriptors.
+
+ You can set these to:
+ - python streams, e.g. open Python file objects, or os.devnull;
+ - filenames, which will be automatically opened for writing; or
+ - `str`, as in the Python string type. If you set these to `str`,
+ output and error will be written to pipes and returned as
+ a string. If both `output` and `error` are set to `str`,
+ then one string is returned containing output concatenated
+ with error.
+
+ input
+
+ Same as output, error, but `str` is not an allowed value.
+
+ Deprecated arguments
+
+ return_output[=False]
+
+ Setting this to True is the same as setting output=str.
+ This argument may be removed in future Spack versions.
+
+ """
+ fail_on_error = kwargs.pop("fail_on_error", True)
+ ignore_errors = kwargs.pop("ignore_errors", ())
+
+ # TODO: This is deprecated. Remove in a future version.
+ return_output = kwargs.pop("return_output", False)
# Default values of None says to keep parent's file descriptors.
- output = kwargs.get("output", None)
- error = kwargs.get("error", None)
- input = kwargs.get("input", None)
+ if return_output:
+ output = str
+ else:
+ output = kwargs.pop("output", None)
+
+ error = kwargs.pop("error", None)
+ input = kwargs.pop("input", None)
+ if input is str:
+ raise ValueError("Cannot use `str` as input stream.")
def streamify(arg, mode):
if isinstance(arg, basestring):
return open(arg, mode), True
+ elif arg is str:
+ return subprocess.PIPE, False
else:
return arg, False
- output, ostream = streamify(output, 'w')
- error, estream = streamify(error, 'w')
- input, istream = streamify(input, 'r')
+ ostream, close_ostream = streamify(output, 'w')
+ estream, close_estream = streamify(error, 'w')
+ istream, close_istream = streamify(input, 'r')
# if they just want to ignore one error code, make it a tuple.
if isinstance(ignore_errors, int):
@@ -92,19 +148,20 @@ class Executable(object):
tty.debug(cmd_line)
try:
- if return_output:
- output = subprocess.PIPE
-
proc = subprocess.Popen(
- cmd, stdin=input, stderr=error, stdout=output)
+ cmd, stdin=istream, stderr=estream, stdout=ostream)
out, err = proc.communicate()
rc = self.returncode = proc.returncode
if fail_on_error and rc != 0 and (rc not in ignore_errors):
raise ProcessError("Command exited with status %d:"
% proc.returncode, cmd_line)
- if return_output:
- return out
+
+ if output is str or error is str:
+ result = ''
+ if output is str: result += out
+ if error is str: result += err
+ return result
except OSError, e:
raise ProcessError(
@@ -119,9 +176,9 @@ class Executable(object):
% (proc.returncode, cmd_line))
finally:
- if ostream: output.close()
- if estream: error.close()
- if istream: input.close()
+ if close_ostream: output.close()
+ if close_estream: error.close()
+ if close_istream: input.close()
def __eq__(self, other):
diff --git a/var/spack/repos/builtin/packages/SuiteSparse/package.py b/var/spack/repos/builtin/packages/SuiteSparse/package.py
new file mode 100644
index 0000000000..6e130d118f
--- /dev/null
+++ b/var/spack/repos/builtin/packages/SuiteSparse/package.py
@@ -0,0 +1,27 @@
+from spack import *
+
+
+class Suitesparse(Package):
+ """
+ SuiteSparse is a suite of sparse matrix algorithms
+ """
+ homepage = 'http://faculty.cse.tamu.edu/davis/suitesparse.html'
+ url = 'http://faculty.cse.tamu.edu/davis/SuiteSparse/SuiteSparse-4.5.1.tar.gz'
+
+ version('4.5.1', 'f0ea9aad8d2d1ffec66a5b6bfeff5319')
+
+ depends_on('blas')
+ depends_on('lapack')
+
+ depends_on('metis@5.1.0', when='@4.5.1')
+
+ def install(self, spec, prefix):
+ # The build system of SuiteSparse is quite old-fashioned
+ # It's basically a plain Makefile which include an header (SuiteSparse_config/SuiteSparse_config.mk)
+ # with a lot of convoluted logic in it.
+ # Any kind of customization will need to go through filtering of that file
+
+ # FIXME : this actually uses the current workaround
+ # FIXME : (blas / lapack always provide libblas and liblapack as aliases)
+ make('install', 'INSTALL=%s' % prefix, 'BLAS=-lblas', 'LAPACK=-llapack')
+
diff --git a/var/spack/repos/builtin/packages/arpack-ng/package.py b/var/spack/repos/builtin/packages/arpack-ng/package.py
new file mode 100644
index 0000000000..0b49d14202
--- /dev/null
+++ b/var/spack/repos/builtin/packages/arpack-ng/package.py
@@ -0,0 +1,57 @@
+from spack import *
+
+
+class ArpackNg(Package):
+ """
+ ARPACK-NG is a collection of Fortran77 subroutines designed to solve large scale eigenvalue problems.
+
+ Important Features:
+
+ * Reverse Communication Interface.
+ * Single and Double Precision Real Arithmetic Versions for Symmetric,
+ Non-symmetric, Standard or Generalized Problems.
+ * Single and Double Precision Complex Arithmetic Versions for Standard or
+ Generalized Problems.
+ * Routines for Banded Matrices - Standard or Generalized Problems.
+ * Routines for The Singular Value Decomposition.
+ * Example driver routines that may be used as templates to implement numerous
+ Shift-Invert strategies for all problem types, data types and precision.
+
+ This project is a joint project between Debian, Octave and Scilab in order to
+ provide a common and maintained version of arpack.
+
+ Indeed, no single release has been published by Rice university for the last
+ few years and since many software (Octave, Scilab, R, Matlab...) forked it and
+ implemented their own modifications, arpack-ng aims to tackle this by providing
+ a common repository and maintained versions.
+
+ arpack-ng is replacing arpack almost everywhere.
+ """
+ homepage = 'https://github.com/opencollab/arpack-ng'
+ url = 'https://github.com/opencollab/arpack-ng/archive/3.3.0.tar.gz'
+
+ version('3.3.0', 'ed3648a23f0a868a43ef44c97a21bad5')
+
+ variant('shared', default=True, description='Enables the build of shared libraries')
+ variant('mpi', default=False, description='Activates MPI support')
+
+ depends_on('blas')
+ depends_on('lapack')
+ depends_on('mpi', when='+mpi')
+
+ def install(self, spec, prefix):
+ # Apparently autotools are not bootstrapped
+ bootstrap = Executable('./bootstrap')
+
+ options = ['--prefix=%s' % prefix]
+
+ if '+mpi' in spec:
+ options.append('--enable-mpi')
+
+ if '~shared' in spec:
+ options.append('--enable-shared=no')
+
+ bootstrap()
+ configure(*options)
+ make()
+ make('install')
diff --git a/var/spack/repos/builtin/packages/boost/boost_11856.patch b/var/spack/repos/builtin/packages/boost/boost_11856.patch
new file mode 100644
index 0000000000..3b4052ca18
--- /dev/null
+++ b/var/spack/repos/builtin/packages/boost/boost_11856.patch
@@ -0,0 +1,34 @@
+--- a/libs/container/src/pool_resource.cpp 2015-11-06 12:49:55.000000000 -0800
++++ b/libs/container/src/pool_resource.cpp 2015-12-22 07:54:36.202131121 -0800
+@@ -32,11 +32,11 @@
+ class pool_data_t
+ : public block_slist_base<>
+ {
+- typedef block_slist_base<> block_slist_base;
++ typedef block_slist_base<> block_slist_base_t;
+
+ public:
+ explicit pool_data_t(std::size_t initial_blocks_per_chunk)
+- : block_slist_base(), next_blocks_per_chunk(initial_blocks_per_chunk)
++ : block_slist_base_t(), next_blocks_per_chunk(initial_blocks_per_chunk)
+ { slist_algo::init_header(&free_slist); }
+
+ void *allocate_block() BOOST_NOEXCEPT
+@@ -59,7 +59,7 @@
+ void release(memory_resource &upstream)
+ {
+ slist_algo::init_header(&free_slist);
+- this->block_slist_base::release(upstream);
++ this->block_slist_base_t::release(upstream);
+ next_blocks_per_chunk = pool_options_minimum_max_blocks_per_chunk;
+ }
+
+@@ -72,7 +72,7 @@
+
+ //Minimum block size is at least max_align, so all pools allocate sizes that are multiple of max_align,
+ //meaning that all blocks are max_align-aligned.
+- char *p = static_cast<char *>(block_slist_base::allocate(blocks_per_chunk*pool_block, mr));
++ char *p = static_cast<char *>(block_slist_base_t::allocate(blocks_per_chunk*pool_block, mr));
+
+ //Create header types. This is no-throw
+ for(std::size_t i = 0, max = blocks_per_chunk; i != max; ++i){
diff --git a/var/spack/repos/builtin/packages/boost/package.py b/var/spack/repos/builtin/packages/boost/package.py
index 3427b74ad6..fb1f5daee7 100644
--- a/var/spack/repos/builtin/packages/boost/package.py
+++ b/var/spack/repos/builtin/packages/boost/package.py
@@ -1,4 +1,5 @@
from spack import *
+import spack
class Boost(Package):
"""Boost provides free peer-reviewed portable C++ source
@@ -44,15 +45,50 @@ class Boost(Package):
version('1.34.1', '2d938467e8a448a2c9763e0a9f8ca7e5')
version('1.34.0', 'ed5b9291ffad776f8757a916e1726ad0')
+ default_install_libs = set(['atomic',
+ 'chrono',
+ 'date_time',
+ 'filesystem',
+ 'graph',
+ 'iostreams',
+ 'locale',
+ 'log',
+ 'math',
+ 'program_options',
+ 'random',
+ 'regex',
+ 'serialization',
+ 'signals',
+ 'system',
+ 'test',
+ 'thread',
+ 'wave'])
+
+ # mpi/python are not installed by default because they pull in many
+ # dependencies and/or because there is a great deal of customization
+ # possible (and it would be difficult to choose sensible defaults)
+ default_noinstall_libs = set(['mpi', 'python'])
+
+ all_libs = default_install_libs | default_noinstall_libs
+
+ for lib in all_libs:
+ variant(lib, default=(lib not in default_noinstall_libs),
+ description="Compile with {0} library".format(lib))
+
variant('debug', default=False, description='Switch to the debug version of Boost')
- variant('python', default=False, description='Activate the component Boost.Python')
- variant('mpi', default=False, description='Activate the component Boost.MPI')
- variant('compression', default=True, description='Activate the compression Boost.iostreams')
+ variant('shared', default=True, description="Additionally build shared libraries")
+ variant('multithreaded', default=True, description="Build multi-threaded versions of libraries")
+ variant('singlethreaded', default=True, description="Build single-threaded versions of libraries")
+ variant('icu_support', default=False, description="Include ICU support (for regex/locale libraries)")
+ depends_on('icu', when='+icu_support')
depends_on('python', when='+python')
depends_on('mpi', when='+mpi')
- depends_on('bzip2', when='+compression')
- depends_on('zlib', when='+compression')
+ depends_on('bzip2', when='+iostreams')
+ depends_on('zlib', when='+iostreams')
+
+ # Patch fix from https://svn.boost.org/trac/boost/ticket/11856
+ patch('boost_11856.patch', when='@1.60.0%gcc@4.4.7')
def url_for_version(self, version):
"""Handle Boost's weird URLs, which write the version two different ways."""
@@ -77,22 +113,20 @@ class Boost(Package):
# fallback to gcc if no toolset found
return 'gcc'
- def determine_bootstrap_options(self, spec, options):
- options.append('--with-toolset=%s' % self.determine_toolset(spec))
+ def determine_bootstrap_options(self, spec, withLibs, options):
+ boostToolsetId = self.determine_toolset(spec)
+ options.append('--with-toolset=%s' % boostToolsetId)
+ options.append("--with-libraries=%s" % ','.join(withLibs))
- without_libs = []
- if '~mpi' in spec:
- without_libs.append('mpi')
- if '~python' in spec:
- without_libs.append('python')
- else:
+ if '+python' in spec:
options.append('--with-python=%s' %
join_path(spec['python'].prefix.bin, 'python'))
- if without_libs:
- options.append('--without-libraries=%s' % ','.join(without_libs))
-
with open('user-config.jam', 'w') as f:
+ compiler_wrapper = join_path(spack.build_env_path, 'c++')
+ f.write("using {0} : : {1} ;\n".format(boostToolsetId,
+ compiler_wrapper))
+
if '+mpi' in spec:
f.write('using mpi : %s ;\n' %
join_path(spec['mpi'].prefix.bin, 'mpicxx'))
@@ -107,12 +141,10 @@ class Boost(Package):
else:
options.append('variant=release')
- if '~compression' in spec:
- options.extend([
- '-s', 'NO_BZIP2=1',
- '-s', 'NO_ZLIB=1'])
+ if '+icu_support' in spec:
+ options.extend(['-s', 'ICU_PATH=%s' % spec['icu'].prefix])
- if '+compression' in spec:
+ if '+iostreams' in spec:
options.extend([
'-s', 'BZIP2_INCLUDE=%s' % spec['bzip2'].prefix.include,
'-s', 'BZIP2_LIBPATH=%s' % spec['bzip2'].prefix.lib,
@@ -120,20 +152,46 @@ class Boost(Package):
'-s', 'ZLIB_LIBPATH=%s' % spec['zlib'].prefix.lib,
])
+ linkTypes = ['static']
+ if '+shared' in spec:
+ linkTypes.append('shared')
+
+ threadingOpts = []
+ if '+multithreaded' in spec:
+ threadingOpts.append('multi')
+ if '+singlethreaded' in spec:
+ threadingOpts.append('single')
+ if not threadingOpts:
+ raise RuntimeError("At least one of {singlethreaded, multithreaded} must be enabled")
+
options.extend([
'toolset=%s' % self.determine_toolset(spec),
- 'link=static,shared',
- 'threading=single,multi',
+ 'link=%s' % ','.join(linkTypes),
'--layout=tagged'])
+
+ return threadingOpts
def install(self, spec, prefix):
+ withLibs = list()
+ for lib in Boost.all_libs:
+ if "+{0}".format(lib) in spec:
+ withLibs.append(lib)
+ if not withLibs:
+ # if no libraries are specified for compilation, then you dont have
+ # to configure/build anything, just copy over to the prefix directory.
+ src = join_path(self.stage.source_path, 'boost')
+ mkdirp(join_path(prefix, 'include'))
+ dst = join_path(prefix, 'include', 'boost')
+ install_tree(src, dst)
+ return
+
# to make Boost find the user-config.jam
env['BOOST_BUILD_PATH'] = './'
bootstrap = Executable('./bootstrap.sh')
bootstrap_options = ['--prefix=%s' % prefix]
- self.determine_bootstrap_options(spec, bootstrap_options)
+ self.determine_bootstrap_options(spec, withLibs, bootstrap_options)
bootstrap(*bootstrap_options)
@@ -143,6 +201,10 @@ class Boost(Package):
b2 = Executable(b2name)
b2_options = ['-j', '%s' % make_jobs]
- self.determine_b2_options(spec, b2_options)
+ threadingOpts = self.determine_b2_options(spec, b2_options)
- b2('install', *b2_options)
+ # In theory it could be done on one call but it fails on
+ # Boost.MPI if the threading options are not separated.
+ for threadingOpt in threadingOpts:
+ b2('install', 'threading=%s' % threadingOpt, *b2_options)
+
diff --git a/var/spack/repos/builtin/packages/caliper/package.py b/var/spack/repos/builtin/packages/caliper/package.py
new file mode 100644
index 0000000000..d51b4a4dd5
--- /dev/null
+++ b/var/spack/repos/builtin/packages/caliper/package.py
@@ -0,0 +1,25 @@
+from spack import *
+
+class Caliper(Package):
+ """
+ Caliper is a generic context annotation system. It gives programmers the
+ ability to provide arbitrary program context information to (performance)
+ tools at runtime.
+ """
+
+ homepage = "https://github.com/LLNL/Caliper"
+ url = ""
+
+ version('master', git='ssh://git@github.com:LLNL/Caliper.git')
+
+ variant('mpi', default=False, description='Enable MPI function wrappers.')
+
+ depends_on('libunwind')
+ depends_on('papi')
+ depends_on('mpi', when='+mpi')
+
+ def install(self, spec, prefix):
+ with working_dir('build', create=True):
+ cmake('..', *std_cmake_args)
+ make()
+ make("install")
diff --git a/var/spack/repos/builtin/packages/cmake/package.py b/var/spack/repos/builtin/packages/cmake/package.py
index f67ae21ebd..e20c1e4aeb 100644
--- a/var/spack/repos/builtin/packages/cmake/package.py
+++ b/var/spack/repos/builtin/packages/cmake/package.py
@@ -28,20 +28,22 @@ class Cmake(Package):
"""A cross-platform, open-source build system. CMake is a family of
tools designed to build, test and package software."""
homepage = 'https://www.cmake.org'
+ url = 'https://cmake.org/files/v3.4/cmake-3.4.3.tar.gz'
- version('2.8.10.2', '097278785da7182ec0aea8769d06860c',
- url = 'http://www.cmake.org/files/v2.8/cmake-2.8.10.2.tar.gz')
-
- version('3.0.2', 'db4c687a31444a929d2fdc36c4dfb95f',
- url = 'http://www.cmake.org/files/v3.0/cmake-3.0.2.tar.gz')
-
- version('3.4.0', 'cd3034e0a44256a0917e254167217fc8',
- url = 'http://cmake.org/files/v3.4/cmake-3.4.0.tar.gz')
+ version('3.4.3', '4cb3ff35b2472aae70f542116d616e63')
+ version('3.4.0', 'cd3034e0a44256a0917e254167217fc8')
+ version('3.3.1', '52638576f4e1e621fed6c3410d3a1b12')
+ version('3.0.2', 'db4c687a31444a929d2fdc36c4dfb95f')
+ version('2.8.10.2', '097278785da7182ec0aea8769d06860c')
variant('ncurses', default=True, description='Enables the build of the ncurses gui')
-
depends_on('ncurses', when='+ncurses')
+ def url_for_version(self, version):
+ """Handle CMake's version-based custom URLs."""
+ return 'https://cmake.org/files/v%s/cmake-%s.tar.gz' % (version.up_to(2), version)
+
+
def install(self, spec, prefix):
configure('--prefix=' + prefix,
'--parallel=' + str(make_jobs),
diff --git a/var/spack/repos/builtin/packages/cmocka/package.py b/var/spack/repos/builtin/packages/cmocka/package.py
new file mode 100644
index 0000000000..7377016a6b
--- /dev/null
+++ b/var/spack/repos/builtin/packages/cmocka/package.py
@@ -0,0 +1,16 @@
+from spack import *
+
+class Cmocka(Package):
+ """Unit-testing framework in pure C"""
+ homepage = "https://cmocka.org/"
+ url = "https://cmocka.org/files/1.0/cmocka-1.0.1.tar.xz"
+
+ version('1.0.1', 'ed861e501a21a92b2af63e466df2015e')
+ parallel = False
+
+ def install(self, spec, prefix):
+ with working_dir('spack-build', create=True):
+ cmake('..', *std_cmake_args)
+
+ make()
+ make("install")
diff --git a/var/spack/repos/builtin/packages/dakota/package.py b/var/spack/repos/builtin/packages/dakota/package.py
new file mode 100644
index 0000000000..5ec82ef83d
--- /dev/null
+++ b/var/spack/repos/builtin/packages/dakota/package.py
@@ -0,0 +1,55 @@
+from spack import *
+
+
+class Dakota(Package):
+ """
+ The Dakota toolkit provides a flexible, extensible interface between analysis codes and iterative systems
+ analysis methods. Dakota contains algorithms for:
+
+ - optimization with gradient and non gradient-based methods;
+ - uncertainty quantification with sampling, reliability, stochastic expansion, and epistemic methods;
+ - parameter estimation with nonlinear least squares methods;
+ - sensitivity/variance analysis with design of experiments and parameter study methods.
+
+ These capabilities may be used on their own or as components within advanced strategies such as hybrid optimization,
+ surrogate-based optimization, mixed integer nonlinear programming, or optimization under uncertainty.
+ """
+
+ homepage = 'https://dakota.sandia.gov/'
+ url = 'https://dakota.sandia.gov/sites/default/files/distributions/public/dakota-6.3-public.src.tar.gz'
+ _url_str = 'https://dakota.sandia.gov/sites/default/files/distributions/public/dakota-{version}-public.src.tar.gz'
+
+ version('6.3', '05a58d209fae604af234c894c3f73f6d')
+
+ variant('debug', default=False, description='Builds a debug version of the libraries')
+ variant('shared', default=True, description='Enables the build of shared libraries')
+ variant('mpi', default=True, description='Activates MPI support')
+
+ depends_on('blas')
+ depends_on('lapack')
+ depends_on('mpi', when='+mpi')
+
+ depends_on('python')
+ depends_on('boost')
+
+ def url_for_version(self, version):
+ return Dakota._url_str.format(version=version)
+
+ def install(self, spec, prefix):
+ options = []
+ options.extend(std_cmake_args)
+
+ options.extend(['-DCMAKE_BUILD_TYPE:STRING=%s' % ('Debug' if '+debug' in spec else 'Release'),
+ '-DBUILD_SHARED_LIBS:BOOL=%s' % ('ON' if '+shared' in spec else 'OFF')])
+
+ if '+mpi' in spec:
+ options.extend(['-DDAKOTA_HAVE_MPI:BOOL=ON',
+ '-DMPI_CXX_COMPILER:STRING=%s' % join_path(spec['mpi'].prefix.bin, 'mpicxx')])
+
+ build_directory = join_path(self.stage.path, 'spack-build')
+ source_directory = self.stage.source_path
+
+ with working_dir(build_directory, create=True):
+ cmake(source_directory, *options)
+ make()
+ make("install")
diff --git a/var/spack/repos/builtin/packages/eigen/package.py b/var/spack/repos/builtin/packages/eigen/package.py
new file mode 100644
index 0000000000..e40046b452
--- /dev/null
+++ b/var/spack/repos/builtin/packages/eigen/package.py
@@ -0,0 +1,69 @@
+##############################################################################
+# Copyright (c) 2013, Lawrence Livermore National Security, LLC.
+# Produced at the Lawrence Livermore National Laboratory.
+#
+# This file is part of Spack.
+# Written by David Beckingsale, david@llnl.gov, All rights reserved.
+# LLNL-CODE-647188
+#
+# For details, see https://github.com/llnl/spack
+# Please also see the LICENSE file for our notice and the LGPL.
+#
+# This program is free software; you can redistribute it and/or modify
+# it under the terms of the GNU General Public License (as published by
+# the Free Software Foundation) version 2.1 dated February 1999.
+#
+# This program is distributed in the hope that it will be useful, but
+# WITHOUT ANY WARRANTY; without even the IMPLIED WARRANTY OF
+# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the terms and
+# conditions of the GNU General Public License for more details.
+#
+# You should have received a copy of the GNU Lesser General Public License
+# along with this program; if not, write to the Free Software Foundation,
+# Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
+##############################################################################
+
+from spack import *
+
+
+class Eigen(Package):
+ """
+ Eigen is a C++ template library for linear algebra: matrices, vectors, numerical solvers, and related algorithms
+ """
+
+ homepage = 'http://eigen.tuxfamily.org/'
+ url = 'http://bitbucket.org/eigen/eigen/get/3.2.7.tar.bz2'
+
+ version('3.2.7', 'cc1bacbad97558b97da6b77c9644f184', url='http://bitbucket.org/eigen/eigen/get/3.2.7.tar.bz2')
+
+ variant('debug', default=False, description='Builds the library in debug mode')
+
+ variant('metis', default=True, description='Enables metis backend')
+ variant('scotch', default=True, description='Enables scotch backend')
+ variant('fftw', default=True, description='Enables FFTW backend')
+ variant('suitesparse', default=True, description='Enables SuiteSparse support')
+
+ # TODO : dependency on googlehash, superlu, adolc missing
+
+ depends_on('metis', when='+metis')
+ depends_on('scotch', when='+scotch')
+ depends_on('fftw', when='+fftw')
+ depends_on('SuiteSparse', when='+suitesparse')
+ depends_on('mpfr@2.3.0:') # Eigen 3.2.7 requires at least 2.3.0
+ depends_on('gmp')
+
+ def install(self, spec, prefix):
+
+ options = []
+ options.extend(std_cmake_args)
+
+ build_directory = join_path(self.stage.path, 'spack-build')
+ source_directory = self.stage.source_path
+
+ if '+debug' in spec:
+ options.append('-DCMAKE_BUILD_TYPE:STRING=Debug')
+
+ with working_dir(build_directory, create=True):
+ cmake(source_directory, *options)
+ make()
+ make("install")
diff --git a/var/spack/repos/builtin/packages/elpa/package.py b/var/spack/repos/builtin/packages/elpa/package.py
new file mode 100644
index 0000000000..2ade5b0b37
--- /dev/null
+++ b/var/spack/repos/builtin/packages/elpa/package.py
@@ -0,0 +1,55 @@
+##############################################################################
+# Copyright (c) 2013, Lawrence Livermore National Security, LLC.
+# Produced at the Lawrence Livermore National Laboratory.
+#
+# This file is part of Spack.
+# Written by Todd Gamblin, tgamblin@llnl.gov, All rights reserved.
+# LLNL-CODE-647188
+#
+# For details, see https://github.com/llnl/spack
+# Please also see the LICENSE file for our notice and the LGPL.
+#
+# This program is free software; you can redistribute it and/or modify
+# it under the terms of the GNU General Public License (as published by
+# the Free Software Foundation) version 2.1 dated February 1999.
+#
+# This program is distributed in the hope that it will be useful, but
+# WITHOUT ANY WARRANTY; without even the IMPLIED WARRANTY OF
+# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the terms and
+# conditions of the GNU General Public License for more details.
+#
+# You should have received a copy of the GNU Lesser General Public License
+# along with this program; if not, write to the Free Software Foundation,
+# Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
+##############################################################################
+
+from spack import *
+
+
+class Elpa(Package):
+ """
+ Eigenvalue solvers for Petaflop-Applications (ELPA)
+ """
+
+ homepage = 'http://elpa.mpcdf.mpg.de/'
+ url = 'http://elpa.mpcdf.mpg.de/elpa-2015.11.001.tar.gz'
+
+ version('2015.11.001', 'de0f35b7ee7c971fd0dca35c900b87e6', url='http://elpa.mpcdf.mpg.de/elpa-2015.11.001.tar.gz')
+
+ variant('openmp', default=False, description='Activates OpenMP support')
+
+ depends_on('mpi')
+ depends_on('blas')
+ depends_on('lapack')
+ depends_on('scalapack')
+
+ def install(self, spec, prefix):
+
+ options = ["--prefix=%s" % prefix]
+
+ if '+openmp' in spec:
+ options.append("--enable-openmp")
+
+ configure(*options)
+ make()
+ make("install")
diff --git a/var/spack/repos/builtin/packages/espresso/package.py b/var/spack/repos/builtin/packages/espresso/package.py
new file mode 100644
index 0000000000..a2bf58f585
--- /dev/null
+++ b/var/spack/repos/builtin/packages/espresso/package.py
@@ -0,0 +1,65 @@
+from spack import *
+
+import os
+
+class Espresso(Package):
+ """
+ QE is an integrated suite of Open-Source computer codes for electronic-structure calculations and materials
+ modeling at the nanoscale. It is based on density-functional theory, plane waves, and pseudopotentials.
+ """
+ homepage = 'http://quantum-espresso.org'
+ url = 'http://www.qe-forge.org/gf/download/frsrelease/204/912/espresso-5.3.0.tar.gz'
+
+ version('5.3.0', '6848fcfaeb118587d6be36bd10b7f2c3')
+
+ variant('mpi', default=True, description='Build Quantum-ESPRESSO with mpi support')
+ variant('openmp', default=False, description='Enables openMP support')
+ variant('scalapack', default=True, description='Enables scalapack support')
+ variant('elpa', default=True, description='Use elpa as an eigenvalue solver')
+
+ depends_on('blas')
+ depends_on('lapack')
+
+ depends_on('mpi', when='+mpi')
+ depends_on('fftw~mpi', when='~mpi')
+ depends_on('fftw+mpi', when='+mpi')
+ depends_on('scalapack', when='+scalapack+mpi') # TODO : + mpi needed to avoid false dependencies installation
+
+ def check_variants(self, spec):
+ error = 'you cannot ask for \'+{variant}\' when \'+mpi\' is not active'
+ if '+scalapack' in spec and '~mpi' in spec:
+ raise RuntimeError(error.format(variant='scalapack'))
+ if '+elpa' in spec and ('~mpi' in spec or '~scalapack' in spec):
+ raise RuntimeError(error.format(variant='elpa'))
+
+ def install(self, spec, prefix):
+ self.check_variants(spec)
+
+ options = ['-prefix=%s' % prefix]
+
+ if '+mpi' in spec:
+ options.append('--enable-parallel')
+
+ if '+openmp' in spec:
+ options.append('--enable-openmp')
+
+ if '+scalapack' in spec:
+ options.append('--with-scalapack=yes')
+
+ if '+elpa' in spec:
+ options.append('--with-elpa=yes')
+
+ # Add a list of directories to search
+ search_list = []
+ for name, dependency_spec in spec.dependencies.iteritems():
+ search_list.extend([dependency_spec.prefix.lib,
+ dependency_spec.prefix.lib64])
+
+ search_list = " ".join(search_list)
+ options.append('LIBDIRS=%s' % search_list)
+ options.append('F90=%s' % os.environ['FC'])
+
+ configure(*options)
+ make('all')
+ make('install')
+
diff --git a/var/spack/repos/builtin/packages/fish/package.py b/var/spack/repos/builtin/packages/fish/package.py
index 1225558705..b5a4a2d209 100644
--- a/var/spack/repos/builtin/packages/fish/package.py
+++ b/var/spack/repos/builtin/packages/fish/package.py
@@ -7,7 +7,8 @@ class Fish(Package):
homepage = "http://fishshell.com/"
url = "http://fishshell.com/files/2.2.0/fish-2.2.0.tar.gz"
- list_url = homepage
+ list_url = "http://fishshell.com/files/"
+ list_depth = 2
version('2.2.0', 'a76339fd14ce2ec229283c53e805faac48c3e99d9e3ede9d82c0554acfc7b77a')
diff --git a/var/spack/repos/builtin/packages/gcc/package.py b/var/spack/repos/builtin/packages/gcc/package.py
index 61b16f3fd8..f8958ee290 100644
--- a/var/spack/repos/builtin/packages/gcc/package.py
+++ b/var/spack/repos/builtin/packages/gcc/package.py
@@ -36,8 +36,6 @@ class Gcc(Package):
list_url = 'http://open-source-box.org/gcc/'
list_depth = 2
- DEPENDS_ON_ISL_PREDICATE = '@5.0:'
-
version('5.3.0', 'c9616fd448f980259c31de613e575719')
version('5.2.0', 'a51bcfeb3da7dd4c623e27207ed43467')
version('4.9.3', '6f831b4d251872736e8e9cc09746f327')
@@ -50,15 +48,14 @@ class Gcc(Package):
version('4.5.4', '27e459c2566b8209ab064570e1b378f7')
variant('gold', default=True, description="Build the gold linker plugin for ld-based LTO")
-
+
depends_on("mpfr")
depends_on("gmp")
- depends_on("mpc") # when @4.5:
+ depends_on("mpc", when='@4.5:')
+ depends_on("isl", when='@5.0:')
depends_on("binutils~libiberty", when='~gold')
depends_on("binutils~libiberty+gold", when='+gold')
- # Save these until we can do optional deps.
- depends_on("isl", when=DEPENDS_ON_ISL_PREDICATE)
#depends_on("ppl")
#depends_on("cloog")
@@ -91,7 +88,7 @@ class Gcc(Package):
"--with-as=%s/bin/as" % spec['binutils'].prefix]
options.extend(binutils_options)
# Isl
- if spec.satisfies(Gcc.DEPENDS_ON_ISL_PREDICATE):
+ if 'isl' in spec:
isl_options = ["--with-isl=%s" % spec['isl'].prefix]
options.extend(isl_options)
@@ -102,7 +99,7 @@ class Gcc(Package):
configure(*options)
make()
make("install")
-
+
self.write_rpath_specs()
@@ -121,7 +118,7 @@ class Gcc(Package):
return
gcc = Executable(join_path(self.prefix.bin, 'gcc'))
- lines = gcc('-dumpspecs', return_output=True).strip().split("\n")
+ lines = gcc('-dumpspecs', output=str).strip().split("\n")
specs_file = join_path(self.spec_dir, 'specs')
with closing(open(specs_file, 'w')) as out:
for line in lines:
diff --git a/var/spack/repos/builtin/packages/hdf/package.py b/var/spack/repos/builtin/packages/hdf/package.py
new file mode 100644
index 0000000000..ac6435f2a2
--- /dev/null
+++ b/var/spack/repos/builtin/packages/hdf/package.py
@@ -0,0 +1,45 @@
+from spack import *
+
+class Hdf(Package):
+ """HDF4 (also known as HDF) is a library and multi-object
+ file format for storing and managing data between machines."""
+
+ homepage = "https://www.hdfgroup.org/products/hdf4/"
+ url = "https://www.hdfgroup.org/ftp/HDF/releases/HDF4.2.11/src/hdf-4.2.11.tar.gz"
+ list_url = "https://www.hdfgroup.org/ftp/HDF/releases/"
+ list_depth = 3
+
+ version('4.2.11', '063f9928f3a19cc21367b71c3b8bbf19')
+
+ variant('szip', default=False, description="Enable szip support")
+
+ depends_on("jpeg")
+ depends_on("szip", when='+szip')
+ depends_on("zlib")
+
+
+ def url_for_version(self, version):
+ return "https://www.hdfgroup.org/ftp/HDF/releases/HDF" + str(version) + "/src/hdf-" + str(version) + ".tar.gz"
+
+
+ def install(self, spec, prefix):
+ config_args = [
+ 'CFLAGS=-fPIC',
+ '--prefix=%s' % prefix,
+ '--with-jpeg=%s' % spec['jpeg'].prefix,
+ '--with-zlib=%s' % spec['zlib'].prefix,
+ '--disable-netcdf', # must be disabled to build NetCDF with HDF4 support
+ '--enable-fortran',
+ '--disable-shared', # fortran and shared libraries are not compatible
+ '--enable-static',
+ '--enable-production'
+ ]
+
+ # SZip support
+ if '+szip' in spec:
+ config_args.append('--with-szlib=%s' % spec['szip'].prefix)
+
+ configure(*config_args)
+
+ make()
+ make("install")
diff --git a/var/spack/repos/builtin/packages/hdf5/package.py b/var/spack/repos/builtin/packages/hdf5/package.py
index 9a40164341..80f79539c0 100644
--- a/var/spack/repos/builtin/packages/hdf5/package.py
+++ b/var/spack/repos/builtin/packages/hdf5/package.py
@@ -1,5 +1,31 @@
+##############################################################################
+# Copyright (c) 2013, Lawrence Livermore National Security, LLC.
+# Produced at the Lawrence Livermore National Laboratory.
+#
+# This file is part of Spack.
+# Written by Todd Gamblin, tgamblin@llnl.gov, All rights reserved.
+# LLNL-CODE-647188
+#
+# For details, see https://github.com/llnl/spack
+# Please also see the LICENSE file for our notice and the LGPL.
+#
+# This program is free software; you can redistribute it and/or modify
+# it under the terms of the GNU General Public License (as published by
+# the Free Software Foundation) version 2.1 dated February 1999.
+#
+# This program is distributed in the hope that it will be useful, but
+# WITHOUT ANY WARRANTY; without even the IMPLIED WARRANTY OF
+# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the terms and
+# conditions of the GNU General Public License for more details.
+#
+# You should have received a copy of the GNU Lesser General Public License
+# along with this program; if not, write to the Free Software Foundation,
+# Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
+##############################################################################
+
from spack import *
+
class Hdf5(Package):
"""HDF5 is a data model, library, and file format for storing and managing
data. It supports an unlimited variety of datatypes, and is designed for
@@ -7,7 +33,7 @@ class Hdf5(Package):
"""
homepage = "http://www.hdfgroup.org/HDF5/"
- url = "http://www.hdfgroup.org/ftp/HDF5/releases/hdf5-1.8.13/src/hdf5-1.8.13.tar.gz"
+ url = "http://www.hdfgroup.org/ftp/HDF5/releases/hdf5-1.8.13/src/hdf5-1.8.13.tar.gz"
list_url = "http://www.hdfgroup.org/ftp/HDF5/releases"
list_depth = 3
@@ -15,26 +41,55 @@ class Hdf5(Package):
version('1.8.15', '03cccb5b33dbe975fdcd8ae9dc021f24')
version('1.8.13', 'c03426e9e77d7766944654280b467289')
+ variant('debug', default=False, description='Builds a debug version of the library')
+
variant('cxx', default=True, description='Enable C++ support')
variant('fortran', default=True, description='Enable Fortran support')
+ variant('unsupported', default=True, description='Enables unsupported configuration options')
+
variant('mpi', default=False, description='Enable MPI support')
- variant('threadsafe', default=False, description='Enable multithreading')
+ variant('szip', default=False, description='Enable szip support')
+ variant('threadsafe', default=False, description='Enable thread-safe capabilities')
depends_on("mpi", when='+mpi')
+ depends_on("szip", when='+szip')
depends_on("zlib")
- # TODO: currently hard-coded to use OpenMPI
+ def validate(self, spec):
+ """
+ Checks if incompatible variants have been activated at the same time
+
+ :param spec: spec of the package
+ :raises RuntimeError: in case of inconsistencies
+ """
+ if '+fortran' in spec and not self.compiler.fc:
+ msg = 'cannot build a fortran variant without a fortran compiler'
+ raise RuntimeError(msg)
+
+ if '+threadsafe' in spec and ('+cxx' in spec or '+fortran' in spec):
+ raise RuntimeError("cannot use variant +threadsafe with either +cxx or +fortran")
+
def install(self, spec, prefix):
+ self.validate(spec)
+ # Handle compilation after spec validation
extra_args = []
+ if '+debug' in spec:
+ extra_args.append('--enable-debug=all')
+ else:
+ extra_args.append('--enable-production')
+
+ if '+unsupported' in spec:
+ extra_args.append("--enable-unsupported")
+
if '+cxx' in spec:
- extra_args.extend([
- '--enable-cxx'
- ])
+ extra_args.append('--enable-cxx')
+
if '+fortran' in spec:
extra_args.extend([
'--enable-fortran',
'--enable-fortran2003'
])
+
if '+mpi' in spec:
# The HDF5 configure script warns if cxx and mpi are enabled
# together. There doesn't seem to be a real reason for this, except
@@ -43,27 +98,29 @@ class Hdf5(Package):
# this is not actually a problem.
extra_args.extend([
"--enable-parallel",
- "--enable-unsupported",
"CC=%s" % spec['mpi'].prefix.bin + "/mpicc",
- "CXX=%s" % spec['mpi'].prefix.bin + "/mpic++",
- "FC=%s" % spec['mpi'].prefix.bin + "/mpifort",
])
- if '+threads' in spec:
- if '+cxx' in spec or '+fortran' in spec:
- die("Cannot use variant +threads with either +cxx or +fortran")
+
+ if '+cxx' in spec:
+ extra_args.append("CXX=%s" % spec['mpi'].prefix.bin + "/mpic++")
+
+ if '+fortran' in spec:
+ extra_args.append("FC=%s" % spec['mpi'].prefix.bin + "/mpifort")
+
+ if '+szip' in spec:
+ extra_args.append("--with-szlib=%s" % spec['szip'].prefix)
+
+ if '+threadsafe' in spec:
extra_args.extend([
'--enable-threadsafe',
'--disable-hl',
- 'CPPFLAGS=-DHDatexit=""',
- 'CFLAGS=-DHDatexit=""'
])
configure(
"--prefix=%s" % prefix,
"--with-zlib=%s" % spec['zlib'].prefix,
- "--enable-shared",
+ "--enable-shared", # TODO : this should be enabled by default, remove it?
*extra_args)
-
make()
make("install")
diff --git a/var/spack/repos/builtin/packages/hpx5/package.py b/var/spack/repos/builtin/packages/hpx5/package.py
new file mode 100644
index 0000000000..3dae3c4170
--- /dev/null
+++ b/var/spack/repos/builtin/packages/hpx5/package.py
@@ -0,0 +1,52 @@
+from spack import *
+import os
+
+class Hpx5(Package):
+ """The HPX-5 Runtime System. HPX-5 (High Performance ParalleX) is an
+ open source, portable, performance-oriented runtime developed at
+ CREST (Indiana University). HPX-5 provides a distributed
+ programming model allowing programs to run unmodified on systems
+ from a single SMP to large clusters and supercomputers with
+ thousands of nodes. HPX-5 supports a wide variety of Intel and ARM
+ platforms. It is being used by a broad range of scientific
+ applications enabling scientists to write code that performs and
+ scales better than contemporary runtimes."""
+ homepage = "http://hpx.crest.iu.edu"
+ url = "http://hpx.crest.iu.edu/release/hpx-2.0.0.tar.gz"
+
+ version('2.0.0', '3d2ff3aab6c46481f9ec65c5b2bfe7a6')
+ version('1.3.0', '2260ecc7f850e71a4d365a43017d8cee')
+ version('1.2.0', '4972005f85566af4afe8b71afbf1480f')
+ version('1.1.0', '646afb460ecb7e0eea713a634933ce4f')
+ version('1.0.0', '8020822adf6090bd59ed7fe465f6c6cb')
+
+ variant('debug', default=False, description='Build a debug version of HPX-5')
+ variant('photon', default=False, description='Enable Photon support')
+ variant('mpi', default=False, description='Enable MPI support')
+
+ depends_on("mpi", when='+mpi')
+ depends_on("mpi", when='+photon')
+
+ def install(self, spec, prefix):
+ extra_args = []
+ if '+debug' in spec:
+ extra_args.extend([
+ '--enable-debug',
+ 'CFLAGS=-g -O0'
+ ])
+ else:
+ extra_args.append('CFLAGS=-O3')
+
+ if '+mpi' in spec:
+ extra_args.append('--enable-mpi')
+
+ if '+photon' in spec:
+ extra_args.extend([
+ '--enable-mpi',
+ '--enable-photon'
+ ])
+
+ os.chdir("./hpx/")
+ configure('--prefix=%s' % prefix, *extra_args)
+ make()
+ make("install")
diff --git a/var/spack/repos/builtin/packages/hwloc/package.py b/var/spack/repos/builtin/packages/hwloc/package.py
index 7ebede76a3..ab7205646e 100644
--- a/var/spack/repos/builtin/packages/hwloc/package.py
+++ b/var/spack/repos/builtin/packages/hwloc/package.py
@@ -14,15 +14,18 @@ class Hwloc(Package):
efficiently."""
homepage = "http://www.open-mpi.org/projects/hwloc/"
url = "http://www.open-mpi.org/software/hwloc/v1.9/downloads/hwloc-1.9.tar.gz"
+ list_url = "http://www.open-mpi.org/software/hwloc/"
+ list_depth = 3
- version('1.11.2', '486169cbe111cdea57be12638828ebbf',
- url='http://www.open-mpi.org/software/hwloc/v1.11/downloads/hwloc-1.11.2.tar.bz2')
- version('1.11.1', '002742efd3a8431f98d6315365a2b543',
- url='http://www.open-mpi.org/software/hwloc/v1.11/downloads/hwloc-1.11.1.tar.bz2')
- version('1.9', '1f9f9155682fe8946a97c08896109508')
+ version('1.11.2', 'e4ca55c2a5c5656da4a4e37c8fc51b23')
+ version('1.11.1', 'feb4e416a1b25963ed565d8b42252fdc')
+ version('1.9', '1f9f9155682fe8946a97c08896109508')
depends_on('libpciaccess')
+ def url_for_version(self, version):
+ return "http://www.open-mpi.org/software/hwloc/v%s/downloads/hwloc-%s.tar.gz" % (version.up_to(2), version)
+
def install(self, spec, prefix):
configure("--prefix=%s" % prefix)
diff --git a/var/spack/repos/builtin/packages/libedit/package.py b/var/spack/repos/builtin/packages/libedit/package.py
index bcd5212b9e..faed8bad37 100644
--- a/var/spack/repos/builtin/packages/libedit/package.py
+++ b/var/spack/repos/builtin/packages/libedit/package.py
@@ -7,6 +7,8 @@ class Libedit(Package):
version('3.1', '43cdb5df3061d78b5e9d59109871b4f6', url="http://thrysoee.dk/editline/libedit-20150325-3.1.tar.gz")
+ depends_on('ncurses')
+
def install(self, spec, prefix):
configure('--prefix=%s' % prefix)
diff --git a/var/spack/repos/builtin/packages/libgpg-error/package.py b/var/spack/repos/builtin/packages/libgpg-error/package.py
index 6c1d1a10a7..dd5fc2408a 100644
--- a/var/spack/repos/builtin/packages/libgpg-error/package.py
+++ b/var/spack/repos/builtin/packages/libgpg-error/package.py
@@ -9,6 +9,7 @@ class LibgpgError(Package):
homepage = "https://www.gnupg.org/related_software/libgpg-error"
url = "ftp://ftp.gnupg.org/gcrypt/libgpg-error/libgpg-error-1.18.tar.bz2"
+ version('1.21', 'ab0b5aba6d0a185b41d07bda804fd8b2')
version('1.18', '12312802d2065774b787cbfc22cc04e9')
def install(self, spec, prefix):
diff --git a/var/spack/repos/builtin/packages/m4/package.py b/var/spack/repos/builtin/packages/m4/package.py
new file mode 100644
index 0000000000..5d76d8866b
--- /dev/null
+++ b/var/spack/repos/builtin/packages/m4/package.py
@@ -0,0 +1,13 @@
+from spack import *
+
+class M4(Package):
+ """GNU M4 is an implementation of the traditional Unix macro processor."""
+ homepage = "https://www.gnu.org/software/m4/m4.html"
+ url = "ftp://ftp.gnu.org/gnu/m4/m4-1.4.17.tar.gz"
+
+ version('1.4.17', 'a5e9954b1dae036762f7b13673a2cf76')
+
+ def install(self, spec, prefix):
+ configure("--prefix=%s" % prefix)
+ make()
+ make("install")
diff --git a/var/spack/repos/builtin/packages/mpfr/package.py b/var/spack/repos/builtin/packages/mpfr/package.py
index 2758b0da2e..a1bd7529cf 100644
--- a/var/spack/repos/builtin/packages/mpfr/package.py
+++ b/var/spack/repos/builtin/packages/mpfr/package.py
@@ -33,7 +33,7 @@ class Mpfr(Package):
version('3.1.3', '5fdfa3cfa5c86514ee4a241a1affa138')
version('3.1.2', 'ee2c3ac63bf0c2359bf08fc3ee094c19')
- depends_on('gmp@4.1.0:')
+ depends_on('gmp')
def install(self, spec, prefix):
configure("--prefix=%s" % prefix)
diff --git a/var/spack/repos/builtin/packages/mpich/package.py b/var/spack/repos/builtin/packages/mpich/package.py
index 00b7dfda75..c517defa83 100644
--- a/var/spack/repos/builtin/packages/mpich/package.py
+++ b/var/spack/repos/builtin/packages/mpich/package.py
@@ -41,21 +41,30 @@ class Mpich(Package):
version('3.1', '5643dd176499bfb7d25079aaff25f2ec')
version('3.0.4', '9c5d5d4fe1e17dd12153f40bc5b6dbc0')
+ variant('verbs', default=False, description='Build support for OpenFabrics verbs.')
+
provides('mpi@:3.0', when='@3:')
provides('mpi@:1.3', when='@1:')
def setup_dependent_environment(self, module, spec, dep_spec):
"""For dependencies, make mpicc's use spack wrapper."""
- os.environ['MPICH_CC'] = 'cc'
- os.environ['MPICH_CXX'] = 'c++'
- os.environ['MPICH_F77'] = 'f77'
- os.environ['MPICH_F90'] = 'f90'
+ os.environ['MPICH_CC'] = os.environ['CC']
+ os.environ['MPICH_CXX'] = os.environ['CXX']
+ os.environ['MPICH_F77'] = os.environ['F77']
+ os.environ['MPICH_F90'] = os.environ['FC']
+ os.environ['MPICH_FC'] = os.environ['FC']
def install(self, spec, prefix):
config_args = ["--prefix=" + prefix,
"--enable-shared"]
+ # Variants
+ if '+verbs' in spec:
+ config_args.append("--with-ibverbs")
+ else:
+ config_args.append("--without-ibverbs")
+
# TODO: Spack should make it so that you can't actually find
# these compilers if they're "disabled" for the current
# compiler configuration.
diff --git a/var/spack/repos/builtin/packages/mumps/Makefile.inc b/var/spack/repos/builtin/packages/mumps/Makefile.inc
new file mode 100644
index 0000000000..2e6a041878
--- /dev/null
+++ b/var/spack/repos/builtin/packages/mumps/Makefile.inc
@@ -0,0 +1,38 @@
+LPORDDIR = $(topdir)/PORD/lib/
+IPORD = -I$(topdir)/PORD/include/
+LPORD = -L$(LPORDDIR) -lpord
+
+ORDERINGSC = $(ORDERINGSF)
+LORDERINGS = $(LMETIS) $(LPORD) $(LSCOTCH)
+IORDERINGSF = $(ISCOTCH)
+IORDERINGSC = $(IMETIS) $(IPORD) $(ISCOTCH)
+
+PLAT =
+LIBEXT = .a
+OUTC = -o
+OUTF = -o
+RM = /bin/rm -f
+AR = ar vr
+RANLIB = ranlib
+
+INCSEQ = -I$(topdir)/libseq
+LIBSEQ = -L$(topdir)/libseq -lmpiseq
+
+INCPAR =
+LIBPAR = $(SCALAP)
+
+LIBOTHERS = -lpthread
+
+#Sequential:
+ifeq ($(MUMPS_TYPE),seq)
+INCS = $(INCSEQ)
+LIBS = $(LIBSEQ)
+LIBSEQNEEDED = libseqneeded
+endif
+
+#Parallel:
+ifeq ($(MUMPS_TYPE),par)
+INCS = $(INCPAR)
+LIBS = $(LIBPAR)
+LIBSEQNEEDED =
+endif
diff --git a/var/spack/repos/builtin/packages/mumps/package.py b/var/spack/repos/builtin/packages/mumps/package.py
new file mode 100644
index 0000000000..44a37903cc
--- /dev/null
+++ b/var/spack/repos/builtin/packages/mumps/package.py
@@ -0,0 +1,139 @@
+from spack import *
+import os
+
+
+class Mumps(Package):
+ """MUMPS: a MUltifrontal Massively Parallel sparse direct Solver"""
+
+ homepage = "http://mumps.enseeiht.fr"
+ url = "http://mumps.enseeiht.fr/MUMPS_5.0.1.tar.gz"
+
+ version('5.0.1', 'b477573fdcc87babe861f62316833db0')
+
+ variant('mpi', default=True, description='Activate the compilation of MUMPS with the MPI support')
+ variant('scotch', default=False, description='Activate Scotch as a possible ordering library')
+ variant('ptscotch', default=False, description='Activate PT-Scotch as a possible ordering library')
+ variant('metis', default=False, description='Activate Metis as a possible ordering library')
+ variant('parmetis', default=False, description='Activate Parmetis as a possible ordering library')
+ variant('double', default=True, description='Activate the compilation of dmumps')
+ variant('float', default=True, description='Activate the compilation of smumps')
+ variant('complex', default=True, description='Activate the compilation of cmumps and/or zmumps')
+ variant('idx64', default=False, description='Use int64_t/integer*8 as default index type')
+
+
+ depends_on('scotch + esmumps', when='~ptscotch+scotch')
+ depends_on('scotch + esmumps + mpi', when='+ptscotch')
+ depends_on('metis', when='~parmetis+metis')
+ depends_on('parmetis', when="+parmetis")
+ depends_on('blas')
+ depends_on('lapack')
+ depends_on('scalapack', when='+mpi')
+ depends_on('mpi', when='+mpi')
+
+ # this function is not a patch function because in case scalapack
+ # is needed it uses self.spec['scalapack'].fc_link set by the
+ # setup_dependent_environment in scalapck. This happen after patch
+ # end before install
+ # def patch(self):
+ def write_makefile_inc(self):
+ if ('+parmetis' in self.spec or '+ptscotch' in self.spec) and '+mpi' not in self.spec:
+ raise RuntimeError('You cannot use the variants parmetis or ptscotch without mpi')
+
+ makefile_conf = ["LIBBLAS = -L%s -lblas" % self.spec['blas'].prefix.lib]
+
+ orderings = ['-Dpord']
+
+ if '+ptscotch' in self.spec or '+scotch' in self.spec:
+ join_lib = ' -l%s' % ('pt' if '+ptscotch' in self.spec else '')
+ makefile_conf.extend(
+ ["ISCOTCH = -I%s" % self.spec['scotch'].prefix.include,
+ "LSCOTCH = -L%s %s%s" % (self.spec['scotch'].prefix.lib,
+ join_lib,
+ join_lib.join(['esmumps', 'scotch', 'scotcherr']))])
+ orderings.append('-Dscotch')
+ if '+ptscotch' in self.spec:
+ orderings.append('-Dptscotch')
+
+ if '+parmetis' in self.spec or '+metis' in self.spec:
+ libname = 'parmetis' if '+parmetis' in self.spec else 'metis'
+ makefile_conf.extend(
+ ["IMETIS = -I%s" % self.spec[libname].prefix.include,
+ "LMETIS = -L%s -l%s" % (self.spec[libname].prefix.lib, libname)])
+
+ orderings.append('-Dmetis')
+ if '+parmetis' in self.spec:
+ orderings.append('-Dparmetis')
+
+ makefile_conf.append("ORDERINGSF = %s" % (' '.join(orderings)))
+
+ # TODO: test this part, it needs a full blas, scalapack and
+ # partitionning environment with 64bit integers
+ if '+idx64' in self.spec:
+ makefile_conf.extend(
+ # the fortran compilation flags most probably are
+ # working only for intel and gnu compilers this is
+ # perhaps something the compiler should provide
+ ['OPTF = -O -DALLOW_NON_INIT %s' % '-fdefault-integer-8' if self.compiler.name == "gcc" else '-i8',
+ 'OPTL = -O ',
+ 'OPTC = -O -DINTSIZE64'])
+ else:
+ makefile_conf.extend(
+ ['OPTF = -O -DALLOW_NON_INIT',
+ 'OPTL = -O ',
+ 'OPTC = -O '])
+
+
+ if '+mpi' in self.spec:
+ makefile_conf.extend(
+ ["CC = %s" % join_path(self.spec['mpi'].prefix.bin, 'mpicc'),
+ "FC = %s" % join_path(self.spec['mpi'].prefix.bin, 'mpif90'),
+ "FL = %s" % join_path(self.spec['mpi'].prefix.bin, 'mpif90'),
+ "SCALAP = %s" % self.spec['scalapack'].fc_link,
+ "MUMPS_TYPE = par"])
+ else:
+ makefile_conf.extend(
+ ["CC = cc",
+ "FC = fc",
+ "FL = fc",
+ "MUMPS_TYPE = seq"])
+
+ # TODO: change the value to the correct one according to the
+ # compiler possible values are -DAdd_, -DAdd__ and/or -DUPPER
+ makefile_conf.append("CDEFS = -DAdd_")
+
+
+ makefile_inc_template = join_path(os.path.dirname(self.module.__file__),
+ 'Makefile.inc')
+ with open(makefile_inc_template, "r") as fh:
+ makefile_conf.extend(fh.read().split('\n'))
+
+ with working_dir('.'):
+ with open("Makefile.inc", "w") as fh:
+ makefile_inc = '\n'.join(makefile_conf)
+ fh.write(makefile_inc)
+
+
+
+ def install(self, spec, prefix):
+ make_libs = []
+
+ # the coice to compile ?examples is to have kind of a sanity
+ # check on the libraries generated.
+ if '+float' in spec:
+ make_libs.append('sexamples')
+ if '+complex' in spec:
+ make_libs.append('cexamples')
+
+ if '+double' in spec:
+ make_libs.append('dexamples')
+ if '+complex' in spec:
+ make_libs.append('zexamples')
+
+ self.write_makefile_inc()
+
+ make(*make_libs)
+
+ install_tree('lib', prefix.lib)
+ install_tree('include', prefix.include)
+ if '~mpi' in spec:
+ install('libseq/libmpiseq.a', prefix.lib)
diff --git a/var/spack/repos/builtin/packages/mvapich2/package.py b/var/spack/repos/builtin/packages/mvapich2/package.py
index 23a11b3171..af5ed1b088 100644
--- a/var/spack/repos/builtin/packages/mvapich2/package.py
+++ b/var/spack/repos/builtin/packages/mvapich2/package.py
@@ -4,15 +4,13 @@ import os
class Mvapich2(Package):
"""MVAPICH2 is an MPI implementation for Infiniband networks."""
homepage = "http://mvapich.cse.ohio-state.edu/"
+ url = "http://mvapich.cse.ohio-state.edu/download/mvapich/mv2/mvapich2-2.2b.tar.gz"
- version('2.2a', 'b8ceb4fc5f5a97add9b3ff1b9cbe39d2',
- url='http://mvapich.cse.ohio-state.edu/download/mvapich/mv2/mvapich2-2.2a.tar.gz')
+ version('2.2b', '5651e8b7a72d7c77ca68da48f3a5d108')
+ version('2.2a', 'b8ceb4fc5f5a97add9b3ff1b9cbe39d2')
+ version('2.0', '9fbb68a4111a8b6338e476dc657388b4')
+ version('1.9', '5dc58ed08fd3142c260b70fe297e127c')
- version('2.0', '9fbb68a4111a8b6338e476dc657388b4',
- url='http://mvapich.cse.ohio-state.edu/download/mvapich/mv2/mvapich2-2.0.tar.gz')
-
- version('1.9', '5dc58ed08fd3142c260b70fe297e127c',
- url="http://mvapich.cse.ohio-state.edu/download/mvapich2/mv2/mvapich2-1.9.tgz")
patch('ad_lustre_rwcontig_open_source.patch', when='@1.9')
provides('mpi@:2.2', when='@1.9') # MVAPICH2-1.9 supports MPI 2.2
@@ -41,16 +39,25 @@ class Mvapich2(Package):
NEMESISIBTCP = 'nemesisibtcp'
NEMESISIB = 'nemesisib'
NEMESIS = 'nemesis'
+ MRAIL = 'mrail'
SUPPORTED_NETWORKS = (PSM, SOCK, NEMESIS, NEMESISIB, NEMESISIBTCP)
variant(PSM, default=False, description='Configures a build for QLogic PSM-CH3')
variant(SOCK, default=False, description='Configures a build for TCP/IP-CH3')
variant(NEMESISIBTCP, default=False, description='Configures a build for both OFA-IB-Nemesis and TCP/IP-Nemesis')
variant(NEMESISIB, default=False, description='Configures a build for OFA-IB-Nemesis')
variant(NEMESIS, default=False, description='Configures a build for TCP/IP-Nemesis')
+ variant(MRAIL, default=False, description='Configures a build for OFA-IB-CH3')
##########
# FIXME : CUDA support is missing
+ def url_for_version(self, version):
+ base_url = "http://mvapich.cse.ohio-state.edu/download"
+ if version < Version('2.0'):
+ return "%s/mvapich2/mv2/mvapich2-%s.tar.gz" % (base_url, version)
+ else:
+ return "%s/mvapich/mv2/mvapich2-%s.tar.gz" % (base_url, version)
+
@staticmethod
def enabled(x):
"""
@@ -117,7 +124,7 @@ class Mvapich2(Package):
if count > 1:
raise RuntimeError('network variants are mutually exclusive (only one can be selected at a time)')
- # From here on I can suppose that ony one variant has been selected
+ # From here on I can suppose that only one variant has been selected
if self.enabled(Mvapich2.PSM) in spec:
network_options = ["--with-device=ch3:psm"]
elif self.enabled(Mvapich2.SOCK) in spec:
@@ -128,7 +135,7 @@ class Mvapich2(Package):
network_options = ["--with-device=ch3:nemesis:ib"]
elif self.enabled(Mvapich2.NEMESIS) in spec:
network_options = ["--with-device=ch3:nemesis"]
- else:
+ elif self.enabled(Mvapich2.MRAIL) in spec:
network_options = ["--with-device=ch3:mrail", "--with-rdma=gen2"]
configure_args.extend(network_options)
@@ -141,7 +148,14 @@ class Mvapich2(Package):
"--enable-romio",
"--disable-silent-rules",
]
- if not self.compiler.f77 and not self.compiler.fc:
+
+ if self.compiler.f77 and self.compiler.fc:
+ configure_args.append("--enable-fortran=all")
+ elif self.compiler.f77:
+ configure_args.append("--enable-fortran=f77")
+ elif self.compiler.fc:
+ configure_args.append("--enable-fortran=fc")
+ else:
configure_args.append("--enable-fortran=none")
# Set the type of the build (debug, release)
diff --git a/var/spack/repos/builtin/packages/netcdf/netcdf-4.3.3-mpi.patch b/var/spack/repos/builtin/packages/netcdf/netcdf-4.3.3-mpi.patch
deleted file mode 100644
index 46dda5fc9d..0000000000
--- a/var/spack/repos/builtin/packages/netcdf/netcdf-4.3.3-mpi.patch
+++ /dev/null
@@ -1,25 +0,0 @@
-diff -Nur netcdf-4.3.3/CMakeLists.txt netcdf-4.3.3.mpi/CMakeLists.txt
---- netcdf-4.3.3/CMakeLists.txt 2015-02-12 16:44:35.000000000 -0500
-+++ netcdf-4.3.3.mpi/CMakeLists.txt 2015-10-14 16:44:41.176300658 -0400
-@@ -753,6 +753,7 @@
- SET(USE_PARALLEL OFF CACHE BOOL "")
- MESSAGE(STATUS "Cannot find HDF5 library built with parallel support. Disabling parallel build.")
- ELSE()
-+ FIND_PACKAGE(MPI REQUIRED)
- SET(USE_PARALLEL ON CACHE BOOL "")
- SET(STATUS_PARALLEL "ON")
- ENDIF()
-diff -Nur netcdf-4.3.3/liblib/CMakeLists.txt netcdf-4.3.3.mpi/liblib/CMakeLists.txt
---- netcdf-4.3.3/liblib/CMakeLists.txt 2015-02-12 16:44:35.000000000 -0500
-+++ netcdf-4.3.3.mpi/liblib/CMakeLists.txt 2015-10-14 16:44:57.757793634 -0400
-@@ -71,6 +71,10 @@
- SET(TLL_LIBS ${TLL_LIBS} ${CURL_LIBRARY})
- ENDIF()
-
-+IF(USE_PARALLEL)
-+ SET(TLL_LIBS ${TLL_LIBS} ${MPI_C_LIBRARIES})
-+ENDIF()
-+
- IF(USE_HDF4)
- SET(TLL_LIBS ${TLL_LIBS} ${HDF4_LIBRARIES})
- ENDIF()
diff --git a/var/spack/repos/builtin/packages/netcdf/package.py b/var/spack/repos/builtin/packages/netcdf/package.py
index 239644d894..41a0d2b6f9 100644
--- a/var/spack/repos/builtin/packages/netcdf/package.py
+++ b/var/spack/repos/builtin/packages/netcdf/package.py
@@ -1,28 +1,86 @@
from spack import *
+
class Netcdf(Package):
"""NetCDF is a set of software libraries and self-describing, machine-independent
- data formats that support the creation, access, and sharing of array-oriented
- scientific data."""
+ data formats that support the creation, access, and sharing of array-oriented
+ scientific data."""
homepage = "http://www.unidata.ucar.edu/software/netcdf/"
url = "ftp://ftp.unidata.ucar.edu/pub/netcdf/netcdf-4.3.3.tar.gz"
+ version('4.4.0', 'cffda0cbd97fdb3a06e9274f7aef438e')
version('4.3.3', '5fbd0e108a54bd82cb5702a73f56d2ae')
- patch('netcdf-4.3.3-mpi.patch')
+ variant('mpi', default=True, description='Enables MPI parallelism')
+ variant('fortran', default=False, description="Download and install NetCDF-Fortran")
+ variant('hdf4', default=False, description="Enable HDF4 support")
# Dependencies:
- depends_on("cmake @2.8.12:")
- # >HDF5
- depends_on("hdf5")
+ depends_on("curl") # required for DAP support
+ depends_on("hdf", when='+hdf4')
+ depends_on("hdf5+mpi~cxx", when='+mpi') # required for NetCDF-4 support
+ depends_on("hdf5~mpi", when='~mpi') # required for NetCDF-4 support
+ depends_on("zlib") # required for NetCDF-4 support
def install(self, spec, prefix):
- with working_dir('spack-build', create=True):
- cmake('..',
- "-DCMAKE_INSTALL_PREFIX:PATH=%s" % prefix,
- "-DENABLE_DAP:BOOL=OFF", # Disable DAP.
- "-DBUILD_SHARED_LIBS:BOOL=OFF") # Don't build shared libraries (use static libs).
-
- make()
- make("install")
+ # Environment variables
+ CPPFLAGS = []
+ LDFLAGS = []
+ LIBS = []
+
+ config_args = [
+ "--prefix=%s" % prefix,
+ "--enable-fsync",
+ "--enable-v2",
+ "--enable-utilities",
+ "--enable-shared",
+ "--enable-static",
+ "--enable-largefile",
+ # necessary for HDF5 support
+ "--enable-netcdf-4",
+ "--enable-dynamic-loading",
+ # necessary for DAP support
+ "--enable-dap"
+ ]
+
+ if '+mpi' in spec:
+ config_args.append('--enable-parallel4')
+
+ CPPFLAGS.append("-I%s/include" % spec['hdf5'].prefix)
+ LDFLAGS.append( "-L%s/lib" % spec['hdf5'].prefix)
+
+ # HDF4 support
+ # As of NetCDF 4.1.3, "--with-hdf4=..." is no longer a valid option
+ # You must use the environment variables CPPFLAGS and LDFLAGS
+ if '+hdf4' in spec:
+ config_args.append("--enable-hdf4")
+ CPPFLAGS.append("-I%s/include" % spec['hdf'].prefix)
+ LDFLAGS.append( "-L%s/lib" % spec['hdf'].prefix)
+ LIBS.append( "-l%s" % "jpeg")
+
+ if 'szip' in spec:
+ CPPFLAGS.append("-I%s/include" % spec['szip'].prefix)
+ LDFLAGS.append( "-L%s/lib" % spec['szip'].prefix)
+ LIBS.append( "-l%s" % "sz")
+
+ # Fortran support
+ # In version 4.2+, NetCDF-C and NetCDF-Fortran have split.
+ # They can be installed separately, but this bootstrap procedure
+ # should be able to install both at the same time.
+ # Note: this is a new experimental feature.
+ if '+fortran' in spec:
+ config_args.append("--enable-remote-fortran-bootstrap")
+
+ config_args.append('CPPFLAGS=%s' % ' '.join(CPPFLAGS))
+ config_args.append('LDFLAGS=%s' % ' '.join(LDFLAGS))
+ config_args.append('LIBS=%s' % ' '.join(LIBS))
+
+ configure(*config_args)
+ make()
+ make("install")
+
+ # After installing NetCDF-C, install NetCDF-Fortran
+ if '+fortran' in spec:
+ make("build-netcdf-fortran")
+ make("install-netcdf-fortran")
diff --git a/var/spack/repos/builtin/packages/netlib-scalapack/package.py b/var/spack/repos/builtin/packages/netlib-scalapack/package.py
new file mode 100644
index 0000000000..22d538560e
--- /dev/null
+++ b/var/spack/repos/builtin/packages/netlib-scalapack/package.py
@@ -0,0 +1,50 @@
+from spack import *
+
+class NetlibScalapack(Package):
+ """ScaLAPACK is a library of high-performance linear algebra routines for parallel distributed memory machines"""
+
+ homepage = "http://www.netlib.org/scalapack/"
+ url = "http://www.netlib.org/scalapack/scalapack-2.0.2.tgz"
+
+ version('2.0.2', '2f75e600a2ba155ed9ce974a1c4b536f')
+ version('2.0.1', '17b8cde589ea0423afe1ec43e7499161')
+ version('2.0.0', '9e76ae7b291be27faaad47cfc256cbfe')
+ # versions before 2.0.0 are not using cmake and requires blacs as
+ # a separated package
+
+ variant('shared', default=True, description='Build the shared library version')
+ variant('fpic', default=False, description="Build with -fpic compiler option")
+
+ provides('scalapack')
+
+ depends_on('mpi')
+ depends_on('lapack')
+
+ def install(self, spec, prefix):
+ options = [
+ "-DBUILD_SHARED_LIBS:BOOL=%s" % ('ON' if '+shared' in spec else 'OFF'),
+ "-DBUILD_STATIC_LIBS:BOOL=%s" % ('OFF' if '+shared' in spec else 'ON'),
+ "-DUSE_OPTIMIZED_LAPACK_BLAS:BOOL=ON", # forces scalapack to use find_package(LAPACK)
+ ]
+
+ if '+fpic' in spec:
+ options.extend([
+ "-DCMAKE_C_FLAGS=-fPIC",
+ "-DCMAKE_Fortran_FLAGS=-fPIC"
+ ])
+
+ options.extend(std_cmake_args)
+
+ with working_dir('spack-build', create=True):
+ cmake('..', *options)
+ make()
+ make("install")
+
+ def setup_dependent_environment(self, module, spec, dependent_spec):
+ # TODO treat OS that are not Linux...
+ lib_suffix = '.so' if '+shared' in spec['scalapack'] else '.a'
+
+ spec['scalapack'].fc_link = '-L%s -lscalapack' % spec['scalapack'].prefix.lib
+ spec['scalapack'].cc_link = spec['scalapack'].fc_link
+ spec['scalapack'].libraries = [join_path(spec['scalapack'].prefix.lib,
+ 'libscalapack%s' % lib_suffix)]
diff --git a/var/spack/repos/builtin/packages/openblas/package.py b/var/spack/repos/builtin/packages/openblas/package.py
index 9c8fa1c694..3c909360a4 100644
--- a/var/spack/repos/builtin/packages/openblas/package.py
+++ b/var/spack/repos/builtin/packages/openblas/package.py
@@ -19,7 +19,9 @@ class Openblas(Package):
with working_dir(prefix.lib):
symlink('libopenblas.a', 'blas.a')
symlink('libopenblas.a', 'libblas.a')
+ symlink('libopenblas.so', 'libblas.so')
# Lapack virtual package should provide liblapack.a
with working_dir(prefix.lib):
symlink('libopenblas.a', 'liblapack.a')
+ symlink('libopenblas.so', 'liblapack.so')
diff --git a/var/spack/repos/builtin/packages/opencv/package.py b/var/spack/repos/builtin/packages/opencv/package.py
new file mode 100644
index 0000000000..99b555323f
--- /dev/null
+++ b/var/spack/repos/builtin/packages/opencv/package.py
@@ -0,0 +1,50 @@
+from spack import *
+
+
+class Opencv(Package):
+ """
+ OpenCV is released under a BSD license and hence it's free for both academic and commercial use. It has C++, C,
+ Python and Java interfaces and supports Windows, Linux, Mac OS, iOS and Android. OpenCV was designed for
+ computational efficiency and with a strong focus on real-time applications. Written in optimized C/C++, the library
+ can take advantage of multi-core processing. Enabled with OpenCL, it can take advantage of the hardware
+ acceleration of the underlying heterogeneous compute platform. Adopted all around the world, OpenCV has more than
+ 47 thousand people of user community and estimated number of downloads exceeding 9 million. Usage ranges from
+ interactive art, to mines inspection, stitching maps on the web or through advanced robotics.
+ """
+ homepage = 'http://opencv.org/'
+ url = 'https://github.com/Itseez/opencv/archive/3.1.0.tar.gz'
+
+ version('3.1.0', '70e1dd07f0aa06606f1bc0e3fa15abd3')
+
+ variant('shared', default=True, description='Enables the build of shared libraries')
+ variant('debug', default=False, description='Builds a debug version of the libraries')
+
+ variant('eigen', default=True, description='Activates support for eigen')
+ variant('ipp', default=True, description='Activates support for IPP')
+
+ depends_on('zlib')
+ depends_on('libpng')
+ depends_on('libjpeg-turbo')
+ depends_on('libtiff')
+
+ depends_on('python')
+ depends_on('py-numpy')
+
+ depends_on('eigen', when='+eigen')
+
+ # FIXME : GUI extensions missing
+ # FIXME : CUDA extensions still missing
+
+ def install(self, spec, prefix):
+ cmake_options = []
+ cmake_options.extend(std_cmake_args)
+
+ cmake_options.extend(['-DCMAKE_BUILD_TYPE:STRING=%s' % ('Debug' if '+debug' in spec else 'Release'),
+ '-DBUILD_SHARED_LIBS:BOOL=%s' % ('ON' if '+shared' in spec else 'OFF'),
+ '-DENABLE_PRECOMPILED_HEADERS:BOOL=OFF',
+ '-DWITH_IPP:BOOL=%s' % ('ON' if '+ipp' in spec else 'OFF')])
+
+ with working_dir('spack_build', create=True):
+ cmake('..', *cmake_options)
+ make('VERBOSE=1')
+ make("install")
diff --git a/var/spack/repos/builtin/packages/openmpi/package.py b/var/spack/repos/builtin/packages/openmpi/package.py
index 463719f9db..e4484af8c5 100644
--- a/var/spack/repos/builtin/packages/openmpi/package.py
+++ b/var/spack/repos/builtin/packages/openmpi/package.py
@@ -17,45 +17,47 @@ class Openmpi(Package):
list_url = "http://www.open-mpi.org/software/ompi/"
list_depth = 3
+ version('1.10.2', 'b2f43d9635d2d52826e5ef9feb97fd4c')
version('1.10.1', 'f0fcd77ed345b7eafb431968124ba16e')
version('1.10.0', '280cf952de68369cebaca886c5ce0304')
- version('1.8.8', '0dab8e602372da1425e9242ae37faf8c')
- version('1.6.5', '03aed2a4aa4d0b27196962a2a65fc475')
+ version('1.8.8', '0dab8e602372da1425e9242ae37faf8c')
+ version('1.6.5', '03aed2a4aa4d0b27196962a2a65fc475')
patch('ad_lustre_rwcontig_open_source.patch', when="@1.6.5")
patch('llnl-platforms.patch', when="@1.6.5")
- patch('configure.patch', when="@1.10.0:")
+ patch('configure.patch', when="@1.10.0:1.10.1")
- variant('psm', default=False, description='Build support for the PSM library.')
+ variant('psm', default=False, description='Build support for the PSM library.')
variant('verbs', default=False, description='Build support for OpenFabrics verbs.')
+ # TODO : variant support for other schedulers is missing
+ variant('tm', default=False, description='Build TM (Torque, PBSPro, and compatible) support')
+
provides('mpi@:2.2', when='@1.6.5')
provides('mpi@:3.0', when='@1.7.5:')
-
depends_on('hwloc')
-
def url_for_version(self, version):
return "http://www.open-mpi.org/software/ompi/v%s/downloads/openmpi-%s.tar.bz2" % (version.up_to(2), version)
-
def setup_dependent_environment(self, module, spec, dep_spec):
"""For dependencies, make mpicc's use spack wrapper."""
- os.environ['OMPI_CC'] = 'cc'
+ os.environ['OMPI_CC'] = 'cc'
os.environ['OMPI_CXX'] = 'c++'
os.environ['OMPI_FC'] = 'f90'
os.environ['OMPI_F77'] = 'f77'
-
def install(self, spec, prefix):
config_args = ["--prefix=%s" % prefix,
"--with-hwloc=%s" % spec['hwloc'].prefix,
- "--with-tm", # necessary for Torque support
"--enable-shared",
"--enable-static"]
# Variants
+ if '+tm' in spec:
+ config_args.append("--with-tm") # necessary for Torque support
+
if '+psm' in spec:
config_args.append("--with-psm")
@@ -85,7 +87,6 @@ class Openmpi(Package):
self.filter_compilers()
-
def filter_compilers(self):
"""Run after install to make the MPI compilers use the
compilers that Spack built the package with.
@@ -94,7 +95,7 @@ class Openmpi(Package):
to Spack's generic cc, c++ and f90. We want them to
be bound to whatever compiler they were built with.
"""
- kwargs = { 'ignore_absent' : True, 'backup' : False, 'string' : False }
+ kwargs = {'ignore_absent': True, 'backup': False, 'string': False}
dir = os.path.join(self.prefix, 'share/openmpi/')
cc_wrappers = ['mpicc-vt-wrapper-data.txt', 'mpicc-wrapper-data.txt',
@@ -132,5 +133,3 @@ class Openmpi(Package):
if not os.path.islink(path):
filter_file('compiler=.*', 'compiler=%s' % self.compiler.fc,
path, **kwargs)
-
-
diff --git a/var/spack/repos/builtin/packages/openssl/package.py b/var/spack/repos/builtin/packages/openssl/package.py
index bbb169ec6b..a225e30f6c 100644
--- a/var/spack/repos/builtin/packages/openssl/package.py
+++ b/var/spack/repos/builtin/packages/openssl/package.py
@@ -10,8 +10,10 @@ class Openssl(Package):
url = "http://www.openssl.org/source/openssl-1.0.1h.tar.gz"
version('1.0.1h', '8d6d684a9430d5cc98a62a5d8fbda8cf')
+ version('1.0.1r', '1abd905e079542ccae948af37e393d28')
version('1.0.2d', '38dd619b2e77cbac69b99f52a053d25a')
version('1.0.2e', '5262bfa25b60ed9de9f28d5d52d77fc5')
+ version('1.0.2f', 'b3bf73f507172be9292ea2a8c28b659d')
depends_on("zlib")
parallel = False
diff --git a/var/spack/repos/builtin/packages/parallel-netcdf/package.py b/var/spack/repos/builtin/packages/parallel-netcdf/package.py
new file mode 100644
index 0000000000..62a8f7ca0b
--- /dev/null
+++ b/var/spack/repos/builtin/packages/parallel-netcdf/package.py
@@ -0,0 +1,20 @@
+from spack import *
+
+class ParallelNetcdf(Package):
+ """Parallel netCDF (PnetCDF) is a library providing high-performance
+ parallel I/O while still maintaining file-format compatibility with
+ Unidata's NetCDF."""
+
+ homepage = "https://trac.mcs.anl.gov/projects/parallel-netcdf"
+ url = "http://cucis.ece.northwestern.edu/projects/PnetCDF/Release/parallel-netcdf-1.6.1.tar.gz"
+
+ version('1.6.1', '62a094eb952f9d1e15f07d56e535052604f1ac34')
+
+ depends_on("m4")
+ depends_on("mpi")
+
+ def install(self, spec, prefix):
+ configure("--prefix=%s" % prefix,
+ "--with-mpi=%s" % spec['mpi'].prefix)
+ make()
+ make("install")
diff --git a/var/spack/repos/builtin/packages/paraview/package.py b/var/spack/repos/builtin/packages/paraview/package.py
index aaab352e66..e43bdd4493 100644
--- a/var/spack/repos/builtin/packages/paraview/package.py
+++ b/var/spack/repos/builtin/packages/paraview/package.py
@@ -14,6 +14,7 @@ class Paraview(Package):
variant('osmesa', default=False, description='Enable OSMesa support')
variant('qt', default=False, description='Enable Qt support')
+ variant('opengl2', default=False, description='Enable OPengl2 backend')
depends_on('python', when='+python')
depends_on('py-numpy', when='+python')
diff --git a/var/spack/repos/builtin/packages/py-mpi4py/package.py b/var/spack/repos/builtin/packages/py-mpi4py/package.py
index 8001689a18..f599205644 100644
--- a/var/spack/repos/builtin/packages/py-mpi4py/package.py
+++ b/var/spack/repos/builtin/packages/py-mpi4py/package.py
@@ -5,7 +5,9 @@ class PyMpi4py(Package):
homepage = "https://pypi.python.org/pypi/mpi4py"
url = "https://pypi.python.org/packages/source/m/mpi4py/mpi4py-1.3.1.tar.gz"
+ version('2.0.0', '4f7d8126d7367c239fd67615680990e3')
version('1.3.1', 'dbe9d22bdc8ed965c23a7ceb6f32fc3c')
+
extends('python')
depends_on('py-setuptools')
depends_on('mpi')
diff --git a/var/spack/repos/builtin/packages/py-phonopy/package.py b/var/spack/repos/builtin/packages/py-phonopy/package.py
new file mode 100644
index 0000000000..6d10fea74f
--- /dev/null
+++ b/var/spack/repos/builtin/packages/py-phonopy/package.py
@@ -0,0 +1,18 @@
+from spack import *
+
+class PyPhonopy(Package):
+ """Phonopy is an open source package for phonon
+ calculations at harmonic and quasi-harmonic levels."""
+ homepage = "http://atztogo.github.io/phonopy/index.html"
+ url = "http://sourceforge.net/projects/phonopy/files/phonopy/phonopy-1.10/phonopy-1.10.0.tar.gz"
+
+ version('1.10.0', '973ed1bcea46e21b9bf747aab9061ff6')
+
+ extends('python')
+ depends_on('py-numpy')
+ depends_on('py-scipy')
+ depends_on('py-matplotlib')
+ depends_on('py-pyyaml')
+
+ def install(self, spec, prefix):
+ python('setup.py', 'install', '--home=%s' % prefix)
diff --git a/var/spack/repos/builtin/packages/py-pyyaml/package.py b/var/spack/repos/builtin/packages/py-pyyaml/package.py
new file mode 100644
index 0000000000..cae42f6e59
--- /dev/null
+++ b/var/spack/repos/builtin/packages/py-pyyaml/package.py
@@ -0,0 +1,13 @@
+from spack import *
+
+class PyPyyaml(Package):
+ """PyYAML is a YAML parser and emitter for Python."""
+ homepage = "http://pyyaml.org/wiki/PyYAML"
+ url = "http://pyyaml.org/download/pyyaml/PyYAML-3.11.tar.gz"
+
+ version('3.11', 'f50e08ef0fe55178479d3a618efe21db')
+
+ extends('python')
+
+ def install(self, spec, prefix):
+ python('setup.py', 'install', '--prefix=%s' % prefix)
diff --git a/var/spack/repos/builtin/packages/py-wheel/package.py b/var/spack/repos/builtin/packages/py-wheel/package.py
new file mode 100644
index 0000000000..3118e74519
--- /dev/null
+++ b/var/spack/repos/builtin/packages/py-wheel/package.py
@@ -0,0 +1,15 @@
+from spack import *
+
+class PyWheel(Package):
+ """A built-package format for Python."""
+
+ homepage = "https://pypi.python.org/pypi/wheel"
+ url = "https://pypi.python.org/packages/source/w/wheel/wheel-0.26.0.tar.gz"
+
+ version('0.26.0', '4cfc6e7e3dc7377d0164914623922a10')
+
+ extends('python')
+ depends_on('py-setuptools')
+
+ def install(self, spec, prefix):
+ python('setup.py', 'install', '--prefix=%s' % prefix)
diff --git a/var/spack/repos/builtin/packages/qhull/package.py b/var/spack/repos/builtin/packages/qhull/package.py
index f6712ced38..8f7c2f31b1 100644
--- a/var/spack/repos/builtin/packages/qhull/package.py
+++ b/var/spack/repos/builtin/packages/qhull/package.py
@@ -8,20 +8,20 @@ class Qhull(Package):
implements the Quickhull algorithm for computing the convex
hull. It handles roundoff errors from floating point
arithmetic. It computes volumes, surface areas, and
- approximations to the convex hull.
-
- Qhull does not support triangulation of non-convex surfaces,
- mesh generation of non-convex objects, medium-sized inputs in
- 9-D and higher, alpha shapes, weighted Voronoi diagrams,
- Voronoi volumes, or constrained Delaunay triangulations."""
+ approximations to the convex hull."""
homepage = "http://www.qhull.org"
+ version('7.2.0', 'e6270733a826a6a7c32b796e005ec3dc',
+ url="http://www.qhull.org/download/qhull-2015-src-7.2.0.tgz")
+
version('1.0', 'd0f978c0d8dfb2e919caefa56ea2953c',
url="http://www.qhull.org/download/qhull-2012.1-src.tgz")
# https://github.com/qhull/qhull/pull/5
- patch('qhull-iterator.patch')
+ patch('qhull-iterator.patch', when='@1.0')
+
+ depends_on('cmake')
def install(self, spec, prefix):
with working_dir('spack-build', create=True):
diff --git a/var/spack/repos/builtin/packages/qt/package.py b/var/spack/repos/builtin/packages/qt/package.py
index e8d843519d..91afa420c1 100644
--- a/var/spack/repos/builtin/packages/qt/package.py
+++ b/var/spack/repos/builtin/packages/qt/package.py
@@ -23,6 +23,7 @@ class Qt(Package):
version('3.3.8b', '9f05b4125cfe477cc52c9742c3c09009',
url="http://download.qt.io/archive/qt/3/qt-x11-free-3.3.8b.tar.gz")
+ variant('mesa', default=False, description='depend on mesa')
# Add patch for compile issues with qt3 found with use in the OpenSpeedShop project
variant('krellpatch', default=False, description="build with openspeedshop based patch.")
patch('qt3krell.patch', when='@3.3.8b+krellpatch')
@@ -48,7 +49,7 @@ class Qt(Package):
# depends_on("icu4c")
# OpenGL hardware acceleration
- depends_on("mesa", when='@4:')
+ depends_on("mesa", when='@4:+mesa')
depends_on("libxcb")
diff --git a/var/spack/repos/builtin/packages/trilinos/package.py b/var/spack/repos/builtin/packages/trilinos/package.py
index 7c43f796a4..edc40476e3 100644
--- a/var/spack/repos/builtin/packages/trilinos/package.py
+++ b/var/spack/repos/builtin/packages/trilinos/package.py
@@ -10,40 +10,44 @@ class Trilinos(Package):
homepage = "https://trilinos.org/"
url = "http://trilinos.csbsju.edu/download/files/trilinos-12.2.1-Source.tar.gz"
+ version('12.4.2', '7c830f7f0f68b8ad324690603baf404e')
version('12.2.1', '6161926ea247863c690e927687f83be9')
version('12.0.1', 'bd99741d047471e127b8296b2ec08017')
version('11.14.3', '2f4f83f8333e4233c57d0f01c4b57426')
version('11.14.2', 'a43590cf896c677890d75bfe75bc6254')
version('11.14.1', '40febc57f76668be8b6a77b7607bb67f')
- variant('mpi', default=True, description='Add a dependency on MPI and enables MPI dependent packages')
+ variant('shared', default=True, description='Enables the build of shared libraries')
+ variant('debug', default=False, description='Builds a debug version of the libraries')
# Everything should be compiled with -fpic
depends_on('blas')
depends_on('lapack')
depends_on('boost')
- depends_on('netcdf')
depends_on('matio')
depends_on('glm')
depends_on('swig')
- depends_on('mpi', when='+mpi')
- def install(self, spec, prefix):
+ # MPI related dependencies
+ depends_on('mpi')
+ depends_on('netcdf+mpi')
+
+ depends_on('python') # Needs py-numpy activated
- options = [
- '-DTrilinos_ENABLE_ALL_PACKAGES:BOOL=ON',
- '-DTrilinos_ENABLE_TESTS:BOOL=OFF',
- '-DTrilinos_ENABLE_EXAMPLES:BOOL=OFF',
- '-DBUILD_SHARED_LIBS:BOOL=ON',
- '-DBLAS_LIBRARY_DIRS:PATH=%s' % spec['blas'].prefix,
- '-DLAPACK_LIBRARY_DIRS:PATH=%s' % spec['lapack'].prefix
- ]
- if '+mpi' in spec:
- mpi_options = ['-DTPL_ENABLE_MPI:BOOL=ON']
- options.extend(mpi_options)
-
- # -DCMAKE_INSTALL_PREFIX and all the likes...
+ def install(self, spec, prefix):
+ options = []
options.extend(std_cmake_args)
+
+ options.extend(['-DTrilinos_ENABLE_ALL_PACKAGES:BOOL=ON',
+ '-DTrilinos_ENABLE_TESTS:BOOL=OFF',
+ '-DTrilinos_ENABLE_EXAMPLES:BOOL=OFF',
+ '-DCMAKE_BUILD_TYPE:STRING=%s' % ('Debug' if '+debug' in spec else 'Release'),
+ '-DBUILD_SHARED_LIBS:BOOL=%s' % ('ON' if '+shared' in spec else 'OFF'),
+ '-DTPL_ENABLE_MPI:STRING=ON',
+ '-DBLAS_LIBRARY_DIRS:PATH=%s' % spec['blas'].prefix,
+ '-DLAPACK_LIBRARY_DIRS:PATH=%s' % spec['lapack'].prefix
+ ])
+
with working_dir('spack-build', create=True):
cmake('..', *options)
make()