| # SPDX-License-Identifier: Apache-2.0 |
| # Copyright 2013-2021 The Meson development team |
| |
| from __future__ import annotations |
| |
| from .base import ExternalDependency, DependencyException, DependencyTypeName |
| from .pkgconfig import PkgConfigDependency |
| from ..mesonlib import (Popen_safe, OptionKey, join_args, version_compare) |
| from ..programs import ExternalProgram |
| from .. import mlog |
| import re |
| import os |
| import json |
| import typing as T |
| |
| if T.TYPE_CHECKING: |
| from ..environment import Environment |
| |
| |
| class DubDependency(ExternalDependency): |
| # dub program and version |
| class_dubbin: T.Optional[T.Tuple[ExternalProgram, str]] = None |
| class_dubbin_searched = False |
| |
| def __init__(self, name: str, environment: 'Environment', kwargs: T.Dict[str, T.Any]): |
| super().__init__(DependencyTypeName('dub'), environment, kwargs, language='d') |
| self.name = name |
| from ..compilers.d import DCompiler, d_feature_args |
| |
| _temp_comp = super().get_compiler() |
| assert isinstance(_temp_comp, DCompiler) |
| self.compiler = _temp_comp |
| |
| if 'required' in kwargs: |
| self.required = kwargs.get('required') |
| |
| if DubDependency.class_dubbin is None and not DubDependency.class_dubbin_searched: |
| DubDependency.class_dubbin = self._check_dub() |
| DubDependency.class_dubbin_searched = True |
| if DubDependency.class_dubbin is None: |
| if self.required: |
| raise DependencyException('DUB not found.') |
| self.is_found = False |
| return |
| |
| (self.dubbin, dubver) = DubDependency.class_dubbin # pylint: disable=unpacking-non-sequence |
| |
| assert isinstance(self.dubbin, ExternalProgram) |
| |
| # Check if Dub version is compatible with Meson |
| if version_compare(dubver, '>1.31.1'): |
| if self.required: |
| raise DependencyException( |
| f"DUB version {dubver} is not compatible with Meson (can't locate artifacts in Dub cache)") |
| self.is_found = False |
| return |
| |
| mlog.debug('Determining dependency {!r} with DUB executable ' |
| '{!r}'.format(name, self.dubbin.get_path())) |
| |
| # if an explicit version spec was stated, use this when querying Dub |
| main_pack_spec = name |
| if 'version' in kwargs: |
| version_spec = kwargs['version'] |
| if isinstance(version_spec, list): |
| version_spec = " ".join(version_spec) |
| main_pack_spec = f'{name}@{version_spec}' |
| |
| # we need to know the target architecture |
| dub_arch = self.compiler.arch |
| |
| # we need to know the build type as well |
| dub_buildtype = str(environment.coredata.get_option(OptionKey('buildtype'))) |
| # MESON types: choices=['plain', 'debug', 'debugoptimized', 'release', 'minsize', 'custom'])), |
| # DUB types: debug (default), plain, release, release-debug, release-nobounds, unittest, profile, profile-gc, |
| # docs, ddox, cov, unittest-cov, syntax and custom |
| if dub_buildtype == 'debugoptimized': |
| dub_buildtype = 'release-debug' |
| elif dub_buildtype == 'minsize': |
| dub_buildtype = 'release' |
| |
| # Ask dub for the package |
| describe_cmd = [ |
| 'describe', main_pack_spec, '--arch=' + dub_arch, |
| '--build=' + dub_buildtype, '--compiler=' + self.compiler.get_exelist()[-1] |
| ] |
| ret, res, err = self._call_dubbin(describe_cmd) |
| |
| if ret != 0: |
| mlog.debug('DUB describe failed: ' + err) |
| if 'locally' in err: |
| fetch_cmd = ['dub', 'fetch', main_pack_spec] |
| mlog.error(mlog.bold(main_pack_spec), 'is not present locally. You may try the following command:') |
| mlog.log(mlog.bold(join_args(fetch_cmd))) |
| self.is_found = False |
| return |
| |
| # A command that might be useful in case of missing DUB package |
| def dub_build_deep_command() -> str: |
| cmd = [ |
| 'dub', 'run', 'dub-build-deep', '--yes', '--', main_pack_spec, |
| '--arch=' + dub_arch, '--compiler=' + self.compiler.get_exelist()[-1], |
| '--build=' + dub_buildtype |
| ] |
| return join_args(cmd) |
| |
| dub_comp_id = self.compiler.get_id().replace('llvm', 'ldc').replace('gcc', 'gdc') |
| description = json.loads(res) |
| |
| self.compile_args = [] |
| self.link_args = self.raw_link_args = [] |
| |
| show_buildtype_warning = False |
| |
| def find_package_target(pkg: T.Dict[str, str]) -> bool: |
| nonlocal show_buildtype_warning |
| # try to find a static library in a DUB folder corresponding to |
| # version, configuration, compiler, arch and build-type |
| # if can find, add to link_args. |
| # link_args order is meaningful, so this function MUST be called in the right order |
| pack_id = f'{pkg["name"]}@{pkg["version"]}' |
| (tgt_file, compatibilities) = self._find_compatible_package_target(description, pkg, dub_comp_id) |
| if tgt_file is None: |
| if not compatibilities: |
| mlog.error(mlog.bold(pack_id), 'not found') |
| elif 'compiler' not in compatibilities: |
| mlog.error(mlog.bold(pack_id), 'found but not compiled with ', mlog.bold(dub_comp_id)) |
| elif dub_comp_id != 'gdc' and 'compiler_version' not in compatibilities: |
| mlog.error(mlog.bold(pack_id), 'found but not compiled with', |
| mlog.bold(f'{dub_comp_id}-{self.compiler.version}')) |
| elif 'arch' not in compatibilities: |
| mlog.error(mlog.bold(pack_id), 'found but not compiled for', mlog.bold(dub_arch)) |
| elif 'platform' not in compatibilities: |
| mlog.error(mlog.bold(pack_id), 'found but not compiled for', |
| mlog.bold(description['platform'].join('.'))) |
| elif 'configuration' not in compatibilities: |
| mlog.error(mlog.bold(pack_id), 'found but not compiled for the', |
| mlog.bold(pkg['configuration']), 'configuration') |
| else: |
| mlog.error(mlog.bold(pack_id), 'not found') |
| |
| mlog.log('You may try the following command to install the necessary DUB libraries:') |
| mlog.log(mlog.bold(dub_build_deep_command())) |
| |
| return False |
| |
| if 'build_type' not in compatibilities: |
| mlog.warning(mlog.bold(pack_id), 'found but not compiled as', mlog.bold(dub_buildtype)) |
| show_buildtype_warning = True |
| |
| self.link_args.append(tgt_file) |
| return True |
| |
| # Main algorithm: |
| # 1. Ensure that the target is a compatible library type (not dynamic) |
| # 2. Find a compatible built library for the main dependency |
| # 3. Do the same for each sub-dependency. |
| # link_args MUST be in the same order than the "linkDependencies" of the main target |
| # 4. Add other build settings (imports, versions etc.) |
| |
| # 1 |
| self.is_found = False |
| packages = {} |
| for pkg in description['packages']: |
| packages[pkg['name']] = pkg |
| |
| if not pkg['active']: |
| continue |
| |
| if pkg['targetType'] == 'dynamicLibrary': |
| mlog.error('DUB dynamic library dependencies are not supported.') |
| self.is_found = False |
| return |
| |
| # check that the main dependency is indeed a library |
| if pkg['name'] == name: |
| self.is_found = True |
| |
| if pkg['targetType'] not in ['library', 'sourceLibrary', 'staticLibrary']: |
| mlog.error(mlog.bold(name), "found but it isn't a library") |
| self.is_found = False |
| return |
| |
| self.version = pkg['version'] |
| self.pkg = pkg |
| |
| # collect all targets |
| targets = {} |
| for tgt in description['targets']: |
| targets[tgt['rootPackage']] = tgt |
| |
| if name not in targets: |
| self.is_found = False |
| if self.pkg['targetType'] == 'sourceLibrary': |
| # source libraries have no associated targets, |
| # but some build settings like import folders must be found from the package object. |
| # Current algo only get these from "buildSettings" in the target object. |
| # Let's save this for a future PR. |
| # (See openssl DUB package for example of sourceLibrary) |
| mlog.error('DUB targets of type', mlog.bold('sourceLibrary'), 'are not supported.') |
| else: |
| mlog.error('Could not find target description for', mlog.bold(main_pack_spec)) |
| |
| if not self.is_found: |
| mlog.error(f'Could not find {name} in DUB description') |
| return |
| |
| # Current impl only supports static libraries |
| self.static = True |
| |
| # 2 |
| if not find_package_target(self.pkg): |
| self.is_found = False |
| return |
| |
| # 3 |
| for link_dep in targets[name]['linkDependencies']: |
| pkg = packages[link_dep] |
| if not find_package_target(pkg): |
| self.is_found = False |
| return |
| |
| if show_buildtype_warning: |
| mlog.log('If it is not suitable, try the following command and reconfigure Meson with', mlog.bold('--clearcache')) |
| mlog.log(mlog.bold(dub_build_deep_command())) |
| |
| # 4 |
| bs = targets[name]['buildSettings'] |
| |
| for flag in bs['dflags']: |
| self.compile_args.append(flag) |
| |
| for path in bs['importPaths']: |
| self.compile_args.append('-I' + path) |
| |
| for path in bs['stringImportPaths']: |
| if 'import_dir' not in d_feature_args[self.compiler.id]: |
| break |
| flag = d_feature_args[self.compiler.id]['import_dir'] |
| self.compile_args.append(f'{flag}={path}') |
| |
| for ver in bs['versions']: |
| if 'version' not in d_feature_args[self.compiler.id]: |
| break |
| flag = d_feature_args[self.compiler.id]['version'] |
| self.compile_args.append(f'{flag}={ver}') |
| |
| if bs['mainSourceFile']: |
| self.compile_args.append(bs['mainSourceFile']) |
| |
| # pass static libraries |
| # linkerFiles are added during step 3 |
| # for file in bs['linkerFiles']: |
| # self.link_args.append(file) |
| |
| for file in bs['sourceFiles']: |
| # sourceFiles may contain static libraries |
| if file.endswith('.lib') or file.endswith('.a'): |
| self.link_args.append(file) |
| |
| for flag in bs['lflags']: |
| self.link_args.append(flag) |
| |
| is_windows = self.env.machines.host.is_windows() |
| if is_windows: |
| winlibs = ['kernel32', 'user32', 'gdi32', 'winspool', 'shell32', 'ole32', |
| 'oleaut32', 'uuid', 'comdlg32', 'advapi32', 'ws2_32'] |
| |
| for lib in bs['libs']: |
| if os.name != 'nt': |
| # trying to add system libraries by pkg-config |
| pkgdep = PkgConfigDependency(lib, environment, {'required': 'true', 'silent': 'true'}) |
| if pkgdep.is_found: |
| for arg in pkgdep.get_compile_args(): |
| self.compile_args.append(arg) |
| for arg in pkgdep.get_link_args(): |
| self.link_args.append(arg) |
| for arg in pkgdep.get_link_args(raw=True): |
| self.raw_link_args.append(arg) |
| continue |
| |
| if is_windows and lib in winlibs: |
| self.link_args.append(lib + '.lib') |
| continue |
| |
| # fallback |
| self.link_args.append('-l'+lib) |
| |
| # This function finds the target of the provided JSON package, built for the right |
| # compiler, architecture, configuration... |
| # It returns (target|None, {compatibilities}) |
| # If None is returned for target, compatibilities will list what other targets were found without full compatibility |
| def _find_compatible_package_target(self, jdesc: T.Dict[str, str], jpack: T.Dict[str, str], dub_comp_id: str) -> T.Tuple[str, T.Set[str]]: |
| dub_build_path = os.path.join(jpack['path'], '.dub', 'build') |
| |
| if not os.path.exists(dub_build_path): |
| return (None, None) |
| |
| # try to find a dir like library-debug-linux.posix-x86_64-ldc_2081-EF934983A3319F8F8FF2F0E107A363BA |
| |
| # fields are: |
| # - configuration |
| # - build type |
| # - platform |
| # - architecture |
| # - compiler id (dmd, ldc, gdc) |
| # - compiler version or frontend id or frontend version? |
| |
| conf = jpack['configuration'] |
| build_type = jdesc['buildType'] |
| platforms = jdesc['platform'] |
| archs = jdesc['architecture'] |
| |
| # Get D frontend version implemented in the compiler, or the compiler version itself |
| # gdc doesn't support this |
| comp_versions = [] |
| |
| if dub_comp_id != 'gdc': |
| comp_versions.append(self.compiler.version) |
| |
| ret, res = self._call_compbin(['--version'])[0:2] |
| if ret != 0: |
| mlog.error('Failed to run {!r}', mlog.bold(dub_comp_id)) |
| return (None, None) |
| d_ver_reg = re.search('v[0-9].[0-9][0-9][0-9].[0-9]', res) # Ex.: v2.081.2 |
| |
| if d_ver_reg is not None: |
| frontend_version = d_ver_reg.group() |
| frontend_id = frontend_version.rsplit('.', 1)[0].replace( |
| 'v', '').replace('.', '') # Fix structure. Ex.: 2081 |
| comp_versions.extend([frontend_version, frontend_id]) |
| |
| compatibilities: T.Set[str] = set() |
| |
| # build_type is not in check_list because different build types might be compatible. |
| # We do show a WARNING that the build type is not the same. |
| # It might be critical in release builds, and acceptable otherwise |
| check_list = ('configuration', 'platform', 'arch', 'compiler', 'compiler_version') |
| |
| for entry in os.listdir(dub_build_path): |
| |
| target = os.path.join(dub_build_path, entry, jpack['targetFileName']) |
| if not os.path.exists(target): |
| # unless Dub and Meson are racing, the target file should be present |
| # when the directory is present |
| mlog.debug("WARNING: Could not find a Dub target: " + target) |
| continue |
| |
| # we build a new set for each entry, because if this target is returned |
| # we want to return only the compatibilities associated to this target |
| # otherwise we could miss the WARNING about build_type |
| comps = set() |
| |
| if conf in entry: |
| comps.add('configuration') |
| |
| if build_type in entry: |
| comps.add('build_type') |
| |
| if all(platform in entry for platform in platforms): |
| comps.add('platform') |
| |
| if all(arch in entry for arch in archs): |
| comps.add('arch') |
| |
| if dub_comp_id in entry: |
| comps.add('compiler') |
| |
| if dub_comp_id == 'gdc' or any(cv in entry for cv in comp_versions): |
| comps.add('compiler_version') |
| |
| if all(key in comps for key in check_list): |
| return (target, comps) |
| else: |
| compatibilities = set.union(compatibilities, comps) |
| |
| return (None, compatibilities) |
| |
| def _call_dubbin(self, args: T.List[str], env: T.Optional[T.Dict[str, str]] = None) -> T.Tuple[int, str, str]: |
| assert isinstance(self.dubbin, ExternalProgram) |
| p, out, err = Popen_safe(self.dubbin.get_command() + args, env=env) |
| return p.returncode, out.strip(), err.strip() |
| |
| def _call_compbin(self, args: T.List[str], env: T.Optional[T.Dict[str, str]] = None) -> T.Tuple[int, str, str]: |
| p, out, err = Popen_safe(self.compiler.get_exelist() + args, env=env) |
| return p.returncode, out.strip(), err.strip() |
| |
| def _check_dub(self) -> T.Optional[T.Tuple[ExternalProgram, str]]: |
| |
| def find() -> T.Optional[T.Tuple[ExternalProgram, str]]: |
| dubbin = ExternalProgram('dub', silent=True) |
| |
| if not dubbin.found(): |
| return None |
| |
| try: |
| p, out = Popen_safe(dubbin.get_command() + ['--version'])[0:2] |
| if p.returncode != 0: |
| mlog.warning('Found dub {!r} but couldn\'t run it' |
| ''.format(' '.join(dubbin.get_command()))) |
| return None |
| |
| except (FileNotFoundError, PermissionError): |
| return None |
| |
| vermatch = re.search(r'DUB version (\d+\.\d+\.\d+.*), ', out.strip()) |
| if vermatch: |
| dubver = vermatch.group(1) |
| else: |
| mlog.warning(f"Found dub {' '.join(dubbin.get_command())} but couldn't parse version in {out.strip()}") |
| return None |
| |
| return (dubbin, dubver) |
| |
| found = find() |
| |
| if found is None: |
| mlog.log('Found DUB:', mlog.red('NO')) |
| else: |
| (dubbin, dubver) = found |
| mlog.log('Found DUB:', mlog.bold(dubbin.get_path()), |
| '(version %s)' % dubver) |
| |
| return found |