| # SPDX-License-Identifier: Apache-2.0 |
| # Copyright 2012-2017 The Meson development team |
| # Copyright © 2023-2024 Intel Corporation |
| |
| from __future__ import annotations |
| |
| from collections import OrderedDict |
| from dataclasses import dataclass |
| from enum import Enum, unique |
| from functools import lru_cache |
| from pathlib import PurePath, Path |
| from textwrap import dedent |
| import itertools |
| import json |
| import os |
| import pickle |
| import re |
| import subprocess |
| import typing as T |
| |
| from . import backends |
| from .. import modules |
| from .. import environment, mesonlib |
| from .. import build |
| from .. import mlog |
| from .. import compilers |
| from ..arglist import CompilerArgs |
| from ..compilers import Compiler |
| from ..linkers import ArLikeLinker, RSPFileSyntax |
| from ..mesonlib import ( |
| File, LibType, MachineChoice, MesonBugException, MesonException, OrderedSet, PerMachine, |
| ProgressBar, quote_arg |
| ) |
| from ..mesonlib import get_compiler_for_source, has_path_sep, OptionKey |
| from .backends import CleanTrees |
| from ..build import GeneratedList, InvalidArguments |
| |
| if T.TYPE_CHECKING: |
| from typing_extensions import Literal |
| |
| from .._typing import ImmutableListProtocol |
| from ..build import ExtractedObjects, LibTypes |
| from ..interpreter import Interpreter |
| from ..linkers.linkers import DynamicLinker, StaticLinker |
| from ..compilers.cs import CsCompiler |
| from ..compilers.fortran import FortranCompiler |
| |
| CommandArgOrStr = T.List[T.Union['NinjaCommandArg', str]] |
| RUST_EDITIONS = Literal['2015', '2018', '2021'] |
| |
| |
| FORTRAN_INCLUDE_PAT = r"^\s*#?include\s*['\"](\w+\.\w+)['\"]" |
| FORTRAN_MODULE_PAT = r"^\s*\bmodule\b\s+(\w+)\s*(?:!+.*)*$" |
| FORTRAN_SUBMOD_PAT = r"^\s*\bsubmodule\b\s*\((\w+:?\w+)\)\s*(\w+)" |
| FORTRAN_USE_PAT = r"^\s*use,?\s*(?:non_intrinsic)?\s*(?:::)?\s*(\w+)" |
| |
| def cmd_quote(arg: str) -> str: |
| # see: https://docs.microsoft.com/en-us/windows/desktop/api/shellapi/nf-shellapi-commandlinetoargvw#remarks |
| |
| # backslash escape any existing double quotes |
| # any existing backslashes preceding a quote are doubled |
| arg = re.sub(r'(\\*)"', lambda m: '\\' * (len(m.group(1)) * 2 + 1) + '"', arg) |
| # any terminal backslashes likewise need doubling |
| arg = re.sub(r'(\\*)$', lambda m: '\\' * (len(m.group(1)) * 2), arg) |
| # and double quote |
| arg = f'"{arg}"' |
| |
| return arg |
| |
| # How ninja executes command lines differs between Unix and Windows |
| # (see https://ninja-build.org/manual.html#ref_rule_command) |
| if mesonlib.is_windows(): |
| quote_func = cmd_quote |
| execute_wrapper = ['cmd', '/c'] # unused |
| rmfile_prefix = ['del', '/f', '/s', '/q', '{}', '&&'] |
| else: |
| quote_func = quote_arg |
| execute_wrapper = [] |
| rmfile_prefix = ['rm', '-f', '{}', '&&'] |
| |
| def gcc_rsp_quote(s: str) -> str: |
| # see: the function buildargv() in libiberty |
| # |
| # this differs from sh-quoting in that a backslash *always* escapes the |
| # following character, even inside single quotes. |
| |
| s = s.replace('\\', '\\\\') |
| |
| return quote_func(s) |
| |
| def get_rsp_threshold() -> int: |
| '''Return a conservative estimate of the commandline size in bytes |
| above which a response file should be used. May be overridden for |
| debugging by setting environment variable MESON_RSP_THRESHOLD.''' |
| |
| if mesonlib.is_windows(): |
| # Usually 32k, but some projects might use cmd.exe, |
| # and that has a limit of 8k. |
| limit = 8192 |
| else: |
| # On Linux, ninja always passes the commandline as a single |
| # big string to /bin/sh, and the kernel limits the size of a |
| # single argument; see MAX_ARG_STRLEN |
| limit = 131072 |
| # Be conservative |
| limit = limit // 2 |
| return int(os.environ.get('MESON_RSP_THRESHOLD', limit)) |
| |
| # a conservative estimate of the command-line length limit |
| rsp_threshold = get_rsp_threshold() |
| |
| # ninja variables whose value should remain unquoted. The value of these ninja |
| # variables (or variables we use them in) is interpreted directly by ninja |
| # (e.g. the value of the depfile variable is a pathname that ninja will read |
| # from, etc.), so it must not be shell quoted. |
| raw_names = {'DEPFILE_UNQUOTED', 'DESC', 'pool', 'description', 'targetdep', 'dyndep'} |
| |
| NINJA_QUOTE_BUILD_PAT = re.compile(r"[$ :\n]") |
| NINJA_QUOTE_VAR_PAT = re.compile(r"[$ \n]") |
| |
| def ninja_quote(text: str, is_build_line: bool = False) -> str: |
| if is_build_line: |
| quote_re = NINJA_QUOTE_BUILD_PAT |
| else: |
| quote_re = NINJA_QUOTE_VAR_PAT |
| # Fast path for when no quoting is necessary |
| if not quote_re.search(text): |
| return text |
| if '\n' in text: |
| errmsg = f'''Ninja does not support newlines in rules. The content was: |
| |
| {text} |
| |
| Please report this error with a test case to the Meson bug tracker.''' |
| raise MesonException(errmsg) |
| return quote_re.sub(r'$\g<0>', text) |
| |
| |
| @dataclass |
| class TargetDependencyScannerInfo: |
| |
| """Information passed to the depscanner about a target. |
| |
| :param private_dir: The private scratch directory for the target. |
| :param source2object: A mapping of source file names to the objects that |
| will be created from them. |
| :param sources: a list of sources mapping them to the language rules to use |
| to scan them. |
| """ |
| |
| private_dir: str |
| source2object: T.Dict[str, str] |
| sources: T.List[T.Tuple[str, Literal['cpp', 'fortran']]] |
| |
| |
| @unique |
| class Quoting(Enum): |
| both = 0 |
| notShell = 1 |
| notNinja = 2 |
| none = 3 |
| |
| class NinjaCommandArg: |
| def __init__(self, s: str, quoting: Quoting = Quoting.both) -> None: |
| self.s = s |
| self.quoting = quoting |
| |
| def __str__(self) -> str: |
| return self.s |
| |
| @staticmethod |
| def list(l: T.List[str], q: Quoting) -> T.List[NinjaCommandArg]: |
| return [NinjaCommandArg(i, q) for i in l] |
| |
| @dataclass |
| class NinjaComment: |
| comment: str |
| |
| def write(self, outfile: T.TextIO) -> None: |
| for l in self.comment.split('\n'): |
| outfile.write('# ') |
| outfile.write(l) |
| outfile.write('\n') |
| outfile.write('\n') |
| |
| class NinjaRule: |
| def __init__(self, rule: str, command: CommandArgOrStr, args: CommandArgOrStr, |
| description: str, rspable: bool = False, deps: T.Optional[str] = None, |
| depfile: T.Optional[str] = None, extra: T.Optional[str] = None, |
| rspfile_quote_style: RSPFileSyntax = RSPFileSyntax.GCC): |
| |
| def strToCommandArg(c: T.Union[NinjaCommandArg, str]) -> NinjaCommandArg: |
| if isinstance(c, NinjaCommandArg): |
| return c |
| |
| # deal with common cases here, so we don't have to explicitly |
| # annotate the required quoting everywhere |
| if c == '&&': |
| # shell constructs shouldn't be shell quoted |
| return NinjaCommandArg(c, Quoting.notShell) |
| if c.startswith('$'): |
| varp = re.match(r'\$\{?(\w*)\}?', c) |
| assert varp is not None, 'for mypy' |
| var: str = varp.group(1) |
| if var not in raw_names: |
| # ninja variables shouldn't be ninja quoted, and their value |
| # is already shell quoted |
| return NinjaCommandArg(c, Quoting.none) |
| else: |
| # shell quote the use of ninja variables whose value must |
| # not be shell quoted (as it also used by ninja) |
| return NinjaCommandArg(c, Quoting.notNinja) |
| |
| return NinjaCommandArg(c) |
| |
| self.name = rule |
| self.command: T.List[NinjaCommandArg] = [strToCommandArg(c) for c in command] # includes args which never go into a rspfile |
| self.args: T.List[NinjaCommandArg] = [strToCommandArg(a) for a in args] # args which will go into a rspfile, if used |
| self.description = description |
| self.deps = deps # depstyle 'gcc' or 'msvc' |
| self.depfile = depfile |
| self.extra = extra |
| self.rspable = rspable # if a rspfile can be used |
| self.refcount = 0 |
| self.rsprefcount = 0 |
| self.rspfile_quote_style = rspfile_quote_style |
| |
| if self.depfile == '$DEPFILE': |
| self.depfile += '_UNQUOTED' |
| |
| @staticmethod |
| def _quoter(x: NinjaCommandArg, qf: T.Callable[[str], str] = quote_func) -> str: |
| if x.quoting == Quoting.none: |
| return x.s |
| elif x.quoting == Quoting.notNinja: |
| return qf(x.s) |
| elif x.quoting == Quoting.notShell: |
| return ninja_quote(x.s) |
| return ninja_quote(qf(str(x))) |
| |
| def write(self, outfile: T.TextIO) -> None: |
| rspfile_args = self.args |
| rspfile_quote_func: T.Callable[[str], str] |
| if self.rspfile_quote_style is RSPFileSyntax.MSVC: |
| rspfile_quote_func = cmd_quote |
| rspfile_args = [NinjaCommandArg('$in_newline', arg.quoting) if arg.s == '$in' else arg for arg in rspfile_args] |
| else: |
| rspfile_quote_func = gcc_rsp_quote |
| |
| def rule_iter() -> T.Iterable[str]: |
| if self.refcount: |
| yield '' |
| if self.rsprefcount: |
| yield '_RSP' |
| |
| for rsp in rule_iter(): |
| outfile.write(f'rule {self.name}{rsp}\n') |
| if rsp == '_RSP': |
| outfile.write(' command = {} @$out.rsp\n'.format(' '.join([self._quoter(x) for x in self.command]))) |
| outfile.write(' rspfile = $out.rsp\n') |
| outfile.write(' rspfile_content = {}\n'.format(' '.join([self._quoter(x, rspfile_quote_func) for x in rspfile_args]))) |
| else: |
| outfile.write(' command = {}\n'.format(' '.join([self._quoter(x) for x in self.command + self.args]))) |
| if self.deps: |
| outfile.write(f' deps = {self.deps}\n') |
| if self.depfile: |
| outfile.write(f' depfile = {self.depfile}\n') |
| outfile.write(f' description = {self.description}\n') |
| if self.extra: |
| for l in self.extra.split('\n'): |
| outfile.write(' ') |
| outfile.write(l) |
| outfile.write('\n') |
| outfile.write('\n') |
| |
| def length_estimate(self, infiles: str, outfiles: str, |
| elems: T.List[T.Tuple[str, T.List[str]]]) -> int: |
| # determine variables |
| # this order of actions only approximates ninja's scoping rules, as |
| # documented at: https://ninja-build.org/manual.html#ref_scope |
| ninja_vars = dict(elems) |
| if self.deps is not None: |
| ninja_vars['deps'] = [self.deps] |
| if self.depfile is not None: |
| ninja_vars['depfile'] = [self.depfile] |
| ninja_vars['in'] = [infiles] |
| ninja_vars['out'] = [outfiles] |
| |
| # expand variables in command |
| command = ' '.join([self._quoter(x) for x in self.command + self.args]) |
| estimate = len(command) |
| for m in re.finditer(r'(\${\w+}|\$\w+)?[^$]*', command): |
| if m.start(1) != -1: |
| estimate -= m.end(1) - m.start(1) + 1 |
| chunk = m.group(1) |
| if chunk[1] == '{': |
| chunk = chunk[2:-1] |
| else: |
| chunk = chunk[1:] |
| chunk = ninja_vars.get(chunk, []) # undefined ninja variables are empty |
| estimate += len(' '.join(chunk)) |
| |
| # determine command length |
| return estimate |
| |
| class NinjaBuildElement: |
| def __init__(self, all_outputs: T.Set[str], outfilenames, rulename, infilenames, implicit_outs=None): |
| self.implicit_outfilenames = implicit_outs or [] |
| if isinstance(outfilenames, str): |
| self.outfilenames = [outfilenames] |
| else: |
| self.outfilenames = outfilenames |
| assert isinstance(rulename, str) |
| self.rulename = rulename |
| if isinstance(infilenames, str): |
| self.infilenames = [infilenames] |
| else: |
| self.infilenames = infilenames |
| self.deps = OrderedSet() |
| self.orderdeps = OrderedSet() |
| self.elems = [] |
| self.all_outputs = all_outputs |
| self.output_errors = '' |
| |
| def add_dep(self, dep: T.Union[str, T.List[str]]) -> None: |
| if isinstance(dep, list): |
| self.deps.update(dep) |
| else: |
| self.deps.add(dep) |
| |
| def add_orderdep(self, dep): |
| if isinstance(dep, list): |
| self.orderdeps.update(dep) |
| else: |
| self.orderdeps.add(dep) |
| |
| def add_item(self, name: str, elems: T.Union[str, T.List[str, CompilerArgs]]) -> None: |
| # Always convert from GCC-style argument naming to the naming used by the |
| # current compiler. Also filter system include paths, deduplicate, etc. |
| if isinstance(elems, CompilerArgs): |
| elems = elems.to_native() |
| if isinstance(elems, str): |
| elems = [elems] |
| self.elems.append((name, elems)) |
| |
| if name == 'DEPFILE': |
| self.elems.append((name + '_UNQUOTED', elems)) |
| |
| def _should_use_rspfile(self): |
| # 'phony' is a rule built-in to ninja |
| if self.rulename == 'phony': |
| return False |
| |
| if not self.rule.rspable: |
| return False |
| |
| infilenames = ' '.join([ninja_quote(i, True) for i in self.infilenames]) |
| outfilenames = ' '.join([ninja_quote(i, True) for i in self.outfilenames]) |
| |
| return self.rule.length_estimate(infilenames, |
| outfilenames, |
| self.elems) >= rsp_threshold |
| |
| def count_rule_references(self): |
| if self.rulename != 'phony': |
| if self._should_use_rspfile(): |
| self.rule.rsprefcount += 1 |
| else: |
| self.rule.refcount += 1 |
| |
| def write(self, outfile): |
| if self.output_errors: |
| raise MesonException(self.output_errors) |
| ins = ' '.join([ninja_quote(i, True) for i in self.infilenames]) |
| outs = ' '.join([ninja_quote(i, True) for i in self.outfilenames]) |
| implicit_outs = ' '.join([ninja_quote(i, True) for i in self.implicit_outfilenames]) |
| if implicit_outs: |
| implicit_outs = ' | ' + implicit_outs |
| use_rspfile = self._should_use_rspfile() |
| if use_rspfile: |
| rulename = self.rulename + '_RSP' |
| mlog.debug(f'Command line for building {self.outfilenames} is long, using a response file') |
| else: |
| rulename = self.rulename |
| line = f'build {outs}{implicit_outs}: {rulename} {ins}' |
| if len(self.deps) > 0: |
| line += ' | ' + ' '.join([ninja_quote(x, True) for x in sorted(self.deps)]) |
| if len(self.orderdeps) > 0: |
| orderdeps = [str(x) for x in self.orderdeps] |
| line += ' || ' + ' '.join([ninja_quote(x, True) for x in sorted(orderdeps)]) |
| line += '\n' |
| # This is the only way I could find to make this work on all |
| # platforms including Windows command shell. Slash is a dir separator |
| # on Windows, too, so all characters are unambiguous and, more importantly, |
| # do not require quoting, unless explicitly specified, which is necessary for |
| # the csc compiler. |
| line = line.replace('\\', '/') |
| if mesonlib.is_windows(): |
| # Support network paths as backslash, otherwise they are interpreted as |
| # arguments for compile/link commands when using MSVC |
| line = ' '.join( |
| (l.replace('//', '\\\\', 1) if l.startswith('//') else l) |
| for l in line.split(' ') |
| ) |
| outfile.write(line) |
| |
| if use_rspfile: |
| if self.rule.rspfile_quote_style is RSPFileSyntax.MSVC: |
| qf = cmd_quote |
| else: |
| qf = gcc_rsp_quote |
| else: |
| qf = quote_func |
| |
| for e in self.elems: |
| (name, elems) = e |
| should_quote = name not in raw_names |
| line = f' {name} = ' |
| newelems = [] |
| for i in elems: |
| if not should_quote or i == '&&': # Hackety hack hack |
| newelems.append(ninja_quote(i)) |
| else: |
| newelems.append(ninja_quote(qf(i))) |
| line += ' '.join(newelems) |
| line += '\n' |
| outfile.write(line) |
| outfile.write('\n') |
| |
| def check_outputs(self): |
| for n in self.outfilenames: |
| if n in self.all_outputs: |
| self.output_errors = f'Multiple producers for Ninja target "{n}". Please rename your targets.' |
| self.all_outputs.add(n) |
| |
| @dataclass |
| class RustDep: |
| |
| name: str |
| |
| # equal to the order value of the `RustCrate` |
| crate: int |
| |
| def to_json(self) -> T.Dict[str, object]: |
| return { |
| "crate": self.crate, |
| "name": self.name, |
| } |
| |
| @dataclass |
| class RustCrate: |
| |
| # When the json file is written, the list of Crates will be sorted by this |
| # value |
| order: int |
| |
| display_name: str |
| root_module: str |
| edition: RUST_EDITIONS |
| deps: T.List[RustDep] |
| cfg: T.List[str] |
| is_proc_macro: bool |
| |
| # This is set to True for members of this project, and False for all |
| # subprojects |
| is_workspace_member: bool |
| proc_macro_dylib_path: T.Optional[str] = None |
| |
| def to_json(self) -> T.Dict[str, object]: |
| ret: T.Dict[str, object] = { |
| "display_name": self.display_name, |
| "root_module": self.root_module, |
| "edition": self.edition, |
| "cfg": self.cfg, |
| "is_proc_macro": self.is_proc_macro, |
| "deps": [d.to_json() for d in self.deps], |
| } |
| |
| if self.is_proc_macro: |
| assert self.proc_macro_dylib_path is not None, "This shouldn't happen" |
| ret["proc_macro_dylib_path"] = self.proc_macro_dylib_path |
| |
| return ret |
| |
| |
| class NinjaBackend(backends.Backend): |
| |
| def __init__(self, build: T.Optional[build.Build], interpreter: T.Optional[Interpreter]): |
| super().__init__(build, interpreter) |
| self.name = 'ninja' |
| self.ninja_filename = 'build.ninja' |
| self.fortran_deps = {} |
| self.all_outputs: T.Set[str] = set() |
| self.introspection_data = {} |
| self.created_llvm_ir_rule = PerMachine(False, False) |
| self.rust_crates: T.Dict[str, RustCrate] = {} |
| self.implicit_meson_outs = [] |
| self._uses_dyndeps = False |
| # nvcc chokes on thin archives: |
| # nvlink fatal : Could not open input file 'libfoo.a.p' |
| # nvlink fatal : elfLink internal error |
| # hence we disable them if 'cuda' is enabled globally. See also |
| # - https://github.com/mesonbuild/meson/pull/9453 |
| # - https://github.com/mesonbuild/meson/issues/9479#issuecomment-953485040 |
| self.allow_thin_archives = PerMachine[bool](True, True) |
| if self.environment: |
| for for_machine in MachineChoice: |
| if 'cuda' in self.environment.coredata.compilers[for_machine]: |
| mlog.debug('cuda enabled globally, disabling thin archives for {}, since nvcc/nvlink cannot handle thin archives natively'.format(for_machine)) |
| self.allow_thin_archives[for_machine] = False |
| |
| def create_phony_target(self, dummy_outfile: str, rulename: str, phony_infilename: str) -> NinjaBuildElement: |
| ''' |
| We need to use aliases for targets that might be used as directory |
| names to workaround a Ninja bug that breaks `ninja -t clean`. |
| This is used for 'reserved' targets such as 'test', 'install', |
| 'benchmark', etc, and also for RunTargets. |
| https://github.com/mesonbuild/meson/issues/1644 |
| ''' |
| if dummy_outfile.startswith('meson-internal__'): |
| raise AssertionError(f'Invalid usage of create_phony_target with {dummy_outfile!r}') |
| |
| to_name = f'meson-internal__{dummy_outfile}' |
| elem = NinjaBuildElement(self.all_outputs, dummy_outfile, 'phony', to_name) |
| self.add_build(elem) |
| |
| return NinjaBuildElement(self.all_outputs, to_name, rulename, phony_infilename) |
| |
| def detect_vs_dep_prefix(self, tempfilename): |
| '''VS writes its dependency in a locale dependent format. |
| Detect the search prefix to use.''' |
| # TODO don't hard-code host |
| for compiler in self.environment.coredata.compilers.host.values(): |
| # Have to detect the dependency format |
| |
| # IFort / masm on windows is MSVC like, but doesn't have /showincludes |
| if compiler.language in {'fortran', 'masm'}: |
| continue |
| if compiler.id == 'pgi' and mesonlib.is_windows(): |
| # for the purpose of this function, PGI doesn't act enough like MSVC |
| return open(tempfilename, 'a', encoding='utf-8') |
| if compiler.get_argument_syntax() == 'msvc': |
| break |
| else: |
| # None of our compilers are MSVC, we're done. |
| return open(tempfilename, 'a', encoding='utf-8') |
| filebase = 'incdetect.' + compilers.lang_suffixes[compiler.language][0] |
| filename = os.path.join(self.environment.get_scratch_dir(), |
| filebase) |
| with open(filename, 'w', encoding='utf-8') as f: |
| f.write(dedent('''\ |
| #include<stdio.h> |
| int dummy; |
| ''')) |
| |
| # The output of cl dependency information is language |
| # and locale dependent. Any attempt at converting it to |
| # Python strings leads to failure. We _must_ do this detection |
| # in raw byte mode and write the result in raw bytes. |
| pc = subprocess.Popen(compiler.get_exelist() + |
| ['/showIncludes', '/c', filebase], |
| cwd=self.environment.get_scratch_dir(), |
| stdout=subprocess.PIPE, stderr=subprocess.PIPE) |
| (stdout, stderr) = pc.communicate() |
| |
| # We want to match 'Note: including file: ' in the line |
| # 'Note: including file: d:\MyDir\include\stdio.h', however |
| # different locales have different messages with a different |
| # number of colons. Match up to the drive name 'd:\'. |
| # When used in cross compilation, the path separator is a |
| # forward slash rather than a backslash so handle both; i.e. |
| # the path is /MyDir/include/stdio.h. |
| # With certain cross compilation wrappings of MSVC, the paths |
| # use backslashes, but without the leading drive name, so |
| # allow the path to start with any path separator, i.e. |
| # \MyDir\include\stdio.h. |
| matchre = re.compile(rb"^(.*\s)([a-zA-Z]:[\\/]|[\\\/]).*stdio.h$") |
| |
| def detect_prefix(out): |
| for line in re.split(rb'\r?\n', out): |
| match = matchre.match(line) |
| if match: |
| with open(tempfilename, 'ab') as binfile: |
| binfile.write(b'msvc_deps_prefix = ' + match.group(1) + b'\n') |
| return open(tempfilename, 'a', encoding='utf-8') |
| return None |
| |
| # Some cl wrappers (e.g. Squish Coco) output dependency info |
| # to stderr rather than stdout |
| result = detect_prefix(stdout) or detect_prefix(stderr) |
| if result: |
| return result |
| |
| raise MesonException(f'Could not determine vs dep dependency prefix string. output: {stderr} {stdout}') |
| |
| def generate(self, capture: bool = False, vslite_ctx: T.Optional[T.Dict] = None) -> T.Optional[T.Dict]: |
| if vslite_ctx: |
| # We don't yet have a use case where we'd expect to make use of this, |
| # so no harm in catching and reporting something unexpected. |
| raise MesonBugException('We do not expect the ninja backend to be given a valid \'vslite_ctx\'') |
| ninja = environment.detect_ninja_command_and_version(log=True) |
| if self.environment.coredata.get_option(OptionKey('vsenv')): |
| builddir = Path(self.environment.get_build_dir()) |
| try: |
| # For prettier printing, reduce to a relative path. If |
| # impossible (e.g., because builddir and cwd are on |
| # different Windows drives), skip and use the full path. |
| builddir = builddir.relative_to(Path.cwd()) |
| except ValueError: |
| pass |
| meson_command = mesonlib.join_args(mesonlib.get_meson_command()) |
| mlog.log() |
| mlog.log('Visual Studio environment is needed to run Ninja. It is recommended to use Meson wrapper:') |
| mlog.log(f'{meson_command} compile -C {builddir}') |
| if ninja is None: |
| raise MesonException('Could not detect Ninja v1.8.2 or newer') |
| (self.ninja_command, self.ninja_version) = ninja |
| outfilename = os.path.join(self.environment.get_build_dir(), self.ninja_filename) |
| tempfilename = outfilename + '~' |
| with open(tempfilename, 'w', encoding='utf-8') as outfile: |
| outfile.write(f'# This is the build file for project "{self.build.get_project()}"\n') |
| outfile.write('# It is autogenerated by the Meson build system.\n') |
| outfile.write('# Do not edit by hand.\n\n') |
| outfile.write('ninja_required_version = 1.8.2\n\n') |
| |
| num_pools = self.environment.coredata.optstore.get_value('backend_max_links') |
| if num_pools > 0: |
| outfile.write(f'''pool link_pool |
| depth = {num_pools} |
| |
| ''') |
| |
| with self.detect_vs_dep_prefix(tempfilename) as outfile: |
| self.generate_rules() |
| |
| self.build_elements = [] |
| self.generate_phony() |
| self.add_build_comment(NinjaComment('Build rules for targets')) |
| |
| # Optionally capture compile args per target, for later use (i.e. VisStudio project's NMake intellisense include dirs, defines, and compile options). |
| if capture: |
| captured_compile_args_per_target = {} |
| for target in self.build.get_targets().values(): |
| if isinstance(target, build.BuildTarget): |
| captured_compile_args_per_target[target.get_id()] = self.generate_common_compile_args_per_src_type(target) |
| |
| for t in ProgressBar(self.build.get_targets().values(), desc='Generating targets'): |
| self.generate_target(t) |
| mlog.log_timestamp("Targets generated") |
| self.add_build_comment(NinjaComment('Test rules')) |
| self.generate_tests() |
| mlog.log_timestamp("Tests generated") |
| self.add_build_comment(NinjaComment('Install rules')) |
| self.generate_install() |
| mlog.log_timestamp("Install generated") |
| self.generate_dist() |
| mlog.log_timestamp("Dist generated") |
| key = OptionKey('b_coverage') |
| if (key in self.environment.coredata.optstore and |
| self.environment.coredata.optstore.get_value(key)): |
| gcovr_exe, gcovr_version, lcov_exe, lcov_version, genhtml_exe, llvm_cov_exe = environment.find_coverage_tools(self.environment.coredata) |
| mlog.debug(f'Using {gcovr_exe} ({gcovr_version}), {lcov_exe} and {llvm_cov_exe} for code coverage') |
| if gcovr_exe or (lcov_exe and genhtml_exe): |
| self.add_build_comment(NinjaComment('Coverage rules')) |
| self.generate_coverage_rules(gcovr_exe, gcovr_version, llvm_cov_exe) |
| mlog.log_timestamp("Coverage rules generated") |
| else: |
| # FIXME: since we explicitly opted in, should this be an error? |
| # The docs just say these targets will be created "if possible". |
| mlog.warning('Need gcovr or lcov/genhtml to generate any coverage reports') |
| self.add_build_comment(NinjaComment('Suffix')) |
| self.generate_utils() |
| mlog.log_timestamp("Utils generated") |
| self.generate_ending() |
| |
| self.write_rules(outfile) |
| self.write_builds(outfile) |
| |
| default = 'default all\n\n' |
| outfile.write(default) |
| # Only overwrite the old build file after the new one has been |
| # fully created. |
| os.replace(tempfilename, outfilename) |
| mlog.cmd_ci_include(outfilename) # For CI debugging |
| # Refresh Ninja's caches. https://github.com/ninja-build/ninja/pull/1685 |
| # Cannot use when running with dyndeps: https://github.com/ninja-build/ninja/issues/1952 |
| if mesonlib.version_compare(self.ninja_version, '>=1.10.0') and os.path.exists(os.path.join(self.environment.build_dir, '.ninja_log')) and not self._uses_dyndeps: |
| subprocess.call(self.ninja_command + ['-t', 'restat'], cwd=self.environment.build_dir) |
| subprocess.call(self.ninja_command + ['-t', 'cleandead'], cwd=self.environment.build_dir) |
| self.generate_compdb() |
| self.generate_rust_project_json() |
| |
| if capture: |
| return captured_compile_args_per_target |
| |
| def generate_rust_project_json(self) -> None: |
| """Generate a rust-analyzer compatible rust-project.json file.""" |
| if not self.rust_crates: |
| return |
| with open(os.path.join(self.environment.get_build_dir(), 'rust-project.json'), |
| 'w', encoding='utf-8') as f: |
| json.dump( |
| { |
| "sysroot_src": os.path.join(self.environment.coredata.compilers.host['rust'].get_sysroot(), |
| 'lib/rustlib/src/rust/library/'), |
| "crates": [c.to_json() for c in self.rust_crates.values()], |
| }, |
| f, indent=4) |
| |
| # http://clang.llvm.org/docs/JSONCompilationDatabase.html |
| def generate_compdb(self): |
| rules = [] |
| # TODO: Rather than an explicit list here, rules could be marked in the |
| # rule store as being wanted in compdb |
| for for_machine in MachineChoice: |
| for compiler in self.environment.coredata.compilers[for_machine].values(): |
| rules += [f"{rule}{ext}" for rule in [self.compiler_to_rule_name(compiler)] |
| for ext in ['', '_RSP']] |
| rules += [f"{rule}{ext}" for rule in [self.compiler_to_pch_rule_name(compiler)] |
| for ext in ['', '_RSP']] |
| compdb_options = ['-x'] if mesonlib.version_compare(self.ninja_version, '>=1.9') else [] |
| ninja_compdb = self.ninja_command + ['-t', 'compdb'] + compdb_options + rules |
| builddir = self.environment.get_build_dir() |
| try: |
| jsondb = subprocess.check_output(ninja_compdb, cwd=builddir) |
| with open(os.path.join(builddir, 'compile_commands.json'), 'wb') as f: |
| f.write(jsondb) |
| except Exception: |
| mlog.warning('Could not create compilation database.', fatal=False) |
| |
| # Get all generated headers. Any source file might need them so |
| # we need to add an order dependency to them. |
| def get_generated_headers(self, target): |
| if hasattr(target, 'cached_generated_headers'): |
| return target.cached_generated_headers |
| header_deps = [] |
| # XXX: Why don't we add deps to CustomTarget headers here? |
| for genlist in target.get_generated_sources(): |
| if isinstance(genlist, (build.CustomTarget, build.CustomTargetIndex)): |
| continue |
| for src in genlist.get_outputs(): |
| if self.environment.is_header(src): |
| header_deps.append(self.get_target_generated_dir(target, genlist, src)) |
| if 'vala' in target.compilers and not isinstance(target, build.Executable): |
| vala_header = File.from_built_file(self.get_target_dir(target), target.vala_header) |
| header_deps.append(vala_header) |
| # Recurse and find generated headers |
| for dep in itertools.chain(target.link_targets, target.link_whole_targets): |
| if isinstance(dep, (build.StaticLibrary, build.SharedLibrary)): |
| header_deps += self.get_generated_headers(dep) |
| if isinstance(target, build.CompileTarget): |
| header_deps.extend(target.get_generated_headers()) |
| target.cached_generated_headers = header_deps |
| return header_deps |
| |
| def get_target_generated_sources(self, target: build.BuildTarget) -> T.MutableMapping[str, File]: |
| """ |
| Returns a dictionary with the keys being the path to the file |
| (relative to the build directory) and the value being the File object |
| representing the same path. |
| """ |
| srcs: T.MutableMapping[str, File] = OrderedDict() |
| for gensrc in target.get_generated_sources(): |
| for s in gensrc.get_outputs(): |
| rel_src = self.get_target_generated_dir(target, gensrc, s) |
| srcs[rel_src] = File.from_built_relative(rel_src) |
| return srcs |
| |
| def get_target_sources(self, target: build.BuildTarget) -> T.MutableMapping[str, File]: |
| srcs: T.MutableMapping[str, File] = OrderedDict() |
| for s in target.get_sources(): |
| # BuildTarget sources are always mesonlib.File files which are |
| # either in the source root, or generated with configure_file and |
| # in the build root |
| if not isinstance(s, File): |
| raise InvalidArguments(f'All sources in target {s!r} must be of type mesonlib.File') |
| f = s.rel_to_builddir(self.build_to_src) |
| srcs[f] = s |
| return srcs |
| |
| def get_target_source_can_unity(self, target, source): |
| if isinstance(source, File): |
| source = source.fname |
| if self.environment.is_llvm_ir(source) or \ |
| self.environment.is_assembly(source): |
| return False |
| suffix = os.path.splitext(source)[1][1:].lower() |
| for lang in backends.LANGS_CANT_UNITY: |
| if lang not in target.compilers: |
| continue |
| if suffix in target.compilers[lang].file_suffixes: |
| return False |
| return True |
| |
| def create_target_source_introspection(self, target: build.Target, comp: compilers.Compiler, parameters, sources, generated_sources, |
| unity_sources: T.Optional[T.List[mesonlib.FileOrString]] = None): |
| ''' |
| Adds the source file introspection information for a language of a target |
| |
| Internal introspection storage format: |
| self.introspection_data = { |
| '<target ID>': { |
| <id tuple>: { |
| 'language: 'lang', |
| 'compiler': ['comp', 'exe', 'list'], |
| 'parameters': ['UNIQUE', 'parameter', 'list'], |
| 'sources': [], |
| 'generated_sources': [], |
| } |
| } |
| } |
| ''' |
| tid = target.get_id() |
| lang = comp.get_language() |
| tgt = self.introspection_data[tid] |
| # Find an existing entry or create a new one |
| id_hash = (lang, tuple(parameters)) |
| src_block = tgt.get(id_hash, None) |
| if src_block is None: |
| # Convert parameters |
| if isinstance(parameters, CompilerArgs): |
| parameters = parameters.to_native(copy=True) |
| parameters = comp.compute_parameters_with_absolute_paths(parameters, self.build_dir) |
| # The new entry |
| src_block = { |
| 'language': lang, |
| 'compiler': comp.get_exelist(), |
| 'parameters': parameters, |
| 'sources': [], |
| 'generated_sources': [], |
| 'unity_sources': [], |
| } |
| tgt[id_hash] = src_block |
| |
| def compute_path(file: mesonlib.FileOrString) -> str: |
| """ Make source files absolute """ |
| if isinstance(file, File): |
| return file.absolute_path(self.source_dir, self.build_dir) |
| return os.path.normpath(os.path.join(self.build_dir, file)) |
| |
| src_block['sources'].extend(compute_path(x) for x in sources) |
| src_block['generated_sources'].extend(compute_path(x) for x in generated_sources) |
| if unity_sources: |
| src_block['unity_sources'].extend(compute_path(x) for x in unity_sources) |
| |
| def create_target_linker_introspection(self, target: build.Target, linker: T.Union[Compiler, StaticLinker], parameters): |
| tid = target.get_id() |
| tgt = self.introspection_data[tid] |
| lnk_hash = tuple(parameters) |
| lnk_block = tgt.get(lnk_hash, None) |
| if lnk_block is None: |
| if isinstance(parameters, CompilerArgs): |
| parameters = parameters.to_native(copy=True) |
| |
| if isinstance(linker, Compiler): |
| linkers = linker.get_linker_exelist() |
| else: |
| linkers = linker.get_exelist() |
| |
| lnk_block = { |
| 'linker': linkers, |
| 'parameters': parameters, |
| } |
| tgt[lnk_hash] = lnk_block |
| |
| def generate_target(self, target): |
| try: |
| if isinstance(target, build.BuildTarget): |
| os.makedirs(self.get_target_private_dir_abs(target)) |
| except FileExistsError: |
| pass |
| if isinstance(target, build.CustomTarget): |
| self.generate_custom_target(target) |
| if isinstance(target, build.RunTarget): |
| self.generate_run_target(target) |
| compiled_sources: T.List[str] = [] |
| source2object: T.Dict[str, str] = {} |
| name = target.get_id() |
| if name in self.processed_targets: |
| return |
| self.processed_targets.add(name) |
| # Initialize an empty introspection source list |
| self.introspection_data[name] = {} |
| # Generate rules for all dependency targets |
| self.process_target_dependencies(target) |
| |
| self.generate_shlib_aliases(target, self.get_target_dir(target)) |
| |
| # If target uses a language that cannot link to C objects, |
| # just generate for that language and return. |
| if isinstance(target, build.Jar): |
| self.generate_jar_target(target) |
| return |
| if target.uses_rust(): |
| self.generate_rust_target(target) |
| return |
| if 'cs' in target.compilers: |
| self.generate_cs_target(target) |
| return |
| if 'swift' in target.compilers: |
| self.generate_swift_target(target) |
| return |
| |
| # CompileTarget compiles all its sources and does not do a final link. |
| # This is, for example, a preprocessor. |
| is_compile_target = isinstance(target, build.CompileTarget) |
| |
| # Preexisting target C/C++ sources to be built; dict of full path to |
| # source relative to build root and the original File object. |
| target_sources: T.MutableMapping[str, File] |
| |
| # GeneratedList and CustomTarget sources to be built; dict of the full |
| # path to source relative to build root and the generating target/list |
| generated_sources: T.MutableMapping[str, File] |
| |
| # List of sources that have been transpiled from a DSL (like Vala) into |
| # a language that is handled below, such as C or C++ |
| transpiled_sources: T.List[str] |
| |
| if 'vala' in target.compilers: |
| # Sources consumed by valac are filtered out. These only contain |
| # C/C++ sources, objects, generated libs, and unknown sources now. |
| target_sources, generated_sources, \ |
| transpiled_sources = self.generate_vala_compile(target) |
| elif 'cython' in target.compilers: |
| target_sources, generated_sources, \ |
| transpiled_sources = self.generate_cython_transpile(target) |
| else: |
| target_sources = self.get_target_sources(target) |
| generated_sources = self.get_target_generated_sources(target) |
| transpiled_sources = [] |
| self.scan_fortran_module_outputs(target) |
| # Generate rules for GeneratedLists |
| self.generate_generator_list_rules(target) |
| |
| # Generate rules for building the remaining source files in this target |
| outname = self.get_target_filename(target) |
| obj_list = [] |
| is_unity = target.is_unity |
| header_deps = [] |
| unity_src = [] |
| unity_deps = [] # Generated sources that must be built before compiling a Unity target. |
| header_deps += self.get_generated_headers(target) |
| |
| if is_unity: |
| # Warn about incompatible sources if a unity build is enabled |
| langs = set(target.compilers.keys()) |
| langs_cant = langs.intersection(backends.LANGS_CANT_UNITY) |
| if langs_cant: |
| langs_are = langs = ', '.join(langs_cant).upper() |
| langs_are += ' are' if len(langs_cant) > 1 else ' is' |
| msg = f'{langs_are} not supported in Unity builds yet, so {langs} ' \ |
| f'sources in the {target.name!r} target will be compiled normally' |
| mlog.log(mlog.red('FIXME'), msg) |
| |
| # Get a list of all generated headers that will be needed while building |
| # this target's sources (generated sources and preexisting sources). |
| # This will be set as dependencies of all the target's sources. At the |
| # same time, also deal with generated sources that need to be compiled. |
| generated_source_files: T.List[File] = [] |
| for rel_src in generated_sources.keys(): |
| raw_src = File.from_built_relative(rel_src) |
| if self.environment.is_source(rel_src): |
| if is_unity and self.get_target_source_can_unity(target, rel_src): |
| unity_deps.append(raw_src) |
| abs_src = os.path.join(self.environment.get_build_dir(), rel_src) |
| unity_src.append(abs_src) |
| else: |
| generated_source_files.append(raw_src) |
| elif self.environment.is_object(rel_src): |
| obj_list.append(rel_src) |
| elif self.environment.is_library(rel_src) or modules.is_module_library(rel_src): |
| pass |
| elif is_compile_target: |
| generated_source_files.append(raw_src) |
| else: |
| # Assume anything not specifically a source file is a header. This is because |
| # people generate files with weird suffixes (.inc, .fh) that they then include |
| # in their source files. |
| header_deps.append(raw_src) |
| |
| # For D language, the object of generated source files are added |
| # as order only deps because other files may depend on them |
| d_generated_deps = [] |
| |
| # These are the generated source files that need to be built for use by |
| # this target. We create the Ninja build file elements for this here |
| # because we need `header_deps` to be fully generated in the above loop. |
| for src in generated_source_files: |
| if self.environment.is_llvm_ir(src): |
| o, s = self.generate_llvm_ir_compile(target, src) |
| else: |
| o, s = self.generate_single_compile(target, src, True, order_deps=header_deps) |
| compiled_sources.append(s) |
| source2object[s] = o |
| obj_list.append(o) |
| if s.split('.')[-1] in compilers.lang_suffixes['d']: |
| d_generated_deps.append(o) |
| |
| use_pch = self.target_uses_pch(target) |
| if use_pch and target.has_pch(): |
| pch_objects = self.generate_pch(target, header_deps=header_deps) |
| else: |
| pch_objects = [] |
| |
| o, od = self.flatten_object_list(target) |
| obj_targets = [t for t in od if t.uses_fortran()] |
| obj_list.extend(o) |
| |
| fortran_order_deps = [File(True, *os.path.split(self.get_target_filename(t))) for t in obj_targets] |
| fortran_inc_args: T.List[str] = [] |
| if target.uses_fortran(): |
| fortran_inc_args = mesonlib.listify([target.compilers['fortran'].get_include_args( |
| self.get_target_private_dir(t), is_system=False) for t in obj_targets]) |
| |
| # Generate compilation targets for sources generated by transpilers. |
| # |
| # Do not try to unity-build the generated source files, as these |
| # often contain duplicate symbols and will fail to compile properly. |
| # |
| # Gather all generated source files and header before generating the |
| # compilation rules, to be able to add correct dependencies on the |
| # generated headers. |
| transpiled_source_files = [] |
| for src in transpiled_sources: |
| raw_src = File.from_built_relative(src) |
| # Generated targets are ordered deps because the must exist |
| # before the sources compiling them are used. After the first |
| # compile we get precise dependency info from dep files. |
| # This should work in all cases. If it does not, then just |
| # move them from orderdeps to proper deps. |
| if self.environment.is_header(src): |
| header_deps.append(raw_src) |
| else: |
| transpiled_source_files.append(raw_src) |
| for src in transpiled_source_files: |
| o, s = self.generate_single_compile(target, src, True, [], header_deps) |
| obj_list.append(o) |
| |
| # Generate compile targets for all the preexisting sources for this target |
| for src in target_sources.values(): |
| if not self.environment.is_header(src) or is_compile_target: |
| if self.environment.is_llvm_ir(src): |
| o, s = self.generate_llvm_ir_compile(target, src) |
| obj_list.append(o) |
| elif is_unity and self.get_target_source_can_unity(target, src): |
| abs_src = os.path.join(self.environment.get_build_dir(), |
| src.rel_to_builddir(self.build_to_src)) |
| unity_src.append(abs_src) |
| else: |
| o, s = self.generate_single_compile(target, src, False, [], |
| header_deps + d_generated_deps + fortran_order_deps, |
| fortran_inc_args) |
| obj_list.append(o) |
| compiled_sources.append(s) |
| source2object[s] = o |
| |
| if is_unity: |
| for src in self.generate_unity_files(target, unity_src): |
| o, s = self.generate_single_compile(target, src, True, unity_deps + header_deps + d_generated_deps, |
| fortran_order_deps, fortran_inc_args, unity_src) |
| obj_list.append(o) |
| compiled_sources.append(s) |
| source2object[s] = o |
| if is_compile_target: |
| # Skip the link stage for this special type of target |
| return |
| linker, stdlib_args = self.determine_linker_and_stdlib_args(target) |
| if isinstance(target, build.StaticLibrary) and target.prelink: |
| final_obj_list = self.generate_prelink(target, obj_list) |
| else: |
| final_obj_list = obj_list |
| elem = self.generate_link(target, outname, final_obj_list, linker, pch_objects, stdlib_args=stdlib_args) |
| self.generate_dependency_scan_target(target, compiled_sources, source2object, generated_source_files, fortran_order_deps) |
| self.add_build(elem) |
| #In AIX, we archive shared libraries. If the instance is a shared library, we add a command to archive the shared library |
| #object and create the build element. |
| if isinstance(target, build.SharedLibrary) and self.environment.machines[target.for_machine].is_aix(): |
| if target.aix_so_archive: |
| elem = NinjaBuildElement(self.all_outputs, linker.get_archive_name(outname), 'AIX_LINKER', [outname]) |
| self.add_build(elem) |
| |
| def should_use_dyndeps_for_target(self, target: 'build.BuildTarget') -> bool: |
| if mesonlib.version_compare(self.ninja_version, '<1.10.0'): |
| return False |
| if 'fortran' in target.compilers: |
| return True |
| if 'cpp' not in target.compilers: |
| return False |
| if '-fmodules-ts' in target.extra_args['cpp']: |
| return True |
| # Currently only the preview version of Visual Studio is supported. |
| cpp = target.compilers['cpp'] |
| if cpp.get_id() != 'msvc': |
| return False |
| cppversion = target.get_option(OptionKey('std', machine=target.for_machine, lang='cpp')) |
| if cppversion not in ('latest', 'c++latest', 'vc++latest'): |
| return False |
| if not mesonlib.current_vs_supports_modules(): |
| return False |
| if mesonlib.version_compare(cpp.version, '<19.28.28617'): |
| return False |
| return True |
| |
| def generate_dependency_scan_target(self, target: build.BuildTarget, |
| compiled_sources: T.List[str], |
| source2object: T.Dict[str, str], |
| generated_source_files: T.List[mesonlib.File], |
| object_deps: T.List['mesonlib.FileOrString']) -> None: |
| if not self.should_use_dyndeps_for_target(target): |
| return |
| self._uses_dyndeps = True |
| depscan_file = self.get_dep_scan_file_for(target) |
| pickle_base = target.name + '.dat' |
| pickle_file = os.path.join(self.get_target_private_dir(target), pickle_base).replace('\\', '/') |
| pickle_abs = os.path.join(self.get_target_private_dir_abs(target), pickle_base).replace('\\', '/') |
| rule_name = 'depscan' |
| scan_sources = list(self.select_sources_to_scan(compiled_sources)) |
| |
| scaninfo = TargetDependencyScannerInfo( |
| self.get_target_private_dir(target), source2object, scan_sources) |
| |
| write = True |
| if os.path.exists(pickle_abs): |
| with open(pickle_abs, 'rb') as p: |
| old = pickle.load(p) |
| write = old != scaninfo |
| |
| if write: |
| with open(pickle_abs, 'wb') as p: |
| pickle.dump(scaninfo, p) |
| |
| elem = NinjaBuildElement(self.all_outputs, depscan_file, rule_name, pickle_file) |
| # Add any generated outputs to the order deps of the scan target, so |
| # that those sources are present |
| for g in generated_source_files: |
| elem.orderdeps.add(g.relative_name()) |
| elem.orderdeps.update(object_deps) |
| self.add_build(elem) |
| |
| def select_sources_to_scan(self, compiled_sources: T.List[str] |
| ) -> T.Iterable[T.Tuple[str, Literal['cpp', 'fortran']]]: |
| # in practice pick up C++ and Fortran files. If some other language |
| # requires scanning (possibly Java to deal with inner class files) |
| # then add them here. |
| for source in compiled_sources: |
| ext = os.path.splitext(source)[1][1:] |
| if ext.lower() in compilers.lang_suffixes['cpp'] or ext == 'C': |
| yield source, 'cpp' |
| elif ext.lower() in compilers.lang_suffixes['fortran']: |
| yield source, 'fortran' |
| |
| def process_target_dependencies(self, target): |
| for t in target.get_dependencies(): |
| if t.get_id() not in self.processed_targets: |
| self.generate_target(t) |
| |
| def custom_target_generator_inputs(self, target): |
| for s in target.sources: |
| if isinstance(s, build.GeneratedList): |
| self.generate_genlist_for_target(s, target) |
| |
| def unwrap_dep_list(self, target): |
| deps = [] |
| for i in target.get_dependencies(): |
| # FIXME, should not grab element at zero but rather expand all. |
| if isinstance(i, list): |
| i = i[0] |
| # Add a dependency on all the outputs of this target |
| for output in i.get_outputs(): |
| deps.append(os.path.join(self.get_target_dir(i), output)) |
| return deps |
| |
| def generate_custom_target(self, target: build.CustomTarget): |
| self.custom_target_generator_inputs(target) |
| (srcs, ofilenames, cmd) = self.eval_custom_target_command(target) |
| deps = self.unwrap_dep_list(target) |
| deps += self.get_target_depend_files(target) |
| if target.build_always_stale: |
| deps.append('PHONY') |
| if target.depfile is None: |
| rulename = 'CUSTOM_COMMAND' |
| else: |
| rulename = 'CUSTOM_COMMAND_DEP' |
| elem = NinjaBuildElement(self.all_outputs, ofilenames, rulename, srcs) |
| elem.add_dep(deps) |
| for d in target.extra_depends: |
| # Add a dependency on all the outputs of this target |
| for output in d.get_outputs(): |
| elem.add_dep(os.path.join(self.get_target_dir(d), output)) |
| |
| cmd, reason = self.as_meson_exe_cmdline(target.command[0], cmd[1:], |
| extra_bdeps=target.get_transitive_build_target_deps(), |
| capture=ofilenames[0] if target.capture else None, |
| feed=srcs[0] if target.feed else None, |
| env=target.env, |
| verbose=target.console) |
| if reason: |
| cmd_type = f' (wrapped by meson {reason})' |
| else: |
| cmd_type = '' |
| if target.depfile is not None: |
| depfile = target.get_dep_outname(elem.infilenames) |
| rel_dfile = os.path.join(self.get_target_dir(target), depfile) |
| abs_pdir = os.path.join(self.environment.get_build_dir(), self.get_target_dir(target)) |
| os.makedirs(abs_pdir, exist_ok=True) |
| elem.add_item('DEPFILE', rel_dfile) |
| if target.console: |
| elem.add_item('pool', 'console') |
| full_name = Path(target.subdir, target.name).as_posix() |
| elem.add_item('COMMAND', cmd) |
| elem.add_item('description', target.description.format(full_name) + cmd_type) |
| self.add_build(elem) |
| self.processed_targets.add(target.get_id()) |
| |
| def build_run_target_name(self, target): |
| if target.subproject != '': |
| subproject_prefix = f'{target.subproject}@@' |
| else: |
| subproject_prefix = '' |
| return f'{subproject_prefix}{target.name}' |
| |
| def generate_run_target(self, target: build.RunTarget): |
| target_name = self.build_run_target_name(target) |
| if not target.command: |
| # This is an alias target, it has no command, it just depends on |
| # other targets. |
| elem = NinjaBuildElement(self.all_outputs, target_name, 'phony', []) |
| else: |
| target_env = self.get_run_target_env(target) |
| _, _, cmd = self.eval_custom_target_command(target) |
| meson_exe_cmd, reason = self.as_meson_exe_cmdline(target.command[0], cmd[1:], |
| env=target_env, |
| verbose=True) |
| cmd_type = f' (wrapped by meson {reason})' if reason else '' |
| elem = self.create_phony_target(target_name, 'CUSTOM_COMMAND', []) |
| elem.add_item('COMMAND', meson_exe_cmd) |
| elem.add_item('description', f'Running external command {target.name}{cmd_type}') |
| elem.add_item('pool', 'console') |
| deps = self.unwrap_dep_list(target) |
| deps += self.get_target_depend_files(target) |
| elem.add_dep(deps) |
| self.add_build(elem) |
| self.processed_targets.add(target.get_id()) |
| |
| def generate_coverage_command(self, elem, outputs: T.List[str], gcovr_exe: T.Optional[str], llvm_cov_exe: T.Optional[str]): |
| targets = self.build.get_targets().values() |
| use_llvm_cov = False |
| exe_args = [] |
| if gcovr_exe is not None: |
| exe_args += ['--gcov', gcovr_exe] |
| if llvm_cov_exe is not None: |
| exe_args += ['--llvm-cov', llvm_cov_exe] |
| |
| for target in targets: |
| if not hasattr(target, 'compilers'): |
| continue |
| for compiler in target.compilers.values(): |
| if compiler.get_id() == 'clang' and not compiler.info.is_darwin(): |
| use_llvm_cov = True |
| break |
| elem.add_item('COMMAND', self.environment.get_build_command() + |
| ['--internal', 'coverage'] + |
| outputs + |
| [self.environment.get_source_dir(), |
| os.path.join(self.environment.get_source_dir(), |
| self.build.get_subproject_dir()), |
| self.environment.get_build_dir(), |
| self.environment.get_log_dir()] + |
| exe_args + |
| (['--use-llvm-cov'] if use_llvm_cov else [])) |
| |
| def generate_coverage_rules(self, gcovr_exe: T.Optional[str], gcovr_version: T.Optional[str], llvm_cov_exe: T.Optional[str]): |
| e = self.create_phony_target('coverage', 'CUSTOM_COMMAND', 'PHONY') |
| self.generate_coverage_command(e, [], gcovr_exe, llvm_cov_exe) |
| e.add_item('description', 'Generates coverage reports') |
| self.add_build(e) |
| self.generate_coverage_legacy_rules(gcovr_exe, gcovr_version, llvm_cov_exe) |
| |
| def generate_coverage_legacy_rules(self, gcovr_exe: T.Optional[str], gcovr_version: T.Optional[str], llvm_cov_exe: T.Optional[str]): |
| e = self.create_phony_target('coverage-html', 'CUSTOM_COMMAND', 'PHONY') |
| self.generate_coverage_command(e, ['--html'], gcovr_exe, llvm_cov_exe) |
| e.add_item('description', 'Generates HTML coverage report') |
| self.add_build(e) |
| |
| if gcovr_exe: |
| e = self.create_phony_target('coverage-xml', 'CUSTOM_COMMAND', 'PHONY') |
| self.generate_coverage_command(e, ['--xml'], gcovr_exe, llvm_cov_exe) |
| e.add_item('description', 'Generates XML coverage report') |
| self.add_build(e) |
| |
| e = self.create_phony_target('coverage-text', 'CUSTOM_COMMAND', 'PHONY') |
| self.generate_coverage_command(e, ['--text'], gcovr_exe, llvm_cov_exe) |
| e.add_item('description', 'Generates text coverage report') |
| self.add_build(e) |
| |
| if mesonlib.version_compare(gcovr_version, '>=4.2'): |
| e = self.create_phony_target('coverage-sonarqube', 'CUSTOM_COMMAND', 'PHONY') |
| self.generate_coverage_command(e, ['--sonarqube'], gcovr_exe, llvm_cov_exe) |
| e.add_item('description', 'Generates Sonarqube XML coverage report') |
| self.add_build(e) |
| |
| def generate_install(self): |
| self.create_install_data_files() |
| elem = self.create_phony_target('install', 'CUSTOM_COMMAND', 'PHONY') |
| elem.add_dep('all') |
| elem.add_item('DESC', 'Installing files.') |
| elem.add_item('COMMAND', self.environment.get_build_command() + ['install', '--no-rebuild']) |
| elem.add_item('pool', 'console') |
| self.add_build(elem) |
| |
| def generate_tests(self): |
| self.serialize_tests() |
| cmd = self.environment.get_build_command(True) + ['test', '--no-rebuild'] |
| if not self.environment.coredata.get_option(OptionKey('stdsplit')): |
| cmd += ['--no-stdsplit'] |
| if self.environment.coredata.get_option(OptionKey('errorlogs')): |
| cmd += ['--print-errorlogs'] |
| elem = self.create_phony_target('test', 'CUSTOM_COMMAND', ['all', 'PHONY']) |
| elem.add_item('COMMAND', cmd) |
| elem.add_item('DESC', 'Running all tests.') |
| elem.add_item('pool', 'console') |
| self.add_build(elem) |
| |
| # And then benchmarks. |
| cmd = self.environment.get_build_command(True) + [ |
| 'test', '--benchmark', '--logbase', |
| 'benchmarklog', '--num-processes=1', '--no-rebuild'] |
| elem = self.create_phony_target('benchmark', 'CUSTOM_COMMAND', ['all', 'PHONY']) |
| elem.add_item('COMMAND', cmd) |
| elem.add_item('DESC', 'Running benchmark suite.') |
| elem.add_item('pool', 'console') |
| self.add_build(elem) |
| |
| def generate_rules(self): |
| self.rules = [] |
| self.ruledict = {} |
| |
| self.add_rule_comment(NinjaComment('Rules for module scanning.')) |
| self.generate_scanner_rules() |
| self.add_rule_comment(NinjaComment('Rules for compiling.')) |
| self.generate_compile_rules() |
| self.add_rule_comment(NinjaComment('Rules for linking.')) |
| self.generate_static_link_rules() |
| self.generate_dynamic_link_rules() |
| self.add_rule_comment(NinjaComment('Other rules')) |
| # Ninja errors out if you have deps = gcc but no depfile, so we must |
| # have two rules for custom commands. |
| self.add_rule(NinjaRule('CUSTOM_COMMAND', ['$COMMAND'], [], '$DESC', |
| extra='restat = 1')) |
| self.add_rule(NinjaRule('CUSTOM_COMMAND_DEP', ['$COMMAND'], [], '$DESC', |
| deps='gcc', depfile='$DEPFILE', |
| extra='restat = 1')) |
| self.add_rule(NinjaRule('COPY_FILE', self.environment.get_build_command() + ['--internal', 'copy'], |
| ['$in', '$out'], 'Copying $in to $out')) |
| |
| c = self.environment.get_build_command() + \ |
| ['--internal', |
| 'regenerate', |
| self.environment.get_source_dir(), |
| # Ninja always runs from the build_dir. This includes cases where the user moved the |
| # build directory and invalidated most references. Make sure it still regenerates. |
| '.'] |
| self.add_rule(NinjaRule('REGENERATE_BUILD', |
| c, [], |
| 'Regenerating build files.', |
| extra='generator = 1')) |
| |
| def add_rule_comment(self, comment: NinjaComment) -> None: |
| self.rules.append(comment) |
| |
| def add_build_comment(self, comment: NinjaComment) -> None: |
| self.build_elements.append(comment) |
| |
| def add_rule(self, rule: NinjaRule) -> None: |
| if rule.name in self.ruledict: |
| raise MesonException(f'Tried to add rule {rule.name} twice.') |
| self.rules.append(rule) |
| self.ruledict[rule.name] = rule |
| |
| def add_build(self, build: NinjaBuildElement) -> None: |
| build.check_outputs() |
| self.build_elements.append(build) |
| |
| if build.rulename != 'phony': |
| # reference rule |
| if build.rulename in self.ruledict: |
| build.rule = self.ruledict[build.rulename] |
| else: |
| mlog.warning(f"build statement for {build.outfilenames} references nonexistent rule {build.rulename}") |
| |
| def write_rules(self, outfile: T.TextIO) -> None: |
| for b in self.build_elements: |
| if isinstance(b, NinjaBuildElement): |
| b.count_rule_references() |
| |
| for r in self.rules: |
| r.write(outfile) |
| |
| def write_builds(self, outfile: T.TextIO) -> None: |
| for b in ProgressBar(self.build_elements, desc='Writing build.ninja'): |
| b.write(outfile) |
| mlog.log_timestamp("build.ninja generated") |
| |
| def generate_phony(self) -> None: |
| self.add_build_comment(NinjaComment('Phony build target, always out of date')) |
| elem = NinjaBuildElement(self.all_outputs, 'PHONY', 'phony', '') |
| self.add_build(elem) |
| |
| def generate_jar_target(self, target: build.Jar): |
| fname = target.get_filename() |
| outname_rel = os.path.join(self.get_target_dir(target), fname) |
| src_list = target.get_sources() |
| resources = target.get_java_resources() |
| compiler = target.compilers['java'] |
| c = 'c' |
| m = 'm' |
| e = '' |
| f = 'f' |
| main_class = target.get_main_class() |
| if main_class != '': |
| e = 'e' |
| |
| # Add possible java generated files to src list |
| generated_sources = self.get_target_generated_sources(target) |
| gen_src_list = [] |
| for rel_src in generated_sources.keys(): |
| raw_src = File.from_built_relative(rel_src) |
| if rel_src.endswith('.java'): |
| gen_src_list.append(raw_src) |
| |
| compile_args = self.determine_java_compile_args(target, compiler) |
| class_list = self.generate_java_compile(src_list + gen_src_list, target, compiler, compile_args) |
| class_dep_list = [os.path.join(self.get_target_private_dir(target), i) for i in class_list] |
| manifest_path = os.path.join(self.get_target_private_dir(target), 'META-INF', 'MANIFEST.MF') |
| manifest_fullpath = os.path.join(self.environment.get_build_dir(), manifest_path) |
| os.makedirs(os.path.dirname(manifest_fullpath), exist_ok=True) |
| with open(manifest_fullpath, 'w', encoding='utf-8') as manifest: |
| if any(target.link_targets): |
| manifest.write('Class-Path: ') |
| cp_paths = [os.path.join(self.get_target_dir(l), l.get_filename()) for l in target.link_targets] |
| manifest.write(' '.join(cp_paths)) |
| manifest.write('\n') |
| jar_rule = 'java_LINKER' |
| commands = [c + m + e + f] |
| commands.append(manifest_path) |
| if e != '': |
| commands.append(main_class) |
| commands.append(self.get_target_filename(target)) |
| # Java compilation can produce an arbitrary number of output |
| # class files for a single source file. Thus tell jar to just |
| # grab everything in the final package. |
| commands += ['-C', self.get_target_private_dir(target), '.'] |
| elem = NinjaBuildElement(self.all_outputs, outname_rel, jar_rule, []) |
| elem.add_dep(class_dep_list) |
| if resources: |
| # Copy all resources into the root of the jar. |
| elem.add_orderdep(self.__generate_sources_structure(Path(self.get_target_private_dir(target)), resources)[0]) |
| elem.add_item('ARGS', commands) |
| self.add_build(elem) |
| # Create introspection information |
| self.create_target_source_introspection(target, compiler, compile_args, src_list, gen_src_list) |
| |
| def generate_cs_resource_tasks(self, target): |
| args = [] |
| deps = [] |
| for r in target.resources: |
| rel_sourcefile = os.path.join(self.build_to_src, target.subdir, r) |
| if r.endswith('.resources'): |
| a = '-resource:' + rel_sourcefile |
| elif r.endswith('.txt') or r.endswith('.resx'): |
| ofilebase = os.path.splitext(os.path.basename(r))[0] + '.resources' |
| ofilename = os.path.join(self.get_target_private_dir(target), ofilebase) |
| elem = NinjaBuildElement(self.all_outputs, ofilename, "CUSTOM_COMMAND", rel_sourcefile) |
| elem.add_item('COMMAND', ['resgen', rel_sourcefile, ofilename]) |
| elem.add_item('DESC', f'Compiling resource {rel_sourcefile}') |
| self.add_build(elem) |
| deps.append(ofilename) |
| a = '-resource:' + ofilename |
| else: |
| raise InvalidArguments(f'Unknown resource file {r}.') |
| args.append(a) |
| return args, deps |
| |
| def generate_cs_target(self, target: build.BuildTarget): |
| fname = target.get_filename() |
| outname_rel = os.path.join(self.get_target_dir(target), fname) |
| src_list = target.get_sources() |
| compiler = target.compilers['cs'] |
| rel_srcs = [os.path.normpath(s.rel_to_builddir(self.build_to_src)) for s in src_list] |
| deps = [] |
| commands = self.generate_basic_compiler_args(target, compiler) |
| commands += target.extra_args['cs'] |
| if isinstance(target, build.Executable): |
| commands.append('-target:exe') |
| elif isinstance(target, build.SharedLibrary): |
| commands.append('-target:library') |
| else: |
| raise MesonException('Unknown C# target type.') |
| (resource_args, resource_deps) = self.generate_cs_resource_tasks(target) |
| commands += resource_args |
| deps += resource_deps |
| commands += compiler.get_output_args(outname_rel) |
| for l in target.link_targets: |
| lname = os.path.join(self.get_target_dir(l), l.get_filename()) |
| commands += compiler.get_link_args(lname) |
| deps.append(lname) |
| if '-g' in commands: |
| outputs = [outname_rel, outname_rel + '.mdb'] |
| else: |
| outputs = [outname_rel] |
| generated_sources = self.get_target_generated_sources(target) |
| generated_rel_srcs = [] |
| for rel_src in generated_sources.keys(): |
| if rel_src.lower().endswith('.cs'): |
| generated_rel_srcs.append(os.path.normpath(rel_src)) |
| deps.append(os.path.normpath(rel_src)) |
| |
| for dep in target.get_external_deps(): |
| commands.extend_direct(dep.get_link_args()) |
| |
| elem = NinjaBuildElement(self.all_outputs, outputs, self.compiler_to_rule_name(compiler), rel_srcs + generated_rel_srcs) |
| elem.add_dep(deps) |
| elem.add_item('ARGS', commands) |
| self.add_build(elem) |
| |
| self.generate_generator_list_rules(target) |
| self.create_target_source_introspection(target, compiler, commands, rel_srcs, generated_rel_srcs) |
| |
| def determine_java_compile_args(self, target, compiler): |
| args = [] |
| args = self.generate_basic_compiler_args(target, compiler) |
| args += target.get_java_args() |
| args += compiler.get_output_args(self.get_target_private_dir(target)) |
| args += target.get_classpath_args() |
| curdir = target.get_subdir() |
| sourcepath = os.path.join(self.build_to_src, curdir) + os.pathsep |
| sourcepath += os.path.normpath(curdir) + os.pathsep |
| for i in target.include_dirs: |
| for idir in i.get_incdirs(): |
| sourcepath += os.path.join(self.build_to_src, i.curdir, idir) + os.pathsep |
| args += ['-sourcepath', sourcepath] |
| return args |
| |
| def generate_java_compile(self, srcs, target, compiler, args): |
| deps = [os.path.join(self.get_target_dir(l), l.get_filename()) for l in target.link_targets] |
| generated_sources = self.get_target_generated_sources(target) |
| for rel_src in generated_sources.keys(): |
| if rel_src.endswith('.java'): |
| deps.append(rel_src) |
| |
| rel_srcs = [] |
| plain_class_paths = [] |
| rel_objs = [] |
| for src in srcs: |
| rel_src = src.rel_to_builddir(self.build_to_src) |
| rel_srcs.append(rel_src) |
| |
| plain_class_path = src.fname[:-4] + 'class' |
| plain_class_paths.append(plain_class_path) |
| rel_obj = os.path.join(self.get_target_private_dir(target), plain_class_path) |
| rel_objs.append(rel_obj) |
| element = NinjaBuildElement(self.all_outputs, rel_objs, self.compiler_to_rule_name(compiler), rel_srcs) |
| element.add_dep(deps) |
| element.add_item('ARGS', args) |
| element.add_item('FOR_JAR', self.get_target_filename(target)) |
| self.add_build(element) |
| return plain_class_paths |
| |
| def generate_java_link(self): |
| rule = 'java_LINKER' |
| command = ['jar', '$ARGS'] |
| description = 'Creating JAR $out' |
| self.add_rule(NinjaRule(rule, command, [], description)) |
| |
| def determine_dep_vapis(self, target): |
| """ |
| Peek into the sources of BuildTargets we're linking with, and if any of |
| them was built with Vala, assume that it also generated a .vapi file of |
| the same name as the BuildTarget and return the path to it relative to |
| the build directory. |
| """ |
| result = OrderedSet() |
| for dep in itertools.chain(target.link_targets, target.link_whole_targets): |
| if not dep.is_linkable_target(): |
| continue |
| for i in dep.sources: |
| if hasattr(i, 'fname'): |
| i = i.fname |
| if i.split('.')[-1] in compilers.lang_suffixes['vala']: |
| vapiname = dep.vala_vapi |
| fullname = os.path.join(self.get_target_dir(dep), vapiname) |
| result.add(fullname) |
| break |
| return list(result) |
| |
| def split_vala_sources(self, t: build.BuildTarget) -> \ |
| T.Tuple[T.MutableMapping[str, File], T.MutableMapping[str, File], |
| T.Tuple[T.MutableMapping[str, File], T.MutableMapping]]: |
| """ |
| Splits the target's sources into .vala, .gs, .vapi, and other sources. |
| Handles both preexisting and generated sources. |
| |
| Returns a tuple (vala, vapi, others) each of which is a dictionary with |
| the keys being the path to the file (relative to the build directory) |
| and the value being the object that generated or represents the file. |
| """ |
| vala: T.MutableMapping[str, File] = OrderedDict() |
| vapi: T.MutableMapping[str, File] = OrderedDict() |
| others: T.MutableMapping[str, File] = OrderedDict() |
| othersgen: T.MutableMapping[str, File] = OrderedDict() |
| # Split preexisting sources |
| for s in t.get_sources(): |
| # BuildTarget sources are always mesonlib.File files which are |
| # either in the source root, or generated with configure_file and |
| # in the build root |
| if not isinstance(s, File): |
| raise InvalidArguments(f'All sources in target {t!r} must be of type mesonlib.File, not {s!r}') |
| f = s.rel_to_builddir(self.build_to_src) |
| if s.endswith(('.vala', '.gs')): |
| srctype = vala |
| elif s.endswith('.vapi'): |
| srctype = vapi |
| else: |
| srctype = others |
| srctype[f] = s |
| # Split generated sources |
| for gensrc in t.get_generated_sources(): |
| for s in gensrc.get_outputs(): |
| f = self.get_target_generated_dir(t, gensrc, s) |
| if s.endswith(('.vala', '.gs')): |
| srctype = vala |
| elif s.endswith('.vapi'): |
| srctype = vapi |
| # Generated non-Vala (C/C++) sources. Won't be used for |
| # generating the Vala compile rule below. |
| else: |
| srctype = othersgen |
| # Duplicate outputs are disastrous |
| if f in srctype and srctype[f] is not gensrc: |
| msg = 'Duplicate output {0!r} from {1!r} {2!r}; ' \ |
| 'conflicts with {0!r} from {4!r} {3!r}' \ |
| ''.format(f, type(gensrc).__name__, gensrc.name, |
| srctype[f].name, type(srctype[f]).__name__) |
| raise InvalidArguments(msg) |
| # Store 'somefile.vala': GeneratedList (or CustomTarget) |
| srctype[f] = gensrc |
| return vala, vapi, (others, othersgen) |
| |
| def generate_vala_compile(self, target: build.BuildTarget) -> \ |
| T.Tuple[T.MutableMapping[str, File], T.MutableMapping[str, File], T.List[str]]: |
| """Vala is compiled into C. Set up all necessary build steps here.""" |
| (vala_src, vapi_src, other_src) = self.split_vala_sources(target) |
| extra_dep_files = [] |
| if not vala_src: |
| raise InvalidArguments(f'Vala library {target.name!r} has no Vala or Genie source files.') |
| |
| valac = target.compilers['vala'] |
| c_out_dir = self.get_target_private_dir(target) |
| # C files generated by valac |
| vala_c_src: T.List[str] = [] |
| # Files generated by valac |
| valac_outputs: T.List = [] |
| # All sources that are passed to valac on the commandline |
| all_files = list(vapi_src) |
| # Passed as --basedir |
| srcbasedir = os.path.join(self.build_to_src, target.get_subdir()) |
| for (vala_file, gensrc) in vala_src.items(): |
| all_files.append(vala_file) |
| # Figure out where the Vala compiler will write the compiled C file |
| # |
| # If the Vala file is in a subdir of the build dir (in our case |
| # because it was generated/built by something else), and is also |
| # a subdir of --basedir (because the builddir is in the source |
| # tree, and the target subdir is the source root), the subdir |
| # components from the source root till the private builddir will be |
| # duplicated inside the private builddir. Otherwise, just the |
| # basename will be used. |
| # |
| # If the Vala file is outside the build directory, the paths from |
| # the --basedir till the subdir will be duplicated inside the |
| # private builddir. |
| if isinstance(gensrc, (build.CustomTarget, build.GeneratedList)) or gensrc.is_built: |
| vala_c_file = os.path.splitext(os.path.basename(vala_file))[0] + '.c' |
| # Check if the vala file is in a subdir of --basedir |
| abs_srcbasedir = os.path.join(self.environment.get_source_dir(), target.get_subdir()) |
| abs_vala_file = os.path.join(self.environment.get_build_dir(), vala_file) |
| if PurePath(os.path.commonpath((abs_srcbasedir, abs_vala_file))) == PurePath(abs_srcbasedir): |
| vala_c_subdir = PurePath(abs_vala_file).parent.relative_to(abs_srcbasedir) |
| vala_c_file = os.path.join(str(vala_c_subdir), vala_c_file) |
| else: |
| path_to_target = os.path.join(self.build_to_src, target.get_subdir()) |
| if vala_file.startswith(path_to_target): |
| vala_c_file = os.path.splitext(os.path.relpath(vala_file, path_to_target))[0] + '.c' |
| else: |
| vala_c_file = os.path.splitext(os.path.basename(vala_file))[0] + '.c' |
| # All this will be placed inside the c_out_dir |
| vala_c_file = os.path.join(c_out_dir, vala_c_file) |
| vala_c_src.append(vala_c_file) |
| valac_outputs.append(vala_c_file) |
| |
| args = self.generate_basic_compiler_args(target, valac) |
| args += valac.get_colorout_args(target.get_option(OptionKey('b_colorout'))) |
| # Tell Valac to output everything in our private directory. Sadly this |
| # means it will also preserve the directory components of Vala sources |
| # found inside the build tree (generated sources). |
| args += ['--directory', c_out_dir] |
| args += ['--basedir', srcbasedir] |
| if target.is_linkable_target(): |
| # Library name |
| args += ['--library', target.name] |
| # Outputted header |
| hname = os.path.join(self.get_target_dir(target), target.vala_header) |
| args += ['--header', hname] |
| if target.is_unity: |
| # Without this the declarations will get duplicated in the .c |
| # files and cause a build failure when all of them are |
| # #include-d in one .c file. |
| # https://github.com/mesonbuild/meson/issues/1969 |
| args += ['--use-header'] |
| valac_outputs.append(hname) |
| # Outputted vapi file |
| vapiname = os.path.join(self.get_target_dir(target), target.vala_vapi) |
| # Force valac to write the vapi and gir files in the target build dir. |
| # Without this, it will write it inside c_out_dir |
| args += ['--vapi', os.path.join('..', target.vala_vapi)] |
| valac_outputs.append(vapiname) |
| # Install header and vapi to default locations if user requests this |
| if len(target.install_dir) > 1 and target.install_dir[1] is True: |
| target.install_dir[1] = self.environment.get_includedir() |
| if len(target.install_dir) > 2 and target.install_dir[2] is True: |
| target.install_dir[2] = os.path.join(self.environment.get_datadir(), 'vala', 'vapi') |
| # Generate GIR if requested |
| if isinstance(target.vala_gir, str): |
| girname = os.path.join(self.get_target_dir(target), target.vala_gir) |
| args += ['--gir', os.path.join('..', target.vala_gir)] |
| valac_outputs.append(girname) |
| # Install GIR to default location if requested by user |
| if len(target.install_dir) > 3 and target.install_dir[3] is True: |
| target.install_dir[3] = os.path.join(self.environment.get_datadir(), 'gir-1.0') |
| # Detect gresources and add --gresources/--gresourcesdir arguments for each |
| gres_dirs = [] |
| for gensrc in other_src[1].values(): |
| if isinstance(gensrc, modules.GResourceTarget): |
| gres_xml, = self.get_custom_target_sources(gensrc) |
| args += ['--gresources=' + gres_xml] |
| for source_dir in gensrc.source_dirs: |
| gres_dirs += [os.path.join(self.get_target_dir(gensrc), source_dir)] |
| # Ensure that resources are built before vala sources |
| # This is required since vala code using [GtkTemplate] effectively depends on .ui files |
| # GResourceHeaderTarget is not suitable due to lacking depfile |
| gres_c, = gensrc.get_outputs() |
| extra_dep_files += [os.path.join(self.get_target_dir(gensrc), gres_c)] |
| for gres_dir in OrderedSet(gres_dirs): |
| args += [f'--gresourcesdir={gres_dir}'] |
| dependency_vapis = self.determine_dep_vapis(target) |
| extra_dep_files += dependency_vapis |
| extra_dep_files.extend(self.get_target_depend_files(target)) |
| args += target.get_extra_args('vala') |
| element = NinjaBuildElement(self.all_outputs, valac_outputs, |
| self.compiler_to_rule_name(valac), |
| all_files + dependency_vapis) |
| element.add_item('ARGS', args) |
| element.add_dep(extra_dep_files) |
| self.add_build(element) |
| self.create_target_source_introspection(target, valac, args, all_files, []) |
| return other_src[0], other_src[1], vala_c_src |
| |
| def generate_cython_transpile(self, target: build.BuildTarget) -> \ |
| T.Tuple[T.MutableMapping[str, File], T.MutableMapping[str, File], T.List[str]]: |
| """Generate rules for transpiling Cython files to C or C++""" |
| |
| static_sources: T.MutableMapping[str, File] = OrderedDict() |
| generated_sources: T.MutableMapping[str, File] = OrderedDict() |
| cython_sources: T.List[str] = [] |
| |
| cython = target.compilers['cython'] |
| |
| args: T.List[str] = [] |
| args += cython.get_always_args() |
| args += cython.get_debug_args(target.get_option(OptionKey('debug'))) |
| args += cython.get_optimization_args(target.get_option(OptionKey('optimization'))) |
| args += cython.get_option_compile_args(target.get_options()) |
| args += self.build.get_global_args(cython, target.for_machine) |
| args += self.build.get_project_args(cython, target.subproject, target.for_machine) |
| args += target.get_extra_args('cython') |
| |
| ext = target.get_option(OptionKey('language', machine=target.for_machine, lang='cython')) |
| |
| pyx_sources = [] # Keep track of sources we're adding to build |
| |
| for src in target.get_sources(): |
| if src.endswith('.pyx'): |
| output = os.path.join(self.get_target_private_dir(target), f'{src}.{ext}') |
| element = NinjaBuildElement( |
| self.all_outputs, [output], |
| self.compiler_to_rule_name(cython), |
| [src.absolute_path(self.environment.get_source_dir(), self.environment.get_build_dir())]) |
| element.add_item('ARGS', args) |
| self.add_build(element) |
| # TODO: introspection? |
| cython_sources.append(output) |
| pyx_sources.append(element) |
| else: |
| static_sources[src.rel_to_builddir(self.build_to_src)] = src |
| |
| header_deps = [] # Keep track of generated headers for those sources |
| for gen in target.get_generated_sources(): |
| for ssrc in gen.get_outputs(): |
| if isinstance(gen, GeneratedList): |
| ssrc = os.path.join(self.get_target_private_dir(target), ssrc) |
| else: |
| ssrc = os.path.join(gen.get_subdir(), ssrc) |
| if ssrc.endswith('.pyx'): |
| output = os.path.join(self.get_target_private_dir(target), f'{ssrc}.{ext}') |
| element = NinjaBuildElement( |
| self.all_outputs, [output], |
| self.compiler_to_rule_name(cython), |
| [ssrc]) |
| element.add_item('ARGS', args) |
| self.add_build(element) |
| pyx_sources.append(element) |
| # TODO: introspection? |
| cython_sources.append(output) |
| else: |
| generated_sources[ssrc] = mesonlib.File.from_built_file(gen.get_subdir(), ssrc) |
| # Following logic in L883-900 where we determine whether to add generated source |
| # as a header(order-only) dep to the .so compilation rule |
| if not self.environment.is_source(ssrc) and \ |
| not self.environment.is_object(ssrc) and \ |
| not self.environment.is_library(ssrc) and \ |
| not modules.is_module_library(ssrc): |
| header_deps.append(ssrc) |
| for source in pyx_sources: |
| source.add_orderdep(header_deps) |
| |
| return static_sources, generated_sources, cython_sources |
| |
| def _generate_copy_target(self, src: 'mesonlib.FileOrString', output: Path) -> None: |
| """Create a target to copy a source file from one location to another.""" |
| if isinstance(src, File): |
| instr = src.absolute_path(self.environment.source_dir, self.environment.build_dir) |
| else: |
| instr = src |
| elem = NinjaBuildElement(self.all_outputs, [str(output)], 'COPY_FILE', [instr]) |
| elem.add_orderdep(instr) |
| self.add_build(elem) |
| |
| def __generate_sources_structure(self, root: Path, structured_sources: build.StructuredSources) -> T.Tuple[T.List[str], T.Optional[str]]: |
| first_file: T.Optional[str] = None |
| orderdeps: T.List[str] = [] |
| for path, files in structured_sources.sources.items(): |
| for file in files: |
| if isinstance(file, File): |
| out = root / path / Path(file.fname).name |
| orderdeps.append(str(out)) |
| self._generate_copy_target(file, out) |
| if first_file is None: |
| first_file = str(out) |
| else: |
| for f in file.get_outputs(): |
| out = root / path / f |
| orderdeps.append(str(out)) |
| self._generate_copy_target(str(Path(file.subdir) / f), out) |
| if first_file is None: |
| first_file = str(out) |
| return orderdeps, first_file |
| |
| def _add_rust_project_entry(self, name: str, main_rust_file: str, args: CompilerArgs, |
| from_subproject: bool, proc_macro_dylib_path: T.Optional[str], |
| deps: T.List[RustDep]) -> None: |
| raw_edition: T.Optional[str] = mesonlib.first(reversed(args), lambda x: x.startswith('--edition')) |
| edition: RUST_EDITIONS = '2015' if not raw_edition else raw_edition.split('=')[-1] |
| |
| cfg: T.List[str] = [] |
| arg_itr: T.Iterator[str] = iter(args) |
| for arg in arg_itr: |
| if arg == '--cfg': |
| cfg.append(next(arg_itr)) |
| elif arg.startswith('--cfg'): |
| cfg.append(arg[len('--cfg'):]) |
| |
| crate = RustCrate( |
| len(self.rust_crates), |
| name, |
| main_rust_file, |
| edition, |
| deps, |
| cfg, |
| is_workspace_member=not from_subproject, |
| is_proc_macro=proc_macro_dylib_path is not None, |
| proc_macro_dylib_path=proc_macro_dylib_path, |
| ) |
| |
| self.rust_crates[name] = crate |
| |
| def _get_rust_dependency_name(self, target: build.BuildTarget, dependency: LibTypes) -> str: |
| # Convert crate names with dashes to underscores by default like |
| # cargo does as dashes can't be used as parts of identifiers |
| # in Rust |
| return target.rust_dependency_map.get(dependency.name, dependency.name).replace('-', '_') |
| |
| def generate_rust_target(self, target: build.BuildTarget) -> None: |
| rustc = target.compilers['rust'] |
| # Rust compiler takes only the main file as input and |
| # figures out what other files are needed via import |
| # statements and magic. |
| base_proxy = target.get_options() |
| args = rustc.compiler_args() |
| # Compiler args for compiling this target |
| args += compilers.get_base_compile_args(base_proxy, rustc, self.environment) |
| self.generate_generator_list_rules(target) |
| |
| # dependencies need to cause a relink, they're not just for ordering |
| deps: T.List[str] = [] |
| |
| # Dependencies for rust-project.json |
| project_deps: T.List[RustDep] = [] |
| |
| orderdeps: T.List[str] = [] |
| |
| main_rust_file = None |
| if target.structured_sources: |
| if target.structured_sources.needs_copy(): |
| _ods, main_rust_file = self.__generate_sources_structure(Path( |
| self.get_target_private_dir(target)) / 'structured', target.structured_sources) |
| orderdeps.extend(_ods) |
| else: |
| # The only way to get here is to have only files in the "root" |
| # positional argument, which are all generated into the same |
| # directory |
| g = target.structured_sources.first_file() |
| |
| if isinstance(g, File): |
| main_rust_file = g.rel_to_builddir(self.build_to_src) |
| elif isinstance(g, GeneratedList): |
| main_rust_file = os.path.join(self.get_target_private_dir(target), g.get_outputs()[0]) |
| else: |
| main_rust_file = os.path.join(g.get_subdir(), g.get_outputs()[0]) |
| |
| for f in target.structured_sources.as_list(): |
| if isinstance(f, File): |
| orderdeps.append(f.rel_to_builddir(self.build_to_src)) |
| else: |
| orderdeps.extend([os.path.join(self.build_to_src, f.subdir, s) |
| for s in f.get_outputs()]) |
| |
| for i in target.get_sources(): |
| if not rustc.can_compile(i): |
| raise InvalidArguments(f'Rust target {target.get_basename()} contains a non-rust source file.') |
| if main_rust_file is None: |
| main_rust_file = i.rel_to_builddir(self.build_to_src) |
| for g in target.get_generated_sources(): |
| for i in g.get_outputs(): |
| if not rustc.can_compile(i): |
| raise InvalidArguments(f'Rust target {target.get_basename()} contains a non-rust source file.') |
| if isinstance(g, GeneratedList): |
| fname = os.path.join(self.get_target_private_dir(target), i) |
| else: |
| fname = os.path.join(g.get_subdir(), i) |
| if main_rust_file is None: |
| main_rust_file = fname |
| orderdeps.append(fname) |
| if main_rust_file is None: |
| raise RuntimeError('A Rust target has no Rust sources. This is weird. Also a bug. Please report') |
| target_name = os.path.join(target.subdir, target.get_filename()) |
| cratetype = target.rust_crate_type |
| args.extend(['--crate-type', cratetype]) |
| |
| # If we're dynamically linking, add those arguments |
| # |
| # Rust is super annoying, calling -C link-arg foo does not work, it has |
| # to be -C link-arg=foo |
| if cratetype in {'bin', 'dylib'}: |
| args.extend(rustc.get_linker_always_args()) |
| |
| args += self.generate_basic_compiler_args(target, rustc) |
| # Rustc replaces - with _. spaces or dots are not allowed, so we replace them with underscores |
| args += ['--crate-name', target.name.replace('-', '_').replace(' ', '_').replace('.', '_')] |
| depfile = os.path.join(target.subdir, target.name + '.d') |
| args += ['--emit', f'dep-info={depfile}', '--emit', f'link={target_name}'] |
| args += ['--out-dir', self.get_target_private_dir(target)] |
| args += ['-C', 'metadata=' + target.get_id()] |
| args += target.get_extra_args('rust') |
| |
| # Rustc always use non-debug Windows runtime. Inject the one selected |
| # by Meson options instead. |
| # https://github.com/rust-lang/rust/issues/39016 |
| if not isinstance(target, build.StaticLibrary): |
| try: |
| buildtype = target.get_option(OptionKey('buildtype')) |
| crt = target.get_option(OptionKey('b_vscrt')) |
| args += rustc.get_crt_link_args(crt, buildtype) |
| except (KeyError, AttributeError): |
| pass |
| |
| if mesonlib.version_compare(rustc.version, '>= 1.67.0'): |
| verbatim = '+verbatim' |
| else: |
| verbatim = '' |
| |
| def _link_library(libname: str, static: bool, bundle: bool = False): |
| type_ = 'static' if static else 'dylib' |
| modifiers = [] |
| if not bundle and static: |
| modifiers.append('-bundle') |
| if verbatim: |
| modifiers.append(verbatim) |
| if modifiers: |
| type_ += ':' + ','.join(modifiers) |
| args.append(f'-l{type_}={libname}') |
| |
| linkdirs = mesonlib.OrderedSet() |
| external_deps = target.external_deps.copy() |
| target_deps = target.get_dependencies() |
| for d in target_deps: |
| linkdirs.add(d.subdir) |
| deps.append(self.get_dependency_filename(d)) |
| if isinstance(d, build.StaticLibrary): |
| external_deps.extend(d.external_deps) |
| if d.uses_rust_abi(): |
| if d not in itertools.chain(target.link_targets, target.link_whole_targets): |
| # Indirect Rust ABI dependency, we only need its path in linkdirs. |
| continue |
| # specify `extern CRATE_NAME=OUTPUT_FILE` for each Rust |
| # dependency, so that collisions with libraries in rustc's |
| # sysroot don't cause ambiguity |
| d_name = self._get_rust_dependency_name(target, d) |
| args += ['--extern', '{}={}'.format(d_name, os.path.join(d.subdir, d.filename))] |
| project_deps.append(RustDep(d_name, self.rust_crates[d.name].order)) |
| continue |
| |
| # Link a C ABI library |
| |
| # Pass native libraries directly to the linker with "-C link-arg" |
| # because rustc's "-l:+verbatim=" is not portable and we cannot rely |
| # on linker to find the right library without using verbatim filename. |
| # For example "-lfoo" won't find "foo.so" in the case name_prefix set |
| # to "", or would always pick the shared library when both "libfoo.so" |
| # and "libfoo.a" are available. |
| # See https://doc.rust-lang.org/rustc/command-line-arguments.html#linking-modifiers-verbatim. |
| # |
| # However, rustc static linker (rlib and staticlib) requires using |
| # "-l" argument and does not rely on platform specific dynamic linker. |
| lib = self.get_target_filename_for_linking(d) |
| link_whole = d in target.link_whole_targets |
| if isinstance(target, build.StaticLibrary) or (isinstance(target, build.Executable) and rustc.get_crt_static()): |
| static = isinstance(d, build.StaticLibrary) |
| libname = os.path.basename(lib) if verbatim else d.name |
| _link_library(libname, static, bundle=link_whole) |
| elif link_whole: |
| link_whole_args = rustc.linker.get_link_whole_for([lib]) |
| args += [f'-Clink-arg={a}' for a in link_whole_args] |
| else: |
| args.append(f'-Clink-arg={lib}') |
| |
| for e in external_deps: |
| for a in e.get_link_args(): |
| if a in rustc.native_static_libs: |
| # Exclude link args that rustc already add by default |
| pass |
| elif a.startswith('-L'): |
| args.append(a) |
| elif a.endswith(('.dll', '.so', '.dylib', '.a', '.lib')) and isinstance(target, build.StaticLibrary): |
| dir_, lib = os.path.split(a) |
| linkdirs.add(dir_) |
| if not verbatim: |
| lib, ext = os.path.splitext(lib) |
| if lib.startswith('lib'): |
| lib = lib[3:] |
| static = a.endswith(('.a', '.lib')) |
| _link_library(lib, static) |
| else: |
| args.append(f'-Clink-arg={a}') |
| |
| for d in linkdirs: |
| d = d or '.' |
| args.append(f'-L{d}') |
| |
| # Because of the way rustc links, this must come after any potential |
| # library need to link with their stdlibs (C++ and Fortran, for example) |
| args.extend(f'-Clink-arg={a}' for a in target.get_used_stdlib_args('rust')) |
| |
| has_shared_deps = any(isinstance(dep, build.SharedLibrary) for dep in target_deps) |
| has_rust_shared_deps = any(dep.uses_rust() |
| and dep.rust_crate_type == 'dylib' |
| for dep in target_deps) |
| |
| if cratetype in {'dylib', 'proc-macro'} or has_rust_shared_deps: |
| # add prefer-dynamic if any of the Rust libraries we link |
| # against are dynamic or this is a dynamic library itself, |
| # otherwise we'll end up with multiple implementations of libstd. |
| args += ['-C', 'prefer-dynamic'] |
| |
| if isinstance(target, build.SharedLibrary) or has_shared_deps: |
| # build the usual rpath arguments as well... |
| |
| # Set runtime-paths so we can run executables without needing to set |
| # LD_LIBRARY_PATH, etc in the environment. Doesn't work on Windows. |
| if has_path_sep(target.name): |
| # Target names really should not have slashes in them, but |
| # unfortunately we did not check for that and some downstream projects |
| # now have them. Once slashes are forbidden, remove this bit. |
| target_slashname_workaround_dir = os.path.join(os.path.dirname(target.name), |
| self.get_target_dir(target)) |
| else: |
| target_slashname_workaround_dir = self.get_target_dir(target) |
| rpath_args, target.rpath_dirs_to_remove = ( |
| rustc.build_rpath_args(self.environment, |
| self.environment.get_build_dir(), |
| target_slashname_workaround_dir, |
| self.determine_rpath_dirs(target), |
| target.build_rpath, |
| target.install_rpath)) |
| # ... but then add rustc's sysroot to account for rustup |
| # installations |
| for rpath_arg in rpath_args: |
| args += ['-C', 'link-arg=' + rpath_arg + ':' + os.path.join(rustc.get_sysroot(), 'lib')] |
| |
| proc_macro_dylib_path = None |
| if getattr(target, 'rust_crate_type', '') == 'proc-macro': |
| proc_macro_dylib_path = os.path.abspath(os.path.join(target.subdir, target.get_filename())) |
| |
| self._add_rust_project_entry(target.name, |
| os.path.abspath(os.path.join(self.environment.build_dir, main_rust_file)), |
| args, |
| bool(target.subproject), |
| proc_macro_dylib_path, |
| project_deps) |
| |
| compiler_name = self.compiler_to_rule_name(rustc) |
| element = NinjaBuildElement(self.all_outputs, target_name, compiler_name, main_rust_file) |
| if orderdeps: |
| element.add_orderdep(orderdeps) |
| if deps: |
| element.add_dep(deps) |
| element.add_item('ARGS', args) |
| element.add_item('targetdep', depfile) |
| element.add_item('cratetype', cratetype) |
| self.add_build(element) |
| if isinstance(target, build.SharedLibrary): |
| self.generate_shsym(target) |
| self.create_target_source_introspection(target, rustc, args, [main_rust_file], []) |
| |
| @staticmethod |
| def get_rule_suffix(for_machine: MachineChoice) -> str: |
| return PerMachine('_FOR_BUILD', '')[for_machine] |
| |
| @classmethod |
| def get_compiler_rule_name(cls, lang: str, for_machine: MachineChoice, mode: str = 'COMPILER') -> str: |
| return f'{lang}_{mode}{cls.get_rule_suffix(for_machine)}' |
| |
| @classmethod |
| def compiler_to_rule_name(cls, compiler: Compiler) -> str: |
| return cls.get_compiler_rule_name(compiler.get_language(), compiler.for_machine, compiler.mode) |
| |
| @classmethod |
| def compiler_to_pch_rule_name(cls, compiler: Compiler) -> str: |
| return cls.get_compiler_rule_name(compiler.get_language(), compiler.for_machine, 'PCH') |
| |
| def swift_module_file_name(self, target): |
| return os.path.join(self.get_target_private_dir(target), |
| self.target_swift_modulename(target) + '.swiftmodule') |
| |
| def target_swift_modulename(self, target): |
| return target.name |
| |
| def determine_swift_dep_modules(self, target): |
| result = [] |
| for l in target.link_targets: |
| if self.is_swift_target(l): |
| result.append(self.swift_module_file_name(l)) |
| return result |
| |
| def get_swift_link_deps(self, target): |
| result = [] |
| for l in target.link_targets: |
| result.append(self.get_target_filename(l)) |
| return result |
| |
| def split_swift_generated_sources(self, target): |
| all_srcs = self.get_target_generated_sources(target) |
| srcs = [] |
| others = [] |
| for i in all_srcs: |
| if i.endswith('.swift'): |
| srcs.append(i) |
| else: |
| others.append(i) |
| return srcs, others |
| |
| def generate_swift_target(self, target): |
| module_name = self.target_swift_modulename(target) |
| swiftc = target.compilers['swift'] |
| abssrc = [] |
| relsrc = [] |
| abs_headers = [] |
| header_imports = [] |
| for i in target.get_sources(): |
| if swiftc.can_compile(i): |
| rels = i.rel_to_builddir(self.build_to_src) |
| abss = os.path.normpath(os.path.join(self.environment.get_build_dir(), rels)) |
| relsrc.append(rels) |
| abssrc.append(abss) |
| elif self.environment.is_header(i): |
| relh = i.rel_to_builddir(self.build_to_src) |
| absh = os.path.normpath(os.path.join(self.environment.get_build_dir(), relh)) |
| abs_headers.append(absh) |
| header_imports += swiftc.get_header_import_args(absh) |
| else: |
| raise InvalidArguments(f'Swift target {target.get_basename()} contains a non-swift source file.') |
| os.makedirs(self.get_target_private_dir_abs(target), exist_ok=True) |
| compile_args = self.generate_basic_compiler_args(target, swiftc) |
| compile_args += swiftc.get_compile_only_args() |
| compile_args += swiftc.get_module_args(module_name) |
| for i in reversed(target.get_include_dirs()): |
| basedir = i.get_curdir() |
| for d in i.get_incdirs(): |
| if d not in ('', '.'): |
| expdir = os.path.join(basedir, d) |
| else: |
| expdir = basedir |
| srctreedir = os.path.normpath(os.path.join(self.environment.get_build_dir(), self.build_to_src, expdir)) |
| sargs = swiftc.get_include_args(srctreedir, False) |
| compile_args += sargs |
| compile_args += target.get_extra_args('swift') |
| link_args = swiftc.get_output_args(os.path.join(self.environment.get_build_dir(), self.get_target_filename(target))) |
| link_args += self.build.get_project_link_args(swiftc, target.subproject, target.for_machine) |
| link_args += self.build.get_global_link_args(swiftc, target.for_machine) |
| rundir = self.get_target_private_dir(target) |
| out_module_name = self.swift_module_file_name(target) |
| in_module_files = self.determine_swift_dep_modules(target) |
| abs_module_dirs = self.determine_swift_dep_dirs(target) |
| module_includes = [] |
| for x in abs_module_dirs: |
| module_includes += swiftc.get_include_args(x, False) |
| link_deps = self.get_swift_link_deps(target) |
| abs_link_deps = [os.path.join(self.environment.get_build_dir(), x) for x in link_deps] |
| for d in target.link_targets: |
| reldir = self.get_target_dir(d) |
| if reldir == '': |
| reldir = '.' |
| link_args += ['-L', os.path.normpath(os.path.join(self.environment.get_build_dir(), reldir))] |
| (rel_generated, _) = self.split_swift_generated_sources(target) |
| abs_generated = [os.path.join(self.environment.get_build_dir(), x) for x in rel_generated] |
| # We need absolute paths because swiftc needs to be invoked in a subdir |
| # and this is the easiest way about it. |
| objects = [] # Relative to swift invocation dir |
| rel_objects = [] # Relative to build.ninja |
| for i in abssrc + abs_generated: |
| base = os.path.basename(i) |
| oname = os.path.splitext(base)[0] + '.o' |
| objects.append(oname) |
| rel_objects.append(os.path.join(self.get_target_private_dir(target), oname)) |
| |
| rulename = self.compiler_to_rule_name(swiftc) |
| |
| # Swiftc does not seem to be able to emit objects and module files in one go. |
| elem = NinjaBuildElement(self.all_outputs, rel_objects, rulename, abssrc) |
| elem.add_dep(in_module_files + rel_generated) |
| elem.add_dep(abs_headers) |
| elem.add_item('ARGS', compile_args + header_imports + abs_generated + module_includes) |
| elem.add_item('RUNDIR', rundir) |
| self.add_build(elem) |
| elem = NinjaBuildElement(self.all_outputs, out_module_name, rulename, abssrc) |
| elem.add_dep(in_module_files + rel_generated) |
| elem.add_item('ARGS', compile_args + abs_generated + module_includes + swiftc.get_mod_gen_args()) |
| elem.add_item('RUNDIR', rundir) |
| self.add_build(elem) |
| if isinstance(target, build.StaticLibrary): |
| elem = self.generate_link(target, self.get_target_filename(target), |
| rel_objects, self.build.static_linker[target.for_machine]) |
| self.add_build(elem) |
| elif isinstance(target, build.Executable): |
| elem = NinjaBuildElement(self.all_outputs, self.get_target_filename(target), rulename, []) |
| elem.add_dep(rel_objects) |
| elem.add_dep(link_deps) |
| elem.add_item('ARGS', link_args + swiftc.get_std_exe_link_args() + objects + abs_link_deps) |
| elem.add_item('RUNDIR', rundir) |
| self.add_build(elem) |
| else: |
| raise MesonException('Swift supports only executable and static library targets.') |
| # Introspection information |
| self.create_target_source_introspection(target, swiftc, compile_args + header_imports + module_includes, relsrc, rel_generated) |
| |
| def _rsp_options(self, tool: T.Union['Compiler', 'StaticLinker', 'DynamicLinker']) -> T.Dict[str, T.Union[bool, RSPFileSyntax]]: |
| """Helper method to get rsp options. |
| |
| rsp_file_syntax() is only guaranteed to be implemented if |
| can_linker_accept_rsp() returns True. |
| """ |
| options = {'rspable': tool.can_linker_accept_rsp()} |
| if options['rspable']: |
| options['rspfile_quote_style'] = tool.rsp_file_syntax() |
| return options |
| |
| def generate_static_link_rules(self): |
| num_pools = self.environment.coredata.optstore.get_value('backend_max_links') |
| if 'java' in self.environment.coredata.compilers.host: |
| self.generate_java_link() |
| for for_machine in MachineChoice: |
| static_linker = self.build.static_linker[for_machine] |
| if static_linker is None: |
| continue |
| rule = 'STATIC_LINKER{}'.format(self.get_rule_suffix(for_machine)) |
| cmdlist: T.List[T.Union[str, NinjaCommandArg]] = [] |
| args = ['$in'] |
| # FIXME: Must normalize file names with pathlib.Path before writing |
| # them out to fix this properly on Windows. See: |
| # https://github.com/mesonbuild/meson/issues/1517 |
| # https://github.com/mesonbuild/meson/issues/1526 |
| if isinstance(static_linker, ArLikeLinker) and not mesonlib.is_windows(): |
| # `ar` has no options to overwrite archives. It always appends, |
| # which is never what we want. Delete an existing library first if |
| # it exists. https://github.com/mesonbuild/meson/issues/1355 |
| cmdlist = execute_wrapper + [c.format('$out') for c in rmfile_prefix] |
| cmdlist += static_linker.get_exelist() |
| cmdlist += ['$LINK_ARGS'] |
| cmdlist += NinjaCommandArg.list(static_linker.get_output_args('$out'), Quoting.none) |
| # The default ar on MacOS (at least through version 12), does not |
| # add extern'd variables to the symbol table by default, and |
| # requires that apple's ranlib be called with a special flag |
| # instead after linking |
| if static_linker.id == 'applear': |
| # This is a bit of a hack, but we assume that that we won't need |
| # an rspfile on MacOS, otherwise the arguments are passed to |
| # ranlib, not to ar |
| cmdlist.extend(args) |
| args = [] |
| # Ensure that we use the user-specified ranlib if any, and |
| # fallback to just picking up some ranlib otherwise |
| ranlib = self.environment.lookup_binary_entry(for_machine, 'ranlib') |
| if ranlib is None: |
| ranlib = ['ranlib'] |
| cmdlist.extend(['&&'] + ranlib + ['-c', '$out']) |
| description = 'Linking static target $out' |
| if num_pools > 0: |
| pool = 'pool = link_pool' |
| else: |
| pool = None |
| |
| options = self._rsp_options(static_linker) |
| self.add_rule(NinjaRule(rule, cmdlist, args, description, **options, extra=pool)) |
| |
| def generate_dynamic_link_rules(self): |
| num_pools = self.environment.coredata.optstore.get_value('backend_max_links') |
| for for_machine in MachineChoice: |
| complist = self.environment.coredata.compilers[for_machine] |
| for langname, compiler in complist.items(): |
| if langname in {'java', 'vala', 'rust', 'cs', 'cython'}: |
| continue |
| rule = '{}_LINKER{}'.format(langname, self.get_rule_suffix(for_machine)) |
| command = compiler.get_linker_exelist() |
| args = ['$ARGS'] + NinjaCommandArg.list(compiler.get_linker_output_args('$out'), Quoting.none) + ['$in', '$LINK_ARGS'] |
| description = 'Linking target $out' |
| if num_pools > 0: |
| pool = 'pool = link_pool' |
| else: |
| pool = None |
| |
| options = self._rsp_options(compiler) |
| self.add_rule(NinjaRule(rule, command, args, description, **options, extra=pool)) |
| if self.environment.machines[for_machine].is_aix(): |
| rule = 'AIX_LINKER{}'.format(self.get_rule_suffix(for_machine)) |
| description = 'Archiving AIX shared library' |
| cmdlist = compiler.get_command_to_archive_shlib() |
| args = [] |
| options = {} |
| self.add_rule(NinjaRule(rule, cmdlist, args, description, **options, extra=None)) |
| |
| args = self.environment.get_build_command() + \ |
| ['--internal', |
| 'symbolextractor', |
| self.environment.get_build_dir(), |
| '$in', |
| '$IMPLIB', |
| '$out'] |
| symrule = 'SHSYM' |
| symcmd = args + ['$CROSS'] |
| syndesc = 'Generating symbol file $out' |
| synstat = 'restat = 1' |
| self.add_rule(NinjaRule(symrule, symcmd, [], syndesc, extra=synstat)) |
| |
| def generate_java_compile_rule(self, compiler): |
| rule = self.compiler_to_rule_name(compiler) |
| command = compiler.get_exelist() + ['$ARGS', '$in'] |
| description = 'Compiling Java sources for $FOR_JAR' |
| self.add_rule(NinjaRule(rule, command, [], description)) |
| |
| def generate_cs_compile_rule(self, compiler: 'CsCompiler') -> None: |
| rule = self.compiler_to_rule_name(compiler) |
| command = compiler.get_exelist() |
| args = ['$ARGS', '$in'] |
| description = 'Compiling C Sharp target $out' |
| self.add_rule(NinjaRule(rule, command, args, description, |
| rspable=mesonlib.is_windows(), |
| rspfile_quote_style=compiler.rsp_file_syntax())) |
| |
| def generate_vala_compile_rules(self, compiler): |
| rule = self.compiler_to_rule_name(compiler) |
| command = compiler.get_exelist() + ['$ARGS', '$in'] |
| description = 'Compiling Vala source $in' |
| self.add_rule(NinjaRule(rule, command, [], description, extra='restat = 1')) |
| |
| def generate_cython_compile_rules(self, compiler: 'Compiler') -> None: |
| rule = self.compiler_to_rule_name(compiler) |
| description = 'Compiling Cython source $in' |
| command = compiler.get_exelist() |
| |
| depargs = compiler.get_dependency_gen_args('$out', '$DEPFILE') |
| depfile = '$out.dep' if depargs else None |
| |
| args = depargs + ['$ARGS', '$in'] |
| args += NinjaCommandArg.list(compiler.get_output_args('$out'), Quoting.none) |
| self.add_rule(NinjaRule(rule, command + args, [], |
| description, |
| depfile=depfile, |
| extra='restat = 1')) |
| |
| def generate_rust_compile_rules(self, compiler): |
| rule = self.compiler_to_rule_name(compiler) |
| command = compiler.get_exelist() + ['$ARGS', '$in'] |
| description = 'Compiling Rust source $in' |
| depfile = '$targetdep' |
| depstyle = 'gcc' |
| self.add_rule(NinjaRule(rule, command, [], description, deps=depstyle, |
| depfile=depfile)) |
| |
| def generate_swift_compile_rules(self, compiler): |
| rule = self.compiler_to_rule_name(compiler) |
| full_exe = self.environment.get_build_command() + [ |
| '--internal', |
| 'dirchanger', |
| '$RUNDIR', |
| ] |
| invoc = full_exe + compiler.get_exelist() |
| command = invoc + ['$ARGS', '$in'] |
| description = 'Compiling Swift source $in' |
| self.add_rule(NinjaRule(rule, command, [], description)) |
| |
| def use_dyndeps_for_fortran(self) -> bool: |
| '''Use the new Ninja feature for scanning dependencies during build, |
| rather than up front. Remove this and all old scanning code once Ninja |
| minimum version is bumped to 1.10.''' |
| return mesonlib.version_compare(self.ninja_version, '>=1.10.0') |
| |
| def generate_fortran_dep_hack(self, crstr: str) -> None: |
| if self.use_dyndeps_for_fortran(): |
| return |
| rule = f'FORTRAN_DEP_HACK{crstr}' |
| if mesonlib.is_windows(): |
| cmd = ['cmd', '/C'] |
| else: |
| cmd = ['true'] |
| self.add_rule_comment(NinjaComment('''Workaround for these issues: |
| https://groups.google.com/forum/#!topic/ninja-build/j-2RfBIOd_8 |
| https://gcc.gnu.org/bugzilla/show_bug.cgi?id=47485''')) |
| self.add_rule(NinjaRule(rule, cmd, [], 'Dep hack', extra='restat = 1')) |
| |
| def generate_llvm_ir_compile_rule(self, compiler): |
| if self.created_llvm_ir_rule[compiler.for_machine]: |
| return |
| rule = self.get_compiler_rule_name('llvm_ir', compiler.for_machine) |
| command = compiler.get_exelist() |
| args = ['$ARGS'] + NinjaCommandArg.list(compiler.get_output_args('$out'), Quoting.none) + compiler.get_compile_only_args() + ['$in'] |
| description = 'Compiling LLVM IR object $in' |
| |
| options = self._rsp_options(compiler) |
| |
| self.add_rule(NinjaRule(rule, command, args, description, **options)) |
| self.created_llvm_ir_rule[compiler.for_machine] = True |
| |
| def generate_compile_rule_for(self, langname, compiler): |
| if langname == 'java': |
| self.generate_java_compile_rule(compiler) |
| return |
| if langname == 'cs': |
| if self.environment.machines.matches_build_machine(compiler.for_machine): |
| self.generate_cs_compile_rule(compiler) |
| return |
| if langname == 'vala': |
| self.generate_vala_compile_rules(compiler) |
| return |
| if langname == 'rust': |
| self.generate_rust_compile_rules(compiler) |
| return |
| if langname == 'swift': |
| if self.environment.machines.matches_build_machine(compiler.for_machine): |
| self.generate_swift_compile_rules(compiler) |
| return |
| if langname == 'cython': |
| self.generate_cython_compile_rules(compiler) |
| return |
| crstr = self.get_rule_suffix(compiler.for_machine) |
| options = self._rsp_options(compiler) |
| if langname == 'fortran': |
| self.generate_fortran_dep_hack(crstr) |
| # gfortran does not update the modification time of *.mod files, therefore restat is needed. |
| # See also: https://github.com/ninja-build/ninja/pull/2275 |
| options['extra'] = 'restat = 1' |
| rule = self.compiler_to_rule_name(compiler) |
| if langname == 'cuda': |
| # for cuda, we manually escape target name ($out) as $CUDA_ESCAPED_TARGET because nvcc doesn't support `-MQ` flag |
| depargs = NinjaCommandArg.list(compiler.get_dependency_gen_args('$CUDA_ESCAPED_TARGET', '$DEPFILE'), Quoting.none) |
| else: |
| depargs = NinjaCommandArg.list(compiler.get_dependency_gen_args('$out', '$DEPFILE'), Quoting.none) |
| command = compiler.get_exelist() |
| args = ['$ARGS'] + depargs + NinjaCommandArg.list(compiler.get_output_args('$out'), Quoting.none) + compiler.get_compile_only_args() + ['$in'] |
| description = f'Compiling {compiler.get_display_language()} object $out' |
| if compiler.get_argument_syntax() == 'msvc': |
| deps = 'msvc' |
| depfile = None |
| else: |
| deps = 'gcc' |
| depfile = '$DEPFILE' |
| self.add_rule(NinjaRule(rule, command, args, description, **options, |
| deps=deps, depfile=depfile)) |
| |
| def generate_pch_rule_for(self, langname, compiler): |
| if langname not in {'c', 'cpp'}: |
| return |
| rule = self.compiler_to_pch_rule_name(compiler) |
| depargs = compiler.get_dependency_gen_args('$out', '$DEPFILE') |
| |
| if compiler.get_argument_syntax() == 'msvc': |
| output = [] |
| else: |
| output = NinjaCommandArg.list(compiler.get_output_args('$out'), Quoting.none) |
| |
| if 'mwcc' in compiler.id: |
| output[0].s = '-precompile' |
| command = compiler.get_exelist() + ['$ARGS'] + depargs + output + ['$in'] # '-c' must be removed |
| else: |
| command = compiler.get_exelist() + ['$ARGS'] + depargs + output + compiler.get_compile_only_args() + ['$in'] |
| description = 'Precompiling header $in' |
| if compiler.get_argument_syntax() == 'msvc': |
| deps = 'msvc' |
| depfile = None |
| else: |
| deps = 'gcc' |
| depfile = '$DEPFILE' |
| self.add_rule(NinjaRule(rule, command, [], description, deps=deps, |
| depfile=depfile)) |
| |
| def generate_scanner_rules(self): |
| rulename = 'depscan' |
| if rulename in self.ruledict: |
| # Scanning command is the same for native and cross compilation. |
| return |
| command = self.environment.get_build_command() + \ |
| ['--internal', 'depscan'] |
| args = ['$picklefile', '$out', '$in'] |
| description = 'Module scanner.' |
| rule = NinjaRule(rulename, command, args, description) |
| self.add_rule(rule) |
| |
| def generate_compile_rules(self): |
| for for_machine in MachineChoice: |
| clist = self.environment.coredata.compilers[for_machine] |
| for langname, compiler in clist.items(): |
| if compiler.get_id() == 'clang': |
| self.generate_llvm_ir_compile_rule(compiler) |
| self.generate_compile_rule_for(langname, compiler) |
| self.generate_pch_rule_for(langname, compiler) |
| for mode in compiler.get_modes(): |
| self.generate_compile_rule_for(langname, mode) |
| |
| def generate_generator_list_rules(self, target): |
| # CustomTargets have already written their rules and |
| # CustomTargetIndexes don't actually get generated, so write rules for |
| # GeneratedLists here |
| for genlist in target.get_generated_sources(): |
| if isinstance(genlist, (build.CustomTarget, build.CustomTargetIndex)): |
| continue |
| self.generate_genlist_for_target(genlist, target) |
| |
| def replace_paths(self, target, args, override_subdir=None): |
| if override_subdir: |
| source_target_dir = os.path.join(self.build_to_src, override_subdir) |
| else: |
| source_target_dir = self.get_target_source_dir(target) |
| relout = self.get_target_private_dir(target) |
| args = [x.replace("@SOURCE_DIR@", self.build_to_src).replace("@BUILD_DIR@", relout) |
| for x in args] |
| args = [x.replace("@CURRENT_SOURCE_DIR@", source_target_dir) for x in args] |
| args = [x.replace("@SOURCE_ROOT@", self.build_to_src).replace("@BUILD_ROOT@", '.') |
| for x in args] |
| args = [x.replace('\\', '/') for x in args] |
| return args |
| |
| def generate_genlist_for_target(self, genlist: build.GeneratedList, target: build.BuildTarget) -> None: |
| for x in genlist.depends: |
| if isinstance(x, build.GeneratedList): |
| self.generate_genlist_for_target(x, target) |
| generator = genlist.get_generator() |
| subdir = genlist.subdir |
| exe = generator.get_exe() |
| infilelist = genlist.get_inputs() |
| outfilelist = genlist.get_outputs() |
| extra_dependencies = self.get_target_depend_files(genlist) |
| for i, curfile in enumerate(infilelist): |
| if len(generator.outputs) == 1: |
| sole_output = os.path.join(self.get_target_private_dir(target), outfilelist[i]) |
| else: |
| sole_output = f'{curfile}' |
| infilename = curfile.rel_to_builddir(self.build_to_src, self.get_target_private_dir(target)) |
| base_args = generator.get_arglist(infilename) |
| outfiles = genlist.get_outputs_for(curfile) |
| outfiles = [os.path.join(self.get_target_private_dir(target), of) for of in outfiles] |
| if generator.depfile is None: |
| rulename = 'CUSTOM_COMMAND' |
| args = base_args |
| else: |
| rulename = 'CUSTOM_COMMAND_DEP' |
| depfilename = generator.get_dep_outname(infilename) |
| depfile = os.path.join(self.get_target_private_dir(target), depfilename) |
| args = [x.replace('@DEPFILE@', depfile) for x in base_args] |
| args = [x.replace("@INPUT@", infilename).replace('@OUTPUT@', sole_output) |
| for x in args] |
| args = self.replace_outputs(args, self.get_target_private_dir(target), outfilelist) |
| # We have consumed output files, so drop them from the list of remaining outputs. |
| if len(generator.outputs) > 1: |
| outfilelist = outfilelist[len(generator.outputs):] |
| args = self.replace_paths(target, args, override_subdir=subdir) |
| cmdlist, reason = self.as_meson_exe_cmdline(exe, |
| self.replace_extra_args(args, genlist), |
| capture=outfiles[0] if generator.capture else None, |
| env=genlist.env) |
| abs_pdir = os.path.join(self.environment.get_build_dir(), self.get_target_dir(target)) |
| os.makedirs(abs_pdir, exist_ok=True) |
| |
| elem = NinjaBuildElement(self.all_outputs, outfiles, rulename, infilename) |
| elem.add_dep([self.get_target_filename(x) for x in generator.depends]) |
| if generator.depfile is not None: |
| elem.add_item('DEPFILE', depfile) |
| if len(extra_dependencies) > 0: |
| elem.add_dep(extra_dependencies) |
| |
| if len(generator.outputs) == 1: |
| what = f'{sole_output!r}' |
| else: |
| # since there are multiple outputs, we log the source that caused the rebuild |
| what = f'from {sole_output!r}' |
| if reason: |
| reason = f' (wrapped by meson {reason})' |
| elem.add_item('DESC', f'Generating {what}{reason}') |
| |
| if isinstance(exe, build.BuildTarget): |
| elem.add_dep(self.get_target_filename(exe)) |
| elem.add_item('COMMAND', cmdlist) |
| self.add_build(elem) |
| |
| def scan_fortran_module_outputs(self, target): |
| """ |
| Find all module and submodule made available in a Fortran code file. |
| """ |
| if self.use_dyndeps_for_fortran(): |
| return |
| compiler = None |
| # TODO other compilers |
| for lang, c in self.environment.coredata.compilers.host.items(): |
| if lang == 'fortran': |
| compiler = c |
| break |
| if compiler is None: |
| self.fortran_deps[target.get_basename()] = {} |
| return |
| |
| modre = re.compile(FORTRAN_MODULE_PAT, re.IGNORECASE) |
| submodre = re.compile(FORTRAN_SUBMOD_PAT, re.IGNORECASE) |
| module_files = {} |
| submodule_files = {} |
| for s in target.get_sources(): |
| # FIXME, does not work for Fortran sources generated by |
| # custom_target() and generator() as those are run after |
| # the configuration (configure_file() is OK) |
| if not compiler.can_compile(s): |
| continue |
| filename = s.absolute_path(self.environment.get_source_dir(), |
| self.environment.get_build_dir()) |
| # Fortran keywords must be ASCII. |
| with open(filename, encoding='ascii', errors='ignore') as f: |
| for line in f: |
| modmatch = modre.match(line) |
| if modmatch is not None: |
| modname = modmatch.group(1).lower() |
| if modname in module_files: |
| raise InvalidArguments( |
| f'Namespace collision: module {modname} defined in ' |
| f'two files {module_files[modname]} and {s}.') |
| module_files[modname] = s |
| else: |
| submodmatch = submodre.match(line) |
| if submodmatch is not None: |
| # '_' is arbitrarily used to distinguish submod from mod. |
| parents = submodmatch.group(1).lower().split(':') |
| submodname = parents[0] + '_' + submodmatch.group(2).lower() |
| |
| if submodname in submodule_files: |
| raise InvalidArguments( |
| f'Namespace collision: submodule {submodname} defined in ' |
| f'two files {submodule_files[submodname]} and {s}.') |
| submodule_files[submodname] = s |
| |
| self.fortran_deps[target.get_basename()] = {**module_files, **submodule_files} |
| |
| def get_fortran_deps(self, compiler: FortranCompiler, src: Path, target) -> T.List[str]: |
| """ |
| Find all module and submodule needed by a Fortran target |
| """ |
| if self.use_dyndeps_for_fortran(): |
| return [] |
| |
| dirname = Path(self.get_target_private_dir(target)) |
| tdeps = self.fortran_deps[target.get_basename()] |
| srcdir = Path(self.source_dir) |
| |
| mod_files = _scan_fortran_file_deps(src, srcdir, dirname, tdeps, compiler) |
| return mod_files |
| |
| def get_no_stdlib_link_args(self, target, linker): |
| if hasattr(linker, 'language') and linker.language in self.build.stdlibs[target.for_machine]: |
| return linker.get_no_stdlib_link_args() |
| return [] |
| |
| def get_compile_debugfile_args(self, compiler, target, objfile): |
| # The way MSVC uses PDB files is documented exactly nowhere so |
| # the following is what we have been able to decipher via |
| # reverse engineering. |
| # |
| # Each object file gets the path of its PDB file written |
| # inside it. This can be either the final PDB (for, say, |
| # foo.exe) or an object pdb (for foo.obj). If the former, then |
| # each compilation step locks the pdb file for writing, which |
| # is a bottleneck and object files from one target cannot be |
| # used in a different target. The latter seems to be the |
| # sensible one (and what Unix does) but there is a catch. If |
| # you try to use precompiled headers MSVC will error out |
| # because both source and pch pdbs go in the same file and |
| # they must be the same. |
| # |
| # This means: |
| # |
| # - pch files must be compiled anew for every object file (negating |
| # the entire point of having them in the first place) |
| # - when using pch, output must go to the target pdb |
| # |
| # Since both of these are broken in some way, use the one that |
| # works for each target. This unfortunately means that you |
| # can't combine pch and object extraction in a single target. |
| # |
| # PDB files also lead to filename collisions. A target foo.exe |
| # has a corresponding foo.pdb. A shared library foo.dll _also_ |
| # has pdb file called foo.pdb. So will a static library |
| # foo.lib, which clobbers both foo.pdb _and_ the dll file's |
| # export library called foo.lib (by default, currently we name |
| # them libfoo.a to avoid this issue). You can give the files |
| # unique names such as foo_exe.pdb but VC also generates a |
| # bunch of other files which take their names from the target |
| # basename (i.e. "foo") and stomp on each other. |
| # |
| # CMake solves this problem by doing two things. First of all |
| # static libraries do not generate pdb files at |
| # all. Presumably you don't need them and VC is smart enough |
| # to look up the original data when linking (speculation, not |
| # tested). The second solution is that you can only have |
| # target named "foo" as an exe, shared lib _or_ static |
| # lib. This makes filename collisions not happen. The downside |
| # is that you can't have an executable foo that uses a shared |
| # library libfoo.so, which is a common idiom on Unix. |
| # |
| # If you feel that the above is completely wrong and all of |
| # this is actually doable, please send patches. |
| |
| if target.has_pch(): |
| tfilename = self.get_target_debug_filename_abs(target) |
| if not tfilename: |
| tfilename = self.get_target_filename_abs(target) |
| return compiler.get_compile_debugfile_args(tfilename, pch=True) |
| else: |
| return compiler.get_compile_debugfile_args(objfile, pch=False) |
| |
| def get_link_debugfile_name(self, linker: T.Union[Compiler, StaticLinker], target: build.BuildTarget) -> T.Optional[str]: |
| filename = self.get_target_debug_filename(target) |
| if filename: |
| return linker.get_link_debugfile_name(filename) |
| return None |
| |
| def get_link_debugfile_args(self, linker: T.Union[Compiler, StaticLinker], target: build.BuildTarget) -> T.List[str]: |
| filename = self.get_target_debug_filename(target) |
| if filename: |
| return linker.get_link_debugfile_args(filename) |
| return [] |
| |
| def generate_llvm_ir_compile(self, target, src: mesonlib.FileOrString): |
| base_proxy = target.get_options() |
| compiler = get_compiler_for_source(target.compilers.values(), src) |
| commands = compiler.compiler_args() |
| # Compiler args for compiling this target |
| commands += compilers.get_base_compile_args(base_proxy, compiler, self.environment) |
| if isinstance(src, File): |
| if src.is_built: |
| src_filename = os.path.join(src.subdir, src.fname) |
| else: |
| src_filename = src.fname |
| elif os.path.isabs(src): |
| src_filename = os.path.basename(src) |
| else: |
| src_filename = src |
| obj_basename = self.canonicalize_filename(src_filename) |
| rel_obj = os.path.join(self.get_target_private_dir(target), obj_basename) |
| rel_obj += '.' + self.environment.machines[target.for_machine].get_object_suffix() |
| commands += self.get_compile_debugfile_args(compiler, target, rel_obj) |
| if isinstance(src, File): |
| if src.is_built: |
| rel_src = src.fname |
| else: |
| rel_src = src.rel_to_builddir(self.build_to_src) |
| else: |
| raise InvalidArguments(f'Invalid source type: {src!r}') |
| # Write the Ninja build command |
| compiler_name = self.get_compiler_rule_name('llvm_ir', compiler.for_machine) |
| element = NinjaBuildElement(self.all_outputs, rel_obj, compiler_name, rel_src) |
| element.add_item('ARGS', commands) |
| self.add_build(element) |
| return (rel_obj, rel_src) |
| |
| @lru_cache(maxsize=None) |
| def generate_inc_dir(self, compiler: 'Compiler', d: str, basedir: str, is_system: bool) -> \ |
| T.Tuple['ImmutableListProtocol[str]', 'ImmutableListProtocol[str]']: |
| # Avoid superfluous '/.' at the end of paths when d is '.' |
| if d not in ('', '.'): |
| expdir = os.path.normpath(os.path.join(basedir, d)) |
| else: |
| expdir = basedir |
| srctreedir = os.path.normpath(os.path.join(self.build_to_src, expdir)) |
| sargs = compiler.get_include_args(srctreedir, is_system) |
| # There may be include dirs where a build directory has not been |
| # created for some source dir. For example if someone does this: |
| # |
| # inc = include_directories('foo/bar/baz') |
| # |
| # But never subdir()s into the actual dir. |
| if os.path.isdir(os.path.join(self.environment.get_build_dir(), expdir)): |
| bargs = compiler.get_include_args(expdir, is_system) |
| else: |
| bargs = [] |
| return (sargs, bargs) |
| |
| def _generate_single_compile(self, target: build.BuildTarget, compiler: Compiler) -> CompilerArgs: |
| commands = self._generate_single_compile_base_args(target, compiler) |
| commands += self._generate_single_compile_target_args(target, compiler) |
| return commands |
| |
| def _generate_single_compile_base_args(self, target: build.BuildTarget, compiler: 'Compiler') -> 'CompilerArgs': |
| base_proxy = target.get_options() |
| # Create an empty commands list, and start adding arguments from |
| # various sources in the order in which they must override each other |
| commands = compiler.compiler_args() |
| # Start with symbol visibility. |
| commands += compiler.gnu_symbol_visibility_args(target.gnu_symbol_visibility) |
| # Add compiler args for compiling this target derived from 'base' build |
| # options passed on the command-line, in default_options, etc. |
| # These have the lowest priority. |
| commands += compilers.get_base_compile_args(base_proxy, |
| compiler, self.environment) |
| return commands |
| |
| @lru_cache(maxsize=None) |
| def _generate_single_compile_target_args(self, target: build.BuildTarget, compiler: Compiler) -> ImmutableListProtocol[str]: |
| # Add compiler args and include paths from several sources; defaults, |
| # build options, external dependencies, etc. |
| commands = self.generate_basic_compiler_args(target, compiler) |
| # Add custom target dirs as includes automatically, but before |
| # target-specific include directories. |
| if target.implicit_include_directories: |
| commands += self.get_custom_target_dir_include_args(target, compiler) |
| # Add include dirs from the `include_directories:` kwarg on the target |
| # and from `include_directories:` of internal deps of the target. |
| # |
| # Target include dirs should override internal deps include dirs. |
| # This is handled in BuildTarget.process_kwargs() |
| # |
| # Include dirs from internal deps should override include dirs from |
| # external deps and must maintain the order in which they are specified. |
| # Hence, we must reverse the list so that the order is preserved. |
| for i in reversed(target.get_include_dirs()): |
| basedir = i.get_curdir() |
| # We should iterate include dirs in reversed orders because |
| # -Ipath will add to begin of array. And without reverse |
| # flags will be added in reversed order. |
| for d in reversed(i.get_incdirs()): |
| # Add source subdir first so that the build subdir overrides it |
| (compile_obj, includeargs) = self.generate_inc_dir(compiler, d, basedir, i.is_system) |
| commands += compile_obj |
| commands += includeargs |
| for d in i.get_extra_build_dirs(): |
| commands += compiler.get_include_args(d, i.is_system) |
| # Add per-target compile args, f.ex, `c_args : ['-DFOO']`. We set these |
| # near the end since these are supposed to override everything else. |
| commands += self.escape_extra_args(target.get_extra_args(compiler.get_language())) |
| |
| # D specific additional flags |
| if compiler.language == 'd': |
| commands += compiler.get_feature_args(target.d_features, self.build_to_src) |
| |
| # Add source dir and build dir. Project-specific and target-specific |
| # include paths must override per-target compile args, include paths |
| # from external dependencies, internal dependencies, and from |
| # per-target `include_directories:` |
| # |
| # We prefer headers in the build dir over the source dir since, for |
| # instance, the user might have an srcdir == builddir Autotools build |
| # in their source tree. Many projects that are moving to Meson have |
| # both Meson and Autotools in parallel as part of the transition. |
| if target.implicit_include_directories: |
| commands += self.get_source_dir_include_args(target, compiler) |
| if target.implicit_include_directories: |
| commands += self.get_build_dir_include_args(target, compiler) |
| # Finally add the private dir for the target to the include path. This |
| # must override everything else and must be the final path added. |
| commands += compiler.get_include_args(self.get_target_private_dir(target), False) |
| return commands |
| |
| # Returns a dictionary, mapping from each compiler src type (e.g. 'c', 'cpp', etc.) to a list of compiler arg strings |
| # used for that respective src type. |
| # Currently used for the purpose of populating VisualStudio intellisense fields but possibly useful in other scenarios. |
| def generate_common_compile_args_per_src_type(self, target: build.BuildTarget) -> dict[str, list[str]]: |
| src_type_to_args = {} |
| |
| use_pch = self.target_uses_pch(target) |
| |
| for src_type_str in target.compilers.keys(): |
| compiler = target.compilers[src_type_str] |
| commands = self._generate_single_compile_base_args(target, compiler) |
| |
| # Include PCH header as first thing as it must be the first one or it will be |
| # ignored by gcc https://gcc.gnu.org/bugzilla/show_bug.cgi?id=100462 |
| if use_pch and 'mw' not in compiler.id: |
| commands += self.get_pch_include_args(compiler, target) |
| |
| commands += self._generate_single_compile_target_args(target, compiler) |
| |
| # Metrowerks compilers require PCH include args to come after intraprocedural analysis args |
| if use_pch and 'mw' in compiler.id: |
| commands += self.get_pch_include_args(compiler, target) |
| |
| commands = commands.compiler.compiler_args(commands) |
| |
| src_type_to_args[src_type_str] = commands.to_native() |
| return src_type_to_args |
| |
| def generate_single_compile(self, target: build.BuildTarget, src, |
| is_generated: bool = False, header_deps=None, |
| order_deps: T.Optional[T.List['mesonlib.FileOrString']] = None, |
| extra_args: T.Optional[T.List[str]] = None, |
| unity_sources: T.Optional[T.List[mesonlib.FileOrString]] = None, |
| ) -> T.Tuple[str, str]: |
| """ |
| Compiles C/C++, ObjC/ObjC++, Fortran, and D sources |
| """ |
| header_deps = header_deps if header_deps is not None else [] |
| order_deps = order_deps if order_deps is not None else [] |
| |
| if isinstance(src, str) and src.endswith('.h'): |
| raise AssertionError(f'BUG: sources should not contain headers {src!r}') |
| |
| compiler = get_compiler_for_source(target.compilers.values(), src) |
| commands = self._generate_single_compile_base_args(target, compiler) |
| |
| # Include PCH header as first thing as it must be the first one or it will be |
| # ignored by gcc https://gcc.gnu.org/bugzilla/show_bug.cgi?id=100462 |
| use_pch = self.target_uses_pch(target) |
| if use_pch and 'mw' not in compiler.id: |
| commands += self.get_pch_include_args(compiler, target) |
| |
| commands += self._generate_single_compile_target_args(target, compiler) |
| |
| # Metrowerks compilers require PCH include args to come after intraprocedural analysis args |
| if use_pch and 'mw' in compiler.id: |
| commands += self.get_pch_include_args(compiler, target) |
| |
| commands = commands.compiler.compiler_args(commands) |
| |
| # Create introspection information |
| if is_generated is False: |
| self.create_target_source_introspection(target, compiler, commands, [src], [], unity_sources) |
| else: |
| self.create_target_source_introspection(target, compiler, commands, [], [src], unity_sources) |
| |
| build_dir = self.environment.get_build_dir() |
| if isinstance(src, File): |
| rel_src = src.rel_to_builddir(self.build_to_src) |
| if os.path.isabs(rel_src): |
| # Source files may not be from the source directory if they originate in source-only libraries, |
| # so we can't assert that the absolute path is anywhere in particular. |
| if src.is_built: |
| assert rel_src.startswith(build_dir) |
| rel_src = rel_src[len(build_dir) + 1:] |
| elif is_generated: |
| raise AssertionError(f'BUG: broken generated source file handling for {src!r}') |
| else: |
| raise InvalidArguments(f'Invalid source type: {src!r}') |
| obj_basename = self.object_filename_from_source(target, src) |
| rel_obj = os.path.join(self.get_target_private_dir(target), obj_basename) |
| dep_file = compiler.depfile_for_object(rel_obj) |
| |
| # Add MSVC debug file generation compile flags: /Fd /FS |
| commands += self.get_compile_debugfile_args(compiler, target, rel_obj) |
| |
| # PCH handling |
| if self.target_uses_pch(target): |
| pchlist = target.get_pch(compiler.language) |
| else: |
| pchlist = [] |
| if not pchlist: |
| pch_dep = [] |
| elif compiler.id == 'intel': |
| pch_dep = [] |
| else: |
| arr = [] |
| i = os.path.join(self.get_target_private_dir(target), compiler.get_pch_name(pchlist[0])) |
| arr.append(i) |
| pch_dep = arr |
| |
| compiler_name = self.compiler_to_rule_name(compiler) |
| extra_deps = [] |
| if compiler.get_language() == 'fortran': |
| # Can't read source file to scan for deps if it's generated later |
| # at build-time. Skip scanning for deps, and just set the module |
| # outdir argument instead. |
| # https://github.com/mesonbuild/meson/issues/1348 |
| if not is_generated: |
| abs_src = Path(build_dir) / rel_src |
| extra_deps += self.get_fortran_deps(compiler, abs_src, target) |
| if not self.use_dyndeps_for_fortran(): |
| # Dependency hack. Remove once multiple outputs in Ninja is fixed: |
| # https://groups.google.com/forum/#!topic/ninja-build/j-2RfBIOd_8 |
| for modname, srcfile in self.fortran_deps[target.get_basename()].items(): |
| modfile = os.path.join(self.get_target_private_dir(target), |
| compiler.module_name_to_filename(modname)) |
| |
| if srcfile == src: |
| crstr = self.get_rule_suffix(target.for_machine) |
| depelem = NinjaBuildElement(self.all_outputs, |
| modfile, |
| 'FORTRAN_DEP_HACK' + crstr, |
| rel_obj) |
| self.add_build(depelem) |
| commands += compiler.get_module_outdir_args(self.get_target_private_dir(target)) |
| if extra_args is not None: |
| commands.extend(extra_args) |
| |
| element = NinjaBuildElement(self.all_outputs, rel_obj, compiler_name, rel_src) |
| self.add_header_deps(target, element, header_deps) |
| for d in extra_deps: |
| element.add_dep(d) |
| for d in order_deps: |
| if isinstance(d, File): |
| d = d.rel_to_builddir(self.build_to_src) |
| elif not self.has_dir_part(d): |
| d = os.path.join(self.get_target_private_dir(target), d) |
| element.add_orderdep(d) |
| element.add_dep(pch_dep) |
| for i in self.get_fortran_orderdeps(target, compiler): |
| element.add_orderdep(i) |
| if dep_file: |
| element.add_item('DEPFILE', dep_file) |
| if compiler.get_language() == 'cuda': |
| # for cuda, we manually escape target name ($out) as $CUDA_ESCAPED_TARGET because nvcc doesn't support `-MQ` flag |
| def quote_make_target(targetName: str) -> str: |
| # this escape implementation is taken from llvm |
| result = '' |
| for (i, c) in enumerate(targetName): |
| if c in {' ', '\t'}: |
| # Escape the preceding backslashes |
| for j in range(i - 1, -1, -1): |
| if targetName[j] == '\\': |
| result += '\\' |
| else: |
| break |
| # Escape the space/tab |
| result += '\\' |
| elif c == '$': |
| result += '$' |
| elif c == '#': |
| result += '\\' |
| result += c |
| return result |
| element.add_item('CUDA_ESCAPED_TARGET', quote_make_target(rel_obj)) |
| element.add_item('ARGS', commands) |
| |
| self.add_dependency_scanner_entries_to_element(target, compiler, element, src) |
| self.add_build(element) |
| assert isinstance(rel_obj, str) |
| assert isinstance(rel_src, str) |
| return (rel_obj, rel_src.replace('\\', '/')) |
| |
| def add_dependency_scanner_entries_to_element(self, target: build.BuildTarget, compiler, element, src): |
| if not self.should_use_dyndeps_for_target(target): |
| return |
| if isinstance(target, build.CompileTarget): |
| return |
| extension = os.path.splitext(src.fname)[1][1:] |
| if extension != 'C': |
| extension = extension.lower() |
| if not (extension in compilers.lang_suffixes['fortran'] or extension in compilers.lang_suffixes['cpp']): |
| return |
| dep_scan_file = self.get_dep_scan_file_for(target) |
| element.add_item('dyndep', dep_scan_file) |
| element.add_orderdep(dep_scan_file) |
| |
| def get_dep_scan_file_for(self, target: build.BuildTarget) -> str: |
| return os.path.join(self.get_target_private_dir(target), 'depscan.dd') |
| |
| def add_header_deps(self, target, ninja_element, header_deps): |
| for d in header_deps: |
| if isinstance(d, File): |
| d = d.rel_to_builddir(self.build_to_src) |
| elif not self.has_dir_part(d): |
| d = os.path.join(self.get_target_private_dir(target), d) |
| ninja_element.add_dep(d) |
| |
| def has_dir_part(self, fname: mesonlib.FileOrString) -> bool: |
| # FIXME FIXME: The usage of this is a terrible and unreliable hack |
| if isinstance(fname, File): |
| return fname.subdir != '' |
| return has_path_sep(fname) |
| |
| # Fortran is a bit weird (again). When you link against a library, just compiling a source file |
| # requires the mod files that are output when single files are built. To do this right we would need to |
| # scan all inputs and write out explicit deps for each file. That is stoo slow and too much effort so |
| # instead just have an ordered dependency on the library. This ensures all required mod files are created. |
| # The real deps are then detected via dep file generation from the compiler. This breaks on compilers that |
| # produce incorrect dep files but such is life. |
| def get_fortran_orderdeps(self, target, compiler): |
| if compiler.language != 'fortran': |
| return [] |
| return [ |
| os.path.join(self.get_target_dir(lt), lt.get_filename()) |
| for lt in itertools.chain(target.link_targets, target.link_whole_targets) |
| ] |
| |
| def generate_msvc_pch_command(self, target, compiler, pch): |
| header = pch[0] |
| pchname = compiler.get_pch_name(header) |
| dst = os.path.join(self.get_target_private_dir(target), pchname) |
| |
| commands = [] |
| commands += self.generate_basic_compiler_args(target, compiler) |
| |
| if len(pch) == 1: |
| # Auto generate PCH. |
| source = self.create_msvc_pch_implementation(target, compiler.get_language(), pch[0]) |
| pch_header_dir = os.path.dirname(os.path.join(self.build_to_src, target.get_source_subdir(), header)) |
| commands += compiler.get_include_args(pch_header_dir, False) |
| else: |
| source = os.path.join(self.build_to_src, target.get_source_subdir(), pch[1]) |
| |
| just_name = os.path.basename(header) |
| (objname, pch_args) = compiler.gen_pch_args(just_name, source, dst) |
| commands += pch_args |
| commands += self._generate_single_compile(target, compiler) |
| commands += self.get_compile_debugfile_args(compiler, target, objname) |
| dep = dst + '.' + compiler.get_depfile_suffix() |
| |
| link_objects = [objname] if compiler.should_link_pch_object() else [] |
| |
| return commands, dep, dst, link_objects, source |
| |
| def generate_gcc_pch_command(self, target, compiler, pch): |
| commands = self._generate_single_compile(target, compiler) |
| if pch.split('.')[-1] == 'h' and compiler.language == 'cpp': |
| # Explicitly compile pch headers as C++. If Clang is invoked in C++ mode, it actually warns if |
| # this option is not set, and for gcc it also makes sense to use it. |
| commands += ['-x', 'c++-header'] |
| dst = os.path.join(self.get_target_private_dir(target), |
| os.path.basename(pch) + '.' + compiler.get_pch_suffix()) |
| dep = dst + '.' + compiler.get_depfile_suffix() |
| return commands, dep, dst, [] # Gcc does not create an object file during pch generation. |
| |
| def generate_mwcc_pch_command(self, target, compiler, pch): |
| commands = self._generate_single_compile(target, compiler) |
| dst = os.path.join(self.get_target_private_dir(target), |
| os.path.basename(pch) + '.' + compiler.get_pch_suffix()) |
| dep = os.path.splitext(dst)[0] + '.' + compiler.get_depfile_suffix() |
| return commands, dep, dst, [] # mwcc compilers do not create an object file during pch generation. |
| |
| def generate_pch(self, target, header_deps=None): |
| header_deps = header_deps if header_deps is not None else [] |
| pch_objects = [] |
| for lang in ['c', 'cpp']: |
| pch = target.get_pch(lang) |
| if not pch: |
| continue |
| if not has_path_sep(pch[0]) or not has_path_sep(pch[-1]): |
| msg = f'Precompiled header of {target.get_basename()!r} must not be in the same ' \ |
| 'directory as source, please put it in a subdirectory.' |
| raise InvalidArguments(msg) |
| compiler: Compiler = target.compilers[lang] |
| if compiler.get_argument_syntax() == 'msvc': |
| (commands, dep, dst, objs, src) = self.generate_msvc_pch_command(target, compiler, pch) |
| extradep = os.path.join(self.build_to_src, target.get_source_subdir(), pch[0]) |
| elif compiler.id == 'intel': |
| # Intel generates on target generation |
| continue |
| elif 'mwcc' in compiler.id: |
| src = os.path.join(self.build_to_src, target.get_source_subdir(), pch[0]) |
| (commands, dep, dst, objs) = self.generate_mwcc_pch_command(target, compiler, pch[0]) |
| extradep = None |
| else: |
| src = os.path.join(self.build_to_src, target.get_source_subdir(), pch[0]) |
| (commands, dep, dst, objs) = self.generate_gcc_pch_command(target, compiler, pch[0]) |
| extradep = None |
| pch_objects += objs |
| rulename = self.compiler_to_pch_rule_name(compiler) |
| elem = NinjaBuildElement(self.all_outputs, objs + [dst], rulename, src) |
| if extradep is not None: |
| elem.add_dep(extradep) |
| self.add_header_deps(target, elem, header_deps) |
| elem.add_item('ARGS', commands) |
| elem.add_item('DEPFILE', dep) |
| self.add_build(elem) |
| return pch_objects |
| |
| def get_target_shsym_filename(self, target): |
| # Always name the .symbols file after the primary build output because it always exists |
| targetdir = self.get_target_private_dir(target) |
| return os.path.join(targetdir, target.get_filename() + '.symbols') |
| |
| def generate_shsym(self, target): |
| target_file = self.get_target_filename(target) |
| symname = self.get_target_shsym_filename(target) |
| elem = NinjaBuildElement(self.all_outputs, symname, 'SHSYM', target_file) |
| # The library we will actually link to, which is an import library on Windows (not the DLL) |
| elem.add_item('IMPLIB', self.get_target_filename_for_linking(target)) |
| if self.environment.is_cross_build(): |
| elem.add_item('CROSS', '--cross-host=' + self.environment.machines[target.for_machine].system) |
| self.add_build(elem) |
| |
| def get_import_filename(self, target): |
| return os.path.join(self.get_target_dir(target), target.import_filename) |
| |
| def get_target_type_link_args(self, target, linker): |
| commands = [] |
| if isinstance(target, build.Executable): |
| # Currently only used with the Swift compiler to add '-emit-executable' |
| commands += linker.get_std_exe_link_args() |
| # If export_dynamic, add the appropriate linker arguments |
| if target.export_dynamic: |
| commands += linker.gen_export_dynamic_link_args(self.environment) |
| # If implib, and that's significant on this platform (i.e. Windows using either GCC or Visual Studio) |
| if target.import_filename: |
| commands += linker.gen_import_library_args(self.get_import_filename(target)) |
| if target.pie: |
| commands += linker.get_pie_link_args() |
| if target.vs_module_defs and hasattr(linker, 'gen_vs_module_defs_args'): |
| commands += linker.gen_vs_module_defs_args(target.vs_module_defs.rel_to_builddir(self.build_to_src)) |
| elif isinstance(target, build.SharedLibrary): |
| if isinstance(target, build.SharedModule): |
| commands += linker.get_std_shared_module_link_args(target.get_options()) |
| else: |
| commands += linker.get_std_shared_lib_link_args() |
| # All shared libraries are PIC |
| commands += linker.get_pic_args() |
| if not isinstance(target, build.SharedModule) or target.force_soname: |
| # Add -Wl,-soname arguments on Linux, -install_name on OS X |
| commands += linker.get_soname_args( |
| self.environment, target.prefix, target.name, target.suffix, |
| target.soversion, target.darwin_versions) |
| # This is only visited when building for Windows using either GCC or Visual Studio |
| if target.vs_module_defs and hasattr(linker, 'gen_vs_module_defs_args'): |
| commands += linker.gen_vs_module_defs_args(target.vs_module_defs.rel_to_builddir(self.build_to_src)) |
| # This is only visited when building for Windows using either GCC or Visual Studio |
| if target.import_filename: |
| commands += linker.gen_import_library_args(self.get_import_filename(target)) |
| elif isinstance(target, build.StaticLibrary): |
| produce_thin_archive = self.allow_thin_archives[target.for_machine] and not target.should_install() |
| commands += linker.get_std_link_args(self.environment, produce_thin_archive) |
| else: |
| raise RuntimeError('Unknown build target type.') |
| return commands |
| |
| def get_target_type_link_args_post_dependencies(self, target, linker): |
| commands = [] |
| if isinstance(target, build.Executable): |
| # If win_subsystem is significant on this platform, add the appropriate linker arguments. |
| # Unfortunately this can't be done in get_target_type_link_args, because some misguided |
| # libraries (such as SDL2) add -mwindows to their link flags. |
| m = self.environment.machines[target.for_machine] |
| |
| if m.is_windows() or m.is_cygwin(): |
| commands += linker.get_win_subsystem_args(target.win_subsystem) |
| return commands |
| |
| def get_link_whole_args(self, linker, target): |
| use_custom = False |
| if linker.id == 'msvc': |
| # Expand our object lists manually if we are on pre-Visual Studio 2015 Update 2 |
| # (incidentally, the "linker" here actually refers to cl.exe) |
| if mesonlib.version_compare(linker.version, '<19.00.23918'): |
| use_custom = True |
| |
| if use_custom: |
| objects_from_static_libs: T.List[ExtractedObjects] = [] |
| for dep in target.link_whole_targets: |
| l = dep.extract_all_objects(False) |
| objects_from_static_libs += self.determine_ext_objs(l, '') |
| objects_from_static_libs.extend(self.flatten_object_list(dep)[0]) |
| |
| return objects_from_static_libs |
| else: |
| target_args = self.build_target_link_arguments(linker, target.link_whole_targets) |
| return linker.get_link_whole_for(target_args) if target_args else [] |
| |
| @lru_cache(maxsize=None) |
| def guess_library_absolute_path(self, linker, libname, search_dirs, patterns) -> Path: |
| from ..compilers.c import CCompiler |
| for d in search_dirs: |
| for p in patterns: |
| trial = CCompiler._get_trials_from_pattern(p, d, libname) |
| if not trial: |
| continue |
| trial = CCompiler._get_file_from_list(self.environment, trial) |
| if not trial: |
| continue |
| # Return the first result |
| return trial |
| |
| def guess_external_link_dependencies(self, linker, target, commands, internal): |
| # Ideally the linker would generate dependency information that could be used. |
| # But that has 2 problems: |
| # * currently ld cannot create dependency information in a way that ninja can use: |
| # https://sourceware.org/bugzilla/show_bug.cgi?id=22843 |
| # * Meson optimizes libraries from the same build using the symbol extractor. |
| # Just letting ninja use ld generated dependencies would undo this optimization. |
| search_dirs = OrderedSet() |
| libs = OrderedSet() |
| absolute_libs = [] |
| |
| build_dir = self.environment.get_build_dir() |
| # the following loop sometimes consumes two items from command in one pass |
| it = iter(linker.native_args_to_unix(commands)) |
| for item in it: |
| if item in internal and not item.startswith('-'): |
| continue |
| |
| if item.startswith('-L'): |
| if len(item) > 2: |
| path = item[2:] |
| else: |
| try: |
| path = next(it) |
| except StopIteration: |
| mlog.warning("Generated linker command has -L argument without following path") |
| break |
| if not os.path.isabs(path): |
| path = os.path.join(build_dir, path) |
| search_dirs.add(path) |
| elif item.startswith('-l'): |
| if len(item) > 2: |
| lib = item[2:] |
| else: |
| try: |
| lib = next(it) |
| except StopIteration: |
| mlog.warning("Generated linker command has '-l' argument without following library name") |
| break |
| libs.add(lib) |
| elif os.path.isabs(item) and self.environment.is_library(item) and os.path.isfile(item): |
| absolute_libs.append(item) |
| |
| guessed_dependencies = [] |
| # TODO The get_library_naming requirement currently excludes link targets that use d or fortran as their main linker |
| try: |
| static_patterns = linker.get_library_naming(self.environment, LibType.STATIC, strict=True) |
| shared_patterns = linker.get_library_naming(self.environment, LibType.SHARED, strict=True) |
| search_dirs = tuple(search_dirs) + tuple(linker.get_library_dirs(self.environment)) |
| for libname in libs: |
| # be conservative and record most likely shared and static resolution, because we don't know exactly |
| # which one the linker will prefer |
| staticlibs = self.guess_library_absolute_path(linker, libname, |
| search_dirs, static_patterns) |
| sharedlibs = self.guess_library_absolute_path(linker, libname, |
| search_dirs, shared_patterns) |
| if staticlibs: |
| guessed_dependencies.append(staticlibs.resolve().as_posix()) |
| if sharedlibs: |
| guessed_dependencies.append(sharedlibs.resolve().as_posix()) |
| except (mesonlib.MesonException, AttributeError) as e: |
| if 'get_library_naming' not in str(e): |
| raise |
| |
| return guessed_dependencies + absolute_libs |
| |
| def generate_prelink(self, target, obj_list): |
| assert isinstance(target, build.StaticLibrary) |
| prelink_name = os.path.join(self.get_target_private_dir(target), target.name + '-prelink.o') |
| elem = NinjaBuildElement(self.all_outputs, [prelink_name], 'CUSTOM_COMMAND', obj_list) |
| |
| prelinker = target.get_prelinker() |
| cmd = prelinker.exelist[:] |
| cmd += prelinker.get_prelink_args(prelink_name, obj_list) |
| |
| cmd = self.replace_paths(target, cmd) |
| elem.add_item('COMMAND', cmd) |
| elem.add_item('description', f'Prelinking {prelink_name}.') |
| self.add_build(elem) |
| return [prelink_name] |
| |
| def generate_link(self, target: build.BuildTarget, outname, obj_list, linker: T.Union['Compiler', 'StaticLinker'], extra_args=None, stdlib_args=None): |
| extra_args = extra_args if extra_args is not None else [] |
| stdlib_args = stdlib_args if stdlib_args is not None else [] |
| implicit_outs = [] |
| if isinstance(target, build.StaticLibrary): |
| linker_base = 'STATIC' |
| else: |
| linker_base = linker.get_language() # Fixme. |
| if isinstance(target, build.SharedLibrary): |
| self.generate_shsym(target) |
| crstr = self.get_rule_suffix(target.for_machine) |
| linker_rule = linker_base + '_LINKER' + crstr |
| # Create an empty commands list, and start adding link arguments from |
| # various sources in the order in which they must override each other |
| # starting from hard-coded defaults followed by build options and so on. |
| # |
| # Once all the linker options have been passed, we will start passing |
| # libraries and library paths from internal and external sources. |
| commands = linker.compiler_args() |
| # First, the trivial ones that are impossible to override. |
| # |
| # Add linker args for linking this target derived from 'base' build |
| # options passed on the command-line, in default_options, etc. |
| # These have the lowest priority. |
| if isinstance(target, build.StaticLibrary): |
| commands += linker.get_base_link_args(target.get_options()) |
| else: |
| commands += compilers.get_base_link_args(target.get_options(), |
| linker, |
| isinstance(target, build.SharedModule), |
| self.environment.get_build_dir()) |
| # Add -nostdlib if needed; can't be overridden |
| commands += self.get_no_stdlib_link_args(target, linker) |
| # Add things like /NOLOGO; usually can't be overridden |
| commands += linker.get_linker_always_args() |
| # Add buildtype linker args: optimization level, etc. |
| commands += linker.get_optimization_link_args(target.get_option(OptionKey('optimization'))) |
| # Add /DEBUG and the pdb filename when using MSVC |
| if target.get_option(OptionKey('debug')): |
| commands += self.get_link_debugfile_args(linker, target) |
| debugfile = self.get_link_debugfile_name(linker, target) |
| if debugfile is not None: |
| implicit_outs += [debugfile] |
| # Add link args specific to this BuildTarget type, such as soname args, |
| # PIC, import library generation, etc. |
| commands += self.get_target_type_link_args(target, linker) |
| # Archives that are copied wholesale in the result. Must be before any |
| # other link targets so missing symbols from whole archives are found in those. |
| if not isinstance(target, build.StaticLibrary): |
| commands += self.get_link_whole_args(linker, target) |
| |
| if not isinstance(target, build.StaticLibrary): |
| # Add link args added using add_project_link_arguments() |
| commands += self.build.get_project_link_args(linker, target.subproject, target.for_machine) |
| # Add link args added using add_global_link_arguments() |
| # These override per-project link arguments |
| commands += self.build.get_global_link_args(linker, target.for_machine) |
| # Link args added from the env: LDFLAGS. We want these to override |
| # all the defaults but not the per-target link args. |
| commands += self.environment.coredata.get_external_link_args(target.for_machine, linker.get_language()) |
| |
| # Now we will add libraries and library paths from various sources |
| |
| # Set runtime-paths so we can run executables without needing to set |
| # LD_LIBRARY_PATH, etc in the environment. Doesn't work on Windows. |
| if has_path_sep(target.name): |
| # Target names really should not have slashes in them, but |
| # unfortunately we did not check for that and some downstream projects |
| # now have them. Once slashes are forbidden, remove this bit. |
| target_slashname_workaround_dir = os.path.join( |
| os.path.dirname(target.name), |
| self.get_target_dir(target)) |
| else: |
| target_slashname_workaround_dir = self.get_target_dir(target) |
| (rpath_args, target.rpath_dirs_to_remove) = ( |
| linker.build_rpath_args(self.environment, |
| self.environment.get_build_dir(), |
| target_slashname_workaround_dir, |
| self.determine_rpath_dirs(target), |
| target.build_rpath, |
| target.install_rpath)) |
| commands += rpath_args |
| |
| # Add link args to link to all internal libraries (link_with:) and |
| # internal dependencies needed by this target. |
| if linker_base == 'STATIC': |
| # Link arguments of static libraries are not put in the command |
| # line of the library. They are instead appended to the command |
| # line where the static library is used. |
| dependencies = [] |
| else: |
| dependencies = target.get_dependencies() |
| internal = self.build_target_link_arguments(linker, dependencies) |
| commands += internal |
| # Only non-static built targets need link args and link dependencies |
| if not isinstance(target, build.StaticLibrary): |
| # For 'automagic' deps: Boost and GTest. Also dependency('threads'). |
| # pkg-config puts the thread flags itself via `Cflags:` |
| |
| commands += linker.get_target_link_args(target) |
| # External deps must be last because target link libraries may depend on them. |
| for dep in target.get_external_deps(): |
| # Extend without reordering or de-dup to preserve `-L -l` sets |
| # https://github.com/mesonbuild/meson/issues/1718 |
| commands.extend_preserving_lflags(linker.get_dependency_link_args(dep)) |
| for d in target.get_dependencies(): |
| if isinstance(d, build.StaticLibrary): |
| for dep in d.get_external_deps(): |
| commands.extend_preserving_lflags(linker.get_dependency_link_args(dep)) |
| |
| # Add link args specific to this BuildTarget type that must not be overridden by dependencies |
| commands += self.get_target_type_link_args_post_dependencies(target, linker) |
| |
| # Add link args for c_* or cpp_* build options. Currently this only |
| # adds c_winlibs and cpp_winlibs when building for Windows. This needs |
| # to be after all internal and external libraries so that unresolved |
| # symbols from those can be found here. This is needed when the |
| # *_winlibs that we want to link to are static mingw64 libraries. |
| if isinstance(linker, Compiler): |
| # The static linker doesn't know what language it is building, so we |
| # don't know what option. Fortunately, it doesn't care to see the |
| # language-specific options either. |
| # |
| # We shouldn't check whether we are making a static library, because |
| # in the LTO case we do use a real compiler here. |
| commands += linker.get_option_link_args(target.get_options()) |
| |
| dep_targets = [] |
| dep_targets.extend(self.guess_external_link_dependencies(linker, target, commands, internal)) |
| |
| # Add libraries generated by custom targets |
| custom_target_libraries = self.get_custom_target_provided_libraries(target) |
| commands += extra_args |
| commands += custom_target_libraries |
| commands += stdlib_args # Standard library arguments go last, because they never depend on anything. |
| dep_targets.extend([self.get_dependency_filename(t) for t in dependencies]) |
| dep_targets.extend([self.get_dependency_filename(t) |
| for t in target.link_depends]) |
| elem = NinjaBuildElement(self.all_outputs, outname, linker_rule, obj_list, implicit_outs=implicit_outs) |
| elem.add_dep(dep_targets + custom_target_libraries) |
| elem.add_item('LINK_ARGS', commands) |
| self.create_target_linker_introspection(target, linker, commands) |
| return elem |
| |
| def get_dependency_filename(self, t): |
| if isinstance(t, build.SharedLibrary): |
| return self.get_target_shsym_filename(t) |
| elif isinstance(t, mesonlib.File): |
| if t.is_built: |
| return t.relative_name() |
| else: |
| return t.absolute_path(self.environment.get_source_dir(), |
| self.environment.get_build_dir()) |
| return self.get_target_filename(t) |
| |
| def generate_shlib_aliases(self, target, outdir): |
| for alias, to, tag in target.get_aliases(): |
| aliasfile = os.path.join(outdir, alias) |
| abs_aliasfile = os.path.join(self.environment.get_build_dir(), outdir, alias) |
| try: |
| os.remove(abs_aliasfile) |
| except Exception: |
| pass |
| try: |
| os.symlink(to, abs_aliasfile) |
| except NotImplementedError: |
| mlog.debug("Library versioning disabled because symlinks are not supported.") |
| except OSError: |
| mlog.debug("Library versioning disabled because we do not have symlink creation privileges.") |
| else: |
| self.implicit_meson_outs.append(aliasfile) |
| |
| def generate_custom_target_clean(self, trees: T.List[str]) -> str: |
| e = self.create_phony_target('clean-ctlist', 'CUSTOM_COMMAND', 'PHONY') |
| d = CleanTrees(self.environment.get_build_dir(), trees) |
| d_file = os.path.join(self.environment.get_scratch_dir(), 'cleantrees.dat') |
| e.add_item('COMMAND', self.environment.get_build_command() + ['--internal', 'cleantrees', d_file]) |
| e.add_item('description', 'Cleaning custom target directories') |
| self.add_build(e) |
| # Write out the data file passed to the script |
| with open(d_file, 'wb') as ofile: |
| pickle.dump(d, ofile) |
| return 'clean-ctlist' |
| |
| def generate_gcov_clean(self) -> None: |
| gcno_elem = self.create_phony_target('clean-gcno', 'CUSTOM_COMMAND', 'PHONY') |
| gcno_elem.add_item('COMMAND', mesonlib.get_meson_command() + ['--internal', 'delwithsuffix', '.', 'gcno']) |
| gcno_elem.add_item('description', 'Deleting gcno files') |
| self.add_build(gcno_elem) |
| |
| gcda_elem = self.create_phony_target('clean-gcda', 'CUSTOM_COMMAND', 'PHONY') |
| gcda_elem.add_item('COMMAND', mesonlib.get_meson_command() + ['--internal', 'delwithsuffix', '.', 'gcda']) |
| gcda_elem.add_item('description', 'Deleting gcda files') |
| self.add_build(gcda_elem) |
| |
| def get_user_option_args(self): |
| cmds = [] |
| for k, v in self.environment.coredata.optstore.items(): |
| if k.is_project(): |
| cmds.append('-D' + str(k) + '=' + (v.value if isinstance(v.value, str) else str(v.value).lower())) |
| # The order of these arguments must be the same between runs of Meson |
| # to ensure reproducible output. The order we pass them shouldn't |
| # affect behavior in any other way. |
| return sorted(cmds) |
| |
| def generate_dist(self) -> None: |
| elem = self.create_phony_target('dist', 'CUSTOM_COMMAND', 'PHONY') |
| elem.add_item('DESC', 'Creating source packages') |
| elem.add_item('COMMAND', self.environment.get_build_command() + ['dist']) |
| elem.add_item('pool', 'console') |
| self.add_build(elem) |
| |
| def generate_scanbuild(self) -> None: |
| if not environment.detect_scanbuild(): |
| return |
| if 'scan-build' in self.all_outputs: |
| return |
| cmd = self.environment.get_build_command() + \ |
| ['--internal', 'scanbuild', self.environment.source_dir, self.environment.build_dir, self.build.get_subproject_dir()] + \ |
| self.environment.get_build_command() + ['setup'] + self.get_user_option_args() |
| elem = self.create_phony_target('scan-build', 'CUSTOM_COMMAND', 'PHONY') |
| elem.add_item('COMMAND', cmd) |
| elem.add_item('pool', 'console') |
| self.add_build(elem) |
| |
| def generate_clangtool(self, name: str, extra_arg: T.Optional[str] = None) -> None: |
| target_name = 'clang-' + name |
| extra_args = [] |
| if extra_arg: |
| target_name += f'-{extra_arg}' |
| extra_args.append(f'--{extra_arg}') |
| if not os.path.exists(os.path.join(self.environment.source_dir, '.clang-' + name)) and \ |
| not os.path.exists(os.path.join(self.environment.source_dir, '_clang-' + name)): |
| return |
| if target_name in self.all_outputs: |
| return |
| cmd = self.environment.get_build_command() + \ |
| ['--internal', 'clang' + name, self.environment.source_dir, self.environment.build_dir] + \ |
| extra_args |
| elem = self.create_phony_target(target_name, 'CUSTOM_COMMAND', 'PHONY') |
| elem.add_item('COMMAND', cmd) |
| elem.add_item('pool', 'console') |
| self.add_build(elem) |
| |
| def generate_clangformat(self) -> None: |
| if not environment.detect_clangformat(): |
| return |
| self.generate_clangtool('format') |
| self.generate_clangtool('format', 'check') |
| |
| def generate_clangtidy(self) -> None: |
| import shutil |
| if not shutil.which('clang-tidy'): |
| return |
| self.generate_clangtool('tidy') |
| self.generate_clangtool('tidy', 'fix') |
| |
| def generate_tags(self, tool: str, target_name: str) -> None: |
| import shutil |
| if not shutil.which(tool): |
| return |
| if target_name in self.all_outputs: |
| return |
| cmd = self.environment.get_build_command() + \ |
| ['--internal', 'tags', tool, self.environment.source_dir] |
| elem = self.create_phony_target(target_name, 'CUSTOM_COMMAND', 'PHONY') |
| elem.add_item('COMMAND', cmd) |
| elem.add_item('pool', 'console') |
| self.add_build(elem) |
| |
| # For things like scan-build and other helper tools we might have. |
| def generate_utils(self) -> None: |
| self.generate_scanbuild() |
| self.generate_clangformat() |
| self.generate_clangtidy() |
| self.generate_tags('etags', 'TAGS') |
| self.generate_tags('ctags', 'ctags') |
| self.generate_tags('cscope', 'cscope') |
| cmd = self.environment.get_build_command() + ['--internal', 'uninstall'] |
| elem = self.create_phony_target('uninstall', 'CUSTOM_COMMAND', 'PHONY') |
| elem.add_item('COMMAND', cmd) |
| elem.add_item('pool', 'console') |
| self.add_build(elem) |
| |
| def generate_ending(self) -> None: |
| for targ, deps in [ |
| ('all', self.get_build_by_default_targets()), |
| ('meson-test-prereq', self.get_testlike_targets()), |
| ('meson-benchmark-prereq', self.get_testlike_targets(True))]: |
| targetlist = [] |
| # These must also be built by default. |
| # XXX: Sometime in the future these should be built only before running tests. |
| if targ == 'all': |
| targetlist.extend(['meson-test-prereq', 'meson-benchmark-prereq']) |
| for t in deps.values(): |
| # Add the first output of each target to the 'all' target so that |
| # they are all built |
| #Add archive file if shared library in AIX for build all. |
| if isinstance(t, build.SharedLibrary) and t.aix_so_archive: |
| if self.environment.machines[t.for_machine].is_aix(): |
| linker, stdlib_args = self.determine_linker_and_stdlib_args(t) |
| t.get_outputs()[0] = linker.get_archive_name(t.get_outputs()[0]) |
| targetlist.append(os.path.join(self.get_target_dir(t), t.get_outputs()[0])) |
| |
| elem = NinjaBuildElement(self.all_outputs, targ, 'phony', targetlist) |
| self.add_build(elem) |
| |
| elem = self.create_phony_target('clean', 'CUSTOM_COMMAND', 'PHONY') |
| elem.add_item('COMMAND', self.ninja_command + ['-t', 'clean']) |
| elem.add_item('description', 'Cleaning') |
| |
| # If we have custom targets in this project, add all their outputs to |
| # the list that is passed to the `cleantrees.py` script. The script |
| # will manually delete all custom_target outputs that are directories |
| # instead of files. This is needed because on platforms other than |
| # Windows, Ninja only deletes directories while cleaning if they are |
| # empty. https://github.com/mesonbuild/meson/issues/1220 |
| ctlist = [] |
| for t in self.build.get_targets().values(): |
| if isinstance(t, build.CustomTarget): |
| # Create a list of all custom target outputs |
| for o in t.get_outputs(): |
| ctlist.append(os.path.join(self.get_target_dir(t), o)) |
| if ctlist: |
| elem.add_dep(self.generate_custom_target_clean(ctlist)) |
| |
| if OptionKey('b_coverage') in self.environment.coredata.optstore and \ |
| self.environment.coredata.optstore.get_value('b_coverage'): |
| self.generate_gcov_clean() |
| elem.add_dep('clean-gcda') |
| elem.add_dep('clean-gcno') |
| self.add_build(elem) |
| |
| deps = self.get_regen_filelist() |
| elem = NinjaBuildElement(self.all_outputs, 'build.ninja', 'REGENERATE_BUILD', deps) |
| elem.add_item('pool', 'console') |
| self.add_build(elem) |
| |
| # If these files used to be explicitly created, they need to appear on the build graph somehow, |
| # otherwise cleandead deletes them. See https://github.com/ninja-build/ninja/issues/2299 |
| if self.implicit_meson_outs: |
| elem = NinjaBuildElement(self.all_outputs, 'meson-implicit-outs', 'phony', self.implicit_meson_outs) |
| self.add_build(elem) |
| |
| elem = NinjaBuildElement(self.all_outputs, 'reconfigure', 'REGENERATE_BUILD', 'PHONY') |
| elem.add_item('pool', 'console') |
| self.add_build(elem) |
| |
| elem = NinjaBuildElement(self.all_outputs, deps, 'phony', '') |
| self.add_build(elem) |
| |
| def get_introspection_data(self, target_id: str, target: build.Target) -> T.List[T.Dict[str, T.Union[bool, str, T.List[T.Union[str, T.Dict[str, T.Union[str, T.List[str], bool]]]]]]]: |
| data = self.introspection_data.get(target_id) |
| if not data: |
| return super().get_introspection_data(target_id, target) |
| |
| return list(data.values()) |
| |
| |
| def _scan_fortran_file_deps(src: Path, srcdir: Path, dirname: Path, tdeps, compiler) -> T.List[str]: |
| """ |
| scan a Fortran file for dependencies. Needs to be distinct from target |
| to allow for recursion induced by `include` statements.er |
| |
| It makes a number of assumptions, including |
| |
| * `use`, `module`, `submodule` name is not on a continuation line |
| |
| Regex |
| ----- |
| |
| * `incre` works for `#include "foo.f90"` and `include "foo.f90"` |
| * `usere` works for legacy and Fortran 2003 `use` statements |
| * `submodre` is for Fortran >= 2008 `submodule` |
| """ |
| |
| incre = re.compile(FORTRAN_INCLUDE_PAT, re.IGNORECASE) |
| usere = re.compile(FORTRAN_USE_PAT, re.IGNORECASE) |
| submodre = re.compile(FORTRAN_SUBMOD_PAT, re.IGNORECASE) |
| |
| mod_files = [] |
| src = Path(src) |
| with src.open(encoding='ascii', errors='ignore') as f: |
| for line in f: |
| # included files |
| incmatch = incre.match(line) |
| if incmatch is not None: |
| incfile = src.parent / incmatch.group(1) |
| # NOTE: src.parent is most general, in particular for CMake subproject with Fortran file |
| # having an `include 'foo.f'` statement. |
| if incfile.suffix.lower()[1:] in compiler.file_suffixes: |
| mod_files.extend(_scan_fortran_file_deps(incfile, srcdir, dirname, tdeps, compiler)) |
| # modules |
| usematch = usere.match(line) |
| if usematch is not None: |
| usename = usematch.group(1).lower() |
| if usename == 'intrinsic': # this keeps the regex simpler |
| continue |
| if usename not in tdeps: |
| # The module is not provided by any source file. This |
| # is due to: |
| # a) missing file/typo/etc |
| # b) using a module provided by the compiler, such as |
| # OpenMP |
| # There's no easy way to tell which is which (that I |
| # know of) so just ignore this and go on. Ideally we |
| # would print a warning message to the user but this is |
| # a common occurrence, which would lead to lots of |
| # distracting noise. |
| continue |
| srcfile = srcdir / tdeps[usename].fname |
| if not srcfile.is_file(): |
| if srcfile.name != src.name: # generated source file |
| pass |
| else: # subproject |
| continue |
| elif srcfile.samefile(src): # self-reference |
| continue |
| |
| mod_name = compiler.module_name_to_filename(usename) |
| mod_files.append(str(dirname / mod_name)) |
| else: # submodules |
| submodmatch = submodre.match(line) |
| if submodmatch is not None: |
| parents = submodmatch.group(1).lower().split(':') |
| assert len(parents) in {1, 2}, ( |
| 'submodule ancestry must be specified as' |
| f' ancestor:parent but Meson found {parents}') |
| |
| ancestor_child = '_'.join(parents) |
| if ancestor_child not in tdeps: |
| raise MesonException("submodule {} relies on ancestor module {} that was not found.".format(submodmatch.group(2).lower(), ancestor_child.split('_', maxsplit=1)[0])) |
| submodsrcfile = srcdir / tdeps[ancestor_child].fname |
| if not submodsrcfile.is_file(): |
| if submodsrcfile.name != src.name: # generated source file |
| pass |
| else: # subproject |
| continue |
| elif submodsrcfile.samefile(src): # self-reference |
| continue |
| mod_name = compiler.module_name_to_filename(ancestor_child) |
| mod_files.append(str(dirname / mod_name)) |
| return mod_files |