| # SPDX-License-Identifier: Apache-2.0 |
| # Copyright 2019 The Meson development team |
| |
| from __future__ import annotations |
| |
| from .common import CMakeException, CMakeBuildFile, CMakeConfiguration |
| import typing as T |
| from .. import mlog |
| from pathlib import Path |
| import json |
| import re |
| |
| STRIP_KEYS = ['cmake', 'reply', 'backtrace', 'backtraceGraph', 'version'] |
| |
| class CMakeFileAPI: |
| def __init__(self, build_dir: Path): |
| self.build_dir = build_dir |
| self.api_base_dir = self.build_dir / '.cmake' / 'api' / 'v1' |
| self.request_dir = self.api_base_dir / 'query' / 'client-meson' |
| self.reply_dir = self.api_base_dir / 'reply' |
| self.cmake_sources: T.List[CMakeBuildFile] = [] |
| self.cmake_configurations: T.List[CMakeConfiguration] = [] |
| self.project_version = '' |
| self.kind_resolver_map = { |
| 'codemodel': self._parse_codemodel, |
| 'cache': self._parse_cache, |
| 'cmakeFiles': self._parse_cmakeFiles, |
| } |
| |
| def get_cmake_sources(self) -> T.List[CMakeBuildFile]: |
| return self.cmake_sources |
| |
| def get_cmake_configurations(self) -> T.List[CMakeConfiguration]: |
| return self.cmake_configurations |
| |
| def get_project_version(self) -> str: |
| return self.project_version |
| |
| def setup_request(self) -> None: |
| self.request_dir.mkdir(parents=True, exist_ok=True) |
| |
| query = { |
| 'requests': [ |
| {'kind': 'codemodel', 'version': {'major': 2, 'minor': 0}}, |
| {'kind': 'cache', 'version': {'major': 2, 'minor': 0}}, |
| {'kind': 'cmakeFiles', 'version': {'major': 1, 'minor': 0}}, |
| ] |
| } |
| |
| query_file = self.request_dir / 'query.json' |
| query_file.write_text(json.dumps(query, indent=2), encoding='utf-8') |
| |
| def load_reply(self) -> None: |
| if not self.reply_dir.is_dir(): |
| raise CMakeException('No response from the CMake file API') |
| |
| root = None |
| reg_index = re.compile(r'^index-.*\.json$') |
| for i in self.reply_dir.iterdir(): |
| if reg_index.match(i.name): |
| root = i |
| break |
| |
| if not root: |
| raise CMakeException('Failed to find the CMake file API index') |
| |
| index = self._reply_file_content(root) # Load the root index |
| index = self._strip_data(index) # Avoid loading duplicate files |
| index = self._resolve_references(index) # Load everything |
| index = self._strip_data(index) # Strip unused data (again for loaded files) |
| |
| # Debug output |
| debug_json = self.build_dir / '..' / 'fileAPI.json' |
| debug_json = debug_json.resolve() |
| debug_json.write_text(json.dumps(index, indent=2), encoding='utf-8') |
| mlog.cmd_ci_include(debug_json.as_posix()) |
| |
| # parse the JSON |
| for i in index['objects']: |
| assert isinstance(i, dict) |
| assert 'kind' in i |
| assert i['kind'] in self.kind_resolver_map |
| |
| self.kind_resolver_map[i['kind']](i) |
| |
| def _parse_codemodel(self, data: T.Dict[str, T.Any]) -> None: |
| assert 'configurations' in data |
| assert 'paths' in data |
| |
| source_dir = data['paths']['source'] |
| build_dir = data['paths']['build'] |
| |
| # The file API output differs quite a bit from the server |
| # output. It is more flat than the server output and makes |
| # heavy use of references. Here these references are |
| # resolved and the resulting data structure is identical |
| # to the CMake serve output. |
| |
| def helper_parse_dir(dir_entry: T.Dict[str, T.Any]) -> T.Tuple[Path, Path]: |
| src_dir = Path(dir_entry.get('source', '.')) |
| bld_dir = Path(dir_entry.get('build', '.')) |
| src_dir = src_dir if src_dir.is_absolute() else source_dir / src_dir |
| bld_dir = bld_dir if bld_dir.is_absolute() else build_dir / bld_dir |
| src_dir = src_dir.resolve() |
| bld_dir = bld_dir.resolve() |
| |
| return src_dir, bld_dir |
| |
| def parse_sources(comp_group: T.Dict[str, T.Any], tgt: T.Dict[str, T.Any]) -> T.Tuple[T.List[Path], T.List[Path], T.List[int]]: |
| gen = [] |
| src = [] |
| idx = [] |
| |
| src_list_raw = tgt.get('sources', []) |
| for i in comp_group.get('sourceIndexes', []): |
| if i >= len(src_list_raw) or 'path' not in src_list_raw[i]: |
| continue |
| if src_list_raw[i].get('isGenerated', False): |
| gen += [Path(src_list_raw[i]['path'])] |
| else: |
| src += [Path(src_list_raw[i]['path'])] |
| idx += [i] |
| |
| return src, gen, idx |
| |
| def parse_target(tgt: T.Dict[str, T.Any]) -> T.Dict[str, T.Any]: |
| src_dir, bld_dir = helper_parse_dir(cnf.get('paths', {})) |
| |
| # Parse install paths (if present) |
| install_paths = [] |
| if 'install' in tgt: |
| prefix = Path(tgt['install']['prefix']['path']) |
| install_paths = [prefix / x['path'] for x in tgt['install']['destinations']] |
| install_paths = list(set(install_paths)) |
| |
| # On the first look, it looks really nice that the CMake devs have |
| # decided to use arrays for the linker flags. However, this feeling |
| # soon turns into despair when you realize that there only one entry |
| # per type in most cases, and we still have to do manual string splitting. |
| link_flags = [] |
| link_libs = [] |
| for i in tgt.get('link', {}).get('commandFragments', []): |
| if i['role'] == 'flags': |
| link_flags += [i['fragment']] |
| elif i['role'] == 'libraries': |
| link_libs += [i['fragment']] |
| elif i['role'] == 'libraryPath': |
| link_flags += ['-L{}'.format(i['fragment'])] |
| elif i['role'] == 'frameworkPath': |
| link_flags += ['-F{}'.format(i['fragment'])] |
| for i in tgt.get('archive', {}).get('commandFragments', []): |
| if i['role'] == 'flags': |
| link_flags += [i['fragment']] |
| |
| # TODO The `dependencies` entry is new in the file API. |
| # maybe we can make use of that in addition to the |
| # implicit dependency detection |
| tgt_data = { |
| 'artifacts': [Path(x.get('path', '')) for x in tgt.get('artifacts', [])], |
| 'sourceDirectory': src_dir, |
| 'buildDirectory': bld_dir, |
| 'name': tgt.get('name', ''), |
| 'fullName': tgt.get('nameOnDisk', ''), |
| 'hasInstallRule': 'install' in tgt, |
| 'installPaths': install_paths, |
| 'linkerLanguage': tgt.get('link', {}).get('language', 'CXX'), |
| 'linkLibraries': ' '.join(link_libs), # See previous comment block why we join the array |
| 'linkFlags': ' '.join(link_flags), # See previous comment block why we join the array |
| 'type': tgt.get('type', 'EXECUTABLE'), |
| 'fileGroups': [], |
| } |
| |
| processed_src_idx = [] |
| for cg in tgt.get('compileGroups', []): |
| # Again, why an array, when there is usually only one element |
| # and arguments are separated with spaces... |
| flags = [] |
| for i in cg.get('compileCommandFragments', []): |
| flags += [i['fragment']] |
| |
| cg_data = { |
| 'defines': [x.get('define', '') for x in cg.get('defines', [])], |
| 'compileFlags': ' '.join(flags), |
| 'language': cg.get('language', 'C'), |
| 'isGenerated': None, # Set later, flag is stored per source file |
| 'sources': [], |
| 'includePath': cg.get('includes', []), |
| } |
| |
| normal_src, generated_src, src_idx = parse_sources(cg, tgt) |
| if normal_src: |
| cg_data = dict(cg_data) |
| cg_data['isGenerated'] = False |
| cg_data['sources'] = normal_src |
| tgt_data['fileGroups'] += [cg_data] |
| if generated_src: |
| cg_data = dict(cg_data) |
| cg_data['isGenerated'] = True |
| cg_data['sources'] = generated_src |
| tgt_data['fileGroups'] += [cg_data] |
| processed_src_idx += src_idx |
| |
| # Object libraries have no compile groups, only source groups. |
| # So we add all the source files to a dummy source group that were |
| # not found in the previous loop |
| normal_src = [] |
| generated_src = [] |
| for idx, src in enumerate(tgt.get('sources', [])): |
| if idx in processed_src_idx: |
| continue |
| |
| if src.get('isGenerated', False): |
| generated_src += [src['path']] |
| else: |
| normal_src += [src['path']] |
| |
| if normal_src: |
| tgt_data['fileGroups'] += [{ |
| 'isGenerated': False, |
| 'sources': normal_src, |
| }] |
| if generated_src: |
| tgt_data['fileGroups'] += [{ |
| 'isGenerated': True, |
| 'sources': generated_src, |
| }] |
| return tgt_data |
| |
| def parse_project(pro: T.Dict[str, T.Any]) -> T.Dict[str, T.Any]: |
| # Only look at the first directory specified in directoryIndexes |
| # TODO Figure out what the other indexes are there for |
| p_src_dir = source_dir |
| p_bld_dir = build_dir |
| try: |
| p_src_dir, p_bld_dir = helper_parse_dir(cnf['directories'][pro['directoryIndexes'][0]]) |
| except (IndexError, KeyError): |
| pass |
| |
| pro_data = { |
| 'name': pro.get('name', ''), |
| 'sourceDirectory': p_src_dir, |
| 'buildDirectory': p_bld_dir, |
| 'targets': [], |
| } |
| |
| for ref in pro.get('targetIndexes', []): |
| tgt = {} |
| try: |
| tgt = cnf['targets'][ref] |
| except (IndexError, KeyError): |
| pass |
| pro_data['targets'] += [parse_target(tgt)] |
| |
| return pro_data |
| |
| for cnf in data.get('configurations', []): |
| cnf_data = { |
| 'name': cnf.get('name', ''), |
| 'projects': [], |
| } |
| |
| for pro in cnf.get('projects', []): |
| cnf_data['projects'] += [parse_project(pro)] |
| |
| self.cmake_configurations += [CMakeConfiguration(cnf_data)] |
| |
| def _parse_cmakeFiles(self, data: T.Dict[str, T.Any]) -> None: |
| assert 'inputs' in data |
| assert 'paths' in data |
| |
| src_dir = Path(data['paths']['source']) |
| |
| for i in data['inputs']: |
| path = Path(i['path']) |
| path = path if path.is_absolute() else src_dir / path |
| self.cmake_sources += [CMakeBuildFile(path, i.get('isCMake', False), i.get('isGenerated', False))] |
| |
| def _parse_cache(self, data: T.Dict[str, T.Any]) -> None: |
| assert 'entries' in data |
| |
| for e in data['entries']: |
| if e['name'] == 'CMAKE_PROJECT_VERSION': |
| self.project_version = e['value'] |
| |
| def _strip_data(self, data: T.Any) -> T.Any: |
| if isinstance(data, list): |
| for idx, i in enumerate(data): |
| data[idx] = self._strip_data(i) |
| |
| elif isinstance(data, dict): |
| new = {} |
| for key, val in data.items(): |
| if key not in STRIP_KEYS: |
| new[key] = self._strip_data(val) |
| data = new |
| |
| return data |
| |
| def _resolve_references(self, data: T.Any) -> T.Any: |
| if isinstance(data, list): |
| for idx, i in enumerate(data): |
| data[idx] = self._resolve_references(i) |
| |
| elif isinstance(data, dict): |
| # Check for the "magic" reference entry and insert |
| # it into the root data dict |
| if 'jsonFile' in data: |
| data.update(self._reply_file_content(data['jsonFile'])) |
| |
| for key, val in data.items(): |
| data[key] = self._resolve_references(val) |
| |
| return data |
| |
| def _reply_file_content(self, filename: Path) -> T.Dict[str, T.Any]: |
| real_path = self.reply_dir / filename |
| if not real_path.exists(): |
| raise CMakeException(f'File "{real_path}" does not exist') |
| |
| data = json.loads(real_path.read_text(encoding='utf-8')) |
| assert isinstance(data, dict) |
| for i in data.keys(): |
| assert isinstance(i, str) |
| return data |