| # Copyright (C) 2019 The Android Open Source Project |
| # |
| # Licensed under the Apache License, Version 2.0 (the "License"); |
| # you may not use this file except in compliance with the License. |
| # You may obtain a copy of the License at |
| # |
| # http://www.apache.org/licenses/LICENSE-2.0 |
| # |
| # Unless required by applicable law or agreed to in writing, software |
| # distributed under the License is distributed on an "AS IS" BASIS, |
| # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| # See the License for the specific language governing permissions and |
| # limitations under the License. |
| |
| # A collection of utilities for extracting build rule information from GN |
| # projects. |
| |
| from __future__ import print_function |
| import collections |
| from compat import iteritems |
| import errno |
| import filecmp |
| import json |
| import os |
| import re |
| import shutil |
| import subprocess |
| import sys |
| from typing import Dict |
| from typing import Optional |
| from typing import Set |
| from typing import Tuple |
| |
| BUILDFLAGS_TARGET = '//gn:gen_buildflags' |
| GEN_VERSION_TARGET = '//src/base:version_gen_h' |
| TARGET_TOOLCHAIN = '//gn/standalone/toolchain:gcc_like_host' |
| HOST_TOOLCHAIN = '//gn/standalone/toolchain:gcc_like_host' |
| LINKER_UNIT_TYPES = ('executable', 'shared_library', 'static_library') |
| |
| # TODO(primiano): investigate these, they require further componentization. |
| ODR_VIOLATION_IGNORE_TARGETS = { |
| '//test/cts:perfetto_cts_deps', |
| '//:perfetto_integrationtests', |
| } |
| |
| |
| def _check_command_output(cmd, cwd): |
| try: |
| output = subprocess.check_output(cmd, stderr=subprocess.STDOUT, cwd=cwd) |
| except subprocess.CalledProcessError as e: |
| print( |
| 'Command "{}" failed in {}:'.format(' '.join(cmd), cwd), |
| file=sys.stderr) |
| print(e.output.decode(), file=sys.stderr) |
| sys.exit(1) |
| else: |
| return output.decode() |
| |
| |
| def repo_root(): |
| """Returns an absolute path to the repository root.""" |
| return os.path.join( |
| os.path.realpath(os.path.dirname(__file__)), os.path.pardir) |
| |
| |
| def _tool_path(name, system_buildtools=False): |
| # Pass-through to use name if the caller requests to use the system |
| # toolchain. |
| if system_buildtools: |
| return [name] |
| wrapper = os.path.abspath( |
| os.path.join(repo_root(), 'tools', 'run_buildtools_binary.py')) |
| return ['python3', wrapper, name] |
| |
| |
| def prepare_out_directory(gn_args, |
| name, |
| root=repo_root(), |
| system_buildtools=False): |
| """Creates the JSON build description by running GN. |
| |
| Returns (path, desc) where |path| is the location of the output directory |
| and |desc| is the JSON build description. |
| """ |
| out = os.path.join(root, 'out', name) |
| try: |
| os.makedirs(out) |
| except OSError as e: |
| if e.errno != errno.EEXIST: |
| raise |
| _check_command_output( |
| _tool_path('gn', system_buildtools) + |
| ['gen', out, '--args=%s' % gn_args], |
| cwd=repo_root()) |
| return out |
| |
| |
| def load_build_description(out, system_buildtools=False): |
| """Creates the JSON build description by running GN.""" |
| desc = _check_command_output( |
| _tool_path('gn', system_buildtools) + |
| ['desc', out, '--format=json', '--all-toolchains', '//*'], |
| cwd=repo_root()) |
| return json.loads(desc) |
| |
| |
| def create_build_description(gn_args, root=repo_root()): |
| """Prepares a GN out directory and loads the build description from it. |
| |
| The temporary out directory is automatically deleted. |
| """ |
| out = prepare_out_directory(gn_args, 'tmp.gn_utils', root=root) |
| try: |
| return load_build_description(out) |
| finally: |
| shutil.rmtree(out) |
| |
| |
| def build_targets(out, targets, quiet=False, system_buildtools=False): |
| """Runs ninja to build a list of GN targets in the given out directory. |
| |
| Compiling these targets is required so that we can include any generated |
| source files in the amalgamated result. |
| """ |
| targets = [t.replace('//', '') for t in targets] |
| with open(os.devnull, 'w') as devnull: |
| stdout = devnull if quiet else None |
| cmd = _tool_path('ninja', system_buildtools) + targets |
| subprocess.check_call(cmd, cwd=os.path.abspath(out), stdout=stdout) |
| |
| |
| def compute_source_dependencies(out, system_buildtools=False): |
| """For each source file, computes a set of headers it depends on.""" |
| ninja_deps = _check_command_output( |
| _tool_path('ninja', system_buildtools) + ['-t', 'deps'], cwd=out) |
| deps = {} |
| current_source = None |
| for line in ninja_deps.split('\n'): |
| filename = os.path.relpath(os.path.join(out, line.strip()), repo_root()) |
| if not line or line[0] != ' ': |
| current_source = None |
| continue |
| elif not current_source: |
| # We're assuming the source file is always listed before the |
| # headers. |
| assert os.path.splitext(line)[1] in ['.c', '.cc', '.cpp', '.S'] |
| current_source = filename |
| deps[current_source] = [] |
| else: |
| assert current_source |
| deps[current_source].append(filename) |
| return deps |
| |
| |
| def label_to_path(label): |
| """Turn a GN output label (e.g., //some_dir/file.cc) into a path.""" |
| assert label.startswith('//') |
| return label[2:] |
| |
| |
| def label_without_toolchain(label): |
| """Strips the toolchain from a GN label. |
| |
| Return a GN label (e.g //buildtools:protobuf(//gn/standalone/toolchain: |
| gcc_like_host) without the parenthesised toolchain part. |
| """ |
| return label.split('(')[0] |
| |
| |
| def label_to_target_name_with_path(label): |
| """ |
| Turn a GN label into a target name involving the full path. |
| e.g., //src/perfetto:tests -> src_perfetto_tests |
| """ |
| name = re.sub(r'^//:?', '', label) |
| name = re.sub(r'[^a-zA-Z0-9_]', '_', name) |
| return name |
| |
| |
| def gen_buildflags(gn_args, target_file): |
| """Generates the perfetto_build_flags.h for the given config. |
| |
| target_file: the path, relative to the repo root, where the generated |
| buildflag header will be copied into. |
| """ |
| tmp_out = prepare_out_directory(gn_args, 'tmp.gen_buildflags') |
| build_targets(tmp_out, [BUILDFLAGS_TARGET], quiet=True) |
| src = os.path.join(tmp_out, 'gen', 'build_config', 'perfetto_build_flags.h') |
| shutil.copy(src, os.path.join(repo_root(), target_file)) |
| shutil.rmtree(tmp_out) |
| |
| |
| def check_or_commit_generated_files(tmp_files, check): |
| """Checks that gen files are unchanged or renames them to the final location |
| |
| Takes in input a list of 'xxx.swp' files that have been written. |
| If check == False, it renames xxx.swp -> xxx. |
| If check == True, it just checks that the contents of 'xxx.swp' == 'xxx'. |
| Returns 0 if no diff was detected, 1 otherwise (to be used as exit code). |
| """ |
| res = 0 |
| for tmp_file in tmp_files: |
| assert (tmp_file.endswith('.swp')) |
| target_file = os.path.relpath(tmp_file[:-4]) |
| if check: |
| if not filecmp.cmp(tmp_file, target_file): |
| sys.stderr.write('%s needs to be regenerated\n' % target_file) |
| res = 1 |
| os.unlink(tmp_file) |
| else: |
| os.rename(tmp_file, target_file) |
| return res |
| |
| |
| class ODRChecker(object): |
| """Detects ODR violations in linker units |
| |
| When we turn GN source sets into Soong & Bazel file groups, there is the risk |
| to create ODR violations by including the same file group into different |
| linker unit (this is because other build systems don't have a concept |
| equivalent to GN's source_set). This class navigates the transitive |
| dependencies (mostly static libraries) of a target and detects if multiple |
| paths end up including the same file group. This is to avoid situations like: |
| |
| traced.exe -> base(file group) |
| traced.exe -> libperfetto(static lib) -> base(file group) |
| """ |
| |
| def __init__(self, gn: 'GnParser', target_name: str): |
| self.gn = gn |
| self.root = gn.get_target(target_name) |
| self.source_sets: Dict[str, Set[str]] = collections.defaultdict(set) |
| self.deps_visited = set() |
| self.source_set_hdr_only = {} |
| |
| self._visit(target_name) |
| num_violations = 0 |
| if target_name in ODR_VIOLATION_IGNORE_TARGETS: |
| return |
| for sset, paths in self.source_sets.items(): |
| if self.is_header_only(sset): |
| continue |
| if len(paths) != 1: |
| num_violations += 1 |
| print( |
| 'ODR violation in target %s, multiple paths include %s:\n %s' % |
| (target_name, sset, '\n '.join(paths)), |
| file=sys.stderr) |
| if num_violations > 0: |
| raise Exception('%d ODR violations detected. Build generation aborted' % |
| num_violations) |
| |
| def _visit(self, target_name: str, parent_path=''): |
| target = self.gn.get_target(target_name) |
| path = ((parent_path + ' > ') if parent_path else '') + target_name |
| if not target: |
| raise Exception('Cannot find target %s' % target_name) |
| for ssdep in target.transitive_source_set_deps(): |
| name_and_path = '%s (via %s)' % (target_name, path) |
| self.source_sets[ssdep.name].add(name_and_path) |
| deps = set(target.non_proto_or_source_set_deps()).union( |
| target.transitive_proto_deps()) - self.deps_visited |
| for dep in deps: |
| if dep.type == 'executable': |
| continue # Execs are strong boundaries and don't cause ODR violations. |
| # static_library dependencies should reset the path. It doesn't matter if |
| # we get to a source file via: |
| # source_set1 > static_lib > source.cc OR |
| # source_set1 > source_set2 > static_lib > source.cc |
| # This is NOT an ODR violation because source.cc is linked from the same |
| # static library |
| next_parent_path = path if dep.type != 'static_library' else '' |
| self.deps_visited.add(dep.name) |
| self._visit(dep.name, next_parent_path) |
| |
| def is_header_only(self, source_set_name: str): |
| cached = self.source_set_hdr_only.get(source_set_name) |
| if cached is not None: |
| return cached |
| target = self.gn.get_target(source_set_name) |
| if target.type != 'source_set': |
| raise TypeError('%s is not a source_set' % source_set_name) |
| res = all(src.endswith('.h') for src in target.sources) |
| self.source_set_hdr_only[source_set_name] = res |
| return res |
| |
| |
| class GnParser(object): |
| """A parser with some cleverness for GN json desc files |
| |
| The main goals of this parser are: |
| 1) Deal with the fact that other build systems don't have an equivalent |
| notion to GN's source_set. Conversely to Bazel's and Soong's filegroups, |
| GN source_sets expect that dependencies, cflags and other source_set |
| properties propagate up to the linker unit (static_library, executable or |
| shared_library). This parser simulates the same behavior: when a |
| source_set is encountered, some of its variables (cflags and such) are |
| copied up to the dependent targets. This is to allow gen_xxx to create |
| one filegroup for each source_set and then squash all the other flags |
| onto the linker unit. |
| 2) Detect and special-case protobuf targets, figuring out the protoc-plugin |
| being used. |
| """ |
| |
| class Target(object): |
| """Reperesents A GN target. |
| |
| Maked properties are propagated up the dependency chain when a |
| source_set dependency is encountered. |
| """ |
| |
| def __init__(self, name, type): |
| self.name = name # e.g. //src/ipc:ipc |
| |
| VALID_TYPES = ('static_library', 'shared_library', 'executable', 'group', |
| 'action', 'source_set', 'proto_library', 'generated_file') |
| assert (type in VALID_TYPES) |
| self.type = type |
| self.testonly = False |
| self.toolchain = None |
| |
| # These are valid only for type == proto_library. |
| # This is typically: 'proto', 'protozero', 'ipc'. |
| self.proto_plugin: Optional[str] = None |
| self.proto_paths = set() |
| self.proto_exports = set() |
| |
| self.sources = set() |
| # TODO(primiano): consider whether the public section should be part of |
| # bubbled-up sources. |
| self.public_headers = set() # 'public' |
| |
| # These are valid only for type == 'action' |
| self.data = set() |
| self.inputs = set() |
| self.outputs = set() |
| self.script = None |
| self.args = [] |
| self.custom_action_type = None |
| self.python_main = None |
| |
| # These variables are propagated up when encountering a dependency |
| # on a source_set target. |
| self.cflags = set() |
| self.defines = set() |
| self.deps: Set[GnParser.Target] = set() |
| self.transitive_deps: Set[GnParser.Target] = set() |
| self.libs = set() |
| self.include_dirs = set() |
| self.ldflags = set() |
| |
| # Deps on //gn:xxx have this flag set to True. These dependencies |
| # are special because they pull third_party code from buildtools/. |
| # We don't want to keep recursing into //buildtools in generators, |
| # this flag is used to stop the recursion and create an empty |
| # placeholder target once we hit //gn:protoc or similar. |
| self.is_third_party_dep_ = False |
| |
| def non_proto_or_source_set_deps(self): |
| return set(d for d in self.deps |
| if d.type != 'proto_library' and d.type != 'source_set') |
| |
| def proto_deps(self): |
| return set(d for d in self.deps if d.type == 'proto_library') |
| |
| def transitive_proto_deps(self): |
| return set(d for d in self.transitive_deps if d.type == 'proto_library') |
| |
| def transitive_cpp_proto_deps(self): |
| return set( |
| d for d in self.transitive_deps if d.type == 'proto_library' and |
| d.proto_plugin != 'descriptor' and d.proto_plugin != 'source_set') |
| |
| def transitive_source_set_deps(self): |
| return set(d for d in self.transitive_deps if d.type == 'source_set') |
| |
| def __lt__(self, other): |
| if isinstance(other, self.__class__): |
| return self.name < other.name |
| raise TypeError( |
| '\'<\' not supported between instances of \'%s\' and \'%s\'' % |
| (type(self).__name__, type(other).__name__)) |
| |
| def __repr__(self): |
| return json.dumps({ |
| k: (list(sorted(v)) if isinstance(v, set) else v) |
| for (k, v) in iteritems(self.__dict__) |
| }, |
| indent=4, |
| sort_keys=True) |
| |
| def update(self, other): |
| for key in ('cflags', 'data', 'defines', 'deps', 'include_dirs', |
| 'ldflags', 'transitive_deps', 'libs', 'proto_paths'): |
| self.__dict__[key].update(other.__dict__.get(key, [])) |
| |
| def __init__(self, gn_desc): |
| self.gn_desc_ = gn_desc |
| self.all_targets = {} |
| self.linker_units = {} # Executables, shared or static libraries. |
| self.source_sets = {} |
| self.actions = {} |
| self.proto_libs = {} |
| |
| def get_target(self, gn_target_name: str) -> Target: |
| """Returns a Target object from the fully qualified GN target name. |
| |
| It bubbles up variables from source_set dependencies as described in the |
| class-level comments. |
| """ |
| target = self.all_targets.get(gn_target_name) |
| if target is not None: |
| return target # Target already processed. |
| |
| desc = self.gn_desc_[gn_target_name] |
| target = GnParser.Target(gn_target_name, desc['type']) |
| target.testonly = desc.get('testonly', False) |
| target.toolchain = desc.get('toolchain', None) |
| self.all_targets[gn_target_name] = target |
| |
| # We should never have GN targets directly depend on buidtools. They |
| # should hop via //gn:xxx, so we can give generators an opportunity to |
| # override them. |
| assert (not gn_target_name.startswith('//buildtools')) |
| |
| # Don't descend further into third_party targets. Genrators are supposed |
| # to either ignore them or route to other externally-provided targets. |
| if gn_target_name.startswith('//gn'): |
| target.is_third_party_dep_ = True |
| return target |
| |
| proto_target_type, proto_desc = self.get_proto_target_type(target) |
| if proto_target_type: |
| assert proto_desc |
| self.proto_libs[target.name] = target |
| target.type = 'proto_library' |
| target.proto_plugin = proto_target_type |
| target.proto_paths.update(self.get_proto_paths(proto_desc)) |
| target.proto_exports.update(self.get_proto_exports(proto_desc)) |
| target.sources.update(proto_desc.get('sources', [])) |
| assert (all(x.endswith('.proto') for x in target.sources)) |
| elif target.type == 'source_set': |
| self.source_sets[gn_target_name] = target |
| target.sources.update(desc.get('sources', [])) |
| target.inputs.update(desc.get('inputs', [])) |
| elif target.type in LINKER_UNIT_TYPES: |
| self.linker_units[gn_target_name] = target |
| target.sources.update(desc.get('sources', [])) |
| elif target.type == 'action': |
| self.actions[gn_target_name] = target |
| target.data.update(desc.get('metadata', {}).get('perfetto_data', [])) |
| target.inputs.update(desc.get('inputs', [])) |
| target.sources.update(desc.get('sources', [])) |
| outs = [re.sub('^//out/.+?/gen/', '', x) for x in desc['outputs']] |
| target.outputs.update(outs) |
| target.script = desc['script'] |
| # Args are typically relative to the root build dir (../../xxx) |
| # because root build dir is typically out/xxx/). |
| target.args = [re.sub('^../../', '//', x) for x in desc['args']] |
| action_types = desc.get('metadata', |
| {}).get('perfetto_action_type_for_generator', []) |
| target.custom_action_type = action_types[0] if len( |
| action_types) > 0 else None |
| python_main = desc.get('metadata', {}).get('perfetto_python_main', []) |
| target.python_main = python_main[0] if python_main else None |
| |
| # Default for 'public' is //* - all headers in 'sources' are public. |
| # TODO(primiano): if a 'public' section is specified (even if empty), then |
| # the rest of 'sources' is considered inaccessible by gn. Consider |
| # emulating that, so that generated build files don't end up with overly |
| # accessible headers. |
| public_headers = [x for x in desc.get('public', []) if x != '*'] |
| target.public_headers.update(public_headers) |
| |
| target.cflags.update(desc.get('cflags', []) + desc.get('cflags_cc', [])) |
| target.libs.update(desc.get('libs', [])) |
| target.ldflags.update(desc.get('ldflags', [])) |
| target.defines.update(desc.get('defines', [])) |
| target.include_dirs.update(desc.get('include_dirs', [])) |
| |
| # Recurse in dependencies. |
| for dep_name in desc.get('deps', []): |
| dep = self.get_target(dep_name) |
| |
| # generated_file targets only exist for GN builds: we can safely ignore |
| # them. |
| if dep.type == 'generated_file': |
| continue |
| |
| # When a proto_library depends on an action, that is always the "_gen" |
| # rule of the action which is "private" to the proto_library rule. |
| # therefore, just ignore it for dep tracking purposes. |
| if dep.type == 'action' and proto_target_type is not None: |
| target_no_toolchain = label_without_toolchain(target.name) |
| dep_no_toolchain = label_without_toolchain(dep.name) |
| assert (dep_no_toolchain == f'{target_no_toolchain}_gen') |
| continue |
| |
| # Non-third party groups are only used for bubbling cflags etc so don't |
| # add a dep. |
| if dep.type == 'group' and not dep.is_third_party_dep_: |
| target.update(dep) # Bubble up groups's cflags/ldflags etc. |
| continue |
| |
| # Linker units act as a hard boundary making all their internal deps |
| # opaque to the outside world. For this reason, do not propogate deps |
| # transitively across them. |
| if dep.type in LINKER_UNIT_TYPES: |
| target.deps.add(dep) |
| continue |
| |
| if dep.type == 'source_set': |
| target.update(dep) # Bubble up source set's cflags/ldflags etc. |
| elif dep.type == 'proto_library': |
| target.proto_paths.update(dep.proto_paths) |
| |
| target.deps.add(dep) |
| target.transitive_deps.add(dep) |
| target.transitive_deps.update(dep.transitive_deps) |
| |
| return target |
| |
| def get_proto_exports(self, proto_desc): |
| # exports in metadata will be available for source_set targets. |
| metadata = proto_desc.get('metadata', {}) |
| return metadata.get('exports', []) |
| |
| def get_proto_paths(self, proto_desc): |
| # import_dirs in metadata will be available for source_set targets. |
| metadata = proto_desc.get('metadata', {}) |
| return metadata.get('import_dirs', []) |
| |
| def get_proto_target_type(self, target: Target |
| ) -> Tuple[Optional[str], Optional[Dict]]: |
| """ Checks if the target is a proto library and return the plugin. |
| |
| Returns: |
| (None, None): if the target is not a proto library. |
| (plugin, proto_desc) where |plugin| is 'proto' in the default (lite) |
| case or 'protozero' or 'ipc' or 'descriptor'; |proto_desc| is the GN |
| json desc of the target with the .proto sources (_gen target for |
| non-descriptor types or the target itself for descriptor type). |
| """ |
| parts = target.name.split('(', 1) |
| name = parts[0] |
| toolchain = '(' + parts[1] if len(parts) > 1 else '' |
| |
| # Descriptor targets don't have a _gen target; instead we look for the |
| # characteristic flag in the args of the target itself. |
| desc = self.gn_desc_.get(target.name) |
| if '--descriptor_set_out' in desc.get('args', []): |
| return 'descriptor', desc |
| |
| # Source set proto targets have a non-empty proto_library_sources in the |
| # metadata of the description. |
| metadata = desc.get('metadata', {}) |
| if 'proto_library_sources' in metadata: |
| return 'source_set', desc |
| |
| # In all other cases, we want to look at the _gen target as that has the |
| # important information. |
| gen_desc = self.gn_desc_.get('%s_gen%s' % (name, toolchain)) |
| if gen_desc is None or gen_desc['type'] != 'action': |
| return None, None |
| args = gen_desc.get('args', []) |
| if '/protoc' not in args[0]: |
| return None, None |
| plugin = 'proto' |
| for arg in (arg for arg in args if arg.startswith('--plugin=')): |
| # |arg| at this point looks like: |
| # --plugin=protoc-gen-plugin=gcc_like_host/protozero_plugin |
| # or |
| # --plugin=protoc-gen-plugin=protozero_plugin |
| plugin = arg.split('=')[-1].split('/')[-1].replace('_plugin', '') |
| return plugin, gen_desc |