blob: eb1a5c7c63001b9c86114acda1c5856e906ed541 [file] [log] [blame]
# Copyright (C) 2019 The Android Open Source Project
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
# A collection of utilities for extracting build rule information from GN
# projects.
from __future__ import print_function
import collections
from compat import iteritems
import errno
import filecmp
import json
import os
import re
import shutil
import subprocess
import sys
from typing import Dict
from typing import Optional
from typing import Set
from typing import Tuple
BUILDFLAGS_TARGET = '//gn:gen_buildflags'
GEN_VERSION_TARGET = '//src/base:version_gen_h'
TARGET_TOOLCHAIN = '//gn/standalone/toolchain:gcc_like_host'
HOST_TOOLCHAIN = '//gn/standalone/toolchain:gcc_like_host'
LINKER_UNIT_TYPES = ('executable', 'shared_library', 'static_library')
# TODO(primiano): investigate these, they require further componentization.
ODR_VIOLATION_IGNORE_TARGETS = {
'//test/cts:perfetto_cts_deps',
'//:perfetto_integrationtests',
}
def _check_command_output(cmd, cwd):
try:
output = subprocess.check_output(cmd, stderr=subprocess.STDOUT, cwd=cwd)
except subprocess.CalledProcessError as e:
print(
'Command "{}" failed in {}:'.format(' '.join(cmd), cwd),
file=sys.stderr)
print(e.output.decode(), file=sys.stderr)
sys.exit(1)
else:
return output.decode()
def repo_root():
"""Returns an absolute path to the repository root."""
return os.path.join(
os.path.realpath(os.path.dirname(__file__)), os.path.pardir)
def _tool_path(name, system_buildtools=False):
# Pass-through to use name if the caller requests to use the system
# toolchain.
if system_buildtools:
return [name]
wrapper = os.path.abspath(
os.path.join(repo_root(), 'tools', 'run_buildtools_binary.py'))
return ['python3', wrapper, name]
def prepare_out_directory(gn_args,
name,
root=repo_root(),
system_buildtools=False):
"""Creates the JSON build description by running GN.
Returns (path, desc) where |path| is the location of the output directory
and |desc| is the JSON build description.
"""
out = os.path.join(root, 'out', name)
try:
os.makedirs(out)
except OSError as e:
if e.errno != errno.EEXIST:
raise
_check_command_output(
_tool_path('gn', system_buildtools) +
['gen', out, '--args=%s' % gn_args],
cwd=repo_root())
return out
def load_build_description(out, system_buildtools=False):
"""Creates the JSON build description by running GN."""
desc = _check_command_output(
_tool_path('gn', system_buildtools) +
['desc', out, '--format=json', '--all-toolchains', '//*'],
cwd=repo_root())
return json.loads(desc)
def create_build_description(gn_args, root=repo_root()):
"""Prepares a GN out directory and loads the build description from it.
The temporary out directory is automatically deleted.
"""
out = prepare_out_directory(gn_args, 'tmp.gn_utils', root=root)
try:
return load_build_description(out)
finally:
shutil.rmtree(out)
def build_targets(out, targets, quiet=False, system_buildtools=False):
"""Runs ninja to build a list of GN targets in the given out directory.
Compiling these targets is required so that we can include any generated
source files in the amalgamated result.
"""
targets = [t.replace('//', '') for t in targets]
with open(os.devnull, 'w', newline='\n') as devnull:
stdout = devnull if quiet else None
cmd = _tool_path('ninja', system_buildtools) + targets
subprocess.check_call(cmd, cwd=os.path.abspath(out), stdout=stdout)
def compute_source_dependencies(out, system_buildtools=False):
"""For each source file, computes a set of headers it depends on."""
ninja_deps = _check_command_output(
_tool_path('ninja', system_buildtools) + ['-t', 'deps'], cwd=out)
deps = {}
current_source = None
for line in ninja_deps.split('\n'):
filename = os.path.relpath(os.path.join(out, line.strip()), repo_root())
# Sanitizer builds may have a dependency of ignorelist.txt. Just skip it.
if filename.endswith('gn/standalone/sanitizers/ignorelist.txt'):
continue
if not line or line[0] != ' ':
current_source = None
continue
elif not current_source:
# We're assuming the source file is always listed before the
# headers.
assert os.path.splitext(line)[1] in ['.c', '.cc', '.cpp', '.S']
current_source = filename
deps[current_source] = []
else:
assert current_source
deps[current_source].append(filename)
return deps
def label_to_path(label):
"""Turn a GN output label (e.g., //some_dir/file.cc) into a path."""
assert label.startswith('//')
return label[2:]
def label_without_toolchain(label):
"""Strips the toolchain from a GN label.
Return a GN label (e.g //buildtools:protobuf(//gn/standalone/toolchain:
gcc_like_host) without the parenthesised toolchain part.
"""
return label.split('(')[0]
def label_to_target_name_with_path(label):
"""
Turn a GN label into a target name involving the full path.
e.g., //src/perfetto:tests -> src_perfetto_tests
"""
name = re.sub(r'^//:?', '', label)
name = re.sub(r'[^a-zA-Z0-9_]', '_', name)
return name
def gen_buildflags(gn_args, target_file):
"""Generates the perfetto_build_flags.h for the given config.
target_file: the path, relative to the repo root, where the generated
buildflag header will be copied into.
"""
tmp_out = prepare_out_directory(gn_args, 'tmp.gen_buildflags')
build_targets(tmp_out, [BUILDFLAGS_TARGET], quiet=True)
src = os.path.join(tmp_out, 'gen', 'build_config', 'perfetto_build_flags.h')
shutil.copy(src, os.path.join(repo_root(), target_file))
shutil.rmtree(tmp_out)
def check_or_commit_generated_files(tmp_files, check):
"""Checks that gen files are unchanged or renames them to the final location
Takes in input a list of 'xxx.swp' files that have been written.
If check == False, it renames xxx.swp -> xxx.
If check == True, it just checks that the contents of 'xxx.swp' == 'xxx'.
Returns 0 if no diff was detected, 1 otherwise (to be used as exit code).
"""
res = 0
for tmp_file in tmp_files:
assert (tmp_file.endswith('.swp'))
target_file = os.path.relpath(tmp_file[:-4])
if check:
if not filecmp.cmp(tmp_file, target_file):
sys.stderr.write('%s needs to be regenerated\n' % target_file)
res = 1
os.unlink(tmp_file)
else:
os.replace(tmp_file, target_file)
return res
class ODRChecker(object):
"""Detects ODR violations in linker units
When we turn GN source sets into Soong & Bazel file groups, there is the risk
to create ODR violations by including the same file group into different
linker unit (this is because other build systems don't have a concept
equivalent to GN's source_set). This class navigates the transitive
dependencies (mostly static libraries) of a target and detects if multiple
paths end up including the same file group. This is to avoid situations like:
traced.exe -> base(file group)
traced.exe -> libperfetto(static lib) -> base(file group)
"""
def __init__(self, gn: 'GnParser', target_name: str):
self.gn = gn
self.root = gn.get_target(target_name)
self.source_sets: Dict[str, Set[str]] = collections.defaultdict(set)
self.deps_visited = set()
self.source_set_hdr_only = {}
self._visit(target_name)
num_violations = 0
if target_name in ODR_VIOLATION_IGNORE_TARGETS:
return
for sset, paths in self.source_sets.items():
if self.is_header_only(sset):
continue
if len(paths) != 1:
num_violations += 1
print(
'ODR violation in target %s, multiple paths include %s:\n %s' %
(target_name, sset, '\n '.join(paths)),
file=sys.stderr)
if num_violations > 0:
raise Exception('%d ODR violations detected. Build generation aborted' %
num_violations)
def _visit(self, target_name: str, parent_path=''):
target = self.gn.get_target(target_name)
path = ((parent_path + ' > ') if parent_path else '') + target_name
if not target:
raise Exception('Cannot find target %s' % target_name)
for ssdep in target.transitive_source_set_deps():
name_and_path = '%s (via %s)' % (target_name, path)
self.source_sets[ssdep.name].add(name_and_path)
deps = set(target.non_proto_or_source_set_deps()).union(
target.transitive_proto_deps()) - self.deps_visited
for dep in deps:
if dep.type == 'executable':
continue # Execs are strong boundaries and don't cause ODR violations.
# static_library dependencies should reset the path. It doesn't matter if
# we get to a source file via:
# source_set1 > static_lib > source.cc OR
# source_set1 > source_set2 > static_lib > source.cc
# This is NOT an ODR violation because source.cc is linked from the same
# static library
next_parent_path = path if dep.type != 'static_library' else ''
self.deps_visited.add(dep.name)
self._visit(dep.name, next_parent_path)
def is_header_only(self, source_set_name: str):
cached = self.source_set_hdr_only.get(source_set_name)
if cached is not None:
return cached
target = self.gn.get_target(source_set_name)
if target.type != 'source_set':
raise TypeError('%s is not a source_set' % source_set_name)
res = all(src.endswith('.h') for src in target.sources)
self.source_set_hdr_only[source_set_name] = res
return res
class GnParser(object):
"""A parser with some cleverness for GN json desc files
The main goals of this parser are:
1) Deal with the fact that other build systems don't have an equivalent
notion to GN's source_set. Conversely to Bazel's and Soong's filegroups,
GN source_sets expect that dependencies, cflags and other source_set
properties propagate up to the linker unit (static_library, executable or
shared_library). This parser simulates the same behavior: when a
source_set is encountered, some of its variables (cflags and such) are
copied up to the dependent targets. This is to allow gen_xxx to create
one filegroup for each source_set and then squash all the other flags
onto the linker unit.
2) Detect and special-case protobuf targets, figuring out the protoc-plugin
being used.
"""
class Target(object):
"""Reperesents A GN target.
Maked properties are propagated up the dependency chain when a
source_set dependency is encountered.
"""
def __init__(self, name, type):
self.name = name # e.g. //src/ipc:ipc
VALID_TYPES = ('static_library', 'shared_library', 'executable', 'group',
'action', 'source_set', 'proto_library', 'generated_file')
assert (type in VALID_TYPES)
self.type = type
self.testonly = False
self.toolchain = None
# These are valid only for type == proto_library.
# This is typically: 'proto', 'protozero', 'ipc'.
self.proto_plugin: Optional[str] = None
self.proto_paths = set()
self.proto_exports = set()
self.sources = set()
# TODO(primiano): consider whether the public section should be part of
# bubbled-up sources.
self.public_headers = set() # 'public'
# These are valid only for type == 'action'
self.data = set()
self.inputs = set()
self.outputs = set()
self.script = None
self.args = []
self.custom_action_type = None
self.python_main = None
# These variables are propagated up when encountering a dependency
# on a source_set target.
self.cflags = set()
self.defines = set()
self.deps: Set[GnParser.Target] = set()
self.transitive_deps: Set[GnParser.Target] = set()
self.libs = set()
self.include_dirs = set()
self.ldflags = set()
# Deps on //gn:xxx have this flag set to True. These dependencies
# are special because they pull third_party code from buildtools/.
# We don't want to keep recursing into //buildtools in generators,
# this flag is used to stop the recursion and create an empty
# placeholder target once we hit //gn:protoc or similar.
self.is_third_party_dep_ = False
def non_proto_or_source_set_deps(self):
return set(d for d in self.deps
if d.type != 'proto_library' and d.type != 'source_set')
def proto_deps(self):
return set(d for d in self.deps if d.type == 'proto_library')
def transitive_proto_deps(self):
return set(d for d in self.transitive_deps if d.type == 'proto_library')
def transitive_cpp_proto_deps(self):
return set(
d for d in self.transitive_deps if d.type == 'proto_library' and
d.proto_plugin != 'descriptor' and d.proto_plugin != 'source_set')
def transitive_source_set_deps(self):
return set(d for d in self.transitive_deps if d.type == 'source_set')
def __lt__(self, other):
if isinstance(other, self.__class__):
return self.name < other.name
raise TypeError(
'\'<\' not supported between instances of \'%s\' and \'%s\'' %
(type(self).__name__, type(other).__name__))
def __repr__(self):
return json.dumps(
{
k: (list(sorted(v)) if isinstance(v, set) else v)
for (k, v) in iteritems(self.__dict__)
},
indent=4,
sort_keys=True)
def update(self, other):
for key in ('cflags', 'data', 'defines', 'deps', 'include_dirs',
'ldflags', 'transitive_deps', 'libs', 'proto_paths'):
self.__dict__[key].update(other.__dict__.get(key, []))
def __init__(self, gn_desc):
self.gn_desc_ = gn_desc
self.all_targets = {}
self.linker_units = {} # Executables, shared or static libraries.
self.source_sets = {}
self.actions = {}
self.proto_libs = {}
def get_target(self, gn_target_name: str) -> Target:
"""Returns a Target object from the fully qualified GN target name.
It bubbles up variables from source_set dependencies as described in the
class-level comments.
"""
target = self.all_targets.get(gn_target_name)
if target is not None:
return target # Target already processed.
desc = self.gn_desc_.get(gn_target_name)
if not desc:
return None
target = GnParser.Target(gn_target_name, desc['type'])
target.testonly = desc.get('testonly', False)
target.toolchain = desc.get('toolchain', None)
self.all_targets[gn_target_name] = target
# We should never have GN targets directly depend on buidtools. They
# should hop via //gn:xxx, so we can give generators an opportunity to
# override them.
assert (not gn_target_name.startswith('//buildtools'))
# Don't descend further into third_party targets. Genrators are supposed
# to either ignore them or route to other externally-provided targets.
if gn_target_name.startswith('//gn'):
target.is_third_party_dep_ = True
return target
proto_target_type, proto_desc = self.get_proto_target_type(target)
if proto_target_type:
assert proto_desc
self.proto_libs[target.name] = target
target.type = 'proto_library'
target.proto_plugin = proto_target_type
target.proto_paths.update(self.get_proto_paths(proto_desc))
target.proto_exports.update(self.get_proto_exports(proto_desc))
target.sources.update(
self.get_proto_sources(proto_target_type, proto_desc))
assert (all(x.endswith('.proto') for x in target.sources))
elif target.type == 'source_set':
self.source_sets[gn_target_name] = target
target.sources.update(desc.get('sources', []))
target.inputs.update(desc.get('inputs', []))
elif target.type in LINKER_UNIT_TYPES:
self.linker_units[gn_target_name] = target
target.sources.update(desc.get('sources', []))
elif target.type == 'action':
self.actions[gn_target_name] = target
target.data.update(desc.get('metadata', {}).get('perfetto_data', []))
target.inputs.update(desc.get('inputs', []))
target.sources.update(desc.get('sources', []))
outs = [re.sub('^//out/.+?/gen/', '', x) for x in desc['outputs']]
target.outputs.update(outs)
target.script = desc['script']
# Args are typically relative to the root build dir (../../xxx)
# because root build dir is typically out/xxx/).
target.args = [re.sub('^../../', '//', x) for x in desc['args']]
action_types = desc.get('metadata',
{}).get('perfetto_action_type_for_generator', [])
target.custom_action_type = action_types[0] if len(
action_types) > 0 else None
python_main = desc.get('metadata', {}).get('perfetto_python_main', [])
target.python_main = python_main[0] if python_main else None
# Default for 'public' is //* - all headers in 'sources' are public.
# TODO(primiano): if a 'public' section is specified (even if empty), then
# the rest of 'sources' is considered inaccessible by gn. Consider
# emulating that, so that generated build files don't end up with overly
# accessible headers.
public_headers = [x for x in desc.get('public', []) if x != '*']
target.public_headers.update(public_headers)
target.cflags.update(desc.get('cflags', []) + desc.get('cflags_cc', []))
target.libs.update(desc.get('libs', []))
target.ldflags.update(desc.get('ldflags', []))
target.defines.update(desc.get('defines', []))
target.include_dirs.update(desc.get('include_dirs', []))
# Recurse in dependencies.
for dep_name in desc.get('deps', []):
dep = self.get_target(dep_name)
# generated_file targets only exist for GN builds: we can safely ignore
# them.
if dep.type == 'generated_file':
continue
# When a proto_library depends on an action, that is always the "_gen"
# rule of the action which is "private" to the proto_library rule.
# therefore, just ignore it for dep tracking purposes.
if dep.type == 'action' and proto_target_type is not None:
target_no_toolchain = label_without_toolchain(target.name)
dep_no_toolchain = label_without_toolchain(dep.name)
assert (dep_no_toolchain == f'{target_no_toolchain}_gen')
continue
# Non-third party groups are only used for bubbling cflags etc so don't
# add a dep.
if dep.type == 'group' and not dep.is_third_party_dep_:
target.update(dep) # Bubble up groups's cflags/ldflags etc.
continue
# Linker units act as a hard boundary making all their internal deps
# opaque to the outside world. For this reason, do not propogate deps
# transitively across them.
if dep.type in LINKER_UNIT_TYPES:
target.deps.add(dep)
continue
if dep.type == 'source_set':
target.update(dep) # Bubble up source set's cflags/ldflags etc.
elif dep.type == 'proto_library':
target.proto_paths.update(dep.proto_paths)
target.deps.add(dep)
target.transitive_deps.add(dep)
target.transitive_deps.update(dep.transitive_deps)
return target
def get_proto_exports(self, proto_desc):
# exports in metadata will be available for source_set targets.
metadata = proto_desc.get('metadata', {})
return metadata.get('exports', [])
def get_proto_paths(self, proto_desc):
metadata = proto_desc.get('metadata', {})
return metadata.get('proto_import_dirs', [])
def get_proto_sources(self, proto_target_type, proto_desc):
if proto_target_type == 'source_set':
metadata = proto_desc.get('metadata', {})
return metadata.get('proto_library_sources', [])
return proto_desc.get('sources', [])
def get_proto_target_type(
self, target: Target) -> Tuple[Optional[str], Optional[Dict]]:
""" Checks if the target is a proto library and return the plugin.
Returns:
(None, None): if the target is not a proto library.
(plugin, proto_desc) where |plugin| is 'proto' in the default (lite)
case or 'protozero' or 'ipc' or 'descriptor'; |proto_desc| is the GN
json desc of the target with the .proto sources (_gen target for
non-descriptor types or the target itself for descriptor type).
"""
parts = target.name.split('(', 1)
name = parts[0]
toolchain = '(' + parts[1] if len(parts) > 1 else ''
# Descriptor targets don't have a _gen target; instead we look for the
# characteristic flag in the args of the target itself.
desc = self.gn_desc_.get(target.name)
if '--descriptor_set_out' in desc.get('args', []):
return 'descriptor', desc
# Source set proto targets have a non-empty proto_library_sources in the
# metadata of the description.
metadata = desc.get('metadata', {})
if 'proto_library_sources' in metadata:
return 'source_set', desc
# In all other cases, we want to look at the _gen target as that has the
# important information.
gen_desc = self.gn_desc_.get('%s_gen%s' % (name, toolchain))
if gen_desc is None or gen_desc['type'] != 'action':
return None, None
args = gen_desc.get('args', [])
if '/protoc' not in args[0]:
return None, None
plugin = 'proto'
for arg in (arg for arg in args if arg.startswith('--plugin=')):
# |arg| at this point looks like:
# --plugin=protoc-gen-plugin=gcc_like_host/protozero_plugin
# or
# --plugin=protoc-gen-plugin=protozero_plugin
plugin = arg.split('=')[-1].split('/')[-1].replace('_plugin', '')
return plugin, gen_desc