| // Protocol Buffers - Google's data interchange format |
| // Copyright 2008 Google Inc. All rights reserved. |
| // |
| // Use of this source code is governed by a BSD-style |
| // license that can be found in the LICENSE file or at |
| // https://developers.google.com/open-source/licenses/bsd |
| |
| // Author: kenton@google.com (Kenton Varda) |
| // Based on original Protocol Buffers design by |
| // Sanjay Ghemawat, Jeff Dean, and others. |
| |
| #include "google/protobuf/compiler/cpp/message.h" |
| |
| #include <algorithm> |
| #include <cmath> |
| #include <cstddef> |
| #include <cstdint> |
| #include <functional> |
| #include <iterator> |
| #include <limits> |
| #include <memory> |
| #include <string> |
| #include <utility> |
| #include <vector> |
| |
| #include "absl/container/flat_hash_map.h" |
| #include "absl/container/flat_hash_set.h" |
| #include "absl/log/absl_check.h" |
| #include "absl/log/absl_log.h" |
| #include "absl/strings/ascii.h" |
| #include "absl/strings/escaping.h" |
| #include "absl/strings/str_cat.h" |
| #include "absl/strings/str_format.h" |
| #include "absl/strings/str_join.h" |
| #include "absl/strings/string_view.h" |
| #include "google/protobuf/compiler/cpp/enum.h" |
| #include "google/protobuf/compiler/cpp/extension.h" |
| #include "google/protobuf/compiler/cpp/field.h" |
| #include "google/protobuf/compiler/cpp/helpers.h" |
| #include "google/protobuf/compiler/cpp/names.h" |
| #include "google/protobuf/compiler/cpp/options.h" |
| #include "google/protobuf/compiler/cpp/padding_optimizer.h" |
| #include "google/protobuf/compiler/cpp/parse_function_generator.h" |
| #include "google/protobuf/compiler/cpp/tracker.h" |
| #include "google/protobuf/descriptor.h" |
| #include "google/protobuf/descriptor.pb.h" |
| #include "google/protobuf/io/printer.h" |
| #include "google/protobuf/wire_format.h" |
| #include "google/protobuf/wire_format_lite.h" |
| |
| |
| // Must be included last. |
| #include "google/protobuf/port_def.inc" |
| |
| namespace google { |
| namespace protobuf { |
| namespace compiler { |
| namespace cpp { |
| namespace { |
| using ::google::protobuf::internal::WireFormat; |
| using ::google::protobuf::internal::WireFormatLite; |
| using ::google::protobuf::internal::cpp::HasHasbit; |
| using ::google::protobuf::internal::cpp::Utf8CheckMode; |
| using Semantic = ::google::protobuf::io::AnnotationCollector::Semantic; |
| using Sub = ::google::protobuf::io::Printer::Sub; |
| |
| static constexpr int kNoHasbit = -1; |
| |
| // Create an expression that evaluates to |
| // "for all i, (_has_bits_[i] & masks[i]) == masks[i]" |
| // masks is allowed to be shorter than _has_bits_, but at least one element of |
| // masks must be non-zero. |
| std::string ConditionalToCheckBitmasks( |
| const std::vector<uint32_t>& masks, bool return_success = true, |
| absl::string_view has_bits_var = "_impl_._has_bits_") { |
| std::vector<std::string> parts; |
| for (int i = 0; i < masks.size(); i++) { |
| if (masks[i] == 0) continue; |
| std::string m = absl::StrCat("0x", absl::Hex(masks[i], absl::kZeroPad8)); |
| // Each xor evaluates to 0 if the expected bits are present. |
| parts.push_back( |
| absl::StrCat("((", has_bits_var, "[", i, "] & ", m, ") ^ ", m, ")")); |
| } |
| ABSL_CHECK(!parts.empty()); |
| // If we have multiple parts, each expected to be 0, then bitwise-or them. |
| std::string result = |
| parts.size() == 1 |
| ? parts[0] |
| : absl::StrCat("(", absl::StrJoin(parts, "\n | "), ")"); |
| return result + (return_success ? " == 0" : " != 0"); |
| } |
| |
| void PrintPresenceCheck(const FieldDescriptor* field, |
| const std::vector<int>& has_bit_indices, io::Printer* p, |
| int* cached_has_word_index) { |
| if (!field->options().weak()) { |
| int has_bit_index = has_bit_indices[field->index()]; |
| if (*cached_has_word_index != (has_bit_index / 32)) { |
| *cached_has_word_index = (has_bit_index / 32); |
| p->Emit({{"index", *cached_has_word_index}}, |
| R"cc( |
| cached_has_bits = $has_bits$[$index$]; |
| )cc"); |
| } |
| p->Emit({{"mask", absl::StrFormat("0x%08xu", 1u << (has_bit_index % 32))}}, |
| R"cc( |
| if (cached_has_bits & $mask$) { |
| )cc"); |
| } else { |
| p->Emit(R"cc( |
| if (has_$name$()) { |
| )cc"); |
| } |
| } |
| |
| struct FieldOrderingByNumber { |
| inline bool operator()(const FieldDescriptor* a, |
| const FieldDescriptor* b) const { |
| return a->number() < b->number(); |
| } |
| }; |
| |
| // Sort the fields of the given Descriptor by number into a new[]'d array |
| // and return it. |
| std::vector<const FieldDescriptor*> SortFieldsByNumber( |
| const Descriptor* descriptor) { |
| std::vector<const FieldDescriptor*> fields(descriptor->field_count()); |
| for (int i = 0; i < descriptor->field_count(); i++) { |
| fields[i] = descriptor->field(i); |
| } |
| std::sort(fields.begin(), fields.end(), FieldOrderingByNumber()); |
| return fields; |
| } |
| |
| // Functor for sorting extension ranges by their "start" field number. |
| struct ExtensionRangeSorter { |
| bool operator()(const Descriptor::ExtensionRange* left, |
| const Descriptor::ExtensionRange* right) const { |
| return left->start_number() < right->start_number(); |
| } |
| }; |
| |
| bool IsPOD(const FieldDescriptor* field) { |
| if (field->is_repeated() || field->is_extension()) return false; |
| switch (field->cpp_type()) { |
| case FieldDescriptor::CPPTYPE_ENUM: |
| case FieldDescriptor::CPPTYPE_INT32: |
| case FieldDescriptor::CPPTYPE_INT64: |
| case FieldDescriptor::CPPTYPE_UINT32: |
| case FieldDescriptor::CPPTYPE_UINT64: |
| case FieldDescriptor::CPPTYPE_FLOAT: |
| case FieldDescriptor::CPPTYPE_DOUBLE: |
| case FieldDescriptor::CPPTYPE_BOOL: |
| return true; |
| case FieldDescriptor::CPPTYPE_STRING: |
| return false; |
| default: |
| return false; |
| } |
| } |
| |
| // Finds runs of fields for which `predicate` is true. |
| // RunMap maps from fields that start each run to the number of fields in that |
| // run. This is optimized for the common case that there are very few runs in |
| // a message and that most of the eligible fields appear together. |
| using RunMap = absl::flat_hash_map<const FieldDescriptor*, size_t>; |
| RunMap FindRuns(const std::vector<const FieldDescriptor*>& fields, |
| const std::function<bool(const FieldDescriptor*)>& predicate) { |
| RunMap runs; |
| const FieldDescriptor* last_start = nullptr; |
| |
| for (auto field : fields) { |
| if (predicate(field)) { |
| if (last_start == nullptr) { |
| last_start = field; |
| } |
| |
| runs[last_start]++; |
| } else { |
| last_start = nullptr; |
| } |
| } |
| return runs; |
| } |
| |
| // Emits an if-statement with a condition that evaluates to true if |field| is |
| // considered non-default (will be sent over the wire), for message types |
| // without true field presence. Should only be called if |
| // !HasHasbit(field). |
| bool EmitFieldNonDefaultCondition(io::Printer* p, const std::string& prefix, |
| const FieldDescriptor* field) { |
| ABSL_CHECK(!HasHasbit(field)); |
| auto v = p->WithVars({{ |
| {"prefix", prefix}, |
| {"name", FieldName(field)}, |
| }}); |
| // Merge and serialize semantics: primitive fields are merged/serialized only |
| // if non-zero (numeric) or non-empty (string). |
| if (!field->is_repeated() && !field->containing_oneof()) { |
| if (field->cpp_type() == FieldDescriptor::CPPTYPE_STRING) { |
| p->Emit(R"cc( |
| if (!$prefix$_internal_$name$().empty()) { |
| )cc"); |
| } else if (field->cpp_type() == FieldDescriptor::CPPTYPE_MESSAGE) { |
| // Message fields still have has_$name$() methods. |
| p->Emit(R"cc( |
| if ($prefix$_internal_has_$name$()) { |
| )cc"); |
| } else if (field->cpp_type() == FieldDescriptor::CPPTYPE_FLOAT) { |
| p->Emit(R"cc( |
| static_assert(sizeof(::uint32_t) == sizeof(float), |
| "Code assumes ::uint32_t and float are the same size."); |
| float tmp_$name$ = $prefix$_internal_$name$(); |
| ::uint32_t raw_$name$; |
| memcpy(&raw_$name$, &tmp_$name$, sizeof(tmp_$name$)); |
| if (raw_$name$ != 0) { |
| )cc"); |
| } else if (field->cpp_type() == FieldDescriptor::CPPTYPE_DOUBLE) { |
| p->Emit(R"cc( |
| static_assert(sizeof(::uint64_t) == sizeof(double), |
| "Code assumes ::uint64_t and double are the same size."); |
| double tmp_$name$ = $prefix$_internal_$name$(); |
| ::uint64_t raw_$name$; |
| memcpy(&raw_$name$, &tmp_$name$, sizeof(tmp_$name$)); |
| if (raw_$name$ != 0) { |
| )cc"); |
| } else { |
| p->Emit(R"cc( |
| if ($prefix$_internal_$name$() != 0) { |
| )cc"); |
| } |
| return true; |
| } else if (field->real_containing_oneof()) { |
| p->Emit(R"cc( |
| if ($has_field$) { |
| )cc"); |
| return true; |
| } |
| return false; |
| } |
| |
| bool HasInternalHasMethod(const FieldDescriptor* field) { |
| return !HasHasbit(field) && |
| field->cpp_type() == FieldDescriptor::CPPTYPE_MESSAGE; |
| } |
| |
| // Collects map entry message type information. |
| void CollectMapInfo( |
| const Options& options, const Descriptor* descriptor, |
| absl::flat_hash_map<absl::string_view, std::string>* variables) { |
| ABSL_CHECK(IsMapEntryMessage(descriptor)); |
| absl::flat_hash_map<absl::string_view, std::string>& vars = *variables; |
| const FieldDescriptor* key = descriptor->map_key(); |
| const FieldDescriptor* val = descriptor->map_value(); |
| vars["key_cpp"] = PrimitiveTypeName(options, key->cpp_type()); |
| switch (val->cpp_type()) { |
| case FieldDescriptor::CPPTYPE_MESSAGE: |
| vars["val_cpp"] = FieldMessageTypeName(val, options); |
| break; |
| case FieldDescriptor::CPPTYPE_ENUM: |
| vars["val_cpp"] = ClassName(val->enum_type(), true); |
| break; |
| default: |
| vars["val_cpp"] = PrimitiveTypeName(options, val->cpp_type()); |
| } |
| vars["key_wire_type"] = absl::StrCat( |
| "TYPE_", absl::AsciiStrToUpper(DeclaredTypeMethodName(key->type()))); |
| vars["val_wire_type"] = absl::StrCat( |
| "TYPE_", absl::AsciiStrToUpper(DeclaredTypeMethodName(val->type()))); |
| } |
| |
| |
| // Returns true to make the message serialize in order, decided by the following |
| // factors in the order of precedence. |
| // --options().message_set_wire_format() == true |
| // --the message is in the allowlist (true) |
| // --GOOGLE_PROTOBUF_SHUFFLE_SERIALIZE is defined (false) |
| // --a ranage of message names that are allowed to stay in order (true) |
| bool ShouldSerializeInOrder(const Descriptor* descriptor, |
| const Options& options) { |
| return true; |
| } |
| |
| bool IsCrossFileMapField(const FieldDescriptor* field) { |
| if (!field->is_map()) { |
| return false; |
| } |
| |
| const Descriptor* d = field->message_type(); |
| const FieldDescriptor* value = d->FindFieldByNumber(2); |
| |
| return IsCrossFileMessage(value); |
| } |
| |
| bool IsCrossFileMaybeMap(const FieldDescriptor* field) { |
| if (IsCrossFileMapField(field)) { |
| return true; |
| } |
| |
| return IsCrossFileMessage(field); |
| } |
| |
| bool HasNonSplitOptionalString(const Descriptor* desc, const Options& options) { |
| for (const auto* field : FieldRange(desc)) { |
| if (IsString(field) && !field->is_repeated() && |
| !field->real_containing_oneof() && !ShouldSplit(field, options)) { |
| return true; |
| } |
| } |
| return false; |
| } |
| |
| struct FieldChunk { |
| FieldChunk(bool has_hasbit, bool is_rarely_present, bool should_split) |
| : has_hasbit(has_hasbit), |
| is_rarely_present(is_rarely_present), |
| should_split(should_split) {} |
| |
| bool has_hasbit; |
| bool is_rarely_present; |
| bool should_split; |
| |
| std::vector<const FieldDescriptor*> fields; |
| }; |
| |
| using ChunkIterator = std::vector<FieldChunk>::const_iterator; |
| |
| // Breaks down a single chunk of fields into a few chunks that share attributes |
| // controlled by "equivalent" predicate. Returns an array of chunks. |
| template <typename Predicate> |
| std::vector<FieldChunk> CollectFields( |
| const std::vector<const FieldDescriptor*>& fields, const Options& options, |
| const Predicate& equivalent) { |
| std::vector<FieldChunk> chunks; |
| for (auto field : fields) { |
| if (chunks.empty() || !equivalent(chunks.back().fields.back(), field)) { |
| chunks.emplace_back(HasHasbit(field), IsRarelyPresent(field, options), |
| ShouldSplit(field, options)); |
| } |
| chunks.back().fields.push_back(field); |
| } |
| return chunks; |
| } |
| |
| template <typename Predicate> |
| ChunkIterator FindNextUnequalChunk(ChunkIterator start, ChunkIterator end, |
| const Predicate& equal) { |
| auto it = start; |
| while (++it != end) { |
| if (!equal(*start, *it)) { |
| return it; |
| } |
| } |
| return end; |
| } |
| |
| // Returns true if two chunks may be grouped for hasword check to skip multiple |
| // cold fields at once. They have to share the following traits: |
| // - whether they have hasbits |
| // - whether they are rarely present |
| // - whether they are split |
| bool MayGroupChunksForHaswordsCheck(const FieldChunk& a, const FieldChunk& b) { |
| return a.has_hasbit == b.has_hasbit && |
| a.is_rarely_present == b.is_rarely_present && |
| a.should_split == b.should_split; |
| } |
| |
| // Returns a bit mask based on has_bit index of "fields" that are typically on |
| // the same chunk. It is used in a group presence check where _has_bits_ is |
| // masked to tell if any thing in "fields" is present. |
| uint32_t GenChunkMask(const std::vector<const FieldDescriptor*>& fields, |
| const std::vector<int>& has_bit_indices) { |
| ABSL_CHECK(!fields.empty()); |
| int first_index_offset = has_bit_indices[fields.front()->index()] / 32; |
| uint32_t chunk_mask = 0; |
| for (auto field : fields) { |
| // "index" defines where in the _has_bits_ the field appears. |
| int index = has_bit_indices[field->index()]; |
| ABSL_CHECK_EQ(first_index_offset, index / 32); |
| chunk_mask |= static_cast<uint32_t>(1) << (index % 32); |
| } |
| ABSL_CHECK_NE(0, chunk_mask); |
| return chunk_mask; |
| } |
| |
| // Returns a bit mask based on has_bit index of "fields" in chunks in [it, end). |
| // Assumes that all chunks share the same hasbit word. |
| uint32_t GenChunkMask(ChunkIterator it, ChunkIterator end, |
| const std::vector<int>& has_bit_indices) { |
| ABSL_CHECK(it != end); |
| |
| int first_index_offset = has_bit_indices[it->fields.front()->index()] / 32; |
| uint32_t chunk_mask = 0; |
| do { |
| ABSL_CHECK_EQ(first_index_offset, |
| has_bit_indices[it->fields.front()->index()] / 32); |
| chunk_mask |= GenChunkMask(it->fields, has_bit_indices); |
| } while (++it != end); |
| return chunk_mask; |
| } |
| |
| // Return the number of bits set in n, a non-negative integer. |
| static int popcnt(uint32_t n) { |
| int result = 0; |
| while (n != 0) { |
| result += (n & 1); |
| n = n / 2; |
| } |
| return result; |
| } |
| |
| // Returns true if it emits conditional check against hasbit words. This is |
| // useful to skip multiple fields that are unlikely present based on profile |
| // (go/pdproto). |
| bool MaybeEmitHaswordsCheck(ChunkIterator it, ChunkIterator end, |
| const Options& options, |
| const std::vector<int>& has_bit_indices, |
| int cached_has_word_index, const std::string& from, |
| io::Printer* p) { |
| if (!it->has_hasbit || !IsProfileDriven(options) || |
| std::distance(it, end) < 2 || !it->is_rarely_present) { |
| return false; |
| } |
| |
| auto hasbit_word = [&has_bit_indices](const FieldDescriptor* field) { |
| return has_bit_indices[field->index()] / 32; |
| }; |
| auto is_same_hasword = [&](const FieldChunk& a, const FieldChunk& b) { |
| return hasbit_word(a.fields.front()) == hasbit_word(b.fields.front()); |
| }; |
| |
| struct HasWordMask { |
| int word; |
| uint32_t mask; |
| }; |
| |
| std::vector<HasWordMask> hasword_masks; |
| while (it != end) { |
| auto next = FindNextUnequalChunk(it, end, is_same_hasword); |
| hasword_masks.push_back({hasbit_word(it->fields.front()), |
| GenChunkMask(it, next, has_bit_indices)}); |
| it = next; |
| } |
| |
| // Emit has_bit check for each has_bit_dword index. |
| p->Emit( |
| {{"cond", |
| [&] { |
| int first_word = hasword_masks.front().word; |
| for (const auto& m : hasword_masks) { |
| uint32_t mask = m.mask; |
| int this_word = m.word; |
| if (this_word != first_word) { |
| p->Emit(R"cc( |
| || |
| )cc"); |
| } |
| auto v = p->WithVars({{"mask", absl::StrFormat("0x%08xu", mask)}}); |
| if (this_word == cached_has_word_index) { |
| p->Emit("(cached_has_bits & $mask$) != 0"); |
| } else { |
| p->Emit({{"from", from}, {"word", this_word}}, |
| "($from$_impl_._has_bits_[$word$] & $mask$) != 0"); |
| } |
| } |
| }}}, |
| R"cc( |
| if (PROTOBUF_PREDICT_FALSE($cond$)) { |
| )cc"); |
| p->Indent(); |
| return true; |
| } |
| |
| absl::flat_hash_map<absl::string_view, std::string> ClassVars( |
| const Descriptor* desc, Options opts) { |
| absl::flat_hash_map<absl::string_view, std::string> vars = MessageVars(desc); |
| |
| vars.emplace("pkg", Namespace(desc, opts)); |
| vars.emplace("Msg", ClassName(desc, false)); |
| vars.emplace("pkg::Msg", QualifiedClassName(desc, opts)); |
| vars.emplace("pkg.Msg", desc->full_name()); |
| |
| // Old-style names, to be removed once all usages are gone in this and other |
| // files. |
| vars.emplace("classname", ClassName(desc, false)); |
| vars.emplace("classtype", QualifiedClassName(desc, opts)); |
| vars.emplace("full_name", desc->full_name()); |
| vars.emplace("superclass", SuperClassName(desc, opts)); |
| |
| for (auto& pair : UnknownFieldsVars(desc, opts)) { |
| vars.emplace(pair); |
| } |
| |
| return vars; |
| } |
| |
| } // anonymous namespace |
| |
| // =================================================================== |
| |
| MessageGenerator::MessageGenerator( |
| const Descriptor* descriptor, |
| const absl::flat_hash_map<absl::string_view, std::string>&, |
| int index_in_file_messages, const Options& options, |
| MessageSCCAnalyzer* scc_analyzer) |
| : descriptor_(descriptor), |
| index_in_file_messages_(index_in_file_messages), |
| options_(options), |
| field_generators_(descriptor), |
| scc_analyzer_(scc_analyzer) { |
| |
| if (!message_layout_helper_) { |
| message_layout_helper_ = std::make_unique<PaddingOptimizer>(); |
| } |
| |
| // Compute optimized field order to be used for layout and initialization |
| // purposes. |
| for (auto field : FieldRange(descriptor_)) { |
| if (IsWeak(field, options_)) { |
| ++num_weak_fields_; |
| continue; |
| } |
| |
| if (!field->real_containing_oneof()) { |
| optimized_order_.push_back(field); |
| } |
| } |
| |
| const size_t initial_size = optimized_order_.size(); |
| message_layout_helper_->OptimizeLayout(&optimized_order_, options_, |
| scc_analyzer_); |
| ABSL_CHECK_EQ(initial_size, optimized_order_.size()); |
| |
| // This message has hasbits iff one or more fields need one. |
| for (auto field : optimized_order_) { |
| if (HasHasbit(field)) { |
| if (has_bit_indices_.empty()) { |
| has_bit_indices_.resize(descriptor_->field_count(), kNoHasbit); |
| } |
| has_bit_indices_[field->index()] = max_has_bit_index_++; |
| } |
| if (IsStringInlined(field, options_)) { |
| if (inlined_string_indices_.empty()) { |
| inlined_string_indices_.resize(descriptor_->field_count(), kNoHasbit); |
| // The bitset[0] is for arena dtor tracking. Donating states start from |
| // bitset[1]; |
| ++max_inlined_string_index_; |
| } |
| |
| inlined_string_indices_[field->index()] = max_inlined_string_index_++; |
| } |
| } |
| field_generators_.Build(options_, scc_analyzer_, has_bit_indices_, |
| inlined_string_indices_); |
| |
| for (int i = 0; i < descriptor->field_count(); i++) { |
| if (descriptor->field(i)->is_required()) { |
| ++num_required_fields_; |
| } |
| } |
| |
| parse_function_generator_ = std::make_unique<ParseFunctionGenerator>( |
| descriptor_, max_has_bit_index_, has_bit_indices_, |
| inlined_string_indices_, options_, scc_analyzer_, variables_); |
| } |
| |
| size_t MessageGenerator::HasBitsSize() const { |
| return (max_has_bit_index_ + 31) / 32; |
| } |
| |
| size_t MessageGenerator::InlinedStringDonatedSize() const { |
| return (max_inlined_string_index_ + 31) / 32; |
| } |
| |
| absl::flat_hash_map<absl::string_view, std::string> |
| MessageGenerator::HasBitVars(const FieldDescriptor* field) const { |
| int has_bit_index = HasBitIndex(field); |
| ABSL_CHECK_NE(has_bit_index, kNoHasbit); |
| return { |
| {"has_array_index", absl::StrCat(has_bit_index / 32)}, |
| {"has_mask", absl::StrFormat("0x%08xu", 1u << (has_bit_index % 32))}, |
| }; |
| } |
| |
| int MessageGenerator::HasBitIndex(const FieldDescriptor* field) const { |
| return has_bit_indices_.empty() ? kNoHasbit |
| : has_bit_indices_[field->index()]; |
| } |
| |
| int MessageGenerator::HasByteIndex(const FieldDescriptor* field) const { |
| int hasbit = HasBitIndex(field); |
| return hasbit == kNoHasbit ? kNoHasbit : hasbit / 8; |
| } |
| |
| int MessageGenerator::HasWordIndex(const FieldDescriptor* field) const { |
| int hasbit = HasBitIndex(field); |
| return hasbit == kNoHasbit ? kNoHasbit : hasbit / 32; |
| } |
| |
| void MessageGenerator::AddGenerators( |
| std::vector<std::unique_ptr<EnumGenerator>>* enum_generators, |
| std::vector<std::unique_ptr<ExtensionGenerator>>* extension_generators) { |
| for (int i = 0; i < descriptor_->enum_type_count(); i++) { |
| enum_generators->emplace_back( |
| std::make_unique<EnumGenerator>(descriptor_->enum_type(i), options_)); |
| enum_generators_.push_back(enum_generators->back().get()); |
| } |
| for (int i = 0; i < descriptor_->extension_count(); i++) { |
| extension_generators->emplace_back(std::make_unique<ExtensionGenerator>( |
| descriptor_->extension(i), options_, scc_analyzer_)); |
| extension_generators_.push_back(extension_generators->back().get()); |
| } |
| } |
| |
| void MessageGenerator::GenerateFieldAccessorDeclarations(io::Printer* p) { |
| auto v = p->WithVars(MessageVars(descriptor_)); |
| |
| // optimized_fields_ does not contain fields where |
| // field->real_containing_oneof() |
| // so we need to iterate over those as well. |
| // |
| // We place the non-oneof fields in optimized_order_, as that controls the |
| // order of the _has_bits_ entries and we want GDB's pretty ps to be |
| // able to infer these indices from the k[FIELDNAME]FieldNumber order. |
| std::vector<const FieldDescriptor*> ordered_fields; |
| ordered_fields.reserve(descriptor_->field_count()); |
| ordered_fields.insert(ordered_fields.begin(), optimized_order_.begin(), |
| optimized_order_.end()); |
| |
| for (auto field : FieldRange(descriptor_)) { |
| if (!field->real_containing_oneof() && !field->options().weak()) { |
| continue; |
| } |
| ordered_fields.push_back(field); |
| } |
| |
| if (!ordered_fields.empty()) { |
| p->Emit({{ |
| "kFields", |
| [&] { |
| for (auto field : ordered_fields) { |
| auto v = p->WithVars(FieldVars(field, options_)); |
| p->Emit({Sub("kField", FieldConstantName(field)) |
| .AnnotatedAs(field)}, |
| R"cc( |
| $kField$ = $number$, |
| )cc"); |
| } |
| }, |
| }}, |
| R"cc( |
| enum : int { |
| $kFields$, |
| }; |
| )cc"); |
| } |
| for (auto field : ordered_fields) { |
| auto name = FieldName(field); |
| |
| auto v = p->WithVars(FieldVars(field, options_)); |
| auto t = p->WithVars(MakeTrackerCalls(field, options_)); |
| p->Emit({{"field_comment", FieldComment(field, options_)}, |
| Sub("const_impl", "const;").WithSuffix(";"), |
| Sub("impl", ";").WithSuffix(";"), |
| {"sizer", |
| [&] { |
| if (!field->is_repeated()) return; |
| p->Emit({Sub("name_size", absl::StrCat(name, "_size")) |
| .AnnotatedAs(field)}, |
| R"cc( |
| $deprecated_attr $int $name_size$() $const_impl$; |
| )cc"); |
| |
| p->Emit({Sub("_internal_name_size", |
| absl::StrCat("_internal_", name, "_size")) |
| .AnnotatedAs(field)}, |
| R"cc( |
| private: |
| int $_internal_name_size$() const; |
| |
| public: |
| )cc"); |
| }}, |
| {"hazzer", |
| [&] { |
| if (!field->has_presence()) return; |
| p->Emit({Sub("has_name", absl::StrCat("has_", name)) |
| .AnnotatedAs(field)}, |
| R"cc( |
| $deprecated_attr $bool $has_name$() $const_impl$; |
| )cc"); |
| }}, |
| {"internal_hazzer", |
| [&] { |
| if (field->is_repeated() || !HasInternalHasMethod(field)) { |
| return; |
| } |
| p->Emit({Sub("_internal_has_name", |
| absl::StrCat("_internal_has_", name)) |
| .AnnotatedAs(field)}, |
| R"cc( |
| private: |
| bool $_internal_has_name$() const; |
| |
| public: |
| )cc"); |
| }}, |
| {"clearer", |
| [&] { |
| p->Emit({Sub("clear_name", absl::StrCat("clear_", name)) |
| .AnnotatedAs({ |
| field, |
| Semantic::kSet, |
| })}, |
| R"cc( |
| $deprecated_attr $void $clear_name$() $impl$; |
| )cc"); |
| }}, |
| {"accessors", |
| [&] { |
| field_generators_.get(field).GenerateAccessorDeclarations(p); |
| }}}, |
| R"cc( |
| // $field_comment$ |
| $sizer$; |
| $hazzer$; |
| $internal_hazzer$; |
| $clearer$; |
| $accessors$; |
| )cc"); |
| } |
| |
| if (descriptor_->extension_range_count() > 0) { |
| // Generate accessors for extensions. |
| // We use "_proto_TypeTraits" as a type name below because "TypeTraits" |
| // causes problems if the class has a nested message or enum type with that |
| // name and "_TypeTraits" is technically reserved for the C++ library since |
| // it starts with an underscore followed by a capital letter. |
| // |
| // For similar reason, we use "_field_type" and "_is_packed" as parameter |
| // names below, so that "field_type" and "is_packed" can be used as field |
| // names. |
| p->Emit(R"cc( |
| template <typename _proto_TypeTraits, $pbi$::FieldType _field_type, |
| bool _is_packed> |
| inline bool HasExtension( |
| const $pbi$::ExtensionIdentifier<$Msg$, _proto_TypeTraits, |
| _field_type, _is_packed>& id) const { |
| $annotate_extension_has$; |
| return $extensions$.Has(id.number()); |
| } |
| |
| template <typename _proto_TypeTraits, $pbi$::FieldType _field_type, |
| bool _is_packed> |
| inline void ClearExtension( |
| const $pbi$::ExtensionIdentifier<$Msg$, _proto_TypeTraits, |
| _field_type, _is_packed>& id) { |
| $extensions$.ClearExtension(id.number()); |
| $annotate_extension_clear$; |
| } |
| |
| template <typename _proto_TypeTraits, $pbi$::FieldType _field_type, |
| bool _is_packed> |
| inline int ExtensionSize( |
| const $pbi$::ExtensionIdentifier<$Msg$, _proto_TypeTraits, |
| _field_type, _is_packed>& id) const { |
| $annotate_extension_repeated_size$; |
| return $extensions$.ExtensionSize(id.number()); |
| } |
| |
| template <typename _proto_TypeTraits, $pbi$::FieldType _field_type, |
| bool _is_packed, |
| std::enable_if_t<!_proto_TypeTraits::kLifetimeBound, int> = 0> |
| inline typename _proto_TypeTraits::Singular::ConstType GetExtension( |
| const $pbi$::ExtensionIdentifier<$Msg$, _proto_TypeTraits, |
| _field_type, _is_packed>& id) const { |
| $annotate_extension_get$; |
| return _proto_TypeTraits::Get(id.number(), $extensions$, id.default_value()); |
| } |
| |
| template <typename _proto_TypeTraits, $pbi$::FieldType _field_type, |
| bool _is_packed, |
| std::enable_if_t<_proto_TypeTraits::kLifetimeBound, int> = 0> |
| inline typename _proto_TypeTraits::Singular::ConstType GetExtension( |
| const $pbi$::ExtensionIdentifier<$Msg$, _proto_TypeTraits, |
| _field_type, _is_packed>& id) const |
| ABSL_ATTRIBUTE_LIFETIME_BOUND { |
| $annotate_extension_get$; |
| return _proto_TypeTraits::Get(id.number(), $extensions$, id.default_value()); |
| } |
| |
| template <typename _proto_TypeTraits, $pbi$::FieldType _field_type, |
| bool _is_packed> |
| inline typename _proto_TypeTraits::Singular::MutableType MutableExtension( |
| const $pbi$::ExtensionIdentifier<$Msg$, _proto_TypeTraits, |
| _field_type, _is_packed>& id) |
| ABSL_ATTRIBUTE_LIFETIME_BOUND { |
| $annotate_extension_mutable$; |
| return _proto_TypeTraits::Mutable(id.number(), _field_type, &$extensions$); |
| } |
| |
| template <typename _proto_TypeTraits, $pbi$::FieldType _field_type, |
| bool _is_packed> |
| inline void SetExtension( |
| const $pbi$::ExtensionIdentifier<$Msg$, _proto_TypeTraits, |
| _field_type, _is_packed>& id, |
| typename _proto_TypeTraits::Singular::ConstType value) { |
| _proto_TypeTraits::Set(id.number(), _field_type, value, &$extensions$); |
| $annotate_extension_set$; |
| } |
| |
| template <typename _proto_TypeTraits, $pbi$::FieldType _field_type, |
| bool _is_packed> |
| inline void SetAllocatedExtension( |
| const $pbi$::ExtensionIdentifier<$Msg$, _proto_TypeTraits, |
| _field_type, _is_packed>& id, |
| typename _proto_TypeTraits::Singular::MutableType value) { |
| _proto_TypeTraits::SetAllocated(id.number(), _field_type, value, |
| &$extensions$); |
| $annotate_extension_set$; |
| } |
| template <typename _proto_TypeTraits, $pbi$::FieldType _field_type, |
| bool _is_packed> |
| inline void UnsafeArenaSetAllocatedExtension( |
| const $pbi$::ExtensionIdentifier<$Msg$, _proto_TypeTraits, |
| _field_type, _is_packed>& id, |
| typename _proto_TypeTraits::Singular::MutableType value) { |
| _proto_TypeTraits::UnsafeArenaSetAllocated(id.number(), _field_type, |
| value, &$extensions$); |
| $annotate_extension_set$; |
| } |
| template <typename _proto_TypeTraits, $pbi$::FieldType _field_type, |
| bool _is_packed> |
| PROTOBUF_NODISCARD inline |
| typename _proto_TypeTraits::Singular::MutableType |
| ReleaseExtension( |
| const $pbi$::ExtensionIdentifier<$Msg$, _proto_TypeTraits, |
| _field_type, _is_packed>& id) { |
| $annotate_extension_release$; |
| return _proto_TypeTraits::Release(id.number(), _field_type, &$extensions$); |
| } |
| template <typename _proto_TypeTraits, $pbi$::FieldType _field_type, |
| bool _is_packed> |
| inline typename _proto_TypeTraits::Singular::MutableType |
| UnsafeArenaReleaseExtension( |
| const $pbi$::ExtensionIdentifier<$Msg$, _proto_TypeTraits, |
| _field_type, _is_packed>& id) { |
| $annotate_extension_release$; |
| return _proto_TypeTraits::UnsafeArenaRelease(id.number(), _field_type, |
| &$extensions$); |
| } |
| |
| template <typename _proto_TypeTraits, $pbi$::FieldType _field_type, |
| bool _is_packed, |
| std::enable_if_t<!_proto_TypeTraits::kLifetimeBound, int> = 0> |
| inline typename _proto_TypeTraits::Repeated::ConstType GetExtension( |
| const $pbi$::ExtensionIdentifier<$Msg$, _proto_TypeTraits, |
| _field_type, _is_packed>& id, |
| int index) const { |
| $annotate_repeated_extension_get$; |
| return _proto_TypeTraits::Get(id.number(), $extensions$, index); |
| } |
| |
| template <typename _proto_TypeTraits, $pbi$::FieldType _field_type, |
| bool _is_packed, |
| std::enable_if_t<_proto_TypeTraits::kLifetimeBound, int> = 0> |
| inline typename _proto_TypeTraits::Repeated::ConstType GetExtension( |
| const $pbi$::ExtensionIdentifier<$Msg$, _proto_TypeTraits, |
| _field_type, _is_packed>& id, |
| int index) const ABSL_ATTRIBUTE_LIFETIME_BOUND { |
| $annotate_repeated_extension_get$; |
| return _proto_TypeTraits::Get(id.number(), $extensions$, index); |
| } |
| |
| template <typename _proto_TypeTraits, $pbi$::FieldType _field_type, |
| bool _is_packed> |
| inline typename _proto_TypeTraits::Repeated::MutableType MutableExtension( |
| const $pbi$::ExtensionIdentifier<$Msg$, _proto_TypeTraits, |
| _field_type, _is_packed>& id, |
| int index) ABSL_ATTRIBUTE_LIFETIME_BOUND { |
| $annotate_repeated_extension_mutable$; |
| return _proto_TypeTraits::Mutable(id.number(), index, &$extensions$); |
| } |
| |
| template <typename _proto_TypeTraits, $pbi$::FieldType _field_type, |
| bool _is_packed> |
| inline void SetExtension( |
| const $pbi$::ExtensionIdentifier<$Msg$, _proto_TypeTraits, |
| _field_type, _is_packed>& id, |
| int index, typename _proto_TypeTraits::Repeated::ConstType value) { |
| _proto_TypeTraits::Set(id.number(), index, value, &$extensions$); |
| $annotate_repeated_extension_set$; |
| } |
| |
| template <typename _proto_TypeTraits, $pbi$::FieldType _field_type, |
| bool _is_packed> |
| inline typename _proto_TypeTraits::Repeated::MutableType AddExtension( |
| const $pbi$::ExtensionIdentifier<$Msg$, _proto_TypeTraits, |
| _field_type, _is_packed>& id) |
| ABSL_ATTRIBUTE_LIFETIME_BOUND { |
| typename _proto_TypeTraits::Repeated::MutableType to_add = |
| _proto_TypeTraits::Add(id.number(), _field_type, &$extensions$); |
| $annotate_repeated_extension_add_mutable$; |
| return to_add; |
| } |
| |
| template <typename _proto_TypeTraits, $pbi$::FieldType _field_type, |
| bool _is_packed> |
| inline void AddExtension( |
| const $pbi$::ExtensionIdentifier<$Msg$, _proto_TypeTraits, |
| _field_type, _is_packed>& id, |
| typename _proto_TypeTraits::Repeated::ConstType value) { |
| _proto_TypeTraits::Add(id.number(), _field_type, _is_packed, value, |
| &$extensions$); |
| $annotate_repeated_extension_add$; |
| } |
| |
| template <typename _proto_TypeTraits, $pbi$::FieldType _field_type, |
| bool _is_packed> |
| inline const typename _proto_TypeTraits::Repeated::RepeatedFieldType& |
| GetRepeatedExtension( |
| const $pbi$::ExtensionIdentifier<$Msg$, _proto_TypeTraits, |
| _field_type, _is_packed>& id) const |
| ABSL_ATTRIBUTE_LIFETIME_BOUND { |
| $annotate_repeated_extension_list$; |
| return _proto_TypeTraits::GetRepeated(id.number(), $extensions$); |
| } |
| |
| template <typename _proto_TypeTraits, $pbi$::FieldType _field_type, |
| bool _is_packed> |
| inline typename _proto_TypeTraits::Repeated::RepeatedFieldType* |
| MutableRepeatedExtension( |
| const $pbi$::ExtensionIdentifier<$Msg$, _proto_TypeTraits, |
| _field_type, _is_packed>& id) |
| ABSL_ATTRIBUTE_LIFETIME_BOUND { |
| $annotate_repeated_extension_list_mutable$; |
| return _proto_TypeTraits::MutableRepeated(id.number(), _field_type, |
| _is_packed, &$extensions$); |
| } |
| )cc"); |
| |
| // Generate MessageSet specific APIs for proto2 MessageSet. |
| // For testing purposes we don't check for bridge.MessageSet, so |
| // we don't use IsProto2MessageSet |
| if (descriptor_->options().message_set_wire_format() && |
| !options_.opensource_runtime && !options_.lite_implicit_weak_fields) { |
| // Special-case MessageSet. |
| p->Emit(R"cc( |
| GOOGLE_PROTOBUF_EXTENSION_MESSAGE_SET_ACCESSORS($Msg$); |
| )cc"); |
| } |
| } |
| |
| for (auto oneof : OneOfRange(descriptor_)) { |
| p->Emit({{"oneof_name", oneof->name()}, |
| Sub{"clear_oneof_name", absl::StrCat("clear_", oneof->name())} |
| .AnnotatedAs({oneof, Semantic::kSet}), |
| {"OneOfName", UnderscoresToCamelCase(oneof->name(), true)}}, |
| R"cc( |
| void $clear_oneof_name$(); |
| $OneOfName$Case $oneof_name$_case() const; |
| )cc"); |
| } |
| } |
| |
| void MessageGenerator::GenerateSingularFieldHasBits( |
| const FieldDescriptor* field, io::Printer* p) { |
| auto t = p->WithVars(MakeTrackerCalls(field, options_)); |
| if (field->options().weak()) { |
| p->Emit( |
| R"cc( |
| inline bool $classname$::has_$name$() const { |
| $annotate_has$; |
| return $weak_field_map$.Has($number$); |
| } |
| )cc"); |
| return; |
| } |
| if (HasHasbit(field)) { |
| auto v = p->WithVars(HasBitVars(field)); |
| p->Emit( |
| {Sub{"ASSUME", |
| [&] { |
| if (field->cpp_type() == FieldDescriptor::CPPTYPE_MESSAGE && |
| !IsLazy(field, options_, scc_analyzer_)) { |
| // We maintain the invariant that for a submessage x, has_x() |
| // returning true implies that x_ is not null. By giving this |
| // information to the compiler, we allow it to eliminate |
| // unnecessary null checks later on. |
| p->Emit( |
| R"cc(PROTOBUF_ASSUME(!value || $field$ != nullptr);)cc"); |
| } |
| }} |
| .WithSuffix(";")}, |
| R"cc( |
| inline bool $classname$::has_$name$() const { |
| $annotate_has$; |
| bool value = ($has_bits$[$has_array_index$] & $has_mask$) != 0; |
| $ASSUME$; |
| return value; |
| } |
| )cc"); |
| } else if (field->cpp_type() == FieldDescriptor::CPPTYPE_MESSAGE) { |
| // Message fields have a has_$name$() method. |
| if (IsLazy(field, options_, scc_analyzer_)) { |
| p->Emit(R"cc( |
| inline bool $classname$::_internal_has_$name$() const { |
| return !$field$.IsCleared(); |
| } |
| )cc"); |
| } else { |
| p->Emit(R"cc( |
| inline bool $classname$::_internal_has_$name$() const { |
| return this != internal_default_instance() && $field$ != nullptr; |
| } |
| )cc"); |
| } |
| p->Emit(R"cc( |
| inline bool $classname$::has_$name$() const { |
| $annotate_has$; |
| return _internal_has_$name$(); |
| } |
| )cc"); |
| } |
| } |
| |
| void MessageGenerator::GenerateOneofHasBits(io::Printer* p) { |
| for (const auto* oneof : OneOfRange(descriptor_)) { |
| p->Emit( |
| { |
| {"oneof_index", oneof->index()}, |
| {"oneof_name", oneof->name()}, |
| {"cap_oneof_name", absl::AsciiStrToUpper(oneof->name())}, |
| }, |
| R"cc( |
| inline bool $classname$::has_$oneof_name$() const { |
| return $oneof_name$_case() != $cap_oneof_name$_NOT_SET; |
| } |
| inline void $classname$::clear_has_$oneof_name$() { |
| $oneof_case$[$oneof_index$] = $cap_oneof_name$_NOT_SET; |
| } |
| )cc"); |
| } |
| } |
| |
| void MessageGenerator::GenerateOneofMemberHasBits(const FieldDescriptor* field, |
| io::Printer* p) { |
| // Singular field in a oneof |
| // N.B.: Without field presence, we do not use has-bits or generate |
| // has_$name$() methods, but oneofs still have set_has_$name$(). |
| // Oneofs also have private _internal_has_$name$() a helper method. |
| if (field->has_presence()) { |
| auto t = p->WithVars(MakeTrackerCalls(field, options_)); |
| p->Emit(R"cc( |
| inline bool $classname$::has_$name$() const { |
| $annotate_has$; |
| return $has_field$; |
| } |
| )cc"); |
| } |
| if (HasInternalHasMethod(field)) { |
| p->Emit(R"cc( |
| inline bool $classname$::_internal_has_$name$() const { |
| return $has_field$; |
| } |
| )cc"); |
| } |
| // set_has_$name$() for oneof fields is always private; hence should not be |
| // annotated. |
| p->Emit(R"cc( |
| inline void $classname$::set_has_$name$() { |
| $oneof_case$[$oneof_index$] = k$field_name$; |
| } |
| )cc"); |
| } |
| |
| void MessageGenerator::GenerateFieldClear(const FieldDescriptor* field, |
| bool is_inline, io::Printer* p) { |
| auto t = p->WithVars(MakeTrackerCalls(field, options_)); |
| p->Emit({{"inline", is_inline ? "inline" : ""}, |
| {"body", |
| [&] { |
| if (field->real_containing_oneof()) { |
| // Clear this field only if it is the active field in this |
| // oneof, otherwise ignore |
| p->Emit( |
| {{"clearing_code", |
| [&] { |
| field_generators_.get(field).GenerateClearingCode(p); |
| }}}, |
| R"cc( |
| if ($has_field$) { |
| $clearing_code$; |
| clear_has_$oneof_name$(); |
| } |
| )cc"); |
| } else { |
| // TODO: figure out if early return breaks tracking |
| if (ShouldSplit(field, options_)) { |
| p->Emit(R"cc( |
| if (PROTOBUF_PREDICT_TRUE(IsSplitMessageDefault())) |
| return; |
| )cc"); |
| } |
| field_generators_.get(field).GenerateClearingCode(p); |
| if (HasHasbit(field)) { |
| auto v = p->WithVars(HasBitVars(field)); |
| p->Emit(R"cc( |
| $has_bits$[$has_array_index$] &= ~$has_mask$; |
| )cc"); |
| } |
| } |
| }}}, |
| R"cc( |
| $inline $void $classname$::clear_$name$() { |
| PROTOBUF_TSAN_WRITE(&_impl_._tsan_detect_race); |
| $body$; |
| $annotate_clear$; |
| } |
| )cc"); |
| } |
| |
| void MessageGenerator::GenerateFieldAccessorDefinitions(io::Printer* p) { |
| p->Emit("// $classname$\n\n"); |
| |
| for (auto field : FieldRange(descriptor_)) { |
| PrintFieldComment(Formatter{p}, field, options_); |
| |
| auto v = p->WithVars(FieldVars(field, options_)); |
| auto t = p->WithVars(MakeTrackerCalls(field, options_)); |
| if (field->is_repeated()) { |
| p->Emit(R"cc( |
| inline int $classname$::_internal_$name$_size() const { |
| return _internal_$name$().size(); |
| } |
| inline int $classname$::$name$_size() const { |
| $annotate_size$; |
| return _internal_$name$_size(); |
| } |
| )cc"); |
| } else if (field->real_containing_oneof()) { |
| GenerateOneofMemberHasBits(field, p); |
| } else { |
| GenerateSingularFieldHasBits(field, p); |
| } |
| |
| if (!IsCrossFileMaybeMap(field)) { |
| GenerateFieldClear(field, true, p); |
| } |
| // Generate type-specific accessors. |
| field_generators_.get(field).GenerateInlineAccessorDefinitions(p); |
| |
| p->Emit("\n"); |
| } |
| |
| GenerateOneofHasBits(p); |
| } |
| |
| void MessageGenerator::GenerateMapEntryClassDefinition(io::Printer* p) { |
| Formatter format(p); |
| absl::flat_hash_map<absl::string_view, std::string> vars; |
| CollectMapInfo(options_, descriptor_, &vars); |
| vars["lite"] = |
| HasDescriptorMethods(descriptor_->file(), options_) ? "" : "Lite"; |
| auto v = p->WithVars(std::move(vars)); |
| // Templatize constexpr constructor as a workaround for a bug in gcc 12 |
| // (warning in gcc 13). |
| p->Emit(R"cc( |
| class $classname$ final |
| : public ::$proto_ns$::internal::MapEntry$lite$< |
| $classname$, $key_cpp$, $val_cpp$, |
| ::$proto_ns$::internal::WireFormatLite::$key_wire_type$, |
| ::$proto_ns$::internal::WireFormatLite::$val_wire_type$> { |
| public: |
| using SuperType = ::$proto_ns$::internal::MapEntry$lite$< |
| $classname$, $key_cpp$, $val_cpp$, |
| ::$proto_ns$::internal::WireFormatLite::$key_wire_type$, |
| ::$proto_ns$::internal::WireFormatLite::$val_wire_type$>; |
| $classname$(); |
| template <typename = void> |
| explicit PROTOBUF_CONSTEXPR $classname$( |
| ::$proto_ns$::internal::ConstantInitialized); |
| explicit $classname$(::$proto_ns$::Arena* arena); |
| static const $classname$* internal_default_instance() { |
| return reinterpret_cast<const $classname$*>( |
| &_$classname$_default_instance_); |
| } |
| )cc"); |
| auto utf8_check = internal::cpp::GetUtf8CheckMode( |
| descriptor_->field(0), GetOptimizeFor(descriptor_->file(), options_) == |
| FileOptions::LITE_RUNTIME); |
| if (descriptor_->field(0)->type() == FieldDescriptor::TYPE_STRING && |
| utf8_check != Utf8CheckMode::kNone) { |
| if (utf8_check == Utf8CheckMode::kStrict) { |
| format( |
| " static bool ValidateKey(std::string* s) {\n" |
| " return ::$proto_ns$::internal::WireFormatLite::" |
| "VerifyUtf8String(s->data(), static_cast<int>(s->size()), " |
| "::$proto_ns$::internal::WireFormatLite::PARSE, \"$1$\");\n" |
| " }\n", |
| descriptor_->field(0)->full_name()); |
| } else { |
| ABSL_CHECK(utf8_check == Utf8CheckMode::kVerify); |
| format( |
| " static bool ValidateKey(std::string* s) {\n" |
| "#ifndef NDEBUG\n" |
| " ::$proto_ns$::internal::WireFormatLite::VerifyUtf8String(\n" |
| " s->data(), static_cast<int>(s->size()), " |
| "::$proto_ns$::internal::" |
| "WireFormatLite::PARSE, \"$1$\");\n" |
| "#else\n" |
| " (void) s;\n" |
| "#endif\n" |
| " return true;\n" |
| " }\n", |
| descriptor_->field(0)->full_name()); |
| } |
| } else { |
| format(" static bool ValidateKey(void*) { return true; }\n"); |
| } |
| if (descriptor_->field(1)->type() == FieldDescriptor::TYPE_STRING && |
| utf8_check != Utf8CheckMode::kNone) { |
| if (utf8_check == Utf8CheckMode::kStrict) { |
| format( |
| " static bool ValidateValue(std::string* s) {\n" |
| " return ::$proto_ns$::internal::WireFormatLite::" |
| "VerifyUtf8String(s->data(), static_cast<int>(s->size()), " |
| "::$proto_ns$::internal::WireFormatLite::PARSE, \"$1$\");\n" |
| " }\n", |
| descriptor_->field(1)->full_name()); |
| } else { |
| ABSL_CHECK(utf8_check == Utf8CheckMode::kVerify); |
| format( |
| " static bool ValidateValue(std::string* s) {\n" |
| "#ifndef NDEBUG\n" |
| " ::$proto_ns$::internal::WireFormatLite::VerifyUtf8String(\n" |
| " s->data(), static_cast<int>(s->size()), " |
| "::$proto_ns$::internal::" |
| "WireFormatLite::PARSE, \"$1$\");\n" |
| "#else\n" |
| " (void) s;\n" |
| "#endif\n" |
| " return true;\n" |
| " }\n", |
| descriptor_->field(1)->full_name()); |
| } |
| } else { |
| format(" static bool ValidateValue(void*) { return true; }\n"); |
| } |
| if (HasDescriptorMethods(descriptor_->file(), options_)) { |
| format(" ::$proto_ns$::Metadata GetMetadata() const final;\n"); |
| } |
| format( |
| " friend struct ::$tablename$;\n" |
| "};\n"); |
| } |
| |
| void MessageGenerator::GenerateImplDefinition(io::Printer* p) { |
| // Prepare decls for _cached_size_ and _has_bits_. Their position in the |
| // output will be determined later. |
| bool need_to_emit_cached_size = !HasSimpleBaseClass(descriptor_, options_); |
| const size_t sizeof_has_bits = HasBitsSize(); |
| |
| // To minimize padding, data members are divided into three sections: |
| // (1) members assumed to align to 8 bytes |
| // (2) members corresponding to message fields, re-ordered to optimize |
| // alignment. |
| // (3) members assumed to align to 4 bytes. |
| p->Emit( |
| {{"extension_set", |
| [&] { |
| if (descriptor_->extension_range_count() == 0) return; |
| |
| p->Emit(R"cc( |
| ::$proto_ns$::internal::ExtensionSet _extensions_; |
| )cc"); |
| }}, |
| {"tracker", |
| [&] { |
| if (!HasTracker(descriptor_, options_)) return; |
| |
| p->Emit(R"cc( |
| static ::$proto_ns$::AccessListener<$Msg$> _tracker_; |
| )cc"); |
| }}, |
| {"inlined_string_donated", |
| [&] { |
| // Generate _inlined_string_donated_ for inlined string type. |
| // TODO: To avoid affecting the locality of |
| // `_has_bits_`, should this be below or above `_has_bits_`? |
| if (inlined_string_indices_.empty()) return; |
| |
| p->Emit({{"donated_size", InlinedStringDonatedSize()}}, |
| R"cc( |
| ::$proto_ns$::internal::HasBits<$donated_size$> |
| _inlined_string_donated_; |
| )cc"); |
| }}, |
| {"has_bits", |
| [&] { |
| if (has_bit_indices_.empty()) return; |
| |
| // _has_bits_ is frequently accessed, so to reduce code size and |
| // improve speed, it should be close to the start of the object. |
| // Placing _cached_size_ together with _has_bits_ improves cache |
| // locality despite potential alignment padding. |
| p->Emit({{"sizeof_has_bits", sizeof_has_bits}}, R"cc( |
| ::$proto_ns$::internal::HasBits<$sizeof_has_bits$> _has_bits_; |
| )cc"); |
| if (need_to_emit_cached_size) { |
| p->Emit(R"cc( |
| mutable ::$proto_ns$::internal::CachedSize _cached_size_; |
| )cc"); |
| need_to_emit_cached_size = false; |
| } |
| }}, |
| {"field_members", |
| [&] { |
| // Emit some private and static members |
| for (auto field : optimized_order_) { |
| field_generators_.get(field).GenerateStaticMembers(p); |
| if (!ShouldSplit(field, options_)) { |
| field_generators_.get(field).GeneratePrivateMembers(p); |
| } |
| } |
| }}, |
| {"decl_split", |
| [&] { |
| if (!ShouldSplit(descriptor_, options_)) return; |
| p->Emit({{"split_field", |
| [&] { |
| for (auto field : optimized_order_) { |
| if (!ShouldSplit(field, options_)) continue; |
| field_generators_.get(field).GeneratePrivateMembers(p); |
| } |
| }}}, |
| R"cc( |
| struct Split { |
| $split_field$; |
| using InternalArenaConstructable_ = void; |
| using DestructorSkippable_ = void; |
| }; |
| static_assert(std::is_trivially_copy_constructible<Split>::value); |
| static_assert(std::is_trivially_destructible<Split>::value); |
| Split* _split_; |
| )cc"); |
| }}, |
| {"oneof_members", |
| [&] { |
| // For each oneof generate a union |
| for (auto oneof : OneOfRange(descriptor_)) { |
| // explicit empty constructor is needed when union contains |
| // ArenaStringPtr members for string fields. |
| p->Emit( |
| {{"camel_oneof_name", |
| UnderscoresToCamelCase(oneof->name(), true)}, |
| {"oneof_name", oneof->name()}, |
| {"oneof_fields", |
| [&] { |
| for (auto field : FieldRange(oneof)) { |
| field_generators_.get(field).GeneratePrivateMembers(p); |
| } |
| }}}, |
| R"cc( |
| union $camel_oneof_name$Union { |
| constexpr $camel_oneof_name$Union() : _constinit_{} {} |
| ::$proto_ns$::internal::ConstantInitialized _constinit_; |
| $oneof_fields$; |
| } $oneof_name$_; |
| )cc"); |
| for (auto field : FieldRange(oneof)) { |
| field_generators_.get(field).GenerateStaticMembers(p); |
| } |
| } |
| }}, |
| {"cached_size_if_no_hasbits", |
| [&] { |
| if (!need_to_emit_cached_size) return; |
| |
| need_to_emit_cached_size = false; |
| p->Emit(R"cc( |
| mutable ::$proto_ns$::internal::CachedSize _cached_size_; |
| )cc"); |
| }}, |
| {"oneof_case", |
| [&] { |
| // Generate _oneof_case_. |
| if (descriptor_->real_oneof_decl_count() == 0) return; |
| |
| p->Emit({{"count", descriptor_->real_oneof_decl_count()}}, |
| R"cc( |
| $uint32$ _oneof_case_[$count$]; |
| )cc"); |
| }}, |
| {"weak_field_map", |
| [&] { |
| if (num_weak_fields_ == 0) return; |
| |
| p->Emit(R"cc( |
| ::$proto_ns$::internal::WeakFieldMap _weak_field_map_; |
| )cc"); |
| }}, |
| {"any_metadata", |
| [&] { |
| // Generate _any_metadata_ for the Any type. |
| if (!IsAnyMessage(descriptor_)) return; |
| |
| p->Emit(R"cc( |
| ::$proto_ns$::internal::AnyMetadata _any_metadata_; |
| )cc"); |
| }}, |
| {"union_impl", |
| [&] { |
| // Only create the _impl_ field if it contains data. |
| if (!HasImplData(descriptor_, options_)) return; |
| |
| // clang-format off |
| p->Emit(R"cc(union { Impl_ _impl_; };)cc"); |
| // clang-format on |
| }}}, |
| R"cc( |
| struct Impl_ { |
| //~ TODO: check if/when there is a need for an |
| //~ outline dtor. |
| inline explicit constexpr Impl_( |
| ::$proto_ns$::internal::ConstantInitialized) noexcept; |
| inline explicit Impl_($pbi$::InternalVisibility visibility, |
| ::$proto_ns$::Arena* arena); |
| inline explicit Impl_($pbi$::InternalVisibility visibility, |
| ::$proto_ns$::Arena* arena, const Impl_& from); |
| //~ Members assumed to align to 8 bytes: |
| $extension_set$; |
| $tracker$; |
| $inlined_string_donated$; |
| $has_bits$; |
| //~ Field members: |
| $field_members$; |
| $decl_split$; |
| $oneof_members$; |
| //~ Members assumed to align to 4 bytes: |
| $cached_size_if_no_hasbits$; |
| $oneof_case$; |
| $weak_field_map$; |
| $any_metadata$; |
| //~ For detecting when concurrent accessor calls cause races. |
| PROTOBUF_TSAN_DECLARE_MEMBER |
| }; |
| $union_impl$; |
| )cc"); |
| |
| ABSL_DCHECK(!need_to_emit_cached_size); |
| } |
| |
| void MessageGenerator::GenerateAnyMethodDefinition(io::Printer* p) { |
| ABSL_DCHECK(IsAnyMessage(descriptor_)); |
| |
| p->Emit({{"any_methods", |
| [&] { |
| if (HasDescriptorMethods(descriptor_->file(), options_)) { |
| p->Emit( |
| R"cc( |
| bool PackFrom(const ::$proto_ns$::Message& message) { |
| $DCHK$_NE(&message, this); |
| return $any_metadata$.PackFrom(GetArena(), message); |
| } |
| bool PackFrom(const ::$proto_ns$::Message& message, |
| ::absl::string_view type_url_prefix) { |
| $DCHK$_NE(&message, this); |
| return $any_metadata$.PackFrom(GetArena(), message, type_url_prefix); |
| } |
| bool UnpackTo(::$proto_ns$::Message* message) const { |
| return $any_metadata$.UnpackTo(message); |
| } |
| static bool GetAnyFieldDescriptors( |
| const ::$proto_ns$::Message& message, |
| const ::$proto_ns$::FieldDescriptor** type_url_field, |
| const ::$proto_ns$::FieldDescriptor** value_field); |
| template < |
| typename T, |
| class = typename std::enable_if<!std::is_convertible< |
| T, const ::$proto_ns$::Message&>::value>::type> |
| bool PackFrom(const T& message) { |
| return $any_metadata$.PackFrom<T>(GetArena(), message); |
| } |
| template < |
| typename T, |
| class = typename std::enable_if<!std::is_convertible< |
| T, const ::$proto_ns$::Message&>::value>::type> |
| bool PackFrom(const T& message, |
| ::absl::string_view type_url_prefix) { |
| return $any_metadata$.PackFrom<T>(GetArena(), message, type_url_prefix); |
| } |
| template < |
| typename T, |
| class = typename std::enable_if<!std::is_convertible< |
| T, const ::$proto_ns$::Message&>::value>::type> |
| bool UnpackTo(T* message) const { |
| return $any_metadata$.UnpackTo<T>(message); |
| } |
| )cc"); |
| } else { |
| p->Emit( |
| R"cc( |
| template <typename T> |
| bool PackFrom(const T& message) { |
| return $any_metadata$.PackFrom(GetArena(), message); |
| } |
| template <typename T> |
| bool PackFrom(const T& message, |
| ::absl::string_view type_url_prefix) { |
| return $any_metadata$.PackFrom(GetArena(), message, type_url_prefix); |
| } |
| template <typename T> |
| bool UnpackTo(T* message) const { |
| return $any_metadata$.UnpackTo(message); |
| } |
| )cc"); |
| } |
| }}}, |
| R"cc( |
| // implements Any |
| // ----------------------------------------------- |
| |
| $any_methods$; |
| |
| template <typename T> |
| bool Is() const { |
| return $any_metadata$.Is<T>(); |
| } |
| static bool ParseAnyTypeUrl(::absl::string_view type_url, |
| std::string* full_type_name); |
| )cc"); |
| } |
| |
| void MessageGenerator::GenerateClassDefinition(io::Printer* p) { |
| if (!ShouldGenerateClass(descriptor_, options_)) return; |
| |
| auto v = p->WithVars(ClassVars(descriptor_, options_)); |
| auto t = p->WithVars(MakeTrackerCalls(descriptor_, options_)); |
| Formatter format(p); |
| |
| if (IsMapEntryMessage(descriptor_)) { |
| GenerateMapEntryClassDefinition(p); |
| return; |
| } |
| |
| auto annotation = p->WithAnnotations({{"classname", descriptor_}}); |
| p->Emit( |
| {{"decl_dtor", |
| [&] { |
| if (HasSimpleBaseClass(descriptor_, options_)) return; |
| |
| p->Emit(R"cc( |
| ~$classname$() override; |
| )cc"); |
| }}, |
| {"decl_annotate", |
| [&] { |
| }}, |
| {"decl_verify_func", |
| [&] { |
| }}, |
| {"descriptor_accessor", |
| [&] { |
| // Only generate this member if it's not disabled. |
| if (!HasDescriptorMethods(descriptor_->file(), options_) || |
| descriptor_->options().no_standard_descriptor_accessor()) { |
| return; |
| } |
| |
| p->Emit(R"cc( |
| static const ::$proto_ns$::Descriptor* descriptor() { |
| return GetDescriptor(); |
| } |
| )cc"); |
| }}, |
| {"get_descriptor", |
| [&] { |
| // These shadow non-static methods of the same names in Message. |
| // We redefine them here because calls directly on the generated |
| // class can be statically analyzed -- we know what descriptor |
| // types are being requested. It also avoids a vtable dispatch. |
| // |
| // We would eventually like to eliminate the methods in Message, |
| // and having this separate also lets us track calls to the base |
| // class methods separately. |
| if (!HasDescriptorMethods(descriptor_->file(), options_)) return; |
| |
| p->Emit(R"cc( |
| static const ::$proto_ns$::Descriptor* GetDescriptor() { |
| return default_instance().GetMetadata().descriptor; |
| } |
| static const ::$proto_ns$::Reflection* GetReflection() { |
| return default_instance().GetMetadata().reflection; |
| } |
| )cc"); |
| }}, |
| {"decl_oneof", |
| [&] { |
| // Generate enum values for every field in oneofs. One list is |
| // generated for each oneof with an additional *_NOT_SET value. |
| for (auto oneof : OneOfRange(descriptor_)) { |
| p->Emit( |
| {{"oneof_camel_name", |
| UnderscoresToCamelCase(oneof->name(), true)}, |
| {"oneof_field", |
| [&] { |
| for (auto field : FieldRange(oneof)) { |
| p->Emit( |
| { |
| {"oneof_constant", OneofCaseConstantName(field)}, |
| {"field_number", field->number()}, |
| }, |
| R"cc( |
| $oneof_constant$ = $field_number$, |
| )cc"); |
| } |
| }}, |
| {"upper_oneof_name", absl::AsciiStrToUpper(oneof->name())}}, |
| R"cc( |
| enum $oneof_camel_name$Case { |
| $oneof_field$, |
| $upper_oneof_name$_NOT_SET = 0, |
| }; |
| )cc"); |
| } |
| }}, |
| {"index_in_file_messages", index_in_file_messages_}, |
| {"decl_any_methods", |
| [&] { |
| if (!IsAnyMessage(descriptor_)) return; |
| |
| GenerateAnyMethodDefinition(p); |
| }}, |
| {"generated_methods", |
| [&] { |
| if (!HasGeneratedMethods(descriptor_->file(), options_)) return; |
| |
| if (HasDescriptorMethods(descriptor_->file(), options_)) { |
| if (!HasSimpleBaseClass(descriptor_, options_)) { |
| // Use Message's built-in MergeFrom and CopyFrom when the |
| // passed-in argument is a generic Message instance, and |
| // only define the custom MergeFrom and CopyFrom |
| // instances when the source of the merge/copy is known |
| // to be the same class as the destination. |
| p->Emit(R"cc( |
| using $superclass$::CopyFrom; |
| void CopyFrom(const $classname$& from); |
| using $superclass$::MergeFrom; |
| void MergeFrom(const $classname$& from) { $classname$::MergeImpl(*this, from); } |
| |
| private: |
| static void MergeImpl(::$proto_ns$::Message& to_msg, |
| const ::$proto_ns$::Message& from_msg); |
| |
| public: |
| )cc"); |
| } else { |
| p->Emit(R"cc( |
| using $superclass$::CopyFrom; |
| inline void CopyFrom(const $classname$& from) { |
| $superclass$::CopyImpl(*this, from); |
| } |
| using $superclass$::MergeFrom; |
| void MergeFrom(const $classname$& from) { |
| $superclass$::MergeImpl(*this, from); |
| } |
| |
| public: |
| )cc"); |
| } |
| } else { |
| p->Emit(R"cc( |
| void CheckTypeAndMergeFrom( |
| const ::$proto_ns$::MessageLite& from) final; |
| void CopyFrom(const $classname$& from); |
| void MergeFrom(const $classname$& from); |
| )cc"); |
| } |
| |
| if (!HasSimpleBaseClass(descriptor_, options_)) { |
| p->Emit(R"cc( |
| ABSL_ATTRIBUTE_REINITIALIZES void Clear() final; |
| bool IsInitialized() const final; |
| |
| ::size_t ByteSizeLong() const final; |
| )cc"); |
| |
| parse_function_generator_->GenerateMethodDecls(p); |
| |
| p->Emit(R"cc( |
| $uint8$* _InternalSerialize( |
| $uint8$* target, |
| ::$proto_ns$::io::EpsCopyOutputStream* stream) const final; |
| )cc"); |
| } |
| }}, |
| {"internal_field_number", |
| [&] { |
| if (!options_.field_listener_options.inject_field_listener_events) |
| return; |
| |
| p->Emit({{"field_count", descriptor_->field_count()}}, R"cc( |
| static constexpr int _kInternalFieldNumber = $field_count$; |
| )cc"); |
| }}, |
| {"decl_non_simple_base", |
| [&] { |
| if (HasSimpleBaseClass(descriptor_, options_)) return; |
| p->Emit( |
| R"cc( |
| int GetCachedSize() const { return $cached_size$.Get(); } |
| |
| private: |
| void SharedCtor(::$proto_ns$::Arena* arena); |
| void SharedDtor(); |
| void InternalSwap($classname$* other); |
| )cc"); |
| }}, |
| {"arena_dtor", |
| [&] { |
| switch (NeedsArenaDestructor()) { |
| case ArenaDtorNeeds::kOnDemand: |
| p->Emit(R"cc( |
| private: |
| static void ArenaDtor(void* object); |
| static void OnDemandRegisterArenaDtor( |
| MessageLite& msg, ::$proto_ns$::Arena& arena) { |
| auto& this_ = static_cast<$classname$&>(msg); |
| if ((this_.$inlined_string_donated_array$[0] & 0x1u) == 0) { |
| return; |
| } |
| this_.$inlined_string_donated_array$[0] &= 0xFFFFFFFEu; |
| arena.OwnCustomDestructor(&this_, &$classname$::ArenaDtor); |
| } |
| )cc"); |
| break; |
| case ArenaDtorNeeds::kRequired: |
| p->Emit(R"cc( |
| private: |
| static void ArenaDtor(void* object); |
| )cc"); |
| break; |
| case ArenaDtorNeeds::kNone: |
| break; |
| } |
| }}, |
| {"get_class_data", |
| [&] { |
| if (HasSimpleBaseClass(descriptor_, options_)) return; |
| |
| p->Emit(R"cc( |
| const ::$proto_ns$::MessageLite::ClassData* GetClassData() |
| const final; |
| )cc"); |
| }}, |
| {"get_metadata", |
| [&] { |
| if (!HasDescriptorMethods(descriptor_->file(), options_)) return; |
| |
| p->Emit(R"cc( |
| ::$proto_ns$::Metadata GetMetadata() const final; |
| )cc"); |
| }}, |
| {"decl_split_methods", |
| [&] { |
| if (!ShouldSplit(descriptor_, options_)) return; |
| p->Emit({{"default_name", DefaultInstanceName(descriptor_, options_, |
| /*split=*/true)}}, |
| R"cc( |
| private: |
| inline bool IsSplitMessageDefault() const { |
| return $split$ == reinterpret_cast<const Impl_::Split*>(&$default_name$); |
| } |
| PROTOBUF_NOINLINE void PrepareSplitMessageForWrite(); |
| |
| public: |
| )cc"); |
| }}, |
| {"nested_types", |
| [&] { |
| // Import all nested message classes into this class's scope with |
| // typedefs. |
| for (int i = 0; i < descriptor_->nested_type_count(); i++) { |
| const Descriptor* nested_type = descriptor_->nested_type(i); |
| if (!IsMapEntryMessage(nested_type)) { |
| p->Emit( |
| { |
| Sub{"nested_full_name", ClassName(nested_type, false)} |
| .AnnotatedAs(nested_type), |
| Sub{"nested_name", ResolveKeyword(nested_type->name())} |
| .AnnotatedAs(nested_type), |
| }, |
| R"cc( |
| using $nested_name$ = $nested_full_name$; |
| )cc"); |
| } |
| } |
| }}, |
| {"nested_enums", |
| [&] { |
| // Import all nested enums and their values into this class's |
| // scope with typedefs and constants. |
| for (int i = 0; i < descriptor_->enum_type_count(); i++) { |
| enum_generators_[i]->GenerateSymbolImports(p); |
| } |
| }}, |
| {"decl_field_accessors", |
| [&] { |
| // Generate accessor methods for all fields. |
| GenerateFieldAccessorDeclarations(p); |
| }}, |
| {"decl_extension_ids", |
| [&] { |
| // Declare extension identifiers. |
| for (int i = 0; i < descriptor_->extension_count(); i++) { |
| extension_generators_[i]->GenerateDeclaration(p); |
| } |
| }}, |
| {"proto2_message_sets", |
| [&] { |
| }}, |
| {"decl_set_has", |
| [&] { |
| for (auto field : FieldRange(descriptor_)) { |
| // set_has_***() generated in all oneofs. |
| if (!field->is_repeated() && !field->options().weak() && |
| field->real_containing_oneof()) { |
| p->Emit({{"field_name", FieldName(field)}}, R"cc( |
| void set_has_$field_name$(); |
| )cc"); |
| } |
| } |
| }}, |
| {"decl_oneof_has", |
| [&] { |
| // Generate oneof function declarations |
| for (auto oneof : OneOfRange(descriptor_)) { |
| p->Emit({{"oneof_name", oneof->name()}}, R"cc( |
| inline bool has_$oneof_name$() const; |
| inline void clear_has_$oneof_name$(); |
| )cc"); |
| } |
| }}, |
| {"decl_data", |
| [&] { |
| if (HasGeneratedMethods(descriptor_->file(), options_)) { |
| parse_function_generator_->GenerateDataDecls(p); |
| } |
| }}, |
| {"decl_impl", [&] { GenerateImplDefinition(p); }}, |
| {"split_friend", |
| [&] { |
| if (!ShouldSplit(descriptor_, options_)) return; |
| |
| p->Emit({{"split_default", DefaultInstanceType(descriptor_, options_, |
| /*split=*/true)}}, |
| R"cc( |
| friend struct $split_default$; |
| )cc"); |
| }}}, |
| R"cc( |
| class $dllexport_decl $$classname$ final : public $superclass$ |
| /* @@protoc_insertion_point(class_definition:$full_name$) */ { |
| public: |
| inline $classname$() : $classname$(nullptr) {} |
| $decl_dtor$; |
| //~ Templatize constexpr constructor as a workaround for a bug in |
| //~ gcc 12 (warning in gcc 13). |
| template <typename = void> |
| explicit PROTOBUF_CONSTEXPR $classname$( |
| ::$proto_ns$::internal::ConstantInitialized); |
| |
| inline $classname$(const $classname$& from) : $classname$(nullptr, from) {} |
| $classname$($classname$&& from) noexcept : $classname$() { |
| *this = ::std::move(from); |
| } |
| |
| inline $classname$& operator=(const $classname$& from) { |
| CopyFrom(from); |
| return *this; |
| } |
| inline $classname$& operator=($classname$&& from) noexcept { |
| if (this == &from) return *this; |
| if (GetArena() == from.GetArena() |
| #ifdef PROTOBUF_FORCE_COPY_IN_MOVE |
| && GetArena() != nullptr |
| #endif // !PROTOBUF_FORCE_COPY_IN_MOVE |
| ) { |
| InternalSwap(&from); |
| } else { |
| CopyFrom(from); |
| } |
| return *this; |
| } |
| $decl_annotate$; |
| $decl_verify_func$; |
| |
| inline const $unknown_fields_type$& unknown_fields() const |
| ABSL_ATTRIBUTE_LIFETIME_BOUND { |
| $annotate_unknown_fields$; |
| return $unknown_fields$; |
| } |
| inline $unknown_fields_type$* mutable_unknown_fields() |
| ABSL_ATTRIBUTE_LIFETIME_BOUND { |
| $annotate_mutable_unknown_fields$; |
| return $mutable_unknown_fields$; |
| } |
| |
| $descriptor_accessor$; |
| $get_descriptor$; |
| static const $classname$& default_instance() { |
| return *internal_default_instance(); |
| } |
| $decl_oneof$; |
| //~ TODO make this private, while still granting other |
| //~ protos access. |
| static inline const $classname$* internal_default_instance() { |
| return reinterpret_cast<const $classname$*>( |
| &_$classname$_default_instance_); |
| } |
| static constexpr int kIndexInFileMessages = $index_in_file_messages$; |
| $decl_any_methods$; |
| friend void swap($classname$& a, $classname$& b) { a.Swap(&b); } |
| inline void Swap($classname$* other) { |
| if (other == this) return; |
| #ifdef PROTOBUF_FORCE_COPY_IN_SWAP |
| if (GetArena() != nullptr && GetArena() == other->GetArena()) { |
| #else // PROTOBUF_FORCE_COPY_IN_SWAP |
| if (GetArena() == other->GetArena()) { |
| #endif // !PROTOBUF_FORCE_COPY_IN_SWAP |
| InternalSwap(other); |
| } else { |
| $pbi$::GenericSwap(this, other); |
| } |
| } |
| void UnsafeArenaSwap($classname$* other) { |
| if (other == this) return; |
| $DCHK$(GetArena() == other->GetArena()); |
| InternalSwap(other); |
| } |
| |
| // implements Message ---------------------------------------------- |
| |
| $classname$* New(::$proto_ns$::Arena* arena = nullptr) const final { |
| return $superclass$::DefaultConstruct<$classname$>(arena); |
| } |
| $generated_methods$; |
| $internal_field_number$; |
| $decl_non_simple_base$; |
| //~ Friend AnyMetadata so that it can call this FullMessageName() |
| //~ method. |
| private: |
| friend class ::$proto_ns$::internal::AnyMetadata; |
| static ::absl::string_view FullMessageName() { return "$full_name$"; } |
| |
| //~ TODO Make this private! Currently people are |
| //~ deriving from protos to give access to this constructor, |
| //~ breaking the invariants we rely on. |
| protected: |
| explicit $classname$(::$proto_ns$::Arena* arena); |
| $classname$(::$proto_ns$::Arena* arena, const $classname$& from); |
| $arena_dtor$; |
| $get_class_data$; |
| |
| public: |
| $get_metadata$; |
| $decl_split_methods$; |
| // nested types ---------------------------------------------------- |
| $nested_types$; |
| $nested_enums$; |
| |
| // accessors ------------------------------------------------------- |
| $decl_field_accessors$; |
| $decl_extension_ids$; |
| $proto2_message_sets$; |
| // @@protoc_insertion_point(class_scope:$full_name$) |
| //~ Generate private members. |
| private: |
| //~ TODO: Remove hack to track field access and remove |
| //~ this class. |
| class _Internal; |
| $decl_set_has$; |
| $decl_oneof_has$; |
| $decl_data$; |
| friend class ::$proto_ns$::MessageLite; |
| friend class ::$proto_ns$::Arena; |
| template <typename T> |
| friend class ::$proto_ns$::Arena::InternalHelper; |
| using InternalArenaConstructable_ = void; |
| using DestructorSkippable_ = void; |
| $decl_impl$; |
| $split_friend$; |
| //~ The TableStruct struct needs access to the private parts, in |
| //~ order to construct the offsets of all members. |
| friend struct ::$tablename$; |
| }; |
| )cc"); |
| } // NOLINT(readability/fn_size) |
| |
| void MessageGenerator::GenerateInlineMethods(io::Printer* p) { |
| auto v = p->WithVars(ClassVars(descriptor_, options_)); |
| auto t = p->WithVars(MakeTrackerCalls(descriptor_, options_)); |
| if (IsMapEntryMessage(descriptor_)) return; |
| GenerateFieldAccessorDefinitions(p); |
| |
| // Generate oneof_case() functions. |
| for (auto oneof : OneOfRange(descriptor_)) { |
| p->Emit( |
| { |
| Sub{"oneof_name", absl::StrCat(oneof->name(), "_case")}.AnnotatedAs( |
| oneof), |
| {"OneofName", |
| absl::StrCat(UnderscoresToCamelCase(oneof->name(), true), "Case")}, |
| {"oneof_index", oneof->index()}, |
| }, |
| R"cc( |
| inline $classname$::$OneofName$ $classname$::$oneof_name$() const { |
| return $classname$::$OneofName$($oneof_case$[$oneof_index$]); |
| } |
| )cc"); |
| } |
| } |
| |
| void MessageGenerator::GenerateSchema(io::Printer* p, int offset, |
| int has_offset) { |
| has_offset = !has_bit_indices_.empty() || IsMapEntryMessage(descriptor_) |
| ? offset + has_offset |
| : -1; |
| int inlined_string_indices_offset; |
| if (inlined_string_indices_.empty()) { |
| inlined_string_indices_offset = -1; |
| } else { |
| ABSL_DCHECK_NE(has_offset, -1); |
| ABSL_DCHECK(!IsMapEntryMessage(descriptor_)); |
| inlined_string_indices_offset = has_offset + has_bit_indices_.size(); |
| } |
| |
| auto v = p->WithVars(ClassVars(descriptor_, options_)); |
| p->Emit( |
| { |
| {"offset", offset}, |
| {"has_offset", has_offset}, |
| {"string_offsets", inlined_string_indices_offset}, |
| }, |
| R"cc( |
| {$offset$, $has_offset$, $string_offsets$, sizeof($classtype$)}, |
| )cc"); |
| } |
| |
| void MessageGenerator::GenerateClassMethods(io::Printer* p) { |
| if (!ShouldGenerateClass(descriptor_, options_)) return; |
| |
| auto v = p->WithVars(ClassVars(descriptor_, options_)); |
| auto t = p->WithVars(MakeTrackerCalls(descriptor_, options_)); |
| Formatter format(p); |
| if (IsMapEntryMessage(descriptor_)) { |
| format( |
| "$classname$::$classname$() {}\n" |
| "$classname$::$classname$(::$proto_ns$::Arena* arena)\n" |
| " : SuperType(arena) {}\n"); |
| if (HasDescriptorMethods(descriptor_->file(), options_)) { |
| if (!descriptor_->options().map_entry()) { |
| format( |
| "::$proto_ns$::Metadata $classname$::GetMetadata() const {\n" |
| "$annotate_reflection$" |
| " return ::_pbi::AssignDescriptors(\n" |
| " &$desc_table$_getter, &$desc_table$_once,\n" |
| " $file_level_metadata$[$1$]);\n" |
| "}\n", |
| index_in_file_messages_); |
| } else { |
| format( |
| "::$proto_ns$::Metadata $classname$::GetMetadata() const {\n" |
| " return ::_pbi::AssignDescriptors(\n" |
| " &$desc_table$_getter, &$desc_table$_once,\n" |
| " $file_level_metadata$[$1$]);\n" |
| "}\n", |
| index_in_file_messages_); |
| } |
| } |
| return; |
| } |
| |
| if (IsAnyMessage(descriptor_)) { |
| if (HasDescriptorMethods(descriptor_->file(), options_)) { |
| format( |
| "bool $classname$::GetAnyFieldDescriptors(\n" |
| " const ::$proto_ns$::Message& message,\n" |
| " const ::$proto_ns$::FieldDescriptor** type_url_field,\n" |
| " const ::$proto_ns$::FieldDescriptor** value_field) {\n" |
| " return ::_pbi::GetAnyFieldDescriptors(\n" |
| " message, type_url_field, value_field);\n" |
| "}\n"); |
| } |
| format( |
| "bool $classname$::ParseAnyTypeUrl(\n" |
| " ::absl::string_view type_url,\n" |
| " std::string* full_type_name) {\n" |
| " return ::_pbi::ParseAnyTypeUrl(type_url, full_type_name);\n" |
| "}\n" |
| "\n"); |
| } |
| |
| format( |
| "class $classname$::_Internal {\n" |
| " public:\n"); |
| format.Indent(); |
| if (!has_bit_indices_.empty()) { |
| format( |
| "using HasBits = " |
| "decltype(std::declval<$classname$>().$has_bits$);\n" |
| "static constexpr ::int32_t kHasBitsOffset =\n" |
| " 8 * PROTOBUF_FIELD_OFFSET($classname$, _impl_._has_bits_);\n"); |
| } |
| if (descriptor_->real_oneof_decl_count() > 0) { |
| format( |
| "static constexpr ::int32_t kOneofCaseOffset =\n" |
| " PROTOBUF_FIELD_OFFSET($classtype$, $oneof_case$);\n"); |
| } |
| for (auto field : FieldRange(descriptor_)) { |
| auto t = p->WithVars(MakeTrackerCalls(field, options_)); |
| field_generators_.get(field).GenerateInternalAccessorDeclarations(p); |
| } |
| if (num_required_fields_ > 0) { |
| const std::vector<uint32_t> masks_for_has_bits = RequiredFieldsBitMask(); |
| format( |
| "static bool MissingRequiredFields(const HasBits& has_bits) " |
| "{\n" |
| " return $1$;\n" |
| "}\n", |
| ConditionalToCheckBitmasks(masks_for_has_bits, false, "has_bits")); |
| } |
| |
| format.Outdent(); |
| format("};\n\n"); |
| for (auto field : FieldRange(descriptor_)) { |
| field_generators_.get(field).GenerateInternalAccessorDefinitions(p); |
| } |
| |
| // Generate non-inline field definitions. |
| for (auto field : FieldRange(descriptor_)) { |
| auto v = p->WithVars(FieldVars(field, options_)); |
| auto t = p->WithVars(MakeTrackerCalls(field, options_)); |
| field_generators_.get(field).GenerateNonInlineAccessorDefinitions(p); |
| if (IsCrossFileMaybeMap(field)) { |
| GenerateFieldClear(field, false, p); |
| } |
| } |
| |
| GenerateStructors(p); |
| format("\n"); |
| |
| if (descriptor_->real_oneof_decl_count() > 0) { |
| GenerateOneofClear(p); |
| format("\n"); |
| } |
| |
| GenerateClassData(p); |
| |
| if (HasGeneratedMethods(descriptor_->file(), options_)) { |
| GenerateClear(p); |
| format("\n"); |
| |
| if (!HasSimpleBaseClass(descriptor_, options_)) { |
| parse_function_generator_->GenerateMethodImpls(p); |
| format("\n"); |
| |
| parse_function_generator_->GenerateDataDefinitions(p); |
| } |
| |
| GenerateSerializeWithCachedSizesToArray(p); |
| format("\n"); |
| |
| GenerateByteSize(p); |
| format("\n"); |
| |
| GenerateMergeFrom(p); |
| format("\n"); |
| |
| GenerateClassSpecificMergeImpl(p); |
| format("\n"); |
| |
| GenerateCopyFrom(p); |
| format("\n"); |
| |
| GenerateIsInitialized(p); |
| format("\n"); |
| } |
| |
| if (ShouldSplit(descriptor_, options_)) { |
| format( |
| "void $classname$::PrepareSplitMessageForWrite() {\n" |
| " if (PROTOBUF_PREDICT_TRUE(IsSplitMessageDefault())) {\n" |
| " void* chunk = $pbi$::CreateSplitMessageGeneric(" |
| "GetArena(), &$1$, sizeof(Impl_::Split), this, &$2$);\n" |
| " $split$ = reinterpret_cast<Impl_::Split*>(chunk);\n" |
| " }\n" |
| "}\n", |
| DefaultInstanceName(descriptor_, options_, /*split=*/true), |
| DefaultInstanceName(descriptor_, options_, /*split=*/false)); |
| } |
| |
| GenerateVerify(p); |
| |
| GenerateSwap(p); |
| format("\n"); |
| |
| if (HasDescriptorMethods(descriptor_->file(), options_)) { |
| if (!descriptor_->options().map_entry()) { |
| format( |
| "::$proto_ns$::Metadata $classname$::GetMetadata() const {\n" |
| "$annotate_reflection$" |
| " return ::_pbi::AssignDescriptors(\n" |
| " &$desc_table$_getter, &$desc_table$_once,\n" |
| " $file_level_metadata$[$1$]);\n" |
| "}\n", |
| index_in_file_messages_); |
| } else { |
| format( |
| "::$proto_ns$::Metadata $classname$::GetMetadata() const {\n" |
| " return ::_pbi::AssignDescriptors(\n" |
| " &$desc_table$_getter, &$desc_table$_once,\n" |
| " $file_level_metadata$[$1$]);\n" |
| "}\n", |
| index_in_file_messages_); |
| } |
| } |
| |
| if (HasTracker(descriptor_, options_)) { |
| format( |
| "::$proto_ns$::AccessListener<$classtype$> " |
| "$1$::$tracker$(&FullMessageName);\n", |
| ClassName(descriptor_)); |
| } |
| } |
| |
| std::pair<size_t, size_t> MessageGenerator::GenerateOffsets(io::Printer* p) { |
| auto v = p->WithVars(ClassVars(descriptor_, options_)); |
| auto t = p->WithVars(MakeTrackerCalls(descriptor_, options_)); |
| Formatter format(p); |
| |
| if (!has_bit_indices_.empty() || IsMapEntryMessage(descriptor_)) { |
| format("PROTOBUF_FIELD_OFFSET($classtype$, $has_bits$),\n"); |
| } else { |
| format("~0u, // no _has_bits_\n"); |
| } |
| format("PROTOBUF_FIELD_OFFSET($classtype$, _internal_metadata_),\n"); |
| if (descriptor_->extension_range_count() > 0) { |
| format("PROTOBUF_FIELD_OFFSET($classtype$, $extensions$),\n"); |
| } else { |
| format("~0u, // no _extensions_\n"); |
| } |
| if (descriptor_->real_oneof_decl_count() > 0) { |
| format("PROTOBUF_FIELD_OFFSET($classtype$, $oneof_case$[0]),\n"); |
| } else { |
| format("~0u, // no _oneof_case_\n"); |
| } |
| if (num_weak_fields_ > 0) { |
| format("PROTOBUF_FIELD_OFFSET($classtype$, $weak_field_map$),\n"); |
| } else { |
| format("~0u, // no _weak_field_map_\n"); |
| } |
| if (!inlined_string_indices_.empty()) { |
| format( |
| "PROTOBUF_FIELD_OFFSET($classtype$, " |
| "$inlined_string_donated_array$),\n"); |
| } else { |
| format("~0u, // no _inlined_string_donated_\n"); |
| } |
| if (ShouldSplit(descriptor_, options_)) { |
| format( |
| "PROTOBUF_FIELD_OFFSET($classtype$, $split$),\n" |
| "sizeof($classtype$::Impl_::Split),\n"); |
| } else { |
| format( |
| "~0u, // no _split_\n" |
| "~0u, // no sizeof(Split)\n"); |
| } |
| const int kNumGenericOffsets = 8; // the number of fixed offsets above |
| const size_t offsets = kNumGenericOffsets + descriptor_->field_count() + |
| descriptor_->real_oneof_decl_count(); |
| size_t entries = offsets; |
| for (auto field : FieldRange(descriptor_)) { |
| // TODO: We should not have an entry in the offset table for fields |
| // that do not use them. |
| if (field->options().weak() || field->real_containing_oneof()) { |
| // Mark the field to prevent unintentional access through reflection. |
| // Don't use the top bit because that is for unused fields. |
| format("::_pbi::kInvalidFieldOffsetTag"); |
| } else { |
| format("PROTOBUF_FIELD_OFFSET($classtype$$1$, $2$)", |
| ShouldSplit(field, options_) ? "::Impl_::Split" : "", |
| ShouldSplit(field, options_) |
| ? absl::StrCat(FieldName(field), "_") |
| : FieldMemberName(field, /*cold=*/false)); |
| } |
| |
| // Some information about a field is in the pdproto profile. The profile is |
| // only available at compile time. So we embed such information in the |
| // offset of the field, so that the information is available when |
| // reflectively accessing the field at run time. |
| // |
| // We embed whether the field is cold to the MSB of the offset, and whether |
| // the field is eagerly verified lazy or inlined string to the LSB of the |
| // offset. |
| |
| if (ShouldSplit(field, options_)) { |
| format(" | ::_pbi::kSplitFieldOffsetMask /*split*/"); |
| } |
| if (IsEagerlyVerifiedLazy(field, options_, scc_analyzer_)) { |
| format(" | 0x1u /*eagerly verified lazy*/"); |
| } else if (IsStringInlined(field, options_)) { |
| format(" | 0x1u /*inlined*/"); |
| } |
| format(",\n"); |
| } |
| |
| int count = 0; |
| for (auto oneof : OneOfRange(descriptor_)) { |
| format("PROTOBUF_FIELD_OFFSET($classtype$, _impl_.$1$_),\n", oneof->name()); |
| count++; |
| } |
| ABSL_CHECK_EQ(count, descriptor_->real_oneof_decl_count()); |
| |
| if (IsMapEntryMessage(descriptor_)) { |
| entries += 2; |
| format( |
| "0,\n" |
| "1,\n"); |
| } else if (!has_bit_indices_.empty()) { |
| entries += has_bit_indices_.size(); |
| for (int i = 0; i < has_bit_indices_.size(); i++) { |
| const std::string index = |
| has_bit_indices_[i] >= 0 ? absl::StrCat(has_bit_indices_[i]) : "~0u"; |
| format("$1$,\n", index); |
| } |
| } |
| if (!inlined_string_indices_.empty()) { |
| entries += inlined_string_indices_.size(); |
| for (int inlined_string_index : inlined_string_indices_) { |
| const std::string index = |
| inlined_string_index >= 0 |
| ? absl::StrCat(inlined_string_index, ", // inlined_string_index") |
| : "~0u,"; |
| format("$1$\n", index); |
| } |
| } |
| |
| return std::make_pair(entries, offsets); |
| } |
| |
| void MessageGenerator::GenerateZeroInitFields(io::Printer* p) const { |
| using Iterator = decltype(optimized_order_.begin()); |
| const FieldDescriptor* first = nullptr; |
| auto emit_pending_zero_fields = [&](Iterator end) { |
| if (first != nullptr) { |
| const FieldDescriptor* last = end[-1]; |
| if (first != last) { |
| p->Emit({{"first", FieldName(first)}, |
| {"last", FieldName(last)}, |
| {"Impl", "Impl_"}, |
| {"impl", "_impl_"}}, |
| R"cc( |
| ::memset(reinterpret_cast<char *>(&$impl$) + |
| offsetof($Impl$, $first$_), |
| 0, |
| offsetof($Impl$, $last$_) - |
| offsetof($Impl$, $first$_) + |
| sizeof($Impl$::$last$_)); |
| )cc"); |
| } else { |
| p->Emit({{"field", FieldMemberName(first, false)}}, |
| R"cc( |
| $field$ = {}; |
| )cc"); |
| } |
| first = nullptr; |
| } |
| }; |
| |
| auto it = optimized_order_.begin(); |
| auto end = optimized_order_.end(); |
| for (; it != end && !ShouldSplit(*it, options_); ++it) { |
| auto const& generator = field_generators_.get(*it); |
| if (generator.has_trivial_zero_default()) { |
| if (first == nullptr) first = *it; |
| } else { |
| emit_pending_zero_fields(it); |
| } |
| } |
| emit_pending_zero_fields(it); |
| } |
| |
| namespace { |
| |
| class MemberInitSeparator { |
| public: |
| explicit MemberInitSeparator(io::Printer* printer) : printer_(printer) {} |
| MemberInitSeparator(const MemberInitSeparator&) = delete; |
| |
| ~MemberInitSeparator() { |
| if (separators_) printer_->Outdent(); |
| } |
| |
| void operator()() { |
| if (separators_) { |
| printer_->Emit(",\n"); |
| } else { |
| printer_->Emit(": "); |
| printer_->Indent(); |
| separators_ = true; |
| } |
| } |
| |
| private: |
| bool separators_ = false; |
| io::Printer* const printer_; |
| }; |
| |
| } // namespace |
| |
| void MessageGenerator::GenerateImplMemberInit(io::Printer* p, |
| InitType init_type) { |
| ABSL_DCHECK(!HasSimpleBaseClass(descriptor_, options_)); |
| |
| auto indent = p->WithIndent(); |
| MemberInitSeparator separator(p); |
| |
| auto init_extensions = [&] { |
| if (descriptor_->extension_range_count() > 0 && |
| init_type != InitType::kConstexpr) { |
| separator(); |
| p->Emit("_extensions_{visibility, arena}"); |
| } |
| }; |
| |
| auto init_inlined_string_indices = [&] { |
| if (!inlined_string_indices_.empty()) { |
| bool dtor_on_demand = NeedsArenaDestructor() == ArenaDtorNeeds::kOnDemand; |
| auto values = [&] { |
| for (size_t i = 0; i < InlinedStringDonatedSize(); ++i) { |
| p->Emit(i ? ", ~0u" : dtor_on_demand ? "~0u" : "0xFFFFFFFEu"); |
| } |
| }; |
| separator(); |
| p->Emit({{"values", values}}, "_inlined_string_donated_{$values$}"); |
| } |
| }; |
| |
| auto init_has_bits = [&] { |
| if (!has_bit_indices_.empty()) { |
| if (init_type == InitType::kArenaCopy) { |
| separator(); |
| p->Emit("_has_bits_{from._has_bits_}"); |
| } |
| separator(); |
| p->Emit("_cached_size_{0}"); |
| } |
| }; |
| |
| auto init_fields = [&] { |
| for (auto* field : optimized_order_) { |
| if (ShouldSplit(field, options_)) continue; |
| |
| auto const& generator = field_generators_.get(field); |
| switch (init_type) { |
| case InitType::kConstexpr: |
| separator(); |
| generator.GenerateMemberConstexprConstructor(p); |
| break; |
| case InitType::kArena: |
| if (!generator.has_trivial_zero_default()) { |
| separator(); |
| generator.GenerateMemberConstructor(p); |
| } |
| break; |
| case InitType::kArenaCopy: |
| if (!generator.has_trivial_value()) { |
| separator(); |
| generator.GenerateMemberCopyConstructor(p); |
| } |
| break; |
| } |
| } |
| }; |
| |
| auto init_split = [&] { |
| if (ShouldSplit(descriptor_, options_)) { |
| separator(); |
| p->Emit({{"name", DefaultInstanceName(descriptor_, options_, true)}}, |
| "_split_{const_cast<Split*>(&$name$._instance)}"); |
| } |
| }; |
| |
| auto init_oneofs = [&] { |
| for (auto oneof : OneOfRange(descriptor_)) { |
| separator(); |
| p->Emit({{"name", oneof->name()}}, "$name$_{}"); |
| } |
| }; |
| |
| auto init_cached_size_if_no_hasbits = [&] { |
| if (has_bit_indices_.empty()) { |
| separator(); |
| p->Emit("_cached_size_{0}"); |
| } |
| }; |
| |
| auto init_oneof_cases = [&] { |
| if (int count = descriptor_->real_oneof_decl_count()) { |
| separator(); |
| if (init_type == InitType::kArenaCopy) { |
| auto cases = [&] { |
| for (int i = 0; i < count; i++) { |
| p->Emit({{"index", i}, {"comma", i ? ", " : ""}}, |
| "$comma$from._oneof_case_[$index$]"); |
| } |
| }; |
| p->Emit({{"cases", cases}}, "_oneof_case_{$cases$}"); |
| } else { |
| p->Emit("_oneof_case_{}"); |
| } |
| } |
| }; |
| |
| auto init_weak_field_map = [&] { |
| if (num_weak_fields_ && init_type != InitType::kConstexpr) { |
| separator(); |
| if (init_type == InitType::kArenaCopy) { |
| p->Emit("_weak_field_map_{visibility, arena, from._weak_field_map_}"); |
| } else { |
| p->Emit("_weak_field_map_{visibility, arena}"); |
| } |
| } |
| }; |
| |
| auto init_any_metadata = [&] { |
| if (IsAnyMessage(descriptor_)) { |
| separator(); |
| p->Emit("_any_metadata_{&type_url_, &value_}"); |
| } |
| }; |
| |
| // Initialization order of the various fields inside `_impl_(...)` |
| init_extensions(); |
| init_inlined_string_indices(); |
| init_has_bits(); |
| init_fields(); |
| init_split(); |
| init_oneofs(); |
| init_cached_size_if_no_hasbits(); |
| init_oneof_cases(); |
| init_weak_field_map(); |
| init_any_metadata(); |
| } |
| |
| void MessageGenerator::GenerateSharedConstructorCode(io::Printer* p) { |
| if (HasSimpleBaseClass(descriptor_, options_)) return; |
| |
| // Generate Impl_::Imp_(visibility, Arena*); |
| p->Emit({{"init_impl", [&] { GenerateImplMemberInit(p, InitType::kArena); }}, |
| {"zero_init", [&] { GenerateZeroInitFields(p); }}}, |
| R"cc( |
| inline PROTOBUF_NDEBUG_INLINE $classname$::Impl_::Impl_( |
| $pbi$::InternalVisibility visibility, |
| ::$proto_ns$::Arena* arena) |
| //~ |
| $init_impl$ {} |
| |
| inline void $classname$::SharedCtor(::_pb::Arena* arena) { |
| new (&_impl_) Impl_(internal_visibility(), arena); |
| $zero_init$; |
| } |
| )cc"); |
| } |
| |
| void MessageGenerator::GenerateInitDefaultSplitInstance(io::Printer* p) { |
| if (!ShouldSplit(descriptor_, options_)) return; |
| |
| auto v = p->WithVars(ClassVars(descriptor_, options_)); |
| auto t = p->WithVars(MakeTrackerCalls(descriptor_, options_)); |
| p->Emit("\n"); |
| for (const auto* field : optimized_order_) { |
| if (ShouldSplit(field, options_)) { |
| field_generators_.get(field).GenerateConstexprAggregateInitializer(p); |
| } |
| } |
| } |
| |
| void MessageGenerator::GenerateSharedDestructorCode(io::Printer* p) { |
| if (HasSimpleBaseClass(descriptor_, options_)) return; |
| auto emit_field_dtors = [&](bool split_fields) { |
| // Write the destructors for each field except oneof members. |
| // optimized_order_ does not contain oneof fields. |
| for (const auto* field : optimized_order_) { |
| if (ShouldSplit(field, options_) != split_fields) continue; |
| field_generators_.get(field).GenerateDestructorCode(p); |
| } |
| }; |
| p->Emit( |
| { |
| {"field_dtors", [&] { emit_field_dtors(/* split_fields= */ false); }}, |
| {"split_field_dtors", |
| [&] { |
| if (!ShouldSplit(descriptor_, options_)) return; |
| p->Emit( |
| { |
| {"split_field_dtors_impl", |
| [&] { emit_field_dtors(/* split_fields= */ true); }}, |
| }, |
| R"cc( |
| if (PROTOBUF_PREDICT_FALSE(!IsSplitMessageDefault())) { |
| auto* $cached_split_ptr$ = $split$; |
| $split_field_dtors_impl$; |
| delete $cached_split_ptr$; |
| } |
| )cc"); |
| }}, |
| {"oneof_field_dtors", |
| [&] { |
| for (const auto* oneof : OneOfRange(descriptor_)) { |
| p->Emit({{"name", oneof->name()}}, |
| R"cc( |
| if (has_$name$()) { |
| clear_$name$(); |
| } |
| )cc"); |
| } |
| }}, |
| {"weak_fields_dtor", |
| [&] { |
| if (num_weak_fields_ == 0) return; |
| // Generate code to destruct oneofs. Clearing should do the work. |
| p->Emit(R"cc( |
| $weak_field_map$.ClearAll(); |
| )cc"); |
| }}, |
| {"impl_dtor", [&] { p->Emit("_impl_.~Impl_();\n"); }}, |
| }, |
| R"cc( |
| inline void $classname$::SharedDtor() { |
| $DCHK$(GetArena() == nullptr); |
| $field_dtors$; |
| $split_field_dtors$; |
| $oneof_field_dtors$; |
| $weak_fields_dtor$; |
| $impl_dtor$; |
| } |
| )cc"); |
| } |
| |
| ArenaDtorNeeds MessageGenerator::NeedsArenaDestructor() const { |
| if (HasSimpleBaseClass(descriptor_, options_)) return ArenaDtorNeeds::kNone; |
| ArenaDtorNeeds needs = ArenaDtorNeeds::kNone; |
| for (const auto* field : FieldRange(descriptor_)) { |
| needs = |
| std::max(needs, field_generators_.get(field).NeedsArenaDestructor()); |
| } |
| return needs; |
| } |
| |
| void MessageGenerator::GenerateArenaDestructorCode(io::Printer* p) { |
| ABSL_CHECK(NeedsArenaDestructor() > ArenaDtorNeeds::kNone); |
| auto emit_field_dtors = [&](bool split_fields) { |
| // Write the destructors for each field except oneof members. |
| // optimized_order_ does not contain oneof fields. |
| for (const auto* field : optimized_order_) { |
| if (ShouldSplit(field, options_) != split_fields) continue; |
| field_generators_.get(field).GenerateArenaDestructorCode(p); |
| } |
| }; |
| bool needs_arena_dtor_split = false; |
| for (const auto* field : optimized_order_) { |
| if (!ShouldSplit(field, options_)) continue; |
| if (field_generators_.get(field).NeedsArenaDestructor() > |
| ArenaDtorNeeds::kNone) { |
| needs_arena_dtor_split = true; |
| break; |
| } |
| } |
| |
| // This code is placed inside a static method, rather than an ordinary one, |
| // since that simplifies Arena's destructor list (ordinary function pointers |
| // rather than member function pointers). _this is the object being |
| // destructed. |
| p->Emit( |
| { |
| {"field_dtors", [&] { emit_field_dtors(/* split_fields= */ false); }}, |
| {"split_field_dtors", |
| [&] { |
| if (!ShouldSplit(descriptor_, options_)) return; |
| if (!needs_arena_dtor_split) { |
| return; |
| } |
| p->Emit( |
| { |
| {"split_field_dtors_impl", |
| [&] { emit_field_dtors(/* split_fields= */ true); }}, |
| }, |
| R"cc( |
| if (PROTOBUF_PREDICT_FALSE( |
| !_this->IsSplitMessageDefault())) { |
| $split_field_dtors_impl$; |
| } |
| )cc"); |
| }}, |
| {"oneof_field_dtors", |
| [&] { |
| for (const auto* oneof : OneOfRange(descriptor_)) { |
| for (const auto* field : FieldRange(oneof)) { |
| field_generators_.get(field).GenerateArenaDestructorCode(p); |
| } |
| } |
| }}, |
| }, |
| R"cc( |
| void $classname$::ArenaDtor(void* object) { |
| $classname$* _this = reinterpret_cast<$classname$*>(object); |
| $field_dtors$; |
| $split_field_dtors$; |
| $oneof_field_dtors$; |
| } |
| )cc"); |
| } |
| |
| void MessageGenerator::GenerateConstexprConstructor(io::Printer* p) { |
| if (!ShouldGenerateClass(descriptor_, options_)) return; |
| |
| auto v = p->WithVars(ClassVars(descriptor_, options_)); |
| auto t = p->WithVars(MakeTrackerCalls(descriptor_, options_)); |
| auto c = p->WithVars({{"constexpr", "PROTOBUF_CONSTEXPR"}}); |
| Formatter format(p); |
| |
| if (IsMapEntryMessage(descriptor_) || !HasImplData(descriptor_, options_)) { |
| p->Emit(R"cc( |
| //~ Templatize constexpr constructor as a workaround for a bug in gcc 12 |
| //~ (warning in gcc 13). |
| template <typename> |
| $constexpr$ $classname$::$classname$(::_pbi::ConstantInitialized) {} |
| )cc"); |
| return; |
| } |
| |
| // Generate Impl_::Imp_(::_pbi::ConstantInitialized); |
| // We use separate p->Emit() calls for LF and #ifdefs as they result in |
| // awkward layout and more awkward indenting of the function statement. |
| p->Emit("\n"); |
| p->Emit({{"init", [&] { GenerateImplMemberInit(p, InitType::kConstexpr); }}}, |
| R"cc( |
| inline constexpr $classname$::Impl_::Impl_( |
| ::_pbi::ConstantInitialized) noexcept |
| //~ |
| $init$ {} |
| )cc"); |
| p->Emit("\n"); |
| |
| p->Emit( |
| R"cc( |
| template <typename> |
| $constexpr$ $classname$::$classname$(::_pbi::ConstantInitialized) |
| : _impl_(::_pbi::ConstantInitialized()) {} |
| )cc"); |
| } |
| |
| bool MessageGenerator::ImplHasCopyCtor() const { |
| if (ShouldSplit(descriptor_, options_)) return false; |
| if (HasSimpleBaseClass(descriptor_, options_)) return false; |
| if (descriptor_->extension_range_count() > 0) return false; |
| if (descriptor_->real_oneof_decl_count() > 0) return false; |
| if (num_weak_fields_ > 0) return false; |
| |
| // If the message contains only scalar fields (ints and enums), |
| // then we can copy the entire impl_ section with a single statement. |
| for (const auto* field : optimized_order_) { |
| if (field->is_repeated()) return false; |
| if (field->is_extension()) return false; |
| switch (field->cpp_type()) { |
| case FieldDescriptor::CPPTYPE_ENUM: |
| case FieldDescriptor::CPPTYPE_INT32: |
| case FieldDescriptor::CPPTYPE_INT64: |
| case FieldDescriptor::CPPTYPE_UINT32: |
| case FieldDescriptor::CPPTYPE_UINT64: |
| case FieldDescriptor::CPPTYPE_FLOAT: |
| case FieldDescriptor::CPPTYPE_DOUBLE: |
| case FieldDescriptor::CPPTYPE_BOOL: |
| break; |
| default: |
| return false; |
| } |
| } |
| return true; |
| } |
| |
| void MessageGenerator::GenerateCopyInitFields(io::Printer* p) const { |
| auto begin = optimized_order_.begin(); |
| auto end = optimized_order_.end(); |
| const FieldDescriptor* first = nullptr; |
| |
| auto emit_pending_copy_fields = [&](decltype(end) itend, bool split) { |
| if (first != nullptr) { |
| const FieldDescriptor* last = itend[-1]; |
| if (first != last) { |
| p->Emit({{"first", FieldName(first)}, |
| {"last", FieldName(last)}, |
| {"Impl", split ? "Impl_::Split" : "Impl_"}, |
| {"pdst", split ? "_impl_._split_" : "&_impl_"}, |
| {"psrc", split ? "from._impl_._split_" : "&from._impl_"}}, |
| R"cc( |
| ::memcpy(reinterpret_cast<char *>($pdst$) + |
| offsetof($Impl$, $first$_), |
| reinterpret_cast<const char *>($psrc$) + |
| offsetof($Impl$, $first$_), |
| offsetof($Impl$, $last$_) - |
| offsetof($Impl$, $first$_) + |
| sizeof($Impl$::$last$_)); |
| )cc"); |
| } else { |
| p->Emit({{"field", FieldMemberName(first, split)}}, |
| R"cc( |
| $field$ = from.$field$; |
| )cc"); |
| } |
| first = nullptr; |
| } |
| }; |
| |
| int has_bit_word_index = -1; |
| auto load_has_bits = [&](const FieldDescriptor* field) { |
| if (has_bit_indices_.empty()) return; |
| int has_bit_index = has_bit_indices_[field->index()]; |
| if (has_bit_word_index != has_bit_index / 32) { |
| p->Emit({{"declare", has_bit_word_index < 0 ? "::uint32_t " : ""}, |
| {"index", has_bit_index / 32}}, |
| "$declare$cached_has_bits = _impl_._has_bits_[$index$];\n"); |
| has_bit_word_index = has_bit_index / 32; |
| } |
| }; |
| |
| auto has_message = [&](const FieldDescriptor* field) { |
| if (has_bit_indices_.empty()) { |
| p->Emit("from.$field$ != nullptr"); |
| } else { |
| int index = has_bit_indices_[field->index()]; |
| std::string mask = absl::StrFormat("0x%08xu", 1u << (index % 32)); |
| p->Emit({{"mask", mask}}, "cached_has_bits & $mask$"); |
| } |
| }; |
| |
| auto emit_copy_message = [&](const FieldDescriptor* field) { |
| load_has_bits(field); |
| p->Emit({{"has_msg", [&] { has_message(field); }}, |
| {"submsg", FieldMessageTypeName(field, options_)}}, |
| R"cc( |
| $field$ = ($has_msg$) ? $superclass$::CopyConstruct<$submsg$>( |
| arena, *from.$field$) |
| : nullptr; |
| )cc"); |
| }; |
| |
| auto generate_copy_fields = [&] { |
| for (auto it = begin; it != end; ++it) { |
| const auto& gen = field_generators_.get(*it); |
| auto v = p->WithVars(FieldVars(*it, options_)); |
| |
| // Non trivial field values are copy constructed |
| if (!gen.has_trivial_value() || gen.should_split()) { |
| emit_pending_copy_fields(it, false); |
| continue; |
| } |
| |
| if (gen.is_message()) { |
| emit_pending_copy_fields(it, false); |
| emit_copy_message(*it); |
| } else if (first == nullptr) { |
| first = *it; |
| } |
| } |
| emit_pending_copy_fields(end, false); |
| }; |
| |
| auto generate_copy_split_fields = [&] { |
| for (auto it = begin; it != end; ++it) { |
| const auto& gen = field_generators_.get(*it); |
| auto v = p->WithVars(FieldVars(*it, options_)); |
| |
| if (!gen.should_split()) { |
| emit_pending_copy_fields(it, true); |
| continue; |
| } |
| |
| if (gen.is_trivial()) { |
| if (first == nullptr) first = *it; |
| } else { |
| emit_pending_copy_fields(it, true); |
| gen.GenerateCopyConstructorCode(p); |
| } |
| } |
| emit_pending_copy_fields(end, true); |
| }; |
| |
| auto generate_copy_oneof_fields = [&]() { |
| for (const auto* oneof : OneOfRange(descriptor_)) { |
| p->Emit( |
| {{"name", oneof->name()}, |
| {"NAME", absl::AsciiStrToUpper(oneof->name())}, |
| {"cases", |
| [&] { |
| for (const auto* field : FieldRange(oneof)) { |
| p->Emit( |
| {{"Name", UnderscoresToCamelCase(field->name(), true)}, |
| {"field", FieldMemberName(field, /*split=*/false)}, |
| {"body", |
| [&] { |
| field_generators_.get(field).GenerateOneofCopyConstruct( |
| p); |
| }}}, |
| R"cc( |
| case k$Name$: |
| $body$; |
| break; |
| )cc"); |
| } |
| }}}, |
| R"cc( |
| switch ($name$_case()) { |
| case $NAME$_NOT_SET: |
| break; |
| $cases$; |
| } |
| )cc"); |
| } |
| }; |
| |
| if (descriptor_->extension_range_count() > 0) { |
| p->Emit(R"cc( |
| _impl_._extensions_.MergeFrom(this, from._impl_._extensions_); |
| )cc"); |
| } |
| |
| p->Emit({{"copy_fields", generate_copy_fields}, |
| {"copy_oneof_fields", generate_copy_oneof_fields}}, |
| R"cc( |
| $copy_fields$; |
| $copy_oneof_fields$; |
| )cc"); |
| |
| if (ShouldSplit(descriptor_, options_)) { |
| p->Emit({{"copy_split_fields", generate_copy_split_fields}}, |
| R"cc( |
| if (PROTOBUF_PREDICT_FALSE(!from.IsSplitMessageDefault())) { |
| PrepareSplitMessageForWrite(); |
| $copy_split_fields$; |
| } |
| )cc"); |
| } |
| } |
| |
| void MessageGenerator::GenerateArenaEnabledCopyConstructor(io::Printer* p) { |
| if (!HasSimpleBaseClass(descriptor_, options_)) { |
| // Generate Impl_::Imp_(visibility, Arena*, const& from); |
| p->Emit( |
| {{"init", [&] { GenerateImplMemberInit(p, InitType::kArenaCopy); }}}, |
| R"cc( |
| inline PROTOBUF_NDEBUG_INLINE $classname$::Impl_::Impl_( |
| $pbi$::InternalVisibility visibility, ::$proto_ns$::Arena* arena, |
| const Impl_& from) |
| //~ |
| $init$ {} |
| )cc"); |
| p->Emit("\n"); |
| } |
| |
| auto copy_construct_impl = [&] { |
| if (!HasSimpleBaseClass(descriptor_, options_)) { |
| p->Emit(R"cc( |
| new (&_impl_) Impl_(internal_visibility(), arena, from._impl_); |
| )cc"); |
| } |
| }; |
| |
| auto force_allocation = [&] { |
| if (ShouldForceAllocationOnConstruction(descriptor_, options_)) { |
| p->Emit(R"cc( |
| //~ force alignment |
| #ifdef PROTOBUF_FORCE_ALLOCATION_ON_CONSTRUCTION |
| $mutable_unknown_fields$; |
| #endif // PROTOBUF_FORCE_ALLOCATION_ON_CONSTRUCTION |
| )cc"); |
| } |
| }; |
| |
| auto maybe_register_arena_dtor = [&] { |
| if (NeedsArenaDestructor() == ArenaDtorNeeds::kRequired) { |
| p->Emit(R"cc( |
| if (arena != nullptr) { |
| arena->OwnCustomDestructor(this, &$classname$::ArenaDtor); |
| } |
| )cc"); |
| } |
| }; |
| |
| p->Emit({{"copy_construct_impl", copy_construct_impl}, |
| {"copy_init_fields", [&] { GenerateCopyInitFields(p); }}, |
| {"force_allocation", force_allocation}, |
| {"maybe_register_arena_dtor", maybe_register_arena_dtor}}, |
| R"cc( |
| $classname$::$classname$( |
| //~ force alignment |
| ::$proto_ns$::Arena* arena, |
| //~ force alignment |
| const $classname$& from) |
| : $superclass$(arena) { |
| $classname$* const _this = this; |
| (void)_this; |
| _internal_metadata_.MergeFrom<$unknown_fields_type$>( |
| from._internal_metadata_); |
| $copy_construct_impl$; |
| $copy_init_fields$; |
| $force_allocation$; |
| $maybe_register_arena_dtor$; |
| |
| // @@protoc_insertion_point(copy_constructor:$full_name$) |
| } |
| )cc"); |
| } |
| |
| void MessageGenerator::GenerateStructors(io::Printer* p) { |
| p->Emit( |
| { |
| {"superclass", SuperClassName(descriptor_, options_)}, |
| {"ctor_body", |
| [&] { |
| if (HasSimpleBaseClass(descriptor_, options_)) return; |
| p->Emit(R"cc(SharedCtor(arena);)cc"); |
| if (NeedsArenaDestructor() == ArenaDtorNeeds::kRequired) { |
| p->Emit(R"cc( |
| if (arena != nullptr) { |
| arena->OwnCustomDestructor(this, &$classname$::ArenaDtor); |
| } |
| )cc"); |
| } |
| }}, |
| }, |
| R"cc( |
| $classname$::$classname$(::$proto_ns$::Arena* arena) |
| : $superclass$(arena) { |
| $ctor_body$; |
| // @@protoc_insertion_point(arena_constructor:$full_name$) |
| } |
| )cc"); |
| |
| // Generate the copy constructor. |
| if (UsingImplicitWeakFields(descriptor_->file(), options_)) { |
| // If we are in lite mode and using implicit weak fields, we generate a |
| // one-liner copy constructor that delegates to MergeFrom. This saves some |
| // code size and also cuts down on the complexity of implicit weak fields. |
| // We might eventually want to do this for all lite protos. |
| p->Emit(R"cc( |
| $classname$::$classname$( |
| //~ Force alignment |
| ::$proto_ns$::Arena* arena, const $classname$& from) |
| : $classname$(arena) { |
| MergeFrom(from); |
| } |
| )cc"); |
| } else if (ImplHasCopyCtor()) { |
| p->Emit(R"cc( |
| $classname$::$classname$( |
| //~ Force alignment |
| ::$proto_ns$::Arena* arena, const $classname$& from) |
| : $classname$(arena) { |
| MergeFrom(from); |
| } |
| )cc"); |
| } else { |
| GenerateArenaEnabledCopyConstructor(p); |
| } |
| |
| // Generate the shared constructor code. |
| GenerateSharedConstructorCode(p); |
| |
| // Generate the destructor. |
| if (HasSimpleBaseClass(descriptor_, options_)) { |
| // For messages using simple base classes, having no destructor |
| // allows our vtable to share the same destructor as every other |
| // message with a simple base class. This works only as long as |
| // we have no fields needing destruction, of course. (No strings |
| // or extensions) |
| } else { |
| p->Emit( |
| R"cc( |
| $classname$::~$classname$() { |
| // @@protoc_insertion_point(destructor:$full_name$) |
| _internal_metadata_.Delete<$unknown_fields_type$>(); |
| SharedDtor(); |
| } |
| )cc"); |
| } |
| |
| // Generate the shared destructor code. |
| GenerateSharedDestructorCode(p); |
| |
| // Generate the arena-specific destructor code. |
| if (NeedsArenaDestructor() > ArenaDtorNeeds::kNone) { |
| GenerateArenaDestructorCode(p); |
| } |
| } |
| |
| void MessageGenerator::GenerateSourceInProto2Namespace(io::Printer* p) { |
| auto v = p->WithVars(ClassVars(descriptor_, options_)); |
| auto t = p->WithVars(MakeTrackerCalls(descriptor_, options_)); |
| Formatter format(p); |
| if (ShouldGenerateExternSpecializations(options_) && |
| ShouldGenerateClass(descriptor_, options_)) { |
| p->Emit(R"cc( |
| template void* Arena::DefaultConstruct<$classtype$>(Arena*); |
| )cc"); |
| if (!IsMapEntryMessage(descriptor_)) { |
| p->Emit(R"cc( |
| template void* Arena::CopyConstruct<$classtype$>(Arena*, const void*); |
| )cc"); |
| } |
| } |
| } |
| |
| void MessageGenerator::GenerateClear(io::Printer* p) { |
| if (HasSimpleBaseClass(descriptor_, options_)) return; |
| Formatter format(p); |
| |
| // The maximum number of bytes we will memset to zero without checking their |
| // hasbit to see if a zero-init is necessary. |
| const int kMaxUnconditionalPrimitiveBytesClear = 4; |
| |
| format( |
| "PROTOBUF_NOINLINE void $classname$::Clear() {\n" |
| "// @@protoc_insertion_point(message_clear_start:$full_name$)\n"); |
| format.Indent(); |
| |
| format("PROTOBUF_TSAN_WRITE(&_impl_._tsan_detect_race);\n"); |
| |
| format( |
| // TODO: It would be better to avoid emitting this if it is not used, |
| // rather than emitting a workaround for the resulting warning. |
| "$uint32$ cached_has_bits = 0;\n" |
| "// Prevent compiler warnings about cached_has_bits being unused\n" |
| "(void) cached_has_bits;\n\n"); |
| |
| if (descriptor_->extension_range_count() > 0) { |
| format("$extensions$.Clear();\n"); |
| } |
| |
| // Collect fields into chunks. Each chunk may have an if() condition that |
| // checks all hasbits in the chunk and skips it if none are set. |
| int zero_init_bytes = 0; |
| for (const auto& field : optimized_order_) { |
| if (CanClearByZeroing(field)) { |
| zero_init_bytes += EstimateAlignmentSize(field); |
| } |
| } |
| bool merge_zero_init = zero_init_bytes > kMaxUnconditionalPrimitiveBytesClear; |
| int chunk_count = 0; |
| |
| std::vector<FieldChunk> chunks = CollectFields( |
| optimized_order_, options_, |
| [&](const FieldDescriptor* a, const FieldDescriptor* b) -> bool { |
| chunk_count++; |
| // This predicate guarantees that there is only a single zero-init |
| // (memset) per chunk, and if present it will be at the beginning. |
| bool same = |
| HasByteIndex(a) == HasByteIndex(b) && |
| a->is_repeated() == b->is_repeated() && |
| IsLikelyPresent(a, options_) == IsLikelyPresent(b, options_) && |
| ShouldSplit(a, options_) == ShouldSplit(b, options_) && |
| (CanClearByZeroing(a) == CanClearByZeroing(b) || |
| (CanClearByZeroing(a) && (chunk_count == 1 || merge_zero_init))); |
| if (!same) chunk_count = 0; |
| return same; |
| }); |
| |
| auto it = chunks.begin(); |
| auto end = chunks.end(); |
| int cached_has_word_index = -1; |
| while (it != end) { |
| auto next = FindNextUnequalChunk(it, end, MayGroupChunksForHaswordsCheck); |
| bool has_haswords_check = MaybeEmitHaswordsCheck( |
| it, next, options_, has_bit_indices_, cached_has_word_index, "", p); |
| bool has_default_split_check = !it->fields.empty() && it->should_split; |
| if (has_default_split_check) { |
| // Some fields are cleared without checking has_bit. So we add the |
| // condition here to avoid writing to the default split instance. |
| format("if (!IsSplitMessageDefault()) {\n"); |
| format.Indent(); |
| } |
| while (it != next) { |
| const std::vector<const FieldDescriptor*>& fields = it->fields; |
| bool chunk_is_split = it->should_split; |
| ABSL_CHECK_EQ(has_default_split_check, chunk_is_split); |
| |
| const FieldDescriptor* memset_start = nullptr; |
| const FieldDescriptor* memset_end = nullptr; |
| bool saw_non_zero_init = false; |
| |
| for (const auto& field : fields) { |
| if (CanClearByZeroing(field)) { |
| ABSL_CHECK(!saw_non_zero_init); |
| if (!memset_start) memset_start = field; |
| memset_end = field; |
| } else { |
| saw_non_zero_init = true; |
| } |
| } |
| |
| // Whether we wrap this chunk in: |
| // if (cached_has_bits & <chunk hasbits) { /* chunk. */ } |
| // We can omit the if() for chunk size 1, or if our fields do not have |
| // hasbits. I don't understand the rationale for the last part of the |
| // condition, but it matches the old logic. |
| const bool check_has_byte = |
| HasBitIndex(fields.front()) != kNoHasbit && fields.size() > 1 && |
| !IsLikelyPresent(fields.back(), options_) && |
| (memset_end != fields.back() || merge_zero_init); |
| |
| if (check_has_byte) { |
| // Emit an if() that will let us skip the whole chunk if none are set. |
| uint32_t chunk_mask = GenChunkMask(fields, has_bit_indices_); |
| std::string chunk_mask_str = |
| absl::StrCat(absl::Hex(chunk_mask, absl::kZeroPad8)); |
| |
| // Check (up to) 8 has_bits at a time if we have more than one field in |
| // this chunk. Due to field layout ordering, we may check |
| // _has_bits_[last_chunk * 8 / 32] multiple times. |
| ABSL_DCHECK_LE(2, popcnt(chunk_mask)); |
| ABSL_DCHECK_GE(8, popcnt(chunk_mask)); |
| |
| if (cached_has_word_index != HasWordIndex(fields.front())) { |
| cached_has_word_index = HasWordIndex(fields.front()); |
| format("cached_has_bits = $has_bits$[$1$];\n", cached_has_word_index); |
| } |
| format("if (cached_has_bits & 0x$1$u) {\n", chunk_mask_str); |
| format.Indent(); |
| } |
| |
| if (memset_start) { |
| if (memset_start == memset_end) { |
| // For clarity, do not memset a single field. |
| field_generators_.get(memset_start).GenerateMessageClearingCode(p); |
| } else { |
| ABSL_CHECK_EQ(chunk_is_split, ShouldSplit(memset_start, options_)); |
| ABSL_CHECK_EQ(chunk_is_split, ShouldSplit(memset_end, options_)); |
| format( |
| "::memset(&$1$, 0, static_cast<::size_t>(\n" |
| " reinterpret_cast<char*>(&$2$) -\n" |
| " reinterpret_cast<char*>(&$1$)) + sizeof($2$));\n", |
| FieldMemberName(memset_start, chunk_is_split), |
| FieldMemberName(memset_end, chunk_is_split)); |
| } |
| } |
| |
| // Clear all non-zero-initializable fields in the chunk. |
| for (const auto& field : fields) { |
| if (CanClearByZeroing(field)) continue; |
| // It's faster to just overwrite primitive types, but we should only |
| // clear strings and messages if they were set. |
| // |
| // TODO: Let the CppFieldGenerator decide this somehow. |
| bool have_enclosing_if = |
| HasBitIndex(field) != kNoHasbit && |
| (field->cpp_type() == FieldDescriptor::CPPTYPE_MESSAGE || |
| field->cpp_type() == FieldDescriptor::CPPTYPE_STRING); |
| |
| if (have_enclosing_if) { |
| PrintPresenceCheck(field, has_bit_indices_, p, |
| &cached_has_word_index); |
| format.Indent(); |
| } |
| |
| field_generators_.get(field).GenerateMessageClearingCode(p); |
| |
| if (have_enclosing_if) { |
| format.Outdent(); |
| format("}\n"); |
| } |
| } |
| |
| if (check_has_byte) { |
| format.Outdent(); |
| format("}\n"); |
| } |
| |
| // To next chunk. |
| ++it; |
| } |
| |
| if (has_default_split_check) { |
| format.Outdent(); |
| format("}\n"); |
| } |
| if (has_haswords_check) { |
| p->Outdent(); |
| p->Emit(R"cc( |
| } |
| )cc"); |
| |
| // Reset here as it may have been updated in just closed if statement. |
| cached_has_word_index = -1; |
| } |
| } |
| // Step 4: Unions. |
| for (auto oneof : OneOfRange(descriptor_)) { |
| format("clear_$1$();\n", oneof->name()); |
| } |
| |
| if (num_weak_fields_) { |
| format("$weak_field_map$.ClearAll();\n"); |
| } |
| |
| // We don't clear donated status. |
| |
| if (!has_bit_indices_.empty()) { |
| // Step 5: Everything else. |
| format("$has_bits$.Clear();\n"); |
| } |
| |
| format("_internal_metadata_.Clear<$unknown_fields_type$>();\n"); |
| |
| format.Outdent(); |
| format("}\n"); |
| } |
| |
| void MessageGenerator::GenerateOneofClear(io::Printer* p) { |
| // Generated function clears the active field and union case (e.g. foo_case_). |
| int i = 0; |
| for (auto oneof : OneOfRange(descriptor_)) { |
| Formatter format(p); |
| auto v = p->WithVars({{"oneofname", oneof->name()}}); |
| |
| format( |
| "void $classname$::clear_$oneofname$() {\n" |
| "// @@protoc_insertion_point(one_of_clear_start:$full_name$)\n"); |
| format.Indent(); |
| format("PROTOBUF_TSAN_WRITE(&_impl_._tsan_detect_race);\n"); |
| format("switch ($oneofname$_case()) {\n"); |
| format.Indent(); |
| for (auto field : FieldRange(oneof)) { |
| format("case k$1$: {\n", UnderscoresToCamelCase(field->name(), true)); |
| format.Indent(); |
| // We clear only allocated objects in oneofs |
| if (!IsStringOrMessage(field)) { |
| format("// No need to clear\n"); |
| } else { |
| field_generators_.get(field).GenerateClearingCode(p); |
| } |
| format("break;\n"); |
| format.Outdent(); |
| format("}\n"); |
| } |
| format( |
| "case $1$_NOT_SET: {\n" |
| " break;\n" |
| "}\n", |
| absl::AsciiStrToUpper(oneof->name())); |
| format.Outdent(); |
| format( |
| "}\n" |
| "$oneof_case$[$1$] = $2$_NOT_SET;\n", |
| i, absl::AsciiStrToUpper(oneof->name())); |
| format.Outdent(); |
| format( |
| "}\n" |
| "\n"); |
| i++; |
| } |
| } |
| |
| void MessageGenerator::GenerateSwap(io::Printer* p) { |
| if (HasSimpleBaseClass(descriptor_, options_)) return; |
| Formatter format(p); |
| |
| format( |
| "void $classname$::InternalSwap($classname$* PROTOBUF_RESTRICT other) " |
| "{\n"); |
| format.Indent(); |
| format("using std::swap;\n"); |
| |
| if (HasGeneratedMethods(descriptor_->file(), options_)) { |
| if (descriptor_->extension_range_count() > 0) { |
| format( |
| "$extensions$.InternalSwap(&other->$extensions$);" |
| "\n"); |
| } |
| |
| if (HasNonSplitOptionalString(descriptor_, options_)) { |
| p->Emit(R"cc( |
| auto* arena = GetArena(); |
| ABSL_DCHECK_EQ(arena, other->GetArena()); |
| )cc"); |
| } |
| format("_internal_metadata_.InternalSwap(&other->_internal_metadata_);\n"); |
| |
| if (!has_bit_indices_.empty()) { |
| for (int i = 0; i < HasBitsSize(); ++i) { |
| format("swap($has_bits$[$1$], other->$has_bits$[$1$]);\n", i); |
| } |
| } |
| |
| // If possible, we swap several fields at once, including padding. |
| const RunMap runs = |
| FindRuns(optimized_order_, [this](const FieldDescriptor* field) { |
| return !ShouldSplit(field, options_) && |
| HasTrivialSwap(field, options_, scc_analyzer_); |
| }); |
| |
| for (size_t i = 0; i < optimized_order_.size(); ++i) { |
| const FieldDescriptor* field = optimized_order_[i]; |
| if (ShouldSplit(field, options_)) { |
| continue; |
| } |
| const auto it = runs.find(field); |
| |
| // We only apply the memswap technique to runs of more than one field, as |
| // `swap(field_, other.field_)` is better than |
| // `memswap<...>(&field_, &other.field_)` for generated code readability. |
| if (it != runs.end() && it->second > 1) { |
| // Use a memswap, then skip run_length fields. |
| const size_t run_length = it->second; |
| const std::string first_field_name = |
| FieldMemberName(field, /*cold=*/false); |
| const std::string last_field_name = FieldMemberName( |
| optimized_order_[i + run_length - 1], /*cold=*/false); |
| |
| auto v = p->WithVars({ |
| {"first", first_field_name}, |
| {"last", last_field_name}, |
| }); |
| |
| format( |
| "$pbi$::memswap<\n" |
| " PROTOBUF_FIELD_OFFSET($classname$, $last$)\n" |
| " + sizeof($classname$::$last$)\n" |
| " - PROTOBUF_FIELD_OFFSET($classname$, $first$)>(\n" |
| " reinterpret_cast<char*>(&$first$),\n" |
| " reinterpret_cast<char*>(&other->$first$));\n"); |
| |
| i += run_length - 1; |
| // ++i at the top of the loop. |
| } else { |
| field_generators_.get(field).GenerateSwappingCode(p); |
| } |
| } |
| if (ShouldSplit(descriptor_, options_)) { |
| format("swap($split$, other->$split$);\n"); |
| } |
| |
| for (auto oneof : OneOfRange(descriptor_)) { |
| format("swap(_impl_.$1$_, other->_impl_.$1$_);\n", oneof->name()); |
| } |
| |
| for (int i = 0; i < descriptor_->real_oneof_decl_count(); i++) { |
| format("swap($oneof_case$[$1$], other->$oneof_case$[$1$]);\n", i); |
| } |
| |
| if (num_weak_fields_) { |
| format( |
| "$weak_field_map$.UnsafeArenaSwap(&other->$weak_field_map$)" |
| ";\n"); |
| } |
| |
| if (!inlined_string_indices_.empty()) { |
| for (size_t i = 0; i < InlinedStringDonatedSize(); ++i) { |
| format( |
| "swap($inlined_string_donated_array$[$1$], " |
| "other->$inlined_string_donated_array$[$1$]);\n", |
| i); |
| } |
| } |
| } else { |
| format("GetReflection()->Swap(this, other);"); |
| } |
| |
| format.Outdent(); |
| format("}\n"); |
| } |
| |
| void MessageGenerator::GenerateClassData(io::Printer* p) { |
| Formatter format(p); |
| if (HasSimpleBaseClass(descriptor_, options_)) return; |
| |
| const auto class_data_members = [&] { |
| p->Emit( |
| { |
| {"merge_impl", |
| [&] { |
| // TODO: This check is not needed once we migrate |
| // CheckTypeAndMergeFrom to ClassData fully. |
| if (HasDescriptorMethods(descriptor_->file(), options_)) { |
| p->Emit(R"cc( |
| $classname$::MergeImpl, |
| )cc"); |
| } else { |
| p->Emit(R"cc( |
| nullptr, // MergeImpl |
| )cc"); |
| } |
| }}, |
| {"on_demand_register_arena_dtor", |
| [&] { |
| if (NeedsArenaDestructor() == ArenaDtorNeeds::kOnDemand) { |
| p->Emit(R"cc( |
| $classname$::OnDemandRegisterArenaDtor, |
| )cc"); |
| } else { |
| p->Emit(R"cc( |
| nullptr, // OnDemandRegisterArenaDtor |
| )cc"); |
| } |
| }}, |
| {"descriptor_methods", |
| [&] { |
| if (HasDescriptorMethods(descriptor_->file(), options_)) { |
| p->Emit(R"cc( |
| &::$proto_ns$::Message::kDescriptorMethods, |
| )cc"); |
| } else { |
| p->Emit(R"cc( |
| nullptr, // DescriptorMethods |
| )cc"); |
| } |
| }}, |
| }, |
| R"cc( |
| $merge_impl$, $on_demand_register_arena_dtor$, $descriptor_methods$, |
| PROTOBUF_FIELD_OFFSET($classname$, $cached_size$), |
| )cc"); |
| }; |
| |
| if (HasDescriptorMethods(descriptor_->file(), options_)) { |
| p->Emit( |
| {{"class_data_members", class_data_members}}, |
| R"cc( |
| const ::$proto_ns$::MessageLite::ClassData* |
| $classname$::GetClassData() const { |
| PROTOBUF_CONSTINIT static const ::$proto_ns$::MessageLite::ClassData |
| _data_ = { |
| $class_data_members$, |
| }; |
| return &_data_; |
| } |
| )cc"); |
| } else { |
| p->Emit( |
| { |
| {"class_data_members", class_data_members}, |
| {"type_size", descriptor_->full_name().size() + 1}, |
| }, |
| R"cc( |
| const ::$proto_ns$::MessageLite::ClassData* |
| $classname$::GetClassData() const { |
| struct ClassData_ { |
| ::$proto_ns$::MessageLite::ClassData header; |
| char type_name[$type_size$]; |
| }; |
| PROTOBUF_CONSTINIT static const ClassData_ _data_ = { |
| { |
| $class_data_members$, |
| }, |
| "$full_name$", |
| }; |
| |
| return &_data_.header; |
| } |
| )cc"); |
| } |
| } |
| |
| void MessageGenerator::GenerateMergeFrom(io::Printer* p) { |
| Formatter format(p); |
| |
| if (!HasDescriptorMethods(descriptor_->file(), options_)) { |
| // Generate CheckTypeAndMergeFrom(). |
| format( |
| "void $classname$::CheckTypeAndMergeFrom(\n" |
| " const ::$proto_ns$::MessageLite& from) {\n" |
| " MergeFrom(*::_pbi::DownCast<const $classname$*>(\n" |
| " &from));\n" |
| "}\n"); |
| } |
| } |
| |
| bool MessageGenerator::RequiresArena(GeneratorFunction function) const { |
| for (const FieldDescriptor* field : FieldRange(descriptor_)) { |
| if (field_generators_.get(field).RequiresArena(function)) { |
| return true; |
| } |
| } |
| return false; |
| } |
| |
| void MessageGenerator::GenerateClassSpecificMergeImpl(io::Printer* p) { |
| if (HasSimpleBaseClass(descriptor_, options_)) return; |
| // Generate the class-specific MergeFrom, which avoids the ABSL_CHECK and |
| // cast. |
| Formatter format(p); |
| if (!HasDescriptorMethods(descriptor_->file(), options_)) { |
| // For messages that don't inherit from Message, just implement MergeFrom |
| // directly. |
| format( |
| "void $classname$::MergeFrom(const $classname$& from) {\n" |
| " $classname$* const _this = this;\n"); |
| } else { |
| format( |
| "void $classname$::MergeImpl(::$proto_ns$::Message& to_msg, const " |
| "::$proto_ns$::Message& from_msg) {\n" |
| " auto* const _this = static_cast<$classname$*>(&to_msg);\n" |
| " auto& from = static_cast<const $classname$&>(from_msg);\n"); |
| } |
| format.Indent(); |
| if (RequiresArena(GeneratorFunction::kMergeFrom)) { |
| p->Emit(R"cc( |
| ::$proto_ns$::Arena* arena = _this->GetArena(); |
| )cc"); |
| } |
| format( |
| "$annotate_mergefrom$" |
| "// @@protoc_insertion_point(class_specific_merge_from_start:" |
| "$full_name$)\n"); |
| format("$DCHK$_NE(&from, _this);\n"); |
| |
| format( |
| "$uint32$ cached_has_bits = 0;\n" |
| "(void) cached_has_bits;\n\n"); |
| |
| if (ShouldSplit(descriptor_, options_)) { |
| format( |
| "if (PROTOBUF_PREDICT_FALSE(!from.IsSplitMessageDefault())) {\n" |
| " _this->PrepareSplitMessageForWrite();\n" |
| "}\n"); |
| } |
| |
| std::vector<FieldChunk> chunks = CollectFields( |
| optimized_order_, options_, |
| [&](const FieldDescriptor* a, const FieldDescriptor* b) -> bool { |
| return HasByteIndex(a) == HasByteIndex(b) && |
| IsLikelyPresent(a, options_) == IsLikelyPresent(b, options_) && |
| ShouldSplit(a, options_) == ShouldSplit(b, options_); |
| }); |
| |
| auto it = chunks.begin(); |
| auto end = chunks.end(); |
| // cached_has_word_index maintains that: |
| // cached_has_bits = from._has_bits_[cached_has_word_index] |
| // for cached_has_word_index >= 0 |
| int cached_has_word_index = -1; |
| while (it != end) { |
| auto next = FindNextUnequalChunk(it, end, MayGroupChunksForHaswordsCheck); |
| bool has_haswords_check = |
| MaybeEmitHaswordsCheck(it, next, options_, has_bit_indices_, |
| cached_has_word_index, "from.", p); |
| |
| while (it != next) { |
| const std::vector<const FieldDescriptor*>& fields = it->fields; |
| const bool cache_has_bits = HasByteIndex(fields.front()) != kNoHasbit; |
| const bool check_has_byte = cache_has_bits && fields.size() > 1 && |
| !IsLikelyPresent(fields.back(), options_); |
| |
| if (cache_has_bits && |
| cached_has_word_index != HasWordIndex(fields.front())) { |
| cached_has_word_index = HasWordIndex(fields.front()); |
| format("cached_has_bits = from.$has_bits$[$1$];\n", |
| cached_has_word_index); |
| } |
| |
| if (check_has_byte) { |
| // Emit an if() that will let us skip the whole chunk if none are set. |
| uint32_t chunk_mask = GenChunkMask(fields, has_bit_indices_); |
| std::string chunk_mask_str = |
| absl::StrCat(absl::Hex(chunk_mask, absl::kZeroPad8)); |
| |
| // Check (up to) 8 has_bits at a time if we have more than one field in |
| // this chunk. Due to field layout ordering, we may check |
| // _has_bits_[last_chunk * 8 / 32] multiple times. |
| ABSL_DCHECK_LE(2, popcnt(chunk_mask)); |
| ABSL_DCHECK_GE(8, popcnt(chunk_mask)); |
| |
| format("if (cached_has_bits & 0x$1$u) {\n", chunk_mask_str); |
| format.Indent(); |
| } |
| |
| // Go back and emit merging code for each of the fields we processed. |
| for (const auto* field : fields) { |
| const auto& generator = field_generators_.get(field); |
| |
| if (field->is_repeated()) { |
| generator.GenerateMergingCode(p); |
| } else if (field->is_optional() && !HasHasbit(field)) { |
| // Merge semantics without true field presence: primitive fields are |
| // merged only if non-zero (numeric) or non-empty (string). |
| bool have_enclosing_if = |
| EmitFieldNonDefaultCondition(p, "from.", field); |
| if (have_enclosing_if) format.Indent(); |
| generator.GenerateMergingCode(p); |
| if (have_enclosing_if) { |
| format.Outdent(); |
| format("}\n"); |
| } |
| } else if (field->options().weak() || |
| cached_has_word_index != HasWordIndex(field)) { |
| // Check hasbit, not using cached bits. |
| auto v = p->WithVars(HasBitVars(field)); |
| format( |
| "if ((from.$has_bits$[$has_array_index$] & $has_mask$) != 0) " |
| "{\n"); |
| format.Indent(); |
| generator.GenerateMergingCode(p); |
| format.Outdent(); |
| format("}\n"); |
| } else { |
| // Check hasbit, using cached bits. |
| ABSL_CHECK(HasHasbit(field)); |
| int has_bit_index = has_bit_indices_[field->index()]; |
| const std::string mask = absl::StrCat( |
| absl::Hex(1u << (has_bit_index % 32), absl::kZeroPad8)); |
| format("if (cached_has_bits & 0x$1$u) {\n", mask); |
| format.Indent(); |
| |
| if (check_has_byte && IsPOD(field)) { |
| generator.GenerateCopyConstructorCode(p); |
| } else { |
| generator.GenerateMergingCode(p); |
| } |
| |
| format.Outdent(); |
| format("}\n"); |
| } |
| } |
| |
| if (check_has_byte) { |
| format.Outdent(); |
| format("}\n"); |
| } |
| |
| // To next chunk. |
| ++it; |
| } |
| |
| if (has_haswords_check) { |
| p->Outdent(); |
| p->Emit(R"cc( |
| } |
| )cc"); |
| |
| // Reset here as it may have been updated in just closed if statement. |
| cached_has_word_index = -1; |
| } |
| } |
| |
| if (HasBitsSize() == 1) { |
| // Optimization to avoid a load. Assuming that most messages have fewer than |
| // 32 fields, this seems useful. |
| p->Emit(R"cc( |
| _this->$has_bits$[0] |= cached_has_bits; |
| )cc"); |
| } else if (HasBitsSize() > 1) { |
| p->Emit(R"cc( |
| _this->$has_bits$.Or(from.$has_bits$); |
| )cc"); |
| } |
| |
| // Merge oneof fields. Oneof field requires oneof case check. |
| for (auto oneof : OneOfRange(descriptor_)) { |
| format("switch (from.$1$_case()) {\n", oneof->name()); |
| format.Indent(); |
| for (auto field : FieldRange(oneof)) { |
| format("case k$1$: {\n", UnderscoresToCamelCase(field->name(), true)); |
| format.Indent(); |
| field_generators_.get(field).GenerateMergingCode(p); |
| format("break;\n"); |
| format.Outdent(); |
| format("}\n"); |
| } |
| format( |
| "case $1$_NOT_SET: {\n" |
| " break;\n" |
| "}\n", |
| absl::AsciiStrToUpper(oneof->name())); |
| format.Outdent(); |
| format("}\n"); |
| } |
| if (num_weak_fields_) { |
| format( |
| "_this->$weak_field_map$.MergeFrom(from.$weak_field_map$);" |
| "\n"); |
| } |
| |
| // Merging of extensions and unknown fields is done last, to maximize |
| // the opportunity for tail calls. |
| if (descriptor_->extension_range_count() > 0) { |
| format( |
| "_this->$extensions$.MergeFrom(internal_default_instance(), " |
| "from.$extensions$);\n"); |
| } |
| |
| format( |
| "_this->_internal_metadata_.MergeFrom<$unknown_fields_type$>(from._" |
| "internal_" |
| "metadata_);\n"); |
| |
| format.Outdent(); |
| format("}\n"); |
| } |
| |
| void MessageGenerator::GenerateCopyFrom(io::Printer* p) { |
| if (HasSimpleBaseClass(descriptor_, options_)) return; |
| Formatter format(p); |
| if (HasDescriptorMethods(descriptor_->file(), options_)) { |
| // We don't override the generalized CopyFrom (aka that which |
| // takes in the Message base class as a parameter); instead we just |
| // let the base Message::CopyFrom take care of it. The base MergeFrom |
| // knows how to quickly confirm the types exactly match, and if so, will |
| // use GetClassData() to get the address of Message::CopyWithSourceCheck, |
| // which calls Clear() and then MergeFrom(), as well as making sure that |
| // clearing the destination message doesn't alter the source, when in debug |
| // builds. Most callers avoid this by passing a "from" message that is the |
| // same type as the message being merged into, rather than a generic |
| // Message. |
| } |
| |
| // Generate the class-specific CopyFrom. |
| format( |
| "void $classname$::CopyFrom(const $classname$& from) {\n" |
| "// @@protoc_insertion_point(class_specific_copy_from_start:" |
| "$full_name$)\n"); |
| format.Indent(); |
| |
| format("if (&from == this) return;\n"); |
| |
| if (!options_.opensource_runtime && HasMessageFieldOrExtension(descriptor_)) { |
| // This check is disabled in the opensource release because we're |
| // concerned that many users do not define NDEBUG in their release builds. |
| // It is also disabled if a message has neither message fields nor |
| // extensions, as it's impossible to copy from its descendant. |
| // |
| // Note that IsDescendant is implemented by reflection and not available for |
| // lite runtime. In that case, check if the size of the source has changed |
| // after Clear. |
| if (HasDescriptorMethods(descriptor_->file(), options_)) { |
| format( |
| "$DCHK$(!::_pbi::IsDescendant(*this, from))\n" |
| " << \"Source of CopyFrom cannot be a descendant of the " |
| "target.\";\n" |
| "Clear();\n"); |
| } else { |
| format( |
| "#ifndef NDEBUG\n" |
| "::size_t from_size = from.ByteSizeLong();\n" |
| "#endif\n" |
| "Clear();\n" |
| "#ifndef NDEBUG\n" |
| "$CHK$_EQ(from_size, from.ByteSizeLong())\n" |
| " << \"Source of CopyFrom changed when clearing target. Either \"\n" |
| " \"source is a nested message in target (not allowed), or \"\n" |
| " \"another thread is modifying the source.\";\n" |
| "#endif\n"); |
| } |
| } else { |
| format("Clear();\n"); |
| } |
| format("MergeFrom(from);\n"); |
| |
| format.Outdent(); |
| format("}\n"); |
| } |
| |
| void MessageGenerator::GenerateVerify(io::Printer* p) { |
| } |
| |
| void MessageGenerator::GenerateSerializeOneofFields( |
| io::Printer* p, const std::vector<const FieldDescriptor*>& fields) { |
| ABSL_CHECK(!fields.empty()); |
| if (fields.size() == 1) { |
| GenerateSerializeOneField(p, fields[0], -1); |
| return; |
| } |
| // We have multiple mutually exclusive choices. Emit a switch statement. |
| const OneofDescriptor* oneof = fields[0]->containing_oneof(); |
| p->Emit({{"name", oneof->name()}, |
| {"cases", |
| [&] { |
| for (const auto* field : fields) { |
| p->Emit({{"Name", UnderscoresToCamelCase(field->name(), true)}, |
| {"body", |
| [&] { |
| field_generators_.get(field) |
| .GenerateSerializeWithCachedSizesToArray(p); |
| }}}, |
| R"cc( |
| case k$Name$: { |
| $body$; |
| break; |
| } |
| )cc"); |
| } |
| }}}, |
| R"cc( |
| switch ($name$_case()) { |
| $cases$; |
| default: |
| break; |
| } |
| )cc"); |
| } |
| |
| void MessageGenerator::GenerateSerializeOneField(io::Printer* p, |
| const FieldDescriptor* field, |
| int cached_has_bits_index) { |
| auto v = p->WithVars(FieldVars(field, options_)); |
| auto emit_body = [&] { |
| field_generators_.get(field).GenerateSerializeWithCachedSizesToArray(p); |
| }; |
| |
| if (field->options().weak()) { |
| emit_body(); |
| p->Emit("\n"); |
| return; |
| } |
| |
| PrintFieldComment(Formatter{p}, field, options_); |
| if (HasHasbit(field)) { |
| p->Emit( |
| { |
| {"body", emit_body}, |
| {"cond", |
| [&] { |
| int has_bit_index = HasBitIndex(field); |
| auto v = p->WithVars(HasBitVars(field)); |
| // Attempt to use the state of cached_has_bits, if possible. |
| if (cached_has_bits_index == has_bit_index / 32) { |
| p->Emit("cached_has_bits & $has_mask$"); |
| } else { |
| p->Emit("($has_bits$[$has_array_index$] & $has_mask$) != 0"); |
| } |
| }}, |
| }, |
| R"cc( |
| if ($cond$) { |
| $body$; |
| } |
| )cc"); |
| } else if (field->is_optional()) { |
| bool have_enclosing_if = EmitFieldNonDefaultCondition(p, "this->", field); |
| if (have_enclosing_if) p->Indent(); |
| emit_body(); |
| if (have_enclosing_if) { |
| p->Outdent(); |
| p->Emit(R"cc( |
| } |
| )cc"); |
| } |
| } else { |
| emit_body(); |
| } |
| p->Emit("\n"); |
| } |
| |
| void MessageGenerator::GenerateSerializeOneExtensionRange(io::Printer* p, |
| int start, int end) { |
| auto v = p->WithVars(variables_); |
| p->Emit({{"start", start}, {"end", end}}, |
| R"cc( |
| // Extension range [$start$, $end$) |
| target = $extensions$._InternalSerialize( |
| internal_default_instance(), $start$, $end$, target, stream); |
| )cc"); |
| } |
| |
| void MessageGenerator::GenerateSerializeWithCachedSizesToArray(io::Printer* p) { |
| if (HasSimpleBaseClass(descriptor_, options_)) return; |
| if (descriptor_->options().message_set_wire_format()) { |
| // Special-case MessageSet. |
| p->Emit(R"cc( |
| $uint8$* $classname$::_InternalSerialize( |
| $uint8$* target, |
| ::$proto_ns$::io::EpsCopyOutputStream* stream) const { |
| $annotate_serialize$ target = |
| $extensions$.InternalSerializeMessageSetWithCachedSizesToArray( |
| internal_default_instance(), target, stream); |
| target = ::_pbi::InternalSerializeUnknownMessageSetItemsToArray( |
| $unknown_fields$, target, stream); |
| return target; |
| } |
| )cc"); |
| return; |
| } |
| |
| p->Emit( |
| { |
| {"debug_cond", ShouldSerializeInOrder(descriptor_, options_) |
| ? "1" |
| : "defined(NDEBUG)"}, |
| {"ndebug", [&] { GenerateSerializeWithCachedSizesBody(p); }}, |
| {"debug", [&] { GenerateSerializeWithCachedSizesBodyShuffled(p); }}, |
| {"ifdef", |
| [&] { |
| if (ShouldSerializeInOrder(descriptor_, options_)) { |
| p->Emit("$ndebug$"); |
| } else { |
| p->Emit(R"cc( |
| //~ force indenting level |
| #ifdef NDEBUG |
| $ndebug$; |
| #else // NDEBUG |
| $debug$; |
| #endif // !NDEBUG |
| )cc"); |
| } |
| }}, |
| }, |
| R"cc( |
| $uint8$* $classname$::_InternalSerialize( |
| $uint8$* target, |
| ::$proto_ns$::io::EpsCopyOutputStream* stream) const { |
| $annotate_serialize$; |
| // @@protoc_insertion_point(serialize_to_array_start:$full_name$) |
| $ifdef$; |
| // @@protoc_insertion_point(serialize_to_array_end:$full_name$) |
| return target; |
| } |
| )cc"); |
| } |
| |
| void MessageGenerator::GenerateSerializeWithCachedSizesBody(io::Printer* p) { |
| if (HasSimpleBaseClass(descriptor_, options_)) return; |
| // If there are multiple fields in a row from the same oneof then we |
| // coalesce them and emit a switch statement. This is more efficient |
| // because it lets the C++ compiler know this is a "at most one can happen" |
| // situation. If we emitted "if (has_x()) ...; if (has_y()) ..." the C++ |
| // compiler's emitted code might check has_y() even when has_x() is true. |
| class LazySerializerEmitter { |
| public: |
| LazySerializerEmitter(MessageGenerator* mg, io::Printer* p) |
| : mg_(mg), p_(p), cached_has_bit_index_(kNoHasbit) {} |
| |
| ~LazySerializerEmitter() { Flush(); } |
| |
| // If conditions allow, try to accumulate a run of fields from the same |
| // oneof, and handle them at the next Flush(). |
| void Emit(const FieldDescriptor* field) { |
| if (!field->has_presence() || MustFlush(field)) { |
| Flush(); |
| } |
| if (field->real_containing_oneof()) { |
| v_.push_back(field); |
| } else { |
| // TODO: Defer non-oneof fields similarly to oneof fields. |
| if (HasHasbit(field) && field->has_presence()) { |
| // We speculatively load the entire _has_bits_[index] contents, even |
| // if it is for only one field. Deferring non-oneof emitting would |
| // allow us to determine whether this is going to be useful. |
| int has_bit_index = mg_->has_bit_indices_[field->index()]; |
| if (cached_has_bit_index_ != has_bit_index / 32) { |
| // Reload. |
| int new_index = has_bit_index / 32; |
| p_->Emit({{"index", new_index}}, |
| R"cc( |
| cached_has_bits = _impl_._has_bits_[$index$]; |
| )cc"); |
| cached_has_bit_index_ = new_index; |
| } |
| } |
| |
| mg_->GenerateSerializeOneField(p_, field, cached_has_bit_index_); |
| } |
| } |
| |
| void EmitIfNotNull(const FieldDescriptor* field) { |
| if (field != nullptr) { |
| Emit(field); |
| } |
| } |
| |
| void Flush() { |
| if (!v_.empty()) { |
| mg_->GenerateSerializeOneofFields(p_, v_); |
| v_.clear(); |
| } |
| } |
| |
| private: |
| // If we have multiple fields in v_ then they all must be from the same |
| // oneof. Would adding field to v_ break that invariant? |
| bool MustFlush(const FieldDescriptor* field) { |
| return !v_.empty() && |
| v_[0]->containing_oneof() != field->containing_oneof(); |
| } |
| |
| MessageGenerator* mg_; |
| io::Printer* p_; |
| std::vector<const FieldDescriptor*> v_; |
| |
| // cached_has_bit_index_ maintains that: |
| // cached_has_bits = from._has_bits_[cached_has_bit_index_] |
| // for cached_has_bit_index_ >= 0 |
| int cached_has_bit_index_; |
| }; |
| |
| class LazyExtensionRangeEmitter { |
| public: |
| LazyExtensionRangeEmitter(MessageGenerator* mg, io::Printer* p) |
| : mg_(mg), p_(p) {} |
| |
| void AddToRange(const Descriptor::ExtensionRange* range) { |
| if (!has_current_range_) { |
| min_start_ = range->start_number(); |
| max_end_ = range->end_number(); |
| has_current_range_ = true; |
| } else { |
| min_start_ = std::min(min_start_, range->start_number()); |
| max_end_ = std::max(max_end_, range->end_number()); |
| } |
| } |
| |
| void Flush() { |
| if (has_current_range_) { |
| mg_->GenerateSerializeOneExtensionRange(p_, min_start_, max_end_); |
| } |
| has_current_range_ = false; |
| } |
| |
| private: |
| MessageGenerator* mg_; |
| io::Printer* p_; |
| bool has_current_range_ = false; |
| int min_start_ = 0; |
| int max_end_ = 0; |
| }; |
| |
| // We need to track the largest weak field, because weak fields are serialized |
| // differently than normal fields. The WeakFieldMap::FieldWriter will |
| // serialize all weak fields that are ordinally between the last serialized |
| // weak field and the current field. In order to guarantee that all weak |
| // fields are serialized, we need to make sure to emit the code to serialize |
| // the largest weak field present at some point. |
| class LargestWeakFieldHolder { |
| public: |
| const FieldDescriptor* Release() { |
| const FieldDescriptor* result = field_; |
| field_ = nullptr; |
| return result; |
| } |
| void ReplaceIfLarger(const FieldDescriptor* field) { |
| if (field_ == nullptr || field_->number() < field->number()) { |
| field_ = field; |
| } |
| } |
| |
| private: |
| const FieldDescriptor* field_ = nullptr; |
| }; |
| |
| std::vector<const FieldDescriptor*> ordered_fields = |
| SortFieldsByNumber(descriptor_); |
| |
| std::vector<const Descriptor::ExtensionRange*> sorted_extensions; |
| sorted_extensions.reserve(descriptor_->extension_range_count()); |
| for (int i = 0; i < descriptor_->extension_range_count(); ++i) { |
| sorted_extensions.push_back(descriptor_->extension_range(i)); |
| } |
| std::sort(sorted_extensions.begin(), sorted_extensions.end(), |
| ExtensionRangeSorter()); |
| p->Emit( |
| { |
| {"handle_weak_fields", |
| [&] { |
| if (num_weak_fields_ == 0) return; |
| p->Emit(R"cc( |
| ::_pbi::WeakFieldMap::FieldWriter field_writer($weak_field_map$); |
| )cc"); |
| }}, |
| {"handle_lazy_fields", |
| [&] { |
| // Merge fields and extension ranges, sorted by field number. |
| LazySerializerEmitter e(this, p); |
| LazyExtensionRangeEmitter re(this, p); |
| LargestWeakFieldHolder largest_weak_field; |
| int i, j; |
| for (i = 0, j = 0; |
| i < ordered_fields.size() || j < sorted_extensions.size();) { |
| if ((j == sorted_extensions.size()) || |
| (i < descriptor_->field_count() && |
| ordered_fields[i]->number() < |
| sorted_extensions[j]->start_number())) { |
| const FieldDescriptor* field = ordered_fields[i++]; |
| re.Flush(); |
| if (field->options().weak()) { |
| largest_weak_field.ReplaceIfLarger(field); |
| PrintFieldComment(Formatter{p}, field, options_); |
| } else { |
| e.EmitIfNotNull(largest_weak_field.Release()); |
| e.Emit(field); |
| } |
| } else { |
| e.EmitIfNotNull(largest_weak_field.Release()); |
| e.Flush(); |
| re.AddToRange(sorted_extensions[j++]); |
| } |
| } |
| re.Flush(); |
| e.EmitIfNotNull(largest_weak_field.Release()); |
| }}, |
| {"handle_unknown_fields", |
| [&] { |
| if (UseUnknownFieldSet(descriptor_->file(), options_)) { |
| p->Emit(R"cc( |
| target = |
| ::_pbi::WireFormat::InternalSerializeUnknownFieldsToArray( |
| $unknown_fields$, target, stream); |
| )cc"); |
| } else { |
| p->Emit(R"cc( |
| target = stream->WriteRaw( |
| $unknown_fields$.data(), |
| static_cast<int>($unknown_fields$.size()), target); |
| )cc"); |
| } |
| }}, |
| }, |
| R"cc( |
| $handle_weak_fields$; |
| $uint32$ cached_has_bits = 0; |
| (void)cached_has_bits; |
| |
| $handle_lazy_fields$; |
| if (PROTOBUF_PREDICT_FALSE($have_unknown_fields$)) { |
| $handle_unknown_fields$; |
| } |
| )cc"); |
| } |
| |
| void MessageGenerator::GenerateSerializeWithCachedSizesBodyShuffled( |
| io::Printer* p) { |
| std::vector<const FieldDescriptor*> ordered_fields = |
| SortFieldsByNumber(descriptor_); |
| |
| std::vector<const Descriptor::ExtensionRange*> sorted_extensions; |
| sorted_extensions.reserve(descriptor_->extension_range_count()); |
| for (int i = 0; i < descriptor_->extension_range_count(); ++i) { |
| sorted_extensions.push_back(descriptor_->extension_range(i)); |
| } |
| std::sort(sorted_extensions.begin(), sorted_extensions.end(), |
| ExtensionRangeSorter()); |
| |
| int num_fields = ordered_fields.size() + sorted_extensions.size(); |
| constexpr int kLargePrime = 1000003; |
| ABSL_CHECK_LT(num_fields, kLargePrime) |
| << "Prime offset must be greater than the number of fields to ensure " |
| "those are coprime."; |
| p->Emit( |
| { |
| {"last_field", num_fields - 1}, |
| {"field_writer", |
| [&] { |
| if (num_weak_fields_ == 0) return; |
| p->Emit(R"cc( |
| ::_pbi::WeakFieldMap::FieldWriter field_writer($weak_field_map$); |
| )cc"); |
| }}, |
| {"ordered_cases", |
| [&] { |
| size_t index = 0; |
| for (const auto* f : ordered_fields) { |
| p->Emit({{"index", index++}, |
| {"body", [&] { GenerateSerializeOneField(p, f, -1); }}}, |
| R"cc( |
| case $index$: { |
| $body$; |
| break; |
| } |
| )cc"); |
| } |
| }}, |
| {"extension_cases", |
| [&] { |
| size_t index = ordered_fields.size(); |
| for (const auto* r : sorted_extensions) { |
| p->Emit({{"index", index++}, |
| {"body", |
| [&] { |
| GenerateSerializeOneExtensionRange( |
| p, r->start_number(), r->end_number()); |
| }}}, |
| R"cc( |
| case $index$: { |
| $body$; |
| break; |
| } |
| )cc"); |
| } |
| }}, |
| {"handle_unknown_fields", |
| [&] { |
| if (UseUnknownFieldSet(descriptor_->file(), options_)) { |
| p->Emit(R"cc( |
| target = |
| ::_pbi::WireFormat::InternalSerializeUnknownFieldsToArray( |
| $unknown_fields$, target, stream); |
| )cc"); |
| } else { |
| p->Emit(R"cc( |
| target = stream->WriteRaw( |
| $unknown_fields$.data(), |
| static_cast<int>($unknown_fields$.size()), target); |
| )cc"); |
| } |
| }}, |
| }, |
| R"cc( |
| $field_writer$; |
| for (int i = $last_field$; i >= 0; i--) { |
| switch (i) { |
| $ordered_cases$; |
| $extension_cases$; |
| default: { |
| $DCHK$(false) << "Unexpected index: " << i; |
| } |
| } |
| } |
| if (PROTOBUF_PREDICT_FALSE($have_unknown_fields$)) { |
| $handle_unknown_fields$; |
| } |
| )cc"); |
| } |
| |
| std::vector<uint32_t> MessageGenerator::RequiredFieldsBitMask() const { |
| const int array_size = HasBitsSize(); |
| std::vector<uint32_t> masks(array_size, 0); |
| |
| for (auto field : FieldRange(descriptor_)) { |
| if (!field->is_required()) { |
| continue; |
| } |
| |
| const int has_bit_index = has_bit_indices_[field->index()]; |
| masks[has_bit_index / 32] |= static_cast<uint32_t>(1) |
| << (has_bit_index % 32); |
| } |
| return masks; |
| } |
| |
| void MessageGenerator::GenerateByteSize(io::Printer* p) { |
| if (HasSimpleBaseClass(descriptor_, options_)) return; |
| |
| if (descriptor_->options().message_set_wire_format()) { |
| // Special-case MessageSet. |
| p->Emit( |
| R"cc( |
| PROTOBUF_NOINLINE ::size_t $classname$::ByteSizeLong() const { |
| $annotate_bytesize$; |
| // @@protoc_insertion_point(message_set_byte_size_start:$full_name$) |
| ::size_t total_size = $extensions$.MessageSetByteSize(); |
| if ($have_unknown_fields$) { |
| total_size += ::_pbi::ComputeUnknownMessageSetItemsSize($unknown_fields$); |
| } |
| $cached_size$.Set(::_pbi::ToCachedSize(total_size)); |
| return total_size; |
| } |
| )cc"); |
| return; |
| } |
| |
| Formatter format(p); |
| format( |
| "::size_t $classname$::ByteSizeLong() const {\n" |
| "$annotate_bytesize$" |
| "// @@protoc_insertion_point(message_byte_size_start:$full_name$)\n"); |
| format.Indent(); |
| format( |
| "::size_t total_size = 0;\n" |
| "\n"); |
| |
| if (descriptor_->extension_range_count() > 0) { |
| format( |
| "total_size += $extensions$.ByteSize();\n" |
| "\n"); |
| } |
| |
| std::vector<FieldChunk> chunks = CollectFields( |
| optimized_order_, options_, |
| [&](const FieldDescriptor* a, const FieldDescriptor* b) -> bool { |
| return a->label() == b->label() && HasByteIndex(a) == HasByteIndex(b) && |
| IsLikelyPresent(a, options_) == IsLikelyPresent(b, options_) && |
| ShouldSplit(a, options_) == ShouldSplit(b, options_); |
| }); |
| |
| auto it = chunks.begin(); |
| auto end = chunks.end(); |
| int cached_has_word_index = -1; |
| |
| format( |
| "$uint32$ cached_has_bits = 0;\n" |
| "// Prevent compiler warnings about cached_has_bits being unused\n" |
| "(void) cached_has_bits;\n\n"); |
| |
| while (it != end) { |
| auto next = FindNextUnequalChunk(it, end, MayGroupChunksForHaswordsCheck); |
| bool has_haswords_check = MaybeEmitHaswordsCheck( |
| it, next, options_, has_bit_indices_, cached_has_word_index, "", p); |
| |
| while (it != next) { |
| const std::vector<const FieldDescriptor*>& fields = it->fields; |
| const bool check_has_byte = fields.size() > 1 && |
| HasWordIndex(fields[0]) != kNoHasbit && |
| !IsLikelyPresent(fields.back(), options_); |
| |
| if (check_has_byte) { |
| // Emit an if() that will let us skip the whole chunk if none are set. |
| uint32_t chunk_mask = GenChunkMask(fields, has_bit_indices_); |
| std::string chunk_mask_str = |
| absl::StrCat(absl::Hex(chunk_mask, absl::kZeroPad8)); |
| |
| // Check (up to) 8 has_bits at a time if we have more than one field in |
| // this chunk. Due to field layout ordering, we may check |
| // _has_bits_[last_chunk * 8 / 32] multiple times. |
| ABSL_DCHECK_LE(2, popcnt(chunk_mask)); |
| ABSL_DCHECK_GE(8, popcnt(chunk_mask)); |
| |
| if (cached_has_word_index != HasWordIndex(fields.front())) { |
| cached_has_word_index = HasWordIndex(fields.front()); |
| format("cached_has_bits = $has_bits$[$1$];\n", cached_has_word_index); |
| } |
| format("if (cached_has_bits & 0x$1$u) {\n", chunk_mask_str); |
| format.Indent(); |
| } |
| |
| // Go back and emit checks for each of the fields we processed. |
| for (const auto* field : fields) { |
| bool have_enclosing_if = false; |
| |
| PrintFieldComment(format, field, options_); |
| |
| if (field->is_repeated()) { |
| // No presence check is required. |
| } else if (HasHasbit(field)) { |
| PrintPresenceCheck(field, has_bit_indices_, p, |
| &cached_has_word_index); |
| have_enclosing_if = true; |
| } else { |
| // Without field presence: field is serialized only if it has a |
| // non-default value. |
| have_enclosing_if = EmitFieldNonDefaultCondition(p, "this->", field); |
| } |
| |
| if (have_enclosing_if) format.Indent(); |
| |
| field_generators_.get(field).GenerateByteSize(p); |
| |
| if (have_enclosing_if) { |
| format.Outdent(); |
| format( |
| "}\n" |
| "\n"); |
| } |
| } |
| |
| if (check_has_byte) { |
| format.Outdent(); |
| format("}\n"); |
| } |
| |
| // To next chunk. |
| ++it; |
| } |
| |
| if (has_haswords_check) { |
| p->Outdent(); |
| p->Emit(R"cc( |
| } |
| )cc"); |
| |
| // Reset here as it may have been updated in just closed if statement. |
| cached_has_word_index = -1; |
| } |
| } |
| |
| // Fields inside a oneof don't use _has_bits_ so we count them in a separate |
| // pass. |
| for (auto oneof : OneOfRange(descriptor_)) { |
| format("switch ($1$_case()) {\n", oneof->name()); |
| format.Indent(); |
| for (auto field : FieldRange(oneof)) { |
| PrintFieldComment(format, field, options_); |
| format("case k$1$: {\n", UnderscoresToCamelCase(field->name(), true)); |
| format.Indent(); |
| field_generators_.get(field).GenerateByteSize(p); |
| format("break;\n"); |
| format.Outdent(); |
| format("}\n"); |
| } |
| format( |
| "case $1$_NOT_SET: {\n" |
| " break;\n" |
| "}\n", |
| absl::AsciiStrToUpper(oneof->name())); |
| format.Outdent(); |
| format("}\n"); |
| } |
| |
| if (num_weak_fields_) { |
| // TagSize + MessageSize |
| format("total_size += $weak_field_map$.ByteSizeLong();\n"); |
| } |
| |
| if (UseUnknownFieldSet(descriptor_->file(), options_)) { |
| // We go out of our way to put the computation of the uncommon path of |
| // unknown fields in tail position. This allows for better code generation |
| // of this function for simple protos. |
| format( |
| "return MaybeComputeUnknownFieldsSize(total_size, &$cached_size$);\n"); |
| } else { |
| format("if (PROTOBUF_PREDICT_FALSE($have_unknown_fields$)) {\n"); |
| format(" total_size += $unknown_fields$.size();\n"); |
| format("}\n"); |
| |
| // We update _cached_size_ even though this is a const method. Because |
| // const methods might be called concurrently this needs to be atomic |
| // operations or the program is undefined. In practice, since any |
| // concurrent writes will be writing the exact same value, normal writes |
| // will work on all common processors. We use a dedicated wrapper class to |
| // abstract away the underlying atomic. This makes it easier on platforms |
| // where even relaxed memory order might have perf impact to replace it with |
| // ordinary loads and stores. |
| p->Emit(R"cc( |
| $cached_size$.Set(::_pbi::ToCachedSize(total_size)); |
| return total_size; |
| )cc"); |
| } |
| |
| format.Outdent(); |
| format("}\n"); |
| } |
| |
| void MessageGenerator::GenerateIsInitialized(io::Printer* p) { |
| if (HasSimpleBaseClass(descriptor_, options_)) return; |
| |
| auto has_required_field = [&](const auto* oneof) { |
| for (const auto* field : FieldRange(oneof)) { |
| if (field->cpp_type() == FieldDescriptor::CPPTYPE_MESSAGE && |
| !ShouldIgnoreRequiredFieldCheck(field, options_) && |
| scc_analyzer_->HasRequiredFields(field->message_type())) { |
| return true; |
| } |
| } |
| return false; |
| }; |
| |
| p->Emit( |
| { |
| {"test_extensions", |
| [&] { |
| if (descriptor_->extension_range_count() == 0) return; |
| p->Emit(R"cc( |
| if (!$extensions$.IsInitialized(internal_default_instance())) { |
| return false; |
| } |
| )cc"); |
| }}, |
| {"test_required_fields", |
| [&] { |
| if (num_required_fields_ == 0) return; |
| p->Emit(R"cc( |
| if (_Internal::MissingRequiredFields($has_bits$)) { |
| return false; |
| } |
| )cc"); |
| }}, |
| {"test_ordinary_fields", |
| [&] { |
| for (const auto* field : optimized_order_) { |
| field_generators_.get(field).GenerateIsInitialized(p); |
| } |
| }}, |
| {"test_weak_fields", |
| [&] { |
| if (num_weak_fields_ == 0) return; |
| p->Emit(R"cc( |
| if (!$weak_field_map$.IsInitialized()) return false; |
| )cc"); |
| }}, |
| {"test_oneof_fields", |
| [&] { |
| for (const auto* oneof : OneOfRange(descriptor_)) { |
| if (!has_required_field(oneof)) continue; |
| p->Emit({{"name", oneof->name()}, |
| {"NAME", absl::AsciiStrToUpper(oneof->name())}, |
| {"cases", |
| [&] { |
| for (const auto* field : FieldRange(oneof)) { |
| p->Emit({{"Name", UnderscoresToCamelCase( |
| field->name(), true)}, |
| {"body", |
| [&] { |
| field_generators_.get(field) |
| .GenerateIsInitialized(p); |
| }}}, |
| R"cc( |
| case k$Name$: { |
| $body$; |
| break; |
| } |
| )cc"); |
| } |
| }}}, |
| R"cc( |
| switch ($name$_case()) { |
| $cases$; |
| case $NAME$_NOT_SET: { |
| break; |
| } |
| } |
| )cc"); |
| } |
| }}, |
| }, |
| R"cc( |
| PROTOBUF_NOINLINE bool $classname$::IsInitialized() const { |
| $test_extensions$; |
| $test_required_fields$; |
| $test_ordinary_fields$; |
| $test_weak_fields$; |
| $test_oneof_fields$; |
| return true; |
| } |
| )cc"); |
| } |
| |
| } // namespace cpp |
| } // namespace compiler |
| } // namespace protobuf |
| } // namespace google |
| |
| #include "google/protobuf/port_undef.inc" |