// Protocol Buffers - Google's data interchange format // Copyright 2008 Google Inc. All rights reserved. // https://developers.google.com/protocol-buffers/ // // Redistribution and use in source and binary forms, with or without // modification, are permitted provided that the following conditions are // met: // // * Redistributions of source code must retain the above copyright // notice, this list of conditions and the following disclaimer. // * Redistributions in binary form must reproduce the above // copyright notice, this list of conditions and the following disclaimer // in the documentation and/or other materials provided with the // distribution. // * Neither the name of Google Inc. nor the names of its // contributors may be used to endorse or promote products derived from // this software without specific prior written permission. // // THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS // "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT // LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR // A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT // OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, // SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT // LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, // DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY // THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT // (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE // OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. // Author: kenton@google.com (Kenton Varda) // Based on original Protocol Buffers design by // Sanjay Ghemawat, Jeff Dean, and others. #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include // Must be included last. #include namespace google { namespace protobuf { namespace compiler { namespace cpp { using internal::WireFormat; using internal::WireFormatLite; namespace { static constexpr int kNoHasbit = -1; // Create an expression that evaluates to // "for all i, (_has_bits_[i] & masks[i]) == masks[i]" // masks is allowed to be shorter than _has_bits_, but at least one element of // masks must be non-zero. std::string ConditionalToCheckBitmasks( const std::vector& masks, bool return_success = true, StringPiece has_bits_var = "_impl_._has_bits_") { std::vector parts; for (int i = 0; i < masks.size(); i++) { if (masks[i] == 0) continue; std::string m = StrCat("0x", strings::Hex(masks[i], strings::ZERO_PAD_8)); // Each xor evaluates to 0 if the expected bits are present. parts.push_back( StrCat("((", has_bits_var, "[", i, "] & ", m, ") ^ ", m, ")")); } GOOGLE_CHECK(!parts.empty()); // If we have multiple parts, each expected to be 0, then bitwise-or them. std::string result = parts.size() == 1 ? parts[0] : StrCat("(", Join(parts, "\n | "), ")"); return result + (return_success ? " == 0" : " != 0"); } void PrintPresenceCheck(const Formatter& format, const FieldDescriptor* field, const std::vector& has_bit_indices, io::Printer* printer, int* cached_has_word_index) { if (!field->options().weak()) { int has_bit_index = has_bit_indices[field->index()]; if (*cached_has_word_index != (has_bit_index / 32)) { *cached_has_word_index = (has_bit_index / 32); format("cached_has_bits = $has_bits$[$1$];\n", *cached_has_word_index); } const std::string mask = StrCat(strings::Hex(1u << (has_bit_index % 32), strings::ZERO_PAD_8)); format("if (cached_has_bits & 0x$1$u) {\n", mask); } else { format("if (has_$1$()) {\n", FieldName(field)); } format.Indent(); } struct FieldOrderingByNumber { inline bool operator()(const FieldDescriptor* a, const FieldDescriptor* b) const { return a->number() < b->number(); } }; // Sort the fields of the given Descriptor by number into a new[]'d array // and return it. std::vector SortFieldsByNumber( const Descriptor* descriptor) { std::vector fields(descriptor->field_count()); for (int i = 0; i < descriptor->field_count(); i++) { fields[i] = descriptor->field(i); } std::sort(fields.begin(), fields.end(), FieldOrderingByNumber()); return fields; } // Functor for sorting extension ranges by their "start" field number. struct ExtensionRangeSorter { bool operator()(const Descriptor::ExtensionRange* left, const Descriptor::ExtensionRange* right) const { return left->start < right->start; } }; bool IsPOD(const FieldDescriptor* field) { if (field->is_repeated() || field->is_extension()) return false; switch (field->cpp_type()) { case FieldDescriptor::CPPTYPE_ENUM: case FieldDescriptor::CPPTYPE_INT32: case FieldDescriptor::CPPTYPE_INT64: case FieldDescriptor::CPPTYPE_UINT32: case FieldDescriptor::CPPTYPE_UINT64: case FieldDescriptor::CPPTYPE_FLOAT: case FieldDescriptor::CPPTYPE_DOUBLE: case FieldDescriptor::CPPTYPE_BOOL: return true; case FieldDescriptor::CPPTYPE_STRING: return false; default: return false; } } // Helper for the code that emits the SharedCtor() and InternalSwap() methods. // Anything that is a POD or a "normal" message (represented by a pointer) can // be manipulated as raw bytes. bool CanBeManipulatedAsRawBytes(const FieldDescriptor* field, const Options& options, MessageSCCAnalyzer* scc_analyzer) { bool ret = CanInitializeByZeroing(field); // Non-repeated, non-lazy message fields are simply raw pointers, so we can // swap them or use memset to initialize these in SharedCtor. We cannot use // this in Clear, as we need to potentially delete the existing value. ret = ret || (!field->is_repeated() && !IsLazy(field, options, scc_analyzer) && field->cpp_type() == FieldDescriptor::CPPTYPE_MESSAGE); return ret; } bool StrContains(const std::string& haystack, const std::string& needle) { return haystack.find(needle) != std::string::npos; } // Finds runs of fields for which `predicate` is true. // RunMap maps from fields that start each run to the number of fields in that // run. This is optimized for the common case that there are very few runs in // a message and that most of the eligible fields appear together. using RunMap = std::unordered_map; RunMap FindRuns(const std::vector& fields, const std::function& predicate) { RunMap runs; const FieldDescriptor* last_start = nullptr; for (auto field : fields) { if (predicate(field)) { if (last_start == nullptr) { last_start = field; } runs[last_start]++; } else { last_start = nullptr; } } return runs; } // Emits an if-statement with a condition that evaluates to true if |field| is // considered non-default (will be sent over the wire), for message types // without true field presence. Should only be called if // !HasHasbit(field). bool EmitFieldNonDefaultCondition(io::Printer* printer, const std::string& prefix, const FieldDescriptor* field) { GOOGLE_CHECK(!HasHasbit(field)); Formatter format(printer); format.Set("prefix", prefix); format.Set("name", FieldName(field)); // Merge and serialize semantics: primitive fields are merged/serialized only // if non-zero (numeric) or non-empty (string). if (!field->is_repeated() && !field->containing_oneof()) { if (field->cpp_type() == FieldDescriptor::CPPTYPE_STRING) { format("if (!$prefix$_internal_$name$().empty()) {\n"); } else if (field->cpp_type() == FieldDescriptor::CPPTYPE_MESSAGE) { // Message fields still have has_$name$() methods. format("if ($prefix$_internal_has_$name$()) {\n"); } else if (field->cpp_type() == FieldDescriptor::CPPTYPE_FLOAT) { format( "static_assert(sizeof(uint32_t) == sizeof(float), \"Code assumes " "uint32_t and float are the same size.\");\n" "float tmp_$name$ = $prefix$_internal_$name$();\n" "uint32_t raw_$name$;\n" "memcpy(&raw_$name$, &tmp_$name$, sizeof(tmp_$name$));\n" "if (raw_$name$ != 0) {\n"); } else if (field->cpp_type() == FieldDescriptor::CPPTYPE_DOUBLE) { format( "static_assert(sizeof(uint64_t) == sizeof(double), \"Code assumes " "uint64_t and double are the same size.\");\n" "double tmp_$name$ = $prefix$_internal_$name$();\n" "uint64_t raw_$name$;\n" "memcpy(&raw_$name$, &tmp_$name$, sizeof(tmp_$name$));\n" "if (raw_$name$ != 0) {\n"); } else { format("if ($prefix$_internal_$name$() != 0) {\n"); } format.Indent(); return true; } else if (field->real_containing_oneof()) { format("if (_internal_has_$name$()) {\n"); format.Indent(); return true; } return false; } // Does the given field have a has_$name$() method? bool HasHasMethod(const FieldDescriptor* field) { if (!IsProto3(field->file())) { // In proto1/proto2, every field has a has_$name$() method. return true; } // For message types without true field presence, only fields with a message // type or inside an one-of have a has_$name$() method. return field->cpp_type() == FieldDescriptor::CPPTYPE_MESSAGE || field->has_optional_keyword() || field->real_containing_oneof(); } // Collects map entry message type information. void CollectMapInfo(const Options& options, const Descriptor* descriptor, std::map* variables) { GOOGLE_CHECK(IsMapEntryMessage(descriptor)); std::map& vars = *variables; const FieldDescriptor* key = descriptor->map_key(); const FieldDescriptor* val = descriptor->map_value(); vars["key_cpp"] = PrimitiveTypeName(options, key->cpp_type()); switch (val->cpp_type()) { case FieldDescriptor::CPPTYPE_MESSAGE: vars["val_cpp"] = FieldMessageTypeName(val, options); break; case FieldDescriptor::CPPTYPE_ENUM: vars["val_cpp"] = ClassName(val->enum_type(), true); break; default: vars["val_cpp"] = PrimitiveTypeName(options, val->cpp_type()); } vars["key_wire_type"] = "TYPE_" + ToUpper(DeclaredTypeMethodName(key->type())); vars["val_wire_type"] = "TYPE_" + ToUpper(DeclaredTypeMethodName(val->type())); } // Does the given field have a private (internal helper only) has_$name$() // method? bool HasPrivateHasMethod(const FieldDescriptor* field) { // Only for oneofs in message types with no field presence. has_$name$(), // based on the oneof case, is still useful internally for generated code. return IsProto3(field->file()) && field->real_containing_oneof(); } // TODO(ckennelly): Cull these exclusions if/when these protos do not have // their methods overridden by subclasses. bool ShouldMarkClassAsFinal(const Descriptor* descriptor, const Options& options) { return true; } // Returns true to make the message serialize in order, decided by the following // factors in the order of precedence. // --options().message_set_wire_format() == true // --the message is in the allowlist (true) // --GOOGLE_PROTOBUF_SHUFFLE_SERIALIZE is defined (false) // --a ranage of message names that are allowed to stay in order (true) bool ShouldSerializeInOrder(const Descriptor* descriptor, const Options& options) { return true; } bool IsCrossFileMapField(const FieldDescriptor* field) { if (!field->is_map()) { return false; } const Descriptor* d = field->message_type(); const FieldDescriptor* value = d->FindFieldByNumber(2); return IsCrossFileMessage(value); } bool IsCrossFileMaybeMap(const FieldDescriptor* field) { if (IsCrossFileMapField(field)) { return true; } return IsCrossFileMessage(field); } bool IsRequired(const std::vector& v) { return v.front()->is_required(); } bool HasNonSplitOptionalString(const Descriptor* desc, const Options& options) { for (const auto* field : FieldRange(desc)) { if (IsString(field, options) && !field->is_repeated() && !field->real_containing_oneof() && !ShouldSplit(field, options)) { return true; } } return false; } // Collects neighboring fields based on a given criteria (equivalent predicate). template std::vector> CollectFields( const std::vector& fields, const Predicate& equivalent) { std::vector> chunks; for (auto field : fields) { if (chunks.empty() || !equivalent(chunks.back().back(), field)) { chunks.emplace_back(); } chunks.back().push_back(field); } return chunks; } // Returns a bit mask based on has_bit index of "fields" that are typically on // the same chunk. It is used in a group presence check where _has_bits_ is // masked to tell if any thing in "fields" is present. uint32_t GenChunkMask(const std::vector& fields, const std::vector& has_bit_indices) { GOOGLE_CHECK(!fields.empty()); int first_index_offset = has_bit_indices[fields.front()->index()] / 32; uint32_t chunk_mask = 0; for (auto field : fields) { // "index" defines where in the _has_bits_ the field appears. int index = has_bit_indices[field->index()]; GOOGLE_CHECK_EQ(first_index_offset, index / 32); chunk_mask |= static_cast(1) << (index % 32); } GOOGLE_CHECK_NE(0, chunk_mask); return chunk_mask; } // Return the number of bits set in n, a non-negative integer. static int popcnt(uint32_t n) { int result = 0; while (n != 0) { result += (n & 1); n = n / 2; } return result; } // For a run of cold chunks, opens and closes an external if statement that // checks multiple has_bits words to skip bulk of cold fields. class ColdChunkSkipper { public: ColdChunkSkipper( const Descriptor* descriptor, const Options& options, const std::vector>& chunks, const std::vector& has_bit_indices, const double cold_threshold) : chunks_(chunks), has_bit_indices_(has_bit_indices), access_info_map_(options.access_info_map), cold_threshold_(cold_threshold) { SetCommonVars(options, &variables_); SetCommonMessageDataVariables(descriptor, &variables_); } // May open an external if check for a batch of cold fields. "from" is the // prefix to _has_bits_ to allow MergeFrom to use "from._has_bits_". // Otherwise, it should be "". void OnStartChunk(int chunk, int cached_has_word_index, const std::string& from, io::Printer* printer); bool OnEndChunk(int chunk, io::Printer* printer); private: bool IsColdChunk(int chunk); int HasbitWord(int chunk, int offset) { return has_bit_indices_[chunks_[chunk][offset]->index()] / 32; } const std::vector>& chunks_; const std::vector& has_bit_indices_; const AccessInfoMap* access_info_map_; const double cold_threshold_; std::map variables_; int limit_chunk_ = -1; }; // Tuning parameters for ColdChunkSkipper. const double kColdRatio = 0.005; bool ColdChunkSkipper::IsColdChunk(int chunk) { // Mark this variable as used until it is actually used (void)cold_threshold_; return false; } void ColdChunkSkipper::OnStartChunk(int chunk, int cached_has_word_index, const std::string& from, io::Printer* printer) { Formatter format(printer, variables_); if (!access_info_map_) { return; } else if (chunk < limit_chunk_) { // We are already inside a run of cold chunks. return; } else if (!IsColdChunk(chunk)) { // We can't start a run of cold chunks. return; } // Find the end of consecutive cold chunks. limit_chunk_ = chunk; while (limit_chunk_ < chunks_.size() && IsColdChunk(limit_chunk_)) { limit_chunk_++; } if (limit_chunk_ <= chunk + 1) { // Require at least two chunks to emit external has_bit checks. limit_chunk_ = -1; return; } // Emit has_bit check for each has_bit_dword index. format("if (PROTOBUF_PREDICT_FALSE("); int first_word = HasbitWord(chunk, 0); while (chunk < limit_chunk_) { uint32_t mask = 0; int this_word = HasbitWord(chunk, 0); // Generate mask for chunks on the same word. for (; chunk < limit_chunk_ && HasbitWord(chunk, 0) == this_word; chunk++) { for (auto field : chunks_[chunk]) { int hasbit_index = has_bit_indices_[field->index()]; // Fields on a chunk must be in the same word. GOOGLE_CHECK_EQ(this_word, hasbit_index / 32); mask |= 1 << (hasbit_index % 32); } } if (this_word != first_word) { format(" ||\n "); } format.Set("mask", strings::Hex(mask, strings::ZERO_PAD_8)); if (this_word == cached_has_word_index) { format("(cached_has_bits & 0x$mask$u) != 0"); } else { format("($1$_impl_._has_bits_[$2$] & 0x$mask$u) != 0", from, this_word); } } format(")) {\n"); format.Indent(); } bool ColdChunkSkipper::OnEndChunk(int chunk, io::Printer* printer) { Formatter format(printer, variables_); if (chunk != limit_chunk_ - 1) { return false; } format.Outdent(); format("}\n"); return true; } void MaySetAnnotationVariable(const Options& options, StringPiece annotation_name, StringPiece injector_template_prefix, StringPiece injector_template_suffix, std::map* variables) { if (options.field_listener_options.forbidden_field_listener_events.count( std::string(annotation_name))) return; (*variables)[StrCat("annotate_", annotation_name)] = strings::Substitute( StrCat(injector_template_prefix, injector_template_suffix), (*variables)["classtype"]); } void GenerateExtensionAnnotations( const Descriptor* descriptor, const Options& options, std::map* variables) { const std::map accessor_annotations_to_hooks = { {"annotate_extension_has", "OnHasExtension"}, {"annotate_extension_clear", "OnClearExtension"}, {"annotate_extension_repeated_size", "OnExtensionSize"}, {"annotate_extension_get", "OnGetExtension"}, {"annotate_extension_mutable", "OnMutableExtension"}, {"annotate_extension_set", "OnSetExtension"}, {"annotate_extension_release", "OnReleaseExtension"}, {"annotate_repeated_extension_get", "OnGetExtension"}, {"annotate_repeated_extension_mutable", "OnMutableExtension"}, {"annotate_repeated_extension_set", "OnSetExtension"}, {"annotate_repeated_extension_add", "OnAddExtension"}, {"annotate_repeated_extension_add_mutable", "OnAddMutableExtension"}, {"annotate_repeated_extension_list", "OnListExtension"}, {"annotate_repeated_extension_list_mutable", "OnMutableListExtension"}, }; for (const auto& annotation : accessor_annotations_to_hooks) { (*variables)[annotation.first] = ""; } if (!HasTracker(descriptor, options)) { return; } StringPiece tracker = (*variables)["tracker"]; StringPiece extensions = (*variables)["extensions"]; for (const auto& annotation : accessor_annotations_to_hooks) { const std::string& annotation_name = annotation.first; const std::string& listener_call = annotation.second; if (!StrContains(annotation_name, "repeated") && !StrContains(annotation_name, "size") && !StrContains(annotation_name, "clear")) { // Primitive fields accessors. // "Has" is here as users calling "has" on a repeated field is a mistake. (*variables)[annotation_name] = StrCat( " ", tracker, ".", listener_call, "(this, id.number(), _proto_TypeTraits::GetPtr(id.number(), ", extensions, ", id.default_value_ref()));"); } else if (StrContains(annotation_name, "repeated") && !StrContains(annotation_name, "list") && !StrContains(annotation_name, "size")) { // Repeated index accessors. std::string str_index = "index"; if (StrContains(annotation_name, "add")) { str_index = StrCat(extensions, ".ExtensionSize(id.number()) - 1"); } (*variables)[annotation_name] = StrCat(" ", tracker, ".", listener_call, "(this, id.number(), " "_proto_TypeTraits::GetPtr(id.number(), ", extensions, ", ", str_index, "));"); } else if (StrContains(annotation_name, "list") || StrContains(annotation_name, "size")) { // Repeated full accessors. (*variables)[annotation_name] = StrCat( " ", tracker, ".", listener_call, "(this, id.number(), _proto_TypeTraits::GetRepeatedPtr(id.number(), ", extensions, "));"); } else { // Generic accessors such as "clear". // TODO(b/190614678): Generalize clear from both repeated and non repeated // calls, currently their underlying memory interfaces are very different. // Or think of removing clear callback as no usages are needed and no // memory exist after calling clear(). } } } } // anonymous namespace // =================================================================== MessageGenerator::MessageGenerator( const Descriptor* descriptor, const std::map& vars, int index_in_file_messages, const Options& options, MessageSCCAnalyzer* scc_analyzer) : descriptor_(descriptor), index_in_file_messages_(index_in_file_messages), classname_(ClassName(descriptor, false)), options_(options), field_generators_(descriptor, options, scc_analyzer), max_has_bit_index_(0), max_inlined_string_index_(0), num_weak_fields_(0), scc_analyzer_(scc_analyzer), variables_(vars) { if (!message_layout_helper_) { message_layout_helper_.reset(new PaddingOptimizer()); } SetCommonMessageDataVariables(descriptor, &variables_); // Variables that apply to this class variables_["classname"] = classname_; variables_["classtype"] = QualifiedClassName(descriptor_, options); variables_["full_name"] = descriptor_->full_name(); variables_["superclass"] = SuperClassName(descriptor_, options_); variables_["annotate_serialize"] = ""; variables_["annotate_deserialize"] = ""; variables_["annotate_reflection"] = ""; variables_["annotate_bytesize"] = ""; variables_["annotate_mergefrom"] = ""; if (HasTracker(descriptor_, options_)) { const std::string injector_template = StrCat(" ", variables_["tracker"], "."); MaySetAnnotationVariable(options, "serialize", injector_template, "OnSerialize(this);\n", &variables_); MaySetAnnotationVariable(options, "deserialize", injector_template, "OnDeserialize(this);\n", &variables_); // TODO(danilak): Ideally annotate_reflection should not exist and we need // to annotate all reflective calls on our own, however, as this is a cause // for side effects, i.e. reading values dynamically, we want the users know // that dynamic access can happen. MaySetAnnotationVariable(options, "reflection", injector_template, "OnGetMetadata();\n", &variables_); MaySetAnnotationVariable(options, "bytesize", injector_template, "OnByteSize(this);\n", &variables_); MaySetAnnotationVariable(options, "mergefrom", injector_template, "OnMergeFrom(_this, &from);\n", &variables_); } GenerateExtensionAnnotations(descriptor_, options_, &variables_); SetUnknownFieldsVariable(descriptor_, options_, &variables_); // Compute optimized field order to be used for layout and initialization // purposes. for (auto field : FieldRange(descriptor_)) { if (IsFieldStripped(field, options_)) { continue; } if (IsWeak(field, options_)) { num_weak_fields_++; } else if (!field->real_containing_oneof()) { optimized_order_.push_back(field); } } message_layout_helper_->OptimizeLayout(&optimized_order_, options_, scc_analyzer_); // This message has hasbits iff one or more fields need one. for (auto field : optimized_order_) { if (HasHasbit(field)) { if (has_bit_indices_.empty()) { has_bit_indices_.resize(descriptor_->field_count(), kNoHasbit); } has_bit_indices_[field->index()] = max_has_bit_index_++; } if (IsStringInlined(field, options_)) { if (inlined_string_indices_.empty()) { inlined_string_indices_.resize(descriptor_->field_count(), kNoHasbit); // The bitset[0] is for arena dtor tracking. Donating states start from // bitset[1]; max_inlined_string_index_++; } inlined_string_indices_[field->index()] = max_inlined_string_index_++; } } if (!has_bit_indices_.empty()) { field_generators_.SetHasBitIndices(has_bit_indices_); } if (!inlined_string_indices_.empty()) { field_generators_.SetInlinedStringIndices(inlined_string_indices_); } num_required_fields_ = 0; for (int i = 0; i < descriptor->field_count(); i++) { if (descriptor->field(i)->is_required()) { ++num_required_fields_; } } parse_function_generator_.reset(new ParseFunctionGenerator( descriptor_, max_has_bit_index_, has_bit_indices_, inlined_string_indices_, options_, scc_analyzer_, variables_)); } MessageGenerator::~MessageGenerator() = default; size_t MessageGenerator::HasBitsSize() const { return (max_has_bit_index_ + 31) / 32; } size_t MessageGenerator::InlinedStringDonatedSize() const { return (max_inlined_string_index_ + 31) / 32; } int MessageGenerator::HasBitIndex(const FieldDescriptor* field) const { return has_bit_indices_.empty() ? kNoHasbit : has_bit_indices_[field->index()]; } int MessageGenerator::HasByteIndex(const FieldDescriptor* field) const { int hasbit = HasBitIndex(field); return hasbit == kNoHasbit ? kNoHasbit : hasbit / 8; } int MessageGenerator::HasWordIndex(const FieldDescriptor* field) const { int hasbit = HasBitIndex(field); return hasbit == kNoHasbit ? kNoHasbit : hasbit / 32; } void MessageGenerator::AddGenerators( std::vector>* enum_generators, std::vector>* extension_generators) { for (int i = 0; i < descriptor_->enum_type_count(); i++) { enum_generators->emplace_back( new EnumGenerator(descriptor_->enum_type(i), variables_, options_)); enum_generators_.push_back(enum_generators->back().get()); } for (int i = 0; i < descriptor_->extension_count(); i++) { extension_generators->emplace_back(new ExtensionGenerator( descriptor_->extension(i), options_, scc_analyzer_)); extension_generators_.push_back(extension_generators->back().get()); } } void MessageGenerator::GenerateFieldAccessorDeclarations(io::Printer* printer) { Formatter format(printer, variables_); // optimized_fields_ does not contain fields where // field->real_containing_oneof() // so we need to iterate over those as well. // // We place the non-oneof fields in optimized_order_, as that controls the // order of the _has_bits_ entries and we want GDB's pretty printers to be // able to infer these indices from the k[FIELDNAME]FieldNumber order. std::vector ordered_fields; ordered_fields.reserve(descriptor_->field_count()); ordered_fields.insert(ordered_fields.begin(), optimized_order_.begin(), optimized_order_.end()); for (auto field : FieldRange(descriptor_)) { if (!field->real_containing_oneof() && !field->options().weak() && !IsFieldStripped(field, options_)) { continue; } ordered_fields.push_back(field); } if (!ordered_fields.empty()) { format("enum : int {\n"); for (auto field : ordered_fields) { Formatter::SaveState save(&format); std::map vars; SetCommonFieldVariables(field, &vars, options_); format.AddMap(vars); format(" ${1$$2$$}$ = $number$,\n", field, FieldConstantName(field)); } format("};\n"); } for (auto field : ordered_fields) { PrintFieldComment(format, field); Formatter::SaveState save(&format); std::map vars; SetCommonFieldVariables(field, &vars, options_); format.AddMap(vars); if (field->is_repeated()) { format("$deprecated_attr$int ${1$$name$_size$}$() const$2$\n", field, !IsFieldStripped(field, options_) ? ";" : " {__builtin_trap();}"); if (!IsFieldStripped(field, options_)) { format( "private:\n" "int ${1$_internal_$name$_size$}$() const;\n" "public:\n", field); } } else if (HasHasMethod(field)) { format("$deprecated_attr$bool ${1$has_$name$$}$() const$2$\n", field, !IsFieldStripped(field, options_) ? ";" : " {__builtin_trap();}"); if (!IsFieldStripped(field, options_)) { format( "private:\n" "bool _internal_has_$name$() const;\n" "public:\n"); } } else if (HasPrivateHasMethod(field)) { if (!IsFieldStripped(field, options_)) { format( "private:\n" "bool ${1$_internal_has_$name$$}$() const;\n" "public:\n", field); } } format("$deprecated_attr$void ${1$clear_$name$$}$()$2$\n", field, !IsFieldStripped(field, options_) ? ";" : "{__builtin_trap();}"); // Generate type-specific accessor declarations. field_generators_.get(field).GenerateAccessorDeclarations(printer); format("\n"); } if (descriptor_->extension_range_count() > 0) { // Generate accessors for extensions. // We use "_proto_TypeTraits" as a type name below because "TypeTraits" // causes problems if the class has a nested message or enum type with that // name and "_TypeTraits" is technically reserved for the C++ library since // it starts with an underscore followed by a capital letter. // // For similar reason, we use "_field_type" and "_is_packed" as parameter // names below, so that "field_type" and "is_packed" can be used as field // names. format(R"( template inline bool HasExtension( const ::PROTOBUF_NAMESPACE_ID::internal::ExtensionIdentifier< $classname$, _proto_TypeTraits, _field_type, _is_packed>& id) const { $annotate_extension_has$ return $extensions$.Has(id.number()); } template inline void ClearExtension( const ::PROTOBUF_NAMESPACE_ID::internal::ExtensionIdentifier< $classname$, _proto_TypeTraits, _field_type, _is_packed>& id) { $extensions$.ClearExtension(id.number()); $annotate_extension_clear$ } template inline int ExtensionSize( const ::PROTOBUF_NAMESPACE_ID::internal::ExtensionIdentifier< $classname$, _proto_TypeTraits, _field_type, _is_packed>& id) const { $annotate_extension_repeated_size$ return $extensions$.ExtensionSize(id.number()); } template inline typename _proto_TypeTraits::Singular::ConstType GetExtension( const ::PROTOBUF_NAMESPACE_ID::internal::ExtensionIdentifier< $classname$, _proto_TypeTraits, _field_type, _is_packed>& id) const { $annotate_extension_get$ return _proto_TypeTraits::Get(id.number(), $extensions$, id.default_value()); } template inline typename _proto_TypeTraits::Singular::MutableType MutableExtension( const ::PROTOBUF_NAMESPACE_ID::internal::ExtensionIdentifier< $classname$, _proto_TypeTraits, _field_type, _is_packed>& id) { $annotate_extension_mutable$ return _proto_TypeTraits::Mutable(id.number(), _field_type, &$extensions$); } template inline void SetExtension( const ::PROTOBUF_NAMESPACE_ID::internal::ExtensionIdentifier< $classname$, _proto_TypeTraits, _field_type, _is_packed>& id, typename _proto_TypeTraits::Singular::ConstType value) { _proto_TypeTraits::Set(id.number(), _field_type, value, &$extensions$); $annotate_extension_set$ } template inline void SetAllocatedExtension( const ::PROTOBUF_NAMESPACE_ID::internal::ExtensionIdentifier< $classname$, _proto_TypeTraits, _field_type, _is_packed>& id, typename _proto_TypeTraits::Singular::MutableType value) { _proto_TypeTraits::SetAllocated(id.number(), _field_type, value, &$extensions$); $annotate_extension_set$ } template inline void UnsafeArenaSetAllocatedExtension( const ::PROTOBUF_NAMESPACE_ID::internal::ExtensionIdentifier< $classname$, _proto_TypeTraits, _field_type, _is_packed>& id, typename _proto_TypeTraits::Singular::MutableType value) { _proto_TypeTraits::UnsafeArenaSetAllocated(id.number(), _field_type, value, &$extensions$); $annotate_extension_set$ } template PROTOBUF_NODISCARD inline typename _proto_TypeTraits::Singular::MutableType ReleaseExtension( const ::PROTOBUF_NAMESPACE_ID::internal::ExtensionIdentifier< $classname$, _proto_TypeTraits, _field_type, _is_packed>& id) { $annotate_extension_release$ return _proto_TypeTraits::Release(id.number(), _field_type, &$extensions$); } template inline typename _proto_TypeTraits::Singular::MutableType UnsafeArenaReleaseExtension( const ::PROTOBUF_NAMESPACE_ID::internal::ExtensionIdentifier< $classname$, _proto_TypeTraits, _field_type, _is_packed>& id) { $annotate_extension_release$ return _proto_TypeTraits::UnsafeArenaRelease(id.number(), _field_type, &$extensions$); } template inline typename _proto_TypeTraits::Repeated::ConstType GetExtension( const ::PROTOBUF_NAMESPACE_ID::internal::ExtensionIdentifier< $classname$, _proto_TypeTraits, _field_type, _is_packed>& id, int index) const { $annotate_repeated_extension_get$ return _proto_TypeTraits::Get(id.number(), $extensions$, index); } template inline typename _proto_TypeTraits::Repeated::MutableType MutableExtension( const ::PROTOBUF_NAMESPACE_ID::internal::ExtensionIdentifier< $classname$, _proto_TypeTraits, _field_type, _is_packed>& id, int index) { $annotate_repeated_extension_mutable$ return _proto_TypeTraits::Mutable(id.number(), index, &$extensions$); } template inline void SetExtension( const ::PROTOBUF_NAMESPACE_ID::internal::ExtensionIdentifier< $classname$, _proto_TypeTraits, _field_type, _is_packed>& id, int index, typename _proto_TypeTraits::Repeated::ConstType value) { _proto_TypeTraits::Set(id.number(), index, value, &$extensions$); $annotate_repeated_extension_set$ } template inline typename _proto_TypeTraits::Repeated::MutableType AddExtension( const ::PROTOBUF_NAMESPACE_ID::internal::ExtensionIdentifier< $classname$, _proto_TypeTraits, _field_type, _is_packed>& id) { typename _proto_TypeTraits::Repeated::MutableType to_add = _proto_TypeTraits::Add(id.number(), _field_type, &$extensions$); $annotate_repeated_extension_add_mutable$ return to_add; } template inline void AddExtension( const ::PROTOBUF_NAMESPACE_ID::internal::ExtensionIdentifier< $classname$, _proto_TypeTraits, _field_type, _is_packed>& id, typename _proto_TypeTraits::Repeated::ConstType value) { _proto_TypeTraits::Add(id.number(), _field_type, _is_packed, value, &$extensions$); $annotate_repeated_extension_add$ } template inline const typename _proto_TypeTraits::Repeated::RepeatedFieldType& GetRepeatedExtension( const ::PROTOBUF_NAMESPACE_ID::internal::ExtensionIdentifier< $classname$, _proto_TypeTraits, _field_type, _is_packed>& id) const { $annotate_repeated_extension_list$ return _proto_TypeTraits::GetRepeated(id.number(), $extensions$); } template inline typename _proto_TypeTraits::Repeated::RepeatedFieldType* MutableRepeatedExtension( const ::PROTOBUF_NAMESPACE_ID::internal::ExtensionIdentifier< $classname$, _proto_TypeTraits, _field_type, _is_packed>& id) { $annotate_repeated_extension_list_mutable$ return _proto_TypeTraits::MutableRepeated(id.number(), _field_type, _is_packed, &$extensions$); } )"); // Generate MessageSet specific APIs for proto2 MessageSet. // For testing purposes we don't check for bridge.MessageSet, so // we don't use IsProto2MessageSet if (descriptor_->options().message_set_wire_format() && !options_.opensource_runtime && !options_.lite_implicit_weak_fields) { // Special-case MessageSet format("GOOGLE_PROTOBUF_EXTENSION_MESSAGE_SET_ACCESSORS($classname$)\n"); } } for (auto oneof : OneOfRange(descriptor_)) { Formatter::SaveState saver(&format); format.Set("oneof_name", oneof->name()); format.Set("camel_oneof_name", UnderscoresToCamelCase(oneof->name(), true)); format( "void ${1$clear_$oneof_name$$}$();\n" "$camel_oneof_name$Case $oneof_name$_case() const;\n", oneof); } } void MessageGenerator::GenerateSingularFieldHasBits( const FieldDescriptor* field, Formatter format) { if (IsFieldStripped(field, options_)) { format( "inline bool $classname$::has_$name$() const { " "__builtin_trap(); }\n"); return; } if (field->options().weak()) { format( "inline bool $classname$::has_$name$() const {\n" "$annotate_has$" " return $weak_field_map$.Has($number$);\n" "}\n"); return; } if (HasHasbit(field)) { int has_bit_index = HasBitIndex(field); GOOGLE_CHECK_NE(has_bit_index, kNoHasbit); format.Set("has_array_index", has_bit_index / 32); format.Set("has_mask", strings::Hex(1u << (has_bit_index % 32), strings::ZERO_PAD_8)); format( "inline bool $classname$::_internal_has_$name$() const {\n" " bool value = " "($has_bits$[$has_array_index$] & 0x$has_mask$u) != 0;\n"); if (field->cpp_type() == FieldDescriptor::CPPTYPE_MESSAGE && !IsLazy(field, options_, scc_analyzer_)) { // We maintain the invariant that for a submessage x, has_x() returning // true implies that x_ is not null. By giving this information to the // compiler, we allow it to eliminate unnecessary null checks later on. format(" PROTOBUF_ASSUME(!value || $field$ != nullptr);\n"); } format( " return value;\n" "}\n" "inline bool $classname$::has_$name$() const {\n" "$annotate_has$" " return _internal_has_$name$();\n" "}\n"); } else if (field->cpp_type() == FieldDescriptor::CPPTYPE_MESSAGE) { // Message fields have a has_$name$() method. if (IsLazy(field, options_, scc_analyzer_)) { format( "inline bool $classname$::_internal_has_$name$() const {\n" " return !$field$.IsCleared();\n" "}\n"); } else { format( "inline bool $classname$::_internal_has_$name$() const {\n" " return this != internal_default_instance() " "&& $field$ != nullptr;\n" "}\n"); } format( "inline bool $classname$::has_$name$() const {\n" "$annotate_has$" " return _internal_has_$name$();\n" "}\n"); } } void MessageGenerator::GenerateOneofHasBits(io::Printer* printer) { Formatter format(printer, variables_); for (auto oneof : OneOfRange(descriptor_)) { format.Set("oneof_name", oneof->name()); format.Set("oneof_index", oneof->index()); format.Set("cap_oneof_name", ToUpper(oneof->name())); format( "inline bool $classname$::has_$oneof_name$() const {\n" " return $oneof_name$_case() != $cap_oneof_name$_NOT_SET;\n" "}\n" "inline void $classname$::clear_has_$oneof_name$() {\n" " $oneof_case$[$oneof_index$] = $cap_oneof_name$_NOT_SET;\n" "}\n"); } } void MessageGenerator::GenerateOneofMemberHasBits(const FieldDescriptor* field, const Formatter& format) { if (IsFieldStripped(field, options_)) { if (HasHasMethod(field)) { format( "inline bool $classname$::has_$name$() const { " "__builtin_trap(); }\n"); } format( "inline void $classname$::set_has_$name$() { __builtin_trap(); " "}\n"); return; } // Singular field in a oneof // N.B.: Without field presence, we do not use has-bits or generate // has_$name$() methods, but oneofs still have set_has_$name$(). // Oneofs also have has_$name$() but only as a private helper // method, so that generated code is slightly cleaner (vs. comparing // _oneof_case_[index] against a constant everywhere). // // If has_$name$() is private, there is no need to add an internal accessor. // Only annotate public accessors. if (HasHasMethod(field)) { format( "inline bool $classname$::_internal_has_$name$() const {\n" " return $oneof_name$_case() == k$field_name$;\n" "}\n" "inline bool $classname$::has_$name$() const {\n" "$annotate_has$" " return _internal_has_$name$();\n" "}\n"); } else if (HasPrivateHasMethod(field)) { format( "inline bool $classname$::_internal_has_$name$() const {\n" " return $oneof_name$_case() == k$field_name$;\n" "}\n"); } // set_has_$name$() for oneof fields is always private; hence should not be // annotated. format( "inline void $classname$::set_has_$name$() {\n" " $oneof_case$[$oneof_index$] = k$field_name$;\n" "}\n"); } void MessageGenerator::GenerateFieldClear(const FieldDescriptor* field, bool is_inline, Formatter format) { if (IsFieldStripped(field, options_)) { format("void $classname$::clear_$name$() { __builtin_trap(); }\n"); return; } // Generate clear_$name$(). if (is_inline) { format("inline "); } format("void $classname$::clear_$name$() {\n"); format.Indent(); if (field->real_containing_oneof()) { // Clear this field only if it is the active field in this oneof, // otherwise ignore format("if (_internal_has_$name$()) {\n"); format.Indent(); field_generators_.get(field).GenerateClearingCode(format.printer()); format("clear_has_$oneof_name$();\n"); format.Outdent(); format("}\n"); } else { if (ShouldSplit(field, options_)) { format("if (IsSplitMessageDefault()) return;\n"); } field_generators_.get(field).GenerateClearingCode(format.printer()); if (HasHasbit(field)) { int has_bit_index = HasBitIndex(field); format.Set("has_array_index", has_bit_index / 32); format.Set("has_mask", strings::Hex(1u << (has_bit_index % 32), strings::ZERO_PAD_8)); format("$has_bits$[$has_array_index$] &= ~0x$has_mask$u;\n"); } } format("$annotate_clear$"); format.Outdent(); format("}\n"); } void MessageGenerator::GenerateFieldAccessorDefinitions(io::Printer* printer) { Formatter format(printer, variables_); format("// $classname$\n\n"); for (auto field : FieldRange(descriptor_)) { PrintFieldComment(format, field); if (IsFieldStripped(field, options_)) { continue; } std::map vars; SetCommonFieldVariables(field, &vars, options_); Formatter::SaveState saver(&format); format.AddMap(vars); // Generate has_$name$() or $name$_size(). if (field->is_repeated()) { if (IsFieldStripped(field, options_)) { format( "inline int $classname$::$name$_size() const { " "__builtin_trap(); }\n"); } else { format( "inline int $classname$::_internal_$name$_size() const {\n" " return $field$$1$.size();\n" "}\n" "inline int $classname$::$name$_size() const {\n" "$annotate_size$" " return _internal_$name$_size();\n" "}\n", IsImplicitWeakField(field, options_, scc_analyzer_) && field->message_type() ? ".weak" : ""); } } else if (field->real_containing_oneof()) { format.Set("field_name", UnderscoresToCamelCase(field->name(), true)); format.Set("oneof_name", field->containing_oneof()->name()); format.Set("oneof_index", StrCat(field->containing_oneof()->index())); GenerateOneofMemberHasBits(field, format); } else { // Singular field. GenerateSingularFieldHasBits(field, format); } if (!IsCrossFileMaybeMap(field)) { GenerateFieldClear(field, true, format); } // Generate type-specific accessors. if (!IsFieldStripped(field, options_)) { field_generators_.get(field).GenerateInlineAccessorDefinitions(printer); } format("\n"); } // Generate has_$name$() and clear_has_$name$() functions for oneofs. GenerateOneofHasBits(printer); } void MessageGenerator::GenerateClassDefinition(io::Printer* printer) { Formatter format(printer, variables_); format.Set("class_final", ShouldMarkClassAsFinal(descriptor_, options_) ? "final" : ""); if (IsMapEntryMessage(descriptor_)) { std::map vars; CollectMapInfo(options_, descriptor_, &vars); vars["lite"] = HasDescriptorMethods(descriptor_->file(), options_) ? "" : "Lite"; format.AddMap(vars); format( "class $classname$ : public " "::$proto_ns$::internal::MapEntry$lite$<$classname$, \n" " $key_cpp$, $val_cpp$,\n" " ::$proto_ns$::internal::WireFormatLite::$key_wire_type$,\n" " ::$proto_ns$::internal::WireFormatLite::$val_wire_type$> {\n" "public:\n" " typedef ::$proto_ns$::internal::MapEntry$lite$<$classname$, \n" " $key_cpp$, $val_cpp$,\n" " ::$proto_ns$::internal::WireFormatLite::$key_wire_type$,\n" " ::$proto_ns$::internal::WireFormatLite::$val_wire_type$> " "SuperType;\n" " $classname$();\n" " explicit PROTOBUF_CONSTEXPR $classname$(\n" " ::$proto_ns$::internal::ConstantInitialized);\n" " explicit $classname$(::$proto_ns$::Arena* arena);\n" " void MergeFrom(const $classname$& other);\n" " static const $classname$* internal_default_instance() { return " "reinterpret_cast(&_$classname$_default_instance_); }\n"); auto utf8_check = GetUtf8CheckMode(descriptor_->field(0), options_); if (descriptor_->field(0)->type() == FieldDescriptor::TYPE_STRING && utf8_check != Utf8CheckMode::kNone) { if (utf8_check == Utf8CheckMode::kStrict) { format( " static bool ValidateKey(std::string* s) {\n" " return ::$proto_ns$::internal::WireFormatLite::" "VerifyUtf8String(s->data(), static_cast(s->size()), " "::$proto_ns$::internal::WireFormatLite::PARSE, \"$1$\");\n" " }\n", descriptor_->field(0)->full_name()); } else { GOOGLE_CHECK(utf8_check == Utf8CheckMode::kVerify); format( " static bool ValidateKey(std::string* s) {\n" "#ifndef NDEBUG\n" " ::$proto_ns$::internal::WireFormatLite::VerifyUtf8String(\n" " s->data(), static_cast(s->size()), " "::$proto_ns$::internal::" "WireFormatLite::PARSE, \"$1$\");\n" "#else\n" " (void) s;\n" "#endif\n" " return true;\n" " }\n", descriptor_->field(0)->full_name()); } } else { format(" static bool ValidateKey(void*) { return true; }\n"); } if (descriptor_->field(1)->type() == FieldDescriptor::TYPE_STRING && utf8_check != Utf8CheckMode::kNone) { if (utf8_check == Utf8CheckMode::kStrict) { format( " static bool ValidateValue(std::string* s) {\n" " return ::$proto_ns$::internal::WireFormatLite::" "VerifyUtf8String(s->data(), static_cast(s->size()), " "::$proto_ns$::internal::WireFormatLite::PARSE, \"$1$\");\n" " }\n", descriptor_->field(1)->full_name()); } else { GOOGLE_CHECK(utf8_check == Utf8CheckMode::kVerify); format( " static bool ValidateValue(std::string* s) {\n" "#ifndef NDEBUG\n" " ::$proto_ns$::internal::WireFormatLite::VerifyUtf8String(\n" " s->data(), static_cast(s->size()), " "::$proto_ns$::internal::" "WireFormatLite::PARSE, \"$1$\");\n" "#else\n" " (void) s;\n" "#endif\n" " return true;\n" " }\n", descriptor_->field(1)->full_name()); } } else { format(" static bool ValidateValue(void*) { return true; }\n"); } if (HasDescriptorMethods(descriptor_->file(), options_)) { format( " using ::$proto_ns$::Message::MergeFrom;\n" "" " ::$proto_ns$::Metadata GetMetadata() const final;\n"); } format( " friend struct ::$tablename$;\n" "};\n"); return; } format( "class $dllexport_decl $${1$$classname$$}$$ class_final$ :\n" " public $superclass$ /* @@protoc_insertion_point(" "class_definition:$full_name$) */ {\n", descriptor_); format(" public:\n"); format.Indent(); if (EnableMessageOwnedArena(descriptor_, options_)) { format( "inline $classname$() : $classname$(" "::$proto_ns$::Arena::InternalCreateMessageOwnedArena(), true) {}\n"); } else if (EnableMessageOwnedArenaTrial(descriptor_, options_)) { format( "inline $classname$() : $classname$(InMoaTrial() ? " "::$proto_ns$::Arena::InternalCreateMessageOwnedArena() : nullptr, " "InMoaTrial()) {}\n"); } else { format("inline $classname$() : $classname$(nullptr) {}\n"); } if (!HasSimpleBaseClass(descriptor_, options_)) { format("~$classname$() override;\n"); } format( "explicit PROTOBUF_CONSTEXPR " "$classname$(::$proto_ns$::internal::ConstantInitialized);\n" "\n" "$classname$(const $classname$& from);\n" "$classname$($classname$&& from) noexcept\n" " : $classname$() {\n" " *this = ::std::move(from);\n" "}\n" "\n" "inline $classname$& operator=(const $classname$& from) {\n" " CopyFrom(from);\n" " return *this;\n" "}\n" "inline $classname$& operator=($classname$&& from) noexcept {\n" " if (this == &from) return *this;\n" " if (GetOwningArena() == from.GetOwningArena()\n" "#ifdef PROTOBUF_FORCE_COPY_IN_MOVE\n" " && GetOwningArena() != nullptr\n" "#endif // !PROTOBUF_FORCE_COPY_IN_MOVE\n" " ) {\n" " InternalSwap(&from);\n" " } else {\n" " CopyFrom(from);\n" " }\n" " return *this;\n" "}\n" "\n"); if (PublicUnknownFieldsAccessors(descriptor_)) { format( "inline const $unknown_fields_type$& unknown_fields() const {\n" " return $unknown_fields$;\n" "}\n" "inline $unknown_fields_type$* mutable_unknown_fields() {\n" " return $mutable_unknown_fields$;\n" "}\n" "\n"); } // Only generate this member if it's not disabled. if (HasDescriptorMethods(descriptor_->file(), options_) && !descriptor_->options().no_standard_descriptor_accessor()) { format( "static const ::$proto_ns$::Descriptor* descriptor() {\n" " return GetDescriptor();\n" "}\n"); } if (HasDescriptorMethods(descriptor_->file(), options_)) { // These shadow non-static methods of the same names in Message. We // redefine them here because calls directly on the generated class can be // statically analyzed -- we know what descriptor types are being requested. // It also avoids a vtable dispatch. // // We would eventually like to eliminate the methods in Message, and having // this separate also lets us track calls to the base class methods // separately. format( "static const ::$proto_ns$::Descriptor* GetDescriptor() {\n" " return default_instance().GetMetadata().descriptor;\n" "}\n" "static const ::$proto_ns$::Reflection* GetReflection() {\n" " return default_instance().GetMetadata().reflection;\n" "}\n"); } format( "static const $classname$& default_instance() {\n" " return *internal_default_instance();\n" "}\n"); // Generate enum values for every field in oneofs. One list is generated for // each oneof with an additional *_NOT_SET value. for (auto oneof : OneOfRange(descriptor_)) { format("enum $1$Case {\n", UnderscoresToCamelCase(oneof->name(), true)); format.Indent(); for (auto field : FieldRange(oneof)) { format("$1$ = $2$,\n", OneofCaseConstantName(field), // 1 field->number()); // 2 } format("$1$_NOT_SET = 0,\n", ToUpper(oneof->name())); format.Outdent(); format( "};\n" "\n"); } // TODO(gerbens) make this private, while still granting other protos access. format( "static inline const $classname$* internal_default_instance() {\n" " return reinterpret_cast(\n" " &_$classname$_default_instance_);\n" "}\n" "static constexpr int kIndexInFileMessages =\n" " $1$;\n" "\n", index_in_file_messages_); if (IsAnyMessage(descriptor_, options_)) { format( "// implements Any -----------------------------------------------\n" "\n"); if (HasDescriptorMethods(descriptor_->file(), options_)) { format( "bool PackFrom(const ::$proto_ns$::Message& message) {\n" " $DCHK$_NE(&message, this);\n" " return $any_metadata$.PackFrom(GetArena(), message);\n" "}\n" "bool PackFrom(const ::$proto_ns$::Message& message,\n" " ::PROTOBUF_NAMESPACE_ID::ConstStringParam " "type_url_prefix) {\n" " $DCHK$_NE(&message, this);\n" " return $any_metadata$.PackFrom(GetArena(), message, " "type_url_prefix);\n" "}\n" "bool UnpackTo(::$proto_ns$::Message* message) const {\n" " return $any_metadata$.UnpackTo(message);\n" "}\n" "static bool GetAnyFieldDescriptors(\n" " const ::$proto_ns$::Message& message,\n" " const ::$proto_ns$::FieldDescriptor** type_url_field,\n" " const ::$proto_ns$::FieldDescriptor** value_field);\n" "template " "::value>::type>\n" "bool PackFrom(const T& message) {\n" " return $any_metadata$.PackFrom(GetArena(), message);\n" "}\n" "template " "::value>::type>\n" "bool PackFrom(const T& message,\n" " ::PROTOBUF_NAMESPACE_ID::ConstStringParam " "type_url_prefix) {\n" " return $any_metadata$.PackFrom(GetArena(), message, " "type_url_prefix);" "}\n" "template " "::value>::type>\n" "bool UnpackTo(T* message) const {\n" " return $any_metadata$.UnpackTo(message);\n" "}\n"); } else { format( "template \n" "bool PackFrom(const T& message) {\n" " return $any_metadata$.PackFrom(GetArena(), message);\n" "}\n" "template \n" "bool PackFrom(const T& message,\n" " ::PROTOBUF_NAMESPACE_ID::ConstStringParam " "type_url_prefix) {\n" " return $any_metadata$.PackFrom(GetArena(), message, " "type_url_prefix);\n" "}\n" "template \n" "bool UnpackTo(T* message) const {\n" " return $any_metadata$.UnpackTo(message);\n" "}\n"); } format( "template bool Is() const {\n" " return $any_metadata$.Is();\n" "}\n" "static bool ParseAnyTypeUrl(::PROTOBUF_NAMESPACE_ID::ConstStringParam " "type_url,\n" " std::string* full_type_name);\n"); } format( "friend void swap($classname$& a, $classname$& b) {\n" " a.Swap(&b);\n" "}\n" "inline void Swap($classname$* other) {\n" " if (other == this) return;\n" "#ifdef PROTOBUF_FORCE_COPY_IN_SWAP\n" " if (GetOwningArena() != nullptr &&\n" " GetOwningArena() == other->GetOwningArena()) {\n " "#else // PROTOBUF_FORCE_COPY_IN_SWAP\n" " if (GetOwningArena() == other->GetOwningArena()) {\n" "#endif // !PROTOBUF_FORCE_COPY_IN_SWAP\n" " InternalSwap(other);\n" " } else {\n" " ::PROTOBUF_NAMESPACE_ID::internal::GenericSwap(this, other);\n" " }\n" "}\n" "void UnsafeArenaSwap($classname$* other) {\n" " if (other == this) return;\n" " $DCHK$(GetOwningArena() == other->GetOwningArena());\n" " InternalSwap(other);\n" "}\n"); format( "\n" "// implements Message ----------------------------------------------\n" "\n" "$classname$* New(::$proto_ns$::Arena* arena = nullptr) const final {\n" " return CreateMaybeMessage<$classname$>(arena);\n" "}\n"); // For instances that derive from Message (rather than MessageLite), some // methods are virtual and should be marked as final. format.Set("full_final", HasDescriptorMethods(descriptor_->file(), options_) ? "final" : ""); if (HasGeneratedMethods(descriptor_->file(), options_)) { if (HasDescriptorMethods(descriptor_->file(), options_)) { if (!HasSimpleBaseClass(descriptor_, options_)) { format( // Use Message's built-in MergeFrom and CopyFrom when the passed-in // argument is a generic Message instance, and only define the // custom MergeFrom and CopyFrom instances when the source of the // merge/copy is known to be the same class as the destination. "using $superclass$::CopyFrom;\n" "void CopyFrom(const $classname$& from);\n" "" "using $superclass$::MergeFrom;\n" "void MergeFrom(" " const $classname$& from) {\n" " $classname$::MergeImpl(*this, from);\n" "}\n" "private:\n" "static void MergeImpl(::$proto_ns$::Message& to_msg, const " "::$proto_ns$::Message& from_msg);\n" "public:\n"); } else { format( "using $superclass$::CopyFrom;\n" "inline void CopyFrom(const $classname$& from) {\n" " $superclass$::CopyImpl(*this, from);\n" "}\n" "" "using $superclass$::MergeFrom;\n" "void MergeFrom(const $classname$& from) {\n" " $superclass$::MergeImpl(*this, from);\n" "}\n" "public:\n"); } } else { format( "void CheckTypeAndMergeFrom(const ::$proto_ns$::MessageLite& from)" " final;\n" "void CopyFrom(const $classname$& from);\n" "void MergeFrom(const $classname$& from);\n"); } if (!HasSimpleBaseClass(descriptor_, options_)) { format( "PROTOBUF_ATTRIBUTE_REINITIALIZES void Clear() final;\n" "bool IsInitialized() const final;\n" "\n" "size_t ByteSizeLong() const final;\n"); parse_function_generator_->GenerateMethodDecls(printer); format( "$uint8$* _InternalSerialize(\n" " $uint8$* target, ::$proto_ns$::io::EpsCopyOutputStream* stream) " "const final;\n"); } } if (options_.field_listener_options.inject_field_listener_events) { format("static constexpr int _kInternalFieldNumber = $1$;\n", descriptor_->field_count()); } if (!HasSimpleBaseClass(descriptor_, options_)) { format( "int GetCachedSize() const final { return " "$cached_size$.Get(); }" "\n\nprivate:\n" "void SharedCtor(::$proto_ns$::Arena* arena, bool is_message_owned);\n" "void SharedDtor();\n" "void SetCachedSize(int size) const$ full_final$;\n" "void InternalSwap($classname$* other);\n"); } format( // Friend AnyMetadata so that it can call this FullMessageName() method. "\nprivate:\n" "friend class ::$proto_ns$::internal::AnyMetadata;\n" "static $1$ FullMessageName() {\n" " return \"$full_name$\";\n" "}\n", options_.opensource_runtime ? "::PROTOBUF_NAMESPACE_ID::StringPiece" : "::StringPiece"); format( // TODO(gerbens) Make this private! Currently people are deriving from // protos to give access to this constructor, breaking the invariants // we rely on. "protected:\n" "explicit $classname$(::$proto_ns$::Arena* arena,\n" " bool is_message_owned = false);\n"); switch (NeedsArenaDestructor()) { case ArenaDtorNeeds::kOnDemand: format( "private:\n" "static void ArenaDtor(void* object);\n" "inline void OnDemandRegisterArenaDtor(::$proto_ns$::Arena* arena) " "override {\n" " if (arena == nullptr || ($inlined_string_donated_array$[0] & " "0x1u) " "== " "0) {\n" " return;\n" " }\n" " $inlined_string_donated_array$[0] &= 0xFFFFFFFEu;\n" " arena->OwnCustomDestructor(this, &$classname$::ArenaDtor);\n" "}\n"); break; case ArenaDtorNeeds::kRequired: format( "private:\n" "static void ArenaDtor(void* object);\n"); break; case ArenaDtorNeeds::kNone: break; } format( "public:\n" "\n"); if (HasDescriptorMethods(descriptor_->file(), options_)) { if (HasGeneratedMethods(descriptor_->file(), options_)) { format( "static const ClassData _class_data_;\n" "const ::$proto_ns$::Message::ClassData*" "GetClassData() const final;\n" "\n"); } format( "::$proto_ns$::Metadata GetMetadata() const final;\n" "\n"); } else { format( "std::string GetTypeName() const final;\n" "\n"); } if (ShouldSplit(descriptor_, options_)) { format( "private:\n" "inline bool IsSplitMessageDefault() const {\n" " return $split$ == reinterpret_cast(&$1$);\n" "}\n" "PROTOBUF_NOINLINE void PrepareSplitMessageForWrite();\n" "public:\n", DefaultInstanceName(descriptor_, options_, /*split=*/true)); } format( "// nested types ----------------------------------------------------\n" "\n"); // Import all nested message classes into this class's scope with typedefs. for (int i = 0; i < descriptor_->nested_type_count(); i++) { const Descriptor* nested_type = descriptor_->nested_type(i); if (!IsMapEntryMessage(nested_type)) { format.Set("nested_full_name", ClassName(nested_type, false)); format.Set("nested_name", ResolveKeyword(nested_type->name())); format("typedef ${1$$nested_full_name$$}$ ${1$$nested_name$$}$;\n", nested_type); } } if (descriptor_->nested_type_count() > 0) { format("\n"); } // Import all nested enums and their values into this class's scope with // typedefs and constants. for (int i = 0; i < descriptor_->enum_type_count(); i++) { enum_generators_[i]->GenerateSymbolImports(printer); format("\n"); } format( "// accessors -------------------------------------------------------\n" "\n"); // Generate accessor methods for all fields. GenerateFieldAccessorDeclarations(printer); // Declare extension identifiers. for (int i = 0; i < descriptor_->extension_count(); i++) { extension_generators_[i]->GenerateDeclaration(printer); } format("// @@protoc_insertion_point(class_scope:$full_name$)\n"); // Generate private members. format.Outdent(); format(" private:\n"); format.Indent(); // TODO(seongkim): Remove hack to track field access and remove this class. format("class _Internal;\n"); for (auto field : FieldRange(descriptor_)) { // set_has_***() generated in all oneofs. if (!field->is_repeated() && !field->options().weak() && field->real_containing_oneof()) { format("void set_has_$1$();\n", FieldName(field)); } } format("\n"); // Generate oneof function declarations for (auto oneof : OneOfRange(descriptor_)) { format( "inline bool has_$1$() const;\n" "inline void clear_has_$1$();\n\n", oneof->name()); } if (HasGeneratedMethods(descriptor_->file(), options_) && !descriptor_->options().message_set_wire_format() && num_required_fields_ > 1) { format( "// helper for ByteSizeLong()\n" "size_t RequiredFieldsByteSizeFallback() const;\n\n"); } if (HasGeneratedMethods(descriptor_->file(), options_)) { parse_function_generator_->GenerateDataDecls(printer); } // Prepare decls for _cached_size_ and _has_bits_. Their position in the // output will be determined later. bool need_to_emit_cached_size = !HasSimpleBaseClass(descriptor_, options_); const std::string cached_size_decl = "mutable ::$proto_ns$::internal::CachedSize _cached_size_;\n"; const size_t sizeof_has_bits = HasBitsSize(); const std::string has_bits_decl = sizeof_has_bits == 0 ? "" : StrCat("::$proto_ns$::internal::HasBits<", sizeof_has_bits, "> _has_bits_;\n"); format( "template friend class " "::$proto_ns$::Arena::InternalHelper;\n" "typedef void InternalArenaConstructable_;\n" "typedef void DestructorSkippable_;\n"); // To minimize padding, data members are divided into three sections: // (1) members assumed to align to 8 bytes // (2) members corresponding to message fields, re-ordered to optimize // alignment. // (3) members assumed to align to 4 bytes. format("struct Impl_ {\n"); format.Indent(); // Members assumed to align to 8 bytes: if (descriptor_->extension_range_count() > 0) { format( "::$proto_ns$::internal::ExtensionSet _extensions_;\n" "\n"); } if (HasTracker(descriptor_, options_)) { format("static ::$proto_ns$::AccessListener<$1$> _tracker_;\n", ClassName(descriptor_)); } // Generate _inlined_string_donated_ for inlined string type. // TODO(congliu): To avoid affecting the locality of `_has_bits_`, should this // be below or above `_has_bits_`? if (!inlined_string_indices_.empty()) { format("::$proto_ns$::internal::HasBits<$1$> _inlined_string_donated_;\n", InlinedStringDonatedSize()); } if (!has_bit_indices_.empty()) { // _has_bits_ is frequently accessed, so to reduce code size and improve // speed, it should be close to the start of the object. Placing // _cached_size_ together with _has_bits_ improves cache locality despite // potential alignment padding. format(has_bits_decl.c_str()); if (need_to_emit_cached_size) { format(cached_size_decl.c_str()); need_to_emit_cached_size = false; } } // Field members: // Emit some private and static members for (auto field : optimized_order_) { const FieldGenerator& generator = field_generators_.get(field); generator.GenerateStaticMembers(printer); if (!ShouldSplit(field, options_)) { generator.GeneratePrivateMembers(printer); } } if (ShouldSplit(descriptor_, options_)) { format("struct Split {\n"); format.Indent(); for (auto field : optimized_order_) { if (!ShouldSplit(field, options_)) continue; const FieldGenerator& generator = field_generators_.get(field); generator.GeneratePrivateMembers(printer); } format.Outdent(); format( " typedef void InternalArenaConstructable_;\n" " typedef void DestructorSkippable_;\n" "};\n" "Split* _split_;\n"); } // For each oneof generate a union for (auto oneof : OneOfRange(descriptor_)) { std::string camel_oneof_name = UnderscoresToCamelCase(oneof->name(), true); format("union $1$Union {\n", camel_oneof_name); format.Indent(); format( // explicit empty constructor is needed when union contains // ArenaStringPtr members for string fields. "constexpr $1$Union() : _constinit_{} {}\n" " ::$proto_ns$::internal::ConstantInitialized _constinit_;\n", camel_oneof_name); for (auto field : FieldRange(oneof)) { if (!IsFieldStripped(field, options_)) { field_generators_.get(field).GeneratePrivateMembers(printer); } } format.Outdent(); format("} $1$_;\n", oneof->name()); for (auto field : FieldRange(oneof)) { if (!IsFieldStripped(field, options_)) { field_generators_.get(field).GenerateStaticMembers(printer); } } } // Members assumed to align to 4 bytes: if (need_to_emit_cached_size) { format(cached_size_decl.c_str()); need_to_emit_cached_size = false; } // Generate _oneof_case_. if (descriptor_->real_oneof_decl_count() > 0) { format( "$uint32$ _oneof_case_[$1$];\n" "\n", descriptor_->real_oneof_decl_count()); } if (num_weak_fields_) { format("::$proto_ns$::internal::WeakFieldMap _weak_field_map_;\n"); } // Generate _any_metadata_ for the Any type. if (IsAnyMessage(descriptor_, options_)) { format("::$proto_ns$::internal::AnyMetadata _any_metadata_;\n"); } format.Outdent(); format("};\n"); // Only create the _impl_ field if it contains data. if (HasImplData(descriptor_, options_)) { format("union { Impl_ _impl_; };\n"); } if (ShouldSplit(descriptor_, options_)) { format( "static Impl_::Split* CreateSplitMessage(" "::$proto_ns$::Arena* arena);\n"); format("friend struct $1$;\n", DefaultInstanceType(descriptor_, options_, /*split=*/true)); } // The TableStruct struct needs access to the private parts, in order to // construct the offsets of all members. format("friend struct ::$tablename$;\n"); format.Outdent(); format("};"); GOOGLE_DCHECK(!need_to_emit_cached_size); } // NOLINT(readability/fn_size) void MessageGenerator::GenerateInlineMethods(io::Printer* printer) { if (IsMapEntryMessage(descriptor_)) return; GenerateFieldAccessorDefinitions(printer); // Generate oneof_case() functions. for (auto oneof : OneOfRange(descriptor_)) { Formatter format(printer, variables_); format.Set("camel_oneof_name", UnderscoresToCamelCase(oneof->name(), true)); format.Set("oneof_name", oneof->name()); format.Set("oneof_index", oneof->index()); format( "inline $classname$::$camel_oneof_name$Case $classname$::" "${1$$oneof_name$_case$}$() const {\n" " return $classname$::$camel_oneof_name$Case(" "$oneof_case$[$oneof_index$]);\n" "}\n", oneof); } } void MessageGenerator::GenerateSchema(io::Printer* printer, int offset, int has_offset) { Formatter format(printer, variables_); has_offset = !has_bit_indices_.empty() || IsMapEntryMessage(descriptor_) ? offset + has_offset : -1; int inlined_string_indices_offset; if (inlined_string_indices_.empty()) { inlined_string_indices_offset = -1; } else { GOOGLE_DCHECK_NE(has_offset, -1); GOOGLE_DCHECK(!IsMapEntryMessage(descriptor_)); inlined_string_indices_offset = has_offset + has_bit_indices_.size(); } format("{ $1$, $2$, $3$, sizeof($classtype$)},\n", offset, has_offset, inlined_string_indices_offset); } void MessageGenerator::GenerateClassMethods(io::Printer* printer) { Formatter format(printer, variables_); if (IsMapEntryMessage(descriptor_)) { format( "$classname$::$classname$() {}\n" "$classname$::$classname$(::$proto_ns$::Arena* arena)\n" " : SuperType(arena) {}\n" "void $classname$::MergeFrom(const $classname$& other) {\n" " MergeFromInternal(other);\n" "}\n"); if (HasDescriptorMethods(descriptor_->file(), options_)) { if (!descriptor_->options().map_entry()) { format( "::$proto_ns$::Metadata $classname$::GetMetadata() const {\n" "$annotate_reflection$" " return ::_pbi::AssignDescriptors(\n" " &$desc_table$_getter, &$desc_table$_once,\n" " $file_level_metadata$[$1$]);\n" "}\n", index_in_file_messages_); } else { format( "::$proto_ns$::Metadata $classname$::GetMetadata() const {\n" " return ::_pbi::AssignDescriptors(\n" " &$desc_table$_getter, &$desc_table$_once,\n" " $file_level_metadata$[$1$]);\n" "}\n", index_in_file_messages_); } } return; } if (IsAnyMessage(descriptor_, options_)) { if (HasDescriptorMethods(descriptor_->file(), options_)) { format( "bool $classname$::GetAnyFieldDescriptors(\n" " const ::$proto_ns$::Message& message,\n" " const ::$proto_ns$::FieldDescriptor** type_url_field,\n" " const ::$proto_ns$::FieldDescriptor** value_field) {\n" " return ::_pbi::GetAnyFieldDescriptors(\n" " message, type_url_field, value_field);\n" "}\n"); } format( "bool $classname$::ParseAnyTypeUrl(\n" " ::PROTOBUF_NAMESPACE_ID::ConstStringParam type_url,\n" " std::string* full_type_name) {\n" " return ::_pbi::ParseAnyTypeUrl(type_url, full_type_name);\n" "}\n" "\n"); } format( "class $classname$::_Internal {\n" " public:\n"); format.Indent(); if (!has_bit_indices_.empty()) { format( "using HasBits = " "decltype(std::declval<$classname$>().$has_bits$);\n"); } for (auto field : FieldRange(descriptor_)) { field_generators_.get(field).GenerateInternalAccessorDeclarations(printer); if (IsFieldStripped(field, options_)) { continue; } if (HasHasbit(field)) { int has_bit_index = HasBitIndex(field); GOOGLE_CHECK_NE(has_bit_index, kNoHasbit) << field->full_name(); format( "static void set_has_$1$(HasBits* has_bits) {\n" " (*has_bits)[$2$] |= $3$u;\n" "}\n", FieldName(field), has_bit_index / 32, (1u << (has_bit_index % 32))); } } if (num_required_fields_ > 0) { const std::vector masks_for_has_bits = RequiredFieldsBitMask(); format( "static bool MissingRequiredFields(const HasBits& has_bits) " "{\n" " return $1$;\n" "}\n", ConditionalToCheckBitmasks(masks_for_has_bits, false, "has_bits")); } format.Outdent(); format("};\n\n"); for (auto field : FieldRange(descriptor_)) { if (!IsFieldStripped(field, options_)) { field_generators_.get(field).GenerateInternalAccessorDefinitions( printer); } } // Generate non-inline field definitions. for (auto field : FieldRange(descriptor_)) { if (IsFieldStripped(field, options_)) { continue; } field_generators_.get(field).GenerateNonInlineAccessorDefinitions(printer); if (IsCrossFileMaybeMap(field)) { Formatter::SaveState saver(&format); std::map vars; SetCommonFieldVariables(field, &vars, options_); if (field->real_containing_oneof()) { SetCommonOneofFieldVariables(field, &vars); } format.AddMap(vars); GenerateFieldClear(field, false, format); } } GenerateStructors(printer); format("\n"); if (descriptor_->real_oneof_decl_count() > 0) { GenerateOneofClear(printer); format("\n"); } if (HasGeneratedMethods(descriptor_->file(), options_)) { GenerateClear(printer); format("\n"); if (!HasSimpleBaseClass(descriptor_, options_)) { parse_function_generator_->GenerateMethodImpls(printer); format("\n"); parse_function_generator_->GenerateDataDefinitions(printer); } GenerateSerializeWithCachedSizesToArray(printer); format("\n"); GenerateByteSize(printer); format("\n"); GenerateMergeFrom(printer); format("\n"); GenerateClassSpecificMergeImpl(printer); format("\n"); GenerateCopyFrom(printer); format("\n"); GenerateIsInitialized(printer); format("\n"); } if (ShouldSplit(descriptor_, options_)) { format( "void $classname$::PrepareSplitMessageForWrite() {\n" " if (IsSplitMessageDefault()) {\n" " $split$ = CreateSplitMessage(GetArenaForAllocation());\n" " }\n" "}\n"); } GenerateVerify(printer); GenerateSwap(printer); format("\n"); if (HasDescriptorMethods(descriptor_->file(), options_)) { if (!descriptor_->options().map_entry()) { format( "::$proto_ns$::Metadata $classname$::GetMetadata() const {\n" "$annotate_reflection$" " return ::_pbi::AssignDescriptors(\n" " &$desc_table$_getter, &$desc_table$_once,\n" " $file_level_metadata$[$1$]);\n" "}\n", index_in_file_messages_); } else { format( "::$proto_ns$::Metadata $classname$::GetMetadata() const {\n" " return ::_pbi::AssignDescriptors(\n" " &$desc_table$_getter, &$desc_table$_once,\n" " $file_level_metadata$[$1$]);\n" "}\n", index_in_file_messages_); } } else { format( "std::string $classname$::GetTypeName() const {\n" " return \"$full_name$\";\n" "}\n" "\n"); } if (HasTracker(descriptor_, options_)) { format( "::$proto_ns$::AccessListener<$classtype$> " "$1$::$tracker$(&FullMessageName);\n", ClassName(descriptor_)); } } std::pair MessageGenerator::GenerateOffsets( io::Printer* printer) { Formatter format(printer, variables_); if (!has_bit_indices_.empty() || IsMapEntryMessage(descriptor_)) { format("PROTOBUF_FIELD_OFFSET($classtype$, $has_bits$),\n"); } else { format("~0u, // no _has_bits_\n"); } format("PROTOBUF_FIELD_OFFSET($classtype$, _internal_metadata_),\n"); if (descriptor_->extension_range_count() > 0) { format("PROTOBUF_FIELD_OFFSET($classtype$, $extensions$),\n"); } else { format("~0u, // no _extensions_\n"); } if (descriptor_->real_oneof_decl_count() > 0) { format("PROTOBUF_FIELD_OFFSET($classtype$, $oneof_case$[0]),\n"); } else { format("~0u, // no _oneof_case_\n"); } if (num_weak_fields_ > 0) { format("PROTOBUF_FIELD_OFFSET($classtype$, $weak_field_map$),\n"); } else { format("~0u, // no _weak_field_map_\n"); } if (!inlined_string_indices_.empty()) { format( "PROTOBUF_FIELD_OFFSET($classtype$, " "$inlined_string_donated_array$),\n"); } else { format("~0u, // no _inlined_string_donated_\n"); } const int kNumGenericOffsets = 6; // the number of fixed offsets above const size_t offsets = kNumGenericOffsets + descriptor_->field_count() + descriptor_->real_oneof_decl_count(); size_t entries = offsets; for (auto field : FieldRange(descriptor_)) { if (IsFieldStripped(field, options_)) { format("~0u, // stripped\n"); continue; } // TODO(sbenza): We should not have an entry in the offset table for fields // that do not use them. if (field->options().weak() || field->real_containing_oneof()) { // Mark the field to prevent unintentional access through reflection. // Don't use the top bit because that is for unused fields. format("::_pbi::kInvalidFieldOffsetTag"); } else { format("PROTOBUF_FIELD_OFFSET($classtype$$1$, $2$)", ShouldSplit(field, options_) ? "::Impl_::Split" : "", ShouldSplit(field, options_) ? FieldName(field) + "_" : FieldMemberName(field, /*cold=*/false)); } // Some information about a field is in the pdproto profile. The profile is // only available at compile time. So we embed such information in the // offset of the field, so that the information is available when // reflectively accessing the field at run time. // // Embed whether the field is eagerly verified lazy or inlined string to the // LSB of the offset. if (IsEagerlyVerifiedLazy(field, options_, scc_analyzer_)) { format(" | 0x1u // eagerly verified lazy\n"); } else if (IsStringInlined(field, options_)) { format(" | 0x1u // inlined\n"); } format(",\n"); } int count = 0; for (auto oneof : OneOfRange(descriptor_)) { format("PROTOBUF_FIELD_OFFSET($classtype$, _impl_.$1$_),\n", oneof->name()); count++; } GOOGLE_CHECK_EQ(count, descriptor_->real_oneof_decl_count()); if (IsMapEntryMessage(descriptor_)) { entries += 2; format( "0,\n" "1,\n"); } else if (!has_bit_indices_.empty()) { entries += has_bit_indices_.size(); for (int i = 0; i < has_bit_indices_.size(); i++) { const std::string index = has_bit_indices_[i] >= 0 ? StrCat(has_bit_indices_[i]) : "~0u"; format("$1$,\n", index); } } if (!inlined_string_indices_.empty()) { entries += inlined_string_indices_.size(); for (int inlined_string_index : inlined_string_indices_) { const std::string index = inlined_string_index >= 0 ? StrCat(inlined_string_index, ", // inlined_string_index") : "~0u,"; format("$1$\n", index); } } return std::make_pair(entries, offsets); } void MessageGenerator::GenerateSharedConstructorCode(io::Printer* printer) { if (HasSimpleBaseClass(descriptor_, options_)) return; Formatter format(printer, variables_); format( "inline void $classname$::SharedCtor(\n" " ::_pb::Arena* arena, bool is_message_owned) {\n" " (void)arena;\n" " (void)is_message_owned;\n"); format.Indent(); // Impl_ _impl_. format("new (&_impl_) Impl_{"); format.Indent(); const char* field_sep = " "; const auto put_sep = [&] { format("\n$1$ ", field_sep); field_sep = ","; }; // Note: any fields without move/copy constructors can't be explicitly // aggregate initialized pre-C++17. if (descriptor_->extension_range_count() > 0) { put_sep(); format("/*decltype($extensions$)*/{::_pbi::ArenaInitialized(), arena}"); } if (!inlined_string_indices_.empty()) { put_sep(); format("decltype($inlined_string_donated_array$){}"); } bool need_to_emit_cached_size = !HasSimpleBaseClass(descriptor_, options_); if (!has_bit_indices_.empty()) { put_sep(); format("decltype($has_bits$){}"); if (need_to_emit_cached_size) { put_sep(); format("/*decltype($cached_size$)*/{}"); need_to_emit_cached_size = false; } } // Initialize member variables with arena constructor. for (auto field : optimized_order_) { GOOGLE_DCHECK(!IsFieldStripped(field, options_)); if (ShouldSplit(field, options_)) { continue; } put_sep(); field_generators_.get(field).GenerateAggregateInitializer(printer); } if (ShouldSplit(descriptor_, options_)) { put_sep(); format("decltype($split$){reinterpret_cast(&$1$)}", DefaultInstanceName(descriptor_, options_, /*split=*/true)); } for (auto oneof : OneOfRange(descriptor_)) { put_sep(); format("decltype(_impl_.$1$_){}", oneof->name()); } if (need_to_emit_cached_size) { put_sep(); format("/*decltype($cached_size$)*/{}"); } if (descriptor_->real_oneof_decl_count() != 0) { put_sep(); format("/*decltype($oneof_case$)*/{}"); } if (num_weak_fields_ > 0) { put_sep(); format("decltype($weak_field_map$){arena}"); } if (IsAnyMessage(descriptor_, options_)) { put_sep(); // AnyMetadata has no move constructor. format("/*decltype($any_metadata$)*/{&_impl_.type_url_, &_impl_.value_}"); } format.Outdent(); format("\n};\n"); if (!inlined_string_indices_.empty()) { // Donate inline string fields. format.Indent(); // The last bit is the tracking bit for registering ArenaDtor. The bit is 1 // means ArenaDtor is not registered on construction, and on demand register // is needed. format("if (arena != nullptr) {\n"); if (NeedsArenaDestructor() == ArenaDtorNeeds::kOnDemand) { format( " if (!is_message_owned) {\n" " $inlined_string_donated_array$[0] = ~0u;\n" " } else {\n" // We should not register ArenaDtor for MOA. " $inlined_string_donated_array$[0] = 0xFFFFFFFEu;\n" " }\n"); } else { format(" $inlined_string_donated_array$[0] = 0xFFFFFFFEu;\n"); } for (size_t i = 1; i < InlinedStringDonatedSize(); ++i) { format(" $inlined_string_donated_array$[$1$] = ~0u;\n", i); } format("}\n"); format.Outdent(); } for (const FieldDescriptor* field : optimized_order_) { if (ShouldSplit(field, options_)) { continue; } field_generators_.get(field).GenerateConstructorCode(printer); } for (auto oneof : OneOfRange(descriptor_)) { format("clear_has_$1$();\n", oneof->name()); } format.Outdent(); format("}\n\n"); } void MessageGenerator::GenerateCreateSplitMessage(io::Printer* printer) { Formatter format(printer, variables_); format( "$classname$::Impl_::Split* " "$classname$::CreateSplitMessage(::$proto_ns$::Arena* arena) {\n"); format.Indent(); const char* field_sep = " "; const auto put_sep = [&] { format("\n$1$ ", field_sep); field_sep = ","; }; format( "const size_t size = sizeof(Impl_::Split);\n" "void* chunk = (arena == nullptr) ?\n" " ::operator new(size) :\n" " arena->AllocateAligned(size, alignof(Impl_::Split));\n" "Impl_::Split* ptr = reinterpret_cast(chunk);\n" "new (ptr) Impl_::Split{"); format.Indent(); for (const FieldDescriptor* field : optimized_order_) { GOOGLE_DCHECK(!IsFieldStripped(field, options_)); if (ShouldSplit(field, options_)) { put_sep(); field_generators_.get(field).GenerateAggregateInitializer(printer); } } format.Outdent(); format("};\n"); for (const FieldDescriptor* field : optimized_order_) { GOOGLE_DCHECK(!IsFieldStripped(field, options_)); if (ShouldSplit(field, options_)) { field_generators_.get(field).GenerateCreateSplitMessageCode(printer); } } format("return ptr;\n"); format.Outdent(); format("}\n"); } void MessageGenerator::GenerateInitDefaultSplitInstance(io::Printer* printer) { if (!ShouldSplit(descriptor_, options_)) return; Formatter format(printer, variables_); const char* field_sep = " "; const auto put_sep = [&] { format("\n$1$ ", field_sep); field_sep = ","; }; for (const auto* field : optimized_order_) { if (ShouldSplit(field, options_)) { put_sep(); field_generators_.get(field).GenerateConstexprAggregateInitializer( printer); } } } void MessageGenerator::GenerateSharedDestructorCode(io::Printer* printer) { if (HasSimpleBaseClass(descriptor_, options_)) return; Formatter format(printer, variables_); format("inline void $classname$::SharedDtor() {\n"); format.Indent(); format("$DCHK$(GetArenaForAllocation() == nullptr);\n"); if (descriptor_->extension_range_count() > 0) { format("$extensions$.~ExtensionSet();\n"); } // Write the destructors for each field except oneof members. // optimized_order_ does not contain oneof fields. for (auto field : optimized_order_) { if (ShouldSplit(field, options_)) { continue; } field_generators_.get(field).GenerateDestructorCode(printer); } if (ShouldSplit(descriptor_, options_)) { format("if (!IsSplitMessageDefault()) {\n"); format.Indent(); format("auto* $cached_split_ptr$ = $split$;\n"); for (auto field : optimized_order_) { if (ShouldSplit(field, options_)) { field_generators_.get(field).GenerateDestructorCode(printer); } } format("delete $cached_split_ptr$;\n"); format.Outdent(); format("}\n"); } // Generate code to destruct oneofs. Clearing should do the work. for (auto oneof : OneOfRange(descriptor_)) { format( "if (has_$1$()) {\n" " clear_$1$();\n" "}\n", oneof->name()); } if (num_weak_fields_) { format("$weak_field_map$.ClearAll();\n"); } if (IsAnyMessage(descriptor_, options_)) { format("$any_metadata$.~AnyMetadata();\n"); } format.Outdent(); format( "}\n" "\n"); } ArenaDtorNeeds MessageGenerator::NeedsArenaDestructor() const { if (HasSimpleBaseClass(descriptor_, options_)) return ArenaDtorNeeds::kNone; ArenaDtorNeeds needs = ArenaDtorNeeds::kNone; for (const auto* field : FieldRange(descriptor_)) { if (IsFieldStripped(field, options_)) continue; needs = std::max(needs, field_generators_.get(field).NeedsArenaDestructor()); } return needs; } void MessageGenerator::GenerateArenaDestructorCode(io::Printer* printer) { GOOGLE_CHECK(NeedsArenaDestructor() > ArenaDtorNeeds::kNone); Formatter format(printer, variables_); // Generate the ArenaDtor() method. Track whether any fields actually produced // code that needs to be called. format("void $classname$::ArenaDtor(void* object) {\n"); format.Indent(); // This code is placed inside a static method, rather than an ordinary one, // since that simplifies Arena's destructor list (ordinary function pointers // rather than member function pointers). _this is the object being // destructed. format("$classname$* _this = reinterpret_cast< $classname$* >(object);\n"); // Process non-oneof fields first. for (auto field : optimized_order_) { if (IsFieldStripped(field, options_) || ShouldSplit(field, options_)) continue; const FieldGenerator& fg = field_generators_.get(field); fg.GenerateArenaDestructorCode(printer); } if (ShouldSplit(descriptor_, options_)) { format("if (!_this->IsSplitMessageDefault()) {\n"); format.Indent(); for (auto field : optimized_order_) { if (IsFieldStripped(field, options_) || !ShouldSplit(field, options_)) continue; const FieldGenerator& fg = field_generators_.get(field); fg.GenerateArenaDestructorCode(printer); } format.Outdent(); format("}\n"); } // Process oneof fields. for (auto oneof : OneOfRange(descriptor_)) { for (auto field : FieldRange(oneof)) { if (IsFieldStripped(field, options_)) continue; field_generators_.get(field).GenerateArenaDestructorCode(printer); } } format.Outdent(); format("}\n"); } void MessageGenerator::GenerateConstexprConstructor(io::Printer* printer) { Formatter format(printer, variables_); if (IsMapEntryMessage(descriptor_) || !HasImplData(descriptor_, options_)) { format( "PROTOBUF_CONSTEXPR $classname$::$classname$(\n" " ::_pbi::ConstantInitialized) {}\n"); return; } format( "PROTOBUF_CONSTEXPR $classname$::$classname$(\n" " ::_pbi::ConstantInitialized)"); bool need_to_emit_cached_size = !HasSimpleBaseClass(descriptor_, options_); format(": _impl_{"); format.Indent(); const char* field_sep = " "; const auto put_sep = [&] { format("\n$1$ ", field_sep); field_sep = ","; }; if (descriptor_->extension_range_count() > 0) { put_sep(); format("/*decltype($extensions$)*/{}"); } if (!inlined_string_indices_.empty()) { put_sep(); format("/*decltype($inlined_string_donated_array$)*/{}"); } if (!has_bit_indices_.empty()) { put_sep(); format("/*decltype($has_bits$)*/{}"); if (need_to_emit_cached_size) { put_sep(); format("/*decltype($cached_size$)*/{}"); need_to_emit_cached_size = false; } } for (auto field : optimized_order_) { if (ShouldSplit(field, options_)) { continue; } put_sep(); field_generators_.get(field).GenerateConstexprAggregateInitializer( printer); } if (ShouldSplit(descriptor_, options_)) { put_sep(); format("/*decltype($split$)*/&$1$._instance", DefaultInstanceName(descriptor_, options_, /*split=*/true)); } for (auto oneof : OneOfRange(descriptor_)) { put_sep(); format("/*decltype(_impl_.$1$_)*/{}", oneof->name()); } if (need_to_emit_cached_size) { put_sep(); format("/*decltype($cached_size$)*/{}"); } if (descriptor_->real_oneof_decl_count() != 0) { put_sep(); format("/*decltype($oneof_case$)*/{}"); } if (num_weak_fields_) { put_sep(); format("/*decltype($weak_field_map$)*/{}"); } if (IsAnyMessage(descriptor_, options_)) { put_sep(); format( "/*decltype($any_metadata$)*/{&_impl_.type_url_, " "&_impl_.value_}"); } format.Outdent(); format("} {}\n"); } void MessageGenerator::GenerateCopyConstructorBody(io::Printer* printer) const { Formatter format(printer, variables_); const RunMap runs = FindRuns(optimized_order_, [this](const FieldDescriptor* field) { return IsPOD(field) && !ShouldSplit(field, options_); }); std::string pod_template = "::memcpy(&$first$, &from.$first$,\n" " static_cast(reinterpret_cast(&$last$) -\n" " reinterpret_cast(&$first$)) + sizeof($last$));\n"; if (ShouldSplit(descriptor_, options_)) { format("if (!from.IsSplitMessageDefault()) {\n"); format.Indent(); format("_this->PrepareSplitMessageForWrite();\n"); for (auto field : optimized_order_) { if (ShouldSplit(field, options_)) { field_generators_.get(field).GenerateCopyConstructorCode(printer); } } format.Outdent(); format("}\n"); } for (size_t i = 0; i < optimized_order_.size(); ++i) { const FieldDescriptor* field = optimized_order_[i]; if (ShouldSplit(field, options_)) { continue; } const auto it = runs.find(field); // We only apply the memset technique to runs of more than one field, as // assignment is better than memset for generated code clarity. if (it != runs.end() && it->second > 1) { // Use a memset, then skip run_length fields. const size_t run_length = it->second; const std::string first_field_name = FieldMemberName(field, /*cold=*/false); const std::string last_field_name = FieldMemberName(optimized_order_[i + run_length - 1], /*cold=*/false); format.Set("first", first_field_name); format.Set("last", last_field_name); format(pod_template.c_str()); i += run_length - 1; // ++i at the top of the loop. } else { field_generators_.get(field).GenerateCopyConstructorCode(printer); } } } void MessageGenerator::GenerateStructors(io::Printer* printer) { Formatter format(printer, variables_); format( "$classname$::$classname$(::$proto_ns$::Arena* arena,\n" " bool is_message_owned)\n" " : $1$(arena, is_message_owned) {\n", SuperClassName(descriptor_, options_)); if (!HasSimpleBaseClass(descriptor_, options_)) { format(" SharedCtor(arena, is_message_owned);\n"); if (NeedsArenaDestructor() == ArenaDtorNeeds::kRequired) { format( " if (arena != nullptr && !is_message_owned) {\n" " arena->OwnCustomDestructor(this, &$classname$::ArenaDtor);\n" " }\n"); } } format( " // @@protoc_insertion_point(arena_constructor:$full_name$)\n" "}\n"); std::map vars; SetUnknownFieldsVariable(descriptor_, options_, &vars); format.AddMap(vars); // Generate the copy constructor. if (UsingImplicitWeakFields(descriptor_->file(), options_)) { // If we are in lite mode and using implicit weak fields, we generate a // one-liner copy constructor that delegates to MergeFrom. This saves some // code size and also cuts down on the complexity of implicit weak fields. // We might eventually want to do this for all lite protos. format( "$classname$::$classname$(const $classname$& from)\n" " : $classname$() {\n" " MergeFrom(from);\n" "}\n"); } else { format( "$classname$::$classname$(const $classname$& from)\n" " : $superclass$() {\n"); format.Indent(); format("$classname$* const _this = this; (void)_this;\n"); if (HasImplData(descriptor_, options_)) { const char* field_sep = " "; const auto put_sep = [&] { format("\n$1$ ", field_sep); field_sep = ","; }; format("new (&_impl_) Impl_{"); format.Indent(); if (descriptor_->extension_range_count() > 0) { put_sep(); format("/*decltype($extensions$)*/{}"); } if (!inlined_string_indices_.empty()) { // Do not copy inlined_string_donated_, because this is not an arena // constructor. put_sep(); format("decltype($inlined_string_donated_array$){}"); } bool need_to_emit_cached_size = !HasSimpleBaseClass(descriptor_, options_); if (!has_bit_indices_.empty()) { put_sep(); format("decltype($has_bits$){from.$has_bits$}"); if (need_to_emit_cached_size) { put_sep(); format("/*decltype($cached_size$)*/{}"); need_to_emit_cached_size = false; } } // Initialize member variables with arena constructor. for (auto field : optimized_order_) { if (ShouldSplit(field, options_)) { continue; } put_sep(); field_generators_.get(field).GenerateCopyAggregateInitializer(printer); } if (ShouldSplit(descriptor_, options_)) { put_sep(); format("decltype($split$){reinterpret_cast(&$1$)}", DefaultInstanceName(descriptor_, options_, /*split=*/true)); } for (auto oneof : OneOfRange(descriptor_)) { put_sep(); format("decltype(_impl_.$1$_){}", oneof->name()); } if (need_to_emit_cached_size) { put_sep(); format("/*decltype($cached_size$)*/{}"); } if (descriptor_->real_oneof_decl_count() != 0) { put_sep(); format("/*decltype($oneof_case$)*/{}"); } if (num_weak_fields_ > 0) { put_sep(); format("decltype($weak_field_map$){from.$weak_field_map$}"); } if (IsAnyMessage(descriptor_, options_)) { put_sep(); format( "/*decltype($any_metadata$)*/{&_impl_.type_url_, &_impl_.value_}"); } format.Outdent(); format("};\n\n"); } format( "_internal_metadata_.MergeFrom<$unknown_fields_type$>(from._internal_" "metadata_);\n"); if (descriptor_->extension_range_count() > 0) { format( "$extensions$.MergeFrom(internal_default_instance(), " "from.$extensions$);\n"); } GenerateCopyConstructorBody(printer); // Copy oneof fields. Oneof field requires oneof case check. for (auto oneof : OneOfRange(descriptor_)) { format( "clear_has_$1$();\n" "switch (from.$1$_case()) {\n", oneof->name()); format.Indent(); for (auto field : FieldRange(oneof)) { format("case k$1$: {\n", UnderscoresToCamelCase(field->name(), true)); format.Indent(); if (!IsFieldStripped(field, options_)) { field_generators_.get(field).GenerateMergingCode(printer); } format("break;\n"); format.Outdent(); format("}\n"); } format( "case $1$_NOT_SET: {\n" " break;\n" "}\n", ToUpper(oneof->name())); format.Outdent(); format("}\n"); } format.Outdent(); format( " // @@protoc_insertion_point(copy_constructor:$full_name$)\n" "}\n" "\n"); } // Generate the shared constructor code. GenerateSharedConstructorCode(printer); if (ShouldSplit(descriptor_, options_)) { GenerateCreateSplitMessage(printer); } // Generate the destructor. if (!HasSimpleBaseClass(descriptor_, options_)) { format( "$classname$::~$classname$() {\n" " // @@protoc_insertion_point(destructor:$full_name$)\n"); format( " if (auto *arena = " "_internal_metadata_.DeleteReturnArena<$unknown_fields_type$>()) {\n" " (void)arena;\n"); if (NeedsArenaDestructor() > ArenaDtorNeeds::kNone) { format(" ArenaDtor(this);\n"); } format( " return;\n" " }\n"); format( " SharedDtor();\n" "}\n" "\n"); } else { // For messages using simple base classes, having no destructor // allows our vtable to share the same destructor as every other // message with a simple base class. This works only as long as // we have no fields needing destruction, of course. (No strings // or extensions) } // Generate the shared destructor code. GenerateSharedDestructorCode(printer); // Generate the arena-specific destructor code. if (NeedsArenaDestructor() > ArenaDtorNeeds::kNone) { GenerateArenaDestructorCode(printer); } if (!HasSimpleBaseClass(descriptor_, options_)) { // Generate SetCachedSize. format( "void $classname$::SetCachedSize(int size) const {\n" " $cached_size$.Set(size);\n" "}\n"); } } void MessageGenerator::GenerateSourceInProto2Namespace(io::Printer* printer) { Formatter format(printer, variables_); format( "template<> " "PROTOBUF_NOINLINE $classtype$*\n" "Arena::CreateMaybeMessage< $classtype$ >(Arena* arena) {\n" " return Arena::CreateMessageInternal< $classtype$ >(arena);\n" "}\n"); } void MessageGenerator::GenerateClear(io::Printer* printer) { if (HasSimpleBaseClass(descriptor_, options_)) return; Formatter format(printer, variables_); // The maximum number of bytes we will memset to zero without checking their // hasbit to see if a zero-init is necessary. const int kMaxUnconditionalPrimitiveBytesClear = 4; format( "void $classname$::Clear() {\n" "// @@protoc_insertion_point(message_clear_start:$full_name$)\n"); format.Indent(); format( // TODO(jwb): It would be better to avoid emitting this if it is not used, // rather than emitting a workaround for the resulting warning. "$uint32$ cached_has_bits = 0;\n" "// Prevent compiler warnings about cached_has_bits being unused\n" "(void) cached_has_bits;\n\n"); if (descriptor_->extension_range_count() > 0) { format("$extensions$.Clear();\n"); } // Collect fields into chunks. Each chunk may have an if() condition that // checks all hasbits in the chunk and skips it if none are set. int zero_init_bytes = 0; for (const auto& field : optimized_order_) { if (CanInitializeByZeroing(field)) { zero_init_bytes += EstimateAlignmentSize(field); } } bool merge_zero_init = zero_init_bytes > kMaxUnconditionalPrimitiveBytesClear; int chunk_count = 0; std::vector> chunks = CollectFields( optimized_order_, [&](const FieldDescriptor* a, const FieldDescriptor* b) -> bool { chunk_count++; // This predicate guarantees that there is only a single zero-init // (memset) per chunk, and if present it will be at the beginning. bool same = HasByteIndex(a) == HasByteIndex(b) && a->is_repeated() == b->is_repeated() && ShouldSplit(a, options_) == ShouldSplit(b, options_) && (CanInitializeByZeroing(a) == CanInitializeByZeroing(b) || (CanInitializeByZeroing(a) && (chunk_count == 1 || merge_zero_init))); if (!same) chunk_count = 0; return same; }); ColdChunkSkipper cold_skipper(descriptor_, options_, chunks, has_bit_indices_, kColdRatio); int cached_has_word_index = -1; for (int chunk_index = 0; chunk_index < chunks.size(); chunk_index++) { std::vector& chunk = chunks[chunk_index]; cold_skipper.OnStartChunk(chunk_index, cached_has_word_index, "", printer); const FieldDescriptor* memset_start = nullptr; const FieldDescriptor* memset_end = nullptr; bool saw_non_zero_init = false; bool chunk_is_cold = !chunk.empty() && ShouldSplit(chunk.front(), options_); for (const auto& field : chunk) { if (CanInitializeByZeroing(field)) { GOOGLE_CHECK(!saw_non_zero_init); if (!memset_start) memset_start = field; memset_end = field; } else { saw_non_zero_init = true; } } // Whether we wrap this chunk in: // if (cached_has_bits & 1 && (memset_end != chunk.back() || merge_zero_init); if (have_outer_if) { // Emit an if() that will let us skip the whole chunk if none are set. uint32_t chunk_mask = GenChunkMask(chunk, has_bit_indices_); std::string chunk_mask_str = StrCat(strings::Hex(chunk_mask, strings::ZERO_PAD_8)); // Check (up to) 8 has_bits at a time if we have more than one field in // this chunk. Due to field layout ordering, we may check // _has_bits_[last_chunk * 8 / 32] multiple times. GOOGLE_DCHECK_LE(2, popcnt(chunk_mask)); GOOGLE_DCHECK_GE(8, popcnt(chunk_mask)); if (cached_has_word_index != HasWordIndex(chunk.front())) { cached_has_word_index = HasWordIndex(chunk.front()); format("cached_has_bits = $has_bits$[$1$];\n", cached_has_word_index); } format("if (cached_has_bits & 0x$1$u) {\n", chunk_mask_str); format.Indent(); } if (chunk_is_cold) { format("if (!IsSplitMessageDefault()) {\n"); format.Indent(); } if (memset_start) { if (memset_start == memset_end) { // For clarity, do not memset a single field. field_generators_.get(memset_start) .GenerateMessageClearingCode(printer); } else { GOOGLE_CHECK_EQ(chunk_is_cold, ShouldSplit(memset_start, options_)); GOOGLE_CHECK_EQ(chunk_is_cold, ShouldSplit(memset_end, options_)); format( "::memset(&$1$, 0, static_cast(\n" " reinterpret_cast(&$2$) -\n" " reinterpret_cast(&$1$)) + sizeof($2$));\n", FieldMemberName(memset_start, chunk_is_cold), FieldMemberName(memset_end, chunk_is_cold)); } } // Clear all non-zero-initializable fields in the chunk. for (const auto& field : chunk) { if (CanInitializeByZeroing(field)) continue; // It's faster to just overwrite primitive types, but we should only // clear strings and messages if they were set. // // TODO(kenton): Let the CppFieldGenerator decide this somehow. bool have_enclosing_if = HasBitIndex(field) != kNoHasbit && (field->cpp_type() == FieldDescriptor::CPPTYPE_MESSAGE || field->cpp_type() == FieldDescriptor::CPPTYPE_STRING); if (have_enclosing_if) { PrintPresenceCheck(format, field, has_bit_indices_, printer, &cached_has_word_index); } field_generators_.get(field).GenerateMessageClearingCode(printer); if (have_enclosing_if) { format.Outdent(); format("}\n"); } } if (chunk_is_cold) { format.Outdent(); format("}\n"); } if (have_outer_if) { format.Outdent(); format("}\n"); } if (cold_skipper.OnEndChunk(chunk_index, printer)) { // Reset here as it may have been updated in just closed if statement. cached_has_word_index = -1; } } // Step 4: Unions. for (auto oneof : OneOfRange(descriptor_)) { format("clear_$1$();\n", oneof->name()); } if (num_weak_fields_) { format("$weak_field_map$.ClearAll();\n"); } // We don't clear donated status. if (!has_bit_indices_.empty()) { // Step 5: Everything else. format("$has_bits$.Clear();\n"); } std::map vars; SetUnknownFieldsVariable(descriptor_, options_, &vars); format.AddMap(vars); format("_internal_metadata_.Clear<$unknown_fields_type$>();\n"); format.Outdent(); format("}\n"); } void MessageGenerator::GenerateOneofClear(io::Printer* printer) { // Generated function clears the active field and union case (e.g. foo_case_). int i = 0; for (auto oneof : OneOfRange(descriptor_)) { Formatter format(printer, variables_); format.Set("oneofname", oneof->name()); format( "void $classname$::clear_$oneofname$() {\n" "// @@protoc_insertion_point(one_of_clear_start:$full_name$)\n"); format.Indent(); format("switch ($oneofname$_case()) {\n"); format.Indent(); for (auto field : FieldRange(oneof)) { format("case k$1$: {\n", UnderscoresToCamelCase(field->name(), true)); format.Indent(); // We clear only allocated objects in oneofs if (!IsStringOrMessage(field) || IsFieldStripped(field, options_)) { format("// No need to clear\n"); } else { field_generators_.get(field).GenerateClearingCode(printer); } format("break;\n"); format.Outdent(); format("}\n"); } format( "case $1$_NOT_SET: {\n" " break;\n" "}\n", ToUpper(oneof->name())); format.Outdent(); format( "}\n" "$oneof_case$[$1$] = $2$_NOT_SET;\n", i, ToUpper(oneof->name())); format.Outdent(); format( "}\n" "\n"); i++; } } void MessageGenerator::GenerateSwap(io::Printer* printer) { if (HasSimpleBaseClass(descriptor_, options_)) return; Formatter format(printer, variables_); format("void $classname$::InternalSwap($classname$* other) {\n"); format.Indent(); format("using std::swap;\n"); if (HasGeneratedMethods(descriptor_->file(), options_)) { if (descriptor_->extension_range_count() > 0) { format( "$extensions$.InternalSwap(&other->$extensions$);" "\n"); } std::map vars; SetUnknownFieldsVariable(descriptor_, options_, &vars); format.AddMap(vars); if (HasNonSplitOptionalString(descriptor_, options_)) { format( "auto* lhs_arena = GetArenaForAllocation();\n" "auto* rhs_arena = other->GetArenaForAllocation();\n"); } format("_internal_metadata_.InternalSwap(&other->_internal_metadata_);\n"); if (!has_bit_indices_.empty()) { for (int i = 0; i < HasBitsSize(); ++i) { format("swap($has_bits$[$1$], other->$has_bits$[$1$]);\n", i); } } // If possible, we swap several fields at once, including padding. const RunMap runs = FindRuns(optimized_order_, [this](const FieldDescriptor* field) { return !ShouldSplit(field, options_) && CanBeManipulatedAsRawBytes(field, options_, scc_analyzer_); }); for (size_t i = 0; i < optimized_order_.size(); ++i) { const FieldDescriptor* field = optimized_order_[i]; if (ShouldSplit(field, options_)) { continue; } const auto it = runs.find(field); // We only apply the memswap technique to runs of more than one field, as // `swap(field_, other.field_)` is better than // `memswap<...>(&field_, &other.field_)` for generated code readability. if (it != runs.end() && it->second > 1) { // Use a memswap, then skip run_length fields. const size_t run_length = it->second; const std::string first_field_name = FieldMemberName(field, /*cold=*/false); const std::string last_field_name = FieldMemberName( optimized_order_[i + run_length - 1], /*cold=*/false); format.Set("first", first_field_name); format.Set("last", last_field_name); format( "::PROTOBUF_NAMESPACE_ID::internal::memswap<\n" " PROTOBUF_FIELD_OFFSET($classname$, $last$)\n" " + sizeof($classname$::$last$)\n" " - PROTOBUF_FIELD_OFFSET($classname$, $first$)>(\n" " reinterpret_cast(&$first$),\n" " reinterpret_cast(&other->$first$));\n"); i += run_length - 1; // ++i at the top of the loop. } else { field_generators_.get(field).GenerateSwappingCode(printer); } } if (ShouldSplit(descriptor_, options_)) { format("swap($split$, other->$split$);\n"); } for (auto oneof : OneOfRange(descriptor_)) { format("swap(_impl_.$1$_, other->_impl_.$1$_);\n", oneof->name()); } for (int i = 0; i < descriptor_->real_oneof_decl_count(); i++) { format("swap($oneof_case$[$1$], other->$oneof_case$[$1$]);\n", i); } if (num_weak_fields_) { format( "$weak_field_map$.UnsafeArenaSwap(&other->$weak_field_map$)" ";\n"); } if (!inlined_string_indices_.empty()) { for (size_t i = 0; i < InlinedStringDonatedSize(); ++i) { format( "swap($inlined_string_donated_array$[$1$], " "other->$inlined_string_donated_array$[$1$]);\n", i); } } } else { format("GetReflection()->Swap(this, other);"); } format.Outdent(); format("}\n"); } void MessageGenerator::GenerateMergeFrom(io::Printer* printer) { Formatter format(printer, variables_); if (!HasSimpleBaseClass(descriptor_, options_)) { if (HasDescriptorMethods(descriptor_->file(), options_)) { // We don't override the generalized MergeFrom (aka that which // takes in the Message base class as a parameter); instead we just // let the base Message::MergeFrom take care of it. The base MergeFrom // knows how to quickly confirm the types exactly match, and if so, will // use GetClassData() to retrieve the address of MergeImpl, which calls // the fast MergeFrom overload. Most callers avoid all this by passing // a "from" message that is the same type as the message being merged // into, rather than a generic Message. format( "const ::$proto_ns$::Message::ClassData " "$classname$::_class_data_ = {\n" " ::$proto_ns$::Message::CopyWithSourceCheck,\n" " $classname$::MergeImpl\n" "};\n" "const ::$proto_ns$::Message::ClassData*" "$classname$::GetClassData() const { return &_class_data_; }\n" "\n"); } else { // Generate CheckTypeAndMergeFrom(). format( "void $classname$::CheckTypeAndMergeFrom(\n" " const ::$proto_ns$::MessageLite& from) {\n" " MergeFrom(*::_pbi::DownCast(\n" " &from));\n" "}\n"); } } else { // In the simple case, we just define ClassData that vectors back to the // simple implementation of Copy and Merge. format( "const ::$proto_ns$::Message::ClassData " "$classname$::_class_data_ = {\n" " $superclass$::CopyImpl,\n" " $superclass$::MergeImpl,\n" "};\n" "const ::$proto_ns$::Message::ClassData*" "$classname$::GetClassData() const { return &_class_data_; }\n" "\n" "\n"); } } void MessageGenerator::GenerateClassSpecificMergeImpl(io::Printer* printer) { if (HasSimpleBaseClass(descriptor_, options_)) return; // Generate the class-specific MergeFrom, which avoids the GOOGLE_CHECK and cast. Formatter format(printer, variables_); if (!HasDescriptorMethods(descriptor_->file(), options_)) { // For messages that don't inherit from Message, just implement MergeFrom // directly. format( "void $classname$::MergeFrom(const $classname$& from) {\n" " $classname$* const _this = this;\n"); } else { format( "void $classname$::MergeImpl(::$proto_ns$::Message& to_msg, const " "::$proto_ns$::Message& from_msg) {\n" " auto* const _this = static_cast<$classname$*>(&to_msg);\n" " auto& from = static_cast(from_msg);\n"); } format.Indent(); format( "$annotate_mergefrom$" "// @@protoc_insertion_point(class_specific_merge_from_start:" "$full_name$)\n"); format("$DCHK$_NE(&from, _this);\n"); format( "$uint32$ cached_has_bits = 0;\n" "(void) cached_has_bits;\n\n"); if (ShouldSplit(descriptor_, options_)) { format( "if (!from.IsSplitMessageDefault()) {\n" " _this->PrepareSplitMessageForWrite();\n" "}\n"); } std::vector> chunks = CollectFields( optimized_order_, [&](const FieldDescriptor* a, const FieldDescriptor* b) -> bool { return HasByteIndex(a) == HasByteIndex(b) && ShouldSplit(a, options_) == ShouldSplit(b, options_); }); ColdChunkSkipper cold_skipper(descriptor_, options_, chunks, has_bit_indices_, kColdRatio); // cached_has_word_index maintains that: // cached_has_bits = from._has_bits_[cached_has_word_index] // for cached_has_word_index >= 0 int cached_has_word_index = -1; for (int chunk_index = 0; chunk_index < chunks.size(); chunk_index++) { const std::vector& chunk = chunks[chunk_index]; bool have_outer_if = chunk.size() > 1 && HasByteIndex(chunk.front()) != kNoHasbit; cold_skipper.OnStartChunk(chunk_index, cached_has_word_index, "from.", printer); if (have_outer_if) { // Emit an if() that will let us skip the whole chunk if none are set. uint32_t chunk_mask = GenChunkMask(chunk, has_bit_indices_); std::string chunk_mask_str = StrCat(strings::Hex(chunk_mask, strings::ZERO_PAD_8)); // Check (up to) 8 has_bits at a time if we have more than one field in // this chunk. Due to field layout ordering, we may check // _has_bits_[last_chunk * 8 / 32] multiple times. GOOGLE_DCHECK_LE(2, popcnt(chunk_mask)); GOOGLE_DCHECK_GE(8, popcnt(chunk_mask)); if (cached_has_word_index != HasWordIndex(chunk.front())) { cached_has_word_index = HasWordIndex(chunk.front()); format("cached_has_bits = from.$has_bits$[$1$];\n", cached_has_word_index); } format("if (cached_has_bits & 0x$1$u) {\n", chunk_mask_str); format.Indent(); } // Go back and emit merging code for each of the fields we processed. bool deferred_has_bit_changes = false; for (const auto field : chunk) { const FieldGenerator& generator = field_generators_.get(field); if (field->is_repeated()) { generator.GenerateMergingCode(printer); } else if (field->is_optional() && !HasHasbit(field)) { // Merge semantics without true field presence: primitive fields are // merged only if non-zero (numeric) or non-empty (string). bool have_enclosing_if = EmitFieldNonDefaultCondition(printer, "from.", field); generator.GenerateMergingCode(printer); if (have_enclosing_if) { format.Outdent(); format("}\n"); } } else if (field->options().weak() || cached_has_word_index != HasWordIndex(field)) { // Check hasbit, not using cached bits. GOOGLE_CHECK(HasHasbit(field)); format("if (from._internal_has_$1$()) {\n", FieldName(field)); format.Indent(); generator.GenerateMergingCode(printer); format.Outdent(); format("}\n"); } else { // Check hasbit, using cached bits. GOOGLE_CHECK(HasHasbit(field)); int has_bit_index = has_bit_indices_[field->index()]; const std::string mask = StrCat( strings::Hex(1u << (has_bit_index % 32), strings::ZERO_PAD_8)); format("if (cached_has_bits & 0x$1$u) {\n", mask); format.Indent(); if (have_outer_if && IsPOD(field)) { // Defer hasbit modification until the end of chunk. // This can reduce the number of loads/stores by up to 7 per 8 fields. deferred_has_bit_changes = true; generator.GenerateCopyConstructorCode(printer); } else { generator.GenerateMergingCode(printer); } format.Outdent(); format("}\n"); } } if (have_outer_if) { if (deferred_has_bit_changes) { // Flush the has bits for the primitives we deferred. GOOGLE_CHECK_LE(0, cached_has_word_index); format("_this->$has_bits$[$1$] |= cached_has_bits;\n", cached_has_word_index); } format.Outdent(); format("}\n"); } if (cold_skipper.OnEndChunk(chunk_index, printer)) { // Reset here as it may have been updated in just closed if statement. cached_has_word_index = -1; } } // Merge oneof fields. Oneof field requires oneof case check. for (auto oneof : OneOfRange(descriptor_)) { format("switch (from.$1$_case()) {\n", oneof->name()); format.Indent(); for (auto field : FieldRange(oneof)) { format("case k$1$: {\n", UnderscoresToCamelCase(field->name(), true)); format.Indent(); if (!IsFieldStripped(field, options_)) { field_generators_.get(field).GenerateMergingCode(printer); } format("break;\n"); format.Outdent(); format("}\n"); } format( "case $1$_NOT_SET: {\n" " break;\n" "}\n", ToUpper(oneof->name())); format.Outdent(); format("}\n"); } if (num_weak_fields_) { format( "_this->$weak_field_map$.MergeFrom(from.$weak_field_map$);" "\n"); } // Merging of extensions and unknown fields is done last, to maximize // the opportunity for tail calls. if (descriptor_->extension_range_count() > 0) { format( "_this->$extensions$.MergeFrom(internal_default_instance(), " "from.$extensions$);\n"); } format( "_this->_internal_metadata_.MergeFrom<$unknown_fields_type$>(from._" "internal_" "metadata_);\n"); format.Outdent(); format("}\n"); } void MessageGenerator::GenerateCopyFrom(io::Printer* printer) { if (HasSimpleBaseClass(descriptor_, options_)) return; Formatter format(printer, variables_); if (HasDescriptorMethods(descriptor_->file(), options_)) { // We don't override the generalized CopyFrom (aka that which // takes in the Message base class as a parameter); instead we just // let the base Message::CopyFrom take care of it. The base MergeFrom // knows how to quickly confirm the types exactly match, and if so, will // use GetClassData() to get the address of Message::CopyWithSourceCheck, // which calls Clear() and then MergeFrom(), as well as making sure that // clearing the destination message doesn't alter the source, when in debug // builds. Most callers avoid this by passing a "from" message that is the // same type as the message being merged into, rather than a generic // Message. } // Generate the class-specific CopyFrom. format( "void $classname$::CopyFrom(const $classname$& from) {\n" "// @@protoc_insertion_point(class_specific_copy_from_start:" "$full_name$)\n"); format.Indent(); format("if (&from == this) return;\n"); if (!options_.opensource_runtime && HasMessageFieldOrExtension(descriptor_)) { // This check is disabled in the opensource release because we're // concerned that many users do not define NDEBUG in their release builds. // It is also disabled if a message has neither message fields nor // extensions, as it's impossible to copy from its descendant. // // Note that FailIfCopyFromDescendant is implemented by reflection and not // available for lite runtime. In that case, check if the size of the source // has changed after Clear. format("#ifndef NDEBUG\n"); if (HasDescriptorMethods(descriptor_->file(), options_)) { format("FailIfCopyFromDescendant(*this, from);\n"); } else { format("size_t from_size = from.ByteSizeLong();\n"); } format( "#endif\n" "Clear();\n"); if (!HasDescriptorMethods(descriptor_->file(), options_)) { format( "#ifndef NDEBUG\n" "$CHK$_EQ(from_size, from.ByteSizeLong())\n" " << \"Source of CopyFrom changed when clearing target. Either \"\n" " \"source is a nested message in target (not allowed), or \"\n" " \"another thread is modifying the source.\";\n" "#endif\n"); } } else { format("Clear();\n"); } format("MergeFrom(from);\n"); format.Outdent(); format("}\n"); } void MessageGenerator::GenerateVerify(io::Printer* printer) { } void MessageGenerator::GenerateSerializeOneofFields( io::Printer* printer, const std::vector& fields) { Formatter format(printer, variables_); GOOGLE_CHECK(!fields.empty()); if (fields.size() == 1) { GenerateSerializeOneField(printer, fields[0], -1); return; } // We have multiple mutually exclusive choices. Emit a switch statement. const OneofDescriptor* oneof = fields[0]->containing_oneof(); format("switch ($1$_case()) {\n", oneof->name()); format.Indent(); for (auto field : fields) { format("case k$1$: {\n", UnderscoresToCamelCase(field->name(), true)); format.Indent(); field_generators_.get(field).GenerateSerializeWithCachedSizesToArray( printer); format("break;\n"); format.Outdent(); format("}\n"); } format.Outdent(); // Doing nothing is an option. format( " default: ;\n" "}\n"); } void MessageGenerator::GenerateSerializeOneField(io::Printer* printer, const FieldDescriptor* field, int cached_has_bits_index) { Formatter format(printer, variables_); if (!field->options().weak()) { // For weakfields, PrintFieldComment is called during iteration. PrintFieldComment(format, field); } bool have_enclosing_if = false; if (field->options().weak()) { } else if (HasHasbit(field)) { // Attempt to use the state of cached_has_bits, if possible. int has_bit_index = HasBitIndex(field); if (cached_has_bits_index == has_bit_index / 32) { const std::string mask = StrCat(strings::Hex(1u << (has_bit_index % 32), strings::ZERO_PAD_8)); format("if (cached_has_bits & 0x$1$u) {\n", mask); } else { format("if (_internal_has_$1$()) {\n", FieldName(field)); } format.Indent(); have_enclosing_if = true; } else if (field->is_optional() && !HasHasbit(field)) { have_enclosing_if = EmitFieldNonDefaultCondition(printer, "this->", field); } field_generators_.get(field).GenerateSerializeWithCachedSizesToArray(printer); if (have_enclosing_if) { format.Outdent(); format("}\n"); } format("\n"); } void MessageGenerator::GenerateSerializeOneExtensionRange( io::Printer* printer, const Descriptor::ExtensionRange* range) { std::map vars = variables_; vars["start"] = StrCat(range->start); vars["end"] = StrCat(range->end); Formatter format(printer, vars); format("// Extension range [$start$, $end$)\n"); format( "target = $extensions$._InternalSerialize(\n" "internal_default_instance(), $start$, $end$, target, stream);\n\n"); } void MessageGenerator::GenerateSerializeWithCachedSizesToArray( io::Printer* printer) { if (HasSimpleBaseClass(descriptor_, options_)) return; Formatter format(printer, variables_); if (descriptor_->options().message_set_wire_format()) { // Special-case MessageSet. format( "$uint8$* $classname$::_InternalSerialize(\n" " $uint8$* target, ::$proto_ns$::io::EpsCopyOutputStream* stream) " "const {\n" "$annotate_serialize$" " target = $extensions$." "InternalSerializeMessageSetWithCachedSizesToArray(\n" // "internal_default_instance(), target, stream);\n"); std::map vars; SetUnknownFieldsVariable(descriptor_, options_, &vars); format.AddMap(vars); format( " target = ::_pbi::" "InternalSerializeUnknownMessageSetItemsToArray(\n" " $unknown_fields$, target, stream);\n"); format( " return target;\n" "}\n"); return; } format( "$uint8$* $classname$::_InternalSerialize(\n" " $uint8$* target, ::$proto_ns$::io::EpsCopyOutputStream* stream) " "const {\n" "$annotate_serialize$"); format.Indent(); format("// @@protoc_insertion_point(serialize_to_array_start:$full_name$)\n"); if (!ShouldSerializeInOrder(descriptor_, options_)) { format.Outdent(); format("#ifdef NDEBUG\n"); format.Indent(); } GenerateSerializeWithCachedSizesBody(printer); if (!ShouldSerializeInOrder(descriptor_, options_)) { format.Outdent(); format("#else // NDEBUG\n"); format.Indent(); GenerateSerializeWithCachedSizesBodyShuffled(printer); format.Outdent(); format("#endif // !NDEBUG\n"); format.Indent(); } format("// @@protoc_insertion_point(serialize_to_array_end:$full_name$)\n"); format.Outdent(); format( " return target;\n" "}\n"); } void MessageGenerator::GenerateSerializeWithCachedSizesBody( io::Printer* printer) { if (HasSimpleBaseClass(descriptor_, options_)) return; Formatter format(printer, variables_); // If there are multiple fields in a row from the same oneof then we // coalesce them and emit a switch statement. This is more efficient // because it lets the C++ compiler know this is a "at most one can happen" // situation. If we emitted "if (has_x()) ...; if (has_y()) ..." the C++ // compiler's emitted code might check has_y() even when has_x() is true. class LazySerializerEmitter { public: LazySerializerEmitter(MessageGenerator* mg, io::Printer* printer) : mg_(mg), format_(printer), eager_(IsProto3(mg->descriptor_->file())), cached_has_bit_index_(kNoHasbit) {} ~LazySerializerEmitter() { Flush(); } // If conditions allow, try to accumulate a run of fields from the same // oneof, and handle them at the next Flush(). void Emit(const FieldDescriptor* field) { if (eager_ || MustFlush(field)) { Flush(); } if (!field->real_containing_oneof()) { // TODO(ckennelly): Defer non-oneof fields similarly to oneof fields. if (!field->options().weak() && !field->is_repeated() && !eager_) { // We speculatively load the entire _has_bits_[index] contents, even // if it is for only one field. Deferring non-oneof emitting would // allow us to determine whether this is going to be useful. int has_bit_index = mg_->has_bit_indices_[field->index()]; if (cached_has_bit_index_ != has_bit_index / 32) { // Reload. int new_index = has_bit_index / 32; format_("cached_has_bits = _impl_._has_bits_[$1$];\n", new_index); cached_has_bit_index_ = new_index; } } mg_->GenerateSerializeOneField(format_.printer(), field, cached_has_bit_index_); } else { v_.push_back(field); } } void EmitIfNotNull(const FieldDescriptor* field) { if (field != nullptr) { Emit(field); } } void Flush() { if (!v_.empty()) { mg_->GenerateSerializeOneofFields(format_.printer(), v_); v_.clear(); } } private: // If we have multiple fields in v_ then they all must be from the same // oneof. Would adding field to v_ break that invariant? bool MustFlush(const FieldDescriptor* field) { return !v_.empty() && v_[0]->containing_oneof() != field->containing_oneof(); } MessageGenerator* mg_; Formatter format_; const bool eager_; std::vector v_; // cached_has_bit_index_ maintains that: // cached_has_bits = from._has_bits_[cached_has_bit_index_] // for cached_has_bit_index_ >= 0 int cached_has_bit_index_; }; class LazyExtensionRangeEmitter { public: LazyExtensionRangeEmitter(MessageGenerator* mg, io::Printer* printer) : mg_(mg), format_(printer) {} void AddToRange(const Descriptor::ExtensionRange* range) { if (!has_current_range_) { current_combined_range_ = *range; has_current_range_ = true; } else { current_combined_range_.start = std::min(current_combined_range_.start, range->start); current_combined_range_.end = std::max(current_combined_range_.end, range->end); } } void Flush() { if (has_current_range_) { mg_->GenerateSerializeOneExtensionRange(format_.printer(), ¤t_combined_range_); } has_current_range_ = false; } private: MessageGenerator* mg_; Formatter format_; bool has_current_range_ = false; Descriptor::ExtensionRange current_combined_range_; }; // We need to track the largest weak field, because weak fields are serialized // differently than normal fields. The WeakFieldMap::FieldWriter will // serialize all weak fields that are ordinally between the last serialized // weak field and the current field. In order to guarantee that all weak // fields are serialized, we need to make sure to emit the code to serialize // the largest weak field present at some point. class LargestWeakFieldHolder { public: const FieldDescriptor* Release() { const FieldDescriptor* result = field_; field_ = nullptr; return result; } void ReplaceIfLarger(const FieldDescriptor* field) { if (field_ == nullptr || field_->number() < field->number()) { field_ = field; } } private: const FieldDescriptor* field_ = nullptr; }; std::vector ordered_fields = SortFieldsByNumber(descriptor_); std::vector sorted_extensions; sorted_extensions.reserve(descriptor_->extension_range_count()); for (int i = 0; i < descriptor_->extension_range_count(); ++i) { sorted_extensions.push_back(descriptor_->extension_range(i)); } std::sort(sorted_extensions.begin(), sorted_extensions.end(), ExtensionRangeSorter()); if (num_weak_fields_) { format( "::_pbi::WeakFieldMap::FieldWriter field_writer(" "$weak_field_map$);\n"); } format( "$uint32$ cached_has_bits = 0;\n" "(void) cached_has_bits;\n\n"); // Merge the fields and the extension ranges, both sorted by field number. { LazySerializerEmitter e(this, printer); LazyExtensionRangeEmitter re(this, printer); LargestWeakFieldHolder largest_weak_field; int i, j; for (i = 0, j = 0; i < ordered_fields.size() || j < sorted_extensions.size();) { if ((j == sorted_extensions.size()) || (i < descriptor_->field_count() && ordered_fields[i]->number() < sorted_extensions[j]->start)) { const FieldDescriptor* field = ordered_fields[i++]; if (IsFieldStripped(field, options_)) { continue; } re.Flush(); if (field->options().weak()) { largest_weak_field.ReplaceIfLarger(field); PrintFieldComment(format, field); } else { e.EmitIfNotNull(largest_weak_field.Release()); e.Emit(field); } } else { e.EmitIfNotNull(largest_weak_field.Release()); e.Flush(); re.AddToRange(sorted_extensions[j++]); } } re.Flush(); e.EmitIfNotNull(largest_weak_field.Release()); } std::map vars; SetUnknownFieldsVariable(descriptor_, options_, &vars); format.AddMap(vars); format("if (PROTOBUF_PREDICT_FALSE($have_unknown_fields$)) {\n"); format.Indent(); if (UseUnknownFieldSet(descriptor_->file(), options_)) { format( "target = " "::_pbi::WireFormat::" "InternalSerializeUnknownFieldsToArray(\n" " $unknown_fields$, target, stream);\n"); } else { format( "target = stream->WriteRaw($unknown_fields$.data(),\n" " static_cast($unknown_fields$.size()), target);\n"); } format.Outdent(); format("}\n"); } void MessageGenerator::GenerateSerializeWithCachedSizesBodyShuffled( io::Printer* printer) { Formatter format(printer, variables_); std::vector ordered_fields = SortFieldsByNumber(descriptor_); ordered_fields.erase( std::remove_if(ordered_fields.begin(), ordered_fields.end(), [this](const FieldDescriptor* f) { return !IsFieldUsed(f, options_); }), ordered_fields.end()); std::vector sorted_extensions; sorted_extensions.reserve(descriptor_->extension_range_count()); for (int i = 0; i < descriptor_->extension_range_count(); ++i) { sorted_extensions.push_back(descriptor_->extension_range(i)); } std::sort(sorted_extensions.begin(), sorted_extensions.end(), ExtensionRangeSorter()); int num_fields = ordered_fields.size() + sorted_extensions.size(); constexpr int kLargePrime = 1000003; GOOGLE_CHECK_LT(num_fields, kLargePrime) << "Prime offset must be greater than the number of fields to ensure " "those are coprime."; if (num_weak_fields_) { format( "::_pbi::WeakFieldMap::FieldWriter field_writer(" "$weak_field_map$);\n"); } format("for (int i = $1$; i >= 0; i-- ) {\n", num_fields - 1); format.Indent(); format("switch(i) {\n"); format.Indent(); int index = 0; for (const auto* f : ordered_fields) { format("case $1$: {\n", index++); format.Indent(); GenerateSerializeOneField(printer, f, -1); format("break;\n"); format.Outdent(); format("}\n"); } for (const auto* r : sorted_extensions) { format("case $1$: {\n", index++); format.Indent(); GenerateSerializeOneExtensionRange(printer, r); format("break;\n"); format.Outdent(); format("}\n"); } format( "default: {\n" " $DCHK$(false) << \"Unexpected index: \" << i;\n" "}\n"); format.Outdent(); format("}\n"); format.Outdent(); format("}\n"); std::map vars; SetUnknownFieldsVariable(descriptor_, options_, &vars); format.AddMap(vars); format("if (PROTOBUF_PREDICT_FALSE($have_unknown_fields$)) {\n"); format.Indent(); if (UseUnknownFieldSet(descriptor_->file(), options_)) { format( "target = " "::_pbi::WireFormat::" "InternalSerializeUnknownFieldsToArray(\n" " $unknown_fields$, target, stream);\n"); } else { format( "target = stream->WriteRaw($unknown_fields$.data(),\n" " static_cast($unknown_fields$.size()), target);\n"); } format.Outdent(); format("}\n"); } std::vector MessageGenerator::RequiredFieldsBitMask() const { const int array_size = HasBitsSize(); std::vector masks(array_size, 0); for (auto field : FieldRange(descriptor_)) { if (!field->is_required()) { continue; } const int has_bit_index = has_bit_indices_[field->index()]; masks[has_bit_index / 32] |= static_cast(1) << (has_bit_index % 32); } return masks; } void MessageGenerator::GenerateByteSize(io::Printer* printer) { if (HasSimpleBaseClass(descriptor_, options_)) return; Formatter format(printer, variables_); if (descriptor_->options().message_set_wire_format()) { // Special-case MessageSet. std::map vars; SetUnknownFieldsVariable(descriptor_, options_, &vars); format.AddMap(vars); format( "size_t $classname$::ByteSizeLong() const {\n" "$annotate_bytesize$" "// @@protoc_insertion_point(message_set_byte_size_start:$full_name$)\n" " size_t total_size = $extensions$.MessageSetByteSize();\n" " if ($have_unknown_fields$) {\n" " total_size += ::_pbi::\n" " ComputeUnknownMessageSetItemsSize($unknown_fields$);\n" " }\n" " int cached_size = " "::_pbi::ToCachedSize(total_size);\n" " SetCachedSize(cached_size);\n" " return total_size;\n" "}\n"); return; } if (num_required_fields_ > 1) { // Emit a function (rarely used, we hope) that handles the required fields // by checking for each one individually. format( "size_t $classname$::RequiredFieldsByteSizeFallback() const {\n" "// @@protoc_insertion_point(required_fields_byte_size_fallback_start:" "$full_name$)\n"); format.Indent(); format("size_t total_size = 0;\n"); for (auto field : optimized_order_) { if (field->is_required()) { format( "\n" "if (_internal_has_$1$()) {\n", FieldName(field)); format.Indent(); PrintFieldComment(format, field); field_generators_.get(field).GenerateByteSize(printer); format.Outdent(); format("}\n"); } } format( "\n" "return total_size;\n"); format.Outdent(); format("}\n"); } format( "size_t $classname$::ByteSizeLong() const {\n" "$annotate_bytesize$" "// @@protoc_insertion_point(message_byte_size_start:$full_name$)\n"); format.Indent(); format( "size_t total_size = 0;\n" "\n"); if (descriptor_->extension_range_count() > 0) { format( "total_size += $extensions$.ByteSize();\n" "\n"); } std::map vars; SetUnknownFieldsVariable(descriptor_, options_, &vars); format.AddMap(vars); // Handle required fields (if any). We expect all of them to be // present, so emit one conditional that checks for that. If they are all // present then the fast path executes; otherwise the slow path executes. if (num_required_fields_ > 1) { // The fast path works if all required fields are present. const std::vector masks_for_has_bits = RequiredFieldsBitMask(); format("if ($1$) { // All required fields are present.\n", ConditionalToCheckBitmasks(masks_for_has_bits)); format.Indent(); // Oneof fields cannot be required, so optimized_order_ contains all of the // fields that we need to potentially emit. for (auto field : optimized_order_) { if (!field->is_required()) continue; PrintFieldComment(format, field); field_generators_.get(field).GenerateByteSize(printer); format("\n"); } format.Outdent(); format( "} else {\n" // the slow path " total_size += RequiredFieldsByteSizeFallback();\n" "}\n"); } else { // num_required_fields_ <= 1: no need to be tricky for (auto field : optimized_order_) { if (!field->is_required()) continue; PrintFieldComment(format, field); format("if (_internal_has_$1$()) {\n", FieldName(field)); format.Indent(); field_generators_.get(field).GenerateByteSize(printer); format.Outdent(); format("}\n"); } } std::vector> chunks = CollectFields( optimized_order_, [&](const FieldDescriptor* a, const FieldDescriptor* b) -> bool { return a->label() == b->label() && HasByteIndex(a) == HasByteIndex(b) && ShouldSplit(a, options_) == ShouldSplit(b, options_); }); // Remove chunks with required fields. chunks.erase(std::remove_if(chunks.begin(), chunks.end(), IsRequired), chunks.end()); ColdChunkSkipper cold_skipper(descriptor_, options_, chunks, has_bit_indices_, kColdRatio); int cached_has_word_index = -1; format( "$uint32$ cached_has_bits = 0;\n" "// Prevent compiler warnings about cached_has_bits being unused\n" "(void) cached_has_bits;\n\n"); for (int chunk_index = 0; chunk_index < chunks.size(); chunk_index++) { const std::vector& chunk = chunks[chunk_index]; const bool have_outer_if = chunk.size() > 1 && HasWordIndex(chunk[0]) != kNoHasbit; cold_skipper.OnStartChunk(chunk_index, cached_has_word_index, "", printer); if (have_outer_if) { // Emit an if() that will let us skip the whole chunk if none are set. uint32_t chunk_mask = GenChunkMask(chunk, has_bit_indices_); std::string chunk_mask_str = StrCat(strings::Hex(chunk_mask, strings::ZERO_PAD_8)); // Check (up to) 8 has_bits at a time if we have more than one field in // this chunk. Due to field layout ordering, we may check // _has_bits_[last_chunk * 8 / 32] multiple times. GOOGLE_DCHECK_LE(2, popcnt(chunk_mask)); GOOGLE_DCHECK_GE(8, popcnt(chunk_mask)); if (cached_has_word_index != HasWordIndex(chunk.front())) { cached_has_word_index = HasWordIndex(chunk.front()); format("cached_has_bits = $has_bits$[$1$];\n", cached_has_word_index); } format("if (cached_has_bits & 0x$1$u) {\n", chunk_mask_str); format.Indent(); } // Go back and emit checks for each of the fields we processed. for (int j = 0; j < chunk.size(); j++) { const FieldDescriptor* field = chunk[j]; const FieldGenerator& generator = field_generators_.get(field); bool have_enclosing_if = false; bool need_extra_newline = false; PrintFieldComment(format, field); if (field->is_repeated()) { // No presence check is required. need_extra_newline = true; } else if (HasHasbit(field)) { PrintPresenceCheck(format, field, has_bit_indices_, printer, &cached_has_word_index); have_enclosing_if = true; } else { // Without field presence: field is serialized only if it has a // non-default value. have_enclosing_if = EmitFieldNonDefaultCondition(printer, "this->", field); } generator.GenerateByteSize(printer); if (have_enclosing_if) { format.Outdent(); format( "}\n" "\n"); } if (need_extra_newline) { format("\n"); } } if (have_outer_if) { format.Outdent(); format("}\n"); } if (cold_skipper.OnEndChunk(chunk_index, printer)) { // Reset here as it may have been updated in just closed if statement. cached_has_word_index = -1; } } // Fields inside a oneof don't use _has_bits_ so we count them in a separate // pass. for (auto oneof : OneOfRange(descriptor_)) { format("switch ($1$_case()) {\n", oneof->name()); format.Indent(); for (auto field : FieldRange(oneof)) { PrintFieldComment(format, field); format("case k$1$: {\n", UnderscoresToCamelCase(field->name(), true)); format.Indent(); if (!IsFieldStripped(field, options_)) { field_generators_.get(field).GenerateByteSize(printer); } format("break;\n"); format.Outdent(); format("}\n"); } format( "case $1$_NOT_SET: {\n" " break;\n" "}\n", ToUpper(oneof->name())); format.Outdent(); format("}\n"); } if (num_weak_fields_) { // TagSize + MessageSize format("total_size += $weak_field_map$.ByteSizeLong();\n"); } if (UseUnknownFieldSet(descriptor_->file(), options_)) { // We go out of our way to put the computation of the uncommon path of // unknown fields in tail position. This allows for better code generation // of this function for simple protos. format( "return MaybeComputeUnknownFieldsSize(total_size, &$cached_size$);\n"); } else { format("if (PROTOBUF_PREDICT_FALSE($have_unknown_fields$)) {\n"); format(" total_size += $unknown_fields$.size();\n"); format("}\n"); // We update _cached_size_ even though this is a const method. Because // const methods might be called concurrently this needs to be atomic // operations or the program is undefined. In practice, since any // concurrent writes will be writing the exact same value, normal writes // will work on all common processors. We use a dedicated wrapper class to // abstract away the underlying atomic. This makes it easier on platforms // where even relaxed memory order might have perf impact to replace it with // ordinary loads and stores. format( "int cached_size = ::_pbi::ToCachedSize(total_size);\n" "SetCachedSize(cached_size);\n" "return total_size;\n"); } format.Outdent(); format("}\n"); } void MessageGenerator::GenerateIsInitialized(io::Printer* printer) { if (HasSimpleBaseClass(descriptor_, options_)) return; Formatter format(printer, variables_); format("bool $classname$::IsInitialized() const {\n"); format.Indent(); if (descriptor_->extension_range_count() > 0) { format( "if (!$extensions$.IsInitialized()) {\n" " return false;\n" "}\n\n"); } if (num_required_fields_ > 0) { format( "if (_Internal::MissingRequiredFields($has_bits$))" " return false;\n"); } // Now check that all non-oneof embedded messages are initialized. for (auto field : optimized_order_) { field_generators_.get(field).GenerateIsInitialized(printer); } if (num_weak_fields_) { // For Weak fields. format("if (!$weak_field_map$.IsInitialized()) return false;\n"); } // Go through the oneof fields, emitting a switch if any might have required // fields. for (auto oneof : OneOfRange(descriptor_)) { bool has_required_fields = false; for (auto field : FieldRange(oneof)) { if (field->cpp_type() == FieldDescriptor::CPPTYPE_MESSAGE && !ShouldIgnoreRequiredFieldCheck(field, options_) && scc_analyzer_->HasRequiredFields(field->message_type())) { has_required_fields = true; break; } } if (!has_required_fields) { continue; } format("switch ($1$_case()) {\n", oneof->name()); format.Indent(); for (auto field : FieldRange(oneof)) { format("case k$1$: {\n", UnderscoresToCamelCase(field->name(), true)); format.Indent(); if (!IsFieldStripped(field, options_)) { field_generators_.get(field).GenerateIsInitialized(printer); } format("break;\n"); format.Outdent(); format("}\n"); } format( "case $1$_NOT_SET: {\n" " break;\n" "}\n", ToUpper(oneof->name())); format.Outdent(); format("}\n"); } format.Outdent(); format( " return true;\n" "}\n"); } } // namespace cpp } // namespace compiler } // namespace protobuf } // namespace google #include