The new log system adds capability to produce structured logs in the Relaxed Extended JSON 2.0.0 format. The new API requires names to be given to variables, forming field names for the variables in structured JSON logs. Named variables are called attributes in the log system. Human readable log messages are built with a libfmt inspired API, where attributes are inserted using replacement fields instead of being streamed together using the streaming operator <<
.
Log lines are composed primarily of a message (msg
) and attributes (attr.*
fields).
As you write log messages, keep the following in mind: A big thing that makes (J|B)SON unique as a data format is the ability to provide rich field names.
What makes logv2 machine readable is that we write an intact Extended BSON format.
But, what makes these lines human readable is that the msg
provides a simple, clear context for interpreting well-formed field names and values in the attr
subdocument.
For maximum readability, a log message additionally has the least amount of repetition possible, and shares attribute names with other related log lines.
The msg
field predicates a reader's interpretation of the log line. It should be crafted with care and attention.
- Concisely describe what the log line is reporting, providing enough context necessary for interpreting attribute field names and values
- Avoid unnecessary punctuation and do not conclude with punctuation
- For new log messages, do not use a format string/substitution for new log messages
- For updating existing log messages, provide both a format string/substitution, and a substitution-free message string
The attr
subdocument includes important metrics/statistics about the logged event for the purposes of debugging or performance analysis. These variables should be named very well, as though intended for a very human-readable portion of the codebase (like config variable declaration, abstract class definitions, etc.)
For attr
fields, do the following:
The bar for understanding should be:
- Someone with reasonable understanding of mongod behavior should understand immediately what is being logged
- Someone with reasonable troubleshooting skill should be able to extract doc- or code-searchable phrases to learn about what is being logged
Exception: Do not add a unit suffix when logging a Duration type. The system automatically adds this unit.
To describe the execution time of an operation using our preferred method: Specify an attr
name of “duration” and provide a value using the Milliseconds Duration type. The log system will automatically append "Millis" to the attribute name.
Alternatively, specify an attr
name of “durationMillis” and provide the number of milliseconds as an integer type.
Importantly: downstream analysis tools will rely on this convention, as a replacement for the "[0-9]+ms$" format of prior logs.
When logging the below information, do so with these specific terms:
- namespace - instead of "ns"
- db - instead of "database"
- error - when an error occurs, instead of "status". Use this for objects of type Status and DBException
- reason - to provide rationale for an event/action when "error" isn't appropriate
For new log lines:
LOGV2(1041, "Transition to PRIMARY complete");
{ ... , "id": 1041, "msg": "Transition to PRIMARY complete", "attr": {} }
LOGV2(1042, "Slow query", "duration"_attr = getDurationMillis());
{ ..., "id": 1042, "msg": "Slow query", "attr": { "durationMillis": 1000 } }
For updating existing log lines:
LOGV2(1040,
"Replica set state transition from {oldState} to {newState} on this node",
"Replica set state transition on this node",
"oldState"_attr = getOldState(), "newState"_attr = getNewState());
{ ..., "id": 1040, "msg": "Replica set state transition on this node", "attr": { "oldState": "SECONARY", "newState": "PRIMARY" } }
The log system is made available with the following header:
#include "mongo/logv2/log.h"
To be able to include it a default log component needs to be defined in the cpp file before including log.h
:
#define MONGO_LOGV2_DEFAULT_COMPONENT ::mongo::logv2::LogComponent::kDefault
Logging is performed using function style macros:
LOGV2(ID, format-string, "name0"_attr = var0, ..., "nameN"_attr = varN);
LOGV2(ID, format-string, message-string, "name0"_attr = var0, ..., "nameN"_attr = varN);
The ID is a signed 32bit integer in the same number space as the error code numbers. It is used to uniquely identify a log statement. If changing existing code, using a new ID is strongly advised to avoid any parsing ambiguity.
The format string contains the description of the log event with libfmt style replacement fields optionally embedded within it. The format string must comply with the format syntax from libfmt. The purpose of embedding the replacement fields is to be able to create a human readable message used by the text output format or a tool that converts JSON logs to a human readable format.
Replacement fields are placed in the format string with curly braces {}
. Everything not surrounded with curly braces is part of the message text. Curly brace characters can be output by escaping them using double braces: {{
or }}
.
Attributes are created with the _attr
user-defined literal. The intermediate object that gets instantiated provides the assignment operator =
for assigning a value to the attribute.
Attributes are associated with replacement fields in the format string by name or index, using names is strongly recommended. When using unnamed replacement fields, attributes map to replacement fields in the order they appear in the format string.
It is allowed to have more attributes than replacement fields in a log statement. However, having fewer attributes than replacement fields is not allowed.
As shown above there is also an API taking both a format string and a message string. This is an API to help with the transition from text output to JSON output. JSON logs have no need for embedded replacement fields in the description, if written in a short and descriptive manner providing context for the attribute names. But a format string may still be needed to provide good JSON to human readable text conversion. See the JSON output format and style guide below for more information.
Both the format string and the message string must be compile time constants. This is to avoid dynamic attribute names in the log output and to be able to add compile time verification of log statements in the future. If the string needs to be shared with anything else (like constructing a Status object) you can use this pattern:
static constexpr char str[] = "the string";
LOGV2(1000, "Logging event, no replacement fields is OK");
const BSONObj& slowOperation = ...;
Milliseconds time = ...;
LOGV2(1001, "Operation {op} is slow, took: {duration}", "op"_attr = slowOperation, "duration"_attr = time);
LOGV2(1002, "Replication state change", "from"_attr = getOldState(), "to"_attr = getNewState());
To override the default component, a separate logging API can be used that takes a LogOptions
structure:
LOGV2_OPTIONS(options, message-string, attr0, ..., attrN);
LogOptions
can be constructed with a LogComponent
to avoid verbosity in the log statement.
LOGV2_OPTIONS(1003, {LogComponent::kCommand}, "Log event to specified component");
LOGV2
is the logging macro for the default informational (0) severity. To log to different severities there are separate logging macros to be used, they all take paramaters like LOGV2
:
LOGV2_WARNING
LOGV2_ERROR
LOGV2_FATAL
LOGV2_FATAL_NOTRACE
LOGV2_FATAL_CONTINUE
There is also variations that take LogOptions
if needed:
LOGV2_WARNING_OPTIONS
LOGV2_ERROR_OPTIONS
LOGV2_FATAL_OPTIONS
Fatal level log statements using LOGV2_FATAL
perform fassert
after logging, using the provided ID as assert id. LOGV2_FATAL_NOTRACE
perform fassertNoTrace
and LOGV2_FATAL_CONTINUE
does not fassert
allowing for continued execution. LOGV2_FATAL_CONTINUE
is meant to be used when a fatal error has occured but a different way of halting execution is desired such as std::terminate
or fassertFailedWithStatus
.
LOGV2_FATAL_OPTIONS
performs fassert
by default like LOGV2_FATAL
but this can be changed by setting the FatalMode
on the LogOptions
.
Debug-level logging is slightly different where an additional parameter (as integer) required to indicate the desired debug level:
LOGV2_DEBUG(ID, debug-level, format-string, attr0, ..., attrN);
LOGV2_DEBUG(ID, debug-level, format-string, message-string, attr0, ..., attrN);
LOGV2_DEBUG_OPTIONS(ID, debug-level, options, format-string, attr0, ..., attrN);
LOGV2_DEBUG_OPTIONS(ID, debug-level, options, format-string, message-string, attr0, ..., attrN);
Status status = ...;
int remainingAttempts = ...;
LOGV2_ERROR(1004, "Initial sync failed. {remaining} attempts left. Reason: {reason}", "reason"_attr = status, "remaining"_attr = remainingAttempts);
Log tags are replacing the Tee from the old log system as the way to indicate that the log should also be written to a RamLog
(accessible with the getLog
command).
Tags are added to a log statement with the options API similarly to how non-default components are specified by constructing a LogOptions
.
Multiple tags can be attached to a log statement using the bitwise or operator |
.
LOGV2_WARNING_OPTIONS(1005, {LogTag::kStartupWarnings}, "XFS filesystem is recommended with WiredTiger");
Sometimes there is a need to add attributes depending on runtime conditionals. To support this there is the DynamicAttributes
class that has an add
method to add named attributes one by one. This class is meant to be used when you have this specific requirement and is not the general logging API.
When finished, it is logged using the regular logging API but the DynamicAttributes
instance is passed as the first attribute parameter. Mixing _attr
literals with the DynamicAttributes
is not supported.
When using the DynamicAttributes
you need to be careful about parameter lifetimes. The DynamicAttributes
binds attributes by reference and the reference must be valid when passing the DynamicAttributes
to the log statement.
DynamicAttributes attrs;
attrs.add("str", "StringData value"_sd);
if (condition) {
// getExtraInfo() returns a reference that is valid until the LOGV2 call below.
// Be careful of functions returning by value
attrs.add("extra", getExtraInfo());
}
LOGV2(1030, "dynamic attributes", attrs);
Many basic types have built in support:
- Boolean
- Integral types
- Single char is logged as integer
- Enums
- Logged as their underlying integral type
- Floating point types
- long double is prohibited
- String types
- std::string
- StringData
- const char*
- Duration types
- Special formatting, see below
- BSON types
- BSONObj
- BSONArray
- BSONElement
- BSON appendable types
BSONObjBuilder::append
overload available
- boost::optional of any loggable type
To make a user defined type loggable it needs a serialization member function that the log system can bind to.
The system binds and uses serialization functions by looking for functions in the following priority order:
Member functions:
void serialize(BSONObjBuilder*) const
BSONObj toBSON() const
BSONArray toBSONArray() const
Non-member functions:
toBSON(const T& val)
(non-member function)
Member functions:
void serialize(fmt::memory_buffer&) const
std::string toString() const
Non-member functions:
toString(const T& val)
(non-member function)
Enums will only try to bind a toString(const T& val)
non-member function. If one is not available the enum value will be logged as its underlying integral type.
In order to offer structured serialization and output, a type would need to supply a structured serialization function (functions 1 to 4 above), otherwise if only stringification is provided the output will be an escaped string.
NOTE: No operator<<
overload is used even if available
class UserDefinedType {
public:
void serialize(BSONObjBuilder* builder) const {
builder->append("str"_sd, _str);
builder->append("int"_sd, _int);
}
private:
std::string _str;
int32_t _int;
};
STL containers and data structures that have STL like interfaces are loggable as long as they contain loggable elements (built-in, user-defined or other containers).
Sequential containers like std::vector
, std::deque
and std::list
are loggable and the elements get formatted as JSON array in structured output.
Associative containers such as std::map
and stdx::unordered_map
loggable with the requirement that they key is of a string type. The structured format is a JSON object where the field names are the key.
Ranges is loggable via helpers to indicate what type of range it is
seqLog(begin, end)
mapLog(begin, end)
seqLog indicates that it is a sequential range where the iterators point to loggable value directly.
mapLog indicates that it is a range coming from an associative container where the iterators point to a key-value pair.
std::array<int, 20> arrayOfInts = ...;
LOGV2(1010, "log container directly: {values}", "values"_attr = arrayOfInts);
LOGV2(1011, "log iterator range: {values}", "values"_attr = seqLog(arrayOfInts.begin(), arrayOfInts.end());
LOGV2(1012, "log first five elements: {values}", "values"_attr = seqLog(arrayOfInts.data(), arrayOfInts.data() + 5);
StringMap<BSONObj> bsonMap = ...;
LOGV2(1013, "log map directly: {values}", "values"_attr = bsonMap);
LOGV2(1014, "log map iterator range: {values}", "values"_attr = mapLog(bsonMap.begin(), bsonMap.end());
Logging of containers uses BSONObj
as an internal representation and uint64_t
is not a supported type with BSONObjBuilder::append()
. As a user you can use boost::transform_iterator
to cast the uint64_t
to a supported type.
std::vector<uint64_t> vec = ...;
// If we know casting to signed is safe
auto asSigned = [](uint64_t i) { return static_cast<int64_t>(i); };
LOGV2(2000, "as signed array: {values}", "values"_attr = seqLog(
boost::make_transform_iterator(vec.begin(), asSigned),
boost::make_transform_iterator(vec.end(), asSigned)
));
// Otherwise we can log as any of these types instead of using asSigned
auto asDecimal128 = [](uint64_t i) { return Decimal128(i); };
auto asString = [](uint64_t i) { return std::to_string(i); };
Duration types have special formatting to match existing practices in the server code base. Their resulting format depends on the context they are logged.
When durations are formatted as JSON or BSON a unit suffix is added to the attribute name when building the field name. The value will be count of the duration as a number.
When logging containers with durations there is no attribute per duration instance that can have the suffix added. In this case durations are instead formatted as a BSON object.
"duration"_attr = Milliseconds(10)
JSON format:
"durationMillis": 10
std::vector<Nanoseconds> nanos = {Nanoseconds(200), Nanoseconds(400)};
"samples"_attr = nanos
JSON format:
"samples": [{"durationNanos": 200}, {"durationNanos": 400}]
Code that emits a high severity log statement may also need to emit a uassert
after the log. There is the UserAssertAfterLog
logging option that allows you to re-use the log statement to do the formatting required for the uassert
. The assertion id can be either the logging ID by passing UserAssertAfterLog
with no arguments or the assertion id can set by constructing UserAssertAfterLog
with an ErrorCodes::Error
.
The assertion reason string will be a plain text formatted log (replacement fields filled in format-string). If replacement fields are not provided in the message string, attribute values will be missing from the assertion message.
LOGV2_ERROR_OPTIONS(1050000, {UserAssertAfterLog()}, "Assertion after log");
Would emit a uassert
after performing the log that is equivalent to:
uasserted(1050000, "Assertion after log");
Using a named error code:
LOGV2_ERROR_OPTIONS(1050, {UserAssertAfterLog(ErrorCodes::DataCorruptionDetected)}, "Data corruption detected for {recordId}, "recordId"_attr=RecordId(123456));
Would emit a uassert
after performing the log that is equivalent to:
uasserted(ErrorCodes::DataCorruptionDetected, "Data corruption detected for RecordId(123456)");
Produces structured logs of the Relaxed Extended JSON 2.0.0 format. Below is an example of a log statement in C++ and a pretty-printed JSON output:
BSONObjBuilder builder;
builder.append("first"_sd, 1);
builder.append("second"_sd, "str");
std::vector<int> vec = {1, 2, 3};
LOGV2_ERROR(1020, "Example (b: {bson}), (vec: {vector})",
"bson"_attr = builder.obj(),
"vector"_attr = vec,
"optional"_attr = boost::none);
{
"t": {
"$date": "2020-01-06T19:10:54.246Z"
},
"s": "E",
"c": "NETWORK",
"ctx": "conn1",
"id": 23453,
"msg": "Example (b: {bson}), (vec: {vector})",
"attr": {
"bson": {
"first": 1,
"second": "str"
},
"vector": [1, 2, 3],
"optional": null
}
}
Structured logging brings significant potential for log analysis to the codebase that isn't present with earlier logging facilities. This is an improvement that facilitates many future improvements.
Not only that, logv2 removes most parsing/post-processing concerns for automated downstream consumption of logs.
Maintaining multiple output formats for even a single version would present serious overhead for both support and engineering. This dual support would last for years given the adoption curve, and effectively creates four formats (old, new, new-old, and newer).
By making a full cutover in a single release, we are in a much better position.
Human readability suffers significantly when attr
field names are included both in the attr
subdocument and within msg
string. This is a powerful feature that we don't want to exclude entirely, but it makes sense to lean on it only when absolutely necessary.