From 7f936e91861895f9bd02d82b0b720088bde36d7c Mon Sep 17 00:00:00 2001 From: Olivier Michallat Date: Wed, 12 Mar 2025 15:46:28 -0700 Subject: [PATCH 1/3] HCD-92 Upgrade Jamm version to 0.4.0 Backport of 600f4d9a69 from the main-5.0 branch. --- bin/cassandra.in.sh | 2 +- build.xml | 2 +- conf/cassandra-env.sh | 2 +- redhat/cassandra.in.sh | 2 +- .../cassandra/audit/BinAuditLogger.java | 7 +- .../cassandra/cql3/ColumnIdentifier.java | 2 +- .../apache/cassandra/db/BufferClustering.java | 2 +- .../apache/cassandra/db/rows/BufferCell.java | 2 +- .../apache/cassandra/db/rows/CellPath.java | 2 +- .../cassandra/db/tries/InMemoryTrie.java | 5 +- .../apache/cassandra/fql/FullQueryLogger.java | 91 +++++++--- .../sai/disk/v1/kdtree/BKDPostingsIndex.java | 8 +- .../apache/cassandra/utils/ObjectSizes.java | 135 +++++++------- .../microbench/tries/ComparisonReadBench.java | 4 +- .../org/apache/cassandra/db/CellSpecTest.java | 2 +- .../db/memtable/MemtableSizeTestBase.java | 10 +- .../cassandra/utils/ObjectSizesTest.java | 164 +++++++++--------- 17 files changed, 249 insertions(+), 193 deletions(-) diff --git a/bin/cassandra.in.sh b/bin/cassandra.in.sh index 61eed888dc52..97d2d65cdab2 100644 --- a/bin/cassandra.in.sh +++ b/bin/cassandra.in.sh @@ -79,7 +79,7 @@ if [ -f "$CASSANDRA_HOME"/lib/jsr223/scala/scala-compiler.jar ] ; then fi # set JVM javaagent opts to avoid warnings/errors -JAVA_AGENT="$JAVA_AGENT -javaagent:$CASSANDRA_HOME/lib/jamm-0.3.2.jar" +JAVA_AGENT="$JAVA_AGENT -javaagent:$CASSANDRA_HOME/lib/jamm-0.4.0.jar" # Added sigar-bin to the java.library.path CASSANDRA-7838 JAVA_OPTS="$JAVA_OPTS:-Djava.library.path=$CASSANDRA_HOME/lib/sigar-bin" diff --git a/build.xml b/build.xml index 50765afbd9ad..9959aa891d16 100644 --- a/build.xml +++ b/build.xml @@ -153,7 +153,7 @@ - + diff --git a/conf/cassandra-env.sh b/conf/cassandra-env.sh index 168e81bc20cd..9915ee686623 100644 --- a/conf/cassandra-env.sh +++ b/conf/cassandra-env.sh @@ -191,7 +191,7 @@ fi JVM_OPTS="$JVM_OPTS -XX:CompileCommandFile=$CASSANDRA_CONF/hotspot_compiler" # add the jamm javaagent -JVM_OPTS="$JVM_OPTS -javaagent:$CASSANDRA_HOME/lib/jamm-0.3.2.jar" +JVM_OPTS="$JVM_OPTS -javaagent:$CASSANDRA_HOME/lib/jamm-0.4.0.jar" # set jvm HeapDumpPath with CASSANDRA_HEAPDUMP_DIR if [ "x$CASSANDRA_HEAPDUMP_DIR" != "x" ]; then diff --git a/redhat/cassandra.in.sh b/redhat/cassandra.in.sh index c06cf72a3fa2..fed5d4384ec3 100644 --- a/redhat/cassandra.in.sh +++ b/redhat/cassandra.in.sh @@ -40,7 +40,7 @@ CLASSPATH="$CLASSPATH:$EXTRA_CLASSPATH" # set JVM javaagent opts to avoid warnings/errors -JAVA_AGENT="$JAVA_AGENT -javaagent:$CASSANDRA_HOME/lib/jamm-0.3.2.jar" +JAVA_AGENT="$JAVA_AGENT -javaagent:$CASSANDRA_HOME/lib/jamm-0.4.0.jar" # diff --git a/src/java/org/apache/cassandra/audit/BinAuditLogger.java b/src/java/org/apache/cassandra/audit/BinAuditLogger.java index 95a53f1ed475..a28a60b577e3 100644 --- a/src/java/org/apache/cassandra/audit/BinAuditLogger.java +++ b/src/java/org/apache/cassandra/audit/BinAuditLogger.java @@ -96,6 +96,11 @@ public void log(AuditLogEntry auditLogEntry) @VisibleForTesting public static class Message extends BinLog.ReleaseableWriteMarshallable implements WeightedQueue.Weighable { + /** + * The shallow size of a {@code Message} object. + */ + private static final long EMPTY_SIZE = ObjectSizes.measure(new Message("")); + private final String message; public Message(String message) @@ -128,7 +133,7 @@ public void release() @Override public int weight() { - return Ints.checkedCast(ObjectSizes.sizeOf(message)); + return Ints.checkedCast(EMPTY_SIZE + ObjectSizes.sizeOf(message)); } } } diff --git a/src/java/org/apache/cassandra/cql3/ColumnIdentifier.java b/src/java/org/apache/cassandra/cql3/ColumnIdentifier.java index d8c0dbc97b8b..5480636ebfd5 100644 --- a/src/java/org/apache/cassandra/cql3/ColumnIdentifier.java +++ b/src/java/org/apache/cassandra/cql3/ColumnIdentifier.java @@ -221,7 +221,7 @@ public long unsharedHeapSize() public long unsharedHeapSizeExcludingData() { return EMPTY_SIZE - + ObjectSizes.sizeOnHeapExcludingData(bytes) + + ObjectSizes.sizeOnHeapExcludingDataOf(bytes) + ObjectSizes.sizeOf(text); } diff --git a/src/java/org/apache/cassandra/db/BufferClustering.java b/src/java/org/apache/cassandra/db/BufferClustering.java index 0423c26ccdf3..625b9387449f 100644 --- a/src/java/org/apache/cassandra/db/BufferClustering.java +++ b/src/java/org/apache/cassandra/db/BufferClustering.java @@ -53,7 +53,7 @@ public long unsharedHeapSizeExcludingData() { if (this == Clustering.EMPTY || this == Clustering.STATIC_CLUSTERING) return 0; - return EMPTY_SIZE + ObjectSizes.sizeOnHeapExcludingData(values); + return EMPTY_SIZE + ObjectSizes.sizeOnHeapExcludingDataOf(values); } public static BufferClustering make(ByteBuffer... values) diff --git a/src/java/org/apache/cassandra/db/rows/BufferCell.java b/src/java/org/apache/cassandra/db/rows/BufferCell.java index fc85b3973aa0..1280a3fded25 100644 --- a/src/java/org/apache/cassandra/db/rows/BufferCell.java +++ b/src/java/org/apache/cassandra/db/rows/BufferCell.java @@ -145,6 +145,6 @@ public Cell clone(ByteBufferCloner cloner) public long unsharedHeapSizeExcludingData() { - return EMPTY_SIZE + ObjectSizes.sizeOnHeapExcludingData(value) + (path == null ? 0 : path.unsharedHeapSizeExcludingData()); + return EMPTY_SIZE + ObjectSizes.sizeOnHeapExcludingDataOf(value) + (path == null ? 0 : path.unsharedHeapSizeExcludingData()); } } diff --git a/src/java/org/apache/cassandra/db/rows/CellPath.java b/src/java/org/apache/cassandra/db/rows/CellPath.java index aacccf30ba13..9354754546bc 100644 --- a/src/java/org/apache/cassandra/db/rows/CellPath.java +++ b/src/java/org/apache/cassandra/db/rows/CellPath.java @@ -128,7 +128,7 @@ public CellPath clone(ByteBufferCloner cloner) public long unsharedHeapSizeExcludingData() { - return EMPTY_SIZE + ObjectSizes.sizeOnHeapExcludingData(value); + return EMPTY_SIZE + ObjectSizes.sizeOnHeapExcludingDataOf(value); } } diff --git a/src/java/org/apache/cassandra/db/tries/InMemoryTrie.java b/src/java/org/apache/cassandra/db/tries/InMemoryTrie.java index f6c496bea2e4..bd2142010886 100644 --- a/src/java/org/apache/cassandra/db/tries/InMemoryTrie.java +++ b/src/java/org/apache/cassandra/db/tries/InMemoryTrie.java @@ -34,6 +34,7 @@ import org.apache.cassandra.utils.ObjectSizes; import org.apache.cassandra.utils.concurrent.OpOrder; import org.github.jamm.MemoryLayoutSpecification; +import org.github.jamm.MemoryMeterStrategy; /** * In-memory trie built for fast modification and reads executing concurrently with writes from a single mutator thread. @@ -1705,7 +1706,7 @@ private long usedBufferSpace() private long usedObjectSpace() { - return (contentCount - objectAllocator.indexCountInPipeline()) * MemoryLayoutSpecification.SPEC.getReferenceSize(); + return (contentCount - objectAllocator.indexCountInPipeline()) * MemoryMeterStrategy.MEMORY_LAYOUT.getReferenceSize(); } /** @@ -1731,7 +1732,7 @@ public long unusedReservedOnHeapMemory() AtomicReferenceArray contentArray = contentArrays[leadBit]; int contentOverhead = ((contentArray != null ? contentArray.length() : 0) - ofs); contentOverhead += objectAllocator.indexCountInPipeline(); - contentOverhead *= MemoryLayoutSpecification.SPEC.getReferenceSize(); + contentOverhead *= MemoryMeterStrategy.MEMORY_LAYOUT.getReferenceSize(); return bufferOverhead + contentOverhead; } diff --git a/src/java/org/apache/cassandra/fql/FullQueryLogger.java b/src/java/org/apache/cassandra/fql/FullQueryLogger.java index 8a145c66e9ca..7a3fe11ebbc8 100644 --- a/src/java/org/apache/cassandra/fql/FullQueryLogger.java +++ b/src/java/org/apache/cassandra/fql/FullQueryLogger.java @@ -84,9 +84,6 @@ public class FullQueryLogger implements QueryEvents.Listener private static final int EMPTY_LIST_SIZE = Ints.checkedCast(ObjectSizes.measureDeep(new ArrayList<>(0))); private static final int EMPTY_BYTEBUF_SIZE; - private static final int OBJECT_HEADER_SIZE = MemoryLayoutSpecification.SPEC.getObjectHeaderSize(); - private static final int OBJECT_REFERENCE_SIZE = MemoryLayoutSpecification.SPEC.getReferenceSize(); - public static final FullQueryLogger instance = new FullQueryLogger(); volatile BinLog binLog; @@ -331,6 +328,11 @@ public void executeSuccess(CQLStatement statement, String query, QueryOptions op public static class Query extends AbstractLogEntry { + /** + * The shallow size of a {@code Query} object. + */ + private static final long EMPTY_SIZE = ObjectSizes.measure(new Query()); + private final String query; public Query(String query, QueryOptions queryOptions, QueryState queryState, long queryStartTime) @@ -339,6 +341,14 @@ public Query(String query, QueryOptions queryOptions, QueryState queryState, lon this.query = query; } + /** + * Constructor only use to compute this class shallow size. + */ + private Query() + { + this.query = null; + } + @Override protected String type() { @@ -355,12 +365,21 @@ public void writeMarshallablePayload(WireOut wire) @Override public int weight() { - return Ints.checkedCast(ObjectSizes.sizeOf(query)) + super.weight(); + // Object deep size = Object' shallow size + query field deep size + deep size of the parent fields + return Ints.checkedCast(EMPTY_SIZE + ObjectSizes.sizeOf(query) + super.fieldsSize()); } } public static class Batch extends AbstractLogEntry { + /** + * The shallow size of a {@code Batch} object (which includes primitive fields). + */ + private static final long EMPTY_SIZE = ObjectSizes.measure(new Batch()); + + /** + * The weight is pre-computed in the constructor and represent the object deep size. + */ private final int weight; private final BatchStatement.Type batchType; private final List queries; @@ -379,25 +398,37 @@ public Batch(BatchStatement.Type batchType, this.values = values; this.batchType = batchType; - int weight = super.weight(); - - // weight, queries, values, batch type - weight += Integer.BYTES + // cached weight - 2 * EMPTY_LIST_SIZE + // queries + values lists - 3 * OBJECT_REFERENCE_SIZE; // batchType and two lists references + // We assume that all the lists are ArrayLists and that the size of each underlying array is the one of the list + // (which is obviously wrong but not worst than the previous computation that was ignoring part of the arrays size in the computation). + long queriesSize = EMPTY_LIST_SIZE + ObjectSizes.sizeOfReferenceArray(queries.size()); for (String query : queries) - weight += ObjectSizes.sizeOf(checkNotNull(query)) + OBJECT_REFERENCE_SIZE; + queriesSize += ObjectSizes.sizeOf(checkNotNull(query)); + long valuesSize = EMPTY_LIST_SIZE + ObjectSizes.sizeOfReferenceArray(values.size()); for (List subValues : values) { - weight += EMPTY_LIST_SIZE + OBJECT_REFERENCE_SIZE; - - for (ByteBuffer value : subValues) - weight += ObjectSizes.sizeOnHeapOf(value) + OBJECT_REFERENCE_SIZE; + valuesSize += EMPTY_LIST_SIZE + ObjectSizes.sizeOfReferenceArray(subValues.size()); + for (ByteBuffer subValue : subValues) + valuesSize += ObjectSizes.sizeOnHeapOf(subValue); } - this.weight = weight; + // No need to add the batch type which is an enum. + this.weight = Ints.checkedCast(EMPTY_SIZE // Shallow size object + + super.fieldsSize() // deep size of the parent fields (non-primitives as they are included in the shallow size) + + queriesSize // deep size queries field + + valuesSize); // deep size values field + } + + /** + * Constructor only use to compute this class shallow size. + */ + private Batch() + { + this.weight = 0; + this.batchType = null; + this.queries = null; + this.values = null; } @Override @@ -482,6 +513,18 @@ private static abstract class AbstractLogEntry extends BinLog.ReleaseableWriteMa } } + /** + * Constructor only use to compute sub-classes shallow size. + */ + private AbstractLogEntry() + { + this.queryStartTime = 0; + this.protocolVersion = 0; + this.queryOptionsBuffer = null; + this.generatedTimestamp = 0; + this.generatedNowInSeconds = 0; + this.keyspace = null; + } @Override protected long version() { @@ -507,16 +550,14 @@ public void release() queryOptionsBuffer.release(); } - @Override - public int weight() + /** + * Returns the sum of the non-primitive fields' deep sizes. + * @return the sum of the non-primitive fields' deep sizes. + */ + protected long fieldsSize() { - return OBJECT_HEADER_SIZE - + Long.BYTES // queryStartTime - + Integer.BYTES // protocolVersion - + OBJECT_REFERENCE_SIZE + EMPTY_BYTEBUF_SIZE + queryOptionsBuffer.capacity() // queryOptionsBuffer - + Long.BYTES // generatedTimestamp - + Integer.BYTES // generatedNowInSeconds - + OBJECT_REFERENCE_SIZE + Ints.checkedCast(ObjectSizes.sizeOf(keyspace)); // keyspace + return EMPTY_BYTEBUF_SIZE + queryOptionsBuffer.capacity() // queryOptionsBuffer + + ObjectSizes.sizeOf(keyspace); // keyspace } } diff --git a/src/java/org/apache/cassandra/index/sai/disk/v1/kdtree/BKDPostingsIndex.java b/src/java/org/apache/cassandra/index/sai/disk/v1/kdtree/BKDPostingsIndex.java index 6fcc44051b70..3aa840e0d1fc 100644 --- a/src/java/org/apache/cassandra/index/sai/disk/v1/kdtree/BKDPostingsIndex.java +++ b/src/java/org/apache/cassandra/index/sai/disk/v1/kdtree/BKDPostingsIndex.java @@ -23,6 +23,7 @@ import com.carrotsearch.hppc.IntLongMap; import org.apache.cassandra.index.sai.disk.io.IndexInputReader; import org.apache.cassandra.io.util.FileHandle; +import org.apache.cassandra.utils.ObjectSizes; import org.github.jamm.MemoryLayoutSpecification; import static com.google.common.base.Preconditions.checkArgument; @@ -57,13 +58,6 @@ class BKDPostingsIndex } } - public long memoryUsage() - { - // IntLongHashMap uses two arrays: one for keys, one for values. - return MemoryLayoutSpecification.sizeOfArray(index.size(), 4L) - + MemoryLayoutSpecification.sizeOfArray(index.size(), 8L); - } - /** * Returns true if given node ID has an auxiliary posting list. */ diff --git a/src/java/org/apache/cassandra/utils/ObjectSizes.java b/src/java/org/apache/cassandra/utils/ObjectSizes.java index 468522c6ff06..4da1e2783f02 100644 --- a/src/java/org/apache/cassandra/utils/ObjectSizes.java +++ b/src/java/org/apache/cassandra/utils/ObjectSizes.java @@ -21,23 +21,26 @@ import java.nio.ByteBuffer; -import org.github.jamm.MemoryLayoutSpecification; import org.github.jamm.MemoryMeter; +import org.github.jamm.MemoryMeter.ByteBufferMode; +import org.github.jamm.MemoryMeter.Guess; + +import static org.github.jamm.MemoryMeterStrategy.MEMORY_LAYOUT; +import static org.github.jamm.utils.ArrayMeasurementUtils.computeArraySize; /** - * A convenience class for wrapping access to MemoryMeter + * A convenience class for wrapping access to MemoryMeter. Should be used instead of using a {@code MemoryMeter} directly. + * {@code MemoryMeter} can be used directly for testing as it allow a more fine tuned configuration for comparison. */ public class ObjectSizes { - private static final MemoryMeter meter = new MemoryMeter().omitSharedBufferOverhead() - .withGuessing(MemoryMeter.Guess.FALLBACK_UNSAFE) - .ignoreKnownSingletons(); - - private static final long EMPTY_HEAP_BUFFER_SIZE = measure(ByteBufferUtil.EMPTY_BYTE_BUFFER); - private static final long EMPTY_BYTE_ARRAY_SIZE = measure(new byte[0]); - private static final long EMPTY_STRING_SIZE = measure(""); + private static final MemoryMeter meter = MemoryMeter.builder().withGuessing(Guess.INSTRUMENTATION_AND_SPECIFICATION, + Guess.UNSAFE) + .build(); - private static final long DIRECT_BUFFER_HEAP_SIZE = measure(ByteBuffer.allocateDirect(0)); + private static final long HEAP_BUFFER_SHALLOW_SIZE = measure(ByteBufferUtil.EMPTY_BYTE_BUFFER); + private static final long DIRECT_BUFFER_SHALLOW_SIZE = measure(ByteBuffer.allocateDirect(0)); + private static final long DIRECT_BUFFER_DEEP_SIZE = measureDeep(ByteBuffer.allocateDirect(0)); /** * Memory a byte array consumes @@ -47,10 +50,7 @@ public class ObjectSizes */ public static long sizeOfArray(byte[] bytes) { - if (bytes == null) - return 0; - - return sizeOfArray(bytes.length, 1); + return meter.measureArray(bytes); } /** @@ -61,10 +61,7 @@ public static long sizeOfArray(byte[] bytes) */ public static long sizeOfArray(long[] longs) { - if (longs == null) - return 0; - - return sizeOfArray(longs.length, 8); + return meter.measureArray(longs); } /** @@ -75,10 +72,7 @@ public static long sizeOfArray(long[] longs) */ public static long sizeOfArray(int[] ints) { - if (ints == null) - return 0; - - return sizeOfArray(ints.length, 4); + return meter.measureArray(ints); } /** @@ -89,7 +83,7 @@ public static long sizeOfArray(int[] ints) */ public static long sizeOfReferenceArray(int length) { - return sizeOfArray(length, MemoryLayoutSpecification.SPEC.getReferenceSize()); + return sizeOfArray(length, MEMORY_LAYOUT.getReferenceSize()); } /** @@ -100,15 +94,12 @@ public static long sizeOfReferenceArray(int length) */ public static long sizeOfArray(Object[] objects) { - if (objects == null) - return 0; - - return sizeOfReferenceArray(objects.length); + return meter.measureArray(objects); } - private static long sizeOfArray(int length, long elementSize) + private static long sizeOfArray(int length, int elementSize) { - return MemoryLayoutSpecification.sizeOfArray(length, elementSize); + return computeArraySize(MEMORY_LAYOUT.getArrayHeaderSize(), length, elementSize, MEMORY_LAYOUT.getObjectAlignment()); } /** @@ -129,65 +120,91 @@ public static long sizeOnHeapOf(ByteBuffer[] array) /** * Amount of non-data heap memory consumed by the array of byte buffers. It sums memory consumed - * by the array itself and for each included byte buffer using {@link #sizeOnHeapExcludingData(ByteBuffer)}. + * by the array itself and for each included byte buffer using {@link #sizeOnHeapExcludingDataOf(ByteBuffer)}. */ - public static long sizeOnHeapExcludingData(ByteBuffer[] array) + public static long sizeOnHeapExcludingDataOf(ByteBuffer[] array) { if (array == null) return 0; long sum = sizeOfArray(array); for (ByteBuffer b : array) - sum += sizeOnHeapExcludingData(b); + sum += sizeOnHeapExcludingDataOf(b); return sum; } /** - * @return heap memory consumed by the byte buffer. If it is a slice, it counts the data size, but it does not - * include the internal array overhead. + * Measures the heap memory used by the specified byte buffer. If the buffer is a slab only the data size will be + * counted but not the internal overhead. A SLAB is assumed to be created by: {@code buffer.duplicate().position(start).limit(end)} without the use of {@code slice()}. + *

This method makes a certain amount of assumptions: + *

    + *
  • That slabs are always created using: {@code buffer.duplicate().position(start).limit(end)} and not through slice
  • + *
  • That the input buffers are not read-only buffers
  • + *
  • That the direct buffers that are not slab are not duplicates
  • + *
+ * Non-respect of those assumptions can lead to an invalid value being returned. + * @param buffer the buffer to measure + * @return the heap memory used by the specified byte buffer. */ public static long sizeOnHeapOf(ByteBuffer buffer) { if (buffer == null) return 0; - if (buffer.isDirect()) - return DIRECT_BUFFER_HEAP_SIZE; + assert !buffer.isReadOnly(); + - int arrayLen = buffer.array().length; - int bufLen = buffer.remaining(); + // We assume here that slabs are always created using: buffer.duplicate().position(start).limit(end) and not through slice + if (ByteBufferMode.SLAB_ALLOCATION_NO_SLICE.isSlab(buffer)) + { + if (buffer.isDirect()) + return DIRECT_BUFFER_SHALLOW_SIZE; // We ignore the underlying buffer - // if we're only referencing a sub-portion of the ByteBuffer, don't count the array overhead (assume it is SLAB - // allocated - the overhead amortized over all the allocations is negligible and better to undercount than over) - if (arrayLen > bufLen) - return EMPTY_HEAP_BUFFER_SIZE + bufLen; + return HEAP_BUFFER_SHALLOW_SIZE + buffer.remaining(); // We ignore the array overhead + } - return EMPTY_HEAP_BUFFER_SIZE + (arrayLen == 0 ? EMPTY_BYTE_ARRAY_SIZE : sizeOfArray(arrayLen, 1)); + if (buffer.isDirect()) + return DIRECT_BUFFER_DEEP_SIZE; // That might not be true if the buffer is a view of another buffer so we could undercount + + return HEAP_BUFFER_SHALLOW_SIZE + meter.measureArray(buffer.array()); } /** - * @return non-data heap memory consumed by the byte buffer. If it is a slice, it does not include the internal - * array overhead. + * Measures the heap memory used by the specified byte buffer excluding the data. If the buffer shallow size will be counted. + * A SLAB is assumed to be created by: {@code buffer.duplicate().position(start).limit(end)} without the use of {@code slice()}. + *

This method makes a certain amount of assumptions: + *

    + *
  • That slabs are always created using: {@code buffer.duplicate().position(start).limit(end)} and not through slice
  • + *
  • That the input buffers are not read-only buffers
  • + *
  • That the direct buffers that are not slab are not duplicates
  • + *
+ * Non-respect of those assumptions can lead to an invalid value being returned. T + * @param buffer the buffer to measure + * @return the heap memory used by the specified byte buffer excluding the data.. */ - public static long sizeOnHeapExcludingData(ByteBuffer buffer) + public static long sizeOnHeapExcludingDataOf(ByteBuffer buffer) { if (buffer == null) return 0; - if (buffer.isDirect()) - return DIRECT_BUFFER_HEAP_SIZE; + assert !buffer.isReadOnly(); - int arrayLen = buffer.array().length; - int bufLen = buffer.remaining(); + // We assume here that slabs are always created using: buffer.duplicate().position(start).limit(end) and not through slice + if (ByteBufferMode.SLAB_ALLOCATION_NO_SLICE.isSlab(buffer)) + { + if (buffer.isDirect()) + return DIRECT_BUFFER_SHALLOW_SIZE; // We ignore the underlying buffer - // if we're only referencing a sub-portion of the ByteBuffer, don't count the array overhead (assume it is SLAB - // allocated - the overhead amortized over all the allocations is negligible and better to undercount than over) - if (arrayLen > bufLen) - return EMPTY_HEAP_BUFFER_SIZE; + return HEAP_BUFFER_SHALLOW_SIZE; // We ignore the array overhead + } - // If buffers are dedicated, account for byte array size and any padding overhead - return EMPTY_HEAP_BUFFER_SIZE + (arrayLen == 0 ? EMPTY_BYTE_ARRAY_SIZE : (sizeOfArray(arrayLen, 1) - arrayLen)); + + if (buffer.isDirect()) + return DIRECT_BUFFER_DEEP_SIZE; // That might not be true if the buffer is a view of another buffer so we could undercount + + byte[] bytes = buffer.array(); + return HEAP_BUFFER_SHALLOW_SIZE + meter.measureArray(bytes) - bytes.length; } /** @@ -196,13 +213,9 @@ public static long sizeOnHeapExcludingData(ByteBuffer buffer) * @param str String to calculate memory size of * @return Total in-memory size of the String */ - // TODO hard coding this to 2 isn't necessarily correct in Java 11 public static long sizeOf(String str) { - if (str == null) - return 0; - - return EMPTY_STRING_SIZE + sizeOfArray(str.length(), Character.BYTES); + return meter.measureStringDeep(str); } /** diff --git a/test/microbench/org/apache/cassandra/test/microbench/tries/ComparisonReadBench.java b/test/microbench/org/apache/cassandra/test/microbench/tries/ComparisonReadBench.java index 7cefbacda9fe..5c5c38f405fa 100644 --- a/test/microbench/org/apache/cassandra/test/microbench/tries/ComparisonReadBench.java +++ b/test/microbench/org/apache/cassandra/test/microbench/tries/ComparisonReadBench.java @@ -69,7 +69,9 @@ public class ComparisonReadBench { // Note: To see a printout of the usage for each object, add .printVisitedTree() here (most useful with smaller number of // partitions). - static MemoryMeter meter = new MemoryMeter().withGuessing(Guess.FALLBACK_UNSAFE); + static MemoryMeter meter = MemoryMeter.builder() + .withGuessing(Guess.INSTRUMENTATION_AND_SPECIFICATION, Guess.UNSAFE) + .build(); public enum TrieAllocation { SHORT_LIVED, diff --git a/test/unit/org/apache/cassandra/db/CellSpecTest.java b/test/unit/org/apache/cassandra/db/CellSpecTest.java index df6d7dcba0fc..0b0bdf46741a 100644 --- a/test/unit/org/apache/cassandra/db/CellSpecTest.java +++ b/test/unit/org/apache/cassandra/db/CellSpecTest.java @@ -84,7 +84,7 @@ public void unsharedHeapSizeExcludingData() private static long valuePtrSize(Object value) { if (value instanceof ByteBuffer) - return ObjectSizes.sizeOnHeapExcludingData((ByteBuffer) value); + return ObjectSizes.sizeOnHeapExcludingDataOf((ByteBuffer) value); else if (value instanceof byte[]) return ObjectSizes.sizeOfArray((byte[]) value) - ((byte[]) value).length; throw new IllegalArgumentException("Unsupported type by valuePtrSize: " + value.getClass()); diff --git a/test/unit/org/apache/cassandra/db/memtable/MemtableSizeTestBase.java b/test/unit/org/apache/cassandra/db/memtable/MemtableSizeTestBase.java index 52ab207d0a54..a027936718f4 100644 --- a/test/unit/org/apache/cassandra/db/memtable/MemtableSizeTestBase.java +++ b/test/unit/org/apache/cassandra/db/memtable/MemtableSizeTestBase.java @@ -34,16 +34,20 @@ import org.apache.cassandra.db.Keyspace; import org.apache.cassandra.utils.FBUtilities; import org.github.jamm.MemoryMeter; +import org.github.jamm.MemoryMeter.Guess; // Note: This test can be run in idea with the allocation type configured in the test yaml and memtable using the // value memtableClass is initialized with. public abstract class MemtableSizeTestBase extends CQLTester { // The meter in ObjectSizes uses omitSharedBufferOverhead which counts off-heap data too - // Note: To see a printout of the usage for each object, add .enableDebug() here (most useful with smaller number of + // Note: To see a printout of the usage for each object, add .printVisitedTree() here (most useful with smaller number of // partitions). - static final MemoryMeter meter = new MemoryMeter().ignoreKnownSingletons() - .withGuessing(MemoryMeter.Guess.FALLBACK_UNSAFE); + static MemoryMeter meter = MemoryMeter.builder() + .withGuessing(Guess.INSTRUMENTATION_AND_SPECIFICATION, + Guess.UNSAFE) +// .printVisitedTreeUpTo(1000) + .build(); static String keyspace; String table; diff --git a/test/unit/org/apache/cassandra/utils/ObjectSizesTest.java b/test/unit/org/apache/cassandra/utils/ObjectSizesTest.java index a4c77bccc411..e7ac87a22241 100644 --- a/test/unit/org/apache/cassandra/utils/ObjectSizesTest.java +++ b/test/unit/org/apache/cassandra/utils/ObjectSizesTest.java @@ -24,126 +24,122 @@ import org.github.jamm.MemoryLayoutSpecification; import org.github.jamm.MemoryMeter; +import org.github.jamm.MemoryMeter.Guess; -import static org.assertj.core.api.Assertions.assertThat; +import static org.github.jamm.MemoryMeter.ByteBufferMode.SLAB_ALLOCATION_NO_SLICE; +import static org.junit.Assert.assertEquals; public class ObjectSizesTest { - private static final MemoryMeter meter = new MemoryMeter().withGuessing(MemoryMeter.Guess.FALLBACK_UNSAFE).omitSharedBufferOverhead().ignoreKnownSingletons(); - - private static final long EMPTY_HEAP_BUFFER_RAW_SIZE = meter.measure(ByteBuffer.allocate(0)); - private static final long EMPTY_OFFHEAP_BUFFER_RAW_SIZE = meter.measure(ByteBuffer.allocateDirect(0)); - private static final ByteBuffer[] EMPTY_BYTE_BUFFER_ARRAY = new ByteBuffer[0]; - - public static final long REF_ARRAY_0_SIZE = MemoryLayoutSpecification.sizeOfArray(0, MemoryLayoutSpecification.SPEC.getReferenceSize()); - public static final long REF_ARRAY_1_SIZE = MemoryLayoutSpecification.sizeOfArray(1, MemoryLayoutSpecification.SPEC.getReferenceSize()); - public static final long REF_ARRAY_2_SIZE = MemoryLayoutSpecification.sizeOfArray(2, MemoryLayoutSpecification.SPEC.getReferenceSize()); - - public static final long BYTE_ARRAY_0_SIZE = MemoryLayoutSpecification.sizeOfArray(0, 1); - public static final long BYTE_ARRAY_10_SIZE = MemoryLayoutSpecification.sizeOfArray(10, 1); - public static final long BYTE_ARRAY_10_EXCEPT_DATA_SIZE = MemoryLayoutSpecification.sizeOfArray(10, 1) - 10; - - private ByteBuffer buf10 = ByteBuffer.allocate(10); - private ByteBuffer prefixBuf8 = buf10.duplicate(); - private ByteBuffer suffixBuf9 = buf10.duplicate(); - private ByteBuffer infixBuf7 = buf10.duplicate(); - - { - prefixBuf8.limit(8); - - suffixBuf9.position(1); - suffixBuf9 = suffixBuf9.slice(); - - infixBuf7.limit(8); - infixBuf7.position(1); - infixBuf7 = infixBuf7.slice(); - } + // We use INSTRUMENTATION as principal strategy as it is our reference strategy + private static final MemoryMeter meter = MemoryMeter.builder() + .withGuessing(Guess.INSTRUMENTATION, Guess.UNSAFE) + .build(); @Test public void testSizeOnHeapExcludingData() { - // empty array of byte buffers - ByteBuffer[] buffers = EMPTY_BYTE_BUFFER_ARRAY; - assertThat(ObjectSizes.sizeOnHeapExcludingData(buffers)).isEqualTo(REF_ARRAY_0_SIZE); - // single empty heap buffer - buffers = new ByteBuffer[]{ ByteBuffer.allocate(0) }; - assertThat(ObjectSizes.sizeOnHeapExcludingData(buffers)).isEqualTo(REF_ARRAY_1_SIZE + EMPTY_HEAP_BUFFER_RAW_SIZE + BYTE_ARRAY_0_SIZE); + checkBufferSizeExcludingData(ByteBuffer.allocate(0), 0); // single non-empty heap buffer - buffers = new ByteBuffer[]{ buf10 }; - assertThat(ObjectSizes.sizeOnHeapExcludingData(buffers)).isEqualTo(REF_ARRAY_1_SIZE + EMPTY_HEAP_BUFFER_RAW_SIZE + BYTE_ARRAY_10_EXCEPT_DATA_SIZE); + checkBufferSizeExcludingData(ByteBuffer.allocate(10), 10); // single empty direct buffer - buffers = new ByteBuffer[]{ ByteBuffer.allocateDirect(0) }; - assertThat(ObjectSizes.sizeOnHeapExcludingData(buffers)).isEqualTo(REF_ARRAY_1_SIZE + EMPTY_OFFHEAP_BUFFER_RAW_SIZE); + checkBufferSizeExcludingData(ByteBuffer.allocateDirect(0), 0); // single non-empty direct buffer - buffers = new ByteBuffer[]{ ByteBuffer.allocateDirect(10) }; - assertThat(ObjectSizes.sizeOnHeapExcludingData(buffers)).isEqualTo(REF_ARRAY_1_SIZE + EMPTY_OFFHEAP_BUFFER_RAW_SIZE); + checkBufferSizeExcludingData(ByteBuffer.allocateDirect(10), 0); + + // heap buffer being a prefix slab + ByteBuffer buffer = (ByteBuffer) ByteBuffer.allocate(10).duplicate().limit(8); + checkBufferSizeExcludingData(buffer, 8); + + // heap buffer being a suffix slab + buffer = (ByteBuffer) ByteBuffer.allocate(10).duplicate().position(1); + checkBufferSizeExcludingData(buffer, 9); + + // heap buffer being an infix slab + buffer = (ByteBuffer) ByteBuffer.allocate(10).duplicate().position(1).limit(8); + checkBufferSizeExcludingData(buffer, 7); + } + + private void checkBufferSizeExcludingData(ByteBuffer buffer, int dataSize) + { + assertEquals(meter.measureDeep(buffer, SLAB_ALLOCATION_NO_SLICE) - dataSize, ObjectSizes.sizeOnHeapExcludingDataOf(buffer)); + } + + @Test + public void testSizeOnHeapExcludingDataArray() + { + checkBufferSizeExcludingDataArray(0, new ByteBuffer[0]); - // two different empty byte buffers - buffers = new ByteBuffer[]{ ByteBuffer.allocate(0), ByteBuffer.allocateDirect(0) }; - assertThat(ObjectSizes.sizeOnHeapExcludingData(buffers)).isEqualTo(REF_ARRAY_2_SIZE + EMPTY_HEAP_BUFFER_RAW_SIZE + BYTE_ARRAY_0_SIZE + EMPTY_OFFHEAP_BUFFER_RAW_SIZE); + // single heap buffer + checkBufferSizeExcludingDataArray(0, ByteBuffer.allocate(0)); - // two different non-empty byte buffers - buffers = new ByteBuffer[]{ buf10, ByteBuffer.allocateDirect(500) }; - assertThat(ObjectSizes.sizeOnHeapExcludingData(buffers)).isEqualTo(REF_ARRAY_2_SIZE + EMPTY_HEAP_BUFFER_RAW_SIZE + BYTE_ARRAY_10_EXCEPT_DATA_SIZE + EMPTY_OFFHEAP_BUFFER_RAW_SIZE); + // multiple buffers + checkBufferSizeExcludingDataArray(10, ByteBuffer.allocate(0), ByteBuffer.allocate(10), ByteBuffer.allocateDirect(10)); - // heap buffer being a prefix slice of other buffer - buffers = new ByteBuffer[]{ prefixBuf8 }; - assertThat(ObjectSizes.sizeOnHeapExcludingData(buffers)).isEqualTo(REF_ARRAY_1_SIZE + EMPTY_HEAP_BUFFER_RAW_SIZE); + // heap buffer being a prefix slab + ByteBuffer prefix = (ByteBuffer) ByteBuffer.allocate(10).duplicate().limit(8); - // heap buffer being a suffix slice of other buffer - buffers = new ByteBuffer[]{ suffixBuf9 }; - assertThat(ObjectSizes.sizeOnHeapExcludingData(buffers)).isEqualTo(REF_ARRAY_1_SIZE + EMPTY_HEAP_BUFFER_RAW_SIZE); + // heap buffer being a suffix slab + ByteBuffer suffix = (ByteBuffer) ByteBuffer.allocate(10).duplicate().position(1); + checkBufferSizeExcludingDataArray(8 + 9, prefix, suffix); + } - // heap buffer being an infix slice of other buffer - buffers = new ByteBuffer[]{ infixBuf7 }; - assertThat(ObjectSizes.sizeOnHeapExcludingData(buffers)).isEqualTo(REF_ARRAY_1_SIZE + EMPTY_HEAP_BUFFER_RAW_SIZE); + private void checkBufferSizeExcludingDataArray(int dataSize, ByteBuffer... buffers) + { + assertEquals(meter.measureDeep(buffers, SLAB_ALLOCATION_NO_SLICE) - dataSize, ObjectSizes.sizeOnHeapExcludingDataOf(buffers)); } @Test public void testSizeOnHeapOf() { - // empty array of byte buffers - ByteBuffer[] buffers = EMPTY_BYTE_BUFFER_ARRAY; - assertThat(ObjectSizes.sizeOnHeapOf(buffers)).isEqualTo(REF_ARRAY_0_SIZE); - // single empty heap buffer - buffers = new ByteBuffer[]{ ByteBuffer.allocate(0) }; - assertThat(ObjectSizes.sizeOnHeapOf(buffers)).isEqualTo(REF_ARRAY_1_SIZE + EMPTY_HEAP_BUFFER_RAW_SIZE + BYTE_ARRAY_0_SIZE); + checkBufferSize(ByteBuffer.allocate(0)); // single non-empty heap buffer - buffers = new ByteBuffer[]{ buf10 }; - assertThat(ObjectSizes.sizeOnHeapOf(buffers)).isEqualTo(REF_ARRAY_1_SIZE + EMPTY_HEAP_BUFFER_RAW_SIZE + BYTE_ARRAY_10_SIZE); + checkBufferSize(ByteBuffer.allocate(10)); // single empty direct buffer - buffers = new ByteBuffer[]{ ByteBuffer.allocateDirect(0) }; - assertThat(ObjectSizes.sizeOnHeapOf(buffers)).isEqualTo(REF_ARRAY_1_SIZE + EMPTY_OFFHEAP_BUFFER_RAW_SIZE); + checkBufferSize(ByteBuffer.allocateDirect(0)); // single non-empty direct buffer - buffers = new ByteBuffer[]{ ByteBuffer.allocateDirect(10) }; - assertThat(ObjectSizes.sizeOnHeapOf(buffers)).isEqualTo(REF_ARRAY_1_SIZE + EMPTY_OFFHEAP_BUFFER_RAW_SIZE); + checkBufferSize(ByteBuffer.allocateDirect(10)); + + // heap buffer being a prefix slab + ByteBuffer buffer = (ByteBuffer) ByteBuffer.allocate(10).duplicate().limit(8); + checkBufferSize(buffer); - // two different empty byte buffers - buffers = new ByteBuffer[]{ ByteBuffer.allocate(0), ByteBuffer.allocateDirect(0) }; - assertThat(ObjectSizes.sizeOnHeapOf(buffers)).isEqualTo(REF_ARRAY_2_SIZE + EMPTY_HEAP_BUFFER_RAW_SIZE + BYTE_ARRAY_0_SIZE + EMPTY_OFFHEAP_BUFFER_RAW_SIZE); + // heap buffer being a suffix slab + buffer = (ByteBuffer) ByteBuffer.allocate(10).duplicate().position(1); + checkBufferSize(buffer); - // two different non-empty byte buffers - buffers = new ByteBuffer[]{ buf10, ByteBuffer.allocateDirect(500) }; - assertThat(ObjectSizes.sizeOnHeapOf(buffers)).isEqualTo(REF_ARRAY_2_SIZE + EMPTY_HEAP_BUFFER_RAW_SIZE + BYTE_ARRAY_10_SIZE + EMPTY_OFFHEAP_BUFFER_RAW_SIZE); + // heap buffer being an infix slab + buffer = (ByteBuffer) ByteBuffer.allocate(10).duplicate().position(1).limit(8); + checkBufferSize(buffer); + } + + private void checkBufferSize(ByteBuffer buffer) + { + assertEquals(meter.measureDeep(buffer, SLAB_ALLOCATION_NO_SLICE), ObjectSizes.sizeOnHeapOf(buffer)); + } - // heap buffer being a prefix slice of other buffer - buffers = new ByteBuffer[]{ prefixBuf8 }; - assertThat(ObjectSizes.sizeOnHeapOf(buffers)).isEqualTo(REF_ARRAY_1_SIZE + EMPTY_HEAP_BUFFER_RAW_SIZE + 8); + @Test + public void testSizeOnHeapOfArray() + { + checkBufferArraySize(new ByteBuffer[0]); - // heap buffer being a suffix slice of other buffer - buffers = new ByteBuffer[]{ suffixBuf9 }; - assertThat(ObjectSizes.sizeOnHeapOf(buffers)).isEqualTo(REF_ARRAY_1_SIZE + EMPTY_HEAP_BUFFER_RAW_SIZE + 9); + // single heap buffer + checkBufferArraySize(ByteBuffer.allocate(0)); - // heap buffer being an infix slice of other buffer - buffers = new ByteBuffer[]{ infixBuf7 }; - assertThat(ObjectSizes.sizeOnHeapOf(buffers)).isEqualTo(REF_ARRAY_1_SIZE + EMPTY_HEAP_BUFFER_RAW_SIZE + 7); + // multiple buffers + checkBufferArraySize(ByteBuffer.allocate(0), ByteBuffer.allocate(10), ByteBuffer.allocateDirect(10)); + } + + private void checkBufferArraySize(ByteBuffer... buffers) + { + assertEquals(meter.measureDeep(buffers, SLAB_ALLOCATION_NO_SLICE), ObjectSizes.sizeOnHeapOf(buffers)); } } \ No newline at end of file From 738ac537c5d00a7dabc166d46fb90785305a9ac2 Mon Sep 17 00:00:00 2001 From: Olivier Michallat Date: Thu, 13 Mar 2025 14:31:48 -0700 Subject: [PATCH 2/3] Fix compile issue and test failure --- .../org/apache/cassandra/cache/CachingRebuffererTest.java | 5 ++++- .../org/apache/cassandra/db/ClusteringHeapSizeTest.java | 6 ------ 2 files changed, 4 insertions(+), 7 deletions(-) diff --git a/test/unit/org/apache/cassandra/cache/CachingRebuffererTest.java b/test/unit/org/apache/cassandra/cache/CachingRebuffererTest.java index d33cddb3fd00..676a5e5a07ca 100644 --- a/test/unit/org/apache/cassandra/cache/CachingRebuffererTest.java +++ b/test/unit/org/apache/cassandra/cache/CachingRebuffererTest.java @@ -47,6 +47,7 @@ import org.apache.cassandra.utils.memory.BufferPool; import org.apache.cassandra.utils.memory.BufferPools; import org.github.jamm.MemoryMeter; +import org.github.jamm.MemoryMeter.Guess; import org.mockito.Mockito; import static org.junit.Assert.assertEquals; @@ -159,7 +160,9 @@ public ReaderType type() } Rebufferer rebufferer = ChunkCache.instance.maybeWrap(new EmptyAllocatingChunkReader()).instantiateRebufferer(); - final MemoryMeter memoryMeter = new MemoryMeter().withGuessing(MemoryMeter.Guess.FALLBACK_UNSAFE); + final MemoryMeter memoryMeter = MemoryMeter.builder().withGuessing(Guess.INSTRUMENTATION_AND_SPECIFICATION, + Guess.UNSAFE) + .build(); final long initialHeap = memoryMeter.measureDeep(ChunkCache.instance); System.out.println("initial deepSize = " + FBUtilities.prettyPrintMemory(initialHeap)); diff --git a/test/unit/org/apache/cassandra/db/ClusteringHeapSizeTest.java b/test/unit/org/apache/cassandra/db/ClusteringHeapSizeTest.java index f9f1f1a63964..2c0a2d38b165 100644 --- a/test/unit/org/apache/cassandra/db/ClusteringHeapSizeTest.java +++ b/test/unit/org/apache/cassandra/db/ClusteringHeapSizeTest.java @@ -45,12 +45,6 @@ public ClusteringHeapSizeTest(Clustering clustering) public void unsharedHeap() { long measureDeep = ObjectSizes.measureDeep(clustering); - if (clustering instanceof BufferClustering) - { - // jamm (used in measureDeep) uses .remaining() where as .sizeOnHeapOf() done in unsharedHeapSize actually looks at memory cost - // without assuming the array is shared (unless capacity > remaining); so account for that - measureDeep += ObjectSizes.measureDeep(new byte[0]); - } long unsharedHeapSize = clustering.unsharedHeapSize(); double allowedDiff = 0.1; // 10% is seen as "close enough" From 2fb7ff2c0c683db56ac186a078022c00c737133d Mon Sep 17 00:00:00 2001 From: Olivier Michallat Date: Mon, 17 Mar 2025 17:39:27 -0700 Subject: [PATCH 3/3] Fix SkipListMemtable.estimateRowOverhead --- .../cassandra/db/memtable/SkipListMemtable.java | 2 +- src/java/org/apache/cassandra/utils/ObjectSizes.java | 12 ++++++++++++ 2 files changed, 13 insertions(+), 1 deletion(-) diff --git a/src/java/org/apache/cassandra/db/memtable/SkipListMemtable.java b/src/java/org/apache/cassandra/db/memtable/SkipListMemtable.java index f9a4befdf815..d6256ce996b2 100644 --- a/src/java/org/apache/cassandra/db/memtable/SkipListMemtable.java +++ b/src/java/org/apache/cassandra/db/memtable/SkipListMemtable.java @@ -273,7 +273,7 @@ private static int estimateRowOverhead(final int count) final Object val = new Object(); for (int i = 0 ; i < count ; i++) partitions.put(cloner.clone(new BufferDecoratedKey(new LongToken(i), HeapCloner.instance.allocate(1))), val); - double avgSize = ObjectSizes.measureDeep(partitions) / (double) count; + double avgSize = ObjectSizes.measureDeepOmitShared(partitions) / (double) count; rowOverhead = (int) ((avgSize - Math.floor(avgSize)) < 0.05 ? Math.floor(avgSize) : Math.ceil(avgSize)); rowOverhead -= ObjectSizes.measureDeep(new LongToken(0)); if (!(allocator instanceof NativeAllocator)) diff --git a/src/java/org/apache/cassandra/utils/ObjectSizes.java b/src/java/org/apache/cassandra/utils/ObjectSizes.java index 4da1e2783f02..b849f3fefef5 100644 --- a/src/java/org/apache/cassandra/utils/ObjectSizes.java +++ b/src/java/org/apache/cassandra/utils/ObjectSizes.java @@ -229,6 +229,18 @@ public static long measureDeep(Object pojo) return meter.measureDeep(pojo); } + /** + * @param pojo the object to measure + * @return The size on the heap of the instance and all retained heap referenced by it, excluding portions of + * ByteBuffer that are not directly referenced by it but including any other referenced that may also be retained + * by other objects. This also includes bytes referenced in direct byte buffers, and may double-count memory if + * it is referenced by multiple ByteBuffer copies. + */ + public static long measureDeepOmitShared(Object pojo) + { + return meter.measureDeep(pojo, ByteBufferMode.SLAB_ALLOCATION_NO_SLICE); + } + /** * @param pojo the object to measure * @return the size on the heap of the instance only, excluding any referenced objects