/*
* Copyright (C) 2012, 2016 higherfrequencytrading.com
* Copyright (C) 2016 Roman Leventov
*
* This program is free software: you can redistribute it and/or modify
* it under the terms of the GNU Lesser General Public License as published by
* the Free Software Foundation, either version 3 of the License.
*
* This program is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
* GNU Lesser General Public License for more details.
*
* You should have received a copy of the GNU Lesser General Public License
* along with this program. If not, see <http://www.gnu.org/licenses/>.
*/
package net.openhft.chronicle.map;
import net.openhft.chronicle.algo.MemoryUnit;
import net.openhft.chronicle.algo.hashing.LongHashFunction;
import net.openhft.chronicle.bytes.Byteable;
import net.openhft.chronicle.bytes.Bytes;
import net.openhft.chronicle.core.Jvm;
import net.openhft.chronicle.core.OS;
import net.openhft.chronicle.hash.ChronicleHashBuilder;
import net.openhft.chronicle.hash.ChronicleHashCorruption;
import net.openhft.chronicle.hash.ChronicleHashRecoveryFailedException;
import net.openhft.chronicle.hash.impl.*;
import net.openhft.chronicle.hash.impl.stage.entry.ChecksumStrategy;
import net.openhft.chronicle.hash.impl.util.CanonicalRandomAccessFiles;
import net.openhft.chronicle.hash.impl.util.Throwables;
import net.openhft.chronicle.hash.impl.util.math.PoissonDistribution;
import net.openhft.chronicle.hash.serialization.*;
import net.openhft.chronicle.hash.serialization.impl.SerializationBuilder;
import net.openhft.chronicle.map.replication.MapRemoteOperations;
import net.openhft.chronicle.set.ChronicleSetBuilder;
import net.openhft.chronicle.values.ValueModel;
import net.openhft.chronicle.values.Values;
import net.openhft.chronicle.wire.TextWire;
import net.openhft.chronicle.wire.Wire;
import org.jetbrains.annotations.NotNull;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
import java.io.File;
import java.io.FileNotFoundException;
import java.io.IOException;
import java.io.RandomAccessFile;
import java.nio.ByteBuffer;
import java.nio.channels.FileChannel;
import java.nio.channels.FileLock;
import java.util.HashMap;
import java.util.Map;
import java.util.Objects;
import java.util.concurrent.ConcurrentHashMap;
import java.util.concurrent.TimeUnit;
import static java.lang.Double.isNaN;
import static java.lang.Math.round;
import static java.nio.ByteOrder.LITTLE_ENDIAN;
import static net.openhft.chronicle.core.Maths.*;
import static net.openhft.chronicle.hash.impl.CompactOffHeapLinearHashTable.*;
import static net.openhft.chronicle.hash.impl.SizePrefixedBlob.*;
import static net.openhft.chronicle.hash.impl.VanillaChronicleHash.throwRecoveryOrReturnIOException;
import static net.openhft.chronicle.hash.impl.util.FileIOUtils.readFully;
import static net.openhft.chronicle.hash.impl.util.FileIOUtils.writeFully;
import static net.openhft.chronicle.hash.impl.util.Objects.builderEquals;
import static net.openhft.chronicle.map.ChronicleHashCorruptionImpl.format;
import static net.openhft.chronicle.map.ChronicleHashCorruptionImpl.report;
import static net.openhft.chronicle.map.DefaultSpi.mapEntryOperations;
import static net.openhft.chronicle.map.DefaultSpi.mapRemoteOperations;
import static net.openhft.chronicle.map.VanillaChronicleMap.alignAddr;
/**
* {@code ChronicleMapBuilder} manages {@link ChronicleMap} configurations; could be used as a
* classic builder and/or factory. This means that in addition to the standard builder usage
* pattern: <pre>{@code
* ChronicleMap<Key, Value> map = ChronicleMapOnHeapUpdatableBuilder
* .of(Key.class, Value.class)
* // ... other configurations
* .create();}</pre>
* it could be prepared and used to create many similar maps: <pre>{@code
* ChronicleMapBuilder<Key, Value> builder = ChronicleMapBuilder
* .of(Key.class, Value.class)
* .entries(..)
* // ... other configurations
*
* ChronicleMap<Key, Value> map1 = builder.create();
* ChronicleMap<Key, Value> map2 = builder.create();}</pre>
* i. e. created {@code ChronicleMap} instances don't depend on the builder.
*
* <p>{@code ChronicleMapBuilder} is mutable, see a note in {@link ChronicleHashBuilder} interface
* documentation.
*
* <p>Later in this documentation, "ChronicleMap" means "ChronicleMaps, created by {@code
* ChronicleMapBuilder}", unless specified different, because theoretically someone might provide
* {@code ChronicleMap} implementations with completely different properties.
*
* <p>In addition to the key and value types, you <i>must</i> configure {@linkplain #entries(long)
* number of entries} you are going to insert into the created map <i>at most</i>. See {@link
* #entries(long)} method documentation for more information on this.
*
* <p>If you key or value type is not constantly sized and known to {@code ChronicleHashBuilder}, i.
* e. it is not a boxed primitive, {@linkplain net.openhft.chronicle.values.Values value interface},
* or {@link Byteable}, you <i>must</i> provide the {@code ChronicleHashBuilder} with some
* information about you keys or values: if they are constantly-sized, call {@link
* #constantKeySizeBySample(Object)}, otherwise {@link #averageKey(Object)} or {@link
* #averageKeySize(double)} method, and accordingly for values.
*
* <p><a name="jvm-configurations"></a>
* There are some JVM-level configurations, which are not stored in the ChronicleMap's persistence
* file (or the other way to say this: they are not parts of <a
* href="https://github.com/OpenHFT/Chronicle-Map/tree/master/spec">the Chronicle Map data store
* specification</a>) and have to be configured explicitly for each created on-heap {@code
* ChronicleMap} instance, even if it is a view of an existing Chronicle Map data store. On the
* other hand, JVM-level configurations could be different for different views of the same Chronicle
* Map data store. The list of JVM-level configurations:
* <ul>
* <li>{@link #name(String)}</li>
* <li>{@link #putReturnsNull(boolean)}</li>
* <li>{@link #removeReturnsNull(boolean)}</li>
* <li>{@link #entryOperations(MapEntryOperations)}</li>
* <li>{@link #mapMethods(MapMethods)}</li>
* <li>{@link #defaultValueProvider(DefaultValueProvider)}</li>
* </ul>
*
* @param <K> key type of the maps, produced by this builder
* @param <V> value type of the maps, produced by this builder
* @see ChronicleHashBuilder
* @see ChronicleMap
* @see ChronicleSetBuilder
*/
public final class ChronicleMapBuilder<K, V> implements
ChronicleHashBuilder<K, ChronicleMap<K, V>, ChronicleMapBuilder<K, V>> {
private static final int UNDEFINED_ALIGNMENT_CONFIG = -1;
private static final int NO_ALIGNMENT = 1;
/**
* If want to increase this number, note {@link OldDeletedEntriesCleanupThread} uses array
* to store all segment indexes -- so it could be current JVM max array size,
* not Integer.MAX_VALUE (which is an obvious limitation, as many APIs and internals use int
* type for representing segment index).
*
* Anyway, unlikely anyone ever need more than 1 billion segments.
*/
private static final int MAX_SEGMENTS = (1 << 30);
private static final Logger LOG =
LoggerFactory.getLogger(ChronicleMapBuilder.class.getName());
private static final double UNDEFINED_DOUBLE_CONFIG = Double.NaN;
private static final ConcurrentHashMap<File, Void> fileLockingControl =
new ConcurrentHashMap<>(128);
private static int MAX_BOOTSTRAPPING_HEADER_SIZE = (int) MemoryUnit.KILOBYTES.toBytes(16);
private static final Logger chronicleMapLogger = LoggerFactory.getLogger(ChronicleMap.class);
private static final ChronicleHashCorruption.Listener defaultChronicleMapCorruptionListener =
corruption -> {
if (corruption.exception() != null) {
chronicleMapLogger.error(corruption.message(), corruption.exception());
} else {
chronicleMapLogger.error(corruption.message());
}
};
private String name;
SerializationBuilder<K> keyBuilder;
SerializationBuilder<V> valueBuilder;
K averageKey;
V averageValue;
/**
* Default timeout is 1 minute. Even loopback tests converge often in the course of seconds,
* let alone WAN replication over many nodes might take tens of seconds.
* <p/>
* TODO review
*/
long cleanupTimeout = 1;
TimeUnit cleanupTimeoutUnit = TimeUnit.MINUTES;
boolean cleanupRemovedEntries = true;
//////////////////////////////
// Configuration fields
DefaultValueProvider<K, V> defaultValueProvider = DefaultSpi.defaultValueProvider();
byte replicationIdentifier = -1;
MapMethods<K, V, ?> methods = DefaultSpi.mapMethods();
MapEntryOperations<K, V, ?> entryOperations = mapEntryOperations();
MapRemoteOperations<K, V, ?> remoteOperations = mapRemoteOperations();
// not final because of cloning
private ChronicleMapBuilderPrivateAPI<K, V> privateAPI =
new ChronicleMapBuilderPrivateAPI<>(this);
// used when configuring the number of segments.
private int minSegments = -1;
private int actualSegments = -1;
// used when reading the number of entries per
private long entriesPerSegment = -1L;
private long actualChunksPerSegmentTier = -1L;
private double averageKeySize = UNDEFINED_DOUBLE_CONFIG;
private K sampleKey;
private double averageValueSize = UNDEFINED_DOUBLE_CONFIG;
private V sampleValue;
private int actualChunkSize = 0;
private int worstAlignment = -1;
private int maxChunksPerEntry = -1;
private int alignment = UNDEFINED_ALIGNMENT_CONFIG;
private long entries = -1L;
private double maxBloatFactor = 1.0;
private boolean allowSegmentTiering = true;
private double nonTieredSegmentsPercentile = 0.99999;
private boolean aligned64BitMemoryOperationsAtomic = OS.is64Bit();
private ChecksumEntries checksumEntries = ChecksumEntries.IF_PERSISTED;
private boolean putReturnsNull = false;
private boolean removeReturnsNull = false;
private boolean replicated;
private boolean persisted;
ChronicleMapBuilder(Class<K> keyClass, Class<V> valueClass) {
keyBuilder = new SerializationBuilder<>(keyClass);
valueBuilder = new SerializationBuilder<>(valueClass);
}
private static boolean isDefined(double config) {
return !isNaN(config);
}
private static long toLong(double v) {
long l = round(v);
if (l != v)
throw new IllegalArgumentException("Integer argument expected, given " + v);
return l;
}
private static long roundUp(double v) {
return round(Math.ceil(v));
}
private static long roundDown(double v) {
return (long) v;
}
//////////////////////////////
// Instance fields
/**
* When Chronicle Maps are created using {@link #createPersistedTo(File)} or
* {@link #recoverPersistedTo(File, boolean)} or {@link
* #createOrRecoverPersistedTo(File, boolean)} methods, file lock on the Chronicle Map's file is
* acquired, that shouldn't be done from concurrent threads within the same JVM process. So
* creation of Chronicle Maps persisted to the same File should be synchronized across JVM's
* threads. Simple way would be to synchronize on some static (lock) object, but would serialize
* all Chronicle Maps creations (persisted to any files), ConcurrentHashMap#compute() gives more
* scalability. ConcurrentHashMap is used effectively for lock striping only, because the
* entries are not even landing the map, because compute() always returns null.
*/
private static void fileLockedIO(
File file, FileChannel fileChannel, FileIOAction fileIOAction) throws IOException {
fileLockingControl.compute(file, (k, v) -> {
try {
try (FileLock ignored = fileChannel.lock()) {
fileIOAction.fileIOAction();
}
return null;
} catch (IOException e) {
throw Jvm.rethrow(e);
}
});
}
/**
* Returns a new {@code ChronicleMapBuilder} instance which is able to {@linkplain #create()
* create} maps with the specified key and value classes.
*
* @param keyClass class object used to infer key type and discover it's properties via
* reflection
* @param valueClass class object used to infer value type and discover it's properties via
* reflection
* @param <K> key type of the maps, created by the returned builder
* @param <V> value type of the maps, created by the returned builder
* @return a new builder for the given key and value classes
*/
public static <K, V> ChronicleMapBuilder<K, V> of(
@NotNull Class<K> keyClass, @NotNull Class<V> valueClass) {
return new ChronicleMapBuilder<>(keyClass, valueClass);
}
private static void checkSegments(long segments) {
if (segments <= 0) {
throw new IllegalArgumentException("segments should be positive, " +
segments + " given");
}
if (segments > MAX_SEGMENTS) {
throw new IllegalArgumentException("Max segments is " + MAX_SEGMENTS + ", " +
segments + " given");
}
}
private static String pretty(int value) {
return value > 0 ? value + "" : "not configured";
}
private static String pretty(Object obj) {
return obj != null ? obj + "" : "not configured";
}
private static void checkSizeIsStaticallyKnown(SerializationBuilder builder, String role) {
if (builder.sizeIsStaticallyKnown) {
throw new IllegalStateException("Size of " + builder.tClass +
" instances is constant and statically known, shouldn't be specified via " +
"average" + role + "Size() or average" + role + "() methods");
}
}
private static void checkAverageSize(double averageSize, String role) {
if (averageSize <= 0 || isNaN(averageSize) ||
Double.isInfinite(averageSize)) {
throw new IllegalArgumentException("Average " + role + " size must be a positive, " +
"finite number");
}
}
private static double averageSizeStoringLength(
SerializationBuilder builder, double averageSize) {
SizeMarshaller sizeMarshaller = builder.sizeMarshaller();
if (averageSize == round(averageSize))
return sizeMarshaller.storingLength(round(averageSize));
long lower = roundDown(averageSize);
long upper = lower + 1;
int lowerStoringLength = sizeMarshaller.storingLength(lower);
int upperStoringLength = sizeMarshaller.storingLength(upper);
if (lowerStoringLength == upperStoringLength)
return lowerStoringLength;
return lower * (upper - averageSize) + upper * (averageSize - lower);
}
static int greatestCommonDivisor(int a, int b) {
if (b == 0) return a;
return greatestCommonDivisor(b, a % b);
}
private static int maxDefaultChunksPerAverageEntry(boolean replicated) {
// When replicated, having 8 chunks (=> 8 bits in bitsets) per entry seems more wasteful
// because when replicated we have bit sets per each remote node, not only allocation
// bit set as when non-replicated
return replicated ? 4 : 8;
}
private static int estimateSegmentsForEntries(long size) {
if (size > 200 << 20)
return 256;
if (size >= 1 << 20)
return 128;
if (size >= 128 << 10)
return 64;
if (size >= 16 << 10)
return 32;
if (size >= 4 << 10)
return 16;
if (size >= 1 << 10)
return 8;
return 1;
}
private static long headerChecksum(ByteBuffer headerBuffer, int headerSize) {
return LongHashFunction.xx_r39().hashBytes(headerBuffer, SIZE_WORD_OFFSET, headerSize + 4);
}
private static void writeNotComplete(
FileChannel fileChannel, ByteBuffer headerBuffer, int headerSize) throws IOException {
//noinspection PointlessBitwiseExpression
headerBuffer.putInt(SIZE_WORD_OFFSET, NOT_COMPLETE | DATA | headerSize);
headerBuffer.clear().position(SIZE_WORD_OFFSET).limit(SIZE_WORD_OFFSET + 4);
writeFully(fileChannel, SIZE_WORD_OFFSET, headerBuffer);
}
/**
* @return ByteBuffer, with self bootstrapping header in [position, limit) range
*/
private static <K, V> ByteBuffer writeHeader(
FileChannel fileChannel, VanillaChronicleMap<K, V, ?> map) throws IOException {
ByteBuffer headerBuffer = ByteBuffer.allocate(
SELF_BOOTSTRAPPING_HEADER_OFFSET + MAX_BOOTSTRAPPING_HEADER_SIZE);
headerBuffer.order(LITTLE_ENDIAN);
Bytes<ByteBuffer> headerBytes = Bytes.wrapForWrite(headerBuffer);
headerBytes.writePosition(SELF_BOOTSTRAPPING_HEADER_OFFSET);
Wire wire = new TextWire(headerBytes);
wire.getValueOut().typedMarshallable(map);
int headerLimit = (int) headerBytes.writePosition();
int headerSize = headerLimit - SELF_BOOTSTRAPPING_HEADER_OFFSET;
// First set readiness bit to READY, to compute checksum correctly
//noinspection PointlessBitwiseExpression
headerBuffer.putInt(SIZE_WORD_OFFSET, READY | DATA | headerSize);
long checksum = headerChecksum(headerBuffer, headerSize);
headerBuffer.putLong(HEADER_OFFSET, checksum);
// Set readiness bit to NOT_COMPLETE, because the Chronicle Map instance is not actually
// ready yet
//noinspection PointlessBitwiseExpression
headerBuffer.putInt(SIZE_WORD_OFFSET, NOT_COMPLETE | DATA | headerSize);
// Write the size-prefixed blob to the file
headerBuffer.position(0).limit(headerLimit);
writeFully(fileChannel, 0, headerBuffer);
headerBuffer.position(SELF_BOOTSTRAPPING_HEADER_OFFSET);
return headerBuffer;
}
private static void commitChronicleMapReady(
VanillaChronicleHash map, RandomAccessFile raf, ByteBuffer headerBuffer, int headerSize)
throws IOException {
FileChannel fileChannel = raf.getChannel();
// see https://higherfrequencytrading.atlassian.net/browse/HCOLL-396
map.msync();
//noinspection PointlessBitwiseExpression
headerBuffer.putInt(SIZE_WORD_OFFSET, READY | DATA | headerSize);
headerBuffer.clear().position(SIZE_WORD_OFFSET).limit(SIZE_WORD_OFFSET + 4);
writeFully(fileChannel, SIZE_WORD_OFFSET, headerBuffer);
}
@Override
public ChronicleMapBuilder<K, V> clone() {
try {
@SuppressWarnings("unchecked")
ChronicleMapBuilder<K, V> result =
(ChronicleMapBuilder<K, V>) super.clone();
result.keyBuilder = keyBuilder.clone();
result.valueBuilder = valueBuilder.clone();
result.privateAPI = new ChronicleMapBuilderPrivateAPI<>(result);
return result;
} catch (CloneNotSupportedException e) {
throw new AssertionError(e);
}
}
/**
* @deprecated don't use private API in the client code
*/
@Override
@Deprecated
public Object privateAPI() {
return privateAPI;
}
/**
* {@inheritDoc}
* <a href="#jvm-configurations">Read more about JVM-level configurations</a>.
*/
@Override
public ChronicleMapBuilder<K, V> name(String name) {
this.name = name;
return this;
}
String name() {
return this.name;
}
/**
* {@inheritDoc}
*
* <p>Example: if keys in your map(s) are English words in {@link String} form, average English
* word length is 5.1, configure average key size of 6: <pre>{@code
* ChronicleMap<String, LongValue> wordFrequencies = ChronicleMapBuilder
* .of(String.class, LongValue.class)
* .entries(50000)
* .averageKeySize(6)
* .create();}</pre>
* (Note that 6 is chosen as average key size in bytes despite strings in Java are UTF-16
* encoded (and each character takes 2 bytes on-heap), because default off-heap {@link String}
* encoding is UTF-8 in {@code ChronicleMap}.)
*
* @param averageKeySize the average size of the key
* @throws IllegalStateException {@inheritDoc}
* @throws IllegalArgumentException {@inheritDoc}
* @see #averageKey(Object)
* @see #constantKeySizeBySample(Object)
* @see #averageValueSize(double)
* @see #actualChunkSize(int)
*/
@Override
public ChronicleMapBuilder<K, V> averageKeySize(double averageKeySize) {
checkSizeIsStaticallyKnown(keyBuilder, "Key");
checkAverageSize(averageKeySize, "key");
this.averageKeySize = averageKeySize;
averageKey = null;
sampleKey = null;
return this;
}
/**
* {@inheritDoc}
*
* @param averageKey the average (by footprint in serialized form) key, is going to be put
* into the hash containers, created by this builder
* @throws NullPointerException {@inheritDoc}
* @see #averageKeySize(double)
* @see #constantKeySizeBySample(Object)
* @see #averageValue(Object)
* @see #actualChunkSize(int)
*/
@Override
public ChronicleMapBuilder<K, V> averageKey(K averageKey) {
Objects.requireNonNull(averageKey);
checkSizeIsStaticallyKnown(keyBuilder, "Key");
this.averageKey = averageKey;
sampleKey = null;
averageKeySize = UNDEFINED_DOUBLE_CONFIG;
return this;
}
/**
* {@inheritDoc}
*
* <p>For example, if your keys are Git commit hashes:<pre>{@code
* Map<byte[], String> gitCommitMessagesByHash =
* ChronicleMapBuilder.of(byte[].class, String.class)
* .constantKeySizeBySample(new byte[20])
* .create();}</pre>
*
* @see #averageKeySize(double)
* @see #averageKey(Object)
* @see #constantValueSizeBySample(Object)
*/
@Override
public ChronicleMapBuilder<K, V> constantKeySizeBySample(K sampleKey) {
this.sampleKey = sampleKey;
averageKey = null;
averageKeySize = UNDEFINED_DOUBLE_CONFIG;
return this;
}
private double averageKeySize() {
if (!isDefined(averageKeySize))
throw new AssertionError();
return averageKeySize;
}
/**
* Configures the average number of bytes, taken by serialized form of values, put into maps,
* created by this builder. However, in many cases {@link #averageValue(Object)} might be easier
* to use and more reliable. If value size is always the same, call {@link
* #constantValueSizeBySample(Object)} method instead of this one.
*
* <p>{@code ChronicleHashBuilder} implementation heuristically chooses {@linkplain
* #actualChunkSize(int) the actual chunk size} based on this configuration and the key size,
* that, however, might result to quite high internal fragmentation, i. e. losses because only
* integral number of chunks could be allocated for the entry. If you want to avoid this, you
* should manually configure the actual chunk size in addition to this average value size
* configuration, which is anyway needed.
*
* <p>If values are of boxed primitive type or {@link Byteable} subclass, i. e. if value size is
* known statically, it is automatically accounted and shouldn't be specified by user.
*
* <p>Calling this method clears any previous {@link #constantValueSizeBySample(Object)} and
* {@link #averageValue(Object)} configurations.
*
* @param averageValueSize number of bytes, taken by serialized form of values
* @return this builder back
* @throws IllegalStateException if value size is known statically and shouldn't be
* configured by user
* @throws IllegalArgumentException if the given {@code averageValueSize} is non-positive
* @see #averageValue(Object)
* @see #constantValueSizeBySample(Object)
* @see #averageKeySize(double)
* @see #actualChunkSize(int)
*/
public ChronicleMapBuilder<K, V> averageValueSize(double averageValueSize) {
checkSizeIsStaticallyKnown(valueBuilder, "Value");
checkAverageSize(averageValueSize, "value");
this.averageValueSize = averageValueSize;
averageValue = null;
sampleValue = null;
return this;
}
/**
* Configures the average number of bytes, taken by serialized form of values, put into maps,
* created by this builder, by serializing the given {@code averageValue} using the configured
* {@link #valueMarshallers(SizedReader, SizedWriter) value marshallers}. In some cases, {@link
* #averageValueSize(double)} might be easier to use, than constructing the "average value".
* If value size is always the same, call {@link #constantValueSizeBySample(Object)} method
* instead of this one.
*
* <p>{@code ChronicleHashBuilder} implementation heuristically chooses {@linkplain
* #actualChunkSize(int) the actual chunk size} based on this configuration and the key size,
* that, however, might result to quite high internal fragmentation, i. e. losses because only
* integral number of chunks could be allocated for the entry. If you want to avoid this, you
* should manually configure the actual chunk size in addition to this average value size
* configuration, which is anyway needed.
*
* <p>If values are of boxed primitive type or {@link Byteable} subclass, i. e. if value size is
* known statically, it is automatically accounted and shouldn't be specified by user.
*
* <p>Calling this method clears any previous {@link #constantValueSizeBySample(Object)}
* and {@link #averageValueSize(double)} configurations.
*
* @param averageValue the average (by footprint in serialized form) value, is going to be put
* into the maps, created by this builder
* @return this builder back
* @throws NullPointerException if the given {@code averageValue} is {@code null}
* @see #averageValueSize(double)
* @see #constantValueSizeBySample(Object)
* @see #averageKey(Object)
* @see #actualChunkSize(int)
*/
public ChronicleMapBuilder<K, V> averageValue(V averageValue) {
Objects.requireNonNull(averageValue);
checkSizeIsStaticallyKnown(valueBuilder, "Value");
this.averageValue = averageValue;
sampleValue = null;
averageValueSize = UNDEFINED_DOUBLE_CONFIG;
return this;
}
/**
* Configures the constant number of bytes, taken by serialized form of values, put into maps,
* created by this builder. This is done by providing the {@code sampleValue}, all values should
* take the same number of bytes in serialized form, as this sample object.
*
* <p>If values are of boxed primitive type or {@link Byteable} subclass, i. e. if value size is
* known statically, it is automatically accounted and this method shouldn't be called.
*
* <p>If value size varies, method {@link #averageValue(Object)} or {@link
* #averageValueSize(double)} should be called instead of this one.
*
* <p>Calling this method clears any previous {@link #averageValue(Object)} and
* {@link #averageValueSize(double)} configurations.
*
* @param sampleValue the sample value
* @return this builder back
* @see #averageValueSize(double)
* @see #averageValue(Object)
* @see #constantKeySizeBySample(Object)
*/
public ChronicleMapBuilder<K, V> constantValueSizeBySample(V sampleValue) {
this.sampleValue = sampleValue;
averageValue = null;
averageValueSize = UNDEFINED_DOUBLE_CONFIG;
return this;
}
double averageValueSize() {
if (!isDefined(averageValueSize))
throw new AssertionError();
return averageValueSize;
}
private <E> double averageKeyOrValueSize(
double configuredSize, SerializationBuilder<E> builder, E average) {
if (isDefined(configuredSize))
return configuredSize;
if (builder.constantSizeMarshaller())
return builder.constantSize();
if (average != null) {
return builder.serializationSize(average);
}
return Double.NaN;
}
/**
* {@inheritDoc}
*
* @throws IllegalStateException is sizes of both keys and values of maps created by this
* builder are constant, hence chunk size shouldn't be configured by user
* @see #entryAndValueOffsetAlignment(int)
* @see #entries(long)
* @see #maxChunksPerEntry(int)
*/
@Override
public ChronicleMapBuilder<K, V> actualChunkSize(int actualChunkSize) {
if (constantlySizedEntries()) {
throw new IllegalStateException("Sizes of key type: " + keyBuilder.tClass + " and " +
"value type: " + valueBuilder.tClass + " are both constant, " +
"so chunk size shouldn't be specified manually");
}
if (actualChunkSize <= 0)
throw new IllegalArgumentException("Chunk size must be positive");
this.actualChunkSize = actualChunkSize;
return this;
}
SerializationBuilder<K> keyBuilder() {
return keyBuilder;
}
private EntrySizeInfo entrySizeInfo() {
double size = 0;
double keySize = averageKeySize();
size += averageSizeStoringLength(keyBuilder, keySize);
size += keySize;
if (replicated)
size += ReplicatedChronicleMap.ADDITIONAL_ENTRY_BYTES;
if (checksumEntries())
size += ChecksumStrategy.CHECKSUM_STORED_BYTES;
double valueSize = averageValueSize();
size += averageSizeStoringLength(valueBuilder, valueSize);
int alignment = valueAlignment();
int worstAlignment;
if (worstAlignmentComputationRequiresValueSize(alignment)) {
long constantSizeBeforeAlignment = toLong(size);
if (constantlySizedValues()) {
// see tierEntrySpaceInnerOffset()
long totalDataSize = constantSizeBeforeAlignment + constantValueSize();
worstAlignment = (int) (alignAddr(totalDataSize, alignment) - totalDataSize);
} else {
determineAlignment:
if (actualChunkSize > 0) {
worstAlignment = worstAlignmentAssumingChunkSize(constantSizeBeforeAlignment,
actualChunkSize);
} else {
int chunkSize = 8;
worstAlignment = worstAlignmentAssumingChunkSize(
constantSizeBeforeAlignment, chunkSize);
if (size + worstAlignment + valueSize >=
maxDefaultChunksPerAverageEntry(replicated) * chunkSize) {
break determineAlignment;
}
chunkSize = 4;
worstAlignment = worstAlignmentAssumingChunkSize(
constantSizeBeforeAlignment, chunkSize);
}
}
} else {
// assume worst case, we always lose most possible bytes for alignment
worstAlignment = worstAlignmentWithoutValueSize(alignment);
}
size += worstAlignment;
size += valueSize;
return new EntrySizeInfo(size, worstAlignment);
}
private boolean worstAlignmentComputationRequiresValueSize(int alignment) {
return alignment != NO_ALIGNMENT &&
constantlySizedKeys() && valueBuilder.constantStoringLengthSizeMarshaller();
}
private int worstAlignmentWithoutValueSize(int alignment) {
return alignment - 1;
}
int segmentEntrySpaceInnerOffset() {
// This is needed, if chunkSize = constant entry size is not aligned, for entry alignment
// to be always the same, we should _misalign_ the first chunk.
if (!constantlySizedEntries())
return 0;
return (int) (constantValueSize() % valueAlignment());
}
private long constantValueSize() {
return valueBuilder.constantSize();
}
boolean constantlySizedKeys() {
return keyBuilder.constantSizeMarshaller() || sampleKey != null;
}
private int worstAlignmentAssumingChunkSize(
long constantSizeBeforeAlignment, int chunkSize) {
int alignment = valueAlignment();
long firstAlignment = alignAddr(constantSizeBeforeAlignment, alignment) -
constantSizeBeforeAlignment;
int gcdOfAlignmentAndChunkSize = greatestCommonDivisor(alignment, chunkSize);
if (gcdOfAlignmentAndChunkSize == alignment)
return (int) firstAlignment;
// assume worst by now because we cannot predict alignment in VanillaCM.entrySize() method
// before allocation
long worstAlignment = firstAlignment;
while (worstAlignment + gcdOfAlignmentAndChunkSize < alignment)
worstAlignment += gcdOfAlignmentAndChunkSize;
return (int) worstAlignment;
}
int worstAlignment() {
if (worstAlignment >= 0)
return worstAlignment;
int alignment = valueAlignment();
if (!worstAlignmentComputationRequiresValueSize(alignment))
return worstAlignment = worstAlignmentWithoutValueSize(alignment);
return worstAlignment = entrySizeInfo().worstAlignment;
}
void worstAlignment(int worstAlignment) {
assert worstAlignment >= 0;
this.worstAlignment = worstAlignment;
}
long chunkSize() {
if (actualChunkSize > 0)
return actualChunkSize;
double averageEntrySize = entrySizeInfo().averageEntrySize;
if (constantlySizedEntries())
return toLong(averageEntrySize);
int maxChunkSize = 1 << 30;
for (long chunkSize = 4; chunkSize <= maxChunkSize; chunkSize *= 2L) {
if (maxDefaultChunksPerAverageEntry(replicated) * chunkSize > averageEntrySize)
return chunkSize;
}
return maxChunkSize;
}
boolean constantlySizedEntries() {
return constantlySizedKeys() && constantlySizedValues();
}
double averageChunksPerEntry() {
if (constantlySizedEntries())
return 1.0;
long chunkSize = chunkSize();
// assuming we always has worst internal fragmentation. This affects total segment
// entry space which is allocated lazily on Linux (main target platform)
// so we can afford this
return (entrySizeInfo().averageEntrySize + chunkSize - 1) / chunkSize;
}
@Override
public ChronicleMapBuilder<K, V> maxChunksPerEntry(int maxChunksPerEntry) {
if (maxChunksPerEntry < 1)
throw new IllegalArgumentException("maxChunksPerEntry should be >= 1, " +
maxChunksPerEntry + " given");
this.maxChunksPerEntry = maxChunksPerEntry;
return this;
}
int maxChunksPerEntry() {
if (constantlySizedEntries())
return 1;
long actualChunksPerSegmentTier = actualChunksPerSegmentTier();
int result = (int) Math.min(actualChunksPerSegmentTier, (long) Integer.MAX_VALUE);
if (this.maxChunksPerEntry > 0)
result = Math.min(this.maxChunksPerEntry, result);
return result;
}
boolean constantlySizedValues() {
return valueBuilder.constantSizeMarshaller() || sampleValue != null;
}
/**
* Configures alignment of address in memory of entries and independently of address in memory
* of values within entries ((i. e. final addresses in native memory are multiples of the given
* alignment) for ChronicleMaps, created by this builder.
*
* <p>Useful when values of the map are updated intensively, particularly fields with volatile
* access, because it doesn't work well if the value crosses cache lines. Also, on some
* (nowadays rare) architectures any misaligned memory access is more expensive than aligned.
*
* <p>If values couldn't reference off-heap memory (i. e. it is not {@link Byteable} or a value
* interface), alignment configuration makes no sense.
*
* <p>Default is {@link ValueModel#recommendedOffsetAlignment()} if the value type is a value
* interface, otherwise 1 (that is effectively no alignment) or chosen heuristically (configure
* explicitly for being sure and to compare performance in your case).
*
* @param alignment the new alignment of the maps constructed by this builder
* @return this builder back
* @throws IllegalStateException if values of maps, created by this builder, couldn't reference
* off-heap memory
*/
public ChronicleMapBuilder<K, V> entryAndValueOffsetAlignment(int alignment) {
if (alignment <= 0) {
throw new IllegalArgumentException("Alignment should be positive integer, " +
alignment + " given");
}
if (!isPowerOf2(alignment)) {
throw new IllegalArgumentException("Alignment should be a power of 2, " + alignment +
" given");
}
this.alignment = alignment;
return this;
}
int valueAlignment() {
if (alignment != UNDEFINED_ALIGNMENT_CONFIG)
return alignment;
try {
if (Values.isValueInterfaceOrImplClass(valueBuilder.tClass)) {
return ValueModel.acquire(valueBuilder.tClass).recommendedOffsetAlignment();
} else {
return NO_ALIGNMENT;
}
} catch (Exception e) {
return NO_ALIGNMENT;
}
}
@Override
public ChronicleMapBuilder<K, V> entries(long entries) {
if (entries <= 0L)
throw new IllegalArgumentException("Entries should be positive, " + entries + " given");
this.entries = entries;
return this;
}
long entries() {
if (entries < 0) {
throw new IllegalStateException("If in-memory Chronicle Map is created or persisted\n" +
"to a file for the first time (i. e. not accessing existing file),\n" +
"ChronicleMapBuilder.entries() must be configured.\n" +
"See Chronicle Map 3 tutorial and javadocs for more information");
}
return entries;
}
@Override
public ChronicleMapBuilder<K, V> entriesPerSegment(long entriesPerSegment) {
if (entriesPerSegment <= 0L)
throw new IllegalArgumentException("Entries per segment should be positive, " +
entriesPerSegment + " given");
this.entriesPerSegment = entriesPerSegment;
return this;
}
long entriesPerSegment() {
long entriesPerSegment;
if (this.entriesPerSegment > 0L) {
entriesPerSegment = this.entriesPerSegment;
} else {
int actualSegments = actualSegments();
double averageEntriesPerSegment = entries() * 1.0 / actualSegments;
if (actualSegments > 1) {
entriesPerSegment = PoissonDistribution.inverseCumulativeProbability(
averageEntriesPerSegment, nonTieredSegmentsPercentile);
} else {
// if there is only 1 segment, there is no source of variance in segments filling
entriesPerSegment = roundUp(averageEntriesPerSegment);
}
}
boolean actualChunksDefined = actualChunksPerSegmentTier > 0;
if (!actualChunksDefined) {
double averageChunksPerEntry = averageChunksPerEntry();
if (entriesPerSegment * averageChunksPerEntry >
MAX_TIER_CHUNKS)
throw new IllegalStateException("Max chunks per segment tier is " +
MAX_TIER_CHUNKS + " configured entries() and actualSegments() so that " +
"there should be " + entriesPerSegment + " entries per segment tier, " +
"while average chunks per entry is " + averageChunksPerEntry);
}
if (entriesPerSegment > MAX_TIER_ENTRIES)
throw new IllegalStateException("shouldn't be more than " +
MAX_TIER_ENTRIES + " entries per segment");
return entriesPerSegment;
}
@Override
public ChronicleMapBuilder<K, V> actualChunksPerSegmentTier(long actualChunksPerSegmentTier) {
if (actualChunksPerSegmentTier <= 0 || actualChunksPerSegmentTier > MAX_TIER_CHUNKS)
throw new IllegalArgumentException("Actual chunks per segment tier should be in [1, " +
MAX_TIER_CHUNKS + "], range, " + actualChunksPerSegmentTier + " given");
this.actualChunksPerSegmentTier = actualChunksPerSegmentTier;
return this;
}
private void checkActualChunksPerSegmentTierIsConfiguredOnlyIfOtherLowLevelConfigsAreManual() {
if (actualChunksPerSegmentTier > 0) {
if (entriesPerSegment <= 0 || (actualChunkSize <= 0 && !constantlySizedEntries()) ||
actualSegments <= 0)
throw new IllegalStateException("Actual chunks per segment tier could be " +
"configured only if other three low level configs are manual: " +
"entriesPerSegment(), actualSegments() and actualChunkSize(), unless " +
"both keys and value sizes are constant");
}
}
private void checkActualChunksPerSegmentGreaterOrEqualToEntries() {
if (actualChunksPerSegmentTier > 0 && entriesPerSegment > 0 &&
entriesPerSegment > actualChunksPerSegmentTier) {
throw new IllegalStateException("Entries per segment couldn't be greater than " +
"actual chunks per segment tier. Entries: " + entriesPerSegment + ", " +
"chunks: " + actualChunksPerSegmentTier + " is configured");
}
}
long actualChunksPerSegmentTier() {
if (actualChunksPerSegmentTier > 0)
return actualChunksPerSegmentTier;
return chunksPerSegmentTier(entriesPerSegment());
}
private long chunksPerSegmentTier(long entriesPerSegment) {
return roundUp(entriesPerSegment * averageChunksPerEntry());
}
@Override
public ChronicleMapBuilder<K, V> minSegments(int minSegments) {
checkSegments(minSegments);
this.minSegments = minSegments;
return this;
}
int minSegments() {
return Math.max(estimateSegments(), minSegments);
}
private int estimateSegments() {
return (int) Math.min(nextPower2(entries() / 32, 1), estimateSegmentsBasedOnSize());
}
//TODO review because this heuristic doesn't seem to perform well
private int estimateSegmentsBasedOnSize() {
// the idea is that if values are huge, operations on them (and simply ser/deser)
// could take long time, so we want more segment to minimize probablity that
// two or more concurrent write ops will go to the same segment, and then all but one of
// these threads will wait for long time.
int segmentsForEntries = estimateSegmentsForEntries(entries());
double averageValueSize = averageValueSize();
return averageValueSize >= 1000000
? segmentsForEntries * 16
: averageValueSize >= 100000
? segmentsForEntries * 8
: averageValueSize >= 10000
? segmentsForEntries * 4
: averageValueSize >= 1000
? segmentsForEntries * 2
: segmentsForEntries;
}
@Override
public ChronicleMapBuilder<K, V> actualSegments(int actualSegments) {
checkSegments(actualSegments);
this.actualSegments = actualSegments;
return this;
}
int actualSegments() {
if (actualSegments > 0)
return actualSegments;
if (entriesPerSegment > 0) {
return (int) segmentsGivenEntriesPerSegmentFixed(entriesPerSegment);
}
// Try to fit 4 bytes per hash lookup slot, then 8. Trying to apply small slot
// size (=> segment size, because slot size depends on segment size) not only because
// they take less memory per entry (if entries are of KBs or MBs, it doesn't matter), but
// also because if segment size is small, slot and free list are likely to lie on a single
// memory page, reducing number of memory pages to update, if Chronicle Map is persisted.
// Actually small segments are all ways better: many segments => better parallelism, lesser
// pauses for per-key operations, if parallel/background operation blocks the segment for
// the whole time while it operates on it (like iteration, probably replication background
// thread will require some level of full segment lock, however currently if doesn't, in
// future durability background thread could update slot states), because smaller segments
// contain less entries/slots and are processed faster.
//
// The only problem with small segments is that due to probability theory, if there are
// a lot of segments each of little number of entries, difference between most filled
// and least filled segment in the Chronicle Map grows. (Number of entries in a segment is
// Poisson-distributed with mean = average number of entries per segment.) It is meaningful,
// because segment tiering is exceptional mechanism, only very few segments should be
// tiered, if any, normally. So, we are required to allocate unnecessarily many entries per
// each segment. To compensate this at least on linux, don't accept segment sizes that with
// the given entry sizes, lead to too small total segment sizes in native memory pages,
// see comment in tryHashLookupSlotSize()
long segments = tryHashLookupSlotSize(4);
if (segments > 0)
return (int) segments;
int maxHashLookupEntrySize = aligned64BitMemoryOperationsAtomic() ? 8 : 4;
long maxEntriesPerSegment =
findMaxEntriesPerSegmentToFitHashLookupSlotSize(maxHashLookupEntrySize);
long maxSegments = trySegments(maxEntriesPerSegment, MAX_SEGMENTS);
if (maxSegments > 0L)
return (int) maxSegments;
throw new IllegalStateException("Max segments is " + MAX_SEGMENTS + ", configured so much" +
" entries (" + entries() + ") or average chunks per entry is too high (" +
averageChunksPerEntry() + ") that builder automatically decided to use " +
(-maxSegments) + " segments");
}
private long tryHashLookupSlotSize(int hashLookupSlotSize) {
long entriesPerSegment = findMaxEntriesPerSegmentToFitHashLookupSlotSize(
hashLookupSlotSize);
long entrySpaceSize = roundUp(entriesPerSegment * entrySizeInfo().averageEntrySize);
// Not to lose too much on linux because of "poor distribution" entry over-allocation.
// This condition should likely filter cases when we target very small hash lookup
// size + entry size is small.
// * 5 => segment will lose not more than 20% of memory, 10% on average
if (entrySpaceSize < OS.pageSize() * 5L)
return -1;
return trySegments(entriesPerSegment, MAX_SEGMENTS);
}
private long findMaxEntriesPerSegmentToFitHashLookupSlotSize(
int targetHashLookupSlotSize) {
long entriesPerSegment = 1L << 62;
long step = entriesPerSegment / 2L;
while (step > 0L) {
if (hashLookupSlotBytes(entriesPerSegment) > targetHashLookupSlotSize)
entriesPerSegment -= step;
step /= 2L;
}
return entriesPerSegment - 1L;
}
private int hashLookupSlotBytes(long entriesPerSegment) {
int valueBits = valueBits(chunksPerSegmentTier(entriesPerSegment));
int keyBits = keyBits(entriesPerSegment, valueBits);
return entrySize(keyBits, valueBits);
}
private long trySegments(long entriesPerSegment, int maxSegments) {
long segments = segmentsGivenEntriesPerSegmentFixed(entriesPerSegment);
segments = nextPower2(Math.max(segments, minSegments()), 1L);
return segments <= maxSegments ? segments : -segments;
}
private long segmentsGivenEntriesPerSegmentFixed(long entriesPerSegment) {
double precision = 1.0 / averageChunksPerEntry();
long entriesPerSegmentShouldBe =
roundDown(PoissonDistribution.meanByCumulativeProbabilityAndValue(
nonTieredSegmentsPercentile, entriesPerSegment, precision));
long segments = divideRoundUp(entries(), entriesPerSegmentShouldBe);
checkSegments(segments);
if (minSegments > 0)
segments = Math.max(minSegments, segments);
return segments;
}
long tierHashLookupCapacity() {
long entriesPerSegment = entriesPerSegment();
long capacity = CompactOffHeapLinearHashTable.capacityFor(entriesPerSegment);
if (actualSegments() > 1) {
// if there is only 1 segment, there is no source of variance in segments filling
long maxEntriesPerTier = PoissonDistribution.inverseCumulativeProbability(
entriesPerSegment, nonTieredSegmentsPercentile);
while (maxEntriesPerTier > MAX_LOAD_FACTOR * capacity) {
capacity *= 2;
}
}
return capacity;
}
int segmentHeaderSize() {
int segments = actualSegments();
long pageSize = OS.pageSize();
if (segments * (64 * 3) < (2 * pageSize)) // i. e. <= 42 segments, if page size is 4K
return 64 * 3; // cache line per header, plus one CL to the left, plus one to the right
if (segments * (64 * 2) < (3 * pageSize)) // i. e. <= 96 segments, if page size is 4K
return 64 * 2;
// reduce false sharing unless we have a lot of segments.
return segments <= 16 * 1024 ? 64 : 32;
}
/**
* Configures if the maps created by this {@code ChronicleMapBuilder} should return {@code null}
* instead of previous mapped values on {@link ChronicleMap#put(Object, Object)
* ChornicleMap.put(key, value)} calls.
*
* <p>{@link Map#put(Object, Object) Map.put()} returns the previous value, functionality
* which is rarely used but fairly cheap for simple in-process, on-heap implementations like
* {@link HashMap}. But an off-heap collection has to create a new object and deserialize
* the data from off-heap memory. A collection hiding remote queries over the network should
* send the value back in addition to that. It's expensive for something you probably don't use.
*
* <p>This is a <a href="#jvm-configurations">JVM-level configuration</a>.
*
* <p>By default, {@code ChronicleMap} conforms the general {@code Map} contract and returns the
* previous mapped value on {@code put()} calls.
*
* @param putReturnsNull {@code true} if you want {@link ChronicleMap#put(Object, Object)
* ChronicleMap.put()} to not return the value that was replaced but
* instead return {@code null}
* @return this builder back
* @see #removeReturnsNull(boolean)
*/
public ChronicleMapBuilder<K, V> putReturnsNull(boolean putReturnsNull) {
this.putReturnsNull = putReturnsNull;
return this;
}
boolean putReturnsNull() {
return putReturnsNull;
}
/**
* Configures if the maps created by this {@code ChronicleMapBuilder} should return {@code null}
* instead of the last mapped value on {@link ChronicleMap#remove(Object)
* ChronicleMap.remove(key)} calls.
*
* <p>{@link Map#remove(Object) Map.remove()} returns the previous value, functionality which is
* rarely used but fairly cheap for simple in-process, on-heap implementations like {@link
* HashMap}. But an off-heap collection has to create a new object and deserialize the data
* from off-heap memory. A collection hiding remote queries over the network should send
* the value back in addition to that. It's expensive for something you probably don't use.
*
* <p>This is a <a href="#jvm-configurations">JVM-level configuration</a>.
*
* <p>By default, {@code ChronicleMap} conforms the general {@code Map} contract and returns the
* mapped value on {@code remove()} calls.
*
* @param removeReturnsNull {@code true} if you want {@link ChronicleMap#remove(Object)
* ChronicleMap.remove()} to not return the value of the removed entry
* but instead return {@code null}
* @return this builder back
* @see #putReturnsNull(boolean)
*/
public ChronicleMapBuilder<K, V> removeReturnsNull(boolean removeReturnsNull) {
this.removeReturnsNull = removeReturnsNull;
return this;
}
boolean removeReturnsNull() {
return removeReturnsNull;
}
@Override
public ChronicleMapBuilder<K, V> maxBloatFactor(double maxBloatFactor) {
if (isNaN(maxBloatFactor) || maxBloatFactor < 1.0 || maxBloatFactor > 1_000.0) {
throw new IllegalArgumentException("maxBloatFactor should be in [1.0, 1_000.0] " +
"bounds, " + maxBloatFactor + " given");
}
this.maxBloatFactor = maxBloatFactor;
return this;
}
@Override
public ChronicleMapBuilder<K, V> allowSegmentTiering(boolean allowSegmentTiering) {
this.allowSegmentTiering = allowSegmentTiering;
return this;
}
@Override
public ChronicleMapBuilder<K, V> nonTieredSegmentsPercentile(
double nonTieredSegmentsPercentile) {
if (isNaN(nonTieredSegmentsPercentile) ||
0.5 <= nonTieredSegmentsPercentile || nonTieredSegmentsPercentile >= 1.0) {
throw new IllegalArgumentException("nonTieredSegmentsPercentile should be in (0.5, " +
"1.0) range, " + nonTieredSegmentsPercentile + " is given");
}
this.nonTieredSegmentsPercentile = nonTieredSegmentsPercentile;
return this;
}
long maxExtraTiers() {
if (!allowSegmentTiering)
return 0;
int actualSegments = actualSegments();
// maxBloatFactor is scale, so we do (- 1.0) to compute _extra_ tiers
return round((maxBloatFactor - 1.0) * actualSegments)
// but to mitigate slight misconfiguration, and uneven distribution of entries
// between segments, add 1.0 x actualSegments
+ actualSegments;
}
@Override
public String toString() {
return "ChronicleMapBuilder{" +
", actualSegments=" + pretty(actualSegments) +
", minSegments=" + pretty(minSegments) +
", entriesPerSegment=" + pretty(entriesPerSegment) +
", actualChunksPerSegmentTier=" + pretty(actualChunksPerSegmentTier) +
", averageKeySize=" + pretty(averageKeySize) +
", sampleKeyForConstantSizeComputation=" + pretty(sampleKey) +
", averageValueSize=" + pretty(averageValueSize) +
", sampleValueForConstantSizeComputation=" + pretty(sampleValue) +
", actualChunkSize=" + pretty(actualChunkSize) +
", valueAlignment=" + valueAlignment() +
", entries=" + entries() +
", putReturnsNull=" + putReturnsNull() +
", removeReturnsNull=" + removeReturnsNull() +
", keyBuilder=" + keyBuilder +
", valueBuilder=" + valueBuilder +
'}';
}
@SuppressWarnings("EqualsWhichDoesntCheckParameterClass")
@Override
public boolean equals(Object o) {
return builderEquals(this, o);
}
@Override
public int hashCode() {
return toString().hashCode();
}
ChronicleMapBuilder<K, V> removedEntryCleanupTimeout(
long removedEntryCleanupTimeout, TimeUnit unit) {
if (unit.toMillis(removedEntryCleanupTimeout) < 1) {
throw new IllegalArgumentException("timeout should be >= 1 millisecond, " +
removedEntryCleanupTimeout + " " + unit + " is given");
}
cleanupTimeout = removedEntryCleanupTimeout;
cleanupTimeoutUnit = unit;
return this;
}
ChronicleMapBuilder<K, V> cleanupRemovedEntries(boolean cleanupRemovedEntries) {
this.cleanupRemovedEntries = cleanupRemovedEntries;
return this;
}
@Override
public ChronicleMapBuilder<K, V> keyReaderAndDataAccess(
SizedReader<K> keyReader, @NotNull DataAccess<K> keyDataAccess) {
keyBuilder.reader(keyReader);
keyBuilder.dataAccess(keyDataAccess);
return this;
}
@Override
public ChronicleMapBuilder<K, V> keyMarshallers(
@NotNull SizedReader<K> keyReader, @NotNull SizedWriter<? super K> keyWriter) {
keyBuilder.reader(keyReader);
keyBuilder.writer(keyWriter);
return this;
}
@Override
public <M extends SizedReader<K> & SizedWriter<? super K>>
ChronicleMapBuilder<K, V> keyMarshaller(@NotNull M sizedMarshaller) {
return keyMarshallers(sizedMarshaller, sizedMarshaller);
}
@Override
public ChronicleMapBuilder<K, V> keyMarshallers(
@NotNull BytesReader<K> keyReader, @NotNull BytesWriter<? super K> keyWriter) {
keyBuilder.reader(keyReader);
keyBuilder.writer(keyWriter);
return this;
}
@Override
public <M extends BytesReader<K> & BytesWriter<? super K>>
ChronicleMapBuilder<K, V> keyMarshaller(@NotNull M marshaller) {
return keyMarshallers(marshaller, marshaller);
}
@Override
public ChronicleMapBuilder<K, V> keySizeMarshaller(@NotNull SizeMarshaller keySizeMarshaller) {
keyBuilder.sizeMarshaller(keySizeMarshaller);
return this;
}
@Override
public ChronicleMapBuilder<K, V> aligned64BitMemoryOperationsAtomic(
boolean aligned64BitMemoryOperationsAtomic) {
this.aligned64BitMemoryOperationsAtomic = aligned64BitMemoryOperationsAtomic;
return this;
}
@Override
public ChronicleMapBuilder<K, V> checksumEntries(boolean checksumEntries) {
this.checksumEntries = checksumEntries ? ChecksumEntries.YES : ChecksumEntries.NO;
return this;
}
boolean checksumEntries() {
switch (checksumEntries) {
case NO: return false;
case YES: return true;
case IF_PERSISTED: return persisted;
default: throw new AssertionError();
}
}
boolean aligned64BitMemoryOperationsAtomic() {
return aligned64BitMemoryOperationsAtomic;
}
/**
* Configures the {@code DataAccess} and {@code SizedReader} used to serialize and deserialize
* values to and from off-heap memory in maps, created by this builder.
*
* @param valueReader the new bytes → value object reader strategy
* @param valueDataAccess the new strategy of accessing the values' bytes for writing
* @return this builder back
* @see #valueMarshallers(SizedReader, SizedWriter)
* @see ChronicleHashBuilder#keyReaderAndDataAccess(SizedReader, DataAccess)
*/
public ChronicleMapBuilder<K, V> valueReaderAndDataAccess(
SizedReader<V> valueReader, @NotNull DataAccess<V> valueDataAccess) {
valueBuilder.reader(valueReader);
valueBuilder.dataAccess(valueDataAccess);
return this;
}
/**
* Configures the marshallers, used to serialize/deserialize values to/from off-heap memory in
* maps, created by this builder.
*
* @param valueReader the new bytes → value object reader strategy
* @param valueWriter the new value object → bytes writer strategy
* @return this builder back
* @see #valueReaderAndDataAccess(SizedReader, DataAccess)
* @see #valueSizeMarshaller(SizeMarshaller)
* @see ChronicleHashBuilder#keyMarshallers(SizedReader, SizedWriter)
*/
public ChronicleMapBuilder<K, V> valueMarshallers(
@NotNull SizedReader<V> valueReader, @NotNull SizedWriter<? super V> valueWriter) {
valueBuilder.reader(valueReader);
valueBuilder.writer(valueWriter);
return this;
}
/**
* Shortcut for {@link #valueMarshallers(SizedReader, SizedWriter)
* valueMarshallers(sizedMarshaller, sizedMarshaller)}.
*/
public <M extends SizedReader<V> & SizedWriter<? super V>>
ChronicleMapBuilder<K, V> valueMarshaller(@NotNull M sizedMarshaller) {
return valueMarshallers(sizedMarshaller, sizedMarshaller);
}
/**
* Configures the marshallers, used to serialize/deserialize values to/from off-heap memory in
* maps, created by this builder.
*
* @param valueReader the new bytes → value object reader strategy
* @param valueWriter the new value object → bytes writer strategy
* @return this builder back
* @see #valueReaderAndDataAccess(SizedReader, DataAccess)
* @see #valueSizeMarshaller(SizeMarshaller)
* @see ChronicleHashBuilder#keyMarshallers(BytesReader, BytesWriter)
*/
public ChronicleMapBuilder<K, V> valueMarshallers(
@NotNull BytesReader<V> valueReader, @NotNull BytesWriter<? super V> valueWriter) {
valueBuilder.reader(valueReader);
valueBuilder.writer(valueWriter);
return this;
}
/**
* Shortcut for {@link #valueMarshallers(BytesReader, BytesWriter)
* valueMarshallers(marshaller, marshaller)}.
*/
public <M extends BytesReader<V> & BytesWriter<? super V>>
ChronicleMapBuilder<K, V> valueMarshaller(@NotNull M marshaller) {
return valueMarshallers(marshaller, marshaller);
}
/**
* Configures the marshaller used to serialize actual value sizes to off-heap memory in maps,
* created by this builder.
*
* <p>Default value size marshaller is so-called "stop bit encoding" marshalling, unless {@link
* #constantValueSizeBySample(Object)} or the builder statically knows the value size is
* constant -- special constant size marshalling is used by default in these cases.
*
* @param valueSizeMarshaller the new marshaller, used to serialize actual value sizes to
* off-heap memory
* @return this builder back
* @see #keySizeMarshaller(SizeMarshaller)
*/
public ChronicleMapBuilder<K, V> valueSizeMarshaller(
@NotNull SizeMarshaller valueSizeMarshaller) {
valueBuilder.sizeMarshaller(valueSizeMarshaller);
return this;
}
/**
* Specifies the function to obtain a value for the key during {@link ChronicleMap#acquireUsing
* acquireUsing()} calls, if the key is absent in the map, created by this builder.
*
* <p>This is a <a href="#jvm-configurations">JVM-level configuration</a>.
*
* @param defaultValueProvider the strategy to obtain a default value by the absent key
* @return this builder back
*/
public ChronicleMapBuilder<K, V> defaultValueProvider(
@NotNull DefaultValueProvider<K, V> defaultValueProvider) {
Objects.requireNonNull(defaultValueProvider);
this.defaultValueProvider = defaultValueProvider;
return this;
}
ChronicleMapBuilder<K, V> replication(byte identifier) {
if (identifier <= 0)
throw new IllegalArgumentException("Identifier must be positive, " + identifier +
" given");
this.replicationIdentifier = identifier;
return this;
}
@Override
public ChronicleMap<K, V> createPersistedTo(File file) throws IOException {
// clone() to make this builder instance thread-safe, because createWithFile() method
// computes some state based on configurations, but doesn't synchronize on configuration
// changes.
return clone().createWithFile(file, false, false, null);
}
@Override
public ChronicleMap<K, V> createOrRecoverPersistedTo(File file) throws IOException {
return createOrRecoverPersistedTo(file, true);
}
@Override
public ChronicleMap<K, V> createOrRecoverPersistedTo(File file, boolean sameLibraryVersion)
throws IOException {
return createOrRecoverPersistedTo(file, sameLibraryVersion,
defaultChronicleMapCorruptionListener);
}
@Override
public ChronicleMap<K, V> createOrRecoverPersistedTo(
File file, boolean sameLibraryVersion,
ChronicleHashCorruption.Listener corruptionListener) throws IOException {
if (file.exists()) {
return recoverPersistedTo(file, sameLibraryVersion, corruptionListener);
} else {
return createPersistedTo(file);
}
}
@Override
public ChronicleMap<K, V> recoverPersistedTo(
File file, boolean sameBuilderConfigAndLibraryVersion) throws IOException {
return recoverPersistedTo(file, sameBuilderConfigAndLibraryVersion,
defaultChronicleMapCorruptionListener);
}
@Override
public ChronicleMap<K, V> recoverPersistedTo(
File file, boolean sameBuilderConfigAndLibraryVersion,
ChronicleHashCorruption.Listener corruptionListener) throws IOException {
return clone().createWithFile(file, true, sameBuilderConfigAndLibraryVersion,
corruptionListener);
}
@Override
public ChronicleMap<K, V> create() {
// clone() to make this builder instance thread-safe, because createWithoutFile() method
// computes some state based on configurations, but doesn't synchronize on configuration
// changes.
return clone().createWithoutFile();
}
private ChronicleMap<K, V> createWithFile(
File file, boolean recover, boolean overrideBuilderConfig,
ChronicleHashCorruption.Listener corruptionListener) throws IOException {
if (overrideBuilderConfig && !recover)
throw new AssertionError("recover -> overrideBuilderConfig");
replicated = replicationIdentifier != -1;
persisted = true;
// It's important to canonicalize the file, because CanonicalRandomAccessFiles.acquire()
// relies on java.io.File equality, which doesn't account symlinks itself.
file = file.getCanonicalFile();
if (!file.exists()) {
if (recover)
throw new FileNotFoundException("file " + file + " should exist for recovery");
//noinspection ResultOfMethodCallIgnored
file.createNewFile();
}
RandomAccessFile raf = CanonicalRandomAccessFiles.acquire(file);
ChronicleHashResources resources = new PersistedChronicleHashResources(file);
try {
VanillaChronicleMap<K, V, ?> result;
if (raf.length() > 0) {
result = openWithExistingFile(file, raf, resources, recover, overrideBuilderConfig,
corruptionListener);
} else {
// Single-element arrays allow to modify variables within lambda
@SuppressWarnings("unchecked")
VanillaChronicleMap<K, V, ?>[] map = new VanillaChronicleMap[1];
ByteBuffer[] headerBuffer = new ByteBuffer[1];
boolean[] newFile = new boolean[1];
FileChannel fileChannel = raf.getChannel();
fileLockedIO(file, fileChannel, () -> {
if (raf.length() == 0) {
map[0] = newMap();
headerBuffer[0] = writeHeader(fileChannel, map[0]);
newFile[0] = true;
} else {
newFile[0] = false;
}
});
if (newFile[0]) {
int headerSize = headerBuffer[0].remaining();
result = createWithNewFile(map[0], file, raf, resources, headerBuffer[0],
headerSize);
} else {
result = openWithExistingFile(file, raf, resources, recover,
overrideBuilderConfig, corruptionListener);
}
}
prepareMapPublication(result);
return result;
} catch (Throwable throwable) {
try {
try {
resources.setChronicleHashIdentityString(
"ChronicleHash{name=" + name + ", file=" + file + "}");
} catch (Throwable t) {
throwable.addSuppressed(t);
} finally {
resources.releaseManually();
}
} catch (Throwable t) {
throwable.addSuppressed(t);
}
throw Throwables.propagateNotWrapping(throwable, IOException.class);
}
}
private void prepareMapPublication(VanillaChronicleMap map) throws IOException {
establishReplication(map);
map.setResourcesName();
map.registerCleaner();
// Ensure safe publication of the ChronicleMap
OS.memory().storeFence();
map.addToOnExitHook();
}
/**
* @return size of the self bootstrapping header
*/
private int waitUntilReady(RandomAccessFile raf, File file, boolean recover)
throws IOException {
FileChannel fileChannel = raf.getChannel();
ByteBuffer sizeWordBuffer = ByteBuffer.allocate(4);
sizeWordBuffer.order(LITTLE_ENDIAN);
// 60 * 10, 100 ms wait = 1 minute total wait
int attempts = 60 * 10;
int lastReadHeaderSize = -1;
for (int attempt = 0; attempt < attempts; attempt++) {
if (raf.length() >= SELF_BOOTSTRAPPING_HEADER_OFFSET) {
sizeWordBuffer.clear();
readFully(fileChannel, SIZE_WORD_OFFSET, sizeWordBuffer);
if (sizeWordBuffer.remaining() == 0) {
int sizeWord = sizeWordBuffer.getInt(0);
lastReadHeaderSize = SizePrefixedBlob.extractSize(sizeWord);
if (SizePrefixedBlob.isReady(sizeWord))
return lastReadHeaderSize;
}
// The only possible reason why not 4 bytes are read, is that the file is
// truncated between length() and read() calls, then continue to wait
}
try {
Thread.sleep(100);
} catch (InterruptedException e) {
Thread.currentThread().interrupt();
if (recover) {
if (lastReadHeaderSize == -1) {
throw new ChronicleHashRecoveryFailedException(e);
} else {
return lastReadHeaderSize;
}
} else {
throw new IOException(e);
}
}
}
if (recover) {
if (lastReadHeaderSize == -1) {
throw new ChronicleHashRecoveryFailedException(
"File header is not recoverable, file=" + file);
} else {
return lastReadHeaderSize;
}
} else {
throw new IOException("Unable to wait until the file=" + file +
" is ready, likely the process which created the file crashed or hung " +
"for more than 1 minute");
}
}
/**
* @return ByteBuffer, in [position, limit) range the self bootstrapping header is read
*/
private static ByteBuffer readSelfBootstrappingHeader(
File file, RandomAccessFile raf, int headerSize, boolean recover,
ChronicleHashCorruption.Listener corruptionListener,
ChronicleHashCorruptionImpl corruption) throws IOException {
if (raf.length() < headerSize + SELF_BOOTSTRAPPING_HEADER_OFFSET) {
throw throwRecoveryOrReturnIOException(file,
"The file is shorter than the header size: " + headerSize +
", file size: " + raf.length(), recover);
}
FileChannel fileChannel = raf.getChannel();
ByteBuffer headerBuffer = ByteBuffer.allocate(
SELF_BOOTSTRAPPING_HEADER_OFFSET + headerSize);
headerBuffer.order(LITTLE_ENDIAN);
readFully(fileChannel, 0, headerBuffer);
if (headerBuffer.remaining() > 0) {
throw throwRecoveryOrReturnIOException(file, "Unable to read the header fully, " +
headerBuffer.remaining() + " is remaining to read, likely the file was " +
"truncated", recover);
}
int sizeWord = headerBuffer.getInt(SIZE_WORD_OFFSET);
if (!SizePrefixedBlob.isReady(sizeWord)) {
if (recover) {
report(corruptionListener, corruption, -1, () ->
format("file={}: size-prefixed blob readiness bit is set to NOT_COMPLETE",
file)
);
// the bit will be overwritten to READY in the end of recovery procedure, so nothing
// to fix right here
} else {
throw new IOException("file=" + file+ ": sizeWord is not ready: " + sizeWord);
}
}
headerBuffer.position(SELF_BOOTSTRAPPING_HEADER_OFFSET);
return headerBuffer;
}
private static boolean checkSumSelfBootstrappingHeader(
ByteBuffer headerBuffer, int headerSize) {
long checkSum = headerChecksum(headerBuffer, headerSize);
long storedChecksum = headerBuffer.getLong(HEADER_OFFSET);
return storedChecksum == checkSum;
}
private VanillaChronicleMap<K, V, ?> createWithNewFile(
VanillaChronicleMap<K, V, ?> map, File file, RandomAccessFile raf,
ChronicleHashResources resources, ByteBuffer headerBuffer, int headerSize)
throws IOException {
map.initBeforeMapping(file, raf, headerBuffer.limit(), false);
map.createMappedStoreAndSegments(resources);
commitChronicleMapReady(map, raf, headerBuffer, headerSize);
return map;
}
private VanillaChronicleMap<K, V, ?> openWithExistingFile(
File file, RandomAccessFile raf, ChronicleHashResources resources,
boolean recover, boolean overrideBuilderConfig,
ChronicleHashCorruption.Listener corruptionListener)
throws IOException {
ChronicleHashCorruptionImpl corruption = recover ? new ChronicleHashCorruptionImpl() : null;
try {
int headerSize = waitUntilReady(raf, file, recover);
FileChannel fileChannel = raf.getChannel();
ByteBuffer headerBuffer = readSelfBootstrappingHeader(
file, raf, headerSize, recover, corruptionListener, corruption);
if (headerSize != headerBuffer.remaining())
throw new AssertionError();
boolean headerCorrect = checkSumSelfBootstrappingHeader(headerBuffer, headerSize);
boolean headerWritten = false;
if (!headerCorrect) {
if (overrideBuilderConfig) {
VanillaChronicleMap<K, V, ?> mapObjectForHeaderOverwrite = newMap();
headerBuffer = writeHeader(fileChannel, mapObjectForHeaderOverwrite);
headerSize = headerBuffer.remaining();
headerWritten = true;
} else {
throw throwRecoveryOrReturnIOException(file,
"Self Bootstrapping Header checksum doesn't match the stored checksum",
recover);
}
}
Bytes<ByteBuffer> headerBytes = Bytes.wrapForRead(headerBuffer);
headerBytes.readPosition(headerBuffer.position());
headerBytes.readLimit(headerBuffer.limit());
Wire wire = new TextWire(headerBytes);
VanillaChronicleMap<K, V, ?> map = wire.getValueIn().typedMarshallable();
map.initBeforeMapping(file, raf, headerBuffer.limit(), recover);
long dataStoreSize = map.globalMutableState().getDataStoreSize();
if (!recover && dataStoreSize > file.length()) {
throw new IOException("The file " + file + " the map is serialized from " +
"has unexpected length " + file.length() + ", probably corrupted. " +
"Data store size is " + dataStoreSize);
}
map.initTransientsFromBuilder(this);
if (!recover) {
map.createMappedStoreAndSegments(resources);
} else {
if (!headerWritten)
writeNotComplete(fileChannel, headerBuffer, headerSize);
map.recover(resources, corruptionListener, corruption);
commitChronicleMapReady(map, raf, headerBuffer, headerSize);
}
return map;
} catch (Throwable t) {
if (recover && !(t instanceof IOException) &&
!(t instanceof ChronicleHashRecoveryFailedException)) {
throw new ChronicleHashRecoveryFailedException(t);
}
throw Throwables.propagateNotWrapping(t, IOException.class);
}
}
private ChronicleMap<K, V> createWithoutFile() {
replicated = replicationIdentifier != -1;
persisted = false;
ChronicleHashResources resources = new InMemoryChronicleHashResources();
try {
VanillaChronicleMap<K, V, ?> map = newMap();
map.createInMemoryStoreAndSegments(resources);
prepareMapPublication(map);
return map;
} catch (Throwable throwable) {
try {
try {
resources.setChronicleHashIdentityString(
"ChronicleHash{name=" + name + ", file=null}");
} catch (Throwable t) {
throwable.addSuppressed(t);
} finally {
resources.releaseManually();
}
} catch (Throwable t) {
throwable.addSuppressed(t);
}
throw Throwables.propagate(throwable);
}
}
private VanillaChronicleMap<K, V, ?> newMap()
throws IOException {
preMapConstruction();
if (replicated) {
return new ReplicatedChronicleMap<>(this);
} else {
return new VanillaChronicleMap<>(this);
}
}
private void preMapConstruction() {
averageKeySize = preMapConstruction(
keyBuilder, averageKeySize, averageKey, sampleKey, "Key");
averageValueSize = preMapConstruction(
valueBuilder, averageValueSize, averageValue, sampleValue, "Value");
stateChecks();
}
private <E> double preMapConstruction(
SerializationBuilder<E> builder, double configuredAverageSize, E average, E sample,
String dim) {
if (sample != null) {
return builder.constantSizeBySample(sample);
} else {
double result = averageKeyOrValueSize(configuredAverageSize, builder, average);
if (!isNaN(result) || allLowLevelConfigurationsAreManual()) {
return result;
} else {
throw new IllegalStateException(dim + " size in serialized form must " +
"be configured in ChronicleMap, at least approximately.\nUse builder" +
".average" + dim + "()/.constant" + dim + "SizeBySample()/" +
".average" + dim + "Size() methods to configure the size");
}
}
}
private void stateChecks() {
checkActualChunksPerSegmentTierIsConfiguredOnlyIfOtherLowLevelConfigsAreManual();
checkActualChunksPerSegmentGreaterOrEqualToEntries();
}
private boolean allLowLevelConfigurationsAreManual() {
return actualSegments > 0 && entriesPerSegment > 0 && actualChunksPerSegmentTier > 0
&& actualChunkSize > 0;
}
private void establishReplication(
VanillaChronicleMap<K, V, ?> map) throws IOException {
if (map instanceof ReplicatedChronicleMap) {
ReplicatedChronicleMap result = (ReplicatedChronicleMap) map;
if (cleanupRemovedEntries)
establishCleanupThread(result);
}
}
private void establishCleanupThread(ReplicatedChronicleMap map) {
OldDeletedEntriesCleanupThread cleanupThread = new OldDeletedEntriesCleanupThread(map);
map.addCloseable(cleanupThread);
cleanupThread.start();
}
/**
* Inject your SPI code around basic {@code ChronicleMap}'s operations with entries:
* removing entries, replacing entries' value and inserting new entries.
*
* <p>This affects behaviour of ordinary map.put(), map.remove(), etc. calls, as well as removes
* and replacing values <i>during iterations</i>, <i>remote map calls</i> and
* <i>internal replication operations</i>.
*
* <p>This is a <a href="#jvm-configurations">JVM-level configuration</a>.
*
* @return this builder back
*/
public ChronicleMapBuilder<K, V> entryOperations(MapEntryOperations<K, V, ?> entryOperations) {
Objects.requireNonNull(entryOperations);
this.entryOperations = entryOperations;
return this;
}
/**
* Inject your SPI around logic of all {@code ChronicleMap}'s operations with individual keys:
* from {@link ChronicleMap#containsKey} to {@link ChronicleMap#acquireUsing} and
* {@link ChronicleMap#merge}.
*
* <p>This affects behaviour of ordinary map calls, as well as <i>remote calls</i>.
*
* <p>This is a <a href="#jvm-configurations">JVM-level configuration</a>.
*
* @return this builder back
*/
public ChronicleMapBuilder<K, V> mapMethods(MapMethods<K, V, ?> mapMethods) {
Objects.requireNonNull(mapMethods);
this.methods = mapMethods;
return this;
}
ChronicleMapBuilder<K, V> remoteOperations(
MapRemoteOperations<K, V, ?> remoteOperations) {
Objects.requireNonNull(remoteOperations);
this.remoteOperations = remoteOperations;
return this;
}
enum ChecksumEntries {YES, NO, IF_PERSISTED}
interface FileIOAction {
void fileIOAction() throws IOException;
}
static class EntrySizeInfo {
final double averageEntrySize;
final int worstAlignment;
EntrySizeInfo(double averageEntrySize, int worstAlignment) {
this.averageEntrySize = averageEntrySize;
this.worstAlignment = worstAlignment;
}
}
}