- Make single-table design easy and error-prone
- Schema definition with serialization and deserialization
- Utility tools for expressions/conditions building, retries, paging, and more...
It's as simple as:
@Value
@DynamoTable(type = Primary, partitionKeys = "accountId", rangePrefix = "account")
@DynamoTable(type = Gsi, indexNumber = 1, partitionKeys = {"apiKey"}, rangePrefix = "accountByApiKey")
@DynamoTable(type = Gsi, indexNumber = 2, partitionKeys = {"oauthGuid"}, rangePrefix = "accountByOauthGuid")
class Account {
@NonNull
String accountId;
@NonNull
String email;
@ToString.Exclude
String apiKey;
}
...
// Initialize schema
SingleTable singleTable = SingleTable.builder()
.dynamoDoc(dynamoDoc)
.tablePrefix("project").build();
TableSchema<Account> schema = singleTable.parseTableSchema(Account.class);
// Insert new account
Account account = new Account("8426", "matus@example.com", null);
schema.table().putItem(new PutItemSpec().withItem(schema.toItem(account)));
// Fetch other account
Optional<Account> otherAccountOpt = Optional.ofNullable(schema.fromItem(schema.table().getItem(
schema.primaryKey(Map.of("accountId","9473")))));
Okay, it could be simpler...
We support both AWS SDK v1 and v2. The table below shows the compatibility matrix:
Dynamo AWS SDK | SingleTable | Docs |
---|---|---|
Java 2.x | 2.x.x | This README |
Java 1.x | 0.x.x | 0.x.x Branch |
<dependency>
<groupId>io.dataspray</groupId>
<artifactId>single-table</artifactId>
<version>${single-table.version}</version>
</dependency>
Latest release in Maven Central is here.
- Getting started
- Create our table
- Insert an item
- Update and Condition expressions builder
- Select an item
- Query ranges with paging
- Scan records of specific type
- Upsert (Update or create if missing))
- Filter records
In our examples, we skip the steps of initializing SingleTable
and parsing our schema
. Here is how you can do this:
SingleTable singleTable = SingleTable.builder()
.tablePrefix("project")
.build();
TableSchema<Account> accountSchema = singleTable.parseTableSchema(Account.class);
IndexSchema<Account> accountByApiKeySchema = singleTable.parseGlobalSecondaryIndexSchema(1, Account.class);
Our library assumes the table is created with partition and range keys with particular names. (pk
, sk
, pkgsi1
,
...) Use our tool to create a valid table.
Note you need to indicate how many LSIs and GSIs you would like to create. This depends on how many you are using in your schemas. But don't worry you can always add more later.
singleTable.createTableIfNotExists(2, 2);
Alternatively, you can create the DynamoDB table and all indexes via AWS CDK stack:
singleTable.createCdkTable(this, "my-stack-name", 2, 2);
client.putItem(PutItemRequest.builder()
.tableName(schema.tableName())
.item(schema.toAttrMap(myAccount))
.build())
ExpressionBuilder expressionBuilder = schema.expressionBuilder();
// Apply conditions
expressionBuilder
// Item exists
.conditionExists()
// Particular field exists
.conditionFieldExists("cancelDate")
// Particular field equals a value
.conditionFieldEquals("isCancelled", false);
// Modify data
expressionBuilder
// Overwrite field
.set("apiKey", apiKey)
// Increment field value
.setIncrement("votersCount", 1);
// Add to a set
.add("transactionIds", ImmutableSet.of("4234", "5312"))
// Remove entry from a json field
.remove(ImmutableList.of("entryJson", entryId, "isMoved"));
Expression expression = expressionBuilder.build();
// For PUTs
client.updateItem(expression.toUpdateItemRequestBuilder()
.key(schema.primaryKey(expectedData))
.build());
// For other requests
expression.updateExpression().ifPresent(builder::updateExpression);
expression.conditionExpression().ifPresent(builder::conditionExpression);
expression.expressionAttributeNames().ifPresent(builder::expressionAttributeNames);
expression.expressionAttributeValues().ifPresent(builder::expressionAttributeValues);
Account account = schema.fromAttrMap(client.getItem(b -> b
.tableName(schema.tableName())
.key(schema.primaryKey(Map.of(
"accountId","account-id-123"))))
.item());
You may want to wrap it in an optional if you prefer not to work with nulls:
Optional<Account> accountOpt = Optional.ofNullable(schema.fromAttrMap(client.getItem(b -> b
.tableName(schema.tableName())
.key(schema.primaryKey(Map.of(
"accountId","account-id-123"))))
.item()));
In this example, we will be querying all range keys for a given partition key.
On every request, we check if there are more results with getLastEvaluatedKey
and then providing this cursor back
using withExclusiveStartKey
to continue quering where we left off.
Optional<String> cursor = Optional.empty();
do {
// Prepare request
QueryRequest.Builder builder = QueryRequest.builder()
.tableName(schema.tableName())
// Query by partition key
.keyConditions(schema.attrMapToConditions(schema.partitionKey(Map.of(
"partitionKey", partitionKey))))
.limit(2);
cursor.ifPresent(exclusiveStartKey -> builder.exclusiveStartKey(schema.toExclusiveStartKey(exclusiveStartKey)));
// Perform request
QueryResponse response = client.query(builder.build());
// Retrieve next cursor
cursor = schema.serializeLastEvaluatedKey(response.lastEvaluatedKey());
// Process results
response.items().stream()
.map(schema::fromAttrMap)
.forEachOrdered(processor::process);
} while (cursor.isPresent());
You may have Cats and Dogs inside your single-table design and you want to retrieve all the Cats without having to also iterate over all the dogs.
One way to do this is using a DynamoDB technique
called sharding. To
apply this technique to our Cats, instead of having a catId
as the parition key, we will instead have cat-XXX
where
XXX will be a deterministic shard partition number based on the catId
. The catId
will be stored as a range key
instead.
Our schema can look like this:
@DynamoTable(type = Primary, shardKeys = {"catId"}, shardPrefix = "cat", shardCount = 100, rangePrefix = "cat", rangeKeys = "catId")
public class Cat {
@NonNull String catId;
}
And our usage would be:
String catId = "A18D5B00";
Cat myCat = new Cat(catId);
// Insertion is same as before, sharding is done under the hood
schema.table().putItem(PutItemRequest.builder()
.tableName(primary.tableName())
.item(schema.toAttrMap(myCat))
.build());
// Retrieving cat is also same as before
Cat otherCat = schema.fromAttrMap(client.getItem(GetItemRequest.builder()
.tableName(schema.tableName())
.key(schema.primaryKey(Map.of(
"catId", catId)))
.build())
.item());
// Finally let's query some cats without an entire table scan
ShardPageResult<Cat> result = singleTable.fetchShardNextPage(
client,
schema,
/* Pagination token */ Optional.empty(),
/* page size */ 100);
processCats(result.getItems());
// Finally let's dump all our cats using pagination
Optional<String> cursorOpt = Optional.empty();
do {
ShardPageResult<Cat> result = singleTable.fetchShardNextPage(
client,
schema,
cursorOpt,
/* page size */ 100);
cursorOpt = result.getCursorOpt();
processCats(result.getItems());
} while (cursorOpt.isPresent());
Upserts are tricky in DynamoDB as there is no native support. Luckily we can do this ourselves by effectively overwriting the entire record whether it exists or not and for particular fields, add logic how to compute the new value based on previous value.
int catCountDiff = 4; // We want to increment by this amount
HashMap<String, String> userCounterNameMap = Maps.newHashMap();
HashMap<String, Object> userCounterValueMap = Maps.newHashMap();
userCounterNameMap.put("#catCount", "catCount");
userCounterValueMap.put(":diff", catCountDiff);
userCounterValueMap.put(":zero", 0L);
String upsertExpression = schema.upsertExpression(
new CatCounter(bagId, catCountDiff),
userCounterNameMap,
userCounterValueMap,
// Indicate we are computing catCount ourselves
ImmutableSet.of("catCount"),
// Compute catCount by adding existing value (or zero) to our catCountDiff
", #catCount = if_not_exists(#catCount, :zero) + :diff");
In this case, we have overwritten the CatCounter
entirely except the catCount
field. The field we are manually
calculating by adding the previous value if exists to our catCountDiff
One way to retrieve a subset of records is to replicate them into a secondary index conditionally based on the existence of a field.
Currently there isn't a way to do this with this library, but would be fairly trivial to add, contributions are welcome.