dynamo-builder
A type-safe DynamoDB query builder for TypeScript. This library is inspired by Beyonce's library and extends the functionality further giving client the ability to configure table partition keys schema design
Features included in this library:
-
Low boilerplate. Define your tables, partitions, indexes and models in YAML and the codegens TypeScript definitions for you.
-
Store heterogeneous models in the same table. Unlike most DynamoDB libraries, this doesn't force you into a 1 model per table paradigm. It supports storing related models in the same table partition, which allows you to "precompute joins" and retrieve those models with a single roundtrip query to the db.
-
Type-safe API. dynamo-builder's API is type-safe. It's aware of which models live under your partition and sort keys (even for global secondary indexes). When you
get
,batchGet
orquery
, the result types are automatically inferred. And when you apply filters on yourquery
the attribute names are automatically type-checked.
Local Development
To test out dev changes here in the consuming package, run:
npm run build npm pack
Then copy the tgz produced into the root of the consuming project and refer to it from the package/json of consuming package as:
"dynamo-builder": "file:dynamo-builder-1.0.0.tgz",
remember to run install in consuming package too.
Usage
1. Install
First install dynamo-builder - npm install dynamo-builder
2. Define your models
Define your tables
, models
and partitions
in YAML:
tables:
# We have a single DynamoDB Table named "Library".
Library:
partitionKeyName: pk #pk is the default value of partitionKeyName of the table
sortKeyName: sk #sk is the default value of sortKeyName of the table
# Let's add two models to our Library table: Author and Book.
models:
Author:
id: string
name: string
Book:
id: string
authorId: string
name: string
# Now, imagine we want to be able to retrieve an Author + all their Books
# in a single DynamoDB Query operation.
# To do that, we need a specific Author and all their Books to live under the same partition key.
# How about we use "Author-$id" as the partition key? Great, let's go with that.
# dynamo-builder calls a group of models that share the same partition key a "patition".
# Let's define one now, and name it "Authors"
partitions:
Authors:
# All dynamo-builder partition keys are prefixed (to help you avoid collisions)
# We said above we want our final partition key to be "Author-$id",
# so we set: "Author" as our prefix here
partitionKeyPrefix: Author
# And, now we can put a given Author and all their Books into the same partition
models:
Author:
partitionKey: [$id] # "Author-$id"
sortKey: [Author, $id]
Book:
partitionKey: [$authorId] # "Author-$authorId"
sortKey: [Book, $id]
partitionKey
and sortKey
syntax
A note on dynamo-builder expects you to specify your partition and sort keys using arrays, e.g. [Author, $id]
. The first element in this example is interpreted as a string literal, while the second substitutes the value of a specific model instance's id
field. In addition, it prefixes partition keys with the partitionKeyPrefix
set on the "partition" configured your the YAML file.
In our example above, we set the Author
partiion's partitionKeyPrefix
to "Author"
and the Author
model's partitionKey
field to [$id]
. Thus the full partition key at runtime is Author-$id
(it uses -
as a delimiter by default, you can override the implementation by passing in delimiter key in table definition).
Supported values for delimiter: "-", "#"
tables:
Library:
delimiter: "#"
models:
...
partitions:
...
If you'd like to form a composite partition or sort key using multiple model fields, that is supported as well, e.g. [$id, $name]
.
Global secondary indexes
If your table(s) have GSI's you can specify them like this:
tables:
Library:
models:
...
partitions:
...
gsis:
byName: # must match your GSI's name
partitionKey: $name # name field must exist on at least one model
sortKey: $id # same here
Note: library currently assumes that your GSI indexes project all model attributes, which will be reflected in the return types of your queries.
External types
You can specify external types you need to import like so:
Author:
...
address: Address from author/Address
Which transforms into import { Address } from "author/address"
3. Codegen TypeScript classes for your models, partition keys and sort keys
npx dynamo-builder --in src/models.yaml --out src/generated/models.ts
4. Create your DynamoDB table(s)
import { LibraryTable } from "generated/models"
const dynamo = new DynamoDB({ endpoint: "...", region: "..."})
await dynamo
.createTable(LibraryTable.asCreateTableInput("PAY_PER_REQUEST"))
.promise()
5. Write type-safe queries
Now you can write partition-aware, type safe queries with abandon:
Get yourself a Beyonce
import { Beyonce } from "dynamo-builder"
import { DynamoDB } from "aws-sdk"
import { LibraryTable } from "generated/models"
const beyonce = new Beyonce(LibraryTable, dynamo)
Then import the generated models
import {
AuthorModel,
BookModel,
} from "generated/models"
Queries
Put
const author = AuthorModel.create({
id: "1",
name: "Jane Austen"
})
await beyonce.put(author)
Get
const author = await beyonce.get(AuthorModel.key({ id: "1" }))
Note: the key prefix
("Author" from our earlier example) will be automatically appeneded.
Update
Beyoncé supports type-safe partial updates on items, without having to read the item from the db first. And it works, even through nested attributes:
const updatedAuthor = await beyonce.update(AuthorModel.key({ id: "1" }), (author) => {
author.name = "Jack London",
author.details.description = "American novelist"
delete author.details.someDeprecatedField
})
Here author
is an intelligent proxy object (thus we avoid having to read the full item from the DB prior to updating it).
And beyonce.update(...)
returns the full Author
, with the updated fields.
Query
Beyoncé supports type-safe query
operations that either return a single model type or all model types that live under a given partition key.
Querying for a specific model type
You can query
for a single type of model like so:
import { BookModel } from "generated/models"
// Get all Books for an Author
const results = await beyonce
.query(BookModel.partitionKey({ authorId: "1" }))
.exec() // returns { Book: Book[] }
To reduce the amount of data retrieved by DynamoDB, Beyoncé automatically applies a KeyConditionExpression
that uses the sortKey
prefix provided in your model definitions. For example, if the YAML definition for the Book
model contains sortKey:[Book, $id]
-- then the generated KeyConditionExpression
will contain a clause like #partitionKey = :partitionKey AND begins_with(#sortKey, Book)
.
Query for all models in a partition
You can also query for all models that live in a partition, like so:
import { AuthorPartition } from "generated/models"
// Get an Author + their books
const results = await beyonce
.query(AuthorPartition.key({ id: "1" }))
.exec() // returns { Author: Author[], Book: Book[] }
Note that, in this case the generated KeyconditionExpression
will not include a clause for the sort key since DynamoDB does not support OR-ing key conditions.
Filtering Queries
You can filter results from a query like so:
// Get an Author + filter on their books
const authorWithFilteredBooks = await beyonce
.query(AuthorPartition.key({ id: "1" }))
.attributeNotExists("title") // type-safe fields
.or("title", "=", "Brave New World") // type safe fields + operators
.exec()
Paginating Queries
When you call .exec()
Beyoncé will automatically page through all the results and return them to you.
If you would like to step through pages manually (e.g to throttle reads) -- use the .iterator()
method instead:
const iterator = beyonce
.query(AuthorPartition.key({ id: "1" }))
.iterator({ pageSize: 1 })
// Step through each page 1 by 1
for await (const { items, errors } of iterator) {
// ...
}
The errors
field above contains any exceptions thrown while attempting to load the next iterator "page".
So it's up to you, the caller to decide if you want to continue walking the iterator, or give up and exit.
Important: When an error is encountered within the iterator, you might get a partial result
that contains one or more items
and one or more errors
. Thus, you should always check errors.length
.
Cursors
Each time you call .next()
on the iterator, you'll also get a cursor
back, which you can use to create a new iterator
that picks up where you left off
const iterator1 = beyonce
.query(AuthorPartition.key({ id: "1" }))
.iterator({ pageSize: 1 })
const firstPage = await iterator1.next()
const { items, cursor } = firstPage.value // do something with these
// Later...
const iterator2 = beyonce
.query(AuthorPartition.key({ id: "1" }))
.iterator({ cursor, pageSize: 1 })
const secondPage = await iterator2.next()
QueryGSI
import { byNameGSI } from "generated/models"
const prideAndPrejudice = await beyonce
.queryGSI(byNameGSI.name, byNameGSI.key("Jane Austen"))
.where("title", "=", "Pride and Prejudice")
.exec()
Scan
You can scan
every record in your DynamoDB table using an API that closely mirrors the query
API. For example:
import { AuthorPartition } from "generated/models"
// Scan through everything in the table and load it into memory (not recommended for prod)
const results = await beyonce
.scan()
.exec() // returns { Author: Author[], Book: Book[] }
const iterator = beyonce
.scan()
.iterator({ pageSize: 1 })
// Step through each page 1 by 1
for await (const { items } of iterator) {
// ...
}
Parallel Scans
You can perform "parallel scans" by passing a parallel
config operation to the .scan
method,
like so:
// Somewhere inside of Worker 1
const segment1 = beyonce
.scan({ parallel: { segmentId: 0, totalSegments: 2 }})
.iterator()
for await (const results of segment1) {
// ...
}
// Somewhere inside of Worker 2
const segment2 = beyonce
.scan({ parallel: { segmentId: 1, totalSegments: 2 }})
.iterator()
for await (const results of segment2) {
// ...
}
These options mirror the underlying DynamoDB API
BatchGet
You can retrieve records in bulk via batchGet
. DynamoDB allows retrieving a maximum of 100 items per
batchGet
query. So, if you ask for more than 100 keys in a single Beyonce batchGet
call, Beyonce will automatically split
DynamoDB calls into N concurrent requests and join the results for you.
// Batch get several items
const { items, unprocessedKeys } = await beyonce.batchGet({
keys: [
// Get 2 authors
AuthorModel.key({ id: "1" }),
AuthorModel.key({ id: "2" }),
// And a specific book from each
Book.key({ authorId: "1", id: "1" })
Book.key({ authorId: "2" id: "2" })
]
})
// And the return type is:
// { author: Author[], book: Book[] }
const { Author, Book } = items
If the unprocessedKeys
array isn't empty, you can retry
via:
await beyonce.batchGet({ keys: unprocessedKeys })
BatchWrite
You can batch put/delete records using batchWrite
. If any operations can't be processed,
you'll get a populated unprocessedPuts
array and/or an unprocessedDeletes
array back.
// Batch put or delete several items at once
const author1 = AuthorModel.create({
id: "1",
name: "Jane Austen"
})
const author2 = AuthorModel.create({
id: "2",
name: "Charles Dickens"
})
const {
unprocessedPuts,
unprocessedDeletes
} = await beyonce.batchWrite({ putItems: [author1], deleteItems: [Author.key({ id: author2.id })] })
BatchWriteWithTransaction
If you'd like to batch pute/delete records in an atomic transaction, you can use batchWriteWithTransaction
.
And all operations will either succeed, or fail.
await beyonce.batchWriteWithTransaction({ putItems: [author1], deleteItems: [Author.key({ id: author2.id })] })
You can also pass a string clientRequestToken
to batchWriteWithTransaction
to force your operations to
be idempotent, per the AWS docs.
Consistent Reads
Beyonce supports consistent reads via an optional parameter on get
, batchGet
and query
, e.g. get(..., { consistentRead: true })
.
And if you'd like to always make consistent reads by default, you can set this as the default when you create a Beyonce instance:
new Beyonce(table, dynamo, { consistentReads: true })
Note: When you enable consistentReads on a Beyonce instance, you can override it on a per-operation basis by setting the method level consistentRead
option.
Things beyonce should do, but doesn't (yet)
- Support the full range of Dynamo filter expressions
An aside on storing heterogenous models in the same table
When using DynamoDB, you often want to "pre-compute" joins by sticking a set of heterogeneous models into the same table, under the same partition key. This allows for retrieving related records using a single query instead of N.
Unfortunately most existing DynamoDB libraries, like DynamoDBMapper, don't support this use case as they follow the SQL convention sticking each model into a separte table.
For example, we might want to fetch an Author
+ all their Book
s in a single query. And we'd accomplish that by sticking both models
under the same partition key - e.g. author-${id}
.
AWS's guidelines, take this to the extreme:
...most well-designed applications require only one table
Keep in mind that the primary reason they recommened this is to avoid forcing the application-layer to perform in-memory joins. Due to Amazon's scale, they are highly motivated to minimize the number of roundtrip db calls.
You are probably not Amazon scale. And thus probably don't need to shove everything into a single table.
But you might want to keep a few related models in the same table, under the same partition key and fetch those models in a type-safe way. Beyonce makes that easy.
Misc
You can enable AWS XRay tracing like so:
const beyonce = new Beyonce(
LibraryTable,
dynamo,
{ xRayTracingEnabled: true }
)