Skip to content

A schema-based DynamoDB modeling tool, high-level API, and type-generator built to supercharge single-table designs⚡

License

Notifications You must be signed in to change notification settings

Nerdware-LLC/ddb-single-table

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

ddb-single-table banner

A schema-based DynamoDB modeling tool, high-level API, and type-generator
built to supercharge single-table designs!⚡

Marshalling ✅ Validation ✅ Where-style query API ✅ and more.
Fully-typed support for ESM and CommonJS

npm package   Test Workflow   Codecov   pre-commit   semantic-release   License: MIT


🚀 Getting Started

  1. Install the package:

    npm install @nerdware/ddb-single-table
  2. Create your table:

    import { Table } from "@nerdware/ddb-single-table";
    
    // OR const { Table } = require("@nerdware/ddb-single-table");
    
    export const myTable = new Table({
      tableName: "my-table-name",
      // The `tableKeysSchema` includes all table and index keys:
      tableKeysSchema: {
        pk: {
          type: "string", // keys can be "string", "number", or "Buffer"
          required: true,
          isHashKey: true,
        },
        sk: {
          type: "string",
          required: true,
          isRangeKey: true,
          index: {
            // This index allows queries using "sk" as the hash key
            name: "Overloaded_SK_GSI",
            rangeKey: "data",
            global: true,
            project: true, // project all attributes
            throughput: { read: 5, write: 5 },
          },
        },
        data: {
          type: "string",
          required: true,
          index: {
            // This index allows queries using "data" as the hash key
            name: "Overloaded_Data_GSI",
            rangeKey: "sk",
            global: true,
            project: true, // project all attributes
            throughput: { read: 5, write: 5 },
          },
        },
      } as const, // For TypeScript, all schema must end with `as const`
      // You can provide your own DDB client instance or configs for a new one:
      ddbClient: {
        // This example shows how to connect to dynamodb-local:
        region: "local",
        endpoint: "http://localhost:8000",
        // All AWS SDK client auth methods are supported.
        // Since this example is using dynamodb-local, we simply use
        // hard-coded "local" credentials, but for production you would
        // obviously want to use a more secure method like an IAM role.
        credentials: {
          accessKeyId: "local",
          secretAccessKey: "local",
        },
      },
    });
  3. Create a model, and generate item-typings from its schema:

    import { myTable } from "./path/to/myTable.ts";
    import { isValid } from "./path/to/some/validators.ts";
    import type { ItemTypeFromSchema } from "@nerdware/ddb-single-table";
    
    const UserModel = myTable.createModel({
      pk: {
        type: "string",
        alias: "id", // <-- Each Model can have custom aliases for keys
        default: ({ createdAt }: { createdAt: Date }) => {
         return `USER#${createdAt.getTime()}`
        },
        validate: (id: string) => /^USER#\d{10,}$/.test(id),
        required: true,
      },
      sk: {
        type: "string",
        default: (userItem: { pk: string }) => {
           return `#DATA#${userItem.pk}`
        },
        validate: (sk: string) => /^#DATA#USER#\d{10,}$/.test(sk)
        required: true,
      },
      data: {
        type: "string",
        alias: "email",
        validate: (value: string) => isValid.email(value),
        required: true,
      },
      profile: {
       type: "map", // Nested attributes ftw!
       schema: {
         displayName: { type: "string", required: true },
         businessName: { type: "string" },
         photoUrl: { type: "string" },
         // You can nest attributes up to the DynamoDB max depth of 32
       },
      },
      checklist: {
        type: "array",
        required: false,
        schema: [
          {
            type: "map",
            schema: {
              id: {
                // Nested attributes have the same awesome schema capabilities!
                type: "string",
                default: (userItem: { sk: string }) => {
                 return `FOO_CHECKLIST_ID#${userItem.sk}#${Date.now()}`
                },
                validate: (id: string) => isValid.checklistID(id),
                required: true,
               },
               description: { type: "string", required: true },
               isCompleted: { type: "boolean", required: true, default: false },
             },
           },
         ],
       },
       /* By default, 'createdAt' and 'updatedAt' attributes are created for
       each Model (unless explicitly disabled). Here's an example with these
       attributes explicitly provided:                                    */
       createdAt: {
         type: "Date",
         required: true,
         default: () => new Date()
       },
       updatedAt: {
         type: "Date",
         required: true,
         default: () => new Date(),
         /* transformValue offers powerful hooks which allow you to modify values TO
         and/or FROM the db. Each attribute can define its own transformValue hooks.*/
         transformValue: {
           toDB: () => new Date(), /* <-- For data traveling TO the db (write ops).
           transformValue can also include a `fromDB` fn to transform values
              coming FROM the db. If specified, your `fromDB` transforms are
              applied for both write and read operations. */
         },
       },
    } as const); // <-- Don't forget to add `as const`!
    
    // The `ItemTypeFromSchema` type is a helper type which converts
    // your schema into a Typescript type for your model's items.
    export type UserItem = ItemTypeFromSchema<typeof UserModel.schema>;
  4. Use your model and generated types:

    import { UserModel, type UserItem } from "./path/to/UserModel.ts";
    
    // Create a new user:
    const newUser = await UserModel.createItem({
      email: "human_person@example.com",
      profile: {
        displayName: "Guy McHumanPerson",
        businessName: "Definitely Not a Penguin in a Human Costume, LLC",
        photoUrl: "s3://my-bucket-name/path/to/human/photo.jpg",
      },
      checklist: [
        { description: "Find fish to eat" },
        { description: "Return human costume by 5pm" },
      ],
    });
    
    // You can use explicit type annotations, or allow TS to infer types.
    // For example, the line below yields the same as the above example:
    //   const newUser: UserItem = await UserModel.createItem(...);
    
    // The `newUser` is of type `UserItem`, with all keys aliased as specified:
    const { id, sk, email, profile, checklist, createdAt, updatedAt }: UserItem = {
      ...newUser,
    };
    
    // You can also use the model to query for items using `where` syntax:
    const usersWhoAreDefinitelyHuman = await UserModel.query({
      where: {
        email: {
          beginsWith: "human_", // <-- All DDB operators are supported!
        },
      },
    });
    
    // There are a lot more features I've yet to document, but hopefully
    // this is enough to get you started! Pull requests are welcome! 🐧
  5. Profit! 💰🥳🎉

✨ Key Features

  • Easy-to-use declarative API for managing DDB tables, connections, and models
  • Auto-generated typings for model items
  • Custom attribute aliases for each model
  • Create attributes/properties from combinations of other attributes/properties
  • Type checking and conversions for all DDB attribute types
  • Validation checks for individual properties
  • Validation checks for entire objects
  • Where-style query API
  • Default values
  • Property-level get/set modifiers
  • Schema-level get/set modifiers
  • Required/nullable property assertions
  • Easy access to a streamlined DynamoDB client (more info here)
  • Automatic retries for batch operations using exponential backoff (more info here)
  • Support for transactions - group up to 100 operations into a single atomic transaction!

Batch Retries with Exponential Backoff

As recommended by AWS, DDB-ST will automatically retry batch operations which either return unprocessed requests (e.g., UnprocessedKeys for BatchGetItem), or result in a retryable error. All retries are implemented using a configurable exponential backoff strategy which adheres to AWS best practices.

Click here for a detailed overview of the exponential backoff strategy.
  1. First request: no delay

  2. Second request: delay initialDelay milliseconds (default: 100)

  3. All subsequent request delays are equal to the previous delay multiplied by the timeMultiplier (default: 2), until either:

    • The maxRetries limit is reached (default: 10), or
    • The maxDelay limit is reached (default: 3500, or 3.5 seconds)

    Ergo, the base delay calculation can be summarized as follows:

    delay in milliseconds = initialDelay * timeMultiplier^attemptNumber

    If useJitter is true (default: false), the delay is randomized by applying the following to the base delay:

    Math.round( Math.random() * delay )

    Note that the determination as to whether the delay exceeds the maxDelay is made BEFORE the jitter is applied.

❓ FAQ

Q: Why "single-table-first"?

A: Single-table design patterns can yield both greater IO and cost performance, while also reducing the amount of infrastructure that needs to be provisioned and maintained. For a technical breakdown as to why this is the case, check out this fantastic presentation from one of the designers of DynamoDB speaking at AWS re:Invent.

Q: How does DDB-ST interact with the underlying DynamoDB client?

A: DDB-ST provides a single streamlined abstraction over both the document and vanilla DynamoDB clients:

  • CRUD actions use the document client to provide built-in marshalling/unmarshalling of DDB-attribute objects.
  • Utility actions like DescribeTable which aren't included in the document client use the vanilla client.
  • To ensure client resources like socket connections are cleaned up, a listener is attached to the process "exit" event which calls the vanilla client's destroy() method. Note that although the document client does expose the same method, calling it on the doc-client results in a no-op.

Q: What version of the AWS SDK does DDB-ST use?

A: Version 3. For the specific minor/patch release, please refer to the package.json.

🤝 Contributing

Pull requests are welcome! Before you begin, please check existing GitHub Issues and Pull Requests to see if your idea is already in the pipeline. If not, here's a guide on how to contribute to this project. Thank you!

📝 License

ddb-single-table is open-source software licensed under an MIT License.

💬 Contact

Trevor Anderson — Trevor@Nerdware.cloud@TeeRevTweets

Check out Nerdware on YouTubeTrevor Anderson's LinkedInTrevor Anderson's TwitterEmail Trevor Anderson

Dare Mighty Things.