Tags Archives: dax

AWS DynamoDB

DynamoDB is a fully managed, highly-available database with replication across multiple AZs

 

NoSQL – not a relational database! – just simple key:value

 

single digit millisecond performance.

 

 

scales to massive workloads

 

100TBs of storage

 

fast and consistence, low latency retrieval

 

integrated with iam for security authorization and admin
enables event-driven programming via DynamoDB Streams

 

low cost and auto-scaling

 

important for exam: 

no db provisioning needed you just create a table

 

 

Has two able classes: standard and infrequent access IA table class

 

 

Basics of DynamoDB

 

 

v important – also exam q:
made up of tables – there is NO SUCH THING AS A DATABASE – you only need to create tables!
the db already exists!

 

 

you just create tables – 

 

each table has a primary key – must be set at creation time
each table can have an infinite number of items ie rows

each item has attributes can be added over time or they can be null

this is much easier to do at any time than with a conventional relational db.

 

max size of an item is 400KB so not good for large objects

 

data types supported are:

 

scalar types: string, number, binary, boolean, null
document types: list, map
set types: string set, number set, binary set

 

tables:

 

have primary key – containing partition key (eg user_id) and sort key (eg game_id)
and
attributes: eg score, result – these are items which are NOT the primary key

 

 

Is a great choice for a schema where you need to rapidly evolve the schema, better than the others for this.

 

read/write capacity modes

 

control how you manage your table capacity – ie the read and write throughput

 

2 Capacity Modes for DynamoDB

 

important for exam:

 

Provisioned mode – is default,  provisioned in advance

 

we have 2 possible capacity modes:

 

provisioned mode, where you specify number of read and writes per second, this is the default.

 

very good for predictable modes.

 

You have to plan the capacity for this in advance, you pay for provisioned read capacity units RCUs and write capacity units (WCUs)

 

these “capacity units” are used to set your desired capacity for your database! – set these in the web dashboard of dynamodb when you are provisioning.

 

you can also add auto scaling mode for both rcu and wcu

 

and

 

on-demand mode – much more expensive, but better for unpredictable workloads or sudden transaction spikes where provisioned mode cant scale sufficiently 

 

on-demand mode:

read write automatically scales up and down acc to workload, so no capacity planning needed
you pay for your use, is more expensive 2-3x more

 

good for unpredictable workloads which can be large or small varying

 

need to know for exam!

 

remember always you just create tables, never a database with dynamodb!

 

you can specify read and write capacity autoscaling separately

 

 

DynamoDB Accelerator DAX

 

 

a fully manageable, highly available seamless in-memory cache for dynamodb

 

helps solve read congestion by means of memory caching

 

microseconds latency for cached data

 

does not require any application logic modification – so it is fully compatible with existing dynamodb api

 

applications can access the database via DAX for much faster reads

 

TTL is 5 mins default for dax data

 

dax is different from elasticache in that it is meant for dynamodb and does not change the api
good for individual object cache

 

elasticache
good for storing aggregation result where further processing required

 

 

DynamoDB Stream

 

very easy – is an ordered stream of data that represents items when created or updated or deleted

 

can then be sent on to kinesis datastreams
can be read by lambda or kinesis client library apps

 

data is retained for 24 hrs

 

 

use cases:

 

reacting to changes in real time eg welcome emails to new users
to do analytics
to insert into derivative tables or into elastic serarch
or implement cross region replication

 

 

Summary of DynamoDB Streams

 

So, to summarize DynamoDB Streams

 

Application -> create/update/delete actions to the TABLE -> Dynamo DB Streams

 

and from the TABLE -> Kinesis Data Streams -> Kinesis Data Firehose

 

Kinesis Data Firehose

-> for analytics -> Redshift
-> for archiving -> S3
-> for indexing -> OpenSearch

 

 

 

DynamoDB Global Tables

 

 

are cross-region

 

two way or multi-way replication

 

to provide for low latency across regions

 

it is
active-active replication

 

this means apps can read and write to the table from any region

 

must enable dynamodb streams for this

 

TTL expiry: automatically deletes items after an expiry timestamp

 

eg 1 month for each row in the table

 

 

DynamoDB Indexes – only need to know at high level for exam

 

basically, these allow you to query on attributes other than the primary key

 

2 types:

 

gsi – global secondary and lsi – local secondary

 

(all you need to know for now)

 

by default you query on the primary key, but you can use gsi or lsi to query on other attributes.

 

 

Transactions in DynamoDB

 

 

these allow you to write to 2 tables at the same time:

 

the transaction however MUST write to both or else none – in order to keep the tables accurate – so data consistency is maintained

 

 

 

 

 

 

 

 

Continue Reading