Rust ORM for ScyllaDB
Use monstrous tandem of scylla and charybdis for your next project
⚠️ WIP: This project is currently in an experimental stage. It's not recommended to use it in production yet.
Charybdis is a ORM layer on top of scylla_rust_driver focused on easy of use and performance
Announcements:
-
Queries are now configurable
With0.4.0release we have provided users with ability to configure each query before execution -
Breaking changes
-
Operations:
find,insert,update,deletenow returnCharybdisQuerythat can be configured before execution.let mut user = user.find_by_primary_key.consistency.execute; -
Callbacks: We now have only single
Callbackstrait that is used for all operation that can accept extension. In case extension is not needed, we can use()or Option<()> and provideNoneas extension argument. -
Batch Operations: Batch is now coupled with Model and it's created by calling
Model::batch()method. It can also be configured before execution.let batch = batch.consistency.chunked_insert.await?;
-
Usage considerations:
- Provide and expressive API for CRUD & Complex Query operations on model as a whole
- Provide easy way to work with subset of model fields by using automatically generated
partial_<model>!macro - Provide easy way to run complex queries by using automatically generated
find_<model>!macro - Automatic migration tool that analyzes the
src/model/*.rsfiles and runs migrations according to differences between the model definition and database
Performance consideration:
- It's build by beta release, so it uses builtin support for
async/awaitin traits that will be stabilized in Rust1.75 - It uses prepared statements (shard/token aware) -> bind values
- It expects
CachingSessionas a session arg for operations - Queries are macro generated str constants (no concatenation at runtime)
- By using
find_<model>!macro we can run complex queries that are generated at compile time as&'static str - Although it has expressive API it's thin layer on top of scylla_rust_driver, and it does not introduce any significant overhead
Table of Contents
- Charybdis Models
- Automatic migration with
charybdis-migrate - Basic Operations
- Configuration Options
- Batch Operations
- Partial Model
- Callbacks
- Collection
- Ignored fields
- Roadmap
Charybdis Models
Define Tables
Declare model as a struct within src/models dir:
// src/models/user.rs
use charybdis_model;
use ;
(Note we use src/models as automatic migration tool expects that dir)
Define UDT
src/models/udts
// src/models/udts/address.rs
use charybdis_udt_model;
use Text;
Define Materialized Views
src/models/materialized_views
// src/models/materialized_views/users_by_username.rs
use charybdis_view_model;
use ;
Resulting auto-generated migration query will be:
CREATE MATERIALIZED VIEW IF NOT EXISTS users_by_email
AS SELECT created_at, updated_at, username, email, id
FROM users
WHERE email IS NOT NULL AND id IS NOT NULL
PRIMARY KEY (email, id)
Automatic migration
charybdis-migrate tool that enables automatic migration to database without need to write migrations by hand.
It expects src/models files and generates migrations based on differences between model definitions and database.
It supports following operations:
- Create new tables
- Create new columns
- Drop columns
- Change field types (drop and recreate column
--drop-and-replaceflag) - Create secondary indexes
- Drop secondary indexes
- Create UDTs (
src/models/udts) - Create materialized views (
src/models/materialized_views) - Table options
⚠️ If table exists, table options will result in alter table query that withoutCLUSTERING ORDERandCOMPACT STORAGEoptions.
Model dropping is not added. If you don't define model within src/model dir
it will leave db structure as it is.
)
⚠️ If you are working with existing datasets, before running migration you need to make sure that your model
definitions structure matches the database in respect to table names, column names, column types, partition keys,
clustering keys and secondary indexes so you don't alter structure accidentally.
If structure is matched, it will not run any migrations. As mentioned above,
in case there is no model definition for table, it will not drop it. In future,
we will add modelize command that will generate src/models files from existing data source.
Global secondary indexes
Local secondary Indexes
They are defined as array of tuples
- first element is array of partition keys
- second element is array of clustering keys
)]
resulting query will be: CREATE INDEX ON menus((location), dish_type);
Basic Operations:
For each operation you need to bring respective trait into scope. They are defined
in charybdis::operations module.
Insert
-
use ; async
Find
-
Find by primary key
let user = User ; let user = user.find_by_primary_key.execute.await?; -
Find by partition key
let users = User .find_by_partition_key.execute.await; -
Find by primary key associated
let users = find_by_primary_key_value.execute.await; -
Macro generated find helpers
Lets say we have model:
We have macro generated functions for up to 3 fields from primary key. Note that if complete primary key is provided, we get single typed result. So in case of our User model, we would get:
find_by_date.execute .execute .executeAnd for our user model we would have
find_by_id.execute -
Custom filtering:
Lets use our
Postmodel as an example:We get automatically generated
find_post!macro that follows conventionfind_<struct_name>!. It can be used to create custom queries.Following will return stream of
Postmodels, and query will be constructed at compile time as&'static str.// automatically generated macro rule let posts = find_post!.execute.await?;We can also use
find_first_post!macro to get single result:let post = find_first_post!.execute.await?;If we just need the
Queryand not the result, we can usefind_post_query!macro:let query = find_post_query!
Update
-
let user = from_json; user.username = "scylla".to_string; user.email = "some@email.com"; user.update.execute.await; -
Collection:
- Let's use our
Usermodel as an example: push_to_<field_name>andpull_from_<field_name>methods are generated for each collection field.let user: User; user.push_tags.execute.await; user.pull_tags.execute.await; user.push_post_ids.execute.await; user.pull_post_ids.execute.await;
- Let's use our
-
Counter
- Let's define post_counter model:
- We can use
increment_<field_name>anddecrement_<field_name>methods to update counter fields.let post_counter: PostCounter; post_counter.increment_likes.execute.await; post_counter.decrement_likes.execute.await; post_counter.increment_comments.execute.await; post_counter.decrement_comments.execute.await;
- Let's define post_counter model:
Delete
-
let user = from_json; user.delete.execute.await; -
Macro generated delete helpers
Lets use our
Postmodel as an example:We have macro generated functions for up to 3 fields from primary key.
delete_by_date.execute.await?; delete_by_date_and_category_id.execute.await?; delete_by_date_and_category_id_and_title.execute.await?;
Configuration
Every operation returns CharybdisQuery that can be configured before execution with method chaining.
let user: User = find_by_id
.consistency
.timeout
.execute
.await?;
let result: QueryResult = user.update.consistency.execute.await?;
Supported configuration options:
consistencyserial_consistencytimestamptimeoutpage_sizetimestamp
Batch
CharybdisModelBatch operations are used to perform multiple operations in a single batch.
-
Batch Operations
let users: ; let batch = batch; // inserts batch.append_inserts; // or updates batch.append_updates; // or deletes batch.append_deletes; batch.execute.await?; -
Chunked Batch Operations
Chunked batch operations are used to operate on large amount of data in chunks.
let users: ; let chunk_size = 100; batch.chunked_inserts.await?; batch.chunked_updates.await?; batch.chunked_deletes.await?; -
Batch Configuration
Batch operations can be configured before execution with method chaining.
let batch = batch .consistency .retry_policy .chunked_inserts.await?; .await?; -
Statements Batch
We can use batch statements to perform collection operations in batch:
let batch = batch; let users: ; for user in users batch.execute.await;
Partial Model:
-
Use auto generated
partial_<model>!macro to run operations on subset of the model fields. This macro generates a new struct with same structure as the original model, but only with provided fields. Macro is automatically generated by#[charybdis_model]. It follows conventionpartial_<struct_name>!.// auto-generated macro - available in crate::models::user partial_user!; let id = new_v4; let user = UpdateUsernameUser ; // we can have same operations as on base model // INSERT into users (id, username) VALUES (?, ?) user.insert.execute.await; // UPDATE users SET username = ? WHERE id = ? user.update.execute.await; // DELETE FROM users WHERE id = ? user.delete.execute.await; // get partial PartUser let partial_user = user.find_by_primary_key.await?; // get native user model by primary key let user = user.as_native.find_by_primary_key.execute.await?; -
Partial Model Considerations:
partial_<model>requires#[derive(Default)]on original modelpartial_<model>require complete primary key in definition- All derives that are defined bellow
#charybdis_modelmacro will be automatically added to partial model. partial_<model>struct implements same field attributes as original model, so if we have#[serde(rename = "rootId")]on original model field, it will be present on partial model field.
-
As Native
In case we need to run operations on native model, we can use
as_nativemethod:partial_user!; let mut update_user_username = UpdateUser ; let native_user: User = update_user_username.as_native.find_by_primary_key.execute.await?; // action that requires native model authorize_user;as_nativeworks by returning new instance of native model with fields from partial model. For other fields it uses default values. -
Recommended naming convention is
Purpose+Original Struct Name. E.g:UpdateAdresssUser,UpdateDescriptionPost.
Callbacks
Callbacks are convenient way to run additional logic on model before or after certain operations. E.g.
- we can use
before_insertto set default values and/or validate model before insert. - we can use
after_updateto update other data sources, e.g. elastic search.
Implementation:
- Let's say we define custom extension that will be used to
update elastic document on every post update:
- Now we can implement Callback that will utilize this extension:
-
Triggering Callbacks
In order to trigger callback we use<operation>_cb. method:insert_cb,update_cb,delete_cbaccording traits. This enables us to have clear distinction betweeninsertand insert with callbacks (insert_cb).use ; post.insert_cb.execute.await; post.update_cb.execute.await; post.delete_cb.execute.await;
Collections
-
For each collection field that is defined as
List<T> orSet`, we get following collection queries:PUSH_<field_name>_QUERYstatic strPULL_<field_name>_QUERYstatic strpush_<field_name>methodpull_<field_name>method
-
Define Model:
-
Generated Collection Queries:
PUSH_TAGS_QUERY; PULL_TAGS_QUERY; PUSH_POST_IDS_QUERY; PULL_POST_IDS_QUERY;Generated query will expect value as first bind value and primary key fields as next bind values.
Now we could use this constant within Batch operations.
let batch = batch; let users: ; for user in users batch.execute.await; -
Generated Collection Methods:
push_to_<field_name>andpull_from_<field_name>methods are generated for each collection field.let user: User; user.push_tags.execute.await; user.pull_tags.execute.await; user.push_post_ids.execute.await; user.pull_post_ids.execute.await;
Ignored fields
We can ignore fields by using #[charybdis(ignore)] attribute:
So field organization will be ignored in all operations and
default value will be used when deserializing from other data sources.
It can be used to hold data that is not persisted in database.
Roadmap:
- Add tests
- Write
modelizecommand to generatesrc/models/*structs from existing database