Add caching with LRU cache under mutex

Caches responses of each GET handler in a separate capacity-limited cache (as a
custom clone-able `CachedResponse` struct). Subsequent requests will build a
`Response` from the cached bytes instead of re-querying the database and
re-serializing the JSON. This greatly speeds up the list endpoints and
`get_interior_ref_list`.

Also caches the api-key-to-id mapping for `Owner`s in order to speed up frequent
authentications.

Each create handler clears the entire list response cache. Each delete handler
also clears the entire list response cache and deletes the cached response for
that key. Deleting an owner also deletes their entry in the
`owner_ids_by_api_key` cache.
This commit is contained in:
2020-08-01 00:25:04 -04:00
parent 68b04b4f4c
commit 519fcb4c5a
14 changed files with 470 additions and 123 deletions

35
src/caches/mod.rs Normal file
View File

@@ -0,0 +1,35 @@
use std::fmt::Debug;
use uuid::Uuid;
use crate::models::ListParams;
mod cache;
mod cached_response;
pub use cache::Cache;
pub use cached_response::CachedResponse;
#[derive(Debug, Clone)]
pub struct Caches {
pub owner_ids_by_api_key: Cache<Uuid, i32>,
pub shop: Cache<i32, CachedResponse>,
pub owner: Cache<i32, CachedResponse>,
pub interior_ref_list: Cache<i32, CachedResponse>,
pub list_shops: Cache<ListParams, CachedResponse>,
pub list_owners: Cache<ListParams, CachedResponse>,
pub list_interior_ref_lists: Cache<ListParams, CachedResponse>,
}
impl Caches {
pub fn initialize() -> Self {
Caches {
owner_ids_by_api_key: Cache::new("owner_ids_by_api_key", 100).log_keys(false),
shop: Cache::new("shop", 100),
owner: Cache::new("owner", 100),
interior_ref_list: Cache::new("interior_ref_list", 100),
list_shops: Cache::new("list_shops", 100),
list_owners: Cache::new("list_owners", 100),
list_interior_ref_lists: Cache::new("list_interior_ref_lists", 100),
}
}
}