281 lines
8.7 KiB
Rust
281 lines
8.7 KiB
Rust
// Copyright 2020 The Matrix.org Foundation C.I.C.
|
|
//
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
//! Collection of small in-memory stores that can be used to cache Olm objects.
|
|
//!
|
|
//! Note: You'll only be interested in these if you are implementing a custom
|
|
//! `CryptoStore`.
|
|
|
|
use std::{collections::HashMap, sync::Arc};
|
|
|
|
use dashmap::DashMap;
|
|
use matrix_sdk_common::locks::Mutex;
|
|
use ruma::{DeviceId, DeviceIdBox, RoomId, UserId};
|
|
|
|
use crate::{
|
|
identities::ReadOnlyDevice,
|
|
olm::{InboundGroupSession, Session},
|
|
};
|
|
|
|
/// In-memory store for Olm Sessions.
|
|
#[derive(Debug, Default, Clone)]
|
|
pub struct SessionStore {
|
|
entries: Arc<DashMap<String, Arc<Mutex<Vec<Session>>>>>,
|
|
}
|
|
|
|
impl SessionStore {
|
|
/// Create a new empty Session store.
|
|
pub fn new() -> Self {
|
|
Self::default()
|
|
}
|
|
|
|
/// Add a session to the store.
|
|
///
|
|
/// Returns true if the session was added, false if the session was
|
|
/// already in the store.
|
|
pub async fn add(&self, session: Session) -> bool {
|
|
let sessions_lock = self
|
|
.entries
|
|
.entry(session.sender_key.to_string())
|
|
.or_insert_with(|| Arc::new(Mutex::new(Vec::new())));
|
|
|
|
let mut sessions = sessions_lock.lock().await;
|
|
|
|
if !sessions.contains(&session) {
|
|
sessions.push(session);
|
|
true
|
|
} else {
|
|
false
|
|
}
|
|
}
|
|
|
|
/// Get all the sessions that belong to the given sender key.
|
|
pub fn get(&self, sender_key: &str) -> Option<Arc<Mutex<Vec<Session>>>> {
|
|
#[allow(clippy::map_clone)]
|
|
self.entries.get(sender_key).map(|s| s.clone())
|
|
}
|
|
|
|
/// Add a list of sessions belonging to the sender key.
|
|
pub fn set_for_sender(&self, sender_key: &str, sessions: Vec<Session>) {
|
|
self.entries.insert(sender_key.to_owned(), Arc::new(Mutex::new(sessions)));
|
|
}
|
|
}
|
|
|
|
#[derive(Debug, Default, Clone)]
|
|
/// In-memory store that holds inbound group sessions.
|
|
pub struct GroupSessionStore {
|
|
#[allow(clippy::type_complexity)]
|
|
entries: Arc<DashMap<RoomId, HashMap<String, HashMap<String, InboundGroupSession>>>>,
|
|
}
|
|
|
|
impl GroupSessionStore {
|
|
/// Create a new empty store.
|
|
pub fn new() -> Self {
|
|
Self::default()
|
|
}
|
|
|
|
/// Add an inbound group session to the store.
|
|
///
|
|
/// Returns true if the session was added, false if the session was
|
|
/// already in the store.
|
|
pub fn add(&self, session: InboundGroupSession) -> bool {
|
|
self.entries
|
|
.entry((&*session.room_id).clone())
|
|
.or_insert_with(HashMap::new)
|
|
.entry(session.sender_key.to_string())
|
|
.or_insert_with(HashMap::new)
|
|
.insert(session.session_id().to_owned(), session)
|
|
.is_none()
|
|
}
|
|
|
|
/// Get all the group sessions the store knows about.
|
|
pub fn get_all(&self) -> Vec<InboundGroupSession> {
|
|
self.entries
|
|
.iter()
|
|
.flat_map(|d| {
|
|
d.value()
|
|
.values()
|
|
.flat_map(|t| t.values().cloned().collect::<Vec<InboundGroupSession>>())
|
|
.collect::<Vec<InboundGroupSession>>()
|
|
})
|
|
.collect()
|
|
}
|
|
|
|
/// Get a inbound group session from our store.
|
|
///
|
|
/// # Arguments
|
|
/// * `room_id` - The room id of the room that the session belongs to.
|
|
///
|
|
/// * `sender_key` - The sender key that sent us the session.
|
|
///
|
|
/// * `session_id` - The unique id of the session.
|
|
pub fn get(
|
|
&self,
|
|
room_id: &RoomId,
|
|
sender_key: &str,
|
|
session_id: &str,
|
|
) -> Option<InboundGroupSession> {
|
|
self.entries
|
|
.get(room_id)
|
|
.and_then(|m| m.get(sender_key).and_then(|m| m.get(session_id).cloned()))
|
|
}
|
|
}
|
|
|
|
/// In-memory store holding the devices of users.
|
|
#[derive(Clone, Debug, Default)]
|
|
pub struct DeviceStore {
|
|
entries: Arc<DashMap<UserId, DashMap<Box<DeviceId>, ReadOnlyDevice>>>,
|
|
}
|
|
|
|
impl DeviceStore {
|
|
/// Create a new empty device store.
|
|
pub fn new() -> Self {
|
|
Self::default()
|
|
}
|
|
|
|
/// Add a device to the store.
|
|
///
|
|
/// Returns true if the device was already in the store, false otherwise.
|
|
pub fn add(&self, device: ReadOnlyDevice) -> bool {
|
|
let user_id = device.user_id();
|
|
self.entries
|
|
.entry(user_id.to_owned())
|
|
.or_insert_with(DashMap::new)
|
|
.insert(device.device_id().into(), device)
|
|
.is_none()
|
|
}
|
|
|
|
/// Get the device with the given device_id and belonging to the given user.
|
|
pub fn get(&self, user_id: &UserId, device_id: &DeviceId) -> Option<ReadOnlyDevice> {
|
|
self.entries.get(user_id).and_then(|m| m.get(device_id).map(|d| d.value().clone()))
|
|
}
|
|
|
|
/// Remove the device with the given device_id and belonging to the given
|
|
/// user.
|
|
///
|
|
/// Returns the device if it was removed, None if it wasn't in the store.
|
|
pub fn remove(&self, user_id: &UserId, device_id: &DeviceId) -> Option<ReadOnlyDevice> {
|
|
self.entries.get(user_id).and_then(|m| m.remove(device_id)).map(|(_, d)| d)
|
|
}
|
|
|
|
/// Get a read-only view over all devices of the given user.
|
|
pub fn user_devices(&self, user_id: &UserId) -> HashMap<DeviceIdBox, ReadOnlyDevice> {
|
|
self.entries
|
|
.entry(user_id.clone())
|
|
.or_insert_with(DashMap::new)
|
|
.iter()
|
|
.map(|i| (i.key().to_owned(), i.value().clone()))
|
|
.collect()
|
|
}
|
|
}
|
|
|
|
#[cfg(test)]
|
|
mod test {
|
|
use ruma::room_id;
|
|
|
|
use crate::{
|
|
identities::device::test::get_device,
|
|
olm::{test::get_account_and_session, InboundGroupSession},
|
|
store::caches::{DeviceStore, GroupSessionStore, SessionStore},
|
|
};
|
|
|
|
#[tokio::test]
|
|
async fn test_session_store() {
|
|
let (_, session) = get_account_and_session().await;
|
|
|
|
let store = SessionStore::new();
|
|
|
|
assert!(store.add(session.clone()).await);
|
|
assert!(!store.add(session.clone()).await);
|
|
|
|
let sessions = store.get(&session.sender_key).unwrap();
|
|
let sessions = sessions.lock().await;
|
|
|
|
let loaded_session = &sessions[0];
|
|
|
|
assert_eq!(&session, loaded_session);
|
|
}
|
|
|
|
#[tokio::test]
|
|
async fn test_session_store_bulk_storing() {
|
|
let (_, session) = get_account_and_session().await;
|
|
|
|
let store = SessionStore::new();
|
|
store.set_for_sender(&session.sender_key, vec![session.clone()]);
|
|
|
|
let sessions = store.get(&session.sender_key).unwrap();
|
|
let sessions = sessions.lock().await;
|
|
|
|
let loaded_session = &sessions[0];
|
|
|
|
assert_eq!(&session, loaded_session);
|
|
}
|
|
|
|
#[tokio::test]
|
|
async fn test_group_session_store() {
|
|
let (account, _) = get_account_and_session().await;
|
|
let room_id = room_id!("!test:localhost");
|
|
|
|
let (outbound, _) =
|
|
account.create_group_session_pair_with_defaults(&room_id).await.unwrap();
|
|
|
|
assert_eq!(0, outbound.message_index().await);
|
|
assert!(!outbound.shared());
|
|
outbound.mark_as_shared();
|
|
assert!(outbound.shared());
|
|
|
|
let inbound = InboundGroupSession::new(
|
|
"test_key",
|
|
"test_key",
|
|
&room_id,
|
|
outbound.session_key().await,
|
|
None,
|
|
)
|
|
.unwrap();
|
|
|
|
let store = GroupSessionStore::new();
|
|
store.add(inbound.clone());
|
|
|
|
let loaded_session = store.get(&room_id, "test_key", outbound.session_id()).unwrap();
|
|
assert_eq!(inbound, loaded_session);
|
|
}
|
|
|
|
#[tokio::test]
|
|
async fn test_device_store() {
|
|
let device = get_device();
|
|
let store = DeviceStore::new();
|
|
|
|
assert!(store.add(device.clone()));
|
|
assert!(!store.add(device.clone()));
|
|
|
|
let loaded_device = store.get(device.user_id(), device.device_id()).unwrap();
|
|
|
|
assert_eq!(device, loaded_device);
|
|
|
|
let user_devices = store.user_devices(device.user_id());
|
|
|
|
assert_eq!(&**user_devices.keys().next().unwrap(), device.device_id());
|
|
assert_eq!(user_devices.values().next().unwrap(), &device);
|
|
|
|
let loaded_device = user_devices.get(device.device_id()).unwrap();
|
|
|
|
assert_eq!(&device, loaded_device);
|
|
|
|
store.remove(device.user_id(), device.device_id());
|
|
|
|
let loaded_device = store.get(device.user_id(), device.device_id());
|
|
assert!(loaded_device.is_none());
|
|
}
|
|
}
|