@@ -0,0 +1,539 @@ |
| | 1 | +//! Encrypted chunk storage for ZephyrFS |
| | 2 | +//! |
| | 3 | +//! Provides storage layer functionality for encrypted chunks while maintaining |
| | 4 | +//! zero-knowledge security. Storage nodes never see plaintext data. |
| | 5 | + |
| | 6 | +use anyhow::{Context, Result}; |
| | 7 | +use rocksdb::{DB, Options, WriteBatch}; |
| | 8 | +use serde::{Deserialize, Serialize}; |
| | 9 | +use sha2::{Digest, Sha256}; |
| | 10 | +use std::collections::HashMap; |
| | 11 | +use std::path::Path; |
| | 12 | +use std::sync::Arc; |
| | 13 | +use tokio::sync::RwLock; |
| | 14 | +use tracing::{debug, info, warn}; |
| | 15 | + |
| | 16 | +use crate::crypto::{EncryptedData, ContentId}; |
| | 17 | + |
| | 18 | +/// Metadata for encrypted chunks stored in the system |
| | 19 | +/// |
| | 20 | +/// Zero-knowledge: Only stores encrypted data and hashes, no plaintext metadata |
| | 21 | +#[derive(Debug, Clone, Serialize, Deserialize)] |
| | 22 | +pub struct EncryptedChunkMetadata { |
| | 23 | + /// Content hash of the encrypted chunk (for deduplication) |
| | 24 | + pub encrypted_hash: String, |
| | 25 | + |
| | 26 | + /// Size of the encrypted chunk in bytes |
| | 27 | + pub encrypted_size: u64, |
| | 28 | + |
| | 29 | + /// Timestamp when chunk was stored |
| | 30 | + pub stored_at: u64, |
| | 31 | + |
| | 32 | + /// Reference count (how many encrypted files reference this chunk) |
| | 33 | + pub ref_count: u32, |
| | 34 | + |
| | 35 | + /// Verification checksum for integrity (of encrypted data) |
| | 36 | + pub checksum: String, |
| | 37 | + |
| | 38 | + /// Content addressing hash (encrypted, for lookup) |
| | 39 | + pub content_id: Option<String>, |
| | 40 | + |
| | 41 | + /// Encryption nonce (safe to store) |
| | 42 | + pub nonce: [u8; 12], |
| | 43 | + |
| | 44 | + /// Additional authenticated data (encrypted metadata) |
| | 45 | + pub aad: Vec<u8>, |
| | 46 | + |
| | 47 | + /// Key derivation path (safe to store, no keys) |
| | 48 | + pub key_path: Vec<u32>, |
| | 49 | +} |
| | 50 | + |
| | 51 | +/// Enhanced file metadata that includes encryption information |
| | 52 | +/// |
| | 53 | +/// Zero-knowledge: Stores encrypted metadata and access patterns |
| | 54 | +#[derive(Debug, Clone, Serialize, Deserialize)] |
| | 55 | +pub struct EncryptedFileMetadata { |
| | 56 | + /// Original filename (encrypted) |
| | 57 | + pub encrypted_name: Vec<u8>, |
| | 58 | + |
| | 59 | + /// Encrypted file size info |
| | 60 | + pub encrypted_size_info: Vec<u8>, |
| | 61 | + |
| | 62 | + /// File hash of encrypted data |
| | 63 | + pub encrypted_file_hash: String, |
| | 64 | + |
| | 65 | + /// List of encrypted chunk IDs |
| | 66 | + pub encrypted_chunk_ids: Vec<String>, |
| | 67 | + |
| | 68 | + /// Timestamp (can be plaintext for sorting) |
| | 69 | + pub created_at: u64, |
| | 70 | + pub modified_at: u64, |
| | 71 | + |
| | 72 | + /// Encryption metadata |
| | 73 | + pub encryption_metadata: EncryptionMetadata, |
| | 74 | + |
| | 75 | + /// Access control capabilities (encrypted) |
| | 76 | + pub capabilities: Vec<u8>, |
| | 77 | +} |
| | 78 | + |
| | 79 | +/// Encryption-specific metadata stored with files |
| | 80 | +/// |
| | 81 | +/// Zero-knowledge: No sensitive key material stored |
| | 82 | +#[derive(Debug, Clone, Serialize, Deserialize)] |
| | 83 | +pub struct EncryptionMetadata { |
| | 84 | + /// Encryption algorithm version |
| | 85 | + pub version: u32, |
| | 86 | + |
| | 87 | + /// Number of encrypted segments |
| | 88 | + pub segment_count: u32, |
| | 89 | + |
| | 90 | + /// Chunk size used for encryption (MB) |
| | 91 | + pub chunk_size_mb: u32, |
| | 92 | + |
| | 93 | + /// Content addressing algorithm |
| | 94 | + pub content_hash_algorithm: String, |
| | 95 | + |
| | 96 | + /// Verification hash algorithm |
| | 97 | + pub verification_hash_algorithm: String, |
| | 98 | + |
| | 99 | + /// Master nonce for file-level operations |
| | 100 | + pub master_nonce: [u8; 12], |
| | 101 | + |
| | 102 | + /// Encrypted content verification data |
| | 103 | + pub encrypted_content_verification: Vec<u8>, |
| | 104 | +} |
| | 105 | + |
| | 106 | +/// Capability token for secure file access |
| | 107 | +/// |
| | 108 | +/// Zero-knowledge: Contains encrypted access permissions and keys |
| | 109 | +#[derive(Debug, Clone, Serialize, Deserialize)] |
| | 110 | +pub struct FileCapability { |
| | 111 | + /// Unique capability ID |
| | 112 | + pub capability_id: String, |
| | 113 | + |
| | 114 | + /// File ID this capability grants access to |
| | 115 | + pub file_id: String, |
| | 116 | + |
| | 117 | + /// Encrypted access permissions (read, write, share, etc.) |
| | 118 | + pub encrypted_permissions: Vec<u8>, |
| | 119 | + |
| | 120 | + /// Encrypted key material for this capability |
| | 121 | + pub encrypted_key_material: Vec<u8>, |
| | 122 | + |
| | 123 | + /// Capability expiration timestamp (optional) |
| | 124 | + pub expires_at: Option<u64>, |
| | 125 | + |
| | 126 | + /// Created timestamp |
| | 127 | + pub created_at: u64, |
| | 128 | + |
| | 129 | + /// Capability signature (for verification) |
| | 130 | + pub signature: Vec<u8>, |
| | 131 | +} |
| | 132 | + |
| | 133 | +/// Enhanced chunk store that handles encrypted chunks |
| | 134 | +/// |
| | 135 | +/// Zero-knowledge: Never processes or sees plaintext data |
| | 136 | +pub struct EncryptedChunkStore { |
| | 137 | + /// Underlying chunk storage |
| | 138 | + db: Arc<DB>, |
| | 139 | + |
| | 140 | + /// Metadata cache for encrypted chunks |
| | 141 | + metadata_cache: Arc<RwLock<HashMap<String, EncryptedChunkMetadata>>>, |
| | 142 | + |
| | 143 | + /// File metadata storage |
| | 144 | + file_metadata_cache: Arc<RwLock<HashMap<String, EncryptedFileMetadata>>>, |
| | 145 | + |
| | 146 | + /// Capability storage |
| | 147 | + capability_cache: Arc<RwLock<HashMap<String, FileCapability>>>, |
| | 148 | + |
| | 149 | + /// Storage statistics |
| | 150 | + stats: Arc<RwLock<EncryptedStorageStats>>, |
| | 151 | +} |
| | 152 | + |
| | 153 | +#[derive(Debug, Default, Clone)] |
| | 154 | +pub struct EncryptedStorageStats { |
| | 155 | + pub total_encrypted_chunks: u64, |
| | 156 | + pub total_encrypted_size: u64, |
| | 157 | + pub total_encrypted_files: u64, |
| | 158 | + pub active_capabilities: u64, |
| | 159 | + pub cache_hits: u64, |
| | 160 | + pub cache_misses: u64, |
| | 161 | + pub deduplication_savings: u64, |
| | 162 | +} |
| | 163 | + |
| | 164 | +impl EncryptedChunkStore { |
| | 165 | + /// Create a new encrypted chunk store |
| | 166 | + /// |
| | 167 | + /// Zero-knowledge: Configures storage to handle only encrypted data |
| | 168 | + pub fn new<P: AsRef<Path>>(db_path: P) -> Result<Self> { |
| | 169 | + info!("Initializing EncryptedChunkStore for zero-knowledge storage"); |
| | 170 | + |
| | 171 | + let mut opts = Options::default(); |
| | 172 | + opts.create_if_missing(true); |
| | 173 | + opts.set_paranoid_checks(true); |
| | 174 | + opts.set_use_fsync(true); |
| | 175 | + |
| | 176 | + let db = DB::open(&opts, db_path) |
| | 177 | + .context("Failed to open encrypted chunk database")?; |
| | 178 | + |
| | 179 | + let store = Self { |
| | 180 | + db: Arc::new(db), |
| | 181 | + metadata_cache: Arc::new(RwLock::new(HashMap::new())), |
| | 182 | + file_metadata_cache: Arc::new(RwLock::new(HashMap::new())), |
| | 183 | + capability_cache: Arc::new(RwLock::new(HashMap::new())), |
| | 184 | + stats: Arc::new(RwLock::new(EncryptedStorageStats::default())), |
| | 185 | + }; |
| | 186 | + |
| | 187 | + store.load_stats_from_db()?; |
| | 188 | + info!("EncryptedChunkStore initialized with zero-knowledge architecture"); |
| | 189 | + Ok(store) |
| | 190 | + } |
| | 191 | + |
| | 192 | + /// Store an encrypted chunk with deduplication |
| | 193 | + /// |
| | 194 | + /// Zero-knowledge: Only handles encrypted data, maintains content-based deduplication |
| | 195 | + pub async fn store_encrypted_chunk(&self, chunk_id: &str, encrypted_data: &EncryptedData) -> Result<String> { |
| | 196 | + debug!("Storing encrypted chunk: {} ({} bytes)", chunk_id, encrypted_data.ciphertext.len()); |
| | 197 | + |
| | 198 | + // Calculate hash of encrypted content for deduplication |
| | 199 | + let mut hasher = Sha256::new(); |
| | 200 | + hasher.update(&encrypted_data.ciphertext); |
| | 201 | + hasher.update(&encrypted_data.nonce); |
| | 202 | + hasher.update(&encrypted_data.aad); |
| | 203 | + let encrypted_hash = hex::encode(hasher.finalize()); |
| | 204 | + |
| | 205 | + // Check if this encrypted chunk already exists (deduplication) |
| | 206 | + if let Some(existing_metadata) = self.get_encrypted_chunk_metadata(&encrypted_hash).await? { |
| | 207 | + // Increment reference count |
| | 208 | + let mut metadata = existing_metadata; |
| | 209 | + metadata.ref_count += 1; |
| | 210 | + self.update_encrypted_chunk_metadata(&encrypted_hash, &metadata).await?; |
| | 211 | + |
| | 212 | + debug!("Deduplicated encrypted chunk: {} (ref_count: {})", encrypted_hash, metadata.ref_count); |
| | 213 | + return Ok(encrypted_hash); |
| | 214 | + } |
| | 215 | + |
| | 216 | + // Create checksum for integrity verification |
| | 217 | + let checksum = self.calculate_encrypted_checksum(&encrypted_data.ciphertext, &encrypted_hash); |
| | 218 | + |
| | 219 | + let metadata = EncryptedChunkMetadata { |
| | 220 | + encrypted_hash: encrypted_hash.clone(), |
| | 221 | + encrypted_size: encrypted_data.ciphertext.len() as u64, |
| | 222 | + stored_at: std::time::SystemTime::now() |
| | 223 | + .duration_since(std::time::UNIX_EPOCH)? |
| | 224 | + .as_secs(), |
| | 225 | + ref_count: 1, |
| | 226 | + checksum, |
| | 227 | + content_id: None, // Set by caller if needed |
| | 228 | + nonce: encrypted_data.nonce, |
| | 229 | + aad: encrypted_data.aad.clone(), |
| | 230 | + key_path: encrypted_data.key_path.clone(), |
| | 231 | + }; |
| | 232 | + |
| | 233 | + // Store encrypted chunk data and metadata atomically |
| | 234 | + let mut batch = WriteBatch::default(); |
| | 235 | + |
| | 236 | + // Store the encrypted ciphertext |
| | 237 | + let chunk_key = format!("chunk:{}", encrypted_hash); |
| | 238 | + batch.put(&chunk_key, &encrypted_data.ciphertext); |
| | 239 | + |
| | 240 | + // Store metadata |
| | 241 | + let metadata_key = format!("meta:{}", encrypted_hash); |
| | 242 | + let metadata_bytes = bincode::serialize(&metadata) |
| | 243 | + .context("Failed to serialize encrypted chunk metadata")?; |
| | 244 | + batch.put(&metadata_key, &metadata_bytes); |
| | 245 | + |
| | 246 | + self.db.write(batch) |
| | 247 | + .context("Failed to store encrypted chunk atomically")?; |
| | 248 | + |
| | 249 | + // Update cache and stats |
| | 250 | + { |
| | 251 | + let mut cache = self.metadata_cache.write().await; |
| | 252 | + cache.insert(encrypted_hash.clone(), metadata); |
| | 253 | + } |
| | 254 | + |
| | 255 | + { |
| | 256 | + let mut stats = self.stats.write().await; |
| | 257 | + stats.total_encrypted_chunks += 1; |
| | 258 | + stats.total_encrypted_size += encrypted_data.ciphertext.len() as u64; |
| | 259 | + } |
| | 260 | + |
| | 261 | + info!("Stored new encrypted chunk: {}", encrypted_hash); |
| | 262 | + Ok(encrypted_hash) |
| | 263 | + } |
| | 264 | + |
| | 265 | + /// Retrieve an encrypted chunk by hash |
| | 266 | + /// |
| | 267 | + /// Zero-knowledge: Returns encrypted data without any decryption |
| | 268 | + pub async fn retrieve_encrypted_chunk(&self, encrypted_hash: &str) -> Result<Option<EncryptedData>> { |
| | 269 | + debug!("Retrieving encrypted chunk: {}", encrypted_hash); |
| | 270 | + |
| | 271 | + // Get metadata first |
| | 272 | + let metadata = match self.get_encrypted_chunk_metadata(encrypted_hash).await? { |
| | 273 | + Some(meta) => meta, |
| | 274 | + None => { |
| | 275 | + debug!("Encrypted chunk not found: {}", encrypted_hash); |
| | 276 | + return Ok(None); |
| | 277 | + } |
| | 278 | + }; |
| | 279 | + |
| | 280 | + // Retrieve encrypted ciphertext |
| | 281 | + let chunk_key = format!("chunk:{}", encrypted_hash); |
| | 282 | + let ciphertext = match self.db.get(&chunk_key) |
| | 283 | + .context("Failed to read encrypted chunk from database")? { |
| | 284 | + Some(data) => data, |
| | 285 | + None => { |
| | 286 | + warn!("Encrypted chunk data missing for hash: {}", encrypted_hash); |
| | 287 | + return Ok(None); |
| | 288 | + } |
| | 289 | + }; |
| | 290 | + |
| | 291 | + // Verify integrity |
| | 292 | + let computed_checksum = self.calculate_encrypted_checksum(&ciphertext, encrypted_hash); |
| | 293 | + if computed_checksum != metadata.checksum { |
| | 294 | + return Err(anyhow::anyhow!( |
| | 295 | + "Encrypted chunk integrity verification failed for {}", encrypted_hash |
| | 296 | + )); |
| | 297 | + } |
| | 298 | + |
| | 299 | + // Reconstruct EncryptedData |
| | 300 | + let encrypted_data = EncryptedData { |
| | 301 | + segment_index: 0, // Will be set by caller |
| | 302 | + ciphertext, |
| | 303 | + nonce: metadata.nonce, |
| | 304 | + aad: metadata.aad, |
| | 305 | + key_path: metadata.key_path, |
| | 306 | + }; |
| | 307 | + |
| | 308 | + { |
| | 309 | + let mut stats = self.stats.write().await; |
| | 310 | + stats.cache_hits += 1; |
| | 311 | + } |
| | 312 | + |
| | 313 | + Ok(Some(encrypted_data)) |
| | 314 | + } |
| | 315 | + |
| | 316 | + /// Store encrypted file metadata |
| | 317 | + /// |
| | 318 | + /// Zero-knowledge: All sensitive metadata is encrypted |
| | 319 | + pub async fn store_encrypted_file_metadata(&self, file_id: &str, metadata: &EncryptedFileMetadata) -> Result<()> { |
| | 320 | + debug!("Storing encrypted file metadata: {}", file_id); |
| | 321 | + |
| | 322 | + let metadata_key = format!("file:{}", file_id); |
| | 323 | + let metadata_bytes = bincode::serialize(metadata) |
| | 324 | + .context("Failed to serialize encrypted file metadata")?; |
| | 325 | + |
| | 326 | + self.db.put(&metadata_key, &metadata_bytes) |
| | 327 | + .context("Failed to store encrypted file metadata")?; |
| | 328 | + |
| | 329 | + // Update cache |
| | 330 | + { |
| | 331 | + let mut cache = self.file_metadata_cache.write().await; |
| | 332 | + cache.insert(file_id.to_string(), metadata.clone()); |
| | 333 | + } |
| | 334 | + |
| | 335 | + { |
| | 336 | + let mut stats = self.stats.write().await; |
| | 337 | + stats.total_encrypted_files += 1; |
| | 338 | + } |
| | 339 | + |
| | 340 | + Ok(()) |
| | 341 | + } |
| | 342 | + |
| | 343 | + /// Retrieve encrypted file metadata |
| | 344 | + /// |
| | 345 | + /// Zero-knowledge: Returns encrypted metadata without decryption |
| | 346 | + pub async fn get_encrypted_file_metadata(&self, file_id: &str) -> Result<Option<EncryptedFileMetadata>> { |
| | 347 | + // Check cache first |
| | 348 | + { |
| | 349 | + let cache = self.file_metadata_cache.read().await; |
| | 350 | + if let Some(metadata) = cache.get(file_id) { |
| | 351 | + return Ok(Some(metadata.clone())); |
| | 352 | + } |
| | 353 | + } |
| | 354 | + |
| | 355 | + let metadata_key = format!("file:{}", file_id); |
| | 356 | + let metadata_bytes = match self.db.get(&metadata_key) |
| | 357 | + .context("Failed to read encrypted file metadata")? { |
| | 358 | + Some(data) => data, |
| | 359 | + None => return Ok(None), |
| | 360 | + }; |
| | 361 | + |
| | 362 | + let metadata: EncryptedFileMetadata = bincode::deserialize(&metadata_bytes) |
| | 363 | + .context("Failed to deserialize encrypted file metadata")?; |
| | 364 | + |
| | 365 | + // Update cache |
| | 366 | + { |
| | 367 | + let mut cache = self.file_metadata_cache.write().await; |
| | 368 | + cache.insert(file_id.to_string(), metadata.clone()); |
| | 369 | + } |
| | 370 | + |
| | 371 | + Ok(Some(metadata)) |
| | 372 | + } |
| | 373 | + |
| | 374 | + /// Store a file capability for secure access control |
| | 375 | + /// |
| | 376 | + /// Zero-knowledge: Capability contains encrypted permissions and keys |
| | 377 | + pub async fn store_capability(&self, capability: &FileCapability) -> Result<()> { |
| | 378 | + debug!("Storing file capability: {}", capability.capability_id); |
| | 379 | + |
| | 380 | + let cap_key = format!("cap:{}", capability.capability_id); |
| | 381 | + let cap_bytes = bincode::serialize(capability) |
| | 382 | + .context("Failed to serialize capability")?; |
| | 383 | + |
| | 384 | + self.db.put(&cap_key, &cap_bytes) |
| | 385 | + .context("Failed to store capability")?; |
| | 386 | + |
| | 387 | + // Update cache and stats |
| | 388 | + { |
| | 389 | + let mut cache = self.capability_cache.write().await; |
| | 390 | + cache.insert(capability.capability_id.clone(), capability.clone()); |
| | 391 | + } |
| | 392 | + |
| | 393 | + { |
| | 394 | + let mut stats = self.stats.write().await; |
| | 395 | + stats.active_capabilities += 1; |
| | 396 | + } |
| | 397 | + |
| | 398 | + Ok(()) |
| | 399 | + } |
| | 400 | + |
| | 401 | + /// Retrieve a file capability |
| | 402 | + pub async fn get_capability(&self, capability_id: &str) -> Result<Option<FileCapability>> { |
| | 403 | + // Check cache first |
| | 404 | + { |
| | 405 | + let cache = self.capability_cache.read().await; |
| | 406 | + if let Some(capability) = cache.get(capability_id) { |
| | 407 | + return Ok(Some(capability.clone())); |
| | 408 | + } |
| | 409 | + } |
| | 410 | + |
| | 411 | + let cap_key = format!("cap:{}", capability_id); |
| | 412 | + let cap_bytes = match self.db.get(&cap_key) |
| | 413 | + .context("Failed to read capability")? { |
| | 414 | + Some(data) => data, |
| | 415 | + None => return Ok(None), |
| | 416 | + }; |
| | 417 | + |
| | 418 | + let capability: FileCapability = bincode::deserialize(&cap_bytes) |
| | 419 | + .context("Failed to deserialize capability")?; |
| | 420 | + |
| | 421 | + // Check expiration |
| | 422 | + if let Some(expires_at) = capability.expires_at { |
| | 423 | + let now = std::time::SystemTime::now() |
| | 424 | + .duration_since(std::time::UNIX_EPOCH)? |
| | 425 | + .as_secs(); |
| | 426 | + if now > expires_at { |
| | 427 | + debug!("Capability expired: {}", capability_id); |
| | 428 | + return Ok(None); |
| | 429 | + } |
| | 430 | + } |
| | 431 | + |
| | 432 | + // Update cache |
| | 433 | + { |
| | 434 | + let mut cache = self.capability_cache.write().await; |
| | 435 | + cache.insert(capability_id.to_string(), capability.clone()); |
| | 436 | + } |
| | 437 | + |
| | 438 | + Ok(Some(capability)) |
| | 439 | + } |
| | 440 | + |
| | 441 | + /// Get storage statistics |
| | 442 | + pub async fn get_encrypted_stats(&self) -> EncryptedStorageStats { |
| | 443 | + let stats = self.stats.read().await; |
| | 444 | + (*stats).clone() |
| | 445 | + } |
| | 446 | + |
| | 447 | + /// Helper methods |
| | 448 | + async fn get_encrypted_chunk_metadata(&self, encrypted_hash: &str) -> Result<Option<EncryptedChunkMetadata>> { |
| | 449 | + // Check cache first |
| | 450 | + { |
| | 451 | + let cache = self.metadata_cache.read().await; |
| | 452 | + if let Some(metadata) = cache.get(encrypted_hash) { |
| | 453 | + return Ok(Some(metadata.clone())); |
| | 454 | + } |
| | 455 | + } |
| | 456 | + |
| | 457 | + let metadata_key = format!("meta:{}", encrypted_hash); |
| | 458 | + let metadata_bytes = match self.db.get(&metadata_key)? { |
| | 459 | + Some(data) => data, |
| | 460 | + None => return Ok(None), |
| | 461 | + }; |
| | 462 | + |
| | 463 | + let metadata: EncryptedChunkMetadata = bincode::deserialize(&metadata_bytes) |
| | 464 | + .context("Failed to deserialize encrypted chunk metadata")?; |
| | 465 | + |
| | 466 | + Ok(Some(metadata)) |
| | 467 | + } |
| | 468 | + |
| | 469 | + async fn update_encrypted_chunk_metadata(&self, encrypted_hash: &str, metadata: &EncryptedChunkMetadata) -> Result<()> { |
| | 470 | + let metadata_key = format!("meta:{}", encrypted_hash); |
| | 471 | + let metadata_bytes = bincode::serialize(metadata)?; |
| | 472 | + |
| | 473 | + self.db.put(&metadata_key, &metadata_bytes)?; |
| | 474 | + |
| | 475 | + // Update cache |
| | 476 | + { |
| | 477 | + let mut cache = self.metadata_cache.write().await; |
| | 478 | + cache.insert(encrypted_hash.to_string(), metadata.clone()); |
| | 479 | + } |
| | 480 | + |
| | 481 | + Ok(()) |
| | 482 | + } |
| | 483 | + |
| | 484 | + fn calculate_encrypted_checksum(&self, ciphertext: &[u8], hash: &str) -> String { |
| | 485 | + let mut hasher = Sha256::new(); |
| | 486 | + hasher.update(ciphertext); |
| | 487 | + hasher.update(hash.as_bytes()); |
| | 488 | + hasher.update(b"zephyrfs-encrypted-chunk-v1"); |
| | 489 | + hex::encode(hasher.finalize()) |
| | 490 | + } |
| | 491 | + |
| | 492 | + fn load_stats_from_db(&self) -> Result<()> { |
| | 493 | + // Implementation would scan database to calculate stats |
| | 494 | + // For now, we'll initialize with defaults |
| | 495 | + Ok(()) |
| | 496 | + } |
| | 497 | +} |
| | 498 | + |
| | 499 | +#[cfg(test)] |
| | 500 | +mod tests { |
| | 501 | + use super::*; |
| | 502 | + use tempfile::tempdir; |
| | 503 | + |
| | 504 | + #[tokio::test] |
| | 505 | + async fn test_encrypted_chunk_store_creation() { |
| | 506 | + let temp_dir = tempdir().unwrap(); |
| | 507 | + let store = EncryptedChunkStore::new(temp_dir.path()).await.unwrap(); |
| | 508 | + let stats = store.get_encrypted_stats().await; |
| | 509 | + |
| | 510 | + assert_eq!(stats.total_encrypted_chunks, 0); |
| | 511 | + assert_eq!(stats.total_encrypted_files, 0); |
| | 512 | + assert_eq!(stats.active_capabilities, 0); |
| | 513 | + } |
| | 514 | + |
| | 515 | + #[tokio::test] |
| | 516 | + async fn test_encrypted_chunk_deduplication() { |
| | 517 | + let temp_dir = tempdir().unwrap(); |
| | 518 | + let store = EncryptedChunkStore::new(temp_dir.path()).await.unwrap(); |
| | 519 | + |
| | 520 | + let encrypted_data = EncryptedData { |
| | 521 | + segment_index: 0, |
| | 522 | + ciphertext: vec![1, 2, 3, 4, 5], |
| | 523 | + nonce: [0u8; 12], |
| | 524 | + aad: vec![], |
| | 525 | + key_path: vec![0, 1], |
| | 526 | + }; |
| | 527 | + |
| | 528 | + // Store same encrypted chunk twice |
| | 529 | + let hash1 = store.store_encrypted_chunk("chunk1", &encrypted_data).await.unwrap(); |
| | 530 | + let hash2 = store.store_encrypted_chunk("chunk2", &encrypted_data).await.unwrap(); |
| | 531 | + |
| | 532 | + // Should be deduplicated (same hash) |
| | 533 | + assert_eq!(hash1, hash2); |
| | 534 | + |
| | 535 | + // Should have reference count of 2 |
| | 536 | + let metadata = store.get_encrypted_chunk_metadata(&hash1).await.unwrap().unwrap(); |
| | 537 | + assert_eq!(metadata.ref_count, 2); |
| | 538 | + } |
| | 539 | +} |