mirror of
https://github.com/rustfs/rustfs.git
synced 2026-01-16 17:20:33 +00:00
* Initial plan * feat: add concurrency-aware buffer sizing and hot object caching for GetObject - Implement adaptive buffer sizing based on concurrent request load - Add per-request tracking with automatic cleanup using RAII guards - Implement hot object cache (LRU) for frequently accessed small files (<= 10MB) - Add disk I/O semaphore to prevent saturation under extreme load - Integrate concurrency module into GetObject implementation - Buffer sizes now adapt: low concurrency uses large buffers for throughput, high concurrency uses smaller buffers for fairness and memory efficiency - Add comprehensive metrics collection for monitoring performance Co-authored-by: houseme <4829346+houseme@users.noreply.github.com> * docs: add comprehensive documentation and tests for concurrent GetObject optimization - Add detailed technical documentation explaining the solution - Document root cause analysis and solution architecture - Include performance expectations and testing recommendations - Add integration tests for concurrency tracking and buffer sizing - Add cache behavior tests - Include benchmark tests for concurrent request handling Co-authored-by: houseme <4829346+houseme@users.noreply.github.com> * fix: address code review issues in concurrency module - Fix race condition in cache size tracking by using consistent atomic operations within lock - Correct buffer sizing logic: 1-2 requests use 100%, 3-4 use 75%, 5-8 use 50%, >8 use 40% - Improve error message for semaphore acquire failure - Document limitation of streaming cache implementation (not yet implemented) - Add TODO for proper streaming cache with suggested approaches - Update tests to match corrected buffer sizing thresholds Co-authored-by: houseme <4829346+houseme@users.noreply.github.com> * docs: add comprehensive implementation summary for concurrent GetObject optimization - Executive summary of problem and solution - Detailed architecture documentation for each optimization - Integration points and code examples - Complete testing strategy and performance validation steps - Monitoring and observability guidelines with Prometheus queries - Deployment considerations and rollout strategy - Future enhancement roadmap - Success criteria and quantitative metrics Co-authored-by: houseme <4829346+houseme@users.noreply.github.com> * fix * fix * refactor: optimize cache with lru 0.16.2 read-first pattern and add advanced features - Implement optimized read-first cache access using peek() to reduce write lock contention - Add batch cache operations: get_cached_batch() for efficient multi-object retrieval - Add cache utility methods: is_cached(), remove_cached(), get_hot_keys() - Implement warm_cache() for pre-populating cache on startup - Add get_advanced_buffer_size() with file size and sequential read optimization - Enhance test suite with 8 new comprehensive tests covering: - Batch operations and cache warming - Hot keys tracking and analysis - Cache removal and LRU behavior verification - Concurrent cache access performance - Advanced buffer sizing strategies - Improve documentation and code comments in English throughout Co-authored-by: houseme <4829346+houseme@users.noreply.github.com> * docs: add final comprehensive optimization summary - Complete technical documentation of all optimizations - Detailed performance analysis and metrics - Production deployment guide with examples - Comprehensive API reference and usage patterns - Migration guide and future enhancement roadmap - All documentation in professional English Co-authored-by: houseme <4829346+houseme@users.noreply.github.com> * fix * fix * add moka crate for cache * feat: replace LRU with Moka cache and add comprehensive metrics - Replace lru crate with moka 0.12.11 for superior concurrent performance - Implement lock-free cache with automatic TTL/TTI expiration - Add size-based eviction using Moka's weigher function - Integrate comprehensive metrics collection throughout GetObject flow: * Cache hit/miss tracking with per-key access counts * Request concurrency gauges * Disk permit wait time histograms * Total request duration tracking * Response size and buffer size histograms - Deep integration with ecfs.rs GetObject operation - Add hit rate calculation method - Enhanced CacheStats with hit/miss counters - Lock-free concurrent reads for better scalability Moka advantages over LRU: - True lock-free concurrent access - Built-in TTL and TTI support - Automatic size-based eviction - Better performance under high concurrency - Native async support Co-authored-by: houseme <4829346+houseme@users.noreply.github.com> * docs: add comprehensive Moka cache migration and metrics documentation - Complete technical documentation of LRU to Moka migration - Detailed performance comparison and benchmarks - Comprehensive metrics catalog with 15+ Prometheus metrics - Prometheus query examples for monitoring - Dashboard and alerting recommendations - Migration guide with code examples - Troubleshooting guide for common issues - Future enhancement roadmap Co-authored-by: houseme <4829346+houseme@users.noreply.github.com> * fix * fix * refactor: update tests for Moka cache implementation - Completely refactor test suite to align with Moka-based concurrency.rs - Add Clone derive to ConcurrencyManager for test convenience - Update all tests to handle Moka's async behavior with proper delays - Add new tests: * test_cache_hit_rate - validate hit rate calculation * test_ttl_expiration - verify TTL configuration * test_is_cached_no_side_effects - ensure contains doesn't affect LRU * bench_concurrent_cache_performance - benchmark concurrent access - Updated existing tests: * test_moka_cache_operations - renamed and updated for Moka API * test_moka_cache_eviction - validate automatic eviction * test_hot_keys_tracking - improved assertions for sorted results * test_concurrent_cache_access - validate lock-free performance - All tests now include appropriate sleep delays for Moka's async processing - Enhanced documentation and assertions for better test clarity - Total: 18 comprehensive integration tests Co-authored-by: houseme <4829346+houseme@users.noreply.github.com> * docs: add comprehensive Moka test suite documentation - Complete test suite documentation for all 18 tests - Detailed test patterns and best practices for Moka - Debugging guide for common test failures - Performance expectations and CI integration - 100% API coverage validation - Future enhancement roadmap Co-authored-by: houseme <4829346+houseme@users.noreply.github.com> * fix * fmt * docs: enhance code comments and test documentation throughout - Significantly improve module-level documentation in concurrency.rs - Add detailed explanations of performance characteristics and expected improvements - Document the solution to issue #911 with before/after metrics - Enhance constant documentation with purpose and impact explanations - Add comprehensive RAII guard documentation with thread safety notes - Improve test documentation with detailed test strategy and rationale - Add Moka-specific test patterns and timing guidance - Include performance expectations for each test category - Document why each test matters for solving the original issue - All documentation written in professional English - Follow Rust documentation best practices with examples Co-authored-by: houseme <4829346+houseme@users.noreply.github.com> * remove lru crate * upgrade version * fix: resolve test errors by correcting module structure and test assertions - Fix test import paths to use crate:: instead of rustfs:: (binary-only crate) - Keep test file in src/storage/ instead of tests/ (no lib.rs exists) - Add #[cfg(test)] guard to mod declaration in storage/mod.rs - Fix Arc type annotations for Moka's ConcurrencyManager in concurrent tests - Correct test_buffer_size_bounds assertions to match actual implementation: * Minimum buffer is 32KB for files <100KB, 64KB otherwise * Maximum buffer respects base_buffer_size when concurrency is low * Buffer sizing doesn't cap at file size, only at min/max constraints - All 17 integration tests now pass successfully Co-authored-by: houseme <4829346+houseme@users.noreply.github.com> * fix: modify `TimeoutLayer::new` to `TimeoutLayer::with_status_code` and improve docker health check * fix * feat: implement cache writeback for small objects in GetObject - Add cache writeback logic for objects meeting caching criteria: * No range/part request (full object retrieval) * Object size known and <= 10MB (max_object_size threshold) * Not encrypted (SSE-C or managed encryption) - Read eligible objects into memory and cache via background task - Serve response from in-memory data for immediate client response - Add metrics counter for cache writeback operations - Add 3 new tests for cache writeback functionality: * test_cache_writeback_flow - validates round-trip caching * test_cache_writeback_size_limit - ensures large objects aren't cached * test_cache_writeback_concurrent - validates thread-safe concurrent writes - Update test suite documentation (now 20 comprehensive tests) Co-authored-by: houseme <4829346+houseme@users.noreply.github.com> * improve code for const * cargo clippy * feat: add cache enable/disable configuration via environment variable - Add is_cache_enabled() method to ConcurrencyManager - Read RUSTFS_OBJECT_CACHE_ENABLE env var (default: false) at startup - Update ecfs.rs to check is_cache_enabled() before cache lookup and writeback - Cache lookup and writeback now respect the enable flag - Add test_cache_enable_configuration test - Constants already exist in rustfs_config: * ENV_OBJECT_CACHE_ENABLE = "RUSTFS_OBJECT_CACHE_ENABLE" * DEFAULT_OBJECT_CACHE_ENABLE = false - Total: 21 comprehensive tests passing Co-authored-by: houseme <4829346+houseme@users.noreply.github.com> * fix * fmt * fix * fix * feat: implement comprehensive CachedGetObject response cache with metadata - Add CachedGetObject struct with full response metadata fields: * body, content_length, content_type, e_tag, last_modified * expires, cache_control, content_disposition, content_encoding * storage_class, version_id, delete_marker, tag_count, etc. - Add dual cache architecture in HotObjectCache: * Legacy simple byte cache for backward compatibility * New response cache for complete GetObject responses - Add ConcurrencyManager methods for response caching: * get_cached_object() - retrieve cached response with metadata * put_cached_object() - store complete response * invalidate_cache() - invalidate on write operations * invalidate_cache_versioned() - invalidate both version and latest * make_cache_key() - generate cache keys with version support * max_object_size() - get cache threshold - Add builder pattern for CachedGetObject construction - Add 6 new tests for response cache functionality (27 total): * test_cached_get_object_basic - basic operations * test_cached_get_object_versioned - version key handling * test_cache_invalidation - write operation invalidation * test_cache_invalidation_versioned - versioned invalidation * test_cached_get_object_size_limit - size enforcement * test_max_object_size - threshold accessor All 27 tests pass successfully. Co-authored-by: houseme <4829346+houseme@users.noreply.github.com> * feat: integrate CachedGetObject cache in ecfs.rs with full metadata and cache invalidation Integration of CachedGetObject response cache in ecfs.rs: 1. get_object: Cache lookup uses get_cached_object() with full metadata - Returns complete response with e_tag, last_modified, content_type, etc. - Parses last_modified from RFC3339 string - Supports versioned cache keys via make_cache_key() 2. get_object: Cache writeback uses put_cached_object() with metadata - Stores content_type, e_tag, last_modified in CachedGetObject - Background writeback via tokio::spawn() 3. Cache invalidation added to write operations: - put_object: invalidate_cache_versioned() after store.put_object() - put_object_extract: invalidate_cache_versioned() after each file extraction - copy_object: invalidate_cache_versioned() after store.copy_object() - delete_object: invalidate_cache_versioned() after store.delete_object() - delete_objects: invalidate_cache_versioned() for each deleted object - complete_multipart_upload: invalidate_cache_versioned() after completion 4. Fixed test_adaptive_buffer_sizing to be more robust with parallel tests All 27 tests pass. Co-authored-by: houseme <4829346+houseme@users.noreply.github.com> * fix: add error logging for time parsing in cache lookup and writeback - Add warning log when RFC3339 parsing fails in cache lookup - Add warning log when time formatting fails in cache writeback - Improves debugging for cache-related issues Co-authored-by: houseme <4829346+houseme@users.noreply.github.com> * fix * fix * upgrade version * fmt * add http: 4138 and add max_object_size key * fix * fix * fix * modify metrics key * add * upgrade crates version and improve docker observability * feat: implement adaptive I/O strategy based on disk permit wait time Implements adaptive I/O strategy for GetObject operations based on disk permit wait times, as requested in issue #911. Key changes: 1. New types in concurrency.rs: - IoLoadLevel enum: Low (<10ms), Medium (10-50ms), High (50-200ms), Critical (>200ms) - IoStrategy struct: contains optimized I/O parameters based on load level - IoLoadMetrics: rolling window metrics for load tracking 2. New ConcurrencyManager methods: - calculate_io_strategy(): calculates adaptive strategy from permit wait duration - record_permit_wait(): records observation for load tracking - smoothed_load_level(): returns averaged load level for stability - io_load_stats(): returns (avg_wait, p95_wait, max_wait, count) for monitoring - adaptive_buffer_size(): convenience method for buffer sizing 3. Integration in ecfs.rs get_object: - Calculate IoStrategy after acquiring disk permit - Use strategy buffer_size instead of static sizing - Consider strategy.cache_writeback_enabled in cache decision - Record new metrics: io.load.level gauge, io.buffer.multiplier gauge, io.strategy.selected counter by load level 4. New tests (4 added, 31 total): - test_io_load_level_classification: validates load level thresholds - test_io_strategy_buffer_sizing: validates buffer multipliers - test_calculate_io_strategy: validates manager strategy calculation - test_io_load_stats: validates stats tracking All 31 tests pass. Clippy clean. Formatted. Co-authored-by: houseme <4829346+houseme@users.noreply.github.com> * fix * fix * docs: add comprehensive architecture design and fix cache hit event notification 1. Added CONCURRENCY_ARCHITECTURE.md with complete design documentation: - Executive summary and problem statement - Architecture overview with request flow diagram - Detailed module analysis for concurrency.rs and ecfs.rs - Critical analysis of helper.complete() for cache hits - Adaptive I/O strategy design with algorithm - Cache architecture with CachedGetObject structure - Metrics and monitoring with Prometheus queries - Performance characteristics and future enhancements 2. Fixed critical issue: Cache hit path now calls helper.complete() - S3 bucket notifications (s3:GetObject events) now trigger for cache hits - Event-driven workflows (Lambda, SNS) work correctly for all object access - Maintains audit trail for both cache hits and misses All 31 tests pass. Co-authored-by: houseme <4829346+houseme@users.noreply.github.com> * fix: set object info and version_id on helper before complete() for cache hits When serving from cache, properly configure the OperationHelper before calling complete() to ensure S3 bucket notifications include complete object metadata: 1. Build ObjectInfo from cached metadata: - bucket, name, size, actual_size - etag, mod_time, version_id, delete_marker - storage_class, content_type, content_encoding - user_metadata (user_defined) 2. Set helper.object(event_info).version_id(version_id_str) before complete() 3. Updated CONCURRENCY_ARCHITECTURE.md with: - Complete code example for cache hit event notification - Explanation of why ObjectInfo is required - Documentation of version_id handling This ensures: - Lambda triggers receive proper object metadata for cache hits - SNS/SQS notifications include complete information - Audit logs contain accurate object details - Version-specific event routing works correctly All 31 tests pass. Co-authored-by: houseme <4829346+houseme@users.noreply.github.com> * fix * improve code * fmt --------- Co-authored-by: copilot-swe-agent[bot] <198982749+Copilot@users.noreply.github.com> Co-authored-by: houseme <4829346+houseme@users.noreply.github.com> Co-authored-by: houseme <housemecn@gmail.com>
320 lines
10 KiB
Markdown
320 lines
10 KiB
Markdown
# Concurrent GetObject Performance Optimization
|
|
|
|
## Problem Statement
|
|
|
|
When multiple concurrent GetObject requests are made to RustFS, performance degrades exponentially:
|
|
|
|
| Concurrency Level | Single Request Latency | Performance Impact |
|
|
|------------------|----------------------|-------------------|
|
|
| 1 request | 59ms | Baseline |
|
|
| 2 requests | 110ms | 1.9x slower |
|
|
| 4 requests | 200ms | 3.4x slower |
|
|
|
|
## Root Cause Analysis
|
|
|
|
The performance degradation was caused by several factors:
|
|
|
|
1. **Fixed Buffer Sizing**: Using `DEFAULT_READ_BUFFER_SIZE` (1MB) for all requests, regardless of concurrent load
|
|
- High memory contention under concurrent load
|
|
- Inefficient cache utilization
|
|
- CPU context switching overhead
|
|
|
|
2. **No Concurrency Control**: Unlimited concurrent disk reads causing I/O saturation
|
|
- Disk I/O queue depth exceeded optimal levels
|
|
- Increased seek times on traditional disks
|
|
- Resource contention between requests
|
|
|
|
3. **Lack of Caching**: Repeated reads of the same objects
|
|
- No reuse of frequently accessed data
|
|
- Unnecessary disk I/O for hot objects
|
|
|
|
## Solution Architecture
|
|
|
|
### 1. Concurrency-Aware Adaptive Buffer Sizing
|
|
|
|
The system now dynamically adjusts buffer sizes based on the current number of concurrent GetObject requests:
|
|
|
|
```rust
|
|
let optimal_buffer_size = get_concurrency_aware_buffer_size(file_size, base_buffer_size);
|
|
```
|
|
|
|
#### Buffer Sizing Strategy
|
|
|
|
| Concurrent Requests | Buffer Size Multiplier | Typical Buffer | Rationale |
|
|
|--------------------|----------------------|----------------|-----------|
|
|
| 1-2 (Low) | 1.0x (100%) | 512KB-1MB | Maximize throughput with large buffers |
|
|
| 3-4 (Medium) | 0.75x (75%) | 256KB-512KB | Balance throughput and fairness |
|
|
| 5-8 (High) | 0.5x (50%) | 128KB-256KB | Improve fairness, reduce memory pressure |
|
|
| 9+ (Very High) | 0.4x (40%) | 64KB-128KB | Ensure fair scheduling, minimize memory |
|
|
|
|
#### Benefits
|
|
- **Reduced memory pressure**: Smaller buffers under high concurrency prevent memory exhaustion
|
|
- **Better cache utilization**: More requests fit in CPU cache with smaller buffers
|
|
- **Improved fairness**: Prevents large requests from starving smaller ones
|
|
- **Adaptive performance**: Automatically tunes for different workload patterns
|
|
|
|
### 2. Hot Object Caching (LRU)
|
|
|
|
Implemented an intelligent LRU cache for frequently accessed small objects:
|
|
|
|
```rust
|
|
pub struct HotObjectCache {
|
|
max_object_size: usize, // Default: 10MB
|
|
max_cache_size: usize, // Default: 100MB
|
|
cache: RwLock<lru::LruCache<String, Arc<CachedObject>>>,
|
|
}
|
|
```
|
|
|
|
#### Caching Policy
|
|
- **Eligible objects**: Size ≤ 10MB, complete object reads (no ranges)
|
|
- **Eviction**: LRU (Least Recently Used)
|
|
- **Capacity**: Up to 1000 objects, 100MB total
|
|
- **Exclusions**: Encrypted objects, partial reads, multipart
|
|
|
|
#### Benefits
|
|
- **Reduced disk I/O**: Cache hits eliminate disk reads entirely
|
|
- **Lower latency**: Memory access is 100-1000x faster than disk
|
|
- **Higher throughput**: Free up disk bandwidth for cache misses
|
|
- **Better scalability**: Cache hit ratio improves with concurrent load
|
|
|
|
### 3. Disk I/O Concurrency Control
|
|
|
|
Added a semaphore to limit maximum concurrent disk reads:
|
|
|
|
```rust
|
|
disk_read_semaphore: Arc<Semaphore> // Default: 64 permits
|
|
```
|
|
|
|
#### Benefits
|
|
- **Prevents I/O saturation**: Limits queue depth to optimal levels
|
|
- **Predictable latency**: Avoids exponential latency increase
|
|
- **Protects disk health**: Reduces excessive seek operations
|
|
- **Graceful degradation**: Queues requests rather than thrashing
|
|
|
|
### 4. Request Tracking and Monitoring
|
|
|
|
Implemented RAII-based request tracking with automatic cleanup:
|
|
|
|
```rust
|
|
pub struct GetObjectGuard {
|
|
start_time: Instant,
|
|
}
|
|
|
|
impl Drop for GetObjectGuard {
|
|
fn drop(&mut self) {
|
|
ACTIVE_GET_REQUESTS.fetch_sub(1, Ordering::Relaxed);
|
|
// Record metrics
|
|
}
|
|
}
|
|
```
|
|
|
|
#### Metrics Collected
|
|
- `rustfs_concurrent_get_requests`: Current concurrent request count
|
|
- `rustfs_get_object_requests_completed`: Total completed requests
|
|
- `rustfs_get_object_duration_seconds`: Request duration histogram
|
|
- `rustfs_object_cache_hits`: Cache hit count
|
|
- `rustfs_object_cache_misses`: Cache miss count
|
|
- `rustfs_buffer_size_bytes`: Buffer size distribution
|
|
|
|
## Performance Expectations
|
|
|
|
### Expected Improvements
|
|
|
|
Based on the optimizations, we expect:
|
|
|
|
| Concurrency Level | Before | After (Expected) | Improvement |
|
|
|------------------|--------|------------------|-------------|
|
|
| 1 request | 59ms | 55-60ms | Similar (baseline) |
|
|
| 2 requests | 110ms | 65-75ms | ~40% faster |
|
|
| 4 requests | 200ms | 80-100ms | ~50% faster |
|
|
| 8 requests | 400ms | 100-130ms | ~65% faster |
|
|
| 16 requests | 800ms | 120-160ms | ~75% faster |
|
|
|
|
### Key Performance Characteristics
|
|
|
|
1. **Sub-linear scaling**: Latency increases sub-linearly with concurrency
|
|
2. **Cache benefits**: Hot objects see near-zero latency from cache hits
|
|
3. **Predictable behavior**: Bounded latency even under extreme load
|
|
4. **Memory efficiency**: Lower memory usage under high concurrency
|
|
|
|
## Implementation Details
|
|
|
|
### Integration Points
|
|
|
|
The optimization is integrated at the GetObject handler level:
|
|
|
|
```rust
|
|
async fn get_object(&self, req: S3Request<GetObjectInput>) -> S3Result<S3Response<GetObjectOutput>> {
|
|
// 1. Track request
|
|
let _request_guard = ConcurrencyManager::track_request();
|
|
|
|
// 2. Try cache
|
|
if let Some(cached_data) = manager.get_cached(&cache_key).await {
|
|
return Ok(S3Response::new(output)); // Fast path
|
|
}
|
|
|
|
// 3. Acquire I/O permit
|
|
let _disk_permit = manager.acquire_disk_read_permit().await;
|
|
|
|
// 4. Calculate optimal buffer size
|
|
let optimal_buffer_size = get_concurrency_aware_buffer_size(
|
|
response_content_length,
|
|
base_buffer_size
|
|
);
|
|
|
|
// 5. Stream with optimal buffer
|
|
let body = StreamingBlob::wrap(
|
|
ReaderStream::with_capacity(final_stream, optimal_buffer_size)
|
|
);
|
|
}
|
|
```
|
|
|
|
### Configuration
|
|
|
|
All defaults can be tuned via code changes:
|
|
|
|
```rust
|
|
// In concurrency.rs
|
|
const HIGH_CONCURRENCY_THRESHOLD: usize = 8;
|
|
const MEDIUM_CONCURRENCY_THRESHOLD: usize = 4;
|
|
|
|
// Cache settings
|
|
max_object_size: 10 * MI_B, // 10MB
|
|
max_cache_size: 100 * MI_B, // 100MB
|
|
disk_read_semaphore: Semaphore::new(64), // 64 concurrent reads
|
|
```
|
|
|
|
## Testing Recommendations
|
|
|
|
### 1. Concurrent Load Testing
|
|
|
|
Use the provided Go client to test different concurrency levels:
|
|
|
|
```go
|
|
concurrency := []int{1, 2, 4, 8, 16, 32}
|
|
for _, c := range concurrency {
|
|
// Run test with c concurrent goroutines
|
|
// Measure average latency and P50/P95/P99
|
|
}
|
|
```
|
|
|
|
### 2. Hot Object Testing
|
|
|
|
Test cache effectiveness with repeated reads:
|
|
|
|
```bash
|
|
# Read same object 100 times with 10 concurrent clients
|
|
for i in {1..10}; do
|
|
for j in {1..100}; do
|
|
mc cat rustfs/test/bxx > /dev/null
|
|
done &
|
|
done
|
|
wait
|
|
```
|
|
|
|
### 3. Mixed Workload Testing
|
|
|
|
Simulate real-world scenarios:
|
|
- 70% small objects (<1MB) - should see high cache hit rate
|
|
- 20% medium objects (1-10MB) - partial cache benefit
|
|
- 10% large objects (>10MB) - adaptive buffer sizing benefit
|
|
|
|
### 4. Stress Testing
|
|
|
|
Test system behavior under extreme load:
|
|
```bash
|
|
# 100 concurrent clients, continuous reads
|
|
ab -n 10000 -c 100 http://rustfs:9000/test/bxx
|
|
```
|
|
|
|
## Monitoring and Observability
|
|
|
|
### Key Metrics to Watch
|
|
|
|
1. **Latency Percentiles**
|
|
- P50, P95, P99 request duration
|
|
- Should show sub-linear growth with concurrency
|
|
|
|
2. **Cache Performance**
|
|
- Cache hit ratio (target: >70% for hot objects)
|
|
- Cache memory usage
|
|
- Eviction rate
|
|
|
|
3. **Resource Utilization**
|
|
- Memory usage per concurrent request
|
|
- Disk I/O queue depth
|
|
- CPU utilization
|
|
|
|
4. **Throughput**
|
|
- Requests per second
|
|
- Bytes per second
|
|
- Concurrent request count
|
|
|
|
### Prometheus Queries
|
|
|
|
```promql
|
|
# Average request duration by concurrency level
|
|
histogram_quantile(0.95,
|
|
rate(rustfs_get_object_duration_seconds_bucket[5m])
|
|
)
|
|
|
|
# Cache hit ratio
|
|
sum(rate(rustfs_object_cache_hits[5m]))
|
|
/
|
|
(sum(rate(rustfs_object_cache_hits[5m])) + sum(rate(rustfs_object_cache_misses[5m])))
|
|
|
|
# Concurrent requests over time
|
|
rustfs_concurrent_get_requests
|
|
|
|
# Memory efficiency (bytes per request)
|
|
rustfs_object_cache_size_bytes / rustfs_concurrent_get_requests
|
|
```
|
|
|
|
## Future Enhancements
|
|
|
|
### Potential Improvements
|
|
|
|
1. **Request Prioritization**
|
|
- Prioritize small requests over large ones
|
|
- Age-based priority to prevent starvation
|
|
- QoS classes for different clients
|
|
|
|
2. **Advanced Caching**
|
|
- Partial object caching (hot blocks)
|
|
- Predictive prefetching based on access patterns
|
|
- Distributed cache across multiple nodes
|
|
|
|
3. **I/O Scheduling**
|
|
- Batch similar requests for sequential I/O
|
|
- Deadline-based I/O scheduling
|
|
- NUMA-aware buffer allocation
|
|
|
|
4. **Adaptive Tuning**
|
|
- Machine learning based buffer sizing
|
|
- Dynamic cache size adjustment
|
|
- Workload-aware optimization
|
|
|
|
5. **Compression**
|
|
- Transparent compression for cached objects
|
|
- Adaptive compression based on CPU availability
|
|
- Deduplication for similar objects
|
|
|
|
## References
|
|
|
|
- [Issue #XXX](https://github.com/rustfs/rustfs/issues/XXX): Original performance issue
|
|
- [PR #XXX](https://github.com/rustfs/rustfs/pull/XXX): Implementation PR
|
|
- [MinIO Best Practices](https://min.io/docs/minio/linux/operations/install-deploy-manage/performance-and-optimization.html)
|
|
- [LRU Cache Design](https://leetcode.com/problems/lru-cache/)
|
|
- [Tokio Concurrency Patterns](https://tokio.rs/tokio/tutorial/shared-state)
|
|
|
|
## Conclusion
|
|
|
|
The concurrency-aware optimization addresses the root causes of performance degradation:
|
|
|
|
1. ✅ **Adaptive buffer sizing** reduces memory contention and improves cache utilization
|
|
2. ✅ **Hot object caching** eliminates redundant disk I/O for frequently accessed files
|
|
3. ✅ **I/O concurrency control** prevents disk saturation and ensures predictable latency
|
|
4. ✅ **Comprehensive monitoring** enables performance tracking and tuning
|
|
|
|
These changes should significantly improve performance under concurrent load while maintaining compatibility with existing clients and workloads.
|