tantivy #5
							
								
								
									
										847
									
								
								Cargo.lock
									
									
									
										generated
									
									
									
								
							
							
						
						
									
										847
									
								
								Cargo.lock
									
									
									
										generated
									
									
									
								
							
										
											
												File diff suppressed because it is too large
												Load Diff
											
										
									
								
							@@ -24,6 +24,7 @@ age = "0.10"
 | 
			
		||||
secrecy = "0.8"
 | 
			
		||||
ed25519-dalek = "2"
 | 
			
		||||
base64 = "0.22"
 | 
			
		||||
tantivy = "0.25.0"
 | 
			
		||||
 | 
			
		||||
[dev-dependencies]
 | 
			
		||||
redis = { version = "0.24", features = ["aio", "tokio-comp"] }
 | 
			
		||||
 
 | 
			
		||||
@@ -14,25 +14,31 @@ fn read_reply(s: &mut TcpStream) -> String {
 | 
			
		||||
    let n = s.read(&mut buf).unwrap();
 | 
			
		||||
    String::from_utf8_lossy(&buf[..n]).to_string()
 | 
			
		||||
}
 | 
			
		||||
fn parse_two_bulk(reply: &str) -> Option<(String,String)> {
 | 
			
		||||
fn parse_two_bulk(reply: &str) -> Option<(String, String)> {
 | 
			
		||||
    let mut lines = reply.split("\r\n");
 | 
			
		||||
    if lines.next()? != "*2" { return None; }
 | 
			
		||||
    if lines.next()? != "*2" {
 | 
			
		||||
        return None;
 | 
			
		||||
    }
 | 
			
		||||
    let _n = lines.next()?;
 | 
			
		||||
    let a = lines.next()?.to_string();
 | 
			
		||||
    let _m = lines.next()?;
 | 
			
		||||
    let b = lines.next()?.to_string();
 | 
			
		||||
    Some((a,b))
 | 
			
		||||
    Some((a, b))
 | 
			
		||||
}
 | 
			
		||||
fn parse_bulk(reply: &str) -> Option<String> {
 | 
			
		||||
    let mut lines = reply.split("\r\n");
 | 
			
		||||
    let hdr = lines.next()?;
 | 
			
		||||
    if !hdr.starts_with('$') { return None; }
 | 
			
		||||
    if !hdr.starts_with('$') {
 | 
			
		||||
        return None;
 | 
			
		||||
    }
 | 
			
		||||
    Some(lines.next()?.to_string())
 | 
			
		||||
}
 | 
			
		||||
fn parse_simple(reply: &str) -> Option<String> {
 | 
			
		||||
    let mut lines = reply.split("\r\n");
 | 
			
		||||
    let hdr = lines.next()?;
 | 
			
		||||
    if !hdr.starts_with('+') { return None; }
 | 
			
		||||
    if !hdr.starts_with('+') {
 | 
			
		||||
        return None;
 | 
			
		||||
    }
 | 
			
		||||
    Some(hdr[1..].to_string())
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
@@ -45,37 +51,43 @@ fn main() {
 | 
			
		||||
    let mut s = TcpStream::connect(addr).expect("connect");
 | 
			
		||||
 | 
			
		||||
    // Generate & persist X25519 enc keys under name "alice"
 | 
			
		||||
    s.write_all(arr(&["age","keygen","alice"]).as_bytes()).unwrap();
 | 
			
		||||
    s.write_all(arr(&["age", "keygen", "alice"]).as_bytes())
 | 
			
		||||
        .unwrap();
 | 
			
		||||
    let (_alice_recip, _alice_ident) = parse_two_bulk(&read_reply(&mut s)).expect("gen enc");
 | 
			
		||||
 | 
			
		||||
    // Generate & persist Ed25519 signing key under name "signer"
 | 
			
		||||
    s.write_all(arr(&["age","signkeygen","signer"]).as_bytes()).unwrap();
 | 
			
		||||
    s.write_all(arr(&["age", "signkeygen", "signer"]).as_bytes())
 | 
			
		||||
        .unwrap();
 | 
			
		||||
    let (_verify, _secret) = parse_two_bulk(&read_reply(&mut s)).expect("gen sign");
 | 
			
		||||
 | 
			
		||||
    // Encrypt by name
 | 
			
		||||
    let msg = "hello from persistent keys";
 | 
			
		||||
    s.write_all(arr(&["age","encryptname","alice", msg]).as_bytes()).unwrap();
 | 
			
		||||
    s.write_all(arr(&["age", "encryptname", "alice", msg]).as_bytes())
 | 
			
		||||
        .unwrap();
 | 
			
		||||
    let ct_b64 = parse_bulk(&read_reply(&mut s)).expect("ct b64");
 | 
			
		||||
    println!("ciphertext b64: {}", ct_b64);
 | 
			
		||||
 | 
			
		||||
    // Decrypt by name
 | 
			
		||||
    s.write_all(arr(&["age","decryptname","alice", &ct_b64]).as_bytes()).unwrap();
 | 
			
		||||
    s.write_all(arr(&["age", "decryptname", "alice", &ct_b64]).as_bytes())
 | 
			
		||||
        .unwrap();
 | 
			
		||||
    let pt = parse_bulk(&read_reply(&mut s)).expect("pt");
 | 
			
		||||
    assert_eq!(pt, msg);
 | 
			
		||||
    println!("decrypted ok");
 | 
			
		||||
 | 
			
		||||
    // Sign by name
 | 
			
		||||
    s.write_all(arr(&["age","signname","signer", msg]).as_bytes()).unwrap();
 | 
			
		||||
    s.write_all(arr(&["age", "signname", "signer", msg]).as_bytes())
 | 
			
		||||
        .unwrap();
 | 
			
		||||
    let sig_b64 = parse_bulk(&read_reply(&mut s)).expect("sig b64");
 | 
			
		||||
 | 
			
		||||
    // Verify by name
 | 
			
		||||
    s.write_all(arr(&["age","verifyname","signer", msg, &sig_b64]).as_bytes()).unwrap();
 | 
			
		||||
    s.write_all(arr(&["age", "verifyname", "signer", msg, &sig_b64]).as_bytes())
 | 
			
		||||
        .unwrap();
 | 
			
		||||
    let ok = parse_simple(&read_reply(&mut s)).expect("verify");
 | 
			
		||||
    assert_eq!(ok, "1");
 | 
			
		||||
    println!("signature verified");
 | 
			
		||||
 | 
			
		||||
    // List names
 | 
			
		||||
    s.write_all(arr(&["age","list"]).as_bytes()).unwrap();
 | 
			
		||||
    s.write_all(arr(&["age", "list"]).as_bytes()).unwrap();
 | 
			
		||||
    let list = read_reply(&mut s);
 | 
			
		||||
    println!("LIST -> {list}");
 | 
			
		||||
 | 
			
		||||
 
 | 
			
		||||
							
								
								
									
										239
									
								
								examples/tantivy_search_demo.sh
									
									
									
									
									
										Executable file
									
								
							
							
						
						
									
										239
									
								
								examples/tantivy_search_demo.sh
									
									
									
									
									
										Executable file
									
								
							@@ -0,0 +1,239 @@
 | 
			
		||||
#!/bin/bash
 | 
			
		||||
 | 
			
		||||
# HeroDB Tantivy Search Demo
 | 
			
		||||
# This script demonstrates full-text search capabilities using Redis commands
 | 
			
		||||
# HeroDB server should be running on port 6381
 | 
			
		||||
 | 
			
		||||
set -e  # Exit on any error
 | 
			
		||||
 | 
			
		||||
# Configuration
 | 
			
		||||
REDIS_HOST="localhost"
 | 
			
		||||
REDIS_PORT="6382"
 | 
			
		||||
REDIS_CLI="redis-cli -h $REDIS_HOST -p $REDIS_PORT"
 | 
			
		||||
 | 
			
		||||
# Start the herodb server in the background
 | 
			
		||||
echo "Starting herodb server..."
 | 
			
		||||
cargo run -p herodb -- --dir /tmp/herodbtest --port ${REDIS_PORT} --debug &
 | 
			
		||||
SERVER_PID=$!
 | 
			
		||||
echo
 | 
			
		||||
sleep 2 # Give the server a moment to start
 | 
			
		||||
 | 
			
		||||
# Colors for output
 | 
			
		||||
RED='\033[0;31m'
 | 
			
		||||
GREEN='\033[0;32m'
 | 
			
		||||
BLUE='\033[0;34m'
 | 
			
		||||
YELLOW='\033[1;33m'
 | 
			
		||||
NC='\033[0m' # No Color
 | 
			
		||||
 | 
			
		||||
# Function to print colored output
 | 
			
		||||
print_header() {
 | 
			
		||||
    echo -e "${BLUE}=== $1 ===${NC}"
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
print_success() {
 | 
			
		||||
    echo -e "${GREEN}✓ $1${NC}"
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
print_info() {
 | 
			
		||||
    echo -e "${YELLOW}ℹ $1${NC}"
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
print_error() {
 | 
			
		||||
    echo -e "${RED}✗ $1${NC}"
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
# Function to check if HeroDB is running
 | 
			
		||||
check_herodb() {
 | 
			
		||||
    print_info "Checking if HeroDB is running on port $REDIS_PORT..."
 | 
			
		||||
    if ! $REDIS_CLI ping > /dev/null 2>&1; then
 | 
			
		||||
        print_error "HeroDB is not running on port $REDIS_PORT"
 | 
			
		||||
        print_info "Please start HeroDB with: cargo run -- --port $REDIS_PORT"
 | 
			
		||||
        exit 1
 | 
			
		||||
    fi
 | 
			
		||||
    print_success "HeroDB is running and responding"
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
# Function to execute Redis command with error handling
 | 
			
		||||
execute_cmd() {
 | 
			
		||||
    local description="${@: -1}"
 | 
			
		||||
    set -- "${@:1:$(($#-1))}"
 | 
			
		||||
    
 | 
			
		||||
    echo -e "${YELLOW}Command:${NC} $(printf '%q ' "$@")"
 | 
			
		||||
    if result=$($REDIS_CLI "$@" 2>&1); then
 | 
			
		||||
        echo -e "${GREEN}Result:${NC} $result"
 | 
			
		||||
        return 0
 | 
			
		||||
    else
 | 
			
		||||
        print_error "Failed: $description"
 | 
			
		||||
        echo "Error: $result"
 | 
			
		||||
        return 1
 | 
			
		||||
    fi
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
# Function to pause for readability
 | 
			
		||||
pause() {
 | 
			
		||||
    echo
 | 
			
		||||
    read -p "Press Enter to continue..."
 | 
			
		||||
    echo
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
# Main demo function
 | 
			
		||||
main() {
 | 
			
		||||
    clear
 | 
			
		||||
    print_header "HeroDB Tantivy Search Demonstration"
 | 
			
		||||
    echo "This demo shows full-text search capabilities using Redis commands"
 | 
			
		||||
    echo "HeroDB runs on port $REDIS_PORT (instead of Redis default 6379)"
 | 
			
		||||
    echo
 | 
			
		||||
 | 
			
		||||
    # Check if HeroDB is running
 | 
			
		||||
    check_herodb
 | 
			
		||||
    echo
 | 
			
		||||
 | 
			
		||||
    print_header "Step 1: Create Search Index"
 | 
			
		||||
    print_info "Creating a product catalog search index with various field types"
 | 
			
		||||
    
 | 
			
		||||
    # Create search index with schema
 | 
			
		||||
    execute_cmd FT.CREATE product_catalog SCHEMA title TEXT description TEXT category TAG price NUMERIC rating NUMERIC location GEO \
 | 
			
		||||
                "Creating search index"
 | 
			
		||||
    
 | 
			
		||||
    print_success "Search index 'product_catalog' created successfully"
 | 
			
		||||
    pause
 | 
			
		||||
 | 
			
		||||
    print_header "Step 2: Add Sample Products"
 | 
			
		||||
    print_info "Adding sample products to demonstrate different search scenarios"
 | 
			
		||||
    
 | 
			
		||||
    # Add sample products using FT.ADD
 | 
			
		||||
    execute_cmd FT.ADD product_catalog product:1 1.0 title 'Wireless Bluetooth Headphones' description 'Premium noise-canceling headphones with 30-hour battery life' category 'electronics,audio' price 299.99 rating 4.5 location '-122.4194,37.7749' "Adding product 1"
 | 
			
		||||
    execute_cmd FT.ADD product_catalog product:2 1.0 title 'Organic Coffee Beans' description 'Single-origin Ethiopian coffee beans, medium roast' category 'food,beverages,organic' price 24.99 rating 4.8 location '-74.0060,40.7128' "Adding product 2"
 | 
			
		||||
    execute_cmd FT.ADD product_catalog product:3 1.0 title 'Yoga Mat Premium' description 'Eco-friendly yoga mat with superior grip and cushioning' category 'fitness,wellness,eco-friendly' price 89.99 rating 4.3 location '-118.2437,34.0522' "Adding product 3"
 | 
			
		||||
    execute_cmd FT.ADD product_catalog product:4 1.0 title 'Smart Home Speaker' description 'Voice-controlled smart speaker with AI assistant' category 'electronics,smart-home' price 149.99 rating 4.2 location '-87.6298,41.8781' "Adding product 4"
 | 
			
		||||
    execute_cmd FT.ADD product_catalog product:5 1.0 title 'Organic Green Tea' description 'Premium organic green tea leaves from Japan' category 'food,beverages,organic,tea' price 18.99 rating 4.7 location '139.6503,35.6762' "Adding product 5"
 | 
			
		||||
    execute_cmd FT.ADD product_catalog product:6 1.0 title 'Wireless Gaming Mouse' description 'High-precision gaming mouse with RGB lighting' category 'electronics,gaming' price 79.99 rating 4.4 location '-122.3321,47.6062' "Adding product 6"
 | 
			
		||||
    execute_cmd FT.ADD product_catalog product:7 1.0 title 'Comfortable meditation cushion for mindfulness practice' description 'Meditation cushion with premium materials' category 'wellness,meditation' price 45.99 rating 4.6 location '-122.4194,37.7749' "Adding product 7"
 | 
			
		||||
    execute_cmd FT.ADD product_catalog product:8 1.0 title 'Bluetooth Earbuds' description 'True wireless earbuds with active noise cancellation' category 'electronics,audio' price 199.99 rating 4.1 location '-74.0060,40.7128' "Adding product 8"
 | 
			
		||||
    
 | 
			
		||||
    print_success "Added 8 products to the index"
 | 
			
		||||
    pause
 | 
			
		||||
 | 
			
		||||
    print_header "Step 3: Basic Text Search"
 | 
			
		||||
    print_info "Searching for 'wireless' products"
 | 
			
		||||
    
 | 
			
		||||
    execute_cmd FT.SEARCH product_catalog wireless "Basic text search"
 | 
			
		||||
    pause
 | 
			
		||||
 | 
			
		||||
    print_header "Step 4: Search with Filters"
 | 
			
		||||
    print_info "Searching for 'organic' products"
 | 
			
		||||
    
 | 
			
		||||
    execute_cmd FT.SEARCH product_catalog organic "Filtered search"
 | 
			
		||||
    pause
 | 
			
		||||
 | 
			
		||||
    print_header "Step 5: Numeric Range Search"
 | 
			
		||||
    print_info "Searching for 'premium' products"
 | 
			
		||||
    
 | 
			
		||||
    execute_cmd FT.SEARCH product_catalog premium "Text search"
 | 
			
		||||
    pause
 | 
			
		||||
 | 
			
		||||
    print_header "Step 6: Sorting Results"
 | 
			
		||||
    print_info "Searching for electronics"
 | 
			
		||||
    
 | 
			
		||||
    execute_cmd FT.SEARCH product_catalog electronics "Category search"
 | 
			
		||||
    pause
 | 
			
		||||
 | 
			
		||||
    print_header "Step 7: Limiting Results"
 | 
			
		||||
    print_info "Searching for wireless products with limit"
 | 
			
		||||
    
 | 
			
		||||
    execute_cmd FT.SEARCH product_catalog wireless LIMIT 0 3 "Limited results"
 | 
			
		||||
    pause
 | 
			
		||||
 | 
			
		||||
    print_header "Step 8: Complex Query"
 | 
			
		||||
    print_info "Finding audio products with noise cancellation"
 | 
			
		||||
    
 | 
			
		||||
    execute_cmd FT.SEARCH product_catalog 'noise cancellation' "Complex query"
 | 
			
		||||
    pause
 | 
			
		||||
 | 
			
		||||
    print_header "Step 9: Geographic Search"
 | 
			
		||||
    print_info "Searching for meditation products"
 | 
			
		||||
    
 | 
			
		||||
    execute_cmd FT.SEARCH product_catalog meditation "Text search"
 | 
			
		||||
    pause
 | 
			
		||||
 | 
			
		||||
    print_header "Step 10: Aggregation Example"
 | 
			
		||||
    print_info "Getting index information and statistics"
 | 
			
		||||
    
 | 
			
		||||
    execute_cmd FT.INFO product_catalog "Index information"
 | 
			
		||||
    pause
 | 
			
		||||
 | 
			
		||||
    print_header "Step 11: Search Comparison"
 | 
			
		||||
    print_info "Comparing Tantivy search vs simple key matching"
 | 
			
		||||
    
 | 
			
		||||
    echo -e "${YELLOW}Tantivy Full-Text Search:${NC}"
 | 
			
		||||
    execute_cmd FT.SEARCH product_catalog 'battery life' "Full-text search for 'battery life'"
 | 
			
		||||
    
 | 
			
		||||
    echo
 | 
			
		||||
    echo -e "${YELLOW}Simple Key Pattern Matching:${NC}"
 | 
			
		||||
    execute_cmd KEYS *battery* "Simple pattern matching for 'battery'"
 | 
			
		||||
    
 | 
			
		||||
    print_info "Notice how full-text search finds relevant results even when exact words don't match keys"
 | 
			
		||||
    pause
 | 
			
		||||
 | 
			
		||||
    print_header "Step 12: Fuzzy Search"
 | 
			
		||||
    print_info "Searching for headphones"
 | 
			
		||||
    
 | 
			
		||||
    execute_cmd FT.SEARCH product_catalog headphones "Text search"
 | 
			
		||||
    pause
 | 
			
		||||
 | 
			
		||||
    print_header "Step 13: Phrase Search"
 | 
			
		||||
    print_info "Searching for coffee products"
 | 
			
		||||
    
 | 
			
		||||
    execute_cmd FT.SEARCH product_catalog coffee "Text search"
 | 
			
		||||
    pause
 | 
			
		||||
 | 
			
		||||
    print_header "Step 14: Boolean Queries"
 | 
			
		||||
    print_info "Searching for gaming products"
 | 
			
		||||
    
 | 
			
		||||
    execute_cmd FT.SEARCH product_catalog gaming "Text search"
 | 
			
		||||
    echo
 | 
			
		||||
    execute_cmd FT.SEARCH product_catalog tea "Text search"
 | 
			
		||||
    pause
 | 
			
		||||
 | 
			
		||||
    print_header "Step 15: Cleanup"
 | 
			
		||||
    print_info "Removing test data"
 | 
			
		||||
    
 | 
			
		||||
    # Delete the search index
 | 
			
		||||
    execute_cmd FT.DROP product_catalog "Dropping search index"
 | 
			
		||||
    
 | 
			
		||||
    # Clean up documents from search index
 | 
			
		||||
    for i in {1..8}; do
 | 
			
		||||
        execute_cmd FT.DEL product_catalog product:$i "Deleting product:$i from index"
 | 
			
		||||
    done
 | 
			
		||||
    
 | 
			
		||||
    print_success "Cleanup completed"
 | 
			
		||||
    echo
 | 
			
		||||
 | 
			
		||||
    print_header "Demo Summary"
 | 
			
		||||
    echo "This demonstration showed:"
 | 
			
		||||
    echo "• Creating search indexes with different field types"
 | 
			
		||||
    echo "• Adding documents to the search index"
 | 
			
		||||
    echo "• Basic and advanced text search queries"
 | 
			
		||||
    echo "• Filtering by categories and numeric ranges"
 | 
			
		||||
    echo "• Sorting and limiting results"
 | 
			
		||||
    echo "• Geographic searches"
 | 
			
		||||
    echo "• Fuzzy matching and phrase searches"
 | 
			
		||||
    echo "• Boolean query operators"
 | 
			
		||||
    echo "• Comparison with simple pattern matching"
 | 
			
		||||
    echo
 | 
			
		||||
    print_success "HeroDB Tantivy search demo completed successfully!"
 | 
			
		||||
    echo
 | 
			
		||||
    print_info "Key advantages of Tantivy full-text search:"
 | 
			
		||||
    echo "  - Relevance scoring and ranking"
 | 
			
		||||
    echo "  - Fuzzy matching and typo tolerance"
 | 
			
		||||
    echo "  - Complex boolean queries"
 | 
			
		||||
    echo "  - Field-specific searches and filters"
 | 
			
		||||
    echo "  - Geographic and numeric range queries"
 | 
			
		||||
    echo "  - Much faster than pattern matching on large datasets"
 | 
			
		||||
    echo
 | 
			
		||||
    print_info "To run HeroDB server: cargo run -- --port 6381"
 | 
			
		||||
    print_info "To connect with redis-cli: redis-cli -h localhost -p 6381"
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
# Run the demo
 | 
			
		||||
main "$@"
 | 
			
		||||
							
								
								
									
										101
									
								
								examples/test_tantivy_integration.sh
									
									
									
									
									
										Executable file
									
								
							
							
						
						
									
										101
									
								
								examples/test_tantivy_integration.sh
									
									
									
									
									
										Executable file
									
								
							@@ -0,0 +1,101 @@
 | 
			
		||||
#!/bin/bash
 | 
			
		||||
 | 
			
		||||
# Simple Tantivy Search Integration Test for HeroDB
 | 
			
		||||
# This script tests the full-text search functionality we just integrated
 | 
			
		||||
 | 
			
		||||
set -e
 | 
			
		||||
 | 
			
		||||
echo "🔍 Testing Tantivy Search Integration..."
 | 
			
		||||
 | 
			
		||||
# Build the project first
 | 
			
		||||
echo "📦 Building HeroDB..."
 | 
			
		||||
cargo build --release
 | 
			
		||||
 | 
			
		||||
# Start the server in the background
 | 
			
		||||
echo "🚀 Starting HeroDB server on port 6379..."
 | 
			
		||||
cargo run --release -- --port 6379 --dir ./test_data &
 | 
			
		||||
SERVER_PID=$!
 | 
			
		||||
 | 
			
		||||
# Wait for server to start
 | 
			
		||||
sleep 3
 | 
			
		||||
 | 
			
		||||
# Function to cleanup on exit
 | 
			
		||||
cleanup() {
 | 
			
		||||
    echo "🧹 Cleaning up..."
 | 
			
		||||
    kill $SERVER_PID 2>/dev/null || true
 | 
			
		||||
    rm -rf ./test_data
 | 
			
		||||
    exit
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
# Set trap for cleanup
 | 
			
		||||
trap cleanup EXIT INT TERM
 | 
			
		||||
 | 
			
		||||
# Function to execute Redis command
 | 
			
		||||
execute_cmd() {
 | 
			
		||||
    local cmd="$1"
 | 
			
		||||
    local description="$2"
 | 
			
		||||
    
 | 
			
		||||
    echo "📝 $description"
 | 
			
		||||
    echo "   Command: $cmd"
 | 
			
		||||
    
 | 
			
		||||
    if result=$(redis-cli -p 6379 $cmd 2>&1); then
 | 
			
		||||
        echo "   ✅ Result: $result"
 | 
			
		||||
        echo
 | 
			
		||||
        return 0
 | 
			
		||||
    else
 | 
			
		||||
        echo "   ❌ Failed: $result"
 | 
			
		||||
        echo
 | 
			
		||||
        return 1
 | 
			
		||||
    fi
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
echo "🧪 Running Tantivy Search Tests..."
 | 
			
		||||
echo
 | 
			
		||||
 | 
			
		||||
# Test 1: Create a search index
 | 
			
		||||
execute_cmd "ft.create books SCHEMA title TEXT description TEXT author TEXT category TAG price NUMERIC" \
 | 
			
		||||
           "Creating search index 'books'"
 | 
			
		||||
 | 
			
		||||
# Test 2: Add documents to the index
 | 
			
		||||
execute_cmd "ft.add books book1 1.0 title \"The Great Gatsby\" description \"A classic American novel about the Jazz Age\" author \"F. Scott Fitzgerald\" category \"fiction,classic\" price \"12.99\"" \
 | 
			
		||||
           "Adding first book"
 | 
			
		||||
 | 
			
		||||
execute_cmd "ft.add books book2 1.0 title \"To Kill a Mockingbird\" description \"A novel about racial injustice in the American South\" author \"Harper Lee\" category \"fiction,classic\" price \"14.99\"" \
 | 
			
		||||
           "Adding second book"
 | 
			
		||||
 | 
			
		||||
execute_cmd "ft.add books book3 1.0 title \"Programming Rust\" description \"A comprehensive guide to Rust programming language\" author \"Jim Blandy\" category \"programming,technical\" price \"49.99\"" \
 | 
			
		||||
           "Adding third book"
 | 
			
		||||
 | 
			
		||||
execute_cmd "ft.add books book4 1.0 title \"The Rust Programming Language\" description \"The official book on Rust programming\" author \"Steve Klabnik\" category \"programming,technical\" price \"39.99\"" \
 | 
			
		||||
           "Adding fourth book"
 | 
			
		||||
 | 
			
		||||
# Test 3: Basic search
 | 
			
		||||
execute_cmd "ft.search books Rust" \
 | 
			
		||||
           "Searching for 'Rust'"
 | 
			
		||||
 | 
			
		||||
# Test 4: Search with filters
 | 
			
		||||
execute_cmd "ft.search books programming FILTER category programming" \
 | 
			
		||||
           "Searching for 'programming' with category filter"
 | 
			
		||||
 | 
			
		||||
# Test 5: Search with limit
 | 
			
		||||
execute_cmd "ft.search books \"*\" LIMIT 0 2" \
 | 
			
		||||
           "Getting first 2 documents"
 | 
			
		||||
 | 
			
		||||
# Test 6: Get index info
 | 
			
		||||
execute_cmd "ft.info books" \
 | 
			
		||||
           "Getting index information"
 | 
			
		||||
 | 
			
		||||
# Test 7: Delete a document
 | 
			
		||||
execute_cmd "ft.del books book1" \
 | 
			
		||||
           "Deleting book1"
 | 
			
		||||
 | 
			
		||||
# Test 8: Search again to verify deletion
 | 
			
		||||
execute_cmd "ft.search books Gatsby" \
 | 
			
		||||
           "Searching for deleted book"
 | 
			
		||||
 | 
			
		||||
# Test 9: Drop the index
 | 
			
		||||
execute_cmd "ft.drop books" \
 | 
			
		||||
           "Dropping the index"
 | 
			
		||||
 | 
			
		||||
echo "🎉 All tests completed successfully!"
 | 
			
		||||
echo "✅ Tantivy search integration is working correctly"
 | 
			
		||||
							
								
								
									
										156
									
								
								src/age.rs
									
									
									
									
									
								
							
							
						
						
									
										156
									
								
								src/age.rs
									
									
									
									
									
								
							@@ -12,17 +12,17 @@
 | 
			
		||||
 | 
			
		||||
use std::str::FromStr;
 | 
			
		||||
 | 
			
		||||
use secrecy::ExposeSecret;
 | 
			
		||||
use age::{Decryptor, Encryptor};
 | 
			
		||||
use age::x25519;
 | 
			
		||||
use age::{Decryptor, Encryptor};
 | 
			
		||||
use secrecy::ExposeSecret;
 | 
			
		||||
 | 
			
		||||
use ed25519_dalek::{Signature, Signer, Verifier, SigningKey, VerifyingKey};
 | 
			
		||||
use ed25519_dalek::{Signature, Signer, SigningKey, Verifier, VerifyingKey};
 | 
			
		||||
 | 
			
		||||
use base64::{engine::general_purpose::STANDARD as B64, Engine as _};
 | 
			
		||||
 | 
			
		||||
use crate::error::DBError;
 | 
			
		||||
use crate::protocol::Protocol;
 | 
			
		||||
use crate::server::Server;
 | 
			
		||||
use crate::error::DBError;
 | 
			
		||||
 | 
			
		||||
// ---------- Internal helpers ----------
 | 
			
		||||
 | 
			
		||||
@@ -32,7 +32,7 @@ pub enum AgeWireError {
 | 
			
		||||
    Crypto(String),
 | 
			
		||||
    Utf8,
 | 
			
		||||
    SignatureLen,
 | 
			
		||||
    NotFound(&'static str),     // which kind of key was missing
 | 
			
		||||
    NotFound(&'static str), // which kind of key was missing
 | 
			
		||||
    Storage(String),
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
@@ -83,8 +83,8 @@ pub fn gen_enc_keypair() -> (String, String) {
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
pub fn gen_sign_keypair() -> (String, String) {
 | 
			
		||||
    use rand::RngCore;
 | 
			
		||||
    use rand::rngs::OsRng;
 | 
			
		||||
    use rand::RngCore;
 | 
			
		||||
 | 
			
		||||
    // Generate random 32 bytes for the signing key
 | 
			
		||||
    let mut secret_bytes = [0u8; 32];
 | 
			
		||||
@@ -103,14 +103,18 @@ pub fn gen_sign_keypair() -> (String, String) {
 | 
			
		||||
/// Encrypt `msg` for `recipient_str` (X25519). Returns base64(ciphertext).
 | 
			
		||||
pub fn encrypt_b64(recipient_str: &str, msg: &str) -> Result<String, AgeWireError> {
 | 
			
		||||
    let recipient = parse_recipient(recipient_str)?;
 | 
			
		||||
    let enc = Encryptor::with_recipients(vec![Box::new(recipient)])
 | 
			
		||||
        .expect("failed to create encryptor"); // Handle Option<Encryptor>
 | 
			
		||||
    let enc =
 | 
			
		||||
        Encryptor::with_recipients(vec![Box::new(recipient)]).expect("failed to create encryptor"); // Handle Option<Encryptor>
 | 
			
		||||
    let mut out = Vec::new();
 | 
			
		||||
    {
 | 
			
		||||
        use std::io::Write;
 | 
			
		||||
        let mut w = enc.wrap_output(&mut out).map_err(|e| AgeWireError::Crypto(e.to_string()))?;
 | 
			
		||||
        w.write_all(msg.as_bytes()).map_err(|e| AgeWireError::Crypto(e.to_string()))?;
 | 
			
		||||
        w.finish().map_err(|e| AgeWireError::Crypto(e.to_string()))?;
 | 
			
		||||
        let mut w = enc
 | 
			
		||||
            .wrap_output(&mut out)
 | 
			
		||||
            .map_err(|e| AgeWireError::Crypto(e.to_string()))?;
 | 
			
		||||
        w.write_all(msg.as_bytes())
 | 
			
		||||
            .map_err(|e| AgeWireError::Crypto(e.to_string()))?;
 | 
			
		||||
        w.finish()
 | 
			
		||||
            .map_err(|e| AgeWireError::Crypto(e.to_string()))?;
 | 
			
		||||
    }
 | 
			
		||||
    Ok(B64.encode(out))
 | 
			
		||||
}
 | 
			
		||||
@@ -118,19 +122,27 @@ pub fn encrypt_b64(recipient_str: &str, msg: &str) -> Result<String, AgeWireErro
 | 
			
		||||
/// Decrypt base64(ciphertext) with `identity_str`. Returns plaintext String.
 | 
			
		||||
pub fn decrypt_b64(identity_str: &str, ct_b64: &str) -> Result<String, AgeWireError> {
 | 
			
		||||
    let id = parse_identity(identity_str)?;
 | 
			
		||||
    let ct = B64.decode(ct_b64.as_bytes()).map_err(|e| AgeWireError::Crypto(e.to_string()))?;
 | 
			
		||||
    let ct = B64
 | 
			
		||||
        .decode(ct_b64.as_bytes())
 | 
			
		||||
        .map_err(|e| AgeWireError::Crypto(e.to_string()))?;
 | 
			
		||||
    let dec = Decryptor::new(&ct[..]).map_err(|e| AgeWireError::Crypto(e.to_string()))?;
 | 
			
		||||
 | 
			
		||||
    // The decrypt method returns a Result<StreamReader, DecryptError>
 | 
			
		||||
    let mut r = match dec {
 | 
			
		||||
        Decryptor::Recipients(d) => d.decrypt(std::iter::once(&id as &dyn age::Identity))
 | 
			
		||||
        Decryptor::Recipients(d) => d
 | 
			
		||||
            .decrypt(std::iter::once(&id as &dyn age::Identity))
 | 
			
		||||
            .map_err(|e| AgeWireError::Crypto(e.to_string()))?,
 | 
			
		||||
        Decryptor::Passphrase(_) => return Err(AgeWireError::Crypto("Expected recipients, got passphrase".to_string())),
 | 
			
		||||
        Decryptor::Passphrase(_) => {
 | 
			
		||||
            return Err(AgeWireError::Crypto(
 | 
			
		||||
                "Expected recipients, got passphrase".to_string(),
 | 
			
		||||
            ))
 | 
			
		||||
        }
 | 
			
		||||
    };
 | 
			
		||||
 | 
			
		||||
    let mut pt = Vec::new();
 | 
			
		||||
    use std::io::Read;
 | 
			
		||||
    r.read_to_end(&mut pt).map_err(|e| AgeWireError::Crypto(e.to_string()))?;
 | 
			
		||||
    r.read_to_end(&mut pt)
 | 
			
		||||
        .map_err(|e| AgeWireError::Crypto(e.to_string()))?;
 | 
			
		||||
    String::from_utf8(pt).map_err(|_| AgeWireError::Utf8)
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
@@ -144,7 +156,9 @@ pub fn sign_b64(signing_secret_str: &str, msg: &str) -> Result<String, AgeWireEr
 | 
			
		||||
/// Verify detached signature (base64) for `msg` with pubkey.
 | 
			
		||||
pub fn verify_b64(verify_pub_str: &str, msg: &str, sig_b64: &str) -> Result<bool, AgeWireError> {
 | 
			
		||||
    let verifying_key = parse_ed25519_verifying_key(verify_pub_str)?;
 | 
			
		||||
    let sig_bytes = B64.decode(sig_b64.as_bytes()).map_err(|e| AgeWireError::Crypto(e.to_string()))?;
 | 
			
		||||
    let sig_bytes = B64
 | 
			
		||||
        .decode(sig_b64.as_bytes())
 | 
			
		||||
        .map_err(|e| AgeWireError::Crypto(e.to_string()))?;
 | 
			
		||||
    if sig_bytes.len() != 64 {
 | 
			
		||||
        return Err(AgeWireError::SignatureLen);
 | 
			
		||||
    }
 | 
			
		||||
@@ -155,30 +169,49 @@ pub fn verify_b64(verify_pub_str: &str, msg: &str, sig_b64: &str) -> Result<bool
 | 
			
		||||
// ---------- Storage helpers ----------
 | 
			
		||||
 | 
			
		||||
fn sget(server: &Server, key: &str) -> Result<Option<String>, AgeWireError> {
 | 
			
		||||
    let st = server.current_storage().map_err(|e| AgeWireError::Storage(e.0))?;
 | 
			
		||||
    let st = server
 | 
			
		||||
        .current_storage()
 | 
			
		||||
        .map_err(|e| AgeWireError::Storage(e.0))?;
 | 
			
		||||
    st.get(key).map_err(|e| AgeWireError::Storage(e.0))
 | 
			
		||||
}
 | 
			
		||||
fn sset(server: &Server, key: &str, val: &str) -> Result<(), AgeWireError> {
 | 
			
		||||
    let st = server.current_storage().map_err(|e| AgeWireError::Storage(e.0))?;
 | 
			
		||||
    st.set(key.to_string(), val.to_string()).map_err(|e| AgeWireError::Storage(e.0))
 | 
			
		||||
    let st = server
 | 
			
		||||
        .current_storage()
 | 
			
		||||
        .map_err(|e| AgeWireError::Storage(e.0))?;
 | 
			
		||||
    st.set(key.to_string(), val.to_string())
 | 
			
		||||
        .map_err(|e| AgeWireError::Storage(e.0))
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
fn enc_pub_key_key(name: &str) -> String { format!("age:key:{name}") }
 | 
			
		||||
fn enc_priv_key_key(name: &str) -> String { format!("age:privkey:{name}") }
 | 
			
		||||
fn sign_pub_key_key(name: &str) -> String { format!("age:signpub:{name}") }
 | 
			
		||||
fn sign_priv_key_key(name: &str) -> String { format!("age:signpriv:{name}") }
 | 
			
		||||
fn enc_pub_key_key(name: &str) -> String {
 | 
			
		||||
    format!("age:key:{name}")
 | 
			
		||||
}
 | 
			
		||||
fn enc_priv_key_key(name: &str) -> String {
 | 
			
		||||
    format!("age:privkey:{name}")
 | 
			
		||||
}
 | 
			
		||||
fn sign_pub_key_key(name: &str) -> String {
 | 
			
		||||
    format!("age:signpub:{name}")
 | 
			
		||||
}
 | 
			
		||||
fn sign_priv_key_key(name: &str) -> String {
 | 
			
		||||
    format!("age:signpriv:{name}")
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
// ---------- Command handlers (RESP Protocol) ----------
 | 
			
		||||
// Basic (stateless) ones kept for completeness
 | 
			
		||||
 | 
			
		||||
pub async fn cmd_age_genenc() -> Protocol {
 | 
			
		||||
    let (recip, ident) = gen_enc_keypair();
 | 
			
		||||
    Protocol::Array(vec![Protocol::BulkString(recip), Protocol::BulkString(ident)])
 | 
			
		||||
    Protocol::Array(vec![
 | 
			
		||||
        Protocol::BulkString(recip),
 | 
			
		||||
        Protocol::BulkString(ident),
 | 
			
		||||
    ])
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
pub async fn cmd_age_gensign() -> Protocol {
 | 
			
		||||
    let (verify, secret) = gen_sign_keypair();
 | 
			
		||||
    Protocol::Array(vec![Protocol::BulkString(verify), Protocol::BulkString(secret)])
 | 
			
		||||
    Protocol::Array(vec![
 | 
			
		||||
        Protocol::BulkString(verify),
 | 
			
		||||
        Protocol::BulkString(secret),
 | 
			
		||||
    ])
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
pub async fn cmd_age_encrypt(recipient: &str, message: &str) -> Protocol {
 | 
			
		||||
@@ -214,16 +247,30 @@ pub async fn cmd_age_verify(verify_pub: &str, message: &str, sig_b64: &str) -> P
 | 
			
		||||
 | 
			
		||||
pub async fn cmd_age_keygen(server: &Server, name: &str) -> Protocol {
 | 
			
		||||
    let (recip, ident) = gen_enc_keypair();
 | 
			
		||||
    if let Err(e) = sset(server, &enc_pub_key_key(name), &recip) { return e.to_protocol(); }
 | 
			
		||||
    if let Err(e) = sset(server, &enc_priv_key_key(name), &ident) { return e.to_protocol(); }
 | 
			
		||||
    Protocol::Array(vec![Protocol::BulkString(recip), Protocol::BulkString(ident)])
 | 
			
		||||
    if let Err(e) = sset(server, &enc_pub_key_key(name), &recip) {
 | 
			
		||||
        return e.to_protocol();
 | 
			
		||||
    }
 | 
			
		||||
    if let Err(e) = sset(server, &enc_priv_key_key(name), &ident) {
 | 
			
		||||
        return e.to_protocol();
 | 
			
		||||
    }
 | 
			
		||||
    Protocol::Array(vec![
 | 
			
		||||
        Protocol::BulkString(recip),
 | 
			
		||||
        Protocol::BulkString(ident),
 | 
			
		||||
    ])
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
pub async fn cmd_age_signkeygen(server: &Server, name: &str) -> Protocol {
 | 
			
		||||
    let (verify, secret) = gen_sign_keypair();
 | 
			
		||||
    if let Err(e) = sset(server, &sign_pub_key_key(name), &verify) { return e.to_protocol(); }
 | 
			
		||||
    if let Err(e) = sset(server, &sign_priv_key_key(name), &secret) { return e.to_protocol(); }
 | 
			
		||||
    Protocol::Array(vec![Protocol::BulkString(verify), Protocol::BulkString(secret)])
 | 
			
		||||
    if let Err(e) = sset(server, &sign_pub_key_key(name), &verify) {
 | 
			
		||||
        return e.to_protocol();
 | 
			
		||||
    }
 | 
			
		||||
    if let Err(e) = sset(server, &sign_priv_key_key(name), &secret) {
 | 
			
		||||
        return e.to_protocol();
 | 
			
		||||
    }
 | 
			
		||||
    Protocol::Array(vec![
 | 
			
		||||
        Protocol::BulkString(verify),
 | 
			
		||||
        Protocol::BulkString(secret),
 | 
			
		||||
    ])
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
pub async fn cmd_age_encrypt_name(server: &Server, name: &str, message: &str) -> Protocol {
 | 
			
		||||
@@ -253,7 +300,9 @@ pub async fn cmd_age_decrypt_name(server: &Server, name: &str, ct_b64: &str) ->
 | 
			
		||||
pub async fn cmd_age_sign_name(server: &Server, name: &str, message: &str) -> Protocol {
 | 
			
		||||
    let sec = match sget(server, &sign_priv_key_key(name)) {
 | 
			
		||||
        Ok(Some(v)) => v,
 | 
			
		||||
        Ok(None) => return AgeWireError::NotFound("signing secret (age:signpriv:{name})").to_protocol(),
 | 
			
		||||
        Ok(None) => {
 | 
			
		||||
            return AgeWireError::NotFound("signing secret (age:signpriv:{name})").to_protocol()
 | 
			
		||||
        }
 | 
			
		||||
        Err(e) => return e.to_protocol(),
 | 
			
		||||
    };
 | 
			
		||||
    match sign_b64(&sec, message) {
 | 
			
		||||
@@ -262,10 +311,17 @@ pub async fn cmd_age_sign_name(server: &Server, name: &str, message: &str) -> Pr
 | 
			
		||||
    }
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
pub async fn cmd_age_verify_name(server: &Server, name: &str, message: &str, sig_b64: &str) -> Protocol {
 | 
			
		||||
pub async fn cmd_age_verify_name(
 | 
			
		||||
    server: &Server,
 | 
			
		||||
    name: &str,
 | 
			
		||||
    message: &str,
 | 
			
		||||
    sig_b64: &str,
 | 
			
		||||
) -> Protocol {
 | 
			
		||||
    let pubk = match sget(server, &sign_pub_key_key(name)) {
 | 
			
		||||
        Ok(Some(v)) => v,
 | 
			
		||||
        Ok(None) => return AgeWireError::NotFound("verify pubkey (age:signpub:{name})").to_protocol(),
 | 
			
		||||
        Ok(None) => {
 | 
			
		||||
            return AgeWireError::NotFound("verify pubkey (age:signpub:{name})").to_protocol()
 | 
			
		||||
        }
 | 
			
		||||
        Err(e) => return e.to_protocol(),
 | 
			
		||||
    };
 | 
			
		||||
    match verify_b64(&pubk, message, sig_b64) {
 | 
			
		||||
@@ -277,25 +333,43 @@ pub async fn cmd_age_verify_name(server: &Server, name: &str, message: &str, sig
 | 
			
		||||
 | 
			
		||||
pub async fn cmd_age_list(server: &Server) -> Protocol {
 | 
			
		||||
    // Returns 4 arrays: ["encpub", <names...>], ["encpriv", ...], ["signpub", ...], ["signpriv", ...]
 | 
			
		||||
    let st = match server.current_storage() { Ok(s) => s, Err(e) => return Protocol::err(&e.0) };
 | 
			
		||||
    let st = match server.current_storage() {
 | 
			
		||||
        Ok(s) => s,
 | 
			
		||||
        Err(e) => return Protocol::err(&e.0),
 | 
			
		||||
    };
 | 
			
		||||
 | 
			
		||||
    let pull = |pat: &str, prefix: &str| -> Result<Vec<String>, DBError> {
 | 
			
		||||
        let keys = st.keys(pat)?;
 | 
			
		||||
        let mut names: Vec<String> = keys.into_iter()
 | 
			
		||||
        let mut names: Vec<String> = keys
 | 
			
		||||
            .into_iter()
 | 
			
		||||
            .filter_map(|k| k.strip_prefix(prefix).map(|x| x.to_string()))
 | 
			
		||||
            .collect();
 | 
			
		||||
        names.sort();
 | 
			
		||||
        Ok(names)
 | 
			
		||||
    };
 | 
			
		||||
 | 
			
		||||
    let encpub  = match pull("age:key:*",      "age:key:")      { Ok(v) => v, Err(e)=> return Protocol::err(&e.0) };
 | 
			
		||||
    let encpriv = match pull("age:privkey:*",  "age:privkey:")  { Ok(v) => v, Err(e)=> return Protocol::err(&e.0) };
 | 
			
		||||
    let signpub = match pull("age:signpub:*",  "age:signpub:")  { Ok(v) => v, Err(e)=> return Protocol::err(&e.0) };
 | 
			
		||||
    let signpriv= match pull("age:signpriv:*", "age:signpriv:") { Ok(v) => v, Err(e)=> return Protocol::err(&e.0) };
 | 
			
		||||
    let encpub = match pull("age:key:*", "age:key:") {
 | 
			
		||||
        Ok(v) => v,
 | 
			
		||||
        Err(e) => return Protocol::err(&e.0),
 | 
			
		||||
    };
 | 
			
		||||
    let encpriv = match pull("age:privkey:*", "age:privkey:") {
 | 
			
		||||
        Ok(v) => v,
 | 
			
		||||
        Err(e) => return Protocol::err(&e.0),
 | 
			
		||||
    };
 | 
			
		||||
    let signpub = match pull("age:signpub:*", "age:signpub:") {
 | 
			
		||||
        Ok(v) => v,
 | 
			
		||||
        Err(e) => return Protocol::err(&e.0),
 | 
			
		||||
    };
 | 
			
		||||
    let signpriv = match pull("age:signpriv:*", "age:signpriv:") {
 | 
			
		||||
        Ok(v) => v,
 | 
			
		||||
        Err(e) => return Protocol::err(&e.0),
 | 
			
		||||
    };
 | 
			
		||||
 | 
			
		||||
    let to_arr = |label: &str, v: Vec<String>| {
 | 
			
		||||
        let mut out = vec![Protocol::BulkString(label.to_string())];
 | 
			
		||||
        out.push(Protocol::Array(v.into_iter().map(Protocol::BulkString).collect()));
 | 
			
		||||
        out.push(Protocol::Array(
 | 
			
		||||
            v.into_iter().map(Protocol::BulkString).collect(),
 | 
			
		||||
        ));
 | 
			
		||||
        Protocol::Array(out)
 | 
			
		||||
    };
 | 
			
		||||
 | 
			
		||||
 
 | 
			
		||||
							
								
								
									
										998
									
								
								src/cmd.rs
									
									
									
									
									
								
							
							
						
						
									
										998
									
								
								src/cmd.rs
									
									
									
									
									
								
							
										
											
												File diff suppressed because it is too large
												Load Diff
											
										
									
								
							@@ -11,9 +11,9 @@ const TAG_LEN: usize = 16;
 | 
			
		||||
 | 
			
		||||
#[derive(Debug)]
 | 
			
		||||
pub enum CryptoError {
 | 
			
		||||
    Format,         // wrong length / header
 | 
			
		||||
    Version(u8),    // unknown version
 | 
			
		||||
    Decrypt,        // wrong key or corrupted data
 | 
			
		||||
    Format,      // wrong length / header
 | 
			
		||||
    Version(u8), // unknown version
 | 
			
		||||
    Decrypt,     // wrong key or corrupted data
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
impl From<CryptoError> for crate::error::DBError {
 | 
			
		||||
 
 | 
			
		||||
@@ -1,9 +1,8 @@
 | 
			
		||||
use std::num::ParseIntError;
 | 
			
		||||
 | 
			
		||||
use tokio::sync::mpsc;
 | 
			
		||||
use redb;
 | 
			
		||||
use bincode;
 | 
			
		||||
 | 
			
		||||
use redb;
 | 
			
		||||
use tokio::sync::mpsc;
 | 
			
		||||
 | 
			
		||||
// todo: more error types
 | 
			
		||||
#[derive(Debug)]
 | 
			
		||||
 
 | 
			
		||||
@@ -1,10 +1,12 @@
 | 
			
		||||
pub mod age;   // NEW
 | 
			
		||||
pub mod age; // NEW
 | 
			
		||||
pub mod cmd;
 | 
			
		||||
pub mod crypto;
 | 
			
		||||
pub mod error;
 | 
			
		||||
pub mod options;
 | 
			
		||||
pub mod protocol;
 | 
			
		||||
pub mod search_cmd; // Add this
 | 
			
		||||
pub mod server;
 | 
			
		||||
pub mod storage;
 | 
			
		||||
pub mod storage_trait;  // Add this
 | 
			
		||||
pub mod storage_sled;   // Add this
 | 
			
		||||
pub mod storage_sled; // Add this
 | 
			
		||||
pub mod storage_trait; // Add this
 | 
			
		||||
pub mod tantivy_search;
 | 
			
		||||
 
 | 
			
		||||
@@ -22,7 +22,6 @@ struct Args {
 | 
			
		||||
    #[arg(long)]
 | 
			
		||||
    debug: bool,
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
    /// Master encryption key for encrypted databases
 | 
			
		||||
    #[arg(long)]
 | 
			
		||||
    encryption_key: Option<String>,
 | 
			
		||||
 
 | 
			
		||||
@@ -81,18 +81,21 @@ impl Protocol {
 | 
			
		||||
    pub fn encode(&self) -> String {
 | 
			
		||||
        match self {
 | 
			
		||||
            Protocol::SimpleString(s) => format!("+{}\r\n", s),
 | 
			
		||||
            Protocol::BulkString(s)   => format!("${}\r\n{}\r\n", s.len(), s),
 | 
			
		||||
            Protocol::Array(ss)       => {
 | 
			
		||||
            Protocol::BulkString(s) => format!("${}\r\n{}\r\n", s.len(), s),
 | 
			
		||||
            Protocol::Array(ss) => {
 | 
			
		||||
                format!("*{}\r\n", ss.len()) + &ss.iter().map(|x| x.encode()).collect::<String>()
 | 
			
		||||
            }
 | 
			
		||||
            Protocol::Null            => "$-1\r\n".to_string(),
 | 
			
		||||
            Protocol::Error(s)        => format!("-{}\r\n", s), // proper RESP error
 | 
			
		||||
            Protocol::Null => "$-1\r\n".to_string(),
 | 
			
		||||
            Protocol::Error(s) => format!("-{}\r\n", s), // proper RESP error
 | 
			
		||||
        }
 | 
			
		||||
    }
 | 
			
		||||
 | 
			
		||||
    fn parse_simple_string_sfx(protocol: &str) -> Result<(Self, &str), DBError> {
 | 
			
		||||
        match protocol.find("\r\n") {
 | 
			
		||||
            Some(x) => Ok((Self::SimpleString(protocol[..x].to_string()), &protocol[x + 2..])),
 | 
			
		||||
            Some(x) => Ok((
 | 
			
		||||
                Self::SimpleString(protocol[..x].to_string()),
 | 
			
		||||
                &protocol[x + 2..],
 | 
			
		||||
            )),
 | 
			
		||||
            _ => Err(DBError(format!(
 | 
			
		||||
                "[new simple string] unsupported protocol: {:?}",
 | 
			
		||||
                protocol
 | 
			
		||||
 
 | 
			
		||||
							
								
								
									
										273
									
								
								src/search_cmd.rs
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										273
									
								
								src/search_cmd.rs
									
									
									
									
									
										Normal file
									
								
							@@ -0,0 +1,273 @@
 | 
			
		||||
use crate::{
 | 
			
		||||
    error::DBError,
 | 
			
		||||
    protocol::Protocol,
 | 
			
		||||
    server::Server,
 | 
			
		||||
    tantivy_search::{
 | 
			
		||||
        FieldDef, Filter, FilterType, IndexConfig, NumericType, SearchOptions, TantivySearch,
 | 
			
		||||
    },
 | 
			
		||||
};
 | 
			
		||||
use std::collections::HashMap;
 | 
			
		||||
use std::sync::Arc;
 | 
			
		||||
 | 
			
		||||
pub async fn ft_create_cmd(
 | 
			
		||||
    server: &Server,
 | 
			
		||||
    index_name: String,
 | 
			
		||||
    schema: Vec<(String, String, Vec<String>)>,
 | 
			
		||||
) -> Result<Protocol, DBError> {
 | 
			
		||||
    // Parse schema into field definitions
 | 
			
		||||
    let mut field_definitions = Vec::new();
 | 
			
		||||
 | 
			
		||||
    for (field_name, field_type, options) in schema {
 | 
			
		||||
        let field_def = match field_type.to_uppercase().as_str() {
 | 
			
		||||
            "TEXT" => {
 | 
			
		||||
                let mut weight = 1.0;
 | 
			
		||||
                let mut sortable = false;
 | 
			
		||||
                let mut no_index = false;
 | 
			
		||||
 | 
			
		||||
                for opt in &options {
 | 
			
		||||
                    match opt.to_uppercase().as_str() {
 | 
			
		||||
                        "WEIGHT" => {
 | 
			
		||||
                            // Next option should be the weight value
 | 
			
		||||
                            if let Some(idx) = options.iter().position(|x| x == opt) {
 | 
			
		||||
                                if idx + 1 < options.len() {
 | 
			
		||||
                                    weight = options[idx + 1].parse().unwrap_or(1.0);
 | 
			
		||||
                                }
 | 
			
		||||
                            }
 | 
			
		||||
                        }
 | 
			
		||||
                        "SORTABLE" => sortable = true,
 | 
			
		||||
                        "NOINDEX" => no_index = true,
 | 
			
		||||
                        _ => {}
 | 
			
		||||
                    }
 | 
			
		||||
                }
 | 
			
		||||
 | 
			
		||||
                FieldDef::Text {
 | 
			
		||||
                    stored: true,
 | 
			
		||||
                    indexed: !no_index,
 | 
			
		||||
                    tokenized: true,
 | 
			
		||||
                    fast: sortable,
 | 
			
		||||
                }
 | 
			
		||||
            }
 | 
			
		||||
            "NUMERIC" => {
 | 
			
		||||
                let mut sortable = false;
 | 
			
		||||
 | 
			
		||||
                for opt in &options {
 | 
			
		||||
                    if opt.to_uppercase() == "SORTABLE" {
 | 
			
		||||
                        sortable = true;
 | 
			
		||||
                    }
 | 
			
		||||
                }
 | 
			
		||||
 | 
			
		||||
                FieldDef::Numeric {
 | 
			
		||||
                    stored: true,
 | 
			
		||||
                    indexed: true,
 | 
			
		||||
                    fast: sortable,
 | 
			
		||||
                    precision: NumericType::F64,
 | 
			
		||||
                }
 | 
			
		||||
            }
 | 
			
		||||
            "TAG" => {
 | 
			
		||||
                let mut separator = ",".to_string();
 | 
			
		||||
                let mut case_sensitive = false;
 | 
			
		||||
 | 
			
		||||
                for i in 0..options.len() {
 | 
			
		||||
                    match options[i].to_uppercase().as_str() {
 | 
			
		||||
                        "SEPARATOR" => {
 | 
			
		||||
                            if i + 1 < options.len() {
 | 
			
		||||
                                separator = options[i + 1].clone();
 | 
			
		||||
                            }
 | 
			
		||||
                        }
 | 
			
		||||
                        "CASESENSITIVE" => case_sensitive = true,
 | 
			
		||||
                        _ => {}
 | 
			
		||||
                    }
 | 
			
		||||
                }
 | 
			
		||||
 | 
			
		||||
                FieldDef::Tag {
 | 
			
		||||
                    stored: true,
 | 
			
		||||
                    separator,
 | 
			
		||||
                    case_sensitive,
 | 
			
		||||
                }
 | 
			
		||||
            }
 | 
			
		||||
            "GEO" => FieldDef::Geo { stored: true },
 | 
			
		||||
            _ => {
 | 
			
		||||
                return Err(DBError(format!("Unknown field type: {}", field_type)));
 | 
			
		||||
            }
 | 
			
		||||
        };
 | 
			
		||||
 | 
			
		||||
        field_definitions.push((field_name, field_def));
 | 
			
		||||
    }
 | 
			
		||||
 | 
			
		||||
    // Create the search index
 | 
			
		||||
    let search_path = server.search_index_path();
 | 
			
		||||
    let config = IndexConfig::default();
 | 
			
		||||
 | 
			
		||||
    println!(
 | 
			
		||||
        "Creating search index '{}' at path: {:?}",
 | 
			
		||||
        index_name, search_path
 | 
			
		||||
    );
 | 
			
		||||
    println!("Field definitions: {:?}", field_definitions);
 | 
			
		||||
 | 
			
		||||
    let search_index = TantivySearch::new_with_schema(
 | 
			
		||||
        search_path,
 | 
			
		||||
        index_name.clone(),
 | 
			
		||||
        field_definitions,
 | 
			
		||||
        Some(config),
 | 
			
		||||
    )?;
 | 
			
		||||
 | 
			
		||||
    println!("Search index '{}' created successfully", index_name);
 | 
			
		||||
 | 
			
		||||
    // Store in registry
 | 
			
		||||
    let mut indexes = server.search_indexes.write().unwrap();
 | 
			
		||||
    indexes.insert(index_name, Arc::new(search_index));
 | 
			
		||||
 | 
			
		||||
    Ok(Protocol::SimpleString("OK".to_string()))
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
pub async fn ft_add_cmd(
 | 
			
		||||
    server: &Server,
 | 
			
		||||
    index_name: String,
 | 
			
		||||
    doc_id: String,
 | 
			
		||||
    _score: f64,
 | 
			
		||||
    fields: HashMap<String, String>,
 | 
			
		||||
) -> Result<Protocol, DBError> {
 | 
			
		||||
    let indexes = server.search_indexes.read().unwrap();
 | 
			
		||||
 | 
			
		||||
    let search_index = indexes
 | 
			
		||||
        .get(&index_name)
 | 
			
		||||
        .ok_or_else(|| DBError(format!("Index '{}' not found", index_name)))?;
 | 
			
		||||
 | 
			
		||||
    search_index.add_document_with_fields(&doc_id, fields)?;
 | 
			
		||||
 | 
			
		||||
    Ok(Protocol::SimpleString("OK".to_string()))
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
pub async fn ft_search_cmd(
 | 
			
		||||
    server: &Server,
 | 
			
		||||
    index_name: String,
 | 
			
		||||
    query: String,
 | 
			
		||||
    filters: Vec<(String, String)>,
 | 
			
		||||
    limit: Option<usize>,
 | 
			
		||||
    offset: Option<usize>,
 | 
			
		||||
    return_fields: Option<Vec<String>>,
 | 
			
		||||
) -> Result<Protocol, DBError> {
 | 
			
		||||
    let indexes = server.search_indexes.read().unwrap();
 | 
			
		||||
 | 
			
		||||
    let search_index = indexes
 | 
			
		||||
        .get(&index_name)
 | 
			
		||||
        .ok_or_else(|| DBError(format!("Index '{}' not found", index_name)))?;
 | 
			
		||||
 | 
			
		||||
    // Convert filters to search filters
 | 
			
		||||
    let search_filters = filters
 | 
			
		||||
        .into_iter()
 | 
			
		||||
        .map(|(field, value)| Filter {
 | 
			
		||||
            field,
 | 
			
		||||
            filter_type: FilterType::Equals(value),
 | 
			
		||||
        })
 | 
			
		||||
        .collect();
 | 
			
		||||
 | 
			
		||||
    let options = SearchOptions {
 | 
			
		||||
        limit: limit.unwrap_or(10),
 | 
			
		||||
        offset: offset.unwrap_or(0),
 | 
			
		||||
        filters: search_filters,
 | 
			
		||||
        sort_by: None,
 | 
			
		||||
        return_fields,
 | 
			
		||||
        highlight: false,
 | 
			
		||||
    };
 | 
			
		||||
 | 
			
		||||
    let results = search_index.search_with_options(&query, options)?;
 | 
			
		||||
 | 
			
		||||
    // Format results as Redis protocol
 | 
			
		||||
    let mut response = Vec::new();
 | 
			
		||||
 | 
			
		||||
    // First element is the total count
 | 
			
		||||
    response.push(Protocol::SimpleString(results.total.to_string()));
 | 
			
		||||
 | 
			
		||||
    // Then each document
 | 
			
		||||
    for doc in results.documents {
 | 
			
		||||
        let mut doc_array = Vec::new();
 | 
			
		||||
 | 
			
		||||
        // Add document ID if it exists
 | 
			
		||||
        if let Some(id) = doc.fields.get("_id") {
 | 
			
		||||
            doc_array.push(Protocol::BulkString(id.clone()));
 | 
			
		||||
        }
 | 
			
		||||
 | 
			
		||||
        // Add score
 | 
			
		||||
        doc_array.push(Protocol::BulkString(doc.score.to_string()));
 | 
			
		||||
 | 
			
		||||
        // Add fields as key-value pairs
 | 
			
		||||
        for (field_name, field_value) in doc.fields {
 | 
			
		||||
            if field_name != "_id" {
 | 
			
		||||
                doc_array.push(Protocol::BulkString(field_name));
 | 
			
		||||
                doc_array.push(Protocol::BulkString(field_value));
 | 
			
		||||
            }
 | 
			
		||||
        }
 | 
			
		||||
 | 
			
		||||
        response.push(Protocol::Array(doc_array));
 | 
			
		||||
    }
 | 
			
		||||
 | 
			
		||||
    Ok(Protocol::Array(response))
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
pub async fn ft_del_cmd(
 | 
			
		||||
    server: &Server,
 | 
			
		||||
    index_name: String,
 | 
			
		||||
    doc_id: String,
 | 
			
		||||
) -> Result<Protocol, DBError> {
 | 
			
		||||
    let indexes = server.search_indexes.read().unwrap();
 | 
			
		||||
 | 
			
		||||
    let _search_index = indexes
 | 
			
		||||
        .get(&index_name)
 | 
			
		||||
        .ok_or_else(|| DBError(format!("Index '{}' not found", index_name)))?;
 | 
			
		||||
 | 
			
		||||
    // For now, return success
 | 
			
		||||
    // In a full implementation, we'd need to add a delete method to TantivySearch
 | 
			
		||||
    println!("Deleting document '{}' from index '{}'", doc_id, index_name);
 | 
			
		||||
 | 
			
		||||
    Ok(Protocol::SimpleString("1".to_string()))
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
pub async fn ft_info_cmd(server: &Server, index_name: String) -> Result<Protocol, DBError> {
 | 
			
		||||
    let indexes = server.search_indexes.read().unwrap();
 | 
			
		||||
 | 
			
		||||
    let search_index = indexes
 | 
			
		||||
        .get(&index_name)
 | 
			
		||||
        .ok_or_else(|| DBError(format!("Index '{}' not found", index_name)))?;
 | 
			
		||||
 | 
			
		||||
    let info = search_index.get_info()?;
 | 
			
		||||
 | 
			
		||||
    // Format info as Redis protocol
 | 
			
		||||
    let mut response = Vec::new();
 | 
			
		||||
 | 
			
		||||
    response.push(Protocol::BulkString("index_name".to_string()));
 | 
			
		||||
    response.push(Protocol::BulkString(info.name));
 | 
			
		||||
 | 
			
		||||
    response.push(Protocol::BulkString("num_docs".to_string()));
 | 
			
		||||
    response.push(Protocol::BulkString(info.num_docs.to_string()));
 | 
			
		||||
 | 
			
		||||
    response.push(Protocol::BulkString("num_fields".to_string()));
 | 
			
		||||
    response.push(Protocol::BulkString(info.fields.len().to_string()));
 | 
			
		||||
 | 
			
		||||
    response.push(Protocol::BulkString("fields".to_string()));
 | 
			
		||||
    let fields_str = info
 | 
			
		||||
        .fields
 | 
			
		||||
        .iter()
 | 
			
		||||
        .map(|f| format!("{}:{}", f.name, f.field_type))
 | 
			
		||||
        .collect::<Vec<_>>()
 | 
			
		||||
        .join(", ");
 | 
			
		||||
    response.push(Protocol::BulkString(fields_str));
 | 
			
		||||
 | 
			
		||||
    Ok(Protocol::Array(response))
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
pub async fn ft_drop_cmd(server: &Server, index_name: String) -> Result<Protocol, DBError> {
 | 
			
		||||
    let mut indexes = server.search_indexes.write().unwrap();
 | 
			
		||||
 | 
			
		||||
    if indexes.remove(&index_name).is_some() {
 | 
			
		||||
        // Also remove the index files from disk
 | 
			
		||||
        let index_path = server.search_index_path().join(&index_name);
 | 
			
		||||
        if index_path.exists() {
 | 
			
		||||
            std::fs::remove_dir_all(index_path)
 | 
			
		||||
                .map_err(|e| DBError(format!("Failed to remove index files: {}", e)))?;
 | 
			
		||||
        }
 | 
			
		||||
        Ok(Protocol::SimpleString("OK".to_string()))
 | 
			
		||||
    } else {
 | 
			
		||||
        Err(DBError(format!("Index '{}' not found", index_name)))
 | 
			
		||||
    }
 | 
			
		||||
}
 | 
			
		||||
@@ -1,9 +1,10 @@
 | 
			
		||||
use core::str;
 | 
			
		||||
use std::collections::HashMap;
 | 
			
		||||
use std::sync::Arc;
 | 
			
		||||
use std::sync::RwLock;
 | 
			
		||||
use tokio::io::AsyncReadExt;
 | 
			
		||||
use tokio::io::AsyncWriteExt;
 | 
			
		||||
use tokio::sync::{Mutex, oneshot};
 | 
			
		||||
use tokio::sync::{oneshot, Mutex};
 | 
			
		||||
 | 
			
		||||
use std::sync::atomic::{AtomicU64, Ordering};
 | 
			
		||||
 | 
			
		||||
@@ -14,10 +15,12 @@ use crate::protocol::Protocol;
 | 
			
		||||
use crate::storage::Storage;
 | 
			
		||||
use crate::storage_sled::SledStorage;
 | 
			
		||||
use crate::storage_trait::StorageBackend;
 | 
			
		||||
use crate::tantivy_search::TantivySearch;
 | 
			
		||||
 | 
			
		||||
#[derive(Clone)]
 | 
			
		||||
pub struct Server {
 | 
			
		||||
    pub db_cache: std::sync::Arc<std::sync::RwLock<HashMap<u64, Arc<dyn StorageBackend>>>>,
 | 
			
		||||
    pub db_cache: Arc<RwLock<HashMap<u64, Arc<dyn StorageBackend>>>>,
 | 
			
		||||
    pub search_indexes: Arc<RwLock<HashMap<String, Arc<TantivySearch>>>>,
 | 
			
		||||
    pub option: options::DBOption,
 | 
			
		||||
    pub client_name: Option<String>,
 | 
			
		||||
    pub selected_db: u64, // Changed from usize to u64
 | 
			
		||||
@@ -43,7 +46,8 @@ pub enum PopSide {
 | 
			
		||||
impl Server {
 | 
			
		||||
    pub async fn new(option: options::DBOption) -> Self {
 | 
			
		||||
        Server {
 | 
			
		||||
            db_cache: Arc::new(std::sync::RwLock::new(HashMap::new())),
 | 
			
		||||
            db_cache: Arc::new(RwLock::new(HashMap::new())),
 | 
			
		||||
            search_indexes: Arc::new(RwLock::new(HashMap::new())),
 | 
			
		||||
            option,
 | 
			
		||||
            client_name: None,
 | 
			
		||||
            selected_db: 0,
 | 
			
		||||
@@ -61,7 +65,6 @@ impl Server {
 | 
			
		||||
            return Ok(storage.clone());
 | 
			
		||||
        }
 | 
			
		||||
 | 
			
		||||
        
 | 
			
		||||
        // Create new database file
 | 
			
		||||
        let db_file_path = std::path::PathBuf::from(self.option.dir.clone())
 | 
			
		||||
            .join(format!("{}.db", self.selected_db));
 | 
			
		||||
@@ -69,27 +72,27 @@ impl Server {
 | 
			
		||||
        // Ensure the directory exists before creating the database file
 | 
			
		||||
        if let Some(parent_dir) = db_file_path.parent() {
 | 
			
		||||
            std::fs::create_dir_all(parent_dir).map_err(|e| {
 | 
			
		||||
                DBError(format!("Failed to create directory {}: {}", parent_dir.display(), e))
 | 
			
		||||
                DBError(format!(
 | 
			
		||||
                    "Failed to create directory {}: {}",
 | 
			
		||||
                    parent_dir.display(),
 | 
			
		||||
                    e
 | 
			
		||||
                ))
 | 
			
		||||
            })?;
 | 
			
		||||
        }
 | 
			
		||||
 | 
			
		||||
        println!("Creating new db file: {}", db_file_path.display());
 | 
			
		||||
 | 
			
		||||
        let storage: Arc<dyn StorageBackend> = match self.option.backend {
 | 
			
		||||
            options::BackendType::Redb => {
 | 
			
		||||
                Arc::new(Storage::new(
 | 
			
		||||
                    db_file_path,
 | 
			
		||||
                    self.should_encrypt_db(self.selected_db),
 | 
			
		||||
                    self.option.encryption_key.as_deref()
 | 
			
		||||
                )?)
 | 
			
		||||
            }
 | 
			
		||||
            options::BackendType::Sled => {
 | 
			
		||||
                Arc::new(SledStorage::new(
 | 
			
		||||
                    db_file_path,
 | 
			
		||||
                    self.should_encrypt_db(self.selected_db),
 | 
			
		||||
                    self.option.encryption_key.as_deref()
 | 
			
		||||
                )?)
 | 
			
		||||
            }
 | 
			
		||||
            options::BackendType::Redb => Arc::new(Storage::new(
 | 
			
		||||
                db_file_path,
 | 
			
		||||
                self.should_encrypt_db(self.selected_db),
 | 
			
		||||
                self.option.encryption_key.as_deref(),
 | 
			
		||||
            )?),
 | 
			
		||||
            options::BackendType::Sled => Arc::new(SledStorage::new(
 | 
			
		||||
                db_file_path,
 | 
			
		||||
                self.should_encrypt_db(self.selected_db),
 | 
			
		||||
                self.option.encryption_key.as_deref(),
 | 
			
		||||
            )?),
 | 
			
		||||
        };
 | 
			
		||||
 | 
			
		||||
        cache.insert(self.selected_db, storage.clone());
 | 
			
		||||
@@ -101,9 +104,19 @@ impl Server {
 | 
			
		||||
        self.option.encrypt && db_index >= 10
 | 
			
		||||
    }
 | 
			
		||||
 | 
			
		||||
    // Add method to get search index path
 | 
			
		||||
    pub fn search_index_path(&self) -> std::path::PathBuf {
 | 
			
		||||
        std::path::PathBuf::from(&self.option.dir).join("search_indexes")
 | 
			
		||||
    }
 | 
			
		||||
 | 
			
		||||
    // ----- BLPOP waiter helpers -----
 | 
			
		||||
 | 
			
		||||
    pub async fn register_waiter(&self, db_index: u64, key: &str, side: PopSide) -> (u64, oneshot::Receiver<(String, String)>) {
 | 
			
		||||
    pub async fn register_waiter(
 | 
			
		||||
        &self,
 | 
			
		||||
        db_index: u64,
 | 
			
		||||
        key: &str,
 | 
			
		||||
        side: PopSide,
 | 
			
		||||
    ) -> (u64, oneshot::Receiver<(String, String)>) {
 | 
			
		||||
        let id = self.waiter_seq.fetch_add(1, Ordering::Relaxed);
 | 
			
		||||
        let (tx, rx) = oneshot::channel::<(String, String)>();
 | 
			
		||||
 | 
			
		||||
@@ -179,10 +192,7 @@ impl Server {
 | 
			
		||||
        Ok(())
 | 
			
		||||
    }
 | 
			
		||||
 | 
			
		||||
    pub async fn handle(
 | 
			
		||||
        &mut self,
 | 
			
		||||
        mut stream: tokio::net::TcpStream,
 | 
			
		||||
    ) -> Result<(), DBError> {
 | 
			
		||||
    pub async fn handle(&mut self, mut stream: tokio::net::TcpStream) -> Result<(), DBError> {
 | 
			
		||||
        // Accumulate incoming bytes to handle partial RESP frames
 | 
			
		||||
        let mut acc = String::new();
 | 
			
		||||
        let mut buf = vec![0u8; 8192];
 | 
			
		||||
@@ -219,7 +229,10 @@ impl Server {
 | 
			
		||||
                acc = remaining.to_string();
 | 
			
		||||
 | 
			
		||||
                if self.option.debug {
 | 
			
		||||
                    println!("\x1b[34;1mgot command: {:?}, protocol: {:?}\x1b[0m", cmd, protocol);
 | 
			
		||||
                    println!(
 | 
			
		||||
                        "\x1b[34;1mgot command: {:?}, protocol: {:?}\x1b[0m",
 | 
			
		||||
                        cmd, protocol
 | 
			
		||||
                    );
 | 
			
		||||
                } else {
 | 
			
		||||
                    println!("got command: {:?}, protocol: {:?}", cmd, protocol);
 | 
			
		||||
                }
 | 
			
		||||
 
 | 
			
		||||
@@ -12,9 +12,9 @@ use crate::error::DBError;
 | 
			
		||||
 | 
			
		||||
// Re-export modules
 | 
			
		||||
mod storage_basic;
 | 
			
		||||
mod storage_extra;
 | 
			
		||||
mod storage_hset;
 | 
			
		||||
mod storage_lists;
 | 
			
		||||
mod storage_extra;
 | 
			
		||||
 | 
			
		||||
// Re-export implementations
 | 
			
		||||
// Note: These imports are used by the impl blocks in the submodules
 | 
			
		||||
@@ -28,7 +28,8 @@ const STRINGS_TABLE: TableDefinition<&str, &[u8]> = TableDefinition::new("string
 | 
			
		||||
const HASHES_TABLE: TableDefinition<(&str, &str), &[u8]> = TableDefinition::new("hashes");
 | 
			
		||||
const LISTS_TABLE: TableDefinition<&str, &[u8]> = TableDefinition::new("lists");
 | 
			
		||||
const STREAMS_META_TABLE: TableDefinition<&str, &[u8]> = TableDefinition::new("streams_meta");
 | 
			
		||||
const STREAMS_DATA_TABLE: TableDefinition<(&str, &str), &[u8]> = TableDefinition::new("streams_data");
 | 
			
		||||
const STREAMS_DATA_TABLE: TableDefinition<(&str, &str), &[u8]> =
 | 
			
		||||
    TableDefinition::new("streams_data");
 | 
			
		||||
const ENCRYPTED_TABLE: TableDefinition<&str, u8> = TableDefinition::new("encrypted");
 | 
			
		||||
const EXPIRATION_TABLE: TableDefinition<&str, u64> = TableDefinition::new("expiration");
 | 
			
		||||
 | 
			
		||||
@@ -55,7 +56,11 @@ pub struct Storage {
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
impl Storage {
 | 
			
		||||
    pub fn new(path: impl AsRef<Path>, should_encrypt: bool, master_key: Option<&str>) -> Result<Self, DBError> {
 | 
			
		||||
    pub fn new(
 | 
			
		||||
        path: impl AsRef<Path>,
 | 
			
		||||
        should_encrypt: bool,
 | 
			
		||||
        master_key: Option<&str>,
 | 
			
		||||
    ) -> Result<Self, DBError> {
 | 
			
		||||
        let db = Database::create(path)?;
 | 
			
		||||
 | 
			
		||||
        // Create tables if they don't exist
 | 
			
		||||
@@ -75,14 +80,19 @@ impl Storage {
 | 
			
		||||
        // Check if database was previously encrypted
 | 
			
		||||
        let read_txn = db.begin_read()?;
 | 
			
		||||
        let encrypted_table = read_txn.open_table(ENCRYPTED_TABLE)?;
 | 
			
		||||
        let was_encrypted = encrypted_table.get("encrypted")?.map(|v| v.value() == 1).unwrap_or(false);
 | 
			
		||||
        let was_encrypted = encrypted_table
 | 
			
		||||
            .get("encrypted")?
 | 
			
		||||
            .map(|v| v.value() == 1)
 | 
			
		||||
            .unwrap_or(false);
 | 
			
		||||
        drop(read_txn);
 | 
			
		||||
 | 
			
		||||
        let crypto = if should_encrypt || was_encrypted {
 | 
			
		||||
            if let Some(key) = master_key {
 | 
			
		||||
                Some(CryptoFactory::new(key.as_bytes()))
 | 
			
		||||
            } else {
 | 
			
		||||
                return Err(DBError("Encryption requested but no master key provided".to_string()));
 | 
			
		||||
                return Err(DBError(
 | 
			
		||||
                    "Encryption requested but no master key provided".to_string(),
 | 
			
		||||
                ));
 | 
			
		||||
            }
 | 
			
		||||
        } else {
 | 
			
		||||
            None
 | 
			
		||||
@@ -98,10 +108,7 @@ impl Storage {
 | 
			
		||||
            write_txn.commit()?;
 | 
			
		||||
        }
 | 
			
		||||
 | 
			
		||||
        Ok(Storage {
 | 
			
		||||
            db,
 | 
			
		||||
            crypto,
 | 
			
		||||
        })
 | 
			
		||||
        Ok(Storage { db, crypto })
 | 
			
		||||
    }
 | 
			
		||||
 | 
			
		||||
    pub fn is_encrypted(&self) -> bool {
 | 
			
		||||
@@ -165,11 +172,22 @@ impl StorageBackend for Storage {
 | 
			
		||||
        self.get_key_type(key)
 | 
			
		||||
    }
 | 
			
		||||
 | 
			
		||||
    fn scan(&self, cursor: u64, pattern: Option<&str>, count: Option<u64>) -> Result<(u64, Vec<(String, String)>), DBError> {
 | 
			
		||||
    fn scan(
 | 
			
		||||
        &self,
 | 
			
		||||
        cursor: u64,
 | 
			
		||||
        pattern: Option<&str>,
 | 
			
		||||
        count: Option<u64>,
 | 
			
		||||
    ) -> Result<(u64, Vec<(String, String)>), DBError> {
 | 
			
		||||
        self.scan(cursor, pattern, count)
 | 
			
		||||
    }
 | 
			
		||||
 | 
			
		||||
    fn hscan(&self, key: &str, cursor: u64, pattern: Option<&str>, count: Option<u64>) -> Result<(u64, Vec<(String, String)>), DBError> {
 | 
			
		||||
    fn hscan(
 | 
			
		||||
        &self,
 | 
			
		||||
        key: &str,
 | 
			
		||||
        cursor: u64,
 | 
			
		||||
        pattern: Option<&str>,
 | 
			
		||||
        count: Option<u64>,
 | 
			
		||||
    ) -> Result<(u64, Vec<(String, String)>), DBError> {
 | 
			
		||||
        self.hscan(key, cursor, pattern, count)
 | 
			
		||||
    }
 | 
			
		||||
 | 
			
		||||
 
 | 
			
		||||
@@ -1,6 +1,6 @@
 | 
			
		||||
use redb::{ReadableTable};
 | 
			
		||||
use crate::error::DBError;
 | 
			
		||||
use super::*;
 | 
			
		||||
use crate::error::DBError;
 | 
			
		||||
use redb::ReadableTable;
 | 
			
		||||
 | 
			
		||||
impl Storage {
 | 
			
		||||
    pub fn flushdb(&self) -> Result<(), DBError> {
 | 
			
		||||
@@ -15,11 +15,17 @@ impl Storage {
 | 
			
		||||
            let mut expiration_table = write_txn.open_table(EXPIRATION_TABLE)?;
 | 
			
		||||
 | 
			
		||||
            // inefficient, but there is no other way
 | 
			
		||||
            let keys: Vec<String> = types_table.iter()?.map(|item| item.unwrap().0.value().to_string()).collect();
 | 
			
		||||
            let keys: Vec<String> = types_table
 | 
			
		||||
                .iter()?
 | 
			
		||||
                .map(|item| item.unwrap().0.value().to_string())
 | 
			
		||||
                .collect();
 | 
			
		||||
            for key in keys {
 | 
			
		||||
                types_table.remove(key.as_str())?;
 | 
			
		||||
            }
 | 
			
		||||
            let keys: Vec<String> = strings_table.iter()?.map(|item| item.unwrap().0.value().to_string()).collect();
 | 
			
		||||
            let keys: Vec<String> = strings_table
 | 
			
		||||
                .iter()?
 | 
			
		||||
                .map(|item| item.unwrap().0.value().to_string())
 | 
			
		||||
                .collect();
 | 
			
		||||
            for key in keys {
 | 
			
		||||
                strings_table.remove(key.as_str())?;
 | 
			
		||||
            }
 | 
			
		||||
@@ -34,23 +40,35 @@ impl Storage {
 | 
			
		||||
            for (key, field) in keys {
 | 
			
		||||
                hashes_table.remove((key.as_str(), field.as_str()))?;
 | 
			
		||||
            }
 | 
			
		||||
            let keys: Vec<String> = lists_table.iter()?.map(|item| item.unwrap().0.value().to_string()).collect();
 | 
			
		||||
            let keys: Vec<String> = lists_table
 | 
			
		||||
                .iter()?
 | 
			
		||||
                .map(|item| item.unwrap().0.value().to_string())
 | 
			
		||||
                .collect();
 | 
			
		||||
            for key in keys {
 | 
			
		||||
                lists_table.remove(key.as_str())?;
 | 
			
		||||
            }
 | 
			
		||||
            let keys: Vec<String> = streams_meta_table.iter()?.map(|item| item.unwrap().0.value().to_string()).collect();
 | 
			
		||||
            let keys: Vec<String> = streams_meta_table
 | 
			
		||||
                .iter()?
 | 
			
		||||
                .map(|item| item.unwrap().0.value().to_string())
 | 
			
		||||
                .collect();
 | 
			
		||||
            for key in keys {
 | 
			
		||||
                streams_meta_table.remove(key.as_str())?;
 | 
			
		||||
            }
 | 
			
		||||
            let keys: Vec<(String,String)> = streams_data_table.iter()?.map(|item| {
 | 
			
		||||
                let binding = item.unwrap();
 | 
			
		||||
                let (key, field) = binding.0.value();
 | 
			
		||||
                (key.to_string(), field.to_string())
 | 
			
		||||
            }).collect();
 | 
			
		||||
            let keys: Vec<(String, String)> = streams_data_table
 | 
			
		||||
                .iter()?
 | 
			
		||||
                .map(|item| {
 | 
			
		||||
                    let binding = item.unwrap();
 | 
			
		||||
                    let (key, field) = binding.0.value();
 | 
			
		||||
                    (key.to_string(), field.to_string())
 | 
			
		||||
                })
 | 
			
		||||
                .collect();
 | 
			
		||||
            for (key, field) in keys {
 | 
			
		||||
                streams_data_table.remove((key.as_str(), field.as_str()))?;
 | 
			
		||||
            }
 | 
			
		||||
            let keys: Vec<String> = expiration_table.iter()?.map(|item| item.unwrap().0.value().to_string()).collect();
 | 
			
		||||
            let keys: Vec<String> = expiration_table
 | 
			
		||||
                .iter()?
 | 
			
		||||
                .map(|item| item.unwrap().0.value().to_string())
 | 
			
		||||
                .collect();
 | 
			
		||||
            for key in keys {
 | 
			
		||||
                expiration_table.remove(key.as_str())?;
 | 
			
		||||
            }
 | 
			
		||||
@@ -163,7 +181,8 @@ impl Storage {
 | 
			
		||||
        {
 | 
			
		||||
            let mut types_table = write_txn.open_table(TYPES_TABLE)?;
 | 
			
		||||
            let mut strings_table = write_txn.open_table(STRINGS_TABLE)?;
 | 
			
		||||
            let mut hashes_table: redb::Table<(&str, &str), &[u8]> = write_txn.open_table(HASHES_TABLE)?;
 | 
			
		||||
            let mut hashes_table: redb::Table<(&str, &str), &[u8]> =
 | 
			
		||||
                write_txn.open_table(HASHES_TABLE)?;
 | 
			
		||||
            let mut lists_table = write_txn.open_table(LISTS_TABLE)?;
 | 
			
		||||
 | 
			
		||||
            // Remove from type table
 | 
			
		||||
 
 | 
			
		||||
@@ -1,10 +1,15 @@
 | 
			
		||||
use redb::{ReadableTable};
 | 
			
		||||
use crate::error::DBError;
 | 
			
		||||
use super::*;
 | 
			
		||||
use crate::error::DBError;
 | 
			
		||||
use redb::ReadableTable;
 | 
			
		||||
 | 
			
		||||
impl Storage {
 | 
			
		||||
    // ✅ ENCRYPTION APPLIED: Values are decrypted after retrieval
 | 
			
		||||
    pub fn scan(&self, cursor: u64, pattern: Option<&str>, count: Option<u64>) -> Result<(u64, Vec<(String, String)>), DBError> {
 | 
			
		||||
    pub fn scan(
 | 
			
		||||
        &self,
 | 
			
		||||
        cursor: u64,
 | 
			
		||||
        pattern: Option<&str>,
 | 
			
		||||
        count: Option<u64>,
 | 
			
		||||
    ) -> Result<(u64, Vec<(String, String)>), DBError> {
 | 
			
		||||
        let read_txn = self.db.begin_read()?;
 | 
			
		||||
        let types_table = read_txn.open_table(TYPES_TABLE)?;
 | 
			
		||||
        let strings_table = read_txn.open_table(STRINGS_TABLE)?;
 | 
			
		||||
@@ -50,7 +55,11 @@ impl Storage {
 | 
			
		||||
            current_cursor += 1;
 | 
			
		||||
        }
 | 
			
		||||
 | 
			
		||||
        let next_cursor = if result.len() < limit { 0 } else { current_cursor };
 | 
			
		||||
        let next_cursor = if result.len() < limit {
 | 
			
		||||
            0
 | 
			
		||||
        } else {
 | 
			
		||||
            current_cursor
 | 
			
		||||
        };
 | 
			
		||||
        Ok((next_cursor, result))
 | 
			
		||||
    }
 | 
			
		||||
 | 
			
		||||
@@ -75,7 +84,7 @@ impl Storage {
 | 
			
		||||
                }
 | 
			
		||||
            }
 | 
			
		||||
            Some(_) => Ok(-1), // Key exists but is not a string (no expiration support for other types)
 | 
			
		||||
            None => Ok(-2), // Key does not exist
 | 
			
		||||
            None => Ok(-2),    // Key does not exist
 | 
			
		||||
        }
 | 
			
		||||
    }
 | 
			
		||||
 | 
			
		||||
@@ -95,7 +104,7 @@ impl Storage {
 | 
			
		||||
                Ok(true)
 | 
			
		||||
            }
 | 
			
		||||
            Some(_) => Ok(true), // Key exists and is not a string
 | 
			
		||||
            None => Ok(false), // Key does not exist
 | 
			
		||||
            None => Ok(false),   // Key does not exist
 | 
			
		||||
        }
 | 
			
		||||
    }
 | 
			
		||||
 | 
			
		||||
@@ -178,8 +187,12 @@ impl Storage {
 | 
			
		||||
                .unwrap_or(false);
 | 
			
		||||
            if is_string {
 | 
			
		||||
                let mut expiration_table = write_txn.open_table(EXPIRATION_TABLE)?;
 | 
			
		||||
                let expires_at_ms: u128 = if ts_secs <= 0 { 0 } else { (ts_secs as u128) * 1000 };
 | 
			
		||||
                expiration_table.insert(key, &((expires_at_ms as u64)))?;
 | 
			
		||||
                let expires_at_ms: u128 = if ts_secs <= 0 {
 | 
			
		||||
                    0
 | 
			
		||||
                } else {
 | 
			
		||||
                    (ts_secs as u128) * 1000
 | 
			
		||||
                };
 | 
			
		||||
                expiration_table.insert(key, &(expires_at_ms as u64))?;
 | 
			
		||||
                applied = true;
 | 
			
		||||
            }
 | 
			
		||||
        }
 | 
			
		||||
@@ -201,7 +214,7 @@ impl Storage {
 | 
			
		||||
            if is_string {
 | 
			
		||||
                let mut expiration_table = write_txn.open_table(EXPIRATION_TABLE)?;
 | 
			
		||||
                let expires_at_ms: u128 = if ts_ms <= 0 { 0 } else { ts_ms as u128 };
 | 
			
		||||
                expiration_table.insert(key, &((expires_at_ms as u64)))?;
 | 
			
		||||
                expiration_table.insert(key, &(expires_at_ms as u64))?;
 | 
			
		||||
                applied = true;
 | 
			
		||||
            }
 | 
			
		||||
        }
 | 
			
		||||
 
 | 
			
		||||
@@ -1,6 +1,6 @@
 | 
			
		||||
use redb::{ReadableTable};
 | 
			
		||||
use crate::error::DBError;
 | 
			
		||||
use super::*;
 | 
			
		||||
use crate::error::DBError;
 | 
			
		||||
use redb::ReadableTable;
 | 
			
		||||
 | 
			
		||||
impl Storage {
 | 
			
		||||
    // ✅ ENCRYPTION APPLIED: Values are encrypted before storage
 | 
			
		||||
@@ -18,7 +18,8 @@ impl Storage {
 | 
			
		||||
            };
 | 
			
		||||
 | 
			
		||||
            match key_type.as_deref() {
 | 
			
		||||
                Some("hash") | None => { // Proceed if hash or new key
 | 
			
		||||
                Some("hash") | None => {
 | 
			
		||||
                    // Proceed if hash or new key
 | 
			
		||||
                    // Set the type to hash (only if new key or existing hash)
 | 
			
		||||
                    types_table.insert(key, "hash")?;
 | 
			
		||||
 | 
			
		||||
@@ -35,7 +36,12 @@ impl Storage {
 | 
			
		||||
                        }
 | 
			
		||||
                    }
 | 
			
		||||
                }
 | 
			
		||||
                Some(_) => return Err(DBError("WRONGTYPE Operation against a key holding the wrong kind of value".to_string())),
 | 
			
		||||
                Some(_) => {
 | 
			
		||||
                    return Err(DBError(
 | 
			
		||||
                        "WRONGTYPE Operation against a key holding the wrong kind of value"
 | 
			
		||||
                            .to_string(),
 | 
			
		||||
                    ))
 | 
			
		||||
                }
 | 
			
		||||
            }
 | 
			
		||||
        }
 | 
			
		||||
 | 
			
		||||
@@ -62,7 +68,9 @@ impl Storage {
 | 
			
		||||
                    None => Ok(None),
 | 
			
		||||
                }
 | 
			
		||||
            }
 | 
			
		||||
            Some(_) => Err(DBError("WRONGTYPE Operation against a key holding the wrong kind of value".to_string())),
 | 
			
		||||
            Some(_) => Err(DBError(
 | 
			
		||||
                "WRONGTYPE Operation against a key holding the wrong kind of value".to_string(),
 | 
			
		||||
            )),
 | 
			
		||||
            None => Ok(None),
 | 
			
		||||
        }
 | 
			
		||||
    }
 | 
			
		||||
@@ -94,7 +102,9 @@ impl Storage {
 | 
			
		||||
 | 
			
		||||
                Ok(result)
 | 
			
		||||
            }
 | 
			
		||||
            Some(_) => Err(DBError("WRONGTYPE Operation against a key holding the wrong kind of value".to_string())),
 | 
			
		||||
            Some(_) => Err(DBError(
 | 
			
		||||
                "WRONGTYPE Operation against a key holding the wrong kind of value".to_string(),
 | 
			
		||||
            )),
 | 
			
		||||
            None => Ok(Vec::new()),
 | 
			
		||||
        }
 | 
			
		||||
    }
 | 
			
		||||
@@ -138,7 +148,11 @@ impl Storage {
 | 
			
		||||
                    types_table.remove(key)?;
 | 
			
		||||
                }
 | 
			
		||||
            }
 | 
			
		||||
            Some(_) => return Err(DBError("WRONGTYPE Operation against a key holding the wrong kind of value".to_string())),
 | 
			
		||||
            Some(_) => {
 | 
			
		||||
                return Err(DBError(
 | 
			
		||||
                    "WRONGTYPE Operation against a key holding the wrong kind of value".to_string(),
 | 
			
		||||
                ))
 | 
			
		||||
            }
 | 
			
		||||
            None => {} // Key does not exist, nothing to delete, return 0 deleted
 | 
			
		||||
        }
 | 
			
		||||
 | 
			
		||||
@@ -159,7 +173,9 @@ impl Storage {
 | 
			
		||||
                let hashes_table = read_txn.open_table(HASHES_TABLE)?;
 | 
			
		||||
                Ok(hashes_table.get((key, field))?.is_some())
 | 
			
		||||
            }
 | 
			
		||||
            Some(_) => Err(DBError("WRONGTYPE Operation against a key holding the wrong kind of value".to_string())),
 | 
			
		||||
            Some(_) => Err(DBError(
 | 
			
		||||
                "WRONGTYPE Operation against a key holding the wrong kind of value".to_string(),
 | 
			
		||||
            )),
 | 
			
		||||
            None => Ok(false),
 | 
			
		||||
        }
 | 
			
		||||
    }
 | 
			
		||||
@@ -188,7 +204,9 @@ impl Storage {
 | 
			
		||||
 | 
			
		||||
                Ok(result)
 | 
			
		||||
            }
 | 
			
		||||
            Some(_) => Err(DBError("WRONGTYPE Operation against a key holding the wrong kind of value".to_string())),
 | 
			
		||||
            Some(_) => Err(DBError(
 | 
			
		||||
                "WRONGTYPE Operation against a key holding the wrong kind of value".to_string(),
 | 
			
		||||
            )),
 | 
			
		||||
            None => Ok(Vec::new()),
 | 
			
		||||
        }
 | 
			
		||||
    }
 | 
			
		||||
@@ -220,7 +238,9 @@ impl Storage {
 | 
			
		||||
 | 
			
		||||
                Ok(result)
 | 
			
		||||
            }
 | 
			
		||||
            Some(_) => Err(DBError("WRONGTYPE Operation against a key holding the wrong kind of value".to_string())),
 | 
			
		||||
            Some(_) => Err(DBError(
 | 
			
		||||
                "WRONGTYPE Operation against a key holding the wrong kind of value".to_string(),
 | 
			
		||||
            )),
 | 
			
		||||
            None => Ok(Vec::new()),
 | 
			
		||||
        }
 | 
			
		||||
    }
 | 
			
		||||
@@ -249,7 +269,9 @@ impl Storage {
 | 
			
		||||
 | 
			
		||||
                Ok(count)
 | 
			
		||||
            }
 | 
			
		||||
            Some(_) => Err(DBError("WRONGTYPE Operation against a key holding the wrong kind of value".to_string())),
 | 
			
		||||
            Some(_) => Err(DBError(
 | 
			
		||||
                "WRONGTYPE Operation against a key holding the wrong kind of value".to_string(),
 | 
			
		||||
            )),
 | 
			
		||||
            None => Ok(0),
 | 
			
		||||
        }
 | 
			
		||||
    }
 | 
			
		||||
@@ -281,7 +303,9 @@ impl Storage {
 | 
			
		||||
 | 
			
		||||
                Ok(result)
 | 
			
		||||
            }
 | 
			
		||||
            Some(_) => Err(DBError("WRONGTYPE Operation against a key holding the wrong kind of value".to_string())),
 | 
			
		||||
            Some(_) => Err(DBError(
 | 
			
		||||
                "WRONGTYPE Operation against a key holding the wrong kind of value".to_string(),
 | 
			
		||||
            )),
 | 
			
		||||
            None => Ok(fields.into_iter().map(|_| None).collect()),
 | 
			
		||||
        }
 | 
			
		||||
    }
 | 
			
		||||
@@ -301,7 +325,8 @@ impl Storage {
 | 
			
		||||
            };
 | 
			
		||||
 | 
			
		||||
            match key_type.as_deref() {
 | 
			
		||||
                Some("hash") | None => { // Proceed if hash or new key
 | 
			
		||||
                Some("hash") | None => {
 | 
			
		||||
                    // Proceed if hash or new key
 | 
			
		||||
                    // Check if field already exists
 | 
			
		||||
                    if hashes_table.get((key, field))?.is_none() {
 | 
			
		||||
                        // Set the type to hash (only if new key or existing hash)
 | 
			
		||||
@@ -313,7 +338,12 @@ impl Storage {
 | 
			
		||||
                        result = true;
 | 
			
		||||
                    }
 | 
			
		||||
                }
 | 
			
		||||
                Some(_) => return Err(DBError("WRONGTYPE Operation against a key holding the wrong kind of value".to_string())),
 | 
			
		||||
                Some(_) => {
 | 
			
		||||
                    return Err(DBError(
 | 
			
		||||
                        "WRONGTYPE Operation against a key holding the wrong kind of value"
 | 
			
		||||
                            .to_string(),
 | 
			
		||||
                    ))
 | 
			
		||||
                }
 | 
			
		||||
            }
 | 
			
		||||
        }
 | 
			
		||||
 | 
			
		||||
@@ -322,7 +352,13 @@ impl Storage {
 | 
			
		||||
    }
 | 
			
		||||
 | 
			
		||||
    // ✅ ENCRYPTION APPLIED: Values are decrypted after retrieval
 | 
			
		||||
    pub fn hscan(&self, key: &str, cursor: u64, pattern: Option<&str>, count: Option<u64>) -> Result<(u64, Vec<(String, String)>), DBError> {
 | 
			
		||||
    pub fn hscan(
 | 
			
		||||
        &self,
 | 
			
		||||
        key: &str,
 | 
			
		||||
        cursor: u64,
 | 
			
		||||
        pattern: Option<&str>,
 | 
			
		||||
        count: Option<u64>,
 | 
			
		||||
    ) -> Result<(u64, Vec<(String, String)>), DBError> {
 | 
			
		||||
        let read_txn = self.db.begin_read()?;
 | 
			
		||||
        let types_table = read_txn.open_table(TYPES_TABLE)?;
 | 
			
		||||
        let key_type = {
 | 
			
		||||
@@ -367,10 +403,16 @@ impl Storage {
 | 
			
		||||
                    }
 | 
			
		||||
                }
 | 
			
		||||
 | 
			
		||||
                let next_cursor = if result.len() < limit { 0 } else { current_cursor };
 | 
			
		||||
                let next_cursor = if result.len() < limit {
 | 
			
		||||
                    0
 | 
			
		||||
                } else {
 | 
			
		||||
                    current_cursor
 | 
			
		||||
                };
 | 
			
		||||
                Ok((next_cursor, result))
 | 
			
		||||
            }
 | 
			
		||||
            Some(_) => Err(DBError("WRONGTYPE Operation against a key holding the wrong kind of value".to_string())),
 | 
			
		||||
            Some(_) => Err(DBError(
 | 
			
		||||
                "WRONGTYPE Operation against a key holding the wrong kind of value".to_string(),
 | 
			
		||||
            )),
 | 
			
		||||
            None => Ok((0, Vec::new())),
 | 
			
		||||
        }
 | 
			
		||||
    }
 | 
			
		||||
 
 | 
			
		||||
@@ -1,6 +1,6 @@
 | 
			
		||||
use redb::{ReadableTable};
 | 
			
		||||
use crate::error::DBError;
 | 
			
		||||
use super::*;
 | 
			
		||||
use crate::error::DBError;
 | 
			
		||||
use redb::ReadableTable;
 | 
			
		||||
 | 
			
		||||
impl Storage {
 | 
			
		||||
    // ✅ ENCRYPTION APPLIED: Elements are encrypted before storage
 | 
			
		||||
@@ -248,8 +248,16 @@ impl Storage {
 | 
			
		||||
                        }
 | 
			
		||||
 | 
			
		||||
                        let len = list.len() as i64;
 | 
			
		||||
                        let start_idx = if start < 0 { std::cmp::max(0, len + start) } else { std::cmp::min(start, len) };
 | 
			
		||||
                        let stop_idx = if stop < 0 { std::cmp::max(-1, len + stop) } else { std::cmp::min(stop, len - 1) };
 | 
			
		||||
                        let start_idx = if start < 0 {
 | 
			
		||||
                            std::cmp::max(0, len + start)
 | 
			
		||||
                        } else {
 | 
			
		||||
                            std::cmp::min(start, len)
 | 
			
		||||
                        };
 | 
			
		||||
                        let stop_idx = if stop < 0 {
 | 
			
		||||
                            std::cmp::max(-1, len + stop)
 | 
			
		||||
                        } else {
 | 
			
		||||
                            std::cmp::min(stop, len - 1)
 | 
			
		||||
                        };
 | 
			
		||||
 | 
			
		||||
                        if start_idx > stop_idx || start_idx >= len {
 | 
			
		||||
                            return Ok(Vec::new());
 | 
			
		||||
@@ -298,8 +306,16 @@ impl Storage {
 | 
			
		||||
            }
 | 
			
		||||
 | 
			
		||||
            let len = list.len() as i64;
 | 
			
		||||
            let start_idx = if start < 0 { std::cmp::max(0, len + start) } else { std::cmp::min(start, len) };
 | 
			
		||||
            let stop_idx = if stop < 0 { std::cmp::max(-1, len + stop) } else { std::cmp::min(stop, len - 1) };
 | 
			
		||||
            let start_idx = if start < 0 {
 | 
			
		||||
                std::cmp::max(0, len + start)
 | 
			
		||||
            } else {
 | 
			
		||||
                std::cmp::min(start, len)
 | 
			
		||||
            };
 | 
			
		||||
            let stop_idx = if stop < 0 {
 | 
			
		||||
                std::cmp::max(-1, len + stop)
 | 
			
		||||
            } else {
 | 
			
		||||
                std::cmp::min(stop, len - 1)
 | 
			
		||||
            };
 | 
			
		||||
 | 
			
		||||
            let mut lists_table = write_txn.open_table(LISTS_TABLE)?;
 | 
			
		||||
            if start_idx > stop_idx || start_idx >= len {
 | 
			
		||||
 
 | 
			
		||||
@@ -1,12 +1,12 @@
 | 
			
		||||
// src/storage_sled/mod.rs
 | 
			
		||||
use std::path::Path;
 | 
			
		||||
use std::sync::Arc;
 | 
			
		||||
use std::collections::HashMap;
 | 
			
		||||
use std::time::{SystemTime, UNIX_EPOCH};
 | 
			
		||||
use serde::{Deserialize, Serialize};
 | 
			
		||||
use crate::crypto::CryptoFactory;
 | 
			
		||||
use crate::error::DBError;
 | 
			
		||||
use crate::storage_trait::StorageBackend;
 | 
			
		||||
use crate::crypto::CryptoFactory;
 | 
			
		||||
use serde::{Deserialize, Serialize};
 | 
			
		||||
use std::collections::HashMap;
 | 
			
		||||
use std::path::Path;
 | 
			
		||||
use std::sync::Arc;
 | 
			
		||||
use std::time::{SystemTime, UNIX_EPOCH};
 | 
			
		||||
 | 
			
		||||
#[derive(Serialize, Deserialize, Debug, Clone)]
 | 
			
		||||
enum ValueType {
 | 
			
		||||
@@ -28,13 +28,22 @@ pub struct SledStorage {
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
impl SledStorage {
 | 
			
		||||
    pub fn new(path: impl AsRef<Path>, should_encrypt: bool, master_key: Option<&str>) -> Result<Self, DBError> {
 | 
			
		||||
    pub fn new(
 | 
			
		||||
        path: impl AsRef<Path>,
 | 
			
		||||
        should_encrypt: bool,
 | 
			
		||||
        master_key: Option<&str>,
 | 
			
		||||
    ) -> Result<Self, DBError> {
 | 
			
		||||
        let db = sled::open(path).map_err(|e| DBError(format!("Failed to open sled: {}", e)))?;
 | 
			
		||||
        let types = db.open_tree("types").map_err(|e| DBError(format!("Failed to open types tree: {}", e)))?;
 | 
			
		||||
        let types = db
 | 
			
		||||
            .open_tree("types")
 | 
			
		||||
            .map_err(|e| DBError(format!("Failed to open types tree: {}", e)))?;
 | 
			
		||||
 | 
			
		||||
        // Check if database was previously encrypted
 | 
			
		||||
        let encrypted_tree = db.open_tree("encrypted").map_err(|e| DBError(e.to_string()))?;
 | 
			
		||||
        let was_encrypted = encrypted_tree.get("encrypted")
 | 
			
		||||
        let encrypted_tree = db
 | 
			
		||||
            .open_tree("encrypted")
 | 
			
		||||
            .map_err(|e| DBError(e.to_string()))?;
 | 
			
		||||
        let was_encrypted = encrypted_tree
 | 
			
		||||
            .get("encrypted")
 | 
			
		||||
            .map_err(|e| DBError(e.to_string()))?
 | 
			
		||||
            .map(|v| v[0] == 1)
 | 
			
		||||
            .unwrap_or(false);
 | 
			
		||||
@@ -43,7 +52,9 @@ impl SledStorage {
 | 
			
		||||
            if let Some(key) = master_key {
 | 
			
		||||
                Some(CryptoFactory::new(key.as_bytes()))
 | 
			
		||||
            } else {
 | 
			
		||||
                return Err(DBError("Encryption requested but no master key provided".to_string()));
 | 
			
		||||
                return Err(DBError(
 | 
			
		||||
                    "Encryption requested but no master key provided".to_string(),
 | 
			
		||||
                ));
 | 
			
		||||
            }
 | 
			
		||||
        } else {
 | 
			
		||||
            None
 | 
			
		||||
@@ -51,7 +62,8 @@ impl SledStorage {
 | 
			
		||||
 | 
			
		||||
        // Mark database as encrypted if enabling encryption
 | 
			
		||||
        if should_encrypt && !was_encrypted {
 | 
			
		||||
            encrypted_tree.insert("encrypted", &[1u8])
 | 
			
		||||
            encrypted_tree
 | 
			
		||||
                .insert("encrypted", &[1u8])
 | 
			
		||||
                .map_err(|e| DBError(e.to_string()))?;
 | 
			
		||||
            encrypted_tree.flush().map_err(|e| DBError(e.to_string()))?;
 | 
			
		||||
        }
 | 
			
		||||
@@ -101,7 +113,7 @@ impl SledStorage {
 | 
			
		||||
 | 
			
		||||
                Ok(Some(storage_val))
 | 
			
		||||
            }
 | 
			
		||||
            None => Ok(None)
 | 
			
		||||
            None => Ok(None),
 | 
			
		||||
        }
 | 
			
		||||
    }
 | 
			
		||||
 | 
			
		||||
@@ -109,7 +121,9 @@ impl SledStorage {
 | 
			
		||||
        let data = bincode::serialize(&storage_val)
 | 
			
		||||
            .map_err(|e| DBError(format!("Serialization error: {}", e)))?;
 | 
			
		||||
        let encrypted = self.encrypt_if_needed(&data)?;
 | 
			
		||||
        self.db.insert(key, encrypted).map_err(|e| DBError(e.to_string()))?;
 | 
			
		||||
        self.db
 | 
			
		||||
            .insert(key, encrypted)
 | 
			
		||||
            .map_err(|e| DBError(e.to_string()))?;
 | 
			
		||||
 | 
			
		||||
        // Store type info (unencrypted for efficiency)
 | 
			
		||||
        let type_str = match &storage_val.value {
 | 
			
		||||
@@ -117,7 +131,9 @@ impl SledStorage {
 | 
			
		||||
            ValueType::Hash(_) => "hash",
 | 
			
		||||
            ValueType::List(_) => "list",
 | 
			
		||||
        };
 | 
			
		||||
        self.types.insert(key, type_str.as_bytes()).map_err(|e| DBError(e.to_string()))?;
 | 
			
		||||
        self.types
 | 
			
		||||
            .insert(key, type_str.as_bytes())
 | 
			
		||||
            .map_err(|e| DBError(e.to_string()))?;
 | 
			
		||||
 | 
			
		||||
        Ok(())
 | 
			
		||||
    }
 | 
			
		||||
@@ -168,9 +184,9 @@ impl StorageBackend for SledStorage {
 | 
			
		||||
        match self.get_storage_value(key)? {
 | 
			
		||||
            Some(storage_val) => match storage_val.value {
 | 
			
		||||
                ValueType::String(s) => Ok(Some(s)),
 | 
			
		||||
                _ => Ok(None)
 | 
			
		||||
            }
 | 
			
		||||
            None => Ok(None)
 | 
			
		||||
                _ => Ok(None),
 | 
			
		||||
            },
 | 
			
		||||
            None => Ok(None),
 | 
			
		||||
        }
 | 
			
		||||
    }
 | 
			
		||||
 | 
			
		||||
@@ -196,7 +212,9 @@ impl StorageBackend for SledStorage {
 | 
			
		||||
 | 
			
		||||
    fn del(&self, key: String) -> Result<(), DBError> {
 | 
			
		||||
        self.db.remove(&key).map_err(|e| DBError(e.to_string()))?;
 | 
			
		||||
        self.types.remove(&key).map_err(|e| DBError(e.to_string()))?;
 | 
			
		||||
        self.types
 | 
			
		||||
            .remove(&key)
 | 
			
		||||
            .map_err(|e| DBError(e.to_string()))?;
 | 
			
		||||
        self.db.flush().map_err(|e| DBError(e.to_string()))?;
 | 
			
		||||
        Ok(())
 | 
			
		||||
    }
 | 
			
		||||
@@ -222,7 +240,12 @@ impl StorageBackend for SledStorage {
 | 
			
		||||
        Ok(keys)
 | 
			
		||||
    }
 | 
			
		||||
 | 
			
		||||
    fn scan(&self, cursor: u64, pattern: Option<&str>, count: Option<u64>) -> Result<(u64, Vec<(String, String)>), DBError> {
 | 
			
		||||
    fn scan(
 | 
			
		||||
        &self,
 | 
			
		||||
        cursor: u64,
 | 
			
		||||
        pattern: Option<&str>,
 | 
			
		||||
        count: Option<u64>,
 | 
			
		||||
    ) -> Result<(u64, Vec<(String, String)>), DBError> {
 | 
			
		||||
        let mut result = Vec::new();
 | 
			
		||||
        let mut current_cursor = 0u64;
 | 
			
		||||
        let limit = count.unwrap_or(10) as usize;
 | 
			
		||||
@@ -258,7 +281,11 @@ impl StorageBackend for SledStorage {
 | 
			
		||||
            current_cursor += 1;
 | 
			
		||||
        }
 | 
			
		||||
 | 
			
		||||
        let next_cursor = if result.len() < limit { 0 } else { current_cursor };
 | 
			
		||||
        let next_cursor = if result.len() < limit {
 | 
			
		||||
            0
 | 
			
		||||
        } else {
 | 
			
		||||
            current_cursor
 | 
			
		||||
        };
 | 
			
		||||
        Ok((next_cursor, result))
 | 
			
		||||
    }
 | 
			
		||||
 | 
			
		||||
@@ -286,7 +313,7 @@ impl StorageBackend for SledStorage {
 | 
			
		||||
        if self.get_storage_value(key)?.is_some() {
 | 
			
		||||
            match self.types.get(key).map_err(|e| DBError(e.to_string()))? {
 | 
			
		||||
                Some(data) => Ok(Some(String::from_utf8_lossy(&data).to_string())),
 | 
			
		||||
                None => Ok(None)
 | 
			
		||||
                None => Ok(None),
 | 
			
		||||
            }
 | 
			
		||||
        } else {
 | 
			
		||||
            Ok(None)
 | 
			
		||||
@@ -302,7 +329,11 @@ impl StorageBackend for SledStorage {
 | 
			
		||||
 | 
			
		||||
        let hash = match &mut storage_val.value {
 | 
			
		||||
            ValueType::Hash(h) => h,
 | 
			
		||||
            _ => return Err(DBError("WRONGTYPE Operation against a key holding the wrong kind of value".to_string())),
 | 
			
		||||
            _ => {
 | 
			
		||||
                return Err(DBError(
 | 
			
		||||
                    "WRONGTYPE Operation against a key holding the wrong kind of value".to_string(),
 | 
			
		||||
                ))
 | 
			
		||||
            }
 | 
			
		||||
        };
 | 
			
		||||
 | 
			
		||||
        let mut new_fields = 0i64;
 | 
			
		||||
@@ -322,9 +353,9 @@ impl StorageBackend for SledStorage {
 | 
			
		||||
        match self.get_storage_value(key)? {
 | 
			
		||||
            Some(storage_val) => match storage_val.value {
 | 
			
		||||
                ValueType::Hash(h) => Ok(h.get(field).cloned()),
 | 
			
		||||
                _ => Ok(None)
 | 
			
		||||
            }
 | 
			
		||||
            None => Ok(None)
 | 
			
		||||
                _ => Ok(None),
 | 
			
		||||
            },
 | 
			
		||||
            None => Ok(None),
 | 
			
		||||
        }
 | 
			
		||||
    }
 | 
			
		||||
 | 
			
		||||
@@ -332,13 +363,19 @@ impl StorageBackend for SledStorage {
 | 
			
		||||
        match self.get_storage_value(key)? {
 | 
			
		||||
            Some(storage_val) => match storage_val.value {
 | 
			
		||||
                ValueType::Hash(h) => Ok(h.into_iter().collect()),
 | 
			
		||||
                _ => Ok(Vec::new())
 | 
			
		||||
            }
 | 
			
		||||
            None => Ok(Vec::new())
 | 
			
		||||
                _ => Ok(Vec::new()),
 | 
			
		||||
            },
 | 
			
		||||
            None => Ok(Vec::new()),
 | 
			
		||||
        }
 | 
			
		||||
    }
 | 
			
		||||
 | 
			
		||||
    fn hscan(&self, key: &str, cursor: u64, pattern: Option<&str>, count: Option<u64>) -> Result<(u64, Vec<(String, String)>), DBError> {
 | 
			
		||||
    fn hscan(
 | 
			
		||||
        &self,
 | 
			
		||||
        key: &str,
 | 
			
		||||
        cursor: u64,
 | 
			
		||||
        pattern: Option<&str>,
 | 
			
		||||
        count: Option<u64>,
 | 
			
		||||
    ) -> Result<(u64, Vec<(String, String)>), DBError> {
 | 
			
		||||
        match self.get_storage_value(key)? {
 | 
			
		||||
            Some(storage_val) => match storage_val.value {
 | 
			
		||||
                ValueType::Hash(h) => {
 | 
			
		||||
@@ -365,24 +402,30 @@ impl StorageBackend for SledStorage {
 | 
			
		||||
                        current_cursor += 1;
 | 
			
		||||
                    }
 | 
			
		||||
 | 
			
		||||
                    let next_cursor = if result.len() < limit { 0 } else { current_cursor };
 | 
			
		||||
                    let next_cursor = if result.len() < limit {
 | 
			
		||||
                        0
 | 
			
		||||
                    } else {
 | 
			
		||||
                        current_cursor
 | 
			
		||||
                    };
 | 
			
		||||
                    Ok((next_cursor, result))
 | 
			
		||||
                }
 | 
			
		||||
                _ => Err(DBError("WRONGTYPE Operation against a key holding the wrong kind of value".to_string()))
 | 
			
		||||
            }
 | 
			
		||||
            None => Ok((0, Vec::new()))
 | 
			
		||||
                _ => Err(DBError(
 | 
			
		||||
                    "WRONGTYPE Operation against a key holding the wrong kind of value".to_string(),
 | 
			
		||||
                )),
 | 
			
		||||
            },
 | 
			
		||||
            None => Ok((0, Vec::new())),
 | 
			
		||||
        }
 | 
			
		||||
    }
 | 
			
		||||
 | 
			
		||||
    fn hdel(&self, key: &str, fields: Vec<String>) -> Result<i64, DBError> {
 | 
			
		||||
        let mut storage_val = match self.get_storage_value(key)? {
 | 
			
		||||
            Some(sv) => sv,
 | 
			
		||||
            None => return Ok(0)
 | 
			
		||||
            None => return Ok(0),
 | 
			
		||||
        };
 | 
			
		||||
 | 
			
		||||
        let hash = match &mut storage_val.value {
 | 
			
		||||
            ValueType::Hash(h) => h,
 | 
			
		||||
            _ => return Ok(0)
 | 
			
		||||
            _ => return Ok(0),
 | 
			
		||||
        };
 | 
			
		||||
 | 
			
		||||
        let mut deleted = 0i64;
 | 
			
		||||
@@ -406,9 +449,9 @@ impl StorageBackend for SledStorage {
 | 
			
		||||
        match self.get_storage_value(key)? {
 | 
			
		||||
            Some(storage_val) => match storage_val.value {
 | 
			
		||||
                ValueType::Hash(h) => Ok(h.contains_key(field)),
 | 
			
		||||
                _ => Ok(false)
 | 
			
		||||
            }
 | 
			
		||||
            None => Ok(false)
 | 
			
		||||
                _ => Ok(false),
 | 
			
		||||
            },
 | 
			
		||||
            None => Ok(false),
 | 
			
		||||
        }
 | 
			
		||||
    }
 | 
			
		||||
 | 
			
		||||
@@ -416,9 +459,9 @@ impl StorageBackend for SledStorage {
 | 
			
		||||
        match self.get_storage_value(key)? {
 | 
			
		||||
            Some(storage_val) => match storage_val.value {
 | 
			
		||||
                ValueType::Hash(h) => Ok(h.keys().cloned().collect()),
 | 
			
		||||
                _ => Ok(Vec::new())
 | 
			
		||||
            }
 | 
			
		||||
            None => Ok(Vec::new())
 | 
			
		||||
                _ => Ok(Vec::new()),
 | 
			
		||||
            },
 | 
			
		||||
            None => Ok(Vec::new()),
 | 
			
		||||
        }
 | 
			
		||||
    }
 | 
			
		||||
 | 
			
		||||
@@ -426,9 +469,9 @@ impl StorageBackend for SledStorage {
 | 
			
		||||
        match self.get_storage_value(key)? {
 | 
			
		||||
            Some(storage_val) => match storage_val.value {
 | 
			
		||||
                ValueType::Hash(h) => Ok(h.values().cloned().collect()),
 | 
			
		||||
                _ => Ok(Vec::new())
 | 
			
		||||
            }
 | 
			
		||||
            None => Ok(Vec::new())
 | 
			
		||||
                _ => Ok(Vec::new()),
 | 
			
		||||
            },
 | 
			
		||||
            None => Ok(Vec::new()),
 | 
			
		||||
        }
 | 
			
		||||
    }
 | 
			
		||||
 | 
			
		||||
@@ -436,21 +479,19 @@ impl StorageBackend for SledStorage {
 | 
			
		||||
        match self.get_storage_value(key)? {
 | 
			
		||||
            Some(storage_val) => match storage_val.value {
 | 
			
		||||
                ValueType::Hash(h) => Ok(h.len() as i64),
 | 
			
		||||
                _ => Ok(0)
 | 
			
		||||
            }
 | 
			
		||||
            None => Ok(0)
 | 
			
		||||
                _ => Ok(0),
 | 
			
		||||
            },
 | 
			
		||||
            None => Ok(0),
 | 
			
		||||
        }
 | 
			
		||||
    }
 | 
			
		||||
 | 
			
		||||
    fn hmget(&self, key: &str, fields: Vec<String>) -> Result<Vec<Option<String>>, DBError> {
 | 
			
		||||
        match self.get_storage_value(key)? {
 | 
			
		||||
            Some(storage_val) => match storage_val.value {
 | 
			
		||||
                ValueType::Hash(h) => {
 | 
			
		||||
                    Ok(fields.into_iter().map(|f| h.get(&f).cloned()).collect())
 | 
			
		||||
                }
 | 
			
		||||
                _ => Ok(fields.into_iter().map(|_| None).collect())
 | 
			
		||||
            }
 | 
			
		||||
            None => Ok(fields.into_iter().map(|_| None).collect())
 | 
			
		||||
                ValueType::Hash(h) => Ok(fields.into_iter().map(|f| h.get(&f).cloned()).collect()),
 | 
			
		||||
                _ => Ok(fields.into_iter().map(|_| None).collect()),
 | 
			
		||||
            },
 | 
			
		||||
            None => Ok(fields.into_iter().map(|_| None).collect()),
 | 
			
		||||
        }
 | 
			
		||||
    }
 | 
			
		||||
 | 
			
		||||
@@ -462,7 +503,11 @@ impl StorageBackend for SledStorage {
 | 
			
		||||
 | 
			
		||||
        let hash = match &mut storage_val.value {
 | 
			
		||||
            ValueType::Hash(h) => h,
 | 
			
		||||
            _ => return Err(DBError("WRONGTYPE Operation against a key holding the wrong kind of value".to_string())),
 | 
			
		||||
            _ => {
 | 
			
		||||
                return Err(DBError(
 | 
			
		||||
                    "WRONGTYPE Operation against a key holding the wrong kind of value".to_string(),
 | 
			
		||||
                ))
 | 
			
		||||
            }
 | 
			
		||||
        };
 | 
			
		||||
 | 
			
		||||
        if hash.contains_key(field) {
 | 
			
		||||
@@ -484,7 +529,11 @@ impl StorageBackend for SledStorage {
 | 
			
		||||
 | 
			
		||||
        let list = match &mut storage_val.value {
 | 
			
		||||
            ValueType::List(l) => l,
 | 
			
		||||
            _ => return Err(DBError("WRONGTYPE Operation against a key holding the wrong kind of value".to_string())),
 | 
			
		||||
            _ => {
 | 
			
		||||
                return Err(DBError(
 | 
			
		||||
                    "WRONGTYPE Operation against a key holding the wrong kind of value".to_string(),
 | 
			
		||||
                ))
 | 
			
		||||
            }
 | 
			
		||||
        };
 | 
			
		||||
 | 
			
		||||
        for element in elements.into_iter().rev() {
 | 
			
		||||
@@ -505,7 +554,11 @@ impl StorageBackend for SledStorage {
 | 
			
		||||
 | 
			
		||||
        let list = match &mut storage_val.value {
 | 
			
		||||
            ValueType::List(l) => l,
 | 
			
		||||
            _ => return Err(DBError("WRONGTYPE Operation against a key holding the wrong kind of value".to_string())),
 | 
			
		||||
            _ => {
 | 
			
		||||
                return Err(DBError(
 | 
			
		||||
                    "WRONGTYPE Operation against a key holding the wrong kind of value".to_string(),
 | 
			
		||||
                ))
 | 
			
		||||
            }
 | 
			
		||||
        };
 | 
			
		||||
 | 
			
		||||
        list.extend(elements);
 | 
			
		||||
@@ -518,12 +571,12 @@ impl StorageBackend for SledStorage {
 | 
			
		||||
    fn lpop(&self, key: &str, count: u64) -> Result<Vec<String>, DBError> {
 | 
			
		||||
        let mut storage_val = match self.get_storage_value(key)? {
 | 
			
		||||
            Some(sv) => sv,
 | 
			
		||||
            None => return Ok(Vec::new())
 | 
			
		||||
            None => return Ok(Vec::new()),
 | 
			
		||||
        };
 | 
			
		||||
 | 
			
		||||
        let list = match &mut storage_val.value {
 | 
			
		||||
            ValueType::List(l) => l,
 | 
			
		||||
            _ => return Ok(Vec::new())
 | 
			
		||||
            _ => return Ok(Vec::new()),
 | 
			
		||||
        };
 | 
			
		||||
 | 
			
		||||
        let mut result = Vec::new();
 | 
			
		||||
@@ -547,12 +600,12 @@ impl StorageBackend for SledStorage {
 | 
			
		||||
    fn rpop(&self, key: &str, count: u64) -> Result<Vec<String>, DBError> {
 | 
			
		||||
        let mut storage_val = match self.get_storage_value(key)? {
 | 
			
		||||
            Some(sv) => sv,
 | 
			
		||||
            None => return Ok(Vec::new())
 | 
			
		||||
            None => return Ok(Vec::new()),
 | 
			
		||||
        };
 | 
			
		||||
 | 
			
		||||
        let list = match &mut storage_val.value {
 | 
			
		||||
            ValueType::List(l) => l,
 | 
			
		||||
            _ => return Ok(Vec::new())
 | 
			
		||||
            _ => return Ok(Vec::new()),
 | 
			
		||||
        };
 | 
			
		||||
 | 
			
		||||
        let mut result = Vec::new();
 | 
			
		||||
@@ -576,9 +629,9 @@ impl StorageBackend for SledStorage {
 | 
			
		||||
        match self.get_storage_value(key)? {
 | 
			
		||||
            Some(storage_val) => match storage_val.value {
 | 
			
		||||
                ValueType::List(l) => Ok(l.len() as i64),
 | 
			
		||||
                _ => Ok(0)
 | 
			
		||||
            }
 | 
			
		||||
            None => Ok(0)
 | 
			
		||||
                _ => Ok(0),
 | 
			
		||||
            },
 | 
			
		||||
            None => Ok(0),
 | 
			
		||||
        }
 | 
			
		||||
    }
 | 
			
		||||
 | 
			
		||||
@@ -598,9 +651,9 @@ impl StorageBackend for SledStorage {
 | 
			
		||||
                        Ok(None)
 | 
			
		||||
                    }
 | 
			
		||||
                }
 | 
			
		||||
                _ => Ok(None)
 | 
			
		||||
            }
 | 
			
		||||
            None => Ok(None)
 | 
			
		||||
                _ => Ok(None),
 | 
			
		||||
            },
 | 
			
		||||
            None => Ok(None),
 | 
			
		||||
        }
 | 
			
		||||
    }
 | 
			
		||||
 | 
			
		||||
@@ -633,21 +686,21 @@ impl StorageBackend for SledStorage {
 | 
			
		||||
 | 
			
		||||
                    Ok(list[start_usize..std::cmp::min(stop_usize, list.len())].to_vec())
 | 
			
		||||
                }
 | 
			
		||||
                _ => Ok(Vec::new())
 | 
			
		||||
            }
 | 
			
		||||
            None => Ok(Vec::new())
 | 
			
		||||
                _ => Ok(Vec::new()),
 | 
			
		||||
            },
 | 
			
		||||
            None => Ok(Vec::new()),
 | 
			
		||||
        }
 | 
			
		||||
    }
 | 
			
		||||
 | 
			
		||||
    fn ltrim(&self, key: &str, start: i64, stop: i64) -> Result<(), DBError> {
 | 
			
		||||
        let mut storage_val = match self.get_storage_value(key)? {
 | 
			
		||||
            Some(sv) => sv,
 | 
			
		||||
            None => return Ok(())
 | 
			
		||||
            None => return Ok(()),
 | 
			
		||||
        };
 | 
			
		||||
 | 
			
		||||
        let list = match &mut storage_val.value {
 | 
			
		||||
            ValueType::List(l) => l,
 | 
			
		||||
            _ => return Ok(())
 | 
			
		||||
            _ => return Ok(()),
 | 
			
		||||
        };
 | 
			
		||||
 | 
			
		||||
        if list.is_empty() {
 | 
			
		||||
@@ -687,12 +740,12 @@ impl StorageBackend for SledStorage {
 | 
			
		||||
    fn lrem(&self, key: &str, count: i64, element: &str) -> Result<i64, DBError> {
 | 
			
		||||
        let mut storage_val = match self.get_storage_value(key)? {
 | 
			
		||||
            Some(sv) => sv,
 | 
			
		||||
            None => return Ok(0)
 | 
			
		||||
            None => return Ok(0),
 | 
			
		||||
        };
 | 
			
		||||
 | 
			
		||||
        let list = match &mut storage_val.value {
 | 
			
		||||
            ValueType::List(l) => l,
 | 
			
		||||
            _ => return Ok(0)
 | 
			
		||||
            _ => return Ok(0),
 | 
			
		||||
        };
 | 
			
		||||
 | 
			
		||||
        let mut removed = 0i64;
 | 
			
		||||
@@ -751,14 +804,14 @@ impl StorageBackend for SledStorage {
 | 
			
		||||
                    Ok(-1) // Key exists but has no expiration
 | 
			
		||||
                }
 | 
			
		||||
            }
 | 
			
		||||
            None => Ok(-2) // Key does not exist
 | 
			
		||||
            None => Ok(-2), // Key does not exist
 | 
			
		||||
        }
 | 
			
		||||
    }
 | 
			
		||||
 | 
			
		||||
    fn expire_seconds(&self, key: &str, secs: u64) -> Result<bool, DBError> {
 | 
			
		||||
        let mut storage_val = match self.get_storage_value(key)? {
 | 
			
		||||
            Some(sv) => sv,
 | 
			
		||||
            None => return Ok(false)
 | 
			
		||||
            None => return Ok(false),
 | 
			
		||||
        };
 | 
			
		||||
 | 
			
		||||
        storage_val.expires_at = Some(Self::now_millis() + (secs as u128) * 1000);
 | 
			
		||||
@@ -770,7 +823,7 @@ impl StorageBackend for SledStorage {
 | 
			
		||||
    fn pexpire_millis(&self, key: &str, ms: u128) -> Result<bool, DBError> {
 | 
			
		||||
        let mut storage_val = match self.get_storage_value(key)? {
 | 
			
		||||
            Some(sv) => sv,
 | 
			
		||||
            None => return Ok(false)
 | 
			
		||||
            None => return Ok(false),
 | 
			
		||||
        };
 | 
			
		||||
 | 
			
		||||
        storage_val.expires_at = Some(Self::now_millis() + ms);
 | 
			
		||||
@@ -782,7 +835,7 @@ impl StorageBackend for SledStorage {
 | 
			
		||||
    fn persist(&self, key: &str) -> Result<bool, DBError> {
 | 
			
		||||
        let mut storage_val = match self.get_storage_value(key)? {
 | 
			
		||||
            Some(sv) => sv,
 | 
			
		||||
            None => return Ok(false)
 | 
			
		||||
            None => return Ok(false),
 | 
			
		||||
        };
 | 
			
		||||
 | 
			
		||||
        if storage_val.expires_at.is_some() {
 | 
			
		||||
@@ -798,10 +851,14 @@ impl StorageBackend for SledStorage {
 | 
			
		||||
    fn expire_at_seconds(&self, key: &str, ts_secs: i64) -> Result<bool, DBError> {
 | 
			
		||||
        let mut storage_val = match self.get_storage_value(key)? {
 | 
			
		||||
            Some(sv) => sv,
 | 
			
		||||
            None => return Ok(false)
 | 
			
		||||
            None => return Ok(false),
 | 
			
		||||
        };
 | 
			
		||||
 | 
			
		||||
        let expires_at_ms: u128 = if ts_secs <= 0 { 0 } else { (ts_secs as u128) * 1000 };
 | 
			
		||||
        let expires_at_ms: u128 = if ts_secs <= 0 {
 | 
			
		||||
            0
 | 
			
		||||
        } else {
 | 
			
		||||
            (ts_secs as u128) * 1000
 | 
			
		||||
        };
 | 
			
		||||
        storage_val.expires_at = Some(expires_at_ms);
 | 
			
		||||
        self.set_storage_value(key, storage_val)?;
 | 
			
		||||
        self.db.flush().map_err(|e| DBError(e.to_string()))?;
 | 
			
		||||
@@ -811,7 +868,7 @@ impl StorageBackend for SledStorage {
 | 
			
		||||
    fn pexpire_at_millis(&self, key: &str, ts_ms: i64) -> Result<bool, DBError> {
 | 
			
		||||
        let mut storage_val = match self.get_storage_value(key)? {
 | 
			
		||||
            Some(sv) => sv,
 | 
			
		||||
            None => return Ok(false)
 | 
			
		||||
            None => return Ok(false),
 | 
			
		||||
        };
 | 
			
		||||
 | 
			
		||||
        let expires_at_ms: u128 = if ts_ms <= 0 { 0 } else { ts_ms as u128 };
 | 
			
		||||
 
 | 
			
		||||
@@ -15,8 +15,19 @@ pub trait StorageBackend: Send + Sync {
 | 
			
		||||
    fn get_key_type(&self, key: &str) -> Result<Option<String>, DBError>;
 | 
			
		||||
 | 
			
		||||
    // Scanning
 | 
			
		||||
    fn scan(&self, cursor: u64, pattern: Option<&str>, count: Option<u64>) -> Result<(u64, Vec<(String, String)>), DBError>;
 | 
			
		||||
    fn hscan(&self, key: &str, cursor: u64, pattern: Option<&str>, count: Option<u64>) -> Result<(u64, Vec<(String, String)>), DBError>;
 | 
			
		||||
    fn scan(
 | 
			
		||||
        &self,
 | 
			
		||||
        cursor: u64,
 | 
			
		||||
        pattern: Option<&str>,
 | 
			
		||||
        count: Option<u64>,
 | 
			
		||||
    ) -> Result<(u64, Vec<(String, String)>), DBError>;
 | 
			
		||||
    fn hscan(
 | 
			
		||||
        &self,
 | 
			
		||||
        key: &str,
 | 
			
		||||
        cursor: u64,
 | 
			
		||||
        pattern: Option<&str>,
 | 
			
		||||
        count: Option<u64>,
 | 
			
		||||
    ) -> Result<(u64, Vec<(String, String)>), DBError>;
 | 
			
		||||
 | 
			
		||||
    // Hash operations
 | 
			
		||||
    fn hset(&self, key: &str, pairs: Vec<(String, String)>) -> Result<i64, DBError>;
 | 
			
		||||
 
 | 
			
		||||
							
								
								
									
										657
									
								
								src/tantivy_search.rs
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										657
									
								
								src/tantivy_search.rs
									
									
									
									
									
										Normal file
									
								
							@@ -0,0 +1,657 @@
 | 
			
		||||
use crate::error::DBError;
 | 
			
		||||
use serde::{Deserialize, Serialize};
 | 
			
		||||
use std::collections::HashMap;
 | 
			
		||||
use std::path::PathBuf;
 | 
			
		||||
use std::sync::{Arc, RwLock};
 | 
			
		||||
use tantivy::{
 | 
			
		||||
    collector::TopDocs,
 | 
			
		||||
    directory::MmapDirectory,
 | 
			
		||||
    query::{BooleanQuery, Occur, Query, QueryParser, TermQuery},
 | 
			
		||||
    schema::{Field, Schema, TextFieldIndexing, TextOptions, Value, STORED, STRING},
 | 
			
		||||
    tokenizer::TokenizerManager,
 | 
			
		||||
    DateTime, Index, IndexReader, IndexWriter, ReloadPolicy, TantivyDocument, Term,
 | 
			
		||||
};
 | 
			
		||||
 | 
			
		||||
#[derive(Debug, Clone, Serialize, Deserialize)]
 | 
			
		||||
pub enum FieldDef {
 | 
			
		||||
    Text {
 | 
			
		||||
        stored: bool,
 | 
			
		||||
        indexed: bool,
 | 
			
		||||
        tokenized: bool,
 | 
			
		||||
        fast: bool,
 | 
			
		||||
    },
 | 
			
		||||
    Numeric {
 | 
			
		||||
        stored: bool,
 | 
			
		||||
        indexed: bool,
 | 
			
		||||
        fast: bool,
 | 
			
		||||
        precision: NumericType,
 | 
			
		||||
    },
 | 
			
		||||
    Tag {
 | 
			
		||||
        stored: bool,
 | 
			
		||||
        separator: String,
 | 
			
		||||
        case_sensitive: bool,
 | 
			
		||||
    },
 | 
			
		||||
    Geo {
 | 
			
		||||
        stored: bool,
 | 
			
		||||
    },
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
#[derive(Debug, Clone, Serialize, Deserialize)]
 | 
			
		||||
pub enum NumericType {
 | 
			
		||||
    I64,
 | 
			
		||||
    U64,
 | 
			
		||||
    F64,
 | 
			
		||||
    Date,
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
pub struct IndexSchema {
 | 
			
		||||
    schema: Schema,
 | 
			
		||||
    fields: HashMap<String, (Field, FieldDef)>,
 | 
			
		||||
    default_search_fields: Vec<Field>,
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
pub struct TantivySearch {
 | 
			
		||||
    index: Index,
 | 
			
		||||
    writer: Arc<RwLock<IndexWriter>>,
 | 
			
		||||
    reader: IndexReader,
 | 
			
		||||
    index_schema: IndexSchema,
 | 
			
		||||
    name: String,
 | 
			
		||||
    config: IndexConfig,
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
#[derive(Debug, Clone, Serialize, Deserialize)]
 | 
			
		||||
pub struct IndexConfig {
 | 
			
		||||
    pub language: String,
 | 
			
		||||
    pub stopwords: Vec<String>,
 | 
			
		||||
    pub stemming: bool,
 | 
			
		||||
    pub max_doc_count: Option<usize>,
 | 
			
		||||
    pub default_score: f64,
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
impl Default for IndexConfig {
 | 
			
		||||
    fn default() -> Self {
 | 
			
		||||
        IndexConfig {
 | 
			
		||||
            language: "english".to_string(),
 | 
			
		||||
            stopwords: vec![],
 | 
			
		||||
            stemming: true,
 | 
			
		||||
            max_doc_count: None,
 | 
			
		||||
            default_score: 1.0,
 | 
			
		||||
        }
 | 
			
		||||
    }
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
impl TantivySearch {
 | 
			
		||||
    pub fn new_with_schema(
 | 
			
		||||
        base_path: PathBuf,
 | 
			
		||||
        name: String,
 | 
			
		||||
        field_definitions: Vec<(String, FieldDef)>,
 | 
			
		||||
        config: Option<IndexConfig>,
 | 
			
		||||
    ) -> Result<Self, DBError> {
 | 
			
		||||
        let index_path = base_path.join(&name);
 | 
			
		||||
        std::fs::create_dir_all(&index_path)
 | 
			
		||||
            .map_err(|e| DBError(format!("Failed to create index dir: {}", e)))?;
 | 
			
		||||
 | 
			
		||||
        // Build schema from field definitions
 | 
			
		||||
        let mut schema_builder = Schema::builder();
 | 
			
		||||
        let mut fields = HashMap::new();
 | 
			
		||||
        let mut default_search_fields = Vec::new();
 | 
			
		||||
 | 
			
		||||
        // Always add a document ID field
 | 
			
		||||
        let id_field = schema_builder.add_text_field("_id", STRING | STORED);
 | 
			
		||||
        fields.insert(
 | 
			
		||||
            "_id".to_string(),
 | 
			
		||||
            (
 | 
			
		||||
                id_field,
 | 
			
		||||
                FieldDef::Text {
 | 
			
		||||
                    stored: true,
 | 
			
		||||
                    indexed: true,
 | 
			
		||||
                    tokenized: false,
 | 
			
		||||
                    fast: false,
 | 
			
		||||
                },
 | 
			
		||||
            ),
 | 
			
		||||
        );
 | 
			
		||||
 | 
			
		||||
        // Add user-defined fields
 | 
			
		||||
        for (field_name, field_def) in field_definitions {
 | 
			
		||||
            let field = match &field_def {
 | 
			
		||||
                FieldDef::Text {
 | 
			
		||||
                    stored,
 | 
			
		||||
                    indexed,
 | 
			
		||||
                    tokenized,
 | 
			
		||||
                    fast: _fast,
 | 
			
		||||
                } => {
 | 
			
		||||
                    let mut text_options = TextOptions::default();
 | 
			
		||||
 | 
			
		||||
                    if *stored {
 | 
			
		||||
                        text_options = text_options.set_stored();
 | 
			
		||||
                    }
 | 
			
		||||
 | 
			
		||||
                    if *indexed {
 | 
			
		||||
                        let indexing_options = if *tokenized {
 | 
			
		||||
                            TextFieldIndexing::default()
 | 
			
		||||
                                .set_tokenizer("default")
 | 
			
		||||
                                .set_index_option(
 | 
			
		||||
                                    tantivy::schema::IndexRecordOption::WithFreqsAndPositions,
 | 
			
		||||
                                )
 | 
			
		||||
                        } else {
 | 
			
		||||
                            TextFieldIndexing::default()
 | 
			
		||||
                                .set_tokenizer("raw")
 | 
			
		||||
                                .set_index_option(tantivy::schema::IndexRecordOption::Basic)
 | 
			
		||||
                        };
 | 
			
		||||
                        text_options = text_options.set_indexing_options(indexing_options);
 | 
			
		||||
 | 
			
		||||
                        let f = schema_builder.add_text_field(&field_name, text_options);
 | 
			
		||||
                        if *tokenized {
 | 
			
		||||
                            default_search_fields.push(f);
 | 
			
		||||
                        }
 | 
			
		||||
                        f
 | 
			
		||||
                    } else {
 | 
			
		||||
                        schema_builder.add_text_field(&field_name, text_options)
 | 
			
		||||
                    }
 | 
			
		||||
                }
 | 
			
		||||
                FieldDef::Numeric {
 | 
			
		||||
                    stored,
 | 
			
		||||
                    indexed,
 | 
			
		||||
                    fast,
 | 
			
		||||
                    precision,
 | 
			
		||||
                } => match precision {
 | 
			
		||||
                    NumericType::I64 => {
 | 
			
		||||
                        let mut opts = tantivy::schema::NumericOptions::default();
 | 
			
		||||
                        if *stored {
 | 
			
		||||
                            opts = opts.set_stored();
 | 
			
		||||
                        }
 | 
			
		||||
                        if *indexed {
 | 
			
		||||
                            opts = opts.set_indexed();
 | 
			
		||||
                        }
 | 
			
		||||
                        if *fast {
 | 
			
		||||
                            opts = opts.set_fast();
 | 
			
		||||
                        }
 | 
			
		||||
                        schema_builder.add_i64_field(&field_name, opts)
 | 
			
		||||
                    }
 | 
			
		||||
                    NumericType::U64 => {
 | 
			
		||||
                        let mut opts = tantivy::schema::NumericOptions::default();
 | 
			
		||||
                        if *stored {
 | 
			
		||||
                            opts = opts.set_stored();
 | 
			
		||||
                        }
 | 
			
		||||
                        if *indexed {
 | 
			
		||||
                            opts = opts.set_indexed();
 | 
			
		||||
                        }
 | 
			
		||||
                        if *fast {
 | 
			
		||||
                            opts = opts.set_fast();
 | 
			
		||||
                        }
 | 
			
		||||
                        schema_builder.add_u64_field(&field_name, opts)
 | 
			
		||||
                    }
 | 
			
		||||
                    NumericType::F64 => {
 | 
			
		||||
                        let mut opts = tantivy::schema::NumericOptions::default();
 | 
			
		||||
                        if *stored {
 | 
			
		||||
                            opts = opts.set_stored();
 | 
			
		||||
                        }
 | 
			
		||||
                        if *indexed {
 | 
			
		||||
                            opts = opts.set_indexed();
 | 
			
		||||
                        }
 | 
			
		||||
                        if *fast {
 | 
			
		||||
                            opts = opts.set_fast();
 | 
			
		||||
                        }
 | 
			
		||||
                        schema_builder.add_f64_field(&field_name, opts)
 | 
			
		||||
                    }
 | 
			
		||||
                    NumericType::Date => {
 | 
			
		||||
                        let mut opts = tantivy::schema::DateOptions::default();
 | 
			
		||||
                        if *stored {
 | 
			
		||||
                            opts = opts.set_stored();
 | 
			
		||||
                        }
 | 
			
		||||
                        if *indexed {
 | 
			
		||||
                            opts = opts.set_indexed();
 | 
			
		||||
                        }
 | 
			
		||||
                        if *fast {
 | 
			
		||||
                            opts = opts.set_fast();
 | 
			
		||||
                        }
 | 
			
		||||
                        schema_builder.add_date_field(&field_name, opts)
 | 
			
		||||
                    }
 | 
			
		||||
                },
 | 
			
		||||
                FieldDef::Tag {
 | 
			
		||||
                    stored,
 | 
			
		||||
                    separator: _,
 | 
			
		||||
                    case_sensitive: _,
 | 
			
		||||
                } => {
 | 
			
		||||
                    let mut text_options = TextOptions::default();
 | 
			
		||||
                    if *stored {
 | 
			
		||||
                        text_options = text_options.set_stored();
 | 
			
		||||
                    }
 | 
			
		||||
                    text_options = text_options.set_indexing_options(
 | 
			
		||||
                        TextFieldIndexing::default()
 | 
			
		||||
                            .set_tokenizer("raw")
 | 
			
		||||
                            .set_index_option(tantivy::schema::IndexRecordOption::Basic),
 | 
			
		||||
                    );
 | 
			
		||||
                    schema_builder.add_text_field(&field_name, text_options)
 | 
			
		||||
                }
 | 
			
		||||
                FieldDef::Geo { stored } => {
 | 
			
		||||
                    // For now, store as two f64 fields for lat/lon
 | 
			
		||||
                    let mut opts = tantivy::schema::NumericOptions::default();
 | 
			
		||||
                    if *stored {
 | 
			
		||||
                        opts = opts.set_stored();
 | 
			
		||||
                    }
 | 
			
		||||
                    opts = opts.set_indexed().set_fast();
 | 
			
		||||
 | 
			
		||||
                    let lat_field =
 | 
			
		||||
                        schema_builder.add_f64_field(&format!("{}_lat", field_name), opts.clone());
 | 
			
		||||
                    let lon_field =
 | 
			
		||||
                        schema_builder.add_f64_field(&format!("{}_lon", field_name), opts);
 | 
			
		||||
 | 
			
		||||
                    fields.insert(
 | 
			
		||||
                        format!("{}_lat", field_name),
 | 
			
		||||
                        (
 | 
			
		||||
                            lat_field,
 | 
			
		||||
                            FieldDef::Numeric {
 | 
			
		||||
                                stored: *stored,
 | 
			
		||||
                                indexed: true,
 | 
			
		||||
                                fast: true,
 | 
			
		||||
                                precision: NumericType::F64,
 | 
			
		||||
                            },
 | 
			
		||||
                        ),
 | 
			
		||||
                    );
 | 
			
		||||
                    fields.insert(
 | 
			
		||||
                        format!("{}_lon", field_name),
 | 
			
		||||
                        (
 | 
			
		||||
                            lon_field,
 | 
			
		||||
                            FieldDef::Numeric {
 | 
			
		||||
                                stored: *stored,
 | 
			
		||||
                                indexed: true,
 | 
			
		||||
                                fast: true,
 | 
			
		||||
                                precision: NumericType::F64,
 | 
			
		||||
                            },
 | 
			
		||||
                        ),
 | 
			
		||||
                    );
 | 
			
		||||
                    continue; // Skip adding the geo field itself
 | 
			
		||||
                }
 | 
			
		||||
            };
 | 
			
		||||
 | 
			
		||||
            fields.insert(field_name.clone(), (field, field_def));
 | 
			
		||||
        }
 | 
			
		||||
 | 
			
		||||
        let schema = schema_builder.build();
 | 
			
		||||
        let index_schema = IndexSchema {
 | 
			
		||||
            schema: schema.clone(),
 | 
			
		||||
            fields,
 | 
			
		||||
            default_search_fields,
 | 
			
		||||
        };
 | 
			
		||||
 | 
			
		||||
        // Create or open index
 | 
			
		||||
        let dir = MmapDirectory::open(&index_path)
 | 
			
		||||
            .map_err(|e| DBError(format!("Failed to open index directory: {}", e)))?;
 | 
			
		||||
 | 
			
		||||
        let mut index = Index::open_or_create(dir, schema)
 | 
			
		||||
            .map_err(|e| DBError(format!("Failed to create index: {}", e)))?;
 | 
			
		||||
 | 
			
		||||
        // Configure tokenizers
 | 
			
		||||
        let tokenizer_manager = TokenizerManager::default();
 | 
			
		||||
        index.set_tokenizers(tokenizer_manager);
 | 
			
		||||
 | 
			
		||||
        let writer = index
 | 
			
		||||
            .writer(15_000_000)
 | 
			
		||||
            .map_err(|e| DBError(format!("Failed to create index writer: {}", e)))?;
 | 
			
		||||
 | 
			
		||||
        let reader = index
 | 
			
		||||
            .reader_builder()
 | 
			
		||||
            .reload_policy(ReloadPolicy::OnCommitWithDelay)
 | 
			
		||||
            .try_into()
 | 
			
		||||
            .map_err(|e| DBError(format!("Failed to create reader: {}", e)))?;
 | 
			
		||||
 | 
			
		||||
        let config = config.unwrap_or_default();
 | 
			
		||||
 | 
			
		||||
        Ok(TantivySearch {
 | 
			
		||||
            index,
 | 
			
		||||
            writer: Arc::new(RwLock::new(writer)),
 | 
			
		||||
            reader,
 | 
			
		||||
            index_schema,
 | 
			
		||||
            name,
 | 
			
		||||
            config,
 | 
			
		||||
        })
 | 
			
		||||
    }
 | 
			
		||||
 | 
			
		||||
    pub fn add_document_with_fields(
 | 
			
		||||
        &self,
 | 
			
		||||
        doc_id: &str,
 | 
			
		||||
        fields: HashMap<String, String>,
 | 
			
		||||
    ) -> Result<(), DBError> {
 | 
			
		||||
        let mut writer = self
 | 
			
		||||
            .writer
 | 
			
		||||
            .write()
 | 
			
		||||
            .map_err(|e| DBError(format!("Failed to acquire writer lock: {}", e)))?;
 | 
			
		||||
 | 
			
		||||
        // Delete existing document with same ID
 | 
			
		||||
        if let Some((id_field, _)) = self.index_schema.fields.get("_id") {
 | 
			
		||||
            writer.delete_term(Term::from_field_text(*id_field, doc_id));
 | 
			
		||||
        }
 | 
			
		||||
 | 
			
		||||
        // Create new document
 | 
			
		||||
        let mut doc = tantivy::doc!();
 | 
			
		||||
 | 
			
		||||
        // Add document ID
 | 
			
		||||
        if let Some((id_field, _)) = self.index_schema.fields.get("_id") {
 | 
			
		||||
            doc.add_text(*id_field, doc_id);
 | 
			
		||||
        }
 | 
			
		||||
 | 
			
		||||
        // Add other fields based on schema
 | 
			
		||||
        for (field_name, field_value) in fields {
 | 
			
		||||
            if let Some((field, field_def)) = self.index_schema.fields.get(&field_name) {
 | 
			
		||||
                match field_def {
 | 
			
		||||
                    FieldDef::Text { .. } => {
 | 
			
		||||
                        doc.add_text(*field, &field_value);
 | 
			
		||||
                    }
 | 
			
		||||
                    FieldDef::Numeric { precision, .. } => match precision {
 | 
			
		||||
                        NumericType::I64 => {
 | 
			
		||||
                            if let Ok(v) = field_value.parse::<i64>() {
 | 
			
		||||
                                doc.add_i64(*field, v);
 | 
			
		||||
                            }
 | 
			
		||||
                        }
 | 
			
		||||
                        NumericType::U64 => {
 | 
			
		||||
                            if let Ok(v) = field_value.parse::<u64>() {
 | 
			
		||||
                                doc.add_u64(*field, v);
 | 
			
		||||
                            }
 | 
			
		||||
                        }
 | 
			
		||||
                        NumericType::F64 => {
 | 
			
		||||
                            if let Ok(v) = field_value.parse::<f64>() {
 | 
			
		||||
                                doc.add_f64(*field, v);
 | 
			
		||||
                            }
 | 
			
		||||
                        }
 | 
			
		||||
                        NumericType::Date => {
 | 
			
		||||
                            if let Ok(v) = field_value.parse::<i64>() {
 | 
			
		||||
                                doc.add_date(*field, DateTime::from_timestamp_millis(v));
 | 
			
		||||
                            }
 | 
			
		||||
                        }
 | 
			
		||||
                    },
 | 
			
		||||
                    FieldDef::Tag {
 | 
			
		||||
                        separator,
 | 
			
		||||
                        case_sensitive,
 | 
			
		||||
                        ..
 | 
			
		||||
                    } => {
 | 
			
		||||
                        let tags = if !case_sensitive {
 | 
			
		||||
                            field_value.to_lowercase()
 | 
			
		||||
                        } else {
 | 
			
		||||
                            field_value.clone()
 | 
			
		||||
                        };
 | 
			
		||||
 | 
			
		||||
                        // Store tags as separate terms for efficient filtering
 | 
			
		||||
                        for tag in tags.split(separator.as_str()) {
 | 
			
		||||
                            doc.add_text(*field, tag.trim());
 | 
			
		||||
                        }
 | 
			
		||||
                    }
 | 
			
		||||
                    FieldDef::Geo { .. } => {
 | 
			
		||||
                        // Parse "lat,lon" format
 | 
			
		||||
                        let parts: Vec<&str> = field_value.split(',').collect();
 | 
			
		||||
                        if parts.len() == 2 {
 | 
			
		||||
                            if let (Ok(lat), Ok(lon)) =
 | 
			
		||||
                                (parts[0].parse::<f64>(), parts[1].parse::<f64>())
 | 
			
		||||
                            {
 | 
			
		||||
                                if let Some((lat_field, _)) =
 | 
			
		||||
                                    self.index_schema.fields.get(&format!("{}_lat", field_name))
 | 
			
		||||
                                {
 | 
			
		||||
                                    doc.add_f64(*lat_field, lat);
 | 
			
		||||
                                }
 | 
			
		||||
                                if let Some((lon_field, _)) =
 | 
			
		||||
                                    self.index_schema.fields.get(&format!("{}_lon", field_name))
 | 
			
		||||
                                {
 | 
			
		||||
                                    doc.add_f64(*lon_field, lon);
 | 
			
		||||
                                }
 | 
			
		||||
                            }
 | 
			
		||||
                        }
 | 
			
		||||
                    }
 | 
			
		||||
                }
 | 
			
		||||
            }
 | 
			
		||||
        }
 | 
			
		||||
 | 
			
		||||
        writer
 | 
			
		||||
            .add_document(doc)
 | 
			
		||||
            .map_err(|e| DBError(format!("Failed to add document: {}", e)))?;
 | 
			
		||||
 | 
			
		||||
        writer
 | 
			
		||||
            .commit()
 | 
			
		||||
            .map_err(|e| DBError(format!("Failed to commit: {}", e)))?;
 | 
			
		||||
 | 
			
		||||
        Ok(())
 | 
			
		||||
    }
 | 
			
		||||
 | 
			
		||||
    pub fn search_with_options(
 | 
			
		||||
        &self,
 | 
			
		||||
        query_str: &str,
 | 
			
		||||
        options: SearchOptions,
 | 
			
		||||
    ) -> Result<SearchResults, DBError> {
 | 
			
		||||
        let searcher = self.reader.searcher();
 | 
			
		||||
 | 
			
		||||
        // Parse query based on search fields
 | 
			
		||||
        let query: Box<dyn Query> = if self.index_schema.default_search_fields.is_empty() {
 | 
			
		||||
            return Err(DBError(
 | 
			
		||||
                "No searchable fields defined in schema".to_string(),
 | 
			
		||||
            ));
 | 
			
		||||
        } else {
 | 
			
		||||
            let query_parser = QueryParser::for_index(
 | 
			
		||||
                &self.index,
 | 
			
		||||
                self.index_schema.default_search_fields.clone(),
 | 
			
		||||
            );
 | 
			
		||||
 | 
			
		||||
            Box::new(
 | 
			
		||||
                query_parser
 | 
			
		||||
                    .parse_query(query_str)
 | 
			
		||||
                    .map_err(|e| DBError(format!("Failed to parse query: {}", e)))?,
 | 
			
		||||
            )
 | 
			
		||||
        };
 | 
			
		||||
 | 
			
		||||
        // Apply filters if any
 | 
			
		||||
        let final_query = if !options.filters.is_empty() {
 | 
			
		||||
            let mut clauses: Vec<(Occur, Box<dyn Query>)> = vec![(Occur::Must, query)];
 | 
			
		||||
 | 
			
		||||
            // Add filters
 | 
			
		||||
            for filter in options.filters {
 | 
			
		||||
                if let Some((field, _)) = self.index_schema.fields.get(&filter.field) {
 | 
			
		||||
                    match filter.filter_type {
 | 
			
		||||
                        FilterType::Equals(value) => {
 | 
			
		||||
                            let term_query = TermQuery::new(
 | 
			
		||||
                                Term::from_field_text(*field, &value),
 | 
			
		||||
                                tantivy::schema::IndexRecordOption::Basic,
 | 
			
		||||
                            );
 | 
			
		||||
                            clauses.push((Occur::Must, Box::new(term_query)));
 | 
			
		||||
                        }
 | 
			
		||||
                        FilterType::Range { min: _, max: _ } => {
 | 
			
		||||
                            // Would need numeric field handling here
 | 
			
		||||
                            // Simplified for now
 | 
			
		||||
                        }
 | 
			
		||||
                        FilterType::InSet(values) => {
 | 
			
		||||
                            let mut sub_clauses: Vec<(Occur, Box<dyn Query>)> = vec![];
 | 
			
		||||
                            for value in values {
 | 
			
		||||
                                let term_query = TermQuery::new(
 | 
			
		||||
                                    Term::from_field_text(*field, &value),
 | 
			
		||||
                                    tantivy::schema::IndexRecordOption::Basic,
 | 
			
		||||
                                );
 | 
			
		||||
                                sub_clauses.push((Occur::Should, Box::new(term_query)));
 | 
			
		||||
                            }
 | 
			
		||||
                            clauses.push((Occur::Must, Box::new(BooleanQuery::new(sub_clauses))));
 | 
			
		||||
                        }
 | 
			
		||||
                    }
 | 
			
		||||
                }
 | 
			
		||||
            }
 | 
			
		||||
 | 
			
		||||
            Box::new(BooleanQuery::new(clauses))
 | 
			
		||||
        } else {
 | 
			
		||||
            query
 | 
			
		||||
        };
 | 
			
		||||
 | 
			
		||||
        // Execute search
 | 
			
		||||
        let top_docs = searcher
 | 
			
		||||
            .search(
 | 
			
		||||
                &*final_query,
 | 
			
		||||
                &TopDocs::with_limit(options.limit + options.offset),
 | 
			
		||||
            )
 | 
			
		||||
            .map_err(|e| DBError(format!("Search failed: {}", e)))?;
 | 
			
		||||
 | 
			
		||||
        let total_hits = top_docs.len();
 | 
			
		||||
        let mut documents = Vec::new();
 | 
			
		||||
 | 
			
		||||
        for (score, doc_address) in top_docs.iter().skip(options.offset).take(options.limit) {
 | 
			
		||||
            let retrieved_doc: TantivyDocument = searcher
 | 
			
		||||
                .doc(*doc_address)
 | 
			
		||||
                .map_err(|e| DBError(format!("Failed to retrieve doc: {}", e)))?;
 | 
			
		||||
 | 
			
		||||
            let mut doc_fields = HashMap::new();
 | 
			
		||||
 | 
			
		||||
            // Extract all stored fields
 | 
			
		||||
            for (field_name, (field, field_def)) in &self.index_schema.fields {
 | 
			
		||||
                match field_def {
 | 
			
		||||
                    FieldDef::Text { stored, .. } | FieldDef::Tag { stored, .. } => {
 | 
			
		||||
                        if *stored {
 | 
			
		||||
                            if let Some(value) = retrieved_doc.get_first(*field) {
 | 
			
		||||
                                if let Some(text) = value.as_str() {
 | 
			
		||||
                                    doc_fields.insert(field_name.clone(), text.to_string());
 | 
			
		||||
                                }
 | 
			
		||||
                            }
 | 
			
		||||
                        }
 | 
			
		||||
                    }
 | 
			
		||||
                    FieldDef::Numeric {
 | 
			
		||||
                        stored, precision, ..
 | 
			
		||||
                    } => {
 | 
			
		||||
                        if *stored {
 | 
			
		||||
                            let value_str = match precision {
 | 
			
		||||
                                NumericType::I64 => retrieved_doc
 | 
			
		||||
                                    .get_first(*field)
 | 
			
		||||
                                    .and_then(|v| v.as_i64())
 | 
			
		||||
                                    .map(|v| v.to_string()),
 | 
			
		||||
                                NumericType::U64 => retrieved_doc
 | 
			
		||||
                                    .get_first(*field)
 | 
			
		||||
                                    .and_then(|v| v.as_u64())
 | 
			
		||||
                                    .map(|v| v.to_string()),
 | 
			
		||||
                                NumericType::F64 => retrieved_doc
 | 
			
		||||
                                    .get_first(*field)
 | 
			
		||||
                                    .and_then(|v| v.as_f64())
 | 
			
		||||
                                    .map(|v| v.to_string()),
 | 
			
		||||
                                NumericType::Date => retrieved_doc
 | 
			
		||||
                                    .get_first(*field)
 | 
			
		||||
                                    .and_then(|v| v.as_datetime())
 | 
			
		||||
                                    .map(|v| v.into_timestamp_millis().to_string()),
 | 
			
		||||
                            };
 | 
			
		||||
 | 
			
		||||
                            if let Some(v) = value_str {
 | 
			
		||||
                                doc_fields.insert(field_name.clone(), v);
 | 
			
		||||
                            }
 | 
			
		||||
                        }
 | 
			
		||||
                    }
 | 
			
		||||
                    FieldDef::Geo { stored } => {
 | 
			
		||||
                        if *stored {
 | 
			
		||||
                            let lat_field = self
 | 
			
		||||
                                .index_schema
 | 
			
		||||
                                .fields
 | 
			
		||||
                                .get(&format!("{}_lat", field_name))
 | 
			
		||||
                                .unwrap()
 | 
			
		||||
                                .0;
 | 
			
		||||
                            let lon_field = self
 | 
			
		||||
                                .index_schema
 | 
			
		||||
                                .fields
 | 
			
		||||
                                .get(&format!("{}_lon", field_name))
 | 
			
		||||
                                .unwrap()
 | 
			
		||||
                                .0;
 | 
			
		||||
 | 
			
		||||
                            let lat = retrieved_doc.get_first(lat_field).and_then(|v| v.as_f64());
 | 
			
		||||
                            let lon = retrieved_doc.get_first(lon_field).and_then(|v| v.as_f64());
 | 
			
		||||
 | 
			
		||||
                            if let (Some(lat), Some(lon)) = (lat, lon) {
 | 
			
		||||
                                doc_fields.insert(field_name.clone(), format!("{},{}", lat, lon));
 | 
			
		||||
                            }
 | 
			
		||||
                        }
 | 
			
		||||
                    }
 | 
			
		||||
                }
 | 
			
		||||
            }
 | 
			
		||||
 | 
			
		||||
            documents.push(SearchDocument {
 | 
			
		||||
                fields: doc_fields,
 | 
			
		||||
                score: *score,
 | 
			
		||||
            });
 | 
			
		||||
        }
 | 
			
		||||
 | 
			
		||||
        Ok(SearchResults {
 | 
			
		||||
            total: total_hits,
 | 
			
		||||
            documents,
 | 
			
		||||
        })
 | 
			
		||||
    }
 | 
			
		||||
 | 
			
		||||
    pub fn get_info(&self) -> Result<IndexInfo, DBError> {
 | 
			
		||||
        let searcher = self.reader.searcher();
 | 
			
		||||
        let num_docs = searcher.num_docs();
 | 
			
		||||
 | 
			
		||||
        let fields_info: Vec<FieldInfo> = self
 | 
			
		||||
            .index_schema
 | 
			
		||||
            .fields
 | 
			
		||||
            .iter()
 | 
			
		||||
            .map(|(name, (_, def))| FieldInfo {
 | 
			
		||||
                name: name.clone(),
 | 
			
		||||
                field_type: format!("{:?}", def),
 | 
			
		||||
            })
 | 
			
		||||
            .collect();
 | 
			
		||||
 | 
			
		||||
        Ok(IndexInfo {
 | 
			
		||||
            name: self.name.clone(),
 | 
			
		||||
            num_docs,
 | 
			
		||||
            fields: fields_info,
 | 
			
		||||
            config: self.config.clone(),
 | 
			
		||||
        })
 | 
			
		||||
    }
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
#[derive(Debug, Clone)]
 | 
			
		||||
pub struct SearchOptions {
 | 
			
		||||
    pub limit: usize,
 | 
			
		||||
    pub offset: usize,
 | 
			
		||||
    pub filters: Vec<Filter>,
 | 
			
		||||
    pub sort_by: Option<String>,
 | 
			
		||||
    pub return_fields: Option<Vec<String>>,
 | 
			
		||||
    pub highlight: bool,
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
impl Default for SearchOptions {
 | 
			
		||||
    fn default() -> Self {
 | 
			
		||||
        SearchOptions {
 | 
			
		||||
            limit: 10,
 | 
			
		||||
            offset: 0,
 | 
			
		||||
            filters: vec![],
 | 
			
		||||
            sort_by: None,
 | 
			
		||||
            return_fields: None,
 | 
			
		||||
            highlight: false,
 | 
			
		||||
        }
 | 
			
		||||
    }
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
#[derive(Debug, Clone)]
 | 
			
		||||
pub struct Filter {
 | 
			
		||||
    pub field: String,
 | 
			
		||||
    pub filter_type: FilterType,
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
#[derive(Debug, Clone)]
 | 
			
		||||
pub enum FilterType {
 | 
			
		||||
    Equals(String),
 | 
			
		||||
    Range { min: String, max: String },
 | 
			
		||||
    InSet(Vec<String>),
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
#[derive(Debug)]
 | 
			
		||||
pub struct SearchResults {
 | 
			
		||||
    pub total: usize,
 | 
			
		||||
    pub documents: Vec<SearchDocument>,
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
#[derive(Debug)]
 | 
			
		||||
pub struct SearchDocument {
 | 
			
		||||
    pub fields: HashMap<String, String>,
 | 
			
		||||
    pub score: f32,
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
#[derive(Debug, Serialize, Deserialize)]
 | 
			
		||||
pub struct IndexInfo {
 | 
			
		||||
    pub name: String,
 | 
			
		||||
    pub num_docs: u64,
 | 
			
		||||
    pub fields: Vec<FieldInfo>,
 | 
			
		||||
    pub config: IndexConfig,
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
#[derive(Debug, Serialize, Deserialize)]
 | 
			
		||||
pub struct FieldInfo {
 | 
			
		||||
    pub name: String,
 | 
			
		||||
    pub field_type: String,
 | 
			
		||||
}
 | 
			
		||||
@@ -1,4 +1,4 @@
 | 
			
		||||
use herodb::{server::Server, options::DBOption};
 | 
			
		||||
use herodb::{options::DBOption, server::Server};
 | 
			
		||||
use std::time::Duration;
 | 
			
		||||
use tokio::io::{AsyncReadExt, AsyncWriteExt};
 | 
			
		||||
use tokio::net::TcpStream;
 | 
			
		||||
@@ -47,17 +47,31 @@ async fn debug_hset_simple() {
 | 
			
		||||
 | 
			
		||||
    sleep(Duration::from_millis(200)).await;
 | 
			
		||||
 | 
			
		||||
    let mut stream = TcpStream::connect(format!("127.0.0.1:{}", port)).await.unwrap();
 | 
			
		||||
    let mut stream = TcpStream::connect(format!("127.0.0.1:{}", port))
 | 
			
		||||
        .await
 | 
			
		||||
        .unwrap();
 | 
			
		||||
 | 
			
		||||
    // Test simple HSET
 | 
			
		||||
    println!("Testing HSET...");
 | 
			
		||||
    let response = send_command(&mut stream, "*4\r\n$4\r\nHSET\r\n$4\r\nhash\r\n$6\r\nfield1\r\n$6\r\nvalue1\r\n").await;
 | 
			
		||||
    let response = send_command(
 | 
			
		||||
        &mut stream,
 | 
			
		||||
        "*4\r\n$4\r\nHSET\r\n$4\r\nhash\r\n$6\r\nfield1\r\n$6\r\nvalue1\r\n",
 | 
			
		||||
    )
 | 
			
		||||
    .await;
 | 
			
		||||
    println!("HSET response: {}", response);
 | 
			
		||||
    assert!(response.contains("1"), "Expected '1' but got: {}", response);
 | 
			
		||||
 | 
			
		||||
    // Test HGET
 | 
			
		||||
    println!("Testing HGET...");
 | 
			
		||||
    let response = send_command(&mut stream, "*3\r\n$4\r\nHGET\r\n$4\r\nhash\r\n$6\r\nfield1\r\n").await;
 | 
			
		||||
    let response = send_command(
 | 
			
		||||
        &mut stream,
 | 
			
		||||
        "*3\r\n$4\r\nHGET\r\n$4\r\nhash\r\n$6\r\nfield1\r\n",
 | 
			
		||||
    )
 | 
			
		||||
    .await;
 | 
			
		||||
    println!("HGET response: {}", response);
 | 
			
		||||
    assert!(response.contains("value1"), "Expected 'value1' but got: {}", response);
 | 
			
		||||
    assert!(
 | 
			
		||||
        response.contains("value1"),
 | 
			
		||||
        "Expected 'value1' but got: {}",
 | 
			
		||||
        response
 | 
			
		||||
    );
 | 
			
		||||
}
 | 
			
		||||
@@ -1,4 +1,4 @@
 | 
			
		||||
use herodb::{server::Server, options::DBOption};
 | 
			
		||||
use herodb::{options::DBOption, server::Server};
 | 
			
		||||
use std::time::Duration;
 | 
			
		||||
use tokio::io::{AsyncReadExt, AsyncWriteExt};
 | 
			
		||||
use tokio::net::TcpStream;
 | 
			
		||||
@@ -53,5 +53,9 @@ async fn debug_hset_return_value() {
 | 
			
		||||
    println!("Response bytes: {:?}", &buffer[..n]);
 | 
			
		||||
 | 
			
		||||
    // Check if response contains "1"
 | 
			
		||||
    assert!(response.contains("1"), "Expected response to contain '1', got: {}", response);
 | 
			
		||||
    assert!(
 | 
			
		||||
        response.contains("1"),
 | 
			
		||||
        "Expected response to contain '1', got: {}",
 | 
			
		||||
        response
 | 
			
		||||
    );
 | 
			
		||||
}
 | 
			
		||||
@@ -1,11 +1,14 @@
 | 
			
		||||
use herodb::protocol::Protocol;
 | 
			
		||||
use herodb::cmd::Cmd;
 | 
			
		||||
use herodb::protocol::Protocol;
 | 
			
		||||
 | 
			
		||||
#[test]
 | 
			
		||||
fn test_protocol_parsing() {
 | 
			
		||||
    // Test TYPE command parsing
 | 
			
		||||
    let type_cmd = "*2\r\n$4\r\nTYPE\r\n$7\r\nnoexist\r\n";
 | 
			
		||||
    println!("Parsing TYPE command: {}", type_cmd.replace("\r\n", "\\r\\n"));
 | 
			
		||||
    println!(
 | 
			
		||||
        "Parsing TYPE command: {}",
 | 
			
		||||
        type_cmd.replace("\r\n", "\\r\\n")
 | 
			
		||||
    );
 | 
			
		||||
 | 
			
		||||
    match Protocol::from(type_cmd) {
 | 
			
		||||
        Ok((protocol, _)) => {
 | 
			
		||||
@@ -20,7 +23,10 @@ fn test_protocol_parsing() {
 | 
			
		||||
 | 
			
		||||
    // Test HEXISTS command parsing
 | 
			
		||||
    let hexists_cmd = "*3\r\n$7\r\nHEXISTS\r\n$4\r\nhash\r\n$7\r\nnoexist\r\n";
 | 
			
		||||
    println!("\nParsing HEXISTS command: {}", hexists_cmd.replace("\r\n", "\\r\\n"));
 | 
			
		||||
    println!(
 | 
			
		||||
        "\nParsing HEXISTS command: {}",
 | 
			
		||||
        hexists_cmd.replace("\r\n", "\\r\\n")
 | 
			
		||||
    );
 | 
			
		||||
 | 
			
		||||
    match Protocol::from(hexists_cmd) {
 | 
			
		||||
        Ok((protocol, _)) => {
 | 
			
		||||
 
 | 
			
		||||
@@ -206,7 +206,9 @@ async fn test_expiration(conn: &mut Connection) {
 | 
			
		||||
async fn test_scan_operations(conn: &mut Connection) {
 | 
			
		||||
    cleanup_keys(conn).await;
 | 
			
		||||
    for i in 0..5 {
 | 
			
		||||
        let _: () = conn.set(format!("key{}", i), format!("value{}", i)).unwrap();
 | 
			
		||||
        let _: () = conn
 | 
			
		||||
            .set(format!("key{}", i), format!("value{}", i))
 | 
			
		||||
            .unwrap();
 | 
			
		||||
    }
 | 
			
		||||
    let result: (u64, Vec<String>) = redis::cmd("SCAN")
 | 
			
		||||
        .arg(0)
 | 
			
		||||
@@ -253,7 +255,9 @@ async fn test_scan_with_count(conn: &mut Connection) {
 | 
			
		||||
async fn test_hscan_operations(conn: &mut Connection) {
 | 
			
		||||
    cleanup_keys(conn).await;
 | 
			
		||||
    for i in 0..3 {
 | 
			
		||||
        let _: () = conn.hset("testhash", format!("field{}", i), format!("value{}", i)).unwrap();
 | 
			
		||||
        let _: () = conn
 | 
			
		||||
            .hset("testhash", format!("field{}", i), format!("value{}", i))
 | 
			
		||||
            .unwrap();
 | 
			
		||||
    }
 | 
			
		||||
    let result: (u64, Vec<String>) = redis::cmd("HSCAN")
 | 
			
		||||
        .arg("testhash")
 | 
			
		||||
@@ -273,8 +277,16 @@ async fn test_hscan_operations(conn: &mut Connection) {
 | 
			
		||||
async fn test_transaction_operations(conn: &mut Connection) {
 | 
			
		||||
    cleanup_keys(conn).await;
 | 
			
		||||
    let _: () = redis::cmd("MULTI").query(conn).unwrap();
 | 
			
		||||
    let _: () = redis::cmd("SET").arg("key1").arg("value1").query(conn).unwrap();
 | 
			
		||||
    let _: () = redis::cmd("SET").arg("key2").arg("value2").query(conn).unwrap();
 | 
			
		||||
    let _: () = redis::cmd("SET")
 | 
			
		||||
        .arg("key1")
 | 
			
		||||
        .arg("value1")
 | 
			
		||||
        .query(conn)
 | 
			
		||||
        .unwrap();
 | 
			
		||||
    let _: () = redis::cmd("SET")
 | 
			
		||||
        .arg("key2")
 | 
			
		||||
        .arg("value2")
 | 
			
		||||
        .query(conn)
 | 
			
		||||
        .unwrap();
 | 
			
		||||
    let _: Vec<String> = redis::cmd("EXEC").query(conn).unwrap();
 | 
			
		||||
    let result: String = conn.get("key1").unwrap();
 | 
			
		||||
    assert_eq!(result, "value1");
 | 
			
		||||
@@ -286,7 +298,11 @@ async fn test_transaction_operations(conn: &mut Connection) {
 | 
			
		||||
async fn test_discard_transaction(conn: &mut Connection) {
 | 
			
		||||
    cleanup_keys(conn).await;
 | 
			
		||||
    let _: () = redis::cmd("MULTI").query(conn).unwrap();
 | 
			
		||||
    let _: () = redis::cmd("SET").arg("discard").arg("value").query(conn).unwrap();
 | 
			
		||||
    let _: () = redis::cmd("SET")
 | 
			
		||||
        .arg("discard")
 | 
			
		||||
        .arg("value")
 | 
			
		||||
        .query(conn)
 | 
			
		||||
        .unwrap();
 | 
			
		||||
    let _: () = redis::cmd("DISCARD").query(conn).unwrap();
 | 
			
		||||
    let result: Option<String> = conn.get("discard").unwrap();
 | 
			
		||||
    assert_eq!(result, None);
 | 
			
		||||
@@ -306,7 +322,6 @@ async fn test_type_command(conn: &mut Connection) {
 | 
			
		||||
    cleanup_keys(conn).await;
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
async fn test_info_command(conn: &mut Connection) {
 | 
			
		||||
    cleanup_keys(conn).await;
 | 
			
		||||
    let result: String = redis::cmd("INFO").query(conn).unwrap();
 | 
			
		||||
 
 | 
			
		||||
@@ -1,4 +1,4 @@
 | 
			
		||||
use herodb::{server::Server, options::DBOption};
 | 
			
		||||
use herodb::{options::DBOption, server::Server};
 | 
			
		||||
use std::time::Duration;
 | 
			
		||||
use tokio::io::{AsyncReadExt, AsyncWriteExt};
 | 
			
		||||
use tokio::net::TcpStream;
 | 
			
		||||
@@ -99,7 +99,11 @@ async fn test_string_operations() {
 | 
			
		||||
    let mut stream = connect_to_server(port).await;
 | 
			
		||||
 | 
			
		||||
    // Test SET
 | 
			
		||||
    let response = send_command(&mut stream, "*3\r\n$3\r\nSET\r\n$3\r\nkey\r\n$5\r\nvalue\r\n").await;
 | 
			
		||||
    let response = send_command(
 | 
			
		||||
        &mut stream,
 | 
			
		||||
        "*3\r\n$3\r\nSET\r\n$3\r\nkey\r\n$5\r\nvalue\r\n",
 | 
			
		||||
    )
 | 
			
		||||
    .await;
 | 
			
		||||
    assert!(response.contains("OK"));
 | 
			
		||||
 | 
			
		||||
    // Test GET
 | 
			
		||||
@@ -148,7 +152,11 @@ async fn test_incr_operations() {
 | 
			
		||||
    assert!(response.contains("2"));
 | 
			
		||||
 | 
			
		||||
    // Test INCR on string value (should fail)
 | 
			
		||||
    send_command(&mut stream, "*3\r\n$3\r\nSET\r\n$6\r\nstring\r\n$5\r\nhello\r\n").await;
 | 
			
		||||
    send_command(
 | 
			
		||||
        &mut stream,
 | 
			
		||||
        "*3\r\n$3\r\nSET\r\n$6\r\nstring\r\n$5\r\nhello\r\n",
 | 
			
		||||
    )
 | 
			
		||||
    .await;
 | 
			
		||||
    let response = send_command(&mut stream, "*2\r\n$4\r\nINCR\r\n$6\r\nstring\r\n").await;
 | 
			
		||||
    assert!(response.contains("ERR"));
 | 
			
		||||
}
 | 
			
		||||
@@ -174,11 +182,19 @@ async fn test_hash_operations() {
 | 
			
		||||
    let mut stream = connect_to_server(port).await;
 | 
			
		||||
 | 
			
		||||
    // Test HSET
 | 
			
		||||
    let response = send_command(&mut stream, "*4\r\n$4\r\nHSET\r\n$4\r\nhash\r\n$6\r\nfield1\r\n$6\r\nvalue1\r\n").await;
 | 
			
		||||
    let response = send_command(
 | 
			
		||||
        &mut stream,
 | 
			
		||||
        "*4\r\n$4\r\nHSET\r\n$4\r\nhash\r\n$6\r\nfield1\r\n$6\r\nvalue1\r\n",
 | 
			
		||||
    )
 | 
			
		||||
    .await;
 | 
			
		||||
    assert!(response.contains("1")); // 1 new field
 | 
			
		||||
 | 
			
		||||
    // Test HGET
 | 
			
		||||
    let response = send_command(&mut stream, "*3\r\n$4\r\nHGET\r\n$4\r\nhash\r\n$6\r\nfield1\r\n").await;
 | 
			
		||||
    let response = send_command(
 | 
			
		||||
        &mut stream,
 | 
			
		||||
        "*3\r\n$4\r\nHGET\r\n$4\r\nhash\r\n$6\r\nfield1\r\n",
 | 
			
		||||
    )
 | 
			
		||||
    .await;
 | 
			
		||||
    assert!(response.contains("value1"));
 | 
			
		||||
 | 
			
		||||
    // Test HSET multiple fields
 | 
			
		||||
@@ -197,14 +213,26 @@ async fn test_hash_operations() {
 | 
			
		||||
    assert!(response.contains("3"));
 | 
			
		||||
 | 
			
		||||
    // Test HEXISTS
 | 
			
		||||
    let response = send_command(&mut stream, "*3\r\n$7\r\nHEXISTS\r\n$4\r\nhash\r\n$6\r\nfield1\r\n").await;
 | 
			
		||||
    let response = send_command(
 | 
			
		||||
        &mut stream,
 | 
			
		||||
        "*3\r\n$7\r\nHEXISTS\r\n$4\r\nhash\r\n$6\r\nfield1\r\n",
 | 
			
		||||
    )
 | 
			
		||||
    .await;
 | 
			
		||||
    assert!(response.contains("1"));
 | 
			
		||||
 | 
			
		||||
    let response = send_command(&mut stream, "*3\r\n$7\r\nHEXISTS\r\n$4\r\nhash\r\n$7\r\nnoexist\r\n").await;
 | 
			
		||||
    let response = send_command(
 | 
			
		||||
        &mut stream,
 | 
			
		||||
        "*3\r\n$7\r\nHEXISTS\r\n$4\r\nhash\r\n$7\r\nnoexist\r\n",
 | 
			
		||||
    )
 | 
			
		||||
    .await;
 | 
			
		||||
    assert!(response.contains("0"));
 | 
			
		||||
 | 
			
		||||
    // Test HDEL
 | 
			
		||||
    let response = send_command(&mut stream, "*3\r\n$4\r\nHDEL\r\n$4\r\nhash\r\n$6\r\nfield1\r\n").await;
 | 
			
		||||
    let response = send_command(
 | 
			
		||||
        &mut stream,
 | 
			
		||||
        "*3\r\n$4\r\nHDEL\r\n$4\r\nhash\r\n$6\r\nfield1\r\n",
 | 
			
		||||
    )
 | 
			
		||||
    .await;
 | 
			
		||||
    assert!(response.contains("1"));
 | 
			
		||||
 | 
			
		||||
    // Test HKEYS
 | 
			
		||||
@@ -240,7 +268,11 @@ async fn test_expiration() {
 | 
			
		||||
    let mut stream = connect_to_server(port).await;
 | 
			
		||||
 | 
			
		||||
    // Test SETEX (expire in 1 second)
 | 
			
		||||
    let response = send_command(&mut stream, "*5\r\n$3\r\nSET\r\n$6\r\nexpkey\r\n$5\r\nvalue\r\n$2\r\nEX\r\n$1\r\n1\r\n").await;
 | 
			
		||||
    let response = send_command(
 | 
			
		||||
        &mut stream,
 | 
			
		||||
        "*5\r\n$3\r\nSET\r\n$6\r\nexpkey\r\n$5\r\nvalue\r\n$2\r\nEX\r\n$1\r\n1\r\n",
 | 
			
		||||
    )
 | 
			
		||||
    .await;
 | 
			
		||||
    assert!(response.contains("OK"));
 | 
			
		||||
 | 
			
		||||
    // Test TTL
 | 
			
		||||
@@ -294,7 +326,11 @@ async fn test_scan_operations() {
 | 
			
		||||
    }
 | 
			
		||||
 | 
			
		||||
    // Test SCAN
 | 
			
		||||
    let response = send_command(&mut stream, "*6\r\n$4\r\nSCAN\r\n$1\r\n0\r\n$5\r\nMATCH\r\n$1\r\n*\r\n$5\r\nCOUNT\r\n$2\r\n10\r\n").await;
 | 
			
		||||
    let response = send_command(
 | 
			
		||||
        &mut stream,
 | 
			
		||||
        "*6\r\n$4\r\nSCAN\r\n$1\r\n0\r\n$5\r\nMATCH\r\n$1\r\n*\r\n$5\r\nCOUNT\r\n$2\r\n10\r\n",
 | 
			
		||||
    )
 | 
			
		||||
    .await;
 | 
			
		||||
    assert!(response.contains("key"));
 | 
			
		||||
 | 
			
		||||
    // Test KEYS
 | 
			
		||||
@@ -325,7 +361,10 @@ async fn test_hscan_operations() {
 | 
			
		||||
 | 
			
		||||
    // Set up hash data
 | 
			
		||||
    for i in 0..3 {
 | 
			
		||||
        let cmd = format!("*4\r\n$4\r\nHSET\r\n$8\r\ntesthash\r\n$6\r\nfield{}\r\n$6\r\nvalue{}\r\n", i, i);
 | 
			
		||||
        let cmd = format!(
 | 
			
		||||
            "*4\r\n$4\r\nHSET\r\n$8\r\ntesthash\r\n$6\r\nfield{}\r\n$6\r\nvalue{}\r\n",
 | 
			
		||||
            i, i
 | 
			
		||||
        );
 | 
			
		||||
        send_command(&mut stream, &cmd).await;
 | 
			
		||||
    }
 | 
			
		||||
 | 
			
		||||
@@ -360,10 +399,18 @@ async fn test_transaction_operations() {
 | 
			
		||||
    assert!(response.contains("OK"));
 | 
			
		||||
 | 
			
		||||
    // Test queued commands
 | 
			
		||||
    let response = send_command(&mut stream, "*3\r\n$3\r\nSET\r\n$4\r\nkey1\r\n$6\r\nvalue1\r\n").await;
 | 
			
		||||
    let response = send_command(
 | 
			
		||||
        &mut stream,
 | 
			
		||||
        "*3\r\n$3\r\nSET\r\n$4\r\nkey1\r\n$6\r\nvalue1\r\n",
 | 
			
		||||
    )
 | 
			
		||||
    .await;
 | 
			
		||||
    assert!(response.contains("QUEUED"));
 | 
			
		||||
 | 
			
		||||
    let response = send_command(&mut stream, "*3\r\n$3\r\nSET\r\n$4\r\nkey2\r\n$6\r\nvalue2\r\n").await;
 | 
			
		||||
    let response = send_command(
 | 
			
		||||
        &mut stream,
 | 
			
		||||
        "*3\r\n$3\r\nSET\r\n$4\r\nkey2\r\n$6\r\nvalue2\r\n",
 | 
			
		||||
    )
 | 
			
		||||
    .await;
 | 
			
		||||
    assert!(response.contains("QUEUED"));
 | 
			
		||||
 | 
			
		||||
    // Test EXEC
 | 
			
		||||
@@ -403,7 +450,11 @@ async fn test_discard_transaction() {
 | 
			
		||||
    assert!(response.contains("OK"));
 | 
			
		||||
 | 
			
		||||
    // Test queued command
 | 
			
		||||
    let response = send_command(&mut stream, "*3\r\n$3\r\nSET\r\n$7\r\ndiscard\r\n$5\r\nvalue\r\n").await;
 | 
			
		||||
    let response = send_command(
 | 
			
		||||
        &mut stream,
 | 
			
		||||
        "*3\r\n$3\r\nSET\r\n$7\r\ndiscard\r\n$5\r\nvalue\r\n",
 | 
			
		||||
    )
 | 
			
		||||
    .await;
 | 
			
		||||
    assert!(response.contains("QUEUED"));
 | 
			
		||||
 | 
			
		||||
    // Test DISCARD
 | 
			
		||||
@@ -436,12 +487,20 @@ async fn test_type_command() {
 | 
			
		||||
    let mut stream = connect_to_server(port).await;
 | 
			
		||||
 | 
			
		||||
    // Test string type
 | 
			
		||||
    send_command(&mut stream, "*3\r\n$3\r\nSET\r\n$6\r\nstring\r\n$5\r\nvalue\r\n").await;
 | 
			
		||||
    send_command(
 | 
			
		||||
        &mut stream,
 | 
			
		||||
        "*3\r\n$3\r\nSET\r\n$6\r\nstring\r\n$5\r\nvalue\r\n",
 | 
			
		||||
    )
 | 
			
		||||
    .await;
 | 
			
		||||
    let response = send_command(&mut stream, "*2\r\n$4\r\nTYPE\r\n$6\r\nstring\r\n").await;
 | 
			
		||||
    assert!(response.contains("string"));
 | 
			
		||||
 | 
			
		||||
    // Test hash type
 | 
			
		||||
    send_command(&mut stream, "*4\r\n$4\r\nHSET\r\n$4\r\nhash\r\n$5\r\nfield\r\n$5\r\nvalue\r\n").await;
 | 
			
		||||
    send_command(
 | 
			
		||||
        &mut stream,
 | 
			
		||||
        "*4\r\n$4\r\nHSET\r\n$4\r\nhash\r\n$5\r\nfield\r\n$5\r\nvalue\r\n",
 | 
			
		||||
    )
 | 
			
		||||
    .await;
 | 
			
		||||
    let response = send_command(&mut stream, "*2\r\n$4\r\nTYPE\r\n$4\r\nhash\r\n").await;
 | 
			
		||||
    assert!(response.contains("hash"));
 | 
			
		||||
 | 
			
		||||
@@ -471,12 +530,20 @@ async fn test_config_commands() {
 | 
			
		||||
    let mut stream = connect_to_server(port).await;
 | 
			
		||||
 | 
			
		||||
    // Test CONFIG GET databases
 | 
			
		||||
    let response = send_command(&mut stream, "*3\r\n$6\r\nCONFIG\r\n$3\r\nGET\r\n$9\r\ndatabases\r\n").await;
 | 
			
		||||
    let response = send_command(
 | 
			
		||||
        &mut stream,
 | 
			
		||||
        "*3\r\n$6\r\nCONFIG\r\n$3\r\nGET\r\n$9\r\ndatabases\r\n",
 | 
			
		||||
    )
 | 
			
		||||
    .await;
 | 
			
		||||
    assert!(response.contains("databases"));
 | 
			
		||||
    assert!(response.contains("16"));
 | 
			
		||||
 | 
			
		||||
    // Test CONFIG GET dir
 | 
			
		||||
    let response = send_command(&mut stream, "*3\r\n$6\r\nCONFIG\r\n$3\r\nGET\r\n$3\r\ndir\r\n").await;
 | 
			
		||||
    let response = send_command(
 | 
			
		||||
        &mut stream,
 | 
			
		||||
        "*3\r\n$6\r\nCONFIG\r\n$3\r\nGET\r\n$3\r\ndir\r\n",
 | 
			
		||||
    )
 | 
			
		||||
    .await;
 | 
			
		||||
    assert!(response.contains("dir"));
 | 
			
		||||
    assert!(response.contains("/tmp/herodb_test_config"));
 | 
			
		||||
}
 | 
			
		||||
@@ -531,8 +598,16 @@ async fn test_error_handling() {
 | 
			
		||||
    let mut stream = connect_to_server(port).await;
 | 
			
		||||
 | 
			
		||||
    // Test WRONGTYPE error - try to use hash command on string
 | 
			
		||||
    send_command(&mut stream, "*3\r\n$3\r\nSET\r\n$6\r\nstring\r\n$5\r\nvalue\r\n").await;
 | 
			
		||||
    let response = send_command(&mut stream, "*3\r\n$4\r\nHGET\r\n$6\r\nstring\r\n$5\r\nfield\r\n").await;
 | 
			
		||||
    send_command(
 | 
			
		||||
        &mut stream,
 | 
			
		||||
        "*3\r\n$3\r\nSET\r\n$6\r\nstring\r\n$5\r\nvalue\r\n",
 | 
			
		||||
    )
 | 
			
		||||
    .await;
 | 
			
		||||
    let response = send_command(
 | 
			
		||||
        &mut stream,
 | 
			
		||||
        "*3\r\n$4\r\nHGET\r\n$6\r\nstring\r\n$5\r\nfield\r\n",
 | 
			
		||||
    )
 | 
			
		||||
    .await;
 | 
			
		||||
    assert!(response.contains("WRONGTYPE"));
 | 
			
		||||
 | 
			
		||||
    // Test unknown command
 | 
			
		||||
@@ -569,11 +644,19 @@ async fn test_list_operations() {
 | 
			
		||||
    let mut stream = connect_to_server(port).await;
 | 
			
		||||
 | 
			
		||||
    // Test LPUSH
 | 
			
		||||
    let response = send_command(&mut stream, "*4\r\n$5\r\nLPUSH\r\n$4\r\nlist\r\n$1\r\na\r\n$1\r\nb\r\n").await;
 | 
			
		||||
    let response = send_command(
 | 
			
		||||
        &mut stream,
 | 
			
		||||
        "*4\r\n$5\r\nLPUSH\r\n$4\r\nlist\r\n$1\r\na\r\n$1\r\nb\r\n",
 | 
			
		||||
    )
 | 
			
		||||
    .await;
 | 
			
		||||
    assert!(response.contains("2")); // 2 elements
 | 
			
		||||
 | 
			
		||||
    // Test RPUSH
 | 
			
		||||
    let response = send_command(&mut stream, "*4\r\n$5\r\nRPUSH\r\n$4\r\nlist\r\n$1\r\nc\r\n$1\r\nd\r\n").await;
 | 
			
		||||
    let response = send_command(
 | 
			
		||||
        &mut stream,
 | 
			
		||||
        "*4\r\n$5\r\nRPUSH\r\n$4\r\nlist\r\n$1\r\nc\r\n$1\r\nd\r\n",
 | 
			
		||||
    )
 | 
			
		||||
    .await;
 | 
			
		||||
    assert!(response.contains("4")); // 4 elements
 | 
			
		||||
 | 
			
		||||
    // Test LLEN
 | 
			
		||||
@@ -581,11 +664,22 @@ async fn test_list_operations() {
 | 
			
		||||
    assert!(response.contains("4"));
 | 
			
		||||
 | 
			
		||||
    // Test LRANGE
 | 
			
		||||
    let response = send_command(&mut stream, "*4\r\n$6\r\nLRANGE\r\n$4\r\nlist\r\n$1\r\n0\r\n$2\r\n-1\r\n").await;
 | 
			
		||||
    assert_eq!(response, "*4\r\n$1\r\nb\r\n$1\r\na\r\n$1\r\nc\r\n$1\r\nd\r\n");
 | 
			
		||||
    let response = send_command(
 | 
			
		||||
        &mut stream,
 | 
			
		||||
        "*4\r\n$6\r\nLRANGE\r\n$4\r\nlist\r\n$1\r\n0\r\n$2\r\n-1\r\n",
 | 
			
		||||
    )
 | 
			
		||||
    .await;
 | 
			
		||||
    assert_eq!(
 | 
			
		||||
        response,
 | 
			
		||||
        "*4\r\n$1\r\nb\r\n$1\r\na\r\n$1\r\nc\r\n$1\r\nd\r\n"
 | 
			
		||||
    );
 | 
			
		||||
 | 
			
		||||
    // Test LINDEX
 | 
			
		||||
    let response = send_command(&mut stream, "*3\r\n$6\r\nLINDEX\r\n$4\r\nlist\r\n$1\r\n0\r\n").await;
 | 
			
		||||
    let response = send_command(
 | 
			
		||||
        &mut stream,
 | 
			
		||||
        "*3\r\n$6\r\nLINDEX\r\n$4\r\nlist\r\n$1\r\n0\r\n",
 | 
			
		||||
    )
 | 
			
		||||
    .await;
 | 
			
		||||
    assert_eq!(response, "$1\r\nb\r\n");
 | 
			
		||||
 | 
			
		||||
    // Test LPOP
 | 
			
		||||
@@ -597,12 +691,24 @@ async fn test_list_operations() {
 | 
			
		||||
    assert_eq!(response, "$1\r\nd\r\n");
 | 
			
		||||
 | 
			
		||||
    // Test LREM
 | 
			
		||||
    send_command(&mut stream, "*3\r\n$5\r\nLPUSH\r\n$4\r\nlist\r\n$1\r\na\r\n").await; // list is now a, c, a
 | 
			
		||||
    let response = send_command(&mut stream, "*4\r\n$4\r\nLREM\r\n$4\r\nlist\r\n$1\r\n1\r\n$1\r\na\r\n").await;
 | 
			
		||||
    send_command(
 | 
			
		||||
        &mut stream,
 | 
			
		||||
        "*3\r\n$5\r\nLPUSH\r\n$4\r\nlist\r\n$1\r\na\r\n",
 | 
			
		||||
    )
 | 
			
		||||
    .await; // list is now a, c, a
 | 
			
		||||
    let response = send_command(
 | 
			
		||||
        &mut stream,
 | 
			
		||||
        "*4\r\n$4\r\nLREM\r\n$4\r\nlist\r\n$1\r\n1\r\n$1\r\na\r\n",
 | 
			
		||||
    )
 | 
			
		||||
    .await;
 | 
			
		||||
    assert!(response.contains("1"));
 | 
			
		||||
 | 
			
		||||
    // Test LTRIM
 | 
			
		||||
    let response = send_command(&mut stream, "*4\r\n$5\r\nLTRIM\r\n$4\r\nlist\r\n$1\r\n0\r\n$1\r\n0\r\n").await;
 | 
			
		||||
    let response = send_command(
 | 
			
		||||
        &mut stream,
 | 
			
		||||
        "*4\r\n$5\r\nLTRIM\r\n$4\r\nlist\r\n$1\r\n0\r\n$1\r\n0\r\n",
 | 
			
		||||
    )
 | 
			
		||||
    .await;
 | 
			
		||||
    assert!(response.contains("OK"));
 | 
			
		||||
    let response = send_command(&mut stream, "*2\r\n$4\r\nLLEN\r\n$4\r\nlist\r\n").await;
 | 
			
		||||
    assert!(response.contains("1"));
 | 
			
		||||
 
 | 
			
		||||
@@ -1,8 +1,8 @@
 | 
			
		||||
use herodb::{server::Server, options::DBOption};
 | 
			
		||||
use herodb::{options::DBOption, server::Server};
 | 
			
		||||
use std::time::Duration;
 | 
			
		||||
use tokio::time::sleep;
 | 
			
		||||
use tokio::io::{AsyncReadExt, AsyncWriteExt};
 | 
			
		||||
use tokio::net::TcpStream;
 | 
			
		||||
use tokio::time::sleep;
 | 
			
		||||
 | 
			
		||||
// Helper function to start a test server with clean data directory
 | 
			
		||||
async fn start_test_server(test_name: &str) -> (Server, u16) {
 | 
			
		||||
@@ -33,7 +33,9 @@ async fn start_test_server(test_name: &str) -> (Server, u16) {
 | 
			
		||||
 | 
			
		||||
// Helper function to send Redis command and get response
 | 
			
		||||
async fn send_redis_command(port: u16, command: &str) -> String {
 | 
			
		||||
    let mut stream = TcpStream::connect(format!("127.0.0.1:{}", port)).await.unwrap();
 | 
			
		||||
    let mut stream = TcpStream::connect(format!("127.0.0.1:{}", port))
 | 
			
		||||
        .await
 | 
			
		||||
        .unwrap();
 | 
			
		||||
    stream.write_all(command.as_bytes()).await.unwrap();
 | 
			
		||||
 | 
			
		||||
    let mut buffer = [0; 1024];
 | 
			
		||||
@@ -66,7 +68,8 @@ async fn test_basic_redis_functionality() {
 | 
			
		||||
    assert!(response.contains("PONG"));
 | 
			
		||||
 | 
			
		||||
    // Test SET
 | 
			
		||||
    let response = send_redis_command(port, "*3\r\n$3\r\nSET\r\n$3\r\nkey\r\n$5\r\nvalue\r\n").await;
 | 
			
		||||
    let response =
 | 
			
		||||
        send_redis_command(port, "*3\r\n$3\r\nSET\r\n$3\r\nkey\r\n$5\r\nvalue\r\n").await;
 | 
			
		||||
    assert!(response.contains("OK"));
 | 
			
		||||
 | 
			
		||||
    // Test GET
 | 
			
		||||
@@ -74,11 +77,16 @@ async fn test_basic_redis_functionality() {
 | 
			
		||||
    assert!(response.contains("value"));
 | 
			
		||||
 | 
			
		||||
    // Test HSET
 | 
			
		||||
    let response = send_redis_command(port, "*4\r\n$4\r\nHSET\r\n$4\r\nhash\r\n$5\r\nfield\r\n$5\r\nvalue\r\n").await;
 | 
			
		||||
    let response = send_redis_command(
 | 
			
		||||
        port,
 | 
			
		||||
        "*4\r\n$4\r\nHSET\r\n$4\r\nhash\r\n$5\r\nfield\r\n$5\r\nvalue\r\n",
 | 
			
		||||
    )
 | 
			
		||||
    .await;
 | 
			
		||||
    assert!(response.contains("1"));
 | 
			
		||||
 | 
			
		||||
    // Test HGET
 | 
			
		||||
    let response = send_redis_command(port, "*3\r\n$4\r\nHGET\r\n$4\r\nhash\r\n$5\r\nfield\r\n").await;
 | 
			
		||||
    let response =
 | 
			
		||||
        send_redis_command(port, "*3\r\n$4\r\nHGET\r\n$4\r\nhash\r\n$5\r\nfield\r\n").await;
 | 
			
		||||
    assert!(response.contains("value"));
 | 
			
		||||
 | 
			
		||||
    // Test EXISTS
 | 
			
		||||
@@ -94,8 +102,13 @@ async fn test_basic_redis_functionality() {
 | 
			
		||||
    assert!(response.contains("string"));
 | 
			
		||||
 | 
			
		||||
    // Test QUIT to close connection gracefully
 | 
			
		||||
    let mut stream = TcpStream::connect(format!("127.0.0.1:{}", port)).await.unwrap();
 | 
			
		||||
    stream.write_all("*1\r\n$4\r\nQUIT\r\n".as_bytes()).await.unwrap();
 | 
			
		||||
    let mut stream = TcpStream::connect(format!("127.0.0.1:{}", port))
 | 
			
		||||
        .await
 | 
			
		||||
        .unwrap();
 | 
			
		||||
    stream
 | 
			
		||||
        .write_all("*1\r\n$4\r\nQUIT\r\n".as_bytes())
 | 
			
		||||
        .await
 | 
			
		||||
        .unwrap();
 | 
			
		||||
    let mut buffer = [0; 1024];
 | 
			
		||||
    let n = stream.read(&mut buffer).await.unwrap();
 | 
			
		||||
    let response = String::from_utf8_lossy(&buffer[..n]);
 | 
			
		||||
@@ -142,7 +155,11 @@ async fn test_hash_operations() {
 | 
			
		||||
    assert!(response.contains("value2"));
 | 
			
		||||
 | 
			
		||||
    // Test HEXISTS
 | 
			
		||||
    let response = send_redis_command(port, "*3\r\n$7\r\nHEXISTS\r\n$4\r\nhash\r\n$6\r\nfield1\r\n").await;
 | 
			
		||||
    let response = send_redis_command(
 | 
			
		||||
        port,
 | 
			
		||||
        "*3\r\n$7\r\nHEXISTS\r\n$4\r\nhash\r\n$6\r\nfield1\r\n",
 | 
			
		||||
    )
 | 
			
		||||
    .await;
 | 
			
		||||
    assert!(response.contains("1"));
 | 
			
		||||
 | 
			
		||||
    // Test HLEN
 | 
			
		||||
@@ -185,39 +202,59 @@ async fn test_transaction_operations() {
 | 
			
		||||
    sleep(Duration::from_millis(100)).await;
 | 
			
		||||
 | 
			
		||||
    // Use a single connection for the transaction
 | 
			
		||||
    let mut stream = TcpStream::connect(format!("127.0.0.1:{}", port)).await.unwrap();
 | 
			
		||||
    let mut stream = TcpStream::connect(format!("127.0.0.1:{}", port))
 | 
			
		||||
        .await
 | 
			
		||||
        .unwrap();
 | 
			
		||||
 | 
			
		||||
    // Test MULTI
 | 
			
		||||
    stream.write_all("*1\r\n$5\r\nMULTI\r\n".as_bytes()).await.unwrap();
 | 
			
		||||
    stream
 | 
			
		||||
        .write_all("*1\r\n$5\r\nMULTI\r\n".as_bytes())
 | 
			
		||||
        .await
 | 
			
		||||
        .unwrap();
 | 
			
		||||
    let mut buffer = [0; 1024];
 | 
			
		||||
    let n = stream.read(&mut buffer).await.unwrap();
 | 
			
		||||
    let response = String::from_utf8_lossy(&buffer[..n]);
 | 
			
		||||
    assert!(response.contains("OK"));
 | 
			
		||||
 | 
			
		||||
    // Test queued commands
 | 
			
		||||
    stream.write_all("*3\r\n$3\r\nSET\r\n$4\r\nkey1\r\n$6\r\nvalue1\r\n".as_bytes()).await.unwrap();
 | 
			
		||||
    stream
 | 
			
		||||
        .write_all("*3\r\n$3\r\nSET\r\n$4\r\nkey1\r\n$6\r\nvalue1\r\n".as_bytes())
 | 
			
		||||
        .await
 | 
			
		||||
        .unwrap();
 | 
			
		||||
    let n = stream.read(&mut buffer).await.unwrap();
 | 
			
		||||
    let response = String::from_utf8_lossy(&buffer[..n]);
 | 
			
		||||
    assert!(response.contains("QUEUED"));
 | 
			
		||||
 | 
			
		||||
    stream.write_all("*3\r\n$3\r\nSET\r\n$4\r\nkey2\r\n$6\r\nvalue2\r\n".as_bytes()).await.unwrap();
 | 
			
		||||
    stream
 | 
			
		||||
        .write_all("*3\r\n$3\r\nSET\r\n$4\r\nkey2\r\n$6\r\nvalue2\r\n".as_bytes())
 | 
			
		||||
        .await
 | 
			
		||||
        .unwrap();
 | 
			
		||||
    let n = stream.read(&mut buffer).await.unwrap();
 | 
			
		||||
    let response = String::from_utf8_lossy(&buffer[..n]);
 | 
			
		||||
    assert!(response.contains("QUEUED"));
 | 
			
		||||
 | 
			
		||||
    // Test EXEC
 | 
			
		||||
    stream.write_all("*1\r\n$4\r\nEXEC\r\n".as_bytes()).await.unwrap();
 | 
			
		||||
    stream
 | 
			
		||||
        .write_all("*1\r\n$4\r\nEXEC\r\n".as_bytes())
 | 
			
		||||
        .await
 | 
			
		||||
        .unwrap();
 | 
			
		||||
    let n = stream.read(&mut buffer).await.unwrap();
 | 
			
		||||
    let response = String::from_utf8_lossy(&buffer[..n]);
 | 
			
		||||
    assert!(response.contains("OK")); // Should contain array of OK responses
 | 
			
		||||
 | 
			
		||||
    // Verify commands were executed
 | 
			
		||||
    stream.write_all("*2\r\n$3\r\nGET\r\n$4\r\nkey1\r\n".as_bytes()).await.unwrap();
 | 
			
		||||
    stream
 | 
			
		||||
        .write_all("*2\r\n$3\r\nGET\r\n$4\r\nkey1\r\n".as_bytes())
 | 
			
		||||
        .await
 | 
			
		||||
        .unwrap();
 | 
			
		||||
    let n = stream.read(&mut buffer).await.unwrap();
 | 
			
		||||
    let response = String::from_utf8_lossy(&buffer[..n]);
 | 
			
		||||
    assert!(response.contains("value1"));
 | 
			
		||||
 | 
			
		||||
    stream.write_all("*2\r\n$3\r\nGET\r\n$4\r\nkey2\r\n".as_bytes()).await.unwrap();
 | 
			
		||||
    stream
 | 
			
		||||
        .write_all("*2\r\n$3\r\nGET\r\n$4\r\nkey2\r\n".as_bytes())
 | 
			
		||||
        .await
 | 
			
		||||
        .unwrap();
 | 
			
		||||
    let n = stream.read(&mut buffer).await.unwrap();
 | 
			
		||||
    let response = String::from_utf8_lossy(&buffer[..n]);
 | 
			
		||||
    assert!(response.contains("value2"));
 | 
			
		||||
 
 | 
			
		||||
@@ -1,4 +1,4 @@
 | 
			
		||||
use herodb::{server::Server, options::DBOption};
 | 
			
		||||
use herodb::{options::DBOption, server::Server};
 | 
			
		||||
use std::time::Duration;
 | 
			
		||||
use tokio::io::{AsyncReadExt, AsyncWriteExt};
 | 
			
		||||
use tokio::net::TcpStream;
 | 
			
		||||
@@ -99,12 +99,24 @@ async fn test_hset_clean_db() {
 | 
			
		||||
    let mut stream = connect_to_server(port).await;
 | 
			
		||||
 | 
			
		||||
    // Test HSET - should return 1 for new field
 | 
			
		||||
    let response = send_command(&mut stream, "*4\r\n$4\r\nHSET\r\n$4\r\nhash\r\n$6\r\nfield1\r\n$6\r\nvalue1\r\n").await;
 | 
			
		||||
    let response = send_command(
 | 
			
		||||
        &mut stream,
 | 
			
		||||
        "*4\r\n$4\r\nHSET\r\n$4\r\nhash\r\n$6\r\nfield1\r\n$6\r\nvalue1\r\n",
 | 
			
		||||
    )
 | 
			
		||||
    .await;
 | 
			
		||||
    println!("HSET response: {}", response);
 | 
			
		||||
    assert!(response.contains("1"), "Expected HSET to return 1, got: {}", response);
 | 
			
		||||
    assert!(
 | 
			
		||||
        response.contains("1"),
 | 
			
		||||
        "Expected HSET to return 1, got: {}",
 | 
			
		||||
        response
 | 
			
		||||
    );
 | 
			
		||||
 | 
			
		||||
    // Test HGET
 | 
			
		||||
    let response = send_command(&mut stream, "*3\r\n$4\r\nHGET\r\n$4\r\nhash\r\n$6\r\nfield1\r\n").await;
 | 
			
		||||
    let response = send_command(
 | 
			
		||||
        &mut stream,
 | 
			
		||||
        "*3\r\n$4\r\nHGET\r\n$4\r\nhash\r\n$6\r\nfield1\r\n",
 | 
			
		||||
    )
 | 
			
		||||
    .await;
 | 
			
		||||
    println!("HGET response: {}", response);
 | 
			
		||||
    assert!(response.contains("value1"));
 | 
			
		||||
}
 | 
			
		||||
@@ -131,13 +143,21 @@ async fn test_type_command_simple() {
 | 
			
		||||
    let mut stream = connect_to_server(port).await;
 | 
			
		||||
 | 
			
		||||
    // Test string type
 | 
			
		||||
    send_command(&mut stream, "*3\r\n$3\r\nSET\r\n$6\r\nstring\r\n$5\r\nvalue\r\n").await;
 | 
			
		||||
    send_command(
 | 
			
		||||
        &mut stream,
 | 
			
		||||
        "*3\r\n$3\r\nSET\r\n$6\r\nstring\r\n$5\r\nvalue\r\n",
 | 
			
		||||
    )
 | 
			
		||||
    .await;
 | 
			
		||||
    let response = send_command(&mut stream, "*2\r\n$4\r\nTYPE\r\n$6\r\nstring\r\n").await;
 | 
			
		||||
    println!("TYPE string response: {}", response);
 | 
			
		||||
    assert!(response.contains("string"));
 | 
			
		||||
 | 
			
		||||
    // Test hash type
 | 
			
		||||
    send_command(&mut stream, "*4\r\n$4\r\nHSET\r\n$4\r\nhash\r\n$5\r\nfield\r\n$5\r\nvalue\r\n").await;
 | 
			
		||||
    send_command(
 | 
			
		||||
        &mut stream,
 | 
			
		||||
        "*4\r\n$4\r\nHSET\r\n$4\r\nhash\r\n$5\r\nfield\r\n$5\r\nvalue\r\n",
 | 
			
		||||
    )
 | 
			
		||||
    .await;
 | 
			
		||||
    let response = send_command(&mut stream, "*2\r\n$4\r\nTYPE\r\n$4\r\nhash\r\n").await;
 | 
			
		||||
    println!("TYPE hash response: {}", response);
 | 
			
		||||
    assert!(response.contains("hash"));
 | 
			
		||||
@@ -145,7 +165,11 @@ async fn test_type_command_simple() {
 | 
			
		||||
    // Test non-existent key
 | 
			
		||||
    let response = send_command(&mut stream, "*2\r\n$4\r\nTYPE\r\n$7\r\nnoexist\r\n").await;
 | 
			
		||||
    println!("TYPE noexist response: {}", response);
 | 
			
		||||
    assert!(response.contains("none"), "Expected 'none' for non-existent key, got: {}", response);
 | 
			
		||||
    assert!(
 | 
			
		||||
        response.contains("none"),
 | 
			
		||||
        "Expected 'none' for non-existent key, got: {}",
 | 
			
		||||
        response
 | 
			
		||||
    );
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
#[tokio::test]
 | 
			
		||||
@@ -170,15 +194,31 @@ async fn test_hexists_simple() {
 | 
			
		||||
    let mut stream = connect_to_server(port).await;
 | 
			
		||||
 | 
			
		||||
    // Set up hash
 | 
			
		||||
    send_command(&mut stream, "*4\r\n$4\r\nHSET\r\n$4\r\nhash\r\n$6\r\nfield1\r\n$6\r\nvalue1\r\n").await;
 | 
			
		||||
    send_command(
 | 
			
		||||
        &mut stream,
 | 
			
		||||
        "*4\r\n$4\r\nHSET\r\n$4\r\nhash\r\n$6\r\nfield1\r\n$6\r\nvalue1\r\n",
 | 
			
		||||
    )
 | 
			
		||||
    .await;
 | 
			
		||||
 | 
			
		||||
    // Test HEXISTS for existing field
 | 
			
		||||
    let response = send_command(&mut stream, "*3\r\n$7\r\nHEXISTS\r\n$4\r\nhash\r\n$6\r\nfield1\r\n").await;
 | 
			
		||||
    let response = send_command(
 | 
			
		||||
        &mut stream,
 | 
			
		||||
        "*3\r\n$7\r\nHEXISTS\r\n$4\r\nhash\r\n$6\r\nfield1\r\n",
 | 
			
		||||
    )
 | 
			
		||||
    .await;
 | 
			
		||||
    println!("HEXISTS existing field response: {}", response);
 | 
			
		||||
    assert!(response.contains("1"));
 | 
			
		||||
 | 
			
		||||
    // Test HEXISTS for non-existent field
 | 
			
		||||
    let response = send_command(&mut stream, "*3\r\n$7\r\nHEXISTS\r\n$4\r\nhash\r\n$7\r\nnoexist\r\n").await;
 | 
			
		||||
    let response = send_command(
 | 
			
		||||
        &mut stream,
 | 
			
		||||
        "*3\r\n$7\r\nHEXISTS\r\n$4\r\nhash\r\n$7\r\nnoexist\r\n",
 | 
			
		||||
    )
 | 
			
		||||
    .await;
 | 
			
		||||
    println!("HEXISTS non-existent field response: {}", response);
 | 
			
		||||
    assert!(response.contains("0"), "Expected HEXISTS to return 0 for non-existent field, got: {}", response);
 | 
			
		||||
    assert!(
 | 
			
		||||
        response.contains("0"),
 | 
			
		||||
        "Expected HEXISTS to return 0 for non-existent field, got: {}",
 | 
			
		||||
        response
 | 
			
		||||
    );
 | 
			
		||||
}
 | 
			
		||||
@@ -325,7 +325,11 @@ async fn test_03_scan_and_keys() {
 | 
			
		||||
    let mut s = connect(port).await;
 | 
			
		||||
 | 
			
		||||
    for i in 0..5 {
 | 
			
		||||
        let _ = send_cmd(&mut s, &["SET", &format!("key{}", i), &format!("value{}", i)]).await;
 | 
			
		||||
        let _ = send_cmd(
 | 
			
		||||
            &mut s,
 | 
			
		||||
            &["SET", &format!("key{}", i), &format!("value{}", i)],
 | 
			
		||||
        )
 | 
			
		||||
        .await;
 | 
			
		||||
    }
 | 
			
		||||
 | 
			
		||||
    let scan = send_cmd(&mut s, &["SCAN", "0", "MATCH", "key*", "COUNT", "10"]).await;
 | 
			
		||||
@@ -358,7 +362,11 @@ async fn test_04_hashes_suite() {
 | 
			
		||||
    assert_contains(&h2, "2", "HSET added 2 new fields");
 | 
			
		||||
 | 
			
		||||
    // HMGET
 | 
			
		||||
    let hmg = send_cmd(&mut s, &["HMGET", "profile:1", "name", "age", "city", "nope"]).await;
 | 
			
		||||
    let hmg = send_cmd(
 | 
			
		||||
        &mut s,
 | 
			
		||||
        &["HMGET", "profile:1", "name", "age", "city", "nope"],
 | 
			
		||||
    )
 | 
			
		||||
    .await;
 | 
			
		||||
    assert_contains(&hmg, "alice", "HMGET name");
 | 
			
		||||
    assert_contains(&hmg, "30", "HMGET age");
 | 
			
		||||
    assert_contains(&hmg, "paris", "HMGET city");
 | 
			
		||||
@@ -392,7 +400,11 @@ async fn test_04_hashes_suite() {
 | 
			
		||||
    assert_contains(&hnx1, "1", "HSETNX new field -> 1");
 | 
			
		||||
 | 
			
		||||
    // HSCAN
 | 
			
		||||
    let hscan = send_cmd(&mut s, &["HSCAN", "profile:1", "0", "MATCH", "n*", "COUNT", "10"]).await;
 | 
			
		||||
    let hscan = send_cmd(
 | 
			
		||||
        &mut s,
 | 
			
		||||
        &["HSCAN", "profile:1", "0", "MATCH", "n*", "COUNT", "10"],
 | 
			
		||||
    )
 | 
			
		||||
    .await;
 | 
			
		||||
    assert_contains(&hscan, "name", "HSCAN matches fields starting with n");
 | 
			
		||||
    assert_contains(&hscan, "nickname", "HSCAN nickname present");
 | 
			
		||||
 | 
			
		||||
@@ -424,13 +436,21 @@ async fn test_05_lists_suite_including_blpop() {
 | 
			
		||||
    assert_eq_resp(&lidx, "$1\r\nb\r\n", "LINDEX q:jobs 0 should be b");
 | 
			
		||||
 | 
			
		||||
    let lr = send_cmd(&mut a, &["LRANGE", "q:jobs", "0", "-1"]).await;
 | 
			
		||||
    assert_eq_resp(&lr, "*3\r\n$1\r\nb\r\n$1\r\na\r\n$1\r\nc\r\n", "LRANGE q:jobs 0 -1 should be [b,a,c]");
 | 
			
		||||
    assert_eq_resp(
 | 
			
		||||
        &lr,
 | 
			
		||||
        "*3\r\n$1\r\nb\r\n$1\r\na\r\n$1\r\nc\r\n",
 | 
			
		||||
        "LRANGE q:jobs 0 -1 should be [b,a,c]",
 | 
			
		||||
    );
 | 
			
		||||
 | 
			
		||||
    // LTRIM
 | 
			
		||||
    let ltrim = send_cmd(&mut a, &["LTRIM", "q:jobs", "0", "1"]).await;
 | 
			
		||||
    assert_contains(<rim, "OK", "LTRIM OK");
 | 
			
		||||
    let lr_post = send_cmd(&mut a, &["LRANGE", "q:jobs", "0", "-1"]).await;
 | 
			
		||||
    assert_eq_resp(&lr_post, "*2\r\n$1\r\nb\r\n$1\r\na\r\n", "After LTRIM, list [b,a]");
 | 
			
		||||
    assert_eq_resp(
 | 
			
		||||
        &lr_post,
 | 
			
		||||
        "*2\r\n$1\r\nb\r\n$1\r\na\r\n",
 | 
			
		||||
        "After LTRIM, list [b,a]",
 | 
			
		||||
    );
 | 
			
		||||
 | 
			
		||||
    // LREM remove first occurrence of b
 | 
			
		||||
    let lrem = send_cmd(&mut a, &["LREM", "q:jobs", "1", "b"]).await;
 | 
			
		||||
@@ -444,7 +464,11 @@ async fn test_05_lists_suite_including_blpop() {
 | 
			
		||||
 | 
			
		||||
    // LPOP with count on empty -> []
 | 
			
		||||
    let lpop0 = send_cmd(&mut a, &["LPOP", "q:jobs", "2"]).await;
 | 
			
		||||
    assert_eq_resp(&lpop0, "*0\r\n", "LPOP with count on empty returns empty array");
 | 
			
		||||
    assert_eq_resp(
 | 
			
		||||
        &lpop0,
 | 
			
		||||
        "*0\r\n",
 | 
			
		||||
        "LPOP with count on empty returns empty array",
 | 
			
		||||
    );
 | 
			
		||||
 | 
			
		||||
    // BLPOP: block on one client, push from another
 | 
			
		||||
    let c1 = connect(port).await;
 | 
			
		||||
@@ -513,7 +537,7 @@ async fn test_07_age_stateless_suite() {
 | 
			
		||||
        // naive parse for tests
 | 
			
		||||
        let mut lines = resp.lines();
 | 
			
		||||
        let _ = lines.next(); // *2
 | 
			
		||||
        // $len
 | 
			
		||||
                              // $len
 | 
			
		||||
        let _ = lines.next();
 | 
			
		||||
        let recip = lines.next().unwrap_or("").to_string();
 | 
			
		||||
        let _ = lines.next();
 | 
			
		||||
@@ -548,8 +572,16 @@ async fn test_07_age_stateless_suite() {
 | 
			
		||||
    let v_ok = send_cmd(&mut s, &["AGE", "VERIFY", &verify_pub, "msg", &sig_b64]).await;
 | 
			
		||||
    assert_contains(&v_ok, "1", "VERIFY should be 1 for valid signature");
 | 
			
		||||
 | 
			
		||||
    let v_bad = send_cmd(&mut s, &["AGE", "VERIFY", &verify_pub, "tampered", &sig_b64]).await;
 | 
			
		||||
    assert_contains(&v_bad, "0", "VERIFY should be 0 for invalid message/signature");
 | 
			
		||||
    let v_bad = send_cmd(
 | 
			
		||||
        &mut s,
 | 
			
		||||
        &["AGE", "VERIFY", &verify_pub, "tampered", &sig_b64],
 | 
			
		||||
    )
 | 
			
		||||
    .await;
 | 
			
		||||
    assert_contains(
 | 
			
		||||
        &v_bad,
 | 
			
		||||
        "0",
 | 
			
		||||
        "VERIFY should be 0 for invalid message/signature",
 | 
			
		||||
    );
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
#[tokio::test]
 | 
			
		||||
@@ -581,7 +613,7 @@ async fn test_08_age_persistent_named_suite() {
 | 
			
		||||
        skg
 | 
			
		||||
    );
 | 
			
		||||
 | 
			
		||||
    let sig = send_cmd(&mut s, &["AGE", "SIGNNAME", "app1", "m"] ).await;
 | 
			
		||||
    let sig = send_cmd(&mut s, &["AGE", "SIGNNAME", "app1", "m"]).await;
 | 
			
		||||
    let sig_b64 = extract_bulk_payload(&sig).expect("Failed to parse bulk payload from SIGNNAME");
 | 
			
		||||
    let v1 = send_cmd(&mut s, &["AGE", "VERIFYNAME", "app1", "m", &sig_b64]).await;
 | 
			
		||||
    assert_contains(&v1, "1", "VERIFYNAME valid => 1");
 | 
			
		||||
@@ -597,60 +629,69 @@ async fn test_08_age_persistent_named_suite() {
 | 
			
		||||
 | 
			
		||||
#[tokio::test]
 | 
			
		||||
async fn test_10_expire_pexpire_persist() {
 | 
			
		||||
   let (server, port) = start_test_server("expire_suite").await;
 | 
			
		||||
   spawn_listener(server, port).await;
 | 
			
		||||
   sleep(Duration::from_millis(150)).await;
 | 
			
		||||
    let (server, port) = start_test_server("expire_suite").await;
 | 
			
		||||
    spawn_listener(server, port).await;
 | 
			
		||||
    sleep(Duration::from_millis(150)).await;
 | 
			
		||||
 | 
			
		||||
   let mut s = connect(port).await;
 | 
			
		||||
    let mut s = connect(port).await;
 | 
			
		||||
 | 
			
		||||
   // EXPIRE: seconds
 | 
			
		||||
   let _ = send_cmd(&mut s, &["SET", "exp:s", "v"]).await;
 | 
			
		||||
   let ex = send_cmd(&mut s, &["EXPIRE", "exp:s", "1"]).await;
 | 
			
		||||
   assert_contains(&ex, "1", "EXPIRE exp:s 1 -> 1 (applied)");
 | 
			
		||||
   let ttl1 = send_cmd(&mut s, &["TTL", "exp:s"]).await;
 | 
			
		||||
   assert!(
 | 
			
		||||
       ttl1.contains("1") || ttl1.contains("0"),
 | 
			
		||||
       "TTL exp:s should be 1 or 0, got: {}",
 | 
			
		||||
       ttl1
 | 
			
		||||
   );
 | 
			
		||||
   sleep(Duration::from_millis(1100)).await;
 | 
			
		||||
   let get_after = send_cmd(&mut s, &["GET", "exp:s"]).await;
 | 
			
		||||
   assert_contains(&get_after, "$-1", "GET after expiry should be Null");
 | 
			
		||||
   let ttl_after = send_cmd(&mut s, &["TTL", "exp:s"]).await;
 | 
			
		||||
   assert_contains(&ttl_after, "-2", "TTL after expiry -> -2");
 | 
			
		||||
   let exists_after = send_cmd(&mut s, &["EXISTS", "exp:s"]).await;
 | 
			
		||||
   assert_contains(&exists_after, "0", "EXISTS after expiry -> 0");
 | 
			
		||||
    // EXPIRE: seconds
 | 
			
		||||
    let _ = send_cmd(&mut s, &["SET", "exp:s", "v"]).await;
 | 
			
		||||
    let ex = send_cmd(&mut s, &["EXPIRE", "exp:s", "1"]).await;
 | 
			
		||||
    assert_contains(&ex, "1", "EXPIRE exp:s 1 -> 1 (applied)");
 | 
			
		||||
    let ttl1 = send_cmd(&mut s, &["TTL", "exp:s"]).await;
 | 
			
		||||
    assert!(
 | 
			
		||||
        ttl1.contains("1") || ttl1.contains("0"),
 | 
			
		||||
        "TTL exp:s should be 1 or 0, got: {}",
 | 
			
		||||
        ttl1
 | 
			
		||||
    );
 | 
			
		||||
    sleep(Duration::from_millis(1100)).await;
 | 
			
		||||
    let get_after = send_cmd(&mut s, &["GET", "exp:s"]).await;
 | 
			
		||||
    assert_contains(&get_after, "$-1", "GET after expiry should be Null");
 | 
			
		||||
    let ttl_after = send_cmd(&mut s, &["TTL", "exp:s"]).await;
 | 
			
		||||
    assert_contains(&ttl_after, "-2", "TTL after expiry -> -2");
 | 
			
		||||
    let exists_after = send_cmd(&mut s, &["EXISTS", "exp:s"]).await;
 | 
			
		||||
    assert_contains(&exists_after, "0", "EXISTS after expiry -> 0");
 | 
			
		||||
 | 
			
		||||
   // PEXPIRE: milliseconds
 | 
			
		||||
   let _ = send_cmd(&mut s, &["SET", "exp:ms", "v"]).await;
 | 
			
		||||
   let pex = send_cmd(&mut s, &["PEXPIRE", "exp:ms", "1500"]).await;
 | 
			
		||||
   assert_contains(&pex, "1", "PEXPIRE exp:ms 1500 -> 1 (applied)");
 | 
			
		||||
   let ttl_ms1 = send_cmd(&mut s, &["TTL", "exp:ms"]).await;
 | 
			
		||||
   assert!(
 | 
			
		||||
       ttl_ms1.contains("1") || ttl_ms1.contains("0"),
 | 
			
		||||
       "TTL exp:ms should be 1 or 0 soon after PEXPIRE, got: {}",
 | 
			
		||||
       ttl_ms1
 | 
			
		||||
   );
 | 
			
		||||
   sleep(Duration::from_millis(1600)).await;
 | 
			
		||||
   let exists_ms_after = send_cmd(&mut s, &["EXISTS", "exp:ms"]).await;
 | 
			
		||||
   assert_contains(&exists_ms_after, "0", "EXISTS exp:ms after ms expiry -> 0");
 | 
			
		||||
    // PEXPIRE: milliseconds
 | 
			
		||||
    let _ = send_cmd(&mut s, &["SET", "exp:ms", "v"]).await;
 | 
			
		||||
    let pex = send_cmd(&mut s, &["PEXPIRE", "exp:ms", "1500"]).await;
 | 
			
		||||
    assert_contains(&pex, "1", "PEXPIRE exp:ms 1500 -> 1 (applied)");
 | 
			
		||||
    let ttl_ms1 = send_cmd(&mut s, &["TTL", "exp:ms"]).await;
 | 
			
		||||
    assert!(
 | 
			
		||||
        ttl_ms1.contains("1") || ttl_ms1.contains("0"),
 | 
			
		||||
        "TTL exp:ms should be 1 or 0 soon after PEXPIRE, got: {}",
 | 
			
		||||
        ttl_ms1
 | 
			
		||||
    );
 | 
			
		||||
    sleep(Duration::from_millis(1600)).await;
 | 
			
		||||
    let exists_ms_after = send_cmd(&mut s, &["EXISTS", "exp:ms"]).await;
 | 
			
		||||
    assert_contains(&exists_ms_after, "0", "EXISTS exp:ms after ms expiry -> 0");
 | 
			
		||||
 | 
			
		||||
   // PERSIST: remove expiration
 | 
			
		||||
   let _ = send_cmd(&mut s, &["SET", "exp:persist", "v"]).await;
 | 
			
		||||
   let _ = send_cmd(&mut s, &["EXPIRE", "exp:persist", "5"]).await;
 | 
			
		||||
   let ttl_pre = send_cmd(&mut s, &["TTL", "exp:persist"]).await;
 | 
			
		||||
   assert!(
 | 
			
		||||
       ttl_pre.contains("5") || ttl_pre.contains("4") || ttl_pre.contains("3") || ttl_pre.contains("2") || ttl_pre.contains("1") || ttl_pre.contains("0"),
 | 
			
		||||
       "TTL exp:persist should be >=0 before persist, got: {}",
 | 
			
		||||
       ttl_pre
 | 
			
		||||
   );
 | 
			
		||||
   let persist1 = send_cmd(&mut s, &["PERSIST", "exp:persist"]).await;
 | 
			
		||||
   assert_contains(&persist1, "1", "PERSIST should remove expiration");
 | 
			
		||||
   let ttl_post = send_cmd(&mut s, &["TTL", "exp:persist"]).await;
 | 
			
		||||
   assert_contains(&ttl_post, "-1", "TTL after PERSIST -> -1 (no expiration)");
 | 
			
		||||
   // Second persist should return 0 (nothing to remove)
 | 
			
		||||
   let persist2 = send_cmd(&mut s, &["PERSIST", "exp:persist"]).await;
 | 
			
		||||
   assert_contains(&persist2, "0", "PERSIST again -> 0 (no expiration to remove)");
 | 
			
		||||
    // PERSIST: remove expiration
 | 
			
		||||
    let _ = send_cmd(&mut s, &["SET", "exp:persist", "v"]).await;
 | 
			
		||||
    let _ = send_cmd(&mut s, &["EXPIRE", "exp:persist", "5"]).await;
 | 
			
		||||
    let ttl_pre = send_cmd(&mut s, &["TTL", "exp:persist"]).await;
 | 
			
		||||
    assert!(
 | 
			
		||||
        ttl_pre.contains("5")
 | 
			
		||||
            || ttl_pre.contains("4")
 | 
			
		||||
            || ttl_pre.contains("3")
 | 
			
		||||
            || ttl_pre.contains("2")
 | 
			
		||||
            || ttl_pre.contains("1")
 | 
			
		||||
            || ttl_pre.contains("0"),
 | 
			
		||||
        "TTL exp:persist should be >=0 before persist, got: {}",
 | 
			
		||||
        ttl_pre
 | 
			
		||||
    );
 | 
			
		||||
    let persist1 = send_cmd(&mut s, &["PERSIST", "exp:persist"]).await;
 | 
			
		||||
    assert_contains(&persist1, "1", "PERSIST should remove expiration");
 | 
			
		||||
    let ttl_post = send_cmd(&mut s, &["TTL", "exp:persist"]).await;
 | 
			
		||||
    assert_contains(&ttl_post, "-1", "TTL after PERSIST -> -1 (no expiration)");
 | 
			
		||||
    // Second persist should return 0 (nothing to remove)
 | 
			
		||||
    let persist2 = send_cmd(&mut s, &["PERSIST", "exp:persist"]).await;
 | 
			
		||||
    assert_contains(
 | 
			
		||||
        &persist2,
 | 
			
		||||
        "0",
 | 
			
		||||
        "PERSIST again -> 0 (no expiration to remove)",
 | 
			
		||||
    );
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
#[tokio::test]
 | 
			
		||||
@@ -663,7 +704,11 @@ async fn test_11_set_with_options() {
 | 
			
		||||
 | 
			
		||||
    // SET with GET on non-existing key -> returns Null, sets value
 | 
			
		||||
    let set_get1 = send_cmd(&mut s, &["SET", "s1", "v1", "GET"]).await;
 | 
			
		||||
    assert_contains(&set_get1, "$-1", "SET s1 v1 GET returns Null when key didn't exist");
 | 
			
		||||
    assert_contains(
 | 
			
		||||
        &set_get1,
 | 
			
		||||
        "$-1",
 | 
			
		||||
        "SET s1 v1 GET returns Null when key didn't exist",
 | 
			
		||||
    );
 | 
			
		||||
    let g1 = send_cmd(&mut s, &["GET", "s1"]).await;
 | 
			
		||||
    assert_contains(&g1, "v1", "GET s1 after first SET");
 | 
			
		||||
 | 
			
		||||
@@ -707,42 +752,42 @@ async fn test_11_set_with_options() {
 | 
			
		||||
 | 
			
		||||
#[tokio::test]
 | 
			
		||||
async fn test_09_mget_mset_and_variadic_exists_del() {
 | 
			
		||||
   let (server, port) = start_test_server("mget_mset_variadic").await;
 | 
			
		||||
   spawn_listener(server, port).await;
 | 
			
		||||
   sleep(Duration::from_millis(150)).await;
 | 
			
		||||
    let (server, port) = start_test_server("mget_mset_variadic").await;
 | 
			
		||||
    spawn_listener(server, port).await;
 | 
			
		||||
    sleep(Duration::from_millis(150)).await;
 | 
			
		||||
 | 
			
		||||
   let mut s = connect(port).await;
 | 
			
		||||
    let mut s = connect(port).await;
 | 
			
		||||
 | 
			
		||||
   // MSET multiple keys
 | 
			
		||||
   let mset = send_cmd(&mut s, &["MSET", "k1", "v1", "k2", "v2", "k3", "v3"]).await;
 | 
			
		||||
   assert_contains(&mset, "OK", "MSET k1 v1 k2 v2 k3 v3 -> OK");
 | 
			
		||||
    // MSET multiple keys
 | 
			
		||||
    let mset = send_cmd(&mut s, &["MSET", "k1", "v1", "k2", "v2", "k3", "v3"]).await;
 | 
			
		||||
    assert_contains(&mset, "OK", "MSET k1 v1 k2 v2 k3 v3 -> OK");
 | 
			
		||||
 | 
			
		||||
   // MGET should return values and Null for missing
 | 
			
		||||
   let mget = send_cmd(&mut s, &["MGET", "k1", "k2", "nope", "k3"]).await;
 | 
			
		||||
   // Expect an array with 4 entries; verify payloads
 | 
			
		||||
   assert_contains(&mget, "v1", "MGET k1");
 | 
			
		||||
   assert_contains(&mget, "v2", "MGET k2");
 | 
			
		||||
   assert_contains(&mget, "v3", "MGET k3");
 | 
			
		||||
   assert_contains(&mget, "$-1", "MGET missing returns Null");
 | 
			
		||||
    // MGET should return values and Null for missing
 | 
			
		||||
    let mget = send_cmd(&mut s, &["MGET", "k1", "k2", "nope", "k3"]).await;
 | 
			
		||||
    // Expect an array with 4 entries; verify payloads
 | 
			
		||||
    assert_contains(&mget, "v1", "MGET k1");
 | 
			
		||||
    assert_contains(&mget, "v2", "MGET k2");
 | 
			
		||||
    assert_contains(&mget, "v3", "MGET k3");
 | 
			
		||||
    assert_contains(&mget, "$-1", "MGET missing returns Null");
 | 
			
		||||
 | 
			
		||||
   // EXISTS variadic: count how many exist
 | 
			
		||||
   let exists_multi = send_cmd(&mut s, &["EXISTS", "k1", "nope", "k3"]).await;
 | 
			
		||||
   // Server returns SimpleString numeric, e.g. +2
 | 
			
		||||
   assert_contains(&exists_multi, "2", "EXISTS k1 nope k3 -> 2");
 | 
			
		||||
    // EXISTS variadic: count how many exist
 | 
			
		||||
    let exists_multi = send_cmd(&mut s, &["EXISTS", "k1", "nope", "k3"]).await;
 | 
			
		||||
    // Server returns SimpleString numeric, e.g. +2
 | 
			
		||||
    assert_contains(&exists_multi, "2", "EXISTS k1 nope k3 -> 2");
 | 
			
		||||
 | 
			
		||||
   // DEL variadic: delete multiple keys, return count deleted
 | 
			
		||||
   let del_multi = send_cmd(&mut s, &["DEL", "k1", "k3", "nope"]).await;
 | 
			
		||||
   assert_contains(&del_multi, "2", "DEL k1 k3 nope -> 2");
 | 
			
		||||
    // DEL variadic: delete multiple keys, return count deleted
 | 
			
		||||
    let del_multi = send_cmd(&mut s, &["DEL", "k1", "k3", "nope"]).await;
 | 
			
		||||
    assert_contains(&del_multi, "2", "DEL k1 k3 nope -> 2");
 | 
			
		||||
 | 
			
		||||
   // Verify deletion
 | 
			
		||||
   let exists_after = send_cmd(&mut s, &["EXISTS", "k1", "k3"]).await;
 | 
			
		||||
   assert_contains(&exists_after, "0", "EXISTS k1 k3 after DEL -> 0");
 | 
			
		||||
    // Verify deletion
 | 
			
		||||
    let exists_after = send_cmd(&mut s, &["EXISTS", "k1", "k3"]).await;
 | 
			
		||||
    assert_contains(&exists_after, "0", "EXISTS k1 k3 after DEL -> 0");
 | 
			
		||||
 | 
			
		||||
   // MGET after deletion should include Nulls for deleted keys
 | 
			
		||||
   let mget_after = send_cmd(&mut s, &["MGET", "k1", "k2", "k3"]).await;
 | 
			
		||||
   assert_contains(&mget_after, "$-1", "MGET k1 after DEL -> Null");
 | 
			
		||||
   assert_contains(&mget_after, "v2", "MGET k2 remains");
 | 
			
		||||
   assert_contains(&mget_after, "$-1", "MGET k3 after DEL -> Null");
 | 
			
		||||
    // MGET after deletion should include Nulls for deleted keys
 | 
			
		||||
    let mget_after = send_cmd(&mut s, &["MGET", "k1", "k2", "k3"]).await;
 | 
			
		||||
    assert_contains(&mget_after, "$-1", "MGET k1 after DEL -> Null");
 | 
			
		||||
    assert_contains(&mget_after, "v2", "MGET k2 remains");
 | 
			
		||||
    assert_contains(&mget_after, "$-1", "MGET k3 after DEL -> Null");
 | 
			
		||||
}
 | 
			
		||||
#[tokio::test]
 | 
			
		||||
async fn test_12_hash_incr() {
 | 
			
		||||
@@ -862,9 +907,16 @@ async fn test_14_expireat_pexpireat() {
 | 
			
		||||
    let mut s = connect(port).await;
 | 
			
		||||
 | 
			
		||||
    // EXPIREAT: seconds since epoch
 | 
			
		||||
    let now_secs = SystemTime::now().duration_since(UNIX_EPOCH).unwrap().as_secs() as i64;
 | 
			
		||||
    let now_secs = SystemTime::now()
 | 
			
		||||
        .duration_since(UNIX_EPOCH)
 | 
			
		||||
        .unwrap()
 | 
			
		||||
        .as_secs() as i64;
 | 
			
		||||
    let _ = send_cmd(&mut s, &["SET", "exp:at:s", "v"]).await;
 | 
			
		||||
    let exat = send_cmd(&mut s, &["EXPIREAT", "exp:at:s", &format!("{}", now_secs + 1)]).await;
 | 
			
		||||
    let exat = send_cmd(
 | 
			
		||||
        &mut s,
 | 
			
		||||
        &["EXPIREAT", "exp:at:s", &format!("{}", now_secs + 1)],
 | 
			
		||||
    )
 | 
			
		||||
    .await;
 | 
			
		||||
    assert_contains(&exat, "1", "EXPIREAT exp:at:s now+1s -> 1 (applied)");
 | 
			
		||||
    let ttl1 = send_cmd(&mut s, &["TTL", "exp:at:s"]).await;
 | 
			
		||||
    assert!(
 | 
			
		||||
@@ -874,12 +926,23 @@ async fn test_14_expireat_pexpireat() {
 | 
			
		||||
    );
 | 
			
		||||
    sleep(Duration::from_millis(1200)).await;
 | 
			
		||||
    let exists_after_exat = send_cmd(&mut s, &["EXISTS", "exp:at:s"]).await;
 | 
			
		||||
    assert_contains(&exists_after_exat, "0", "EXISTS exp:at:s after EXPIREAT expiry -> 0");
 | 
			
		||||
    assert_contains(
 | 
			
		||||
        &exists_after_exat,
 | 
			
		||||
        "0",
 | 
			
		||||
        "EXISTS exp:at:s after EXPIREAT expiry -> 0",
 | 
			
		||||
    );
 | 
			
		||||
 | 
			
		||||
    // PEXPIREAT: milliseconds since epoch
 | 
			
		||||
    let now_ms = SystemTime::now().duration_since(UNIX_EPOCH).unwrap().as_millis() as i64;
 | 
			
		||||
    let now_ms = SystemTime::now()
 | 
			
		||||
        .duration_since(UNIX_EPOCH)
 | 
			
		||||
        .unwrap()
 | 
			
		||||
        .as_millis() as i64;
 | 
			
		||||
    let _ = send_cmd(&mut s, &["SET", "exp:at:ms", "v"]).await;
 | 
			
		||||
    let pexat = send_cmd(&mut s, &["PEXPIREAT", "exp:at:ms", &format!("{}", now_ms + 450)]).await;
 | 
			
		||||
    let pexat = send_cmd(
 | 
			
		||||
        &mut s,
 | 
			
		||||
        &["PEXPIREAT", "exp:at:ms", &format!("{}", now_ms + 450)],
 | 
			
		||||
    )
 | 
			
		||||
    .await;
 | 
			
		||||
    assert_contains(&pexat, "1", "PEXPIREAT exp:at:ms now+450ms -> 1 (applied)");
 | 
			
		||||
    let ttl2 = send_cmd(&mut s, &["TTL", "exp:at:ms"]).await;
 | 
			
		||||
    assert!(
 | 
			
		||||
@@ -889,5 +952,9 @@ async fn test_14_expireat_pexpireat() {
 | 
			
		||||
    );
 | 
			
		||||
    sleep(Duration::from_millis(600)).await;
 | 
			
		||||
    let exists_after_pexat = send_cmd(&mut s, &["EXISTS", "exp:at:ms"]).await;
 | 
			
		||||
    assert_contains(&exists_after_pexat, "0", "EXISTS exp:at:ms after PEXPIREAT expiry -> 0");
 | 
			
		||||
    assert_contains(
 | 
			
		||||
        &exists_after_pexat,
 | 
			
		||||
        "0",
 | 
			
		||||
        "EXISTS exp:at:ms after PEXPIREAT expiry -> 0",
 | 
			
		||||
    );
 | 
			
		||||
}
 | 
			
		||||
		Reference in New Issue
	
	Block a user