---
title: "Performance Benchmarking and Optimization"
output: rmarkdown::html_vignette
vignette: >
  %\VignetteIndexEntry{Performance Benchmarking and Optimization}
  %\VignetteEngine{knitr::rmarkdown}
  %\VignetteEncoding{UTF-8}
---

```{r, include = FALSE}
knitr::opts_chunk$set(
  collapse = TRUE,
  comment = "#>"
)
```

## Introduction

**New in v0.2.5**: riemtan includes cross-platform parallel processing using the futureverse framework. This vignette demonstrates how to benchmark performance and optimize your workflows for large brain connectome datasets.

## Parallel Processing Overview

riemtan's parallel processing provides:

- **Cross-platform compatibility**: Works on Windows, Mac, and Linux
- **Intelligent auto-detection**: Automatically uses parallelization when beneficial
- **Optional progress reporting**: Monitor long-running computations
- **Full backwards compatibility**: Existing code works unchanged

## Setup

```{r eval=FALSE}
library(riemtan)
library(Matrix)
library(microbenchmark)  # For benchmarking

# Load AIRM metric
data(airm)

# Create example dataset (100 matrices of size 50x50)
set.seed(42)
create_spd_matrix <- function(p) {
  mat <- diag(p) + matrix(rnorm(p*p, 0, 0.1), p, p)
  mat <- (mat + t(mat)) / 2
  mat <- mat + diag(p) * 0.5
  Matrix::pack(Matrix::Matrix(mat, sparse = FALSE))
}

connectomes <- lapply(1:100, function(i) create_spd_matrix(50))
```

## Benchmarking Parallel vs Sequential

### Tangent Computations

Compare parallel vs sequential performance for tangent space projections:

```{r eval=FALSE}
# Create sample
sample <- CSample$new(conns = connectomes, metric_obj = airm)

# Sequential baseline
set_parallel_plan("sequential")
time_seq <- system.time(sample$compute_tangents())
print(paste("Sequential:", round(time_seq[3], 2), "seconds"))

# Parallel with 4 workers
set_parallel_plan("multisession", workers = 4)
time_par4 <- system.time(sample$compute_tangents())
print(paste("Parallel (4 workers):", round(time_par4[3], 2), "seconds"))
print(paste("Speedup:", round(time_seq[3] / time_par4[3], 2), "x"))

# Parallel with 8 workers
set_parallel_plan("multisession", workers = 8)
time_par8 <- system.time(sample$compute_tangents())
print(paste("Parallel (8 workers):", round(time_par8[3], 2), "seconds"))
print(paste("Speedup:", round(time_seq[3] / time_par8[3], 2), "x"))

# Reset
reset_parallel_plan()
```

**Expected results** (for n=100, p=50):
- Sequential: ~15-20 seconds
- Parallel (4 workers): ~4-6 seconds (3-4x speedup)
- Parallel (8 workers): ~2-3 seconds (6-8x speedup)

### Frechet Mean Computation

Benchmark Frechet mean computation with different sample sizes:

```{r eval=FALSE}
# Function to benchmark Frechet mean
benchmark_fmean <- function(n, workers = 1) {
  conns_subset <- connectomes[1:n]
  sample <- CSample$new(conns = conns_subset, metric_obj = airm)

  if (workers == 1) {
    set_parallel_plan("sequential")
  } else {
    set_parallel_plan("multisession", workers = workers)
  }

  time <- system.time(sample$compute_fmean(tol = 0.01, max_iter = 50))
  reset_parallel_plan()

  time[3]
}

# Benchmark different sample sizes
sample_sizes <- c(20, 50, 100, 200)
results <- data.frame(
  n = sample_sizes,
  sequential = sapply(sample_sizes, benchmark_fmean, workers = 1),
  parallel_4 = sapply(sample_sizes, benchmark_fmean, workers = 4),
  parallel_8 = sapply(sample_sizes, benchmark_fmean, workers = 8)
)

# Calculate speedups
results$speedup_4 = results$sequential / results$parallel_4
results$speedup_8 = results$sequential / results$parallel_8

print(results)
```

**Expected speedup patterns**:
- Small samples (n < 50): Minimal benefit (overhead dominates)
- Medium samples (n = 50-100): 2-3x speedup
- Large samples (n > 100): 3-5x speedup

### Parquet I/O Benchmarks

Compare sequential vs parallel Parquet loading:

```{r eval=FALSE}
# Create Parquet dataset
write_connectomes_to_parquet(
  connectomes,
  output_dir = "benchmark_data",
  subject_ids = paste0("subj_", 1:100)
)

# Sequential loading
backend_seq <- create_parquet_backend("benchmark_data")
set_parallel_plan("sequential")
time_load_seq <- system.time({
  conns <- backend_seq$get_all_matrices()
})

# Parallel loading
backend_par <- create_parquet_backend("benchmark_data")
set_parallel_plan("multisession", workers = 4)
time_load_par <- system.time({
  conns <- backend_par$get_all_matrices(parallel = TRUE)
})

print(paste("Sequential load:", round(time_load_seq[3], 2), "seconds"))
print(paste("Parallel load:", round(time_load_par[3], 2), "seconds"))
print(paste("Speedup:", round(time_load_seq[3] / time_load_par[3], 2), "x"))

# Cleanup
reset_parallel_plan()
unlink("benchmark_data", recursive = TRUE)
```

**Expected results**:
- Sequential: ~10-15 seconds
- Parallel (4 workers): ~2-3 seconds (5-7x speedup)

## Scaling Analysis

### Worker Count Scaling

Test how performance scales with number of workers:

```{r eval=FALSE}
# Function to measure scaling
measure_scaling <- function(workers_list) {
  sample <- CSample$new(conns = connectomes, metric_obj = airm)

  times <- sapply(workers_list, function(w) {
    if (w == 1) {
      set_parallel_plan("sequential")
    } else {
      set_parallel_plan("multisession", workers = w)
    }

    time <- system.time(sample$compute_tangents())[3]
    reset_parallel_plan()
    time
  })

  data.frame(
    workers = workers_list,
    time = times,
    speedup = times[1] / times,
    efficiency = (times[1] / times) / workers_list * 100
  )
}

# Test with different worker counts
workers <- c(1, 2, 4, 8)
scaling_results <- measure_scaling(workers)

print(scaling_results)

# Plot scaling (if plotting package available)
if (requireNamespace("ggplot2", quietly = TRUE)) {
  library(ggplot2)

  p <- ggplot(scaling_results, aes(x = workers, y = speedup)) +
    geom_line() +
    geom_point(size = 3) +
    geom_abline(slope = 1, intercept = 0, linetype = "dashed", color = "red") +
    labs(
      title = "Parallel Scaling Performance",
      x = "Number of Workers",
      y = "Speedup",
      subtitle = "Dashed line = ideal linear scaling"
    ) +
    theme_minimal()

  print(p)
}
```

**Interpretation**:
- **Speedup**: Time(sequential) / Time(parallel)
- **Efficiency**: Speedup / Workers × 100%
- **Ideal scaling**: Speedup = Workers (efficiency = 100%)
- **Reality**: Diminishing returns due to overhead (efficiency < 100%)

### Dataset Size Impact

Measure how parallelization benefit changes with dataset size:

```{r eval=FALSE}
# Test different dataset sizes
test_sizes <- c(10, 25, 50, 100, 200)

size_benchmark <- function(n) {
  conns_subset <- lapply(1:n, function(i) create_spd_matrix(50))
  sample <- CSample$new(conns = conns_subset, metric_obj = airm)

  # Sequential
  set_parallel_plan("sequential")
  time_seq <- system.time(sample$compute_tangents())[3]

  # Parallel
  set_parallel_plan("multisession", workers = 4)
  time_par <- system.time(sample$compute_tangents())[3]

  reset_parallel_plan()

  c(sequential = time_seq, parallel = time_par, speedup = time_seq / time_par)
}

results_by_size <- t(sapply(test_sizes, size_benchmark))
results_by_size <- data.frame(n = test_sizes, results_by_size)

print(results_by_size)
```

**Expected pattern**:
- Small datasets (n < 10): Speedup < 1 (overhead penalty)
- Medium datasets (n = 25-50): Speedup ≈ 2-3
- Large datasets (n > 100): Speedup ≈ 3-4 (approaches maximum)

## Optimization Guidelines

### When to Use Parallel Processing

**Always beneficial** (enable by default):
- Large datasets (n ≥ 100 matrices)
- High-dimensional matrices (p ≥ 100)
- Repeated computations (multiple calls to compute methods)

**Sometimes beneficial** (test on your system):
- Medium datasets (n = 25-100)
- Medium-dimensional matrices (p = 50-100)
- Single computations

**Not beneficial** (use sequential):
- Small datasets (n < 10)
- Low-dimensional matrices (p < 20)
- Memory-constrained environments

### Optimal Worker Count

```{r eval=FALSE}
# Conservative (recommended for most users)
n_workers <- parallel::detectCores() - 1
set_parallel_plan("multisession", workers = n_workers)

# Aggressive (maximum performance, may slow system)
n_workers <- parallel::detectCores()
set_parallel_plan("multisession", workers = n_workers)

# Custom (based on benchmarking)
# Test different worker counts and choose optimal
```

**Guidelines**:
- Leave 1-2 cores for system operations
- More workers ≠ always faster (diminishing returns)
- Test on your specific hardware and dataset
- Consider memory: each worker needs its own data copy

### Memory Optimization

```{r eval=FALSE}
# For memory-constrained environments:

# 1. Use Parquet backend with small cache
backend <- create_parquet_backend("large_dataset", cache_size = 5)

# 2. Use moderate worker count
set_parallel_plan("multisession", workers = 2)

# 3. Use batch loading for very large datasets
sample <- CSample$new(backend = backend, metric_obj = airm)
conns_batch <- sample$load_connectomes_batched(
  indices = 1:500,
  batch_size = 50,   # Small batches
  progress = TRUE
)

# 4. Clear cache frequently
backend$clear_cache()
```

### Progress Reporting Overhead

Progress reporting adds small overhead (~1-5%):

```{r eval=FALSE}
# With progress (slightly slower)
set_parallel_plan("multisession", workers = 4)
time_with_progress <- system.time({
  sample$compute_tangents(progress = TRUE)
})[3]

# Without progress (slightly faster)
time_no_progress <- system.time({
  sample$compute_tangents(progress = FALSE)
})[3]

overhead <- (time_with_progress - time_no_progress) / time_no_progress * 100
print(paste("Progress overhead:", round(overhead, 1), "%"))
```

**Recommendation**: Use progress for long-running operations (>10 seconds), disable for fast operations.

## Benchmarking Best Practices

### 1. Use microbenchmark for Accurate Measurements

```{r eval=FALSE}
library(microbenchmark)

# Run multiple times to get stable estimates
mb_result <- microbenchmark(
  sequential = {
    set_parallel_plan("sequential")
    sample$compute_tangents()
  },
  parallel_4 = {
    set_parallel_plan("multisession", workers = 4)
    sample$compute_tangents()
  },
  times = 10  # Run 10 times each
)

print(mb_result)
plot(mb_result)
```

### 2. Control for Caching Effects

```{r eval=FALSE}
# Clear any caches between runs
sample <- CSample$new(conns = connectomes, metric_obj = airm)

# Warm up (first run may be slower)
sample$compute_tangents()

# Now benchmark
time <- system.time(sample$compute_tangents())[3]
```

### 3. Report System Information

```{r eval=FALSE}
# Record system specs with benchmarks
system_info <- list(
  cores = parallel::detectCores(),
  memory = as.numeric(system("wmic ComputerSystem get TotalPhysicalMemory", intern = TRUE)[2]) / 1e9,
  r_version = R.version.string,
  riemtan_version = packageVersion("riemtan"),
  os = Sys.info()["sysname"]
)

print(system_info)
```

## Common Performance Patterns

### Pattern 1: Compute-Bound Operations

Tangent computations, vectorizations scale well:

```{r eval=FALSE}
# Expect near-linear scaling up to physical cores
set_parallel_plan("multisession", workers = 4)
sample$compute_tangents()  # 3-4x speedup
sample$compute_vecs()       # 2-4x speedup
sample$compute_conns()      # 3-4x speedup
```

### Pattern 2: Iteration-Heavy Operations

Frechet mean has sub-linear scaling (synchronization overhead):

```{r eval=FALSE}
# Expect 2-3x speedup (less than compute-bound)
set_parallel_plan("multisession", workers = 4)
sample$compute_fmean()  # 2-3x speedup
```

### Pattern 3: I/O-Bound Operations

Parquet loading can super-scale (disk parallelism):

```{r eval=FALSE}
# May see >4x speedup with 4 workers (parallel disk I/O)
backend <- create_parquet_backend("dataset")
set_parallel_plan("multisession", workers = 4)
conns <- backend$get_all_matrices(parallel = TRUE)  # 5-10x speedup
```

## Platform-Specific Notes

### Windows

```{r eval=FALSE}
# Use multisession (multicore not available)
set_parallel_plan("multisession", workers = 4)

# Expect slightly higher overhead than Unix
# Typical speedup: 70-80% of Unix performance
```

### Mac/Linux

```{r eval=FALSE}
# Can use multicore for lower overhead
set_parallel_plan("multicore", workers = 4)

# Or multisession for better stability
set_parallel_plan("multisession", workers = 4)
```

### HPC Clusters

```{r eval=FALSE}
# Use cluster strategy for distributed computing
library(future)
plan(cluster, workers = c("node1", "node2", "node3", "node4"))

# Or use batchtools for SLURM integration
library(future.batchtools)
plan(batchtools_slurm, workers = 16)
```

## Summary

**Key Takeaways**:

1. **Enable parallelization for large datasets**: n ≥ 100 or p ≥ 100
2. **Use conservative worker count**: `parallel::detectCores() - 1`
3. **Expect 3-8x speedup** for compute-bound operations
4. **Use Parquet + parallel I/O** for maximum performance
5. **Benchmark on your system**: Performance varies by hardware
6. **Reset plan when done**: `reset_parallel_plan()`

**Typical Workflow**:

```{r eval=FALSE}
# Setup
library(riemtan)
set_parallel_plan("multisession", workers = parallel::detectCores() - 1)

# Load data
backend <- create_parquet_backend("large_dataset")
sample <- CSample$new(backend = backend, metric_obj = airm)

# Compute with progress
sample$compute_tangents(progress = TRUE)
sample$compute_fmean(progress = TRUE)

# Cleanup
reset_parallel_plan()
```

For more details, see:
- [Using Parquet Storage vignette](using-parquet.html)
- [futureverse documentation](https://future.futureverse.org/)
- [riemtan package documentation](https://nicoesve.github.io/riemtan/)
