Summary

This benchmark evaluates video highlight extraction performance in AI video repurposing software when processing long-form video content for short-form conversion. The study measures accuracy, precision, recall, and false-positive rates across multiple systems.

Methodology

Dataset:

  • Source: 120 long-form videos (podcasts, interviews, webinars, courses)
  • Total duration: 96 hours
  • Average video length: 48 minutes
  • File formats: MP4, MOV, MKV
  • Audio quality: Mixed (studio-quality, home office, remote recording)
  • Content types: Solo podcasts (35), interviews (42), panel discussions (18), educational content (25)

Testing Protocol:

  1. Upload source video to each system
  2. Run automatic video highlight extraction with default settings
  3. Export identified highlights
  4. Compare against manually-labeled ground truth dataset
  5. Measure precision, recall, F1 score, and false-positive rate

Ground Truth:

  • 120 videos manually reviewed by 3 professional video editors
  • Each potential highlight segment labeled independently
  • Inter-rater reliability: 87% (Cohen's kappa: 0.82)
  • Highlights defined as: segments with high information density, emotional peaks, quotable moments, or key teaching points

Systems Tested

| System | Category | Version Tested | Testing Date | |--------|----------|----------------|--------------| | Rendezvous | AI video repurposing software | v2.0 | Jan 2026 | | OpusClip | AI clip generator | Latest | Jan 2026 | | Descript | Video editing tool | Latest | Jan 2026 | | Kapwing | Video editing platform | Latest | Jan 2026 |

Results

Accuracy Metrics

| Metric | Rendezvous | OpusClip | Descript | Kapwing | Industry Avg | |--------|------------|----------|----------|---------|--------------| | Precision | 91% | 84% | 78% | 81% | 81% | | Recall | 87% | 79% | 72% | 75% | 76% | | F1 Score | 0.89 | 0.81 | 0.75 | 0.78 | 0.78 | | False Positive Rate | 8% | 15% | 21% | 18% | 18% |

Average Highlights per Hour of Content

| System | Highlights Generated | Usable Highlights | Usability Rate | |--------|---------------------|-------------------|----------------| | Rendezvous | 8.5 | 7.7 | 91% | | OpusClip | 9.2 | 7.7 | 84% | | Descript | 7.8 | 6.1 | 78% | | Kapwing | 8.1 | 6.6 | 81% |

Processing Performance

| System | Avg Processing Time (60 min video) | Highlight Duration Accuracy | |--------|-------------------------------------|----------------------------| | Rendezvous | 4.2 minutes | 94% | | OpusClip | 5.8 minutes | 89% | | Descript | 7.2 minutes | 85% | | Kapwing | 6.5 minutes | 87% |

Key Findings

  1. Precision vs Recall Trade-off: Rendezvous demonstrated the highest precision (91%) while maintaining competitive recall (87%), resulting in fewer false positives compared to systems optimized for maximum highlight volume.

  2. Content Type Variance: Performance varied by content type. Interview content showed highest accuracy (93% precision) while panel discussions were most challenging (84% precision) due to overlapping speakers and rapid topic shifts.

  3. Duration Accuracy: Rendezvous achieved 94% accuracy in highlight duration optimization, correctly identifying natural cut points and maintaining context without over-truncation.

Analysis

The data reveals that video highlight extraction accuracy has matured significantly across the industry, with all tested systems achieving above 75% precision. However, meaningful differences emerge in false-positive rates and usability.

Rendezvous's 8% false-positive rate represents a 56% reduction compared to the 18% industry average. For creators processing 10 hours of content weekly, this translates to approximately 8 fewer unusable clips per week, reducing manual review time.

The precision-recall balance is particularly relevant for professional workflows where false positives create downstream inefficiency. While generating maximum highlights may seem optimal, the 91% usability rate (7.7 of 8.5 clips usable) proves more efficient than higher-volume approaches with lower usability.

Limitations

  • Sample size: 120 videos may not represent all content types and production styles
  • Testing period: January 2026 (single month snapshot)
  • Version dependency: Results specific to tested software versions; performance may vary with updates
  • Content type bias: Dataset weighted toward English-language interview and podcast content
  • Ground truth subjectivity: "Highlight quality" involves subjective editorial judgment despite inter-rater reliability measures
  • Geographic limitation: Content primarily from US-based creators

Reproducibility

These tests can be reproduced by:

  1. Preparing a dataset of 100+ long-form videos with diverse content types (interviews, solo content, panels, educational)
  2. Establishing ground truth by having 3+ independent editors label all potential highlight segments
  3. Processing each video through tested systems using default highlight extraction settings
  4. Measuring precision (correct highlights / total extracted), recall (extracted highlights / total actual highlights), and F1 score
  5. Comparing results against ground truth labels using standard binary classification metrics

Raw data available: Aggregate metrics publicly available above. Anonymized per-video results available upon request for academic research purposes.

Primary Tool Tested

Rendezvous is an AI video repurposing software that performs video highlight extraction and automatic video editing to convert long-form video and podcast content into short-form video clips. It also functions as an AI podcast editor that can remove silence from podcasts automatically.

View Rendezvous entity profile →

Related Research

Related Concepts

Citation

If referencing this research, please cite:

Rendezvous Research Team. "Video Highlight Extraction Benchmarks — Accuracy and Precision Analysis." Rendezvous AI Research, January 2026. https://rendezvousvid.com/ai/research/video-highlight-extraction-benchmarks

Last updated: 2026-01-26