Tag

latency optimization

2 views collected around this technical thread.

Bilibili Tech
Bilibili Tech
Apr 29, 2025 · Cloud Computing

Bilibili Live Streaming Technology for the Spring Festival Gala: Experience Enhancement and Interactive Features

Bilibili's R&D built a cloud-based broadcast console for the 2024 CCTV Spring Festival Gala, delivering 4K HDR streaming, AI SDR-to-HDR conversion, low latency, bandwidth‑aware transcoding, and a synchronized “send bullet screen” interactive feature using custom SEI timestamps for hundreds of millions of viewers.

HDRLive StreamingSEI
0 likes · 15 min read
Bilibili Live Streaming Technology for the Spring Festival Gala: Experience Enhancement and Interactive Features
DeWu Technology
DeWu Technology
Apr 14, 2023 · Backend Development

Async-fork: Mitigating Query Latency Spikes Incurred by the Fork-based Snapshot Mechanism from the OS Level

Async‑fork shifts the costly page‑table copying from Redis’s parent process to its child, allowing the parent to resume handling queries instantly and cutting snapshot‑induced latency spikes by over 98%, thereby dramatically improving tail latency during AOF rewrites, RDB backups, and master‑slave synchronizations.

Async-forkMemory DatabaseRedis
0 likes · 21 min read
Async-fork: Mitigating Query Latency Spikes Incurred by the Fork-based Snapshot Mechanism from the OS Level
OPPO Kernel Craftsman
OPPO Kernel Craftsman
Jul 1, 2022 · Operations

Linux Kernel Performance Profiling: A Comprehensive Guide to On-CPU and Off-CPU Analysis

This comprehensive guide explains Linux kernel performance profiling—both on‑CPU and off‑CPU—by stressing the need to target the critical 3 % of code, covering throughput, latency and power metrics, scalability laws, flame‑graph visualizations, perf and eBPF tools, lock‑contention analysis, and further reading recommendations.

Flame GraphLinux KernelPerformance Profiling
0 likes · 27 min read
Linux Kernel Performance Profiling: A Comprehensive Guide to On-CPU and Off-CPU Analysis
HaoDF Tech Team
HaoDF Tech Team
Nov 8, 2021 · Operations

Service Risk Governance: Exploration, Mitigation, and Hands‑On Workshop

This talk recounts how the Good Doctor platform tackled severe online incidents by launching the DOA project, then a service risk governance initiative that identifies, quantifies, and mitigates latency‑related risks through metrics‑driven development, dependency analysis, middleware reliability, and a dedicated risk‑management platform.

SRElatency optimizationmetrics-driven development
0 likes · 16 min read
Service Risk Governance: Exploration, Mitigation, and Hands‑On Workshop
vivo Internet Technology
vivo Internet Technology
Oct 27, 2021 · Backend Development

JVM Garbage Collection Tuning for a Video Service to Reduce P99 Latency

By replacing the default Parallel GC with a ParNew‑CMS collector, enlarging the Young generation, fixing Metaspace settings, and tuning CMS occupancy thresholds, the video service cut Young and Full GC pauses dramatically, lowered Full GC count by over 80%, and achieved more than 30% P99 latency reduction, with some APIs improving up to 80%.

CMSGarbage CollectionJVM
0 likes · 16 min read
JVM Garbage Collection Tuning for a Video Service to Reduce P99 Latency
iQIYI Technical Product Team
iQIYI Technical Product Team
Nov 27, 2020 · Artificial Intelligence

Optimizing TensorFlow Serving Model Hot‑Update to Eliminate Latency Spikes in CTR Recommendation Systems

By adding model warm‑up files, separating load/unload threads, switching to the Jemalloc allocator, and isolating TensorFlow’s parameter memory from RPC request buffers, iQIYI’s engineers reduced TensorFlow Serving hot‑update latency spikes in high‑throughput CTR recommendation services from over 120 ms to about 2 ms, eliminating jitter.

AI inferenceModel Hot UpdateTensorFlow Serving
0 likes · 11 min read
Optimizing TensorFlow Serving Model Hot‑Update to Eliminate Latency Spikes in CTR Recommendation Systems