π₯ CPU and heap profiling and flamegraph visualization tool built on top of @platformatic/react-pprof.
- Dual Profiling: Captures both CPU and heap profiles concurrently for comprehensive performance insights
- Auto-Start Profiling: Profiling starts immediately when using
flame run(default behavior) - Automatic Flamegraph Generation: Interactive HTML flamegraphs are created automatically for both CPU and heap profiles on exit
- Sourcemap Support: Automatically translates transpiled code locations back to original source files (TypeScript, bundled JavaScript, etc.)
- Clear File Path Display: Shows exact paths and browser URLs for generated files
- Manual Control: Optional manual mode with signal-based control using
SIGUSR2 - Interactive Visualization: WebGL-based HTML flamegraphs with zoom, search, and filtering
- CLI Interface: Simple command-line tool for profiling and visualization
- Zero Config: Works out of the box with sensible defaults
npm install -g @platformatic/flame# Start profiling your application (profiling begins immediately)
flame run server.js
# The application runs with CPU and heap profiling active
# When you stop the app (Ctrl-C or normal exit), you'll see:
# π₯ CPU profile written to: cpu-profile-2025-08-27T12-00-00-000Z.pb
# π₯ Heap profile written to: heap-profile-2025-08-27T12-00-00-000Z.pb
# π₯ Generating CPU flamegraph...
# π₯ CPU flamegraph generated: cpu-profile-2025-08-27T12-00-00-000Z.html
# π₯ Generating heap flamegraph...
# π₯ Heap flamegraph generated: heap-profile-2025-08-27T12-00-00-000Z.html
# π₯ Open file:///path/to/cpu-profile-2025-08-27T12-00-00-000Z.html in your browser to view the CPU flamegraph
# π₯ Open file:///path/to/heap-profile-2025-08-27T12-00-00-000Z.html in your browser to view the heap flamegraph# Start profiling in manual mode (requires signals to start)
flame run --manual server.js
# In another terminal, toggle profiling on/off
kill -USR2 <PID>
# Or use the built-in toggle command
flame toggle# Generate HTML flamegraph from pprof file
flame generate cpu-profile-2024-01-01T12-00-00-000Z.pb
# Specify custom output file
flame generate -o my-flamegraph.html profile.pb.gzflame [options] <command>
Commands:
run <script> Run a script with profiling enabled
generate <pprof-file> Generate HTML flamegraph from pprof file
toggle Toggle profiling for running flame processes
Options:
-o, --output <file> Output HTML file (for generate command)
-m, --manual Manual profiling mode (require SIGUSR2 to start)
-d, --delay <value> Delay before starting profiler (ms, 'none', or 'until-started')
-s, --sourcemap-dirs <dirs> Directories to search for sourcemaps (colon/semicolon-separated)
--node-options <options> Node.js CLI options to pass to the profiled process
-h, --help Show help message
-v, --version Show version numberconst { startProfiling, generateFlamegraph, parseProfile } = require('@platformatic/flame')
// Start profiling a script with auto-start (default)
const { pid, toggleProfiler } = startProfiling('server.js', ['--port', '3000'], { autoStart: true })
// Or start in manual mode
const { pid, toggleProfiler } = startProfiling('server.js', ['--port', '3000'], { autoStart: false })
console.log(`Started profiling process ${pid}`)
// Toggle profiling programmatically (useful in manual mode)
toggleProfiler()
// Generate interactive flamegraph from pprof file
await generateFlamegraph('profile.pb.gz', 'flamegraph.html')
// Parse profile data
const profile = await parseProfile('profile.pb')- Auto-Start Mode (Default): Both CPU and heap profiling begin immediately when
flame runstarts your script - Auto-Generation on Exit: Profile (.pb) files and interactive HTML flamegraphs are automatically created for both CPU and heap profiles when the process exits
- Manual Mode: Use
--manualflag to requireSIGUSR2signals for start/stop control (no auto-HTML generation) - Interactive Visualization: The
@platformatic/react-pproflibrary generates interactive WebGL-based HTML flamegraphs for both profile types
When profiling transpiled or bundled applications (TypeScript, Webpack, ESBuild, etc.), the flame tool can automatically translate stack traces from generated code back to your original source files using sourcemaps.
# Profile a TypeScript application with sourcemap support
flame run --sourcemap-dirs=dist server.js
# Search multiple directories (colon or semicolon separated)
flame run --sourcemap-dirs=dist:build:out server.js- The tool searches specified directories for
.map,.js.map,.cjs.map, and.mjs.mapfiles - During profiling, stack frame locations are automatically translated from generated code to source locations
- The resulting flamegraph shows your original source file paths and line numbers
const { startProfiling } = require('@platformatic/flame')
startProfiling('dist/server.js', [], {
autoStart: true,
sourcemapDirs: ['dist', 'build'] // Can be a string or array
})Sourcemap support can also be controlled via environment variables:
# Specify directories (colon/semicolon separated)
export FLAME_SOURCEMAP_DIRS="dist:build"
# Then run normally
flame run server.js- Sourcemaps are loaded at startup and applied during profile capture
- The tool excludes
node_modulesdirectories when searching - If a sourcemap cannot be found or fails to parse, the original generated location is preserved
- Both CPU and heap profiles benefit from sourcemap translation
Profile files are saved with timestamps in the format:
cpu-profile-2024-01-01T12-00-00-000Z.pb
heap-profile-2024-01-01T12-00-00-000Z.pb
Both CPU and heap profiles share the same timestamp for easy correlation. The files are compressed Protocol Buffer format compatible with the pprof ecosystem.
// server.js
const express = require('express')
const app = express()
app.get('/', (req, res) => {
// Your application logic
res.json({ message: 'Hello World' })
})
app.listen(3000)Auto-Start Mode (Recommended):
# Profile the Express app (profiling starts immediately)
flame run server.js
# In another terminal, make some requests while profiling is active
curl http://localhost:3000
curl http://localhost:3000
curl http://localhost:3000
# Stop the server (Ctrl-C) to automatically save profiles and generate HTML flamegraphs
# You'll see the exact file paths and browser URLs in the output:
# π₯ CPU profile written to: cpu-profile-2025-08-27T15-30-45-123Z.pb
# π₯ Heap profile written to: heap-profile-2025-08-27T15-30-45-123Z.pb
# π₯ CPU flamegraph generated: cpu-profile-2025-08-27T15-30-45-123Z.html
# π₯ Heap flamegraph generated: heap-profile-2025-08-27T15-30-45-123Z.html
# π₯ Open file:///path/to/cpu-profile-2025-08-27T15-30-45-123Z.html in your browser to view the CPU flamegraph
# π₯ Open file:///path/to/heap-profile-2025-08-27T15-30-45-123Z.html in your browser to view the heap flamegraphManual Mode:
# Profile the Express app in manual mode
flame run --manual server.js
# In another terminal, start profiling
flame toggle
# Make some requests
curl http://localhost:3000
curl http://localhost:3000
# Stop profiling and save profile
flame toggle
# Generate flamegraph
flame generate cpu-profile-*.pbTwo comprehensive benchmark applications are included for testing profiling overhead and feature Express.js middlewares and Fastify plugins with realistic computation endpoints.
# Run individual apps without profiling
node examples/express-benchmark-app.js
node examples/fastify-benchmark-app.js
# Run comprehensive performance overhead analysis
node examples/run-performance-benchmark.jsBoth apps include:
- Health check endpoint (no computation)
- Light computation endpoint (~1k iterations)
- Medium computation endpoint (~10k iterations)
- Heavy computation endpoint (~100k iterations)
- Multiple middleware layers for realistic overhead testing
- Comprehensive test suites
Based on comprehensive benchmarks with 10 concurrent connections over 10 seconds per endpoint:
Express.js Framework:
| Endpoint | Load Level | Without Profiling | With Profiling | Throughput Overhead | Latency Overhead |
|---|---|---|---|---|---|
| Health Check | Minimal | 13,571 req/s | 13,752 req/s | -1.3% | -6.3% |
| Light Computation | Low | 10,187 req/s | 9,979 req/s | +2.0% | +12.0% |
| Medium Computation | Moderate | 71 req/s | 66 req/s | +6.1% | +6.3% |
| Heavy Computation | High | 295 req/s | 291 req/s | +1.3% | +1.4% |
| Mixed Computation | Very High | 56 req/s | 53 req/s | +5.2% | +5.8% |
Express Summary: Average throughput overhead of 2.7% and latency overhead of 3.9%
Fastify Framework:
| Endpoint | Load Level | Without Profiling | With Profiling | Throughput Overhead | Latency Overhead |
|---|---|---|---|---|---|
| Health Check | Minimal | 41,174 req/s | 38,747 req/s | +5.9% | 0.0% |
| Light Computation | Low | 35,056 req/s | 32,847 req/s | +6.3% | 0.0% |
| Medium Computation | Moderate | 3,235 req/s | 3,126 req/s | +3.4% | +4.2% |
| Heavy Computation | High | 345 req/s | 336 req/s | +2.6% | +2.6% |
| Mixed Computation | Very High | 311 req/s | 304 req/s | +2.3% | +2.3% |
Fastify Summary: Average throughput overhead of 4.1% and latency overhead of 1.8%
CSV data is automatically generated when running node examples/run-performance-benchmark.js for further analysis.
You can run load tests manually:
# Terminal 1: Start the test server with profiling
flame run examples/load-test-server.js
# Terminal 2: Run load test with autocannon
npx autocannon -c 10 -d 10 http://localhost:3000
# Terminal 1: Toggle profiling (start/stop)
kill -USR2 <PID>- Node.js >= 18.0.0
@datadog/pproffor CPU profiling@platformatic/react-pproffor flamegraph generation
MIT