Research
Financial AI performance benchmark
December 2025
WebSocket optimization
By improving the WebSocket architecture, we reduced startup time from ~46 seconds to immediate startup.
Problems
- Early design: Subscribing to every coin over WebSocket when selecting a coin caused a ~46s delay
- Wasted resources: Unnecessary subscriptions consumed bandwidth and CPU
- Slow startup: Poor first-run experience
Mitigations
- API-first analysis: Coin selection uses API-driven analysis instead of subscribing to all symbols up front
- Focused WebSockets: WebSockets are used primarily for position monitoring
- Subscription policy: Keep major symbols always on; add subscriptions only for open positions / active signals
- Batch cleanup: Release unused subscriptions in batches
Results
- Startup time: ~46s → immediate (~99% improvement)
- Resource use: ~70% fewer WebSocket subscriptions
- Memory: Fewer idle connections reduce memory pressure
API call optimization
Layered caching
To minimize API calls, we implemented a multi-layer caching system.
- AI caching: Cache repeated judgments under identical inputs
- Market data caching: Cache prices, volumes, and related market fields
- Coin selection caching: Cache selection results to shorten initial load
- Invalidation: Automatically invalidate when market conditions change materially
API reduction outcomes
- Initial load: ~4 minutes → immediate (~99% improvement)
- API dependency: 100% → ~30% (~70% reduction)
- Cost: Fewer calls reduce rate-limit pressure and operational costs
- Stability: Lower risk of hitting exchange rate limits
Memory efficiency
Database optimization
- Indexes: Add indexes for hot query paths
- Retention: Automatically prune old logs
- Compression: Store large blobs in compressed form where appropriate
Memory management
- WebSocket hygiene: Remove unused subscriptions to save memory
- Cache size caps: Bound caches to predictable memory usage
- GC-friendly patterns: Avoid retaining large graphs longer than necessary
Multi-exchange performance
Concurrent operation
We operate six exchanges concurrently while keeping stable performance without proportional slowdown.
- Isolation: Per-exchange monitoring paths reduce cross-talk
- Resource separation: Contain failures and resource spikes per venue
- Prioritization: Prioritize critical venues when needed
- Load spreading: Spread API calls across time windows to avoid bursts
Public R&D and investor lens
Technical quality
These benchmarks demonstrate engineering quality in a real financial AI stack.
- Startup: ~46s → immediate (~99% improvement)
- API efficiency: ~70% fewer calls
- Resource efficiency: Better memory and network utilization
- Scalability: Six concurrent exchanges without degradation
Operational efficiency
Performance work translates into lower operating cost and higher reliability.
- Cost: Fewer calls reduce rate-limit incidents and wasted cycles
- Stability: Lower chance of hitting exchange limits during volatility
- Scalability: Room to add venues without linear cost explosion
- UX: Fast startup and predictable runtime behavior
Conclusion
The benchmark shows that WebSocket optimization, API minimization, and memory efficiency can improve both operational efficiency and stability.
For public R&D and investors, these metrics support claims of technical quality, operational efficiency, and scalability.