Skip to content
Developer Tools

TraceForge

337x faster raster-to-SVG conversion with GPU-accelerated dual-engine pipeline

337x
Processing Speed
45 min to 8 sec per asset
2,000+
Monthly Volume
Conversions with zero manual intervention
14
Vectorization Models
Presets tuned for different asset types
40-60%
File Size Reduction
Via SVGO optimization pipeline

The Challenge

Design teams at agencies and product companies were spending 45+ minutes per asset manually tracing rasters in Illustrator. Existing automated tools like Adobe's Live Trace produced noisy output requiring extensive cleanup. Batch processing didn't exist—each asset required individual attention. For teams processing hundreds of brand assets during rebrands or design system migrations, this meant weeks of tedious manual work. The core technical problem: raster-to-vector conversion requires understanding image topology, not just edge detection. Single-algorithm approaches either over-simplify (losing detail) or over-trace (creating thousands of unnecessary nodes).

The Approach

Started by benchmarking every open-source vectorization engine available. Potrace excelled at clean geometric shapes (logos, icons) while VTracer handled photographic complexity better. Rather than picking one, I built a dual-engine architecture letting users choose the right tool per asset. The key insight was that GPU-accelerated neural upscaling before vectorization dramatically improves output quality—feeding a 4x upscaled image to Potrace produces cleaner paths than running Potrace on the original. Built the pipeline on FastAPI with async processing, WebSocket progress streaming for long batch operations, and an SVGO post-processing stage that strips metadata and optimizes path data. Added 14 vectorization presets tuned for different asset types: logos, icons, illustrations, technical drawings, and photographs.

Tech Decisions

GPU Pipeline
CUDA + Neural Upscaling

4x neural upscaling before vectorization produces dramatically cleaner SVG paths. GPU acceleration makes this practical at batch scale—CPU-only upscaling would add 3-5 minutes per asset, negating the speed advantage.

Dual Engine
Potrace + VTracer

No single vectorization algorithm handles all image types well. Potrace excels at geometric shapes with clean edges; VTracer handles photographic complexity with gradient regions. Offering both eliminates the one-size-fits-all compromise.

Async Processing
FastAPI + WebSockets

Batch operations processing hundreds of assets need non-blocking execution with real-time progress feedback. FastAPI's native async support with WebSocket streaming gives users immediate visibility into long-running operations.

Technical Challenges

The Solution

TraceForge ships as a self-hosted web application with a React frontend and FastAPI backend. The GPU pipeline handles neural upscaling via CUDA-accelerated models, then routes to either Potrace or VTracer based on user selection or automatic detection. WebSocket connections stream real-time progress for batch operations processing hundreds of assets. The SVGO optimization stage runs 12 plugins that reduce SVG file sizes by 40-60% without visual degradation. Currently processing 2,000+ conversions monthly with zero manual intervention. The entire pipeline runs on a single RTX 3080 with 8-second average processing time per asset—down from 45 minutes of manual work.

Key Takeaways

Reusable Insights
  • GPU parallelism transforms image processing economics—operations that are impractical on CPU become trivial with even a mid-range GPU.
  • Offering multiple algorithms with sensible defaults beats any single one-size-fits-all approach for creative tools.
  • WebSocket progress streaming is table stakes for any operation longer than 3 seconds—users will assume the process is broken without real-time feedback.
  • Post-processing optimization (SVGO) often delivers more practical value than improving the core algorithm—a 40% smaller file loads faster regardless of path quality.

Related Projects

2025 / Developer Tools

Claude Pilot

Recovered 2+ hours daily lost to context-switching between terminal, database clients, and config files. Claude Code power users were drowning in fragmented tooling—no unified view of sessions, memory state, or MCP server health. Architected a native Electron control center with 25 tRPC endpoints managing PostgreSQL, Memgraph, and Qdrant memory systems. 80% test coverage, zero production incidents since launch.

ElectronReactTypeScripttRPCZod
Case Study
2024 / AI/ML

PhotoKeep Pro

Slashed cloud GPU costs by 73% while boosting restoration quality by 4dB over commercial alternatives. A restoration startup was burning $12k/month on fragmented API calls with inconsistent results. Engineered a unified orchestration layer managing 14+ deep learning models (SUPIR, HAT, CodeFormer) with thread-safe VRAM allocation and LRU eviction across 49GB. Now delivering 28.5dB PSNR quality at 99.95% uptime—outperforming Magnific AI and Topaz on blind tests.

PythonFastAPIPyTorchReactTypeScript
Case Study
2024 / AI/ML

PenQWEN

Reduced security assessment setup time from 4 hours to 12 minutes with zero hallucinated commands. Pentesting teams were wasting senior hours on boilerplate reconnaissance while generic LLMs generated dangerous garbage. Built a domain-adapted Qwen2.5 model through two-stage LoRA training: cybersecurity corpus adaptation, then agentic fine-tuning for tool calling and OPSEC. 3.6GB adapters trained on 12GB curated security data now automate 60% of routine enumeration tasks.

PythonPyTorchLoRAQwen2.5Transformers
Case Study
2025 / AI/ML

Voice Cloner

Built a production AI voice platform handling single-voice TTS, multi-speaker conversations, and full audiobook production from manuscript uploads — all on a single RTX 3080. The platform runs Qwen3-TTS 1.7B with 12-second P50 latency, 41+ curated voices, and zero-shot cloning from short reference audio. The Audiobook Studio parses DOCX/PDF/TXT manuscripts into chapters with dialogue detection, assigns character voices, applies pronunciation dictionaries, and exports distribution-ready M4B with chapter markers. Multi-voice conversations support drag-and-drop line ordering, per-line effects (speed, volume, gap), stage directions, multiple takes, ambient audio, and a waveform timeline editor. 99.95% uptime, 0.03% error rate, Stripe subscription billing.

PythonFastAPIPyTorchQwen3-TTSRedis
Case Study

Have a similar challenge?

I help teams solve complex technical problems. Let's discuss your project.

START_CONVERSATION()