
We made our VIN decoder 100x faster. Again
Follow-up to our previous post.
First, the v3 rewrite: SQLite was killing us on batch operations - 1000 VINs meant 4000 queries. We switched to binary indexes and now it's:
- Cold start: 200ms -> 23ms
- Single decode: 30ms -> 0.3ms
- Batch 1000: 4 seconds -> 300ms
Still fully offline, still no API keys.
On the EU data feedback: this is the real problem we've been digging into. Vehicle data is a mess globally, but especially across regions:
-US sources use 37k+ boolean feature keys with values embedded in key names ("12.3\" display": true)
- Canadian sources use nested category structures - better, but incompatible
- EU sources have great mechanical specs but almost no feature data
Same car, three regions, three completely different data contracts. And trim names are chaos:
- a US "Premium Plus" is a Canadian "Progressiv" is a German "45 TFSI quattro S tronic".
We're working on a schema standard (VIS) to normalize this. The goal: decode a VIN anywhere, get the same structured output regardless of source. Will share more when it's ready. As always - fully open source - code here: https://github.com/cardog-ai/corgi/
