All articles tagged with TensorFlow
Cloud latency kills real-time AI. In the wake of the Schrems II ruling, moving inference to the edge isn't just about performance—it's about compliance. Here is the 2020 architecture for deploying quantized TensorFlow models on Norwegian infrastructure.
Stop wrapping your Keras models in Flask. Learn how to deploy TensorFlow Serving via Docker on high-performance NVMe infrastructure for sub-100ms inference times while keeping your data compliant with Norwegian standards.
Stop burning budget on V100s for simple inference. We benchmark the new NVIDIA T4 against the Pascal generation and show you how to deploy mixed-precision models on Ubuntu 18.04 using nvidia-docker2.
Latency kills AI projects. We dissect CPU threading, TensorFlow 1.x configurations, and why NVMe storage is non-negotiable for production models in 2019.
It is 2017, and TensorFlow 1.0 has changed the game. But throwing a Titan X at your model is useless if your I/O is choking the pipeline. Here is how to architecture a training stack that actually saturates the bus, strictly for Norwegian data compliance.
Stop serving models with Flask. Learn how to deploy TensorFlow 1.0 candidates using gRPC and Docker for sub-millisecond inference latency on Norwegian infrastructure.