Edge AI/ML
Model Optimization
TensorFlow Lite Deployment
apiVersion: apps/v1
kind: Deployment
metadata:
name: edge-inference
spec:
replicas: 1
selector:
matchLabels:
app: edge-ml
template:
spec:
containers:
- name: inference
image: tensorflow/serving:latest
resources:
limits:
cpu: "2"
memory: "4Gi"
nvidia.com/gpu: "1"
volumeMounts:
- name: model-store
mountPath: /models
env:
- name: MODEL_NAME
value: edge_model
- name: MODEL_BASE_PATH
value: /modelsONNX Runtime Optimization
Edge Configuration
Model Serving
Triton Inference Server
Best Practices
Last updated