Projects · Python · Model Serving · FastAPI · MLOps

Deployable AI

A lightweight local inference-serving toolkit for registering models and exposing prediction endpoints quickly

2024.06.23 · 1 min read · by Zhenlin Wang · updated 2024-08-30

Introduction

Deployable AI is a small toolkit for serving local model inference quickly. It follows a familiar model-package pattern: register a model artifact, provide an inference script, and expose the model through a predictable API endpoint.

Workflow

The repo centers on three pieces:

Notes

The project is intentionally narrow: it targets JSON request workflows and local development speed. It is useful as a compact serving layer when the goal is to validate model behavior before investing in heavier deployment infrastructure.