Demo
1. Asynchronous Inference Server with FastAPI
First, run the backend inference server
$python inference_server.py
INFO: Started server process [18987]
INFO: Waiting for application startup.
INFO: Application startup complete.
INFO: Uvicorn running on http://127.0.0.1:8000 (Press CTRL+C to quit)
Then, on the different process, run the front end web app via streamlit. Specify the address of the inference server you booted up above by -s option. For example, if the server is booted at http://127.0.0.1:8000, pass 8000.
2. Realtime Inference
This requires a webcam connected to your computer.
First, uncomment the mode you want to try and comment all others out. Following example chose MiDAS to try.
model = BiWAKO.MiDAS("weights/mono_depth_small")
# model = BiWAKO.U2Net("mobile")
# model = BiWAKO.HumanParsing("human_attribute")
Now, just run the script
3. Video Inference
Sample usage of VideoPredictor. Assign the path to input mp4 to video and run the script.