Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Skip to content
#

distributed-inference

Here are 16 public repositories matching this topic...

Official impl. of ACM MM paper "Identity-Aware Attribute Recognition via Real-Time Distributed Inference in Mobile Edge Clouds". A distributed inference model for pedestrian attribute recognition with re-ID in an MEC-enabled camera monitoring system. Jointly training of pedestrian attribute recognition and Re-ID.

  • Updated Apr 26, 2020
  • Python

Turn any Kubernetes Cluster into a private LLM endpoint. One Helm command deploys distributed inference across commodity hardware. Raspberry Pi's, old servers, mixed architectures. OpenAI-Compatible API Powered by llama.cpp RPC

  • Updated Mar 4, 2026
  • Dockerfile

Pool your CUDA + ROCm GPUs into one OpenAI-compatible API. Speculative decoding proxy gives you 2-3x faster inference — for free, using hardware you already own. Stop renting GPU clouds. Be a tightwad.

  • Updated Feb 20, 2026
  • Python

Improve this page

Add a description, image, and links to the distributed-inference topic page so that developers can more easily learn about it.

Curate this topic

Add this topic to your repo

To associate your repository with the distributed-inference topic, visit your repo's landing page and select "manage topics."

Learn more