LiteLLM

AI Data Loss Prevention (DLP) with Microsoft Presidio

Even if you have deployed an enterprise GPT platform as a dedicated workspace for your organization, it remains best practice to avoid storing unnecessary PII in your AI data infrastructure. In fact, many regulated industries & sectors such as education…
Read More

Private Enterprise GPT on Any Cloud with Inference APIs

Are your employees using the consumer versions of ChatGPT or Copilot (formerly Bing Chat) without your knowledge? Especially with hybrid work arrangements, this could be surreptitiously happening on employees’ mobile phones or personal laptops – even if company devices are…
Read More

RAG with any AI Model using Postgres pgVector + LibreChat

The addition of the RAG API microservice to LibreChat in version 0.7.0, the most rapidly trending open source ChatGPT clone, swings the door open to chatting with PDFs and documents using any supported AI model, in a private, self-hosted environment.…
Read More

Serverless Deployment of AI Middleware, LiteLLM, with Google Cloud Run

AI middleware is an emerging term for the layer of the technology stack that facilitates the interfacing of AI end user applications with the Large Language Models and GPU-accelerated machines that drive them. Here are the major sub-categories of this…
Read More

Proxies & Load Balancers for AI LLM Models (AI Middleware)

The Cambrianesque explosion of capable, open Large Language AI Models represents an opportunity to extend virtually any application with AI capabilities, but a strategy for managing multiple AI endpoints is clearly needed. Hosting open models in your own environment requires…
Read More