Baseten Blog | Page 8
A checklist for switching to open source ML models
Transitioning from using ML models via closed source APIs to open source ML models? This checklist provides all necessary resources for the shift.
A guide to LLM inference and performance
Learn if LLM inference is compute or memory bound to fully utilize GPU power. Get insights on better GPU resource utilization.
Pinning ML model revisions for compatibility and security
Pin versions of open source packages like PyPi's transformers to avoid breaking changes or security issues; similarly, pin model revisions for stability.
Deployment and inference for open source text embedding models
Text embedding models convert text into semantic vectors. Numerous open source models cater to search, recommendation, classification & LLM-augmented retrieval.
New in October 2023
All-new model management, a text embedding model that matches OpenAI, and misgif, the most fun you’ll have with AI all week.
Jina AI’s jina-embeddings-v2: an open source text embedding model that matches OpenAI’s ada-002
Jina AI released jina-embeddings-v2-base-en, a text embedding model that matches OpenAI’s ada-002 model in both benchmark performance and context window length.
New in September 2023
Mistral 7B LLM, GPU comparisons, model observability features, and an open source AI event series
NVIDIA A10 vs A100 GPUs for LLM and Stable Diffusion inference
This article compares two popular GPUs—the NVIDIA A10 and A100—for model inference and discusses the option of using multi-GPU instances for larger models.
New in August 2023
Truss' latest update addresses key ML model serving issues. Discover how to speed up SDXL inference to 3s and build ChatGPT-like apps with Llama 2 & Chainlit.