Baseten Blog | Page 9

Product

New in June 2023

LangChain adds Baseten integration, Falcon soars to the top of the LLM leaderboard

Hacks & projects

Three techniques to adapt LLMs for any use case

Prompt engineering, embeddings, vector databases, and fine-tuning are ways to adapt Large Language Models (LLMs) to run on your data for your use case

Community

What I learned from my AI startup’s internal hackathon

See hackathon projects from Baseten for ML infrastructure, inference, user experience, and streaming

ML models

Deploy Falcon-40B on Baseten

Deploy Falcon-40B and Falcon-7B, top-ranked open-source LLMs on HuggingFace, to Baseten's production-ready ML infrastructure.

Product

Deploy open-source models in a couple clicks from Baseten’s model library

An explanation of how Baseten's model library works for deploying and serving popular open-source models.

ML models

Getting started with foundation models

Summarizing foundation models, focusing on data type, scale, in-context learning, and fine-tuning, illustrated with Meta's LLaMA model family.

Product

New in May 2023

Explore new text generation and text-to-speech models, their GPU requirements, and join the community around open-source models.

GPU guides

Understanding NVIDIA’s Datacenter GPU line

This guide helps you navigate NVIDIA’s datacenter GPU lineup and map it to your model serving needs.

GPU guides

Comparing GPUs across architectures and tiers

So what are reliable metrics for comparing GPUs across architectures and tiers? We’ll consider core count, FLOPS, VRAM, and TDP.

1…89101112