Skip to content

Releases: meta-llama/llama-stack

v0.0.63

18 Dec 07:17
Compare
Choose a tag to compare

A small but important bug-fix release to update the URL datatype for the client-SDKs. The issue affected multimodal agentic turns especially.

Full Changelog: v0.0.62...v0.0.63

v0.0.62

18 Dec 02:39
Compare
Choose a tag to compare

What's Changed

A few important updates some of which are backwards incompatible. You must update your run.yamls when upgrading. As always look to templates/<distro>/run.yaml for reference.

  • Make embedding generation go through inference by @dineshyv in #606
  • [/scoring] add ability to define aggregation functions for scoring functions & refactors by @yanxi0830 in #597
  • Update the "InterleavedTextMedia" type by @ashwinb in #635
  • [NEW!] Experimental post-training APIs! #540, #593, etc.

A variety of fixes and enhancements. Some selected ones:

New Contributors

Full Changelog: v0.0.61...v0.0.62

v0.0.61

10 Dec 20:50
e2054d5
Compare
Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v0.0.55...v0.0.61

v0.0.55 release

23 Nov 17:14
Compare
Choose a tag to compare

What's Changed

  • Fix TGI inference adapter
  • Fix llama stack build in 0.0.54 by @dltn in #505
  • Several documentation related improvements
  • Fix opentelemetry adapter by @dineshyv in #510
  • Update Ollama supported llama model list by @hickeyma in #483

Full Changelog: v0.0.54...v0.0.55

Llama Stack 0.0.54 Release

22 Nov 00:36
Compare
Choose a tag to compare

What's Changed

  • Bugfixes release on top of 0.0.53
  • Don't depend on templates.py when print llama stack build messages by @ashwinb in #496
  • Restructure docs by @dineshyv in #494
  • Since we are pushing for HF repos, we should accept them in inference configs by @ashwinb in #497
  • Fix fp8 quantization script. by @liyunlu0618 in #500
  • use logging instead of prints by @dineshyv in #499

New Contributors

Full Changelog: v0.0.53...v0.0.54

Llama Stack 0.0.53 Release

20 Nov 22:18
Compare
Choose a tag to compare

🚀 Initial Release Notes for Llama Stack!

Added

  • Resource-oriented design for models, shields, memory banks, datasets and eval tasks
  • Persistence for registered objects with distribution
  • Ability to persist memory banks created for FAISS
  • PostgreSQL KVStore implementation
  • Environment variable placeholder support in run.yaml files
  • Comprehensive Zero-to-Hero notebooks and quickstart guides
  • Support for quantized models in Ollama
  • Vision models support for Together, Fireworks, Meta-Reference, and Ollama, and vLLM
  • Bedrock distribution with safety shields support
  • Evals API with task registration and scoring functions
  • MMLU and SimpleQA benchmark scoring functions
  • Huggingface dataset provider integration for benchmarks
  • Support for custom dataset registration from local paths
  • Benchmark evaluation CLI tools with visualization tables
  • RAG evaluation scoring functions and metrics
  • Local persistence for datasets and eval tasks

Changed

  • Split safety into distinct providers (llama-guard, prompt-guard, code-scanner)
  • Changed provider naming convention (implsinline, adaptersremote)
  • Updated API signatures for dataset and eval task registration
  • Restructured folder organization for providers
  • Enhanced Docker build configuration
  • Added version prefixing for REST API routes
  • Enhanced evaluation task registration workflow
  • Improved benchmark evaluation output formatting
  • Restructured evals folder organization for better modularity

Removed

  • llama stack configure command

What's Changed

Read more