Skip to content
This repository was archived by the owner on Mar 7, 2025. It is now read-only.

v0.5.0

Choose a tag to compare

@Tomas2D Tomas2D released this 13 Nov 21:46
· 148 commits to main since this release
c2a6cf6

🚀 Features / Enhancements

  • Added integration for LangChain Chat Models; see an example of generation and streaming.
  • Added support for LangChain Model Serialization (saving and loading models); see an example.
  • Added support for the Chat endpoint in Model class; see an example.
  • Added support for new moderation models (HAP, STIGMA, Implicit Hate) - not released on API yet but will be available soon.
  • Added type validation for input_tokens property in generate response.
  • Extend LangChain generation information / LLM Output (token_usage structure, generated tokens, stop_reason, conversation_id, created_at, ...).
  • Add optional raw_response=True/False parameter to generate_stream / generate_as_complete and generate methods to receive a raw response instead of unwrapped results.

🐛 Bug fixes

  • LangChain extension now correctly tokenizes the inputs (previously, the GPT2 tokenizer had been used).
  • Improve general error handling.

Full Changelog: v0.4.1...v0.5.0