Skip to content

Commit

Permalink
Update CHANGELOG.
Browse files Browse the repository at this point in the history
  • Loading branch information
drazvan committed Jun 30, 2023
1 parent 058668c commit 098bfcb
Showing 1 changed file with 11 additions and 9 deletions.
20 changes: 11 additions & 9 deletions CHANGELOG.md
Original file line number Diff line number Diff line change
Expand Up @@ -8,18 +8,20 @@ and this project adheres to [Semantic Versioning](https://semver.org/spec/v2.0.0

### Added

- Support for defining subflows.
- Support for using formatters to change how variables are included in a prompt template.
- Output parsers for prompt templates.
- The `verbose_v1` formatter and output parser to be used for smaller models that don't understand Colang very well in a few-shot manner.
- Support for including context variables in prompt templates.
- Support for chat models i.e. prompting with a sequence of messages.
- Experimental support for allowing the LLM to generate multi-step flows.
- Support for defining [subflows](docs/user_guide/colang-language-syntax-guide.md#subflows).
- Improved support for [customizing LLM prompts](docs/user_guide/advanced/prompt-customization.md)
- Support for using filters to change how variables are included in a prompt template.
- Output parsers for prompt templates.
- The `verbose_v1` formatter and output parser to be used for smaller models that don't understand Colang very well in a few-shot manner.
- Support for including context variables in prompt templates.
- Support for chat models i.e. prompting with a sequence of messages.
- Experimental support for allowing the LLM to generate [multi-step flows](docs/user_guide/configuration-guide.md#multi-step-generation).
- Example of using Llama Index from a guardrails configuration (#40).
- Example for using HuggingFace Endpoint LLMs with a guardrails configuration.
- [Example](examples/llm/hf_endpoint) for using HuggingFace Endpoint LLMs with a guardrails configuration.
- [Example](examples/llm/hf_pipeline_dolly) for using HuggingFace Pipeline LLMs with a guardrails configuration.
- Support to alter LLM parameters passed as `model_kwargs` in LangChain.
- CLI tool for running evaluations on the different steps (e.g., canonical form generation, next steps, bot message) and on existing rails implementation (e.g., moderation, jailbreak, fact-checking, and hallucination).
- Initial evaluation results for `text-davinci-003` and `gpt-3.5-turbo`.
- [Initial evaluation](nemoguardrails/eval/README.md) results for `text-davinci-003` and `gpt-3.5-turbo`.
- The `lowest_temperature` can be set through the guardrails config (to be used for deterministic tasks).

### Changed
Expand Down

0 comments on commit 098bfcb

Please sign in to comment.