diff --git a/.release-please-manifest.json b/.release-please-manifest.json index 355faa1..04733b0 100644 --- a/.release-please-manifest.json +++ b/.release-please-manifest.json @@ -1,3 +1,3 @@ { - ".": "0.12.3" + ".": "0.13.0" } \ No newline at end of file diff --git a/CHANGELOG.md b/CHANGELOG.md index debf3ad..d6a238b 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,118 +1,127 @@ -# Changelog - -## Unreleased - -## 0.12.3 - -* Make `ChatOutput` public - -## 0.12.2 - -* Fix missing import in doc string - -## 0.12.1 - -* Add `Client::chat` method to send chat messages to a model - -## 0.12.0 - -* Add `Client::tokenizer_by_model` to fetch the Tokenizer for a given model name - -## 0.11.0 - -* Add `with_maximum_tokens` method to `Prompt` -* Remove maximum tokens argument from `Prompt::from_text` -* Make maximum tokens optional - -## 0.10.1 - -* Fix: Version number in Cargo.toml - -## 0.10.0 - -* Add the option to have authentication exclusively on a per request basis, without the need to specify a dummy token. -* Rename `Client::new` to `Client::with_authentication`. - -## 0.9.0 - -* Add `How::api_token` to allow specifying API tokens for individual requests. - -## 0.8.0 - -* Add `Error::Unavailable` to decouple service unavailability from 'queue full' 503 responses. - -## 0.7.1 - -* Add `Client::tokenize` and `Client::detokenize`. Thanks to @andreaskoepf - -## 0.7.0 - -* Add `client_timeout` to `How` -* Remove builder-methods from `How` as it introduced an unnecessary level of indirection - -## 0.6.1 - -* Add `explanation` to `Client` for submitting explanation requests to the API -* Add `be_nice`-builder-method to `How` to make maintaining backwards compatibility easier - -## 0.6.0 - -* Add `start_with_one_of` option to `Sampling`, to force the model to start a completion with one of several options. - -## 0.5.7 - -* Add new `Prompt` method `join_consecutive_text_items` to make it easier to construct few shot prompts and other such use cases programmatically, without introducing strange tokenization side effects. - -## 0.5.6 - -* Allow for executing `TaskSemanticEmbedding` without specifying models. - -## 0.5.5 - -* Support for processing images already in memory via `Modality::from_image`. - -## 0.5.4 - -* `Modality::from_image_path` now works with string literals. - -## 0.5.3 - -* Fix version number - -## 0.5.2 - -* Preprocess image on client side - -## 0.5.1 - -* Minimal support for sending multimodal prompts - -## 0.5.0 - -* Removed deprecated function `Client::complete`. It has been replaced with `Client::execute`. -* Introduced `how` parameter to `Client::execute` in order to control whether the request has the nice flag set, or not. - -## 0.4.1 - -* Allow for `Prompt::from_text` to be called with `String` or any other type which implement `Into>`. This enables prompts to take ownership of their values which is practical for use cases there you want to return prompts from functions, which are based on locally generated strings. - -## 0.4.0 - -* Support for stop sequences - -## 0.3.0 - -* Add `TaskSemanticEmbedding`. -* `Completion` renamed to `CompletionOutput` - -## 0.2.0 - -* Add `Error::Busy` to conveniently handle busy models. - -## 0.1.1 - -* Fix: `Client::new` did not work due to a missing `https://` in the base URL. - -## 0.1.0 - -* Initial release allows for sending receiving simple (not all parameters of the HTTP API are supported) completion requests. +# Changelog + +## [0.13.0](https://github.com/Aleph-Alpha/aleph-alpha-client-rs/compare/aleph-alpha-client-v0.12.3...aleph-alpha-client-v0.13.0) (2024-10-24) + +### ⚠ BREAKING CHANGES + +* do not hide role behind enum + +### Features + +* add with_messages method to construct TaskChat ([8fc5c10](https://github.com/Aleph-Alpha/aleph-alpha-client-rs/commit/8fc5c10e4b24c351c70c16d0d7405e2a5cc0f4f8)) +* do not hide role behind enum ([6ef29c5](https://github.com/Aleph-Alpha/aleph-alpha-client-rs/commit/6ef29c5d74824b801850021d3c3be550f0502057)) + +## 0.12.3 + +* Make `ChatOutput` public + +## 0.12.2 + +* Fix missing import in doc string + +## 0.12.1 + +* Add `Client::chat` method to send chat messages to a model + +## 0.12.0 + +* Add `Client::tokenizer_by_model` to fetch the Tokenizer for a given model name + +## 0.11.0 + +* Add `with_maximum_tokens` method to `Prompt` +* Remove maximum tokens argument from `Prompt::from_text` +* Make maximum tokens optional + +## 0.10.1 + +* Fix: Version number in Cargo.toml + +## 0.10.0 + +* Add the option to have authentication exclusively on a per request basis, without the need to specify a dummy token. +* Rename `Client::new` to `Client::with_authentication`. + +## 0.9.0 + +* Add `How::api_token` to allow specifying API tokens for individual requests. + +## 0.8.0 + +* Add `Error::Unavailable` to decouple service unavailability from 'queue full' 503 responses. + +## 0.7.1 + +* Add `Client::tokenize` and `Client::detokenize`. Thanks to @andreaskoepf + +## 0.7.0 + +* Add `client_timeout` to `How` +* Remove builder-methods from `How` as it introduced an unnecessary level of indirection + +## 0.6.1 + +* Add `explanation` to `Client` for submitting explanation requests to the API +* Add `be_nice`-builder-method to `How` to make maintaining backwards compatibility easier + +## 0.6.0 + +* Add `start_with_one_of` option to `Sampling`, to force the model to start a completion with one of several options. + +## 0.5.7 + +* Add new `Prompt` method `join_consecutive_text_items` to make it easier to construct few shot prompts and other such use cases programmatically, without introducing strange tokenization side effects. + +## 0.5.6 + +* Allow for executing `TaskSemanticEmbedding` without specifying models. + +## 0.5.5 + +* Support for processing images already in memory via `Modality::from_image`. + +## 0.5.4 + +* `Modality::from_image_path` now works with string literals. + +## 0.5.3 + +* Fix version number + +## 0.5.2 + +* Preprocess image on client side + +## 0.5.1 + +* Minimal support for sending multimodal prompts + +## 0.5.0 + +* Removed deprecated function `Client::complete`. It has been replaced with `Client::execute`. +* Introduced `how` parameter to `Client::execute` in order to control whether the request has the nice flag set, or not. + +## 0.4.1 + +* Allow for `Prompt::from_text` to be called with `String` or any other type which implement `Into>`. This enables prompts to take ownership of their values which is practical for use cases there you want to return prompts from functions, which are based on locally generated strings. + +## 0.4.0 + +* Support for stop sequences + +## 0.3.0 + +* Add `TaskSemanticEmbedding`. +* `Completion` renamed to `CompletionOutput` + +## 0.2.0 + +* Add `Error::Busy` to conveniently handle busy models. + +## 0.1.1 + +* Fix: `Client::new` did not work due to a missing `https://` in the base URL. + +## 0.1.0 + +* Initial release allows for sending receiving simple (not all parameters of the HTTP API are supported) completion requests. diff --git a/Cargo.toml b/Cargo.toml index c86446f..36c1ab9 100644 --- a/Cargo.toml +++ b/Cargo.toml @@ -1,6 +1,6 @@ [package] name = "aleph-alpha-client" -version = "0.12.3" +version = "0.13.0" edition = "2021" description = "Interact with large language models provided by the Aleph Alpha API in Rust code" license = "MIT"