0.15.0 (2024-12-10)
- add option to ask for special tokens in completion response
- rename start_with to complete_with
- clean up naming of methods to setup clients
- token env is called PHARIA_AI_TOKEN
- replace from_authentication by new from_env method
- add option to ask for special tokens in completion response (1dbcb77)
- clean up naming of methods to setup clients (5f18f38)
- rename base_url env variable to inference_url (b9a2fd2)
- rename start_with to complete_with (edd5590)
- replace from_authentication by new from_env method (d2da859)
- token env is called PHARIA_AI_TOKEN (827b44c)
0.14.0 (2024-11-28)
- Update
tokenizers
dependency to v0.21.0 ed5ea41
0.13.2 (2024-10-28)
- add stream completion method (c513a36)
0.13.1 (2024-10-24)
- add id to release workflow (4fb33c2)
0.13.0 (2024-10-24)
- do not hide role behind enum
- Make
ChatOutput
public
- Fix missing import in doc string
- Add
Client::chat
method to send chat messages to a model
- Add
Client::tokenizer_by_model
to fetch the Tokenizer for a given model name
- Add
with_maximum_tokens
method toPrompt
- Remove maximum tokens argument from
Prompt::from_text
- Make maximum tokens optional
- Fix: Version number in Cargo.toml
- Add the option to have authentication exclusively on a per request basis, without the need to specify a dummy token.
- Rename
Client::new
toClient::with_authentication
.
- Add
How::api_token
to allow specifying API tokens for individual requests.
- Add
Error::Unavailable
to decouple service unavailability from 'queue full' 503 responses.
- Add
Client::tokenize
andClient::detokenize
. Thanks to @andreaskoepf
- Add
client_timeout
toHow
- Remove builder-methods from
How
as it introduced an unnecessary level of indirection
- Add
explanation
toClient
for submitting explanation requests to the API - Add
be_nice
-builder-method toHow
to make maintaining backwards compatibility easier
- Add
start_with_one_of
option toSampling
, to force the model to start a completion with one of several options.
- Add new
Prompt
methodjoin_consecutive_text_items
to make it easier to construct few shot prompts and other such use cases programmatically, without introducing strange tokenization side effects.
- Allow for executing
TaskSemanticEmbedding
without specifying models.
- Support for processing images already in memory via
Modality::from_image
.
Modality::from_image_path
now works with string literals.
- Fix version number
- Preprocess image on client side
- Minimal support for sending multimodal prompts
- Removed deprecated function
Client::complete
. It has been replaced withClient::execute
. - Introduced
how
parameter toClient::execute
in order to control whether the request has the nice flag set, or not.
- Allow for
Prompt::from_text
to be called withString
or any other type which implementInto<Cow<'_, str>>
. This enables prompts to take ownership of their values which is practical for use cases there you want to return prompts from functions, which are based on locally generated strings.
- Support for stop sequences
- Add
TaskSemanticEmbedding
. Completion
renamed toCompletionOutput
- Add
Error::Busy
to conveniently handle busy models.
- Fix:
Client::new
did not work due to a missinghttps://
in the base URL.
- Initial release allows for sending receiving simple (not all parameters of the HTTP API are supported) completion requests.