Skip to content

Releases: chigkim/VOLlama

VOLlama v0.2.0

08 Dec 16:32
Compare
Choose a tag to compare

Changelog

  • Moved shortcuts from rag menus to chat menu
  • Attach URL: CTRL+U
  • Increased Timeout for Ollama and OpenAILike to 1 hour
  • Min_p sampling
  • Minor UI Fixed
  • Provided patch to work with latest dependencies
  • Bug fixes and performance improvements”

Download

VOLlama v0.1.4-beta.3

04 Jun 18:00
Compare
Choose a tag to compare
VOLlama v0.1.4-beta.3 Pre-release
Pre-release

Change log

  • OpenAILike API for OpenAI Compatible platform such as openrouter.ai
  • Moved Set Host from advance menu to Chat Menu > Api settings.
  • If you use Ollama in another machine, please reset the address for Ollama in Chat Menu> API Settings> Ollama > Base Url.

Download

VOLlama v0.1.4-beta.2

26 May 12:08
Compare
Choose a tag to compare
VOLlama v0.1.4-beta.2 Pre-release
Pre-release

Change log

  • Added edit menu and moved clear and copy from chat menu to edit menu.
  • Added models on chat menu.
  • When create a new prompt, it saves automatically.
  • On prompt manager, changed duplicate to new, and replace to save.
  • When opening prompt manager on Mac, it now correctly selects the current prompt.

Download

VOLlama v0.1.4-beta.1

23 May 23:48
Compare
Choose a tag to compare
VOLlama v0.1.4-beta.1 Pre-release
Pre-release

Change log

  • Prompt manager
  • Import Awesome ChatGPT Prompts
  • Partial support for GPT-4O: Throws an error in some cases but ignore the error.
  • Able to attach entire document for long context model.

Download

VOLlama v0.1.3

06 May 01:39
Compare
Choose a tag to compare

Change log

  • Shortcut for API Settings: Control+(or Command)+Shift+A
  • Creates indexes for only supported file types.
  • Smaller package size

Download

VOLlama v0.1.2

27 Apr 12:53
Compare
Choose a tag to compare

Change log

  • Rolled back default context length to 4096 for better model compatibility.
  • No longer create a single file package to speed up load time.

Download

VOLlama v0.1.1

26 Apr 18:01
Compare
Choose a tag to compare

Change log

  • Increased default context length to 8192 to match Llama3.
  • Clearer status message

Download

VOLlama v0.1.0-alpha.18

29 Feb 13:19
Compare
Choose a tag to compare
Pre-release

Change log

  • Clear last message clears last user message as well as model's response.
  • Editing the message history does not move the history cursor to the bottom, so you can keep editing another message.
  • Pressing esc moves history cursor to the bottom for new message.
  • Fixed incorrect history cursor when there's a system message.
  • Changed send button to edit when editing history.
  • Paste last cleared user message into the prompt for editing or simply resending.
  • Edit history: alt+(or option on mac)+up/down
  • Embedding with Ollama
  • Make sure to download nomic-embed-text: ollama pull nomic-embed-text
  • If you have saved index, you need to index and save again.
  • Support for OpenAI and Gemini
  • Better error message if no text found from semantic search.
  • Able to index individual files
  • Fixed bug when index gets reset when changing model
  • Token counter on the status bar for RAG.
  • Fixed response mode not working.
  • Output similarity score for each chunk.
  • Able to save and load index
  • Rag settings include: chunk_size, chunk_overlap, similarity_top_k, similarity_cutoff, show_context, ragResponseMode
  • Different response Modes for RAG
  • Option to display context sent to RAG system
  • Epub reader
  • Global settings improvement
  • Retrieval-Augmented Generation (RAG): See below for instruction on how to use it.
  • Many bug fixes
  • Generation parameters in advance menu
  • Voice and rate gets saved into persistent settings
  • Changed from NSSpeechSynthesizer to AVSpeechSynthesizer to avoid random crash
  • Status bar and stats (Accessible in Windows only)
  • Check for update
  • Fixed message order.
  • Checkbox Speak response in chat menu
  • Start speaking during generation
  • ESC: stop generating and speaking and focus to prompt
  • Configure voice (Windows only)
  • Confirmation before delete a model.
  • Attach image: Control(Command)+I
  • Create multiline prompt with Control(Command)+Enter
  • Set system message in advance menu
  • Delete model in advance menu
  • Persistent Settings for system message and host address
  • Save and recall chat history
  • Focus model list: Control(Command)+L

Retrieval-Augmented Generation

  • Go to Chat menu > Attach url or press Control(Command)+U.
  • Enter https://www.apple.com/apple-vision-pro/
  • Wait until the document is indexed.
  • In the message field, type "/q What can you do with Vision Pro?" without the quotes.
  • Putting "/q " in the beginning of your messsage triggers LlamaIndex to kick in.
  • You can also index a folder with documents in them. It'll index all the documents it can index such as pdf, txt, docs, etc including inside all the sub folders.

Copy model in advance menu

It lets you duplicate an existing model through modelfile, and you can use it like preset with different name and parameters like temperature, repeat penalty, Maximum length to generate, context length, etc. It does not duplicate the model weight files, so you won't wasting your storage space even if you duplicate and create bunch.

See modelfile for more detail.

It is very important for Mac users to turn off smart quotes before opening copy model dialog. If you see left double quotation mark instead of quotation mark in modelfile, that means you have smart quotes on.

  • MacOS 13 Ventura or later: System settings > Keyboard > Edit input sorce > turn off smart quotes
  • Before Ventura: System preferences > keyboard > text > uncheck smart quotes

VOLlama v0.1.0-Alpha.14

24 Feb 21:38
Compare
Choose a tag to compare
Pre-release

Change log

  • Embedding with Ollama
  • Important to download embedding model: ollama pull nomic-embed-text
  • If you have saved index, you need to index and save again.
  • Support for OpenAI and Gemini
  • Better error message if no text found from semantic search.
  • Able to index individual files
  • Fixed bug when index gets reset when changing model
  • Token counter on the status bar for RAG.
  • Fixed response mode not working.
  • Output similarity score for each chunk.
  • Able to save and load index
  • Rag settings include: chunk_size, chunk_overlap, similarity_top_k, similarity_cutoff, show_context, ragResponseMode
  • Different response Modes for RAG
  • Option to display context sent to RAG system
  • Epub reader
  • Global settings improvement
  • Retrieval-Augmented Generation (RAG): See below for instruction on how to use it.
  • Many bug fixes
  • Generation parameters in advance menu
  • Voice and rate gets saved into persistent settings
  • Changed from NSSpeechSynthesizer to AVSpeechSynthesizer to avoid random crash
  • Status bar and stats (Accessible in Windows only)
  • Check for update
  • Fixed message order.
  • Checkbox Speak response in chat menu
  • Start speaking during generation
  • ESC: stop generating and speaking and focus to prompt
  • Configure voice (Windows only)
  • Confirmation before delete a model.
  • Attach image: Control(Command)+I
  • Create multiline prompt with Control(Command)+Enter
  • Set system message in advance menu
  • Delete model in advance menu
  • Persistent Settings for system message and host address
  • Save and recall chat history
  • Focus model list: Control(Command)+L

Retrieval-Augmented Generation

  • Go to Chat menu > Attach url or press Control(Command)+U.
  • Enter https://www.apple.com/apple-vision-pro/
  • Wait until the document is indexed.
  • In the message field, type "/q What can you do with Vision Pro?" without the quotes.
  • Putting "/q " in the beginning of your messsage triggers LlamaIndex to kick in.
  • You can also index a folder with documents in them. It'll index all the documents it can index such as pdf, txt, docs, etc including inside all the sub folders.

Copy model in advance menu

It lets you duplicate an existing model through modelfile, and you can use it like preset with different name and parameters like temperature, repeat penalty, Maximum length to generate, context length, etc. It does not duplicate the model weight files, so you won't wasting your storage space even if you duplicate and create bunch.

See modelfile for more detail.

It is very important for Mac users to turn off smart quotes before opening copy model dialog. If you see left double quotation mark instead of quotation mark in modelfile, that means you have smart quotes on.

  • MacOS 13 Ventura or later: System settings > Keyboard > Edit input sorce > turn off smart quotes
  • Before Ventura: System preferences > keyboard > text > uncheck smart quotes

VOLlama v0.1.0-alpha.13

18 Feb 15:32
Compare
Choose a tag to compare
Pre-release

Change log

  • Better error message for when no text found from semantic search.
  • Rag menu to index individual file
  • Fixed bug when index gets reset when changing model
  • Token counter on the status bar for RAG.
  • Fixed response mode not working.
  • Output similarity score for each chunk.
  • Able to save and load index
  • Rag settings include: chunk_size, chunk_overlap, similarity_top_k, similarity_cutoff, show_context, ragResponseMode
  • Different response Modes for RAG
  • Option to display context sent to RAG system
  • Epub reader
  • Global settings improvement
  • Retrieval-Augmented Generation (RAG): See below for instruction on how to use it.
  • Many bug fixes
  • Generation parameters in advance menu
  • Voice and rate gets saved into persistent settings
  • Changed from NSSpeechSynthesizer to AVSpeechSynthesizer to avoid random crash
  • Status bar and stats (Accessible in Windows only)
  • Check for update
  • Fixed message order.
  • Checkbox Speak response in chat menu
  • Start speaking during generation
  • ESC: stop generating and speaking and focus to prompt
  • Configure voice (Windows only)
  • Confirmation before delete a model.
  • Attach image: Control(Command)+I
  • Create multiline prompt with Control(Command)+Enter
  • Set system message in advance menu
  • Delete model in advance menu
  • Persistent Settings for system message and host address
  • Save and recall chat history
  • Focus model list: Control(Command)+L

Retrieval-Augmented Generation

  • Go to Chat menu > Attach url or press Control(Command)+U.
  • Enter https://www.apple.com/apple-vision-pro/
  • Wait until the document is indexed.
  • In the message field, type "/q What can you do with Vision Pro?" without the quotes.
  • Putting "/q " in the beginning of your messsage triggers LlamaIndex to kick in.
  • You can also index a folder with documents in them. It'll index all the documents it can index such as pdf, txt, docs, etc including inside all the sub folders.

Copy model in advance menu

It lets you duplicate an existing model through modelfile, and you can use it like preset with different name and parameters like temperature, repeat penalty, Maximum length to generate, context length, etc. It does not duplicate the model weight files, so you won't wasting your storage space even if you duplicate and create bunch.

See modelfile for more detail.

It is very important for Mac users to turn off smart quotes before opening copy model dialog. If you see left double quotation mark instead of quotation mark in modelfile, that means you have smart quotes on.

  • MacOS 13 Ventura or later: System settings > Keyboard > Edit input sorce > turn off smart quotes
  • Before Ventura: System preferences > keyboard > text > uncheck smart quotes