Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Effect of batch size on answer quality and speed. #630

Open
Snikch63200 opened this issue Oct 23, 2024 · 1 comment
Open

Effect of batch size on answer quality and speed. #630

Snikch63200 opened this issue Oct 23, 2024 · 1 comment
Labels
question Further information is requested

Comments

@Snikch63200
Copy link

Hello,

I changed batch size from 1 (default) to 8, 32 and saw no changes on paperQA behavioural (answer quality end speed), as follows :

settings=Settings(
    llm=f"openai/mixtral:8x7b",
    llm_config=local_llm_config,
    summary_llm=f"openai/mixtral:8x7b",
    summary_llm_config=local_llm_config,
    verbosity=3,
    temperature=0.7,
    batch_size=32,
    index_directory="indexes",
    paper_directory="./Papers",
    agent=AgentSettings(
                        agent_llm_config=local_llm_config,
                        agent_llm=f"openai/mixtral:8x7b",
                        agent_type="ToolSelector",
                        ),
    )

Is batch size have any impact on paperQA behavioural ?

@dosubot dosubot bot added the question Further information is requested label Oct 23, 2024
Copy link

dosubot bot commented Oct 23, 2024

The batch_size parameter in the paperQA codebase is primarily used for processing text documents in batches during the embedding process. It determines how many documents are processed together in a single batch when calling the embedding function [1]. However, there is no specific information available in the codebase that directly links changes in batch_size to variations in answer quality or speed for the openai/mixtral:8x7b model [1].

The impact of batch_size might not be noticeable in terms of answer quality and speed, especially if the bottleneck is elsewhere in the system or if the model and hardware can handle larger batches efficiently.

To continue talking to Dosu, mention @dosu.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
question Further information is requested
Projects
None yet
Development

No branches or pull requests

1 participant