4 hour - incomplete run of: query_engine = AutoQueryEngine.from_parameters(documents=documents) #170
-
Stepping through https://abrahimzaman360.medium.com/introduction-to-autollm-c8cd31be2a5f in Google Colab, and at this step, using a text with 31K words, I killed the process after 4 hours. I can say the text is oddly formated: “And for me,” said Thorin. “And raspberry jam and apple-tart,” said “And mince-pies and cheese,” said Bofur. If I'm reading the OpenAI billing correctly, I was only charged $0.01, which I find hard to believe but anything is possible. A newbie question I know but is the excessive spacing in the text having an impact on the parsing? It wouldn't in awk, or at least noticeably, but I don't know about here. The query engine was 3.5 turbo if I'm reading the source correctly. Thanks! |
Beta Was this translation helpful? Give feedback.
Replies: 5 comments 5 replies
-
Hi @pdurusau, Thanks for trying our package! Could you share your code for better support? Note: Great news: our new release is out today! It includes updates to enhance user experience, including a progress bar for improved feedback during parsing. This should help in situations like yours. Update the package with Looking forward to your feedback! Best, |
Beta Was this translation helpful? Give feedback.
-
It's the most basic script, btw, I copied out of another colab window, not the one that failed, so I did use .from_defaults(documents=documents) Embarrassingly simple: from getpass import getpass os.environ["OPENAI_API_KEY"] = getpass("Enter your OpenAI API key:") !pip install autollm from google.colab import files from autollm import AutoQueryEngine, read_files_as_documents documents = read_files_as_documents(input_dir="/content") query_engine = AutoQueryEngine.from_defaults(documents=documents) Ah, the file is attached. Thanks for the quick response! Patrick |
Beta Was this translation helpful? Give feedback.
-
Update, same code and file as before, ran 21 hours, 15 minutes before reporting Embedding Token Usage. Then ran until it crashed at 12+ GB of memory in Colab. Would this have something to do with query_engine = AutoQueryEngine.from_defaults( ??? Not likely the best setting for default? |
Beta Was this translation helpful? Give feedback.
-
Hi @pdurusau, Thank you for reaching out and sharing the details of the issue you're experiencing with the AutoQueryEngine.from_defaults in your Google Colab environment. To better understand and assist you, could you please provide some additional information? Any screenshots, log messages, error messages etc.. I don't expect to llm_max_token parameter to be relevant with the current crash. Also you can share a reproducible code if you want. That would very much help me better investiage the issue. |
Beta Was this translation helpful? Give feedback.
-
Thanks! Running the suggestions now! Appreciate the help! |
Beta Was this translation helpful? Give feedback.
Hi @pdurusau,
Thanks for trying our package! Could you share your code for better support?
Note:
AutoQueryEngine.from_parameters
is deprecated. Please usefrom_defaults
now.Great news: our new release is out today! It includes updates to enhance user experience, including a progress bar for improved feedback during parsing. This should help in situations like yours. Update the package with
pip install --upgrade autollm[readers]
to access these enhancements.Looking forward to your feedback!
Best,
Talha