Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add QELM to list. #850

Open
wants to merge 1 commit into
base: main
Choose a base branch
from
Open

Add QELM to list. #850

wants to merge 1 commit into from

Conversation

github-actions[bot]
Copy link
Contributor

Add QELM to list.


Closes #849

@github-actions github-actions bot added the submission Project submission label Dec 11, 2024
@CLAassistant
Copy link

CLAassistant commented Dec 11, 2024

CLA assistant check
All committers have signed the CLA.

Copy link

@Trivarian Trivarian left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Ted was here

Copy link
Member

@frankharkins frankharkins left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thank you!

@@ -0,0 +1,10 @@
name = "QELM"
url = "https://github.com/R-D-BioTech-Alaska/Qelm"
description = "An innovative project that merges the power of quantum computing with natural language processing to create a next-generation language model. Leveraging Qiskit and Qiskit Aer, QELM explores the potential of quantum circuits in enhancing language understanding and generation capabilities. Utilizing Qubits, Qelm can train a model that would normally take up gigabytes worth of data in llm files all the way down to miniature sizes. With this size, llm's can run instantly with no loss of capabilities or intelligence and on small computers instead requiring data centers to run single models. The goal is to create llm's that are instant, smarter and can be utilized anywhere and by anyone."
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

This description needs to be <135 characters, preferably focussing on what users can do with this package. Take a look at the other ecosystem projects for examples.

Also:

Utilizing Qubits, Qelm can train a model that would normally take up gigabytes worth of data in llm files all the way down to miniature sizes. With this size, llm's can run instantly with no loss of capabilities or intelligence and on small computers instead requiring data centers to run single models.

Can you explain what you mean by "minature sizes"? It would also be good for ecosystem users to show benchmarks for the size / performance claims.

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

The best way to describe "miniature sizes" in this context would be as ultra-compact models. So far, we have trained two basic quantum-enhanced language models (QELMs) using quantum-inspired parameters (working on the third). For testing purposes, we utilized CPU threading with 12 threads, ran 2 epochs, and used a learning rate of 0.05. Each training session averaged about 8 hours. Whereas we can leverage gpu, we are utilizing cpu to sort of "bottleneck" it.

Currently, the models produce nonsensical outputs because they were trained on small, basic datasets and have not yet been properly tokenized or configured for inference. Our primary goal was to test the training process and ensure it functioned as expected. We've recently started training with a real dataset based on Global Health Statistics, and our next step is to move on to the RedPajama 1T dataset for more comprehensive testing.

When comparing model sizes, we observed significant differences between a traditional LLM and our QELM. For example, training a small, fabricated dataset with a traditional LLM yielded a model size of approximately 125 MB, whereas the same dataset trained on QELM resulted in a model size of just 238 KB. Despite this compact size, both models shared similar training times and produced equally nonsensical outputs, highlighting the early stage of development for QELM. However the Qelm was able to respond instantly without any cpu throttling, whereas the llm did slightly.

This project is still in its infancy, and as it’s being handled primarily by myself with assistance from a friend, we haven’t reached the bench-marking phase yet. However, we hope to conduct performance benchmarks soon. We are also working on creating a dedicated website for the project.

Additionally, the 135-character input limitation issue has been resolved. If you need further details or have any questions, feel free to reach out. Thank you!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
submission Project submission
Projects
None yet
Development

Successfully merging this pull request may close these issues.

[Submission]: Qelm
5 participants