Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Guarding against LLMs that would learn our repo "by heart" #119

Open
tischi opened this issue Sep 7, 2024 · 1 comment
Open

Guarding against LLMs that would learn our repo "by heart" #119

tischi opened this issue Sep 7, 2024 · 1 comment

Comments

@tischi
Copy link
Collaborator

tischi commented Sep 7, 2024

Even though I think LLMs generally do not work like this, I still wonder whether we could guard against some - otherwise super dumb - LLM to just learn our repo by heart and then achieve great results.

Given the discussions in #118 I wonder whether we could somehow maintain a separate secret branch where we ask the conceptually same questions but just with a slightly modifications?

Maybe:

  • changing the english in the prompt a bit
  • changing the actual values in the input data and the corresponding assertions
  • changing the order of the input and output arguments

It would be a bit of work...but maybe worth it?

What do you think?

@haesleinhuepf
Copy link
Owner

I'm currently working on training such an LLM, because I wanted to know how to achieve this:

and then achieve great results.

When I'm done, I'll share it (+ training scripts) and we can develop a strategy against it.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants