Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Idea for generating test cases #2

Open
ethancaballero opened this issue Apr 7, 2017 · 2 comments
Open

Idea for generating test cases #2

ethancaballero opened this issue Apr 7, 2017 · 2 comments

Comments

@ethancaballero
Copy link

ethancaballero commented Apr 7, 2017

Train seq2seq RNNs to generate syntactically valid inputs to gold code, and then use gold code as oracle to get output for generated input (if the input has correct syntax).

 Inside of try/except, seq2seq receives gold code as input and has to generate an input for gold code that doesn’t set off exception.

 Reward seq2seq +1 if no exception, -1 if exception. & add entropy (or minibatch discrimination?) to loss so it’ll have varied generations.
 To learn which syntax error it made, have seq2seq predict which exception it set off and lessen negative reward if prediction is correct.

 Maybe SL pretrain with already provided test cases (if we have any) before RL stage (entropy in RL loss will ensure RL stage stays varied).

@Avmb
Copy link
Collaborator

Avmb commented Apr 8, 2017

Interesting idea, though I expect this to be a very hard problem.

@ethancaballero
Copy link
Author

ethancaballero commented Jul 10, 2017

This might be more plausible with "Hindsight Experience Replay": https://arxiv.org/abs/1707.01495

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants