Support triu function for tvm.relay.expr.Call
Inputs
#2
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
This PR is to address #3
Having a proper support for triangular upper when the inputs are a list of
CallNode
s will allow us to be a step closer to successfully implementing Qwen 1.5 (0.5B) (See tenstorrent/tt-buda-demos#20).Explanation
When compiling Qwen 1.5 (0.5B) (tenstorrent/tt-buda-demos#37), one of its OP codes is
aten::triu
with its inputs containing nested functions of OP calls.self.trilu
seems to be able to successfully handle these inputs whenmode="upper"
to do triangular upper operation.Issues
1
Op(trilu)
instead ofOp(triu)
After doing
self.trilu(inputs, input_types, mode="triu")
, the resulting output would be:Is this a genuine issue??? Or can it be ignored for now?
2. NaN tensor values for Grayskull e75
When tested on @marty1885's e75, he ran into an error where his tensor values were NaN.
But weirdly @JonathanALevine's e150 was able to successfully compile and run it until running into some errors later.
*This is a draft for now since this is just a workaround and not a proper fix yet.