-
-
Notifications
You must be signed in to change notification settings - Fork 2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
WIP feat:Init commit for rust backend #1180
base: master
Are you sure you want to change the base?
Conversation
aacaf4e
to
afcd7bd
Compare
Signed-off-by: GitHub <[email protected]>
afcd7bd
to
ef3fe9a
Compare
cc @lu-zero |
Co-authored-by: Luca Barbato <[email protected]> Signed-off-by: Aisuko <[email protected]>
Signed-off-by: GitHub <[email protected]>
Signed-off-by: Aisuko <[email protected]>
Signed-off-by: Aisuko <[email protected]>
9b74e4e
to
61bd269
Compare
Co-authored-by: Luca Barbato <[email protected]> Signed-off-by: Aisuko <[email protected]>
Signed-off-by: Aisuko <[email protected]>
ef8a86b
to
a2bb86f
Compare
Signed-off-by: Aisuko <[email protected]>
it seems to look for libtorch and fails to find it. if you use the ndarray backend does it work? |
Will try it and give a feedback Update
|
On the M1 probably the wgpu backend is the nicest to use, but ndarray is the one that does not depend on the host system. |
Signed-off-by: Aisuko <[email protected]>
Signed-off-by: Aisuko <[email protected]>
4c7f5ca
to
fb67c91
Compare
Thanks a lot. I have made some change here. I have been migrated the code which is included
Here I hit an issue on reshaping of the Tensor. So, we can try to implement a simple one instead of getting stuck on the Llama2. |
d9f1f7d
to
da3a0d8
Compare
Signed-off-by: Aisuko <[email protected]>
cb216fa
to
ed95d9c
Compare
Signed-off-by: Aisuko <[email protected]>
ed95d9c
to
660cc49
Compare
// And now the nonlinear scale | ||
let min_log_hz = 1000.0; // beginning of log region (Hz) | ||
let min_log_mel = (min_log_hz - f_min) / f_sp; | ||
let logstep = (6.4f64).ln() / 27.0; // step size for log region |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
those constants are repeated, being always f64 you can just keep them as const
s
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
thank you, will do.
a6ff963
to
b91b79c
Compare
Signed-off-by: Aisuko <[email protected]>
b91b79c
to
d62c701
Compare
Signed-off-by: Aisuko <[email protected]>
Signed-off-by: Aisuko <[email protected]>
❌ Deploy Preview for localai failed.
|
|
||
let tensor3=tensor2.transpose(); | ||
|
||
let tensor41=tensor3.repeat(2, 2); |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
@lu-zero Here, I am going to use wgpu
backend instead of tch
. However, I the repeat function here only support 2 dimensions tensor, (Can only repeat dimension with dim=1) https://github.com/Tracel-AI/burn/blob/b86bc5876149bd73bc59cb5197fd3ee8b92509d4/burn-tensor/src/tensor/ops/tensor.rs#L222C7-L222C7.
I have been tried several solutions, like use swap_dims
and flattern
these internal function of Tensor, but here hard to say it is correct and also causes other issues. Is there a better example for this?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
asking upstream probably it is the best route (sorry for the belated reply, I got very busy and the message got lost in the mailbox)
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
No worries, thanks for your support. I will continue to work on this one after I applied PhD successfully. Currently, sooo busy. But I still want to get this PR to merged.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Once you are more free please contact me, probably a good deal of the issues will be ironed out by upstream meanwhile :)
Description
This PR relates to #939
Notes for Reviewers
Signed commits