Skip to content

Special optimization for torch backend #21313

Open
@pass-lin

Description

@pass-lin

PyTorch provides specialized cudnn operators for many common implementations, such as F.group_norm

In the past speed test benchmarks of the keras-team, we can also find that the torch backend of keras is usually slower than the native torch backend. This may be related to the fact that many implementations in keras do not use the cudnn operators provided by torch.

Is it necessary to make some special optimizations for the torch backend for the existing layer implementation to improve the inference performance of the torch backend.

If necessary, we can submit a PR.

Metadata

Metadata

Labels

stat:contributions welcomeA pull request to fix this issue would be welcome.type:featureThe user is asking for a new feature.

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions