You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Fix broken documentation URLs and add lint-ignore comments for URLs that
fail due to timeouts or authentication requirements.
Fixed 404 broken URLs:
- Updated torch.export documentation URLs from
pytorch.org/docs/stable/export.html to
docs.pytorch.org/docs/stable/user_guide/torch_compiler/export.html
- Fixed ARM Ethos-U backend URL to use new path structure
(/backends/arm-ethos-u/arm-ethos-u-overview.html)
- Fixed XNNPACK internals URL (xnnpack-internals.html →
xnnpack-arch-internals.html)
- Fixed XNNPACK delegate lowering tutorial URL
- Fixed Qualcomm backend URL
Added @lint-ignore for URLs that can't be automatically verified:
- HuggingFace gated model URLs (require authentication, return 401)
- PyTorch HUD URLs (timeout in CI environment, return 000)
Other cleanup:
- Removed outdated paragraph referencing non-existent
test_xnnpack_qnnpack.py file from delegate documentation
Test plan
- Verify link-check / lint-urls CI job passes
- Spot-check a few of the updated URLs manually
Co-authored-by: Mergen Nachin <[email protected]>
The torchaudio [RNNT-emformer](https://pytorch.org/audio/stable/tutorials/online_asr_tutorial.html) model is an Automatic Speech Recognition (ASR) model, comprised of three different submodels: an encoder, a predictor and a joiner.
140
+
The torchaudio [RNNT-emformer](https://docs.pytorch.org/audio/stable/generated/torchaudio.pipelines.EMFORMER_RNNT_BASE_LIBRISPEECH.html) model is an Automatic Speech Recognition (ASR) model, comprised of three different submodels: an encoder, a predictor and a joiner.
141
141
The [predictor](https://github.com/pytorch/executorch/blob/main/examples/cadence/models/rnnt_predictor.py) is a sequence of basic ops (embedding, ReLU, linear, layer norm) and can be exported using:
Copy file name to clipboardExpand all lines: docs/source/backends-cadence.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -197,7 +197,7 @@ In all cases the generated file is called `CadenceDemoModel.pte`.
197
197
198
198
***Speech/Audio Models***:
199
199
200
-
The torchaudio [RNNT-emformer](https://pytorch.org/audio/stable/tutorials/online_asr_tutorial.html) model is an Automatic Speech Recognition (ASR) model, comprised of three different submodels:
200
+
The torchaudio [RNNT-emformer](https://docs.pytorch.org/audio/stable/generated/torchaudio.pipelines.EMFORMER_RNNT_BASE_LIBRISPEECH.html) model is an Automatic Speech Recognition (ASR) model, comprised of three different submodels:
etc.) to prepare a PyTorch program for execution on devices.
27
27
28
28
Program preparation is often simply called AOT (ahead-of-time) because export, transformations and compilations to the program are performed before it is eventually run with the ExecuTorch runtime, written in C++. To have a lightweight runtime and small overhead in execution, we push work as much as possible to AOT.
@@ -33,14 +33,14 @@ Starting from the program source code, below are the steps you would go through
33
33
34
34
* Like all PyTorch use cases, ExecuTorch starts from model authoring, where standard `nn.Module` eager mode PyTorch programs are created.
35
35
* Export-specific helpers are used to represent advanced features like [control
(for example, data dependent dynamic shape constraint).
40
40
41
41
### Export
42
42
43
-
To deploy the program to the device, engineers need to have a graph representation for compiling a model to run on various backends. With [`torch.export()`](https://pytorch.org/docs/main/export.html), an [EXIR](ir-exir.md) (export intermediate representation) is generated with ATen dialect. All AOT compilations are based on this EXIR, but can have multiple dialects along the lowering path as detailed below.
43
+
To deploy the program to the device, engineers need to have a graph representation for compiling a model to run on various backends. With [`torch.export()`](https://docs.pytorch.org/docs/stable/user_guide/torch_compiler/export.html), an [EXIR](ir-exir.md) (export intermediate representation) is generated with ATen dialect. All AOT compilations are based on this EXIR, but can have multiple dialects along the lowering path as detailed below.
44
44
45
45
*_[ATen Dialect](ir-exir.md#aten-dialect)_. PyTorch Edge is based on PyTorch’s Tensor library ATen, which has clear contracts for efficient execution. ATen Dialect is a graph represented by ATen nodes which are fully ATen compliant. Custom operators are allowed, but must be registered with the dispatcher. It’s flatten with no module hierarchy (submodules in a bigger module), but the source code and module hierarchy are preserved in the metadata. This representation is also autograd safe.
46
46
* Optionally, _quantization_, either QAT (quantization-aware training) or PTQ (post training quantization) can be applied to the whole ATen graph before converting to Core ATen. Quantization helps with reducing the model size, which is important for edge devices.
0 commit comments