qaihm-bot commited on
Commit
53a0ca3
1 Parent(s): 82725b9

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +5 -88
README.md CHANGED
@@ -34,10 +34,12 @@ More details on model performance across various devices, can be found
34
  - Model size: 26.7 MB
35
 
36
 
 
 
37
  | Device | Chipset | Target Runtime | Inference Time (ms) | Peak Memory Range (MB) | Precision | Primary Compute Unit | Target Model
38
  | ---|---|---|---|---|---|---|---|
39
- | Samsung Galaxy S23 Ultra (Android 13) | Snapdragon® 8 Gen 2 | TFLite | 8.341 ms | 1 - 2 MB | INT8 | NPU | [FFNet-78S-Quantized.tflite](https://huggingface.co/qualcomm/FFNet-78S-Quantized/blob/main/FFNet-78S-Quantized.tflite)
40
- | Samsung Galaxy S23 Ultra (Android 13) | Snapdragon® 8 Gen 2 | QNN Model Library | 5.979 ms | 6 - 18 MB | INT8 | NPU | [FFNet-78S-Quantized.so](https://huggingface.co/qualcomm/FFNet-78S-Quantized/blob/main/FFNet-78S-Quantized.so)
41
 
42
 
43
  ## Installation
@@ -95,93 +97,8 @@ device. This script does the following:
95
  python -m qai_hub_models.models.ffnet_78s_quantized.export
96
  ```
97
 
98
- ```
99
- Profile Job summary of FFNet-78S-Quantized
100
- --------------------------------------------------
101
- Device: Snapdragon X Elite CRD (11)
102
- Estimated Inference Time: 7.09 ms
103
- Estimated Peak Memory Range: 6.01-6.01 MB
104
- Compute Units: NPU (146) | Total (146)
105
-
106
-
107
- ```
108
- ## How does this work?
109
-
110
- This [export script](https://github.com/quic/ai-hub-models/blob/main/qai_hub_models/models/FFNet-78S-Quantized/export.py)
111
- leverages [Qualcomm® AI Hub](https://aihub.qualcomm.com/) to optimize, validate, and deploy this model
112
- on-device. Lets go through each step below in detail:
113
-
114
- Step 1: **Compile model for on-device deployment**
115
-
116
- To compile a PyTorch model for on-device deployment, we first trace the model
117
- in memory using the `jit.trace` and then call the `submit_compile_job` API.
118
-
119
- ```python
120
- import torch
121
-
122
- import qai_hub as hub
123
- from qai_hub_models.models.ffnet_78s_quantized import Model
124
-
125
- # Load the model
126
- torch_model = Model.from_pretrained()
127
- torch_model.eval()
128
-
129
- # Device
130
- device = hub.Device("Samsung Galaxy S23")
131
-
132
- # Trace model
133
- input_shape = torch_model.get_input_spec()
134
- sample_inputs = torch_model.sample_inputs()
135
 
136
- pt_model = torch.jit.trace(torch_model, [torch.tensor(data[0]) for _, data in sample_inputs.items()])
137
-
138
- # Compile model on a specific device
139
- compile_job = hub.submit_compile_job(
140
- model=pt_model,
141
- device=device,
142
- input_specs=torch_model.get_input_spec(),
143
- )
144
-
145
- # Get target model to run on-device
146
- target_model = compile_job.get_target_model()
147
-
148
- ```
149
-
150
-
151
- Step 2: **Performance profiling on cloud-hosted device**
152
-
153
- After compiling models from step 1. Models can be profiled model on-device using the
154
- `target_model`. Note that this scripts runs the model on a device automatically
155
- provisioned in the cloud. Once the job is submitted, you can navigate to a
156
- provided job URL to view a variety of on-device performance metrics.
157
- ```python
158
- profile_job = hub.submit_profile_job(
159
- model=target_model,
160
- device=device,
161
- )
162
-
163
- ```
164
-
165
- Step 3: **Verify on-device accuracy**
166
-
167
- To verify the accuracy of the model on-device, you can run on-device inference
168
- on sample input data on the same cloud hosted device.
169
- ```python
170
- input_data = torch_model.sample_inputs()
171
- inference_job = hub.submit_inference_job(
172
- model=target_model,
173
- device=device,
174
- inputs=input_data,
175
- )
176
-
177
- on_device_output = inference_job.download_output_data()
178
-
179
- ```
180
- With the output of the model, you can compute like PSNR, relative errors or
181
- spot check the output with expected output.
182
 
183
- **Note**: This on-device profiling and inference requires access to Qualcomm®
184
- AI Hub. [Sign up for access](https://myaccount.qualcomm.com/signup).
185
 
186
 
187
 
@@ -206,7 +123,7 @@ Explore all available models on [Qualcomm® AI Hub](https://aihub.qualcomm.com/)
206
  ## License
207
  - The license for the original implementation of FFNet-78S-Quantized can be found
208
  [here](https://github.com/Qualcomm-AI-research/FFNet/blob/master/LICENSE).
209
- - The license for the compiled assets for on-device deployment can be found [here]({deploy_license_url})
210
 
211
  ## References
212
  * [Simple and Efficient Architectures for Semantic Segmentation](https://arxiv.org/abs/2206.08236)
 
34
  - Model size: 26.7 MB
35
 
36
 
37
+
38
+
39
  | Device | Chipset | Target Runtime | Inference Time (ms) | Peak Memory Range (MB) | Precision | Primary Compute Unit | Target Model
40
  | ---|---|---|---|---|---|---|---|
41
+ | Samsung Galaxy S23 Ultra (Android 13) | Snapdragon® 8 Gen 2 | TFLite | 8.325 ms | 1 - 8 MB | INT8 | NPU | [FFNet-78S-Quantized.tflite](https://huggingface.co/qualcomm/FFNet-78S-Quantized/blob/main/FFNet-78S-Quantized.tflite)
42
+
43
 
44
 
45
  ## Installation
 
97
  python -m qai_hub_models.models.ffnet_78s_quantized.export
98
  ```
99
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
100
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
101
 
 
 
102
 
103
 
104
 
 
123
  ## License
124
  - The license for the original implementation of FFNet-78S-Quantized can be found
125
  [here](https://github.com/Qualcomm-AI-research/FFNet/blob/master/LICENSE).
126
+ - The license for the compiled assets for on-device deployment can be found [here](https://qaihub-public-assets.s3.us-west-2.amazonaws.com/qai-hub-models/Qualcomm+AI+Hub+Proprietary+License.pdf)
127
 
128
  ## References
129
  * [Simple and Efficient Architectures for Semantic Segmentation](https://arxiv.org/abs/2206.08236)