You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
## Description
Add models Benchmarks (memory usage, inference time, model size)
### Type of change
- [ ] Bug fix (non-breaking change which fixes an issue)
- [ ] New feature (non-breaking change which adds functionality)
- [ ] Breaking change (fix or feature that would cause existing
functionality to not work as expected)
- [x] Documentation update (improves or adds clarity to existing
documentation)
### Checklist
- [x] I have performed a self-review of my code
- [x] I have commented my code, particularly in hard-to-understand areas
- [x] I have updated the documentation accordingly
- [x] My changes generate no new warnings
Times presented in the tables are measured as consecutive runs of the model. Initial run times may be up to 2x longer due to model loading and initialization.
8
+
:::
9
+
10
+
## Classification
11
+
12
+
| Model | iPhone 16 Pro (Core ML) [ms]| iPhone 13 Pro (Core ML) [ms]| iPhone SE 3 (Core ML) [ms]| Samsung Galaxy S24 (XNNPACK) [ms]| OnePlus 12 (XNNPACK) [ms]|
Times presented in the tables are measured as consecutive runs of the model. Initial run times may be up to 2x longer due to model loading and initialization.
108
+
:::
109
+
110
+
| Model | iPhone 16 Pro (Core ML) [ms]| iPhone 13 Pro (Core ML) [ms]| iPhone SE 3 (Core ML) [ms]| Samsung Galaxy S24 (XNNPACK) [ms]| OnePlus 12 (XNNPACK) [ms]|
Times presented in the tables are measured as consecutive runs of the model. Initial run times may be up to 2x longer due to model loading and initialization.
146
+
:::
147
+
148
+
| Model | iPhone 16 Pro (XNNPACK) [ms]| iPhone 13 Pro (XNNPACK) [ms]| iPhone SE 3 (XNNPACK) [ms]| Samsung Galaxy S24 (XNNPACK) [ms]| OnePlus 12 (XNNPACK) [ms]|
Times presented in the tables are measured as consecutive runs of the model. Initial run times may be up to 2x longer due to model loading and initialization.
106
+
:::
107
+
108
+
| Model | iPhone 16 Pro (Core ML) [ms]| iPhone 13 Pro (Core ML) [ms]| iPhone SE 3 (Core ML) [ms]| Samsung Galaxy S24 (XNNPACK) [ms]| OnePlus 12 (XNNPACK) [ms]|
Copy file name to clipboardexpand all lines: docs/docs/fundamentals/getting-started.md
+8-2
Original file line number
Diff line number
Diff line change
@@ -7,12 +7,15 @@ import Tabs from '@theme/Tabs';
7
7
import TabItem from '@theme/TabItem';
8
8
9
9
## What is ExecuTorch?
10
-
ExecuTorch is a novel AI framework developed by Meta, designed to streamline deploying PyTorch models on a variety of devices, including mobile phones and microcontrollers. This framework enables exporting models into standalone binaries, allowing them to run locally without requiring API calls. ExecuTorch achieves state-of-the-art performance through optimizations and delegates such as CoreML and XNNPack. It provides a seamless export process with robust debugging options, making it easier to resolve issues if they arise.
10
+
11
+
ExecuTorch is a novel AI framework developed by Meta, designed to streamline deploying PyTorch models on a variety of devices, including mobile phones and microcontrollers. This framework enables exporting models into standalone binaries, allowing them to run locally without requiring API calls. ExecuTorch achieves state-of-the-art performance through optimizations and delegates such as Core ML and XNNPACK. It provides a seamless export process with robust debugging options, making it easier to resolve issues if they arise.
11
12
12
13
## React Native ExecuTorch
14
+
13
15
React Native ExecuTorch is our way of bringing ExecuTorch into the React Native world. Our API is built to be simple, declarative, and efficient. Plus, we’ll provide a set of pre-exported models for common use cases, so you won’t have to worry about handling exports yourself. With just a few lines of JavaScript, you’ll be able to run AI models (even LLMs 👀) right on your device—keeping user data private and saving on cloud costs.
14
16
15
17
## Installation
18
+
16
19
Installation is pretty straightforward, just use your favorite package manager.
17
20
18
21
<Tabs>
@@ -54,12 +57,15 @@ Because we are using ExecuTorch under the hood, you won't be able to build iOS a
54
57
:::
55
58
56
59
Running the app with the library:
60
+
57
61
```bash
58
62
yarn run expo:<ios | android> -d
59
63
```
60
64
61
65
## Good reads
62
-
If you want to dive deeper into ExecuTorch or our previous work with the framework, we highly encourage you to check out the following resources:
66
+
67
+
If you want to dive deeper into ExecuTorch or our previous work with the framework, we highly encourage you to check out the following resources:
-[Native code for iOS](https://medium.com/swmansion/bringing-native-ai-to-your-mobile-apps-with-executorch-part-i-ios-f1562a4556e8?source=user_profile_page---------0-------------250189c98ccf---------------)
65
71
-[Native code for Android](https://medium.com/swmansion/bringing-native-ai-to-your-mobile-apps-with-executorch-part-ii-android-29431b6b9f7f?source=user_profile_page---------2-------------b8e3a5cb1c63---------------)
Copy file name to clipboardexpand all lines: docs/docs/llms/exporting-llama.md
+18-9
Original file line number
Diff line number
Diff line change
@@ -3,32 +3,41 @@ title: Exporting Llama
3
3
sidebar_position: 2
4
4
---
5
5
6
-
In order to make the process of export as simple as possible for you, we created a script that runs a Docker container and exports the model.
6
+
In order to make the process of export as simple as possible for you, we created a script that runs a Docker container and exports the model.
7
7
8
8
## Steps to export Llama
9
+
9
10
### 1. Create an account
10
-
Get a [HuggingFace](https://huggingface.co/) account. This will allow you to download needed files. You can also use the [official Llama website](https://www.llama.com/llama-downloads/).
11
+
12
+
Get a [HuggingFace](https://huggingface.co/) account. This will allow you to download needed files. You can also use the [official Llama website](https://www.llama.com/llama-downloads/).
11
13
12
14
### 2. Select a model
15
+
13
16
Pick the model that suits your needs. Before you download it, you'll need to accept a license. For best performance, we recommend using Spin-Quant or QLoRA versions of the model:
0 commit comments