Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

docs: Add page on decentralized AI inference #3407

Open
wants to merge 7 commits into
base: master
Choose a base branch
from

Conversation

ielashi
Copy link
Contributor

@ielashi ielashi commented Aug 29, 2024

Most of the content is taken from the overview page and moved/reorganized into an inference page that we can evolve over time.

I moved the examples into the inference tab in the sidebar as well, as they are inference examples.

Most of the content is taken from the overview page and moved/reorganized
into an inference page that we can evolve over time.

I moved the examples into the inference tab in the sidebar as well, as
they are inference examples.
@ielashi ielashi requested a review from a team as a code owner August 29, 2024 15:20
Copy link

github-actions bot commented Aug 29, 2024

🤖 Here's your preview: https://fkpod-biaaa-aaaam-abtcq-cai.icp0.io

docs/developer-docs/ai/inference.mdx Outdated Show resolved Hide resolved
docs/developer-docs/ai/inference.mdx Outdated Show resolved Hide resolved
docs/developer-docs/ai/inference.mdx Outdated Show resolved Hide resolved
docs/developer-docs/ai/inference.mdx Outdated Show resolved Hide resolved
docs/developer-docs/ai/inference.mdx Outdated Show resolved Hide resolved
docs/developer-docs/ai/inference.mdx Outdated Show resolved Hide resolved
docs/developer-docs/ai/overview.mdx Outdated Show resolved Hide resolved
docs/developer-docs/ai/inference.mdx Outdated Show resolved Hide resolved
docs/developer-docs/ai/inference.mdx Outdated Show resolved Hide resolved
docs/developer-docs/ai/inference.mdx Outdated Show resolved Hide resolved
@ielashi
Copy link
Contributor Author

ielashi commented Sep 11, 2024

@jessiemongeon1 I should have addressed all your points. PTAL.


<MarkdownChipRow labels={["Intermediate", "Concept", "DeAI" ]} />

It's possible for canister smart contracts to run inference in a number of ways, depending on the decentralization and performance requirements.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
It's possible for canister smart contracts to run inference in a number of ways, depending on the decentralization and performance requirements.
## Overview
Inference in the context of decentralized AI refers to using a trained model to draw conclusions about new data.
It's possible for canister smart contracts to run inference in a number of ways depending on the decentralization and performance requirements.
Canisters can utilize inference run on-chain, on-device, or through HTTPS outcalls.

Comment on lines +40 to +43
An alternative to running the model on-chain would be for the user to download the model from a canister smart contract, and the inference then happens on the user's device.
If the user trusts their own device, then they can trust that the inference ran correctly.
A disadvantage here is that the model needs to be downloaded to the user's device with corresponding drawbacks of less confidentiality of the model and decreased user experience due to increased latency.
ICP supports this use case for practically all existing models because a smart contract on ICP can store models up to 400GiB.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
An alternative to running the model on-chain would be for the user to download the model from a canister smart contract, and the inference then happens on the user's device.
If the user trusts their own device, then they can trust that the inference ran correctly.
A disadvantage here is that the model needs to be downloaded to the user's device with corresponding drawbacks of less confidentiality of the model and decreased user experience due to increased latency.
ICP supports this use case for practically all existing models because a smart contract on ICP can store models up to 400GiB.
An alternative to running the model on-chain would be to download the model from a canister, then run the inference on the local device. If the user trusts their own device, then they can trust that the inference ran correctly.
A disadvantage of this workflow is that the model needs to be downloaded to the user's device, resulting in less confidentiality of the model and decreased user experience due to increased latency.
ICP supports this workflow for most existing models because a smart contract on ICP can store models up to 400GiB.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants