download at zoo.dev/modeling-app/download
A CAD application from the future, brought to you by the Zoo team.
Modeling App is our take on what a modern modelling experience can be. It is applying several lessons learned in the decades since most major CAD tools came into existence:
- All artifacts—including parts and assemblies—should be represented as human-readable code. At the end of the day, your CAD project should be "plain text"
- This makes version control—which is a solved problem in software engineering—trivial for CAD
- All GUI (or point-and-click) interactions should be actions performed on this code representation under the hood
- This unlocks a hybrid approach to modeling. Whether you point-and-click as you always have or you write your own KCL code, you are performing the same action in Modeling App
- Everything graphics has to be built for the GPU
- Most CAD applications have had to retrofit support for GPUs, but our geometry engine is made for GPUs (primarily Nvidia's Vulkan), getting the order of magnitude rendering performance boost with it
- Make the resource-intensive pieces of an application auto-scaling
- One of the bottlenecks of today's hardware design tools is that they all rely on the local machine's resources to do the hardest parts, which include geometry rendering and analysis. Our geometry engine parallelizes rendering and just sends video frames back to the app (seriously, inspect source, it's just a
<video>
element), and our API will offload analysis as we build it in
- One of the bottlenecks of today's hardware design tools is that they all rely on the local machine's resources to do the hardest parts, which include geometry rendering and analysis. Our geometry engine parallelizes rendering and just sends video frames back to the app (seriously, inspect source, it's just a
We are excited about what a small team of people could build in a short time with our API. We welcome you to try our API, build your own applications, or contribute to ours!
Modeling App is a hybrid user interface for CAD modeling. You can point-and-click to design parts (and soon assemblies), but everything you make is really just kcl
code under the hood. All of your CAD models can be checked into source control such as GitHub and responsibly versioned, rolled back, and more.
The 3D view in Modeling App is just a video stream from our hosted geometry engine. The app sends new modeling commands to the engine via WebSockets, which returns back video frames of the view within the engine.
- UI
- Networking
- WebSockets (via KittyCAD TS client)
- Code Editor
- CodeMirror
- Custom WASM LSP Server
- Modeling
We recommend downloading the latest application binary from our Releases page. If you don't see your platform or architecture supported there, please file an issue.
First, install Rust via rustup
. This project uses a lot of Rust compiled to WASM within it. We always use the latest stable version of Rust, so you may need to run rustup update stable
. Then, run:
yarn install
followed by:
yarn build:wasm
or if you have the gh cli installed
./get-latest-wasm-bundle.sh # this will download the latest main wasm bundle
That will build the WASM binary and put in the public
dir (though gitignored).
Finally, to run the web app only, run:
yarn start
If you're not an KittyCAD employee you won't be able to access the dev environment, you should copy everything from .env.production
to .env.development
to make it point to production instead, then when you navigate to localhost:3000
the easiest way to sign in is to paste localStorage.setItem('TOKEN_PERSIST_KEY', "your-token-from-https://zoo.dev/account/api-tokens")
replacing the with a real token from https://zoo.dev/account/api-tokens of course, then navigate to localhost:3000 again. Note that navigating to localhost:3000/signin
removes your token so you will need to set the token again.
The Copilot LSP plugin in the editor requires a Zoo API token to run. In production, we authenticate this with a token via cookie in the browser and device auth token in the desktop environment, but this token is inaccessible in the dev browser version because the cookie is considered "cross-site" (from localhost
to dev.zoo.dev
). There is an optional environment variable called VITE_KC_DEV_TOKEN
that you can populate with a dev token in a .env.development.local
file to not check it into Git, which will use that token instead of other methods for the LSP service.
Chrome is in the process of rolling out a new default which
blocks Third-Party Cookies.
If you're having trouble logging into the modeling-app
, you may need to
enable third-party cookies. You can enable third-party cookies by clicking on
the eye with a slash through it in the URL bar, and clicking on "Enable
Third-Party Cookies".
To spin up the desktop app, yarn install
and yarn build:wasm
need to have been done before hand then
yarn tron:start
This will start the application and hot-reload on changes.
Devtools can be opened with the usual Cmd-Opt-I (Mac) or Ctrl-Shift-I (Linux and Windows).
To build, run yarn tron:package
.
Which commands from setup are one off vs need to be run every time?
The following will need to be run when checking out a new commit and guarantees the build is not stale:
yarn install
yarn build:wasm
yarn start # or yarn build:local && yarn serve for slower but more production-like build
Before you submit a contribution PR to this repo, please ensure that:
- There is a corresponding issue for the changes you want to make, so that discussion of approach can be had before work begins.
- You have separated out refactoring commits from feature commits as much as possible
- You have run all of the following commands locally:
yarn fmt
yarn tsc
yarn test
- Here they are all together:
yarn fmt && yarn tsc && yarn test
It will be used to document changelog discussions and release testing.
https://github.com/KittyCAD/modeling-app/issues/new
Create a new tag and push it to the repo. The semantic-release.sh
script will automatically bump the minor part, which we use the most. For instance going from v0.27.0
to v0.28.0
.
VERSION=$(./scripts/semantic-release.sh)
git tag $VERSION
git push origin --tags
This will trigger the build-apps
workflow, set the version, build & sign the apps, and generate release files as well as updater-test artifacts.
The workflow should be listed right away in this list).
The release builds can be found under the out-{arch}-{platform}
zip files, at the very bottom of the build-apps
summary page for the workflow (triggered by the tag in 2.).
Manually test against this list across Windows, MacOS, Linux and posting results as comments in the issue.
The other build-apps
output in the release build-apps
workflow (triggered by 2.) is updater-test-{arch}-{platform}
. It's a semi-automated process: for macOS, Windows, and Linux, download the corresponding updater-test artifact file, install the app, run it, expect an updater prompt to a dummy v0.255.255, install it and check that the app comes back at that version.
The only difference with these builds is that they point to a different update location on the release bucket, with this dummy v0.255.255 always available. This helps ensuring that the version we release will be able to update to the next one available.
If the prompt doesn't show up, start the app in command line to grab the electron-updater logs. This is likely an issue with the current build that needs addressing (or the updater-test location in the storage bucket).
# Windows (PowerShell)
& 'C:\Program Files\Zoo Modeling App\Zoo Modeling App.exe'
# macOS
/Applications/Zoo\ Modeling\ App.app/Contents/MacOS/Zoo\ Modeling\ App
# Linux
./Zoo Modeling App-{version}-{arch}-linux.AppImage
Head over to https://github.com/KittyCAD/modeling-app/releases/new, pick the newly created tag and type it in the Release title field as well.
Hit Generate release notes as a starting point to discuss the changelog in the issue. Once done, make sure Set as the latest release is checked, and hit Publish release.
A new publish-apps-release
will kick in and you should be able to find it here. On success, the files will be uploaded to the public bucket as well as to the GitHub release, and the announcement on Discord will be sent.
If everything is well and the release is out to the public, the issue tracking the release shall be closed.
Make sure you install cargo fuzz:
$ cargo install cargo-fuzz
$ cd src/wasm-lib/kcl
# list the fuzz targets
$ cargo fuzz list
# run the parser fuzzer
$ cargo +nightly fuzz run parser
For more information on fuzzing you can check out this guide.
You will need a ./e2e/playwright/playwright-secrets.env
file:
$ touch ./e2e/playwright/playwright-secrets.env
$ cat ./e2e/playwright/playwright-secrets.env
token=<dev.zoo.dev/account/api-tokens>
snapshottoken=<your-snapshot-token>
For a portable way to run Playwright you'll need Docker.
After that, open a terminal and run:
docker run --network host --rm --init -it playwright/chrome:playwright-x.xx.x
and in another terminal, run:
PW_TEST_CONNECT_WS_ENDPOINT=ws://127.0.0.1:4444/ yarn playwright test --project="Google Chrome" <test suite>
open a terminal and run:
docker run --network host --rm --init -it playwright/chrome:playwright-1.46.0
and in another terminal, run:
PW_TEST_CONNECT_WS_ENDPOINT=ws://127.0.0.1:4444/ yarn playwright test --project="Google Chrome" e2e/playwright/command-bar-tests.spec.ts
run a specific test change the test from test('...
to test.only('...
(note if you commit this, the tests will instantly fail without running any of the tests)
Gotcha: running the docker container with a mismatched image against your ./node_modules/playwright
will cause a failure. Make sure the versions are matched and up to date.
run headed
yarn playwright test --headed
run with step through debugger
PWDEBUG=1 yarn playwright test
However, if you want a debugger I recommend using VSCode and the playwright
extension, as the above command is a cruder debugger that steps into every function call which is annoying.
With the extension you can set a breakpoint after waitForDefaultPlanesVisibilityChange
in order to skip app loading, then the vscode debugger's "step over" is much better for being able to stay at the right level of abstraction as you debug the code.
If you want to limit to a single browser use --project="webkit"
or firefox
, Google Chrome
Or comment out browsers in playwright.config.ts
.
note chromium has encoder compat issues which is why were testing against the branded 'Google Chrome'
You may consider using the VSCode extension, it's useful for running individual threads, but some some reason the "record a test" is locked to chromium with we can't use. A work around is to us the CI yarn playwright codegen -b wk --load-storage ./store localhost:3000
Where ./store
should look like this
./store
should look like this{
"cookies": [],
"origins": [
{
"origin": "http://localhost:3000",
"localStorage": [
{
"name": "store",
"value": "{\"state\":{\"openPanes\":[\"code\"]},\"version\":0}"
},
{
"name": "persistCode",
"value": ""
},
{
"name": "TOKEN_PERSIST_KEY",
"value": "your-token"
}
]
}
]
}
However because much of our tests involve clicking in the stream at specific locations, it's code-gen looks await page.locator('video').click();
when really we need to use a pixel coord, so I think it's of limited use.
If you already haven't, run the following:
yarn
yarn build:wasm
yarn start
and finally:
yarn test:unit
For individual testing:
yarn test abstractSyntaxTree -t "unexpected closed curly brace" --silent=false
Which will run our suite of Vitest unit and React Testing Library E2E tests, in interactive mode by default.
cd src/wasm-lib
KITTYCAD_API_TOKEN=XXX cargo test -- --test-threads=1
Where XXX
is an API token from the production engine (NOT the dev environment).
We recommend using nextest to run the Rust tests (its faster and is used in CI). Once installed, run the tests using
cd src/wasm-lib
KITTYCAD_API_TOKEN=XXX cargo run nextest
When you see the CI CD fail on jobs you may wonder three things
- Do I have a bug in my code?
- Is the test flaky?
- Is there a bug in
main
?
To answer these questions the following commands will give you confidence to locate the issue.
Part of the CI CD pipeline performs static analysis on the code. Use the following commands to mimic the CI CD jobs.
The following set of commands should get us closer to one and done commands to instantly retest issues.
yarn test-setup
Gotcha, are packages up to date and is the wasm built?
yarn tsc
yarn fmt-check
yarn lint
yarn test:unit:local
Gotcha: Our unit tests have integration tests in them. You need to run a localhost server to run the unit tests.
Playwright Electron
These E2E tests run in electron. There are tests that are skipped if they are ran in a windows, linux, or macos environment. We can use playwright tags to implement test skipping.
yarn test:playwright:electron:local
yarn test:playwright:electron:windows:local
yarn test:playwright:electron:macos:local
yarn test:playwright:electron:ubuntu:local
Why does it say local? The CI CD commands that run in the pipeline cannot be ran locally. A single command will not properly setup the testing environment locally.
The tests are broken into snapshot tests and non-snapshot tests, and they run in that order, they automatically commit new snap shots, so if you see an image commit check it was an intended change. If we have non-determinism in the snapshots such that they are always committing new images, hopefully this annoyance makes us fix them asap, if you notice this happening let Kurt know. But for the odd occasion git reset --hard HEAD~ && git push -f
is your friend.
How to interpret failing playwright tests?
If your tests fail, click through to the action and see that the tests failed on a line that includes await page.getByTestId('loading').waitFor({ state: 'detached' })
, this means the test fail because the stream never started. It's you choice if you want to re-run the test, or ignore the failure.
We run on ubuntu and macos, because safari doesn't work on linux because of the dreaded "no RTCPeerConnection variable" error. But linux runs first and then macos for the same reason that we limit the number of parallel tests to 1 because we limit stream connections per user, so tests would start failing we if let them run together.
If something fails on CI you can download the artifact, unzip it and then open playwright-report/data/<UUID>.zip
with https://trace.playwright.dev/ to see what happened.
Besides following the instructions above and using the playwright docs, our app is weird because of the whole stream thing, which means our testing is weird. Because we've just figured out this stuff and therefore docs might go stale quick here's a 15min vid/tutorial
Screen.Recording.2023-11-21.at.11.37.07.am.mp4
PS: for the debug panel, the following JSON is useful for snapping the camera
{"type":"modeling_cmd_req","cmd_id":"054e5472-e5e9-4071-92d7-1ce3bac61956","cmd":{"type":"default_camera_look_at","center":{"x":15,"y":0,"z":0},"up":{"x":0,"y":0,"z":1},"vantage":{"x":30,"y":30,"z":30}}}
For how to contribute to KCL, see our KCL README.
To display logging (to the terminal or console) set ZOO_LOG=1
. This will log some warnings and simple performance metrics. To view these in test runs, use -- --nocapture
.
To enable memory metrics, build with --features dhat-heap
.