diff --git a/.github/workflows/build-docs-revamp.yml b/.github/workflows/build-docs-revamp.yml index 871ce662ddc07..7d63950f0ae47 100644 --- a/.github/workflows/build-docs-revamp.yml +++ b/.github/workflows/build-docs-revamp.yml @@ -31,6 +31,16 @@ jobs: vercel-token: ${{ secrets.VERCEL_TOKEN }} vercel-org-id: ${{ secrets.VERCEL_ORG_ID }} vercel-project-id: ${{ secrets.VERCEL_DOCS_NEXT_PROJECT_ID }} + github-token: ${{ secrets.GITHUB_TOKEN }} + scope: ${{ secrets.VERCEL_ORG_ID }} + + - name: Publish to Vercel Production + uses: amondnet/vercel-action@v25 + if: github.event_name == 'push' && github.ref == 'refs/heads/docs/revamp' + with: + vercel-token: ${{ secrets.VERCEL_TOKEN }} + vercel-org-id: ${{ secrets.VERCEL_ORG_ID }} + vercel-project-id: ${{ secrets.VERCEL_PROJECT_ID }} vercel-args: "--prod" github-token: ${{ secrets.GITHUB_TOKEN }} scope: ${{ secrets.VERCEL_ORG_ID }} diff --git a/docs/docs-next/docs/guides/data-assets/passing-data-between-assets.md b/docs/docs-next/docs/guides/data-assets/passing-data-between-assets.md index 1a733f75eae3d..509309ccafa72 100644 --- a/docs/docs-next/docs/guides/data-assets/passing-data-between-assets.md +++ b/docs/docs-next/docs/guides/data-assets/passing-data-between-assets.md @@ -96,7 +96,7 @@ Consider this example: This example downloads a zip file from Google Drive, unzips it, and loads the data into a pandas DataFrame. It relies on each asset running on the same file system to perform these operations. -The assets are modeled as tasks, rather than as data assets. For more information on the difference between tasks and data assets, check out the [Thinking in Assets](/concepts/thinking-in-assets) guide. +The assets are modeled as tasks, rather than as data assets. For more information on the difference between tasks and data assets, check out the [Thinking in Assets](/concepts/assets/thinking-in-assets) guide. In this refactor, the `download_files`, `unzip_files`, and `load_data` assets are combined into a single asset, `my_dataset`. This asset downloads the files, unzips them, and loads the data into a data warehouse.