Releases: dagster-io/dagster
1.5.7 / 0.21.7 (libraries)
New
- The
OpExecutionContext
andAssetExecutionContext
now have apartition_keys
property - [dagster-ui] The asset graph layout algorithm has been changed to a much faster one called “tight-tree”
- [dagster-ui] The Runs table filters has a top level filter for partitions
- [dagster-dbt]
dbt-core==1.7.*
is now supported.
Bugfixes
- Fixed an issue where some schedules skipped a tick on the day after a fall Daylight Savings Time transition.
- Fixed a bug that caused backfill policies that execute multiple partitions in a single run not to work with dynamic partitions.
- Fixed a bug that caused an error when
build_schedule_from_partitioned_job
was used with a job with multi-partitioned assets and thepartitions_def
argument wasn’t provided todefine_asset_job
. - We now raise an error early if the empty string is provided as an asset’s group name (Thanks Sierrra!)
- Fixed an issue where custom setup and teardown methods were not properly called on nested Pythonic resources.
- Added a warning message when op or asset config is passed as an argument not named
config
. - [dagster-cloud] Fixed an issue where overriding the default I/O manager could break the Snowflake-dbt insights job.
- [auto-materialize] Fixed an issue where materializing an unpartitioned parent of a dynamic-partitioned asset would only result in the latest dynamic partition of that asset being requested. Now, all partitions will be requested.
- [dagster-embedded-elt] Fixed an issue in
dagster-embedded-elt
where sling’supdated_at
parameter was set to the incorrect type - [dagster-ui] Fixed an issue in the launchpad where selecting a partition wouldn’t correctly overwrite fields using the partition’s specific configuration
Community Contributions
- A docs fix to the testing concepts page, thanks @NicolaiLolansen!
- The schema can now be overridden per asset in DB IO managers, thanks @jrstats!
Experimental
- Improved failure recovery and retry behavior when the daemon that launches auto-materialization runs fails or crashes in the middle of a tick.
- [asset-checks] UI performance for displaying check results is improved
- [asset-checks] Removed noisy experimental warning about
AssetCheckKey
- [op-concurrency] Fixed a bug where concurrency slots were not getting assigned if a run that was assigned a slot was deleted before the slot had actually been claimed during execution.
- [dagster-pipes] The
PipesSubprocessClient
now inherits the environment variables of the parent process in the launched subprocess. - [dagster-pipes] Exceptions are now reported in the event log for framework components and from the external process.
Documentation
- Added a guide for using Dagster Pipes with Databricks
1.5.6 / 0.21.6 (libraries)
New
- [dagster-k8s] The
PipesK8sClient
will now attempt to load the appropriate kubernetes config, and exposes arguments for controlling this process. - [ui] The launch asset backfill modal now offers a preview dialog showing the targeted asset partitions and their backfill policies when partition mapping or varying backfill policies are present.
- [asset-checks] New
load_asset_checks_from_modules
functions for loading asset checks in tandem withload_assets_from_modules
. - Previously, the daemon process would terminate with an error if it believed that a thread might be hanging, which sometimes caused undesirable terminations when doing large backfills or auto-materializing many assets. Now, the daemon process will emit a warning instead of terminate.
- [dagster-dbt]
dagster-dbt project scaffold
now uses~/.dbt/profiles.yml
if aprofiles.yml
is not present in the dbt project directory. - [dagster-dbt]
@dbt_assets
now supportPartitionMapping
usingDagsterDbtTranslator.get_partition_mapping
. - [dagster-dbt] Self dependencies can now be enabled for dbt models that are represented by
@dbt_assets
. To enable this, add the following metadata to your dbt model’s metadata in your dbt project:
meta:
dagster:
has_self_dependency: True
Bugfixes
- Fixed an issue where Dagster imports would throw errors when using
pydantic<2.0.0
but havingpydantic-core
installed. - Previously, asset backfills that targeted partitioned assets with a
LastPartitionMapping
dependency would raise an error. This has been fixed. - Fixed a multipartitions partition mapping bug where a
instance is not available to load partitions
error was raised. - [asset-checks] Fixed an issue with conflicting op names when using
build_asset_with_blocking_check
- [ui] Viewing run logs containing very large messages no longer causes the UI to crash in Safari on macOS
- [ui] Drilling into the graph of a graph-backed asset with 4+ inputs or outputs no longer causes the asset graph to crash with a rendering error.
- [ui] On the backfill details page, clicking to a specific asset correctly shows the partitions that were materialized for that asset when partition mapping is in use.
- [ui] The Asset > Partition page now loads partition details more quickly in cases where calculating the staleness of the partition took a significant amount of time.
- Fixed a bug introduced in
1.5.0
where instances that haven’t been migrated to the latest schema hit an error upon run deletion. - [auto-materialize] Previously, if an asset partition was skipped on a tick for one reason, and then processed and skipped on a later tick for an additional reason, only the most recent skip reason would be tracked. Now, all skip reasons are tracked.
- [dagster-dbt] Fixed an issue where if an
exclude
that didn’t match any dbt nodes was used in@dbt_assets
, an error would be raised. The error is now properly handled. - [dagster-dbt] When invoking
DbtCliResource.cli(...)
in anop
,AssetMaterialization
's instead ofOutput
are now emitted.
Experimental
- Global op concurrency slots are now released in between retries of op execution failures.
Documentation
- Updated the tutorial to reflect best practices and APIs as of Dagster 1.5
Dagster Cloud
- The
report_asset_observation
REST endpoint for reporting runless events is now available.
1.5.5 / 0.21.5 (libraries)
New
- Dagster now supports using Pydantic 2 models for Config and Resources. Pydantic 1.10 continues to be supported.
- Added a
report_asset_observation
REST API endpoint for runless external asset observation events - Dramatically improved the performance of partition-mapping, for basic hourly and daily partitions definitions
- [ui] When viewing a list of runs, you can quickly add the tag in the “Launched by” column as a filter on the list view. Hover over the tag to see the “Add to filter” button.
- [helm] The env vars
DAGSTER_K8S_PG_PASSWORD_SECRET
andDAGSTER_K8S_INSTANCE_CONFIG_MAP
will no longer be set in all pods. - [dagster-pyspark]
build_pyspark_zip
now takes anexclude
parameter that can be used to customize the set of excluded files. - [ui] Links beginning with http://, https:// in unstructured run logs (via context.log) are automatically converted to clickable links
Bugfixes
- Fixed an asset backfill bug where certain asset partitions with no targeted parents would hang indefinitely.
- Fixed a bug where the
source_key_prefix
argument toload_assets_from_current_module
andload_assets_from_package_name
was ignored - Fixed two bugs in
dagster_embedded_elt
where the mode parameter was not being passed to Sling, and only one asset could be created at a time - Fixed a bug with handing default values for Pydantic validators on Windows
- [ui] Clicking an asset with checks shows them in the asset graph sidebar, even if live data on the page is still loading.
- [ui] Reported materialization events are presented more clearly in the asset graph sidebar and in other parts of the Dagster UI.
Deprecations
- [helm] The
pipelineRun
configuration in the Helm chart is now deprecated. The same config can be set underdagster-user-deployments
Community Contributions
- Added
setup_for_execution
andteardown_after_execution
calls to the inner IOManagers of theBranchingIOManager
- thank you @danielgafni! - The
S3FakeResource.upload_fileobj()
signature is now consistent withboto3 S3.Client.upload_fileobj()
- thank you @jeanineharb! dbt_assets
now have an optional name parameter - thank you @AlexanderVR!
Documentation
- Added a link to Dagster University to the docs landing page 🎓
- Improved readability of API docs landing page
- Removed straggling mention of Dagit from the Kubernetes OSS deployment guide
1.5.4 / 0.21.4 (libraries)
New
- Added a
report_asset_check
REST API endpoint for runless external asset check evaluation events. This is available in cloud as well. - The
config
argument is now supported on@graph_multi_asset
- [ui] Improved performance for global search UI, especially for deployments with very large numbers of jobs or assets.
- [dagster-pipes] Add S3 context injector/reader.
- [dagster-dbt] When an exception when running a dbt command, error messages from the underlying dbt invocation are now properly surfaced to the Dagster exception.
- [dagster-dbt] The path to the dbt executable is now configurable in
DbtCliResource
.
Bugfixes
- Fixed a bug introduced in 1.5.3 that caused errors when launching specific Ops in a Job.
- Fixed a bug introduced in 1.5.0 that prevented the
AssetExecutionContext
type annotation for thecontext
parameter in@asset_check
functions. - Fixed an issue where the Dagster scheduler would sometimes fail to retry a tick if there was an error reloading a code location in the middle of the tick.
- [dagster-dbt] Fixed an issue where explicitly passing in
profiles_dir=None
intoDbtCliResource
would cause incorrect validation. - [dagster-dbt] Fixed an issue where partial parsing was not working when reusing existing target paths in subsequent dbt invocations.
- [ui] Fixed an issue where the job partitions UI would show “0 total partitions” if the job consisted of more than 100 assets
Community Contributions
- [dagster-duckdb] The
DuckDBResource
andDuckDBIOManager
accept aconnection_config
configuration that will be passed asconfig
to the DuckDB connection. Thanks @xjhc!
Experimental
- Added events in the run log when a step is blocked by a global op concurrency limit.
- Added a backoff for steps querying for open concurrency slots.
- Auto-materialize logic to skip materializing when (1) a backfill is in progress or (2) parent partitions are required but nonexistent are now refactored to be skip rules.
- [ui] Added 2 new asset graph layout algorithms under user settings that are significantly faster for large graphs (1000+ assets).
Documentation
- Added several pieces of documentation for Dagster Pipes, including:
- A high-level explanation of Pipes
- A tutorial that demonstrates how to use Pipes with a local subprocess
- A reference for using a local subprocess with Pipes
- A detailed explanation of Pipes, including how to customize the process
- API references for Pipes (orchestration-side) and dagster-pipes (external process)
- Added documentation for the new experimental External Assets feature
Dagster Cloud
- Running multiple agents is no longer considered experimental.
- When the agent spins up a new code server while updating a code location, it will now wait until the new code location uploads any changes to Dagster Cloud before allowing the new server to serve requests.
1.5.3 / 0.21.3 (libraries)
New
- Alert policies can now be set on assets + asset checks (currently experimental). Check out the alerting docs for more information.
- Added a new flag
--live-data-poll-rate
that allows configuring how often the UI polls for new asset data when viewing the asset graph, asset catalog, or overview assets page. It defaults to 2000 ms. - Added back the ability to materialize changed and missing assets from the global asset-graph. A dialog will open allowing you to preview and select which assets to materialize.
- Added an experimental AMP Timeline page to give more visibility into the automaterialization daemon. You can enable it under user settings
- Added a
report_asset_materialization
REST API endpoint for creating external asset materialization events. This is available in cloud as well. - [dbt] The
@dbt_assets
decorator now accepts abackfill_policy
argument, for controlling how the assets are backfilled. - [dbt] The
@dbt_assets
decorator now accepts aop_tags
argument, for passing tags to the op underlying the producedAssetsDefinition
. - [pipes] Added
get_materialize_result
&get_asset_check_result
toPipesClientCompletedInvocation
- [dagster-datahub] The
acryl-datahub
pin in thedagster-datahub
package has been removed. - [dagster-databricks] The
PipesDatabricksClient
now performs stdout/stderr forwarding from the Databricks master node to Dagster. - [dagster-dbt] The hostname of the dbt API can now be configured when executing the
dagster-dbt-cloud
CLI. - [dagster-k8s] Added the ability to customize how raw k8s config tags set on an individual Dagster job are merged with raw k8s config set on the
K8sRunLauncher
. See the docs for more information.
Bugfixes
-
Previously, the asset backfill page would display negative counts if failed partitions were manually re-executed. This has been fixed.
-
Fixed an issue where the run list dialog for viewing the runs occupying global op concurrency slots did not expand to fit the content size.
-
Fixed an issue where selecting a partition would clear the launchpad and typing in the launchpad would clear the partition selection
-
Fixed various issues with the asset-graph displaying the wrong graph
-
The IO manager’s
handle_output
method is no longer invoked when observing an observable source asset. -
[ui] Fixed an issue where the run config dialog could not be scrolled.
-
[pipes] Fixed an issue in the
PipesDockerClient
with parsing logs fetched via the docker client. -
[external assets] Fixed an issue in
external_assets_from_specs
where providing multiple specs would error -
[external assets] Correct copy in tooltip to explain why Materialize button is disabled on an external asset.
Breaking Changes
- [pipes] A change has been made to the environment variables used to detect if the external process has been launched with pipes. Update the
dagster-pipes
version used in the external process. - [pipes] The top level function
is_dagster_pipes_process
has been removed from thedagster-pipes
package.
Community Contributions
- Override a method in the azure data lake IO manager (thanks @0xfabioo)!
- Add support of external launch types in ECS run launcher (thanks @cuttius)!
Experimental
- The Python GraphQL client is considered stable and is no longer marked as experimental.
1.5.2 / 0.21.2 (libraries)
Bugfixes
- Previously, asset backfills targeting assets with multi-run backfill policies would raise a "did not submit all run requests" error. This has been fixed.
Dagster Cloud
- The experimental dagster-insights package has receieved some API surface area updates and bugfixes.
1.5.1 / 0.21.1 (libraries)
New
- Dagster now automatically infers a dependency relationship between a time-partitioned asset and a multi-partitioned asset with a time dimension. Previously, this was only inferred when the time dimension was the same in each asset.
- The
EnvVar
utility will now raise an exception if it is used outside of the context of a Dagster resource or config class. Theget_value()
utility will retrieve the value outside of this context. - [ui] The runs page now displays a “terminate all” button at the top, to bulk terminate in-progress runs.
- [ui] Asset Graph - Various performance improvements that make navigating large asset graphs smooth
- [ui] Asset Graph - The graph now only fetches data for assets within the viewport solving timeout issues with large asset graphs
- [ui] Asset Graph Sidebar - The sidebar now shows asset status
- [dagster-dbt] When executing dbt invocations using
DbtCliResource
, an explicittarget_path
can now be specified. - [dagster-dbt] Asset checks can now be enabled by using
DagsterDbtTranslator
andDagsterDbtTranslatorSettings
: see the docs for more information. - [dagster-embedded-elt] Dagster library for embedded ELT
Bugfixes
- [ui] Fixed various issues on the asset details page where partition names would overflow outside their containers
- [ui] Backfill notification - Fixed an issue where the backfill link didn’t take the —path-prefix option into account
- [ui] Fixed an issue where the instance configuration yaml would persist rendering even after navigating away from the page.
- [ui] Fixed issues where config yaml displays could not be scrolled.
- [dagster-webserver] Fixed a performance issue that caused the UI to load slowly
Deprecations
- [dagster-dbt] Enabling asset checks using dbt project metadata has been deprecated.
1.5.0 (core) / 0.21.0 (libraries) "How Will I Know"
Major Changes since 1.4.0 (core) / 0.20.0 (libraries)
Core
-
Improved ergonomics for execution dependencies in assets - We introduced a set of APIs to simplify working with Dagster that don't use the I/O manager system for handling data between assets. I/O manager workflows will not be affected.
AssetDep
type allows you to specify upstream dependencies with partition mappings when using thedeps
parameter of@asset
andAssetSpec
.MaterializeResult
can be optionally returned from an asset to report metadata about the asset when the asset handles any storage requirements within the function body and does not use an I/O manager.AssetSpec
has been added as a new way to declare the assets produced by@multi_asset
. When using , the does not need to return any values to be stored by the I/O manager. Instead, the should handle any storage requirements in the body of the function.
-
Asset checks (experimental) - You can now define, execute, and monitor data quality checks in Dagster [docs].
- The
@asset_check
decorator, as well as thecheck_specs
argument to@asset
and@multi_asset
enable defining asset checks. - Materializing assets from the UI will default to executing their asset checks. You can also execute individual checks.
- When viewing an asset in the asset graph or the asset details page, you can see whether its checks have passed, failed, or haven’t run successfully.
- The
-
Auto materialize customization (experimental) -
AutoMaterializePolicies
can now be customized [docs].- All policies are composed of a set of
AutoMaterializeRule
s which determine if an asset should be materialized or skipped. - To modify the default behavior, rules can be added to or removed from a policy to change the conditions under which assets will be materialized.
- All policies are composed of a set of
dagster-pipes
- Dagster pipes is a new library that implements a protocol for launching compute into external execution environments and consuming streaming logs and Dagster metadata from those environments. See #16319 for more details on the motivation and vision behind Pipes.
- Out-the-box integrations
- Clients: local subprocess, Docker containers, Kubernetes, and Databricks
PipesSubprocessClient
,PipesDocketClient
,PipesK8sClient
,PipesDatabricksClient
- Transport: Unix pipes, Filesystem, s3, dbfs
- Languages: Python
- Clients: local subprocess, Docker containers, Kubernetes, and Databricks
- Dagster pipes is composable with existing launching infrastructure via
open_pipes_session
. One can augment existing invocations rather than replacing them wholesale.
Since 1.4.17 (core) / 0.20.17 (libraries)
New
- [ui] Global Asset Graph performance improvement - the first time you load the graph it will be cached to disk and any subsequent load of the graph should load instantly.
Bugfixes
- Fixed a bug where deleted runs could retain instance-wide op concurrency slots.
Breaking Changes
AssetExecutionContext
is now a subclass ofOpExecutionContext
, not a type alias. The code
def my_helper_function(context: AssetExecutionContext):
...
@op
def my_op(context: OpExecutionContext):
my_helper_function(context)
will cause type checking errors. To migrate, update type hints to respect the new subclassing.
AssetExecutionContext
cannot be used as the type annotation for@op
s run in@jobs
. To migrate, update the type hint in@op
toOpExecutionContext
.@op
s that are used in@graph_assets
may still use theAssetExecutionContext
type hint.
# old
@op
def my_op(context: AssetExecutionContext):
...
# correct
@op
def my_op(context: OpExecutionContext):
...
- [ui] We have removed the option to launch an asset backfill as a single run. To achieve this behavior, add
backfill_policy=BackfillPolicy.single_run()
to your assets.
Community Contributions
has_dynamic_partition
implementation has been optimized. Thanks @edvardlindelof!- [dagster-airbyte] Added an optional
stream_to_asset_map
argument tobuild_airbyte_assets
to support the Airbyte prefix setting with special characters. Thanks @chollinger93! - [dagster-k8s] Moved “labels” to a lower precedence. Thanks @jrouly!
- [dagster-k8s] Improved handling of failed jobs. Thanks @Milias!
- [dagster-databricks] Fixed an issue where
DatabricksPysparkStepLauncher
fails to get logs whenjob_run
doesn’t havecluster_id
at root level. Thanks @PadenZach! - Docs type fix from @sethusabarish, thank you!
Documentation
- Our Partitions documentation has gotten a facelift! We’ve split the original page into several smaller pages, as follows:
- Partitions - An overview of what a partition is, benefits, and how to use it
- Partitioning assets - Details about partitioning assets
- Partitioning ops - Details about partitioning ops
- Testing partitions - As described
Dagster Cloud
- New dagster-insights sub-module - We have released an experimental
dagster_cloud.dagster_insights
module that contains utilities for capturing and submitting external metrics about data operations to Dagster Cloud via an api. Dagster Cloud Insights is a soon-to-be released feature that shows improves visibility into usage and cost metrics such as run duration and Snowflake credits in the Cloud UI.
1.4.17 / 0.20.17 (libraries)
New
- [dagster-dbt]
DbtCliResource
now enforces that the current installed version ofdbt-core
is at least version1.4.0
. - [dagster-dbt]
DbtCliResource
now properly respectsDBT_TARGET_PATH
if it is set by the user. Artifacts from dbt invocations usingDbtCliResource
will now be placed in unique subdirectories ofDBT_TARGET_PATH
.
Bugfixes
- When executing a backfill that targets a range of time partitions in a single run, the
partition_time_window
attribute onOpExecutionContext
andAssetExecutionContext
now returns the time range, instead of raising an error. - Fixed an issue where the asset backfill page raised a GraphQL error for backfills that targeted different partitions per-asset.
- Fixed
job_name
property on the result object ofbuild_hook_context
.
Experimental
AssetSpec
has been added as a new way to declare the assets produced by@multi_asset
.AssetDep
type allows you to specify upstream dependencies with partition mappings when using thedeps
parameter of@asset
andAssetSpec
.- [dagster-ext]
report_asset_check
method added toExtContext
. - [dagster-ext] ext clients now must use
yield from
to forward reported materializations and asset check results to Dagster. Results reported from ext that are not yielded will raise an error.
Documentation
- The Dagster UI documentation got an overhaul! We’ve updated all our screenshots and added a number of previously undocumented pages/features, including:
- The Overview page, aka the Factory Floor
- Job run compute logs
- Global asset lineage
- Overview > Resources
- The Resources documentation has been updated to include additional context about using resources, as well as when to use
os.getenv()
versus Dagster’sEnvVar
. - Information about custom loggers has been moved from the Loggers documentation to its own page, Custom loggers.
1.4.16 / 0.20.16 (libraries)
New
- [ui] When using the search input within Overview pages, if the viewer’s code locations have not yet fully loaded into the app, a loading spinner will now appear to indicate that search results are pending.
Bugfixes
- Fixed an asset backfill bug that caused occasionally caused duplicate runs to be kicked off in response to manual runs upstream.
- Fixed an issue where launching a run from the Launchpad that included many assets would sometimes raise an exception when trying to create the tags for the run.
- [ui] Fixed a bug where clicking to view a job from a run could lead to an empty page in situations where the viewer’s code locations had not yet loaded in the app.
Deprecations
- Deprecated
ExpectationResult
. This will be made irrelevant by upcoming data quality features.
Community Contributions
- Enabled chunked backfill runs to target more than one asset, thanks @ruizh22!
Experimental
- Users can now emit arbitrary asset materializations, observations, and asset check evaluations from sensors via
SensorResult
.