- Fixed tick display in the sensor/schedule timeline view in Dagit.
- Changed the
dagster sensor list and dagster schedule list CLI commands to include schedules and sensors that have never been turned on. - Fixed the backfill progress stats in Dagit which incorrectly capped the number of successful/failed runs.
- Improved query performance in Dagit on pipeline (or job) views, schedule views, and schedules list view by loading partition set data on demand instead of by default.
- Fixed an issue in Dagit where re-executing a pipeline that shares an identical name and graph to a pipeline in another repository could lead to the wrong pipeline being executed.
- Fixed an issue in Dagit where loading a very large DAG in the pipeline overview could sometimes lead to a render loop that repeated the same GraphQL query every few seconds, causing an endless loading state and never rendering the DAG.
- Fixed an issue with
execute_in_process where providing default executor config to a job would cause config errors. - Fixed an issue with default config for jobs where using an
ops config entry in place of solids would cause a config error. - Dynamic outputs are now properly supported while using
adls2_io_manager ModeDefinition now validates the keys of resource_defs at definition time.Failure exceptions no longer bypass the RetryPolicy if one is set.
- Added
serviceAccount.name to the user deployment Helm subchart and schema, thanks @jrouly!
- To account for ECS’ eventual consistency model, the
EcsRunLauncher will now exponentially backoff certain requests for up to a minute while waiting for ECS to reach a consistent state. - Memoization is now available from all execution entrypoints. This means that a pipeline tagged for use with memoization can be launched from dagit, the
launch CLI, and other modes of external execution, whereas before, memoization was only available via execute_pipeline and the execute CLI. - Memoization now works with root input managers. In order to use a root input manager in a pipeline that utilizes memoization, provide a string value to the
version argument on the decorator:
from dagster import root_input_manager
@root_input_manager(version="foo")
def my_root_manager(_):
pass
- The
versioned_fs_io_manager now defaults to using the storage directory of the instance as a base directory. GraphDefinition.to_job now accepts a tags dictionary with non-string values - which will be serialized to JSON. This makes job tags work similarly to pipeline tags and solid tags.
- The guide for migrating to the experimental graph, job, and op APIs now includes an example of how to migrate a pipeline with a composite solid.
- [helm] The compute log manager now defaults to a
NoOpComputeLogManager. It did not make sense to default to the LocalComputeLogManager as pipeline runs are executed in ephemeral jobs, so logs could not be retrieved once these jobs were cleaned up. To have compute logs in a Kubernetes environment, users should configure a compute log manager that uses a cloud provider. - [helm] The K8sRunLauncher now supports environment variables to be passed in from the current container to the launched Kubernetes job.
- [examples] Added a new
dbt_pipeline to the hacker news example repo, which demonstrates how to run a dbt project within a Dagster pipeline. - Changed the default configuration of steps launched by the
k8s_job_executor to match the configuration set in the K8sRunLauncher.
- Fixed an issue where dagster gRPC servers failed to load if they did not have permissions to write to a temporary directory.
- Enabled compression and raised the message receive limit for our gRPC communication. This prevents large pipelines from causing gRPC message limit errors. This limit can now be manually overridden with the
DAGSTER_GRPC_MAX_RX_BYTES environment variable. - Fixed errors with
dagster instance migrate when the asset catalog contains wiped assets. - Fixed an issue where backfill jobs with the “Re-execute from failures” option enabled were not picking up the solid selection from the originating failed run.
- Previously, when using memoization, if every step was memoized already, you would get an error. Now, the run succeeds and runs no steps.
- [dagster-dbt] If you specify
--models, --select, or --exclude flags while configuring the dbt_cli_resource, it will no longer attempt to supply these flags to commands that don’t accept them. - [dagstermill] Fixed an issue where
yield_result wrote output value to the same file path if output names are the same for different solids.
- Added the ability to customize the TTL and backoff limit on Dagster Kubernetes jobs (thanks @Oliver-Sellwood!)
ops can now be used as a config entry in place of solids.- Fixed a GraphQL bug in ECS deployments by making the
EcsRunLauncher more resilient to ECS’ eventual consistency model.
- Fixed hyperlink display to be more visible within source code snippets.
- Added documentation for Run Status Sensor on the Sensors concept page.
- The Dagit web app now has a strict Content Security Policy.
- Introduced a new decorator
[@run_status_sensor](https://docs.dagster.io/_apidocs/schedules-sensors#dagster.run_status_sensor) which defines sensors that react to given PipelineRunStatus. - You can now specify a
solid on build_hook_context. This allows you to access the hook_context.solid parameter.
dagster’s dependency on docstring-parser has been loosened.@pipeline now pulls its description from the doc string on the decorated function if it is provided.- The sensor example generated via
dagster new-project now no longer targets a non-existent mode.
- Thanks for the docs typo fix @cvoegele!
- The “jobs” key is now supported when returning a dict from
@repository functions. GraphDefinition.to_job now supports the description argument.- Jobs with nested Graph structures no longer fail to load in dagit.
- Previously, the ECS reference deployment granted its tasks the
AmazonECS_FullAccess policy. Now, the attached roles has been more narrowly scoped to only allow the daemon and dagit tasks to interact with the ECS actions required by the EcsRunLauncher. - The EcsRunLauncher launches ECS tasks by setting a command override. Previously, if the Task Definition it was using also defined an entrypoint, it would concatenate the entrypoint and the overridden command which would cause launches to fail with
Error: Got unexpected extra arguments. Now, it ignores the entrypoint and launches succeed.
- Fixed a broken link in the sensor testing overview.
- Improved Asset catalog load times in Dagit, for Dagster instances that have fully migrated using
dagster instance migrate. - When using the
ScheduleDefinition constructor to instantiate a schedule definition, if a schedule name is not provided, the name of the schedule will now default to the pipeline name, plus “_schedule”, instead of raising an error.
- Fixed a bug where pipeline definition arguments
description and solid_retry_policy were getting dropped when using a solid_hook decorator on a pipeline definition (#4355). - Fixed an issue where the Dagit frontend wasn’t disabling certain UI elements when launched in read-only mode.
- Fixed a bug where directly invoking an async solid with type annotations would fail, if called from another async function.
- Added a guide to migrating from the existing Pipeline, Mode, Preset, and Solid APIs to the new experimental Graph, Job, and Op APIs. Check out the guide here!
- Fixes implementation issues in
@pipeline_failure_sensor that prevented them from working.