Command-Line Help for sdf

This document contains the help content for the sdf command-line program.

Command Overview:

sdf

SDF: A fast SQL compiler, local development framework, and in-memory analytical database

Usage: sdf [OPTIONS] COMMAND

Subcommands:
  • new — Create a new sdf workspace
  • clean — Remove artifacts that sdf has generated in the past
  • compile — Compile models
  • run — Run models
  • test — Test your models
  • stats — Statistics for your data
  • report — Report code quality
  • check — Check code quality
  • lineage — Display lineage for a given table and/or column
  • push — Push a local workspace to the SDF Service
  • system — System maintenance, install and update
  • auth — Authenticate CLI to services like SDF, AWS, OpenAI, etc
  • man — Display reference material, like the CLI, dialect specific functions, schemas for authoring and interchange
  • init — Initialize a workspace interactively
  • dbt — Initialize an sdf workspace from an existing dbt project
  • exec — Execute custom scripts
Options:
  • --log-level LOG_LEVEL — Set log level

    Possible values: trace, debug, debug-pretty, info, warn, error

  • --log-file LOG_FILE — Creates or replaces the log file

  • --show-all-errors — Don’t suppress errors

sdf new

Create a new sdf workspace

Usage: sdf new [OPTIONS] [PATH]

Arguments:
  • PATH — Create a new sdf workspace at path
Options:
  • --list-samples — List all available samples

    Default value: false

  • --sample SAMPLE — Create a workspace with the sample content at path

  • -s, --show SHOW — Display progress messages

    Default value: progress

    Possible values: progress, none

sdf clean

Remove artifacts that sdf has generated in the past

Usage: sdf clean [OPTIONS]

Options:
  • -e, --environment ENVIRONMENT

  • --path PATH — Remove artifacts at path that sdf has created in the past [default: current workspace directory]

  • -s, --show SHOW — Display progress messages

    Default value: progress

    Possible values: progress, none

sdf compile

Compile models

Usage: sdf compile [OPTIONS] [TARGETS]...

Arguments:
  • TARGETS — Compile only the given source dirs, files or tables [default: all models]
Options:
  • --infer-tables — Infer table schema and generate ddls

    Default value: false

  • -e, --environment ENVIRONMENT — Use this environment

  • -s, --show SHOW — Display messages [default: progress if TARGETS is empty, all otherwise]

    Possible values: all, progress, result, none

  • -q, --query QUERY — Supply a .sql file or provide a sql snippet on the cmd line, e.g. ‘select * from t’

  • --stage STAGE — Run the following stages [default: all stages]

    Possible values: preprocess, parse, lint, resolve, classify, execute

  • --cache CACHE — Controls cache use

    Default value: read-write

    Possible values: read-write, write-only, read-only, none

  • --save SAVE — Controls which assets to save [default: none]

    Possible values: info-schema, assembly, table-deps

  • --targets-only — Processes only specified targets assuming that all the non-target dependencies already exist

    Default value: false

  • --vars VARS — Supply var bindings as a yml file or provide them as string e.g. ‘(key: value)‘

  • --env-vars ENV_VARS — Supply env var bindings as a yml file or provide them as string e.g. ‘(key: value)‘

  • --downstream — Execute cmd not only the given targets but also for all its downstream artifacts

    Default value: false

  • --format FORMAT — Show error tables in this format

    Default value: table

    Possible values: table, csv, tsv, json, nd-json, yml

  • --limit LIMIT — Limiting number of shown rows. Run with —limit 0 to remove limit

  • --prefer-local — When set will minimize remote DB accesses and will rely on local schema information in .sdf.yml files

    Default value: false

  • --no-incremental-mode — Set incremental-mode to false (default is true)

    Default value: false

  • --no-snapshot-mode — Set snapshot-mode to false (default is true)

    Default value: false

  • --warn-non-local-executable

    Default value: false

  • --fix — Fix the linting issues

    Default value: false

  • -w, --warnings WARNINGS — Specify the warnings to apply. Example: -w warning1 -w warning2

Available warnings:

  • all: Turn on all warnings.
  • none: Turn off all warnings.
  • error: Treat warnings as errors.
  • capitalization-columns=upper|lower|pascal|snake|camel|consistent: Set capitalization style for columns.
  • capitalization-tables=upper|lower|pascal|snake|camel|consistent: Set capitalization style for tables.
  • capitalization-aliases=upper|lower|pascal|snake|camel|consistent: Set capitalization style for aliases.
  • type-implicit-conversions: Warn about implicit type conversions.
  • references-consistent: Warn about consistent references.

sdf run

Run models

Usage: sdf run [OPTIONS] [TARGETS]...

Arguments:
  • TARGETS — Run only the given source dirs, files or tables [default: all models]
Options:
  • -e, --environment ENVIRONMENT — Use this environment

  • -s, --show SHOW — Display messages [default: progress if TARGETS is empty, all otherwise]

    Possible values: all, progress, result, none

  • -q, --query QUERY — Supply a .sql file or provide a sql snippet on the cmd line, e.g. ‘select * from t’

  • --stage STAGE — Run the following stages [default: all stages]

    Possible values: preprocess, parse, lint, resolve, classify, execute

  • --cache CACHE — Controls cache use

    Default value: read-write

    Possible values: read-write, write-only, read-only, none

  • --save SAVE — Controls which assets to save [default: none]

    Possible values: info-schema, assembly, table-deps

  • --targets-only — Processes only specified targets assuming that all the non-target dependencies already exist

    Default value: false

  • --vars VARS — Supply var bindings as a yml file or provide them as string e.g. ‘(key: value)‘

  • --env-vars ENV_VARS — Supply env var bindings as a yml file or provide them as string e.g. ‘(key: value)‘

  • --downstream — Execute cmd not only the given targets but also for all its downstream artifacts

    Default value: false

  • --format FORMAT — Show error tables in this format

    Default value: table

    Possible values: table, csv, tsv, json, nd-json, yml

  • --limit LIMIT — Limiting number of shown rows. Run with —limit 0 to remove limit

  • -d, --date DATE — Run command for this date , use YYYY-MM-DD or YYYY-MM-DDTHH:MM format

  • --from FROM — Run command for all dates from from (inclusive), use YYYY-MM-DD or YYYY-MM-DDTHH:MM format

  • --to TO — Run command for all dates to to (exclusive), use YYYY-MM-DD or YYYY-MM-DDTHH:MM format [default: now]

  • --dry-run — Plan command but don’t evaluate it

    Default value: false

sdf test

Test your models

Usage: sdf test [OPTIONS] [TARGETS]...

Arguments:
  • TARGETS — Assess code (data) quality: use source dirs, files or tables to determine which code (data) contracts to run [default: all code (data) contracts]
Options:
  • -e, --environment ENVIRONMENT — Use this environment

  • -s, --show SHOW — Display messages [default: progress if TARGETS is empty, all otherwise]

    Possible values: all, progress, result, none

  • -q, --query QUERY — Supply a .sql file or provide a sql snippet on the cmd line, e.g. ‘select * from t’

  • --stage STAGE — Run the following stages [default: all stages]

    Possible values: preprocess, parse, lint, resolve, classify, execute

  • --cache CACHE — Controls cache use

    Default value: read-write

    Possible values: read-write, write-only, read-only, none

  • --save SAVE — Controls which assets to save [default: none]

    Possible values: info-schema, assembly, table-deps

  • --targets-only — Processes only specified targets assuming that all the non-target dependencies already exist

    Default value: false

  • --vars VARS — Supply var bindings as a yml file or provide them as string e.g. ‘(key: value)‘

  • --env-vars ENV_VARS — Supply env var bindings as a yml file or provide them as string e.g. ‘(key: value)‘

  • --downstream — Execute cmd not only the given targets but also for all its downstream artifacts

    Default value: false

  • --format FORMAT — Show error tables in this format

    Default value: table

    Possible values: table, csv, tsv, json, nd-json, yml

  • --limit LIMIT — Limiting number of shown rows. Run with —limit 0 to remove limit

sdf stats

Statistics for your data

Usage: sdf stats [OPTIONS] [TARGETS]...

Arguments:
  • TARGETS — Profile data quality: use source dirs, files or tables to determine which data statistics to run [default: all data stats]
Options:
  • -e, --environment ENVIRONMENT — Use this environment

  • -s, --show SHOW — Display messages [default: progress if TARGETS is empty, all otherwise]

    Possible values: all, progress, result, none

  • -q, --query QUERY — Supply a .sql file or provide a sql snippet on the cmd line, e.g. ‘select * from t’

  • --stage STAGE — Run the following stages [default: all stages]

    Possible values: preprocess, parse, lint, resolve, classify, execute

  • --cache CACHE — Controls cache use

    Default value: read-write

    Possible values: read-write, write-only, read-only, none

  • --save SAVE — Controls which assets to save [default: none]

    Possible values: info-schema, assembly, table-deps

  • --targets-only — Processes only specified targets assuming that all the non-target dependencies already exist

    Default value: false

  • --vars VARS — Supply var bindings as a yml file or provide them as string e.g. ‘(key: value)‘

  • --env-vars ENV_VARS — Supply env var bindings as a yml file or provide them as string e.g. ‘(key: value)‘

  • --downstream — Execute cmd not only the given targets but also for all its downstream artifacts

    Default value: false

  • --format FORMAT — Show error tables in this format

    Default value: table

    Possible values: table, csv, tsv, json, nd-json, yml

  • --limit LIMIT — Limiting number of shown rows. Run with —limit 0 to remove limit

sdf report

Report code quality

Usage: sdf report [OPTIONS] [TARGETS]...

Arguments:
  • TARGETS — Profile code (data) quality, use source dirs, files or tables to determine which code (data) reports to run [default: all code (data) reports]
Options:
  • -e, --environment ENVIRONMENT — Use this environment

  • -s, --show SHOW — Display messages [default: progress if TARGETS is empty, all otherwise]

    Possible values: all, progress, result, none

  • -q, --query QUERY — Supply a .sql file or provide a sql snippet on the cmd line, e.g. ‘select * from t’

  • --stage STAGE — Run the following stages [default: all stages]

    Possible values: preprocess, parse, lint, resolve, classify, execute

  • --cache CACHE — Controls cache use

    Default value: read-write

    Possible values: read-write, write-only, read-only, none

  • --save SAVE — Controls which assets to save [default: none]

    Possible values: info-schema, assembly, table-deps

  • --targets-only — Processes only specified targets assuming that all the non-target dependencies already exist

    Default value: false

  • --vars VARS — Supply var bindings as a yml file or provide them as string e.g. ‘(key: value)‘

  • --env-vars ENV_VARS — Supply env var bindings as a yml file or provide them as string e.g. ‘(key: value)‘

  • --downstream — Execute cmd not only the given targets but also for all its downstream artifacts

    Default value: false

  • --format FORMAT — Show error tables in this format

    Default value: table

    Possible values: table, csv, tsv, json, nd-json, yml

  • --limit LIMIT — Limiting number of shown rows. Run with —limit 0 to remove limit

sdf check

Check code quality

Usage: sdf check [OPTIONS] [TARGETS]...

Arguments:
  • TARGETS — Check code (data) quality: use source dirs, files or tables to determine which code (data) checks to run [default: all code (data) checks]
Options:
  • -e, --environment ENVIRONMENT — Use this environment

  • -s, --show SHOW — Display messages [default: progress if TARGETS is empty, all otherwise]

    Possible values: all, progress, result, none

  • -q, --query QUERY — Supply a .sql file or provide a sql snippet on the cmd line, e.g. ‘select * from t’

  • --stage STAGE — Run the following stages [default: all stages]

    Possible values: preprocess, parse, lint, resolve, classify, execute

  • --cache CACHE — Controls cache use

    Default value: read-write

    Possible values: read-write, write-only, read-only, none

  • --save SAVE — Controls which assets to save [default: none]

    Possible values: info-schema, assembly, table-deps

  • --targets-only — Processes only specified targets assuming that all the non-target dependencies already exist

    Default value: false

  • --vars VARS — Supply var bindings as a yml file or provide them as string e.g. ‘(key: value)‘

  • --env-vars ENV_VARS — Supply env var bindings as a yml file or provide them as string e.g. ‘(key: value)‘

  • --downstream — Execute cmd not only the given targets but also for all its downstream artifacts

    Default value: false

  • --format FORMAT — Show error tables in this format

    Default value: table

    Possible values: table, csv, tsv, json, nd-json, yml

  • --limit LIMIT — Limiting number of shown rows. Run with —limit 0 to remove limit

  • --prefer-local — When set will minimize remote DB accesses and will rely on local schema information in .sdf.yml files

    Default value: false

sdf lineage

Display lineage for a given table and/or column

Usage: sdf lineage [OPTIONS] [TARGETS]...

Arguments:
  • TARGETS — Target selection: compute the lineage for the given source dirs, files or @tables [default: all models]
Options:
  • --column COLUMN — The column for which to compute lineage

  • --forward — Compute downstream lineage instead of the default upstream lineage

    Default value: false

  • --show-scans — Display scan dependencies in addition to copy and mod dependencies. Unset by default

    Default value: false

  • --max-depth MAX_DEPTH — Limiting the depth of shown lineage tree. Default value of 0 shows full lineage

    Default value: 0

  • -e, --environment ENVIRONMENT — Use this environment

  • -s, --show SHOW — Display messages [default: progress if TARGETS is empty, all otherwise]

    Possible values: all, progress, result, none

  • -q, --query QUERY — Supply a .sql file or provide a sql snippet on the cmd line, e.g. ‘select * from t’

  • --stage STAGE — Run the following stages [default: all stages]

    Possible values: preprocess, parse, lint, resolve, classify, execute

  • --cache CACHE — Controls cache use

    Default value: read-write

    Possible values: read-write, write-only, read-only, none

  • --save SAVE — Controls which assets to save [default: none]

    Possible values: info-schema, assembly, table-deps

  • --targets-only — Processes only specified targets assuming that all the non-target dependencies already exist

    Default value: false

  • --vars VARS — Supply var bindings as a yml file or provide them as string e.g. ‘(key: value)‘

  • --env-vars ENV_VARS — Supply env var bindings as a yml file or provide them as string e.g. ‘(key: value)‘

  • --downstream — Execute cmd not only the given targets but also for all its downstream artifacts

    Default value: false

  • --format FORMAT — Show error tables in this format

    Default value: table

    Possible values: table, csv, tsv, json, nd-json, yml

  • --limit LIMIT — Limiting number of shown rows. Run with —limit 0 to remove limit

sdf push

Push a local workspace to the SDF Service

Usage: sdf push [OPTIONS] [TARGETS]...

Arguments:
  • TARGETS — Target selection: push only the given source dirs, files or tables [default: current workspace directory]
Options:
  • --delete — Delete this workspace from the SDF Service

    Default value: false

  • -y, --yes — Answer yes to all prompts which include credentials’ uploading to the console when using a table provider

    Default value: false

  • -d, --dry-run — No changes will be made to the console when this is set

    Default value: false

  • -e, --environment ENVIRONMENT — Use this environment

  • -s, --show SHOW — Display messages [default: progress if TARGETS is empty, all otherwise]

    Possible values: all, progress, result, none

  • -q, --query QUERY — Supply a .sql file or provide a sql snippet on the cmd line, e.g. ‘select * from t’

  • --stage STAGE — Run the following stages [default: all stages]

    Possible values: preprocess, parse, lint, resolve, classify, execute

  • --cache CACHE — Controls cache use

    Default value: read-write

    Possible values: read-write, write-only, read-only, none

  • --save SAVE — Controls which assets to save [default: none]

    Possible values: info-schema, assembly, table-deps

  • --targets-only — Processes only specified targets assuming that all the non-target dependencies already exist

    Default value: false

  • --vars VARS — Supply var bindings as a yml file or provide them as string e.g. ‘(key: value)‘

  • --env-vars ENV_VARS — Supply env var bindings as a yml file or provide them as string e.g. ‘(key: value)‘

  • --downstream — Execute cmd not only the given targets but also for all its downstream artifacts

    Default value: false

  • --format FORMAT — Show error tables in this format

    Default value: table

    Possible values: table, csv, tsv, json, nd-json, yml

  • --limit LIMIT — Limiting number of shown rows. Run with —limit 0 to remove limit

sdf system

System maintenance, install and update

Usage: sdf system [OPTIONS] COMMAND

Subcommands:
  • update — Update sdf in place to the latest version
  • uninstall — Uninstall sdf from the system
Options:
  • -s, --show SHOW — Display messages

    Default value: progress

    Possible values: progress, none

sdf system update

Update sdf in place to the latest version

Usage: sdf system update [OPTIONS]

Options:
  • --version VERSION — Update sdf to this version [default: latest version]

  • -p, --preview — Update sdf to the latest preview version

    Default value: false

  • --stable — Update sdf to the latest stable version

    Default value: false

sdf system uninstall

Uninstall sdf from the system

Usage: sdf system uninstall

sdf auth

Authenticate CLI to services like SDF, AWS, OpenAI, etc

Usage: sdf auth [OPTIONS] COMMAND

Subcommands:
  • login
  • logout — Log out of SDF Service
  • status — Show status of credentials / tokens
Options:
  • -e, --environment ENVIRONMENT — Use this environment

  • -s, --show SHOW — Display messages [default: progress if TARGETS is empty, all otherwise]

    Possible values: all, progress, result, none

  • -q, --query QUERY — Supply a .sql file or provide a sql snippet on the cmd line, e.g. ‘select * from t’

  • --stage STAGE — Run the following stages [default: all stages]

    Possible values: preprocess, parse, lint, resolve, classify, execute

  • --cache CACHE — Controls cache use

    Default value: read-write

    Possible values: read-write, write-only, read-only, none

  • --save SAVE — Controls which assets to save [default: none]

    Possible values: info-schema, assembly, table-deps

  • --targets-only — Processes only specified targets assuming that all the non-target dependencies already exist

    Default value: false

  • --vars VARS — Supply var bindings as a yml file or provide them as string e.g. ‘(key: value)‘

  • --env-vars ENV_VARS — Supply env var bindings as a yml file or provide them as string e.g. ‘(key: value)‘

  • --downstream — Execute cmd not only the given targets but also for all its downstream artifacts

    Default value: false

  • --format FORMAT — Show error tables in this format

    Default value: table

    Possible values: table, csv, tsv, json, nd-json, yml

  • --limit LIMIT — Limiting number of shown rows. Run with —limit 0 to remove limit

sdf auth login

Usage: sdf auth login [OPTIONS] [COMMAND]

Subcommands:
  • aws — Configure how to authenticate with AWS
  • snowflake — Configure how to authenticate with Snowflake
  • bigquery — Configure how to authenticate with BigQuery
Options:
  • -n, --name NAME — Name of the credential to use
  • --id-token ID_TOKEN — Path to a file containing an OIDC identity token (a JWT)
  • --access-key ACCESS_KEY — Access key for headless authentication
  • --secret-key SECRET_KEY — Secret key for headless authentication
  • --credentials-dir CREDENTIALS_DIR — Path to the file where credentials will be stored (default is platform specific)

sdf auth login aws

Configure how to authenticate with AWS

Usage: sdf auth login aws [OPTIONS]

Options:
  • -n, --name NAME — Name of the credential to use

  • --default-region DEFAULT_REGION — AWS Region to use (default: us-east-1)

    Default value: us-east-1

  • --profile PROFILE — AWS profile to use, as usually defined in ~/.aws/config or ~/.aws/credentials

  • --role-arn ROLE_ARN — ARN of the role to assume

  • --external-id EXTERNAL_ID — External ID to use when assuming the role

  • --use-web-identity — Use web identity to authenticate

    Default value: false

  • --bucket-region-map BUCKET_REGION_MAP — Mapping of bucket names to regions

  • --credentials-dir CREDENTIALS_DIR — Path to the file where credentials will be stored (default is platform specific)

  • --access-key-id ACCESS_KEY_ID — AWS access key id

  • --secret-access-key SECRET_ACCESS_KEY — AWS secret access key

  • --session-token SESSION_TOKEN — AWS session token

sdf auth login snowflake

Configure how to authenticate with Snowflake

Usage: sdf auth login snowflake [OPTIONS] --account-id ACCOUNT_ID --username USERNAME

Options:
  • -n, --name NAME — Name of the credential to use
  • -a, --account-id ACCOUNT_ID — Snowflake account id
  • -U, --username USERNAME — Snowflake username
  • -P, --password PASSWORD — Snowflake password
  • -r, --role ROLE — Snowflake role
  • -W, --warehouse WAREHOUSE — Snowflake warehouse
  • --credentials-dir CREDENTIALS_DIR — Path to the file where credentials will be stored (default is platform specific)
  • --private-key-path PRIVATE_KEY_PATH — Path to the private key file for key pair authentication (mutually exclusive with password and private_key_pem)
  • --private-key-pem PRIVATE_KEY_PEM — The private key in PEM format (mutually exclusive with private_key_path and password)
  • --private-key-passphrase PRIVATE_KEY_PASSPHRASE — The passphrase for the private key, if it’s encrypted

sdf auth login bigquery

Configure how to authenticate with BigQuery

Usage: sdf auth login bigquery [OPTIONS]

Options:
  • -n, --name NAME — Name of the credential to use
  • -p, --project-id PROJECT_ID — GCP project id
  • -E, --client-email CLIENT_EMAIL — client_email of the service account key file
  • -K, --private-key PRIVATE_KEY — private_key of the service account key file
  • -J, --json-path JSON_PATH — path to the json file containing project id, client email, private key
  • --credentials-dir CREDENTIALS_DIR — path to the file where credentials will be stored (default is platform specific)

sdf auth logout

Log out of SDF Service

Usage: sdf auth logout [OPTIONS] [COMMAND]

Subcommands:
  • aws — Logout from AWS
  • openai — Logout from OpenAI
  • snowflake — Logout from Snowflake
  • bigquery — Logout from BigQuery
Options:
  • -n, --name NAME — Name of the credential to use
  • --credentials-dir CREDENTIALS_DIR — Path to the file where credentials are be stored (default is platform specific)

sdf auth logout aws

Logout from AWS

Usage: sdf auth logout aws [OPTIONS]

Options:
  • -n, --name NAME — Name of the credential to use
  • --credentials-dir CREDENTIALS_DIR — Path to the file where credentials are be stored (default is platform specific)

sdf auth logout openai

Logout from OpenAI

Usage: sdf auth logout openai [OPTIONS]

Options:
  • -n, --name NAME — Name of the credential to use
  • --credentials-dir CREDENTIALS_DIR — Path to the file where credentials are be stored (default is platform specific)

sdf auth logout snowflake

Logout from Snowflake

Usage: sdf auth logout snowflake [OPTIONS]

Options:
  • -n, --name NAME — Name of the credential to use
  • --credentials-dir CREDENTIALS_DIR — Path to the file where credentials are be stored (default is platform specific)

sdf auth logout bigquery

Logout from BigQuery

Usage: sdf auth logout bigquery [OPTIONS]

Options:
  • -n, --name NAME — Name of the credential to use
  • --credentials-dir CREDENTIALS_DIR — Path to the file where credentials are be stored (default is platform specific)

sdf auth status

Show status of credentials / tokens

Usage: sdf auth status [OPTIONS] [TARGETS]...

Arguments:
  • TARGETS — Compile only the given source dirs, files or tables [default: all models]
Options:
  • --credentials-dir CREDENTIALS_DIR — Path to the file where credentials will be stored (default is platform specific)

sdf man

Display reference material, like the CLI, dialect specific functions, schemas for authoring and interchange

Usage: sdf man COMMAND

Subcommands:
  • cli — Display SDF’s command line interface
  • functions — Display SDF’s functions definitions
  • definition-schema — Display SDF’s yml blocks as a json schema doc [only: json]
  • event-schema — Display SDF’s trace events as a json schema only [only: json]
  • information-schema — Display SDF’s information schemas [only: sql]
  • error-codes — Display SDF’ error codes [only: markdown]

sdf man cli

Display SDF’s command line interface

Usage: sdf man cli [OPTIONS]

Options:
  • --format FORMAT — Format reference material in this format [only: markdown]

    Possible values: sql, yml, markdown, json

sdf man functions

Display SDF’s functions definitions

Usage: sdf man functions [OPTIONS]

Options:
  • --dialect DIALECT — Dialect for all functions

    Default value: trino

  • --section SECTION — Section to display [ section value must appear in the function registry]

    Default value: all

  • --format FORMAT — Format reference material in this format [default: markdown | yml]

    Possible values: sql, yml, markdown, json

  • --implemented — limit output to only implemented functions

    Default value: false

sdf man definition-schema

Display SDF’s yml blocks as a json schema doc [only: json]

Usage: sdf man definition-schema [OPTIONS]

Options:
  • --format FORMAT — Format reference material in this format [only: json]

    Possible values: sql, yml, markdown, json

sdf man event-schema

Display SDF’s trace events as a json schema only [only: json]

Usage: sdf man event-schema [OPTIONS]

Options:
  • --format FORMAT — Format reference material in this format [only: json]

    Possible values: sql, yml, markdown, json

sdf man information-schema

Display SDF’s information schemas [only: sql]

Usage: sdf man information-schema [OPTIONS]

Options:
  • --format FORMAT — Format reference material in this format [only: sql]

    Possible values: sql, yml, markdown, json

sdf man error-codes

Display SDF’ error codes [only: markdown]

Usage: sdf man error-codes [OPTIONS]

Options:
  • --format FORMAT — Format reference material in this format [only: markdown]

    Possible values: sql, yml, markdown, json

sdf init

Initialize a workspace interactively

Usage: sdf init [PATH]

Arguments:
  • PATH — Create a new sdf workspace at path

sdf dbt

Initialize an sdf workspace from an existing dbt project

Usage: sdf dbt [OPTIONS] COMMAND

Subcommands:
  • init — Initialize a sdf workspace from a dbt project — best effort
  • refresh — Re-initialize a sdf workspace from a dbt project — best effort
Options:
  • -s, --show SHOW — Display messages

    Default value: progress

    Possible values: progress, none

sdf dbt init

Initialize a sdf workspace from a dbt project — best effort

Usage: sdf dbt init [OPTIONS]

Options:
  • --target TARGET — Use this DBT target over the default target in profiles.yml

  • --profiles-dir PROFILES_DIR — Use this DBT profile instead of the defaults at ~/.dbt/profile.yml — (note dbt uses —profile_dir, this CLI uses —profile-dir)

  • --workspace-dir WORKSPACE_DIR — Specifies the workspace directory where we expect to see manifest and dbt project files The SDF workspace file will be placed in the same directory. Default: current directory

  • -s, --save — Save and overwrite the workspace file

    Default value: false

  • -c, --config CONFIG — Supply a config yml file or provide config as yml string e.g. ‘(key: value)‘

sdf dbt refresh

Re-initialize a sdf workspace from a dbt project — best effort

Usage: sdf dbt refresh [OPTIONS]

Options:
  • --target TARGET — Use this DBT target over the default target in profiles.yml

  • --profiles-dir PROFILES_DIR — Use this DBT profile instead of the defaults at ~/.dbt/profile.yml — (note dbt uses —profile_dir, this CLI uses —profile-dir)

  • --workspace-dir WORKSPACE_DIR — Specifies the workspace directory where we expect to see manifest and dbt project files The SDF workspace file will be placed in the same directory. Default: current directory

  • -s, --save — Save and overwrite the workspace file

    Default value: false

  • -c, --config CONFIG — Supply a config yml file or provide config as yml string e.g. ‘(key: value)‘

sdf exec

Execute custom scripts

Usage: sdf exec [OPTIONS] COMMAND

Arguments:
  • COMMAND — The command to be executed
Options:
  • -s, --show SHOW — The verbosity of the output

    Default value: all

    Possible values: all, progress, result, none

  • --path PATH — Execute command at the path given [default: current workspace directory]