To connect to Snowflake you need to install ODBC connector:
Connecting to Snowflake
To connect to Snowflake create new documentation by clicking Add documentation and choosing Database connection.
On the connection screen choose Snowflake as DBMS.
Select authentication mode:
- Password - you will need to provide username and password.
- SSO (browser) - Dataedo will open web browser to let you authenticate with Okta.
- JWT (private key) - authentication with a private key. Learn more
Provide database connection details:
- Host - provide a host name or address where a database is on. E.g. server17, server17.ourdomain.com or 192.168.0.37.
- Port - change the default port of Amazon Redshift instance if required
- User and password - provide your username and password
- Database - type in schema name
Here is a comparison with connection details in DBeaver.
You can save password for later connections by checking Save password option. Password are saved in the repository database.
When connection was successful Dataedo will read objects and show a list of objects found. You can choose which objects to import. You can also use advanced filter to narrow down list of objects.
Confirm list of objects to import by clicking Next.
Next screen with allow you to change default name of the documentation under with your schema will be visible in Dataedo repository.
Click Import to start the import.
When done close import window with Finish button.
Your database schema has been imported to new documentation in the repository.
To sync any changes in the schema in Snowflake and reimport any technical metadata simply choose Import changes option. You will be asked to connect to Snowflake again and changes will be synced from the source.
|Tables, External tables||✅||✅|
|Input arguments (all as a single field)||✅||✅|
|Output results (as a single field)||✅||✅|
|Last updated time||✅|
Dataedo reads comments from following Snowflake objects:
Datedo supports following data profiling in Snowflake:
|Table row count||✅|
|Table sample data||✅|
|Column distribution (unique, non-unique, null, empty values)||✅|
|Min, max values||✅|
|Number of distinct values||✅|
|Top 10/100/1000 values||✅|
|10 random values||✅|
|Views - object level||From dependencies||✅|
|Views - object level||From SQL parsing||TBD|
|Views - column level||From SQL parsing||TBD|
|External tables (Azure, AWS)||TBD||TBD|
|External tables (Azure, AWS) - Stages||From SQL parsing||✅|
|Stages - Pipelines||From SQL parsing||✅|
|Tables - Pipelines||From SQL parsing||✅|
|Tables - COPY INTO||TBD||TBD|
|Tables - from Kafka||TBD||TBD|
INFORMATION_SCHEMAcurrently doesn’t contain data on object dependencies. Information about dependencies is retrieved from the
SNOWFLAKE.ACCOUNT_USAGE.OBJECT_DEPENDENCIESview. By default, the
SNOWFLAKEdatabase is available only to the
ACCOUNTADMINrole. Read more about Enabling Snowflake Database Usage for Other Roles.
Due to Snowflake Connector for .NET limitations:
- Username and password
- Cannot contain ; and = characters
- Database, role and warehouse name
- Cannot contain ; and ‘ or \ characters