Sprinkle

Sprinkle

  • Docs
  • Tutorials
  • API
  • FAQ's
  • Blog
  • Go to sprinkledata.com

›Data Warehouse

Data Warehouse

  • Why the warehouse?
  • Amazon Athena
  • Apache Hive
  • Databricks
  • BigQuery
  • Snowflake
  • Redshift

Storage

  • Why the storage?
  • AWS S3 Bucket
  • Google Cloud Storage
  • Azure Blob Storage

Data Sources

  • Overview and Creating Data Source
  • Ingestion Mode
  • How Sprinkle handles the ingestion if there is a change in schema in the client DB?
  • Flattening JSON columns in DB
  • Column excluding and masking in DB table
  • Ingestion via SSH Tunnel
  • Configurable Destination Schema and table name
  • PostgreSQL
  • Salesforce
  • MySQL
  • MongoDB
  • Mixpanel
  • Hubspot
  • CosmosDB
  • CSV
  • AppsFlyer
  • CleverTap
  • SQL
  • Kafka
  • Amazon Kinesis
  • Azure Event Hub
  • Azure Table Storage
  • Zoho CRM
  • Freshsales
  • Google Analytics
  • GoogleSheet
  • Google Cloud Storage
  • Azure Blob
  • S3
  • Webhook
  • Sendgrid
  • Segment
  • Google Ads
  • Google Analytics MCF
  • Zendesk Support
  • Zendesk Chat
  • Google Search Console
  • Shopify
  • Facebook Ads
  • Mailchimp
  • WebURL
  • Klaviyo
  • SAP S4
  • Intercom
  • Marketo
  • Freshdesk
  • Leadsquared
  • Bigquery
  • MongoDB Atlas
  • Paytm
  • HDFS
  • FTPS
  • FTP

CDC Setup

  • MySQL
  • Postgres
  • Mongo

Transform

  • Schema Browser
  • Overview and Creating Flow
  • Advanced Features in Flow

KPI

    Models

    • Overview
    • Creating Model
    • Joins
    • Hierarchical Filters
    • Default Date Filters
    • Column Description in reports

    Segments

    • Overview
    • Creating Segment
    • Publish segment as table
    • Transpose
    • Show Labels Annotations on Charts
    • Tooltips
    • Fixed Columns
    • Conditional Builders
    • Cumulative Sum and Percentages
    • Embed Segment

    Metric Alerts

    • Overview and Creating Metric Alerts

Dashboards

  • Overview and Creating Dashboard
  • Embed Dashboard
  • Restricting filters
  • Sharing resources

Drill Down

  • Drill Down Feature In Segments And Dashboards
  • Drill Down Hierarchical Dimensions
  • Drill Down Expression Hierarchical Dimensions

Explores

  • Overview and Creating Explore
  • Show Labels Annotations on Charts
  • Tooltips

Machine Learning

  • Jupyter
  • Notebook Setup Guide

Sharing

  • Sharing Segments and Explore Reports
  • Share folders with users or groups

Scheduling

  • Schedule Timeline
  • Autorun

Notifications

  • Email Notifications
  • Slack Notifications

View Activity

  • View Activity

Admin

  • Admin -> usage
  • User Permissions & Restrictions
  • Github Integration

Launch On Cloud

  • AWS
  • Azure
  • Setup Sprinkle

Security

  • Security at Sprinkle
  • GDPR

Feedback

  • Option to take feedback from UI

Release Notes

  • Release Notes

Redshift

Amazon Redshift is a data warehouse product which forms part of the larger cloud-computing platform Amazon Web Services. Amazon Redshift is based on an older version of PostgreSQL and Redshift has made changes to that version.An initial preview beta was released in November 2012 and a full release was made available on February 15, 2013.

Sprinkle now integrates with Redshift and moves your data to your preferred data warehouse.

Now lets see how users can configure Redshift in the Sprinkle platform.

First the users need to go on Admin -> Drivers.

Admin tab consists of a number of subcategories of which Driver is one. The drivers tab routes the user to Storage, Warehouse and Compute tabs.

Warehouse Redshift

On clicking the Create new warehouse, the pop up displays a number of data warehouses. In this case, Redshiftis selected.

Warehouse Redshift

This will take the users to a new page wherein he needs to provide the details as requested.

Warehouse Redshift

Copy the jdbc url for your Redshift cluster with the following steps. More documentation at https://docs.aws.amazon.com/redshift/latest/mgmt/configuring-connections.html#connecting-connection-string.

1. Redshift console -> Clusters -> Choose the cluster
2. Copy the JDBC url from General Information. It will be of the following form jdbc:redshift://xxxxx.xxxxxxxxxx.<region>.redshift.amazonaws.com:5439/<databasename>

The user is required to specify a master user name, master user password and spectrum schema.

The spectrum schema can be created with the following steps. For more detailed information users can visit https://docs.aws.amazon.com/redshift/latest/dg/c-getting-started-using-spectrum.html

  1. Create IAM role with following steps
    1. IAM console -> Create role

    2. Choose AWS service ->Redshift -> Redshift-Customizable

    3. Choose AmazonS3FullAccess and AmazonAthenaFullAccess policies. If you want to use granular permissions, create a new Policy with following permissions.

       {
      
           "Effect": "Allow",
      
           "Action": [
      
               "glue:BatchCreatePartition",
      
               "glue:UpdateDatabase",
      
               "glue:CreateTable",
      
               "glue:DeleteDatabase",
      
               "glue:GetTables",
      
               "glue:GetPartitions",
      
               "glue:BatchDeletePartition",
      
               "glue:UpdateTable",
      
               "glue:BatchGetPartition",
      
               "glue:DeleteTable",
      
               "glue:GetDatabases",
      
               "glue:GetTable",
      
               "glue:GetDatabase",
      
               "glue:GetPartition",
      
               "glue:CreateDatabase",
      
               "glue:BatchDeleteTable",
      
               "glue:CreatePartition",
      
               "glue:DeletePartition",
      
               "glue:UpdatePartition"
           ],
      
           "Resource": "*"
      
       },
      
       {
      
           "Effect": "Allow",
      
           "Action": [
      
               "s3:PutObject",
      
               "s3:GetObject",
      
               "s3:ListBucketMultipartUploads",
      
               "s3:ListBucket",
      
               "s3:DeleteObject",
      
               "s3:GetBucketLocation",
      
               "s3:ListMultipartUploadParts"
           ],
      
           "Resource": [
      
               "arn:aws:s3:::bucketname/*",
      
               "arn:aws:s3:::bucketname"
           ]
       }
      
    4. Create role by specifying name : SprinkleSpectrumRole

    5. Copy the role ARN arn:aws:iam::294374035562:role/SprinkleSpectrumRole

  2. Associate IAM role with Redshift cluster
    1. Redshift console -> Clusters -> Choose the cluster
    2. Actions -> Manage IAM roles
    3. Choose the IAM role created in Step 1 -> Done
  3. Create external schema named sprinkle_spectrum using following command , replace iam_role with ARN copied in step 1
create external schema sprinkle_spectrum 

from data catalog 


database 'dev' 


iam_role 'arn:aws:iam::~~294374035562~~:role/SprinkleSpectrumRole'


create external database if not exists;

Once these details are provided users need to click on “Create”and a new Redshift data warehouse will be created, provided all the details provided are correct. If a user wants to check if the details are correct, he can click on “Test Connection” before clicking on “Create”.

← SnowflakeWhy the storage? →

Product

FeaturesHow it worksIntegrationsDeploymentPricing

Industries

Retail & EcommerceUrban MobilityFinanceEducation

Departments

MarketingOperationsTechnology

Connect

Free trialAbout Us

Actionable Insights. Faster.

Sprinkle offers self-service analytics by unlocking enterprise scale data via simple search and powerful reporting service.


Copyright © 2021 Sprinkle data