LogoLogo
Home PageLoginSQL Generator
  • 🖥️What is Rasgo?
  • 🚀Setting Up Rasgo
    • Connect Rasgo to your Data
  • 🛠️Using Rasgo
    • Modeling your Data
    • Prompt Guide
    • AI Notes
    • AI-Generated Documentation
    • AI Readiness Score
    • Reports
    • Additional Features
    • Admin Settings
  • 🎉What's New
  • Integrations
    • ❄️Snowflake
    • 🔍BigQuery
    • 🔴Redshift
    • 🧱DeltaLake (via Databricks)
    • 💬OpenAI
    • 🅰️Anthropic
    • ✨Gemini
    • ☁️dbt Cloud
  • 🔐API
    • Table Metadata
    • Column Metadata
  • Reference
    • Status Page
    • Frequently Asked Questions
      • Rasgo Architecture
      • Contacting Rasgo Support
      • What does Rasgo do with my data?
  • Rasgo Graveyard
    • PyRasgo 0.3
      • Source Methods
        • publish.source_data()
        • read.source_data()
        • get.data_sources()
        • get.data_source()
      • Feature Methods
        • feature.get_stats()
        • publish.features_from_source_code()
        • publish.feature_from_source()
        • publish.features()
        • read.feature_data()
        • get.feature_attributes()
        • get.features()
        • get.feature()
      • Collection Methods
        • collection.add_attributes()
        • collection.preview()
        • collection.get_compatible_features()
        • read.collection_snapshot_data()
        • read.collection_data()
        • get.collection_attributes()
        • get.collections()
        • get.collection()
      • Features yml file
      • version 0.3
    • Transforms Overview
      • Build your Own Transform
        • Argument Types
        • Make your own Transform
        • SQL Best Practices
        • Utilities
          • cleanse_name()
    • All Transforms
      • Aggregate String
      • Aggregate
      • Apply
      • Bin
      • Cast
      • Clean
      • Conditional Agg
      • Correlation
      • Cumulative Agg
      • Datarobot Score
      • Dateadd
      • Datediff
      • Datepart
      • Datespine Groups
      • Datespine
      • Datetrunc
      • Describe
      • Drop Columns
      • Dropna
      • Encode Values
      • Entropy
      • Extract Sequences
      • Filter
      • Funnel
      • Heatmap
      • Histogram
      • If Then
      • Join
      • Joins
      • Label Encode
      • Lag
      • Latest
      • Lead
      • Linear Regression
      • Market Basket
      • Math
      • Metric Plot
      • Metric
      • Min Max Scaler
      • Moving Avg
      • New Columns
      • One Hot Encode
      • Order
      • Pivot Table
      • Plot
      • Prefix
      • Profile Column
      • Query
      • Rank
      • Ratio With Shrinkage
      • Remove Duplicates
      • Remove Outliers
      • Rename
      • Replace Missing
      • Replace String
      • Reshape
      • Rolling Agg
      • Rsi
      • Sample Class
      • Sample
      • Sankey
      • Scale Columns
      • Select
      • Sliding Slope
      • Standard Scaler
      • Suffix
      • Summarize Flatlines
      • Summarize Islands
      • Summarize
      • Target Encode
      • Text To Sql
      • Timeseries Agg
      • To Date
      • Train Test Split
      • Union
      • Unions
      • Unpivot
      • Uppercase Columns
      • Vlookup
Powered by GitBook
On this page
  • How does Rasgo work with Redshift?
  • Connecting to Redshift
  • Credentials
  • Running Asynchronously
  • Success!

Was this helpful?

  1. Integrations

Redshift

How does Rasgo work with Redshift?

Rasgo is a metadata-only product, meaning all of your actual data stay in your data warehouse and Rasgo queries it there via dynamically generating SQL.

Rasgo performs reads-only operations in your Redshift env:

  • Rasgo reads the information schema for tables and columns it has access to

  • Rasgo dynamically generates and executes SQL on behalf of the user to analyze data

Connecting to Redshift

IP Restrictions

Rasgo will always connect to your account from these IP addresses. Make sure to whitelist them if you have networking restrictions enabled in AWS.

IP Address

54.84.138.60

54.84.66.109

Credentials

Rasgo needs service credentials to authenticate and execute SQL queries in your account.

When connecting Rasgo to your Redshift account, you'll need to provide this info:

Field
Description
Example

Database

Database to create new views and tables in

RASGO

Schema

Schema to create new views and tables in

PUBLIC

Host

Connection string

my-cluster.abc.us-east-2.redshift.amazonaws.com

User

Redshift username

RASGO_USER

Password

Redshift password

i<3rasgo

Running Asynchronously

Rasgo connects to your Redshift account using AWS's python connector. This limits the response time of individual queries to 30 seconds. If a query does not return results in under 30 seconds, the connection will be terminated and Rasgo will return an error message to the end user.

Success!

Configuration is complete! You're ready to start using Rasgo.

PreviousBigQueryNextDeltaLake (via Databricks)

Last updated 1 year ago

Was this helpful?

If your use case requires queries that take longer than 30 seconds, you can configure Rasgo to connect to your Data API to run queries asynchronously. This will require and providing an API key in addition to the above service credentials.

🔴
enabling the API in your AWS account