Queries
Available queries in the Management API
application
Returns the Application specified by the given ID.
Arguments
Returns
The Application object.
Propel Applications represent the web or mobile app you are building. They provide the API credentials that allow your client- or server-side app to access the Propel API. The Application’s Propeller determines the speed and cost of your Metric Queries.
The Application’s unique identifier.
The Application’s unique name.
The Application’s description.
The Application’s Account.
The Account object.
The Account’s unique identifier.
The Application’s Environment.
The Environments object.
Environments are independent and isolated Propel workspaces for development, staging (testing), and production workloads. Environments are hosted in a specific region, initially in us-east-2 only.
The Environment’s unique identifier.
The Environment’s unique name.
The Environment’s description.
The Environment’s creation date and time in UTC.
The Environment’s last modification date and time in UTC.
The Environment’s creator. It can be either a User ID, an Environment ID, or “system” if it was created by Propel.
The Environment’s last modifier. It can be either a User ID, an Environment ID, or “system” if it was modified by Propel.
The Application’s creation date and time in UTC.
The Application’s last modification date and time in UTC.
The Application’s creator. It can be either a User ID, an Application ID, or “system” if it was created by Propel.
The Application’s last modifier. It can be either a User ID, an Application ID, or “system” if it was modified by Propel.
The Application’s OAuth 2.0 client identifier.
The Application’s OAuth 2.0 client secret.
The Application’s Propeller.
A Propeller determines your Application’s query processing power. The larger the Propeller, the faster the queries and the higher the cost. Every Propel Application (and therefore every set of API credentials) has a Propeller that determines the speed and cost of queries.
P1_X_SMALL
: Max records per second: 5,000,000 records per secondP1_SMALL
: Max records per second: 25,000,000 records per secondP1_MEDIUM
: Max records per second: 100,000,000 records per secondP1_LARGE
: Max records per second: 250,000,000 records per secondP1_X_LARGE
: Max records per second: 500,000,000 records per second
The Application’s OAuth 2.0 scopes.
The API operations an Application is authorized to perform.
ADMIN
: Grant read/write access to Data Sources, Data Pools, Metrics and Policies.APPLICATION_ADMIN
: Grant read/write access to Applications.DATA_POOL_QUERY
: Grant read access to query Data Pools.DATA_POOL_READ
: Grant read access to read Data Pools.DATA_POOL_STATS
: Grant read access to fetch column statistics from Data Pools.ENVIRONMENT_ADMIN
: Grant read/write access to Environments.METRIC_QUERY
: Grant read access to query Metrics.METRIC_STATS
: Grant read access to fetch Dimension statistics from Metrics.METRIC_READ
: Grant read access to Metrics.
This does not allow querying Metrics. For that, see METRIC_QUERY
.
A paginated list of Data Pool Access Policies associated with the Application.
Arguments
applicationByName
Returns the Application with the given unique name.
Arguments
Returns
The Application object.
Propel Applications represent the web or mobile app you are building. They provide the API credentials that allow your client- or server-side app to access the Propel API. The Application’s Propeller determines the speed and cost of your Metric Queries.
The Application’s unique identifier.
The Application’s unique name.
The Application’s description.
The Application’s Account.
The Account object.
The Account’s unique identifier.
The Application’s Environment.
The Environments object.
Environments are independent and isolated Propel workspaces for development, staging (testing), and production workloads. Environments are hosted in a specific region, initially in us-east-2 only.
The Environment’s unique identifier.
The Environment’s unique name.
The Environment’s description.
The Environment’s creation date and time in UTC.
The Environment’s last modification date and time in UTC.
The Environment’s creator. It can be either a User ID, an Environment ID, or “system” if it was created by Propel.
The Environment’s last modifier. It can be either a User ID, an Environment ID, or “system” if it was modified by Propel.
The Application’s creation date and time in UTC.
The Application’s last modification date and time in UTC.
The Application’s creator. It can be either a User ID, an Application ID, or “system” if it was created by Propel.
The Application’s last modifier. It can be either a User ID, an Application ID, or “system” if it was modified by Propel.
The Application’s OAuth 2.0 client identifier.
The Application’s OAuth 2.0 client secret.
The Application’s Propeller.
A Propeller determines your Application’s query processing power. The larger the Propeller, the faster the queries and the higher the cost. Every Propel Application (and therefore every set of API credentials) has a Propeller that determines the speed and cost of queries.
P1_X_SMALL
: Max records per second: 5,000,000 records per secondP1_SMALL
: Max records per second: 25,000,000 records per secondP1_MEDIUM
: Max records per second: 100,000,000 records per secondP1_LARGE
: Max records per second: 250,000,000 records per secondP1_X_LARGE
: Max records per second: 500,000,000 records per second
The Application’s OAuth 2.0 scopes.
The API operations an Application is authorized to perform.
ADMIN
: Grant read/write access to Data Sources, Data Pools, Metrics and Policies.APPLICATION_ADMIN
: Grant read/write access to Applications.DATA_POOL_QUERY
: Grant read access to query Data Pools.DATA_POOL_READ
: Grant read access to read Data Pools.DATA_POOL_STATS
: Grant read access to fetch column statistics from Data Pools.ENVIRONMENT_ADMIN
: Grant read/write access to Environments.METRIC_QUERY
: Grant read access to query Metrics.METRIC_STATS
: Grant read access to fetch Dimension statistics from Metrics.METRIC_READ
: Grant read access to Metrics.
This does not allow querying Metrics. For that, see METRIC_QUERY
.
A paginated list of Data Pool Access Policies associated with the Application.
Arguments
applications
Returns the Applications within the Environment.
The applications
query uses cursor-based pagination typical of GraphQL APIs. You can use the pairs of parameters first
and after
or last
and before
to page forward or backward through the results, respectively.
For forward pagination, the first
parameter defines the number of results to return, and the after
parameter defines the cursor to continue from. You should pass the cursor for the last result of the current page to after
.
For backward pagination, the last
parameter defines the number of results to return, and the before
parameter defines the cursor to continue from. You should pass the cursor for the first result of the current page to before
.
Arguments
Returns
The Application connection object.
Learn more about pagination in GraphQL.
The Application connection’s edges.
The Application edge object.
Learn more about pagination in GraphQL.
The edge’s cursor.
The edge’s node.
See Application
The Application connection’s nodes.
The Application object.
Propel Applications represent the web or mobile app you are building. They provide the API credentials that allow your client- or server-side app to access the Propel API. The Application’s Propeller determines the speed and cost of your Metric Queries.
The Application’s unique identifier.
The Application’s unique name.
The Application’s description.
The Application’s Environment.
See Environment
The Application’s creation date and time in UTC.
The Application’s last modification date and time in UTC.
The Application’s creator. It can be either a User ID, an Application ID, or “system” if it was created by Propel.
The Application’s last modifier. It can be either a User ID, an Application ID, or “system” if it was modified by Propel.
The Application’s OAuth 2.0 client identifier.
The Application’s OAuth 2.0 client secret.
The Application’s OAuth 2.0 scopes.
See ApplicationScope
A paginated list of Data Pool Access Policies associated with the Application.
Arguments
The Application connection’s page info.
The page info object used for pagination.
Points to the first item returned in the results. Used when paginating backward.
Points to the last item returned in the results. Used when paginating forward.
A boolean that indicates whether a next page of results exists. Can be used to display a “next page” button in user interfaces, for example.
A boolean that indicates whether a previous page of results exists. Can be used to display a “previous page” button in user interfaces, for example.
dataSource
Returns the Data Source specified by the given ID.
query {
dataSource(id: "DSOXXXXX") {
id
uniqueName
type
tables (first: 100){
nodes {
id
name
columns (first: 100) {
nodes {
name
type
isNullable
supportedDataPoolColumnTypes
}
}
}
}
}
}
Arguments
Returns
The Data Source object.
A Data Source is a connection to your data warehouse. It has the necessary connection details for Propel to access Snowflake or any other supported Data Source.
The Data Source’s unique identifier.
The Data Source’s unique name.
The Data Source’s description.
The Data Source’s Account.
The Account object.
The Account’s unique identifier.
The Data Source’s Environment.
The Environments object.
Environments are independent and isolated Propel workspaces for development, staging (testing), and production workloads. Environments are hosted in a specific region, initially in us-east-2 only.
The Environment’s unique identifier.
The Environment’s unique name.
The Environment’s description.
The Environment’s creation date and time in UTC.
The Environment’s last modification date and time in UTC.
The Environment’s creator. It can be either a User ID, an Environment ID, or “system” if it was created by Propel.
The Environment’s last modifier. It can be either a User ID, an Environment ID, or “system” if it was modified by Propel.
The Data Source’s creation date and time in UTC.
The Data Source’s last modification date and time in UTC.
The Data Source’s creator. It can be either a User ID, an Application ID, or “system” if it was created by Propel.
The Data Source’s last modifier. It can be either a User ID, an Application ID, or “system” if it was modified by Propel.
The Data Source’s type.
The types of Data Sources.
WEBHOOK
: Indicates a Webhook Data Source.TWILIO_SEGMENT
: Indicates a Twilio Segment Data Source.S3
: Indicates an Amazon S3 Data Source.Redshift
: Indicates a Redshift Data Source.POSTGRESQL
: Indicates a PostgreSQL Data Source.KAFKA
: Indicates a Kafka Data Source.Http
: Indicates an Http Data Source.CLICKHOUSE
: Indicates a ClickHouse Data Source.AMAZON_DYNAMODB
: Indicates an Amazon DynamoDB Data Source.AMAZON_DATA_FIREHOSE
: Indicates an Amazon Data Firehose Data Source.Snowflake
: Indicates a Snowflake Data Source.INTERNAL
: Indicates an internal Data Source.
The Data Source’s status.
The status of a Data Source.
CREATED
: The Data Source has been created, but it is not connected yet.CONNECTING
: Propel is attempting to connect the Data Source.CONNECTED
: The Data Source is connected.BROKEN
: The Data Source failed to connect.DELETING
: Propel is deleting the Data Source.
The Data Source’s connection settings.
The tables contained within the Data Source, according to the most recent table introspection.
Arguments
See TableConnection
A list of table introspections performed for the Data Source. You can see how tables and columns changed over time by paging through this list.
Arguments
A list of checks performed on the Data Source during its most recent connection attempt.
The Data Source Check object.
Data Source Checks are executed when setting up your Data Source. They check that Propel will be able to receive data and setup Data Pools.
The exact Checks to perform vary by Data Source. For example, Snowflake-backed Data Sources will have their own specific Checks.
The name of the Data Source Check to be performed.
A description of the Data Source Check to be performed.
The status of the Data Source Check (all checks begin as NOT_STARTED before transitioning to SUCCEEDED or FAILED).
If the Data Source Check failed, this field includes a descriptive error message.
See Error
The time at which the Data Source Check was performed.
If you list Data Pools via the dataPools
field on a Data Source, you will get Data Pools for the Data Source.
The dataPools
field uses cursor-based pagination typical of GraphQL APIs. You can use the pairs of parameters first
and after
or last
and before
to page forward or backward through the results, respectively.
For forward pagination, the first
parameter defines the number of results to return, and the after
parameter defines the cursor to continue from. You should pass the cursor for the last result of the current page to after
.
For backward pagination, the last
parameter defines the number of results to return, and the before
parameter defines the cursor to continue from. You should pass the cursor for the first result of the current page to before
.
Arguments
dataSourceByName
Returns the Data Source specified by the given unique name.
Arguments
Returns
The Data Source object.
A Data Source is a connection to your data warehouse. It has the necessary connection details for Propel to access Snowflake or any other supported Data Source.
The Data Source’s unique identifier.
The Data Source’s unique name.
The Data Source’s description.
The Data Source’s Account.
The Account object.
The Account’s unique identifier.
The Data Source’s Environment.
The Environments object.
Environments are independent and isolated Propel workspaces for development, staging (testing), and production workloads. Environments are hosted in a specific region, initially in us-east-2 only.
The Environment’s unique identifier.
The Environment’s unique name.
The Environment’s description.
The Environment’s creation date and time in UTC.
The Environment’s last modification date and time in UTC.
The Environment’s creator. It can be either a User ID, an Environment ID, or “system” if it was created by Propel.
The Environment’s last modifier. It can be either a User ID, an Environment ID, or “system” if it was modified by Propel.
The Data Source’s creation date and time in UTC.
The Data Source’s last modification date and time in UTC.
The Data Source’s creator. It can be either a User ID, an Application ID, or “system” if it was created by Propel.
The Data Source’s last modifier. It can be either a User ID, an Application ID, or “system” if it was modified by Propel.
The Data Source’s type.
The types of Data Sources.
WEBHOOK
: Indicates a Webhook Data Source.TWILIO_SEGMENT
: Indicates a Twilio Segment Data Source.S3
: Indicates an Amazon S3 Data Source.Redshift
: Indicates a Redshift Data Source.POSTGRESQL
: Indicates a PostgreSQL Data Source.KAFKA
: Indicates a Kafka Data Source.Http
: Indicates an Http Data Source.CLICKHOUSE
: Indicates a ClickHouse Data Source.AMAZON_DYNAMODB
: Indicates an Amazon DynamoDB Data Source.AMAZON_DATA_FIREHOSE
: Indicates an Amazon Data Firehose Data Source.Snowflake
: Indicates a Snowflake Data Source.INTERNAL
: Indicates an internal Data Source.
The Data Source’s status.
The status of a Data Source.
CREATED
: The Data Source has been created, but it is not connected yet.CONNECTING
: Propel is attempting to connect the Data Source.CONNECTED
: The Data Source is connected.BROKEN
: The Data Source failed to connect.DELETING
: Propel is deleting the Data Source.
The Data Source’s connection settings.
The tables contained within the Data Source, according to the most recent table introspection.
Arguments
See TableConnection
A list of table introspections performed for the Data Source. You can see how tables and columns changed over time by paging through this list.
Arguments
A list of checks performed on the Data Source during its most recent connection attempt.
The Data Source Check object.
Data Source Checks are executed when setting up your Data Source. They check that Propel will be able to receive data and setup Data Pools.
The exact Checks to perform vary by Data Source. For example, Snowflake-backed Data Sources will have their own specific Checks.
The name of the Data Source Check to be performed.
A description of the Data Source Check to be performed.
The status of the Data Source Check (all checks begin as NOT_STARTED before transitioning to SUCCEEDED or FAILED).
If the Data Source Check failed, this field includes a descriptive error message.
See Error
The time at which the Data Source Check was performed.
If you list Data Pools via the dataPools
field on a Data Source, you will get Data Pools for the Data Source.
The dataPools
field uses cursor-based pagination typical of GraphQL APIs. You can use the pairs of parameters first
and after
or last
and before
to page forward or backward through the results, respectively.
For forward pagination, the first
parameter defines the number of results to return, and the after
parameter defines the cursor to continue from. You should pass the cursor for the last result of the current page to after
.
For backward pagination, the last
parameter defines the number of results to return, and the before
parameter defines the cursor to continue from. You should pass the cursor for the first result of the current page to before
.
Arguments
dataSources
Returns the Data Sources within the Environment.
A Data Source is a connection to your data warehouse. It has the necessary connection details for Propel to access Snowflake or any other supported Data Source. Environments are independent and isolated Propel workspaces for development, staging (testing), and production workloads.
The dataSources
query uses cursor-based pagination typical of GraphQL APIs. You can use the pairs of parameters first
and after
or last
and before
to page forward or backward through the results, respectively.
For forward pagination, the first
parameter defines the number of results to return, and the after
parameter defines the cursor to continue from. You should pass the cursor for the last result of the current page to after
.
For backward pagination, the last
parameter defines the number of results to return, and the before
parameter defines the cursor to continue from. You should pass the cursor for the first result of the current page to before
.
Arguments
Returns
The Data Source connection object.
Learn more about pagination in GraphQL.
The Data Source connection’s edges.
The Data Source edge object.
Learn more about pagination in GraphQL.
The edge’s cursor.
The edge’s node.
See DataSource
The Data Source connection’s nodes.
The Data Source object.
A Data Source is a connection to your data warehouse. It has the necessary connection details for Propel to access Snowflake or any other supported Data Source.
The Data Source’s unique identifier.
The Data Source’s unique name.
The Data Source’s description.
The Data Source’s Environment.
See Environment
The Data Source’s creation date and time in UTC.
The Data Source’s last modification date and time in UTC.
The Data Source’s creator. It can be either a User ID, an Application ID, or “system” if it was created by Propel.
The Data Source’s last modifier. It can be either a User ID, an Application ID, or “system” if it was modified by Propel.
The Data Source’s type.
See DataSourceType
The Data Source’s status.
See DataSourceStatus
The Data Source’s connection settings.
The tables contained within the Data Source, according to the most recent table introspection.
Arguments
See TableConnection
A list of table introspections performed for the Data Source. You can see how tables and columns changed over time by paging through this list.
Arguments
A list of checks performed on the Data Source during its most recent connection attempt.
See DataSourceCheck
If you list Data Pools via the dataPools
field on a Data Source, you will get Data Pools for the Data Source.
The dataPools
field uses cursor-based pagination typical of GraphQL APIs. You can use the pairs of parameters first
and after
or last
and before
to page forward or backward through the results, respectively.
For forward pagination, the first
parameter defines the number of results to return, and the after
parameter defines the cursor to continue from. You should pass the cursor for the last result of the current page to after
.
For backward pagination, the last
parameter defines the number of results to return, and the before
parameter defines the cursor to continue from. You should pass the cursor for the first result of the current page to before
.
Arguments
The Data Source connection’s page info.
The page info object used for pagination.
Points to the first item returned in the results. Used when paginating backward.
Points to the last item returned in the results. Used when paginating forward.
A boolean that indicates whether a next page of results exists. Can be used to display a “next page” button in user interfaces, for example.
A boolean that indicates whether a previous page of results exists. Can be used to display a “previous page” button in user interfaces, for example.
dataPool
Returns the Data Pool specified by the given ID.
A Data Pool is a cached table hydrated from your data warehouse optimized for high-concurrency and low-latency queries.
Arguments
Returns
The Data Pool object. Data Pools are Propel’s high-speed data store and cache
The Data Pool’s unique identifier.
The Data Pool’s unique name.
The Data Pool’s description.
The Data Pool’s Account.
The Account object.
The Account’s unique identifier.
The Data Pool’s Environment.
The Environments object.
Environments are independent and isolated Propel workspaces for development, staging (testing), and production workloads. Environments are hosted in a specific region, initially in us-east-2 only.
The Environment’s unique identifier.
The Environment’s unique name.
The Environment’s description.
The Environment’s creation date and time in UTC.
The Environment’s last modification date and time in UTC.
The Environment’s creator. It can be either a User ID, an Environment ID, or “system” if it was created by Propel.
The Environment’s last modifier. It can be either a User ID, an Environment ID, or “system” if it was modified by Propel.
The Data Pool’s creation date and time in UTC.
The Data Pool’s last modification date and time in UTC.
The Data Pool’s creator. It can be either a User ID, an Application ID, or “system” if it was created by Propel.
The Data Pool’s last modifier. It can be either a User ID, an Application ID, or “system” if it was modified by Propel.
The Data Pool’s Data Source. See DataSource
The Data Pool’s status.
The status of a Data Pool.
CREATED
: The Data Pool has been created and will be set up soon.PENDING
: Propel is attempting to set up the Data Pool.LIVE
: The Data Pool is set up and serving data. Check its Syncs to monitor data ingestion.SETUP_FAILED
: The Data Pool setup failed. Check its Setup Tasks before re-attempting setup.CONNECTING
CONNECTED
BROKEN
PAUSING
PAUSED
DELETING
: Propel is deleting the Data Pool and all of its associated data.
The Data Pool’s data retention in days (not yet supported).
The name of the Data Pool’s table.
The Data Pool’s primary timestamp column, if any.
A Data Pool’s primary timestamp column. Propel uses the primary timestamp to order and partition your data in Data Pools. It will serve as the time dimension for your Metrics.
The name of the column that represents the primary timestamp.
The primary timestamp column’s type.
The number of records in the Data Pool.
The amount of storage in terabytes used by the Data Pool.
The Data Pool’s columns.
Arguments
The Data Pool column connection object.
Learn more about pagination in GraphQL.
The Data Pool column connection’s edges.
The Data Pool column connection’s nodes.
See DataPoolColumn
The list of measures (numeric columns) in the Data Pool.
Arguments
The Data Pool column connection object.
Learn more about pagination in GraphQL.
The Data Pool column connection’s edges.
The Data Pool column connection’s nodes.
See DataPoolColumn
A list of setup tasks performed on the Data Pool during its most recent setup attempt.
The Data Pool Setup Task object.
Data Pool Setup Tasks are executed when setting up your Data Pool. They ensure Propel will be able to sync records from your Data Source to your Data Pool.
The exact Setup Tasks to perform vary by Data Source. For example, Data Pools pointing to a Snowflake-backed Data Sources will have their own specific Setup Tasks.
The name of the Data Pool Setup Task to be performed.
A description of the Data Pool Setup Task to be performed.
The status of the Data Pool Setup Task (all setup tasks begin as NOT_STARTED before transitioning to SUCCEEDED or FAILED).
If the Data Pool Setup Task failed, this field includes a descriptive error message.
See Error
The time at which the Data Pool Setup Task was completed.
Settings related to Data Pool syncing.
Settings related to Data Pool syncing.
Indicates whether syncing is enabled or disabled.
The syncing interval.
Note that the syncing interval is approximate. For example, setting the syncing interval to EVERY_1_HOUR
does not mean that syncing will occur exactly on the hour. Instead, the syncing interval starts relative to
when the Data Pool goes LIVE
, and Propel will attempt to sync approximately every hour. Additionally,
if you pause or resume syncing, this too can shift the syncing interval around.
The date and time of the most recent Sync in UTC.
The list of Syncs of the Data Pool.
Arguments
The filter to apply when listing the Syncs for a Data Pool.
EMPTY
: Returns only Syncs with empty records.NOT_EMPTY
: Returns only Syncs that contain one or more records.ALL
: Returns all Syncs, regardless of whether they contain records or not.
See SyncConnection
The Deletion Jobs that were historically issued to this Data Pool, sorted by creation time, in descending order.
Arguments
The Add Column Jobs that were historically issued to this Data Pool, sorted by creation time, in descending order.
Arguments
The UpdateDataPoolRecords Jobs that were historically issued to this Data Pool, sorted by creation time, in descending order.
Arguments
Whether the Data Pool has access control enabled or not.
If the Data Pool has access control enabled, Applications must be assigned Data Pool Access Policies in order to query the Data Pool and its Metrics.
A paginated list of Data Pool Access Policies available on the Data Pool.
Arguments
Validates a custom expression against the Data Pool’s available columns. If the provided expression is invalid, the ValidateExpressionResult response will contain a reason explaining why.
Arguments
Response returned by the validateExpression query for validating expressions in Custom Metrics.
Returns whether the expression is valid or not with a reason explaining why.
True if the expression is valid, false otherwise.
The reason for why the expression is not valid in case it isn’t, null otherwise.
The Data Pool’s table settings.
A Data Pool’s table settings.
These describe how the Data Pool’s table is created in ClickHouse.
The ClickHouse table engine for the Data Pool’s table.
See TableEngine
The PARTITION BY clause for the Data Pool’s table.
The PRIMARY KEY clause for the Data Pool’s table.
The ORDER BY clause for the Data Pool’s table.
The TTL clause for the Data Pool’s table.
The Data Pool’s columns that participate in its PARTITION BY clause.
The name of the Data Source column that this Data Pool column derives from.
The Data Pool column’s type. This may differ from the corresponding Data Source column’s type.
See ColumnType
The ClickHouse type. This is the exact representation of the type in ClickHouse.
Whether the column is nullable, meaning whether it accepts a null value.
The name of the Data Source column that this Data Pool column derives from.
deprecated: Start usingcolumnName
insteadThe Data Pool’s columns that participate in its PRIMARY KEY clause.
The name of the Data Source column that this Data Pool column derives from.
The Data Pool column’s type. This may differ from the corresponding Data Source column’s type.
See ColumnType
The ClickHouse type. This is the exact representation of the type in ClickHouse.
Whether the column is nullable, meaning whether it accepts a null value.
The name of the Data Source column that this Data Pool column derives from.
deprecated: Start usingcolumnName
insteadThe Data Pool’s columns that participate in its ORDER BY clause.
The name of the Data Source column that this Data Pool column derives from.
The Data Pool column’s type. This may differ from the corresponding Data Source column’s type.
See ColumnType
The ClickHouse type. This is the exact representation of the type in ClickHouse.
Whether the column is nullable, meaning whether it accepts a null value.
The name of the Data Source column that this Data Pool column derives from.
deprecated: Start usingcolumnName
insteadThe Data Pool’s tenant ID, if configured.
deprecated: Will be removed; use Data Pool Access Policies insteadThe Data Pool’s unique ID column. Propel uses the primary timestamp and a unique ID to compose a primary key for determining whether records should be inserted, deleted, or updated within the Data Pool.
deprecated: Will be removed; use table settings to define the primary key.A Data Pool’s unique ID column. Propel uses the primary timestamp and a unique ID to compose a primary key for determining whether records should be inserted, deleted, or updated within the Data Pool.
The name of the column that represents the unique ID.
dataPoolByName
Returns the Data Pool specified by the given unique name.
A Data Pool is a cached table hydrated from your data warehouse optimized for high-concurrency and low-latency queries.
Arguments
Returns
The Data Pool object. Data Pools are Propel’s high-speed data store and cache
The Data Pool’s unique identifier.
The Data Pool’s unique name.
The Data Pool’s description.
The Data Pool’s Account.
The Account object.
The Account’s unique identifier.
The Data Pool’s Environment.
The Environments object.
Environments are independent and isolated Propel workspaces for development, staging (testing), and production workloads. Environments are hosted in a specific region, initially in us-east-2 only.
The Environment’s unique identifier.
The Environment’s unique name.
The Environment’s description.
The Environment’s creation date and time in UTC.
The Environment’s last modification date and time in UTC.
The Environment’s creator. It can be either a User ID, an Environment ID, or “system” if it was created by Propel.
The Environment’s last modifier. It can be either a User ID, an Environment ID, or “system” if it was modified by Propel.
The Data Pool’s creation date and time in UTC.
The Data Pool’s last modification date and time in UTC.
The Data Pool’s creator. It can be either a User ID, an Application ID, or “system” if it was created by Propel.
The Data Pool’s last modifier. It can be either a User ID, an Application ID, or “system” if it was modified by Propel.
The Data Pool’s Data Source. See DataSource
The Data Pool’s status.
The status of a Data Pool.
CREATED
: The Data Pool has been created and will be set up soon.PENDING
: Propel is attempting to set up the Data Pool.LIVE
: The Data Pool is set up and serving data. Check its Syncs to monitor data ingestion.SETUP_FAILED
: The Data Pool setup failed. Check its Setup Tasks before re-attempting setup.CONNECTING
CONNECTED
BROKEN
PAUSING
PAUSED
DELETING
: Propel is deleting the Data Pool and all of its associated data.
The Data Pool’s data retention in days (not yet supported).
The name of the Data Pool’s table.
The Data Pool’s primary timestamp column, if any.
A Data Pool’s primary timestamp column. Propel uses the primary timestamp to order and partition your data in Data Pools. It will serve as the time dimension for your Metrics.
The name of the column that represents the primary timestamp.
The primary timestamp column’s type.
The number of records in the Data Pool.
The amount of storage in terabytes used by the Data Pool.
The Data Pool’s columns.
Arguments
The Data Pool column connection object.
Learn more about pagination in GraphQL.
The Data Pool column connection’s edges.
The Data Pool column connection’s nodes.
See DataPoolColumn
The list of measures (numeric columns) in the Data Pool.
Arguments
The Data Pool column connection object.
Learn more about pagination in GraphQL.
The Data Pool column connection’s edges.
The Data Pool column connection’s nodes.
See DataPoolColumn
A list of setup tasks performed on the Data Pool during its most recent setup attempt.
The Data Pool Setup Task object.
Data Pool Setup Tasks are executed when setting up your Data Pool. They ensure Propel will be able to sync records from your Data Source to your Data Pool.
The exact Setup Tasks to perform vary by Data Source. For example, Data Pools pointing to a Snowflake-backed Data Sources will have their own specific Setup Tasks.
The name of the Data Pool Setup Task to be performed.
A description of the Data Pool Setup Task to be performed.
The status of the Data Pool Setup Task (all setup tasks begin as NOT_STARTED before transitioning to SUCCEEDED or FAILED).
If the Data Pool Setup Task failed, this field includes a descriptive error message.
See Error
The time at which the Data Pool Setup Task was completed.
Settings related to Data Pool syncing.
Settings related to Data Pool syncing.
Indicates whether syncing is enabled or disabled.
The syncing interval.
Note that the syncing interval is approximate. For example, setting the syncing interval to EVERY_1_HOUR
does not mean that syncing will occur exactly on the hour. Instead, the syncing interval starts relative to
when the Data Pool goes LIVE
, and Propel will attempt to sync approximately every hour. Additionally,
if you pause or resume syncing, this too can shift the syncing interval around.
The date and time of the most recent Sync in UTC.
The list of Syncs of the Data Pool.
Arguments
The filter to apply when listing the Syncs for a Data Pool.
EMPTY
: Returns only Syncs with empty records.NOT_EMPTY
: Returns only Syncs that contain one or more records.ALL
: Returns all Syncs, regardless of whether they contain records or not.
See SyncConnection
The Deletion Jobs that were historically issued to this Data Pool, sorted by creation time, in descending order.
Arguments
The Add Column Jobs that were historically issued to this Data Pool, sorted by creation time, in descending order.
Arguments
The UpdateDataPoolRecords Jobs that were historically issued to this Data Pool, sorted by creation time, in descending order.
Arguments
Whether the Data Pool has access control enabled or not.
If the Data Pool has access control enabled, Applications must be assigned Data Pool Access Policies in order to query the Data Pool and its Metrics.
A paginated list of Data Pool Access Policies available on the Data Pool.
Arguments
Validates a custom expression against the Data Pool’s available columns. If the provided expression is invalid, the ValidateExpressionResult response will contain a reason explaining why.
Arguments
Response returned by the validateExpression query for validating expressions in Custom Metrics.
Returns whether the expression is valid or not with a reason explaining why.
True if the expression is valid, false otherwise.
The reason for why the expression is not valid in case it isn’t, null otherwise.
The Data Pool’s table settings.
A Data Pool’s table settings.
These describe how the Data Pool’s table is created in ClickHouse.
The ClickHouse table engine for the Data Pool’s table.
See TableEngine
The PARTITION BY clause for the Data Pool’s table.
The PRIMARY KEY clause for the Data Pool’s table.
The ORDER BY clause for the Data Pool’s table.
The TTL clause for the Data Pool’s table.
The Data Pool’s columns that participate in its PARTITION BY clause.
The name of the Data Source column that this Data Pool column derives from.
The Data Pool column’s type. This may differ from the corresponding Data Source column’s type.
See ColumnType
The ClickHouse type. This is the exact representation of the type in ClickHouse.
Whether the column is nullable, meaning whether it accepts a null value.
The name of the Data Source column that this Data Pool column derives from.
deprecated: Start usingcolumnName
instead