From 1d689b68accb42ea85b68e02becd709158d493f0 Mon Sep 17 00:00:00 2001 From: atovpeko Date: Wed, 1 Oct 2025 15:50:10 +0300 Subject: [PATCH 1/6] data tiering update --- _partials/_not-supported-for-azure.md | 1 + .../data-tiering/about-data-tiering.md | 16 ++++++++++++++++ .../data-tiering/enabling-data-tiering.md | 16 ++++++++++++++++ use-timescale/data-tiering/index.md | 17 +++++++++++++++++ .../data-tiering/querying-tiered-data.md | 15 +++++++++++++++ .../data-tiering/tiered-data-replicas-forks.md | 17 +++++++++++++++++ 6 files changed, 82 insertions(+) create mode 100644 _partials/_not-supported-for-azure.md diff --git a/_partials/_not-supported-for-azure.md b/_partials/_not-supported-for-azure.md new file mode 100644 index 0000000000..066f410e4f --- /dev/null +++ b/_partials/_not-supported-for-azure.md @@ -0,0 +1 @@ +This feature is on our roadmap for $CLOUD_LONG on Microsoft Azure. Stay tuned! \ No newline at end of file diff --git a/use-timescale/data-tiering/about-data-tiering.md b/use-timescale/data-tiering/about-data-tiering.md index 1c16f8e1a0..77554caa75 100644 --- a/use-timescale/data-tiering/about-data-tiering.md +++ b/use-timescale/data-tiering/about-data-tiering.md @@ -11,9 +11,14 @@ cloud_ui: --- import TieredStorageBilling from "versionContent/_partials/_tiered-storage-billing.mdx"; +import NotSupportedAzure from "versionContent/_partials/_not-supported-for-azure.mdx"; # About storage tiers + + + + The tiered storage architecture in $CLOUD_LONG includes a high-performance storage tier and a low-cost object storage tier. You use the high-performance tier for data that requires quick access, and the object tier for rarely used historical data. Tiering policies move older data asynchronously and periodically from high-performance to low-cost storage, sparing you the need to do it manually. Chunks from a single hypertable, including compressed chunks, can stretch across these two storage tiers. ## High-performance storage @@ -119,6 +124,17 @@ The low-cost storage tier comes with the following limitations: partitioned on more than one dimension. Make sure your hypertables are partitioned on time only, before you enable tiered storage. + + + + + + + + + + + [blog-data-tiering]: https://www.timescale.com/blog/expanding-the-boundaries-of-postgresql-announcing-a-bottomless-consumption-based-object-storage-layer-built-on-amazon-s3/ [querying-tiered-data]: /use-timescale/:currentVersion:/data-tiering/querying-tiered-data/ [parquet]: https://parquet.apache.org/ diff --git a/use-timescale/data-tiering/enabling-data-tiering.md b/use-timescale/data-tiering/enabling-data-tiering.md index e5a50f7e04..9061f293b1 100644 --- a/use-timescale/data-tiering/enabling-data-tiering.md +++ b/use-timescale/data-tiering/enabling-data-tiering.md @@ -11,9 +11,14 @@ cloud_ui: --- import TieredStorageBilling from "versionContent/_partials/_tiered-storage-billing.mdx"; +import NotSupportedAzure from "versionContent/_partials/_not-supported-for-azure.mdx"; # Manage storage and tiering + + + + The tiered storage architecture in $CLOUD_LONG includes a high-performance storage tier and a low-cost object storage tier: - You use [high-performance storage][high-performance-storage] to store and query frequently accessed data. @@ -286,6 +291,17 @@ If you no longer want to use tiered storage for a particular hypertable, drop th + + + + + + + + + + + [data-retention]: /use-timescale/:currentVersion:/data-retention/ [console]: https://console.cloud.timescale.com/dashboard/services [hypertable]: /use-timescale/:currentVersion:/hypertables/ diff --git a/use-timescale/data-tiering/index.md b/use-timescale/data-tiering/index.md index 0d1cd52b5d..bea4d0852a 100644 --- a/use-timescale/data-tiering/index.md +++ b/use-timescale/data-tiering/index.md @@ -7,8 +7,14 @@ keywords: [tiered storage] tags: [storage, data management] --- +import NotSupportedAzure from "versionContent/_partials/_not-supported-for-azure.mdx"; + # Tiered storage + + + + Tiered storage is a [hierarchical storage management architecture][hierarchical-storage] for [time series and analytics][create-service] $SERVICE_SHORT you create in [$CLOUD_LONG](https://console.cloud.timescale.com/). @@ -53,6 +59,17 @@ In this section, you: * [Learn about replicas and forks with tiered data][replicas-and-forks]: understand how tiered storage works with forks and replicas of your $SERVICE_SHORT. + + + + + + + + + + + [about-data-tiering]: /use-timescale/:currentVersion:/data-tiering/about-data-tiering/ [enabling-data-tiering]: /use-timescale/:currentVersion:/data-tiering/enabling-data-tiering/ [replicas-and-forks]: /use-timescale/:currentVersion:/data-tiering/tiered-data-replicas-forks/ diff --git a/use-timescale/data-tiering/querying-tiered-data.md b/use-timescale/data-tiering/querying-tiered-data.md index 9b19ed5686..226bbc3f14 100644 --- a/use-timescale/data-tiering/querying-tiered-data.md +++ b/use-timescale/data-tiering/querying-tiered-data.md @@ -7,8 +7,14 @@ keywords: [ tiered storage, tiering ] tags: [ storage, data management ] --- +import NotSupportedAzure from "versionContent/_partials/_not-supported-for-azure.mdx"; + # Querying tiered data + + + + Once rarely used data is tiered and migrated to the object storage tier, it can still be queried with standard SQL by enabling the `timescaledb.enable_tiered_reads` GUC. By default, the GUC is set to `false`, so that queries do not touch tiered data. @@ -186,3 +192,12 @@ Queries over tiered data are expected to be slower than over local data. However * Text and non-native types (JSON, JSONB, GIS) filtering is slower when querying tiered data. + + + + + + + + + diff --git a/use-timescale/data-tiering/tiered-data-replicas-forks.md b/use-timescale/data-tiering/tiered-data-replicas-forks.md index 645ea13a6d..5da677d32f 100644 --- a/use-timescale/data-tiering/tiered-data-replicas-forks.md +++ b/use-timescale/data-tiering/tiered-data-replicas-forks.md @@ -7,8 +7,14 @@ keywords: [tiered storage] tags: [storage, data management] --- +import NotSupportedAzure from "versionContent/_partials/_not-supported-for-azure.mdx"; + # How tiered data works on replicas and forks + + + + There is one more thing that makes Tiered Storage even more amazing: when you keep data in the low-cost object storage tier, you pay for this data only once, regardless of whether you have a [high-availability replica][ha-replica] or [read replicas][read-replica] running in your $SERVICE_SHORT. We call this the savings multiplication effect of Tiered Storage. @@ -68,6 +74,17 @@ In the case of such a restore, new references are added to the deleted tiered ch Once 14 days pass after soft deleting the data,that is the number of references to the tiered data drop to 0, we hard delete the tiered data. + + + + + + + + + + + [ha-replica]: /use-timescale/:currentVersion:/ha-replicas/high-availability/ [read-replica]: /use-timescale/:currentVersion:/ha-replicas/read-scaling/#read-replicas [operations-forking]: /use-timescale/:currentVersion:/services/service-management/#fork-a-service From 8c608d8df793da7529c6b9b2c510496e753eda4b Mon Sep 17 00:00:00 2001 From: atovpeko Date: Thu, 2 Oct 2025 11:46:36 +0300 Subject: [PATCH 2/6] Updates for unsupported features --- integrations/aws.md | 16 ++++ integrations/cloudwatch.md | 16 ++++ integrations/corporate-data-center.md | 18 +++- integrations/datadog.md | 15 ++++ integrations/google-cloud.md | 16 ++++ integrations/microsoft-azure.md | 15 ++++ integrations/prometheus.md | 17 +++- use-timescale/data-tiering/index.md | 14 ++-- .../metrics-logging/aws-cloudwatch.md | 15 ++++ use-timescale/metrics-logging/datadog.md | 15 ++++ use-timescale/metrics-logging/index.md | 15 ++++ .../metrics-logging/metrics-to-prometheus.md | 17 +++- use-timescale/page-index/page-index.js | 2 +- use-timescale/security/overview.md | 83 +++---------------- use-timescale/security/transit-gateway.md | 15 ++++ use-timescale/security/vpc.md | 15 ++++ 16 files changed, 223 insertions(+), 81 deletions(-) diff --git a/integrations/aws.md b/integrations/aws.md index fdbd599dc2..f00058ce63 100644 --- a/integrations/aws.md +++ b/integrations/aws.md @@ -8,9 +8,14 @@ keywords: [AWS, integrations] import IntegrationPrereqsCloud from "versionContent/_partials/_integration-prereqs-cloud-only.mdx"; import TransitGateway from "versionContent/_partials/_transit-gateway.mdx"; +import NotSupportedAzure from "versionContent/_partials/_not-supported-for-azure.mdx"; # Integrate Amazon Web Services with $CLOUD_LONG + + + + [Amazon Web Services (AWS)][aws] is a comprehensive cloud computing platform that provides on-demand infrastructure, storage, databases, AI, analytics, and security services to help businesses build, deploy, and scale applications in the cloud. This page explains how to integrate your AWS infrastructure with $CLOUD_LONG using [AWS Transit Gateway][aws-transit-gateway]. @@ -33,6 +38,17 @@ To connect to $CLOUD_LONG: You have successfully integrated your AWS infrastructure with $CLOUD_LONG. + + + + + + + + + + + [aws]: https://aws.amazon.com/ [aws-transit-gateway]: https://aws.amazon.com/transit-gateway/ [gtw-setup]: https://docs.aws.amazon.com/vpc/latest/tgw/tgw-getting-started.html diff --git a/integrations/cloudwatch.md b/integrations/cloudwatch.md index ba0b58a60e..43a046aa70 100644 --- a/integrations/cloudwatch.md +++ b/integrations/cloudwatch.md @@ -6,12 +6,18 @@ price_plans: [scale, enterprise] keywords: [integrate] --- +import NotSupportedAzure from "versionContent/_partials/_not-supported-for-azure.mdx"; import IntegrationPrereqsCloud from "versionContent/_partials/_integration-prereqs-cloud-only.mdx"; import CloudWatchExporter from "versionContent/_partials/_cloudwatch-data-exporter.mdx"; import ManageDataExporter from "versionContent/_partials/_manage-a-data-exporter.mdx"; # Integrate Amazon CloudWatch with $CLOUD_LONG + + + + + [Amazon CloudWatch][cloudwatch] is a monitoring and observability service designed to help collect, analyze, and act on data from applications, infrastructure, and services running in AWS and on-premises environments. You can export telemetry data from your $SERVICE_LONGs with the time-series and analytics capability enabled to CloudWatch. The available metrics include CPU usage, RAM usage, and storage. This integration is available for [Scale and Enterprise][pricing-plan-features] pricing tiers. @@ -33,6 +39,16 @@ tool. You create an exporter on the [project level][projects], in the same AWS r + + + + + + + + + + [projects]: /use-timescale/:currentVersion:/security/members/ [pricing-plan-features]: /about/:currentVersion:/pricing-and-account-management/#features-included-in-each-plan [cloudwatch]: https://aws.amazon.com/cloudwatch/ diff --git a/integrations/corporate-data-center.md b/integrations/corporate-data-center.md index c95fed50be..cfef323a9b 100644 --- a/integrations/corporate-data-center.md +++ b/integrations/corporate-data-center.md @@ -8,9 +8,14 @@ keywords: [on-premise, integrations] import IntegrationPrereqsCloud from "versionContent/_partials/_integration-prereqs-cloud-only.mdx"; import TransitGateway from "versionContent/_partials/_transit-gateway.mdx"; +import NotSupportedAzure from "versionContent/_partials/_not-supported-for-azure.mdx"; # Integrate your data center with $CLOUD_LONG + + + + This page explains how to integrate your corporate on-premise infrastructure with $CLOUD_LONG using [AWS Transit Gateway][aws-transit-gateway]. ## Prerequisites @@ -33,7 +38,18 @@ To connect to $CLOUD_LONG: -You have successfully integrated your Microsoft Azure infrastructure with $CLOUD_LONG. +You have successfully integrated your corporate data center with $CLOUD_LONG. + + + + + + + + + + + [aws-transit-gateway]: https://aws.amazon.com/transit-gateway/ [gtw-setup]: https://docs.aws.amazon.com/vpc/latest/tgw/tgw-getting-started.html diff --git a/integrations/datadog.md b/integrations/datadog.md index f484dd2816..2695c0be70 100644 --- a/integrations/datadog.md +++ b/integrations/datadog.md @@ -9,9 +9,14 @@ keywords: [integrate] import IntegrationPrereqsCloud from "versionContent/_partials/_integration-prereqs-cloud-only.mdx"; import DataDogExporter from "versionContent/_partials/_datadog-data-exporter.mdx"; import ManageDataExporter from "versionContent/_partials/_manage-a-data-exporter.mdx"; +import NotSupportedAzure from "versionContent/_partials/_not-supported-for-azure.mdx"; # Integrate Datadog with $CLOUD_LONG + + + + [Datadog][datadog] is a cloud-based monitoring and analytics platform that provides comprehensive visibility into applications, infrastructure, and systems through real-time monitoring, logging, and analytics. @@ -132,6 +137,16 @@ metrics about your $SERVICE_LONGs. Metrics for your $SERVICE_LONG are now visible in Datadog. Check the Datadog $PG integration documentation for a comprehensive list of [metrics][datadog-postgres-metrics] collected. + + + + + + + + + + [datadog]: https://www.datadoghq.com/ [datadog-agent-install]: https://docs.datadoghq.com/getting_started/agent/#installation [datadog-postgres]: https://docs.datadoghq.com/integrations/postgres/ diff --git a/integrations/google-cloud.md b/integrations/google-cloud.md index acb099d606..672604a76d 100644 --- a/integrations/google-cloud.md +++ b/integrations/google-cloud.md @@ -8,9 +8,14 @@ keywords: [Google Cloud, integrations] import IntegrationPrereqsCloud from "versionContent/_partials/_integration-prereqs-cloud-only.mdx"; import TransitGateway from "versionContent/_partials/_transit-gateway.mdx"; +import NotSupportedAzure from "versionContent/_partials/_not-supported-for-azure.mdx"; # Integrate Google Cloud with $CLOUD_LONG + + + + [Google Cloud][google-cloud] is a suite of cloud computing services, offering scalable infrastructure, AI, analytics, databases, security, and developer tools to help businesses build, deploy, and manage applications. This page explains how to integrate your Google Cloud infrastructure with $CLOUD_LONG using [AWS Transit Gateway][aws-transit-gateway]. @@ -37,6 +42,17 @@ To connect to $CLOUD_LONG: You have successfully integrated your Google Cloud infrastructure with $CLOUD_LONG. + + + + + + + + + + + [google-cloud]: https://cloud.google.com/?hl=en [aws-transit-gateway]: https://aws.amazon.com/transit-gateway/ [gtw-setup]: https://docs.aws.amazon.com/vpc/latest/tgw/tgw-getting-started.html diff --git a/integrations/microsoft-azure.md b/integrations/microsoft-azure.md index 6ccf3279d4..c8c1758d1c 100644 --- a/integrations/microsoft-azure.md +++ b/integrations/microsoft-azure.md @@ -8,9 +8,14 @@ keywords: [Azure, integrations] import IntegrationPrereqsCloud from "versionContent/_partials/_integration-prereqs-cloud-only.mdx"; import TransitGateway from "versionContent/_partials/_transit-gateway.mdx"; +import NotSupportedAzure from "versionContent/_partials/_not-supported-for-azure.mdx"; # Integrate Microsoft Azure with $CLOUD_LONG + + + + [Microsoft Azure][azure] is a cloud computing platform and services suite, offering infrastructure, AI, analytics, security, and developer tools to help businesses build, deploy, and manage applications. This page explains how to integrate your Microsoft Azure infrastructure with $CLOUD_LONG using [AWS Transit Gateway][aws-transit-gateway]. @@ -37,6 +42,16 @@ To connect to $CLOUD_LONG: You have successfully integrated your Microsoft Azure infrastructure with $CLOUD_LONG. + + + + + + + + + + [aws-transit-gateway]: https://aws.amazon.com/transit-gateway/ [gtw-setup]: https://docs.aws.amazon.com/vpc/latest/tgw/tgw-getting-started.html [azure]: https://azure.microsoft.com/en-gb/ diff --git a/integrations/prometheus.md b/integrations/prometheus.md index e0cf5932c8..2cce9f9077 100644 --- a/integrations/prometheus.md +++ b/integrations/prometheus.md @@ -7,7 +7,22 @@ keywords: [integrate] --- import PrometheusIntegrate from "versionContent/_partials/_prometheus-integrate.mdx"; +import NotSupportedAzure from "versionContent/_partials/_not-supported-for-azure.mdx"; # Integrate Prometheus with $CLOUD_LONG - \ No newline at end of file + + + + + + + + + + + + + + + diff --git a/use-timescale/data-tiering/index.md b/use-timescale/data-tiering/index.md index bea4d0852a..8cef8b81cc 100644 --- a/use-timescale/data-tiering/index.md +++ b/use-timescale/data-tiering/index.md @@ -1,5 +1,5 @@ --- -title: Tiered Storage +title: Storage on Tiger Cloud excerpt: Save on storage costs by tiering older data to a low-cost bottomless object storage tier. Tiger Cloud tiered storage makes sure you cut costs while having data available for analytical queries products: [cloud] price_plans: [scale, enterprise] @@ -7,9 +7,7 @@ keywords: [tiered storage] tags: [storage, data management] --- -import NotSupportedAzure from "versionContent/_partials/_not-supported-for-azure.mdx"; - -# Tiered storage +# Storage @@ -63,7 +61,9 @@ In this section, you: - +$CLOUD_LONG stores your data in high-performance storage optimized for frequent querying. Based on [AWS EBS gp3][aws-gp3], the high-performance storage provides you with up to 16 TB and 16,000 IOPS. Its [$HYPERCORE row-columnar storage engine][hypercore], designed specifically for real-time analytics, enables you to compress your data by up to 98%, while improving performance. + +Coupled with other optimizations, $CLOUD_LONG high-performance storage makes sure your data is always accessible and your queries run at lightning speed. @@ -77,4 +77,6 @@ In this section, you: [querying-tiered-data]: /use-timescale/:currentVersion:/data-tiering/querying-tiered-data/ [add-retention-policies]: /api/:currentVersion:/continuous-aggregates/add_policies/ [create-service]: /getting-started/:currentVersion:/services/ -[hierarchical-storage]: https://en.wikipedia.org/wiki/Hierarchical_storage_management \ No newline at end of file +[hierarchical-storage]: https://en.wikipedia.org/wiki/Hierarchical_storage_management +[hypercore]: /use-timescale/:currentVersion:/hypercore +[aws-gp3]: https://docs.aws.amazon.com/ebs/latest/userguide/general-purpose.html \ No newline at end of file diff --git a/use-timescale/metrics-logging/aws-cloudwatch.md b/use-timescale/metrics-logging/aws-cloudwatch.md index 6c155ee5fe..c898a7fffa 100644 --- a/use-timescale/metrics-logging/aws-cloudwatch.md +++ b/use-timescale/metrics-logging/aws-cloudwatch.md @@ -10,9 +10,14 @@ tags: [telemetry, monitor] import ManageDataExporter from "versionContent/_partials/_manage-a-data-exporter.mdx"; import PrereqsCloud from "versionContent/_partials/_prereqs-cloud-no-connection.mdx"; import CloudWatchExporter from "versionContent/_partials/_cloudwatch-data-exporter.mdx"; +import NotSupportedAzure from "versionContent/_partials/_not-supported-for-azure.mdx"; # Export telemetry data to AWS Cloudwatch + + + + You can export telemetry data from your $SERVICE_LONGs with the time-series and analytics capability enabled to [Amazon CloudWatch][cloudwatch]. Available metrics include CPU usage, RAM usage, and storage. This integration is available for [Scale or Enterprise][pricing-plan-features] pricing plans. @@ -35,6 +40,16 @@ This section shows you how to attach, monitor, edit, and delete a data exporter. + + + + + + + + + + [cloudwatch]: https://aws.amazon.com/cloudwatch/ [cloudwatch-docs]: https://docs.aws.amazon.com/cloudwatch/index.html [console-integrations]: https://console.cloud.timescale.com/dashboard/integrations diff --git a/use-timescale/metrics-logging/datadog.md b/use-timescale/metrics-logging/datadog.md index 431551cfee..2ee823293d 100644 --- a/use-timescale/metrics-logging/datadog.md +++ b/use-timescale/metrics-logging/datadog.md @@ -10,9 +10,14 @@ tags: [telemetry, monitor] import DataDogExporter from "versionContent/_partials/_datadog-data-exporter.mdx"; import PrereqsCloud from "versionContent/_partials/_prereqs-cloud-no-connection.mdx"; import ManageDataExporter from "versionContent/_partials/_manage-a-data-exporter.mdx"; +import NotSupportedAzure from "versionContent/_partials/_not-supported-for-azure.mdx"; # Export telemetry data to Datadog + + + + You can export telemetry data from your $SERVICE_LONGs with the time-series and analytics capability enabled to [Datadog][datadog]. The available metrics include CPU usage, RAM usage, and storage. This integration is available for [Scale or Enterprise][pricing-plan-features] pricing plans. This page shows you how to create a Datadog exporter in $CONSOLE, and manage the lifecycle of data exporters. @@ -34,6 +39,16 @@ This section shows you how to attach, monitor, edit, and delete a data exporter. + + + + + + + + + + [datadog]: https://www.datadoghq.com [datadog-api-key]: https://docs.datadoghq.com/account_management/api-app-keys/#add-an-api-key-or-client-token [datadog-docs]: https://docs.datadoghq.com/ diff --git a/use-timescale/metrics-logging/index.md b/use-timescale/metrics-logging/index.md index 47000ace7d..0e30a29a14 100644 --- a/use-timescale/metrics-logging/index.md +++ b/use-timescale/metrics-logging/index.md @@ -12,6 +12,10 @@ cloud_ui: # Metrics and logging + + + + Find metrics and logs for your $SERVICE_SHORTs in $CONSOLE, or integrate with third-party monitoring services: * [Monitor][monitor] your $SERVICE_SHORTs in $CONSOLE. @@ -19,6 +23,17 @@ Find metrics and logs for your $SERVICE_SHORTs in $CONSOLE, or integrate with th * Export metrics to [Amazon Cloudwatch][cloudwatch]. * Export metrics to [Prometheus][prometheus]. + + + + +[Monitor][monitor] your $SERVICE_SHORTs in $CONSOLE. + + + + + + [prometheus]: /use-timescale/:currentVersion:/metrics-logging/metrics-to-prometheus/ [datadog]: /use-timescale/:currentVersion:/metrics-logging/datadog/ [cloudwatch]: /use-timescale/:currentVersion:/metrics-logging/aws-cloudwatch/ diff --git a/use-timescale/metrics-logging/metrics-to-prometheus.md b/use-timescale/metrics-logging/metrics-to-prometheus.md index df2e3c2a6d..657611c640 100644 --- a/use-timescale/metrics-logging/metrics-to-prometheus.md +++ b/use-timescale/metrics-logging/metrics-to-prometheus.md @@ -12,7 +12,22 @@ cloud_ui: --- import PrometheusIntegrate from "versionContent/_partials/_prometheus-integrate.mdx"; +import NotSupportedAzure from "versionContent/_partials/_not-supported-for-azure.mdx"; # Export metrics to Prometheus - \ No newline at end of file + + + + + + + + + + + + + + + diff --git a/use-timescale/page-index/page-index.js b/use-timescale/page-index/page-index.js index 871bc94722..579bbd38b6 100644 --- a/use-timescale/page-index/page-index.js +++ b/use-timescale/page-index/page-index.js @@ -381,7 +381,7 @@ module.exports = [ ], }, { - title: "Tiered storage", + title: "Storage", href: "data-tiering", excerpt: "Save on storage costs by tiering older data to a low-cost bottomless object storage tier", diff --git a/use-timescale/security/overview.md b/use-timescale/security/overview.md index 8fea3090a6..f2e022d535 100644 --- a/use-timescale/security/overview.md +++ b/use-timescale/security/overview.md @@ -6,44 +6,15 @@ keywords: [security] tags: [encryption, VPC, privacy] --- -# About security in $CLOUD_LONG - -$COMPANY implements a variety of secure software engineering practices in $TIMESCALE_DB, including code static analysis -for security hardening, automated scanning for dependency vulnerabilities, and code security reviews. -Additionally, $COMPANY has developed the https://github.com/timescale/pgspot open-source extension that we -use to identify security issues with $PG extensions. This has helped tighten our security posture. -$COMPANY products do not have any identified weaknesses. - -This page lists the additional things we do to ensure operational security and to lock down $SERVICE_LONGs. -To see our security features at a glance, see [$COMPANY Security][security-at-timescale]. - -## Role-based access - -$CLOUD_LONG provides role-based access for you to: - -* Administer your $PROJECT_LONG - In $CONSOLE_LONG, users with the Owner, Admin, and Viewer roles have different permissions to manage users and $SERVICE_SHORTs in the $PROJECT_SHORT. -* Manage data in each $SERVICE_SHORT - To restrict access to your data on the database level, you can create other roles on top of the default tsdbadmin role. - -## Data encryption - -Your data on $CLOUD_LONG is encrypted both in transit and at rest. Both active -databases and backups are encrypted. +import SecurityOverview from "versionContent/_partials/_security-overview.mdx"; -$CLOUD_LONG uses AWS as its cloud provider, with all the security that AWS -provides. Data encryption uses the industry-standard AES-256 algorithm. -Cryptographic keys are managed by -[AWS Key Management Service (AWS KMS)][aws-kms]. Keys are never stored in plaintext. +# About security in $CLOUD_LONG -For more information about AWS security, see the AWS documentation on security -in [Amazon Elastic Compute Cloud][ec2-security] and -[Elastic Block Storage][ebs-security]. + -## Networking security + -Customer access to $SERVICE_LONGs is only provided over TLS-encrypted -connections. There is no option to use unencrypted plaintext connections. + ## Networking with Virtual Private Cloud (VPC) peering @@ -52,50 +23,20 @@ $SERVICE_SHORT. $SERVICE_SHORT_CAP addresses are published in public DNS, but th connected to from the customer's peered VPC using private network addresses. VPC peering only enables communication to be initiated from your Customer VPC to -$SERVICE_LONGs running in the $CLOUD_LONG VPC. $CLOUD_LONG cannot initiate -communication with your VPC. To learn how to set up VPC Peering, see +$SERVICE_LONGs running in the $CLOUD_LONG VPC. $CLOUD_LONG cannot initiate +communication with your VPC. To learn how to set up VPC Peering, see [Secure your $SERVICE_LONGs with VPC Peering and AWS PrivateLink][vpc-peering]. -## IP address allow lists - -You can allow only trusted IP addresses to access your $SERVICE_LONGs. You do this by -creating [IP address allow lists][ip-allowlist] and attaching them to your $SERVICE_SHORTs. - -## Operator access - -Normally all the resources required for providing $SERVICE_LONGs are -automatically created, maintained and terminated by the $CLOUD_LONG -infrastructure. No manual operator intervention is required. - -However, the $COMPANY operations team has the capability to securely -log in to the $SERVICE_SHORT virtual machines for troubleshooting purposes. These -accesses are audit logged. - -No customer access to the virtual machine level is provided. - -## Customer data privacy + -Customer data privacy is of utmost importance at $COMPANY. By default, your data is encrypted both in transit and at rest. To do this, -$COMPANY uses various technical mechanisms, processes, and software development -lifecycle practices, to help ensure the security and privacy of your data. + -$COMPANY complies with the European Union's General Data Protection Regulation -(GDPR), and all practices are covered by our -[Privacy Policy][timescale-privacy-policy] -and the [Terms of Service][tsc-tos]. All customer data is -processed in accordance with $COMPANY's GDPR-compliant -[Data Processor Addendum][tsc-data-processor-addendum], -which applies to all $COMPANY customers. + -$COMPANY operators never access customer data, unless explicitly requested by -the customer to troubleshoot a technical issue. The $COMPANY operations team -has mandatory recurring training regarding the applicable policies. + -## HIPAA compliance + -The $CLOUD_LONG [Enterprise plan][pricing-plan-features] is Health Insurance Portability and Accountability Act -(HIPAA) compliant. This allows organizations to securely manage and analyze sensitive healthcare data, ensuring they -meet regulatory requirements while building compliant applications. [timescale-privacy-policy]: https://www.timescale.com/legal/privacy [tsc-tos]: https://www.timescale.com/legal/timescale-cloud-terms-of-service diff --git a/use-timescale/security/transit-gateway.md b/use-timescale/security/transit-gateway.md index 735488b652..76e186d647 100644 --- a/use-timescale/security/transit-gateway.md +++ b/use-timescale/security/transit-gateway.md @@ -11,9 +11,14 @@ cloud_ui: --- import TransitGateway from "versionContent/_partials/_transit-gateway.mdx"; +import NotSupportedAzure from "versionContent/_partials/_not-supported-for-azure.mdx"; # Securely connect to $CLOUD_LONG using AWS Transit Gateway + + + + [AWS Transit Gateway][aws-transit-gateway] enables you to securely connect to your $CLOUD_LONG from AWS, Google Cloud, Microsoft Azure, or any other cloud or on-premise environment. You use AWS Transit Gateway as a traffic controller for your network. Instead of setting up multiple direct connections to different clouds, on-premise data centers, and other AWS services, you connect everything to AWS Transit Gateway. This simplifies your network and makes it easier to manage and scale. @@ -89,6 +94,16 @@ AWS Transit Gateway enables you to connect from almost any environment, this pag You can now securely access your $SERVICE_SHORTs in $CLOUD_LONG. + + + + + + + + + + [aws-transit-gateway]: https://aws.amazon.com/transit-gateway/ [pricing-plans]: /about/:currentVersion:/pricing-and-account-management/ [azure-aws]: https://aws.amazon.com/blogs/modernizing-with-aws/designing-private-network-connectivity-aws-azure/ diff --git a/use-timescale/security/vpc.md b/use-timescale/security/vpc.md index f05ac71110..baa80947a9 100644 --- a/use-timescale/security/vpc.md +++ b/use-timescale/security/vpc.md @@ -10,9 +10,14 @@ cloud_ui: --- import VpcLimitations from "versionContent/_partials/_vpc-limitations.mdx"; +import NotSupportedAzure from "versionContent/_partials/_not-supported-for-azure.mdx"; # Secure your $CLOUD_LONG services with $VPC Peering and AWS PrivateLink + + + + You use Virtual Private Cloud ($VPC) peering to ensure that your $SERVICE_LONGs are only accessible through your secured AWS infrastructure. This reduces the potential attack vector surface and improves security. @@ -215,6 +220,16 @@ Migration takes a few minutes to complete and requires a change to DNS settings $SERVICE_SHORT. The $SERVICE_SHORT is not accessible during this time. If you receive a DNS error, allow some time for DNS propagation. + + + + + + + + + + [aws-dashboard]: https://console.aws.amazon.com/vpc/home#PeeringConnections: [aws-security-groups]: https://console.aws.amazon.com/vpcconsole/home#securityGroups: [console-login]: https://console.cloud.timescale.com/ From 5eb67f755926aecf717a67c1ea589dd18777ee47 Mon Sep 17 00:00:00 2001 From: atovpeko Date: Thu, 2 Oct 2025 11:47:48 +0300 Subject: [PATCH 3/6] Updates for unsupported features --- _partials/_security-overview.md | 77 +++++++++++++++++++++++++++++++++ 1 file changed, 77 insertions(+) create mode 100644 _partials/_security-overview.md diff --git a/_partials/_security-overview.md b/_partials/_security-overview.md new file mode 100644 index 0000000000..23c956bab3 --- /dev/null +++ b/_partials/_security-overview.md @@ -0,0 +1,77 @@ +$COMPANY implements a variety of secure software engineering practices in $TIMESCALE_DB, including code static analysis +for security hardening, automated scanning for dependency vulnerabilities, and code security reviews. +Additionally, $COMPANY has developed the https://github.com/timescale/pgspot open-source extension that we +use to identify security issues with $PG extensions. This has helped tighten our security posture. +$COMPANY products do not have any identified weaknesses. + +This page lists the additional things we do to ensure operational security and to lock down $SERVICE_LONGs. +To see our security features at a glance, see [$COMPANY Security][security-at-timescale]. + +## Role-based access + +$CLOUD_LONG provides role-based access for you to: + +* Administer your $PROJECT_LONG + In $CONSOLE_LONG, users with the Owner, Admin, and Viewer roles have different permissions to manage users and $SERVICE_SHORTs in the $PROJECT_SHORT. +* Manage data in each $SERVICE_SHORT + To restrict access to your data on the database level, you can create other roles on top of the default tsdbadmin role. + +## Data encryption + +Your data on $CLOUD_LONG is encrypted both in transit and at rest. Both active +databases and backups are encrypted. + +$CLOUD_LONG uses AWS as its cloud provider, with all the security that AWS +provides. Data encryption uses the industry-standard AES-256 algorithm. +Cryptographic keys are managed by +[AWS Key Management Service (AWS KMS)][aws-kms]. Keys are never stored in plaintext. + +For more information about AWS security, see the AWS documentation on security +in [Amazon Elastic Compute Cloud][ec2-security] and +[Elastic Block Storage][ebs-security]. + +## IP address allow lists + +You can allow only trusted IP addresses to access your $SERVICE_LONGs. You do this by +creating [IP address allow lists][ip-allowlist] and attaching them to your $SERVICE_SHORTs. + +## Operator access + +Normally all the resources required for providing $SERVICE_LONGs are +automatically created, maintained and terminated by the $CLOUD_LONG +infrastructure. No manual operator intervention is required. + +However, the $COMPANY operations team has the capability to securely +log in to the $SERVICE_SHORT virtual machines for troubleshooting purposes. These +accesses are audit logged. + +No customer access to the virtual machine level is provided. + +## Customer data privacy + +Customer data privacy is of utmost importance at $COMPANY. By default, your data is encrypted both in transit and at rest. To do this, +$COMPANY uses various technical mechanisms, processes, and software development +lifecycle practices, to help ensure the security and privacy of your data. + +$COMPANY complies with the European Union's General Data Protection Regulation +(GDPR), and all practices are covered by our +[Privacy Policy][timescale-privacy-policy] +and the [Terms of Service][tsc-tos]. All customer data is +processed in accordance with $COMPANY's GDPR-compliant +[Data Processor Addendum][tsc-data-processor-addendum], +which applies to all $COMPANY customers. + +$COMPANY operators never access customer data, unless explicitly requested by +the customer to troubleshoot a technical issue. The $COMPANY operations team +has mandatory recurring training regarding the applicable policies. + +## HIPAA compliance + +The $CLOUD_LONG [Enterprise plan][pricing-plan-features] is Health Insurance Portability and Accountability Act +(HIPAA) compliant. This allows organizations to securely manage and analyze sensitive healthcare data, ensuring they +meet regulatory requirements while building compliant applications. + +## Networking security + +Customer access to $SERVICE_LONGs is only provided over TLS-encrypted +connections. There is no option to use unencrypted plaintext connections. \ No newline at end of file From 2bbd9124d86a245c40183e961fb8618511ff74c8 Mon Sep 17 00:00:00 2001 From: atovpeko Date: Sun, 5 Oct 2025 15:50:25 +0300 Subject: [PATCH 4/6] revert security --- _partials/_security-overview.md | 77 ---------------------------- use-timescale/security/overview.md | 81 ++++++++++++++++++++++++++---- 2 files changed, 70 insertions(+), 88 deletions(-) delete mode 100644 _partials/_security-overview.md diff --git a/_partials/_security-overview.md b/_partials/_security-overview.md deleted file mode 100644 index 23c956bab3..0000000000 --- a/_partials/_security-overview.md +++ /dev/null @@ -1,77 +0,0 @@ -$COMPANY implements a variety of secure software engineering practices in $TIMESCALE_DB, including code static analysis -for security hardening, automated scanning for dependency vulnerabilities, and code security reviews. -Additionally, $COMPANY has developed the https://github.com/timescale/pgspot open-source extension that we -use to identify security issues with $PG extensions. This has helped tighten our security posture. -$COMPANY products do not have any identified weaknesses. - -This page lists the additional things we do to ensure operational security and to lock down $SERVICE_LONGs. -To see our security features at a glance, see [$COMPANY Security][security-at-timescale]. - -## Role-based access - -$CLOUD_LONG provides role-based access for you to: - -* Administer your $PROJECT_LONG - In $CONSOLE_LONG, users with the Owner, Admin, and Viewer roles have different permissions to manage users and $SERVICE_SHORTs in the $PROJECT_SHORT. -* Manage data in each $SERVICE_SHORT - To restrict access to your data on the database level, you can create other roles on top of the default tsdbadmin role. - -## Data encryption - -Your data on $CLOUD_LONG is encrypted both in transit and at rest. Both active -databases and backups are encrypted. - -$CLOUD_LONG uses AWS as its cloud provider, with all the security that AWS -provides. Data encryption uses the industry-standard AES-256 algorithm. -Cryptographic keys are managed by -[AWS Key Management Service (AWS KMS)][aws-kms]. Keys are never stored in plaintext. - -For more information about AWS security, see the AWS documentation on security -in [Amazon Elastic Compute Cloud][ec2-security] and -[Elastic Block Storage][ebs-security]. - -## IP address allow lists - -You can allow only trusted IP addresses to access your $SERVICE_LONGs. You do this by -creating [IP address allow lists][ip-allowlist] and attaching them to your $SERVICE_SHORTs. - -## Operator access - -Normally all the resources required for providing $SERVICE_LONGs are -automatically created, maintained and terminated by the $CLOUD_LONG -infrastructure. No manual operator intervention is required. - -However, the $COMPANY operations team has the capability to securely -log in to the $SERVICE_SHORT virtual machines for troubleshooting purposes. These -accesses are audit logged. - -No customer access to the virtual machine level is provided. - -## Customer data privacy - -Customer data privacy is of utmost importance at $COMPANY. By default, your data is encrypted both in transit and at rest. To do this, -$COMPANY uses various technical mechanisms, processes, and software development -lifecycle practices, to help ensure the security and privacy of your data. - -$COMPANY complies with the European Union's General Data Protection Regulation -(GDPR), and all practices are covered by our -[Privacy Policy][timescale-privacy-policy] -and the [Terms of Service][tsc-tos]. All customer data is -processed in accordance with $COMPANY's GDPR-compliant -[Data Processor Addendum][tsc-data-processor-addendum], -which applies to all $COMPANY customers. - -$COMPANY operators never access customer data, unless explicitly requested by -the customer to troubleshoot a technical issue. The $COMPANY operations team -has mandatory recurring training regarding the applicable policies. - -## HIPAA compliance - -The $CLOUD_LONG [Enterprise plan][pricing-plan-features] is Health Insurance Portability and Accountability Act -(HIPAA) compliant. This allows organizations to securely manage and analyze sensitive healthcare data, ensuring they -meet regulatory requirements while building compliant applications. - -## Networking security - -Customer access to $SERVICE_LONGs is only provided over TLS-encrypted -connections. There is no option to use unencrypted plaintext connections. \ No newline at end of file diff --git a/use-timescale/security/overview.md b/use-timescale/security/overview.md index f2e022d535..8cf3d4b4c7 100644 --- a/use-timescale/security/overview.md +++ b/use-timescale/security/overview.md @@ -6,15 +6,44 @@ keywords: [security] tags: [encryption, VPC, privacy] --- -import SecurityOverview from "versionContent/_partials/_security-overview.mdx"; - # About security in $CLOUD_LONG - +$COMPANY implements a variety of secure software engineering practices in $TIMESCALE_DB, including code static analysis +for security hardening, automated scanning for dependency vulnerabilities, and code security reviews. +Additionally, $COMPANY has developed the https://github.com/timescale/pgspot open-source extension that we +use to identify security issues with $PG extensions. This has helped tighten our security posture. +$COMPANY products do not have any identified weaknesses. + +This page lists the additional things we do to ensure operational security and to lock down $SERVICE_LONGs. +To see our security features at a glance, see [$COMPANY Security][security-at-timescale]. + +## Role-based access + +$CLOUD_LONG provides role-based access for you to: + +* Administer your $PROJECT_LONG + In $CONSOLE_LONG, users with the Owner, Admin, and Viewer roles have different permissions to manage users and $SERVICE_SHORTs in the $PROJECT_SHORT. +* Manage data in each $SERVICE_SHORT + To restrict access to your data on the database level, you can create other roles on top of the default tsdbadmin role. + +## Data encryption + +Your data on $CLOUD_LONG is encrypted both in transit and at rest. Both active +databases and backups are encrypted. + +$CLOUD_LONG uses AWS as its cloud provider, with all the security that AWS +provides. Data encryption uses the industry-standard AES-256 algorithm. +Cryptographic keys are managed by +[AWS Key Management Service (AWS KMS)][aws-kms]. Keys are never stored in plaintext. - +For more information about AWS security, see the AWS documentation on security +in [Amazon Elastic Compute Cloud][ec2-security] and +[Elastic Block Storage][ebs-security]. - +## Networking security + +Customer access to $SERVICE_LONGs is only provided over TLS-encrypted +connections. There is no option to use unencrypted plaintext connections. ## Networking with Virtual Private Cloud (VPC) peering @@ -27,16 +56,46 @@ $SERVICE_LONGs running in the $CLOUD_LONG VPC. $CLOUD_LONG cannot initiate communication with your VPC. To learn how to set up VPC Peering, see [Secure your $SERVICE_LONGs with VPC Peering and AWS PrivateLink][vpc-peering]. - +## IP address allow lists + +You can allow only trusted IP addresses to access your $SERVICE_LONGs. You do this by +creating [IP address allow lists][ip-allowlist] and attaching them to your $SERVICE_SHORTs. + +## Operator access + +Normally all the resources required for providing $SERVICE_LONGs are +automatically created, maintained and terminated by the $CLOUD_LONG +infrastructure. No manual operator intervention is required. + +However, the $COMPANY operations team has the capability to securely +log in to the $SERVICE_SHORT virtual machines for troubleshooting purposes. These +accesses are audit logged. + +No customer access to the virtual machine level is provided. + +## Customer data privacy - +Customer data privacy is of utmost importance at $COMPANY. By default, your data is encrypted both in transit and at rest. To do this, +$COMPANY uses various technical mechanisms, processes, and software development +lifecycle practices, to help ensure the security and privacy of your data. - +$COMPANY complies with the European Union's General Data Protection Regulation +(GDPR), and all practices are covered by our +[Privacy Policy][timescale-privacy-policy] +and the [Terms of Service][tsc-tos]. All customer data is +processed in accordance with $COMPANY's GDPR-compliant +[Data Processor Addendum][tsc-data-processor-addendum], +which applies to all $COMPANY customers. - +$COMPANY operators never access customer data, unless explicitly requested by +the customer to troubleshoot a technical issue. The $COMPANY operations team +has mandatory recurring training regarding the applicable policies. - +## HIPAA compliance +The $CLOUD_LONG [Enterprise plan][pricing-plan-features] is Health Insurance Portability and Accountability Act +(HIPAA) compliant. This allows organizations to securely manage and analyze sensitive healthcare data, ensuring they +meet regulatory requirements while building compliant applications. [timescale-privacy-policy]: https://www.timescale.com/legal/privacy [tsc-tos]: https://www.timescale.com/legal/timescale-cloud-terms-of-service @@ -47,4 +106,4 @@ communication with your VPC. To learn how to set up VPC Peering, see [vpc-peering]: /use-timescale/:currentVersion:/security/vpc [security-at-timescale]: https://www.timescale.com/security [ip-allowlist]: /use-timescale/:currentVersion:/security/ip-allow-list/ -[pricing-plan-features]: /about/:currentVersion:/pricing-and-account-management/#features-included-in-each-plan +[pricing-plan-features]: /about/:currentVersion:/pricing-and-account-management/#features-included-in-each-plan \ No newline at end of file From c4d8f967631d5c410565fb7a9ca384e37a676ec2 Mon Sep 17 00:00:00 2001 From: atovpeko Date: Tue, 14 Oct 2025 14:10:56 +0300 Subject: [PATCH 5/6] turn tabs into notes --- _partials/_not-supported-for-azure.md | 6 +++- _partials/_prometheus-integrate.md | 3 ++ integrations/aws.md | 11 ++------ integrations/cloudwatch.md | 14 ++-------- integrations/corporate-data-center.md | 12 +------- integrations/datadog.md | 13 ++------- integrations/google-cloud.md | 10 +------ integrations/microsoft-azure.md | 10 +------ integrations/prometheus.md | 13 --------- .../data-tiering/about-data-tiering.md | 19 +++---------- .../data-tiering/enabling-data-tiering.md | 28 +++++-------------- use-timescale/data-tiering/index.md | 2 +- .../data-tiering/querying-tiered-data.md | 13 ++------- .../tiered-data-replicas-forks.md | 18 ++---------- .../metrics-logging/aws-cloudwatch.md | 14 ++-------- use-timescale/metrics-logging/datadog.md | 14 ++-------- use-timescale/metrics-logging/index.md | 14 ---------- .../metrics-logging/metrics-to-prometheus.md | 15 ---------- use-timescale/security/transit-gateway.md | 18 ++---------- use-timescale/security/vpc.md | 15 ++-------- 20 files changed, 43 insertions(+), 219 deletions(-) diff --git a/_partials/_not-supported-for-azure.md b/_partials/_not-supported-for-azure.md index 066f410e4f..cc108fd357 100644 --- a/_partials/_not-supported-for-azure.md +++ b/_partials/_not-supported-for-azure.md @@ -1 +1,5 @@ -This feature is on our roadmap for $CLOUD_LONG on Microsoft Azure. Stay tuned! \ No newline at end of file + + +This feature is on our roadmap for $CLOUD_LONG on Microsoft Azure. Stay tuned! + + \ No newline at end of file diff --git a/_partials/_prometheus-integrate.md b/_partials/_prometheus-integrate.md index 62307cdb28..d3699dc46b 100644 --- a/_partials/_prometheus-integrate.md +++ b/_partials/_prometheus-integrate.md @@ -1,4 +1,5 @@ import IntegrationPrereqs from "versionContent/_partials/_integration-prereqs.mdx"; +import NotSupportedAzure from "versionContent/_partials/_not-supported-for-azure.mdx"; [Prometheus][prometheus] is an open-source monitoring system with a dimensional data model, flexible query language, and a modern alerting approach. @@ -20,6 +21,8 @@ To follow the steps on this page: - [Install Postgres Exporter][install-exporter]. To reduce latency and potential data transfer costs, install Prometheus and Postgres Exporter on a machine in the same AWS region as your $SERVICE_LONG. + + ## Export $SERVICE_LONG telemetry to Prometheus To export your data, do the following: diff --git a/integrations/aws.md b/integrations/aws.md index 5adf048a09..97378275e8 100644 --- a/integrations/aws.md +++ b/integrations/aws.md @@ -12,9 +12,6 @@ import NotSupportedAzure from "versionContent/_partials/_not-supported-for-azure # Integrate Amazon Web Services with $CLOUD_LONG - - - [Amazon Web Services (AWS)][aws] is a comprehensive cloud computing platform that provides on-demand infrastructure, storage, databases, AI, analytics, and security services to help businesses build, deploy, and scale applications in the cloud. @@ -26,6 +23,8 @@ This page explains how to integrate your AWS infrastructure with $CLOUD_LONG usi - Set up [AWS Transit Gateway][gtw-setup]. + + ## Connect your AWS infrastructure to your $SERVICE_LONGs To connect to $CLOUD_LONG: @@ -38,15 +37,9 @@ To connect to $CLOUD_LONG: You have successfully integrated your AWS infrastructure with $CLOUD_LONG. - - - - - - [aws]: https://aws.amazon.com/ diff --git a/integrations/cloudwatch.md b/integrations/cloudwatch.md index 15517d79f7..795c6a9e27 100644 --- a/integrations/cloudwatch.md +++ b/integrations/cloudwatch.md @@ -13,11 +13,6 @@ import ManageDataExporter from "versionContent/_partials/_manage-a-data-exporter # Integrate Amazon CloudWatch with $CLOUD_LONG - - - - - [Amazon CloudWatch][cloudwatch] is a monitoring and observability service designed to help collect, analyze, and act on data from applications, infrastructure, and services running in AWS and on-premises environments. You can export telemetry data from your $SERVICE_LONGs with the time-series and analytics capability enabled to CloudWatch. The available metrics include CPU usage, RAM usage, and storage. This integration is available for [Scale and Enterprise][pricing-plan-features] pricing tiers. @@ -30,6 +25,8 @@ This pages explains how to export telemetry data from your $SERVICE_LONG into Cl - Sign up for [Amazon CloudWatch][cloudwatch-signup]. + + ## Create a data exporter A $CLOUD_LONG data exporter sends telemetry data from a $SERVICE_LONG to a third-party monitoring @@ -39,15 +36,8 @@ tool. You create an exporter on the [project level][projects], in the same AWS r - - - - - - - [projects]: /use-timescale/:currentVersion:/security/members/ [pricing-plan-features]: /about/:currentVersion:/pricing-and-account-management/#features-included-in-each-plan diff --git a/integrations/corporate-data-center.md b/integrations/corporate-data-center.md index b36f1c4fd8..72e3aaa8c5 100644 --- a/integrations/corporate-data-center.md +++ b/integrations/corporate-data-center.md @@ -12,10 +12,6 @@ import NotSupportedAzure from "versionContent/_partials/_not-supported-for-azure # Integrate your data center with $CLOUD_LONG - - - - This page explains how to integrate your corporate on-premise infrastructure with $CLOUD_LONG using [AWS Transit Gateway][aws-transit-gateway]. ## Prerequisites @@ -23,6 +19,7 @@ This page explains how to integrate your corporate on-premise infrastructure wit - Set up [AWS Transit Gateway][gtw-setup]. + ## Connect your on-premise infrastructure to your $SERVICE_LONGs @@ -40,15 +37,8 @@ To connect to $CLOUD_LONG: You have successfully integrated your corporate data center with $CLOUD_LONG. - - - - - - - [aws-transit-gateway]: https://aws.amazon.com/transit-gateway/ diff --git a/integrations/datadog.md b/integrations/datadog.md index 835b3c1225..c4f050f3ea 100644 --- a/integrations/datadog.md +++ b/integrations/datadog.md @@ -13,10 +13,6 @@ import NotSupportedAzure from "versionContent/_partials/_not-supported-for-azure # Integrate Datadog with $CLOUD_LONG - - - - [Datadog][datadog] is a cloud-based monitoring and analytics platform that provides comprehensive visibility into applications, infrastructure, and systems through real-time monitoring, logging, and analytics. @@ -41,6 +37,8 @@ This page explains how to: - Install [Datadog Agent][datadog-agent-install]. + + ## Monitor $SERVICE_LONG metrics with Datadog Export telemetry data from your $SERVICE_LONGs with the time-series and analytics capability enabled to @@ -137,15 +135,8 @@ metrics about your $SERVICE_LONGs. Metrics for your $SERVICE_LONG are now visible in Datadog. Check the Datadog $PG integration documentation for a comprehensive list of [metrics][datadog-postgres-metrics] collected. - - - - - - - [datadog]: https://www.datadoghq.com/ [datadog-agent-install]: https://docs.datadoghq.com/getting_started/agent/#installation diff --git a/integrations/google-cloud.md b/integrations/google-cloud.md index 079dcfab12..d221232704 100644 --- a/integrations/google-cloud.md +++ b/integrations/google-cloud.md @@ -12,10 +12,6 @@ import NotSupportedAzure from "versionContent/_partials/_not-supported-for-azure # Integrate Google Cloud with $CLOUD_LONG - - - - [Google Cloud][google-cloud] is a suite of cloud computing services, offering scalable infrastructure, AI, analytics, databases, security, and developer tools to help businesses build, deploy, and manage applications. This page explains how to integrate your Google Cloud infrastructure with $CLOUD_LONG using [AWS Transit Gateway][aws-transit-gateway]. @@ -25,6 +21,7 @@ This page explains how to integrate your Google Cloud infrastructure with $CLOUD - Set up [AWS Transit Gateway][gtw-setup]. + ## Connect your Google Cloud infrastructure to your $SERVICE_LONGs @@ -42,15 +39,10 @@ To connect to $CLOUD_LONG: You have successfully integrated your Google Cloud infrastructure with $CLOUD_LONG. - - - - - [google-cloud]: https://cloud.google.com/?hl=en diff --git a/integrations/microsoft-azure.md b/integrations/microsoft-azure.md index 413206b1bd..95b67cd3c7 100644 --- a/integrations/microsoft-azure.md +++ b/integrations/microsoft-azure.md @@ -12,9 +12,6 @@ import NotSupportedAzure from "versionContent/_partials/_not-supported-for-azure # Integrate Microsoft Azure with $CLOUD_LONG - - - [Microsoft Azure][azure] is a cloud computing platform and services suite, offering infrastructure, AI, analytics, security, and developer tools to help businesses build, deploy, and manage applications. @@ -25,6 +22,7 @@ This page explains how to integrate your Microsoft Azure infrastructure with $CL - Set up [AWS Transit Gateway][gtw-setup]. + ## Connect your Microsoft Azure infrastructure to your $SERVICE_LONGs @@ -42,15 +40,9 @@ To connect to $CLOUD_LONG: You have successfully integrated your Microsoft Azure infrastructure with $CLOUD_LONG. - - - - - - [aws-transit-gateway]: https://aws.amazon.com/transit-gateway/ [gtw-setup]: https://docs.aws.amazon.com/vpc/latest/tgw/tgw-getting-started.html diff --git a/integrations/prometheus.md b/integrations/prometheus.md index ccb4367a0c..05f4dbed4f 100644 --- a/integrations/prometheus.md +++ b/integrations/prometheus.md @@ -11,18 +11,5 @@ import NotSupportedAzure from "versionContent/_partials/_not-supported-for-azure # Integrate Prometheus with $CLOUD_LONG - - - - - - - - - - - - - diff --git a/use-timescale/data-tiering/about-data-tiering.md b/use-timescale/data-tiering/about-data-tiering.md index 0dd7e8f2f1..939f67e9c1 100644 --- a/use-timescale/data-tiering/about-data-tiering.md +++ b/use-timescale/data-tiering/about-data-tiering.md @@ -2,7 +2,6 @@ title: About Tiger storage tiers excerpt: Learn how Tiger helps you save on storage costs. The tiered storage architecture includes a high-performance storage tier and a low-cost object storage tier built on Amazon s3 products: [cloud] -price_plans: [scale, enterprise] keywords: [tiered storage] tags: [storage, data management] cloud_ui: @@ -15,10 +14,6 @@ import NotSupportedAzure from "versionContent/_partials/_not-supported-for-azure # About storage tiers - - - - The tiered storage architecture in $CLOUD_LONG includes a high-performance storage tier and a low-cost object storage tier. You use the high-performance tier for data that requires quick access, and the object tier for rarely used historical data. Tiering policies move older data asynchronously and periodically from high-performance to low-cost storage, sparing you the need to do it manually. Chunks from a single hypertable, including compressed chunks, can stretch across these two storage tiers. ![$CLOUD_LONG tiered storage](https://assets.timescale.com/docs/images/timescale-tiered-storage-architecture.png) @@ -36,8 +31,12 @@ $CLOUD_LONG high-performance storage comes in the following types: ## Low-cost storage + + Once you [enable tiered storage][manage-tiering], you can start moving rarely used data to the object tier. The object tier is based on AWS S3 and stores your data in the [Apache Parquet][parquet] format. Within a Parquet file, a set of rows is grouped together to form a row group. Within a row group, values for a single column across multiple rows are stored together. The original size of the data in your $SERVICE_SHORT, compressed or uncompressed, does not correspond directly to its size in S3. A compressed hypertable may even take more space in S3 than it does in $CLOUD_LONG. + + Apache Parquet allows for more efficient scans across longer time periods, and $CLOUD_LONG uses other metadata and query optimizations to reduce the amount of data that needs to be fetched to satisfy a query, such as: - **Chunk skipping**: exclude the chunks that fall outside the query time window. @@ -126,16 +125,6 @@ The low-cost storage tier comes with the following limitations: partitioned on more than one dimension. Make sure your hypertables are partitioned on time only, before you enable tiered storage. - - - - - - - - - - [blog-data-tiering]: https://www.timescale.com/blog/expanding-the-boundaries-of-postgresql-announcing-a-bottomless-consumption-based-object-storage-layer-built-on-amazon-s3/ [querying-tiered-data]: /use-timescale/:currentVersion:/data-tiering/querying-tiered-data/ diff --git a/use-timescale/data-tiering/enabling-data-tiering.md b/use-timescale/data-tiering/enabling-data-tiering.md index 229f715a31..b46e8fa1c2 100644 --- a/use-timescale/data-tiering/enabling-data-tiering.md +++ b/use-timescale/data-tiering/enabling-data-tiering.md @@ -15,10 +15,6 @@ import NotSupportedAzure from "versionContent/_partials/_not-supported-for-azure # Manage storage and tiering - - - - The tiered storage architecture in $CLOUD_LONG includes a high-performance storage tier and a low-cost object storage tier: - You use [high-performance storage][high-performance-storage] to store and query frequently accessed data. @@ -59,7 +55,11 @@ This storage type gives you up to 16 TB of storage and is available under [all $ -This storage type gives you up to 64 TB and 32,000 IOPS, and is available under the [$ENTERPRISE $PRICING_PLAN][pricing-plans]. To get enhanced storage: +This storage type gives you up to 64 TB and 32,000 IOPS, and is available under the [$ENTERPRISE $PRICING_PLAN][pricing-plans]. + + + +To get enhanced storage: @@ -92,6 +92,8 @@ You change from enhanced storage to standard in the same way. If you are using o You enable the low-cost object storage tier in $CONSOLE and then tier the data with policies or manually. + + ### Enable tiered storage You enable tiered storage from the `Overview` tab in $CONSOLE. @@ -108,12 +110,6 @@ You enable tiered storage from the `Overview` tab in $CONSOLE. - - -Data tiering is available in [Scale and Enterprise][pricing-plans] pricing plans only. - - - ### Automate tiering with policies A tiering policy automatically moves any chunks that only contain data @@ -291,16 +287,6 @@ If you no longer want to use tiered storage for a particular hypertable, drop th - - - - - - - - - - [data-retention]: /use-timescale/:currentVersion:/data-retention/ [console]: https://console.cloud.timescale.com/dashboard/services diff --git a/use-timescale/data-tiering/index.md b/use-timescale/data-tiering/index.md index e627b83649..17284d8a4e 100644 --- a/use-timescale/data-tiering/index.md +++ b/use-timescale/data-tiering/index.md @@ -2,7 +2,7 @@ title: Storage on Tiger Cloud excerpt: Save on storage costs by tiering older data to a low-cost bottomless object storage tier. Tiger Cloud tiered storage makes sure you cut costs while having data available for analytical queries products: [cloud] -price_plans: [scale, enterprise] +price_plans: [scale, enterprise, performance] keywords: [tiered storage] tags: [storage, data management] --- diff --git a/use-timescale/data-tiering/querying-tiered-data.md b/use-timescale/data-tiering/querying-tiered-data.md index b6a1710a78..1cc781d5ad 100644 --- a/use-timescale/data-tiering/querying-tiered-data.md +++ b/use-timescale/data-tiering/querying-tiered-data.md @@ -11,10 +11,6 @@ import NotSupportedAzure from "versionContent/_partials/_not-supported-for-azure # Querying tiered data - - - - Once rarely used data is tiered and migrated to the object storage tier, it can still be queried with standard SQL by enabling the `timescaledb.enable_tiered_reads` GUC. By default, the GUC is set to `false`, so that queries do not touch tiered data. @@ -30,6 +26,8 @@ Your hypertable is spread across the tiers, so queries and `JOIN`s work and fetc By default, tiered data is not accessed by queries. Querying tiered data may slow down query performance as the data is not stored locally on the high-performance storage tier. See [Performance considerations](#performance-considerations). + + ## Enable querying tiered data for a single query @@ -192,12 +190,5 @@ Queries over tiered data are expected to be slower than over local data. However * Text and non-native types (JSON, JSONB, GIS) filtering is slower when querying tiered data. - - - - - - - diff --git a/use-timescale/data-tiering/tiered-data-replicas-forks.md b/use-timescale/data-tiering/tiered-data-replicas-forks.md index 18d853126c..4c1491074b 100644 --- a/use-timescale/data-tiering/tiered-data-replicas-forks.md +++ b/use-timescale/data-tiering/tiered-data-replicas-forks.md @@ -9,11 +9,7 @@ tags: [storage, data management] import NotSupportedAzure from "versionContent/_partials/_not-supported-for-azure.mdx"; -# How tiered data works on replicas and forks - - - - +# How tiered data works on replicas and forks There is one more thing that makes Tiered Storage even more amazing: when you keep data in the low-cost object storage tier, you pay for this data only once, regardless of whether you have a [high-availability replica][ha-replica] @@ -25,6 +21,8 @@ When creating one (or more) forks, you won't be billed for data shared with the If you decide to tier more data that's not in the primary, you will pay to store it in the low-cost tier, but you will still see substantial savings by moving that data from the high-performance tier of the fork to the cheaper object storage tier. + + ## How this works behind the scenes Once you tier data to the low-cost object storage tier, we keep a reference to that data on your Database's catalog. @@ -74,16 +72,6 @@ In the case of such a restore, new references are added to the deleted tiered ch Once 14 days pass after soft deleting the data,that is the number of references to the tiered data drop to 0, we hard delete the tiered data. - - - - - - - - - - [ha-replica]: /use-timescale/:currentVersion:/ha-replicas/high-availability/ [read-replica]: /use-timescale/:currentVersion:/ha-replicas/read-scaling/#read-replicas diff --git a/use-timescale/metrics-logging/aws-cloudwatch.md b/use-timescale/metrics-logging/aws-cloudwatch.md index 7d7d36f214..eb4c3b3b1a 100644 --- a/use-timescale/metrics-logging/aws-cloudwatch.md +++ b/use-timescale/metrics-logging/aws-cloudwatch.md @@ -14,10 +14,6 @@ import NotSupportedAzure from "versionContent/_partials/_not-supported-for-azure # Export telemetry data to AWS Cloudwatch - - - - You can export telemetry data from your $SERVICE_LONGs with the time-series and analytics capability enabled to [Amazon CloudWatch][cloudwatch]. Available metrics include CPU usage, RAM usage, and storage. This integration is available for [Scale or Enterprise][pricing-plan-features] pricing plans. @@ -27,6 +23,8 @@ This page shows you how to create an Amazon CloudWatch exporter in $CONSOLE, and + + ## Create a data exporter $CLOUD_LONG data exporters send telemetry data from a $SERVICE_LONG to a third-party monitoring @@ -40,15 +38,7 @@ This section shows you how to attach, monitor, edit, and delete a data exporter. - - - - - - - - [cloudwatch]: https://aws.amazon.com/cloudwatch/ [cloudwatch-docs]: https://docs.aws.amazon.com/cloudwatch/index.html diff --git a/use-timescale/metrics-logging/datadog.md b/use-timescale/metrics-logging/datadog.md index cadcf55a30..c01409c7c8 100644 --- a/use-timescale/metrics-logging/datadog.md +++ b/use-timescale/metrics-logging/datadog.md @@ -14,10 +14,6 @@ import NotSupportedAzure from "versionContent/_partials/_not-supported-for-azure # Export telemetry data to Datadog - - - - You can export telemetry data from your $SERVICE_LONGs with the time-series and analytics capability enabled to [Datadog][datadog]. The available metrics include CPU usage, RAM usage, and storage. This integration is available for [Scale or Enterprise][pricing-plan-features] pricing plans. This page shows you how to create a Datadog exporter in $CONSOLE, and manage the lifecycle of data exporters. @@ -26,6 +22,8 @@ This page shows you how to create a Datadog exporter in $CONSOLE, and manage the + + ## Create a data exporter $CLOUD_LONG data exporters send telemetry data from a $SERVICE_LONG to third-party monitoring @@ -39,15 +37,7 @@ This section shows you how to attach, monitor, edit, and delete a data exporter. - - - - - - - - [datadog]: https://www.datadoghq.com [datadog-api-key]: https://docs.datadoghq.com/account_management/api-app-keys/#add-an-api-key-or-client-token diff --git a/use-timescale/metrics-logging/index.md b/use-timescale/metrics-logging/index.md index 8dffac3656..1d5be16953 100644 --- a/use-timescale/metrics-logging/index.md +++ b/use-timescale/metrics-logging/index.md @@ -12,10 +12,6 @@ cloud_ui: # Metrics and logging - - - - Find metrics and logs for your $SERVICE_SHORTs in $CONSOLE, or integrate with third-party monitoring services: * [Monitor][monitor] your $SERVICE_SHORTs in $CONSOLE. @@ -23,16 +19,6 @@ Find metrics and logs for your $SERVICE_SHORTs in $CONSOLE, or integrate with th * Export metrics to [Amazon Cloudwatch][cloudwatch]. * Export metrics to [Prometheus][prometheus]. - - - - -[Monitor][monitor] your $SERVICE_SHORTs in $CONSOLE. - - - - - [prometheus]: /use-timescale/:currentVersion:/metrics-logging/metrics-to-prometheus/ [datadog]: /use-timescale/:currentVersion:/metrics-logging/datadog/ diff --git a/use-timescale/metrics-logging/metrics-to-prometheus.md b/use-timescale/metrics-logging/metrics-to-prometheus.md index 09c17475a0..66d0ccafee 100644 --- a/use-timescale/metrics-logging/metrics-to-prometheus.md +++ b/use-timescale/metrics-logging/metrics-to-prometheus.md @@ -12,22 +12,7 @@ cloud_ui: --- import PrometheusIntegrate from "versionContent/_partials/_prometheus-integrate.mdx"; -import NotSupportedAzure from "versionContent/_partials/_not-supported-for-azure.mdx"; # Export metrics to Prometheus - - - - - - - - - - - - - - diff --git a/use-timescale/security/transit-gateway.md b/use-timescale/security/transit-gateway.md index 613bc0aa2c..845c430ca4 100644 --- a/use-timescale/security/transit-gateway.md +++ b/use-timescale/security/transit-gateway.md @@ -15,12 +15,10 @@ import NotSupportedAzure from "versionContent/_partials/_not-supported-for-azure # Securely connect to $CLOUD_LONG using AWS Transit Gateway - - - - [AWS Transit Gateway][aws-transit-gateway] enables you to securely connect to your $CLOUD_LONG from AWS, Google Cloud, Microsoft Azure, or any other cloud or on-premise environment. + + You use AWS Transit Gateway as a traffic controller for your network. Instead of setting up multiple direct connections to different clouds, on-premise data centers, and other AWS services, you connect everything to AWS Transit Gateway. This simplifies your network and makes it easier to manage and scale. You can then create a peering connection between your $SERVICE_LONGs and AWS Transit Gateway in $CLOUD_LONG. This means that, no matter how big or complex your infrastructure is, you can connect securely to your $SERVICE_LONGs. @@ -92,17 +90,7 @@ AWS Transit Gateway enables you to connect from almost any environment, this pag -You can now securely access your $SERVICE_SHORTs in $CLOUD_LONG. - - - - - - - - - - +You can now securely access your $SERVICE_SHORTs in $CLOUD_LONG. [aws-transit-gateway]: https://aws.amazon.com/transit-gateway/ [pricing-plans]: /about/:currentVersion:/pricing-and-account-management/ diff --git a/use-timescale/security/vpc.md b/use-timescale/security/vpc.md index 943e0e99e7..09d2263285 100644 --- a/use-timescale/security/vpc.md +++ b/use-timescale/security/vpc.md @@ -14,14 +14,12 @@ import NotSupportedAzure from "versionContent/_partials/_not-supported-for-azure # Secure your $CLOUD_LONG services with $VPC Peering and AWS PrivateLink - - - - You use Virtual Private Cloud ($VPC) peering to ensure that your $SERVICE_LONGs are only accessible through your secured AWS infrastructure. This reduces the potential attack vector surface and improves security. + + The data isolation architecture that ensures a highly secure connection between your apps and $CLOUD_LONG is: @@ -220,15 +218,6 @@ Migration takes a few minutes to complete and requires a change to DNS settings $SERVICE_SHORT. The $SERVICE_SHORT is not accessible during this time. If you receive a DNS error, allow some time for DNS propagation. - - - - - - - - - [aws-dashboard]: https://console.aws.amazon.com/vpc/home#PeeringConnections: [aws-security-groups]: https://console.aws.amazon.com/vpcconsole/home#securityGroups: From cd1b69fe7b5700e58d3226c78a7f1f313f11643f Mon Sep 17 00:00:00 2001 From: atovpeko Date: Tue, 14 Oct 2025 14:13:08 +0300 Subject: [PATCH 6/6] turn tabs into notes --- use-timescale/data-tiering/index.md | 10 +++++----- 1 file changed, 5 insertions(+), 5 deletions(-) diff --git a/use-timescale/data-tiering/index.md b/use-timescale/data-tiering/index.md index 17284d8a4e..001960e480 100644 --- a/use-timescale/data-tiering/index.md +++ b/use-timescale/data-tiering/index.md @@ -1,6 +1,6 @@ --- -title: Storage on Tiger Cloud -excerpt: Save on storage costs by tiering older data to a low-cost bottomless object storage tier. Tiger Cloud tiered storage makes sure you cut costs while having data available for analytical queries +title: Storage on Tiger +excerpt: Save on storage costs by tiering older data to a low-cost bottomless object storage tier. Tiger tiered storage makes sure you cut costs while having data available for analytical queries products: [cloud] price_plans: [scale, enterprise, performance] keywords: [tiered storage] @@ -9,9 +9,9 @@ tags: [storage, data management] # Storage - + - + Tiered storage is a [hierarchical storage management architecture][hierarchical-storage] for [real-time analytics][create-service] $SERVICE_SHORTs you create in [$CLOUD_LONG](https://console.cloud.timescale.com/). @@ -49,7 +49,7 @@ In this section, you: - + $CLOUD_LONG stores your data in high-performance storage optimized for frequent querying. Based on [AWS EBS gp3][aws-gp3], the high-performance storage provides you with up to 16 TB and 16,000 IOPS. Its [$HYPERCORE row-columnar storage engine][hypercore], designed specifically for real-time analytics, enables you to compress your data by up to 98%, while improving performance.