You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/core_concepts/11_persistent_storage/index.mdx
+262-3Lines changed: 262 additions & 3 deletions
Original file line number
Diff line number
Diff line change
@@ -238,7 +238,7 @@ For best performance, [install MinIO locally](https://min.io/docs/minio/kubernet
238
238
239
239
[MinIO](https://min.io/) is an open-source, high-performance, and scalable object storage server that is compatible with Amazon S3 APIs, designed for building private and public cloud storage solutions.
240
240
241
-
Then from Windmill, just [fill the S3 resource type](../../integrations/s3.md).
241
+
Then from Windmill, just [fill the S3 resource type](../../integrations/s3.mdx).
242
242
243
243
#### Azure Blob
244
244
@@ -254,7 +254,7 @@ Then from Windmill, just [fill the S3 resource type](../../integrations/s3.md).
254
254
255
255
### Connect your Windmill workspace to your S3 bucket or your Azure Blob storage
256
256
257
-
Once you've created an [S3 or Azure Blob resource](../../integrations/s3.md) in Windmill, go to the workspace settings > S3 Storage. Select the resource and click Save.
257
+
Once you've created an [S3 or Azure Blob resource](../../integrations/s3.mdx) in Windmill, go to the workspace settings > S3 Storage. Select the resource and click Save.

414
+
415
+
:::info
416
+
Certain file types, typically parquet files, can be directly rendered by Windmill
417
+
:::
418
+
419
+
For more info on how to use files and S3 files in Windmill, see [Handling files and binary data](/docs/core_concepts/files_binary_data).
420
+
421
+
### Windmill embedded integration with Polars and DuckDB for data pipelines
422
+
423
+
ETLs can be easily implemented in Windmill using its integration with Polars and DuckDB for facilitate working with tabular data. In this case, you don't need to manually interact with the S3 bucket, Polars/DuckDB does it natively and in a efficient way. Reading and Writing datasets to S3 can be done seamlessly.
# Write the result of a query to a different parquet file on S3
514
+
conn.execute(
515
+
"""
516
+
COPY (
517
+
SELECT COUNT(*) FROM read_parquet('{input_uri}')
518
+
) TO '{output_uri}' (FORMAT 'parquet');
519
+
""".format(
520
+
input_uri=input_uri, output_uri=output_uri
521
+
)
522
+
)
523
+
524
+
conn.close()
525
+
return S3Object(s3=output_file)
526
+
```
527
+
528
+
</TabItem>
529
+
</Tabs>
530
+
531
+
:::info
532
+
533
+
Polars and DuckDB need to be configured to access S3 within the Windmill script. The job will need to accessed the S3 resources, which either needs to be accessible to the user running the job, or the S3 resource needs to be [set as public in the workspace settings](/docs/core_concepts/persistent_storage#connect-your-windmill-workspace-to-your-s3-bucket-or-your-azure-blob-storage).
534
+
535
+
:::
536
+
537
+
For more info on how Data Pipelines in Windmill, see [Data Pipelines](../27_data_pipelines/index.mdx).
Copy file name to clipboardExpand all lines: docs/core_concepts/18_files_binary_data/index.mdx
+2-2Lines changed: 2 additions & 2 deletions
Original file line number
Diff line number
Diff line change
@@ -193,7 +193,7 @@ Certain file types, typically parquet files, can be directly rendered by Windmil
193
193
194
194
### Windmill embedded integration with Polars and DuckDB for data pipelines
195
195
196
-
ETL can be easily implemented in Windmill using its integration with Polars and DuckDB for facilitate working with tabular data. In this case, you don't need to manually interact with the S3 bucket, Polars/DuckDB does it natively and in a efficient way. Reading and Writing datasets to S3 can be done seamlessly.
196
+
ETLs can be easily implemented in Windmill using its integration with Polars and DuckDB for facilitate working with tabular data. In this case, you don't need to manually interact with the S3 bucket, Polars/DuckDB does it natively and in a efficient way. Reading and Writing datasets to S3 can be done seamlessly.
Polars and DuckDB needs to be configured to access S3 within the Windmill script. The job will need to accessed the S3 resources, which either needs to be accessible to the user running the job, or the S3 resource needs to be [set as public in the workspace settings](/docs/core_concepts/persistent_storage#connect-your-windmill-workspace-to-your-s3-bucket-or-your-azure-blob-storage).
306
+
Polars and DuckDB need to be configured to access S3 within the Windmill script. The job will need to accessed the S3 resources, which either needs to be accessible to the user running the job, or the S3 resource needs to be [set as public in the workspace settings](/docs/core_concepts/persistent_storage#connect-your-windmill-workspace-to-your-s3-bucket-or-your-azure-blob-storage).
0 commit comments