Skip to content

Commit

Permalink
[DOCS] Fix deadlinks in docs (#1220)
Browse files Browse the repository at this point in the history
* Fix image urls

* Fix the sql script download link

* Add Snowflake info

* Fix indention of Snowflake tutorial
  • Loading branch information
jiayuasu committed Jan 28, 2024
1 parent 644a261 commit 4cadcd7
Show file tree
Hide file tree
Showing 4 changed files with 22 additions and 19 deletions.
7 changes: 5 additions & 2 deletions docs/setup/overview.md
Original file line number Diff line number Diff line change
Expand Up @@ -13,6 +13,7 @@
- [x] Spatial DataFrame/SQL on Spark
- [x] Spatial DataStream on Flink
- [x] Spatial Table/SQL on Flink
- [x] Spatial SQL on Snowflake

## Complex spatial objects

Expand All @@ -28,6 +29,8 @@
## Rich spatial analytics tools

- [x] Coordinate Reference System / Spatial Reference System Transformation
- [x] High resolution map generation: [Visualize Spatial DataFrame/RDD](../../tutorial/viz)
- [x] Apache Zeppelin integration
- [x] Apache Zeppelin dashboard integration
- [X] Integrate with a variety of Python tools including Jupyter notebook, GeoPandas, Shapely
- [X] Integrate with a variety of visualization tools including KeplerGL, DeckGL
- [x] High resolution and scalable map generation: [Visualize Spatial DataFrame/RDD](../../tutorial/viz)
- [x] Support Scala, Java, Python, R
18 changes: 9 additions & 9 deletions docs/setup/snowflake/install.md
Original file line number Diff line number Diff line change
Expand Up @@ -20,11 +20,11 @@ A stage is a Snowflake object that maps to a location in a cloud storage provide

In this case, we will create a stage named `ApacheSedona` in the `public` schema of the database created in the previous step. The stage will be used to load Sedona's JAR files into the database. We will choose a `Snowflake managed` stage.

![Create Stage](../../../../image/snowflake/snowflake-1.png)
<img src="./../../../image/snowflake/snowflake-1.png">

After creating the stage, you should be able to see the stage in the database.

![Stage Details](../../../../image/snowflake/snowflake-2.png)
<img src="./../../../image/snowflake/snowflake-2.png">

You can refer to [Snowflake Documentation](https://docs.snowflake.com/en/sql-reference/sql/create-stage.html) to how to create a stage.

Expand All @@ -39,7 +39,7 @@ Then you can upload the 2 JAR files to the stage created in the previous step.

After uploading the 2 JAR files, you should be able to see the 2 JAR files in the stage.

![JAR Details](../../../../image/snowflake/snowflake-3.png)
<img src="./../../../image/snowflake/snowflake-3.png">

You can refer to [Snowflake Documentation](https://docs.snowflake.com/en/sql-reference/sql/put.html) to how to upload files to a stage.

Expand All @@ -49,17 +49,17 @@ A schema is a Snowflake object that maps to a database. You can use a schema to

In this case, we will create a schema named `SEDONA` in the database created in the previous step. The schema will be used to create Sedona's functions.

![Create Schema](../../../../image/snowflake/snowflake-4.png)
<img src="./../../../image/snowflake/snowflake-4.png">

You can find your schema in the database as follows:

![Schema Details](../../../../image/snowflake/snowflake-5.png)
<img src="./../../../image/snowflake/snowflake-5.png">

You can refer to [Snowflake Documentation](https://docs.snowflake.com/en/sql-reference/sql/create-schema.html) to how to create a schema.

## Step 4: Get the SQL script for creating Sedona's functions

You will need to download [sedona-snowflake.sql](../../../../image/snowflake/sedona-snowflake.sql) to create Sedona's functions in the schema created in the previous step.
You will need to download [sedona-snowflake.sql](./../../../image/snowflake/sedona-snowflake.sql) to create Sedona's functions in the schema created in the previous step.

You can also get this SQL script by running the following command:

Expand All @@ -75,11 +75,11 @@ We will create a worksheet in the database created in the previous step, and run

In this case, we will choose the option `Create Worksheet from SQL File`.

![Create Worksheet](../../../../image/snowflake/snowflake-6.png)
<img src="./../../../image/snowflake/snowflake-6.png">

In the worksheet, choose `SEDONA_TEST` as the database, and `PUBLIC` as the schema. The SQL script should be in the worksheet. Then right click the worksheet and choose `Run All`. Snowflake will take 3 minutes to create Sedona's functions.

![Worksheet](../../../../image/snowflake/snowflake-7.png)
<img src="./../../../image/snowflake/snowflake-7.png">

## Step 6: Verify the installation

Expand All @@ -97,4 +97,4 @@ SRID=4326;POINT (1 2)

The worksheet should look like this:

![Worksheet](../../../../image/snowflake/snowflake-8.png)
<img src="./../../../image/snowflake/snowflake-8.png">
14 changes: 7 additions & 7 deletions docs/tutorial/snowflake/sql.md
Original file line number Diff line number Diff line change
Expand Up @@ -424,11 +424,11 @@ SRID=0;POINT(1 2)
1. Sedona Snowflake doesn't support `M` dimension due to the limitation of WKB serialization. Sedona Spark and Sedona Flink support XYZM because it uses our in-house serialization format. Although Sedona Snowflake has functions related to `M` dimension, all `M` values will be ignored.
2. Sedona H3 functions are not supported because Snowflake does not allow embedded C code in UDF.
3. All User Defined Table Functions only work with geometries created by Sedona constructors due to Snowflake current limitation `Data type GEOMETRY is not supported in non-SQL UDTF return type`. This includes:
* ST_MinimumBoundingRadius
* ST_Intersection_Aggr
* ST_SubDivideExplode
* ST_Envelope_Aggr
* ST_Union_Aggr
* ST_Collect
* ST_Dump
* ST_MinimumBoundingRadius
* ST_Intersection_Aggr
* ST_SubDivideExplode
* ST_Envelope_Aggr
* ST_Union_Aggr
* ST_Collect
* ST_Dump
4. Only Sedona ST functions are available in Snowflake. Raster functions (RS functions) are not available in Snowflake yet.
2 changes: 1 addition & 1 deletion mkdocs.yml
Original file line number Diff line number Diff line change
@@ -1,5 +1,5 @@
site_name: Apache Sedona&trade;
site_description: Apache Sedona&trade; is a cluster computing system for processing large-scale spatial data. Sedona extends existing cluster computing systems, such as Apache Spark and Apache Flink, with a set of out-of-the-box distributed Spatial Datasets and Spatial SQL that efficiently load, process, and analyze large-scale spatial data across machines.
site_description: Apache Sedona&trade; is a cluster computing system for processing large-scale spatial data. Sedona extends existing cluster computing systems, such as Apache Spark, Apache Flink, and Snowflake, with a set of out-of-the-box distributed Spatial Datasets and Spatial SQL that efficiently load, process, and analyze large-scale spatial data across machines.
nav:
- Home: index.md
- Setup:
Expand Down

0 comments on commit 4cadcd7

Please sign in to comment.