Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[DOCS] Fix deadlinks in docs #1220

Merged
merged 4 commits into from
Jan 28, 2024
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Jump to
Jump to file
Failed to load files.
Diff view
Diff view
7 changes: 5 additions & 2 deletions docs/setup/overview.md
Original file line number Diff line number Diff line change
Expand Up @@ -13,6 +13,7 @@
- [x] Spatial DataFrame/SQL on Spark
- [x] Spatial DataStream on Flink
- [x] Spatial Table/SQL on Flink
- [x] Spatial SQL on Snowflake

## Complex spatial objects

Expand All @@ -28,6 +29,8 @@
## Rich spatial analytics tools

- [x] Coordinate Reference System / Spatial Reference System Transformation
- [x] High resolution map generation: [Visualize Spatial DataFrame/RDD](../../tutorial/viz)
- [x] Apache Zeppelin integration
- [x] Apache Zeppelin dashboard integration
- [X] Integrate with a variety of Python tools including Jupyter notebook, GeoPandas, Shapely
- [X] Integrate with a variety of visualization tools including KeplerGL, DeckGL
- [x] High resolution and scalable map generation: [Visualize Spatial DataFrame/RDD](../../tutorial/viz)
- [x] Support Scala, Java, Python, R
18 changes: 9 additions & 9 deletions docs/setup/snowflake/install.md
Original file line number Diff line number Diff line change
Expand Up @@ -20,11 +20,11 @@ A stage is a Snowflake object that maps to a location in a cloud storage provide

In this case, we will create a stage named `ApacheSedona` in the `public` schema of the database created in the previous step. The stage will be used to load Sedona's JAR files into the database. We will choose a `Snowflake managed` stage.

![Create Stage](../../../../image/snowflake/snowflake-1.png)
<img src="./../../../image/snowflake/snowflake-1.png">

After creating the stage, you should be able to see the stage in the database.

![Stage Details](../../../../image/snowflake/snowflake-2.png)
<img src="./../../../image/snowflake/snowflake-2.png">

You can refer to [Snowflake Documentation](https://docs.snowflake.com/en/sql-reference/sql/create-stage.html) to how to create a stage.

Expand All @@ -39,7 +39,7 @@ Then you can upload the 2 JAR files to the stage created in the previous step.

After uploading the 2 JAR files, you should be able to see the 2 JAR files in the stage.

![JAR Details](../../../../image/snowflake/snowflake-3.png)
<img src="./../../../image/snowflake/snowflake-3.png">

You can refer to [Snowflake Documentation](https://docs.snowflake.com/en/sql-reference/sql/put.html) to how to upload files to a stage.

Expand All @@ -49,17 +49,17 @@ A schema is a Snowflake object that maps to a database. You can use a schema to

In this case, we will create a schema named `SEDONA` in the database created in the previous step. The schema will be used to create Sedona's functions.

![Create Schema](../../../../image/snowflake/snowflake-4.png)
<img src="./../../../image/snowflake/snowflake-4.png">

You can find your schema in the database as follows:

![Schema Details](../../../../image/snowflake/snowflake-5.png)
<img src="./../../../image/snowflake/snowflake-5.png">

You can refer to [Snowflake Documentation](https://docs.snowflake.com/en/sql-reference/sql/create-schema.html) to how to create a schema.

## Step 4: Get the SQL script for creating Sedona's functions

You will need to download [sedona-snowflake.sql](../../../../image/snowflake/sedona-snowflake.sql) to create Sedona's functions in the schema created in the previous step.
You will need to download [sedona-snowflake.sql](./../../../image/snowflake/sedona-snowflake.sql) to create Sedona's functions in the schema created in the previous step.

You can also get this SQL script by running the following command:

Expand All @@ -75,11 +75,11 @@ We will create a worksheet in the database created in the previous step, and run

In this case, we will choose the option `Create Worksheet from SQL File`.

![Create Worksheet](../../../../image/snowflake/snowflake-6.png)
<img src="./../../../image/snowflake/snowflake-6.png">

In the worksheet, choose `SEDONA_TEST` as the database, and `PUBLIC` as the schema. The SQL script should be in the worksheet. Then right click the worksheet and choose `Run All`. Snowflake will take 3 minutes to create Sedona's functions.

![Worksheet](../../../../image/snowflake/snowflake-7.png)
<img src="./../../../image/snowflake/snowflake-7.png">

## Step 6: Verify the installation

Expand All @@ -97,4 +97,4 @@ SRID=4326;POINT (1 2)

The worksheet should look like this:

![Worksheet](../../../../image/snowflake/snowflake-8.png)
<img src="./../../../image/snowflake/snowflake-8.png">
14 changes: 7 additions & 7 deletions docs/tutorial/snowflake/sql.md
Original file line number Diff line number Diff line change
Expand Up @@ -424,11 +424,11 @@ SRID=0;POINT(1 2)
1. Sedona Snowflake doesn't support `M` dimension due to the limitation of WKB serialization. Sedona Spark and Sedona Flink support XYZM because it uses our in-house serialization format. Although Sedona Snowflake has functions related to `M` dimension, all `M` values will be ignored.
2. Sedona H3 functions are not supported because Snowflake does not allow embedded C code in UDF.
3. All User Defined Table Functions only work with geometries created by Sedona constructors due to Snowflake current limitation `Data type GEOMETRY is not supported in non-SQL UDTF return type`. This includes:
* ST_MinimumBoundingRadius
* ST_Intersection_Aggr
* ST_SubDivideExplode
* ST_Envelope_Aggr
* ST_Union_Aggr
* ST_Collect
* ST_Dump
* ST_MinimumBoundingRadius
* ST_Intersection_Aggr
* ST_SubDivideExplode
* ST_Envelope_Aggr
* ST_Union_Aggr
* ST_Collect
* ST_Dump
4. Only Sedona ST functions are available in Snowflake. Raster functions (RS functions) are not available in Snowflake yet.
2 changes: 1 addition & 1 deletion mkdocs.yml
Original file line number Diff line number Diff line change
@@ -1,5 +1,5 @@
site_name: Apache Sedona&trade;
site_description: Apache Sedona&trade; is a cluster computing system for processing large-scale spatial data. Sedona extends existing cluster computing systems, such as Apache Spark and Apache Flink, with a set of out-of-the-box distributed Spatial Datasets and Spatial SQL that efficiently load, process, and analyze large-scale spatial data across machines.
site_description: Apache Sedona&trade; is a cluster computing system for processing large-scale spatial data. Sedona extends existing cluster computing systems, such as Apache Spark, Apache Flink, and Snowflake, with a set of out-of-the-box distributed Spatial Datasets and Spatial SQL that efficiently load, process, and analyze large-scale spatial data across machines.
nav:
- Home: index.md
- Setup:
Expand Down