Skip to content
Permalink
Browse files
Merge branch 'feature/remove-gp' of https://github.com/lisakowen/incu…
…bator-hawq-docs into develop
  • Loading branch information
dyozie committed Oct 19, 2016
2 parents 19c676a + 1f999f7 commit f6e8b8a23380e5fd42931a15e11e633f6df42015
Show file tree
Hide file tree
Showing 11 changed files with 9 additions and 9 deletions.
@@ -4,5 +4,5 @@ title: Supported Client Applications

Users can connect to HAWQ using various client applications:

- A number of [HAWQ Client Applications](g-greenplum-database-client-applications.html) are provided with your HAWQ installation. The `psql` client application provides an interactive command-line interface to HAWQ.
- A number of [HAWQ Client Applications](g-hawq-database-client-applications.html) are provided with your HAWQ installation. The `psql` client application provides an interactive command-line interface to HAWQ.
- Using standard ODBC/JDBC Application Interfaces, such as ODBC and JDBC, users can connect their client applications to HAWQ.
@@ -11,7 +11,7 @@ This section explains how to configure client connections and authentication for
* <a class="subnav" href="./roles_privs.html">Managing Roles and Privileges</a>
* <a class="subnav" href="./g-establishing-a-database-session.html">Establishing a Database Session</a>
* <a class="subnav" href="./g-supported-client-applications.html">Supported Client Applications</a>
* <a class="subnav" href="./g-greenplum-database-client-applications.html">HAWQ Client Applications</a>
* <a class="subnav" href="./g-hawq-database-client-applications.html">HAWQ Client Applications</a>
* <a class="subnav" href="./g-connecting-with-psql.html">Connecting with psql</a>
* <a class="subnav" href="./g-database-application-interfaces.html">Database Application Interfaces</a>
* <a class="subnav" href="./g-troubleshooting-connection-problems.html">Troubleshooting Connection Problems</a>
@@ -11,6 +11,6 @@ Use the `CREATE WRITABLE EXTERNAL TABLE` command to define the external table an

You can optionally declare a distribution policy for your writable external tables. By default, writable external tables use a random distribution policy. If the source table you are exporting data from has a hash distribution policy, defining the same distribution key column(s) for the writable external table improves unload performance by eliminating the requirement to move rows over the interconnect. If you unload data from a particular table, you can use the `LIKE` clause to copy the column definitions and distribution policy from the source table.

- **[Example - HAWQ file server (gpfdist)](../../datamgmt/load/g-example-greenplum-file-server-gpfdist.html)**
- **[Example - HAWQ file server (gpfdist)](../../datamgmt/load/g-example-hawq-file-server-gpfdist.html)**


@@ -8,7 +8,7 @@ The `gpfdist://` protocol is used in a URI to reference a running `gpfdist` inst

Run `gpfdist` on the host where the external data files reside. `gpfdist` uncompresses `gzip` (`.gz`) and `bzip2` (.`bz2`) files automatically. You can use the wildcard character (\*) or other C-style pattern matching to denote multiple files to read. The files specified are assumed to be relative to the directory that you specified when you started the `gpfdist` instance.

All virtual segments access the external file(s) in parallel, subject to the number of segments set in the `gp_external_max_segments` parameter, the length of the `gpfdist` location list, and the limits specified by the `hawq_rm_nvseg_perquery_limit` and `hawq_rm_nvseg_perquery_perseg_limit` parameters. Use multiple `gpfdist` data sources in a `CREATE EXTERNAL TABLE` statement to scale the external table's scan performance. For more information about configuring `gpfdist`, see [Using the Greenplum Parallel File Server (gpfdist)](g-using-the-greenplum-parallel-file-server--gpfdist-.html#topic13).
All virtual segments access the external file(s) in parallel, subject to the number of segments set in the `gp_external_max_segments` parameter, the length of the `gpfdist` location list, and the limits specified by the `hawq_rm_nvseg_perquery_limit` and `hawq_rm_nvseg_perquery_perseg_limit` parameters. Use multiple `gpfdist` data sources in a `CREATE EXTERNAL TABLE` statement to scale the external table's scan performance. For more information about configuring `gpfdist`, see [Using the HAWQ File Server (gpfdist)](g-using-the-hawq-file-server--gpfdist-.html#topic13).

See the `gpfdist` reference documentation for more information about using `gpfdist` with external tables.

@@ -2,6 +2,6 @@
title: Installing gpfdist
---

`gpfdist` is installed in `$GPHOME/bin` of your HAWQ master host installation. Run `gpfdist` from a machine other than the HAWQ master, such as on a machine devoted to ETL processing. If you want to install `gpfdist` on your ETL server, get it from the *Greenplum Load Tools* package and follow its installation instructions.
You may choose to run `gpfdist` from a machine other than the HAWQ master, such as on a machine devoted to ETL processing. To install `gpfdist` on your ETL server, refer to [Client-Based HAWQ Load Tools](client-loadtools.html) for information related to Linux and Windows load tools installation and configuration.


@@ -24,7 +24,7 @@ HAWQ leverages the parallel architecture of the Hadoop Distributed File System t

- **[Working with File-Based External Tables](../../datamgmt/load/g-working-with-file-based-ext-tables.html)**

- **[Using the Greenplum Parallel File Server (gpfdist)](../../datamgmt/load/g-using-the-greenplum-parallel-file-server--gpfdist-.html)**
- **[Using the HAWQ File Server (gpfdist)](../../datamgmt/load/g-using-the-hawq-file-server--gpfdist-.html)**

- **[Creating and Using Web External Tables](../../datamgmt/load/g-creating-and-using-web-external-tables.html)**

@@ -46,7 +46,7 @@ HAWQ leverages the parallel architecture of the Hadoop Distributed File System t

- **[Optimizing Data Load and Query Performance](../../datamgmt/load/g-optimizing-data-load-and-query-performance.html)**

- **[Unloading Data from HAWQ](../../datamgmt/load/g-unloading-data-from-greenplum-database.html)**
- **[Unloading Data from HAWQ](../../datamgmt/load/g-unloading-data-from-hawq-database.html)**

- **[Transforming XML Data](../../datamgmt/load/g-transforming-xml-data.html)**

@@ -1,5 +1,5 @@
---
title: Using the Greenplum Parallel File Server (gpfdist)
title: Using the HAWQ File Server (gpfdist)
---

The `gpfdist` protocol provides the best performance and is the easiest to set up. `gpfdist` ensures optimum use of all segments in your HAWQ system for external table reads.
@@ -119,7 +119,7 @@ Regular readable external tables can access static flat files or, by using HAWQ

Web external tables access dynamic data sources – either on a web server or by executing OS commands or scripts.

The LOCATION clause specifies the location of the external data. The location string begins with a protocol string that specifies the storage type and protocol used to access the data. The `gpfdist://` protocol specifies data files served by one or more instances of the Greenplum parallel file distribution server `gpfdist`. The `http://` protocol specifies one or more HTTP URLs and is used with web tables. The `pxf://` protocol specifies data accessed through the PXF service, which provides access to data in a Hadoop system. Using the PXF API, you can create PXF plug-ins to provide HAWQ access to any other data source.
The LOCATION clause specifies the location of the external data. The location string begins with a protocol string that specifies the storage type and protocol used to access the data. The `gpfdist://` protocol specifies data files served by one or more instances of the HAWQ file server `gpfdist`. The `http://` protocol specifies one or more HTTP URLs and is used with web tables. The `pxf://` protocol specifies data accessed through the PXF service, which provides access to data in a Hadoop system. Using the PXF API, you can create PXF plug-ins to provide HAWQ access to any other data source.

**Note:** The `file://` protocol is deprecated. Instead, use the `gpfdist://`, `gpfdists://`, or `pxf://` protocol, or the `COPY` command instead.

0 comments on commit f6e8b8a

Please sign in to comment.