Skip to content
Permalink
Browse files
upd topic & xref titles for new PXF, Managing Data section names
  • Loading branch information
lisakowen committed Oct 27, 2016
1 parent 01f3f8e commit 20882b73b24f7a90788d5c5156a1959621c8288d
Showing 6 changed files with 8 additions and 8 deletions.
@@ -274,7 +274,7 @@ Also, make sure that your `CREATE EXTERNAL TABLE` definition has the correct hos

## <a id="usingpxf"></a>Using PXF

HAWQ Extension Framework \(PXF\) is an extensible framework that allows HAWQ to query external system data. The details of how to install and use PXF can be found in [Working with PXF and External Data](../pxf/HawqExtensionFrameworkPXF.html).
HAWQ Extension Framework \(PXF\) is an extensible framework that allows HAWQ to query external system data. The details of how to install and use PXF can be found in [Using PXF with Unmanaged Data](../pxf/HawqExtensionFrameworkPXF.html).

### <a id="usingpxftobackupthetpchdatabase"></a>Using PXF to Back Up the tpch Database

@@ -1,5 +1,5 @@
---
title: Managing Data
title: Managing Data with HAWQ
---

This chapter provides information about manipulating data and concurrent access in HAWQ.
@@ -24,7 +24,7 @@ This chapter provides information about manipulating data and concurrent access

The topics in this section describe methods for loading and writing data into and out of HAWQ, and how to format data files.

- **[Working with PXF and External Data](../pxf/HawqExtensionFrameworkPXF.html)**
- **[Using PXF with Unmanaged Data](../pxf/HawqExtensionFrameworkPXF.html)**

HAWQ Extension Framework (PXF) is an extensible framework that allows HAWQ to query external system data. 

@@ -32,8 +32,8 @@ HAWQ can access data in external files using the HAWQ Extension Framework (PXF).
PXF is an extensible framework that allows HAWQ to access data in external
sources as readable or writable HAWQ tables. PXF has built-in connectors for
accessing data inside HDFS files, Hive tables, and HBase tables. PXF also
integrates with HCatalog to query Hive tables directly. See [Working with PXF
and External Data](../pxf/HawqExtensionFrameworkPXF.html) for more
integrates with HCatalog to query Hive tables directly. See [Using PXF
with Unmanaged Data](../pxf/HawqExtensionFrameworkPXF.html) for more
details.

Users can create custom PXF connectors to access other parallel data stores or
@@ -1,5 +1,5 @@
---
title: Working with PXF and External Data
title: Using PXF with Unmanaged Data
---

HAWQ Extension Framework (PXF) is an extensible framework that allows HAWQ to query external system data. 
@@ -3148,7 +3148,7 @@ The estimated cost for vacuuming a buffer that has to be read from disk. This re

Specifies the cutoff age (in transactions) that `VACUUM` should use to decide whether to replace transaction IDs with *FrozenXID* while scanning a table.

For information about `VACUUM` and transaction ID management, see [Managing Data](../../datamgmt/dml.html#topic1) and the [PostgreSQL documentation](http://www.postgresql.org/docs/8.2/static/routine-vacuuming.html#VACUUM-FOR-WRAPAROUND).
For information about `VACUUM` and transaction ID management, see [Managing Data with HAWQ](../../datamgmt/dml.html#topic1) and the [PostgreSQL documentation](http://www.postgresql.org/docs/8.2/static/routine-vacuuming.html#VACUUM-FOR-WRAPAROUND).

| Value Range | Default | Set Classifications |
|------------------------|-----------|------------------------|
@@ -164,7 +164,7 @@ For writable external tables, specifies the URI location of the `gpfdist` proces

With two `gpfdist` locations listed as in the above example, half of the segments would send their output data to the `data1.out` file and the other half to the `data2.out` file.

For the `pxf` protocol, the `LOCATION` string specifies the \<host\> and \<port\> of the PXF service, the location of the data, and the PXF plug-ins (Java classes) used to convert the data between storage format and HAWQ format. If the \<port\> is omitted, the \<host\> is taken to be the logical name for the high availability name service and the \<port\> is the value of the `pxf_service_port` configuration variable, 51200 by default. The URL parameters `FRAGMENTER`, `ACCESSOR`, and `RESOLVER` are the names of PXF plug-ins (Java classes) that convert between the external data format and HAWQ data format. The `FRAGMENTER` parameter is only used with readable external tables. PXF allows combinations of these parameters to be configured as profiles so that a single `PROFILE` parameter can be specified to access external data, for example `?PROFILE=Hive`. Additional \<custom-options\>` can be added to the LOCATION URI to further describe the external data format or storage options. For details about the plug-ins and profiles provided with PXF and information about creating custom plug-ins for other data sources see [Working with PXF and External Data](../../pxf/HawqExtensionFrameworkPXF.html).</dd>
For the `pxf` protocol, the `LOCATION` string specifies the \<host\> and \<port\> of the PXF service, the location of the data, and the PXF plug-ins (Java classes) used to convert the data between storage format and HAWQ format. If the \<port\> is omitted, the \<host\> is taken to be the logical name for the high availability name service and the \<port\> is the value of the `pxf_service_port` configuration variable, 51200 by default. The URL parameters `FRAGMENTER`, `ACCESSOR`, and `RESOLVER` are the names of PXF plug-ins (Java classes) that convert between the external data format and HAWQ data format. The `FRAGMENTER` parameter is only used with readable external tables. PXF allows combinations of these parameters to be configured as profiles so that a single `PROFILE` parameter can be specified to access external data, for example `?PROFILE=Hive`. Additional \<custom-options\>` can be added to the LOCATION URI to further describe the external data format or storage options. For details about the plug-ins and profiles provided with PXF and information about creating custom plug-ins for other data sources see [Using PXF with Unmanaged Data](../../pxf/HawqExtensionFrameworkPXF.html).</dd>

<dt>EXECUTE '\<command\>' ON ... </dt>
<dd>Allowed for readable web external tables or writable external tables only. For readable web external tables, specifies the OS command to be executed by the segment instances. The \<command\> can be a single OS command or a script. If \<command\> executes a script, that script must reside in the same location on all of the segment hosts and be executable by the HAWQ superuser (`gpadmin`).

0 comments on commit 20882b7

Please sign in to comment.