You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/relational-databases/polybase/polybase-guide.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -57,7 +57,7 @@ PolyBase provides these same functionalities for the following SQL products from
57
57
| Some connectors separate from PolyBase services | The S3-compatible object storage connector, as well as ADSL Gen2, and Azure Blob Storage, are no longer dependent of PolyBase services. PolyBase services must still run to support connectivity with Oracle, Teradata, MongoDB, and Generic ODBC. The PolyBase feature must still be installed on your SQL Server instance. |
58
58
| Parquet file format | PolyBase is now capable of querying data from Parquet files stored on S3-compatible object storage. For more information, see to [Virtualize parquet file in a S3-compatible object storage with PolyBase](polybase-virtualize-parquet-file.md). |
59
59
|Delta table format | PolyBase is now capable of querying data from Delta Table format stored on S3-compatible object storage, Azure Storage Account V2, and Azure Data Lake Storage Gen2. For more information, see to [Virtualize Delta Table format](virtualize-delta.md)|
60
-
|Create External Table as Select (CETAS) | PolyBase can now use CETAS to create an external table and then export, in parallel, the result of a [!INCLUDE[tsql](../../includes/tsql-md.md)] SELECT statement to Azure Data Lake Storage Gen2, Azure Storage Account V2, and S3-compatible object storage. For more information, see [CREATE EXTERNAL TABLE AS SELECT](../../t-sql/statements/create-external-table-as-select-sql.md)|
60
+
|Create External Table as Select (CETAS) | PolyBase can now use CETAS to create an external table and then export, in parallel, the result of a [!INCLUDE[tsql](../../includes/tsql-md.md)] SELECT statement to Azure Data Lake Storage Gen2, Azure Storage Account V2, and S3-compatible object storage. For more information, see [CREATE EXTERNAL TABLE AS SELECT (Transact-SQL)](../../t-sql/statements/create-external-table-as-select-transact-sql.md). |
61
61
62
62
For more new features of [!INCLUDE[sssql22-md](../../includes/sssql22-md.md)], see [What's new in SQL Server 2022?](../../sql-server/what-s-new-in-sql-server-2022.md)
Specifies a text format with column delimiters, also called field terminators.
@@ -171,38 +171,42 @@ Specifies the format of the external data.
171
171
Specifies the data compression method for the external data. When DATA_COMPRESSION isn't specified, the default is uncompressed data.
172
172
To work properly, Gzip compressed files must have the ".gz" file extension.
173
173
174
-
#### [Delimited text](#tab/delimited)
174
+
#### [Delimited text](#tab/delimited)
175
175
The DELIMITEDTEXT format type supports these compression methods:
176
-
- DATA COMPRESSION = 'org.apache.hadoop.io.compress.DefaultCodec'
177
-
- DATA COMPRESSION = 'org.apache.hadoop.io.compress.GzipCodec'
176
+
- DATA COMPRESSION = `org.apache.hadoop.io.compress.DefaultCodec`
177
+
- DATA COMPRESSION = `org.apache.hadoop.io.compress.GzipCodec`
178
178
179
179
#### [RC](#tab/rc)
180
180
The RCFILE format type supports this compression method:
181
-
- DATA COMPRESSION = 'org.apache.hadoop.io.compress.DefaultCodec'
181
+
- DATA COMPRESSION = `org.apache.hadoop.io.compress.DefaultCodec`
182
182
183
183
#### [ORC](#tab/orc)
184
184
The ORC file format type supports these compression methods:
185
-
- DATA COMPRESSION = 'org.apache.hadoop.io.compress.DefaultCodec'
186
-
- DATA COMPRESSION = 'org.apache.hadoop.io.compress.SnappyCodec'
185
+
- DATA COMPRESSION = `org.apache.hadoop.io.compress.DefaultCodec`
186
+
- DATA COMPRESSION = `org.apache.hadoop.io.compress.SnappyCodec`
187
187
188
188
#### [Parquet](#tab/parquet)
189
189
The PARQUET file format type supports the following compression methods:
190
-
- DATA COMPRESSION = 'org.apache.hadoop.io.compress.GzipCodec'
191
-
- DATA COMPRESSION = 'org.apache.hadoop.io.compress.SnappyCodec'
190
+
- DATA COMPRESSION = `org.apache.hadoop.io.compress.GzipCodec`
191
+
- DATA COMPRESSION = `org.apache.hadoop.io.compress.SnappyCodec`
192
192
193
193
#### [JSON](#tab/json)
194
194
The JSON file format type supports the following compression methods:
195
-
- DATA COMPRESSION = 'org.apache.hadoop.io.compress.GzipCodec'
196
-
- DATA COMPRESSION = 'org.apache.hadoop.io.compress.SnappyCodec'
197
-
- DATA COMPRESSION = 'org.apache.hadoop.io.compress.DefaultCodec'
195
+
- DATA COMPRESSION = `org.apache.hadoop.io.compress.GzipCodec`
196
+
- DATA COMPRESSION = `org.apache.hadoop.io.compress.SnappyCodec`
197
+
- DATA COMPRESSION = `org.apache.hadoop.io.compress.DefaultCodec`
198
+
199
+
### [Delta table](#tab/delta)
200
+
N/A
198
201
---
202
+
199
203
### Delimited text format options
200
204
201
205
The format options described in this section are optional and only apply to delimited text files.
202
206
203
207
#### FIELD_TERMINATOR
204
208
`FIELD_TERMINATOR = *field_terminator*`
205
-
Applies only to delimited text files. The field terminator specifies one or more characters that mark the end of each field (column) in the text-delimited file. The default is the pipe character ꞌ|ꞌ. For guaranteed support, we recommend using one or more ascii characters.
209
+
Applies only to delimited text files. The field terminator specifies one or more characters that mark the end of each field (column) in the text-delimited file. The default is the pipe character `|`. For guaranteed support, we recommend using one or more ASCII characters.
206
210
207
211
208
212
Examples:
@@ -218,7 +222,7 @@ Examples:
218
222
#### STRING_DELIMITER
219
223
`STRING_DELIMITER = *string_delimiter*`
220
224
221
-
Specifies the field terminator for data of type string in the text-delimited file. The string delimiter is one or more characters in length and is enclosed with single quotes. The default is the empty string "". For guaranteed support, we recommend using one or more ascii characters.
225
+
Specifies the field terminator for data of type string in the text-delimited file. The string delimiter is one or more characters in length and is enclosed with single quotes. The default is the empty string "". For guaranteed support, we recommend using one or more ASCII characters.
222
226
223
227
224
228
Examples:
@@ -389,7 +393,7 @@ WITH (
389
393
```
390
394
391
395
### B. Create an RCFile external file format
392
-
This example creates an external file format for a RCFile that uses the `serialization/deserialization` method `o`rg.apache.hadoop.hive.serde2.columnar.LazyBinaryColumnarSerDe`. It also specifies to use the Default Codec for the data compression method. If DATA_COMPRESSION isn't specified, the default is no compression.
396
+
This example creates an external file format for a RCFile that uses the `serialization/deserialization` method `org.apache.hadoop.hive.serde2.columnar.LazyBinaryColumnarSerDe`. It also specifies to use the Default Codec for the data compression method. If DATA_COMPRESSION isn't specified, the default is no compression.
393
397
394
398
```sql
395
399
CREATE EXTERNAL FILE FORMAT rcfile1
@@ -442,7 +446,7 @@ WITH (FORMAT_TYPE = DELIMITEDTEXT,
442
446
### F. Create a JSON external file format
443
447
**Applies to:** Azure SQL Edge
444
448
445
-
This example creates an external file format for a JSON file that compresses the data with the org.apache.io.compress.SnappyCodec data compression method. If DATA_COMPRESSION isn't specified, the default is no compression. This example applies to Azure SQL Edge and is currently not supported for other SQL products.
449
+
This example creates an external file format for a JSON file that compresses the data with the `org.apache.io.compress.SnappyCodec` data compression method. If DATA_COMPRESSION isn't specified, the default is no compression. This example applies to Azure SQL Edge and is currently not supported for other SQL products.
0 commit comments