SELECT INTO
Contents
SELECT INTO#
SELECT INTO OUTFILE
语句将表的查询结果导出为一个文件。
Note
LOAD DATA INFILE 语句与SELECT INTO OUTFILE
互补,它用于从指定文件创建表以及加载数据到表中。
Syntax#
SelectIntoStmt
::= SelectStmt 'INTO' 'OUTFILE' filePath SelectIntoOptionList
filePath
::= string_literal
SelectIntoOptionList
::= 'OPTIONS' '(' SelectInfoOptionItem (',' SelectInfoOptionItem)* ')'
SelectInfoOptionItem
::= 'DELIMITER' '=' string_literal
|'HEADER' '=' bool_literal
|'NULL_VALUE' '=' string_literal
|'FORMAT' '=' string_literal
|'MODE' '=' string_literal
SELECT INTO OUTFILE
分为三个部分。
第一部分是一个普通的
SELECT
语句,通过这个SELECT
语句来查询所需要的数据;第二部分是
filePath
,定义将查询的记录导出到哪个文件中,详情见FilePath;第三部分是
SelectIntoOptionList
为可选选项,其可能的取值有:
配置项 |
类型 |
默认值 |
描述 |
---|---|---|---|
delimiter |
String |
, |
列分隔符,默认为‘ |
header |
Boolean |
true |
是否包含表头, 默认为 |
null_value |
String |
null |
NULL填充值,默认填充 |
format |
String |
csv |
输出文件格式: |
mode |
String |
error_if_exists |
输出模式: |
quote |
String |
“” |
输出数据的包围字符串,字符串长度<=1。默认为””,表示输出数据包围字符串为空。当配置包围字符串时,将使用包围字符串包围一个field。例如,我们配置包围字符串为 |
Important
请注意,目前仅有集群版支持quote字符的转义。所以,如果您使用的是单机版,请谨慎选择quote字符,保证原始字符串内并不包含quote字符。
SQL语句模版#
SELECT ... INTO OUTFILE 'file_path' OPTIONS (key = value, ...)
FilePath#
FilePath支持’file://’, ‘hdfs://’, ‘hive://’三种。
Hive 支持#
OpenMLDB 支持导出数据到 Hive,但需要额外的设置和功能限制,详情见 Hive 支持。
Examples#
从表
t1
查询输出到data.csv
文件中,使用,
作为列分隔符
SELECT col1, col2, col3 FROM t1 INTO OUTFILE 'data.csv' OPTIONS ( delimiter = ',' );
从表
t1
查询输出到data.csv
文件中,使用|
作为列分隔符,NULL值的填充值为NA
字符串:
SELECT col1, col2, col3 FROM t1 INTO OUTFILE 'data2.csv' OPTIONS ( delimiter = '|', null_value='NA');
导出表格 t1 到 Hive 数据库
SELECT col1, col2, col3 FROM t1 INTO OUTFILE 'hive://db1.t1';
Q&A#
Q: select into 错误 Found duplicate column(s)?
Exception in thread "main" org.apache.spark.sql.AnalysisException: Found duplicate column(s) when inserting into file:/tmp/out: `c1`;
at org.apache.spark.sql.util.SchemaUtils$.checkColumnNameDuplication(SchemaUtils.scala:90)
at org.apache.spark.sql.execution.datasources.InsertIntoHadoopFsRelationCommand.run(InsertIntoHadoopFsRelationCommand.scala:84)
at org.apache.spark.sql.execution.command.DataWritingCommandExec.sideEffectResult$lzycompute(commands.scala:108)
at org.apache.spark.sql.execution.command.DataWritingCommandExec.sideEffectResult(commands.scala:106)
at org.apache.spark.sql.execution.command.DataWritingCommandExec.doExecute(commands.scala:131)
at org.apache.spark.sql.execution.SparkPlan.$anonfun$execute$1(SparkPlan.scala:175)
at org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$1(SparkPlan.scala:213)
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:210)
at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:171)
at org.apache.spark.sql.execution.QueryExecution.toRdd$lzycompute(QueryExecution.scala:122)
at org.apache.spark.sql.execution.QueryExecution.toRdd(QueryExecution.scala:121)
at org.apache.spark.sql.DataFrameWriter.$anonfun$runCommand$1(DataFrameWriter.scala:944)
A: 查询语句是允许列名重复的。但SELECT INTO
除了查询还需要写入,写入中会检查重复列名。请避免重复列名,可以用c1 as c_new
来重命名列。