|
| 1 | +## 1.格式: |
| 2 | + |
| 3 | +``` |
| 4 | +CREATE TABLE tableName( |
| 5 | + colName colType, |
| 6 | + ... |
| 7 | + )WITH( |
| 8 | + type ='file', |
| 9 | + format = 'csv', |
| 10 | + fieldDelimiter = ',', |
| 11 | + fileName = 'xxxx', |
| 12 | + filePath = 'xx/xxx', |
| 13 | + location = 'local', |
| 14 | + nullLiteral = 'null', |
| 15 | + allowComment = 'true', |
| 16 | + arrayElementDelimiter = ',', |
| 17 | + quoteCharacter = '"', |
| 18 | + escapeCharacter = '\\', |
| 19 | + ignoreParseErrors = 'true', |
| 20 | + hdfsSite = 'xxx/hdfs-site.xml', |
| 21 | + coreSite = 'xxx/core-site.xml', |
| 22 | + hdfsUser = 'root', |
| 23 | + charsetName = 'UTF-8' |
| 24 | + ); |
| 25 | +``` |
| 26 | + |
| 27 | +## 2.支持的格式 |
| 28 | + |
| 29 | +支持 HDFS、 Local 支持 Csv、Json、Arvo 格式文件 |
| 30 | + |
| 31 | +## 3.表结构定义 |
| 32 | + |
| 33 | +|参数名称|含义| |
| 34 | +|----|---| |
| 35 | +| tableName | 在 sql 中使用的名称;即注册到flink-table-env上的名称| |
| 36 | +| colName | 列名称| |
| 37 | +| colType | 列类型 [colType支持的类型](../colType.md)| |
| 38 | + |
| 39 | +## 4.参数 |
| 40 | + |
| 41 | +通用参数设置 |
| 42 | + |
| 43 | +|参数名称|默认值|是否必填|参数说明| |
| 44 | +|----|---|---|---| |
| 45 | +|type|file|是|当前表的类型| |
| 46 | +|format|csv|是|文件格式,仅支持csv,json,Arvo类型| |
| 47 | +|fileName|无|是|文件名| |
| 48 | +|filePath|无|是|文件绝对路径| |
| 49 | +|location|local|是|文件存储介质,仅支持HDFS、Local| |
| 50 | +|charsetName|UTF-8|否|文件编码格式| |
| 51 | + |
| 52 | +### 4.1 Csv 参数设置 |
| 53 | + |
| 54 | +|参数名称|默认值|是否必填|参数说明| |
| 55 | +|----|---|---|---| |
| 56 | +|ignoreParseErrors|true|否|是否忽略解析失败的数据| |
| 57 | +|fieldDelimiter|,|否|csv数据的分割符| |
| 58 | +|nullLiteral|"null"|否|填充csv数据中的null值| |
| 59 | +|allowComments|true|否|| |
| 60 | +|arrayElementDelimiter|,|否|| |
| 61 | +|quoteCharacter|"|否|| |
| 62 | +|escapeCharacter|\|否|| |
| 63 | + |
| 64 | +### 4.2 Arvo 参数说明 |
| 65 | + |
| 66 | +|参数名称|默认值|是否必填|参数说明| |
| 67 | +|----|---|---|---| |
| 68 | +|avroFormat|无|是|在format = 'arvo'的情况下是必填项| |
| 69 | + |
| 70 | +### 4.3 HDFS 参数说明 |
| 71 | + |
| 72 | +|参数名称|默认值|是否必填|参数说明| |
| 73 | +|----|---|---|---| |
| 74 | +|hdfsSite|${HADOOP_CONF_HOME}/hdfs-site.xml|是|hdfs-site.xml所在位置| |
| 75 | +|coreSite|${HADOOP_CONF_HOME}/core-site.xml|是|core-site.xml所在位置| |
| 76 | +|hdfsUser|root|否|HDFS访问用户,默认是[root]用户| |
| 77 | + |
| 78 | +### 4.4 Json 参数说明 |
| 79 | + |
| 80 | +Json无特殊参数 |
| 81 | + |
| 82 | +## 5.样例 |
| 83 | + |
| 84 | +数据展示: |
| 85 | + |
| 86 | +csv |
| 87 | + |
| 88 | +```csv |
| 89 | +712382,1/1/2017 0:00,1/1/2017 0:03,223,7051,Wellesley St E / Yonge St Green P,7089,Church St / Wood St,Member |
| 90 | +``` |
| 91 | + |
| 92 | +json |
| 93 | + |
| 94 | +```json |
| 95 | +{ |
| 96 | + "trip_id": "712382", |
| 97 | + "trip_start_time": "1/1/2017 0:00", |
| 98 | + "trip_stop_time": "1/1/2017 0:03", |
| 99 | + "trip_duration_seconds": "223", |
| 100 | + "from_station_id": "7051", |
| 101 | + "from_station_name": "Wellesley St E / Yonge St Green P", |
| 102 | + "to_station_id": "7089", |
| 103 | + "to_station_name": "Church St / Wood St", |
| 104 | + "user_type": "Member" |
| 105 | +}, |
| 106 | + |
| 107 | +``` |
| 108 | + |
| 109 | +### 5.1 csv |
| 110 | + |
| 111 | +```sql |
| 112 | +CREATE TABLE SourceOne |
| 113 | +( |
| 114 | + trip_id varchar, |
| 115 | + trip_start_time varchar, |
| 116 | + trip_stop_time varchar, |
| 117 | + trip_duration_seconds varchar, |
| 118 | + from_station_id varchar, |
| 119 | + from_station_name varchar, |
| 120 | + to_station_id varchar, |
| 121 | + to_station_name varchar, |
| 122 | + user_type varchar |
| 123 | +) WITH ( |
| 124 | + type = 'file', |
| 125 | + format = 'csv', |
| 126 | + fieldDelimiter = ',', |
| 127 | + fileName = '2017-Q1.csv', |
| 128 | + filePath = '/data', |
| 129 | + location = 'local', |
| 130 | + charsetName = 'UTF-8' |
| 131 | + ); |
| 132 | +``` |
| 133 | + |
| 134 | +### 5.2 json |
| 135 | + |
| 136 | +```sql |
| 137 | +CREATE TABLE SourceOne |
| 138 | +( |
| 139 | + trip_id varchar, |
| 140 | + trip_start_time varchar, |
| 141 | + trip_stop_time varchar, |
| 142 | + trip_duration_seconds varchar, |
| 143 | + from_station_id varchar, |
| 144 | + from_station_name varchar, |
| 145 | + to_station_id varchar, |
| 146 | + to_station_name varchar, |
| 147 | + user_type varchar |
| 148 | +) WITH ( |
| 149 | + type = 'file', |
| 150 | + format = 'json', |
| 151 | + fieldDelimiter = ',', |
| 152 | + fileName = '2017-Q1.json', |
| 153 | + filePath = '/data', |
| 154 | + charsetName = 'UTF-8' |
| 155 | + ); |
| 156 | +``` |
| 157 | + |
| 158 | +### 5.3 HDFS |
| 159 | + |
| 160 | +```sql |
| 161 | +CREATE TABLE SourceOne |
| 162 | +( |
| 163 | + trip_id varchar, |
| 164 | + trip_start_time varchar, |
| 165 | + trip_stop_time varchar, |
| 166 | + trip_duration_seconds varchar, |
| 167 | + from_station_id varchar, |
| 168 | + from_station_name varchar, |
| 169 | + to_station_id varchar, |
| 170 | + to_station_name varchar, |
| 171 | + user_type varchar |
| 172 | +) WITH ( |
| 173 | + type = 'file', |
| 174 | + format = 'json', |
| 175 | + fieldDelimiter = ',', |
| 176 | + fileName = '2017-Q1.json', |
| 177 | + filePath = 'hdfs://ns1/data', |
| 178 | + location = 'hdfs', |
| 179 | + hdfsSite = '/Users/wtz/dtstack/conf/yarn/kudu1/hdfs-site.xml', |
| 180 | + coreSite = '/Users/wtz/dtstack/conf/yarn/kudu1/core-site.xml', |
| 181 | + hdfsUser = 'admin', |
| 182 | + charsetName = 'UTF-8' |
| 183 | + ); |
| 184 | +``` |
0 commit comments