SHOW:
|
|
- or go back to the newest paste.
1 | - | { |
1 | + | { |
2 | - | "convert_data_parquet": { |
2 | + | "convert_data_parquet": { |
3 | - | "desc": "测试", |
3 | + | "desc": "测试", |
4 | - | "strategy": "streaming.core.strategy.SparkStreamingStrategy", |
4 | + | "strategy": "streaming.core.strategy.SparkStreamingStrategy", |
5 | - | "algorithm": [], |
5 | + | "algorithm": [], |
6 | - | "ref": [], |
6 | + | "ref": [], |
7 | - | "compositor": [ |
7 | + | "compositor": [ |
8 | - | { |
8 | + | { |
9 | - | "name": "streaming.core.compositor.spark.source.HDFSSourceCompositor", |
9 | + | "name": "streaming.core.compositor.spark.source.HDFSSourceCompositor", |
10 | - | "params": [ |
10 | + | "params": [ |
11 | - | { |
11 | + | { |
12 | - | "path":"hdfs://....." |
12 | + | "path":"hdfs://....." |
13 | - | } |
13 | + | } |
14 | - | ] |
14 | + | ] |
15 | - | }, |
15 | + | }, |
16 | - | { |
16 | + | { |
17 | - | "name": "streaming.core.compositor.spark.transformation.SingleColumnJSONCompositor", |
17 | + | "name": "streaming.core.compositor.spark.transformation.SingleColumnJSONCompositor", |
18 | - | "params": [ |
18 | + | "params": [ |
19 | - | { |
19 | + | { |
20 | - | "name": "raw" |
20 | + | "name": "raw" |
21 | - | } |
21 | + | } |
22 | - | ] |
22 | + | ] |
23 | - | }, |
23 | + | }, |
24 | - | { |
24 | + | { |
25 | - | "name": "streaming.core.compositor.spark.transformation.JSONTableCompositor", |
25 | + | "name": "streaming.core.compositor.spark.transformation.JSONTableCompositor", |
26 | - | "params": [ |
26 | + | "params": [ |
27 | - | { |
27 | + | { |
28 | - | "tableName": "test" |
28 | + | "tableName": "test" |
29 | - | } |
29 | + | } |
30 | - | ] |
30 | + | ] |
31 | - | }, |
31 | + | }, |
32 | - | { |
32 | + | { |
33 | - | "name": "streaming.core.compositor.spark.transformation.SQLCompositor", |
33 | + | "name": "streaming.core.compositor.spark.transformation.SQLCompositor", |
34 | - | "params": [ |
34 | + | "params": [ |
35 | - | { |
35 | + | { |
36 | - | "sql": "select split(raw,'\\007|\\001')[2] from test limit 100" |
36 | + | "sql": "select split(raw,'\\007|\\001')[2] from test limit 100" |
37 | - | } |
37 | + | } |
38 | - | ] |
38 | + | ] |
39 | - | }, |
39 | + | }, |
40 | - | { |
40 | + | { |
41 | - | "name": "streaming.core.compositor.spark.output.SQLOutputCompositor", |
41 | + | "name": "streaming.core.compositor.spark.output.SQLOutputCompositor", |
42 | - | "params": [ |
42 | + | "params": [ |
43 | - | { |
43 | + | { |
44 | - | "format": "parquet", |
44 | + | "format": "parquet", |
45 | - | "path":"", |
45 | + | "path":"", |
46 | - | "mode": "" |
46 | + | "mode": "" |
47 | - | } |
47 | + | } |
48 | - | ] |
48 | + | ] |
49 | - | } |
49 | + | } |
50 | - | ], |
50 | + | ], |
51 | - | "configParams": { |
51 | + | "configParams": { |
52 | - | } |
52 | + | } |
53 | - | } |
53 | + | } |
54 | } |