|
22 | 22 | import org.apache.flink.client.deployment.application.ApplicationConfiguration;
|
23 | 23 | import org.apache.flink.client.program.ClusterClient;
|
24 | 24 | import org.apache.flink.client.program.ClusterClientProvider;
|
| 25 | +import org.apache.flink.client.program.PackagedProgram; |
| 26 | +import org.apache.flink.client.program.PackagedProgramUtils; |
25 | 27 | import org.apache.flink.configuration.Configuration;
|
26 | 28 | import org.apache.flink.configuration.PipelineOptions;
|
27 | 29 | import org.apache.flink.core.fs.FSDataInputStream;
|
28 | 30 | import org.apache.flink.core.fs.FileSystem;
|
29 | 31 | import org.apache.flink.core.fs.Path;
|
| 32 | +import org.apache.flink.runtime.jobgraph.JobGraph; |
| 33 | +import org.apache.flink.runtime.jobgraph.SavepointRestoreSettings; |
30 | 34 | import org.apache.flink.yarn.YarnClusterClientFactory;
|
31 | 35 | import org.apache.flink.yarn.YarnClusterDescriptor;
|
32 | 36 | import org.apache.flink.yarn.configuration.YarnConfigOptions;
|
|
38 | 42 |
|
39 | 43 | import org.apache.commons.cli.CommandLine;
|
40 | 44 | import org.apache.hadoop.yarn.api.records.ApplicationId;
|
| 45 | +import org.apache.hadoop.yarn.api.records.FinalApplicationStatus; |
41 | 46 | import org.slf4j.Logger;
|
42 | 47 | import org.slf4j.LoggerFactory;
|
43 | 48 |
|
44 | 49 | import java.io.File;
|
| 50 | +import java.util.ArrayList; |
45 | 51 | import java.util.Collections;
|
46 | 52 | import java.util.List;
|
47 | 53 | import java.util.stream.Collectors;
|
@@ -87,23 +93,72 @@ public PipelineExecution.ExecutionInfo deploy(
|
87 | 93 |
|
88 | 94 | ClusterClient<ApplicationId> client = null;
|
89 | 95 | try {
|
90 |
| - ClusterClientProvider<ApplicationId> clusterClientProvider = |
91 |
| - descriptor.deploySessionCluster(specification); |
92 |
| - client = clusterClientProvider.getClusterClient(); |
93 |
| - ApplicationId clusterId = client.getClusterId(); |
94 |
| - LOG.info("Deployment Flink CDC From Cluster ID {}", clusterId); |
95 |
| - return new PipelineExecution.ExecutionInfo( |
96 |
| - clusterId.toString(), "submit job successful"); |
| 96 | + // If applicationId is passed, we get the state of yarn; if not, we create a session |
| 97 | + // cluster. |
| 98 | + String applicationId = flinkConfig.get(YarnConfigOptions.APPLICATION_ID); |
| 99 | + if (applicationId != null) { |
| 100 | + FinalApplicationStatus applicationStatus = |
| 101 | + descriptor |
| 102 | + .getYarnClient() |
| 103 | + .getApplicationReport(ApplicationId.fromString(applicationId)) |
| 104 | + .getFinalApplicationStatus(); |
| 105 | + if (FinalApplicationStatus.UNDEFINED.equals(applicationStatus)) { |
| 106 | + // applicationId is running. |
| 107 | + client = |
| 108 | + descriptor |
| 109 | + .retrieve(ApplicationId.fromString(applicationId)) |
| 110 | + .getClusterClient(); |
| 111 | + } |
| 112 | + } else { |
| 113 | + ClusterClientProvider<ApplicationId> clusterClientProvider = |
| 114 | + descriptor.deploySessionCluster(specification); |
| 115 | + client = clusterClientProvider.getClusterClient(); |
| 116 | + applicationId = String.valueOf(client.getClusterId()); |
| 117 | + } |
| 118 | + LOG.info("Deployment Flink CDC From application ID {}", applicationId); |
| 119 | + // how to get jobGraph |
| 120 | + assert client != null; |
| 121 | + client.submitJob(getJobGraph(flinkConfig, 1)); |
| 122 | + |
| 123 | + return new PipelineExecution.ExecutionInfo(applicationId, "submit job successful"); |
97 | 124 | } catch (Exception e) {
|
98 | 125 | if (client != null) {
|
99 | 126 | client.shutDownCluster();
|
100 | 127 | }
|
101 |
| - throw new RuntimeException("Failed to deploy Flink CDC job", e); |
| 128 | + throw new RuntimeException("Failed to yarn session deploy Flink CDC job", e); |
102 | 129 | } finally {
|
103 | 130 | descriptor.close();
|
104 | 131 | if (client != null) {
|
105 | 132 | client.close();
|
106 | 133 | }
|
107 | 134 | }
|
108 | 135 | }
|
| 136 | + |
| 137 | + /** Get jobGraph from configuration. */ |
| 138 | + private JobGraph getJobGraph(Configuration configuration, int parallelism) throws Exception { |
| 139 | + SavepointRestoreSettings savepointRestoreSettings = |
| 140 | + SavepointRestoreSettings.fromConfiguration(configuration); |
| 141 | + PackagedProgram.Builder builder = |
| 142 | + PackagedProgram.newBuilder() |
| 143 | + .setSavepointRestoreSettings(savepointRestoreSettings) |
| 144 | + .setEntryPointClassName( |
| 145 | + configuration |
| 146 | + .getOptional( |
| 147 | + ApplicationConfiguration.APPLICATION_MAIN_CLASS) |
| 148 | + .get()) |
| 149 | + .setArguments( |
| 150 | + configuration |
| 151 | + .getOptional(ApplicationConfiguration.APPLICATION_ARGS) |
| 152 | + .orElse(new ArrayList<>()) |
| 153 | + .toArray(new String[] {})) |
| 154 | + .setJarFile( |
| 155 | + new File( |
| 156 | + configuration |
| 157 | + .getOptional(PipelineOptions.JARS) |
| 158 | + .orElse(new ArrayList<>()) |
| 159 | + .get(0))); |
| 160 | + PackagedProgram program = builder.build(); |
| 161 | + return PackagedProgramUtils.createJobGraph( |
| 162 | + program, configuration, parallelism, null, false); |
| 163 | + } |
109 | 164 | }
|
0 commit comments