Codota Logo
ExecutionConfig.isSysoutLoggingEnabled
Code IndexAdd Codota to your IDE (free)

How to use
isSysoutLoggingEnabled
method
in
org.apache.flink.api.common.ExecutionConfig

Best Java code snippets using org.apache.flink.api.common.ExecutionConfig.isSysoutLoggingEnabled (Showing top 20 results out of 315)

  • Common ways to obtain ExecutionConfig
private void myMethod () {
ExecutionConfig e =
  • Codota Iconnew ExecutionConfig()
  • Codota IconStreamExecutionEnvironment env;env.getConfig().disableSysoutLogging()
  • Codota IconExecutionEnvironment env;env.getConfig().disableSysoutLogging()
  • Smart code suggestions by Codota
}
origin: apache/flink

@Override
protected PlanExecutor getExecutor() throws Exception {
  // check if we had already started a PlanExecutor. If true, then stop it, because there will
  // be a new jar file available for the user code classes
  if (this.executor != null) {
    this.executor.stop();
  }
  // write generated classes to disk so that they can be shipped to the cluster
  URL jarUrl = flinkILoop.writeFilesToDisk().getAbsoluteFile().toURI().toURL();
  List<URL> allJarFiles = new ArrayList<>(jarFiles);
  allJarFiles.add(jarUrl);
  this.executor = PlanExecutor.createRemoteExecutor(
    host,
    port,
    clientConfiguration,
    allJarFiles,
    globalClasspaths
  );
  executor.setPrintStatusDuringExecution(getConfig().isSysoutLoggingEnabled());
  return executor;
}
origin: apache/flink

protected PlanExecutor getExecutor() throws Exception {
  if (executor == null) {
    executor = PlanExecutor.createRemoteExecutor(host, port, clientConfiguration,
      jarFiles, globalClasspaths);
    executor.setPrintStatusDuringExecution(getConfig().isSysoutLoggingEnabled());
  }
  // if we are using sessions, we keep the executor running
  if (getSessionTimeout() > 0 && !executor.isRunning()) {
    executor.start();
    installShutdownHook();
  }
  return executor;
}
origin: apache/flink

@Override
@PublicEvolving
public void startNewSession() throws Exception {
  if (executor != null) {
    // we need to end the previous session
    executor.stop();
    // create also a new JobID
    jobID = JobID.generate();
  }
  // create a new local executor
  executor = PlanExecutor.createLocalExecutor(configuration);
  executor.setPrintStatusDuringExecution(getConfig().isSysoutLoggingEnabled());
  // if we have a session, start the mini cluster eagerly to have it available across sessions
  if (getSessionTimeout() > 0) {
    executor.start();
    // also install the reaper that will shut it down eventually
    executorReaper = new ExecutorReaper(executor);
  }
}
origin: apache/flink

client.setPrintStatusDuringExecution(executionConfig.isSysoutLoggingEnabled());
origin: apache/flink

assertEquals(disableGenericTypes, copy1.hasGenericTypesDisabled());
assertEquals(objectReuseEnabled, copy1.isObjectReuseEnabled());
assertEquals(sysoutLoggingEnabled, copy1.isSysoutLoggingEnabled());
assertEquals(parallelism, copy1.getParallelism());
origin: org.apache.flink/flink-scala-shell_2.10

@Override
protected PlanExecutor getExecutor() throws Exception {
  // check if we had already started a PlanExecutor. If true, then stop it, because there will
  // be a new jar file available for the user code classes
  if (this.executor != null) {
    this.executor.stop();
  }
  // write generated classes to disk so that they can be shipped to the cluster
  URL jarUrl = flinkILoop.writeFilesToDisk().getAbsoluteFile().toURI().toURL();
  List<URL> allJarFiles = new ArrayList<>(jarFiles);
  allJarFiles.add(jarUrl);
  this.executor = PlanExecutor.createRemoteExecutor(
    host,
    port,
    clientConfiguration,
    allJarFiles,
    globalClasspaths
  );
  executor.setPrintStatusDuringExecution(getConfig().isSysoutLoggingEnabled());
  return executor;
}
origin: org.apache.flink/flink-scala-shell

@Override
protected PlanExecutor getExecutor() throws Exception {
  // check if we had already started a PlanExecutor. If true, then stop it, because there will
  // be a new jar file available for the user code classes
  if (this.executor != null) {
    this.executor.stop();
  }
  // write generated classes to disk so that they can be shipped to the cluster
  URL jarUrl = flinkILoop.writeFilesToDisk().getAbsoluteFile().toURI().toURL();
  List<URL> allJarFiles = new ArrayList<>(jarFiles);
  allJarFiles.add(jarUrl);
  this.executor = PlanExecutor.createRemoteExecutor(
    host,
    port,
    clientConfiguration,
    allJarFiles,
    globalClasspaths
  );
  executor.setPrintStatusDuringExecution(getConfig().isSysoutLoggingEnabled());
  return executor;
}
origin: org.apache.flink/flink-streaming-java_2.10

try {
  client = new StandaloneClusterClient(configuration);
  client.setPrintStatusDuringExecution(getConfig().isSysoutLoggingEnabled());
origin: org.apache.flink/flink-java

protected PlanExecutor getExecutor() throws Exception {
  if (executor == null) {
    executor = PlanExecutor.createRemoteExecutor(host, port, clientConfiguration,
      jarFiles, globalClasspaths);
    executor.setPrintStatusDuringExecution(getConfig().isSysoutLoggingEnabled());
  }
  // if we are using sessions, we keep the executor running
  if (getSessionTimeout() > 0 && !executor.isRunning()) {
    executor.start();
    installShutdownHook();
  }
  return executor;
}
origin: org.apache.beam/beam-runners-flink_2.11

client.setPrintStatusDuringExecution(getConfig().isSysoutLoggingEnabled());
origin: org.apache.flink/flink-streaming-java

client.setPrintStatusDuringExecution(getConfig().isSysoutLoggingEnabled());
origin: org.apache.flink/flink-streaming-java_2.11

client.setPrintStatusDuringExecution(getConfig().isSysoutLoggingEnabled());
origin: dataArtisans/cascading-flink

client.setPrintStatusDuringExecution(env.getConfig().isSysoutLoggingEnabled());
origin: com.alibaba.blink/flink-java

protected PlanExecutor getExecutor() throws Exception {
  if (executor == null) {
    executor = PlanExecutor.createRemoteExecutor(host, port, clientConfiguration,
      jarFiles, globalClasspaths);
    executor.setPrintStatusDuringExecution(getConfig().isSysoutLoggingEnabled());
    executor.setJobListeners(this.getJobListeners());
  }
  // if we are using sessions, we keep the executor running
  if (getSessionTimeout() > 0 && !executor.isRunning()) {
    executor.start();
    installShutdownHook();
  }
  return executor;
}
origin: org.apache.flink/flink-streaming-java_2.10

try {
  exec.start();
  return exec.submitJobAndWait(jobGraph, getConfig().isSysoutLoggingEnabled());
origin: DTStack/flinkx

try {
  exec.start();
  return exec.submitJobAndWait(jobGraph, getConfig().isSysoutLoggingEnabled());
origin: DTStack/flinkStreamSQL

try {
  exec.start();
  return exec.submitJobAndWait(jobGraph, getConfig().isSysoutLoggingEnabled());
origin: org.apache.flink/flink-java

@Override
@PublicEvolving
public void startNewSession() throws Exception {
  if (executor != null) {
    // we need to end the previous session
    executor.stop();
    // create also a new JobID
    jobID = JobID.generate();
  }
  // create a new local executor
  executor = PlanExecutor.createLocalExecutor(configuration);
  executor.setPrintStatusDuringExecution(getConfig().isSysoutLoggingEnabled());
  // if we have a session, start the mini cluster eagerly to have it available across sessions
  if (getSessionTimeout() > 0) {
    executor.start();
    // also install the reaper that will shut it down eventually
    executorReaper = new ExecutorReaper(executor);
  }
}
origin: com.alibaba.blink/flink-java

@Override
@PublicEvolving
public void startNewSession() throws Exception {
  if (executor != null) {
    // we need to end the previous session
    executor.stop();
    // create also a new JobID
    jobID = JobID.generate();
  }
  // create a new local executor
  executor = PlanExecutor.createLocalExecutor(configuration);
  executor.setPrintStatusDuringExecution(getConfig().isSysoutLoggingEnabled());
  // if we have a session, start the mini cluster eagerly to have it available across sessions
  if (getSessionTimeout() > 0) {
    executor.start();
    // also install the reaper that will shut it down eventually
    executorReaper = new ExecutorReaper(executor);
  }
}
origin: DTStack/flinkx

@Override
@PublicEvolving
public void startNewSession() throws Exception {
  if (executor != null) {
    // we need to end the previous session
    executor.stop();
    // create also a new JobID
    jobID = JobID.generate();
  }
  // create a new local executor
  executor = new MyLocalExecutor(configuration);
  executor.setPrintStatusDuringExecution(getConfig().isSysoutLoggingEnabled());
  executor.setClasspaths(classpaths);
  // if we have a session, start the mini cluster eagerly to have it available across sessions
  if (getSessionTimeout() > 0) {
    executor.start();
    // also install the reaper that will shut it down eventually
    executorReaper = new ExecutorReaper(executor);
  }
}
org.apache.flink.api.commonExecutionConfigisSysoutLoggingEnabled

Javadoc

Gets whether progress update messages should be printed to System.out

Popular methods of ExecutionConfig

  • <init>
  • isObjectReuseEnabled
    Returns whether object reuse has been enabled or disabled. @see #enableObjectReuse()
  • disableSysoutLogging
    Disables the printing of progress update messages to System.out
  • getAutoWatermarkInterval
    Returns the interval of the automatic watermark emission.
  • setGlobalJobParameters
    Register a custom, serializable user configuration object.
  • enableObjectReuse
    Enables reusing objects that Flink internally uses for deserialization and passing data to user-code
  • setAutoWatermarkInterval
    Sets the interval of the automatic watermark emission. Watermarks are used throughout the streaming
  • disableObjectReuse
    Disables reusing objects that Flink internally uses for deserialization and passing data to user-cod
  • getRestartStrategy
    Returns the restart strategy which has been set for the current job.
  • registerKryoType
    Registers the given type with the serialization stack. If the type is eventually serialized as a POJ
  • registerTypeWithKryoSerializer
    Registers the given Serializer via its class as a serializer for the given type at the KryoSerialize
  • setRestartStrategy
    Sets the restart strategy to be used for recovery. ExecutionConfig config = env.getConfig();
  • registerTypeWithKryoSerializer,
  • setRestartStrategy,
  • getParallelism,
  • addDefaultKryoSerializer,
  • getGlobalJobParameters,
  • getNumberOfExecutionRetries,
  • getRegisteredKryoTypes,
  • setParallelism,
  • getDefaultKryoSerializerClasses

Popular in Java

  • Making http requests using okhttp
  • findViewById (Activity)
  • getSystemService (Context)
  • getSupportFragmentManager (FragmentActivity)
    Return the FragmentManager for interacting with fragments associated with this activity.
  • Container (java.awt)
    A generic Abstract Window Toolkit(AWT) container object is a component that can contain other AWT co
  • SQLException (java.sql)
    An exception that indicates a failed JDBC operation. It provides the following information about pro
  • Semaphore (java.util.concurrent)
    A counting semaphore. Conceptually, a semaphore maintains a set of permits. Each #acquire blocks if
  • Collectors (java.util.stream)
  • JFileChooser (javax.swing)
  • JList (javax.swing)
Codota Logo
  • Products

    Search for Java codeSearch for JavaScript codeEnterprise
  • IDE Plugins

    IntelliJ IDEAWebStormAndroid StudioEclipseVisual Studio CodePyCharmSublime TextPhpStormVimAtomGoLandRubyMineEmacsJupyter
  • Company

    About UsContact UsCareers
  • Resources

    FAQBlogCodota Academy Plugin user guide Terms of usePrivacy policyJava Code IndexJavascript Code Index
Get Codota for your IDE now