Warning: Use of undefined constan…… 0 [待解決問(wèn)題] 搭建網(wǎng)站時(shí),在網(wǎng)站安裝時(shí),返回的頁(yè)面為 Warning: Use of undefined constant TPL_TEMPLATE_PATH - assumed 'TPL_TEMPLATE_PATH' (this will throw an Error in a future version of PHP) in /opt/lampp/htdocs/YXcms/protected/core.php on line 254 Angel-史迪仔 | 菜鳥(niǎo)二級(jí) | 園豆: 202 提問(wèn)于:2019-05-18 13:37 顯示幫助 使用"Ctrl+Enter"可進(jìn)行快捷提交,評(píng)論支持部分 Markdown 語(yǔ)法:[link](http://example.com) _italic_ **bold** `code`。 < > 分享 分享您的問(wèn)題 清除回答草稿 您需要 登錄 以后才能回答,未注冊(cè)用戶(hù)請(qǐng)先 注冊(cè) 。
然后等一會(huì),這個(gè)界面就掛掉了,跳轉(zhuǎn)到如下界面: 然后控制臺(tái)報(bào)了如下錯(cuò)誤: ------------------------------------------------------------ The program finished with the following exception: org.apache.flink.client.program.ProgramInvocationException: The main method caused an error: org.apache.flink.client.program.ProgramInvocationException: Job failed (JobID: 12b759f143190ee08d831f2fabb4c3f2) at org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java: 335 ) at org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java: 205 ) at org.apache.flink.client.ClientUtils.executeProgram(ClientUtils.java: 138 ) at org.apache.flink.client.cli.CliFrontend.executeProgram(CliFrontend.java: 662 ) at org.apache.flink.client.cli.CliFrontend.run(CliFrontend.java: 210 ) at org.apache.flink.client.cli.CliFrontend.parseParameters(CliFrontend.java: 893 ) at org.apache.flink.client.cli.CliFrontend.lambda$main$ 10 (CliFrontend.java: 966 ) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java: 422 ) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java: 1692 ) at org.apache.flink.runtime.security.HadoopSecurityContext.runSecured(HadoopSecurityContext.java: 41 ) at org.apache.flink.client.cli.CliFrontend.main(CliFrontend.java: 966 ) Caused by: java.util.concurrent.ExecutionException: org.apache.flink.client.program.ProgramInvocationException: Job failed (JobID: 12b759f143190ee08d831f2fabb4c3f2) at java.util.concurrent.CompletableFuture.reportGet(CompletableFuture.java: 357 ) at java.util.concurrent.CompletableFuture. get (CompletableFuture.java: 1895 ) at org.apache.flink.streaming.api.environment.StreamContextEnvironment.execute(StreamContextEnvironment.java: 83 ) at org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.execute(StreamExecutionEnvironment.java: 1620 ) at org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.execute(StreamExecutionEnvironment.java: 1602 ) at org.apache.flink.streaming.api.scala.StreamExecutionEnvironment.execute(StreamExecutionEnvironment.scala: 667 ) at vip.shuai7boy.flink.checkpoint.TestSavepoints$.main(TestSavepoints.scala: 30 ) at vip.shuai7boy.flink.checkpoint.TestSavepoints.main(TestSavepoints.scala) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java: 62 ) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java: 43 ) at java.lang.reflect.Method.invoke(Method.java: 498 ) at org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java: 321 ) ... 11 more Caused by: org.apache.flink.client.program.ProgramInvocationException: Job failed (JobID: 12b759f143190ee08d831f2fabb4c3f2) at org.apache.flink.client.deployment.ClusterClientJobClientAdapter.lambda$ null $ 6 (ClusterClientJobClientAdapter.java: 112 ) at java.util.concurrent.CompletableFuture.uniApply(CompletableFuture.java: 602 ) at java.util.concurrent.CompletableFuture$UniApply.tryFire(CompletableFuture.java: 577 ) at java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java: 474 ) at java.util.concurrent.CompletableFuture.complete(CompletableFuture.java: 1962 ) at org.apache.flink.client.program.rest.RestClusterClient.lambda$pollResourceAsync$ 21 (RestClusterClient.java: 565 ) at java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java: 760 ) at java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java: 736 ) at java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java: 474 ) at java.util.concurrent.CompletableFuture.complete(CompletableFuture.java: 1962 ) at org.apache.flink.runtime.concurrent.FutureUtils.lambda$retryOperationWithDelay$ 8 (FutureUtils.java: 291 ) at java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java: 760 ) at java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java: 736 ) at java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java: 474 ) at java.util.concurrent.CompletableFuture.postFire(CompletableFuture.java: 561 ) at java.util.concurrent.CompletableFuture$UniCompose.tryFire(CompletableFuture.java: 929 ) at java.util.concurrent.CompletableFuture$Completion.run(CompletableFuture.java: 442 ) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java: 1149 ) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java: 624 ) at java.lang.Thread.run(Thread.java: 748 ) Caused by: org.apache.flink.runtime.client.JobExecutionException: Job execution failed. at org.apache.flink.runtime.jobmaster.JobResult.toJobExecutionResult(JobResult.java: 147 ) at org.apache.flink.client.deployment.ClusterClientJobClientAdapter.lambda$ null $ 6 (ClusterClientJobClientAdapter.java: 110 ) ... 19 more Caused by: org.apache.flink.runtime.JobException: Recovery is suppressed by NoRestartBackoffTimeStrategy at org.apache.flink.runtime.executiongraph.failover.flip1.ExecutionFailureHandler.handleFailure(ExecutionFailureHandler.java: 110 ) at org.apache.flink.runtime.executiongraph.failover.flip1.ExecutionFailureHandler.getFailureHandlingResult(ExecutionFailureHandler.java: 76 ) at org.apache.flink.runtime.scheduler.DefaultScheduler.handleTaskFailure(DefaultScheduler.java: 192 ) at org.apache.flink.runtime.scheduler.DefaultScheduler.maybeHandleTaskFailure(DefaultScheduler.java: 186 ) at org.apache.flink.runtime.scheduler.DefaultScheduler.updateTaskExecutionStateInternal(DefaultScheduler.java: 180 ) at org.apache.flink.runtime.scheduler.SchedulerBase.updateTaskExecutionState(SchedulerBase.java: 496 ) at org.apache.flink.runtime.scheduler.UpdateSchedulerNgOnInternalFailuresListener.notifyTaskFailure(UpdateSchedulerNgOnInternalFailuresListener.java: 49 ) at org.apache.flink.runtime.executiongraph.ExecutionGraph.notifySchedulerNgAboutInternalTaskFailure(ExecutionGraph.java: 1703 ) at org.apache.flink.runtime.executiongraph.Execution.processFail(Execution.java: 1252 ) at org.apache.flink.runtime.executiongraph.Execution.processFail(Execution.java: 1220 ) at org.apache.flink.runtime.executiongraph.Execution.markFailed(Execution.java: 1051 ) at org.apache.flink.runtime.executiongraph.ExecutionVertex.markFailed(ExecutionVertex.java: 748 ) at org.apache.flink.runtime.scheduler.DefaultExecutionVertexOperations.markFailed(DefaultExecutionVertexOperations.java: 41 ) at org.apache.flink.runtime.scheduler.DefaultScheduler.handleTaskDeploymentFailure(DefaultScheduler.java: 446 ) at org.apache.flink.runtime.scheduler.DefaultScheduler.lambda$assignResourceOrHandleError$ 5 (DefaultScheduler.java: 433 ) at java.util.concurrent.CompletableFuture.uniHandle(CompletableFuture.java: 822 ) at java.util.concurrent.CompletableFuture$UniHandle.tryFire(CompletableFuture.java: 797 ) at java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java: 474 ) at java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java: 1977 ) at org.apache.flink.runtime.jobmaster.slotpool.SchedulerImpl.lambda$internalAllocateSlot$ 0 (SchedulerImpl.java: 168 ) at java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java: 760 ) at java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java: 736 ) at java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java: 474 ) at java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java: 1977 ) at org.apache.flink.runtime.jobmaster.slotpool.SlotSharingManager$SingleTaskSlot.release(SlotSharingManager.java: 726 ) at org.apache.flink.runtime.jobmaster.slotpool.SlotSharingManager$MultiTaskSlot.release(SlotSharingManager.java: 537 ) at org.apache.flink.runtime.jobmaster.slotpool.SlotSharingManager$MultiTaskSlot.lambda$ new $ 0 (SlotSharingManager.java: 432 ) at java.util.concurrent.CompletableFuture.uniHandle(CompletableFuture.java: 822 ) at java.util.concurrent.CompletableFuture$UniHandle.tryFire(CompletableFuture.java: 797 ) at java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java: 474 ) at java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java: 1977 ) at org.apache.flink.runtime.concurrent.FutureUtils.lambda$forward$ 21 (FutureUtils.java: 1065 ) at java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java: 760 ) at java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java: 736 ) at java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java: 474 ) at java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java: 1977 ) at org.apache.flink.runtime.concurrent.FutureUtils$Timeout.run(FutureUtils.java: 999 ) at org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRunAsync(AkkaRpcActor.java: 402 ) at org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRpcMessage(AkkaRpcActor.java: 195 ) at org.apache.flink.runtime.rpc.akka.FencedAkkaRpcActor.handleRpcMessage(FencedAkkaRpcActor.java: 74 ) at org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleMessage(AkkaRpcActor.java: 152 ) at akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala: 26 ) at akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala: 21 ) at scala.PartialFunction.applyOrElse(PartialFunction.scala: 123 ) at scala.PartialFunction.applyOrElse$(PartialFunction.scala: 122 ) at akka.japi.pf.UnitCaseStatement.applyOrElse(CaseStatements.scala: 21 ) at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala: 171 ) at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala: 172 ) at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala: 172 ) at akka.actor.Actor.aroundReceive(Actor.scala: 517 ) at akka.actor.Actor.aroundReceive$(Actor.scala: 515 ) at akka.actor.AbstractActor.aroundReceive(AbstractActor.scala: 225 ) at akka.actor.ActorCell.receiveMessage(ActorCell.scala: 592 ) at akka.actor.ActorCell.invoke(ActorCell.scala: 561 ) at akka.dispatch.Mailbox.processMailbox(Mailbox.scala: 258 ) at akka.dispatch.Mailbox.run(Mailbox.scala: 225 ) at akka.dispatch.Mailbox.exec(Mailbox.scala: 235 ) at akka.dispatch.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java: 260 ) at akka.dispatch.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java: 1339 ) at akka.dispatch.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java: 1979 ) at akka.dispatch.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java: 107 ) Caused by: org.apache.flink.runtime.jobmanager.scheduler.NoResourceAvailableException: Could not allocate the required slot within slot request timeout. Please make sure that the cluster has enough resources. at org.apache.flink.runtime.scheduler.DefaultScheduler.maybeWrapWithNoResourceAvailableException(DefaultScheduler.java: 452 ) ... 47 more Caused by: java.util.concurrent.CompletionException: java.util.concurrent.TimeoutException at java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java: 292 ) at java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java: 308 ) at java.util.concurrent.CompletableFuture.uniApply(CompletableFuture.java: 593 ) at java.util.concurrent.CompletableFuture$UniApply.tryFire(CompletableFuture.java: 577 ) ... 27 more Caused by: java.util.concurrent.TimeoutException ... 25 more 我的服務(wù)器運(yùn)行情況 一共有四臺(tái)服務(wù)器,jps命令信息如下: 第一臺(tái)服務(wù)器(tuge1): 5794 ResourceManager 5459 NameNode 5689 DFSZKFailoverController 10297 Jps 1834 Application 8123 JobHistoryServer 2686 QuorumPeerMain 第二臺(tái)服務(wù)器(tuge2): 4929 DFSZKFailoverController 4822 NameNode 4748 JournalNode 12429 Jps 4654 QuorumPeerMain 第三臺(tái)服務(wù)器(tuge3): 9700 Jps 4965 JournalNode 5157 NodeManager 5048 DataNode 4877 QuorumPeerMain 第四臺(tái)服務(wù)器(tuge4): 4771 JournalNode 4846 DataNode 4958 NodeManager 11758 Jps PS:我的虛擬機(jī)配置的每臺(tái)服務(wù)器都是2核2G. 我的flink配置情況 flink-conf.yaml配置如下: ################################################################################ # Licensed to the Apache Software Foundation (ASF) under one # or more contributor license agreements. See the NOTICE file # distributed with this work for additional information # regarding copyright ownership. The ASF licenses this file # to you under the Apache License, Version 2.0 (the # " License " ); you may not use this file except in compliance # with the License. You may obtain a copy of the License at # # http: // www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an " AS IS " BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. ################################################################################ # ============================================================================== # Common # ============================================================================== # The external address of the host on which the JobManager runs and can be # reached by the TaskManagers and any clients which want to connect. This setting # is only used in Standalone mode and may be overwritten on the JobManager side # by specifying the --host parameter of the bin/ jobmanager.sh executable. # In high availability mode, if you use the bin/start- cluster.sh script and setup # the conf /masters file, this will be taken care of automatically. Yarn/ Mesos # automatically configure the host name based on the hostname of the node where the #設(shè)置task內(nèi)存 taskmanager.network.memory.fraction: 0.1 taskmanager.network.memory.min: 64mb taskmanager.network.memory.max: 1gb # JobManager runs. jobmanager.rpc.address: tuge1 # The RPC port where the JobManager is reachable. jobmanager.rpc.port: 6123 # The heap size for the JobManager JVM jobmanager.heap.size: 1024m # The total process memory size for the TaskManager. # # Note this accounts for all memory usage within the TaskManager process, including JVM metaspace and other overhead. taskmanager.memory.process.size: 1024m # To exclude JVM metaspace and overhead, please, use total Flink memory size instead of ' taskmanager.memory.process.size ' . # It is not recommended to set both ' taskmanager.memory.process.size ' and Flink memory. # # taskmanager.memory.flink.size: 1280m # The number of task slots that each TaskManager offers. Each slot runs one parallel pipeline. taskmanager.numberOfTaskSlots: 2 # The parallelism used for programs that did not specify and other parallelism. parallelism. default : 1 # The default file system scheme and authority. # # By default file paths without scheme are interpreted relative to the local # root file system ' file:/// ' . Use this to override the default and interpret # relative paths relative to a different file system, # for example ' hdfs://mynamenode:12345 ' # # fs. default - scheme # ============================================================================== # High Availability # ============================================================================== # The high -availability mode. Possible options are ' NONE ' or ' zookeeper ' . # high - availability: zookeeper # The path where metadata for master recovery is persisted. While ZooKeeper stores # the small ground truth for checkpoint and leader election, this location stores # the larger objects, like persisted dataflow graphs. # # Must be a durable file system that is accessible from all nodes # (like HDFS, S3, Ceph, nfs, ...) # high -availability.storageDir: hdfs: // tuge1:9000/ha/ # The list of ZooKeeper quorum peers that coordinate the high - availability # setup. This must be a list of the form: # " host1:clientPort,host2:clientPort,... " ( default clientPort: 2181 ) # high -availability.zookeeper.quorum: tuge1: 2181 ,tuge2: 2181 ,tuge3: 2181 # ACL options are based on https: // zookeeper.apache.org/doc/r3.1.2/zookeeperProgrammers.html#sc_BuiltinACLSchemes # It can be either " creator " (ZOO_CREATE_ALL_ACL) or " open " (ZOO_OPEN_ACL_UNSAFE) # The default value is " open " and it can be changed to " creator " if ZK security is enabled # # high - availability.zookeeper.client.acl: open # ============================================================================== # Fault tolerance and checkpointing # ============================================================================== # The backend that will be used to store operator state checkpoints if # checkpointing is enabled. # # Supported backends are ' jobmanager ' , ' filesystem ' , ' rocksdb ' , or the # < class -name-of-factory> . # # state.backend: filesystem # Directory for checkpoints filesystem, when using any of the default bundled # state backends. # # state.checkpoints.dir: hdfs: // namenode-host:port/flink-checkpoints # Default target directory for savepoints, optional. # state.savepoints.dir: hdfs: // tuge1:9000/flink-checkpoints # Flag to enable /disable incremental checkpoints for backends that # support incremental checkpoints (like the RocksDB state backend). # # state.backend.incremental: false # The failover strategy, i.e., how the job computation recovers from task failures. # Only restart tasks that may have been affected by the task failure, which typically includes # downstream tasks and potentially upstream tasks if their produced data is no longer available for consumption. jobmanager.execution.failover - strategy: region # ============================================================================== # Rest & web frontend # ============================================================================== # The port to which the REST client connects to. If rest.bind - port has # not been specified, then the server will bind to this port as well. # #rest.port: 8081 # The address to which the REST client will connect to # #rest.address: 0.0 . 0.0 # Port range for the REST and web server to bind to. # #rest.bind -port: 8080 - 8090 # The address that the REST & web server binds to # #rest.bind -address: 0.0 . 0.0 # Flag to specify whether job submission is enabled from the web- based # runtime monitor. Uncomment to disable. # web.submit.enable: true # ============================================================================== # Advanced # ============================================================================== # Override the directories for temporary files. If not specified, the # system -specific Java temporary directory (java.io.tmpdir property) is taken. # # For framework setups on Yarn or Mesos, Flink will automatically pick up the # containers ' temp directories without any need for configuration. # # Add a delimited list for multiple directories, using the system directory # delimiter (colon ' : ' on unix) or a comma, e.g.: # /data1/tmp:/data2/tmp:/data3/ tmp # # Note: Each directory entry is read from and written to by a different I/ O # thread. You can include the same directory multiple times in order to create # multiple I /O threads against that directory. This is for example relevant for # high - throughput RAIDs. # io.tmp.dirs: / tmp # The classloading resolve order. Possible values are ' child-first ' (Flink ' s default) # and ' parent-first ' (Java ' s default). # # Child first classloading allows users to use different dependency / library # versions in their application than those in the classpath. Switching back # to ' parent-first ' may help with debugging dependency issues. # # classloader.resolve -order: child- first # The amount of memory going to the network stack. These numbers usually need # no tuning. Adjusting them may be necessary in case of an " Insufficient number # of network buffers " error. The default min is 64MB, the default max is 1GB. # # taskmanager.memory.network.fraction: 0.1 # taskmanager.memory.network.min: 64mb # taskmanager.memory.network.max: 1gb # ============================================================================== # Flink Cluster Security Configuration # ============================================================================== # Kerberos authentication for various components - Hadoop, ZooKeeper, and connectors - # may be enabled in four steps: # 1 . configure the local krb5.conf file # 2 . provide Kerberos credentials (either a keytab or a ticket cache w/ kinit) # 3 . make the credentials available to various JAAS login contexts # 4 . configure the connector to use JAAS/ SASL # The below configure how Kerberos credentials are provided. A keytab will be used instead of # a ticket cache if the keytab path and principal are set . # security.kerberos.login.use -ticket-cache: true # security.kerberos.login.keytab: /path/to/kerberos/ keytab # security.kerberos.login.principal: flink - user # The configuration below defines which JAAS login contexts # security.kerberos.login.contexts: Client,KafkaClient # ============================================================================== # ZK Security Configuration # ============================================================================== # Below configurations are applicable if ZK ensemble is configured for security # Override below configuration to provide custom ZK service name if configured # zookeeper.sasl.service - name: zookeeper # The configuration below must match one of the values set in " security.kerberos.login.contexts " # zookeeper.sasl.login -context- name: Client # ============================================================================== # HistoryServer # ============================================================================== # The HistoryServer is started and stopped via bin/historyserver.sh (start| stop) # Directory to upload completed jobs to. Add this directory to the list of # monitored directories of the HistoryServer as well (see below). #jobmanager.archive.fs.dir: hdfs: /// completed-jobs/ # The address under which the web - based HistoryServer listens. #historyserver.web.address: 0.0 . 0.0 # The port under which the web - based HistoryServer listens. #historyserver.web.port: 8082 # Comma separated list of directories to monitor for completed jobs. #historyserver.archive.fs.dir: hdfs: /// completed-jobs/ # Interval in milliseconds for refreshing the monitored directories. #historyserver.archive.fs.refresh -interval: 10000 yarn.application -attempts: 10 期望結(jié)果 我希望可以正常運(yùn)行。大佬們幫忙看下啥原因~ flink linux yarn shuai7boy | 菜鳥(niǎo)二級(jí) | 園豆: 218 提問(wèn)于:2020-05-28 13:09 顯示幫助 使用"Ctrl+Enter"可進(jìn)行快捷提交,評(píng)論支持部分 Markdown 語(yǔ)法:[link](http://example.com) _italic_ **bold** `code`。 < > 分享 分享您的問(wèn)題 所有回答(1) 0 Caused by: org.apache.flink.runtime.jobmanager.scheduler.NoResourceAvailableException: Could not allocate the required slot within slot request timeout. Please make sure that the cluster has enough resources 你的YARN里面隊(duì)列沒(méi)有分配資源,在YARN里面劃分一個(gè)隊(duì)列,給到CPU核數(shù)和內(nèi)存。然后在提交Flink任務(wù)應(yīng)該就可以了。 例如在文件 fair-scheduler.xml 創(chuàng)建一個(gè)queue_hadoop_01的隊(duì)列: hadoophadoop101024mb,1vcores6144mb,6vcoresfair1.0hadoophadoop600000600000 然后在執(zhí)行Flink提交命令: flink run -m yarn-cluster -yn 2 -yjm 1024 -ytm 1024 -yqu queue_hadoop_01 flink_task.jar 哥不是小蘿莉 | 園豆:202 (菜鳥(niǎo)二級(jí)) | 2020-05-31 15:16 編輯文本 預(yù)覽 上傳圖片 Ctrl+Enter鍵快速提交 清除回答草稿 您需要 登錄 以后才能回答,未注冊(cè)用戶(hù)請(qǐng)先 注冊(cè) 。
winform圖片背景的問(wèn)題。 0 懸賞園豆: 20 [待解決問(wèn)題] winform 通過(guò)控件屬性導(dǎo)入圖片(PNG格式),如按鈕背景圖片、窗體背景、Label控件的背景圖片等。,能正常發(fā)布、運(yùn)行。 但是過(guò)一段時(shí)間后(約3~5小時(shí)之間)會(huì)報(bào)以下錯(cuò)誤,然后界面的圖片就消失了,程序也一直報(bào)錯(cuò),重新啟動(dòng)程序后又能正常使用一段時(shí)間,具體錯(cuò)誤如下: [2020-05-26 13:44:30.161] TextLogger.Error 1 異常文本 【出現(xiàn)時(shí)間】:2020/5/26 13:44:30 【異常方法】:get_Flags 【異常窗口】:System.Drawing.dll 【異常類(lèi)型】:ArgumentException 【異常信息】:Parameter is not valid. 【堆棧調(diào)用】: at System.Drawing.Image.get_Flags() at System.Windows.Forms.ControlPaint.IsImageTransparent(Image backgroundImage) at System.Windows.Forms.Control.set_BackgroundImageLayout(ImageLayout value) at IccOilDS.Fmain.InitializeComponent() in E:\IccOilDS\Fmain.Designer.cs:line 268 at IccOilDS.Fmain..ctor() in E:\IccOilDS\Fmain.cs:line 31 at IccOilDS.Frmckdd.btnPrintCKD_Click(Object sender, EventArgs e) in E:\IccOilDS\Frmckdd.cs:line 927 at System.Windows.Forms.Control.OnClick(EventArgs e) at System.Windows.Forms.Button.OnClick(EventArgs e) at System.Windows.Forms.Button.OnMouseUp(MouseEventArgs mevent) at System.Windows.Forms.Control.WmMouseUp(Message& m, MouseButtons button, Int32 clicks) at System.Windows.Forms.Control.WndProc(Message& m) at System.Windows.Forms.ButtonBase.WndProc(Message& m) at System.Windows.Forms.Button.WndProc(Message& m) at System.Windows.Forms.Control.ControlNativeWindow.OnMessage(Message& m) at System.Windows.Forms.Control.ControlNativeWindow.WndProc(Message& m) at System.Windows.Forms.NativeWindow.Callback(IntPtr hWnd, Int32 msg, IntPtr wparam, IntPtr lparam) 【系統(tǒng)版本】:1.0.7450.20051 【總窗體數(shù)】:4 【總線(xiàn)程數(shù)】:1 [2020-05-26 13:44:35.749] TextLogger.Error 1 異常文本 【出現(xiàn)時(shí)間】:2020/5/26 13:44:35 【異常方法】:CheckErrorStatus 【異常窗口】:System.Drawing.dll 【異常類(lèi)型】:OutOfMemoryException 【異常信息】:Out of memory. 【堆棧調(diào)用】: at System.Drawing.Graphics.CheckErrorStatus(Int32 status) at System.Drawing.Graphics.DrawImage(Image image, Int32 x, Int32 y, Int32 width, Int32 height) at System.Drawing.Graphics.DrawImage(Image image, Rectangle rect) at System.Windows.Forms.PictureBox.OnPaint(PaintEventArgs pe) at System.Windows.Forms.Control.PaintWithErrorHandling(PaintEventArgs e, Int16 layer) at System.Windows.Forms.Control.WmPaint(Message& m) at System.Windows.Forms.Control.WndProc(Message& m) at System.Windows.Forms.Control.ControlNativeWindow.OnMessage(Message& m) at System.Windows.Forms.Control.ControlNativeWindow.WndProc(Message& m) at System.Windows.Forms.NativeWindow.Callback(IntPtr hWnd, Int32 msg, IntPtr wparam, IntPtr lparam) 【系統(tǒng)版本】:1.0.7450.20051 【總窗體數(shù)】:4 【總線(xiàn)程數(shù)】:1 winform net 黑色鍵盤(pán) | 初學(xué)一級(jí) | 園豆: 8 提問(wèn)于:2020-05-27 11:44 顯示幫助 使用"Ctrl+Enter"可進(jìn)行快捷提交,評(píng)論支持部分 Markdown 語(yǔ)法:[link](http://example.com) _italic_ **bold** `code`。 < > 分享 分享您的問(wèn)題 所有回答(2) 0 刪除所有你的業(yè)務(wù)代碼,只留下一個(gè)控件,加入背景圖片?這樣跑跑看。否則你很難證明不是你的其他業(yè)務(wù)代碼或者哪兒的代碼造成的。 如果這樣都能出錯(cuò),你可以去找微軟啊。 如果你的應(yīng)用程序涉及各種媒體類(lèi)型,WPF是一個(gè)特別好的選擇。 比如,如果您需要合并視頻,文檔或圖像序列之間的動(dòng)畫(huà)過(guò)渡,或上述任何組合。 如何你需要?jiǎng)?chuàng)建具有類(lèi)似Web的導(dǎo)航的桌面應(yīng)用程序,WPF也很棒。 WPF的優(yōu)勢(shì)如下: 能夠相對(duì)容易地創(chuàng)建非常豐富的UI。 更輕松的動(dòng)畫(huà)和特效 天然的強(qiáng)大可擴(kuò)展性 強(qiáng)大的樣式(style)和樹(shù)形結(jié)構(gòu) 借助Dx可顯示高級(jí)圖形 (Winform 用的是GDI繪圖) 愛(ài)編程的大叔 | 園豆:30664 (高人七級(jí)) | 2020-05-27 11:58 編輯文本 預(yù)覽 上傳圖片 Ctrl+Enter鍵快速提交 0 大概率是代碼有問(wèn)題,建議發(fā)一個(gè)可復(fù)現(xiàn)問(wèn)題的demo 會(huì)長(zhǎng) | 園豆:8408 (大俠五級(jí)) | 2020-05-27 13:46 編輯文本 預(yù)覽 上傳圖片 Ctrl+Enter鍵快速提交 清除回答草稿 您需要 登錄 以后才能回答,未注冊(cè)用戶(hù)請(qǐng)先 注冊(cè) 。
用Xelatex轉(zhuǎn)化文件為xdv時(shí)報(bào)錯(cuò) 0 懸賞園豆: 10 [待解決問(wèn)題] 報(bào)錯(cuò)Exception: Xelatex error converting to xdv 打開(kāi)log后沒(méi)發(fā)現(xiàn)什么漏掉的包 出現(xiàn) ! Undefined control sequence. __xeCJK_patch_microtype_get_slot: l.4928 } Here is how much of TeX's memory you used: 10816 strings out of 414071 228709 string characters out of 2901731 568659 words of memory out of 3000000 28362 multiletter control sequences out of 15000+200000 532378 words of font info for 29 fonts, out of 3000000 for 9000 1348 hyphenation exceptions out of 8191 57i,0n,60p,381b,167s stack positions out of 5000i,500n,10000p,200000b,50000s No pages of output. 不知道是哪里的問(wèn)題但網(wǎng)上有的說(shuō)法是(原文是轉(zhuǎn)化svg出現(xiàn)錯(cuò)誤)運(yùn)行 regsvr32 MiKTeX209-core-PS.dll regsvr32 MiKTeX209-core.dll regsvr32 MiKTeX209-packagemanager-PS.dll regsvr32 MiKTeX209-packagemanager.dll 但運(yùn)行不起來(lái),顯示“找不到指定模塊” 找遍了各個(gè)盤(pán)也找不到這幾個(gè)文件 請(qǐng)問(wèn)這樣如何解決 嘿呦呦呦 | 初學(xué)一級(jí) | 園豆: 168 提問(wèn)于:2020-08-13 11:35 顯示幫助 使用"Ctrl+Enter"可進(jìn)行快捷提交,評(píng)論支持部分 Markdown 語(yǔ)法:[link](http://example.com) _italic_ **bold** `code`。 < > 分享 分享您的問(wèn)題 清除回答草稿 您需要 登錄 以后才能回答,未注冊(cè)用戶(hù)請(qǐng)先 注冊(cè) 。
第一個(gè)字節(jié)不能讀報(bào)IOException 0 懸賞園豆: 10 [待解決問(wèn)題] 這個(gè)可以舉個(gè)例子嗎,什么情況下第一個(gè)字節(jié)不能讀 java.io.IOException – If the first byte cannot be read for any reason other than the end of the file java c# io ioexception read xiaofang.bk | 初學(xué)一級(jí) | 園豆: 1 提問(wèn)于:2020-06-21 18:33 顯示幫助 使用"Ctrl+Enter"可進(jìn)行快捷提交,評(píng)論支持部分 Markdown 語(yǔ)法:[link](http://example.com) _italic_ **bold** `code`。 < > 分享 分享您的問(wèn)題 清除回答草稿 您需要 登錄 以后才能回答,未注冊(cè)用戶(hù)請(qǐng)先 注冊(cè) 。