本帖最后由 Cherise 于 2015-11-10 11:57 编辑
最近按照CDH的官方文档将集群从5.1.2升级到了5.4.7,相应的Spark从1.0.0升级到了1.3.0。
升级后,spark可正常使用,但遇到了一个很奇怪的问题,在Spark web UI页面无法看到executor的执行日志stderr,只有如下信息:
SLF4J: Class path contains multiple SLF4J bindings.SLF4J: Found binding in [jar:file:/opt/cloudera/parcels/CDH-5.4.7-1.cdh5.4.7.p0.3/jars/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]SLF4J: Found binding in [jar:file:/opt/cloudera/parcels/CDH-5.4.7-1.cdh5.4.7.p0.3/jars/avro-tools-1.7.6-cdh5.4.7.jar!/org/slf4j/impl/StaticLoggerBinder.class]SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
从日志上看有SLF4J包的冲突,后来我将/opt/cloudera/parcels/CDH-5.4.7-1.cdh5.4.7.p0.3/jars/目录下的avro-tools-1.7.6-cdh5.4.7.jar重命名为avro-tools-1.7.6-cdh5.4.7.jar.bak,修改后没有上面的错误了,但却一点日志信息都没有了,当然/var/run/spark/work目录下的日志与之相同。
但在/var/log/spark/spark-worker-cdhdatanode1.log文件中却看到了以下信息:
2015-11-09 16:22:15,165 INFO org.apache.spark.executor.CoarseGrainedExecutorBackend: Registered signal handlers for [TERM, HUP, INT]
2015-11-09 16:22:16,873 INFO org.apache.spark.SecurityManager: Changing view acls to: spark,root
2015-11-09 16:22:16,874 INFO org.apache.spark.SecurityManager: Changing modify acls to: spark,root
2015-11-09 16:22:16,875 INFO org.apache.spark.SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(spark, root); use
rs with modify permissions: Set(spark, root)
2015-11-09 16:22:17,308 INFO akka.event.slf4j.Slf4jLogger: Slf4jLogger started
2015-11-09 16:22:17,425 INFO Remoting: Starting remoting
2015-11-09 16:22:17,677 INFO Remoting: Remoting started; listening on addresses :[akka.tcp://driverPropsFetcher@cdhdatanode1:44604]
2015-11-09 16:22:17,680 INFO Remoting: Remoting now listens on addresses: [akka.tcp://driverPropsFetcher@cdhdatanode1:44604]
2015-11-09 16:22:17,693 INFO org.apache.spark.util.Utils: Successfully started service 'driverPropsFetcher' on port 44604.
2015-11-09 16:22:17,951 INFO akka.remote.RemoteActorRefProvider$RemotingTerminator: Shutting down remote daemon.
2015-11-09 16:22:17,952 INFO akka.remote.RemoteActorRefProvider$RemotingTerminator: Remote daemon shut down; proceeding with flushing remote transports.
2015-11-09 16:22:17,964 INFO org.apache.spark.SecurityManager: Changing view acls to: spark,root
2015-11-09 16:22:17,964 INFO org.apache.spark.SecurityManager: Changing modify acls to: spark,root
2015-11-09 16:22:17,964 INFO org.apache.spark.SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(spark, root); use
rs with modify permissions: Set(spark, root)
2015-11-09 16:22:18,023 INFO Remoting: Remoting shut down
2015-11-09 16:22:18,025 INFO akka.remote.RemoteActorRefProvider$RemotingTerminator: Remoting shut down.
2015-11-09 16:22:18,034 INFO akka.event.slf4j.Slf4jLogger: Slf4jLogger started
2015-11-09 16:22:18,045 INFO Remoting: Starting remoting
2015-11-09 16:22:18,060 INFO org.apache.spark.util.Utils: Successfully started service 'sparkExecutor' on port 40365.
2015-11-09 16:22:18,061 INFO Remoting: Remoting started; listening on addresses :[akka.tcp://sparkExecutor@cdhdatanode1:40365]
2015-11-09 16:22:18,061 INFO Remoting: Remoting now listens on addresses: [akka.tcp://sparkExecutor@cdhdatanode1:40365]
2015-11-09 16:22:18,069 INFO org.apache.spark.util.AkkaUtils: Connecting to MapOutputTracker: akka.tcp://sparkDriver@cdhnamenode:59211/user/MapOutputTracker
2015-11-09 16:22:18,334 INFO org.apache.spark.util.AkkaUtils: Connecting to BlockManagerMaster: akka.tcp://sparkDriver@cdhnamenode:59211/user/BlockManagerMaster
2015-11-09 16:22:18,485 INFO org.apache.spark.storage.DiskBlockManager: Created local directory at /tmp/spark-18da685b-6197-444c-9478-d250f6b59f4a/spark-8bc0a086-ea65-4c4e-9
4ec-c143539591d0/spark-17299c8e-86b5-47fa-b621-f21261677efc/blockmgr-5c385183-0196-4bf6-8f69-31ea24bdfa0c
2015-11-09 16:22:18,514 INFO org.apache.spark.storage.MemoryStore: MemoryStore started with capacity 267.3 MB
2015-11-09 16:22:18,825 INFO org.apache.spark.util.AkkaUtils: Connecting to OutputCommitCoordinator: akka.tcp://sparkDriver@cdhnamenode:59211/user/OutputCommitCoordinator
2015-11-09 16:22:19,010 INFO org.apache.spark.executor.CoarseGrainedExecutorBackend: Connecting to driver: akka.tcp://sparkDriver@cdhnamenode:59211/user/CoarseGrainedScheduler
怎么感觉executor的日志被写入到了worker的日志中??
有没有童鞋遇到过这种情况?还是我升级有问题?没有日志实在是很不方便啊!困惑几天了,望大家不吝赐教,谢谢!
PS:executor的stdout日志是可以正常写入的;将spark组件删除,重新添加spark(standalone)服务也不行
|
|