x_job_hive2hive_json-22_44_22.055.log 28 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379
  1. 2022-07-17 22:44:22.379 [main] INFO VMInfo - VMInfo# operatingSystem class => sun.management.OperatingSystemImpl
  2. 2022-07-17 22:44:22.389 [main] INFO Engine - the machine info =>
  3. osInfo: Oracle Corporation 1.8 25.262-b10
  4. jvmInfo: Linux amd64 3.10.0-957.el7.x86_64
  5. cpu num: 128
  6. totalPhysicalMemory: -0.00G
  7. freePhysicalMemory: -0.00G
  8. maxFileDescriptorCount: -1
  9. currentOpenFileDescriptorCount: -1
  10. GC Names [PS MarkSweep, PS Scavenge]
  11. MEMORY_NAME | allocation_size | init_size
  12. PS Eden Space | 256.00MB | 256.00MB
  13. Code Cache | 240.00MB | 2.44MB
  14. Compressed Class Space | 1,024.00MB | 0.00MB
  15. PS Survivor Space | 42.50MB | 42.50MB
  16. PS Old Gen | 683.00MB | 683.00MB
  17. Metaspace | -0.00MB | 0.00MB
  18. 2022-07-17 22:44:22.408 [main] INFO Engine -
  19. {
  20. "content":[
  21. {
  22. "reader":{
  23. "name":"hdfsreader",
  24. "parameter":{
  25. "column":[
  26. "*"
  27. ],
  28. "defaultFS":"hdfs://hadoop01:8020/",
  29. "encoding":"UTF-8",
  30. "fieldDelimiter":"\t",
  31. "fileType":"text",
  32. "path":"/user/hive/warehouse/user_info/user_info_data.txt"
  33. }
  34. },
  35. "writer":{
  36. "name":"hdfswriter",
  37. "parameter":{
  38. "column":[
  39. {
  40. "name":"user_id",
  41. "type":"string"
  42. },
  43. {
  44. "name":"age",
  45. "type":"int"
  46. }
  47. ],
  48. "compress":"",
  49. "defaultFS":"hdfs://hadoop01:8020/",
  50. "fieldDelimiter":"\t",
  51. "fileName":"user_info_data_1.txt",
  52. "fileType":"text",
  53. "path":"/user/hive/warehouse/user_info/",
  54. "writeMode":"append"
  55. }
  56. }
  57. }
  58. ],
  59. "setting":{
  60. "speed":{
  61. "channel":"1"
  62. }
  63. }
  64. }
  65. 2022-07-17 22:44:22.428 [main] WARN Engine - prioriy set to 0, because NumberFormatException, the value is: null
  66. 2022-07-17 22:44:22.430 [main] INFO PerfTrace - PerfTrace traceId=job_-1, isEnable=false, priority=0
  67. 2022-07-17 22:44:22.430 [main] INFO JobContainer - DataX jobContainer starts job.
  68. 2022-07-17 22:44:22.433 [main] INFO JobContainer - Set jobId = 0
  69. 2022-07-17 22:44:22.457 [job-0] INFO HdfsReader$Job - init() begin...
  70. 2022-07-17 22:44:22.869 [job-0] INFO HdfsReader$Job - hadoopConfig details:{"finalParameters":[]}
  71. 2022-07-17 22:44:22.869 [job-0] INFO HdfsReader$Job - init() ok and end...
  72. 2022-07-17 22:44:24.044 [job-0] INFO JobContainer - jobContainer starts to do prepare ...
  73. 2022-07-17 22:44:24.044 [job-0] INFO JobContainer - DataX Reader.Job [hdfsreader] do prepare work .
  74. 2022-07-17 22:44:24.044 [job-0] INFO HdfsReader$Job - prepare(), start to getAllFiles...
  75. 2022-07-17 22:44:24.045 [job-0] INFO HdfsReader$Job - get HDFS all files in path = [/user/hive/warehouse/user_info/user_info_data.txt]
  76. 2022-07-17 22:44:24.709 [job-0] ERROR HdfsReader$Job - 无法读取路径[/user/hive/warehouse/user_info/user_info_data.txt]下的所有文件,请确认您的配置项fs.defaultFS, path的值是否正确,是否有读写权限,网络是否已断开!
  77. 2022-07-17 22:44:24.715 [job-0] ERROR JobContainer - Exception when job run
  78. com.alibaba.datax.common.exception.DataXException: Code:[HdfsReader-09], Description:[您配置的path格式有误]. - org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.ipc.StandbyException): Operation category READ is not supported in state standby. Visit https://s.apache.org/sbnn-error
  79. at org.apache.hadoop.hdfs.server.namenode.ha.StandbyState.checkOperation(StandbyState.java:88)
  80. at org.apache.hadoop.hdfs.server.namenode.NameNode$NameNodeHAContext.checkOperation(NameNode.java:1962)
  81. at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkOperation(FSNamesystem.java:1421)
  82. at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getListing(FSNamesystem.java:3729)
  83. at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.getListing(NameNodeRpcServer.java:1138)
  84. at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.getListing(ClientNamenodeProtocolServerSideTranslatorPB.java:708)
  85. at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
  86. at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:523)
  87. at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:991)
  88. at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:870)
  89. at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:816)
  90. at java.security.AccessController.doPrivileged(Native Method)
  91. at javax.security.auth.Subject.doAs(Subject.java:422)
  92. at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1875)
  93. at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2680)
  94. at org.apache.hadoop.ipc.Client.call(Client.java:1476)
  95. at org.apache.hadoop.ipc.Client.call(Client.java:1407)
  96. at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:229)
  97. at com.sun.proxy.$Proxy20.getListing(Unknown Source)
  98. at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.getListing(ClientNamenodeProtocolTranslatorPB.java:573)
  99. at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
  100. at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
  101. at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
  102. at java.lang.reflect.Method.invoke(Method.java:498)
  103. at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:187)
  104. at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:102)
  105. at com.sun.proxy.$Proxy21.getListing(Unknown Source)
  106. at org.apache.hadoop.hdfs.DFSClient.listPaths(DFSClient.java:2094)
  107. at org.apache.hadoop.hdfs.DFSClient.listPaths(DFSClient.java:2077)
  108. at org.apache.hadoop.hdfs.DistributedFileSystem.listStatusInternal(DistributedFileSystem.java:791)
  109. at org.apache.hadoop.hdfs.DistributedFileSystem.access$700(DistributedFileSystem.java:106)
  110. at org.apache.hadoop.hdfs.DistributedFileSystem$18.doCall(DistributedFileSystem.java:853)
  111. at org.apache.hadoop.hdfs.DistributedFileSystem$18.doCall(DistributedFileSystem.java:849)
  112. at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
  113. at org.apache.hadoop.hdfs.DistributedFileSystem.listStatus(DistributedFileSystem.java:860)
  114. at com.alibaba.datax.plugin.reader.hdfsreader.DFSUtil.getHDFSAllFilesNORegex(DFSUtil.java:162)
  115. at com.alibaba.datax.plugin.reader.hdfsreader.DFSUtil.getHDFSAllFiles(DFSUtil.java:141)
  116. at com.alibaba.datax.plugin.reader.hdfsreader.DFSUtil.getAllFiles(DFSUtil.java:112)
  117. at com.alibaba.datax.plugin.reader.hdfsreader.HdfsReader$Job.prepare(HdfsReader.java:169)
  118. at com.alibaba.datax.core.job.JobContainer.prepareJobReader(JobContainer.java:715)
  119. at com.alibaba.datax.core.job.JobContainer.prepare(JobContainer.java:308)
  120. at com.alibaba.datax.core.job.JobContainer.start(JobContainer.java:115)
  121. at com.alibaba.datax.core.Engine.start(Engine.java:92)
  122. at com.alibaba.datax.core.Engine.entry(Engine.java:171)
  123. at com.alibaba.datax.core.Engine.main(Engine.java:204)
  124. - org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.ipc.StandbyException): Operation category READ is not supported in state standby. Visit https://s.apache.org/sbnn-error
  125. at org.apache.hadoop.hdfs.server.namenode.ha.StandbyState.checkOperation(StandbyState.java:88)
  126. at org.apache.hadoop.hdfs.server.namenode.NameNode$NameNodeHAContext.checkOperation(NameNode.java:1962)
  127. at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkOperation(FSNamesystem.java:1421)
  128. at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getListing(FSNamesystem.java:3729)
  129. at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.getListing(NameNodeRpcServer.java:1138)
  130. at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.getListing(ClientNamenodeProtocolServerSideTranslatorPB.java:708)
  131. at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
  132. at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:523)
  133. at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:991)
  134. at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:870)
  135. at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:816)
  136. at java.security.AccessController.doPrivileged(Native Method)
  137. at javax.security.auth.Subject.doAs(Subject.java:422)
  138. at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1875)
  139. at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2680)
  140. at org.apache.hadoop.ipc.Client.call(Client.java:1476)
  141. at org.apache.hadoop.ipc.Client.call(Client.java:1407)
  142. at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:229)
  143. at com.sun.proxy.$Proxy20.getListing(Unknown Source)
  144. at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.getListing(ClientNamenodeProtocolTranslatorPB.java:573)
  145. at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
  146. at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
  147. at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
  148. at java.lang.reflect.Method.invoke(Method.java:498)
  149. at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:187)
  150. at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:102)
  151. at com.sun.proxy.$Proxy21.getListing(Unknown Source)
  152. at org.apache.hadoop.hdfs.DFSClient.listPaths(DFSClient.java:2094)
  153. at org.apache.hadoop.hdfs.DFSClient.listPaths(DFSClient.java:2077)
  154. at org.apache.hadoop.hdfs.DistributedFileSystem.listStatusInternal(DistributedFileSystem.java:791)
  155. at org.apache.hadoop.hdfs.DistributedFileSystem.access$700(DistributedFileSystem.java:106)
  156. at org.apache.hadoop.hdfs.DistributedFileSystem$18.doCall(DistributedFileSystem.java:853)
  157. at org.apache.hadoop.hdfs.DistributedFileSystem$18.doCall(DistributedFileSystem.java:849)
  158. at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
  159. at org.apache.hadoop.hdfs.DistributedFileSystem.listStatus(DistributedFileSystem.java:860)
  160. at com.alibaba.datax.plugin.reader.hdfsreader.DFSUtil.getHDFSAllFilesNORegex(DFSUtil.java:162)
  161. at com.alibaba.datax.plugin.reader.hdfsreader.DFSUtil.getHDFSAllFiles(DFSUtil.java:141)
  162. at com.alibaba.datax.plugin.reader.hdfsreader.DFSUtil.getAllFiles(DFSUtil.java:112)
  163. at com.alibaba.datax.plugin.reader.hdfsreader.HdfsReader$Job.prepare(HdfsReader.java:169)
  164. at com.alibaba.datax.core.job.JobContainer.prepareJobReader(JobContainer.java:715)
  165. at com.alibaba.datax.core.job.JobContainer.prepare(JobContainer.java:308)
  166. at com.alibaba.datax.core.job.JobContainer.start(JobContainer.java:115)
  167. at com.alibaba.datax.core.Engine.start(Engine.java:92)
  168. at com.alibaba.datax.core.Engine.entry(Engine.java:171)
  169. at com.alibaba.datax.core.Engine.main(Engine.java:204)
  170. at com.alibaba.datax.common.exception.DataXException.asDataXException(DataXException.java:40) ~[datax-common-0.0.1-SNAPSHOT.jar:na]
  171. at com.alibaba.datax.plugin.reader.hdfsreader.DFSUtil.getHDFSAllFiles(DFSUtil.java:150) ~[hdfsreader-0.0.1-SNAPSHOT.jar:na]
  172. at com.alibaba.datax.plugin.reader.hdfsreader.DFSUtil.getAllFiles(DFSUtil.java:112) ~[hdfsreader-0.0.1-SNAPSHOT.jar:na]
  173. at com.alibaba.datax.plugin.reader.hdfsreader.HdfsReader$Job.prepare(HdfsReader.java:169) ~[hdfsreader-0.0.1-SNAPSHOT.jar:na]
  174. at com.alibaba.datax.core.job.JobContainer.prepareJobReader(JobContainer.java:715) ~[datax-core-0.0.1-SNAPSHOT.jar:na]
  175. at com.alibaba.datax.core.job.JobContainer.prepare(JobContainer.java:308) ~[datax-core-0.0.1-SNAPSHOT.jar:na]
  176. at com.alibaba.datax.core.job.JobContainer.start(JobContainer.java:115) ~[datax-core-0.0.1-SNAPSHOT.jar:na]
  177. at com.alibaba.datax.core.Engine.start(Engine.java:92) [datax-core-0.0.1-SNAPSHOT.jar:na]
  178. at com.alibaba.datax.core.Engine.entry(Engine.java:171) [datax-core-0.0.1-SNAPSHOT.jar:na]
  179. at com.alibaba.datax.core.Engine.main(Engine.java:204) [datax-core-0.0.1-SNAPSHOT.jar:na]
  180. Caused by: org.apache.hadoop.ipc.RemoteException: Operation category READ is not supported in state standby. Visit https://s.apache.org/sbnn-error
  181. at org.apache.hadoop.hdfs.server.namenode.ha.StandbyState.checkOperation(StandbyState.java:88)
  182. at org.apache.hadoop.hdfs.server.namenode.NameNode$NameNodeHAContext.checkOperation(NameNode.java:1962)
  183. at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkOperation(FSNamesystem.java:1421)
  184. at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getListing(FSNamesystem.java:3729)
  185. at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.getListing(NameNodeRpcServer.java:1138)
  186. at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.getListing(ClientNamenodeProtocolServerSideTranslatorPB.java:708)
  187. at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
  188. at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:523)
  189. at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:991)
  190. at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:870)
  191. at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:816)
  192. at java.security.AccessController.doPrivileged(Native Method)
  193. at javax.security.auth.Subject.doAs(Subject.java:422)
  194. at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1875)
  195. at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2680)
  196. at org.apache.hadoop.ipc.Client.call(Client.java:1476) ~[hadoop-common-2.7.1.jar:na]
  197. at org.apache.hadoop.ipc.Client.call(Client.java:1407) ~[hadoop-common-2.7.1.jar:na]
  198. at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:229) ~[hadoop-common-2.7.1.jar:na]
  199. at com.sun.proxy.$Proxy20.getListing(Unknown Source) ~[na:na]
  200. at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.getListing(ClientNamenodeProtocolTranslatorPB.java:573) ~[hadoop-hdfs-2.7.1.jar:na]
  201. at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[na:1.8.0_262]
  202. at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) ~[na:1.8.0_262]
  203. at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[na:1.8.0_262]
  204. at java.lang.reflect.Method.invoke(Method.java:498) ~[na:1.8.0_262]
  205. at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:187) ~[hadoop-common-2.7.1.jar:na]
  206. at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:102) ~[hadoop-common-2.7.1.jar:na]
  207. at com.sun.proxy.$Proxy21.getListing(Unknown Source) ~[na:na]
  208. at org.apache.hadoop.hdfs.DFSClient.listPaths(DFSClient.java:2094) ~[hadoop-hdfs-2.7.1.jar:na]
  209. at org.apache.hadoop.hdfs.DFSClient.listPaths(DFSClient.java:2077) ~[hadoop-hdfs-2.7.1.jar:na]
  210. at org.apache.hadoop.hdfs.DistributedFileSystem.listStatusInternal(DistributedFileSystem.java:791) ~[hadoop-hdfs-2.7.1.jar:na]
  211. at org.apache.hadoop.hdfs.DistributedFileSystem.access$700(DistributedFileSystem.java:106) ~[hadoop-hdfs-2.7.1.jar:na]
  212. at org.apache.hadoop.hdfs.DistributedFileSystem$18.doCall(DistributedFileSystem.java:853) ~[hadoop-hdfs-2.7.1.jar:na]
  213. at org.apache.hadoop.hdfs.DistributedFileSystem$18.doCall(DistributedFileSystem.java:849) ~[hadoop-hdfs-2.7.1.jar:na]
  214. at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-2.7.1.jar:na]
  215. at org.apache.hadoop.hdfs.DistributedFileSystem.listStatus(DistributedFileSystem.java:860) ~[hadoop-hdfs-2.7.1.jar:na]
  216. at com.alibaba.datax.plugin.reader.hdfsreader.DFSUtil.getHDFSAllFilesNORegex(DFSUtil.java:162) ~[hdfsreader-0.0.1-SNAPSHOT.jar:na]
  217. at com.alibaba.datax.plugin.reader.hdfsreader.DFSUtil.getHDFSAllFiles(DFSUtil.java:141) ~[hdfsreader-0.0.1-SNAPSHOT.jar:na]
  218. ... 8 common frames omitted
  219. 2022-07-17 22:44:24.726 [job-0] INFO StandAloneJobContainerCommunicator - Total 0 records, 0 bytes | Speed 0B/s, 0 records/s | Error 0 records, 0 bytes | All Task WaitWriterTime 0.000s | All Task WaitReaderTime 0.000s | Percentage 0.00%
  220. 2022-07-17 22:44:24.731 [job-0] ERROR Engine -
  221. 经DataX智能分析,该任务最可能的错误原因是:
  222. com.alibaba.datax.common.exception.DataXException: Code:[HdfsReader-09], Description:[您配置的path格式有误]. - org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.ipc.StandbyException): Operation category READ is not supported in state standby. Visit https://s.apache.org/sbnn-error
  223. at org.apache.hadoop.hdfs.server.namenode.ha.StandbyState.checkOperation(StandbyState.java:88)
  224. at org.apache.hadoop.hdfs.server.namenode.NameNode$NameNodeHAContext.checkOperation(NameNode.java:1962)
  225. at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkOperation(FSNamesystem.java:1421)
  226. at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getListing(FSNamesystem.java:3729)
  227. at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.getListing(NameNodeRpcServer.java:1138)
  228. at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.getListing(ClientNamenodeProtocolServerSideTranslatorPB.java:708)
  229. at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
  230. at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:523)
  231. at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:991)
  232. at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:870)
  233. at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:816)
  234. at java.security.AccessController.doPrivileged(Native Method)
  235. at javax.security.auth.Subject.doAs(Subject.java:422)
  236. at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1875)
  237. at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2680)
  238. at org.apache.hadoop.ipc.Client.call(Client.java:1476)
  239. at org.apache.hadoop.ipc.Client.call(Client.java:1407)
  240. at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:229)
  241. at com.sun.proxy.$Proxy20.getListing(Unknown Source)
  242. at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.getListing(ClientNamenodeProtocolTranslatorPB.java:573)
  243. at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
  244. at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
  245. at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
  246. at java.lang.reflect.Method.invoke(Method.java:498)
  247. at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:187)
  248. at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:102)
  249. at com.sun.proxy.$Proxy21.getListing(Unknown Source)
  250. at org.apache.hadoop.hdfs.DFSClient.listPaths(DFSClient.java:2094)
  251. at org.apache.hadoop.hdfs.DFSClient.listPaths(DFSClient.java:2077)
  252. at org.apache.hadoop.hdfs.DistributedFileSystem.listStatusInternal(DistributedFileSystem.java:791)
  253. at org.apache.hadoop.hdfs.DistributedFileSystem.access$700(DistributedFileSystem.java:106)
  254. at org.apache.hadoop.hdfs.DistributedFileSystem$18.doCall(DistributedFileSystem.java:853)
  255. at org.apache.hadoop.hdfs.DistributedFileSystem$18.doCall(DistributedFileSystem.java:849)
  256. at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
  257. at org.apache.hadoop.hdfs.DistributedFileSystem.listStatus(DistributedFileSystem.java:860)
  258. at com.alibaba.datax.plugin.reader.hdfsreader.DFSUtil.getHDFSAllFilesNORegex(DFSUtil.java:162)
  259. at com.alibaba.datax.plugin.reader.hdfsreader.DFSUtil.getHDFSAllFiles(DFSUtil.java:141)
  260. at com.alibaba.datax.plugin.reader.hdfsreader.DFSUtil.getAllFiles(DFSUtil.java:112)
  261. at com.alibaba.datax.plugin.reader.hdfsreader.HdfsReader$Job.prepare(HdfsReader.java:169)
  262. at com.alibaba.datax.core.job.JobContainer.prepareJobReader(JobContainer.java:715)
  263. at com.alibaba.datax.core.job.JobContainer.prepare(JobContainer.java:308)
  264. at com.alibaba.datax.core.job.JobContainer.start(JobContainer.java:115)
  265. at com.alibaba.datax.core.Engine.start(Engine.java:92)
  266. at com.alibaba.datax.core.Engine.entry(Engine.java:171)
  267. at com.alibaba.datax.core.Engine.main(Engine.java:204)
  268. - org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.ipc.StandbyException): Operation category READ is not supported in state standby. Visit https://s.apache.org/sbnn-error
  269. at org.apache.hadoop.hdfs.server.namenode.ha.StandbyState.checkOperation(StandbyState.java:88)
  270. at org.apache.hadoop.hdfs.server.namenode.NameNode$NameNodeHAContext.checkOperation(NameNode.java:1962)
  271. at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkOperation(FSNamesystem.java:1421)
  272. at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getListing(FSNamesystem.java:3729)
  273. at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.getListing(NameNodeRpcServer.java:1138)
  274. at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.getListing(ClientNamenodeProtocolServerSideTranslatorPB.java:708)
  275. at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
  276. at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:523)
  277. at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:991)
  278. at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:870)
  279. at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:816)
  280. at java.security.AccessController.doPrivileged(Native Method)
  281. at javax.security.auth.Subject.doAs(Subject.java:422)
  282. at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1875)
  283. at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2680)
  284. at org.apache.hadoop.ipc.Client.call(Client.java:1476)
  285. at org.apache.hadoop.ipc.Client.call(Client.java:1407)
  286. at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:229)
  287. at com.sun.proxy.$Proxy20.getListing(Unknown Source)
  288. at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.getListing(ClientNamenodeProtocolTranslatorPB.java:573)
  289. at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
  290. at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
  291. at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
  292. at java.lang.reflect.Method.invoke(Method.java:498)
  293. at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:187)
  294. at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:102)
  295. at com.sun.proxy.$Proxy21.getListing(Unknown Source)
  296. at org.apache.hadoop.hdfs.DFSClient.listPaths(DFSClient.java:2094)
  297. at org.apache.hadoop.hdfs.DFSClient.listPaths(DFSClient.java:2077)
  298. at org.apache.hadoop.hdfs.DistributedFileSystem.listStatusInternal(DistributedFileSystem.java:791)
  299. at org.apache.hadoop.hdfs.DistributedFileSystem.access$700(DistributedFileSystem.java:106)
  300. at org.apache.hadoop.hdfs.DistributedFileSystem$18.doCall(DistributedFileSystem.java:853)
  301. at org.apache.hadoop.hdfs.DistributedFileSystem$18.doCall(DistributedFileSystem.java:849)
  302. at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
  303. at org.apache.hadoop.hdfs.DistributedFileSystem.listStatus(DistributedFileSystem.java:860)
  304. at com.alibaba.datax.plugin.reader.hdfsreader.DFSUtil.getHDFSAllFilesNORegex(DFSUtil.java:162)
  305. at com.alibaba.datax.plugin.reader.hdfsreader.DFSUtil.getHDFSAllFiles(DFSUtil.java:141)
  306. at com.alibaba.datax.plugin.reader.hdfsreader.DFSUtil.getAllFiles(DFSUtil.java:112)
  307. at com.alibaba.datax.plugin.reader.hdfsreader.HdfsReader$Job.prepare(HdfsReader.java:169)
  308. at com.alibaba.datax.core.job.JobContainer.prepareJobReader(JobContainer.java:715)
  309. at com.alibaba.datax.core.job.JobContainer.prepare(JobContainer.java:308)
  310. at com.alibaba.datax.core.job.JobContainer.start(JobContainer.java:115)
  311. at com.alibaba.datax.core.Engine.start(Engine.java:92)
  312. at com.alibaba.datax.core.Engine.entry(Engine.java:171)
  313. at com.alibaba.datax.core.Engine.main(Engine.java:204)
  314. at com.alibaba.datax.common.exception.DataXException.asDataXException(DataXException.java:40)
  315. at com.alibaba.datax.plugin.reader.hdfsreader.DFSUtil.getHDFSAllFiles(DFSUtil.java:150)
  316. at com.alibaba.datax.plugin.reader.hdfsreader.DFSUtil.getAllFiles(DFSUtil.java:112)
  317. at com.alibaba.datax.plugin.reader.hdfsreader.HdfsReader$Job.prepare(HdfsReader.java:169)
  318. at com.alibaba.datax.core.job.JobContainer.prepareJobReader(JobContainer.java:715)
  319. at com.alibaba.datax.core.job.JobContainer.prepare(JobContainer.java:308)
  320. at com.alibaba.datax.core.job.JobContainer.start(JobContainer.java:115)
  321. at com.alibaba.datax.core.Engine.start(Engine.java:92)
  322. at com.alibaba.datax.core.Engine.entry(Engine.java:171)
  323. at com.alibaba.datax.core.Engine.main(Engine.java:204)
  324. Caused by: org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.ipc.StandbyException): Operation category READ is not supported in state standby. Visit https://s.apache.org/sbnn-error
  325. at org.apache.hadoop.hdfs.server.namenode.ha.StandbyState.checkOperation(StandbyState.java:88)
  326. at org.apache.hadoop.hdfs.server.namenode.NameNode$NameNodeHAContext.checkOperation(NameNode.java:1962)
  327. at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkOperation(FSNamesystem.java:1421)
  328. at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getListing(FSNamesystem.java:3729)
  329. at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.getListing(NameNodeRpcServer.java:1138)
  330. at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.getListing(ClientNamenodeProtocolServerSideTranslatorPB.java:708)
  331. at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
  332. at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:523)
  333. at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:991)
  334. at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:870)
  335. at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:816)
  336. at java.security.AccessController.doPrivileged(Native Method)
  337. at javax.security.auth.Subject.doAs(Subject.java:422)
  338. at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1875)
  339. at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2680)
  340. at org.apache.hadoop.ipc.Client.call(Client.java:1476)
  341. at org.apache.hadoop.ipc.Client.call(Client.java:1407)
  342. at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:229)
  343. at com.sun.proxy.$Proxy20.getListing(Unknown Source)
  344. at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.getListing(ClientNamenodeProtocolTranslatorPB.java:573)
  345. at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
  346. at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
  347. at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
  348. at java.lang.reflect.Method.invoke(Method.java:498)
  349. at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:187)
  350. at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:102)
  351. at com.sun.proxy.$Proxy21.getListing(Unknown Source)
  352. at org.apache.hadoop.hdfs.DFSClient.listPaths(DFSClient.java:2094)
  353. at org.apache.hadoop.hdfs.DFSClient.listPaths(DFSClient.java:2077)
  354. at org.apache.hadoop.hdfs.DistributedFileSystem.listStatusInternal(DistributedFileSystem.java:791)
  355. at org.apache.hadoop.hdfs.DistributedFileSystem.access$700(DistributedFileSystem.java:106)
  356. at org.apache.hadoop.hdfs.DistributedFileSystem$18.doCall(DistributedFileSystem.java:853)
  357. at org.apache.hadoop.hdfs.DistributedFileSystem$18.doCall(DistributedFileSystem.java:849)
  358. at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
  359. at org.apache.hadoop.hdfs.DistributedFileSystem.listStatus(DistributedFileSystem.java:860)
  360. at com.alibaba.datax.plugin.reader.hdfsreader.DFSUtil.getHDFSAllFilesNORegex(DFSUtil.java:162)
  361. at com.alibaba.datax.plugin.reader.hdfsreader.DFSUtil.getHDFSAllFiles(DFSUtil.java:141)
  362. ... 8 more