当前位置: 首页 > 知识库问答 >
问题:

phoenix表select join返回失败后尝试=11

终弘厚
2023-03-14
SELECT r.*, 
       l.productid, 
       l.prodname, 
       l.projnum, 
       l.projname, 
       l.capitalnumb, 
       l.fundname, 
       l.channelno, 
       l.channelname, 
       l.merchantno, 
       l.merchantname, 
       l.staff_name, 
       l.busipers 
FROM   lake.lake_retu_record r 
       LEFT JOIN lake.lake_loan l 
              ON r.loanacno = l.loanacno 
WHERE  r.acdate = '20200912000000' 
LIMIT  1000 

我在网上搜索了很长时间。但这并不能解决问题。请提供帮助或尝试提供一些如何解决此问题的想法。
我试图修改这些参数phoenix.query.queueesize phoenix.query.threadpoolsize phoenix.query.targetconcurrency phoenix.query.maxconcurrency

    20/10/18 10:35:50 WARN client.SyncCoprocessorRpcChannel: Call failed on IOException
    org.apache.hadoop.hbase.client.RetriesExhaustedException: Failed after attempts=11, exceptions:
    Sun Oct 18 10:35:08 CST 2020, RpcRetryingCaller{globalStartTime=1602988506012, pause=100, maxAttempts=11}, java.io.IOException: Call to CQ-HM-ALIFC-test/10.10.10.10:16020 failed on local exception: java.io.IOException: Connection reset by peer
    Sun Oct 18 10:35:08 CST 2020, RpcRetryingCaller{globalStartTime=1602988506012, pause=100, maxAttempts=11}, org.apache.hadoop.hbase.exceptions.ConnectionClosedException: Call to CQ-HM-ALIFC-test/10.10.10.10:16020 failed on local exception: org.apache.hadoop.hbase.exceptions.ConnectionClosedException: Connection closed
    Sun Oct 18 10:35:09 CST 2020, RpcRetryingCaller{globalStartTime=1602988506012, pause=100, maxAttempts=11}, org.apache.hadoop.hbase.exceptions.ConnectionClosedException: Call to CQ-HM-ALIFC-test/10.10.10.10:16020 failed on local exception: org.apache.hadoop.hbase.exceptions.ConnectionClosedException: Connection closed
    Sun Oct 18 10:35:10 CST 2020, RpcRetryingCaller{globalStartTime=1602988506012, pause=100, maxAttempts=11}, org.apache.hadoop.hbase.exceptions.ConnectionClosedException: Call to CQ-HM-ALIFC-test/10.10.10.10:16020 failed on local exception: org.apache.hadoop.hbase.exceptions.ConnectionClosedException: Connection closed
    Sun Oct 18 10:35:11 CST 2020, RpcRetryingCaller{globalStartTime=1602988506012, pause=100, maxAttempts=11}, org.apache.hadoop.hbase.exceptions.ConnectionClosedException: Call to CQ-HM-ALIFC-test/10.10.10.10:16020 failed on local exception: org.apache.hadoop.hbase.exceptions.ConnectionClosedException: Connection closed
    Sun Oct 18 10:35:12 CST 2020, RpcRetryingCaller{globalStartTime=1602988506012, pause=100, maxAttempts=11}, org.apache.hadoop.hbase.exceptions.ConnectionClosedException: Call to CQ-HM-ALIFC-test/10.10.10.10:16020 failed on local exception: org.apache.hadoop.hbase.exceptions.ConnectionClosedException: Connection closed
    Sun Oct 18 10:35:15 CST 2020, RpcRetryingCaller{globalStartTime=1602988506012, pause=100, maxAttempts=11}, org.apache.hadoop.hbase.exceptions.ConnectionClosedException: Call to CQ-HM-ALIFC-test/10.10.10.10:16020 failed on local exception: org.apache.hadoop.hbase.exceptions.ConnectionClosedException: Connection closed
    Sun Oct 18 10:35:19 CST 2020, RpcRetryingCaller{globalStartTime=1602988506012, pause=100, maxAttempts=11}, org.apache.hadoop.hbase.exceptions.ConnectionClosedException: Call to CQ-HM-ALIFC-test/10.10.10.10:16020 failed on local exception: org.apache.hadoop.hbase.exceptions.ConnectionClosedException: Connection closed
    Sun Oct 18 10:35:29 CST 2020, RpcRetryingCaller{globalStartTime=1602988506012, pause=100, maxAttempts=11}, org.apache.hadoop.hbase.exceptions.ConnectionClosedException: Call to CQ-HM-ALIFC-test/10.10.10.10:16020 failed on local exception: org.apache.hadoop.hbase.exceptions.ConnectionClosedException: Connection closed
    Sun Oct 18 10:35:40 CST 2020, RpcRetryingCaller{globalStartTime=1602988506012, pause=100, maxAttempts=11}, org.apache.hadoop.hbase.exceptions.ConnectionClosedException: Call to CQ-HM-ALIFC-test/10.10.10.10:16020 failed on local exception: org.apache.hadoop.hbase.exceptions.ConnectionClosedException: Connection closed
    Sun Oct 18 10:35:50 CST 2020, RpcRetryingCaller{globalStartTime=1602988506012, pause=100, maxAttempts=11}, org.apache.hadoop.hbase.exceptions.ConnectionClosedException: Call to CQ-HM-ALIFC-test/10.10.10.10:16020 failed on local exception: org.apache.hadoop.hbase.exceptions.ConnectionClosedException: Connection closed
    
        at org.apache.hadoop.hbase.client.RpcRetryingCallerImpl.callWithRetries(RpcRetryingCallerImpl.java:145)
        at org.apache.hadoop.hbase.client.RegionCoprocessorRpcChannel.callExecService(RegionCoprocessorRpcChannel.java:91)
        at org.apache.hadoop.hbase.client.SyncCoprocessorRpcChannel.callMethod(SyncCoprocessorRpcChannel.java:52)
        at org.apache.phoenix.coprocessor.generated.ServerCachingProtos$ServerCachingService$Stub.addServerCache(ServerCachingProtos.java:8484)
        at org.apache.phoenix.cache.ServerCacheClient$3.call(ServerCacheClient.java:494)
        at org.apache.phoenix.cache.ServerCacheClient$3.call(ServerCacheClient.java:462)
        at org.apache.hadoop.hbase.client.HTable$12.call(HTable.java:997)
        at java.util.concurrent.FutureTask.run(FutureTask.java:266)
        at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
        at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
        at java.lang.Thread.run(Thread.java:748)
    Caused by: org.apache.hadoop.hbase.exceptions.ConnectionClosedException: Call to CQ-HM-ALIFC-test/10.10.10.10:16020 failed on local exception: org.apache.hadoop.hbase.exceptions.ConnectionClosedException: Connection closed
        at org.apache.hadoop.hbase.ipc.IPCUtil.wrapException(IPCUtil.java:202)
        at org.apache.hadoop.hbase.ipc.AbstractRpcClient.onCallFinished(AbstractRpcClient.java:390)
        at org.apache.hadoop.hbase.ipc.AbstractRpcClient.access$100(AbstractRpcClient.java:95)
        at org.apache.hadoop.hbase.ipc.AbstractRpcClient$3.run(AbstractRpcClient.java:410)
        at org.apache.hadoop.hbase.ipc.AbstractRpcClient$3.run(AbstractRpcClient.java:406)
        at org.apache.hadoop.hbase.ipc.Call.callComplete(Call.java:103)
        at org.apache.hadoop.hbase.ipc.Call.setException(Call.java:118)
        at org.apache.hadoop.hbase.ipc.NettyRpcDuplexHandler.cleanupCalls(NettyRpcDuplexHandler.java:203)
        at org.apache.hadoop.hbase.ipc.NettyRpcDuplexHandler.channelInactive(NettyRpcDuplexHandler.java:211)
        at org.apache.hbase.thirdparty.io.netty.channel.AbstractChannelHandlerContext.invokeChannelInactive(AbstractChannelHandlerContext.java:245)
        at org.apache.hbase.thirdparty.io.netty.channel.AbstractChannelHandlerContext.invokeChannelInactive(AbstractChannelHandlerContext.java:231)
        at org.apache.hbase.thirdparty.io.netty.channel.AbstractChannelHandlerContext.fireChannelInactive(AbstractChannelHandlerContext.java:224)
        at org.apache.hbase.thirdparty.io.netty.handler.codec.ByteToMessageDecoder.channelInputClosed(ByteToMessageDecoder.java:377)
        at org.apache.hbase.thirdparty.io.netty.handler.codec.ByteToMessageDecoder.channelInactive(ByteToMessageDecoder.java:342)
        at org.apache.hbase.thirdparty.io.netty.channel.AbstractChannelHandlerContext.invokeChannelInactive(AbstractChannelHandlerContext.java:245)
        at org.apache.hbase.thirdparty.io.netty.channel.AbstractChannelHandlerContext.invokeChannelInactive(AbstractChannelHandlerContext.java:231)
        at org.apache.hbase.thirdparty.io.netty.channel.AbstractChannelHandlerContext.fireChannelInactive(AbstractChannelHandlerContext.java:224)
        at org.apache.hbase.thirdparty.io.netty.channel.ChannelInboundHandlerAdapter.channelInactive(ChannelInboundHandlerAdapter.java:75)
        at org.apache.hbase.thirdparty.io.netty.handler.timeout.IdleStateHandler.channelInactive(IdleStateHandler.java:277)
        at org.apache.hbase.thirdparty.io.netty.channel.AbstractChannelHandlerContext.invokeChannelInactive(AbstractChannelHandlerContext.java:245)
        at org.apache.hbase.thirdparty.io.netty.channel.AbstractChannelHandlerContext.invokeChannelInactive(AbstractChannelHandlerContext.java:231)
        at org.apache.hbase.thirdparty.io.netty.channel.AbstractChannelHandlerContext.fireChannelInactive(AbstractChannelHandlerContext.java:224)
        at org.apache.hbase.thirdparty.io.netty.channel.DefaultChannelPipeline$HeadContext.channelInactive(DefaultChannelPipeline.java:1354)
        at org.apache.hbase.thirdparty.io.netty.channel.AbstractChannelHandlerContext.invokeChannelInactive(AbstractChannelHandlerContext.java:245)
        at org.apache.hbase.thirdparty.io.netty.channel.AbstractChannelHandlerContext.invokeChannelInactive(AbstractChannelHandlerContext.java:231)
        at org.apache.hbase.thirdparty.io.netty.channel.DefaultChannelPipeline.fireChannelInactive(DefaultChannelPipeline.java:917)
        at org.apache.hbase.thirdparty.io.netty.channel.AbstractChannel$AbstractUnsafe$8.run(AbstractChannel.java:822)
        at org.apache.hbase.thirdparty.io.netty.util.concurrent.AbstractEventExecutor.safeExecute(AbstractEventExecutor.java:163)
        at org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor.runAllTasks(SingleThreadEventExecutor.java:403)
        at org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:463)
        at org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$5.run(SingleThreadEventExecutor.java:858)
        at org.apache.hbase.thirdparty.io.netty.util.concurrent.DefaultThreadFactory$DefaultRunnableDecorator.run(DefaultThreadFactory.java:138)
        ... 1 more
    Caused by: org.apache.hadoop.hbase.exceptions.ConnectionClosedException: Connection closed
        ... 25 more

共有1个答案

叶鸿煊
2023-03-14

解决办法如下:
我找到了异常的主要原因

org.apache.hadoop.hbase.client.RetriesExhaustedException: Failed after attempts=11, exceptions:
Sun Oct 18 10:35:08 CST 2020, RpcRetryingCaller{globalStartTime=1602988506012, pause=100, maxAttempts=11}, java.io.IOException: Call to CQ-HM-ALIFC-test/10.10.10.10:16020 failed on local exception: java.io.IOException: Connection reset by peer

初步判断是RPC有问题,所以当我执行SQL语句时,我打开报告错误的服务器的RegionServer日志。我发现RegionServer生成以下日志:

WARN org.apache.hadoop.hbase.ipc.NettyRpcServer: RPC data length of 353247227 received from 10.10.10.10 is greater than max allowed 268435456. Set "hbase.ipc.max.request.size" on server to override this limit (not recommended)

我增加了参数,解决了问题!

 类似资料:
  • 问题内容: 嗯,我有一个金钱对象,可以将其他金钱对象添加到其中。我在Java中尝试测试我的代码是否还可以,但是随后失败了。 我非常肯定自己的代码是正确的(返回正确的答案),我认为我使用的是错误的方式。T_T 如果要查找是否要进行测试,该怎么使用? 问题答案: 您没有在Money类中重写Object类中的方法。如果是这样,则通过它们的引用比较对象,在这种情况下,引用是不同的。 在这里您可以找到实施规

  • HBase和Hadoop在Docker中是CDH独立模式。ZK、HBASE、Phoenix使用shell都可以操作。java操作zk也可以,但是java不能操作Hbase和phoenix,代码可以。有人能帮我吗,谢谢!

  • 我在HBASE有一张桌子。我和菲尼克斯一起创造的 然后我用spark填充它(我使用hortonworks shc-core连接器): 多谢帮忙。

  • 问题内容: 我试图在Windows 2008 R2中使用Kibana运行ElasticSearch。 我关注了这篇文章:在带有木瓜的Windows服务器上安装logstash 一步一步,但我得到的是: 当我去 我得到: 因此,似乎ElasticSearch正在运行,但是由于某些原因,Kibana无法连接到它。 ElasticSearch日志包含错误: 知道我在做什么错吗? 问题答案: 我也遇到过类

  • 我有一个,它扩展了一个名为的基本所有这些所做的就是创建一个位置服务客户端 和

  • 我有一个返回< code>List的方法。现在我想知道如何正确放置< code>try/catch块。如果我将< code>return语句放在< code>try中,我会得到错误 并非所有代码路径都返回值 如果我放置在之后(就像我目前所做的那样),即使在之后,它也会返回。那么最好的方法应该是什么? 方法如下: