gpt4 book ai didi

hbase - 凤凰和查询超时

转载 作者:行者123 更新时间:2023-12-02 03:21:57 25 4
gpt4 key购买 nike

我有一个非常大的 hbase/phoenix 表,我正在尝试使用 sqlline 进行查询。我收到以下异常。我尝试在客户端和服务器上更改各种设置:

phoenix.query.keepAliveMs
phoenix.query.timeoutMs
dfs.client.socket-timeout

我尝试将以上所有设置为 6000000,但仍然遇到同样的问题。有任何想法吗?看起来原来的异常是 HBaseClient.CallTimeoutException。

sqlline 版本 1.1.80: jdbc:phoenix:nn.lan> 从 OUTPUT_TEST_TABLE 中选择计数 (*);

+------------------------------------------+
| COUNT(1) |
+------------------------------------------+
java.lang.RuntimeException: org.apache.phoenix.exception.PhoenixIOException: org.apache.phoenix.exception.PhoenixIOException: Failed after attempts=36, exceptions:
Tue Sep 22 09:22:40 BST 2015, null, java.net.SocketTimeoutException: callTimeout=60000, callDuration=60302: row '' on table 'OUTPUT_TEST_TABLE' at region=OUTPUT_TEST_TABLE,,1442867748366.5e70471615adc084cd0a60f943a654b0., hostname=d2.lan,16020,1442909697689, seqNum=194461

at sqlline.IncrementalRows.hasNext(IncrementalRows.java:73)
at sqlline.TableOutputFormat.print(TableOutputFormat.java:33)
at sqlline.SqlLine.print(SqlLine.java:1653)
at sqlline.Commands.execute(Commands.java:833)
at sqlline.Commands.sql(Commands.java:732)
at sqlline.SqlLine.dispatch(SqlLine.java:808)
at sqlline.SqlLine.begin(SqlLine.java:681)
at sqlline.SqlLine.start(SqlLine.java:398)
at sqlline.SqlLine.main(SqlLine.java:292)

0: jdbc:phoenix:nn.lan> 15/09/22 09:22:44 WARN client.ScannerCallable: Ignore, probably already closed
java.io.IOException: Call to d2.lan/192.168.0.30:16020 failed on local exception: org.apache.hadoop.hbase.ipc.CallTimeoutException: Call id=100, waitTime=60001, operationTimeout=60000 expired.
at org.apache.hadoop.hbase.ipc.RpcClientImpl.wrapException(RpcClientImpl.java:1232)
at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1200)
at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
at org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$BlockingStub.scan(ClientProtos.java:32651)
at org.apache.hadoop.hbase.client.ScannerCallable.close(ScannerCallable.java:355)
at org.apache.hadoop.hbase.client.ScannerCallable.call(ScannerCallable.java:195)
at org.apache.hadoop.hbase.client.ScannerCallableWithReplicas.call(ScannerCallableWithReplicas.java:142)
at org.apache.hadoop.hbase.client.ScannerCallableWithReplicas.call(ScannerCallableWithReplicas.java:61)
at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithoutRetries(RpcRetryingCaller.java:200)
at org.apache.hadoop.hbase.client.ClientScanner.call(ClientScanner.java:320)
at org.apache.hadoop.hbase.client.ClientScanner.nextScanner(ClientScanner.java:258)
at org.apache.hadoop.hbase.client.ClientScanner.possiblyNextScanner(ClientScanner.java:241)
at org.apache.hadoop.hbase.client.ClientScanner.loadCache(ClientScanner.java:532)
at org.apache.hadoop.hbase.client.ClientScanner.next(ClientScanner.java:364)
at org.apache.phoenix.iterate.ScanningResultIterator.next(ScanningResultIterator.java:55)
at org.apache.phoenix.iterate.TableResultIterator.next(TableResultIterator.java:107)
at org.apache.phoenix.iterate.SpoolingResultIterator.<init>(SpoolingResultIterator.java:125)
at org.apache.phoenix.iterate.SpoolingResultIterator.<init>(SpoolingResultIterator.java:83)
at org.apache.phoenix.iterate.SpoolingResultIterator.<init>(SpoolingResultIterator.java:62)
at org.apache.phoenix.iterate.SpoolingResultIterator$SpoolingResultIteratorFactory.newIterator(SpoolingResultIterator.java:78)
at org.apache.phoenix.iterate.ParallelIterators$1.call(ParallelIterators.java:109)
at org.apache.phoenix.iterate.ParallelIterators$1.call(ParallelIterators.java:100)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at org.apache.phoenix.job.JobManager$InstrumentedJobFutureTask.run(JobManager.java:183)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.ipc.CallTimeoutException: Call id=100, waitTime=60001, operationTimeout=60000 expired.
at org.apache.hadoop.hbase.ipc.Call.checkAndSetTimeout(Call.java:70)
at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1174)
... 26 more

15

最佳答案

尝试更改 HBase 的配置。属性 hbase.rpc.timeout。您必须在客户端增加此属性。

<property>
<name>hbase.rpc.timeout</name>
<value>360000000</value>
</property>

关于hbase - 凤凰和查询超时,我们在Stack Overflow上找到一个类似的问题: https://stackoverflow.com/questions/32712185/

25 4 0
Copyright 2021 - 2024 cfsdn All Rights Reserved 蜀ICP备2022000587号
广告合作:1813099741@qq.com 6ren.com