Skip to content

Commit 22ae9c1

Browse files
committed
Fixes the issue that ShardingSphere cannot connect to HiveServer2 using remote Hive Metastore Server
1 parent 4e5b0ce commit 22ae9c1

File tree

7 files changed

+298
-10
lines changed

7 files changed

+298
-10
lines changed

docs/document/content/user-manual/shardingsphere-jdbc/optional-plugins/hiveserver2/_index.cn.md

+47-2
Original file line numberDiff line numberDiff line change
@@ -40,6 +40,17 @@ ShardingSphere 对 HiveServer2 JDBC Driver 的支持位于可选模块中。
4040
<artifactId>hive-service</artifactId>
4141
<version>4.0.1</version>
4242
</dependency>
43+
<dependency>
44+
<groupId>org.apache.hadoop</groupId>
45+
<artifactId>hadoop-mapreduce-client-core</artifactId>
46+
<version>3.3.6</version>
47+
<exclusions>
48+
<exclusion>
49+
<groupId>*</groupId>
50+
<artifactId>*</artifactId>
51+
</exclusion>
52+
</exclusions>
53+
</dependency>
4354
</dependencies>
4455
```
4556

@@ -81,6 +92,17 @@ ShardingSphere 对 HiveServer2 JDBC Driver 的支持位于可选模块中。
8192
</exclusion>
8293
</exclusions>
8394
</dependency>
95+
<dependency>
96+
<groupId>org.apache.hadoop</groupId>
97+
<artifactId>hadoop-mapreduce-client-core</artifactId>
98+
<version>3.3.6</version>
99+
<exclusions>
100+
<exclusion>
101+
<groupId>*</groupId>
102+
<artifactId>*</artifactId>
103+
</exclusion>
104+
</exclusions>
105+
</dependency>
84106
</dependencies>
85107
```
86108

@@ -427,8 +449,31 @@ ShardingSphere 仅针对 HiveServer2 `4.0.1` 进行集成测试。
427449
### Hadoop 限制
428450

429451
用户仅可使用 Hadoop `3.3.6` 来作为 HiveServer2 JDBC Driver `4.0.1` 的底层 Hadoop 依赖。
430-
HiveServer2 JDBC Driver `4.0.1` 不支持 Hadoop `3.4.1`,
431-
参考 https://github.com/apache/hive/pull/5500 。
452+
HiveServer2 JDBC Driver `4.0.1` 不支持 Hadoop `3.4.1`, 参考 https://github.com/apache/hive/pull/5500 。
453+
454+
对于 HiveServer2 JDBC Driver `org.apache.hive:hive-jdbc:4.0.1` 或 `classifier` 为 `standalone` 的 `org.apache.hive:hive-jdbc:4.0.1`,
455+
实际上并不额外依赖 `org.apache.hadoop:hadoop-mapreduce-client-core:3.3.6`。
456+
457+
但 `org.apache.shardingsphere:shardingsphere-infra-database-hive` 的
458+
`org.apache.shardingsphere.infra.database.hive.metadata.data.loader.HiveMetaDataLoader` 会使用 `org.apache.hadoop.hive.conf.HiveConf`,
459+
这进一步使用了 `org.apache.hadoop:hadoop-mapreduce-client-core:3.3.6` 的 `org.apache.hadoop.mapred.JobConf` 类。
460+
461+
ShardingSphere 仅需要使用 `org.apache.hadoop.mapred.JobConf` 类,
462+
因此排除 `org.apache.hadoop:hadoop-mapreduce-client-core:3.3.6` 的所有额外依赖是合理行为。
463+
464+
```xml
465+
<dependency>
466+
<groupId>org.apache.hadoop</groupId>
467+
<artifactId>hadoop-mapreduce-client-core</artifactId>
468+
<version>3.3.6</version>
469+
<exclusions>
470+
<exclusion>
471+
<groupId>*</groupId>
472+
<artifactId>*</artifactId>
473+
</exclusion>
474+
</exclusions>
475+
</dependency>
476+
```
432477

433478
### SQL 限制
434479

docs/document/content/user-manual/shardingsphere-jdbc/optional-plugins/hiveserver2/_index.en.md

+47-2
Original file line numberDiff line numberDiff line change
@@ -41,6 +41,17 @@ The possible Maven dependencies are as follows.
4141
<artifactId>hive-service</artifactId>
4242
<version>4.0.1</version>
4343
</dependency>
44+
<dependency>
45+
<groupId>org.apache.hadoop</groupId>
46+
<artifactId>hadoop-mapreduce-client-core</artifactId>
47+
<version>3.3.6</version>
48+
<exclusions>
49+
<exclusion>
50+
<groupId>*</groupId>
51+
<artifactId>*</artifactId>
52+
</exclusion>
53+
</exclusions>
54+
</dependency>
4455
</dependencies>
4556
```
4657

@@ -83,6 +94,17 @@ The following is an example of a possible configuration,
8394
</exclusion>
8495
</exclusions>
8596
</dependency>
97+
<dependency>
98+
<groupId>org.apache.hadoop</groupId>
99+
<artifactId>hadoop-mapreduce-client-core</artifactId>
100+
<version>3.3.6</version>
101+
<exclusions>
102+
<exclusion>
103+
<groupId>*</groupId>
104+
<artifactId>*</artifactId>
105+
</exclusion>
106+
</exclusions>
107+
</dependency>
86108
</dependencies>
87109
```
88110

@@ -433,8 +455,31 @@ Reference https://issues.apache.org/jira/browse/HIVE-28418.
433455
### Hadoop Limitations
434456

435457
Users can only use Hadoop `3.3.6` as the underlying Hadoop dependency of HiveServer2 JDBC Driver `4.0.1`.
436-
HiveServer2 JDBC Driver `4.0.1` does not support Hadoop `3.4.1`,
437-
Reference https://github.com/apache/hive/pull/5500.
458+
HiveServer2 JDBC Driver `4.0.1` does not support Hadoop `3.4.1`. Reference https://github.com/apache/hive/pull/5500 .
459+
460+
For HiveServer2 JDBC Driver `org.apache.hive:hive-jdbc:4.0.1` or `org.apache.hive:hive-jdbc:4.0.1` with `classifier` as `standalone`,
461+
there is actually no additional dependency on `org.apache.hadoop:hadoop-mapreduce-client-core:3.3.6`.
462+
463+
But `org.apache.shardingsphere:shardingsphere-infra-database-hive`'s
464+
`org.apache.shardingsphere.infra.database.hive.metadata.data.loader.HiveMetaDataLoader` uses `org.apache.hadoop.hive.conf.HiveConf`,
465+
which further uses `org.apache.hadoop:hadoop-mapreduce-client-core:3.3.6`'s `org.apache.hadoop.mapred.JobConf` class.
466+
467+
ShardingSphere only needs to use the `org.apache.hadoop.mapred.JobConf` class,
468+
so it is reasonable to exclude all additional dependencies of `org.apache.hadoop:hadoop-mapreduce-client-core:3.3.6`.
469+
470+
```xml
471+
<dependency>
472+
<groupId>org.apache.hadoop</groupId>
473+
<artifactId>hadoop-mapreduce-client-core</artifactId>
474+
<version>3.3.6</version>
475+
<exclusions>
476+
<exclusion>
477+
<groupId>*</groupId>
478+
<artifactId>*</artifactId>
479+
</exclusion>
480+
</exclusions>
481+
</dependency>
482+
```
438483

439484
### SQL Limitations
440485

pom.xml

+1
Original file line numberDiff line numberDiff line change
@@ -129,6 +129,7 @@
129129
<clickhouse-jdbc.version>0.6.3</clickhouse-jdbc.version>
130130
<hive.version>4.0.1</hive.version>
131131
<hive-server2-jdbc-driver-thin.version>1.5.0</hive-server2-jdbc-driver-thin.version>
132+
<hadoop.version>3.3.6</hadoop.version>
132133
<presto.version>0.288.1</presto.version>
133134

134135
<hikari-cp.version>4.0.3</hikari-cp.version>

test/native/pom.xml

+11
Original file line numberDiff line numberDiff line change
@@ -190,6 +190,17 @@
190190
</exclusion>
191191
</exclusions>
192192
</dependency>
193+
<dependency>
194+
<groupId>org.apache.hadoop</groupId>
195+
<artifactId>hadoop-mapreduce-client-core</artifactId>
196+
<version>${hadoop.version}</version>
197+
<exclusions>
198+
<exclusion>
199+
<groupId>*</groupId>
200+
<artifactId>*</artifactId>
201+
</exclusion>
202+
</exclusions>
203+
</dependency>
193204
<dependency>
194205
<groupId>org.testcontainers</groupId>
195206
<artifactId>junit-jupiter</artifactId>
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,122 @@
1+
/*
2+
* Licensed to the Apache Software Foundation (ASF) under one or more
3+
* contributor license agreements. See the NOTICE file distributed with
4+
* this work for additional information regarding copyright ownership.
5+
* The ASF licenses this file to You under the Apache License, Version 2.0
6+
* (the "License"); you may not use this file except in compliance with
7+
* the License. You may obtain a copy of the License at
8+
*
9+
* http://www.apache.org/licenses/LICENSE-2.0
10+
*
11+
* Unless required by applicable law or agreed to in writing, software
12+
* distributed under the License is distributed on an "AS IS" BASIS,
13+
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
14+
* See the License for the specific language governing permissions and
15+
* limitations under the License.
16+
*/
17+
18+
package org.apache.shardingsphere.test.natived.jdbc.databases.hive;
19+
20+
import com.zaxxer.hikari.HikariConfig;
21+
import com.zaxxer.hikari.HikariDataSource;
22+
import org.apache.shardingsphere.test.natived.commons.TestShardingService;
23+
import org.awaitility.Awaitility;
24+
import org.junit.jupiter.api.AfterAll;
25+
import org.junit.jupiter.api.BeforeAll;
26+
import org.junit.jupiter.api.Test;
27+
import org.junit.jupiter.api.condition.EnabledInNativeImage;
28+
import org.testcontainers.containers.GenericContainer;
29+
import org.testcontainers.containers.Network;
30+
import org.testcontainers.junit.jupiter.Container;
31+
import org.testcontainers.junit.jupiter.Testcontainers;
32+
33+
import javax.sql.DataSource;
34+
import java.nio.file.Paths;
35+
import java.sql.Connection;
36+
import java.sql.DriverManager;
37+
import java.sql.SQLException;
38+
import java.sql.Statement;
39+
import java.time.Duration;
40+
import java.util.Properties;
41+
42+
import static org.hamcrest.MatcherAssert.assertThat;
43+
import static org.hamcrest.Matchers.is;
44+
import static org.hamcrest.Matchers.nullValue;
45+
46+
@SuppressWarnings({"SqlDialectInspection", "SqlNoDataSourceInspection", "resource"})
47+
@EnabledInNativeImage
48+
@Testcontainers
49+
class StandaloneMetastoreTest {
50+
51+
private static final Network NETWORK = Network.newNetwork();
52+
53+
@Container
54+
public static final GenericContainer<?> HMS_CONTAINER = new GenericContainer<>("apache/hive:4.0.1")
55+
.withEnv("SERVICE_NAME", "metastore")
56+
.withNetwork(NETWORK)
57+
.withNetworkAliases("metastore");
58+
59+
@Container
60+
public static final GenericContainer<?> HS2_CONTAINER = new GenericContainer<>("apache/hive:4.0.1")
61+
.withEnv("SERVICE_NAME", "hiveserver2")
62+
.withEnv("SERVICE_OPTS", "-Dhive.metastore.uris=thrift://metastore:9083")
63+
.withNetwork(NETWORK)
64+
.withExposedPorts(10000)
65+
.dependsOn(HMS_CONTAINER);
66+
67+
private static final String SYSTEM_PROP_KEY_PREFIX = "fixture.test-native.yaml.database.hive.hms.";
68+
69+
// Due to https://issues.apache.org/jira/browse/HIVE-28317 , the `initFile` parameter of HiveServer2 JDBC Driver must be an absolute path.
70+
private static final String ABSOLUTE_PATH = Paths.get("src/test/resources/test-native/sql/test-native-databases-hive-iceberg.sql").toAbsolutePath().toString();
71+
72+
private String jdbcUrlPrefix;
73+
74+
@BeforeAll
75+
static void beforeAll() {
76+
assertThat(System.getProperty(SYSTEM_PROP_KEY_PREFIX + "ds0.jdbc-url"), is(nullValue()));
77+
assertThat(System.getProperty(SYSTEM_PROP_KEY_PREFIX + "ds1.jdbc-url"), is(nullValue()));
78+
assertThat(System.getProperty(SYSTEM_PROP_KEY_PREFIX + "ds2.jdbc-url"), is(nullValue()));
79+
}
80+
81+
@AfterAll
82+
static void afterAll() {
83+
NETWORK.close();
84+
System.clearProperty(SYSTEM_PROP_KEY_PREFIX + "ds0.jdbc-url");
85+
System.clearProperty(SYSTEM_PROP_KEY_PREFIX + "ds1.jdbc-url");
86+
System.clearProperty(SYSTEM_PROP_KEY_PREFIX + "ds2.jdbc-url");
87+
}
88+
89+
@Test
90+
void assertShardingInLocalTransactions() throws SQLException {
91+
jdbcUrlPrefix = "jdbc:hive2://localhost:" + HS2_CONTAINER.getMappedPort(10000) + "/";
92+
DataSource dataSource = createDataSource();
93+
TestShardingService testShardingService = new TestShardingService(dataSource);
94+
testShardingService.processSuccessInHive();
95+
}
96+
97+
private Connection openConnection() throws SQLException {
98+
Properties props = new Properties();
99+
return DriverManager.getConnection(jdbcUrlPrefix, props);
100+
}
101+
102+
private DataSource createDataSource() throws SQLException {
103+
Awaitility.await().atMost(Duration.ofMinutes(1L)).ignoreExceptions().until(() -> {
104+
openConnection().close();
105+
return true;
106+
});
107+
try (
108+
Connection connection = openConnection();
109+
Statement statement = connection.createStatement()) {
110+
statement.executeUpdate("CREATE DATABASE demo_ds_0");
111+
statement.executeUpdate("CREATE DATABASE demo_ds_1");
112+
statement.executeUpdate("CREATE DATABASE demo_ds_2");
113+
}
114+
HikariConfig config = new HikariConfig();
115+
config.setDriverClassName("org.apache.shardingsphere.driver.ShardingSphereDriver");
116+
config.setJdbcUrl("jdbc:shardingsphere:classpath:test-native/yaml/jdbc/databases/hive/standalone-hms.yaml?placeholder-type=system_props");
117+
System.setProperty(SYSTEM_PROP_KEY_PREFIX + "ds0.jdbc-url", jdbcUrlPrefix + "demo_ds_0" + ";initFile=" + ABSOLUTE_PATH);
118+
System.setProperty(SYSTEM_PROP_KEY_PREFIX + "ds1.jdbc-url", jdbcUrlPrefix + "demo_ds_1" + ";initFile=" + ABSOLUTE_PATH);
119+
System.setProperty(SYSTEM_PROP_KEY_PREFIX + "ds2.jdbc-url", jdbcUrlPrefix + "demo_ds_2" + ";initFile=" + ABSOLUTE_PATH);
120+
return new HikariDataSource(config);
121+
}
122+
}

test/native/src/test/java/org/apache/shardingsphere/test/natived/jdbc/databases/hive/ZookeeperServiceDiscoveryTest.java

+6-6
Original file line numberDiff line numberDiff line change
@@ -71,7 +71,7 @@ class ZookeeperServiceDiscoveryTest {
7171
*/
7272
@SuppressWarnings("unused")
7373
@Container
74-
private static final GenericContainer<?> HIVE_SERVER2_1_CONTAINER = new FixedHostPortGenericContainer<>("apache/hive:4.0.1")
74+
private static final GenericContainer<?> HS2_1_CONTAINER = new FixedHostPortGenericContainer<>("apache/hive:4.0.1")
7575
.withNetwork(NETWORK)
7676
.withEnv("SERVICE_NAME", "hiveserver2")
7777
.withEnv("SERVICE_OPTS", "-Dhive.server2.support.dynamic.service.discovery=true" + " "
@@ -116,10 +116,10 @@ void assertShardingInLocalTransactions() throws SQLException {
116116
DataSource dataSource = createDataSource();
117117
TestShardingService testShardingService = new TestShardingService(dataSource);
118118
testShardingService.processSuccessInHive();
119-
HIVE_SERVER2_1_CONTAINER.stop();
119+
HS2_1_CONTAINER.stop();
120120
int randomPortSecond = InstanceSpec.getRandomPort();
121121
try (
122-
GenericContainer<?> hiveServer2SecondContainer = new FixedHostPortGenericContainer<>("apache/hive:4.0.1")
122+
GenericContainer<?> hs2SecondContainer = new FixedHostPortGenericContainer<>("apache/hive:4.0.1")
123123
.withNetwork(NETWORK)
124124
.withEnv("SERVICE_NAME", "hiveserver2")
125125
.withEnv("SERVICE_OPTS", "-Dhive.server2.support.dynamic.service.discovery=true" + " "
@@ -128,8 +128,8 @@ void assertShardingInLocalTransactions() throws SQLException {
128128
+ "-Dhive.server2.thrift.port=" + randomPortSecond)
129129
.withFixedExposedPort(randomPortSecond, randomPortSecond)
130130
.dependsOn(ZOOKEEPER_CONTAINER)) {
131-
hiveServer2SecondContainer.start();
132-
extracted(hiveServer2SecondContainer.getMappedPort(randomPortSecond));
131+
hs2SecondContainer.start();
132+
extracted(hs2SecondContainer.getMappedPort(randomPortSecond));
133133
testShardingService.processSuccessInHive();
134134
}
135135
}
@@ -140,7 +140,7 @@ private Connection openConnection() throws SQLException {
140140
}
141141

142142
private DataSource createDataSource() throws SQLException {
143-
extracted(HIVE_SERVER2_1_CONTAINER.getMappedPort(RANDOM_PORT_FIRST));
143+
extracted(HS2_1_CONTAINER.getMappedPort(RANDOM_PORT_FIRST));
144144
HikariConfig config = new HikariConfig();
145145
config.setDriverClassName("org.apache.shardingsphere.driver.ShardingSphereDriver");
146146
config.setJdbcUrl("jdbc:shardingsphere:classpath:test-native/yaml/jdbc/databases/hive/zsd.yaml?placeholder-type=system_props");
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,64 @@
1+
#
2+
# Licensed to the Apache Software Foundation (ASF) under one or more
3+
# contributor license agreements. See the NOTICE file distributed with
4+
# this work for additional information regarding copyright ownership.
5+
# The ASF licenses this file to You under the Apache License, Version 2.0
6+
# (the "License"); you may not use this file except in compliance with
7+
# the License. You may obtain a copy of the License at
8+
#
9+
# http://www.apache.org/licenses/LICENSE-2.0
10+
#
11+
# Unless required by applicable law or agreed to in writing, software
12+
# distributed under the License is distributed on an "AS IS" BASIS,
13+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
14+
# See the License for the specific language governing permissions and
15+
# limitations under the License.
16+
#
17+
18+
dataSources:
19+
ds_0:
20+
dataSourceClassName: com.zaxxer.hikari.HikariDataSource
21+
driverClassName: org.apache.hive.jdbc.HiveDriver
22+
jdbcUrl: $${fixture.test-native.yaml.database.hive.hms.ds0.jdbc-url::}
23+
ds_1:
24+
dataSourceClassName: com.zaxxer.hikari.HikariDataSource
25+
driverClassName: org.apache.hive.jdbc.HiveDriver
26+
jdbcUrl: $${fixture.test-native.yaml.database.hive.hms.ds1.jdbc-url::}
27+
ds_2:
28+
dataSourceClassName: com.zaxxer.hikari.HikariDataSource
29+
driverClassName: org.apache.hive.jdbc.HiveDriver
30+
jdbcUrl: $${fixture.test-native.yaml.database.hive.hms.ds2.jdbc-url::}
31+
32+
rules:
33+
- !SHARDING
34+
tables:
35+
t_order:
36+
actualDataNodes: <LITERAL>ds_0.t_order, ds_1.t_order, ds_2.t_order
37+
keyGenerateStrategy:
38+
column: order_id
39+
keyGeneratorName: snowflake
40+
t_order_item:
41+
actualDataNodes: <LITERAL>ds_0.t_order_item, ds_1.t_order_item, ds_2.t_order_item
42+
keyGenerateStrategy:
43+
column: order_item_id
44+
keyGeneratorName: snowflake
45+
defaultDatabaseStrategy:
46+
standard:
47+
shardingColumn: user_id
48+
shardingAlgorithmName: inline
49+
shardingAlgorithms:
50+
inline:
51+
type: CLASS_BASED
52+
props:
53+
strategy: STANDARD
54+
algorithmClassName: org.apache.shardingsphere.test.natived.commons.algorithm.ClassBasedInlineShardingAlgorithmFixture
55+
keyGenerators:
56+
snowflake:
57+
type: SNOWFLAKE
58+
auditors:
59+
sharding_key_required_auditor:
60+
type: DML_SHARDING_CONDITIONS
61+
62+
- !BROADCAST
63+
tables:
64+
- t_address

0 commit comments

Comments
 (0)