bp: #43636
This commit is contained in:
@ -137,6 +137,12 @@ under the License.
|
||||
<groupId>org.apache.hadoop</groupId>
|
||||
<artifactId>hadoop-aliyun</artifactId>
|
||||
</dependency>
|
||||
<dependency>
|
||||
<groupId>commons-collections</groupId>
|
||||
<artifactId>commons-collections</artifactId>
|
||||
<version>${commons-collections.version}</version>
|
||||
<scope>test</scope>
|
||||
</dependency>
|
||||
</dependencies>
|
||||
<build>
|
||||
<finalName>doris-fe-common</finalName>
|
||||
|
||||
@ -17,10 +17,14 @@
|
||||
|
||||
package org.apache.doris.common.security.authentication;
|
||||
|
||||
import com.google.common.base.Strings;
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.fs.CommonConfigurationKeysPublic;
|
||||
import org.apache.logging.log4j.LogManager;
|
||||
import org.apache.logging.log4j.Logger;
|
||||
|
||||
public abstract class AuthenticationConfig {
|
||||
private static final Logger LOG = LogManager.getLogger(AuthenticationConfig.class);
|
||||
public static String HADOOP_USER_NAME = "hadoop.username";
|
||||
public static String HADOOP_KERBEROS_PRINCIPAL = "hadoop.kerberos.principal";
|
||||
public static String HADOOP_KERBEROS_KEYTAB = "hadoop.kerberos.keytab";
|
||||
@ -42,6 +46,10 @@ public abstract class AuthenticationConfig {
|
||||
return AuthenticationConfig.getKerberosConfig(conf, HADOOP_KERBEROS_PRINCIPAL, HADOOP_KERBEROS_KEYTAB);
|
||||
}
|
||||
|
||||
public static AuthenticationConfig getSimpleAuthenticationConfig(Configuration conf) {
|
||||
return AuthenticationConfig.createSimpleAuthenticationConfig(conf);
|
||||
}
|
||||
|
||||
/**
|
||||
* get kerberos config from hadoop conf
|
||||
* @param conf config
|
||||
@ -54,17 +62,35 @@ public abstract class AuthenticationConfig {
|
||||
String krbKeytabKey) {
|
||||
String authentication = conf.get(CommonConfigurationKeysPublic.HADOOP_SECURITY_AUTHENTICATION, null);
|
||||
if (AuthType.KERBEROS.getDesc().equals(authentication)) {
|
||||
KerberosAuthenticationConfig krbConfig = new KerberosAuthenticationConfig();
|
||||
krbConfig.setKerberosPrincipal(conf.get(krbPrincipalKey));
|
||||
krbConfig.setKerberosKeytab(conf.get(krbKeytabKey));
|
||||
krbConfig.setConf(conf);
|
||||
krbConfig.setPrintDebugLog(Boolean.parseBoolean(conf.get(DORIS_KRB5_DEBUG, "false")));
|
||||
return krbConfig;
|
||||
} else {
|
||||
// AuthType.SIMPLE
|
||||
SimpleAuthenticationConfig simpleAuthenticationConfig = new SimpleAuthenticationConfig();
|
||||
simpleAuthenticationConfig.setUsername(conf.get(HADOOP_USER_NAME));
|
||||
return simpleAuthenticationConfig;
|
||||
String principalKey = conf.get(krbPrincipalKey);
|
||||
String keytabKey = conf.get(krbKeytabKey);
|
||||
if (!Strings.isNullOrEmpty(principalKey) && !Strings.isNullOrEmpty(keytabKey)) {
|
||||
KerberosAuthenticationConfig krbConfig = new KerberosAuthenticationConfig();
|
||||
krbConfig.setKerberosPrincipal(principalKey);
|
||||
krbConfig.setKerberosKeytab(keytabKey);
|
||||
krbConfig.setConf(conf);
|
||||
krbConfig.setPrintDebugLog(Boolean.parseBoolean(conf.get(DORIS_KRB5_DEBUG, "false")));
|
||||
return krbConfig;
|
||||
} else {
|
||||
// Due to some historical reasons, `core-size.xml` may be stored in path:`fe/conf`,
|
||||
// but this file may only contain `hadoop.security.authentication configuration`,
|
||||
// and no krbPrincipalKey and krbKeytabKey,
|
||||
// which will cause kerberos initialization failure.
|
||||
// Now:
|
||||
// if kerberos is needed, the relevant configuration can be written in the catalog properties,
|
||||
// if kerberos is not needed, to prevent the influence of historical reasons,
|
||||
// the following simple authentication method needs to be used.
|
||||
LOG.warn("{} or {} is null or empty, fallback to simple authentication",
|
||||
krbPrincipalKey, krbKeytabKey);
|
||||
}
|
||||
}
|
||||
return createSimpleAuthenticationConfig(conf);
|
||||
}
|
||||
|
||||
private static AuthenticationConfig createSimpleAuthenticationConfig(Configuration conf) {
|
||||
// AuthType.SIMPLE
|
||||
SimpleAuthenticationConfig simpleAuthenticationConfig = new SimpleAuthenticationConfig();
|
||||
simpleAuthenticationConfig.setUsername(conf.get(HADOOP_USER_NAME));
|
||||
return simpleAuthenticationConfig;
|
||||
}
|
||||
}
|
||||
|
||||
@ -0,0 +1,45 @@
|
||||
// Licensed to the Apache Software Foundation (ASF) under one
|
||||
// or more contributor license agreements. See the NOTICE file
|
||||
// distributed with this work for additional information
|
||||
// regarding copyright ownership. The ASF licenses this file
|
||||
// to you under the Apache License, Version 2.0 (the
|
||||
// "License"); you may not use this file except in compliance
|
||||
// with the License. You may obtain a copy of the License at
|
||||
//
|
||||
// http://www.apache.org/licenses/LICENSE-2.0
|
||||
//
|
||||
// Unless required by applicable law or agreed to in writing,
|
||||
// software distributed under the License is distributed on an
|
||||
// "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
||||
// KIND, either express or implied. See the License for the
|
||||
// specific language governing permissions and limitations
|
||||
// under the License.
|
||||
|
||||
package org.apache.doris.common.security.authentication;
|
||||
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.fs.CommonConfigurationKeysPublic;
|
||||
import org.junit.Assert;
|
||||
import org.junit.Test;
|
||||
|
||||
public class AuthenticationTest {
|
||||
|
||||
@Test
|
||||
public void testAuthConf() {
|
||||
Configuration conf = new Configuration();
|
||||
|
||||
AuthenticationConfig conf1 = AuthenticationConfig.getKerberosConfig(conf);
|
||||
Assert.assertEquals(SimpleAuthenticationConfig.class, conf1.getClass());
|
||||
|
||||
conf.set(CommonConfigurationKeysPublic.HADOOP_SECURITY_AUTHENTICATION, "kerberos");
|
||||
|
||||
AuthenticationConfig conf2 = AuthenticationConfig.getKerberosConfig(conf);
|
||||
Assert.assertEquals(SimpleAuthenticationConfig.class, conf2.getClass());
|
||||
|
||||
conf.set(AuthenticationConfig.HADOOP_KERBEROS_PRINCIPAL, "principal");
|
||||
conf.set(AuthenticationConfig.HADOOP_KERBEROS_KEYTAB, "keytab");
|
||||
|
||||
AuthenticationConfig conf3 = AuthenticationConfig.getKerberosConfig(conf);
|
||||
Assert.assertEquals(KerberosAuthenticationConfig.class, conf3.getClass());
|
||||
}
|
||||
}
|
||||
@ -43,6 +43,7 @@ import java.util.Map;
|
||||
public class S3FileSystem extends ObjFileSystem {
|
||||
|
||||
private static final Logger LOG = LogManager.getLogger(S3FileSystem.class);
|
||||
private HadoopAuthenticator authenticator = null;
|
||||
|
||||
public S3FileSystem(Map<String, String> properties) {
|
||||
super(StorageBackend.StorageType.S3.name(), StorageBackend.StorageType.S3, new S3ObjStorage(properties));
|
||||
@ -77,7 +78,9 @@ public class S3FileSystem extends ObjFileSystem {
|
||||
PropertyConverter.convertToHadoopFSProperties(properties).entrySet().stream()
|
||||
.filter(entry -> entry.getKey() != null && entry.getValue() != null)
|
||||
.forEach(entry -> conf.set(entry.getKey(), entry.getValue()));
|
||||
AuthenticationConfig authConfig = AuthenticationConfig.getKerberosConfig(conf);
|
||||
// S3 does not support Kerberos authentication,
|
||||
// so here we create a simple authentication
|
||||
AuthenticationConfig authConfig = AuthenticationConfig.getSimpleAuthenticationConfig(conf);
|
||||
HadoopAuthenticator authenticator = HadoopAuthenticator.getHadoopAuthenticator(authConfig);
|
||||
try {
|
||||
dfsFileSystem = authenticator.doAs(() -> {
|
||||
@ -87,6 +90,7 @@ public class S3FileSystem extends ObjFileSystem {
|
||||
throw new RuntimeException(e);
|
||||
}
|
||||
});
|
||||
this.authenticator = authenticator;
|
||||
RemoteFSPhantomManager.registerPhantomReference(this);
|
||||
} catch (Exception e) {
|
||||
throw new UserException("Failed to get S3 FileSystem for " + e.getMessage(), e);
|
||||
@ -134,4 +138,9 @@ public class S3FileSystem extends ObjFileSystem {
|
||||
}
|
||||
return Status.OK;
|
||||
}
|
||||
|
||||
@VisibleForTesting
|
||||
public HadoopAuthenticator getAuthenticator() {
|
||||
return authenticator;
|
||||
}
|
||||
}
|
||||
|
||||
@ -489,4 +489,9 @@ public class DFSFileSystem extends RemoteFileSystem {
|
||||
}
|
||||
return Status.OK;
|
||||
}
|
||||
|
||||
@VisibleForTesting
|
||||
public HadoopAuthenticator getAuthenticator() {
|
||||
return authenticator;
|
||||
}
|
||||
}
|
||||
|
||||
@ -0,0 +1,158 @@
|
||||
// Licensed to the Apache Software Foundation (ASF) under one
|
||||
// or more contributor license agreements. See the NOTICE file
|
||||
// distributed with this work for additional information
|
||||
// regarding copyright ownership. The ASF licenses this file
|
||||
// to you under the Apache License, Version 2.0 (the
|
||||
// "License"); you may not use this file except in compliance
|
||||
// with the License. You may obtain a copy of the License at
|
||||
//
|
||||
// http://www.apache.org/licenses/LICENSE-2.0
|
||||
//
|
||||
// Unless required by applicable law or agreed to in writing,
|
||||
// software distributed under the License is distributed on an
|
||||
// "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
||||
// KIND, either express or implied. See the License for the
|
||||
// specific language governing permissions and limitations
|
||||
// under the License.
|
||||
|
||||
package org.apache.doris.fs.remote;
|
||||
|
||||
import org.apache.doris.common.Pair;
|
||||
import org.apache.doris.common.UserException;
|
||||
import org.apache.doris.common.security.authentication.AuthenticationConfig;
|
||||
import org.apache.doris.common.security.authentication.HadoopAuthenticator;
|
||||
import org.apache.doris.common.security.authentication.HadoopKerberosAuthenticator;
|
||||
import org.apache.doris.common.security.authentication.HadoopSimpleAuthenticator;
|
||||
import org.apache.doris.common.util.LocationPath;
|
||||
import org.apache.doris.fs.FileSystemCache;
|
||||
import org.apache.doris.fs.FileSystemType;
|
||||
import org.apache.doris.fs.remote.dfs.DFSFileSystem;
|
||||
|
||||
import com.google.common.base.Strings;
|
||||
import com.google.common.collect.ImmutableMap;
|
||||
import mockit.Mock;
|
||||
import mockit.MockUp;
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.fs.CommonConfigurationKeysPublic;
|
||||
import org.apache.hadoop.fs.LocalFileSystem;
|
||||
import org.apache.hadoop.security.UserGroupInformation;
|
||||
import org.junit.Assert;
|
||||
import org.junit.Test;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.net.URI;
|
||||
import java.security.PrivilegedExceptionAction;
|
||||
import java.util.ArrayList;
|
||||
import java.util.Map;
|
||||
|
||||
public class RemoteFileSystemTest {
|
||||
|
||||
@Test
|
||||
public void testFilesystemAndAuthType() throws UserException {
|
||||
|
||||
// These paths should use s3 filesystem, and use simple auth
|
||||
ArrayList<String> s3Paths = new ArrayList<>();
|
||||
s3Paths.add("s3://a/b/c");
|
||||
s3Paths.add("s3a://a/b/c");
|
||||
s3Paths.add("s3n://a/b/c");
|
||||
s3Paths.add("oss://a/b/c"); // default use s3 filesystem
|
||||
s3Paths.add("gs://a/b/c");
|
||||
s3Paths.add("bos://a/b/c");
|
||||
s3Paths.add("cos://a/b/c");
|
||||
s3Paths.add("cosn://a/b/c");
|
||||
s3Paths.add("lakefs://a/b/c");
|
||||
s3Paths.add("obs://a/b/c");
|
||||
|
||||
// These paths should use dfs filesystem, and auth will be changed by configure
|
||||
ArrayList<String> dfsPaths = new ArrayList<>();
|
||||
dfsPaths.add("ofs://a/b/c");
|
||||
dfsPaths.add("gfs://a/b/c");
|
||||
dfsPaths.add("hdfs://a/b/c");
|
||||
dfsPaths.add("oss://a/b/c"); // if endpoint contains 'oss-dls.aliyuncs', will use dfs filesystem
|
||||
|
||||
new MockUp<UserGroupInformation>(UserGroupInformation.class) {
|
||||
@Mock
|
||||
public <T> T doAs(PrivilegedExceptionAction<T> action) throws IOException, InterruptedException {
|
||||
return (T) new LocalFileSystem();
|
||||
}
|
||||
};
|
||||
|
||||
new MockUp<HadoopKerberosAuthenticator>(HadoopKerberosAuthenticator.class) {
|
||||
@Mock
|
||||
public synchronized UserGroupInformation getUGI() throws IOException {
|
||||
return UserGroupInformation.getCurrentUser();
|
||||
}
|
||||
};
|
||||
|
||||
Configuration confWithoutKerberos = new Configuration();
|
||||
|
||||
Configuration confWithKerberosIncomplete = new Configuration();
|
||||
confWithKerberosIncomplete.set(CommonConfigurationKeysPublic.HADOOP_SECURITY_AUTHENTICATION, "kerberos");
|
||||
|
||||
Configuration confWithKerberos = new Configuration();
|
||||
confWithKerberos.set(CommonConfigurationKeysPublic.HADOOP_SECURITY_AUTHENTICATION, "kerberos");
|
||||
confWithKerberos.set(AuthenticationConfig.HADOOP_KERBEROS_PRINCIPAL, "principal");
|
||||
confWithKerberos.set(AuthenticationConfig.HADOOP_KERBEROS_KEYTAB, "keytab");
|
||||
|
||||
ImmutableMap<String, String> s3props = ImmutableMap.of("s3.endpoint", "http://127.0.0.1");
|
||||
s3props.forEach(confWithKerberos::set);
|
||||
s3props.forEach(confWithoutKerberos::set);
|
||||
s3props.forEach(confWithKerberosIncomplete::set);
|
||||
|
||||
for (String path : s3Paths) {
|
||||
checkS3Filesystem(path, confWithKerberos, s3props);
|
||||
}
|
||||
for (String path : s3Paths) {
|
||||
checkS3Filesystem(path, confWithKerberosIncomplete, s3props);
|
||||
}
|
||||
for (String path : s3Paths) {
|
||||
checkS3Filesystem(path, confWithoutKerberos, s3props);
|
||||
}
|
||||
|
||||
s3props = ImmutableMap.of("s3.endpoint", "oss://xx-oss-dls.aliyuncs/abc");
|
||||
System.setProperty("java.security.krb5.realm", "realm");
|
||||
System.setProperty("java.security.krb5.kdc", "kdc");
|
||||
|
||||
for (String path : dfsPaths) {
|
||||
checkDFSFilesystem(path, confWithKerberos, HadoopKerberosAuthenticator.class.getName(), s3props);
|
||||
}
|
||||
for (String path : dfsPaths) {
|
||||
checkDFSFilesystem(path, confWithKerberosIncomplete, HadoopSimpleAuthenticator.class.getName(), s3props);
|
||||
}
|
||||
for (String path : dfsPaths) {
|
||||
checkDFSFilesystem(path, confWithoutKerberos, HadoopSimpleAuthenticator.class.getName(), s3props);
|
||||
}
|
||||
|
||||
}
|
||||
|
||||
private void checkS3Filesystem(String path, Configuration conf, Map<String, String> m) throws UserException {
|
||||
RemoteFileSystem fs = createFs(path, conf, m);
|
||||
Assert.assertTrue(fs instanceof S3FileSystem);
|
||||
HadoopAuthenticator authenticator = ((S3FileSystem) fs).getAuthenticator();
|
||||
Assert.assertTrue(authenticator instanceof HadoopSimpleAuthenticator);
|
||||
}
|
||||
|
||||
private void checkDFSFilesystem(String path, Configuration conf, String authClass, Map<String, String> m) throws UserException {
|
||||
RemoteFileSystem fs = createFs(path, conf, m);
|
||||
Assert.assertTrue(fs instanceof DFSFileSystem);
|
||||
HadoopAuthenticator authenticator = ((DFSFileSystem) fs).getAuthenticator();
|
||||
Assert.assertEquals(authClass, authenticator.getClass().getName());
|
||||
}
|
||||
|
||||
private RemoteFileSystem createFs(String path, Configuration conf, Map<String, String> m) throws UserException {
|
||||
LocationPath locationPath = new LocationPath(path, m);
|
||||
FileSystemType fileSystemType = locationPath.getFileSystemType();
|
||||
URI uri = locationPath.getPath().toUri();
|
||||
String fsIdent = Strings.nullToEmpty(uri.getScheme()) + "://" + Strings.nullToEmpty(uri.getAuthority());
|
||||
FileSystemCache fileSystemCache = new FileSystemCache();
|
||||
RemoteFileSystem fs = fileSystemCache.getRemoteFileSystem(
|
||||
new FileSystemCache.FileSystemCacheKey(
|
||||
Pair.of(fileSystemType, fsIdent),
|
||||
ImmutableMap.of(),
|
||||
null,
|
||||
conf));
|
||||
fs.nativeFileSystem(path);
|
||||
return fs;
|
||||
}
|
||||
|
||||
}
|
||||
Reference in New Issue
Block a user