|
| 1 | +/* |
| 2 | + * Copyright 2024 Datastrato Pvt Ltd. |
| 3 | + * This software is licensed under the Apache License version 2. |
| 4 | + */ |
| 5 | + |
| 6 | +package com.datastrato.gravitino.catalog.hadoop.integration.test; |
| 7 | + |
| 8 | +import static com.datastrato.gravitino.catalog.hadoop.kerberos.AuthenticationConfig.AUTH_TYPE_KEY; |
| 9 | +import static com.datastrato.gravitino.catalog.hadoop.kerberos.AuthenticationConfig.ENABLE_AUTH_KEY; |
| 10 | +import static com.datastrato.gravitino.catalog.hadoop.kerberos.KerberosConfig.IMPERSONATION_ENABLE_KEY; |
| 11 | +import static com.datastrato.gravitino.catalog.hadoop.kerberos.KerberosConfig.KEY_TAB_URI_KEY; |
| 12 | +import static com.datastrato.gravitino.catalog.hadoop.kerberos.KerberosConfig.PRINCIPAL_KEY; |
| 13 | + |
| 14 | +import com.datastrato.gravitino.Catalog; |
| 15 | +import com.datastrato.gravitino.NameIdentifier; |
| 16 | +import com.datastrato.gravitino.SchemaChange; |
| 17 | +import com.datastrato.gravitino.client.GravitinoAdminClient; |
| 18 | +import com.datastrato.gravitino.client.GravitinoMetalake; |
| 19 | +import com.datastrato.gravitino.client.KerberosTokenProvider; |
| 20 | +import com.datastrato.gravitino.file.Fileset; |
| 21 | +import com.datastrato.gravitino.integration.test.container.ContainerSuite; |
| 22 | +import com.datastrato.gravitino.integration.test.container.HiveContainer; |
| 23 | +import com.datastrato.gravitino.integration.test.util.AbstractIT; |
| 24 | +import com.datastrato.gravitino.integration.test.util.GravitinoITUtils; |
| 25 | +import com.google.common.base.Throwables; |
| 26 | +import com.google.common.collect.ImmutableMap; |
| 27 | +import com.google.common.collect.Maps; |
| 28 | +import java.io.File; |
| 29 | +import java.io.IOException; |
| 30 | +import java.nio.charset.StandardCharsets; |
| 31 | +import java.nio.file.Files; |
| 32 | +import java.util.Map; |
| 33 | +import org.apache.commons.io.FileUtils; |
| 34 | +import org.apache.hadoop.security.UserGroupInformation; |
| 35 | +import org.junit.jupiter.api.AfterAll; |
| 36 | +import org.junit.jupiter.api.Assertions; |
| 37 | +import org.junit.jupiter.api.BeforeAll; |
| 38 | +import org.junit.jupiter.api.Tag; |
| 39 | +import org.junit.jupiter.api.Test; |
| 40 | +import org.slf4j.Logger; |
| 41 | +import org.slf4j.LoggerFactory; |
| 42 | +import sun.security.krb5.KrbException; |
| 43 | + |
| 44 | +@Tag("gravitino-docker-it") |
| 45 | +public class HadoopUserAuthenticationIT extends AbstractIT { |
| 46 | + private static final Logger LOG = LoggerFactory.getLogger(HadoopUserAuthenticationIT.class); |
| 47 | + |
| 48 | + private static final ContainerSuite containerSuite = ContainerSuite.getInstance(); |
| 49 | + |
| 50 | + private static final String SDK_KERBEROS_PRINCIPAL_KEY = "client.kerberos.principal"; |
| 51 | + private static final String SDK_KERBEROS_KEYTAB_KEY = "client.kerberos.keytab"; |
| 52 | + |
| 53 | + private static final String GRAVITINO_CLIENT_PRINCIPAL = "gravitino_client@HADOOPKRB"; |
| 54 | + private static final String GRAVITINO_CLIENT_KEYTAB = "/gravitino_client.keytab"; |
| 55 | + |
| 56 | + private static final String GRAVITINO_SERVER_PRINCIPAL = "HTTP/localhost@HADOOPKRB"; |
| 57 | + private static final String GRAVITINO_SERVER_KEYTAB = "/gravitino_server.keytab"; |
| 58 | + |
| 59 | + private static final String HADOOP_CLIENT_PRINCIPAL = "cli@HADOOPKRB"; |
| 60 | + private static final String HADOOP_CLIENT_KEYTAB = "/client.keytab"; |
| 61 | + |
| 62 | + private static String TMP_DIR; |
| 63 | + |
| 64 | + private static String HDFS_URL; |
| 65 | + |
| 66 | + private static GravitinoAdminClient adminClient; |
| 67 | + |
| 68 | + private static HiveContainer kerberosHiveContainer; |
| 69 | + |
| 70 | + private static final String METALAKE_NAME = |
| 71 | + GravitinoITUtils.genRandomName("CatalogHadoop_metalake"); |
| 72 | + private static final String CATALOG_NAME = |
| 73 | + GravitinoITUtils.genRandomName("CatalogHadoop_catalog"); |
| 74 | + private static final String SCHEMA_NAME = GravitinoITUtils.genRandomName("CatalogHadoop_schema"); |
| 75 | + |
| 76 | + @SuppressWarnings("unused") |
| 77 | + private static final String TABLE_NAME = "test_table"; |
| 78 | + |
| 79 | + @BeforeAll |
| 80 | + public static void startIntegrationTest() throws Exception { |
| 81 | + containerSuite.startKerberosHiveContainer(); |
| 82 | + kerberosHiveContainer = containerSuite.getKerberosHiveContainer(); |
| 83 | + |
| 84 | + File baseDir = new File(System.getProperty("java.io.tmpdir")); |
| 85 | + File file = Files.createTempDirectory(baseDir.toPath(), "test").toFile(); |
| 86 | + file.deleteOnExit(); |
| 87 | + TMP_DIR = file.getAbsolutePath(); |
| 88 | + |
| 89 | + HDFS_URL = String.format("hdfs://%s:9000", kerberosHiveContainer.getContainerIpAddress()); |
| 90 | + |
| 91 | + // Prepare kerberos related-config; |
| 92 | + prepareKerberosConfig(); |
| 93 | + |
| 94 | + // Config kerberos configuration for Gravitino server |
| 95 | + addKerberosConfig(); |
| 96 | + |
| 97 | + // Start Gravitino server |
| 98 | + AbstractIT.startIntegrationTest(); |
| 99 | + } |
| 100 | + |
| 101 | + @AfterAll |
| 102 | + public static void stop() { |
| 103 | + // Reset the UGI |
| 104 | + UserGroupInformation.reset(); |
| 105 | + |
| 106 | + // Clean up the kerberos configuration |
| 107 | + System.clearProperty("java.security.krb5.conf"); |
| 108 | + System.clearProperty("sun.security.krb5.debug"); |
| 109 | + } |
| 110 | + |
| 111 | + private static void prepareKerberosConfig() throws IOException, KrbException { |
| 112 | + // Keytab of the Gravitino SDK client |
| 113 | + kerberosHiveContainer |
| 114 | + .getContainer() |
| 115 | + .copyFileFromContainer("/gravitino_client.keytab", TMP_DIR + GRAVITINO_CLIENT_KEYTAB); |
| 116 | + |
| 117 | + // Keytab of the Gravitino server |
| 118 | + kerberosHiveContainer |
| 119 | + .getContainer() |
| 120 | + .copyFileFromContainer("/gravitino_server.keytab", TMP_DIR + GRAVITINO_SERVER_KEYTAB); |
| 121 | + |
| 122 | + // Keytab of Gravitino server to connector to HDFS |
| 123 | + kerberosHiveContainer |
| 124 | + .getContainer() |
| 125 | + .copyFileFromContainer("/etc/admin.keytab", TMP_DIR + HADOOP_CLIENT_KEYTAB); |
| 126 | + |
| 127 | + String tmpKrb5Path = TMP_DIR + "krb5.conf_tmp"; |
| 128 | + String krb5Path = TMP_DIR + "krb5.conf"; |
| 129 | + kerberosHiveContainer.getContainer().copyFileFromContainer("/etc/krb5.conf", tmpKrb5Path); |
| 130 | + |
| 131 | + // Modify the krb5.conf and change the kdc and admin_server to the container IP |
| 132 | + String ip = containerSuite.getKerberosHiveContainer().getContainerIpAddress(); |
| 133 | + String content = FileUtils.readFileToString(new File(tmpKrb5Path), StandardCharsets.UTF_8); |
| 134 | + content = content.replace("kdc = localhost:88", "kdc = " + ip + ":88"); |
| 135 | + content = content.replace("admin_server = localhost", "admin_server = " + ip + ":749"); |
| 136 | + FileUtils.write(new File(krb5Path), content, StandardCharsets.UTF_8); |
| 137 | + |
| 138 | + LOG.info("Kerberos kdc config:\n{}", content); |
| 139 | + System.setProperty("java.security.krb5.conf", krb5Path); |
| 140 | + System.setProperty("sun.security.krb5.debug", "true"); |
| 141 | + } |
| 142 | + |
| 143 | + private static void addKerberosConfig() { |
| 144 | + AbstractIT.customConfigs.put("gravitino.authenticator", "kerberos"); |
| 145 | + AbstractIT.customConfigs.put( |
| 146 | + "gravitino.authenticator.kerberos.principal", GRAVITINO_SERVER_PRINCIPAL); |
| 147 | + AbstractIT.customConfigs.put( |
| 148 | + "gravitino.authenticator.kerberos.keytab", TMP_DIR + GRAVITINO_SERVER_KEYTAB); |
| 149 | + AbstractIT.customConfigs.put(SDK_KERBEROS_KEYTAB_KEY, TMP_DIR + GRAVITINO_CLIENT_KEYTAB); |
| 150 | + AbstractIT.customConfigs.put(SDK_KERBEROS_PRINCIPAL_KEY, GRAVITINO_CLIENT_PRINCIPAL); |
| 151 | + } |
| 152 | + |
| 153 | + @Test |
| 154 | + public void testUserAuthentication() { |
| 155 | + KerberosTokenProvider provider = |
| 156 | + KerberosTokenProvider.builder() |
| 157 | + .withClientPrincipal(GRAVITINO_CLIENT_PRINCIPAL) |
| 158 | + .withKeyTabFile(new File(TMP_DIR + GRAVITINO_CLIENT_KEYTAB)) |
| 159 | + .build(); |
| 160 | + adminClient = GravitinoAdminClient.builder(serverUri).withKerberosAuth(provider).build(); |
| 161 | + |
| 162 | + GravitinoMetalake[] metalakes = adminClient.listMetalakes(); |
| 163 | + Assertions.assertEquals(0, metalakes.length); |
| 164 | + |
| 165 | + GravitinoMetalake gravitinoMetalake = |
| 166 | + adminClient.createMetalake(METALAKE_NAME, null, ImmutableMap.of()); |
| 167 | + |
| 168 | + // Create a catalog |
| 169 | + Map<String, String> properties = Maps.newHashMap(); |
| 170 | + |
| 171 | + properties.put(ENABLE_AUTH_KEY, "true"); |
| 172 | + properties.put(AUTH_TYPE_KEY, "kerberos"); |
| 173 | + properties.put(IMPERSONATION_ENABLE_KEY, "true"); |
| 174 | + properties.put(KEY_TAB_URI_KEY, TMP_DIR + HADOOP_CLIENT_KEYTAB); |
| 175 | + properties.put(PRINCIPAL_KEY, HADOOP_CLIENT_PRINCIPAL); |
| 176 | + properties.put("location", HDFS_URL + "/user/hadoop/"); |
| 177 | + |
| 178 | + kerberosHiveContainer.executeInContainer("hadoop", "fs", "-mkdir", "/user/hadoop"); |
| 179 | + |
| 180 | + Catalog catalog = |
| 181 | + gravitinoMetalake.createCatalog( |
| 182 | + CATALOG_NAME, Catalog.Type.FILESET, "hadoop", "comment", properties); |
| 183 | + |
| 184 | + // Test create schema |
| 185 | + Exception exception = |
| 186 | + Assertions.assertThrows( |
| 187 | + Exception.class, |
| 188 | + () -> catalog.asSchemas().createSchema(SCHEMA_NAME, "comment", ImmutableMap.of())); |
| 189 | + String exceptionMessage = Throwables.getStackTraceAsString(exception); |
| 190 | + // Make sure real user is 'gravitino_client' |
| 191 | + Assertions.assertTrue( |
| 192 | + exceptionMessage.contains("Permission denied: user=gravitino_client, access=WRITE")); |
| 193 | + |
| 194 | + // Now try to give the user the permission to create schema again |
| 195 | + kerberosHiveContainer.executeInContainer("hadoop", "fs", "-chmod", "-R", "777", "/user/hadoop"); |
| 196 | + Assertions.assertDoesNotThrow( |
| 197 | + () -> catalog.asSchemas().createSchema(SCHEMA_NAME, "comment", ImmutableMap.of())); |
| 198 | + |
| 199 | + catalog |
| 200 | + .asFilesetCatalog() |
| 201 | + .createFileset( |
| 202 | + NameIdentifier.of(METALAKE_NAME, CATALOG_NAME, SCHEMA_NAME, TABLE_NAME), |
| 203 | + "comment", |
| 204 | + Fileset.Type.MANAGED, |
| 205 | + null, |
| 206 | + ImmutableMap.of()); |
| 207 | + |
| 208 | + catalog |
| 209 | + .asFilesetCatalog() |
| 210 | + .dropFileset(NameIdentifier.of(METALAKE_NAME, CATALOG_NAME, SCHEMA_NAME, TABLE_NAME)); |
| 211 | + |
| 212 | + catalog.asSchemas().alterSchema(SCHEMA_NAME, SchemaChange.setProperty("k1", "value1")); |
| 213 | + |
| 214 | + catalog.asSchemas().dropSchema(SCHEMA_NAME, true); |
| 215 | + } |
| 216 | +} |
0 commit comments