Skip to content

Commit a60853f

Browse files
committed
test kms
1 parent 394df55 commit a60853f

File tree

3 files changed

+108
-6
lines changed

3 files changed

+108
-6
lines changed

.github/scripts/install-hdfs.sh

+58-4
Original file line numberDiff line numberDiff line change
@@ -2,7 +2,7 @@
22

33
set -e
44

5-
KERBEROS=${KERBEROS-"false"}
5+
KERBEROS="${KERBEROS-false}"
66
AES=${AES-"false"}
77
if [ "$DATA_TRANSFER_PROTECTION" = "privacy" ]; then
88
KERBEROS="true"
@@ -15,11 +15,18 @@ else
1515
ENCRYPT_DATA_TRANSFER="false"
1616
fi
1717

18+
CONF_KMS_PROVIDER=""
19+
TRANSPARENT_ENCRYPTION=false
20+
if [ "$HADOOP_VERSION" != "2.10.1" ]; then
21+
TRANSPARENT_ENCRYPTION=true
22+
CONF_KMS_PROVIDER="kms://http@localhost:9600/kms"
23+
fi
24+
1825
CONF_AUTHENTICATION="simple"
1926
KERBEROS_REALM="EXAMPLE.COM"
2027
KERBEROS_PRINCIPLE="administrator"
2128
KERBEROS_PASSWORD="password1234"
22-
if [ $KERBEROS = "true" ]; then
29+
if [ "$KERBEROS" = "true" ]; then
2330
CONF_AUTHENTICATION="kerberos"
2431

2532
HOSTNAME=$(hostname)
@@ -50,7 +57,7 @@ EOF
5057
sudo apt-get install -y krb5-user krb5-kdc krb5-admin-server
5158

5259
printf "$KERBEROS_PASSWORD\n$KERBEROS_PASSWORD" | sudo kdb5_util -r "$KERBEROS_REALM" create -s
53-
for p in nn dn $USER gohdfs1 gohdfs2; do
60+
for p in nn dn kms $USER gohdfs1 gohdfs2; do
5461
sudo kadmin.local -q "addprinc -randkey $p/$HOSTNAME@$KERBEROS_REALM"
5562
sudo kadmin.local -q "addprinc -randkey $p/localhost@$KERBEROS_REALM"
5663
sudo kadmin.local -q "xst -k /tmp/$p.keytab $p/$HOSTNAME@$KERBEROS_REALM"
@@ -116,6 +123,10 @@ sudo tee $HADOOP_ROOT/etc/hadoop/core-site.xml <<EOF
116123
<name>hadoop.rpc.protection</name>
117124
<value>$RPC_PROTECTION</value>
118125
</property>
126+
<property>
127+
<name>hadoop.security.key.provider.path</name>
128+
<value>$CONF_KMS_PROVIDER</value>
129+
</property>
119130
</configuration>
120131
EOF
121132

@@ -172,6 +183,41 @@ $HADOOP_ROOT/bin/hdfs namenode -format
172183
sudo groupadd hadoop
173184
sudo usermod -a -G hadoop $USER
174185

186+
sudo tee $HADOOP_ROOT/etc/hadoop/kms-site.xml <<EOF
187+
<configuration>
188+
<property>
189+
<name>hadoop.kms.key.provider.uri</name>
190+
<value>jceks://file@/tmp/hdfs/kms.keystore</value>
191+
</property>
192+
<property>
193+
<name>hadoop.security.keystore.java-keystore-provider.password-file</name>
194+
<value>kms.keystore.password</value>
195+
</property>
196+
<property>
197+
<name>hadoop.kms.authentication.type</name>
198+
<value>$CONF_AUTHENTICATION</value>
199+
</property>
200+
<property>
201+
<name>hadoop.kms.authentication.kerberos.keytab</name>
202+
<value>/tmp/kms.keytab</value>
203+
</property>
204+
<property>
205+
<name>hadoop.kms.authentication.kerberos.principal</name>
206+
<value>kms/localhost@$KERBEROS_REALM</value>
207+
</property>
208+
</configuration>
209+
EOF
210+
211+
sudo tee $HADOOP_ROOT/etc/hadoop/kms.keystore.password <<EOF
212+
123456
213+
EOF
214+
215+
if [ "$TRANSPARENT_ENCRYPTION" = "true" ]; then
216+
echo "Starting KMS..."
217+
rm $HADOOP_ROOT/etc/hadoop/kms-log4j.properties
218+
$HADOOP_ROOT/bin/hadoop kms > /tmp/hdfs/kms.log 2>&1 &
219+
fi
220+
175221
echo "Starting namenode..."
176222
$HADOOP_ROOT/bin/hdfs namenode > /tmp/hdfs/namenode.log 2>&1 &
177223

@@ -183,5 +229,13 @@ sleep 5
183229
echo "Waiting for cluster to exit safe mode..."
184230
$HADOOP_ROOT/bin/hdfs dfsadmin -safemode wait
185231

232+
$HADOOP_ROOT/bin/hadoop fs -mkdir -p /_test/kms
233+
if [ "$TRANSPARENT_ENCRYPTION" = "true" ]; then
234+
echo "Prepare encrypted zone"
235+
$HADOOP_ROOT/bin/hadoop key create key1
236+
$HADOOP_ROOT/bin/hdfs crypto -createZone -keyName key1 -path /_test/kms
237+
fi
238+
186239
echo "HADOOP_CONF_DIR=$(pwd)/$HADOOP_ROOT/etc/hadoop" >> $GITHUB_ENV
187-
echo "$(pwd)/$HADOOP_ROOT/bin" >> $GITHUB_PATH
240+
echo "TRANSPARENT_ENCRYPTION=$TRANSPARENT_ENCRYPTION" >> $GITHUB_ENV
241+
echo "$(pwd)/$HADOOP_ROOT/bin" >> $GITHUB_PATH

.github/workflows/tests.yml

+12-2
Original file line numberDiff line numberDiff line change
@@ -38,8 +38,8 @@ jobs:
3838
go-version: 1.17
3939

4040
# This step installs downloads hadoop and starts a local cluster with one
41-
# namenode and one datanode. It adds the hadoop binaries to GITHUB_PATH
42-
# and HADOOP_CONF_DIR to GITHUB_ENV.
41+
# namenode and one datanode. It adds the hadoop binaries to GITHUB_PATH,
42+
# TRANSPARENT_ENCRYPTION and HADOOP_CONF_DIR to GITHUB_ENV.
4343
- name: install-hdfs.sh
4444
run: ./.github/scripts/install-hdfs.sh
4545
env:
@@ -65,6 +65,16 @@ jobs:
6565
run: |
6666
make test
6767
68+
- name: cat kms.log
69+
if: always()
70+
run: |
71+
if [ -f /tmp/hdfs/kms.log ]
72+
then
73+
cat /tmp/hdfs/kms.log
74+
else
75+
echo "not exists"
76+
fi
77+
6878
- name: cat namenode.log
6979
if: always()
7080
run: cat /tmp/hdfs/namenode.log

cmd/hdfs/test/te.bats

+38
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,38 @@
1+
#!/usr/bin/env bats
2+
3+
load helper
4+
5+
@test "te: upload via native client, ensure we can download" {
6+
if [ "$TRANSPARENT_ENCRYPTION" = "true" ]; then
7+
run $HADOOP_FS -put $ROOT_TEST_DIR/testdata/foo.txt /_test/kms/foo1
8+
assert_success
9+
run $HDFS cat /_test/kms/foo1
10+
assert_output "bar"
11+
else
12+
skip
13+
fi
14+
}
15+
16+
@test "te: ensure native client can download once we uploaded to encrypted zone" {
17+
if [ "$TRANSPARENT_ENCRYPTION" = "true" ]; then
18+
run $HDFS put $ROOT_TEST_DIR/testdata/foo.txt /_test/kms/foo2
19+
assert_success
20+
run $HADOOP_FS -cat /_test/kms/foo2
21+
assert_output "bar"
22+
else
23+
skip
24+
fi
25+
}
26+
27+
@test "te: tail" {
28+
if [ "$TRANSPARENT_ENCRYPTION" = "true" ]; then
29+
run $HDFS put $ROOT_TEST_DIR/testdata/mobydick.txt /_test/kms/
30+
assert_success
31+
run bash -c "$HDFS tail /_test/kms/mobydick.txt > $BATS_TMPDIR/mobydick_test.txt"
32+
assert_success
33+
SHA=`tail $ROOT_TEST_DIR/testdata/mobydick.txt | shasum | awk '{ print $1 }'`
34+
assert_equal $SHA `shasum < $BATS_TMPDIR/mobydick_test.txt | awk '{ print $1 }'`
35+
else
36+
skip
37+
fi
38+
}

0 commit comments

Comments
 (0)