Skip to content

Commit

Permalink
KYLIN-1672 support kylin on cdh 5.7
Browse files Browse the repository at this point in the history
Signed-off-by: Li Yang <liyang@apache.org>
  • Loading branch information
lynne0326 authored and shaofengshi committed Nov 26, 2016
1 parent 0fca071 commit ccf8c16
Show file tree
Hide file tree
Showing 11 changed files with 428 additions and 1,131 deletions.
Empty file modified dev-support/test_all_against_hdp_2_2_4_2_2.sh
100644 → 100755
Empty file.
Original file line number Diff line number Diff line change
Expand Up @@ -77,6 +77,7 @@ public void write(Object key, Object value) throws IOException, InterruptedExcep
outKV[0] = key;
outKV[1] = value;
}

}

@Override
Expand All @@ -99,6 +100,7 @@ public String getStatus() {
throw new NotImplementedException();
}


@Override
public float getProgress() {
throw new NotImplementedException();
Expand Down Expand Up @@ -195,17 +197,17 @@ public String getJobName() {
}

@Override
public RawComparator<?> getSortComparator() {
public boolean userClassesTakesPrecedence() {
throw new NotImplementedException();
}

@Override
public String getJar() {
public RawComparator<?> getSortComparator() {
throw new NotImplementedException();
}

@Override
public RawComparator<?> getGroupingComparator() {
public String getJar() {
throw new NotImplementedException();
}

Expand All @@ -221,7 +223,7 @@ public boolean getTaskCleanupNeeded() {

@Override
public boolean getProfileEnabled() {
throw new NotImplementedException();
return false;
}

@Override
Expand Down Expand Up @@ -308,6 +310,11 @@ public InputSplit getInputSplit() {
public RawComparator<?> getCombinerKeyGroupingComparator() {
throw new NotImplementedException();
}

@Override
public RawComparator<?> getGroupingComparator() {
return null;
}
});
}
}
146 changes: 70 additions & 76 deletions examples/test_case_data/sandbox/core-site.xml
Original file line number Diff line number Diff line change
Expand Up @@ -14,152 +14,146 @@
See the License for the specific language governing permissions and
limitations under the License.
-->
<!--Autogenerated by Cloudera Manager-->
<configuration>

<property>
<name>fs.defaultFS</name>
<value>hdfs://sandbox.hortonworks.com:8020</value>
<final>true</final>
<value>hdfs://quickstart.cloudera:8020</value>
</property>

<property>
<name>fs.trash.interval</name>
<value>360</value>
<value>1</value>
</property>

<property>
<name>ha.failover-controller.active-standby-elector.zk.op.retries</name>
<value>120</value>
<name>io.compression.codecs</name>
<value>org.apache.hadoop.io.compress.DefaultCodec,org.apache.hadoop.io.compress.GzipCodec,org.apache.hadoop.io.compress.BZip2Codec,org.apache.hadoop.io.compress.DeflateCodec,org.apache.hadoop.io.compress.SnappyCodec,org.apache.hadoop.io.compress.Lz4Codec</value>
</property>

<property>
<name>hadoop.http.authentication.simple.anonymous.allowed</name>
<value>true</value>
<name>hadoop.security.authentication</name>
<value>simple</value>
</property>

<property>
<name>hadoop.proxyuser.falcon.groups</name>
<value>users</value>
<name>hadoop.security.authorization</name>
<value>false</value>
</property>
<property>
<name>hadoop.rpc.protection</name>
<value>authentication</value>
</property>
<property>
<name>hadoop.security.auth_to_local</name>
<value>DEFAULT</value>
</property>

<property>
<name>hadoop.proxyuser.falcon.hosts</name>
<name>hadoop.proxyuser.oozie.hosts</name>
<value>*</value>
</property>

<property>
<name>hadoop.proxyuser.hbase.groups</name>
<value>users</value>
<name>hadoop.proxyuser.oozie.groups</name>
<value>*</value>
</property>

<property>
<name>hadoop.proxyuser.hbase.hosts</name>
<name>hadoop.proxyuser.mapred.hosts</name>
<value>*</value>
</property>

<property>
<name>hadoop.proxyuser.hcat.groups</name>
<name>hadoop.proxyuser.mapred.groups</name>
<value>*</value>
</property>

<property>
<name>hadoop.proxyuser.hcat.hosts</name>
<value>sandbox.hortonworks.com</value>
<name>hadoop.proxyuser.flume.hosts</name>
<value>*</value>
</property>

<property>
<name>hadoop.proxyuser.hive.groups</name>
<value>users</value>
<name>hadoop.proxyuser.flume.groups</name>
<value>*</value>
</property>
<property>
<name>hadoop.proxyuser.HTTP.hosts</name>
<value>*</value>
</property>
<property>
<name>hadoop.proxyuser.HTTP.groups</name>
<value>*</value>
</property>

<property>
<name>hadoop.proxyuser.hive.hosts</name>
<value>*</value>
</property>

<property>
<name>hadoop.proxyuser.hue.groups</name>
<name>hadoop.proxyuser.hive.groups</name>
<value>*</value>
</property>

<property>
<name>hadoop.proxyuser.hue.hosts</name>
<value>*</value>
</property>

<property>
<name>hadoop.proxyuser.oozie.groups</name>
<name>hadoop.proxyuser.hue.groups</name>
<value>*</value>
</property>

<property>
<name>hadoop.proxyuser.oozie.hosts</name>
<value>sandbox.hortonworks.com</value>
<name>hadoop.proxyuser.httpfs.hosts</name>
<value>*</value>
</property>

<property>
<name>hadoop.proxyuser.root.groups</name>
<name>hadoop.proxyuser.httpfs.groups</name>
<value>*</value>
</property>

<property>
<name>hadoop.proxyuser.root.hosts</name>
<name>hadoop.proxyuser.hdfs.groups</name>
<value>*</value>
</property>

<property>
<name>hadoop.security.auth_to_local</name>
<value>DEFAULT</value>
<name>hadoop.proxyuser.hdfs.hosts</name>
<value>*</value>
</property>

<property>
<name>hadoop.security.authentication</name>
<value>simple</value>
<name>hadoop.proxyuser.yarn.hosts</name>
<value>*</value>
</property>

<property>
<name>hadoop.security.authorization</name>
<value>false</value>
<name>hadoop.proxyuser.yarn.groups</name>
<value>*</value>
</property>

<property>
<name>io.compression.codecs</name>
<value>org.apache.hadoop.io.compress.GzipCodec,org.apache.hadoop.io.compress.DefaultCodec,org.apache.hadoop.io.compress.SnappyCodec</value>
<name>hadoop.security.group.mapping</name>
<value>org.apache.hadoop.security.ShellBasedUnixGroupsMapping</value>
</property>

<property>
<name>io.file.buffer.size</name>
<value>131072</value>
<name>hadoop.security.instrumentation.requires.admin</name>
<value>false</value>
</property>

<property>
<name>io.serializations</name>
<value>org.apache.hadoop.io.serializer.WritableSerialization</value>
<name>net.topology.script.file.name</name>
<value>/etc/hadoop/conf.cloudera.yarn/topology.py</value>
</property>

<property>
<name>ipc.client.connect.max.retries</name>
<value>50</value>
<name>io.file.buffer.size</name>
<value>65536</value>
</property>

<property>
<name>ipc.client.connection.maxidletime</name>
<value>30000</value>
<name>hadoop.ssl.enabled</name>
<value>false</value>
</property>
<property>
<name>hadoop.ssl.require.client.cert</name>
<value>false</value>
<final>true</final>
</property>

<property>
<name>ipc.client.idlethreshold</name>
<value>8000</value>
<name>hadoop.ssl.keystores.factory.class</name>
<value>org.apache.hadoop.security.ssl.FileBasedKeyStoresFactory</value>
<final>true</final>
</property>

<property>
<name>ipc.server.tcpnodelay</name>
<value>true</value>
<name>hadoop.ssl.server.conf</name>
<value>ssl-server.xml</value>
<final>true</final>
</property>

<property>
<name>mapreduce.jobtracker.webinterface.trusted</name>
<value>false</value>
<name>hadoop.ssl.client.conf</name>
<value>ssl-client.xml</value>
<final>true</final>
</property>

</configuration>
Loading

0 comments on commit ccf8c16

Please sign in to comment.