CDH Client Configuration Environment Variables in Oracle Big Data Appliance v4.0.0

(Doc ID 1965898.1)

Last updated on JANUARY 30, 2015

Applies to:

Big Data Appliance Integrated Software - Version 4.0 and later
Linux x86-64

Symptoms

On Oracle Big Data Appliance (BDA)/Exadata CDH Client Integration, the documentation tells us to reconfigure hadoop-env.sh (unsecured cluster) like this:

export HADOOP_HOME=hadoop-home-dir/share/hadoop/mapreduce1
export HADOOP_MAPRED_HOME=$HADOOP_HOME
export HADOOP_CONF_DIR=yarn-conf-dir
export JAVA_HOME=/usr/java/version
ln -s $HADOOP_HOME/../../../bin-mapreduce1 $HADOOP_HOME/bin
ln -s $HADOOP_HOME/../../../etc/hadoop-mapreduce1 $HADOOP_HOME/conf
rm -f $HADOOP_HOME/lib/slf4j-log4j*jar
alias hadoop=$HADOOP_HOME/bin/hadoop
alias hdfs=$HADOOP_HOME/../../../bin/hdfs

  
Why are the alias and symlink commands there? Also, as this file is run before every hadoop command, it will start spawning "already exists" errors after its first run.

Additionally, all references to the HADOOP_HOME concern the "mapreduce1" subdirectory. Isn't this where the MRv1 things are? Is this normal (exdata will use MRv1) or is this a mixup?

Cause

Sign In with your My Oracle Support account

Don't have a My Oracle Support account? Click to get started

My Oracle Support provides customers with access to over a
Million Knowledge Articles and hundreds of Community platforms