--- conf/hadoop-env.sh.orig 2013-07-23 07:26:38.000000000 +0900
+++ conf/hadoop-env.sh 2013-08-18 23:31:06.000000000 +0900
# The java implementation to use. Required.
-# export JAVA_HOME=/usr/lib/j2sdk1.5-sun
+export JAVA_HOME=@java_home@
+export JAVA_OPTS="-Dfile.encoding=UTF-8 $JAVA_OPTS"
# Extra Java CLASSPATH elements. Optional.
# export HADOOP_CLASSPATH=
# The maximum amount of heap to use, in MB. Default is 1000.
-# export HADOOP_HEAPSIZE=2000
+export HADOOP_HEAPSIZE=2000
# Extra Java runtime options. Empty by default.
# export HADOOP_OPTS=-server
# export HADOOP_SSH_OPTS="-o ConnectTimeout=1 -o SendEnv=HADOOP_CONF_DIR"
# Where log files are stored. $HADOOP_HOME/logs by default.
-# export HADOOP_LOG_DIR=${HADOOP_HOME}/logs
+export HADOOP_LOG_DIR=@hadoop_log_dir@
# File naming remote slave hosts. $HADOOP_HOME/conf/slaves by default.
# export HADOOP_SLAVES=${HADOOP_HOME}/conf/slaves
# NOTE: this should be set to a directory that can only be written to by
# the users that are going to run the hadoop daemons. Otherwise there is
# the potential for a symlink attack.
-# export HADOOP_PID_DIR=/var/hadoop/pids
+export HADOOP_PID_DIR=@hadoop_pid_dir@
# A string representing this instance of hadoop. $USER by default.
# export HADOOP_IDENT_STRING=$USER