Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
5 changes: 0 additions & 5 deletions .gitignore
Original file line number Diff line number Diff line change
Expand Up @@ -52,15 +52,10 @@ patchprocess/
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-ui/src/main/webapp/package-lock.json
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-ui/src/main/webapp/yarn-error.log

# Ignore files generated by HDDS acceptance tests.
hadoop-ozone/acceptance-test/docker-compose.log
hadoop-ozone/acceptance-test/junit-results.xml

#robotframework outputs
log.html
output.xml
report.html

hadoop-hdds/docs/public

.mvn
3 changes: 0 additions & 3 deletions dev-support/bin/dist-layout-stitching
Original file line number Diff line number Diff line change
Expand Up @@ -21,9 +21,6 @@ VERSION=$1
# project.build.directory
BASEDIR=$2

#hdds.version
HDDS_VERSION=$3

function run()
{
declare res
Expand Down
2 changes: 1 addition & 1 deletion dev-support/docker/Dockerfile
Original file line number Diff line number Diff line change
Expand Up @@ -189,7 +189,7 @@ ENV MAVEN_OPTS -Xms256m -Xmx1536m
# YETUS CUT HERE
###

# Hugo static website generator (for new hadoop site and Ozone docs)
# Hugo static website generator for new hadoop site
RUN curl -L -o hugo.deb https://github.com/gohugoio/hugo/releases/download/v0.58.3/hugo_0.58.3_Linux-64bit.deb \
&& dpkg --install hugo.deb \
&& rm hugo.deb
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -596,11 +596,6 @@ function hadoop_bootstrap
YARN_LIB_JARS_DIR=${YARN_LIB_JARS_DIR:-"share/hadoop/yarn/lib"}
MAPRED_DIR=${MAPRED_DIR:-"share/hadoop/mapreduce"}
MAPRED_LIB_JARS_DIR=${MAPRED_LIB_JARS_DIR:-"share/hadoop/mapreduce/lib"}
HDDS_DIR=${HDDS_DIR:-"share/hadoop/hdds"}
HDDS_LIB_JARS_DIR=${HDDS_LIB_JARS_DIR:-"share/hadoop/hdds/lib"}
OZONE_DIR=${OZONE_DIR:-"share/hadoop/ozone"}
OZONE_LIB_JARS_DIR=${OZONE_LIB_JARS_DIR:-"share/hadoop/ozone/lib"}
OZONEFS_DIR=${OZONEFS_DIR:-"share/hadoop/ozonefs"}

HADOOP_TOOLS_HOME=${HADOOP_TOOLS_HOME:-${HADOOP_HOME}}
HADOOP_TOOLS_DIR=${HADOOP_TOOLS_DIR:-"share/hadoop/tools"}
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -390,15 +390,6 @@ export HADOOP_OS_TYPE=${HADOOP_OS_TYPE:-$(uname -s)}
#
# export HDFS_DFSROUTER_OPTS=""

###
# Ozone Manager specific parameters
###
# Specify the JVM options to be used when starting the Ozone Manager.
# These options will be appended to the options specified as HADOOP_OPTS
# and therefore may override any similar flags set in HADOOP_OPTS
#
# export HDFS_OM_OPTS=""

###
# HDFS StorageContainerManager specific parameters
###
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -282,13 +282,6 @@ log4j.appender.NMAUDIT.MaxBackupIndex=${nm.audit.log.maxbackupindex}
#log4j.appender.nodemanagerrequestlog.Filename=${hadoop.log.dir}/jetty-nodemanager-yyyy_mm_dd.log
#log4j.appender.nodemanagerrequestlog.RetainDays=3

#Http Server request logs for Ozone S3Gateway
log4j.logger.http.requests.s3gateway=INFO,s3gatewayrequestlog
log4j.appender.s3gatewayrequestlog=org.apache.hadoop.http.HttpRequestLogAppender
log4j.appender.s3gatewayrequestlog.Filename=${hadoop.log.dir}/jetty-s3gateway-yyyy_mm_dd.log
log4j.appender.s3gatewayrequestlog.RetainDays=3


# WebHdfs request log on datanodes
# Specify -Ddatanode.webhdfs.logger=INFO,HTTPDRFA on datanode startup to
# direct the log to a separate file.
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -99,52 +99,6 @@
{/dn.BPServiceActorInfo}
</table>

{#ozone.enabled}
<div class="page-header"><h1>Ozone: SCM Connections</h1></div>
<table class="table">
<thead>
<tr>
<th>SCM Address</th>
<th>Status</th>
<th>Version</th>
<th>Missed count</th>
<th>Last heartbeat</th>
</tr>
</thead>
{#ozone.SCMServers}
<tr>
<td>{addressString}</td>
<td>{state}</td>
<td>{versionNumber}</td>
<td>{missedCount}s</td>
<td>{lastSuccessfulHeartbeat|elapsed|fmt_time}</td>
</tr>
{/ozone.SCMServers}
</table>

<div class="page-header"><h1>Ozone: Storage locations</h1></div>
<table class="table">
<thead>
<tr>
<th>ID</th>
<th>Capacity</th>
<th>Remaining</th>
<th>SCM used</th>
<th>failed</th>
</tr>
</thead>
{#ozone.LocationReport}
<tr>
<td>{id}</td>
<td>{capacity|fmt_bytes}</td>
<td>{remaining|fmt_bytes}</td>
<td>{scmUsed|fmt_bytes}</td>
<td>{failed}</td>
</tr>
{/ozone.LocationReport}
</table>
{/ozone.enabled}

<div class="page-header"><h1>Volume Information</h1></div>
<table class="table">
<thead>
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -48,29 +48,5 @@ log4j.appender.DNMETRICSRFA.layout.ConversionPattern=%d{ISO8601} %m%n
log4j.appender.DNMETRICSRFA.MaxBackupIndex=1
log4j.appender.DNMETRICSRFA.MaxFileSize=64MB

#
# Add a logger for ozone that is separate from the Datanode.
#
log4j.logger.org.apache.hadoop.ozone=INFO,OZONE,FILE

# Do not log into datanode logs. Remove this line to have single log.
log4j.additivity.org.apache.hadoop.ozone=false

# For development purposes, log both to console and log file.
log4j.appender.OZONE=org.apache.log4j.ConsoleAppender
log4j.appender.OZONE.Threshold=ALL
log4j.appender.OZONE.layout=org.apache.log4j.PatternLayout
log4j.appender.OZONE.layout.ConversionPattern=%d{ISO8601} [%t] %-5p %c{2} (%F:%M(%L)) \
%X{component} %X{function} %X{resource} %X{user} %X{request} - %m%n

# Real ozone logger that writes to ozone.log
log4j.appender.FILE=org.apache.log4j.DailyRollingFileAppender
log4j.appender.FILE.File=${hadoop.log.dir}/ozone.log
log4j.appender.FILE.Threshold=debug
log4j.appender.FILE.layout=org.apache.log4j.PatternLayout
log4j.appender.FILE.layout.ConversionPattern=%d{ISO8601} [%t] %-5p \
(%F:%L) %X{function} %X{resource} %X{user} %X{request} - \
%m%n

# Supress KMS error log
log4j.logger.com.sun.jersey.server.wadl.generators.WadlGeneratorJAXBGrammarGenerator=OFF