On Oracle Big Data Appliance , Critical Node becomes Unresponsive after Hdfs/NFS Gateway is Manually Setup on that Node
Last updated on OCTOBER 11, 2016
Applies to:Big Data Appliance Integrated Software - Version 4.1.0 and later
On Oracle Big Data Appliance, HDFS/NFS Gateway is manually setup on a critical node say node03 where CM, RM, DataNode, Journal Node, Zookeeper, MySQL and MGMT services reside.
Configuring An NFS Gateway on BDA and Client Setup to Access NFS Gateway (Doc ID 1998215.1)
The main purpose it was chosen by customer to setup HDFS/NFS gateway on node03 is that there are large partitions(/u01 and /u02) available on this node which is not used by Datanode. Thus 'Temporary Dump Directory' ( dfs.nfs3.dump.dir) parameter for Hdfs /NFS gateway is set to point to /u02/tmp.
'Temporary Dump Directory' temporarily saves out-of-order writes before writing them to HDFS. This directory is needed because the NFS client often reorders writes, and so sequential writes can arrive at the NFS gateway in random order and need to be saved until they can be ordered correctly. After these out-of-order writes have exceeded 1MB in memory for any given file, they are dumped to the dfs.nfs3.dump.dir (the memory threshold is not currently configurable).
And during the large file transfer from Client (Mainframe) to Hdfs gateway there are so many alerts generated about concerning health of Journal node and ultimately the node becomes inaccessible.
Time: Oct 26, 2015 7:14:04 PM
View Details on bda1node03.infoftps.com
Monitor Startup: false
Role: journalnode (bda1node03)
Role Type: JournalNode
Cluster Display Name: bda1
Service Display Name: hdfs
Service Type: HDFS
Health Test Results:
Health Test Name Event Code Severity Content
Sign In with your My Oracle Support account
Don't have a My Oracle Support account? Click to get started
Million Knowledge Articles and hundreds of Community platforms