Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.

...

...

...

For a simple environment such as called out in Build a Virtualized 5-Node Hadoop 2.0 Cluster, first create a new linux user and set a password for the new account.

Code Block
languagetext
useradd -m -s /bin/bash theNEWusername
passwd theNEWusername

Then create a HDFS home directory for this new user.

Code Block
languagetext
su hdfs
hdfs dfs -mkdir /user/theNEWusername
hdfs dfs -chown theNEWusername /user/theNEWusername

Hortonworks Sandbox

http://127.0.0.1:8888/  (HUE: hue/1111  and   AMBARI: admin/admin)

...

scp -P 2222 example.jar mruser@127.0.0.1:/home/mruser/mrJars (password is "Sprint2000" for linux user and "hadoop" for one created in hue)

Networking & VirtualBox

if you make changes to /etc/sysconfig/network-scripts/ifcfg-eth0, then make sure you rm /etc/udev/rules.d/70-persistent-net.rules

HDP File Locations

Binaries: /usr/lib/SERVICENAME

Configuration: /etc/SERVICENAME/conf

Logs: /var/log/SERVICENAME

Image Removed

Benchmarking & Performance/Scalability Testing

Repo Help

Apache Ambari

Other Stuff

YARN

Kill a hadoop job:

yarn application -kill $ApplicationId

You can get a list of all ApplicationId's doing:

yarn application -list

Random Notes

dfs.datanode.max.transfer.threads - default is 1024, but bump to >= 4096 or >= 16K for HBase