Hadoop: Difference between revisions
Jump to navigation
Jump to search
Line 152: | Line 152: | ||
* [https://stackoverflow.com/questions/1988249/ Bash » Switch user & execute remaining script] | * [https://stackoverflow.com/questions/1988249/ Bash » Switch user & execute remaining script] | ||
* [https://unix.stackexchange.com/questions/291454/ Difference between sudo user vs. root user] | * [https://unix.stackexchange.com/questions/291454/ Difference between sudo user vs. root user] | ||
* [https://jumpcloud.com/blog/how-to-create-a-new-sudo-user-manage-sudo-access-on-ubuntu-20-04 Create a Sudo User & Manage Access] | * [https://jumpcloud.com/blog/how-to-create-a-new-sudo-user-manage-sudo-access-on-ubuntu-20-04 Create a Sudo User & Manage Access] | ||
* [https://unix.stackexchange.com/questions/69314/ Bash » Non-interactive <code>ssh-keygen</code>] | * [https://unix.stackexchange.com/questions/69314/ Bash » Non-interactive <code>ssh-keygen</code>] | ||
* [https://superuser.com/questions/468161/ Bash » Switch user & execute script] | * [https://superuser.com/questions/468161/ Bash » Switch user & execute script] | ||
* [https://unix.stackexchange.com/questions/28526/ Bash » Add a user if it doesn't exist] | * [https://unix.stackexchange.com/questions/28526/ Bash » Add a user if it doesn't exist] | ||
* [[Sed Replace A Multi-Line String]] | |||
* [https://serverfault.com/questions/283129/ SSH Connection Hang Forever] | * [https://serverfault.com/questions/283129/ SSH Connection Hang Forever] | ||
Revision as of 04:53, 24 December 2022
Hadoop is a Java-based programming framework that supports the processing and storage of extremely large datasets on a cluster of inexpensive machines. It was the first major open source project in the big data playing field and is sponsored by the Apache Software Foundation. Hadoop is comprised of four main layers:
- Hadoop Common is the collection of utilities and libraries that support other Hadoop modules.
- HDFS, which stands for Hadoop Distributed File System, is responsible for persisting data to disk.
- YARN, short for Yet Another Resource Negotiator, is the "operating system" for HDFS.
- MapReduce is the original processing model for Hadoop clusters. It distributes work within the cluster or map, then organizes and reduces the results from the nodes into a response to a query. Many other processing models are available for the 3.x version of Hadoop
Configuration
mkdir -p /home/hadoop/hdfs/{datanode,namenode}/
sudo tee -a $HADOOP_HOME/etc/hadoop/core-site.xml >/dev/null <<EOF
<configuration>
<property>
<name>hadoop.tmp.dir</name>
<value>/home/hadoop/tmp</value>
</property>
<property>
<name>fs.default.name</name>
<value>hdfs://0.0.0.0:9000</value>
<description>The default file system URI</description>
</property>
</configuration>
EOF
mkdir -p /home/hadoop/hdfs/{datanode,namenode}
sudo tee -a $HADOOP_HOME/etc/hadoop/hdfs-site.xml >/dev/null <<EOF
<configuration>
<property>
<name>dfs.replication</name>
<value>1</value>
</property>
<property>
<name>dfs.namenode.name.dir</name>
<value>/home/hadoop/hdfs/namenode/</value>
</property>
<property>
<name>dfs.datanode.data.dir</name>
<value>/home/hadoop/hdfs/datanode/</value>
</property>
</configuration>
EOF
sudo tee -a $HADOOP_HOME/etc/hadoop/mapred-site.xml >/dev/null <<EOF
<configuration>
<property>
<name>mapreduce.framework.name</name>
<value>yarn</value>
</property>
</configuration>
EOF
sudo tee -a $HADOOP_HOME/etc/hadoop/yarn-site.xml >/dev/null <<EOF
<configuration>
<property>
<name>yarn.nodemanager.aux-services</name>
<value>mapreduce_shuffle</value>
</property>
</configuration>
EOF
Knowledge
ssh-keygen -b 4096 -t rsa -f ~/.ssh/id_rsa -q -N "[email protected]" readlink -f /usr/bin/java | sed "s:bin/java::" sudo apt-get install pdsh sudo apt-get install ssh | |
| |
su -h hadoop hdfs namenode -format sudo -u haddop -H sh -c "whoami; echo ${HOME}" sh $HADOOP_HOME/sbin/start-dfs.sh http://127.0.0.1:9870 sh $HADOOP_HOME/sbin/start-yarn.sh http://127.0.0.1:8088 | |
| |
sudo apt dist-upgrade sudo do-release-upgrade sudo apt --fix-broken install sudo apt install ubuntu-desktop |
[Service]
User=hadoop
Group=hadoop
Type=forking
SuccessExitStatus=143
|