码迷,mamicode.com
首页 > 其他好文 > 详细

Using the command line to manage files on HDFS--转载

时间:2015-12-24 20:38:36      阅读:270      评论:0      收藏:0      [点我收藏+]

标签:

原文地址:http://zh.hortonworks.com/hadoop-tutorial/using-commandline-manage-files-hdfs/

In this tutorial we will walk through some of the basic HDFS commands you will need to manage files on HDFS. To complete this tutorial you will need a working HDP cluster. The easiest way to have a Hadoop cluster is to download the Hortonworks Sandbox.

Let’s get started.

Step 1: Let’s create a directory in HDFS, upload a file and list.

Let’s look at the syntax first:

hadoop fs -mkdir:
  • It will take path uri’s as argument and creates directory or directories.
    Usage: 
            hadoop fs -mkdir <paths> 
    Example:
            hadoop fs -mkdir /user/hadoop/dir1 /user/hadoop/dir2
            hadoop fs -mkdir hdfs://nn1.example.com/user/hadoop/dir
hadoop fs -ls:
  • Lists the contents of a directory
  • For a file returns stats of a file
    Usage: 
            hadoop fs -ls <args>
    Example: 
            hadoop fs -ls /user/hadoop/dir1 /user/hadoop/dir2
            hadoop fs -ls /user/hadoop/dir1/filename.txt
            hadoop fs -ls hdfs://<hostname>:9000/user/hadoop/dir1/

Let’s use the following commands as follows and execute. You can ssh to the sandbox using Tools like Putty. You could download putty.exe from the internet.

技术分享

Let’s touch a file locally.

$ touch filename.txt

技术分享

Step 2: Now, let’s check how to find out space utilization in a HDFS dir.

hadoop fs -du:
  • Displays sizes of files and directories contained in the given directory or the size of a file if its just a file.
    Usage:  
            hadoop fs -du URI
    Example:
            hadoop fs -du  /user/hadoop/ /user/hadoop/dir1/Sample.txt

技术分享

Step 4:

Now let’s see how to upload and download files from and to Hadoop Data File System(HDFS)
Upload: ( we have already tried this earlier)

hadoop fs -put:
  • Copy single src file, or multiple src files from local file system to the Hadoop data file system
    Usage: 
            hadoop fs -put <localsrc> ... <HDFS_dest_Path>
    Example:
            hadoop fs -put /home/ec2-user/Samplefile.txt ./ambari.repo /user/hadoop/dir3/

Download:
hadoop fs -get:

  • Copies/Downloads files to the local file system
    Usage: 
            hadoop fs -get <hdfs_src> <localdst> 
    Example:
            hadoop fs -get /user/hadoop/dir3/Samplefile.txt /home/

技术分享

Step 5: Let’s look at quickly two advanced features.

hadoop fs -getmerge
  • Takes a source directory files as input and concatenates files in src into the destination local file.
    Usage: 
            hadoop fs -getmerge <src> <localdst> [addnl]
    Example:
            hadoop fs -getmerge /user/hadoop/dir1/  ./Samplefile2.txt
    Option:
            addnl: can be set to enable adding a newline on end of each file
hadoop distcp:
  • Copy file or directories recursively
  • It is a tool used for large inter/intra-cluster copying
  • It uses MapReduce to effect its distribution copy, error handling and recovery, and reporting
    Usage: 
            hadoop distcp <srcurl> <desturl>
    Example:
            hadoop distcp hdfs://<NameNode1>:8020/user/hadoop/dir1/ \ 
                    hdfs://<NameNode2>:8020/user/hadoop/dir2/

You could use the following steps to perform getmerge and discp.
Let’s upload two files for this exercise first:

# touch txt1 txt2
# hadoop fs -put txt1 txt2 /user/hadoop/dir2/
# hadoop fs -ls /user/hadoop/dir2/

技术分享

Step 6:Getting help

You can use Help command to get list of commands supported by Hadoop Data File System(HDFS)

    Example:    
            hadoop fs -help

技术分享

Hope this short tutorial was useful to get the basics of file management.

Using the command line to manage files on HDFS--转载

标签:

原文地址:http://www.cnblogs.com/davidwang456/p/5074126.html

(0)
(0)
   
举报
评论 一句话评论(0
登录后才能评论!
© 2014 mamicode.com 版权所有  联系我们:gaon5@hotmail.com
迷上了代码!