<>一、windows客户端准备
<><1>双击安装hadoop依赖
<><2>放在非中文路径下,双击winutils.exe进行安装
如果安装错误,先安装微软运行库
<><3>新建Java Maven 工程
<><4>maven坐标
<dependencies> <dependency> <groupId>org.apache.hadoop</groupId> <artifactId>
hadoop-client</artifactId> <version>3.1.3</version> </dependency> <dependency> <
groupId>junit</groupId> <artifactId>junit</artifactId> <version>4.12</version>
</dependency> <dependency> <groupId>org.slf4j</groupId> <artifactId>
slf4j-log4j12</artifactId> <version>1.7.30</version> </dependency> </
dependencies>
<><5>在sre/main/resources下新建log4j.properties配置文件
log4j.rootLogger=INFO, stdout
log4j.appender.stdout=org.apache.log4j.ConsoleAppender
log4j.appender.stdout.layout=org.apache.log4j.PatternLayout
log4j.appender.stdout.layout.ConversionPattern=%d %p [%c] - %m%n
log4j.appender.logfile=org.apache.log4j.FileAppender
log4j.appender.logfile.File=target/spring.log
log4j.appender.logfile.layout=org.apache.log4j.PatternLayout
log4j.appender.logfile.layout.ConversionPattern=%d %p [%c] - %m%n
<><6>新建类 HdfsClient
<>二、API操作
<><1>创建文件夹
package com.demo.hdfs; import org.apache.hadoop.conf.Configuration; import org.
apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; import org.junit.
After; import org.junit.Before; import org.junit.Test; import java.io.
IOException; import java.net.URI; import java.net.URISyntaxException; /** *
客户端代码操作 * 1. 获取一个客户端对象 * 2. 执行相关操作 * 3. 关闭资源 */ public class HdfsClient {
private FileSystem fs; @Before public void init() throws URISyntaxException,
InterruptedException, IOException { //连接集群nn地址 URI uri = new URI(
"hdfs://hadoop102:8020"); //创建一个配置文件 Configuration configuration = new
Configuration(); //用户 String user = "root"; //获取客户端对象 fs = FileSystem.get(uri,
configuration, user); } @After public void close() throws IOException { //关闭资源
fs.close(); } @Test public void testMkdir() throws IOException { //创建一个文件夹 fs.
mkdirs(new Path("/apipath1")); } }
<><2>上传
//上传 @Test public void testPut() throws IOException { //参数1 :是否删除元数据 //参数2:
是否允许覆盖 //参数3: 目的路径 fs.copyFromLocalFile(false, true, new Path(
"E:\\study\\BigData\\Hadoop\\脚本\\myhadoop.sh"), new Path("hdfs://hadoop102/code"
)); }
<><3>下载
//下载 @Test public void testGet() throws IOException { //参数1 :是否删除原数据 //参数2:
目的路径 //参数3: 原文件路径 //参数4:是否进行本地校验 ,false 开启,ture 关闭 fs.copyToLocalFile(false, new
Path("hdfs://hadoop102/code/myhadoop.sh"), new Path("E:\\"), true); }
.xxx.crc文件是一种校验方法
<><4>删除
//删除 @Test public void testRm() throws IOException { //参数1 :目的路径 //参数2: 是否递归删除
//1. 删除文件 //fs.delete(new Path("hdfs://hadoop102/code/myhadoop.sh"), true);
//2. 删除空目录 //fs.delete( new Path("hdfs://hadoop102/code/code1"), true); //3.
删除非空目录,如果不是递归删除,则报错 fs.delete(new Path("hdfs://hadoop102/code/code2"), true); }
<><5>文件更名和移动
//文件更名和移动 @Test public void testMv() throws IOException { //参数1 :原文件路径 //参数2:
目的文件路径 //1.文件名称更改 //fs.rename(new Path("hdfs://hadoop102/code/myhadoop.sh"),
new Path("hdfs://hadoop102/code/myhadoop_rename.sh")); //2.文件移动并修改
//fs.rename(new Path("hdfs://hadoop102/code/myhadoop_rename.sh"), new
Path("hdfs://hadoop102/myhadoop_remove_and_rename.sh")); //3. 目录更名 fs.rename(new
Path("hdfs://hadoop102/code"), new Path("hdfs://hadoop102/code_rename")); }
<><6>文件详细信息查看
查看文件名称、权限、长度、块信息
//文件详细信息查看 @Test public void testFileDetail() throws IOException { //参数1 :路径
//参数2: 目的文件路径 //获取所有文件信息 RemoteIterator<LocatedFileStatus> listFiles = fs.
listFiles(new Path("hdfs://hadoop102/"), true); //遍历文件信息 while (listFiles.
hasNext()) { LocatedFileStatus fileStatus = listFiles.next(); System.out.println
("-----------------------------------------"); System.out.println(fileStatus.
toString()); System.out.println("-----------------------------------------"); }
}
toString方法
public String toString() { StringBuilder sb = new StringBuilder(); sb.append(
this.getClass().getSimpleName()); sb.append("{"); sb.append("path=" + this.path)
; sb.append("; isDirectory=" + this.isdir); if (!this.isDirectory()) { sb.append
("; length=" + this.length); sb.append("; replication=" + this.block_replication
); sb.append("; blocksize=" + this.blocksize); } sb.append(";
modification_time=" + this.modification_time); sb.append("; access_time=" + this
.access_time); sb.append("; owner=" + this.owner); sb.append("; group=" + this.
group); sb.append("; permission=" + this.permission); sb.append("; isSymlink=" +
this.isSymlink()); if (this.isSymlink()) { try { sb.append("; symlink=" + this.
getSymlink()); } catch (IOException var3) { throw new RuntimeException(
"Unexpected exception", var3); } } sb.append("; hasAcl=" + this.hasAcl()); sb.
append("; isEncrypted=" + this.isEncrypted()); sb.append("; isErasureCoded=" +
this.isErasureCoded()); sb.append("}"); return sb.toString(); }
<><7>文件和文件夹判断
//判断是文件夹还是文件 @Test public void testFile() throws IOException { //获取所有文件信息
FileStatus[] fileStatuses = fs.listStatus(new Path("hdfs://hadoop102/"));
//遍历文件信息 for (FileStatus fileStatus : fileStatuses) { if (fileStatus.isFile()) {
System.out.println("this is file--" + fileStatus.toString()); System.out.println
("-----------------------------------------"); } else { System.out.println(
"this is path--" + fileStatus.toString()); System.out.println(
"-----------------------------------------"); } } }