飞道的博客

MapReduce入门

217人阅读  评论(0)

MapReduce概述

YARN运行Jar脚本

MapReduce模板

编写代码时,要注意导包的问题。全部选择含Hadoop字眼的包,导错包将会导致程序无法正常运行。

Driver模板

package com.aa.mode;

import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.conf.Configured;
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.io.Text;
import org.apache.hadoop.mapreduce.Job;
import org.apache.hadoop.mapreduce.lib.input.TextInputFormat;
import org.apache.hadoop.mapreduce.lib.output.TextOutputFormat;
import org.apache.hadoop.util.Tool;
import org.apache.hadoop.util.ToolRunner;

public class MRDriverMode extends Configured implements Tool {
   

    //构建、配置、连接、提交Job
    public int run(String[] args) throws Exception{
   

        //构建Job
        //实例化一个MapReduce的Job对象
        Job job = Job.getInstance(this.getConf(), "mode");
        //指定允许jar包运行的类
        job.setJarByClass(MRDriverMode.class);

        //配置Job
        //Input:配置输入
        //指定输入类的类型
        job.setInputFormatClass(TextInputFormat.class);//可以不指定,不指定就是用默认值TextInputFormat
        //指定输入源
        Path inputPath = new Path(args[0]);//使用第一个参数作为程序的输入
        TextInputFormat.setInputPaths(job,inputPath);

        //Map:配置Map...
        job.setMapperClass(MRMapperMode.class);//设置调用Mapper类
        job.setMapOutputKeyClass(Text.class);//设置K2的类型
        job.setMapOutputValueClass(Text.class);//这只V2的类型

        //Shuffle:配置Shuffle...先写null占位,再慢慢改
        //job.setPartitionerClass(null);//设置分区器
        //job.setSortComparatorClass(null);//设置排序器
        //job.setGroupingComparatorClass(null);//设置分组器
        //job.setCombinerClass(null);//设置Map端聚合



        //Reduce:配置Reduce...
        job.setReducerClass(MRReducerMode.class);//设置调用reduce的类
        job.setOutputKeyClass(Text.class);//设置K3的类型
        job.setOutputValueClass(Text.class);//设置V3的类型
        //job.setNumReduceTasks(1);//设置ReduceTask的个数,默认为1个,写不写都可以

        //Output:配置输出
        //指定输出类的类型
        job.setOutputFormatClass(TextOutputFormat.class);//默认就是TextOutputFormat
        //设置输出的路径
        Path outputPath = new Path(args[1]);
        //判断输出路径是否存在,存在就删除
        FileSystem fs = FileSystem.get(this.getConf());
        if(fs.exists(outputPath)){
   
            fs.delete(outputPath,true);//使用true为递归删除,可以删除目录
        }
        TextOutputFormat.setOutputPath(job,outputPath);

        //提交Job
        return job.waitForCompletion(true)?0:-1;//true会显示执行过程。该方法内部会submit
    }


    //程序入口方法
    public static void main(String[] args) throws Exception{
   
        //构建配置连接对象
        Configuration conf = new Configuration();
        //调用run方法
        int status = ToolRunner.run(conf, new MRDriverMode(), args);
        System.out.println("status = " + status);
        //退出程序
        System.exit(status);
    }
}

配置Map与Reduce的参数时,可以先写null占位(防止报错),再慢慢改。

Map模板

package com.aa.mode;

import org.apache.hadoop.io.LongWritable;
import org.apache.hadoop.io.Text;
import org.apache.hadoop.mapreduce.Mapper;

import java.io.IOException;

public class MRMapperMode extends Mapper<LongWritable, Text,Text,Text> {
   //这4个参数为:KEYIN, VALUEIN, KEYOUT, VALUEOUT
    @Override
    protected void map(LongWritable key, Text value, Context context) throws IOException, InterruptedException {
   
        //Map处理的逻辑由实际的需求决定
    }
}

Reduce模板

package com.aa.mode;

import org.apache.hadoop.io.Text;
import org.apache.hadoop.mapreduce.Reducer;

import java.io.IOException;

/**
 *
 *
 */
public class MRReducerMode extends Reducer<Text,Text,Text,Text> {
   //这4个参数为:KEYIN, VALUEIN, KEYOUT, VALUEOUT
    @Override
    protected void reduce(Text key, Iterable<Text> values, Context context) throws IOException, InterruptedException {
   
        //Reduce的处理逻辑由实际的需求决定
    }
}

以后就可以参照这套模板来写MapReduce代码,又回到了Ctrl+C,Ctrl+V改参数的年代。。。

WordCount小项目

Driver类

package com.aa.wordcount;

import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.conf.Configured;
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.io.IntWritable;
import org.apache.hadoop.io.Text;
import org.apache.hadoop.mapreduce.Job;
import org.apache.hadoop.mapreduce.lib.input.TextInputFormat;
import org.apache.hadoop.mapreduce.lib.output.TextOutputFormat;
import org.apache.hadoop.util.Tool;
import org.apache.hadoop.util.ToolRunner;

public class WordCountDriver extends Configured implements Tool {
   

    @Override
    public int run(String[] args) throws Exception {
   

        //构建
        Job job = Job.getInstance(this.getConf(), "userwc");
        job.setJarByClass(WordCountDriver.class);

        //配置
        job.setInputFormatClass(TextInputFormat.class);
        //使用程序的第一个参数作为输入
        TextInputFormat.setInputPaths(job, new Path(args[0]));

        job.setMapperClass(WordCountMapper.class);
        job.setMapOutputKeyClass(Text.class);
        job.setMapOutputValueClass(IntWritable.class);

        job.setReducerClass(WordCountReduce.class);
        job.setOutputKeyClass(Text.class);
        job.setOutputValueClass(IntWritable.class);

        job.setOutputFormatClass(TextOutputFormat.class);

        //使用程序的第二个参数作为输出路径
        Path outputPath = new Path(args[1]);
        FileSystem fs = FileSystem.get(this.getConf());
        if (fs.exists(outputPath)) {
   
            fs.delete(outputPath, true);
        }
        TextOutputFormat.setOutputPath(job, outputPath);
        //提交
        return job.waitForCompletion(true) ? 0 : 1;
    }

    public static void main(String[] args) throws Exception {
   
        Configuration conf = new Configuration();
        int status = ToolRunner.run(conf, new WordCountDriver(), args);
        System.out.println("status = " + status);
        System.exit(status);
    }

}

Mapper类

package com.aa.wordcount;

import org.apache.hadoop.io.IntWritable;
import org.apache.hadoop.io.LongWritable;
import org.apache.hadoop.io.Text;
import org.apache.hadoop.mapreduce.Mapper;

import java.io.IOException;

public class WordCountMapper extends Mapper<LongWritable, Text,Text, IntWritable> {
   
    //输出的K2
    Text outputKey = new Text();
    //输出的V2
    IntWritable outputValue = new IntWritable(1);

    //每一条KV调用一次map

    @Override//直接输入map就可以重写
    protected void map(LongWritable key, Text value, Context context) throws IOException, InterruptedException {
   
        //将每行的内容分割得到每个单词
        String[] words = value.toString().split("\\s+");
        //根据给定的正则表达式的匹配来拆分此字符串,空格、回车、换行等空白符都可以

        //迭代取出每个单词作为K2,iter,反编译看到,数组的增强for循环遍历底层还是普通for循环
        for (String word : words) {
   
            //将当前的单词作为K2
            this.outputKey.set(word);
            //将K2和V2传递到下一步
            context.write(outputKey,outputValue);
        }

    }
}

Reducer类

package com.aa.wordcount;

import org.apache.hadoop.io.IntWritable;
import org.apache.hadoop.io.Text;
import org.apache.hadoop.mapreduce.Reducer;

import java.io.IOException;

public class WordCountReduce extends Reducer<Text, IntWritable, Text, IntWritable> {
   

    //输出V3
    IntWritable outputValue = new IntWritable();
    //由于K3=K2,直接使用即可,不需要新建变量

    @Override
    protected void reduce(Text key, Iterable<IntWritable> values, Context context) throws IOException, InterruptedException {
   
        int sum=0;
        for (IntWritable value : values) {
   
            //取出当前单词所有1,进行累加
            sum+=value.get();
        }
        //给V3赋值
        this.outputValue.set(sum);
        //传递到下一步
        context.write(key,outputValue);
    }
}

集群运行

打开集群

参照集群的开机顺序

启动ZooKeeper:

cd /export/server/zookeeper-3.4.6/
bin/zkServer.sh status
bin/zkServer.sh start
bin/zkServer.sh status

启动HDFS:

start-dfs.sh

启动YARN

start-yarn.sh

使用jps查看进程。

上传文件

上传wc.txt文件:

cd /export/data/

rz上传。
创建HDFS目录:

hdfs dfs -mkdir -p /wordcount/input

复制文件到HDFS:

hdfs dfs -put /export/data/wc.txt /wordcount/input

查看是否成功:

hdfs dfs -ls /wordcount/input

上传Maven编译好的jar包:

cd /export/data/

使用rz上传。

运行脚本

使用YARN运行jar脚本:

yarn jar /export/data/day210426-1.0-SNAPSHOT.jar com.aa.wordcount.WordCountDriver /wordcount/input/wc.txt /wordcount/output1

运行过程:

[root@node1 data]# yarn jar /export/data/day210426-1.0-SNAPSHOT.jar com.aa.wordcount.WordCountDriver /wordcount/input/wc.txt /wordcount/output1       
21/04/26 20:23:27 INFO client.RMProxy: Connecting to ResourceManager at node3/192.168.88.11:8032
21/04/26 20:23:27 INFO input.FileInputFormat: Total input paths to process : 1
21/04/26 20:23:28 INFO mapreduce.JobSubmitter: number of splits:3
21/04/26 20:23:28 INFO mapreduce.JobSubmitter: Submitting tokens for job: job_1619438325526_0002
21/04/26 20:23:28 INFO impl.YarnClientImpl: Submitted application application_1619438325526_0002
21/04/26 20:23:28 INFO mapreduce.Job: The url to track the job: http://node3:8088/proxy/application_1619438325526_0002/
21/04/26 20:23:28 INFO mapreduce.Job: Running job: job_1619438325526_0002
21/04/26 20:23:39 INFO mapreduce.Job: Job job_1619438325526_0002 running in uber mode : false
21/04/26 20:23:39 INFO mapreduce.Job:  map 0% reduce 0%
21/04/26 20:24:00 INFO mapreduce.Job:  map 9% reduce 0%
21/04/26 20:24:03 INFO mapreduce.Job:  map 10% reduce 0%
21/04/26 20:24:06 INFO mapreduce.Job:  map 12% reduce 0%
21/04/26 20:24:07 INFO mapreduce.Job:  map 13% reduce 0%
21/04/26 20:24:09 INFO mapreduce.Job:  map 14% reduce 0%
21/04/26 20:24:10 INFO mapreduce.Job:  map 18% reduce 0%
21/04/26 20:24:13 INFO mapreduce.Job:  map 20% reduce 0%
21/04/26 20:24:15 INFO mapreduce.Job:  map 22% reduce 0%
21/04/26 20:24:16 INFO mapreduce.Job:  map 24% reduce 0%
21/04/26 20:24:18 INFO mapreduce.Job:  map 25% reduce 0%
21/04/26 20:24:19 INFO mapreduce.Job:  map 28% reduce 0%
21/04/26 20:24:21 INFO mapreduce.Job:  map 31% reduce 0%
21/04/26 20:24:22 INFO mapreduce.Job:  map 32% reduce 0%
21/04/26 20:24:24 INFO mapreduce.Job:  map 35% reduce 0%
21/04/26 20:24:25 INFO mapreduce.Job:  map 36% reduce 0%
21/04/26 20:24:27 INFO mapreduce.Job:  map 39% reduce 0%
21/04/26 20:24:28 INFO mapreduce.Job:  map 40% reduce 0%
21/04/26 20:24:31 INFO mapreduce.Job:  map 43% reduce 0%
21/04/26 20:24:32 INFO mapreduce.Job:  map 45% reduce 0%
21/04/26 20:24:34 INFO mapreduce.Job:  map 47% reduce 0%
21/04/26 20:24:35 INFO mapreduce.Job:  map 48% reduce 0%
21/04/26 20:24:37 INFO mapreduce.Job:  map 52% reduce 0%
21/04/26 20:24:39 INFO mapreduce.Job:  map 53% reduce 0%
21/04/26 20:24:40 INFO mapreduce.Job:  map 55% reduce 0%
21/04/26 20:24:42 INFO mapreduce.Job:  map 56% reduce 0%
21/04/26 20:24:43 INFO mapreduce.Job:  map 58% reduce 0%
21/04/26 20:24:45 INFO mapreduce.Job:  map 60% reduce 0%
21/04/26 20:24:46 INFO mapreduce.Job:  map 64% reduce 0%
21/04/26 20:24:49 INFO mapreduce.Job:  map 65% reduce 0%
21/04/26 20:24:52 INFO mapreduce.Job:  map 68% reduce 0%
21/04/26 20:24:54 INFO mapreduce.Job:  map 69% reduce 0%
21/04/26 20:24:55 INFO mapreduce.Job:  map 71% reduce 0%
21/04/26 20:24:57 INFO mapreduce.Job:  map 73% reduce 0%
21/04/26 20:24:58 INFO mapreduce.Job:  map 75% reduce 0%
21/04/26 20:25:01 INFO mapreduce.Job:  map 77% reduce 0%
21/04/26 20:25:02 INFO mapreduce.Job:  map 80% reduce 0%
21/04/26 20:25:04 INFO mapreduce.Job:  map 81% reduce 0%
21/04/26 20:25:05 INFO mapreduce.Job:  map 85% reduce 0%
21/04/26 20:25:07 INFO mapreduce.Job:  map 88% reduce 0%
21/04/26 20:25:08 INFO mapreduce.Job:  map 89% reduce 0%
21/04/26 20:25:10 INFO mapreduce.Job:  map 91% reduce 0%
21/04/26 20:25:11 INFO mapreduce.Job:  map 93% reduce 0%
21/04/26 20:25:13 INFO mapreduce.Job:  map 95% reduce 0%
21/04/26 20:25:14 INFO mapreduce.Job:  map 98% reduce 0%
21/04/26 20:25:17 INFO mapreduce.Job:  map 100% reduce 0%
21/04/26 20:25:23 INFO mapreduce.Job:  map 100% reduce 11%
21/04/26 20:25:26 INFO mapreduce.Job:  map 100% reduce 22%
21/04/26 20:25:29 INFO mapreduce.Job:  map 100% reduce 67%
21/04/26 20:25:32 INFO mapreduce.Job:  map 100% reduce 69%
21/04/26 20:25:35 INFO mapreduce.Job:  map 100% reduce 71%
21/04/26 20:25:38 INFO mapreduce.Job:  map 100% reduce 74%
21/04/26 20:25:41 INFO mapreduce.Job:  map 100% reduce 77%
21/04/26 20:25:44 INFO mapreduce.Job:  map 100% reduce 80%
21/04/26 20:25:47 INFO mapreduce.Job:  map 100% reduce 83%
21/04/26 20:25:50 INFO mapreduce.Job:  map 100% reduce 85%
21/04/26 20:25:53 INFO mapreduce.Job:  map 100% reduce 88%
21/04/26 20:25:56 INFO mapreduce.Job:  map 100% reduce 90%
21/04/26 20:25:59 INFO mapreduce.Job:  map 100% reduce 92%
21/04/26 20:26:02 INFO mapreduce.Job:  map 100% reduce 95%
21/04/26 20:26:05 INFO mapreduce.Job:  map 100% reduce 98%
21/04/26 20:26:08 INFO mapreduce.Job:  map 100% reduce 100%
21/04/26 20:26:08 INFO mapreduce.Job: Job job_1619438325526_0002 completed successfully
21/04/26 20:26:09 INFO mapreduce.Job: Counters: 49
        File System Counters
                FILE: Number of bytes read=1679368168
                FILE: Number of bytes written=2519541603
                FILE: Number of read operations=0
                FILE: Number of large read operations=0
                FILE: Number of write operations=0
                HDFS: Number of bytes read=380018507
                HDFS: Number of bytes written=82
                HDFS: Number of read operations=12
                HDFS: Number of large read operations=0
                HDFS: Number of write operations=2
        Job Counters 
                Launched map tasks=3
                Launched reduce tasks=1
                Data-local map tasks=3
                Total time spent by all maps in occupied slots (ms)=277120
                Total time spent by all reduces in occupied slots (ms)=55989
                Total time spent by all map tasks (ms)=277120
                Total time spent by all reduce tasks (ms)=55989
                Total vcore-milliseconds taken by all map tasks=277120
                Total vcore-milliseconds taken by all reduce tasks=55989
                Total megabyte-milliseconds taken by all map tasks=283770880
                Total megabyte-milliseconds taken by all reduce tasks=57332736
        Map-Reduce Framework
                Map input records=33308000
                Map output records=79388000
                Map output bytes=680908000
                Map output materialized bytes=839684018
                Input split bytes=315
                Combine input records=0
                Combine output records=0
                Reduce input groups=6
                Reduce shuffle bytes=839684018
                Reduce input records=79388000
                Reduce output records=6
                Spilled Records=238164000
                Shuffled Maps =3
                Failed Shuffles=0
                Merged Map outputs=3
                GC time elapsed (ms)=8971
                CPU time spent (ms)=212670
                Physical memory (bytes) snapshot=1000062976
                Virtual memory (bytes) snapshot=8431161344
                Total committed heap usage (bytes)=731906048
        Shuffle Errors
                BAD_ID=0
                CONNECTION=0
                IO_ERROR=0
                WRONG_LENGTH=0
                WRONG_MAP=0
                WRONG_REDUCE=0
        File Input Format Counters 
                Bytes Read=380018192
        File Output Format Counters 
                Bytes Written=82
status = 0

笔者的老爷机很吃力呢:

node1:50070可以看到:

可以看出,自己写的wordcount程序计算出了正确的结果。和官方自带的示例程序相比,可以删除原有目录。

node3:8088还可以看到提交过的MapReduce任务。

参照HDFS集群关机顺序关闭集群。

关闭集群

参照关闭集群的顺序
node1关闭HDFS:

stop-dfs.sh

关闭YARN:

stop-yarn.sh

3台机关闭ZooKeeper:

cd /export/server/zookeeper-3.4.6/
bin/zkServer.sh stop

3台机断电:

poweroff

本地运行

扔集群运行还是比较麻烦的,好在Java可以跨平台,在本地运行也可以测试代码的正确性。

package com.aa.wordcount_local;

import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.conf.Configured;
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.io.IntWritable;
import org.apache.hadoop.io.LongWritable;
import org.apache.hadoop.io.Text;
import org.apache.hadoop.mapreduce.Job;
import org.apache.hadoop.mapreduce.Mapper;
import org.apache.hadoop.mapreduce.Reducer;
import org.apache.hadoop.mapreduce.lib.input.TextInputFormat;
import org.apache.hadoop.mapreduce.lib.output.TextOutputFormat;
import org.apache.hadoop.util.Tool;
import org.apache.hadoop.util.ToolRunner;

import java.io.IOException;

public class WordCountLocal extends Configured implements Tool {
   

    @Override
    public int run(String[] args) throws Exception {
   
        //构建
        Job job = Job.getInstance(this.getConf(), "userwc_local");
        job.setJarByClass(WordCountLocal.class);

        //配置
        job.setInputFormatClass(TextInputFormat.class);
        //使用程序的第一个参数作为输入
        TextInputFormat.setInputPaths(job, new Path(args[0]));

        job.setMapperClass(WCMapper.class);
        job.setMapOutputKeyClass(Text.class);
        job.setMapOutputValueClass(IntWritable.class);

        job.setReducerClass(WCReducer.class);
        job.setOutputKeyClass(Text.class);
        job.setOutputValueClass(IntWritable.class);

        job.setOutputFormatClass(TextOutputFormat.class);
        //使用程序的第二个参数作为输出路径
        Path outputPath = new Path(args[1]);
        FileSystem fs = FileSystem.get(this.getConf());
        if (fs.exists(outputPath)) {
   
            fs.delete(outputPath, true);//true为递归删除
        }
        TextOutputFormat.setOutputPath(job, outputPath);

        return job.waitForCompletion(true) ? 0 : -1;
    }

    public static void main(String[] args) throws Exception {
   
        Configuration conf = new Configuration();
        int status = ToolRunner.run(conf, new WordCountLocal(), args);
        System.out.println("status = " + status);
        System.exit(status);
    }

    public static class WCMapper extends Mapper<LongWritable, Text, Text, IntWritable> {
   
        //输出的K2
        Text outputKey = new Text();
        //输出的V2
        IntWritable outputValue = new IntWritable(1);//由于是平权的,使之恒=1便可以计数

        @Override
        protected void map(LongWritable key, Text value, Context context) throws IOException, InterruptedException {
   
            //将每行的内容分割得到每个单词
            String[] words = value.toString().split("\\s+");
            //迭代取出每个单词作为K2
            for (String word : words) {
   
                //将当前的单词作为K2
                this.outputKey.set(word);
                //将K2和V2传递到下一步
                context.write(outputKey,outputValue);
            }
        }
    }

    public static class WCReducer extends Reducer<Text,IntWritable,Text,IntWritable>{
   
        //K3=K2,不需要重复定义变量
        //输出V3
        IntWritable outputValue = new IntWritable();

        @Override
        protected void reduce(Text key, Iterable<IntWritable> values, Context context) throws IOException, InterruptedException {
   
            int sum=0;
            for (IntWritable value : values) {
   
                sum+=value.get();
            }
            //给V3赋值
            this.outputValue.set(sum);
            //传递到下一步
            context.write(key,this.outputValue);
        }
    }

}

当放到同一个类中时,Mapper和Reducer必须申明为static(别问我为神马)。

本地部署

Hadoop本地环境变量配置
配置成功后可以在本地运行。idea中进行如下设置:

Run→Edit →Configus



修改内容为:

com.aa.wordcount_local.WordCountLocal

传递的2个路径变量使用空格隔开:

E:\bigdata\2021.4.26\wc.txt E:\bigdata\2021.4.26\result

Apply之后即可直接运行。

运行过程:

"C:\Program Files\Java\jdk1.8.0_241\bin\java.exe" "-javaagent:C:\Program Files\JetBrains\IntelliJ IDEA 2019.3.5\lib\idea_rt.jar=59719:C:\Program Files\JetBrains\IntelliJ IDEA 2019.3.5\bin" -Dfile.encoding=UTF-8 -classpath "C:\Program Files\Java\jdk1.8.0_241\jre\lib\charsets.jar;C:\Program Files\Java\jdk1.8.0_241\jre\lib\deploy.jar;C:\Program Files\Java\jdk1.8.0_241\jre\lib\ext\access-bridge-64.jar;C:\Program Files\Java\jdk1.8.0_241\jre\lib\ext\cldrdata.jar;C:\Program Files\Java\jdk1.8.0_241\jre\lib\ext\dnsns.jar;C:\Program Files\Java\jdk1.8.0_241\jre\lib\ext\jaccess.jar;C:\Program Files\Java\jdk1.8.0_241\jre\lib\ext\jfxrt.jar;C:\Program Files\Java\jdk1.8.0_241\jre\lib\ext\localedata.jar;C:\Program Files\Java\jdk1.8.0_241\jre\lib\ext\nashorn.jar;C:\Program Files\Java\jdk1.8.0_241\jre\lib\ext\sunec.jar;C:\Program Files\Java\jdk1.8.0_241\jre\lib\ext\sunjce_provider.jar;C:\Program Files\Java\jdk1.8.0_241\jre\lib\ext\sunmscapi.jar;C:\Program Files\Java\jdk1.8.0_241\jre\lib\ext\sunpkcs11.jar;C:\Program Files\Java\jdk1.8.0_241\jre\lib\ext\zipfs.jar;C:\Program Files\Java\jdk1.8.0_241\jre\lib\javaws.jar;C:\Program Files\Java\jdk1.8.0_241\jre\lib\jce.jar;C:\Program Files\Java\jdk1.8.0_241\jre\lib\jfr.jar;C:\Program Files\Java\jdk1.8.0_241\jre\lib\jfxswt.jar;C:\Program Files\Java\jdk1.8.0_241\jre\lib\jsse.jar;C:\Program Files\Java\jdk1.8.0_241\jre\lib\management-agent.jar;C:\Program Files\Java\jdk1.8.0_241\jre\lib\plugin.jar;C:\Program Files\Java\jdk1.8.0_241\jre\lib\resources.jar;C:\Program Files\Java\jdk1.8.0_241\jre\lib\rt.jar;C:\Users\killer\IdeaProjects\bigdata_demo\day210426\target\classes;C:\Program Files\apache-maven-3.3.9\Maven_Repository\org\apache\hadoop\hadoop-common\2.7.5\hadoop-common-2.7.5.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\org\apache\hadoop\hadoop-annotations\2.7.5\hadoop-annotations-2.7.5.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\com\google\guava\guava\11.0.2\guava-11.0.2.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\commons-cli\commons-cli\1.2\commons-cli-1.2.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\org\apache\commons\commons-math3\3.1.1\commons-math3-3.1.1.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\xmlenc\xmlenc\0.52\xmlenc-0.52.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\commons-httpclient\commons-httpclient\3.1\commons-httpclient-3.1.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\commons-codec\commons-codec\1.4\commons-codec-1.4.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\commons-io\commons-io\2.4\commons-io-2.4.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\commons-net\commons-net\3.1\commons-net-3.1.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\commons-collections\commons-collections\3.2.2\commons-collections-3.2.2.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\javax\servlet\servlet-api\2.5\servlet-api-2.5.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\org\mortbay\jetty\jetty\6.1.26\jetty-6.1.26.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\org\mortbay\jetty\jetty-util\6.1.26\jetty-util-6.1.26.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\org\mortbay\jetty\jetty-sslengine\6.1.26\jetty-sslengine-6.1.26.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\javax\servlet\jsp\jsp-api\2.1\jsp-api-2.1.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\com\sun\jersey\jersey-core\1.9\jersey-core-1.9.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\com\sun\jersey\jersey-json\1.9\jersey-json-1.9.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\org\codehaus\jettison\jettison\1.1\jettison-1.1.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\com\sun\xml\bind\jaxb-impl\2.2.3-1\jaxb-impl-2.2.3-1.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\org\codehaus\jackson\jackson-jaxrs\1.8.3\jackson-jaxrs-1.8.3.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\org\codehaus\jackson\jackson-xc\1.8.3\jackson-xc-1.8.3.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\com\sun\jersey\jersey-server\1.9\jersey-server-1.9.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\asm\asm\3.1\asm-3.1.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\commons-logging\commons-logging\1.1.3\commons-logging-1.1.3.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\log4j\log4j\1.2.17\log4j-1.2.17.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\net\java\dev\jets3t\jets3t\0.9.0\jets3t-0.9.0.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\org\apache\httpcomponents\httpclient\4.1.2\httpclient-4.1.2.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\org\apache\httpcomponents\httpcore\4.1.2\httpcore-4.1.2.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\com\jamesmurty\utils\java-xmlbuilder\0.4\java-xmlbuilder-0.4.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\commons-lang\commons-lang\2.6\commons-lang-2.6.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\commons-configuration\commons-configuration\1.6\commons-configuration-1.6.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\commons-digester\commons-digester\1.8\commons-digester-1.8.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\commons-beanutils\commons-beanutils\1.7.0\commons-beanutils-1.7.0.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\commons-beanutils\commons-beanutils-core\1.8.0\commons-beanutils-core-1.8.0.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\org\slf4j\slf4j-api\1.7.10\slf4j-api-1.7.10.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\org\slf4j\slf4j-log4j12\1.7.10\slf4j-log4j12-1.7.10.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\org\codehaus\jackson\jackson-core-asl\1.9.13\jackson-core-asl-1.9.13.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\org\codehaus\jackson\jackson-mapper-asl\1.9.13\jackson-mapper-asl-1.9.13.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\org\apache\avro\avro\1.7.4\avro-1.7.4.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\com\thoughtworks\paranamer\paranamer\2.3\paranamer-2.3.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\org\xerial\snappy\snappy-java\1.0.4.1\snappy-java-1.0.4.1.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\com\google\protobuf\protobuf-java\2.5.0\protobuf-java-2.5.0.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\com\google\code\gson\gson\2.2.4\gson-2.2.4.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\org\apache\hadoop\hadoop-auth\2.7.5\hadoop-auth-2.7.5.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\org\apache\directory\server\apacheds-kerberos-codec\2.0.0-M15\apacheds-kerberos-codec-2.0.0-M15.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\org\apache\directory\server\apacheds-i18n\2.0.0-M15\apacheds-i18n-2.0.0-M15.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\org\apache\directory\api\api-asn1-api\1.0.0-M20\api-asn1-api-1.0.0-M20.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\org\apache\directory\api\api-util\1.0.0-M20\api-util-1.0.0-M20.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\org\apache\curator\curator-framework\2.7.1\curator-framework-2.7.1.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\com\jcraft\jsch\0.1.54\jsch-0.1.54.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\org\apache\curator\curator-client\2.7.1\curator-client-2.7.1.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\org\apache\curator\curator-recipes\2.7.1\curator-recipes-2.7.1.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\com\google\code\findbugs\jsr305\3.0.0\jsr305-3.0.0.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\org\apache\htrace\htrace-core\3.1.0-incubating\htrace-core-3.1.0-incubating.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\org\apache\zookeeper\zookeeper\3.4.6\zookeeper-3.4.6.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\org\apache\commons\commons-compress\1.4.1\commons-compress-1.4.1.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\org\tukaani\xz\1.0\xz-1.0.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\org\apache\hadoop\hadoop-client\2.7.5\hadoop-client-2.7.5.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\org\apache\hadoop\hadoop-mapreduce-client-app\2.7.5\hadoop-mapreduce-client-app-2.7.5.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\org\apache\hadoop\hadoop-mapreduce-client-common\2.7.5\hadoop-mapreduce-client-common-2.7.5.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\org\apache\hadoop\hadoop-yarn-client\2.7.5\hadoop-yarn-client-2.7.5.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\org\apache\hadoop\hadoop-yarn-server-common\2.7.5\hadoop-yarn-server-common-2.7.5.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\org\apache\hadoop\hadoop-mapreduce-client-shuffle\2.7.5\hadoop-mapreduce-client-shuffle-2.7.5.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\org\apache\hadoop\hadoop-yarn-api\2.7.5\hadoop-yarn-api-2.7.5.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\org\apache\hadoop\hadoop-mapreduce-client-jobclient\2.7.5\hadoop-mapreduce-client-jobclient-2.7.5.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\org\apache\hadoop\hadoop-hdfs\2.7.5\hadoop-hdfs-2.7.5.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\commons-daemon\commons-daemon\1.0.13\commons-daemon-1.0.13.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\io\netty\netty\3.6.2.Final\netty-3.6.2.Final.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\io\netty\netty-all\4.0.23.Final\netty-all-4.0.23.Final.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\xerces\xercesImpl\2.9.1\xercesImpl-2.9.1.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\xml-apis\xml-apis\1.3.04\xml-apis-1.3.04.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\org\fusesource\leveldbjni\leveldbjni-all\1.8\leveldbjni-all-1.8.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\org\apache\hadoop\hadoop-mapreduce-client-core\2.7.5\hadoop-mapreduce-client-core-2.7.5.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\org\apache\hadoop\hadoop-yarn-common\2.7.5\hadoop-yarn-common-2.7.5.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\javax\xml\bind\jaxb-api\2.2.2\jaxb-api-2.2.2.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\javax\xml\stream\stax-api\1.0-2\stax-api-1.0-2.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\javax\activation\activation\1.1\activation-1.1.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\com\sun\jersey\jersey-client\1.9\jersey-client-1.9.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\com\google\inject\guice\3.0\guice-3.0.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\javax\inject\javax.inject\1\javax.inject-1.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\aopalliance\aopalliance\1.0\aopalliance-1.0.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\com\sun\jersey\contribs\jersey-guice\1.9\jersey-guice-1.9.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\com\google\inject\extensions\guice-servlet\3.0\guice-servlet-3.0.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\junit\junit\4.13\junit-4.13.jar;C:\Program Files\apache-maven-3.3.9\Maven_Repository\org\hamcrest\hamcrest-core\1.3\hamcrest-core-1.3.jar" com.aa.wordcount_local.WordCountLocal E:\bigdata\2021.4.26\wc.txt E:\bigdata\2021.4.26\result
21/04/26 21:26:22 INFO Configuration.deprecation: session.id is deprecated. Instead, use dfs.metrics.session-id
21/04/26 21:26:22 INFO jvm.JvmMetrics: Initializing JVM Metrics with processName=JobTracker, sessionId=
21/04/26 21:26:24 WARN mapreduce.JobResourceUploader: No job jar file set.  User classes may not be found. See Job or Job#setJar(String).
21/04/26 21:26:24 INFO input.FileInputFormat: Total input paths to process : 1
21/04/26 21:26:24 INFO mapreduce.JobSubmitter: number of splits:12
21/04/26 21:26:24 INFO mapreduce.JobSubmitter: Submitting tokens for job: job_local588307001_0001
21/04/26 21:26:24 INFO mapreduce.Job: The url to track the job: http://localhost:8080/
21/04/26 21:26:24 INFO mapreduce.Job: Running job: job_local588307001_0001
21/04/26 21:26:24 INFO mapred.LocalJobRunner: OutputCommitter set in config null
21/04/26 21:26:24 INFO output.FileOutputCommitter: File Output Committer Algorithm version is 1
21/04/26 21:26:24 INFO mapred.LocalJobRunner: OutputCommitter is org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter
21/04/26 21:26:24 INFO mapred.LocalJobRunner: Waiting for map tasks
21/04/26 21:26:24 INFO mapred.LocalJobRunner: Starting task: attempt_local588307001_0001_m_000000_0
21/04/26 21:26:24 INFO output.FileOutputCommitter: File Output Committer Algorithm version is 1
21/04/26 21:26:24 INFO util.ProcfsBasedProcessTree: ProcfsBasedProcessTree currently is supported only on Linux.
21/04/26 21:26:24 INFO mapred.Task:  Using ResourceCalculatorProcessTree : org.apache.hadoop.yarn.util.WindowsBasedProcessTree@46db6474
21/04/26 21:26:24 INFO mapred.MapTask: Processing split: file:/E:/bigdata/2021.4.26/wc.txt:0+33554432
21/04/26 21:26:24 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)
21/04/26 21:26:24 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
21/04/26 21:26:24 INFO mapred.MapTask: soft limit at 83886080
21/04/26 21:26:24 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600
21/04/26 21:26:24 INFO mapred.MapTask: kvstart = 26214396; length = 6553600
21/04/26 21:26:24 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
21/04/26 21:26:25 INFO mapreduce.Job: Job job_local588307001_0001 running in uber mode : false
21/04/26 21:26:25 INFO mapreduce.Job:  map 0% reduce 0%
21/04/26 21:26:27 INFO mapred.MapTask: Spilling map output
21/04/26 21:26:27 INFO mapred.MapTask: bufstart = 0; bufend = 29274899; bufvoid = 104857600
21/04/26 21:26:27 INFO mapred.MapTask: kvstart = 26214396(104857584); kvend = 12561604(50246416); length = 13652793/6553600
21/04/26 21:26:27 INFO mapred.MapTask: (EQUATOR) 39760649 kvi 9940156(39760624)
21/04/26 21:26:28 INFO mapred.MapTask: Finished spill 0
21/04/26 21:26:28 INFO mapred.MapTask: (RESET) equator 39760649 kv 9940156(39760624) kvi 7318732(29274928)
21/04/26 21:26:30 INFO mapred.MapTask: Spilling map output
21/04/26 21:26:30 INFO mapred.MapTask: bufstart = 39760649; bufend = 69035531; bufvoid = 104857600
21/04/26 21:26:30 INFO mapred.MapTask: kvstart = 9940156(39760624); kvend = 22501764(90007056); length = 13652793/6553600
21/04/26 21:26:30 INFO mapred.MapTask: (EQUATOR) 79521285 kvi 19880316(79521264)
21/04/26 21:26:30 INFO mapred.LocalJobRunner: map > map
21/04/26 21:26:30 INFO mapred.LocalJobRunner: map > map
21/04/26 21:26:30 INFO mapred.MapTask: Starting flush of map output
21/04/26 21:26:31 INFO mapreduce.Job:  map 6% reduce 0%
21/04/26 21:26:32 INFO mapred.MapTask: Finished spill 1
21/04/26 21:26:32 INFO mapred.MapTask: (RESET) equator 79521285 kv 19880316(79521264) kvi 19146444(76585776)
21/04/26 21:26:32 INFO mapred.MapTask: Spilling map output
21/04/26 21:26:32 INFO mapred.MapTask: bufstart = 79521285; bufend = 81094879; bufvoid = 104857600
21/04/26 21:26:32 INFO mapred.MapTask: kvstart = 19880316(79521264); kvend = 19146448(76585792); length = 733869/6553600
21/04/26 21:26:32 INFO mapred.MapTask: Finished spill 2
21/04/26 21:26:32 INFO mapred.Merger: Merging 3 sorted segments
21/04/26 21:26:32 INFO mapred.Merger: Down to the last merge-pass, with 3 segments left of total size: 74143116 bytes
21/04/26 21:26:33 INFO mapred.LocalJobRunner: map > sort > 
21/04/26 21:26:36 INFO mapred.Task: Task:attempt_local588307001_0001_m_000000_0 is done. And is in the process of committing
21/04/26 21:26:36 INFO mapred.LocalJobRunner: map > sort
21/04/26 21:26:36 INFO mapred.Task: Task 'attempt_local588307001_0001_m_000000_0' done.
21/04/26 21:26:36 INFO mapred.Task: Final Counters for attempt_local588307001_0001_m_000000_0: Counters: 17
	File System Counters
		FILE: Number of bytes read=107703100
		FILE: Number of bytes written=148580113
		FILE: Number of read operations=0
		FILE: Number of large read operations=0
		FILE: Number of write operations=0
	Map-Reduce Framework
		Map input records=2941057
		Map output records=7009866
		Map output bytes=60123375
		Map output materialized bytes=74143113
		Input split bytes=98
		Combine input records=0
		Spilled Records=14019732
		Failed Shuffles=0
		Merged Map outputs=0
		GC time elapsed (ms)=55
		Total committed heap usage (bytes)=1497366528
	File Input Format Counters 
		Bytes Read=33558528
21/04/26 21:26:36 INFO mapred.LocalJobRunner: Finishing task: attempt_local588307001_0001_m_000000_0
21/04/26 21:26:36 INFO mapred.LocalJobRunner: Starting task: attempt_local588307001_0001_m_000001_0
21/04/26 21:26:36 INFO output.FileOutputCommitter: File Output Committer Algorithm version is 1
21/04/26 21:26:36 INFO util.ProcfsBasedProcessTree: ProcfsBasedProcessTree currently is supported only on Linux.
21/04/26 21:26:36 INFO mapred.Task:  Using ResourceCalculatorProcessTree : org.apache.hadoop.yarn.util.WindowsBasedProcessTree@30b25df7
21/04/26 21:26:36 INFO mapred.MapTask: Processing split: file:/E:/bigdata/2021.4.26/wc.txt:33554432+33554432
21/04/26 21:26:36 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)
21/04/26 21:26:36 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
21/04/26 21:26:36 INFO mapred.MapTask: soft limit at 83886080
21/04/26 21:26:36 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600
21/04/26 21:26:36 INFO mapred.MapTask: kvstart = 26214396; length = 6553600
21/04/26 21:26:36 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
21/04/26 21:26:36 INFO mapreduce.Job:  map 100% reduce 0%
21/04/26 21:26:38 INFO mapred.MapTask: Spilling map output
21/04/26 21:26:38 INFO mapred.MapTask: bufstart = 0; bufend = 29274882; bufvoid = 104857600
21/04/26 21:26:38 INFO mapred.MapTask: kvstart = 26214396(104857584); kvend = 12561604(50246416); length = 13652793/6553600
21/04/26 21:26:38 INFO mapred.MapTask: (EQUATOR) 39760641 kvi 9940156(39760624)
21/04/26 21:26:39 INFO mapred.MapTask: Finished spill 0
21/04/26 21:26:39 INFO mapred.MapTask: (RESET) equator 39760641 kv 9940156(39760624) kvi 7318728(29274912)
21/04/26 21:26:41 INFO mapred.MapTask: Spilling map output
21/04/26 21:26:41 INFO mapred.MapTask: bufstart = 39760641; bufend = 69035537; bufvoid = 104857600
21/04/26 21:26:41 INFO mapred.MapTask: kvstart = 9940156(39760624); kvend = 22501764(90007056); length = 13652793/6553600
21/04/26 21:26:41 INFO mapred.MapTask: (EQUATOR) 79521288 kvi 19880316(79521264)
21/04/26 21:26:41 INFO mapred.LocalJobRunner: 
21/04/26 21:26:41 INFO mapred.MapTask: Starting flush of map output
21/04/26 21:26:41 INFO mapreduce.Job:  map 8% reduce 0%
21/04/26 21:26:42 INFO mapred.LocalJobRunner: map > sort
21/04/26 21:26:42 INFO mapreduce.Job:  map 14% reduce 0%
21/04/26 21:26:42 INFO mapred.MapTask: Finished spill 1
21/04/26 21:26:42 INFO mapred.MapTask: (RESET) equator 79521288 kv 19880316(79521264) kvi 19146444(76585776)
21/04/26 21:26:42 INFO mapred.MapTask: Spilling map output
21/04/26 21:26:42 INFO mapred.MapTask: bufstart = 79521288; bufend = 81094874; bufvoid = 104857600
21/04/26 21:26:42 INFO mapred.MapTask: kvstart = 19880316(79521264); kvend = 19146448(76585792); length = 733869/6553600
21/04/26 21:26:42 INFO mapred.MapTask: Finished spill 2
21/04/26 21:26:42 INFO mapred.Merger: Merging 3 sorted segments
21/04/26 21:26:42 INFO mapred.Merger: Down to the last merge-pass, with 3 segments left of total size: 74143105 bytes
21/04/26 21:26:45 INFO mapred.LocalJobRunner: map > sort > 
21/04/26 21:26:45 INFO mapreduce.Job:  map 16% reduce 0%
21/04/26 21:26:46 INFO mapred.Task: Task:attempt_local588307001_0001_m_000001_0 is done. And is in the process of committing
21/04/26 21:26:46 INFO mapred.LocalJobRunner: map > sort
21/04/26 21:26:46 INFO mapred.Task: Task 'attempt_local588307001_0001_m_000001_0' done.
21/04/26 21:26:46 INFO mapred.Task: Final Counters for attempt_local588307001_0001_m_000001_0: Counters: 17
	File System Counters
		FILE: Number of bytes read=215405945
		FILE: Number of bytes written=296866361
		FILE: Number of read operations=0
		FILE: Number of large read operations=0
		FILE: Number of write operations=0
	Map-Reduce Framework
		Map input records=2941058
		Map output records=7009866
		Map output bytes=60123364
		Map output materialized bytes=74143102
		Input split bytes=98
		Combine input records=0
		Spilled Records=14019732
		Failed Shuffles=0
		Merged Map outputs=0
		GC time elapsed (ms)=115
		Total committed heap usage (bytes)=1507852288
	File Input Format Counters 
		Bytes Read=33558528
21/04/26 21:26:46 INFO mapred.LocalJobRunner: Finishing task: attempt_local588307001_0001_m_000001_0
21/04/26 21:26:46 INFO mapred.LocalJobRunner: Starting task: attempt_local588307001_0001_m_000002_0
21/04/26 21:26:46 INFO output.FileOutputCommitter: File Output Committer Algorithm version is 1
21/04/26 21:26:46 INFO util.ProcfsBasedProcessTree: ProcfsBasedProcessTree currently is supported only on Linux.
21/04/26 21:26:46 INFO mapred.Task:  Using ResourceCalculatorProcessTree : org.apache.hadoop.yarn.util.WindowsBasedProcessTree@5a8242f3
21/04/26 21:26:46 INFO mapred.MapTask: Processing split: file:/E:/bigdata/2021.4.26/wc.txt:67108864+33554432
21/04/26 21:26:46 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)
21/04/26 21:26:46 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
21/04/26 21:26:46 INFO mapred.MapTask: soft limit at 83886080
21/04/26 21:26:46 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600
21/04/26 21:26:46 INFO mapred.MapTask: kvstart = 26214396; length = 6553600
21/04/26 21:26:46 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
21/04/26 21:26:46 INFO mapreduce.Job:  map 100% reduce 0%
21/04/26 21:26:48 INFO mapred.MapTask: Spilling map output
21/04/26 21:26:48 INFO mapred.MapTask: bufstart = 0; bufend = 29274882; bufvoid = 104857600
21/04/26 21:26:48 INFO mapred.MapTask: kvstart = 26214396(104857584); kvend = 12561604(50246416); length = 13652793/6553600
21/04/26 21:26:48 INFO mapred.MapTask: (EQUATOR) 39760641 kvi 9940156(39760624)
21/04/26 21:26:49 INFO mapred.MapTask: Finished spill 0
21/04/26 21:26:49 INFO mapred.MapTask: (RESET) equator 39760641 kv 9940156(39760624) kvi 7318728(29274912)
21/04/26 21:26:51 INFO mapred.MapTask: Spilling map output
21/04/26 21:26:51 INFO mapred.MapTask: bufstart = 39760641; bufend = 69035523; bufvoid = 104857600
21/04/26 21:26:51 INFO mapred.MapTask: kvstart = 9940156(39760624); kvend = 22501764(90007056); length = 13652793/6553600
21/04/26 21:26:51 INFO mapred.MapTask: (EQUATOR) 79521281 kvi 19880316(79521264)
21/04/26 21:26:51 INFO mapred.LocalJobRunner: 
21/04/26 21:26:51 INFO mapred.MapTask: Starting flush of map output
21/04/26 21:26:51 INFO mapreduce.Job:  map 17% reduce 0%
21/04/26 21:26:52 INFO mapred.LocalJobRunner: map > sort
21/04/26 21:26:52 INFO mapreduce.Job:  map 22% reduce 0%
21/04/26 21:26:52 INFO mapred.MapTask: Finished spill 1
21/04/26 21:26:52 INFO mapred.MapTask: (RESET) equator 79521281 kv 19880316(79521264) kvi 19146448(76585792)
21/04/26 21:26:52 INFO mapred.MapTask: Spilling map output
21/04/26 21:26:52 INFO mapred.MapTask: bufstart = 79521281; bufend = 81094884; bufvoid = 104857600
21/04/26 21:26:52 INFO mapred.MapTask: kvstart = 19880316(79521264); kvend = 19146452(76585808); length = 733865/6553600
21/04/26 21:26:53 INFO mapred.MapTask: Finished spill 2
21/04/26 21:26:53 INFO mapred.Merger: Merging 3 sorted segments
21/04/26 21:26:53 INFO mapred.Merger: Down to the last merge-pass, with 3 segments left of total size: 74143106 bytes
21/04/26 21:26:55 INFO mapred.LocalJobRunner: map > sort > 
21/04/26 21:26:55 INFO mapreduce.Job:  map 24% reduce 0%
21/04/26 21:26:56 INFO mapred.Task: Task:attempt_local588307001_0001_m_000002_0 is done. And is in the process of committing
21/04/26 21:26:56 INFO mapred.LocalJobRunner: map > sort
21/04/26 21:26:56 INFO mapred.Task: Task 'attempt_local588307001_0001_m_000002_0' done.
21/04/26 21:26:56 INFO mapred.Task: Final Counters for attempt_local588307001_0001_m_000002_0: Counters: 17
	File System Counters
		FILE: Number of bytes read=323108791
		FILE: Number of bytes written=445152611
		FILE: Number of read operations=0
		FILE: Number of large read operations=0
		FILE: Number of write operations=0
	Map-Reduce Framework
		Map input records=2941054
		Map output records=7009865
		Map output bytes=60123367
		Map output materialized bytes=74143103
		Input split bytes=98
		Combine input records=0
		Spilled Records=14019730
		Failed Shuffles=0
		Merged Map outputs=0
		GC time elapsed (ms)=77
		Total committed heap usage (bytes)=1313865728
	File Input Format Counters 
		Bytes Read=33558528
21/04/26 21:26:56 INFO mapred.LocalJobRunner: Finishing task: attempt_local588307001_0001_m_000002_0
21/04/26 21:26:56 INFO mapred.LocalJobRunner: Starting task: attempt_local588307001_0001_m_000003_0
21/04/26 21:26:56 INFO output.FileOutputCommitter: File Output Committer Algorithm version is 1
21/04/26 21:26:56 INFO util.ProcfsBasedProcessTree: ProcfsBasedProcessTree currently is supported only on Linux.
21/04/26 21:26:56 INFO mapred.Task:  Using ResourceCalculatorProcessTree : org.apache.hadoop.yarn.util.WindowsBasedProcessTree@be049c8
21/04/26 21:26:56 INFO mapred.MapTask: Processing split: file:/E:/bigdata/2021.4.26/wc.txt:100663296+33554432
21/04/26 21:26:56 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)
21/04/26 21:26:56 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
21/04/26 21:26:56 INFO mapred.MapTask: soft limit at 83886080
21/04/26 21:26:56 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600
21/04/26 21:26:56 INFO mapred.MapTask: kvstart = 26214396; length = 6553600
21/04/26 21:26:56 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
21/04/26 21:26:57 INFO mapreduce.Job:  map 100% reduce 0%
21/04/26 21:26:58 INFO mapred.MapTask: Spilling map output
21/04/26 21:26:58 INFO mapred.MapTask: bufstart = 0; bufend = 29274885; bufvoid = 104857600
21/04/26 21:26:58 INFO mapred.MapTask: kvstart = 26214396(104857584); kvend = 12561600(50246400); length = 13652797/6553600
21/04/26 21:26:58 INFO mapred.MapTask: (EQUATOR) 39760634 kvi 9940152(39760608)
21/04/26 21:27:00 INFO mapred.MapTask: Finished spill 0
21/04/26 21:27:00 INFO mapred.MapTask: (RESET) equator 39760634 kv 9940152(39760608) kvi 7318728(29274912)
21/04/26 21:27:02 INFO mapred.MapTask: Spilling map output
21/04/26 21:27:02 INFO mapred.MapTask: bufstart = 39760634; bufend = 69035520; bufvoid = 104857600
21/04/26 21:27:02 INFO mapred.MapTask: kvstart = 9940152(39760608); kvend = 22501760(90007040); length = 13652793/6553600
21/04/26 21:27:02 INFO mapred.MapTask: (EQUATOR) 79521272 kvi 19880312(79521248)
21/04/26 21:27:02 INFO mapred.LocalJobRunner: 
21/04/26 21:27:02 INFO mapred.MapTask: Starting flush of map output
21/04/26 21:27:02 INFO mapreduce.Job:  map 25% reduce 0%
21/04/26 21:27:02 INFO mapred.LocalJobRunner: map > sort
21/04/26 21:27:03 INFO mapreduce.Job:  map 31% reduce 0%
21/04/26 21:27:03 INFO mapred.MapTask: Finished spill 1
21/04/26 21:27:03 INFO mapred.MapTask: (RESET) equator 79521272 kv 19880312(79521248) kvi 19146444(76585776)
21/04/26 21:27:03 INFO mapred.MapTask: Spilling map output
21/04/26 21:27:03 INFO mapred.MapTask: bufstart = 79521272; bufend = 81094865; bufvoid = 104857600
21/04/26 21:27:03 INFO mapred.MapTask: kvstart = 19880312(79521248); kvend = 19146448(76585792); length = 733865/6553600
21/04/26 21:27:03 INFO mapred.MapTask: Finished spill 2
21/04/26 21:27:03 INFO mapred.Merger: Merging 3 sorted segments
21/04/26 21:27:03 INFO mapred.Merger: Down to the last merge-pass, with 3 segments left of total size: 74143105 bytes
21/04/26 21:27:05 INFO mapred.LocalJobRunner: map > sort > 
21/04/26 21:27:06 INFO mapreduce.Job:  map 32% reduce 0%
21/04/26 21:27:08 INFO mapred.Task: Task:attempt_local588307001_0001_m_000003_0 is done. And is in the process of committing
21/04/26 21:27:08 INFO mapred.LocalJobRunner: map > sort
21/04/26 21:27:08 INFO mapred.Task: Task 'attempt_local588307001_0001_m_000003_0' done.
21/04/26 21:27:08 INFO mapred.Task: Final Counters for attempt_local588307001_0001_m_000003_0: Counters: 17
	File System Counters
		FILE: Number of bytes read=430811636
		FILE: Number of bytes written=593438859
		FILE: Number of read operations=0
		FILE: Number of large read operations=0
		FILE: Number of write operations=0
	Map-Reduce Framework
		Map input records=2941058
		Map output records=7009866
		Map output bytes=60123364
		Map output materialized bytes=74143102
		Input split bytes=98
		Combine input records=0
		Spilled Records=14019732
		Failed Shuffles=0
		Merged Map outputs=0
		GC time elapsed (ms)=67
		Total committed heap usage (bytes)=1136132096
	File Input Format Counters 
		Bytes Read=33558528
21/04/26 21:27:08 INFO mapred.LocalJobRunner: Finishing task: attempt_local588307001_0001_m_000003_0
21/04/26 21:27:08 INFO mapred.LocalJobRunner: Starting task: attempt_local588307001_0001_m_000004_0
21/04/26 21:27:08 INFO output.FileOutputCommitter: File Output Committer Algorithm version is 1
21/04/26 21:27:08 INFO util.ProcfsBasedProcessTree: ProcfsBasedProcessTree currently is supported only on Linux.
21/04/26 21:27:08 INFO mapred.Task:  Using ResourceCalculatorProcessTree : org.apache.hadoop.yarn.util.WindowsBasedProcessTree@6918296b
21/04/26 21:27:08 INFO mapred.MapTask: Processing split: file:/E:/bigdata/2021.4.26/wc.txt:134217728+33554432
21/04/26 21:27:08 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)
21/04/26 21:27:08 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
21/04/26 21:27:08 INFO mapred.MapTask: soft limit at 83886080
21/04/26 21:27:08 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600
21/04/26 21:27:08 INFO mapred.MapTask: kvstart = 26214396; length = 6553600
21/04/26 21:27:08 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
21/04/26 21:27:08 INFO mapreduce.Job:  map 100% reduce 0%
21/04/26 21:27:10 INFO mapred.MapTask: Spilling map output
21/04/26 21:27:10 INFO mapred.MapTask: bufstart = 0; bufend = 29274888; bufvoid = 104857600
21/04/26 21:27:10 INFO mapred.MapTask: kvstart = 26214396(104857584); kvend = 12561600(50246400); length = 13652797/6553600
21/04/26 21:27:10 INFO mapred.MapTask: (EQUATOR) 39760636 kvi 9940152(39760608)
21/04/26 21:27:11 INFO mapred.MapTask: Finished spill 0
21/04/26 21:27:11 INFO mapred.MapTask: (RESET) equator 39760636 kv 9940152(39760608) kvi 7318728(29274912)
21/04/26 21:27:13 INFO mapred.MapTask: Spilling map output
21/04/26 21:27:13 INFO mapred.MapTask: bufstart = 39760636; bufend = 69035514; bufvoid = 104857600
21/04/26 21:27:13 INFO mapred.MapTask: kvstart = 9940152(39760608); kvend = 22501760(90007040); length = 13652793/6553600
21/04/26 21:27:13 INFO mapred.MapTask: (EQUATOR) 79521269 kvi 19880312(79521248)
21/04/26 21:27:13 INFO mapred.LocalJobRunner: 
21/04/26 21:27:13 INFO mapred.MapTask: Starting flush of map output
21/04/26 21:27:13 INFO mapreduce.Job:  map 33% reduce 0%
21/04/26 21:27:14 INFO mapred.LocalJobRunner: map > sort
21/04/26 21:27:14 INFO mapreduce.Job:  map 39% reduce 0%
21/04/26 21:27:14 INFO mapred.MapTask: Finished spill 1
21/04/26 21:27:14 INFO mapred.MapTask: (RESET) equator 79521269 kv 19880312(79521248) kvi 19146428(76585712)
21/04/26 21:27:14 INFO mapred.MapTask: Spilling map output
21/04/26 21:27:14 INFO mapred.MapTask: bufstart = 79521269; bufend = 81094906; bufvoid = 104857600
21/04/26 21:27:14 INFO mapred.MapTask: kvstart = 19880312(79521248); kvend = 19146432(76585728); length = 733881/6553600
21/04/26 21:27:14 INFO mapred.MapTask: Finished spill 2
21/04/26 21:27:14 INFO mapred.Merger: Merging 3 sorted segments
21/04/26 21:27:14 INFO mapred.Merger: Down to the last merge-pass, with 3 segments left of total size: 74143152 bytes
21/04/26 21:27:17 INFO mapred.LocalJobRunner: map > sort > 
21/04/26 21:27:17 INFO mapreduce.Job:  map 41% reduce 0%
21/04/26 21:27:17 INFO mapred.Task: Task:attempt_local588307001_0001_m_000004_0 is done. And is in the process of committing
21/04/26 21:27:17 INFO mapred.LocalJobRunner: map > sort
21/04/26 21:27:17 INFO mapred.Task: Task 'attempt_local588307001_0001_m_000004_0' done.
21/04/26 21:27:17 INFO mapred.Task: Final Counters for attempt_local588307001_0001_m_000004_0: Counters: 17
	File System Counters
		FILE: Number of bytes read=538514528
		FILE: Number of bytes written=741725201
		FILE: Number of read operations=0
		FILE: Number of large read operations=0
		FILE: Number of write operations=0
	Map-Reduce Framework
		Map input records=2941058
		Map output records=7009870
		Map output bytes=60123403
		Map output materialized bytes=74143149
		Input split bytes=98
		Combine input records=0
		Spilled Records=14019740
		Failed Shuffles=0
		Merged Map outputs=0
		GC time elapsed (ms)=67
		Total committed heap usage (bytes)=953155584
	File Input Format Counters 
		Bytes Read=33558528
21/04/26 21:27:17 INFO mapred.LocalJobRunner: Finishing task: attempt_local588307001_0001_m_000004_0
21/04/26 21:27:17 INFO mapred.LocalJobRunner: Starting task: attempt_local588307001_0001_m_000005_0
21/04/26 21:27:17 INFO output.FileOutputCommitter: File Output Committer Algorithm version is 1
21/04/26 21:27:17 INFO util.ProcfsBasedProcessTree: ProcfsBasedProcessTree currently is supported only on Linux.
21/04/26 21:27:17 INFO mapred.Task:  Using ResourceCalculatorProcessTree : org.apache.hadoop.yarn.util.WindowsBasedProcessTree@4b678c5a
21/04/26 21:27:17 INFO mapred.MapTask: Processing split: file:/E:/bigdata/2021.4.26/wc.txt:167772160+33554432
21/04/26 21:27:17 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)
21/04/26 21:27:17 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
21/04/26 21:27:17 INFO mapred.MapTask: soft limit at 83886080
21/04/26 21:27:17 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600
21/04/26 21:27:17 INFO mapred.MapTask: kvstart = 26214396; length = 6553600
21/04/26 21:27:17 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
21/04/26 21:27:18 INFO mapreduce.Job:  map 100% reduce 0%
21/04/26 21:27:19 INFO mapred.MapTask: Spilling map output
21/04/26 21:27:19 INFO mapred.MapTask: bufstart = 0; bufend = 29274881; bufvoid = 104857600
21/04/26 21:27:19 INFO mapred.MapTask: kvstart = 26214396(104857584); kvend = 12561604(50246416); length = 13652793/6553600
21/04/26 21:27:19 INFO mapred.MapTask: (EQUATOR) 39760640 kvi 9940156(39760624)
21/04/26 21:27:21 INFO mapred.MapTask: Finished spill 0
21/04/26 21:27:21 INFO mapred.MapTask: (RESET) equator 39760640 kv 9940156(39760624) kvi 7318728(29274912)
21/04/26 21:27:22 INFO mapred.MapTask: Spilling map output
21/04/26 21:27:22 INFO mapred.MapTask: bufstart = 39760640; bufend = 69035526; bufvoid = 104857600
21/04/26 21:27:22 INFO mapred.MapTask: kvstart = 9940156(39760624); kvend = 22501764(90007056); length = 13652793/6553600
21/04/26 21:27:22 INFO mapred.MapTask: (EQUATOR) 79521283 kvi 19880316(79521264)
21/04/26 21:27:22 INFO mapred.LocalJobRunner: 
21/04/26 21:27:22 INFO mapred.MapTask: Starting flush of map output
21/04/26 21:27:23 INFO mapreduce.Job:  map 42% reduce 0%
21/04/26 21:27:23 INFO mapred.LocalJobRunner: map > sort
21/04/26 21:27:24 INFO mapred.MapTask: Finished spill 1
21/04/26 21:27:24 INFO mapred.MapTask: (RESET) equator 79521283 kv 19880316(79521264) kvi 19146452(76585808)
21/04/26 21:27:24 INFO mapred.MapTask: Spilling map output
21/04/26 21:27:24 INFO mapred.MapTask: bufstart = 79521283; bufend = 81094858; bufvoid = 104857600
21/04/26 21:27:24 INFO mapred.MapTask: kvstart = 19880316(79521264); kvend = 19146456(76585824); length = 733861/6553600
21/04/26 21:27:24 INFO mapred.MapTask: Finished spill 2
21/04/26 21:27:24 INFO mapred.Merger: Merging 3 sorted segments
21/04/26 21:27:24 INFO mapred.Merger: Down to the last merge-pass, with 3 segments left of total size: 74143079 bytes
21/04/26 21:27:24 INFO mapreduce.Job:  map 47% reduce 0%
21/04/26 21:27:26 INFO mapred.LocalJobRunner: map > sort > 
21/04/26 21:27:27 INFO mapred.Task: Task:attempt_local588307001_0001_m_000005_0 is done. And is in the process of committing
21/04/26 21:27:27 INFO mapred.LocalJobRunner: map > sort
21/04/26 21:27:27 INFO mapred.Task: Task 'attempt_local588307001_0001_m_000005_0' done.
21/04/26 21:27:27 INFO mapred.Task: Final Counters for attempt_local588307001_0001_m_000005_0: Counters: 17
	File System Counters
		FILE: Number of bytes read=646217347
		FILE: Number of bytes written=890011397
		FILE: Number of read operations=0
		FILE: Number of large read operations=0
		FILE: Number of write operations=0
	Map-Reduce Framework
		Map input records=2941056
		Map output records=7009864
		Map output bytes=60123342
		Map output materialized bytes=74143076
		Input split bytes=98
		Combine input records=0
		Spilled Records=14019728
		Failed Shuffles=0
		Merged Map outputs=0
		GC time elapsed (ms)=75
		Total committed heap usage (bytes)=825753600
	File Input Format Counters 
		Bytes Read=33558528
21/04/26 21:27:27 INFO mapred.LocalJobRunner: Finishing task: attempt_local588307001_0001_m_000005_0
21/04/26 21:27:27 INFO mapred.LocalJobRunner: Starting task: attempt_local588307001_0001_m_000006_0
21/04/26 21:27:27 INFO output.FileOutputCommitter: File Output Committer Algorithm version is 1
21/04/26 21:27:27 INFO util.ProcfsBasedProcessTree: ProcfsBasedProcessTree currently is supported only on Linux.
21/04/26 21:27:27 INFO mapred.Task:  Using ResourceCalculatorProcessTree : org.apache.hadoop.yarn.util.WindowsBasedProcessTree@33ada1f
21/04/26 21:27:27 INFO mapred.MapTask: Processing split: file:/E:/bigdata/2021.4.26/wc.txt:201326592+33554432
21/04/26 21:27:27 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)
21/04/26 21:27:27 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
21/04/26 21:27:27 INFO mapred.MapTask: soft limit at 83886080
21/04/26 21:27:27 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600
21/04/26 21:27:27 INFO mapred.MapTask: kvstart = 26214396; length = 6553600
21/04/26 21:27:27 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
21/04/26 21:27:27 INFO mapreduce.Job:  map 100% reduce 0%
21/04/26 21:27:29 INFO mapred.MapTask: Spilling map output
21/04/26 21:27:29 INFO mapred.MapTask: bufstart = 0; bufend = 29274896; bufvoid = 104857600
21/04/26 21:27:29 INFO mapred.MapTask: kvstart = 26214396(104857584); kvend = 12561604(50246416); length = 13652793/6553600
21/04/26 21:27:29 INFO mapred.MapTask: (EQUATOR) 39760648 kvi 9940156(39760624)
21/04/26 21:27:30 INFO mapred.MapTask: Finished spill 0
21/04/26 21:27:30 INFO mapred.MapTask: (RESET) equator 39760648 kv 9940156(39760624) kvi 7318728(29274912)
21/04/26 21:27:32 INFO mapred.MapTask: Spilling map output
21/04/26 21:27:32 INFO mapred.MapTask: bufstart = 39760648; bufend = 69035532; bufvoid = 104857600
21/04/26 21:27:32 INFO mapred.MapTask: kvstart = 9940156(39760624); kvend = 22501764(90007056); length = 13652793/6553600
21/04/26 21:27:32 INFO mapred.MapTask: (EQUATOR) 79521286 kvi 19880316(79521264)
21/04/26 21:27:32 INFO mapred.LocalJobRunner: 
21/04/26 21:27:32 INFO mapred.MapTask: Starting flush of map output
21/04/26 21:27:32 INFO mapreduce.Job:  map 50% reduce 0%
21/04/26 21:27:33 INFO mapred.LocalJobRunner: map > sort
21/04/26 21:27:33 INFO mapred.MapTask: Finished spill 1
21/04/26 21:27:33 INFO mapred.MapTask: (RESET) equator 79521286 kv 19880316(79521264) kvi 19146440(76585760)
21/04/26 21:27:33 INFO mapred.MapTask: Spilling map output
21/04/26 21:27:33 INFO mapred.MapTask: bufstart = 79521286; bufend = 81094880; bufvoid = 104857600
21/04/26 21:27:33 INFO mapred.MapTask: kvstart = 19880316(79521264); kvend = 19146444(76585776); length = 733873/6553600
21/04/26 21:27:33 INFO mapred.MapTask: Finished spill 2
21/04/26 21:27:33 INFO mapred.Merger: Merging 3 sorted segments
21/04/26 21:27:33 INFO mapred.Merger: Down to the last merge-pass, with 3 segments left of total size: 74143117 bytes
21/04/26 21:27:33 INFO mapreduce.Job:  map 56% reduce 0%
21/04/26 21:27:36 INFO mapred.LocalJobRunner: map > sort > 
21/04/26 21:27:36 INFO mapreduce.Job:  map 58% reduce 0%
21/04/26 21:27:37 INFO mapred.Task: Task:attempt_local588307001_0001_m_000006_0 is done. And is in the process of committing
21/04/26 21:27:37 INFO mapred.LocalJobRunner: map > sort
21/04/26 21:27:37 INFO mapred.Task: Task 'attempt_local588307001_0001_m_000006_0' done.
21/04/26 21:27:37 INFO mapred.Task: Final Counters for attempt_local588307001_0001_m_000006_0: Counters: 17
	File System Counters
		FILE: Number of bytes read=753919692
		FILE: Number of bytes written=1038297669
		FILE: Number of read operations=0
		FILE: Number of large read operations=0
		FILE: Number of write operations=0
	Map-Reduce Framework
		Map input records=2941058
		Map output records=7009867
		Map output bytes=60123374
		Map output materialized bytes=74143114
		Input split bytes=98
		Combine input records=0
		Spilled Records=14019734
		Failed Shuffles=0
		Merged Map outputs=0
		GC time elapsed (ms)=71
		Total committed heap usage (bytes)=791150592
	File Input Format Counters 
		Bytes Read=33558528
21/04/26 21:27:37 INFO mapred.LocalJobRunner: Finishing task: attempt_local588307001_0001_m_000006_0
21/04/26 21:27:37 INFO mapred.LocalJobRunner: Starting task: attempt_local588307001_0001_m_000007_0
21/04/26 21:27:37 INFO output.FileOutputCommitter: File Output Committer Algorithm version is 1
21/04/26 21:27:37 INFO util.ProcfsBasedProcessTree: ProcfsBasedProcessTree currently is supported only on Linux.
21/04/26 21:27:37 INFO mapred.Task:  Using ResourceCalculatorProcessTree : org.apache.hadoop.yarn.util.WindowsBasedProcessTree@a1d7db9
21/04/26 21:27:37 INFO mapred.MapTask: Processing split: file:/E:/bigdata/2021.4.26/wc.txt:234881024+33554432
21/04/26 21:27:37 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)
21/04/26 21:27:37 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
21/04/26 21:27:37 INFO mapred.MapTask: soft limit at 83886080
21/04/26 21:27:37 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600
21/04/26 21:27:37 INFO mapred.MapTask: kvstart = 26214396; length = 6553600
21/04/26 21:27:37 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
21/04/26 21:27:37 INFO mapreduce.Job:  map 100% reduce 0%
21/04/26 21:27:39 INFO mapred.MapTask: Spilling map output
21/04/26 21:27:39 INFO mapred.MapTask: bufstart = 0; bufend = 29274882; bufvoid = 104857600
21/04/26 21:27:39 INFO mapred.MapTask: kvstart = 26214396(104857584); kvend = 12561604(50246416); length = 13652793/6553600
21/04/26 21:27:39 INFO mapred.MapTask: (EQUATOR) 39760641 kvi 9940156(39760624)
21/04/26 21:27:40 INFO mapred.MapTask: Finished spill 0
21/04/26 21:27:40 INFO mapred.MapTask: (RESET) equator 39760641 kv 9940156(39760624) kvi 7318728(29274912)
21/04/26 21:27:42 INFO mapred.MapTask: Spilling map output
21/04/26 21:27:42 INFO mapred.MapTask: bufstart = 39760641; bufend = 69035537; bufvoid = 104857600
21/04/26 21:27:42 INFO mapred.MapTask: kvstart = 9940156(39760624); kvend = 22501764(90007056); length = 13652793/6553600
21/04/26 21:27:42 INFO mapred.MapTask: (EQUATOR) 79521288 kvi 19880316(79521264)
21/04/26 21:27:42 INFO mapred.LocalJobRunner: 
21/04/26 21:27:42 INFO mapred.MapTask: Starting flush of map output
21/04/26 21:27:42 INFO mapreduce.Job:  map 58% reduce 0%
21/04/26 21:27:43 INFO mapred.LocalJobRunner: map > sort
21/04/26 21:27:43 INFO mapred.MapTask: Finished spill 1
21/04/26 21:27:43 INFO mapred.MapTask: (RESET) equator 79521288 kv 19880316(79521264) kvi 19146444(76585776)
21/04/26 21:27:43 INFO mapred.MapTask: Spilling map output
21/04/26 21:27:43 INFO mapred.MapTask: bufstart = 79521288; bufend = 81094880; bufvoid = 104857600
21/04/26 21:27:43 INFO mapred.MapTask: kvstart = 19880316(79521264); kvend = 19146448(76585792); length = 733869/6553600
21/04/26 21:27:43 INFO mapred.MapTask: Finished spill 2
21/04/26 21:27:43 INFO mapred.Merger: Merging 3 sorted segments
21/04/26 21:27:43 INFO mapred.Merger: Down to the last merge-pass, with 3 segments left of total size: 74143111 bytes
21/04/26 21:27:43 INFO mapreduce.Job:  map 64% reduce 0%
21/04/26 21:27:46 INFO mapred.LocalJobRunner: map > sort > 
21/04/26 21:27:46 INFO mapreduce.Job:  map 66% reduce 0%
21/04/26 21:27:46 INFO mapred.Task: Task:attempt_local588307001_0001_m_000007_0 is done. And is in the process of committing
21/04/26 21:27:46 INFO mapred.LocalJobRunner: map > sort
21/04/26 21:27:46 INFO mapred.Task: Task 'attempt_local588307001_0001_m_000007_0' done.
21/04/26 21:27:46 INFO mapred.Task: Final Counters for attempt_local588307001_0001_m_000007_0: Counters: 17
	File System Counters
		FILE: Number of bytes read=861622031
		FILE: Number of bytes written=1186583929
		FILE: Number of read operations=0
		FILE: Number of large read operations=0
		FILE: Number of write operations=0
	Map-Reduce Framework
		Map input records=2941058
		Map output records=7009866
		Map output bytes=60123370
		Map output materialized bytes=74143108
		Input split bytes=98
		Combine input records=0
		Spilled Records=14019732
		Failed Shuffles=0
		Merged Map outputs=0
		GC time elapsed (ms)=68
		Total committed heap usage (bytes)=1449656320
	File Input Format Counters 
		Bytes Read=33558528
21/04/26 21:27:46 INFO mapred.LocalJobRunner: Finishing task: attempt_local588307001_0001_m_000007_0
21/04/26 21:27:46 INFO mapred.LocalJobRunner: Starting task: attempt_local588307001_0001_m_000008_0
21/04/26 21:27:46 INFO output.FileOutputCommitter: File Output Committer Algorithm version is 1
21/04/26 21:27:46 INFO util.ProcfsBasedProcessTree: ProcfsBasedProcessTree currently is supported only on Linux.
21/04/26 21:27:46 INFO mapred.Task:  Using ResourceCalculatorProcessTree : org.apache.hadoop.yarn.util.WindowsBasedProcessTree@22664ee0
21/04/26 21:27:46 INFO mapred.MapTask: Processing split: file:/E:/bigdata/2021.4.26/wc.txt:268435456+33554432
21/04/26 21:27:46 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)
21/04/26 21:27:46 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
21/04/26 21:27:46 INFO mapred.MapTask: soft limit at 83886080
21/04/26 21:27:46 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600
21/04/26 21:27:46 INFO mapred.MapTask: kvstart = 26214396; length = 6553600
21/04/26 21:27:46 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
21/04/26 21:27:47 INFO mapreduce.Job:  map 100% reduce 0%
21/04/26 21:27:48 INFO mapred.MapTask: Spilling map output
21/04/26 21:27:48 INFO mapred.MapTask: bufstart = 0; bufend = 29274881; bufvoid = 104857600
21/04/26 21:27:48 INFO mapred.MapTask: kvstart = 26214396(104857584); kvend = 12561604(50246416); length = 13652793/6553600
21/04/26 21:27:48 INFO mapred.MapTask: (EQUATOR) 39760640 kvi 9940156(39760624)
21/04/26 21:27:50 INFO mapred.MapTask: Finished spill 0
21/04/26 21:27:50 INFO mapred.MapTask: (RESET) equator 39760640 kv 9940156(39760624) kvi 7318728(29274912)
21/04/26 21:27:51 INFO mapred.MapTask: Spilling map output
21/04/26 21:27:51 INFO mapred.MapTask: bufstart = 39760640; bufend = 69035523; bufvoid = 104857600
21/04/26 21:27:51 INFO mapred.MapTask: kvstart = 9940156(39760624); kvend = 22501764(90007056); length = 13652793/6553600
21/04/26 21:27:51 INFO mapred.MapTask: (EQUATOR) 79521281 kvi 19880316(79521264)
21/04/26 21:27:51 INFO mapred.LocalJobRunner: 
21/04/26 21:27:51 INFO mapred.MapTask: Starting flush of map output
21/04/26 21:27:52 INFO mapreduce.Job:  map 67% reduce 0%
21/04/26 21:27:52 INFO mapred.LocalJobRunner: map > sort
21/04/26 21:27:53 INFO mapred.MapTask: Finished spill 1
21/04/26 21:27:53 INFO mapred.MapTask: (RESET) equator 79521281 kv 19880316(79521264) kvi 19146440(76585760)
21/04/26 21:27:53 INFO mapred.MapTask: Spilling map output
21/04/26 21:27:53 INFO mapred.MapTask: bufstart = 79521281; bufend = 81094894; bufvoid = 104857600
21/04/26 21:27:53 INFO mapred.MapTask: kvstart = 19880316(79521264); kvend = 19146444(76585776); length = 733873/6553600
21/04/26 21:27:53 INFO mapred.MapTask: Finished spill 2
21/04/26 21:27:53 INFO mapred.Merger: Merging 3 sorted segments
21/04/26 21:27:53 INFO mapred.Merger: Down to the last merge-pass, with 3 segments left of total size: 74143120 bytes
21/04/26 21:27:53 INFO mapreduce.Job:  map 72% reduce 0%
21/04/26 21:27:55 INFO mapred.LocalJobRunner: map > sort > 
21/04/26 21:27:56 INFO mapreduce.Job:  map 74% reduce 0%
21/04/26 21:27:56 INFO mapred.Task: Task:attempt_local588307001_0001_m_000008_0 is done. And is in the process of committing
21/04/26 21:27:56 INFO mapred.LocalJobRunner: map > sort
21/04/26 21:27:56 INFO mapred.Task: Task 'attempt_local588307001_0001_m_000008_0' done.
21/04/26 21:27:56 INFO mapred.Task: Final Counters for attempt_local588307001_0001_m_000008_0: Counters: 17
	File System Counters
		FILE: Number of bytes read=969324379
		FILE: Number of bytes written=1334870207
		FILE: Number of read operations=0
		FILE: Number of large read operations=0
		FILE: Number of write operations=0
	Map-Reduce Framework
		Map input records=2941056
		Map output records=7009867
		Map output bytes=60123377
		Map output materialized bytes=74143117
		Input split bytes=98
		Combine input records=0
		Spilled Records=14019734
		Failed Shuffles=0
		Merged Map outputs=0
		GC time elapsed (ms)=57
		Total committed heap usage (bytes)=1227358208
	File Input Format Counters 
		Bytes Read=33558528
21/04/26 21:27:56 INFO mapred.LocalJobRunner: Finishing task: attempt_local588307001_0001_m_000008_0
21/04/26 21:27:56 INFO mapred.LocalJobRunner: Starting task: attempt_local588307001_0001_m_000009_0
21/04/26 21:27:56 INFO output.FileOutputCommitter: File Output Committer Algorithm version is 1
21/04/26 21:27:56 INFO util.ProcfsBasedProcessTree: ProcfsBasedProcessTree currently is supported only on Linux.
21/04/26 21:27:56 INFO mapred.Task:  Using ResourceCalculatorProcessTree : org.apache.hadoop.yarn.util.WindowsBasedProcessTree@559403aa
21/04/26 21:27:56 INFO mapred.MapTask: Processing split: file:/E:/bigdata/2021.4.26/wc.txt:301989888+33554432
21/04/26 21:27:56 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)
21/04/26 21:27:56 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
21/04/26 21:27:56 INFO mapred.MapTask: soft limit at 83886080
21/04/26 21:27:56 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600
21/04/26 21:27:56 INFO mapred.MapTask: kvstart = 26214396; length = 6553600
21/04/26 21:27:56 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
21/04/26 21:27:57 INFO mapreduce.Job:  map 100% reduce 0%
21/04/26 21:27:58 INFO mapred.MapTask: Spilling map output
21/04/26 21:27:58 INFO mapred.MapTask: bufstart = 0; bufend = 29274888; bufvoid = 104857600
21/04/26 21:27:58 INFO mapred.MapTask: kvstart = 26214396(104857584); kvend = 12561600(50246400); length = 13652797/6553600
21/04/26 21:27:58 INFO mapred.MapTask: (EQUATOR) 39760636 kvi 9940152(39760608)
21/04/26 21:28:00 INFO mapred.MapTask: Finished spill 0
21/04/26 21:28:00 INFO mapred.MapTask: (RESET) equator 39760636 kv 9940152(39760608) kvi 7318728(29274912)
21/04/26 21:28:02 INFO mapred.MapTask: Spilling map output
21/04/26 21:28:02 INFO mapred.MapTask: bufstart = 39760636; bufend = 69035517; bufvoid = 104857600
21/04/26 21:28:02 INFO mapred.MapTask: kvstart = 9940152(39760608); kvend = 22501760(90007040); length = 13652793/6553600
21/04/26 21:28:02 INFO mapred.MapTask: (EQUATOR) 79521270 kvi 19880312(79521248)
21/04/26 21:28:02 INFO mapred.LocalJobRunner: 
21/04/26 21:28:02 INFO mapred.MapTask: Starting flush of map output
21/04/26 21:28:02 INFO mapreduce.Job:  map 75% reduce 0%
21/04/26 21:28:02 INFO mapred.LocalJobRunner: map > sort
21/04/26 21:28:03 INFO mapreduce.Job:  map 81% reduce 0%
21/04/26 21:28:03 INFO mapred.MapTask: Finished spill 1
21/04/26 21:28:03 INFO mapred.MapTask: (RESET) equator 79521270 kv 19880312(79521248) kvi 19146444(76585776)
21/04/26 21:28:03 INFO mapred.MapTask: Spilling map output
21/04/26 21:28:03 INFO mapred.MapTask: bufstart = 79521270; bufend = 81094872; bufvoid = 104857600
21/04/26 21:28:03 INFO mapred.MapTask: kvstart = 19880312(79521248); kvend = 19146448(76585792); length = 733865/6553600
21/04/26 21:28:03 INFO mapred.MapTask: Finished spill 2
21/04/26 21:28:03 INFO mapred.Merger: Merging 3 sorted segments
21/04/26 21:28:03 INFO mapred.Merger: Down to the last merge-pass, with 3 segments left of total size: 74143112 bytes
21/04/26 21:28:05 INFO mapred.LocalJobRunner: map > sort > 
21/04/26 21:28:06 INFO mapreduce.Job:  map 82% reduce 0%
21/04/26 21:28:07 INFO mapred.Task: Task:attempt_local588307001_0001_m_000009_0 is done. And is in the process of committing
21/04/26 21:28:07 INFO mapred.LocalJobRunner: map > sort
21/04/26 21:28:07 INFO mapred.Task: Task 'attempt_local588307001_0001_m_000009_0' done.
21/04/26 21:28:07 INFO mapred.Task: Final Counters for attempt_local588307001_0001_m_000009_0: Counters: 17
	File System Counters
		FILE: Number of bytes read=1077026719
		FILE: Number of bytes written=1483156469
		FILE: Number of read operations=0
		FILE: Number of large read operations=0
		FILE: Number of write operations=0
	Map-Reduce Framework
		Map input records=2941054
		Map output records=7009866
		Map output bytes=60123371
		Map output materialized bytes=74143109
		Input split bytes=98
		Combine input records=0
		Spilled Records=14019732
		Failed Shuffles=0
		Merged Map outputs=0
		GC time elapsed (ms)=61
		Total committed heap usage (bytes)=1042808832
	File Input Format Counters 
		Bytes Read=33558528
21/04/26 21:28:07 INFO mapred.LocalJobRunner: Finishing task: attempt_local588307001_0001_m_000009_0
21/04/26 21:28:07 INFO mapred.LocalJobRunner: Starting task: attempt_local588307001_0001_m_000010_0
21/04/26 21:28:07 INFO output.FileOutputCommitter: File Output Committer Algorithm version is 1
21/04/26 21:28:07 INFO util.ProcfsBasedProcessTree: ProcfsBasedProcessTree currently is supported only on Linux.
21/04/26 21:28:07 INFO mapreduce.Job:  map 100% reduce 0%
21/04/26 21:28:07 INFO mapred.Task:  Using ResourceCalculatorProcessTree : org.apache.hadoop.yarn.util.WindowsBasedProcessTree@abc125
21/04/26 21:28:07 INFO mapred.MapTask: Processing split: file:/E:/bigdata/2021.4.26/wc.txt:335544320+33554432
21/04/26 21:28:07 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)
21/04/26 21:28:07 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
21/04/26 21:28:07 INFO mapred.MapTask: soft limit at 83886080
21/04/26 21:28:07 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600
21/04/26 21:28:07 INFO mapred.MapTask: kvstart = 26214396; length = 6553600
21/04/26 21:28:07 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
21/04/26 21:28:09 INFO mapred.MapTask: Spilling map output
21/04/26 21:28:09 INFO mapred.MapTask: bufstart = 0; bufend = 29274881; bufvoid = 104857600
21/04/26 21:28:09 INFO mapred.MapTask: kvstart = 26214396(104857584); kvend = 12561604(50246416); length = 13652793/6553600
21/04/26 21:28:09 INFO mapred.MapTask: (EQUATOR) 39760640 kvi 9940156(39760624)
21/04/26 21:28:11 INFO mapred.MapTask: Finished spill 0
21/04/26 21:28:11 INFO mapred.MapTask: (RESET) equator 39760640 kv 9940156(39760624) kvi 7318728(29274912)
21/04/26 21:28:12 INFO mapred.MapTask: Spilling map output
21/04/26 21:28:12 INFO mapred.MapTask: bufstart = 39760640; bufend = 69035522; bufvoid = 104857600
21/04/26 21:28:12 INFO mapred.MapTask: kvstart = 9940156(39760624); kvend = 22501764(90007056); length = 13652793/6553600
21/04/26 21:28:12 INFO mapred.MapTask: (EQUATOR) 79521281 kvi 19880316(79521264)
21/04/26 21:28:13 INFO mapred.LocalJobRunner: 
21/04/26 21:28:13 INFO mapred.MapTask: Starting flush of map output
21/04/26 21:28:13 INFO mapred.LocalJobRunner: map > sort
21/04/26 21:28:13 INFO mapreduce.Job:  map 89% reduce 0%
21/04/26 21:28:14 INFO mapred.MapTask: Finished spill 1
21/04/26 21:28:14 INFO mapred.MapTask: (RESET) equator 79521281 kv 19880316(79521264) kvi 19146444(76585776)
21/04/26 21:28:14 INFO mapred.MapTask: Spilling map output
21/04/26 21:28:14 INFO mapred.MapTask: bufstart = 79521281; bufend = 81094888; bufvoid = 104857600
21/04/26 21:28:14 INFO mapred.MapTask: kvstart = 19880316(79521264); kvend = 19146448(76585792); length = 733869/6553600
21/04/26 21:28:14 INFO mapred.MapTask: Finished spill 2
21/04/26 21:28:14 INFO mapred.Merger: Merging 3 sorted segments
21/04/26 21:28:14 INFO mapred.Merger: Down to the last merge-pass, with 3 segments left of total size: 74143111 bytes
21/04/26 21:28:16 INFO mapred.LocalJobRunner: map > sort > 
21/04/26 21:28:16 INFO mapreduce.Job:  map 90% reduce 0%
21/04/26 21:28:18 INFO mapred.Task: Task:attempt_local588307001_0001_m_000010_0 is done. And is in the process of committing
21/04/26 21:28:18 INFO mapred.LocalJobRunner: map > sort
21/04/26 21:28:18 INFO mapred.Task: Task 'attempt_local588307001_0001_m_000010_0' done.
21/04/26 21:28:18 INFO mapred.Task: Final Counters for attempt_local588307001_0001_m_000010_0: Counters: 17
	File System Counters
		FILE: Number of bytes read=1184729058
		FILE: Number of bytes written=1631442729
		FILE: Number of read operations=0
		FILE: Number of large read operations=0
		FILE: Number of write operations=0
	Map-Reduce Framework
		Map input records=2941058
		Map output records=7009866
		Map output bytes=60123370
		Map output materialized bytes=74143108
		Input split bytes=98
		Combine input records=0
		Spilled Records=14019732
		Failed Shuffles=0
		Merged Map outputs=0
		GC time elapsed (ms)=65
		Total committed heap usage (bytes)=855638016
	File Input Format Counters 
		Bytes Read=33558528
21/04/26 21:28:18 INFO mapred.LocalJobRunner: Finishing task: attempt_local588307001_0001_m_000010_0
21/04/26 21:28:18 INFO mapred.LocalJobRunner: Starting task: attempt_local588307001_0001_m_000011_0
21/04/26 21:28:18 INFO output.FileOutputCommitter: File Output Committer Algorithm version is 1
21/04/26 21:28:18 INFO util.ProcfsBasedProcessTree: ProcfsBasedProcessTree currently is supported only on Linux.
21/04/26 21:28:18 INFO mapred.Task:  Using ResourceCalculatorProcessTree : org.apache.hadoop.yarn.util.WindowsBasedProcessTree@613a943c
21/04/26 21:28:18 INFO mapred.MapTask: Processing split: file:/E:/bigdata/2021.4.26/wc.txt:369098752+10911248
21/04/26 21:28:18 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)
21/04/26 21:28:18 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
21/04/26 21:28:18 INFO mapred.MapTask: soft limit at 83886080
21/04/26 21:28:18 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600
21/04/26 21:28:18 INFO mapred.MapTask: kvstart = 26214396; length = 6553600
21/04/26 21:28:18 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
21/04/26 21:28:18 INFO mapreduce.Job:  map 100% reduce 0%
21/04/26 21:28:19 INFO mapred.LocalJobRunner: 
21/04/26 21:28:19 INFO mapred.MapTask: Starting flush of map output
21/04/26 21:28:19 INFO mapred.MapTask: Spilling map output
21/04/26 21:28:19 INFO mapred.MapTask: bufstart = 0; bufend = 19550923; bufvoid = 104857600
21/04/26 21:28:19 INFO mapred.MapTask: kvstart = 26214396(104857584); kvend = 17096516(68386064); length = 9117881/6553600
21/04/26 21:28:19 INFO mapreduce.Job:  map 92% reduce 0%
21/04/26 21:28:20 INFO mapred.MapTask: Finished spill 0
21/04/26 21:28:20 INFO mapred.Task: Task:attempt_local588307001_0001_m_000011_0 is done. And is in the process of committing
21/04/26 21:28:20 INFO mapred.LocalJobRunner: map
21/04/26 21:28:20 INFO mapred.Task: Task 'attempt_local588307001_0001_m_000011_0' done.
21/04/26 21:28:20 INFO mapred.Task: Final Counters for attempt_local588307001_0001_m_000011_0: Counters: 17
	File System Counters
		FILE: Number of bytes read=1195640485
		FILE: Number of bytes written=1655552632
		FILE: Number of read operations=0
		FILE: Number of large read operations=0
		FILE: Number of write operations=0
	Map-Reduce Framework
		Map input records=956375
		Map output records=2279471
		Map output bytes=19550923
		Map output materialized bytes=24109871
		Input split bytes=98
		Combine input records=0
		Spilled Records=2279471
		Failed Shuffles=0
		Merged Map outputs=0
		GC time elapsed (ms)=50
		Total committed heap usage (bytes)=781713408
	File Input Format Counters 
		Bytes Read=10911248
21/04/26 21:28:20 INFO mapred.LocalJobRunner: Finishing task: attempt_local588307001_0001_m_000011_0
21/04/26 21:28:20 INFO mapred.LocalJobRunner: map task executor complete.
21/04/26 21:28:20 INFO mapred.LocalJobRunner: Waiting for reduce tasks
21/04/26 21:28:20 INFO mapred.LocalJobRunner: Starting task: attempt_local588307001_0001_r_000000_0
21/04/26 21:28:20 INFO output.FileOutputCommitter: File Output Committer Algorithm version is 1
21/04/26 21:28:20 INFO util.ProcfsBasedProcessTree: ProcfsBasedProcessTree currently is supported only on Linux.
21/04/26 21:28:20 INFO mapred.Task:  Using ResourceCalculatorProcessTree : org.apache.hadoop.yarn.util.WindowsBasedProcessTree@4ea8102b
21/04/26 21:28:20 INFO mapred.ReduceTask: Using ShuffleConsumerPlugin: org.apache.hadoop.mapreduce.task.reduce.Shuffle@3b3fc2dc
21/04/26 21:28:20 INFO reduce.MergeManagerImpl: MergerManager: memoryLimit=3995913472, maxSingleShuffleLimit=998978368, mergeThreshold=2637303040, ioSortFactor=10, memToMemMergeOutputsThreshold=10
21/04/26 21:28:20 INFO reduce.EventFetcher: attempt_local588307001_0001_r_000000_0 Thread started: EventFetcher for fetching Map Completion Events
21/04/26 21:28:20 INFO reduce.LocalFetcher: localfetcher#1 about to shuffle output of map attempt_local588307001_0001_m_000006_0 decomp: 74143110 len: 74143114 to MEMORY
21/04/26 21:28:20 INFO reduce.InMemoryMapOutput: Read 74143110 bytes from map-output for attempt_local588307001_0001_m_000006_0
21/04/26 21:28:20 INFO reduce.MergeManagerImpl: closeInMemoryFile -> map-output of size: 74143110, inMemoryMapOutputs.size() -> 1, commitMemory -> 0, usedMemory ->74143110
21/04/26 21:28:20 INFO reduce.LocalFetcher: localfetcher#1 about to shuffle output of map attempt_local588307001_0001_m_000009_0 decomp: 74143105 len: 74143109 to MEMORY
21/04/26 21:28:20 INFO reduce.InMemoryMapOutput: Read 74143105 bytes from map-output for attempt_local588307001_0001_m_000009_0
21/04/26 21:28:20 INFO reduce.MergeManagerImpl: closeInMemoryFile -> map-output of size: 74143105, inMemoryMapOutputs.size() -> 2, commitMemory -> 74143110, usedMemory ->148286215
21/04/26 21:28:20 INFO reduce.LocalFetcher: localfetcher#1 about to shuffle output of map attempt_local588307001_0001_m_000002_0 decomp: 74143099 len: 74143103 to MEMORY
21/04/26 21:28:20 INFO reduce.InMemoryMapOutput: Read 74143099 bytes from map-output for attempt_local588307001_0001_m_000002_0
21/04/26 21:28:20 INFO reduce.MergeManagerImpl: closeInMemoryFile -> map-output of size: 74143099, inMemoryMapOutputs.size() -> 3, commitMemory -> 148286215, usedMemory ->222429314
21/04/26 21:28:20 INFO reduce.LocalFetcher: localfetcher#1 about to shuffle output of map attempt_local588307001_0001_m_000005_0 decomp: 74143072 len: 74143076 to MEMORY
21/04/26 21:28:20 INFO mapreduce.Job:  map 100% reduce 0%
21/04/26 21:28:20 INFO reduce.InMemoryMapOutput: Read 74143072 bytes from map-output for attempt_local588307001_0001_m_000005_0
21/04/26 21:28:20 INFO reduce.MergeManagerImpl: closeInMemoryFile -> map-output of size: 74143072, inMemoryMapOutputs.size() -> 4, commitMemory -> 222429314, usedMemory ->296572386
21/04/26 21:28:20 INFO reduce.LocalFetcher: localfetcher#1 about to shuffle output of map attempt_local588307001_0001_m_000008_0 decomp: 74143113 len: 74143117 to MEMORY
21/04/26 21:28:20 INFO reduce.InMemoryMapOutput: Read 74143113 bytes from map-output for attempt_local588307001_0001_m_000008_0
21/04/26 21:28:20 INFO reduce.MergeManagerImpl: closeInMemoryFile -> map-output of size: 74143113, inMemoryMapOutputs.size() -> 5, commitMemory -> 296572386, usedMemory ->370715499
21/04/26 21:28:20 INFO reduce.LocalFetcher: localfetcher#1 about to shuffle output of map attempt_local588307001_0001_m_000011_0 decomp: 24109867 len: 24109871 to MEMORY
21/04/26 21:28:20 INFO reduce.InMemoryMapOutput: Read 24109867 bytes from map-output for attempt_local588307001_0001_m_000011_0
21/04/26 21:28:20 INFO reduce.MergeManagerImpl: closeInMemoryFile -> map-output of size: 24109867, inMemoryMapOutputs.size() -> 6, commitMemory -> 370715499, usedMemory ->394825366
21/04/26 21:28:20 INFO reduce.LocalFetcher: localfetcher#1 about to shuffle output of map attempt_local588307001_0001_m_000001_0 decomp: 74143098 len: 74143102 to MEMORY
21/04/26 21:28:20 INFO reduce.InMemoryMapOutput: Read 74143098 bytes from map-output for attempt_local588307001_0001_m_000001_0
21/04/26 21:28:20 INFO reduce.MergeManagerImpl: closeInMemoryFile -> map-output of size: 74143098, inMemoryMapOutputs.size() -> 7, commitMemory -> 394825366, usedMemory ->468968464
21/04/26 21:28:21 INFO reduce.LocalFetcher: localfetcher#1 about to shuffle output of map attempt_local588307001_0001_m_000004_0 decomp: 74143145 len: 74143149 to MEMORY
21/04/26 21:28:21 INFO reduce.InMemoryMapOutput: Read 74143145 bytes from map-output for attempt_local588307001_0001_m_000004_0
21/04/26 21:28:21 INFO reduce.MergeManagerImpl: closeInMemoryFile -> map-output of size: 74143145, inMemoryMapOutputs.size() -> 8, commitMemory -> 468968464, usedMemory ->543111609
21/04/26 21:28:21 INFO reduce.LocalFetcher: localfetcher#1 about to shuffle output of map attempt_local588307001_0001_m_000007_0 decomp: 74143104 len: 74143108 to MEMORY
21/04/26 21:28:21 INFO reduce.InMemoryMapOutput: Read 74143104 bytes from map-output for attempt_local588307001_0001_m_000007_0
21/04/26 21:28:21 INFO reduce.MergeManagerImpl: closeInMemoryFile -> map-output of size: 74143104, inMemoryMapOutputs.size() -> 9, commitMemory -> 543111609, usedMemory ->617254713
21/04/26 21:28:21 INFO reduce.LocalFetcher: localfetcher#1 about to shuffle output of map attempt_local588307001_0001_m_000010_0 decomp: 74143104 len: 74143108 to MEMORY
21/04/26 21:28:21 INFO reduce.InMemoryMapOutput: Read 74143104 bytes from map-output for attempt_local588307001_0001_m_000010_0
21/04/26 21:28:21 INFO reduce.MergeManagerImpl: closeInMemoryFile -> map-output of size: 74143104, inMemoryMapOutputs.size() -> 10, commitMemory -> 617254713, usedMemory ->691397817
21/04/26 21:28:21 INFO reduce.LocalFetcher: localfetcher#1 about to shuffle output of map attempt_local588307001_0001_m_000000_0 decomp: 74143109 len: 74143113 to MEMORY
21/04/26 21:28:21 INFO reduce.InMemoryMapOutput: Read 74143109 bytes from map-output for attempt_local588307001_0001_m_000000_0
21/04/26 21:28:21 INFO reduce.MergeManagerImpl: closeInMemoryFile -> map-output of size: 74143109, inMemoryMapOutputs.size() -> 11, commitMemory -> 691397817, usedMemory ->765540926
21/04/26 21:28:21 INFO reduce.LocalFetcher: localfetcher#1 about to shuffle output of map attempt_local588307001_0001_m_000003_0 decomp: 74143098 len: 74143102 to MEMORY
21/04/26 21:28:21 INFO reduce.InMemoryMapOutput: Read 74143098 bytes from map-output for attempt_local588307001_0001_m_000003_0
21/04/26 21:28:21 INFO reduce.MergeManagerImpl: closeInMemoryFile -> map-output of size: 74143098, inMemoryMapOutputs.size() -> 12, commitMemory -> 765540926, usedMemory ->839684024
21/04/26 21:28:21 INFO reduce.EventFetcher: EventFetcher is interrupted.. Returning
21/04/26 21:28:21 INFO mapred.LocalJobRunner: 12 / 12 copied.
21/04/26 21:28:21 INFO reduce.MergeManagerImpl: finalMerge called with 12 in-memory map-outputs and 0 on-disk map-outputs
21/04/26 21:28:21 INFO mapred.Merger: Merging 12 sorted segments
21/04/26 21:28:21 INFO mapred.Merger: Down to the last merge-pass, with 12 segments left of total size: 839683988 bytes
21/04/26 21:28:26 INFO mapred.LocalJobRunner: reduce > sort
21/04/26 21:28:26 INFO mapreduce.Job:  map 100% reduce 38%
21/04/26 21:28:29 INFO mapred.LocalJobRunner: reduce > sort
21/04/26 21:28:29 INFO mapreduce.Job:  map 100% reduce 43%
21/04/26 21:28:32 INFO mapred.LocalJobRunner: reduce > sort
21/04/26 21:28:32 INFO mapreduce.Job:  map 100% reduce 48%
21/04/26 21:28:35 INFO mapred.LocalJobRunner: reduce > sort
21/04/26 21:28:35 INFO mapreduce.Job:  map 100% reduce 52%
21/04/26 21:28:38 INFO mapred.LocalJobRunner: reduce > sort
21/04/26 21:28:38 INFO mapreduce.Job:  map 100% reduce 56%
21/04/26 21:28:41 INFO mapred.LocalJobRunner: reduce > sort
21/04/26 21:28:41 INFO mapreduce.Job:  map 100% reduce 59%
21/04/26 21:28:44 INFO mapred.LocalJobRunner: reduce > sort
21/04/26 21:28:44 INFO mapreduce.Job:  map 100% reduce 63%
21/04/26 21:28:46 INFO reduce.MergeManagerImpl: Merged 12 segments, 839684024 bytes to disk to satisfy reduce memory limit
21/04/26 21:28:46 INFO reduce.MergeManagerImpl: Merging 1 files, 839684006 bytes from disk
21/04/26 21:28:46 INFO reduce.MergeManagerImpl: Merging 0 segments, 0 bytes from memory into reduce
21/04/26 21:28:46 INFO mapred.Merger: Merging 1 sorted segments
21/04/26 21:28:46 INFO mapred.Merger: Down to the last merge-pass, with 1 segments left of total size: 839683999 bytes
21/04/26 21:28:46 INFO mapred.LocalJobRunner: reduce > sort
21/04/26 21:28:46 INFO Configuration.deprecation: mapred.skip.on is deprecated. Instead, use mapreduce.job.skiprecords
21/04/26 21:28:47 INFO mapred.LocalJobRunner: reduce > reduce
21/04/26 21:28:47 INFO mapreduce.Job:  map 100% reduce 67%
21/04/26 21:28:50 INFO mapred.LocalJobRunner: reduce > reduce
21/04/26 21:28:50 INFO mapreduce.Job:  map 100% reduce 69%
21/04/26 21:28:53 INFO mapred.LocalJobRunner: reduce > reduce
21/04/26 21:28:53 INFO mapreduce.Job:  map 100% reduce 72%
21/04/26 21:28:56 INFO mapred.LocalJobRunner: reduce > reduce
21/04/26 21:28:56 INFO mapreduce.Job:  map 100% reduce 76%
21/04/26 21:28:59 INFO mapred.LocalJobRunner: reduce > reduce
21/04/26 21:28:59 INFO mapreduce.Job:  map 100% reduce 79%
21/04/26 21:29:02 INFO mapred.LocalJobRunner: reduce > reduce
21/04/26 21:29:02 INFO mapreduce.Job:  map 100% reduce 83%
21/04/26 21:29:05 INFO mapred.LocalJobRunner: reduce > reduce
21/04/26 21:29:05 INFO mapreduce.Job:  map 100% reduce 86%
21/04/26 21:29:08 INFO mapred.LocalJobRunner: reduce > reduce
21/04/26 21:29:08 INFO mapreduce.Job:  map 100% reduce 89%
21/04/26 21:29:11 INFO mapred.LocalJobRunner: reduce > reduce
21/04/26 21:29:11 INFO mapreduce.Job:  map 100% reduce 92%
21/04/26 21:29:14 INFO mapred.LocalJobRunner: reduce > reduce
21/04/26 21:29:14 INFO mapreduce.Job:  map 100% reduce 95%
21/04/26 21:29:17 INFO mapred.LocalJobRunner: reduce > reduce
21/04/26 21:29:17 INFO mapreduce.Job:  map 100% reduce 99%
21/04/26 21:29:18 INFO mapred.Task: Task:attempt_local588307001_0001_r_000000_0 is done. And is in the process of committing
21/04/26 21:29:18 INFO mapred.LocalJobRunner: reduce > reduce
21/04/26 21:29:18 INFO mapred.Task: Task attempt_local588307001_0001_r_000000_0 is allowed to commit now
21/04/26 21:29:18 INFO output.FileOutputCommitter: Saved output of task 'attempt_local588307001_0001_r_000000_0' to file:/E:/bigdata/2021.4.26/result/_temporary/0/task_local588307001_0001_r_000000
21/04/26 21:29:18 INFO mapred.LocalJobRunner: reduce > reduce
21/04/26 21:29:18 INFO mapred.Task: Task 'attempt_local588307001_0001_r_000000_0' done.
21/04/26 21:29:18 INFO mapred.Task: Final Counters for attempt_local588307001_0001_r_000000_0: Counters: 24
	File System Counters
		FILE: Number of bytes read=2875008947
		FILE: Number of bytes written=2495236732
		FILE: Number of read operations=0
		FILE: Number of large read operations=0
		FILE: Number of write operations=0
	Map-Reduce Framework
		Combine input records=0
		Combine output records=0
		Reduce input groups=6
		Reduce shuffle bytes=839684072
		Reduce input records=79388000
		Reduce output records=6
		Spilled Records=79388000
		Shuffled Maps =12
		Failed Shuffles=0
		Merged Map outputs=12
		GC time elapsed (ms)=370
		Total committed heap usage (bytes)=2379218944
	Shuffle Errors
		BAD_ID=0
		CONNECTION=0
		IO_ERROR=0
		WRONG_LENGTH=0
		WRONG_MAP=0
		WRONG_REDUCE=0
	File Output Format Counters 
		Bytes Written=94
21/04/26 21:29:18 INFO mapred.LocalJobRunner: Finishing task: attempt_local588307001_0001_r_000000_0
21/04/26 21:29:18 INFO mapred.LocalJobRunner: reduce task executor complete.
21/04/26 21:29:18 INFO mapreduce.Job:  map 100% reduce 100%
21/04/26 21:29:18 INFO mapreduce.Job: Job job_local588307001_0001 completed successfully
21/04/26 21:29:18 INFO mapreduce.Job: Counters: 30
	File System Counters
		FILE: Number of bytes read=11179032658
		FILE: Number of bytes written=13940914909
		FILE: Number of read operations=0
		FILE: Number of large read operations=0
		FILE: Number of write operations=0
	Map-Reduce Framework
		Map input records=33308000
		Map output records=79388000
		Map output bytes=680908000
		Map output materialized bytes=839684072
		Input split bytes=1176
		Combine input records=0
		Combine output records=0
		Reduce input groups=6
		Reduce shuffle bytes=839684072
		Reduce input records=79388000
		Reduce output records=6
		Spilled Records=235884529
		Shuffled Maps =12
		Failed Shuffles=0
		Merged Map outputs=12
		GC time elapsed (ms)=1198
		Total committed heap usage (bytes)=15761670144
	Shuffle Errors
		BAD_ID=0
		CONNECTION=0
		IO_ERROR=0
		WRONG_LENGTH=0
		WRONG_MAP=0
		WRONG_REDUCE=0
	File Input Format Counters 
		Bytes Read=380055056
	File Output Format Counters 
		Bytes Written=94
status = 0

Process finished with exit code 0

比虚拟机集群还要慢!!!步骤还长。。。简直不能忍:

但结果是正确的。

二手房数量统计小项目

package com.aa.secondhouse;

import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.conf.Configured;
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.io.IntWritable;
import org.apache.hadoop.io.LongWritable;
import org.apache.hadoop.io.Text;
import org.apache.hadoop.mapreduce.Job;
import org.apache.hadoop.mapreduce.Mapper;
import org.apache.hadoop.mapreduce.Reducer;
import org.apache.hadoop.mapreduce.lib.input.TextInputFormat;
import org.apache.hadoop.mapreduce.lib.output.TextOutputFormat;
import org.apache.hadoop.util.Tool;
import org.apache.hadoop.util.ToolRunner;

import java.io.IOException;

public class SecondHouseCount extends Configured implements Tool {
   

    //构建、配置、提交
    @Override
    public int run(String[] strings) throws Exception {
   
        //实例化一个MapReduce的Job对象
        Job job = Job.getInstance(this.getConf(), "second house numb");
        //指定允许Jar包运行的类
        job.setJarByClass(SecondHouseCount.class);

        //InPut
        //指定输入类的类型
        job.setInputFormatClass(TextInputFormat.class);//可以不指定,默认是TextInputFormat
        //指定输入流
        Path inputPath = new Path("E:\\bigdata\\2021.4.26\\secondhouse.csv");//使用本地路径作为程序输入

        TextInputFormat.setInputPaths(job, inputPath);

        //Map:配置Map
        job.setMapperClass(SecondMapper.class);//用来设置Mapper类
        job.setMapOutputKeyClass(Text.class);//设置K2的类型
        job.setMapOutputValueClass(IntWritable.class);//设置V2的类型

        //shuffle:配置shuffle
        //job.setPartitionerClass(null);//设置分区器
        //job.setSortComparatorClass(null);//设置排序器
        //job.setGroupingComparatorClass(null);//设置分组器
        //job.setCombinerClass(null);//设置Map端聚合

        //Reduce:配置Reduce
        job.setReducerClass(SecondReducer.class);//设置调用Reduce的类
        job.setOutputKeyClass(Text.class);//设置K3的类型
        job.setOutputValueClass(IntWritable.class);//设置V3的类型
        job.setNumReduceTasks(1);//设置ReduceTask的个数,默认为1个

        //Output:配置输出
        job.setOutputFormatClass(TextOutputFormat.class);//默认是使用TextOutputFormat
        //设置输出的路径
        Path outputPath = new Path("E:\\bigdata\\2021.4.26\\result1");
        //判断输出路径是否存在,存在就删除
        FileSystem fs = FileSystem.get(this.getConf());
        if (fs.exists(outputPath)) {
   
            fs.delete(outputPath, true);
        }
        TextOutputFormat.setOutputPath(job, outputPath);

        //提交job
        return job.waitForCompletion(true) ? 0 : -1;
    }

    public static void main(String[] args) throws Exception {
   
        //构建配置管理对象
        Configuration conf = new Configuration();
        //通过工具类的run方法调用当前类的实例的run方法
        int status = ToolRunner.run(conf, new SecondHouseCount(), args);
        System.out.println("status = " + status);
        //退出程序
        System.exit(status);
    }

    public static class SecondMapper extends Mapper<LongWritable, Text, Text, IntWritable> {
   
        //输出K2
        Text outputKey = new Text();
        //输出V2
        IntWritable outputValue = new IntWritable(1);//不设置会导致结果=0

        @Override
        protected void map(LongWritable key, Text value, Context context) throws IOException, InterruptedException {
   
            //取出地区
            String region = value.toString().split(",")[3];
            //取出作为K2
            this.outputKey.set(region);
            //输出
            context.write(this.outputKey,this.outputValue);
        }
    }

    public static class SecondReducer extends Reducer<Text,IntWritable,Text,IntWritable>{
   
        //输出V3
        IntWritable outputValue=new IntWritable();

        @Override
        protected void reduce(Text key, Iterable<IntWritable> values, Context context) throws IOException, InterruptedException {
   
            int sum=0;
            for (IntWritable value : values) {
   
                sum+=value.get();
            }
            this.outputValue.set(sum);
            context.write(key,this.outputValue);
        }
    }

}

直接给定绝对路径。。。最好新建一个文件夹,不然这段程序执行后会把第二个路径清空。。。HDD还能抢救一下,SSD欲哭无泪。。。

csv是逗号分割的文本格式,大概长这副德行:

传感器之类的下位机有时候会保存文件为csv。

运行过程:

21/04/26 22:28:27 INFO Configuration.deprecation: session.id is deprecated. Instead, use dfs.metrics.session-id
21/04/26 22:28:27 INFO jvm.JvmMetrics: Initializing JVM Metrics with processName=JobTracker, sessionId=
21/04/26 22:28:28 WARN mapreduce.JobResourceUploader: No job jar file set.  User classes may not be found. See Job or Job#setJar(String).
21/04/26 22:28:28 INFO input.FileInputFormat: Total input paths to process : 1
21/04/26 22:28:28 INFO mapreduce.JobSubmitter: number of splits:1
21/04/26 22:28:29 INFO mapreduce.JobSubmitter: Submitting tokens for job: job_local1236341923_0001
21/04/26 22:28:29 INFO mapreduce.Job: The url to track the job: http://localhost:8080/
21/04/26 22:28:29 INFO mapreduce.Job: Running job: job_local1236341923_0001
21/04/26 22:28:29 INFO mapred.LocalJobRunner: OutputCommitter set in config null
21/04/26 22:28:29 INFO output.FileOutputCommitter: File Output Committer Algorithm version is 1
21/04/26 22:28:29 INFO mapred.LocalJobRunner: OutputCommitter is org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter
21/04/26 22:28:29 INFO mapred.LocalJobRunner: Waiting for map tasks
21/04/26 22:28:29 INFO mapred.LocalJobRunner: Starting task: attempt_local1236341923_0001_m_000000_0
21/04/26 22:28:29 INFO output.FileOutputCommitter: File Output Committer Algorithm version is 1
21/04/26 22:28:29 INFO util.ProcfsBasedProcessTree: ProcfsBasedProcessTree currently is supported only on Linux.
21/04/26 22:28:29 INFO mapred.Task:  Using ResourceCalculatorProcessTree : org.apache.hadoop.yarn.util.WindowsBasedProcessTree@3e639b3a
21/04/26 22:28:29 INFO mapred.MapTask: Processing split: file:/E:/bigdata/2021.4.26/secondhouse.csv:0+2225948
21/04/26 22:28:29 INFO mapred.MapTask: (EQUATOR) 0 kvi 26214396(104857584)
21/04/26 22:28:29 INFO mapred.MapTask: mapreduce.task.io.sort.mb: 100
21/04/26 22:28:29 INFO mapred.MapTask: soft limit at 83886080
21/04/26 22:28:29 INFO mapred.MapTask: bufstart = 0; bufvoid = 104857600
21/04/26 22:28:29 INFO mapred.MapTask: kvstart = 26214396; length = 6553600
21/04/26 22:28:29 INFO mapred.MapTask: Map output collector class = org.apache.hadoop.mapred.MapTask$MapOutputBuffer
21/04/26 22:28:29 INFO mapred.LocalJobRunner: 
21/04/26 22:28:29 INFO mapred.MapTask: Starting flush of map output
21/04/26 22:28:29 INFO mapred.MapTask: Spilling map output
21/04/26 22:28:29 INFO mapred.MapTask: bufstart = 0; bufend = 310211; bufvoid = 104857600
21/04/26 22:28:29 INFO mapred.MapTask: kvstart = 26214396(104857584); kvend = 26101596(104406384); length = 112801/6553600
21/04/26 22:28:29 INFO mapred.MapTask: Finished spill 0
21/04/26 22:28:29 INFO mapred.Task: Task:attempt_local1236341923_0001_m_000000_0 is done. And is in the process of committing
21/04/26 22:28:29 INFO mapred.LocalJobRunner: map
21/04/26 22:28:29 INFO mapred.Task: Task 'attempt_local1236341923_0001_m_000000_0' done.
21/04/26 22:28:29 INFO mapred.Task: Final Counters for attempt_local1236341923_0001_m_000000_0: Counters: 17
	File System Counters
		FILE: Number of bytes read=2226112
		FILE: Number of bytes written=660803
		FILE: Number of read operations=0
		FILE: Number of large read operations=0
		FILE: Number of write operations=0
	Map-Reduce Framework
		Map input records=28201
		Map output records=28201
		Map output bytes=310211
		Map output materialized bytes=366619
		Input split bytes=107
		Combine input records=0
		Spilled Records=28201
		Failed Shuffles=0
		Merged Map outputs=0
		GC time elapsed (ms)=0
		Total committed heap usage (bytes)=385875968
	File Input Format Counters 
		Bytes Read=2225948
21/04/26 22:28:29 INFO mapred.LocalJobRunner: Finishing task: attempt_local1236341923_0001_m_000000_0
21/04/26 22:28:29 INFO mapred.LocalJobRunner: map task executor complete.
21/04/26 22:28:29 INFO mapred.LocalJobRunner: Waiting for reduce tasks
21/04/26 22:28:29 INFO mapred.LocalJobRunner: Starting task: attempt_local1236341923_0001_r_000000_0
21/04/26 22:28:29 INFO output.FileOutputCommitter: File Output Committer Algorithm version is 1
21/04/26 22:28:29 INFO util.ProcfsBasedProcessTree: ProcfsBasedProcessTree currently is supported only on Linux.
21/04/26 22:28:29 INFO mapred.Task:  Using ResourceCalculatorProcessTree : org.apache.hadoop.yarn.util.WindowsBasedProcessTree@50329503
21/04/26 22:28:29 INFO mapred.ReduceTask: Using ShuffleConsumerPlugin: org.apache.hadoop.mapreduce.task.reduce.Shuffle@7a43db10
21/04/26 22:28:29 INFO reduce.MergeManagerImpl: MergerManager: memoryLimit=3995913472, maxSingleShuffleLimit=998978368, mergeThreshold=2637303040, ioSortFactor=10, memToMemMergeOutputsThreshold=10
21/04/26 22:28:29 INFO reduce.EventFetcher: attempt_local1236341923_0001_r_000000_0 Thread started: EventFetcher for fetching Map Completion Events
21/04/26 22:28:29 INFO reduce.LocalFetcher: localfetcher#1 about to shuffle output of map attempt_local1236341923_0001_m_000000_0 decomp: 366615 len: 366619 to MEMORY
21/04/26 22:28:29 INFO reduce.InMemoryMapOutput: Read 366615 bytes from map-output for attempt_local1236341923_0001_m_000000_0
21/04/26 22:28:29 INFO reduce.MergeManagerImpl: closeInMemoryFile -> map-output of size: 366615, inMemoryMapOutputs.size() -> 1, commitMemory -> 0, usedMemory ->366615
21/04/26 22:28:29 INFO reduce.EventFetcher: EventFetcher is interrupted.. Returning
21/04/26 22:28:29 INFO mapred.LocalJobRunner: 1 / 1 copied.
21/04/26 22:28:29 INFO reduce.MergeManagerImpl: finalMerge called with 1 in-memory map-outputs and 0 on-disk map-outputs
21/04/26 22:28:29 INFO mapred.Merger: Merging 1 sorted segments
21/04/26 22:28:29 INFO mapred.Merger: Down to the last merge-pass, with 1 segments left of total size: 366606 bytes
21/04/26 22:28:29 INFO reduce.MergeManagerImpl: Merged 1 segments, 366615 bytes to disk to satisfy reduce memory limit
21/04/26 22:28:29 INFO reduce.MergeManagerImpl: Merging 1 files, 366619 bytes from disk
21/04/26 22:28:29 INFO reduce.MergeManagerImpl: Merging 0 segments, 0 bytes from memory into reduce
21/04/26 22:28:29 INFO mapred.Merger: Merging 1 sorted segments
21/04/26 22:28:29 INFO mapred.Merger: Down to the last merge-pass, with 1 segments left of total size: 366606 bytes
21/04/26 22:28:29 INFO mapred.LocalJobRunner: 1 / 1 copied.
21/04/26 22:28:29 INFO Configuration.deprecation: mapred.skip.on is deprecated. Instead, use mapreduce.job.skiprecords
21/04/26 22:28:29 INFO mapred.Task: Task:attempt_local1236341923_0001_r_000000_0 is done. And is in the process of committing
21/04/26 22:28:29 INFO mapred.LocalJobRunner: 1 / 1 copied.
21/04/26 22:28:29 INFO mapred.Task: Task attempt_local1236341923_0001_r_000000_0 is allowed to commit now
21/04/26 22:28:29 INFO output.FileOutputCommitter: Saved output of task 'attempt_local1236341923_0001_r_000000_0' to file:/E:/bigdata/2021.4.26/result1/_temporary/0/task_local1236341923_0001_r_000000
21/04/26 22:28:29 INFO mapred.LocalJobRunner: reduce > reduce
21/04/26 22:28:29 INFO mapred.Task: Task 'attempt_local1236341923_0001_r_000000_0' done.
21/04/26 22:28:29 INFO mapred.Task: Final Counters for attempt_local1236341923_0001_r_000000_0: Counters: 24
	File System Counters
		FILE: Number of bytes read=2959382
		FILE: Number of bytes written=1027631
		FILE: Number of read operations=0
		FILE: Number of large read operations=0
		FILE: Number of write operations=0
	Map-Reduce Framework
		Combine input records=0
		Combine output records=0
		Reduce input groups=17
		Reduce shuffle bytes=366619
		Reduce input records=28201
		Reduce output records=17
		Spilled Records=28201
		Shuffled Maps =1
		Failed Shuffles=0
		Merged Map outputs=1
		GC time elapsed (ms)=9
		Total committed heap usage (bytes)=385875968
	Shuffle Errors
		BAD_ID=0
		CONNECTION=0
		IO_ERROR=0
		WRONG_LENGTH=0
		WRONG_MAP=0
		WRONG_REDUCE=0
	File Output Format Counters 
		Bytes Written=209
21/04/26 22:28:29 INFO mapred.LocalJobRunner: Finishing task: attempt_local1236341923_0001_r_000000_0
21/04/26 22:28:29 INFO mapred.LocalJobRunner: reduce task executor complete.
21/04/26 22:28:30 INFO mapreduce.Job: Job job_local1236341923_0001 running in uber mode : false
21/04/26 22:28:30 INFO mapreduce.Job:  map 100% reduce 100%
21/04/26 22:28:30 INFO mapreduce.Job: Job job_local1236341923_0001 completed successfully
21/04/26 22:28:30 INFO mapreduce.Job: Counters: 30
	File System Counters
		FILE: Number of bytes read=5185494
		FILE: Number of bytes written=1688434
		FILE: Number of read operations=0
		FILE: Number of large read operations=0
		FILE: Number of write operations=0
	Map-Reduce Framework
		Map input records=28201
		Map output records=28201
		Map output bytes=310211
		Map output materialized bytes=366619
		Input split bytes=107
		Combine input records=0
		Combine output records=0
		Reduce input groups=17
		Reduce shuffle bytes=366619
		Reduce input records=28201
		Reduce output records=17
		Spilled Records=56402
		Shuffled Maps =1
		Failed Shuffles=0
		Merged Map outputs=1
		GC time elapsed (ms)=9
		Total committed heap usage (bytes)=771751936
	Shuffle Errors
		BAD_ID=0
		CONNECTION=0
		IO_ERROR=0
		WRONG_LENGTH=0
		WRONG_MAP=0
		WRONG_REDUCE=0
	File Input Format Counters 
		Bytes Read=2225948
	File Output Format Counters 
		Bytes Written=209
status = 0

Process finished with exit code 0

执行后:

至此,已初步掌握MapReduce。

MapReduce进阶继续介绍MapReduce的更多用法。


转载:https://blog.csdn.net/qq_41990268/article/details/116170764
查看评论
* 以上用户言论只代表其个人观点,不代表本网站的观点或立场