错误:java.lang.UnsatisfiedLinkError:org.apache.hadoop.io.nativeio.NativeIO$Windows

如何解决错误:java.lang.UnsatisfiedLinkError:org.apache.hadoop.io.nativeio.NativeIO$Windows

我是 Java 和 hadoop 的新手。我正在关注本教程 (https://developpaper.com/simple-java-hadoop-mapreduce-program-calculate-average-score-from-package-to-submit-and-run/)。我在提供输入时做了一些小的修改。请参考代码。

public class Score {

    public static class Map extends Mapper<LongWritable,Text,IntWritable> {
        // Implement map function
        public void map(LongWritable key,Text value,Context context) throws IOException,InterruptedException {
            // Convert the data of the input plain text file to string
            String line = value.toString();
            // Split the input data into rows first
            StringTokenizer tokenizerArticle = new StringTokenizer(line,"\n");
            // Process each line separately
            while (tokenizerArticle.hasMoreElements()) {
                // Line by space
                StringTokenizer tokenizerLine = new StringTokenizer(tokenizerArticle.nextToken());
                String strName = tokenizerLine.nextToken(); // student name
                                                            // section
                String strScore = tokenizerLine.nextToken(); // grade section
                Text name = new Text(strName);
                int scoreInt = Integer.parseInt(strScore);
                // Output name and score
                context.write(name,new IntWritable(scoreInt));
            }
        }
    }

    public static class Reduce extends Reducer<Text,IntWritable,IntWritable> {
        // Implement reduce function
        public void reduce(Text key,Iterable<IntWritable> values,Context context)
                throws IOException,InterruptedException {
            int sum = 0;
            int count = 0;
            Iterator<IntWritable> iterator = values.iterator();
            while (iterator.hasNext()) {
                sum += iterator.next().get(); // calculate the total score
                count++; // count the total number of accounts
            }
            Integer average = (int) sum / count; // calculate the average score
            context.write(key,new IntWritable(average));
        }
    }

    public static void main(String[] args) throws IOException,ClassNotFoundException,InterruptedException {
        // TODO Auto-generated method stub
        Configuration conf = new Configuration();
        // " localhost:9000 "It needs to be set according to the actual
        // situation
        conf.set("mapred.job.tracker","localhost:9000");
        // Input directory and output directory in an HDFS file system

        
        String[] ioArgs = new String[] { "score.txt","output" };
        String[] otherArgs = new GenericOptionsParser(conf,ioArgs).getRemainingArgs();
        if (otherArgs.length != 2) {
            System.err.println("Usage: Score Average <in> <out>");
            System.exit(2);
        }

        Job job = new Job(conf,"Score Average");
        job.setJarByClass(Score.class);
        // Set map,combine and reduce processing classes
        job.setMapperClass(Map.class);
        job.setCombinerClass(Reduce.class);
        job.setReducerClass(Reduce.class);
        // Set output type
        job.setOutputKeyClass(Text.class);
        job.setOutputValueClass(IntWritable.class);
        // The input data set is divided into small data blocks splites to
        // provide a RecordReder implementation
        job.setInputFormatClass(TextInputFormat.class);
        // Provide an implementation of recordwriter,responsible for data
        // output
        job.setOutputFormatClass(TextOutputFormat.class);
        // Set input and output directories
        FileInputFormat.addInputPath(job,new Path(otherArgs[0]));
        FileOutputFormat.setOutputPath(job,new Path(otherArgs[1]));
        System.exit(job.waitForCompletion(true) ? 0 : 1);
    }

}

在那里我使用 maven 在 eclips 中尝试了程序源代码。我为 pom.xml 文件添加了依赖项,如下所示。

<project xmlns="http://maven.apache.org/POM/4.0.0" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
  xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 http://maven.apache.org/xsd/maven-4.0.0.xsd">
  <modelVersion>4.0.0</modelVersion>

  <groupId>mapreducedemocode</groupId>
  <artifactId>mapreducedemocode</artifactId>
  <version>0.0.1-SNAPSHOT</version>
  <packaging>jar</packaging>

  <name>mapreducedemocode</name>
  <url>http://maven.apache.org</url>

  <properties>
    <project.build.sourceEncoding>UTF-8</project.build.sourceEncoding>
  </properties>

  <dependencies>
    <dependency>
      <groupId>junit</groupId>
      <artifactId>junit</artifactId>
      <version>3.8.1</version>
      <scope>test</scope>
    </dependency>
    <dependency>
      <groupId>org.apache.hadoop</groupId>
      <artifactId>hadoop-client</artifactId>
      <version>2.7.3</version>
    </dependency>
  </dependencies>
</project>

但是当我运行代码时,它给出了以下异常。

Exception in thread "main" java.lang.UnsatisfiedLinkError: org.apache.hadoop.io.nativeio.NativeIO$Windows.access0(Ljava/lang/String;I)Z
    at org.apache.hadoop.io.nativeio.NativeIO$Windows.access0(Native Method)

谁能帮我解决这个问题。请注意,我将输入文件 ('score.txt) 添加到项目中,并通过将参数添加为 'score.txt output' 来设置运行配置。提供参数是否有问题。

版权声明:本文内容由互联网用户自发贡献,该文观点与技术仅代表作者本人。本站仅提供信息存储空间服务,不拥有所有权,不承担相关法律责任。如发现本站有涉嫌侵权/违法违规的内容, 请发送邮件至 dio@foxmail.com 举报,一经查实,本站将立刻删除。

相关推荐


依赖报错 idea导入项目后依赖报错,解决方案:https://blog.csdn.net/weixin_42420249/article/details/81191861 依赖版本报错:更换其他版本 无法下载依赖可参考:https://blog.csdn.net/weixin_42628809/a
错误1:代码生成器依赖和mybatis依赖冲突 启动项目时报错如下 2021-12-03 13:33:33.927 ERROR 7228 [ main] o.s.b.d.LoggingFailureAnalysisReporter : *************************** APPL
错误1:gradle项目控制台输出为乱码 # 解决方案:https://blog.csdn.net/weixin_43501566/article/details/112482302 # 在gradle-wrapper.properties 添加以下内容 org.gradle.jvmargs=-Df
错误还原:在查询的过程中,传入的workType为0时,该条件不起作用 &lt;select id=&quot;xxx&quot;&gt; SELECT di.id, di.name, di.work_type, di.updated... &lt;where&gt; &lt;if test=&qu
报错如下,gcc版本太低 ^ server.c:5346:31: 错误:‘struct redisServer’没有名为‘server_cpulist’的成员 redisSetCpuAffinity(server.server_cpulist); ^ server.c: 在函数‘hasActiveC
解决方案1 1、改项目中.idea/workspace.xml配置文件,增加dynamic.classpath参数 2、搜索PropertiesComponent,添加如下 &lt;property name=&quot;dynamic.classpath&quot; value=&quot;tru
删除根组件app.vue中的默认代码后报错:Module Error (from ./node_modules/eslint-loader/index.js): 解决方案:关闭ESlint代码检测,在项目根目录创建vue.config.js,在文件中添加 module.exports = { lin
查看spark默认的python版本 [root@master day27]# pyspark /home/software/spark-2.3.4-bin-hadoop2.7/conf/spark-env.sh: line 2: /usr/local/hadoop/bin/hadoop: No s
使用本地python环境可以成功执行 import pandas as pd import matplotlib.pyplot as plt # 设置字体 plt.rcParams[&#39;font.sans-serif&#39;] = [&#39;SimHei&#39;] # 能正确显示负号 p
错误1:Request method ‘DELETE‘ not supported 错误还原:controller层有一个接口,访问该接口时报错:Request method ‘DELETE‘ not supported 错误原因:没有接收到前端传入的参数,修改为如下 参考 错误2:cannot r
错误1:启动docker镜像时报错:Error response from daemon: driver failed programming external connectivity on endpoint quirky_allen 解决方法:重启docker -&gt; systemctl r
错误1:private field ‘xxx‘ is never assigned 按Altʾnter快捷键,选择第2项 参考:https://blog.csdn.net/shi_hong_fei_hei/article/details/88814070 错误2:启动时报错,不能找到主启动类 #
报错如下,通过源不能下载,最后警告pip需升级版本 Requirement already satisfied: pip in c:\users\ychen\appdata\local\programs\python\python310\lib\site-packages (22.0.4) Coll
错误1:maven打包报错 错误还原:使用maven打包项目时报错如下 [ERROR] Failed to execute goal org.apache.maven.plugins:maven-resources-plugin:3.2.0:resources (default-resources)
错误1:服务调用时报错 服务消费者模块assess通过openFeign调用服务提供者模块hires 如下为服务提供者模块hires的控制层接口 @RestController @RequestMapping(&quot;/hires&quot;) public class FeignControl
错误1:运行项目后报如下错误 解决方案 报错2:Failed to execute goal org.apache.maven.plugins:maven-compiler-plugin:3.8.1:compile (default-compile) on project sb 解决方案:在pom.
参考 错误原因 过滤器或拦截器在生效时,redisTemplate还没有注入 解决方案:在注入容器时就生效 @Component //项目运行时就注入Spring容器 public class RedisBean { @Resource private RedisTemplate&lt;String
使用vite构建项目报错 C:\Users\ychen\work&gt;npm init @vitejs/app @vitejs/create-app is deprecated, use npm init vite instead C:\Users\ychen\AppData\Local\npm-