当前位置: 首页 > 知识库问答 >
问题:

异常线程"main"java.lang.NoClassDefFoundError: org/apache/hadoop/跟踪/SpannReceiverHost

有宏峻
2023-03-14

我正在运行Hadoop 2.8。1和蜂巢2.3。0我正在尝试从配置单元中创建的a表中读取值,当前异常为

java.lang.ClassNotFoundException: org.apache.hadoop.tracing.SpanReceiverHost
    at java.net.URLClassLoader.findClass(URLClassLoader.java:381)
    at java.lang.ClassLoader.loadClass(ClassLoader.java:424)
    at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:331)
    at java.lang.ClassLoader.loadClass(ClassLoader.java:357)

这是我用来读取表格的代码

 public static final String HIVEURL = "jdbc:hive2://localhost:10000";
    public static final String DB_NAME = "default";
    public static final String TABLE_NAME = "order_line";

    public static void main(String[] args) throws Exception {
        HiveConf hiveConf = new HiveConf();
        //hiveConf.setVar(HiveConf.ConfVars.METASTOREURIS, HIVEURL);
        HiveMetaStoreClient hiveClient = new HiveMetaStoreClient(hiveConf);

        Job job =Job.getInstance();
        TaskAttemptContext ctx = new TaskAttemptContextImpl(job.getConfiguration(), new TaskAttemptID());
        HCatInputFormat hcif = HCatInputFormat.setInput(job, DB_NAME, TABLE_NAME);


        HCatSchema allCols = hcif.getTableSchema(job.getConfiguration());
        List<HCatFieldSchema> usedList = new ArrayList<>();
        usedList.add(allCols.get(2)); // por ex...
        HCatSchema someCols = new HCatSchema(usedList);
        hcif.setOutputSchema(job, someCols);

        for(InputSplit split: hcif.getSplits(job)) {
            RecordReader<WritableComparable, HCatRecord> rr = hcif.createRecordReader(split,ctx);
            rr.initialize(split, ctx);

            while(rr.nextKeyValue()) {
                HCatRecord record = rr.getCurrentValue();
                // usar record.get(...) para obter a coluna...
                //Object o = record.get(1);
                //System.out.println(o.toString());

            }

            rr.close();
        }

        hiveClient.close();
    }

这里是我使用的Pom文件

组织。阿帕奇。蜂箱hcatalog hive hcatalog core 2.3。0组织。阿帕奇。蜂箱hcatalog蜂巢hcatalog 0.13。1-cdh5。3.5组织。阿帕奇。蜂巢普通蜂巢2.3。0

    <dependency>
        <groupId>org.apache.hive</groupId>
        <artifactId>hive</artifactId>
        <version>0.13.1-cdh5.3.5</version>
    </dependency>

    <dependency>
        <groupId>org.apache.hive</groupId>
        <artifactId>hive-metastore</artifactId>
        <version>2.3.0</version>
    </dependency>

    <!-- https://mvnrepository.com/artifact/org.apache.hadoop/hadoop-common -->
    <dependency>
        <groupId>org.apache.hadoop</groupId>
        <artifactId>hadoop-common</artifactId>
        <version>2.8.1</version>
    </dependency>

    <!-- https://mvnrepository.com/artifact/org.apache.hadoop/hadoop-core -->
    <dependency>
        <groupId>org.apache.hadoop</groupId>
        <artifactId>hadoop-core</artifactId>
        <version>2.6.0-mr1-cdh5.12.1</version>
        <type>pom</type>
    </dependency>

    <dependency>
        <groupId>org.apache.thrift</groupId>
        <artifactId>libthrift</artifactId>
        <version>0.9.3</version>
    </dependency>
</dependencies>

共有2个答案

彭修筠
2023-03-14
<!-- https://mvnrepository.com/artifact/org.apache.hadoop/hadoop-common -->
<!-- org/apache/hadoop/tracing/SpanReceiverHost The Last version is hadoop-common 2.7.7-->
<dependency>
    <groupId>org.apache.hadoop</groupId>
    <artifactId>hadoop-common</artifactId>
    <version>2.7.7</version>
</dependency>
商正浩
2023-03-14

我看不出堆栈跟踪片段中的loadClass实际上是由什么引起的,但是看起来这个类实际上不存在于您正在使用的hadoop-通用的版本2.8.1中。它似乎在2.7.2之后的某个地方消失了

它,或者同名的东西,在hbase源代码中

你有版本的混搭吗?

 类似资料: