• 关于处理hdfs数据库异常的解决办法 Please check the logs or run fsck in order to identify the missing blocks. See the Hadoop FAQ for common causes and potential solutions.


    今天突然发现,我们的flink程序挂掉了,然后先把程序启动起来,发现正常运行,然后去看什么问题,发现我们的集群当中的一台服务器挂掉了,然后重新启动了一台服务器

    然后紧接着,我去看yarn资源管理器看失败的任务发现超时

    Application application_1603766479824_0004 failed 1 times (global limit =2; local limit is =1) due to AM Container for appattempt_1603766479824_0004_000001 exited with exitCode: -100
    Failing this attempt.Diagnostics: Container released on a *lost* nodeFor more detailed output, check the application tracking page: http://ip-172-31-30-217.cn-northwest-1.compute.internal:8088/cluster/app/application_1603766479824_0004 Then click on links to logs of each attempt.
    . Failing the application.

    然后我们去hdfs的界面发现也有问题:

    Please check the logs or run fsck in order to identify the missing blocks. See the Hadoop FAQ for common causes and potential solutions.

     flink程序在写日志的时候,数据块好像有问题。解决这个问题的办法如下

    连接:http://www.julyme.com/20180202/99.html

     但是在执行的时候出现了一点小问题,文件目录权限的问题:

     Permission denied: user=root, access=READ_EXECUTE, inode="/tmp/entity-file-history/done":yarn:hadoo

    解决这个问题的方法是:

    先切回到你的用户下,然后执行 hadoop fsck -delete 命令这样数据块就被修复了

    至此问题得到解决。

  • 相关阅读:
    Jquery 复习练习(01)
    web前段 弹出小例子
    MacBook 显示隐藏文件夹命令
    sqlserver 纵横
    C#获取当前页面的url
    C# Json 转对象
    jquery导航栏
    AJAX
    hao dongxi
    微信网页获取openId
  • 原文地址:https://www.cnblogs.com/gxgd/p/14500107.html
Copyright © 2020-2023  润新知