且构网

分享程序员开发的那些事...
且构网 - 分享程序员编程开发的那些事

EMR Spark-TransportClient:无法发送RPC

更新时间:2021-11-22 04:19:29

最后,我解决了这个问题.这是由于磁盘空间不足.一列hadoop日志显示:

Finally I resolved the problem. It was due to insufficient disk space. One column of hadoop logs showed:

Hadoop纱:1/1本地目录不正确:/var/lib/hadoop-yarn/cache/yarn/nm-local-dir; 1/1日志目录不正确:/var/log/hadoop-yarn/containers

Hadoop YARN: 1/1 local-dirs are bad: /var/lib/hadoop-yarn/cache/yarn/nm-local-dir; 1/1 log-dirs are bad: /var/log/hadoop-yarn/containers

使用Google搜索,发现

Googling it I found http://gethue.com/hadoop-yarn-11-local-dirs-are-bad-varlibhadoop-yarncacheyarnnm-local-dir-11-log-dirs-are-bad-varloghadoop-yarncontainers/

如果遇到此错误,请腾出一些磁盘空间!"

"If you are getting this error, make some disk space!"

要查看此错误,我必须激活EMR中的纱线记录.参见

To see this error I have to activate the yarn logs in EMR. See

http://docs.aws.amazon. com/emr/latest/ManagementGuide/emr-web-interfaces.html

要访问群集ec2实例中的日志端口,我为此更改了安全组

To have access to the logs port in the cluster ec2 instances I changed security groups for it

即:

主实例在这里监听:172.30.12.84:8088和核心 此处的实例:172.30.12.21:8042

master instance was listening here: 172.30.12.84:8088 And core instance here: 172.30.12.21:8042

最后,我解决了在etl.py中更改实例类型,使其他实例具有更大磁盘的问题:

Finally I fixed the problem changing in etl.py the type of instances by other ones with bigger disks:

主版:m3.2xlarge
核心:c3.4xlarge

master: m3.2xlarge
core: c3.4xlarge