国产机的三星竞争对手分析象是苹果三星!是世界级的!不要被对手的阴谋得逞!不跟国产机互喷!

aiven86 的BLOG
用户名:aiven86
文章数:12
访问量:23979
注册日期:
阅读量:5863
阅读量:12276
阅读量:378248
阅读量:1070912
51CTO推荐博文
下载所有源码包进行安装,过程中安装各种依赖包。过程中遇到的错误首先安装Python2.7安装其他依赖包,若报找不到zlib模块,则首先安装zlib-devel,然后重新编译python2.7error: command 'gcc' failed with exit status 1 的解决办法yum install gcc python-devel 需要lxml时执行:yum install python-lxmlkeystone glance quantum horizon swift错误:Traceback (most recent call last): &File &setup.py&, line 22, in &module& & &from mon import setup &File &/tmp/openstack_src/python-swiftclient-1.1.1/swiftclient/__init__.py&, line 5, in &module& & &from client import * &File &/tmp/openstack_src/python-swiftclient-1.1.1/swiftclient/client.py&, line 31, in &module& & &from httplib import HTTPException, HTTPSConnectionImportError: cannot import name HTTPSConnection安装openssl openssl-devel重新编译python2.7./configure make make install安装lxml错误:src/lxml/lxml.etree.c:163157: error: ‘error: command 'gcc' failed with exit status 1安装如下包解决yum install python-devel libxml2-devel libxslt-develImportError: No module named MySQLdb解决:安装MySQLdb如果通过yum安装有可能导致路径不正确,最好源码安装安装时需要依赖mysql-devel安装mysql-server创建数据库并修改权限分别是nova,keystone,glance,quantum:GRANT ALL PRIVILEGES ON *.* TO 'quantum'@'%' IDENTIFIED BY 'quantum';flush PRIVILEGES ;同步数据库nova-manage db syncglance-manage db_synckeystone-manage db_sync创建文件夹:/etc/nova/etc/glance/etc/keystone/etc/quantum/home/creds/novarc novarc_compute/var/log/nova/var/log/glance/var/log/keystone/var/log/quantum/var/lock/nova/var/lock/glance启动服务错误nova-api无法load ec2 from /etc/nova/api-paste.ini解决 安装python-paste-deploy-1.5该版本必须大于1.5sudo。。。。。。File &/usr/local/lib/python2.7/subprocess.py&, line 1249, in _execute_child 19:43:28 TRACE nova.service & & raise child_exception 19:43:28 TRACE nova.service OSError: [Errno 2] No such file or directory解决:安装sudo问题:Traceback (most recent call last):[Sat Aug 25 01:39:10 2012] [error] [client 10.28.163.59] & File &/opt/horizon-2012.2/openstack_dashboard/wsgi/django.wsgi&, line 4, in &module&[Sat Aug 25 01:39:10 2012] [error] [client 10.28.163.59] & & import django.core.handlers.wsgi[Sat Aug 25 01:39:10 2012] [error] [client 10.28.163.59] ImportError: No module named django.core.handlers.wsgi解决:安装mod_wsgi:./configure --with-python=/usr/local/bin/python2.7安装mod_wsgi-3.3过程中的错误/usr/bin/ld: /usr/local/lib/libpython2.7.a(abstract.o): relocation R_X86_64_32 against `.rodata.str1.8' can not be used when ma recompile with -fPIC/usr/local/lib/libpython2.7.a: could not read symbols: Bad valuecollect2: ld returned 1 exit statusapxs:Error: Command failed with rc=65536解决:源码安装mod_wsgi出现如下问题:[python]/usr/bin/ld: .../lib/libpython2.7.a(abstract.o): relocation R_X86_64_32 against 'a local symbol' can not be used when ma recompile with -fPIC &.../lib/libpython2.7.a: could not read symbols: Bad value &collect2: ld returned 1 exit status &网上有很多介绍该问题的解决方法:http://www.cbug.org//multiple-python-versions-cause-shared-library-mess.html#more-85 介绍的最为详细,不料一一试之,问题依然存在!最后在modwsgi官网/p/modwsgi/wiki/InstallationIssues发现如下:When attempting to compile mod_wsgi on a Linux system using an X86 64 bit processor, the following error message can arise:[python]/bin/sh /usr/lib64/apr/build/libtool --silent --mode=link gcc -o \ & &mod_wsgi.la -I/usr/local/include/python2.4 -DNDEBUG &-rpath \ & &/usr/lib64/httpd/modules -module -avoid-version mod_wsgi.lo \ & &-L/usr/local/lib/python2.4/config -lpython2.4 -lpthread -ldl -lutil &/usr/bin/ld: /usr/local/lib/python2.4/config/ &libpython2.4.a(abstract.o): relocation R_X86_64_32 against `a local &symbol' can not be used when ma recompile with -fPIC &/usr/local/lib/python2.4/config/libpython2.4.a: could not read symbols: Bad value &collect2: ld returned 1 exit status &apxs:Error: Command failed with rc=65536. &make: *** [mod_wsgi.la] Error 1This error is believed to be result of the version of Python being used having been originally compiled for the generic X86 32 bit architecture whereas mod_wsgi is being compiled for X86 64 bit architecture. The actual error arises in this case because 'libtool' would appear to be unable to generate a dynamically loadable module for the X86 64 bit architecture from a X86 32 bit static library. Alternatively, the problem is due to 'libtool' on this platform not being able to create a loadable module from a X86 64 bit static library in all cases.If the first issue, the only solution to this problem is to recompile Python for the X86 64 bit architecture. When doing this, it is preferable, and may actually be necessary, to ensure that the '--enable-shared' option is provided to the 'configure' script for Python when it is being compiled and installed.If rebuilding Python to generate a shared library, do make sure that the Python shared library, or a symlink to it appears in the Python 'config' directory of your Python installation. If the shared library doesn't appear here next to the static version of the library, 'libtool' will not be able to find it and will still use the static version of the library. It is understood that the Python build process may not actually do this, so you may have to do it by hand.If the version of Python being used was compiled for X86 64 bit architecture and a shared library does exist, but not in the 'config' directory, then adding the missing symlink may be all that is required.简单翻译之就是说如果遇到此问题,将python重新编译(以x86 64位静态库),即./configure时加上 --enable shared之后编译mod_wsgi会出现如下问题:[python]error while loading shared libraries: libpython2.6.so.1.0: \ & cannot open shared object file: No such file or directory &解决方法很简单:(debian系统)[python]其他系统具体详见:/susuper_/blog/item/b1a3af012eafbfb.html至此编译mod_wsgi,正常,make&&make install 一切顺利困扰了一天的问题终于解决,可以进行下一步了在/etc/httpd/conf/httpd.conf添加LoadModule wsgi_module modules/mod_wsgi.so问题:Memcached cache backend requires either the 'memcache' or 'cmemcache' library解决:easy_install python-memcached问题:FilterError: /usr/bin/env: node: No such file or directory解决:安装nodejs即node.js源码安装:安装好后复制/usr/local/bin/node 到/usr/bindashboard创建相关文件夹:mkdir /opt/horizon-2012.2/staticmkdir /opt/horizon-2012.2/static/dashboardmkdir /opt/horizon-2012.2/static/dashboard/css chown apache:apache -R static配置keystone登录dashboard,无效的用户名和密码执行keystone_data.sh脚本问题:apxs: command not found解决:安装httpd-develnova-computeUnable to load the virtualization driver: No module named libvirt将2.6的pythonpath加入到环境变量PYTHONPATH中可通过python&&&import sys&&&print sys.path查看路径export PYTHONPATH=../.../..:/../../:....keystone安装问题:运行keystone_data.sh必须初始化数据库否则用户名和密码错误ServiceCatalogException: Invalid service catalog service: computekeystone.conf中的catalog配置
了这篇文章
类别:未分类┆阅读(0)┆评论(0)作者:【吴业亮】开发
:http://blog.csdn.net/wylfengyujiancheng
在部署和使用openstack过程中们经常问我关于openstack问题,感谢小伙伴们的信任。其中有很多问题都是重复的,现将一些代表性的问题总结如下:
一、vnc打不开问题
1、检查controller和所以computer节点的nova.conf 文件vnc选项IP是否为controller的管理IP
# cat /etc/nova/nova.conf | grep “novncproxy_base_url“
2、novncproxy_base_url中尽量写IP,写主机名在无dns情况下无法解析导致vnc打不开
二、安装neutron 提示neutron agent-list 失败。
Cpu和内存不够导致,提高硬件配置
三、Ping不通?
1、检查物理链路是否正常,网线,交换机是否正常
2、网卡的状态是否为UP
3、neutron绑定的网卡是否为规划的网卡(各个节点都需检查)
4、neutron服务是否正常
5、是否为up
6、网关是否为up
7、虚拟机是否已经正常启动,且分配到IP
8、创建虚拟机所用的组是否允许icmp协议通过?
四、虚拟机获取不到IP?
1、中有其他的dhcp源干扰
2、neutron所桥接的网卡名称错误,或状态异常,如网线没插好,onboot=no
3、neutron服务异常
五、创建云硬盘失败
Ntp问题,各个节点时间不同步。
六、创建虚拟机失败
此类问题比较多,任何一个组件出问题都会导致虚拟机启动失败。最好的办法是掌握虚拟机创建过程中各个组件的调度过程,从日志分析。no valid host was found,这个是个大类,很多错误都会显示这个,需要具体的看日志。
七、安装cinder报 pvcreate /dev/sdb
&& vgcreate cinder-volumes /dev/sdb failed. ?
系统中没有sdb磁盘,cinder是块服务,需要底端存储。解决办法新增一块磁盘
八、管理网和业务网可以同网段吗?
尽量做到管理和业务分离,如果IP实在不够用记得不要冲突
九、官网上节点需要3张网卡,其它节点需要两张网卡,本方案为什么两块网卡?
本方案将第一块和第二块网卡合并了,如果需要三块网卡也可合并成一块网卡
十、创建的虚拟机运行http服务,外部无法访问?
1、虚拟机的开启
3、安全组没允许80端口通过
十一、重启controller物理机后整个环境不可用
在一些老的物理机或手写的网卡,重启物理机网卡都会启动失败,而消息队列和都是监听的管理网IP,网卡没起来导致消息队列和数据库启动失败,从而导致nova和neutron等一系列的服务启动失败。
解决办法:
通过下列命令查询那些服务启动失败,然后挨个手动启动。
ctl –failed
十二、外部可以访问虚拟机,但ping不通?
网络规划问题
网络问题排查思路:
1、ping 自己 目的:检查网卡状态和配置是否正确
2、ping网关 目的:检查寻址是否正常
3、ping dns 目的:检查dns是否可达
此类问题不清楚的,建议学习的路由和交换
十三、镜像忘记密码怎么办?
1、创建虚拟机时选择秘钥对,可从controller上免密码登录
2、创建虚拟机时直接修改密码
参见:/openstack-mirror-and-password/
十四、dashboard界面总是抛出异常错误?
在点击openstack的dashboard时右上角总是弹出一些错误的提示,再次刷新时又不提示
数据库安装完成后,默认最大连接数是100,一般流量稍微大一点这个连接数是远远不够的
解决办法:
1、修改mairadb的配置文件,将最大连接数改为1500
# echo "max_connections=1500" &&/f.d/f
2、重启数据库
十五、如何制作openstack镜像模板?
参见:http://blog.csdn.net/wylfengyujiancheng/article/details/
Linux模板类似
十六、后端存储如何选型?
上图为官方调查报告,根据公司资金及实力选择。个人比较推荐ceph
十七、在部署计算节点时,脚本不能继续进行,但没有报错
请确保controller节点正常运行且与computer节点网络正常,因为在computer节点上启动nova服务时,需要去controller上注册服务。
十八、消息队列启动失败
检查/etc/hosts中IP是否正确,如果采用脚本安装还需检查installrc和lib/hosts中IP是否正确
十九、数据库启动失败
检查监听IP是否正确,如果采用脚本安装还需检查installrc和lib/hosts中IP是否正确
如都没问题,手动卸载数据库,并删除库文件rm -rf /var/lib/mysql/再次安装
二十、执行source admin-openrc.sh后执行命令报错
环境中有代理,或者其他环境变量干扰。此问题也和开发人员习惯有关,在写代码是从不喜欢写绝对路径,环境中执行了该导致。出现问题是无从查起。
解决办法:
重启物理机、或者将admin-openrc.sh内容拷贝出来在命令行中执行一遍
二十一、通过dashboard界面上传镜像,创建虚拟机失败?
由于网络原因,在传输时文件极其容易损坏,建议现将镜像上传到后台,通过命令行上传。
# source /root/admin-openrc.sh
# openstack image create "cirros"
--file /tmp/cirros-0.3.4-x86_64-disk.img
--disk-format qcow2 --container-format bare
二十二、如何做到公司不同部门走不同网段网络?
1、用admin用户创建两外出网络,不同租户(对应公司部门)创建的虚拟机分配不同的浮动IP网络
2、也可将admin创建的不同网络绑定到对应的物理网卡上,方法后续补充
二十三、如何将官网的openstack源同步到本地?
采用reposync,将所有的源都同步下来
参见http://blog.csdn.net/wylfengyujiancheng/article/details/
二十四、Mongodb启动失败,如何解决?
监听的IP是否正确,如果采用脚本安装还需检查installrc和lib/hosts中IP是否正确
二十五、如何实现虚拟机迁移?
参见:http://blog.csdn.net/wylfengyujiancheng/article/details/
二十六、如何在线拉伸虚拟机大小?
http://blog.csdn.net/wylfengyujiancheng/article/details/
二十七、如何将快照导入到本地?
http://blog.csdn.net/wylfengyujiancheng/article/details/
二十八、cpu、内存和硬盘资源超配
它们默认的超配比为:
CPU: CONF.cpu_allocation_ratio = 16
RAM: CONF.ram_allocation_ratio = 1.5
DISK: CONF.disk_allocation_ratio = 1.0
在生产环境中的利用率建议cpu:vcpu 1:1.5 mem 1:1 disk 1:1
二十九、虚拟机流量全部通过network节点,大规模时如何减轻network的负载
采用dvr,分布式虚拟路由技术
请参见:/u/756/blog/3168
三十、win7镜像放到openstack中出 错误代码0X0000005D
这个该如何解决?
将虚拟化类型改为kvm
# cat /etc/nova/nova.conf | grep ^virt_type
三十一、安装neutron时修改网卡失败
nmcli connection modify ipv4.addresses 192.168.80.253/24 && nmcli connection modify ipv4.method manual && nmcli connection up failed.
请检查各个节点的各个网卡名称是否与实际名称一致,如不一致可用nmtui命令修改
待解决问题:
脚本中有没有安装swift组件?
虽然swift是openstack的元老组件,它和ceph比都有自己的优势和劣势,但本人比较倾向于ceph,脚本中没写swift的安装脚本,但本地源中包含了swift的rpm包。如有兴趣,请小伙伴们补充
/wuyeliang/install_openstack
有没有界面直接调用安装脚本,os安装完成后即登录http界面,通过点点鼠标即可完成部署?
前端技术本人不擅长,如有兴趣,请小伙伴们补充
/wuyeliang/install_openstack
有没有卸载脚本?
暂时没有,如有兴趣,请小伙伴们补充
/wuyeliang/install_openstack
关于高可用和负载均衡脚本?高可用和负载均衡采取哪种架构?
这个是需要花费不少时间暂时本人没这个精力写自动化脚本。如有兴趣,请小伙伴们补充。/wuyeliang/install_openstack
本人建议采用3controller模式。具体如下
数据库如下:
/endoresu/p/5029649.html
http://docs.openstack.org/ha-guide/Hadoop安装遇到的各种异常及解决办法(1) - 服务器技术综合 - 次元立方网 - 电脑知识与技术互动交流平台
Hadoop安装遇到的各种异常及解决办法(1)
11:10:23,665 INFO org.apache.Hadoop.ipc.Client: Retrying connect to server: -hadoop-38/10.10.208.38:9000. Already tried 0 time(s); retry policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS)
11:10:24,667 INFO org.apache.hadoop.ipc.Client: Retrying connect to server: -hadoop-38/10.10.208.38:9000. Already tried 1 time(s); retry policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS)
11:10:25,667 INFO org.apache.hadoop.ipc.Client: Retrying connect to server: Linux-hadoop-38/10.10.208.38:9000. Already tried 2 time(s); retry policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS)
11:10:26,669 INFO org.apache.hadoop.ipc.Client: Retrying connect to server: Linux-hadoop-38/10.10.208.38:9000. Already tried 3 time(s); retry policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS)
11:10:27,670 INFO org.apache.hadoop.ipc.Client: Retrying connect to server: Linux-hadoop-38/10.10.208.38:9000. Already tried 4 time(s); retry policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS)
11:10:28,671 INFO org.apache.hadoop.ipc.Client: Retrying connect to server: Linux-hadoop-38/10.10.208.38:9000. Already tried 5 time(s); retry policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS)
11:10:29,672 INFO org.apache.hadoop.ipc.Client: Retrying connect to server: Linux-hadoop-38/10.10.208.38:9000. Already tried 6 time(s); retry policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS)
11:10:30,674 INFO org.apache.hadoop.ipc.Client: Retrying connect to server: Linux-hadoop-38/10.10.208.38:9000. Already tried 7 time(s); retry policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS)
11:10:31,675 INFO org.apache.hadoop.ipc.Client: Retrying connect to server: Linux-hadoop-38/10.10.208.38:9000. Already tried 8 time(s); retry policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS)
11:10:32,676 INFO org.apache.hadoop.ipc.Client: Retrying connect to server: Linux-hadoop-38/10.10.208.38:9000. Already tried 9 time(s); retry policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS)
11:10:32,677 WARN org.apache.hadoop.hdfs.server.datanode.DataNode: Problem connecting to server: Linux-hadoop-38/10.10.208.38:9000
1、ping Linux-hadoop-38能通,telnet Linux-hadoop-38 9000不能通,说明开启了防火墙
2、去Linux-hadoop-38主机关闭防火墙/etc/init.d/iptables stop,显示:
iptables:清除防火墙规则:[确定]
iptables:将链设置为政策 ACCEPT:filter [确定]
iptables:正在卸载模块:[确定]
11:26:30,788 FATAL org.apache.hadoop.hdfs.server.datanode.DataNode: Initialization failed for block pool Block pool BP--10.10.208.38-8 (storage id DS-7.0.0.1-6048958) service to Linux-hadoop-38/10.10.208.38:9000
java.io.IOException: Incompatible clusterIDs in /usr/local/hadoop/tmp/dfs/data: namenode clusterID = CID-8efaa-440a-b61c-290e4ccfb006; datanode clusterID = clustername
at org.apache.hadoop.hdfs.server.datanode.DataStorage.doTransition(DataStorage.java:391)
at org.apache.hadoop.hdfs.server.datanode.DataStorage.recoverTransitionRead(DataStorage.java:191)
at org.apache.hadoop.hdfs.server.datanode.DataStorage.recoverTransitionRead(DataStorage.java:219)
at org.apache.hadoop.hdfs.server.datanode.DataNode.initStorage(DataNode.java:916)
at org.apache.hadoop.hdfs.server.datanode.DataNode.initBlockPool(DataNode.java:887)
at org.apache.hadoop.hdfs.server.datanode.BPOfferService.verifyAndSetNamespaceInfo(BPOfferService.java:309)
at org.apache.hadoop.hdfs.server.datanode.BPServiceActor.connectToNNAndHandshake(BPServiceActor.java:218)
at org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:660)
at java.lang.Thread.run(Thread.java:662)
1、在hdfs-site.xml配置文件中,配置了dfs.namenode.name.dir,在master中,该配置的目录下有个current文件夹,里面有个VERSION文件,内容如下:
#Thu Mar 13 10:51:23 CST 2014
namespaceID=
clusterID=CID-8efaa-440a-b61c-290e4ccfb006
storageType=NAME_NODE
blockpoolID=BP--10.10.208.38-8
layoutVersion=-40
2、在core-site.xml配置文件中,配置了hadoop.tmp.dir,在slave中,该配置的目录下有个dfs/data/current目录,里面也有一个VERSION文件,内容
#Wed Mar 12 17:23:04 CST 2014
storageID=DS-.10.208.54-6184818
clusterID=clustername
storageType=DATA_NODE
layoutVersion=-40
3、一目了然,两个内容不一样,导致的。删除slave中的错误内容,重启,搞定!
参考资料: /Linux/98.htm
12:34:46,828 FATAL org.apache.hadoop.yarn.server.nodemanager.containermanager.AuxServices: Failed to initialize mapreduce_shuffle
java.lang.RuntimeException: No class defiend for mapreduce_shuffle
at org.apache.hadoop.yarn.server.nodemanager.containermanager.AuxServices.init(AuxServices.java:94)
at org.apache.hadoop.positeService.init(CompositeService.java:58)
at org.apache.hadoop.yarn.server.nodemanager.containermanager.ContainerManagerImpl.init(ContainerManagerImpl.java:181)
at org.apache.hadoop.positeService.init(CompositeService.java:58)
at org.apache.hadoop.yarn.server.nodemanager.NodeManager.init(NodeManager.java:185)
at org.apache.hadoop.yarn.server.nodemanager.NodeManager.initAndStartNodeManager(NodeManager.java:328)
at org.apache.hadoop.yarn.server.nodemanager.NodeManager.main(NodeManager.java:351)
12:34:46,830 FATAL org.apache.hadoop.yarn.server.nodemanager.NodeManager: Error starting NodeManager
java.lang.RuntimeException: No class defiend for mapreduce_shuffle
at org.apache.hadoop.yarn.server.nodemanager.containermanager.AuxServices.init(AuxServices.java:94)
at org.apache.hadoop.positeService.init(CompositeService.java:58)
at org.apache.hadoop.yarn.server.nodemanager.containermanager.ContainerManagerImpl.init(ContainerManagerImpl.java:181)
at org.apache.hadoop.positeService.init(CompositeService.java:58)
at org.apache.hadoop.yarn.server.nodemanager.NodeManager.init(NodeManager.java:185)
at org.apache.hadoop.yarn.server.nodemanager.NodeManager.initAndStartNodeManager(NodeManager.java:328)
at org.apache.hadoop.yarn.server.nodemanager.NodeManager.main(NodeManager.java:351)
12:34:46,846 INFO org.apache.hadoop.yarn.server.nodemanager.containermanager.monitor.ContainersMonitorImpl: ResourceCalculatorPlugin is unavailable on this system. org.apache.hadoop.yarn.server.nodemanager.containermanager.monitor.ContainersMonitorImpl is disabled.
1、yarn-site.xml配置错误:
&property&
&name&yarn.nodemanager.aux-services&/name&
&value&mapreduce_shuffle&/value&
&/property&
2、修改为:
&property&
&name&yarn.nodemanager.aux-services&/name&
&value&mapreduce.shuffle&/value&
&/property&
3、重启服务
WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
参考 /Linux/99.htm
14/03/13 17:25:41 ERROR lzo.GPLNativeCodeLoader: Could not load native gpl library
java.lang.UnsatisfiedLinkError: no gplcompression in java.library.path
at java.lang.ClassLoader.loadLibrary(ClassLoader.java:1734)
at java.lang.Runtime.loadLibrary0(Runtime.java:823)
at java.lang.System.loadLibrary(System.java:1028)
at pression.lzo.GPLNativeCodeLoader.&clinit&(GPLNativeCodeLoader.java:32)
at pression.lzo.LzoCodec.&clinit&(LzoCodec.java:67)
at pression.lzo.LzoIndexer.&init&(LzoIndexer.java:36)
at pression.lzo.LzoIndexer.main(LzoIndexer.java:134)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.apache.hadoop.util.RunJar.main(RunJar.java:208)
14/03/13 17:25:41 ERROR lzo.LzoCodec: Cannot load native-lzo without native-hadoop
14/03/13 17:25:43 INFO lzo.LzoIndexer: [INDEX] LZO Indexing file /test2.lzo, size 0.00 GB...
Exception in thread &main& java.lang.RuntimeException: native-lzo library not available
at pression.lzo.LzopCodec.createDecompressor(LzopCodec.java:91)
at pression.lzo.LzoIndex.createIndex(LzoIndex.java:222)
at pression.lzo.LzoIndexer.indexSingleFile(LzoIndexer.java:117)
at pression.lzo.LzoIndexer.indexInternal(LzoIndexer.java:98)
at pression.lzo.LzoIndexer.index(LzoIndexer.java:52)
at pression.lzo.LzoIndexer.main(LzoIndexer.java:137)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.apache.hadoop.util.RunJar.main(RunJar.java:208)
解决办法:很明显,没有native-lzo
编译安装/编译lzo,/Linux/01.htm
14/03/17 10:23:59 INFO mapreduce.JobSubmitter: Cleaning up the staging area /tmp/hadoop-yarn/staging/hadoop/.staging/job_6_0003
java.lang.IllegalArgumentException: Compression codec pression.lzo.LzoCodec not found.
at org.apache.pressionCodecFactory.getCodecClasses(CompressionCodecFactory.java:134)
at org.apache.pressionCodecFactory.&init&(CompressionCodecFactory.java:174)
at org.apache.hadoop.mapreduce.lib.input.TextInputFormat.isSplitable(TextInputFormat.java:58)
at org.apache.hadoop.mapreduce.lib.input.FileInputFormat.getSplits(FileInputFormat.java:276)
at org.apache.hadoop.mapreduce.JobSubmitter.writeNewSplits(JobSubmitter.java:468)
at org.apache.hadoop.mapreduce.JobSubmitter.writeSplits(JobSubmitter.java:485)
at org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:369)
at org.apache.hadoop.mapreduce.Job$11.run(Job.java:1269)
at org.apache.hadoop.mapreduce.Job$11.run(Job.java:1266)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:396)
at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1408)
at org.apache.hadoop.mapreduce.Job.submit(Job.java:1266)
at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:1287)
at org.apache.hadoop.examples.WordCount.main(WordCount.java:84)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.apache.hadoop.util.ProgramDriver$ProgramDescription.invoke(ProgramDriver.java:72)
at org.apache.hadoop.util.ProgramDriver.driver(ProgramDriver.java:144)
at org.apache.hadoop.examples.ExampleDriver.main(ExampleDriver.java:68)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.apache.hadoop.util.RunJar.main(RunJar.java:208)
Caused by: java.lang.ClassNotFoundException: Class pression.lzo.LzoCodec not found
at org.apache.hadoop.conf.Configuration.getClassByName(Configuration.java:1680)
at org.apache.pressionCodecFactory.getCodecClasses(CompressionCodecFactory.java:127)
... 26 more
临时解决办法:
将/usr/local/hadoop/lib/hadoop-lzo-0.4.10.jar拷贝到/usr/local/jdk/lib下,重启linux
14/03/17 10:35:03 ERROR security.UserGroupInformation: PriviledgedActionException as:Hadoop (auth:SIMPLE) cause:org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.hdfs.server.namenode.SafeModeException): Cannot delete /tmp/hadoop-yarn/staging/hadoop/.staging/job_7_0001. Name node is in safe mode.
The reported blocks 0 needs additional 12 blocks to reach the threshold 0.9990 of total blocks 12. Safe mode will be turned off automatically.
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.deleteInternal(FSNamesystem.java:2905)
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.deleteInt(FSNamesystem.java:2872)
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.delete(FSNamesystem.java:2859)
at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.delete(NameNodeRpcServer.java:642)
at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.delete(ClientNamenodeProtocolServerSideTranslatorPB.java:408)
at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java:44968)
at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:453)
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1002)
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1752)
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1748)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:396)
at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1408)
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1746)
org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.hdfs.server.namenode.SafeModeException): Cannot delete /tmp/hadoop-yarn/staging/hadoop/.staging/job_7_0001. Name node is in safe mode.
The reported blocks 0 needs additional 12 blocks to reach the threshold 0.9990 of total blocks 12. Safe mode will be turned off automatically.
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.deleteInternal(FSNamesystem.java:2905)
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.deleteInt(FSNamesystem.java:2872)
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.delete(FSNamesystem.java:2859)
at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.delete(NameNodeRpcServer.java:642)
at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.delete(ClientNamenodeProtocolServerSideTranslatorPB.java:408)
at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java:44968)
at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:453)
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1002)
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1752)
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1748)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:396)
at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1408)
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1746)
at org.apache.hadoop.ipc.Client.call(Client.java:1238)
at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:202)
at $Proxy9.delete(Unknown Source)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:164)
at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:83)
at $Proxy9.delete(Unknown Source)
at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.delete(ClientNamenodeProtocolTranslatorPB.java:408)
at org.apache.hadoop.hdfs.DFSClient.delete(DFSClient.java:1487)
at org.apache.hadoop.hdfs.DistributedFileSystem.delete(DistributedFileSystem.java:355)
at org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:418)
at org.apache.hadoop.mapreduce.Job$11.run(Job.java:1269)
at org.apache.hadoop.mapreduce.Job$11.run(Job.java:1266)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:396)
at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1408)
at org.apache.hadoop.mapreduce.Job.submit(Job.java:1266)
at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:1287)
at org.apache.hadoop.examples.WordCount.main(WordCount.java:84)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.apache.hadoop.util.ProgramDriver$ProgramDescription.invoke(ProgramDriver.java:72)
at org.apache.hadoop.util.ProgramDriver.driver(ProgramDriver.java:144)
at org.apache.hadoop.examples.ExampleDriver.main(ExampleDriver.java:68)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.apache.hadoop.util.RunJar.main(RunJar.java:208)
解决办法:
明显是安全模式:
原因是reboot机器的时候,防火墙起来了,导致NodeManager启动不起来,导致一直是安全模式。关闭防火墙重启hadoop,ok!
10:35:10,447 INFO org.apache.hadoop.hdfs.server.namenode.NameNode: Caching file names occuring more than 10 times
10:35:10,450 INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: dfs.namenode.safemode.threshold-pct = 0.6033
10:35:10,450 INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: dfs.namenode.safemode.min.datanodes = 0
10:35:10,450 INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: dfs.namenode.safemode.extension = 30000
10:35:10,476 INFO org.apache.hadoop.mon.Storage: Lock on /usr/local/hadoop/hdfs/name/in_use.lock acquired by nodename 9580@Linux-hadoop-38
10:35:10,479 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Stopping NameNode metrics system...
10:35:10,480 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics system stopped.
10:35:10,480 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics system shutdown complete.
10:35:10,480 FATAL org.apache.hadoop.hdfs.server.namenode.NameNode: Exception in namenode join
java.io.IOException: NameNode is not formatted.
at org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:217)
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFSImage(FSNamesystem.java:728)
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFromDisk(FSNamesystem.java:521)
at org.apache.hadoop.hdfs.server.namenode.NameNode.loadNamesystem(NameNode.java:403)
at org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:437)
at org.apache.hadoop.hdfs.server.namenode.NameNode.&init&(NameNode.java:613)
at org.apache.hadoop.hdfs.server.namenode.NameNode.&init&(NameNode.java:598)
at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1169)
at org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1233)
10:35:10,484 INFO org.apache.hadoop.util.ExitUtil: Exiting with status 1
10:35:10,501 INFO org.apache.hadoop.hdfs.server.namenode.NameNode: SHUTDOWN_MSG:
hadoop namenode -format,不是hdfs namenode -format
启动JobHistoryServer
sbin/mr-jobhistory-daemon.sh start historyserver
Datanode denied communication with namenode: DatanodeRegistration 解决办法
Hadoop版本:2.2.0
单机伪同步环境。
启动start-dfs.sh后,发现本地没有datanode进程,查看datanode的日志如下:
23:48:11,357 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Block pool BP--127.0.0.1-3 (storage id DS--127.0.0.1-9655564) service to localhost/192.168.1.101:9000 beginning handshake with NN
23:48:11,381 FATAL org.apache.hadoop.hdfs.server.datanode.DataNode: Initialization failed for block pool Block pool BP--127.0.0.1-3 (storage id DS--127.0.0.1-9655564) service to localhost/192.168.1.101:9000
org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.hdfs.server.protocol.DisallowedDatanodeException):Datanode denied communication with namenode: DatanodeRegistration(0.0.0.0, storageID=DS--127.0.0.1-9655564, infoPort=50075, ipcPort=50020, storageInfo=lv=-47;cid=CID-ba95b66c-d94b--adb486ba5682;nsid=;c=0)
at org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager.registerDatanode(DatanodeManager.java:739)
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.registerDatanode(FSNamesystem.java:3929)
at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.registerDatanode(NameNodeRpcServer.java:948)
at org.apache.hadoop.hdfs.protocolPB.DatanodeProtocolServerSideTranslatorPB.registerDatanode(DatanodeProtocolServerSideTranslatorPB.java:90)
at org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos$DatanodeProtocolService$2.callBlockingMethod(DatanodeProtocolProtos.java:24079)
at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:585)
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:928)
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2048)
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2044)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:394)
at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1491)
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2042)
at org.apache.hadoop.ipc.Client.call(Client.java:1347)
at org.apache.hadoop.ipc.Client.call(Client.java:1300)
at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:206)
at com.sun.proxy.$Proxy9.registerDatanode(Unknown Source)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:186)
at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:102)
at com.sun.proxy.$Proxy9.registerDatanode(Unknown Source)
at org.apache.hadoop.hdfs.protocolPB.DatanodeProtocolClientSideTranslatorPB.registerDatanode(DatanodeProtocolClientSideTranslatorPB.java:146)
at org.apache.hadoop.hdfs.server.datanode.BPServiceActor.register(BPServiceActor.java:623)
at org.apache.hadoop.hdfs.server.datanode.BPServiceActor.connectToNNAndHandshake(BPServiceActor.java:225)
at org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:664)
at java.lang.Thread.run(Thread.java:695)
23:48:11,382 WARN org.apache.hadoop.hdfs.server.datanode.DataNode: Ending block pool service for: Block pool BP--127.0.0.1-3 (storage id DS--127.0.0.1-9655564) service to localhost/192.168.1.101:9000
23:48:11,484 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Removed Block pool BP--127.0.0.1-3 (storage id DS--127.0.0.1-9655564)
23:48:11,484 INFO org.apache.hadoop.hdfs.server.datanode.DataBlockScanner: Removed bpid=BP--127.0.0.1-3 from blockPoolScannerMap
23:48:11,484 INFO org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl: Removing block pool BP--127.0.0.1-3
23:48:13,485 WARN org.apache.hadoop.hdfs.server.datanode.DataNode: Exiting Datanode
23:48:13,486 INFO org.apache.hadoop.util.ExitUtil: Exiting with status 0
23:48:13,487 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: SHUTDOWN_MSG:
/************************************************************
SHUTDOWN_MSG: Shutting down DataNode at localhost/127.0.0.1
************************************************************/
导致这个问题的原因很多,在stackoverflow上有很多解决方案,但都不适用我。经过摸索,通过修改/etc/hosts解决。
/etc/hosts原来为:
127.0.0.1 localhost
由于我本机IP地址为 192.168.1.101,所以将hosts文件修改为这个ip地址:
192.168.1.101 localhost
理论上127.0.0.1也是可行的,但不知道我的为什么不行。
修改后,重新删除dfs.namenode.name.dir和dfs.datanode.data.dir以及hadoop.tmp.dir目录,分别执行:
hdfs namenode -format
start-dfs.sh
start-yarn.sh
后,用jps查看服务,全部正常:
37619 NodeManager
37247 NameNode
37330 DataNode
37536 ResourceManager
37432 SecondaryNameNode
延伸阅读:
目的:通过挂载的方式,可以类似访问本地磁盘的方式一...
本教程为 李华明 编著的iOS-Cocos2d游戏开发系列教程:教程涵盖关于i......
专题主要学习DirectX的初级编程入门学习,对Directx11的入门及初学者有......
&面向对象的JavaScript&这一说法多少有些冗余,因为JavaScript 语言本......
Windows7系统专题 无论是升级操作系统、资料备份、加强资料的安全及管......

我要回帖

更多关于 竞争分析的主要对象是 的文章

 

随机推荐