您的位置:首页 > 其它

HIVE2 :beeline连接设置用户名和密码注意问题

2016-12-19 17:26 357 查看
beeline connect有几种方式,见hive-site.xml,缺省为NONE。

 <property>

    <name>hive.server2.authentication</name>

    <value>NONE</value>

    <description>

      Expects one of [nosasl, none, ldap, kerberos, pam, custom].

      Client authentication types.

        NONE: no authentication check

        LDAP: LDAP/AD based authentication

        KERBEROS: Kerberos/GSSAPI authentication

        CUSTOM: Custom authentication provider

                (Use with property hive.server2.custom.authentication.class)

        PAM: Pluggable authentication module

        NOSASL:  Raw transport

    </description>

  </property>

设置相应用户名和密码

<property>

    <name>hive.server2.thrift.client.user</name>

    <value>root</value>

    <description>Username to use against thrift client</description>

  </property>

  <property>

    <name>hive.server2.thrift.client.password</name>

    <value>123456</value>

    <description>Password to use against thrift client</description>
  </property>

注意这里设置的用户要求对inode="/tmp/hive" 有执行权限,否则会出现下列问题:

Connecting to jdbc:hive2://localhost:10000/default

Enter username for jdbc:hive2://localhost:10000/default: hive

Enter password for jdbc:hive2://localhost:10000/default: **

Error: Failed to open new session: java.lang.RuntimeException: org.apache.hadoop.security.AccessControlException: Permission denied: user=hive, access=EXECUTE, inode="/tmp/hive":root:supergroup:drwxrwx---

    at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.check(FSPermissionChecker.java:319)

    at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkTraverse(FSPermissionChecker.java:259)

    at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:205)

    at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:190)

    at org.apache.hadoop.hdfs.server.namenode.FSDirectory.checkPermission(FSDirectory.java:1698)

    at org.apache.hadoop.hdfs.server.namenode.FSDirStatAndListingOp.getFileInfo(FSDirStatAndListingOp.java:108)

    at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getFileInfo(FSNamesystem.java:3817)

    at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.getFileInfo(NameNodeRpcServer.java:1005)

    at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.getFileInfo(ClientNamenodeProtocolServerSideTranslatorPB.java:843)
内容来自用户分享和网络整理,不保证内容的准确性,如有侵权内容,可联系管理员处理 点击这里给我发消息
标签: 
相关文章推荐