Hive / HCatalog Forum

Question concerning Hive installation

  • #22868

    I currently have a few installations of HDP V1.2. For now I will refer to my installation on a single node.In my log file for hive, I get the following error:
    13/04/22 10:20:48 INFO metastore.HiveMetaStore: 0: get_databases: default
    13/04/22 10:20:48 INFO HiveMetaStore.audit: ugi=hive ip=unknown-ip-addr cmd=get_databases: default
    13/04/22 10:20:48 INFO service.AbstractService: Service:ThriftSQLService is started.
    13/04/22 10:20:48 INFO service.AbstractService: Service:HiveServer2 is started.
    13/04/22 10:20:48 WARN conf.HiveConf: DEPRECATED: Configuration property hive.metastore.local no longer has any effect. Make sure
    to provide a valid value for hive.metastore.uris if you are connecting to a remote metastore.
    13/04/22 10:20:48 INFO thrift.ThriftCLIService: ThriftSQLService listening on 0.0.0.0/0.0.0.0:10000
    13/04/22 10:21:26 WARN conf.HiveConf: DEPRECATED: Configuration property hive.metastore.local no longer has any effect. Make sure
    to provide a valid value for hive.metastore.uris if you are connecting to a remote metastore.
    13/04/22 10:21:26 INFO cli.CLIService: SessionHandle [b7d0bcab-7214-4b87-96b2-430c857901b6]: openSession()
    Hive history file=/tmp/hive/hive_job_log_hive_201304221021_1025778294.txt
    13/04/22 10:21:26 INFO exec.HiveHistory: Hive history file=/tmp/hive/hive_job_log_hive_201304221021_1025778294.txt
    13/04/22 10:21:26 WARN security.ShellBasedUnixGroupsMapping: got exception trying to get groups for user fakeuser
    org.apache.hadoop.util.Shell$ExitCodeException: id: fakeuser: No such user

    at org.apache.hadoop.util.Shell.runCommand(Shell.java:255)
    at org.apache.hadoop.util.Shell.run(Shell.java:182)
    at org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:375)
    .
    .
    .
    Any idea why a user called “fakeuser” is being used? I am trying to make a beeline connection, which is not working and I am just wondering if this is the reason why.

    –Trevor WIlliams

to create new topics or reply. | New User Registration

  • Author
    Replies
  • #22873
    tedr
    Moderator

    Hi Trevor,

    It certainly could be why. Though not being familiar with beeline, I can’t say for sure, but beeline may have been configured to use that user.

    Thanks,
    Ted.

    #22875

    The real question is not about beeline working. I need to find out if the “hive” installation setup is somehow incorrect. Where is ‘fakeuser’ defined? and do I need to change it? Thanks…

    –Trevor Williams

    #23008

    OK. I found what I was looking for in the following file: /etc/hive/conf/hive-site.xml
    Using the following parameters:
    javax.jdo.option.ConnectionUserName
    javax.jdo.option.ConnectionPassword

    –Trevor

    #23031
    Larry Liu
    Moderator

    Hi, Trevor

    It is interesting. In my hive-site.xml, I don’t see fakeuser. Was this user setup on purpose?

    Larry

    #23037

    Hi Larry,

    The problem is I asked two questions in one (which seems to be a big no-no). One was concerning this error I was seeing in the hive log file “every” start up and I was having problems using beeline which is a CLI which can be used to make a connection to a remote HIVE server. The username is hive and the password is input by you during the installation/setup.I could not find where ‘fakeuser’ was specified.so I believe it is probably just a place holder and has no significance to Hive operation. Of course this is an assumption.

    –Trevor

    #23056
    Larry Liu
    Moderator

    Hi, Trevor

    Is there any reference in your /etc/passwd and /etc/group? Can you please also try to grep the fakeuser with all hadoop configuration files?
    For example:

    grep -R fakeuser /etc/*

    It must be somewhere since the error reported from hadoop when running shell command (org.apache.hadoop.util.Shell.runCommand)

    Thanks
    larry

    #23243

    Larry,

    Sorry, No joy. No fakeuser anywhere in the ‘/etc’ tree. It must be in the apache code. Please keep in mind, it is no longer an issue for me.

    –Trevor

    #23245

    Spoke too soon, I di find a reference to fakeuser in:
    /var/lib/ambari-agent/puppet/modules/hdp-hive/files/hiveserver2Smoke.sh
    the line looks like this:
    smokeout=`/usr/lib/hive/bin/beeline -u $1 -n fakeuser -p fakepwd -d org.apache.hive.jdbc.HiveDriver -e ‘!run $2′ 2>&1| awk ‘{print}’|grep Error`
    This is the smoke test for the start of hiveserver2.

    –Trevor

    #23319
    Seth Lyubich
    Moderator

    Hi Trevor,

    I think fakeuser is used only for Smoke test. I am able to see same on my 1.2 HDP installation and pass the smoke test:

    [root@ambari1 files]# grep -i fakeuser *
    hiveserver2Smoke.sh:smokeout=`/usr/lib/hive/bin/beeline -u $1 -n fakeuser -p fakepwd -d org.apache.hive.jdbc.HiveDriver -e ‘!run $2′ 2>&1| awk ‘{print}’|grep Error`
    [root@ambari1 files]# sh hiveserver2Smoke.sh
    Smoke test of hiveserver2 passed

    You can try to test following way:

    start beeline interface
    # /usr/lib/hive/bin/beeline

    Connect to Hive Server2:
    beeline> !connect jdbc:hive2://hiveserver2name:10000 username password org.apache.hive.jdbc.HiveDriver

    Then you cant try command like:
    show tables;

    And see if it works.

    Also you can make sure that port 10000 is bound to process and process is running.

    Please let us know if this is helpful.

    Thanks,
    Seth

You must be to reply to this topic. | Create Account

Support from the Experts

A HDP Support Subscription connects you experts with deep experience running Apache Hadoop in production, at-scale on the most demanding workloads.

Enterprise Support »

Become HDP Certified

Real world training designed by the core architects of Hadoop. Scenario-based training courses are available in-classroom or online from anywhere in the world

Training »

Hortonworks Data Platform
The Hortonworks Data Platform is a 100% open source distribution of Apache Hadoop that is truly enterprise grade having been built, tested and hardened with enterprise rigor.
Get started with Sandbox
Hortonworks Sandbox is a self-contained virtual machine with Apache Hadoop pre-configured alongside a set of hands-on, step-by-step Hadoop tutorials.
Modern Data Architecture
Tackle the challenges of big data. Hadoop integrates with existing EDW, RDBMS and MPP systems to deliver lower cost, higher capacity infrastructure.