HDP on Linux – Installation Forum

Ambari MySQL install failure

  • #52318
    Bruno Sarlandie

    System CentOS 6.5 (Final), Kernel 2.6.32-431.el6.x86_64, 1 server + 2 virtualbox,
    On install of MySQL, (mysql removed before installation) :
    Traceback (most recent call last):
    File “/var/lib/ambari-agent/cache/stacks/HDP/2.0.6/services/HIVE/package/scripts/mysql_server.py”, line 70, in <module>
    File “/usr/lib/python2.6/site-packages/resource_management/libraries/script/script.py”, line 106, in execute
    File “/var/lib/ambari-agent/cache/stacks/HDP/2.0.6/services/HIVE/package/scripts/mysql_server.py”, line 30, in install
    File “/var/lib/ambari-agent/cache/stacks/HDP/2.0.6/services/HIVE/package/scripts/mysql_server.py”, line 51, in configure
    mysql_service(daemon_name=params.daemon_name, action=’stop’)
    File “/var/lib/ambari-agent/cache/stacks/HDP/2.0.6/services/HIVE/package/scripts/mysql_service.py”, line 35, in mysql_service
    File “/usr/lib/python2.6/site-packages/resource_management/core/base.py”, line 148, in __init__
    File “/usr/lib/python2.6/site-packages/resource_management/core/environment.py”, line 149, in run
    self.run_action(resource, action)
    File “/usr/lib/python2.6/site-packages/resource_management/core/environment.py”, line 115, in run_action
    File “/usr/lib/python2.6/site-packages/resource_management/core/providers/system.py”, line 235, in action_run
    wait_for_finish=self.resource.wait_for_finish, timeout=self.resource.timeout)
    File “/usr/lib/python2.6/site-packages/resource_management/core/shell.py”, line 35, in checked_call
    return _call(command, logoutput, True, cwd, env, preexec_fn, user, wait_for_finish, timeout)
    File “/usr/lib/python2.6/site-packages/resource_management/core/shell.py”, line 86, in _call
    File “/usr/lib/python2.6/site-packages/resource_management/core/logger.py”, line 35, in info
    File “/usr/lib/python2.6/site-packages/resource_management/core/logger.py”, line 55, in get_protected_text
    text = text.replace(unprotected_string, protected_string)
    UnicodeDecodeError: ‘ascii’ codec can’t decode byte 0xc3 in position 48: ordinal not in range(128)

to create new topics or reply. | New User Registration

  • Author
  • #52353
    Jeff Sposetti

    Do you happen to setting hive password with any special chars?

    Bruno Sarlandie

    No, just a very simple password like “hivepwd”.
    I removed the “!p” at the end of the password field in the command, and it was ok.
    Many problem identical (UnicodeDecodeError) on module logger.py when the logger output the command (info log).
    The linux server configuration (Centos 6) is in french fr_FR.

    Jeff Sposetti

    Thanks for the follow-up info. Can you post the contents of this file?

    $ more /etc/sysconfig/i18n

    And also the output of this command?

    $ locale

    Vincent Gromakowski

    Same error here on French Centos 6.5
    I didn’t use any special character in Hive

    Ben Han

    It s probably due to locales problems when service mysqls stop is called. Here in France it returns “specials” characters Arrêt
    I managed to pass the step by adding LANG=C to the command in /var/lib/ambari-agent/cache/stacks/HDP/2.0.6/services/HIVE/package/scripts/mysql_service.py”, line 35
    Don’t know how to clean solve the problem for the moment.

You must be to reply to this topic. | Create Account

Support from the Experts

A HDP Support Subscription connects you experts with deep experience running Apache Hadoop in production, at-scale on the most demanding workloads.

Enterprise Support »

Become HDP Certified

Real world training designed by the core architects of Hadoop. Scenario-based training courses are available in-classroom or online from anywhere in the world

Training »

Hortonworks Data Platform
The Hortonworks Data Platform is a 100% open source distribution of Apache Hadoop that is truly enterprise grade having been built, tested and hardened with enterprise rigor.
Get started with Sandbox
Hortonworks Sandbox is a self-contained virtual machine with Apache Hadoop pre-configured alongside a set of hands-on, step-by-step Hadoop tutorials.
Modern Data Architecture
Tackle the challenges of big data. Hadoop integrates with existing EDW, RDBMS and MPP systems to deliver lower cost, higher capacity infrastructure.