hadoop-hdfs-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Elek, Marton (JIRA)" <j...@apache.org>
Subject [jira] [Commented] (HDDS-320) Failed to start container with apache/hadoop-runner image.
Date Fri, 19 Oct 2018 15:09:00 GMT

    [ https://issues.apache.org/jira/browse/HDDS-320?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16656911#comment-16656911
] 

Elek, Marton commented on HDDS-320:
-----------------------------------

The base image of smoketest runs as hadoop user with uid=1000. If you have different uid please
make sure that the following directories and files are writable by anybody:

{code}
 /home/haridas/projects/additional/hadoop/hadoop-ozone/dist/target/ozone-0.4.0-SNAPSHOT/etc/hadoop
 /home/haridas/projects/additional/hadoop/hadoop-ozone/dist/target/ozone-0.4.0-SNAPSHOT/etc/hadoop/*
 /home/haridas/projects/additional/hadoop/hadoop-ozone/dist/target/ozone-0.4.0-SNAPSHOT/logs
{code}

(It may be fixed with adjusting the docker-compose, using a .env user_id parameter)

> Failed to start container with apache/hadoop-runner image.
> ----------------------------------------------------------
>
>                 Key: HDDS-320
>                 URL: https://issues.apache.org/jira/browse/HDDS-320
>             Project: Hadoop Distributed Data Store
>          Issue Type: Bug
>          Components: document
>         Environment: centos 7.4
>            Reporter: Junjie Chen
>            Priority: Minor
>
> Following the doc in hadoop-ozone/doc/content/GettingStarted.md, the docker-compose up
-d step failed, the error list list below:
> [root@VM_16_5_centos ozone]# docker-compose logs
> Attaching to ozone_scm_1, ozone_datanode_1, ozone_ozoneManager_1
> datanode_1      | Traceback (most recent call last):
> datanode_1      |   File "/opt/envtoconf.py", line 104, in <module>
> datanode_1      |     Simple(sys.argv[1:]).main()
> datanode_1      |   File "/opt/envtoconf.py", line 93, in main
> datanode_1      |     self.process_envs()
> datanode_1      |   File "/opt/envtoconf.py", line 67, in process_envs
> datanode_1      |     with open(self.destination_file_path(name, extension) + ".raw",
"w") as myfile:
> datanode_1      | IOError: [Errno 13] Permission denied: '/opt/hadoop/etc/hadoop/log4j.properties.raw'
> datanode_1      | Traceback (most recent call last):
> datanode_1      |   File "/opt/envtoconf.py", line 104, in <module>
> datanode_1      |     Simple(sys.argv[1:]).main()
> datanode_1      |   File "/opt/envtoconf.py", line 93, in main
> datanode_1      |     self.process_envs()
> datanode_1      |   File "/opt/envtoconf.py", line 67, in process_envs
> datanode_1      |     with open(self.destination_file_path(name, extension) + ".raw",
"w") as myfile:
> ozoneManager_1  |     with open(self.destination_file_path(name, extension) + ".raw",
"w") as myfile:
> ozoneManager_1  | IOError: [Errno 13] Permission denied: '/opt/hadoop/etc/hadoop/log4j.properties.raw'
> ozoneManager_1  | Traceback (most recent call last):
> ozoneManager_1  |   File "/opt/envtoconf.py", line 104, in <module>
> ozoneManager_1  |     Simple(sys.argv[1:]).main()
> ozoneManager_1  |   File "/opt/envtoconf.py", line 93, in main
> ozoneManager_1  |     self.process_envs()
> ozoneManager_1  |   File "/opt/envtoconf.py", line 67, in process_envs              
                                              
> ozoneManager_1  |     with open(self.destination_file_path(name, extension) + ".raw",
"w") as myfile:                              
> ozoneManager_1  | IOError: [Errno 13] Permission denied: '/opt/hadoop/etc/hadoop/log4j.properties.raw'
                            
> scm_1           | Traceback (most recent call last):
> scm_1           |   File "/opt/envtoconf.py", line 104, in <module>           
                                                    
> scm_1           |     Simple(sys.argv[1:]).main()
> scm_1           |   File "/opt/envtoconf.py", line 93, in main
> scm_1           |     self.process_envs()
> scm_1           |   File "/opt/envtoconf.py", line 67, in process_envs              
                                              
> scm_1           |     with open(self.destination_file_path(name, extension) + ".raw",
"w") as myfile:                              
> scm_1           | IOError: [Errno 13] Permission denied: '/opt/hadoop/etc/hadoop/log4j.properties.raw'
                            
> scm_1           | Traceback (most recent call last):
> scm_1           |   File "/opt/envtoconf.py", line 104, in <module>           
                                                    
> scm_1           |     Simple(sys.argv[1:]).main()
> scm_1           |   File "/opt/envtoconf.py", line 93, in main
> scm_1           |     self.process_envs()
> scm_1           |   File "/opt/envtoconf.py", line 67, in process_envs              
                                              
> scm_1           |     with open(self.destination_file_path(name, extension) + ".raw",
"w") as myfile:                              
> scm_1           | IOError: [Errno 13] Permission denied: '/opt/hadoop/etc/hadoop/log4j.properties.raw'
                            
> scm_1           | Traceback (most recent call last):
> scm_1           |   File "/opt/envtoconf.py", line 104, in <module>           
                                                    
> scm_1           |     Simple(sys.argv[1:]).main()
> scm_1           |   File "/opt/envtoconf.py", line 93, in main
> scm_1           |     self.process_envs()
> scm_1           |   File "/opt/envtoconf.py", line 67, in process_envs              
                                              
> scm_1           |     with open(self.destination_file_path(name, extension) + ".raw",
"w") as myfile:                              
> scm_1           | IOError: [Errno 13] Permission denied: '/opt/hadoop/etc/hadoop/log4j.properties.raw'
  
> my docker-compose version is:
> docker-compose version 1.22.0, build f46880fe
> docker images:
> apache/hadoop-runner   latest              569314fd9a73        5 weeks ago         646MB
> From the Dockerfile, we can see " chown hadoop /opt" command. It looks like we need a
"-R " here?



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

---------------------------------------------------------------------
To unsubscribe, e-mail: hdfs-issues-unsubscribe@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-help@hadoop.apache.org


Mime
View raw message