Home > Net >  Namenode HDFS Docker
Namenode HDFS Docker

Time:05-18

I want to run the start-all.sh file, as I have formatted the Namenode in HDFS. I am using it with Docker.

To access it I use:

docker exec -u root -it jupyter-stack /bin/bash

The start-all.sh file is located in the path:

/usr/local/spark-2.4.4-bin-hadoop2.7/sbin/

When I access that path and run it I get the following error:

bash: /usr/local/bin/start-all.sh: Permission denied

How can I fix it and be able to run all services?

CodePudding user response:

Your error refers to a different path than that of the script, so seems you should give the absolute path in the command.

However, running multiple services in the same container is bad practice in Docker.

You should use a different container for the namenode (and Datanode) such as ones distributed by user bde2020 - https://github.com/big-data-europe/docker-hadoop

Or you can use less Docker images by using alternative filesystems, such as MinIO, or simply letting Spark read from its local container filesystem

  • Related