ECE Allocator install with new data directory


(Tim Arp) #1

I'm building a ECE cluster at AWS. I'm using the AMIs provided by Elastic.

I'm trying to create my allocator EC2 instances with 4TB of disk. I have done this in cloud formation, with the following parameters.
BlockDeviceMappings:
-
DeviceName: /dev/sdb
Ebs:
VolumeSize: 4000
VolumeType: gp2
DeleteOnTermination: true

In the OS userdata I run the following commands.
"mkfs -t xfs /dev/xvdb\n",
"mkdir /data\n",
"mount /dev/xvdb /data\n",
"chown elastic.elastic /data\n",
"echo '/dev/xvdb /data ext4 defaults,nofail 0 2' >> /etc/fstab\n",

OS looks like this,
[elastic@ip-10-212-98-233 mapper]$ df -h
Filesystem Size Used Avail Use% Mounted on
devtmpfs 16G 0 16G 0% /dev
tmpfs 16G 0 16G 0% /dev/shm
tmpfs 16G 8.3M 16G 1% /run
tmpfs 16G 0 16G 0% /sys/fs/cgroup
/dev/xvda3 19G 1.7G 17G 10% /
/dev/xvda1 1014M 209M 806M 21% /boot
/dev/mapper/lxc-data 19G 1.1G 18G 6% /mnt/data
/dev/xvdb 4.0T 33M 4.0T 1% /data
tmpfs 3.2G 0 3.2G 0% /run/user/1000

I try to install ECE with the following command
./elastic-cloud-enterprise-installer.sh --coordinator-host 10.212.97.168 --host-storage-path /data --debug

This errors out, with nothing useful, debug output is below.
-- Debug Info --
NODE_TYPE - additional
REPOSITORY - docker.elastic.co/cloud-enterprise
RUNNER_REGION - ece-region
RUNNER_HOST_IP - 10.212.98.233
ELASTIC_UID - 1000
NODE_TYPE - additional
RUNNER_AVAILABILITY_ZONE - ece-region-1b
HOST_STORAGE_ROOT_VOLUME_PATH - /
MEMORY_SETTINGS - {}
COORDINATOR_HOST - 10.212.97.168
RUNNER_ID - 10.212.98.233
HOST_STORAGE_PATH - /data
RUNNER_CAPACITY - 26971
CLOUD_ENTERPRISE_VERSION - 1.0.0-beta2
HOST_STORAGE_DEVICE_PATH - /dev/mapper/lxc-data
ELASTIC_GID - 1000

-- Verifying Prerequisites --
Checking host storage path... PASSED
Checking docker version... PASSED
Checking internal ip connectivity... PASSED
Checking coordinator connectivity... PASSED
Checking OS settings... PASSED
Checking OS kernel version... PASSED
Checking Minimum required memory... PASSED
-- Completed Verifying Prerequisites --

  • Running Bootstrap container
  • Monitoring bootstrap process
Errors have caused Elastic Cloud Enterprise installation to fail

I believe the problem is how I'm creating the filesystem or mounting it.  The install worked fine before I started trying to add the large filesystem.  Am I doing this the proper way?  Can someone help?

(Christian Strzadala) #2

Hi @tarp,

Does this install fail constantly? When it fails, can you run the following commands to try and get some more info out of the issue.

  • docker ps -a
  • docker logs elastic-cloud-enterprise-bootstrap-1.0.0-beta2

If you can give us the output of those two commands, that would help us diagnose the issue here.

Thanks,

Christian


(Tim Arp) #3

Hi Christian,

I'm out of the office for about a week, so I can't test this right now.

Can you look at my filesystem creation process and tell me if this is sound? Is there a better way to be doing this with EC2 at AWS using the provided AMIs?

Thanks,
Tim


(Tim Arp) #4

Hi Christian,

Sorry for getting back to you so late, I was on vacation. Yes this happens everytime I try to build the allocator with the larger disk.

Output you requested is here.
[root@ip-10-212-101-176 ~]# docker ps -a
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
d3f0aa5746ae docker.elastic.co/cloud-enterprise/elastic-cloud-enterprise:1.0.0-beta2 "/sbin/my_init" 17 minutes ago Exited (0) 17 minutes ago elastic-cloud-enterprise-bootstrap-1.0.0-beta2

*** /elastic_cloud_apps/bootstrap/bootstrap-additional exited with status 1.
*** Killing all processes...
*** Runit exited with status 0
fail: /etc/service/docker-socket-proxy: runsv not running
fail: /etc/service/entry-point: runsv not running
*** Killing all processes...


(Tim Arp) #5

Hi,
I have corrected this issue with the following steps
shutdown docker service
I copied the existing /mnt/data/* to /data
updated permissions
remounted /data as /mnt/data
updated the /etc/fstab
remount
startup docker

This gave me my 4TB partition and I didn't have to run the installer with any special data flags.


(system) #6

This topic was automatically closed 14 days after the last reply. New replies are no longer allowed.