Skip to content
This repository has been archived by the owner on Nov 23, 2017. It is now read-only.

Error when using a more up-to-date AMI #80

Open
mmistroni opened this issue Feb 4, 2017 · 3 comments
Open

Error when using a more up-to-date AMI #80

mmistroni opened this issue Feb 4, 2017 · 3 comments

Comments

@mmistroni
Copy link

Hi all
I am trying to launch an ec2 cluster using a more up to date AMI:ami-c928c1a9
Here's my command

root@9f2c58d4fbe6:/spark-ec2# ./spark-ec2 -k ec2AccessKey -i ec2AccessKey.pem -s 2 --ami=ami-c928c1a9 --region us-west-2 launch MMTestCluster4

I am launching this from a docker container running Ubuntu 16.06 and i am getting this exception

Connection to ec2-54-187-145-15.us-west-2.compute.amazonaws.com closed.
Deploying files to master...
Warning: Permanently added 'ec2-54-187-145-15.us-west-2.compute.amazonaws.com,54.187.145.15' (ECDSA) to the list of known hosts.
protocol version mismatch -- is your shell clean?
(see the rsync man page for an explanation)
rsync error: protocol incompatibility (code 2) at compat.c(176) [sender=3.1.1]
Traceback (most recent call last):
File "./spark_ec2.py", line 1534, in
main()
File "./spark_ec2.py", line 1526, in main
real_main()
File "./spark_ec2.py", line 1362, in real_main
setup_cluster(conn, master_nodes, slave_nodes, opts, True)
File "./spark_ec2.py", line 846, in setup_cluster
modules=modules
File "./spark_ec2.py", line 1121, in deploy_files
subprocess.check_call(command)
File "/usr/lib/python2.7/subprocess.py", line 541, in check_call
raise CalledProcessError(retcode, cmd)
subprocess.CalledProcessError: Command '['rsync', '-rv', '-e', 'ssh -o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null -i ec2AccessKey.pem', '/tmp/tmp

I am willing to help sort out this issues as i am skilled in python and i am a user of scala/pyton aws API.
Please give me some hints / starting points, and also if possible a test environment as it's goign to cost me loads of money to keep on creating large instances (and then destroy them) on my aws account

thanks in advance and regards
Marco

@shivaram
Copy link
Contributor

shivaram commented Feb 4, 2017

I think the error is happening due to some ssh output that comes up when we try to run rsync - https://www.centos.org/forums/viewtopic.php?t=53369 seems relevant

Also I think the error is happening when we rsync from the client (i.e. in this case your Ubuntu machine) to the master node. So one shouldn't need a big cluster to debug this - launching 1 master, 1 slave with t1.micro might be enough

@mmistroni
Copy link
Author

mmistroni commented Feb 5, 2017 via email

@milad181
Copy link

milad181 commented Apr 12, 2017

Hello,

I have the same issue when I am using a new AMI. Is there any workaround this?
All I want is to run a cluster of spark using the latest Amazon Linux AMI.

Thanks

Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants