-
Notifications
You must be signed in to change notification settings - Fork 10
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Will you provide the ''job submitting'' scripts (i.e., bash shell)?? #6
Comments
hi, Xiandong Thanks for interesting, I wrote a generator here for my paper: It now can (1) support trace jobs generation( from a trace file) and poisson job generation( from user configurations) (2) monitor cluster information via REST API. I do not have a document for it yet, I will do soon. You can refer conf.default as a starting point as well as some code for your own purpose. For more questions, feel free to contact me. Wei Chen |
Thank you so much. Q1: Will you categorize Big-C into Yarn on Docker, Docker on Yarn or a new type? Maybe Docker on Yarn? Q2: Does each time a new job is submitted, a new docker/container is launched? If so, I am concerned about the cost of launching a docker frequently, 10ms per job? |
Hi, Xiandong For Q1, I would rather say Big-c is a yarn on docker. Since yarn now already supports docker containers, we leverage the docker as well as cgroup in our project. For Q2, It depends on many factors, like your hard drive, image size ... I test one of my images(I build jdk, hadoop and spark in this image) on my server (equipped raid-5 hdd): admin@host7:~$ time docker run -d cwei/hadoop:3.0.0 /bin/bash real 0m0.354s It turns out launching a container takes 0.35s which could be a big problem for a 10ms task. By the way, I would love to know if you are using some tiny workloads, since I am trying to optimize the scheduling delay for YARN. |
Thanks for your illustration.
I understand your idea that compared with a job costing 10 mins, 10ms launch time is negligible.
What about SparkStreaming jobs in Hibench where the task durations are relatively small? |
I understand. I think streaming jobs may not be fit for this situation because the spark executors are allocated at the beginning of the job and the resources will also be kept during its lifetime. Wei |
Hi, This is Xiandong(显东)。
Would you provide the ''job submitting'' scripts of your original experiments? I am preparing a similar project but have no idea how to design a "load generator" as described in your paper.
My Email: [email protected], Wechat: qi839395901.
The text was updated successfully, but these errors were encountered: