Hey Kyle,
Can we take a look at your deployment manifest (with all the secrets redacted)?
Zak + Dan, CF OSS Integration team
toggle quoted messageShow quoted text
On Fri, Sep 11, 2015 at 8:55 AM, kyle havlovitz <kylehav(a)gmail.com> wrote: I'm getting an error pushing any app during the staging step. The cf logs returns only this:
2015-09-11T15:24:24.33+0000 [API] OUT Created app with guid
47efa472-e2d5-400b-9135-b1a1dbe3ba88 2015-09-11T15:24:24.41+0000 [API] OUT Updated app with guid
47efa472-e2d5-400b-9135-b1a1dbe3ba88 ({"route"=>"5737c5f5-b017-43da-9013-2b6fe7db03f7"}) 2015-09-11T15:24:29.54+0000 [DEA/0] OUT Got staging request for app
with id 47efa472-e2d5-400b-9135-b1a1dbe3ba88 2015-09-11T15:24:30.71+0000 [API] OUT Updated app with guid
47efa472-e2d5-400b-9135-b1a1dbe3ba88 ({"state"=>"STARTED"}) 2015-09-11T15:24:30.76+0000 [STG/0] OUT -----> Downloaded app
package (4.0K) 2015-09-11T15:25:06.00+0000 [API] ERR encountered error: Staging
error: failed to stage application: 2015-09-11T15:25:06.00+0000 [API] ERR Script exited with status 1
In the warden logs, there are a few suspect messages:
{ "timestamp": 1441985105.8883495,
"message": "Exited with status 1 (35.120s):
[[\"/opt/cloudfoundry/warden/warden/src/closefds/closefds\", \"/opt/cloudfoundry/warden/warden/src/closefds/closefds\"], \"/var/warden/containers/18vf956il5v/bin/iomux-link\", \"-w\", \"/var/warden/containers/18vf956il5v/jobs/8/cursors\", \"/var/warden/containers/18vf956il5v/jobs/8\"]", "log_level": "warn",
"source": "Warden::Container::Linux",
"data": {
"handle": "18vf956il5v",
"stdout": "",
"stderr": ""
},
"thread_id": 69890836968240,
"fiber_id": 69890849112480,
"process_id": 17063,
"file":
"/opt/cloudfoundry/warden/warden/lib/warden/container/spawn.rb", "lineno": 135,
"method": "set_deferred_success"
}
{
"timestamp": 1441985105.94083,
"message": "Exited with status 23 (0.023s):
[[\"/opt/cloudfoundry/warden/warden/src/closefds/closefds\", \"/opt/cloudfoundry/warden/warden/src/closefds/closefds\"], \"rsync\", \"-e\", \"/var/warden/containers/18vf956il5v/bin/wsh --socket /var/warden/containers/18vf956il5v/run/wshd.sock --rsh\", \"-r\", \"-p\", \"--links\", \"vcap(a)container:/tmp/staged/staging_info.yml\", \"/tmp/dea_ng/staging/d20150911-17093-1amg6y8\"]", "log_level": "warn",
"source": "Warden::Container::Linux",
"data": {
"handle": "18vf956il5v",
"stdout": "",
"stderr": "rsync: link_stat \"/tmp/staged/staging_info.yml\"
failed: No such file or directory (2)\nrsync error: some files/attrs were not transferred (see previous errors) (code 23) at main.c(1655) [Receiver=3.1.0]\nrsync: [Receiver] write error: Broken pipe (32)\n" },
"thread_id": 69890836968240,
"fiber_id": 69890849112480,
"process_id": 17063,
"file":
"/opt/cloudfoundry/warden/warden/lib/warden/container/spawn.rb", "lineno": 135,
"method": "set_deferred_success"
}
{
"timestamp": 1441985106.0086887,
"message": "Killing oom-notifier process",
"log_level": "debug",
"source": "Warden::Container::Features::MemLimit::OomNotifier",
"data": {
},
"thread_id": 69890836968240,
"fiber_id": 69890848620580,
"process_id": 17063,
"file":
"/opt/cloudfoundry/warden/warden/lib/warden/container/features/mem_limit.rb", "lineno": 51,
"method": "kill"
}
{
"timestamp": 1441985106.0095143,
"message": "Exited with status 0 (35.427s):
[[\"/opt/cloudfoundry/warden/warden/src/closefds/closefds\", \"/opt/cloudfoundry/warden/warden/src/closefds/closefds\"], \"/opt/cloudfoundry/warden/warden/src/oom/oom\", \"/tmp/warden/cgroup/memory/instance-18vf956il5v\"]", "log_level": "warn",
"source": "Warden::Container::Features::MemLimit::OomNotifier",
"data": {
"stdout": "",
"stderr": ""
},
"thread_id": 69890836968240,
"fiber_id": 69890849112480,
"process_id": 17063,
"file":
"/opt/cloudfoundry/warden/warden/lib/warden/container/spawn.rb", "lineno": 135,
"method": "set_deferred_success"
}
Obviously something is misconfigured, but I'm not sure what. I don't know why the out of memory thing is appearing as the memory being used by the test app i've pushed is tiny (64M app with staticfile buildpack) and the dea config has resource.memory_mb set to 8 gigs and staging.memory_limit_mb set to 1 gigs. Is there some config I'm lacking that's causing this to fail?
|