cf-release v213 fixed a bug where buildpack caches were not getting cleaned
up properly. Are you running a version prior to 213?
If so, you might need to go into the
`shared/run.10.10.2.37.xip.io-cc-droplets` directory and delete the
`buildpack_cache` subdirectory. Deleting the contents of buildpack_cache
will not cause failures in the system, the caches will simply get generated
from scratch the next time apps stage.
If you upgrade to 213 or later, you can just call the api endpoint to clean
out the buildpack cache. You can find the details in the v213 release notes:https://github.com/cloudfoundry/cf-release/releases/tag/v213
Joseph & Natalie
CF Release Integration Team
On Fri, Sep 25, 2015 at 11:53 AM, Matthias Ender <Matthias.Ender(a)sas.com>
ah, it’s not postgres:
Does it keep a copy of each artifact ever pushed?
Or is this part of the event storage, controlled by these properties:
*From:* Aleksey Zalesov [mailto:aleksey.zalesov(a)altoros.com]
*Sent:* Friday, September 25, 2015 12:23 PM
*To:* Discussions about Cloud Foundry projects and the system overall. <
*Subject:* [cf-dev] Re: postgres out of disk space
Does your database occupy this 100 GB? Or something else like logs?
Aleksey Zalesov | CloudFoundry Engineer | Altoros
Tel: (617) 841-2121 ext. 5707 | Toll free: 855-ALTOROS
Fax: (866) 201-3646 | Skype: aleksey_zalesov
www.altoros.com | blog.altoros.com | twitter.com/altoros
On 25 Sep 2015, at 14:22, Matthias Ender <Matthias.Ender(a)sas.com> wrote:
I have a cf-aws-tiny cf-boshrelease deployment, and it’s been running well
for over 4 months.
We have about 40 apps, with a couple of dozen of cf pushes each day.
Yesterday pushing apps became spotty and then impossible, with various
Turned out the 100GB disk for the postgres instance on the data note was
I increased the disk size and things a running again.
But – what happened there? 100G and growing seems like awfully large
database for a rather modest use.
And I’m worried it’ll just happen again in a few months.