Date
1 - 2 of 2
Increasing warden yml network and user pool size
Dmitriy Kalinin
This question should be sent to cf-dev mailing list.
On Wed, Jul 1, 2015 at 4:28 PM, Animesh Singh <animation2007(a)gmail.com> wrote: We are seeing some performance bottlenecks at warden, and at times warden |
|
Animesh Singh
We are seeing some performance bottlenecks at warden, and at times warden
drops all connections under increasing load. We think increasing this network and user pool_size might help. We have tried effecting those changes thorugh CF YML, but they arent getting set. Any clues on how can we get this effective? sudo more ./var/vcap/data/jobs/dea_next/a25eb00c949666d87c19508cc917f1601a5c5ba8-136 0a7f1564ff515d5948677293e3aa209712f4f/config/warden.yml --- server: unix_domain_permissions: 0777 unix_domain_path: /var/vcap/data/warden/warden.sock container_klass: Warden::Container::Linux container_rootfs_path: /var/vcap/packages/rootfs_lucid64 container_depot_path: /var/vcap/data/warden/depot container_rlimits: core: 0 pidfile: /var/vcap/sys/run/warden/warden.pid quota: disk_quota_enabled: true logging: file: /var/vcap/sys/log/warden/warden.log level: info syslog: vcap.warden health_check_server: port: 2345 network: pool_start_address: 10.254.0.0 pool_size: 256 # Interface MTU size # (for OpenStack use 1454 to avoid problems with rubygems with GRE tunneling) mtu: 1400 user: pool_start_uid: 20000 pool_size: 256 Thanks, Animesh |
|