Date   

Re: Moving Diego Repositories

Kris Hicks <khicks@...>
 

I've found when doing this the easiest thing to do is to use sed to remove
existing imports followed by goimports -w on the same files.

KH

On Sunday, June 26, 2016, Amit Gupta <agupta(a)pivotal.io> wrote:

Congrats on making it official!

I assume you folks are going to build some scripts or some "go fmt -r" to
find/change all import paths that need updating. If you find anything
useful that other teams might be able to leverage, sharing would be much
appreciated.

Cheers,
Amit

On Fri, Jun 24, 2016 at 4:20 PM, Onsi Fakhouri <ofakhouri(a)pivotal.io
<javascript:_e(%7B%7D,'cvml','ofakhouri(a)pivotal.io');>> wrote:

That stuff has always technically been foundation code and should have
moved out of pivotal-golang into cloudfoundry-incubator long ago. Back
when we made the org the rules about how orgs are structured weren't quite
so clear.

Onsi

On Fri, Jun 24, 2016 at 4:18 PM, Alex Suraci <asuraci(a)pivotal.io
<javascript:_e(%7B%7D,'cvml','asuraci(a)pivotal.io');>> wrote:

why are we moving the pivotal-golang repos? I thought the point of that
was to be able to create those cheaply for super generic packages that we
just need to have somewhere in cases where the repo incubation foundation
lifecycle crap gets in the way. are we now able to create repos willy-nilly
under the cloudfoundry org?

On Sat, Jun 25, 2016, 6:13 AM James Myers <jmyers(a)pivotal.io
<javascript:_e(%7B%7D,'cvml','jmyers(a)pivotal.io');>> wrote:

Hi all,

We are currently in the process of moving all of the Diego repositories
from `cloudfoundry-incubator` to the `cloudfoundry` organization.

Specifically we are moving the following repositories to the
`cloudfoundry` organization:

- auction
- auctioneer
- bbs
- benchmark-bbs
- buildpack_app_lifecycle
- cacheddownloader
- cf-debug-server
- cf-lager
- cf_http
- consuladapter
- converger
- diego-ssh
- diego-upgrade-stability-tests
- docker_app_lifecycle
- executor
- file-server
- healthcheck
- inigo
- locket
- rep
- route-emitter
- runtime-schema
- vizzini
- diego-cf-compatibility
- diego-ci-pools
- diego-ci
- diego-design-notes
- diego-dev-notes
- diego-dockerfiles
- diego-perf-release
- diego-release
- diego-stress-tests

We are also moving the following from `pivotal-golang` to
`cloudfoundry` as well:

- archiver
- bytefmt
- clock
- eventhub
- lager
- localip
- operationq

We are also renaming the following, and will be updating their package
names accordingly:

- benchmark-bbs -> benchmarkbbs
- buildpack_app_lifecycle -> buildpackapplifecycle
- cf-debug-server -> debugserver
- cf-lager -> cflager
- cf_http -> cfhttp
- docker_app_lifecycle -> dockerapplifecycle
- file-server -> fileserver
- runtime-schema -> runtimeschema

You might be asking yourself, what does this mean for me?

Generally it means the following:

- If you are importing any of the above repos in your golang code, you
should change it from `github.com/cloudfoundry-incubator/REPO_NAME`
<http://github.com/cloudfoundry-incubator/REPO_NAME> to `
code.cloudfoundry.org/REPO_NAME`
<http://code.cloudfoundry.org/REPO_NAME>.

- Update your golang code when you update your dependencies to
reference the new package names marked above.

- If you are consuming the Diego bosh release from bosh.io, you will
need to update the location to
http://bosh.io/releases/github.com/cloudfoundry/diego-release.


Other than that, github redirects should handle most of the issues for
you.

As a side note we are also moving the following deprecated repositories
to the `cloudfoundry-attic`:

- diego-acceptance-tests
- diego-smoke-tests
- receptor
- runtime-metrics-server

Let us know if you have any questions.

Best,

Jim + Andrew, CF Diego Team


Re: Moving Diego Repositories

Amit Kumar Gupta
 

Congrats on making it official!

I assume you folks are going to build some scripts or some "go fmt -r" to
find/change all import paths that need updating. If you find anything
useful that other teams might be able to leverage, sharing would be much
appreciated.

Cheers,
Amit

On Fri, Jun 24, 2016 at 4:20 PM, Onsi Fakhouri <ofakhouri(a)pivotal.io> wrote:

That stuff has always technically been foundation code and should have
moved out of pivotal-golang into cloudfoundry-incubator long ago. Back
when we made the org the rules about how orgs are structured weren't quite
so clear.

Onsi

On Fri, Jun 24, 2016 at 4:18 PM, Alex Suraci <asuraci(a)pivotal.io> wrote:

why are we moving the pivotal-golang repos? I thought the point of that
was to be able to create those cheaply for super generic packages that we
just need to have somewhere in cases where the repo incubation foundation
lifecycle crap gets in the way. are we now able to create repos willy-nilly
under the cloudfoundry org?

On Sat, Jun 25, 2016, 6:13 AM James Myers <jmyers(a)pivotal.io> wrote:

Hi all,

We are currently in the process of moving all of the Diego repositories
from `cloudfoundry-incubator` to the `cloudfoundry` organization.

Specifically we are moving the following repositories to the
`cloudfoundry` organization:

- auction
- auctioneer
- bbs
- benchmark-bbs
- buildpack_app_lifecycle
- cacheddownloader
- cf-debug-server
- cf-lager
- cf_http
- consuladapter
- converger
- diego-ssh
- diego-upgrade-stability-tests
- docker_app_lifecycle
- executor
- file-server
- healthcheck
- inigo
- locket
- rep
- route-emitter
- runtime-schema
- vizzini
- diego-cf-compatibility
- diego-ci-pools
- diego-ci
- diego-design-notes
- diego-dev-notes
- diego-dockerfiles
- diego-perf-release
- diego-release
- diego-stress-tests

We are also moving the following from `pivotal-golang` to `cloudfoundry`
as well:

- archiver
- bytefmt
- clock
- eventhub
- lager
- localip
- operationq

We are also renaming the following, and will be updating their package
names accordingly:

- benchmark-bbs -> benchmarkbbs
- buildpack_app_lifecycle -> buildpackapplifecycle
- cf-debug-server -> debugserver
- cf-lager -> cflager
- cf_http -> cfhttp
- docker_app_lifecycle -> dockerapplifecycle
- file-server -> fileserver
- runtime-schema -> runtimeschema

You might be asking yourself, what does this mean for me?

Generally it means the following:

- If you are importing any of the above repos in your golang code, you
should change it from `github.com/cloudfoundry-incubator/REPO_NAME`
<http://github.com/cloudfoundry-incubator/REPO_NAME> to `
code.cloudfoundry.org/REPO_NAME`
<http://code.cloudfoundry.org/REPO_NAME>.

- Update your golang code when you update your dependencies to reference
the new package names marked above.

- If you are consuming the Diego bosh release from bosh.io, you will
need to update the location to
http://bosh.io/releases/github.com/cloudfoundry/diego-release.


Other than that, github redirects should handle most of the issues for
you.

As a side note we are also moving the following deprecated repositories
to the `cloudfoundry-attic`:

- diego-acceptance-tests
- diego-smoke-tests
- receptor
- runtime-metrics-server

Let us know if you have any questions.

Best,

Jim + Andrew, CF Diego Team


Re: Creating new user UAA

sridhar vennela
 

Excellent!!!. Good luck.


Re: Retrieve __VCAP__ID from instance_ID

Vinod A
 

Installed the sample app and tried it on Bluemix and I don't see CF_INSTANCE_GUID or INSTANCE_GUID in Environment section. How do I know if its Diego or DEA ?.

Thanks,
Vinod


Re: How shoulld I debug a blobstore error?

Ronak Banka
 

You can use xip.io for temporary domain, try generating manifest again and
deploy.

On Saturday, 25 June 2016, Eyal Shalev <eshalev(a)cisco.com> wrote:

As I was not planning on sigining exposing this deployment to the outside,
I was planning on using HAPROXY through it's floating IP.

I did not set a domain, because I have not yet interfaced with my DNS
server. So I just left the domain as "DOMAIN"

When looking at manifest generated from the stuff, i see the following 2
lines repeating sveral times:
private_endpoint: https://blobstore.service.cf.internal
public_endpoint: http://blobstore.DOMAIN

Could it be that the api-node is trying to access blobstore through the
haproxy node?

Is there anyway to avoid setting up a domain for CF? (as this is just a
lab experiment)


Re: Creating new user UAA

Bryan Perino
 

That was the problem. I removed the MySQL profile and it worked. Thanks!


Re: Moving Diego Repositories

Onsi Fakhouri <ofakhouri@...>
 

That stuff has always technically been foundation code and should have
moved out of pivotal-golang into cloudfoundry-incubator long ago. Back
when we made the org the rules about how orgs are structured weren't quite
so clear.

Onsi

On Fri, Jun 24, 2016 at 4:18 PM, Alex Suraci <asuraci(a)pivotal.io> wrote:

why are we moving the pivotal-golang repos? I thought the point of that
was to be able to create those cheaply for super generic packages that we
just need to have somewhere in cases where the repo incubation foundation
lifecycle crap gets in the way. are we now able to create repos willy-nilly
under the cloudfoundry org?

On Sat, Jun 25, 2016, 6:13 AM James Myers <jmyers(a)pivotal.io> wrote:

Hi all,

We are currently in the process of moving all of the Diego repositories
from `cloudfoundry-incubator` to the `cloudfoundry` organization.

Specifically we are moving the following repositories to the
`cloudfoundry` organization:

- auction
- auctioneer
- bbs
- benchmark-bbs
- buildpack_app_lifecycle
- cacheddownloader
- cf-debug-server
- cf-lager
- cf_http
- consuladapter
- converger
- diego-ssh
- diego-upgrade-stability-tests
- docker_app_lifecycle
- executor
- file-server
- healthcheck
- inigo
- locket
- rep
- route-emitter
- runtime-schema
- vizzini
- diego-cf-compatibility
- diego-ci-pools
- diego-ci
- diego-design-notes
- diego-dev-notes
- diego-dockerfiles
- diego-perf-release
- diego-release
- diego-stress-tests

We are also moving the following from `pivotal-golang` to `cloudfoundry`
as well:

- archiver
- bytefmt
- clock
- eventhub
- lager
- localip
- operationq

We are also renaming the following, and will be updating their package
names accordingly:

- benchmark-bbs -> benchmarkbbs
- buildpack_app_lifecycle -> buildpackapplifecycle
- cf-debug-server -> debugserver
- cf-lager -> cflager
- cf_http -> cfhttp
- docker_app_lifecycle -> dockerapplifecycle
- file-server -> fileserver
- runtime-schema -> runtimeschema

You might be asking yourself, what does this mean for me?

Generally it means the following:

- If you are importing any of the above repos in your golang code, you
should change it from `github.com/cloudfoundry-incubator/REPO_NAME`
<http://github.com/cloudfoundry-incubator/REPO_NAME> to `
code.cloudfoundry.org/REPO_NAME` <http://code.cloudfoundry.org/REPO_NAME>
.

- Update your golang code when you update your dependencies to reference
the new package names marked above.

- If you are consuming the Diego bosh release from bosh.io, you will
need to update the location to
http://bosh.io/releases/github.com/cloudfoundry/diego-release.


Other than that, github redirects should handle most of the issues for
you.

As a side note we are also moving the following deprecated repositories
to the `cloudfoundry-attic`:

- diego-acceptance-tests
- diego-smoke-tests
- receptor
- runtime-metrics-server

Let us know if you have any questions.

Best,

Jim + Andrew, CF Diego Team


Re: How shoulld I debug a blobstore error?

Eyal Shalev
 

As I was not planning on sigining exposing this deployment to the outside, I was planning on using HAPROXY through it's floating IP.

I did not set a domain, because I have not yet interfaced with my DNS server. So I just left the domain as "DOMAIN"

When looking at manifest generated from the stuff, i see the following 2 lines repeating sveral times:
private_endpoint: https://blobstore.service.cf.internal
public_endpoint: http://blobstore.DOMAIN

Could it be that the api-node is trying to access blobstore through the haproxy node?

Is there anyway to avoid setting up a domain for CF? (as this is just a lab experiment)


Re: How shoulld I debug a blobstore error?

Ronak Banka
 

Hello Eyal,

Can you paste the domain properties from your final manifest ( system
domain & app domain part).

Thanks
Ronak

On Saturday, 25 June 2016, Eyal Shalev <eshalev(a)cisco.com> wrote:

Hello Amit,
I have removed the lines that you have marked.

Now I am getting a different error...
Process 'consul_agent' running
Process 'cloud_controller_ng' Connection failed
Process 'cloud_controller_worker_local_1' not monitored
Process 'cloud_controller_worker_local_2' not monitored
Process 'nginx_cc' initializing
Process 'cloud_controller_migration' running
Process 'metron_agent' running
Process 'statsd-injector' running
Process 'route_registrar' running
System 'system_localhost' running


The blob store is available, but still the process fails:
[2016-06-24 20:13:17+0000] ------------ STARTING
cloud_controller_worker_ctl at Fri Jun 24 20:13:17 UTC 2016 --------------
[2016-06-24 20:13:17+0000] Removing stale pidfile
[2016-06-24 20:13:17+0000] Checking for blobstore availability
[2016-06-24 20:13:17+0000] Blobstore is available
[2016-06-24 20:13:18+0000] Buildpacks installation failed


and also:
[2016-06-24 20:33:16+0000] ------------ STARTING cloud_controller_ng_ctl
at Fri Jun 24 20:33:16 UTC 2016 --------------
[2016-06-24 20:33:16+0000] Checking for blobstore availability
[2016-06-24 20:33:16+0000] Blobstore is available
[2016-06-24 20:33:38+0000] Killing
/var/vcap/sys/run/cloud_controller_ng/cloud_controller_ng.pid: 28368
[2016-06-24 20:33:38+0000] Stopped
[2016-06-24 20:33:39+0000] ------------ STARTING cloud_controller_ng_ctl
at Fri Jun 24 20:33:39 UTC 2016 --------------
[2016-06-24 20:33:39+0000] Checking for blobstore availability
[2016-06-24 20:33:39+0000] Blobstore is available
[2016-06-24 20:34:02+0000] Killing
/var/vcap/sys/run/cloud_controller_ng/cloud_controller_ng.pid: 28818
[2016-06-24 20:34:03+0000] Stopped

Which brings me to another question:
Do you have a stable old release of CF for openstack? I don't mind
downgrading, if the new releases are unstable. If that is not possible, can
you post a valid cf-stub.yml without the need for any manual removal of
invalid lines? (that way I have a reference to what tried and tested stub
should look like)

Thanks alot for your help,
Eyal


Re: Retrieve __VCAP__ID from instance_ID

Daniel Mikusa
 

On Fri, Jun 24, 2016 at 3:01 PM, Vinod A <vin.app(a)gmail.com> wrote:

ok thanks, will wait for someone to reply. Do you know if the behavior
should be same or has it been changed from DEA to Diego ?.
I know there are some changes to the env variables between DEAs and Diego.
I don't know exactly and don't have a DEA to compare.

https://github.com/cloudfoundry/diego-design-notes/blob/master/migrating-to-diego.md#cf-specific-environment-variables

If you have a DEA, you could push a this test app [1], it dumps the
environment variables, and compare to the app I have running on PWS [2].

[1] - https://github.com/dmikusa-pivotal/cf-ex-php-info
[2] - https://php-info.cfapps.io/info.php

Dan




Thanks,
Vinod


Moving Diego Repositories

James Myers
 

Hi all,

We are currently in the process of moving all of the Diego repositories
from `cloudfoundry-incubator` to the `cloudfoundry` organization.

Specifically we are moving the following repositories to the `cloudfoundry`
organization:

- auction
- auctioneer
- bbs
- benchmark-bbs
- buildpack_app_lifecycle
- cacheddownloader
- cf-debug-server
- cf-lager
- cf_http
- consuladapter
- converger
- diego-ssh
- diego-upgrade-stability-tests
- docker_app_lifecycle
- executor
- file-server
- healthcheck
- inigo
- locket
- rep
- route-emitter
- runtime-schema
- vizzini
- diego-cf-compatibility
- diego-ci-pools
- diego-ci
- diego-design-notes
- diego-dev-notes
- diego-dockerfiles
- diego-perf-release
- diego-release
- diego-stress-tests

We are also moving the following from `pivotal-golang` to `cloudfoundry` as
well:

- archiver
- bytefmt
- clock
- eventhub
- lager
- localip
- operationq

We are also renaming the following, and will be updating their package
names accordingly:

- benchmark-bbs -> benchmarkbbs
- buildpack_app_lifecycle -> buildpackapplifecycle
- cf-debug-server -> debugserver
- cf-lager -> cflager
- cf_http -> cfhttp
- docker_app_lifecycle -> dockerapplifecycle
- file-server -> fileserver
- runtime-schema -> runtimeschema

You might be asking yourself, what does this mean for me?

Generally it means the following:

- If you are importing any of the above repos in your golang code, you
should change it from `github.com/cloudfoundry-incubator/REPO_NAME` to `
code.cloudfoundry.org/REPO_NAME`.

- Update your golang code when you update your dependencies to reference
the new package names marked above.

- If you are consuming the Diego bosh release from bosh.io, you will need
to update the location to
http://bosh.io/releases/github.com/cloudfoundry/diego-release.


Other than that, github redirects should handle most of the issues for you.

As a side note we are also moving the following deprecated repositories to
the `cloudfoundry-attic`:

- diego-acceptance-tests
- diego-smoke-tests
- receptor
- runtime-metrics-server

Let us know if you have any questions.

Best,

Jim + Andrew, CF Diego Team


Re: How shoulld I debug a blobstore error?

Eyal Shalev
 

Hello Amit,
I have removed the lines that you have marked.

Now I am getting a different error...
Process 'consul_agent' running
Process 'cloud_controller_ng' Connection failed
Process 'cloud_controller_worker_local_1' not monitored
Process 'cloud_controller_worker_local_2' not monitored
Process 'nginx_cc' initializing
Process 'cloud_controller_migration' running
Process 'metron_agent' running
Process 'statsd-injector' running
Process 'route_registrar' running
System 'system_localhost' running


The blob store is available, but still the process fails:
[2016-06-24 20:13:17+0000] ------------ STARTING cloud_controller_worker_ctl at Fri Jun 24 20:13:17 UTC 2016 --------------
[2016-06-24 20:13:17+0000] Removing stale pidfile
[2016-06-24 20:13:17+0000] Checking for blobstore availability
[2016-06-24 20:13:17+0000] Blobstore is available
[2016-06-24 20:13:18+0000] Buildpacks installation failed


and also:
[2016-06-24 20:33:16+0000] ------------ STARTING cloud_controller_ng_ctl at Fri Jun 24 20:33:16 UTC 2016 --------------
[2016-06-24 20:33:16+0000] Checking for blobstore availability
[2016-06-24 20:33:16+0000] Blobstore is available
[2016-06-24 20:33:38+0000] Killing /var/vcap/sys/run/cloud_controller_ng/cloud_controller_ng.pid: 28368
[2016-06-24 20:33:38+0000] Stopped
[2016-06-24 20:33:39+0000] ------------ STARTING cloud_controller_ng_ctl at Fri Jun 24 20:33:39 UTC 2016 --------------
[2016-06-24 20:33:39+0000] Checking for blobstore availability
[2016-06-24 20:33:39+0000] Blobstore is available
[2016-06-24 20:34:02+0000] Killing /var/vcap/sys/run/cloud_controller_ng/cloud_controller_ng.pid: 28818
[2016-06-24 20:34:03+0000] Stopped

Which brings me to another question:
Do you have a stable old release of CF for openstack? I don't mind downgrading, if the new releases are unstable. If that is not possible, can you post a valid cf-stub.yml without the need for any manual removal of invalid lines? (that way I have a reference to what tried and tested stub should look like)

Thanks alot for your help,
Eyal


Re: Retrieve __VCAP__ID from instance_ID

Vinod A
 

ok thanks, will wait for someone to reply. Do you know if the behavior should be same or has it been changed from DEA to Diego ?.

Thanks,
Vinod


Re: How shoulld I debug a blobstore error?

Amit Kumar Gupta
 

Yes sir, please try removing them and regenerating your manifest before
redeploying.

Cheers,
Amit

On Friday, June 24, 2016, Eyal Shalev <eshalev(a)cisco.com> wrote:

Hello Amit,
I used the latest stub from GIT. As I was working last week, I could swear
I saw the stub change as I was working...
I guess that that explains it.
Should I remove the lines in the highlighted block?

Thanks,
Eyal


Re: How shoulld I debug a blobstore error?

Eyal Shalev
 

Hello Amit,
I used the latest stub from GIT. As I was working last week, I could swear I saw the stub change as I was working...
I guess that that explains it.
Should I remove the lines in the highlighted block?

Thanks,
Eyal


Re: Creating new user UAA

sridhar vennela
 

Hi Bryan,

UAA needs a database to store user accounts. By default it uses in-memory db, In your case, it is trying to connect MySQL DB. I am not sure. How are you running you UAA server? Are you using below command? If yes, you need to have MySQL running on your local machine

./gradlew -Dspring.profiles.active=default,mysql run


[2016-06-24 02:30:56.279] cloudfoundry-identity-server/uaa - ???? [localhost-startStop-1] .... WARN --- XmlWebApplicationContext: Caused by: java.net.ConnectException: Connection refused
[2016-06-24 02:30:56.279] cloudfoundry-identity-server/uaa - ???? [localhost-startStop-1] .... WARN --- XmlWebApplicationContext: at java.net.PlainSocketImpl.socketConnect(Native Method)
[2016-06-24 02:30:56.279] cloudfoundry-identity-server/uaa - ???? [localhost-startStop-1] .... WARN --- XmlWebApplicationContext: at java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:350)
[2016-06-24 02:30:56.279] cloudfoundry-identity-server/uaa - ???? [localhost-startStop-1] .... WARN --- XmlWebApplicationContext: at java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:206)
[2016-06-24 02:30:56.279] cloudfoundry-identity-server/uaa - ???? [localhost-startStop-1] .... WARN --- XmlWebApplicationContext: at java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:188)
[2016-06-24 02:30:56.279] cloudfoundry-identity-server/uaa - ???? [localhost-startStop-1] .... WARN --- XmlWebApplicationContext: at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:392)
[2016-06-24 02:30:56.279] cloudfoundry-identity-server/uaa - ???? [localhost-startStop-1] .... WARN --- XmlWebApplicationContext: at java.net.Socket.connect(Socket.java:589)
[2016-06-24 02:30:56.279] cloudfoundry-identity-server/uaa - ???? [localhost-startStop-1] .... WARN --- XmlWebApplicationContext: at java.net.Socket.connect(Socket.java:538)
[2016-06-24 02:30:56.279] cloudfoundry-identity-server/uaa - ???? [localhost-startStop-1] .... WARN --- XmlWebApplicationContext: at org.mariadb.jdbc.internal.mysql.MySQLProtocol.connect(MySQLProtocol.java:387)
[2016-06-24 02:30:56.279] cloudfoundry-identity-server/uaa - ???? [localhost-startStop-1] .... WARN --- XmlWebApplicationContext: at org.mariadb.jdbc.internal.mysql.MySQLProtocol.connect(MySQLProtocol.java:709)
[2016-06-24 02:30:56.279] cloudfoundry-identity-server/uaa - ???? [localhost-startStop-1] .... WARN --- XmlWebApplicationContext: ... 89 more


Re: How shoulld I debug a blobstore error?

Amit Kumar Gupta
 

Those stubs are out of date. Our OpenStack stubs may remain out of date
for a bit while we find more resource to support automated testing of the
stubs against and OpenStack environment. In the mean time, you should be
fine to just delete those offending lines. Please try that and let me know
if it works.

Best,
Amit

On Fri, Jun 24, 2016 at 10:40 AM, Pramod Mandagere <nagapramod(a)gmail.com>
wrote:

Hi Amit,
We have the exact same issue and our cf-stub looking exactly like the one
u pointed out. Stuck with this issue,
Regards,
Pramod

On Fri, Jun 24, 2016 at 10:26 AM Amit Gupta <agupta(a)pivotal.io> wrote:

Hi Eyal,

No need to tear things down. By any chance, did you generate your
manifest using a stub? Do you have something that looks like this in your
manifest?


<https://github.com/cloudfoundry/cf-release/blob/3290ed8/spec/fixtures/openstack/cf-stub.yml#L155-L187>
https://github.com/cloudfoundry/cf-release/blob/3290ed8/spec/fixtures/openstack/cf-stub.yml#L155-L187

Best,
Amit

On Fri, Jun 24, 2016 at 6:54 AM, Eyal Shalev <eshalev(a)cisco.com> wrote:

Hello, I am trying to deploy cloud foundry on openstack.

running "bosh deploy" gives me the following error:
Started updating job consul_z1 > consul_z1/0
(d248a6da-476a-4fb1-bc08-01ce23c94eb4) (canary). Done (00:01:03)
Started updating job ha_proxy_z1 > ha_proxy_z1/0
(f416b7c5-24d3-4d58-bbf9-066740971065) (canary). Done (00:00:43)
Started updating job nats_z1 > nats_z1/0
(074d00e8-96b3-4b7c-af1a-347549ff51dd) (canary). Done (00:00:44)
Started updating job etcd_z1 > etcd_z1/0
(ef27b75d-e0d0-4a95-8ec1-4b0c9371fa93) (canary). Done (00:01:23)
Started updating job stats_z1 > stats_z1/0
(20c1f5e9-3b7a-4c84-a940-3628cadf942c) (canary). Done (00:00:47)
Started updating job blobstore_z1 > blobstore_z1/0
(28f3cf80-502b-4e83-8dff-ba3477894a13) (canary). Done (00:01:20)
Started updating job postgres_z1 > postgres_z1/0
(601262d5-97d2-436c-b67d-1e63680f9dc6) (canary). Done (00:01:16)
Started updating job uaa_z1 > uaa_z1/0
(9792327b-05c6-4bd8-8434-8a48fa364a91) (canary). Done (00:00:52)
Started updating job api_z1 > api_z1/0
(03f63cee-dfc1-4a93-9ddc-1bf129eb7396) (canary). Failed: 'api_z1/0
(03f63cee-dfc1-4a93-9ddc-1bf129eb7396)' is not running after update. Review
logs for failed jobs: cloud_controller_ng, cloud_controller_worker_local_1,
cloud_controller_worker_local_2, nginx_cc, cloud_controller_worker_1
(00:11:11)

Error 400007: 'api_z1/0 (03f63cee-dfc1-4a93-9ddc-1bf129eb7396)' is not
running after update. Review logs for failed jobs: cloud_controller_ng,
cloud_controller_worker_local_1, cloud_controller_worker_local_2, nginx_cc,
cloud_controller_worker_1



"bosh vms" command seem to show blobstore be up and running:

+---------------------------------------------------------------------------+---------+-----+-----------+---------------+
| VM
| State | AZ | VM Type | IPs |

+---------------------------------------------------------------------------+---------+-----+-----------+---------------+
| api_z1/0 (03f63cee-dfc1-4a93-9ddc-1bf129eb7396)
| failing | n/a | large_z1 | 192.168.10.54 |
| blobstore_z1/0 (28f3cf80-502b-4e83-8dff-ba3477894a13)
| running | n/a | medium_z1 | 192.168.10.52 |
| consul_z1/0 (d248a6da-476a-4fb1-bc08-01ce23c94eb4)
| running | n/a | small_z1 | 192.168.10.76 |
| doppler_z1/0 (ec0232f4-e736-4bbf-9fe4-b35b3d0c673d)
| running | n/a | medium_z1 | 192.168.10.57 |
| etcd_z1/0 (ef27b75d-e0d0-4a95-8ec1-4b0c9371fa93)
| running | n/a | medium_z1 | 192.168.10.72 |
| ha_proxy_z1/0 (f416b7c5-24d3-4d58-bbf9-066740971065)
| running | n/a | router_z1 | 192.168.10.64 |
|
| | | | XX.XX.XX.XX |
| hm9000_z1/0 (c50fbd8e-1c8f-4b96-a1c8-a427fdabba79)
| running | n/a | medium_z1 | 192.168.10.55 |
| loggregator_trafficcontroller_z1/0
(e187f9c1-dc5d-4d9b-850d-ca6118513228) | running | n/a | small_z1 |
192.168.10.58 |
| nats_z1/0 (074d00e8-96b3-4b7c-af1a-347549ff51dd)
| running | n/a | medium_z1 | 192.168.10.66 |
| postgres_z1/0 (601262d5-97d2-436c-b67d-1e63680f9dc6)
| running | n/a | medium_z1 | 192.168.10.68 |
| router_z1/0 (4da988b9-5cbb-4a6f-b082-5c4ed25e6a7c)
| running | n/a | router_z1 | 192.168.10.69 |
| runner_z1/0 (4bce5874-2edc-4a0f-a074-0a136933de65)
| running | n/a | runner_z1 | 192.168.10.56 |
| stats_z1/0 (20c1f5e9-3b7a-4c84-a940-3628cadf942c)
| running | n/a | small_z1 | 192.168.10.51 |
| uaa_z1/0 (9792327b-05c6-4bd8-8434-8a48fa364a91)
| running | n/a | medium_z1 | 192.168.10.53 |

+---------------------------------------------------------------------------+---------+-----+-----------+---------------+

The vcap job logs also seem to show it thinks that the blobstore node is
healthey,

However, when I ssh to the api node, I see problems:
"monit summary" yields:
Process 'cloud_controller_ng' not monitored
Process 'cloud_controller_worker_local_1' not monitored
Process 'cloud_controller_worker_local_2' not monitored
Process 'nginx_cc' not monitored
Process 'cloud_controller_migration' running
Process 'cloud_controller_clock' running
Process 'cloud_controller_worker_1' running
Process 'metron_agent' running
Process 'statsd-injector' running
Process 'route_registrar' running
System 'system_localhost' running


, /var/vcap/sys/log/cloud_controller_ng/cloud_controller_ng_ctl.log
looks like this:
[2016-06-24 12:32:51+0000] Checking for blobstore availability
[2016-06-24 12:33:38+0000] ------------ STARTING cloud_controller_ng_ctl
at Fri Jun 24 12:33:38 UTC 2016 --------------
[2016-06-24 12:33:38+0000] Checking for blobstore availability
[2016-06-24 12:34:24+0000] ------------ STARTING cloud_controller_ng_ctl
at Fri Jun 24 12:34:24 UTC 2016 --------------
[2016-06-24 12:34:24+0000] Checking for blobstore availability
[2016-06-24 12:34:41+0000] Blobstore is not available

Other jobs on the same node are failing as well, all with similar
messages both in stderr and stdout logs.
However, I do not see any additional information, such as an IP address
or URL that I can test.

Additionally on the blobstore node, both "monit" and the vcap job logs
seem to think that it is healthy.

How should I go about debugging this error? (I am looking for more
verbose information on the error "blobstore is not available")

I do not mind tearing down this deployment and starting over with more
verbose flags to expose the problem.
It seems to occur consitantly.

PS: my stemcell is
bosh-stemcell-3232.6-openstack-kvm-ubuntu-trusty-go_agent.tgz , and I have
not done anything special to configure blobstore on cf-stub besides adding
TLS Ca's and certs. Also, I could find no instructions on how to add these
certs to cf-stub, so I just added some self-generated keys & certs.



Re: How shoulld I debug a blobstore error?

Pramod Mandagere
 

Hi Amit,
We have the exact same issue and our cf-stub looking exactly like the one u
pointed out. Stuck with this issue,
Regards,
Pramod

On Fri, Jun 24, 2016 at 10:26 AM Amit Gupta <agupta(a)pivotal.io> wrote:

Hi Eyal,

No need to tear things down. By any chance, did you generate your
manifest using a stub? Do you have something that looks like this in your
manifest?


<https://github.com/cloudfoundry/cf-release/blob/3290ed8/spec/fixtures/openstack/cf-stub.yml#L155-L187>
https://github.com/cloudfoundry/cf-release/blob/3290ed8/spec/fixtures/openstack/cf-stub.yml#L155-L187

Best,
Amit

On Fri, Jun 24, 2016 at 6:54 AM, Eyal Shalev <eshalev(a)cisco.com> wrote:

Hello, I am trying to deploy cloud foundry on openstack.

running "bosh deploy" gives me the following error:
Started updating job consul_z1 > consul_z1/0
(d248a6da-476a-4fb1-bc08-01ce23c94eb4) (canary). Done (00:01:03)
Started updating job ha_proxy_z1 > ha_proxy_z1/0
(f416b7c5-24d3-4d58-bbf9-066740971065) (canary). Done (00:00:43)
Started updating job nats_z1 > nats_z1/0
(074d00e8-96b3-4b7c-af1a-347549ff51dd) (canary). Done (00:00:44)
Started updating job etcd_z1 > etcd_z1/0
(ef27b75d-e0d0-4a95-8ec1-4b0c9371fa93) (canary). Done (00:01:23)
Started updating job stats_z1 > stats_z1/0
(20c1f5e9-3b7a-4c84-a940-3628cadf942c) (canary). Done (00:00:47)
Started updating job blobstore_z1 > blobstore_z1/0
(28f3cf80-502b-4e83-8dff-ba3477894a13) (canary). Done (00:01:20)
Started updating job postgres_z1 > postgres_z1/0
(601262d5-97d2-436c-b67d-1e63680f9dc6) (canary). Done (00:01:16)
Started updating job uaa_z1 > uaa_z1/0
(9792327b-05c6-4bd8-8434-8a48fa364a91) (canary). Done (00:00:52)
Started updating job api_z1 > api_z1/0
(03f63cee-dfc1-4a93-9ddc-1bf129eb7396) (canary). Failed: 'api_z1/0
(03f63cee-dfc1-4a93-9ddc-1bf129eb7396)' is not running after update. Review
logs for failed jobs: cloud_controller_ng, cloud_controller_worker_local_1,
cloud_controller_worker_local_2, nginx_cc, cloud_controller_worker_1
(00:11:11)

Error 400007: 'api_z1/0 (03f63cee-dfc1-4a93-9ddc-1bf129eb7396)' is not
running after update. Review logs for failed jobs: cloud_controller_ng,
cloud_controller_worker_local_1, cloud_controller_worker_local_2, nginx_cc,
cloud_controller_worker_1



"bosh vms" command seem to show blobstore be up and running:

+---------------------------------------------------------------------------+---------+-----+-----------+---------------+
| VM
| State | AZ | VM Type | IPs |

+---------------------------------------------------------------------------+---------+-----+-----------+---------------+
| api_z1/0 (03f63cee-dfc1-4a93-9ddc-1bf129eb7396)
| failing | n/a | large_z1 | 192.168.10.54 |
| blobstore_z1/0 (28f3cf80-502b-4e83-8dff-ba3477894a13)
| running | n/a | medium_z1 | 192.168.10.52 |
| consul_z1/0 (d248a6da-476a-4fb1-bc08-01ce23c94eb4)
| running | n/a | small_z1 | 192.168.10.76 |
| doppler_z1/0 (ec0232f4-e736-4bbf-9fe4-b35b3d0c673d)
| running | n/a | medium_z1 | 192.168.10.57 |
| etcd_z1/0 (ef27b75d-e0d0-4a95-8ec1-4b0c9371fa93)
| running | n/a | medium_z1 | 192.168.10.72 |
| ha_proxy_z1/0 (f416b7c5-24d3-4d58-bbf9-066740971065)
| running | n/a | router_z1 | 192.168.10.64 |
|
| | | | XX.XX.XX.XX |
| hm9000_z1/0 (c50fbd8e-1c8f-4b96-a1c8-a427fdabba79)
| running | n/a | medium_z1 | 192.168.10.55 |
| loggregator_trafficcontroller_z1/0
(e187f9c1-dc5d-4d9b-850d-ca6118513228) | running | n/a | small_z1 |
192.168.10.58 |
| nats_z1/0 (074d00e8-96b3-4b7c-af1a-347549ff51dd)
| running | n/a | medium_z1 | 192.168.10.66 |
| postgres_z1/0 (601262d5-97d2-436c-b67d-1e63680f9dc6)
| running | n/a | medium_z1 | 192.168.10.68 |
| router_z1/0 (4da988b9-5cbb-4a6f-b082-5c4ed25e6a7c)
| running | n/a | router_z1 | 192.168.10.69 |
| runner_z1/0 (4bce5874-2edc-4a0f-a074-0a136933de65)
| running | n/a | runner_z1 | 192.168.10.56 |
| stats_z1/0 (20c1f5e9-3b7a-4c84-a940-3628cadf942c)
| running | n/a | small_z1 | 192.168.10.51 |
| uaa_z1/0 (9792327b-05c6-4bd8-8434-8a48fa364a91)
| running | n/a | medium_z1 | 192.168.10.53 |

+---------------------------------------------------------------------------+---------+-----+-----------+---------------+

The vcap job logs also seem to show it thinks that the blobstore node is
healthey,

However, when I ssh to the api node, I see problems:
"monit summary" yields:
Process 'cloud_controller_ng' not monitored
Process 'cloud_controller_worker_local_1' not monitored
Process 'cloud_controller_worker_local_2' not monitored
Process 'nginx_cc' not monitored
Process 'cloud_controller_migration' running
Process 'cloud_controller_clock' running
Process 'cloud_controller_worker_1' running
Process 'metron_agent' running
Process 'statsd-injector' running
Process 'route_registrar' running
System 'system_localhost' running


, /var/vcap/sys/log/cloud_controller_ng/cloud_controller_ng_ctl.log looks
like this:
[2016-06-24 12:32:51+0000] Checking for blobstore availability
[2016-06-24 12:33:38+0000] ------------ STARTING cloud_controller_ng_ctl
at Fri Jun 24 12:33:38 UTC 2016 --------------
[2016-06-24 12:33:38+0000] Checking for blobstore availability
[2016-06-24 12:34:24+0000] ------------ STARTING cloud_controller_ng_ctl
at Fri Jun 24 12:34:24 UTC 2016 --------------
[2016-06-24 12:34:24+0000] Checking for blobstore availability
[2016-06-24 12:34:41+0000] Blobstore is not available

Other jobs on the same node are failing as well, all with similar
messages both in stderr and stdout logs.
However, I do not see any additional information, such as an IP address
or URL that I can test.

Additionally on the blobstore node, both "monit" and the vcap job logs
seem to think that it is healthy.

How should I go about debugging this error? (I am looking for more
verbose information on the error "blobstore is not available")

I do not mind tearing down this deployment and starting over with more
verbose flags to expose the problem.
It seems to occur consitantly.

PS: my stemcell is
bosh-stemcell-3232.6-openstack-kvm-ubuntu-trusty-go_agent.tgz , and I have
not done anything special to configure blobstore on cf-stub besides adding
TLS Ca's and certs. Also, I could find no instructions on how to add these
certs to cf-stub, so I just added some self-generated keys & certs.



Re: How shoulld I debug a blobstore error?

Amit Kumar Gupta
 

Hi Eyal,

No need to tear things down. By any chance, did you generate your manifest
using a stub? Do you have something that looks like this in your manifest?

<https://github.com/cloudfoundry/cf-release/blob/3290ed8/spec/fixtures/openstack/cf-stub.yml#L155-L187>
https://github.com/cloudfoundry/cf-release/blob/3290ed8/spec/fixtures/openstack/cf-stub.yml#L155-L187

Best,
Amit

On Fri, Jun 24, 2016 at 6:54 AM, Eyal Shalev <eshalev(a)cisco.com> wrote:

Hello, I am trying to deploy cloud foundry on openstack.

running "bosh deploy" gives me the following error:
Started updating job consul_z1 > consul_z1/0
(d248a6da-476a-4fb1-bc08-01ce23c94eb4) (canary). Done (00:01:03)
Started updating job ha_proxy_z1 > ha_proxy_z1/0
(f416b7c5-24d3-4d58-bbf9-066740971065) (canary). Done (00:00:43)
Started updating job nats_z1 > nats_z1/0
(074d00e8-96b3-4b7c-af1a-347549ff51dd) (canary). Done (00:00:44)
Started updating job etcd_z1 > etcd_z1/0
(ef27b75d-e0d0-4a95-8ec1-4b0c9371fa93) (canary). Done (00:01:23)
Started updating job stats_z1 > stats_z1/0
(20c1f5e9-3b7a-4c84-a940-3628cadf942c) (canary). Done (00:00:47)
Started updating job blobstore_z1 > blobstore_z1/0
(28f3cf80-502b-4e83-8dff-ba3477894a13) (canary). Done (00:01:20)
Started updating job postgres_z1 > postgres_z1/0
(601262d5-97d2-436c-b67d-1e63680f9dc6) (canary). Done (00:01:16)
Started updating job uaa_z1 > uaa_z1/0
(9792327b-05c6-4bd8-8434-8a48fa364a91) (canary). Done (00:00:52)
Started updating job api_z1 > api_z1/0
(03f63cee-dfc1-4a93-9ddc-1bf129eb7396) (canary). Failed: 'api_z1/0
(03f63cee-dfc1-4a93-9ddc-1bf129eb7396)' is not running after update. Review
logs for failed jobs: cloud_controller_ng, cloud_controller_worker_local_1,
cloud_controller_worker_local_2, nginx_cc, cloud_controller_worker_1
(00:11:11)

Error 400007: 'api_z1/0 (03f63cee-dfc1-4a93-9ddc-1bf129eb7396)' is not
running after update. Review logs for failed jobs: cloud_controller_ng,
cloud_controller_worker_local_1, cloud_controller_worker_local_2, nginx_cc,
cloud_controller_worker_1



"bosh vms" command seem to show blobstore be up and running:

+---------------------------------------------------------------------------+---------+-----+-----------+---------------+
| VM
| State | AZ | VM Type | IPs |

+---------------------------------------------------------------------------+---------+-----+-----------+---------------+
| api_z1/0 (03f63cee-dfc1-4a93-9ddc-1bf129eb7396)
| failing | n/a | large_z1 | 192.168.10.54 |
| blobstore_z1/0 (28f3cf80-502b-4e83-8dff-ba3477894a13)
| running | n/a | medium_z1 | 192.168.10.52 |
| consul_z1/0 (d248a6da-476a-4fb1-bc08-01ce23c94eb4)
| running | n/a | small_z1 | 192.168.10.76 |
| doppler_z1/0 (ec0232f4-e736-4bbf-9fe4-b35b3d0c673d)
| running | n/a | medium_z1 | 192.168.10.57 |
| etcd_z1/0 (ef27b75d-e0d0-4a95-8ec1-4b0c9371fa93)
| running | n/a | medium_z1 | 192.168.10.72 |
| ha_proxy_z1/0 (f416b7c5-24d3-4d58-bbf9-066740971065)
| running | n/a | router_z1 | 192.168.10.64 |
|
| | | | XX.XX.XX.XX |
| hm9000_z1/0 (c50fbd8e-1c8f-4b96-a1c8-a427fdabba79)
| running | n/a | medium_z1 | 192.168.10.55 |
| loggregator_trafficcontroller_z1/0
(e187f9c1-dc5d-4d9b-850d-ca6118513228) | running | n/a | small_z1 |
192.168.10.58 |
| nats_z1/0 (074d00e8-96b3-4b7c-af1a-347549ff51dd)
| running | n/a | medium_z1 | 192.168.10.66 |
| postgres_z1/0 (601262d5-97d2-436c-b67d-1e63680f9dc6)
| running | n/a | medium_z1 | 192.168.10.68 |
| router_z1/0 (4da988b9-5cbb-4a6f-b082-5c4ed25e6a7c)
| running | n/a | router_z1 | 192.168.10.69 |
| runner_z1/0 (4bce5874-2edc-4a0f-a074-0a136933de65)
| running | n/a | runner_z1 | 192.168.10.56 |
| stats_z1/0 (20c1f5e9-3b7a-4c84-a940-3628cadf942c)
| running | n/a | small_z1 | 192.168.10.51 |
| uaa_z1/0 (9792327b-05c6-4bd8-8434-8a48fa364a91)
| running | n/a | medium_z1 | 192.168.10.53 |

+---------------------------------------------------------------------------+---------+-----+-----------+---------------+

The vcap job logs also seem to show it thinks that the blobstore node is
healthey,

However, when I ssh to the api node, I see problems:
"monit summary" yields:
Process 'cloud_controller_ng' not monitored
Process 'cloud_controller_worker_local_1' not monitored
Process 'cloud_controller_worker_local_2' not monitored
Process 'nginx_cc' not monitored
Process 'cloud_controller_migration' running
Process 'cloud_controller_clock' running
Process 'cloud_controller_worker_1' running
Process 'metron_agent' running
Process 'statsd-injector' running
Process 'route_registrar' running
System 'system_localhost' running


, /var/vcap/sys/log/cloud_controller_ng/cloud_controller_ng_ctl.log looks
like this:
[2016-06-24 12:32:51+0000] Checking for blobstore availability
[2016-06-24 12:33:38+0000] ------------ STARTING cloud_controller_ng_ctl
at Fri Jun 24 12:33:38 UTC 2016 --------------
[2016-06-24 12:33:38+0000] Checking for blobstore availability
[2016-06-24 12:34:24+0000] ------------ STARTING cloud_controller_ng_ctl
at Fri Jun 24 12:34:24 UTC 2016 --------------
[2016-06-24 12:34:24+0000] Checking for blobstore availability
[2016-06-24 12:34:41+0000] Blobstore is not available

Other jobs on the same node are failing as well, all with similar messages
both in stderr and stdout logs.
However, I do not see any additional information, such as an IP address or
URL that I can test.

Additionally on the blobstore node, both "monit" and the vcap job logs
seem to think that it is healthy.

How should I go about debugging this error? (I am looking for more verbose
information on the error "blobstore is not available")

I do not mind tearing down this deployment and starting over with more
verbose flags to expose the problem.
It seems to occur consitantly.

PS: my stemcell is
bosh-stemcell-3232.6-openstack-kvm-ubuntu-trusty-go_agent.tgz , and I have
not done anything special to configure blobstore on cf-stub besides adding
TLS Ca's and certs. Also, I could find no instructions on how to add these
certs to cf-stub, so I just added some self-generated keys & certs.



Re: Consul Encryption in CF v234+

Long Nguyen
 

Hi there!

 We found that if you monit stop all the consul nodes before upgrading and adding ssl. The deployment should upgrade without any issues.

Thanks,
Long

On June 23, 2016 at 11:56:04 AM, Hiort, Carsten, Springer DE (carsten.hiort(a)springer.com) wrote:

Hi,

CF v234 enforces the use of SSL for Consul. We are currently wondering if there is a supposed upgrade path.
When you switch to SSL and the Consul cluster get’s upgraded all machines that are not yet upgraded will be blind with respect to service discovery/ DNS through Consul. This particularly affects the DEAs as they are not able to figure out where to get the droplets from when staging causing a 500 when cf pushing. I did already try deploying the Certs on 231 with require_ssl=false but then setting require_ssl true or upgrading to v234+ will still rest in this situation.
Any thoughts highly appreciated!


Thanks,

Carsten

---

Carsten Hiort
Platform Engineer
Platform Engineering 

SpringerNature
Abraham-Lincoln-Str. 46, 65189 Wiesbaden, Germany
T   +49 611 7878665
M +49 175 2965802
carsten.hiort(a)springernature.com
www.springernature.com

Springer Nature is one of the world’s leading global research, educational and professional publishers, created in May 2015 through the combination of Nature Publishing Group,
Palgrave Macmillan, Macmillan Education and Springer Science+Business Media

Springer Fachmedien Wiesbaden GmbH
Registered Office: Wiesbaden | Amtsgericht Wiesbaden, HRB 9754
Directors: Armin Gross, Joachim Krieger, Dr. Niels Peter Thomas