Buildbot tutorial failure - worker can't connect to master - buildbot

I'm trying out buildbot, and following the tutorial. I successfully installed and started the master, as shown from the twistd.log partial file below.
2018-11-14 08:32:52-0600 [-] BuildMaster is running
2018-11-14 08:32:52-0600 [-] buildbotNetUsageData: sending
{'installid': '38122a5be5d93ebdf0ef631a1584f36f418d12c0', 'versions':
{'Python': '3.6.3', 'Buildbot': '1.5.0', 'Twisted': '18.9.0'},
'platform': {'platform': 'Darwin-18.2.0-x86_64-i386-64bit', 'system':
'Darwin', 'machine': 'x86_64', 'processor': 'i386',
'python_implementation': 'CPython', 'version': 'Darwin Kernel Version
18.2.0:', 'distro': '10.14.2'}, 'plugins':
{'buildbot/worker/base/Worker': 1, 'buildbot/config/BuilderConfig': 1,
'buildbot/schedulers/basic/SingleBranchScheduler': 1,
'buildbot/schedulers/forcesched/ForceScheduler': 1,
'buildbot/changes/gitpoller/GitPoller': 1,
'buildbot/steps/source/git/Git': 1,
'buildbot/steps/shell/ShellCommand': 1}, 'db': 'sqlite', 'mq':
'simple', 'www_plugins': ['waterfall_view', 'console_view', 'grid_view']}
After installing and setting up the worker, the worker fails to connect to the master, and I'm also unable to load localhost:8010 as the tutorial indicates should be running. I've verified that I created the worker with the same worker name and password as the master.cfg file specifies. Any thoughts on what I missed?
A partial worker twistd.log is below:
2018-11-14 08:32:34-0600 [-] Connecting to localhost:9989
2018-11-14 08:32:34-0600 [Uninitialized] Connection to localhost:9989 failed: Connection Refused
2018-11-14 08:32:34-0600 [Uninitialized] <twisted.internet.tcp.Connector object at 0x10d8e6080> will retry in 62 seconds
2018-11-14 08:32:34-0600 [Uninitialized] Retry attempt 4/inf
2018-11-14 08:32:34-0600 [-] Stopping factory <buildbot_worker.pb.BotFactory object at 0x10d8d6c50>
2018-11-14 08:33:37-0600 [-] Starting factory <buildbot_worker.pb.BotFactory object at 0x10d8d6c50>
2018-11-14 08:33:37-0600 [-] Connecting to localhost:9989
2018-11-14 08:33:37-0600 [Uninitialized] Connection to localhost:9989 failed: Connection Refused

Please, try use buildbot reconfig (http://docs.buildbot.net/1.5.0/manual/cfg-intro.html?highlight=reconfig)
or buildbot upgrade-master (http://docs.buildbot.net/1.5.0/manual/cmdline.html?highlight=upgrade-master#)
Also, You can find a reference to reconfig in the Buildbot Quick Tour: (http://docs.buildbot.net/1.5.0/tutorial/tour.html)
You should use this after each changes in your Buildbot decision (workers, schedules, notifiers, factories, etc...)

Related

Mongooseim 3.6.0 postgress connection Issue

Hi I am new to mongooseim , I am planing to setup mongooseim in my local setup with postgres database connection. I have installed mongooseim 3.6.0 in ubuntu 14.04 machine. And create database in postgres & add the schema for that.Then I have done the following changes in mongooseim.cfg file.
{outgoing_pools, [
{rdbms, global, default, [{workers, 1}], [{server, {psql, "localhost", 5432, "mongooseim", "postgres", "password"}}]}
]}.
And this one.
{rdbms_server_type, pgsql}.
These are the changes I have done for default config file. Then When I restart the server it gives this error. postgres server running & user credentials are working.
020-06-05 18:17:46.815 [info] <0.249.0> msg: "Starting reporters with []\n", options: []
2020-06-05 18:17:47.034 [notice] <0.130.0>#lager_file_backend:143 Changed loglevel of /var/log/mongooseim/ejabberd.log to info
2020-06-05 18:17:47.136 [info] <0.43.0> Application mnesia exited with reason: stopped
2020-06-05 18:17:47.621 [error] <0.593.0>#mongoose_rdbms_psql:connect CRASH REPORT Process <0.593.0> with 0 neighbours crashed with reason: call to undefined function mongoose_rdbms_psql:connect({psql,"server",5432,"mongooseim","postgres","password"}, 5000)
2020-06-05 18:17:47.621 [error] <0.592.0>#mongoose_rdbms_psql:connect Supervisor 'wpool_pool-mongoose_wpool$rdbms$global$default-process-sup' had child 'wpool_pool-mongoose_wpool$rdbms$global$default-1' started with wpool_process:start_link('wpool_pool-mongoose_wpool$rdbms$global$default-1', mongoose_rdbms, [{server,{psql,"server",5432,"mongooseim","postgres","password"}}], [{queue_manager,'wpool_pool-mongoose_wpool$rdbms$global$default-queue-manager'},{time_checker,'wpool_pool-mongoose_wpool$rdbms$global$default-time-checker'},...]) at undefined exit with reason call to undefined function mongoose_rdbms_psql:connect({psql,"server",5432,"mongooseim","postgres","password"}, 5000) in context start_error
2020-06-05 18:17:47.622 [error] <0.589.0> Supervisor 'mongoose_wpool$rdbms$global$default' had child 'wpool_pool-mongoose_wpool$rdbms$global$default-process-sup' started with wpool_process_sup:start_link('mongoose_wpool$rdbms$global$default', 'wpool_pool-mongoose_wpool$rdbms$global$default-process-sup', [{queue_manager,'wpool_pool-mongoose_wpool$rdbms$global$default-queue-manager'},{time_checker,'wpool_pool-mongoose_wpool$rdbms$global$default-time-checker'},...]) at undefined exit with reason {shutdown,{failed_to_start_child,'wpool_pool-mongoose_wpool$rdbms$global$default-1',{undef,[{mongoose_rdbms_psql,connect,[{psql,"server",5432,"mongooseim","postgres","password"},5000],[]},{mongoose_rdbms,connect,4,[{file,"/root/deb/mongooseim/_build/prod/lib/mongooseim/src/rdbms/mongoose_rdbms.erl"},{line,668}]},{mongoose_rdbms,init,1,[{file,"/root/deb/mongooseim/_build/prod/lib/mongooseim/src/rdbms/mongoose_rdbms.erl"},{line,431}]},{wpool_process,init,1,[{file,"/root/deb/mongooseim/_build/..."},...]},...]}}} in context start_error
2020-06-05 18:17:47.622 [error] <0.583.0>#mongoose_wpool_mgr:handle_call:105 Pool not started: {error,{{shutdown,{failed_to_start_child,'wpool_pool-mongoose_wpool$rdbms$global$default-process-sup',{shutdown,{failed_to_start_child,'wpool_pool-mongoose_wpool$rdbms$global$default-1',{undef,[{mongoose_rdbms_psql,connect,[{psql,"server",5432,"mongooseim","postgres","password"},5000],[]},{mongoose_rdbms,connect,4,[{file,"/root/deb/mongooseim/_build/prod/lib/mongooseim/src/rdbms/mongoose_rdbms.erl"},{line,668}]},{mongoose_rdbms,init,1,[{file,"/root/deb/mongooseim/_build/prod/lib/mongooseim/src/rdbms/mongoose_rdbms.erl"},{line,431}]},{wpool_process,init,1,[{file,"/root/deb/mongooseim/_build/default/lib/worker_pool/src/wpool_process.erl"},{line,85}]},{gen_server,init_it,2,[{file,"gen_server.erl"},{line,374}]},{gen_server,init_it,6,[{file,"gen_server.erl"},{line,342}]},{proc_lib,init_p_do_apply,3,[{file,"proc_lib.erl"},{line,249}]}]}}}}},{child,undefined,'mongoose_wpool$rdbms$global$default',{wpool,start_pool,['mongoose_wpool$rdbms$global$default',[{worker,{mongoose_rdbms,[{server,{psql,"server",5432,"mongooseim","postgres","password"}}]}},{pool_sup_shutdown,infinity},{workers,1}]]},temporary,infinity,supervisor,[wpool]}}}
2020-06-05 18:17:47.678 [warning] <0.615.0>#service_mongoose_system_metrics:report_transparency:129 We are gathering the MongooseIM system's metrics to analyse the trends and needs of our users, improve MongooseIM, and know where to focus our efforts. For more info on how to customise, read, enable, and disable these metrics visit:
- MongooseIM docs -
https://mongooseim.readthedocs.io/en/latest/operation-and-maintenance/System-Metrics-Privacy-Policy/
- MongooseIM GitHub page - https://github.com/esl/MongooseIM
The last sent report is also written to a file /var/log/mongooseim/system_metrics_report.json
2020-06-05 18:17:48.404 [warning] <0.1289.0>#nkpacket_stun:get_stun_servers:231 Current NAT is changing ports!

Rundeck 3.x error for remote execution - Failed: AuthenticationFailure: Authentication failure connecting to node. Could not authenticate

Getting this error in the rundeck 3.x (latest) console when trying to do a remote uptime to the host in question. Can ssh into the host from the rundeck server as rundeck user and root and have set the necessary public keys there and in key storage on rundeck server.
For resources.xml properties file, what should the settings be as that is where it is discovered.
The error in the rundeck server /var/log/rundeck/service.log is:
[2020-01-21 01:15:50.826] ERROR ExecutionUtilService --- [eduler_Worker-1] Execution failed: 29 in project TestProject: [Workflow result\
: , step failures: {1=Dispatch failed on 1 nodes: [some-random-host: AuthenticationFailure: Authentication failure connecting to node: "\
some-random-host". Could not authenticate. + {dataContext=MultiDataContextImpl(map={ContextView(node:some-random-host)=BaseDataContext{{\
exec={exitCode=-1}}}, ContextView(step:1, node:some-random-host)=BaseDataContext{{exec={exitCode=-1}}}}, base=null)} ]}, Node failures: \
{some-random-host=[AuthenticationFailure: Authentication failure connecting to node: "some-random-host". Could not authenticate. + {data\
Context=MultiDataContextImpl(map={ContextView(node:some-random-host)=BaseDataContext{{exec={exitCode=-1}}}, ContextView(step:1, node:som\
e-random-host)=BaseDataContext{{exec={exitCode=-1}}}}, base=null)} ]}, status: failed]
Thanks.
Make sure that you have the remote node well configured (and well referenced in your node definition), you have a good guide here.

Error while using persistent datasource using mongodb ini hyperledger composer

I am trying to use persistent datasource using mongoDB in hyperledger composer on a UBUNTU droplet
but after starting the rest server and den after issuing a command docker logs -f rest i am getting the following error(i have provided a link to the image)
webuser#ubuntu16:~$ docker logs -f rest
[2018-08-29T12:38:31.278Z] PM2 log: Launching in no daemon mode
[2018-08-29T12:38:31.351Z] PM2 log: Starting execution sequence in -fork mode- for app name:composer-rest-server id:0
[2018-08-29T12:38:31.359Z] PM2 log: App name:composer-rest-server id:0 online
WARNING: NODE_APP_INSTANCE value of '0' did not match any instance config file names.
WARNING: See https://github.com/lorenwest/node-config/wiki/Strict-Mode
Discovering types from business network definition ...
(node:15) DeprecationWarning: current URL string parser is deprecated, and will be removed in a future version. To use the new parser, pass option { useNewUrlParser: true } to MongoClient.connect.
Connection fails: Error: Error trying to ping. Error: Failed to connect before the deadline
It will be retried for the next request.
Exception: Error: Error trying to ping. Error: Failed to connect before the deadline
Error: Error trying to ping. Error: Failed to connect before the deadline
at _checkRuntimeVersions.then.catch (/home/composer/.npm-global/lib/node_modules/composer-rest-server/node_modules/composer-connector-hlfv1/lib/hlfconnection.js:806:34)
at <anonymous>
[2018-08-29T12:38:41.021Z] PM2 log: App [composer-rest-server] with id [0] and pid [15], exited with code [1] via signal [SIGINT]
[2018-08-29T12:38:41.024Z] PM2 log: Starting execution sequence in -fork mode- for app name:composer-rest-server id:0
[2018-08-29T12:38:41.028Z] PM2 log: App name:composer-rest-server id:0 online
WARNING: NODE_APP_INSTANCE value of '0' did not match any instance config file names.
WARNING: See https://github.com/lorenwest/node-config/wiki/Strict-Mode
Discovering types from business network definition ...
(node:40) DeprecationWarning: current URL string parser is deprecated, and will be removed in a future version. To use the new parser, pass option { useNewUrlParser: true } to MongoClient.connect.
Connection fails: Error: Error trying to ping. Error: Failed to connect before the deadline
It will be retried for the next request.
I don't understand what is the problem and what wrong I am doing because I have followed all the steps in the hyperledger composer document with success....
Is it because I am using it on ubuntu droplet....?? anyone help
EDIT
I followed all the steps mentioned in this tutorial
but instead of using google authentication i am using github authentication.
Also i have changed my local host to the ip of my ubuntu droplet in connection.json file and also in this command
sed -e 's/localhost:7051/peer0.org1.example.com:7051/' -e 's/localhost:7053/peer0.org1.example.com:7053/' -e 's/localhost:7054/ca.org1.example.com:7054/' -e 's/localhost:7050/orderer.example.com:7050/' < $HOME/.composer/cards/restadmin#trade-network/connection.json > /tmp/connection.json && cp -p /tmp/connection.json $HOME/.composer/cards/restadmin#trade-network/
bt yet with no success! i get the following error now.....
webuser#ubuntu16:~$ docker logs rest
[2018-08-30T05:03:02.916Z] PM2 log: Launching in no daemon mode
[2018-08-30T05:03:02.989Z] PM2 log: Starting execution sequence in -fork mode- for app name:composer-rest-server id:0
[2018-08-30T05:03:02.997Z] PM2 log: App name:composer-rest-server id:0 online
WARNING: NODE_APP_INSTANCE value of '0' did not match any instance config file names.
WARNING: See https://github.com/lorenwest/node-config/wiki/Strict-Mode
Discovering types from business network definition ...
(node:15) DeprecationWarning: current URL string parser is deprecated, and will be removed in a future version. To use the new parser, pass option { useNewUrlParser: true } to MongoClient.connect.
Discovering the Returning Transactions..
Discovered types from business network definition
Generating schemas for all types in business network definition ...
Generated schemas for all types in business network definition
Adding schemas for all types to Loopback ...
Added schemas for all types to Loopback
SyntaxError: Unexpected string in JSON at position 92
at JSON.parse ()
at Promise.then (/home/composer/.npm-global/lib/node_modules/composer-rest-server/server/server.js:141:34)
at
at process._tickDomainCallback (internal/process/next_tick.js:228:7)
[2018-08-30T05:03:09.942Z] PM2 log: App [composer-rest-server] with id [0] and pid [15], exited with code 1 via signal [SIGINT]
This error Error trying to ping. Error: Failed to connect before the deadline means that the composer-rest-server in the container cannot see/connect to the underlying Fabric at the URLs in the connection.json of the card that you are using to start the REST server.
There are a number of reasons why:
The Fabric is not started
You are using a Business Network Card that has localhost in the URLs of the connection.json, and localhost just re-directs back into the rest container.
Your rest container is started on a different Docker network bridge to your Fabric containers and cannot connect to the Fabric.
Have you followed this tutorial in the Composer documentation? If followed completely it will avoid the 3 problems mentioned above.

Spymemcached issue with 2 nodes configured

I am using Ketama algorithm's spymemcached for my project. I do have two memcached servers running as part of HA (high availability) and my configurations are
hibernate.cache.use_second_level_cache=true
hibernate.cache.use_query_cache=true
hibernate.cache.region.factory_class=kr.pe.kwonnam.hibernate4memcached.Hibernate4MemcachedRegionFactory
hibernate.cache.default_cache_concurrency_strategy=NONSTRICT_READ_WRITE
hibernate.cache.region_prefix=myProjectCache
hibernate.cache.use_structured_entries=false
h4m.adapter.class=kr.pe.kwonnam.hibernate4memcached.spymemcached.SpyMemcachedAdapter
h4m.timestamper.class=kr.pe.kwonnam.hibernate4memcached.timestamper.HibernateCacheTimestamperMemcachedImpl
h4m.adapter.spymemcached.hosts=host1:11211,host2:11211
h4m.adapter.spymemcached.hashalgorithm=KETAMA_HASH
h4m.adapter.spymemcached.operation.timeout.millis=5000
h4m.adapter.spymemcached.transcoder=kr.pe.kwonnam.hibernate4memcached.spymemcached.KryoTranscoder
h4m.adapter.spymemcached.cachekey.prefix=myProject
h4m.adapter.spymemcached.kryotranscoder.compression.threashold.bytes=20000
# 10 minutes
h4m.expiry.seconds=600
# a day
h4m.expiry.seconds.validatorCache.org.hibernate.cache.spi.UpdateTimestampsCache=86400
# 1 hour
h4m.expiry.seconds.validatorCache.org.hibernate.cache.internal.StandardQueryCache=3600
# 30 minutes
h4m.expiry.seconds.myProjectCache.database1=1800
h4m.expiry.seconds.myProjectCache.database2=1800
Configurations are followed as per the link below :
SpyMemcachedAdapter
Both nodes host1 and host2 are reachable, up and running.
Issue :
As part of testing HA , when I bring down one memcached (host1) my application is connecting to host2 but only after trying to connect host1 (which will be timedout - as host1 is made down) for every request. Which will result in too much of time taken
Below is the exception thrown for every request
2017-07-07 17:27:31.915 [SimpleAsyncTaskExecutor-6] ERROR u.c.o.sProcessor - TransId:004579 - Exception occurred while processing request :Timeout waiting for value: waited 5,000 ms. Node status: Connection Status { /host1:11211 active: false, authed: true, last read: 247,290 ms ago /host2:11211 active: true, authed: true, last read: 5 ms ago }
2017-07-07 17:28:54.666 INFO net.spy.memcached.MemcachedConnection: Reconnecting due to failure to connect to {QA sa=/host1:11211, #Rops=0, #Wops=214, #iq=0, topRop=null, topWop=Cmd: 5 Opaque: 341143 Key: myProject.myProjectCache.databse1# Amount: 0 Default: 1499444604639 Exp: 2592000, toWrite=0, interested=0}
java.net.ConnectException: Connection refused
at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
at sun.nio.ch.SocketChannelImpl.finishConnect(Unknown Source)
at net.spy.memcached.MemcachedConnection.handleIO(MemcachedConnection.java:677)
at net.spy.memcached.MemcachedConnection.handleIO(MemcachedConnection.java:436)
at net.spy.memcached.MemcachedConnection.run(MemcachedConnection.java:1446)
2017-07-07 17:28:54.666 WARN net.spy.memcached.MemcachedConnection: Closing, and reopening {QA sa=/host1:11211, #Rops=0, #Wops=214, #iq=0, topRop=null, topWop=Cmd: 5 Opaque: 341143 Key: myProject.myProjectCache.databse1# Amount: 0 Default: 1499444604639 Exp: 2592000, toWrite=0, interested=0}, attempt 14.
2017-07-07 17:28:54.841 WARN net.spy.memcached.MemcachedConnection: Could not redistribute to another node, retrying primary node for myProject.myProjectCache.databse1#-1:my.co.org.myProject.dao.entity.databse1.tablexyz#14744.
Am using memcached for the first time, not sure whether this is the behavior of spymemcached? Or am I missing something in my configurations? Or
by changing timeout configurations will it resolve time taken to process the request?
Any suggestions/help much appreciated.
If you are using DefaultConnectionFactory which uses OOTB ConnectionFactoryBuilder then the reconnect will happen after failed operation count has reached timeoutExceptionThreshold (in version 2.7 of spymemcached) is initialized to 998. So if you create your own ConnectionFactory and change the timeoutExceptionThreshold to lower value then you should see the automatic recovery.
Hope this helps.

punjab installed and runs, but can not connect to http://mydomain:5280/xmpp-httpbind

I have Punjab installed on Ubuntu, which has apache and is a web
server. However, when I launch the page http://mydomain:5280/xmpp-httpbind,
I got the following error msg :
" Network Error (tcp_error)
communication error occurred: "Operation timed out
The Web Server may be down, too busy, or experiencing other problems preventing it from responding to requests. You may wish to try again at a later time."
The twisted.log does not say very much:
2011-08-12 10:02:33+0200 [-] Log opened.
2011-08-12 10:02:33+0200 [-] twistd 10.2.0 (/usr/bin/python 2.7.1) starting up.
2011-08-12 10:02:33+0200 [-] reactor class: twisted.internet.selectreactor.SelectReactor.
2011-08-12 10:02:33+0200 [-] twisted.web.server.Site starting on 5280
2011-08-12 10:02:33+0200 [-] Starting factory <twisted.web.server.Site instance at 0x19c0ef0>
The important settings in the punjab.tac file is:
root.putChild('http-bind', resource.IResource(bosh))
site = server.Site(root)
application = service.Application("punjab")
internet.TCPServer(5280, site).setServiceParent(application)
I am sitting behind a company proxy. Is this the reason? Thanks for
your kind help!
/Stella
did you try connecting to http://mydomain:5280/http-bind?