Unable to fetch data to dashboard using netflix hystrix turbine - hystrix

We are using netflix turbine to fetch data from cluster and display on hystrix dashboard.
I have one cluster called PROD with only one node(192.168.33.17) in which java application (hystrix) is running on port 8989.
We have turbine node where we have following configuration:
config.properties file:
turbine.aggregator.clusterConfig=PROD
turbine.instanceUrlSuffix.PROD=/hystrix.stream
turbine.ConfigPropertyBasedDiscovery.PROD.instances=192.168.33.17:8989/hystrix-examples-webapp
InstanceDiscovery.impl=com.netflix.turbine.discovery.ConfigPropertyBasedDiscovery
and placing config.properties file specified using a JVM property such as
java -Darchaius.configurationSource.additionalUrls=file:///home/vagrant/config.properties -jar start.jar
When we are directly accessing node data using
curl http://192.168.33.17:8989/hystrix-examples-webapp/hystrix.stream
then response is proper and we are able to see on the dashboard.
But when we doing the same using turbine we can't see any response on dashboard, although we are getting ping response for
curl http://192.168.33.13:8080/turbine/turbine.stream
response :ping
with turbine log:
2016-08-22 17:40:11 INFO com.netflix.turbine.discovery.InstanceObservable$1:260 [InstanceObservable$1] [run]: Retrieved hosts from InstanceDiscovery: 1
2016-08-22 17:40:11 INFO com.netflix.turbine.discovery.InstanceObservable$1:272 [InstanceObservable$1] [run]: Found hosts that have been previously terminated: 0
2016-08-22 17:40:11 INFO com.netflix.turbine.discovery.InstanceObservable$1:289 [InstanceObservable$1] [run]: Hosts up:1, hosts down: 0
2016-08-22 17:41:11 INFO com.netflix.turbine.discovery.InstanceObservable$1:260 [InstanceObservable$1] [run]: Retrieved hosts from InstanceDiscovery: 1
2016-08-22 17:41:11 INFO com.netflix.turbine.discovery.InstanceObservable$1:272 [InstanceObservable$1] [run]: Found hosts that have been previously terminated: 0
2016-08-22 17:41:11 INFO com.netflix.turbine.discovery.InstanceObservable$1:289 [InstanceObservable$1] [run]: Hosts up:1, hosts down: 0
2016-08-22 17:42:11 INFO com.netflix.turbine.discovery.InstanceObservable$1:260 [InstanceObservable$1] [run]: Retrieved hosts from InstanceDiscovery: 1
2016-08-22 17:42:11 INFO com.netflix.turbine.discovery.InstanceObservable$1:272 [InstanceObservable$1] [run]: Found hosts that have been previously terminated: 0
2016-08-22 17:42:11 INFO com.netflix.turbine.discovery.InstanceObservable$1:289 [InstanceObservable$1] [run]: Hosts up:1, hosts down: 0
2016-08-22 17:43:11 INFO com.netflix.turbine.discovery.InstanceObservable$1:260 [InstanceObservable$1] [run]: Retrieved hosts from InstanceDiscovery: 1
2016-08-22 17:43:11 INFO com.netflix.turbine.discovery.InstanceObservable$1:272 [InstanceObservable$1] [run]: Found hosts that have been previously terminated: 0
2016-08-22 17:43:11 INFO com.netflix.turbine.discovery.InstanceObservable$1:289 [InstanceObservable$1] [run]: Hosts up:1, hosts down: 0
2016-08-22 17:44:11 INFO com.netflix.turbine.discovery.InstanceObservable$1:260 [InstanceObservable$1] [run]: Retrieved hosts from InstanceDiscovery: 1
2016-08-22 17:44:11 INFO com.netflix.turbine.discovery.InstanceObservable$1:272 [InstanceObservable$1] [run]: Found hosts that have been previously terminated: 0
2016-08-22 17:44:11 INFO com.netflix.turbine.discovery.InstanceObservable$1:289 [InstanceObservable$1] [run]: Hosts up:1, hosts down: 0
20
So, are we missing any configuration ?

Related

Remote connection to [null] failed with java.net.NoRouteToHostException: No route to host in taskmanager

When I start my apache flink 1.10 taskmanager service in kubernetes(v1.15.2) cluster,it shows logs like this:
2020-05-01 08:34:55,847 INFO org.apache.flink.runtime.taskexecutor.TaskExecutor - Could not resolve ResourceManager address akka.tcp://flink#flink-jobmanager:6123/user/resourcemanager, retrying in 10000 ms: Could not connect to rpc endpoint under address akka.tcp://flink#flink-jobmanager:6123/user/resourcemanager..
2020-05-01 08:34:55,847 WARN akka.remote.transport.netty.NettyTransport - Remote connection to [null] failed with java.net.NoRouteToHostException: No route to host
2020-05-01 08:34:55,848 WARN akka.remote.ReliableDeliverySupervisor - Association with remote system [akka.tcp://flink#flink-jobmanager:6123] has failed, address is now gated for [50] ms. Reason: [Association failed with [akka.tcp://flink#flink-jobmanager:6123]] Caused by: [java.net.NoRouteToHostException: No route to host]
2020-05-01 08:35:08,874 WARN akka.remote.transport.netty.NettyTransport - Remote connection to [null] failed with java.net.NoRouteToHostException: No route to host
2020-05-01 08:35:08,877 WARN akka.remote.ReliableDeliverySupervisor - Association with remote system [akka.tcp://flink#flink-jobmanager:6123] has failed, address is now gated for [50] ms. Reason: [Association failed with [akka.tcp://flink#flink-jobmanager:6123]] Caused by: [java.net.NoRouteToHostException: No route to host]
2020-05-01 08:35:08,878 INFO org.apache.flink.runtime.taskexecutor.TaskExecutor - Could not resolve ResourceManager address akka.tcp://flink#flink-jobmanager:6123/user/resourcemanager, retrying in 10000 ms: Could not connect to rpc endpoint under address akka.tcp://flink#flink-jobmanager:6123/user/resourcemanager..
2020-05-01 08:35:21,907 WARN akka.remote.transport.netty.NettyTransport - Remote connection to [null] failed with java.net.NoRouteToHostException: No route to host
and the taskmanager could not registered success, and I logged into taskmanager and find out I could success ping jobmanager liket this:
flink#flink-taskmanager-54d85f57c7-nl9cf:~$ ping flink-jobmanager
PING flink-jobmanager.dabai-fat.svc.cluster.local (10.254.58.171) 56(84) bytes of data.
64 bytes from flink-jobmanager.dabai-fat.svc.cluster.local (10.254.58.171): icmp_seq=1 ttl=64 time=0.045 ms
64 bytes from flink-jobmanager.dabai-fat.svc.cluster.local (10.254.58.171): icmp_seq=2 ttl=64 time=0.076 ms
64 bytes from flink-jobmanager.dabai-fat.svc.cluster.local (10.254.58.171): icmp_seq=3 ttl=64 time=0.079 ms
so why this would happen and what should I do to fix it?
Try to install nmap in your kubernetes taskmanger's pod container:
apt-get udpate
apt-get install nmap -y
then scan the jobmanager and make sure the pod's expose port 6123 is accessable(in my case ,I found could not access the port 6123 from current pod).
nmap -T4 <your-jobmanager's-pod-ip>
Hope this help.

SonarQube don't start

I'm trying to install SonarQube on my CentOS7 server.
When i try to access via web I see this:
web_page
It's supposed to show a SonarQube page right??
Here i put my logs and configs:
web.log:
2019.08.29 17:24:54 INFO web[][o.s.s.a.EmbeddedTomcat] HTTP connector enabled on port 9000
2019.08.29 17:25:05 INFO web[][o.s.p.ProcessEntryPoint] Starting web
2019.08.29 17:25:05 INFO web[][o.a.t.u.n.NioSelectorPool] Using a shared selector for servlet write/read
2019.08.29 17:25:06 INFO web[][o.e.plugins] [Kurt Wagner] modules [], plugins [], sites []
2019.08.29 17:25:06 INFO web[][o.s.s.e.EsClientProvider] Connected to local Elasticsearch: [127.0.0.1:9001]
2019.08.29 17:25:06 INFO web[][o.s.s.p.LogServerVersion] SonarQube Server / 6.4.0.25310 / ad64a17b531c0e1f6fef0ce7e4d0d0b060977754
2019.08.29 17:25:06 INFO web[][o.sonar.db.Database] Create JDBC data source for jdbc:postgresql://localhost/sonar
2019.08.29 17:25:06 ERROR web[][o.s.s.p.Platform] Web server startup failed
java.lang.IllegalStateException: Can not connect to database. Please check connectivity and settings (see the properties prefixed by 'sonar.jdbc.').
at org.sonar.db.DefaultDatabase.checkConnection(DefaultDatabase.java:108)
at org.sonar.db.DefaultDatabase.start(DefaultDatabase.java:75)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.picocontainer.lifecycle.ReflectionLifecycleStrategy.invokeMethod(ReflectionLifecycleStrategy.java:110)
at org.picocontainer.lifecycle.ReflectionLifecycleStrategy.start(ReflectionLifecycleStrategy.java:89)
at org.sonar.core.platform.ComponentContainer$1.start(ComponentContainer.java:320)
at org.picocontainer.injectors.AbstractInjectionFactory$LifecycleAdapter.start(AbstractInjectionFactory.java:84)
at org.picocontainer.behaviors.AbstractBehavior.start(AbstractBehavior.java:169)
at org.picocontainer.behaviors.Stored$RealComponentLifecycle.start(Stored.java:132)
at org.picocontainer.behaviors.Stored.start(Stored.java:110)
at org.picocontainer.DefaultPicoContainer.potentiallyStartAdapter(DefaultPicoContainer.java:1016)
at org.picocontainer.DefaultPicoContainer.startAdapters(DefaultPicoContainer.java:1009)
at org.picocontainer.DefaultPicoContainer.start(DefaultPicoContainer.java:767)
at org.sonar.core.platform.ComponentContainer.startComponents(ComponentContainer.java:143)
at org.sonar.server.platform.platformlevel.PlatformLevel.start(PlatformLevel.java:88)
at org.sonar.server.platform.Platform.start(Platform.java:231)
at org.sonar.server.platform.Platform.startLevel1Container(Platform.java:190)
at org.sonar.server.platform.Platform.init(Platform.java:86)
at org.sonar.server.platform.web.PlatformServletContextListener.contextInitialized(PlatformServletContextListener.java:43)
at org.apache.catalina.core.StandardContext.listenerStart(StandardContext.java:4727)
at org.apache.catalina.core.StandardContext.startInternal(StandardContext.java:5189)
at org.apache.catalina.util.LifecycleBase.start(LifecycleBase.java:150)
at org.apache.catalina.core.ContainerBase$StartChild.call(ContainerBase.java:1419)
at org.apache.catalina.core.ContainerBase$StartChild.call(ContainerBase.java:1409)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:748)
Caused by: org.apache.commons.dbcp.SQLNestedException: Cannot create PoolableConnectionFactory (FATAL: password authentication failed for user "sonar")
at org.apache.commons.dbcp.BasicDataSource.createPoolableConnectionFactory(BasicDataSource.java:1549)
at org.apache.commons.dbcp.BasicDataSource.createDataSource(BasicDataSource.java:1388)
at org.apache.commons.dbcp.BasicDataSource.getConnection(BasicDataSource.java:1044)
at org.sonar.db.profiling.NullConnectionInterceptor.getConnection(NullConnectionInterceptor.java:31)
at org.sonar.db.profiling.ProfiledDataSource.getConnection(ProfiledDataSource.java:323)
at org.sonar.db.DefaultDatabase.checkConnection(DefaultDatabase.java:106)
... 30 common frames omitted
Caused by: org.postgresql.util.PSQLException: FATAL: password authentication failed for user "sonar"
at org.postgresql.core.v3.ConnectionFactoryImpl.doAuthentication(ConnectionFactoryImpl.java:451)
at org.postgresql.core.v3.ConnectionFactoryImpl.openConnectionImpl(ConnectionFactoryImpl.java:223)
at org.postgresql.core.ConnectionFactory.openConnection(ConnectionFactory.java:66)
at org.postgresql.jdbc.PgConnection.<init>(PgConnection.java:211)
at org.postgresql.Driver.makeConnection(Driver.java:407)
at org.postgresql.Driver.connect(Driver.java:275)
at org.apache.commons.dbcp.DriverConnectionFactory.createConnection(DriverConnectionFactory.java:38)
at org.apache.commons.dbcp.PoolableConnectionFactory.makeObject(PoolableConnectionFactory.java:582)
at org.apache.commons.dbcp.BasicDataSource.validateConnectionFactory(BasicDataSource.java:1556)
at org.apache.commons.dbcp.BasicDataSource.createPoolableConnectionFactory(BasicDataSource.java:1545)
... 35 common frames omitted
sonar.log:
2019.08.29 17:28:37 INFO app[][o.s.a.SchedulerImpl] Process [web] is stopped
2019.08.29 17:28:37 INFO app[][o.s.a.SchedulerImpl] Process [es] is stopped
2019.08.29 17:28:37 INFO app[][o.s.a.SchedulerImpl] SonarQube is stopped
<-- Wrapper Stopped
--> Wrapper Started as Daemon
Launching a JVM...
Wrapper (Version 3.2.3) http://wrapper.tanukisoftware.org
Copyright 1999-2006 Tanuki Software, Inc. All Rights Reserved.
2019.08.29 17:28:40 INFO app[][o.s.a.AppFileSystem] Cleaning or creating temp directory /opt/sonarqube/temp
2019.08.29 17:28:40 INFO app[][o.s.a.p.JavaProcessLauncherImpl] Launch process[es]: /usr/java/jdk1.8.0_131/jre/bin/java -Djava.awt.headless=true -Xmx1G -Xms256m -Xss256k -Djna.nosys=true -XX:+UseParNewGC -XX:+UseConcMarkSweepGC -XX:CMSInitiatingOccupancyFraction=75 -XX:+UseCMSInitiatingOccupancyOnly -XX:+HeapDumpOnOutOfMemoryError -Djava.io.tmpdir=/opt/sonarqube/temp -cp ./lib/common/*:./lib/search/* org.sonar.search.SearchServer /opt/sonarqube/temp/sq-process6725369398987844378properties
2019.08.29 17:28:45 INFO app[][o.s.a.SchedulerImpl] Process[es] is up
2019.08.29 17:28:45 INFO app[][o.s.a.p.JavaProcessLauncherImpl] Launch process[web]: /usr/java/jdk1.8.0_131/jre/bin/java -Djava.awt.headless=true -Dfile.encoding=UTF-8 -Xmx512m -Xms128m -XX:+HeapDumpOnOutOfMemoryError -Djava.io.tmpdir=/opt/sonarqube/temp -cp ./lib/common/*:./lib/server/*:/opt/sonarqube/lib/jdbc/postgresql/postgresql-9.4.1209.jre7.jar org.sonar.server.app.WebServer /opt/sonarqube/temp/sq-process3428525909039640490properties
systemctl status httpd:
[oksmart#CLOUDSVRSONAR01 logs]$ systemctl status httpd -l
● httpd.service - The Apache HTTP Server
Loaded: loaded (/usr/lib/systemd/system/httpd.service; enabled; vendor preset: disabled)
Active: active (running) since Thu 2019-08-29 16:59:00 CEST; 33min ago
Docs: man:httpd(8)
man:apachectl(8)
Process: 5697 ExecStop=/bin/kill -WINCH ${MAINPID} (code=exited, status=0/SUCCESS)
Main PID: 5834 (httpd)
Status: "Total requests: 50; Current requests/sec: 0; Current traffic: 0 B/sec"
CGroup: /system.slice/httpd.service
├─5834 /usr/sbin/httpd -DFOREGROUND
├─5835 /usr/sbin/httpd -DFOREGROUND
├─5836 /usr/sbin/httpd -DFOREGROUND
├─5837 /usr/sbin/httpd -DFOREGROUND
├─5839 /usr/sbin/httpd -DFOREGROUND
├─5900 /usr/sbin/httpd -DFOREGROUND
├─5901 /usr/sbin/httpd -DFOREGROUND
├─5903 /usr/sbin/httpd -DFOREGROUND
├─5904 /usr/sbin/httpd -DFOREGROUND
├─5905 /usr/sbin/httpd -DFOREGROUND
└─5906 /usr/sbin/httpd -DFOREGROUND
Aug 29 16:59:00 CLOUDSVRSONAR01 systemd[1]: Starting The Apache HTTP Server...
Aug 29 16:59:00 CLOUDSVRSONAR01 httpd[5834]: AH00558: httpd: Could not reliably determine the server's fully qualified domain name, using fe80::250:56ff:fe01:114a. Set the 'ServerName' directive globally to suppress this message
Aug 29 16:59:00 CLOUDSVRSONAR01 systemd[1]: Started The Apache HTTP Server.
/etc/httpd/conf.d/sonar.oksmart.es.conf :
<VirtualHost *:80>
ServerName localhost
# ProxyPreserveHost On
# ProxyPass / http://localhost:9000/
# ProxyPassReverse / http://localhost:9000/
TransferLog /var/log/httpd/sonar.oksmart.es_access.log
ErrorLog /var/log/httpd/sonar.oksmart.es_error.log
</VirtualHost>
NOTE: I commented the proxy options because if i comment out those lines, i get an error on the web page.
sonar.sh :
DEF_APP_NAME="SonarQube"
DEF_APP_LONG_NAME="SonarQube"
APP_NAME="${DEF_APP_NAME}"
APP_LONG_NAME="${DEF_APP_LONG_NAME}"
WRAPPER_CMD="./wrapper"
WRAPPER_CONF="../../conf/wrapper.conf"
PRIORITY=
PIDDIR="."
RUN_AS_USER=root
...
sonar.service:
[Unit]
Description=SonarQube service
After=syslog.target network.target
[Service]
Type=forking
ExecStart=/opt/sonarqube/bin/linux-x86-64/sonar.sh start
ExecStop=/opt/sonarqube/bin/linux-x86-64/sonar.sh stop
User=root
Group=sonar
Restart=always
[Install]
WantedBy=multi-user.target
Any idea of what am i doing wrong??
Firewall is disabled.
Thanks all!
Looks like the web container failed to startup because of database authentication.
Besides that, you commented out your proxy config, that's why you're seeing default Apache homepage.

Replace zookeeper server from zookeeper ensemble (with SolrCloud)

I have a SolrCloud cluster (6.6) setup with external Zookeeper Ensemble (3.4.8) of 5 nodes. Recently, one machine (ip1:port1) that run 1 Zookeeper with id=1 went down. This is what I've done to replace zookeeper:
Start zookeeper in another machine with the same id (=1).
Change zoo.cfg in 4 live zookeeper to match new zookeeper server and restart.
Update ZK_HOST variable in solr.in.sh to match new zookeeper server.
Restart solr.
After that, my solr cluster seemed to functioning well, but in solr.log, it looked like solr client and zookeeper servers still try to connect to the old zookeeper:
Solr log
2017-12-01 15:04:38.782 WARN (Timer-0-SendThread(ip1:port1)) [ ] o.a.z.ClientCnxn Client session timed out, have not heard from server in 30029ms for sessionid 0x0
2017-12-01 15:04:40.807 WARN (Timer-0-SendThread(ip1:port1)) [ ] o.a.z.ClientCnxn Client session timed out, have not heard from server in 31030ms for sessionid 0x0
Zookeeper log:
2017-12-01 13:53:57,972 [myid:] - INFO [main-SendThread(ip1:port1):ClientCnxn$SendThread#1032] - Opening socket connection to server ip1:port1. Will not attempt to authenticate using SASL (unknown error)
2017-12-01 13:54:03,972 [myid:] - WARN [main-SendThread(ip1:port1):ClientCnxn$SendThread#1162] - Session 0x0 for server null, unexpected error, closing socket connection and attempting reconnect
java.net.NoRouteToHostException: No route to host
at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:744)
at org.apache.zookeeper.ClientCnxnSocketNIO.doTransport(ClientCnxnSocketNIO.java:361)
at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1141)
2017-12-01 13:54:05,074 [myid:] - INFO [main-SendThread(ip1:port1):ClientCnxn$SendThread#1032] - Opening socket connection to server ip1:port1. Will not attempt to authenticate using SASL (unknown error)
2017-12-01 13:54:06,974 [myid:] - WARN [main-SendThread(ip1:port1):ClientCnxn$SendThread#1162] - Session 0x0 for server null, unexpected error, closing socket connection and attempting reconnect
I've done some search in add/remove zookeeper but didn't find a document for it. My zookeeper version (3.4.7) is not supported for dynamic reconfiguration (which is in zookeeper 3.5).
Is there a way I can manually remove/add zookeeper server from ensemble?
Thanks for your attention!

JBoss Configuration in RHEL 7.4 for missing data source

I am trying to deploy a .war file called LicenseManager.war. The data server is hosted on another machine with an ip address. I am the database drivers have been correctly imported. After trying to copy and paste the LicenseManager.war file in deployments folder, the deployment fails and a LicenseManager.war.failed is created. The following is the server log :
2017-09-07 14:38:47,802 INFO [org.infinispan.configuration.cache.EvictionConfigurationBuilder] (ServerService Thread Pool -- 59) ISPN000152: Passivation configured without an eviction policy being selected. Only manually evicted entities will be passivated.
2017-09-07 14:38:47,825 INFO [org.infinispan.configuration.cache.EvictionConfigurationBuilder] (ServerService Thread Pool -- 59) ISPN000152: Passivation configured without an eviction policy being selected. Only manually evicted entities will be passivated.
2017-09-07 14:38:47,859 INFO [org.infinispan.configuration.cache.EvictionConfigurationBuilder] (ServerService Thread Pool -- 59) ISPN000152: Passivation configured without an eviction policy being selected. Only manually evicted entities will be passivated.
2017-09-07 14:38:47,859 INFO [org.infinispan.configuration.cache.EvictionConfigurationBuilder] (ServerService Thread Pool -- 59) ISPN000152: Passivation configured without an eviction policy being selected. Only manually evicted entities will be passivated.
2017-09-07 14:38:48,415 INFO [org.jboss.weld.deployer] (MSC service thread 1-1) WFLYWELD0006: Starting Services for CDI deployment: LicenseManager.war
2017-09-07 14:38:48,602 INFO [org.jboss.weld.Version] (MSC service thread 1-1) WELD-000900: 2.3.3 (redhat)
2017-09-07 14:38:48,851 INFO [org.jboss.weld.deployer] (MSC service thread 1-1) WFLYWELD0009: Starting weld service for deployment LicenseManager.war
2017-09-07 14:38:49,303 INFO [org.jboss.as.clustering.infinispan] (ServerService Thread Pool -- 59) WFLYCLINF0002: Started client-mappings cache from ejb container
2017-09-07 14:38:49,642 ERROR [org.jboss.as.controller.management-operation] (Controller Boot Thread) WFLYCTL0013: Operation ("deploy") failed - address: ([("deployment" => "LicenseManager.war")]) - failure description: {"WFLYCTL0180: Services with missing/unavailable dependencies" => [
"jboss.persistenceunit.\"LicenseManager.war#LicenseManager\" is missing [jboss.naming.context.java.jboss.manageIDDatasource]",
"jboss.persistenceunit.\"LicenseManager.war#LicenseManager\".__FIRST_PHASE__ is missing [jboss.naming.context.java.jboss.manageIDDatasource]"
]}
2017-09-07 14:38:49,734 INFO [org.jboss.as.server] (ServerService Thread Pool -- 34) WFLYSRV0010: Deployed "LicenseManager.war" (runtime-name : "LicenseManager.war")
2017-09-07 14:38:49,755 INFO [org.jboss.as.controller] (Controller Boot Thread) WFLYCTL0183: Service status report
WFLYCTL0184: New missing/unsatisfied dependencies:
service jboss.naming.context.java.jboss.manageIDDatasource (missing) dependents: [service jboss.persistenceunit."LicenseManager.war#LicenseManager", service jboss.persistenceunit."LicenseManager.war#LicenseManager".__FIRST_PHASE__]
2017-09-07 14:38:50,197 INFO [org.jboss.as] (Controller Boot Thread) WFLYSRV0060: Http management interface listening on http://0.0.0.0:9990/management
2017-09-07 14:38:50,198 INFO [org.jboss.as] (Controller Boot Thread) WFLYSRV0051: Admin console listening on http://0.0.0.0:9990
2017-09-07 14:38:50,198 ERROR [org.jboss.as] (Controller Boot Thread) WFLYSRV0026: JBoss EAP 7.0.0.GA (WildFly Core 2.1.2.Final-redhat-1) started (with errors) in 18730ms - Started 578 of 983 services (135 services failed or missing dependencies, 385 services are lazy, passive or on-demand)
2017-09-07 14:38:50,410 INFO [org.jboss.as.clustering.infinispan] (ServerService Thread Pool -- 59) WFLYCLINF0003: Stopped client-mappings cache from ejb container
2017-09-07 14:38:50,541 INFO [org.jboss.as.server.deployment] (MSC service thread 1-2) WFLYSRV0028: Stopped deployment LicenseManager.war (runtime-name: LicenseManager.war) in 301ms
2017-09-07 14:38:50,724 INFO [org.jboss.as.repository] (DeploymentScanner-threads - 1) WFLYDR0002: Content removed from location /home/aboshi/EAP-7.0.0/standalone/data/content/7e/7e872f95551f1c9ff56e422fcc5a18b44a4e12/content
2017-09-07 14:38:50,724 INFO [org.jboss.as.server] (DeploymentScanner-threads - 1) WFLYSRV0009: Undeployed "LicenseManager.war" (runtime-name: "LicenseManager.war")
2017-09-07 14:38:50,783 INFO [org.jboss.as.controller] (DeploymentScanner-threads - 1) WFLYCTL0183: Service status report
WFLYCTL0184: New missing/unsatisfied dependencies:
service jboss.deployment.unit."LicenseManager.war".WeldBootstrapService (missing) dependents: [service jboss.deployment.unit."LicenseManager.war".component."com.sun.faces.taglib.html_basic.CommandButtonTag".WeldInstantiator, service jboss.deployment.unit."LicenseManager.war".component."com.sun.faces.taglib.jsf_core.RegexValidatorTag".WeldInstantiator, service jboss.deployment.unit."LicenseManager.war".component."com.sun.faces.taglib.jsf_core.SelectItemTag".WeldInstantiator, service jboss.deployment.unit."LicenseManager.war".component.LicenseTimerBean.WeldInterceptorBindingsService, WFLYCTL0208: ... and 63 more ]
service jboss.deployment.unit."LicenseManager.war".WeldStartService (missing) dependents: [service jboss.deployment.unit."LicenseManager.war".component."com.sun.faces.taglib.html_basic.CommandButtonTag".WeldInstantiator, service jboss.deployment.unit."LicenseManager.war".component."com.sun.faces.taglib.jsf_core.RegexValidatorTag".WeldInstantiator, service jboss.deployment.unit."LicenseManager.war".component."com.sun.faces.taglib.jsf_core.SelectItemTag".WeldInstantiator, service jboss.deployment.unit."LicenseManager.war".component.LicenseTimerBean.WeldInterceptorBindingsService, WFLYCTL0208: ... and 64 more ]
service jboss.deployment.unit."LicenseManager.war".beanmanager (missing) dependents: [service jboss.persistenceunit."LicenseManager.war#LicenseManager"]
service jboss.deployment.unit."LicenseManager.war".component.LicenseManagerAction.CREATE (missing) dependents: [service jboss.deployment.unit."LicenseManager.war".component.LicenseManagerAction.START] service jboss.deployment.unit."LicenseManager.war".component.LicenseManagerAction.JndiBindingsService (missing) dependents: [service jboss.deployment.unit."LicenseManager.war".jndiDependencyService]
service
Also, one of the main errors is :
6:34:19,720 ERROR [org.jboss.as.controller.management-operation] (DeploymentScanner-threads - 2) WFLYCTL0013: Operation ("deploy") failed - address: ([("deployment" => "LicenseManager.war")]) - failure description: {"WFLYCTL0180: Services with missing/unavailable dependencies" => [
"jboss.persistenceunit.\"LicenseManager.war#LicenseManager\" is missing [jboss.naming.context.java.jboss.manageIDDatasource]",
"jboss.persistenceunit.\"LicenseManager.war#LicenseManager\".__FIRST_PHASE__ is missing [jboss.naming.context.java.jboss.manageIDDatasource]"
]}
This is way too much to understand. Also, a similar configuration ran successfully on a windows 7 OS where again, the database was external. But this one is on RHEL 7.4. Can some please help me understand what is going on? Thank you very much!!
Regards,
Abhi Joe
It seems JBoss is not finding the datasource where it expects it:
jboss.naming.context.java.jboss.manageIDDatasource
You might first turn logging to 'debug' then start the server-- grep for any lines that deal with manageIDDatasource.
It's quite possible the deployment of the datasource wasn't as smooth as could be. Look for examples of doing this, then when there are no errors things should be better. (If there are still questions about the validity of the datasource, then maybe write a small servlet to retrieve and test it)

Google app engine: (devappserver2.py:835) No default module found.

I am trying to run a service with in google app engine. I used this command to run the local app engine development server gor Go and run the admin service.
goapp serve -admin_port 8081 admin
But it gives me error of "No default module found. Ignoring."
parminder_sohal#compact-window-137215:~/conference-central/default$ goapp serve -admin_port 8081 admin
INFO 2016-07-15 21:00:25,850 devappserver2.py:769] Skipping SDK update check.
INFO 2016-07-15 21:00:25,949 api_server.py:205] Starting API server at: http://0.0.0.0:57478
INFO 2016-07-15 21:00:25,974 dispatcher.py:197] Starting module "admin" running at: http://0.0.0.0:8080
INFO 2016-07-15 21:00:25,975 admin_server.py:116] Starting admin server at: http://0.0.0.0:8081
WARNING 2016-07-15 21:00:25,975 devappserver2.py:835] No default module found. Ignoring.

Resources