Websphere 7.0, EJB3.0
Is there a way to avoid default EJB timer persistence. Looks like timers survive server crashes and restarts.
Thanks!
Non persistent EJB timers added in Java EE 6 (WebSphere 8 and up).
If you need non persistent timers in WebSphere 7 you should probably need to use AlarmManager
The preferable way with EJB-3.0 is to clear all timers on server shutdown/startup to avoid timers persisted due to failure.
You can place code in appropriate methods - onApplicationEnd or onApplicationStart by implementing ApplicationListener interface, alternatively you can refer here, but former is better.
Else, can run the command to explicitly cancel timers as mentioned in documentation.
cancelEJBTimers
This command cancels and removes from persistent storage EJB timers based on the specified filter criteria.
The syntax for this command is: cancelEJBTimers server filter [options] filter: -all | -timer | -app [-mod [-bean ]]
-all
-timer timer id
-app application name
-mod module name
-bean bean name
options: -host host name
-port portnumber
-conntype connector type
-user userid
-password password
-quiet
-logfile filename
-replacelog
-trace
-help
In EJB-3.1, with TimerConfig you can configure it with setPersistent(isPersistent).
Related
According to https://cloud.google.com/trace/docs/setup/php, App Engine flexible environment for PHP can run a daemon that sends trace spans to Stackdriver in the background rather than as part of the request processing (which could cause increased response latency).
I am running Kubernetes Engine, but would still like to send trace requests in the background. Therefore:
Is it possible to run that batch daemon myself?
Out of curiosity, how does the Stackdriver PHP Exporter pass these spans to the daemon? I tried to search for that in the source code, but could not find out how it is done.
If #1 is not possible, is there another way to perform span sending in the background?
Stackdriver Trace with Google Cloud Run seems to cover a similar topic, but does not address how to run the daemon manually.
In case anyone else is looking for this, I was able to run the batch daemon as follows:
sudo -u www-data -E vendor/bin/google-cloud-batch daemon
Note that the daemon itself must be run as the same user as your “serving” PHP processes in order to access SysV memory shared between both, hence the sudo.
You will also need the PHP sysv and pcntl extensions enabled.
I have Nagios setup for Passive monitoring and we don't use RRD data for any purpose. Is there a way to stop data storage and calculation at RRD. (intent here is to save system resource by stopping RRD calculation)
There's a setting in the main nagios.cfg for process_performance_data, set that to 0 and restart Nagios.
https://assets.nagios.com/downloads/nagioscore/docs/nagioscore/3/en/configmain.html
has anyone from you pulled server hardware data from Nagios to build an inventory? Basically I am trying to create an inventory of an existing servers and hardware components monitored with nagios (i.e. hostname, CPU, MEM, HDD and etc.) There are other ways to do it, but maybe there is a plugin or a way to pull it directly from nagios?
Thanks
You could run this on the command line against your configuration files:
grep "host_name" /usr/local/nagios/etc/objects/hosts.cfg >>output.log
That will output a list of all parameters associated with the host_name line within your configuration.
There is a new mk_inventory plugin for nagios which will pull HW info from the machine
Hi I have just installed a clean copy of Nagios and Check_MK. But I don't understand how they work together. Nagios uses nrpe to connect to clients and performs checks. This means that some Nagios plugins have to sit on the client and return results from when they are called. But how does Check_MK tie into Nagios. Does it use check_mk_agent to replace all the Nagios plugins to perform its checks? Also does the Nagios configurations all have to be fully configured with all the clients already in place to be checked and then ported to Check_MK interface (wato) or can the clients be added to Check_MK without being present in the Nagios configurations. This is where my confusion lies and I cant find a concrete answer to this question anywhere. Please help.
Check_MK uses Nagios core for theses tasks:
Manage Check results
Triggering of alarms
Manage planned downtimes
Test host availability
Detect network failures
As you can see at the bottom of this page: http://mathias-kettner.com/checkmk_monitoring_system.html
Check_MK needs both: client side monitoring agent and server side monitoring system.
The server side monitoring system calls the agent of the host and passes the check results to the monitoring core (usually Nagios but there is also an new core just for Check_MK).
What makes Check_MK different from other passive Checks (like NRPE) is that the results for all checks is send to the monitoring system in one package. If you run the agent on a host in a shell it will return something like this:
➜ ~ check_mk_agent
<<<df>>>
/dev/mapper/MyStorage-rootvol ext4 15350768 13206900 1341052 91% /
dev devtmpfs 4022348 0 4022348 0% /dev
plus many more lines ....
So the server side part of Check_MK splits these packages into single checks so that the Nagios core can handle them.
So Check_MK wont replace your existing checks, it doesn't care about them. It will just add more.
You don't necessary need WATO to configure Check_MK. WATO is just an interface for the configuration. Configuration can also be done with plain text files. You should start with WATO and take a look at the configuration it has generated.
I want monitor my ntpd services and ntpd service is used udp protocol。nagios have a method that is named check_udp,As follows:
[root#localhost]# ./check_udp -H 127.0.0.1 -p 123
With UDP checks, a send/expect string must be specified.
I want to know how to use check_udp...
I suggest saving yourself some effort and use one of the many NTP nagios plugins like the one below from the official nagios plugins package.
http://nagiosplugins.org/man/check_ntp_peer