How do I turn SQL logging on in Postgres 8.2? - database

I've got the following settings in my postgres conf:
log_destination = 'stderr'
redirect_stderr = on
log_directory = '/tmp/psqlog'
log_statement = 'all'
And yet no logs are logged. What am I missing here? There is reference on the internet to a variable called "logging_collector", but when I try and set that, postgres dies on startup with a FATAL: unknown variable.
This is on MacOS 10.4.
Ta.

I believe that you need to change log_destination to "syslog" or a specific directory. Output that goes to stderr will just get tossed out. Here's the link to the doc page, but I'll see if I can find an example postgresql.conf somewhere http://www.postgresql.org/docs/8.2/static/runtime-config-logging.html
This mailing list entry provides some info on setting up logging with syslog http://archives.postgresql.org/pgsql-admin/2004-03/msg00381.php
Also, if you're building postgres from source, you might have better luck using a os x package from Fink or MacPorts. Doing all of the configuration yourself can be tricky for beginners, but the packages normally give you a good base to work from.

Related

SQLAlchemy/pgAdmin: Error: password authentication failed for user "root"

I'm conducting a study and I need to store some data. I found an open source data scraper and parser online: https://github.com/hicsail/materials
I've followed some instructions (some of which but not all came from here) and
installed Postgres, created a docker-compose.yml file, and created a config file:
Above is the config file, and this is the .yml file
I started by going into the pgAdmin folder and running "docker-compose up", after which, this was the result:
I'm not sure if the "no privileges flag" means anything. Anyway, after this, I opened up a localhost:5050 in my browser and logged into pgAdmin.
I named the database "materials" as this was what it was supposed to be named.
Same thing with the username and password; both were named "root". However, when I run the command to parse the data, I'm getting this error:
I've been stuck on this for a long time now, and I can't seem to find any solution. This is running in a python2.7 conda environment, as per the requirements. These were the other installed libraries (I'm not 100% sure those were the exact versions, but I tried to get them as close as possible).
psycopg2==2.7.3.1
requests>=2.20.0
SQLAlchemy==1.0.9
wheel==0.24.0
If I need to clarify anything please let me know.
Thanks.
I found an answer for this. I had to go into services, then to postgres, and stop it from running. I then had to kill the port "5432" and run it again.

Credentails Manager service is stopped

Hi,
I am getting the below error when trying to update my view . Also, creds manager service is stopped and not getting started.
As long as the credmanager.exe is not started, none of the view (snapshot or dynamic) will work. So you need to debug that first.
Typically, you would check the ClearCase logs, and the Windows event, to see what the issue is. See "how to fix or investigate 'Operation “view_ws_is_ws_view” failed'?", using cleartool getlog.
That kind of error ("This application has failed to start becauselibatriaks.dllwas not found. Re-installing the application may fix this problem") was seen with rather old versions of ClearCase (7.0.1).
See for instance this thread.
If the DLL is present and the directory is in your path you should never
get this error.
Check to see how far ...\rational\clearcase\bin is down your SYSTEM path, and make sure the DLL's are in there.
ccgzip.exe will be called (by the client process) when checking in any files that fall back to the "compressed_file" element type. It may be called by the view server during file opens and checkouts to construct cleartext if needed. The former uses your user path, the latter will normally use the SYSTEM path.
We've seen odd behavior on Windows when the path gets >500 characters
long, though there seems to be no "official" limit to the length this environment variable (%PATH%) can grow to.
The error message you see is likely a path issue. Whether that's related to the other issue is unknown...
Is the albd server process running? The credential manager service is flagged to depend on this service. If that service fails to start, or terminates, the credential manager service will also fail.
A service startup failure should be in the Application or System Event logs. and from there the "Troubleshooting albd startup failures" technote on ibm.com would likely be a good place to start.

pnp4nagios not logging performance data for new host

We've just updated Nagios from 3.5.x to the current version (4.0.7) and subsequently added a new host for monitoring.
The new host shows as 'Down' in Nagios, and this seems to be related to the fact that pnp4nagios is not logging performance data (the individual checks for users, http etc are all find).
Initially there was an error that the directory
/usr/local/pnp4nagios/var/perfdata/newhost.com
that contains the xml setup and rrd files for the new host was missing), so I manually created this directory, but now it complains that the files are missing.
Does anyone know the appropriate steps to overcome this issue?
Thanks,
Toby
PS I'd tag this 'pnp4nagios', but that tag doesn't exist and I can't create them
UPDATE
It's possible that pnp4nagios is a red herring/symptom. Looking more closely I realise that Nagios actually believes the host is down, even though all services are up. The host status information is '(Host check timed out after 30.01 seconds)'...does this make any more sense?
It's indeed very unlikely that pnp4nagios has something to do with your host being down. pnp actually exports output and performance data to feed the rrd database and xml files (via npcd module or evenhandler command).
The fact that nagios reports the host check timed out after 30 sec means that :
- you have a problem with your host check command, please double-check the syntax
- this check command times out after a certain timelapse (most likely defined in nagios.conf) because the plugin was still running.
I'd recommend running this command from the server's prompt. You want to do something like :
/path/to/libexec/check_command -H ipaddress -args
For example:
/usr/local/libexec/nagios/check_ping -H 192.168.1.1 -w 200,40% -c 500,80% -timeout 120
See if something might be hanging. Having the output would be helpful.
Once your host check returns correct output and performance data to nagios, pnp will hopefuly do the rest.
In the unlikely event it helps anyone, pnp4nagios was indeed a red herring. The problem was that ping wasn't enabled for the host being checked, and this is the test for whether a host is up or not. Hence this was failing, despite other services being reported as working.

cakephp: warning 512 /tmp/cache/ not writable on shared host justhost

When I go to www.merryflowers.com/webroot/ i'm getting the following warnings. Based on the guidance i got from my previous post (cakephp: configuring cakephp on shared host justhost), I right clicked on the app/tmp/ (on the remote server) and all the folders within that and set the permission to be writable (ie. 777). But I'm still getting the same warnings.
Since i'm using windows 7 (chmod doesn't work), I also tried CACLS on the command prompt for tmp folder. Since i'm not familiar with CACLS, i don't know the exact command to make tmp writable to all. Can someone please help me out. Thank you.
Warning (512): /home/aquinto1/public_html/merryflowers.com/tmp/cache/ is not writable [CORE/cake/libs/cache/file.php, line 278].php, line 429
Warning (512): /models/ is not writable [CORE/cake/libs/cache/file.php, line 278]
Warning (512): /persistent/ is not writable [CORE/cake/libs/cache/file.php, line 278]
Is your site hosted locally on your Windows machine, like through XAMPP or WAMP, etc? Those are *nix paths, not Windows paths.
Did you FTP to your sites - like, with an FTP client - and change the permissions? Doing this through FTP clients isn't always 100% reliable. It looks like you changed the perms on /tmp, but they didn't cascade to /tmp/cache, etc. like you thought. Try setting them all one by one.
According to your other post - cakephp: configuring cakephp on shared host justhost - your site is set up with remote hosting. I looked at their service briefly, from the looks of them, you can probably remote (aka, "shell" or "ssh") into your server and get access to the command line. A lot of webhosts provide this these days, although you may have to specifically request they enable it for you.
On a Windows machine, you can use PuTTY to shell into your remote server: http://www.chiark.greenend.org.uk/~sgtatham/putty/
HTH. :)

For some reason when I link against Oracle GSS starts behaving differently (Solaris)

My test program works fine. I can create a client and a server and run them against each other. I can set my KRB5_CONFIG environment variable and use a local configuration for testing.
For some reason when I place the code in our production software it fails. Even if I strip our main() function to just calling gss_import_name() with a hard coded name I end up with the message "Cannot open configuration file".
If I run truss then I see a lot of Oracle going on. It tries to open lots of different Oracle trace files. It also tries to open
/krb5/krb.conf
instead of the file I specify.
It's as if Oracle is giving us the wrong gss, or maybe some other option in our huge and complex build system. I note -L/usr/lib/sparcv9 though this is after my -lgss now if that matters (too long since I worked in C on a regular basis!). The libgss.so.1 in that directory is larger than the one in /usr/lib - though putting that option into my test program's link command does not break it.
Any help?
Thanks
- Richard
This fixed what appeared to be a similar problem for us:
export KRB5_CONFIG=/etc/krb5.conf
It does appear likely that Oracle sets this env var incorrectly if it's not already set.
$ grep -r KRB5_CONFIG $ORACLE_HOME
Binary file /usr/lib/oracle/11.1.0.1/client64/lib/libclntsh.so matches
Binary file /usr/lib/oracle/11.1.0.1/client64/lib/libclntsh.so.11.1 matches
$ grep -r '/krb5/krb.conf' $ORACLE_HOME
Binary file /usr/lib/oracle/11.1.0.1/client64/lib/libclntsh.so matches
Binary file /usr/lib/oracle/11.1.0.1/client64/lib/libclntsh.so.11.1 matches
I found that the Oracle libraries contained an implementation of GSS. To make my code work I ensured I linked "-lgss" before linking any of the Oracle libraries.
I've not tested to see if this upsets Oracle in single sign-on, because we use Oracle with user name and password. That works fine.
I ran in to the very same issue with Oracle 11.2.0.4.0 on HP-UX 11.31 and wasted almost an entire day for that. Indeed, the crappy Oracle lib peforms a putenv with /opt/krb5/krb.conf and the tip from Richard Corfield makes the app even crash. The only workaround is to create a symbolic link. I have created a service request with Oracle for that issue.
Update (2014-06-02): I have received an update from Oracle. They confirmed the bug. It seems like there is a private GSS-API which is redefining symbols.
Bug 10184681 - ORACLE NEEDS TO USE VERSIONED SYMBOLS TO AVOID EXTERNAL SYMBOL CONFLICTS
This issue has been open since 2010-10. Terrible.

Resources