I'm start using Atom with the package Data-Atom for SQL Server connection.
But I keep getting this error when I execute a query:
Results:
"Error (ETIMEOUT) - Timeout: Request failed to complete in 15000ms"
But on Visual Studio the same query executes without errors, so the server is ok I presume.
There is a way to resolve this?
Update: The connection options weren't being passed through to node-mssql but after this PR is merged they will be.
In my attempt to set the timeout to one hour I added the below to data-atom-connections.cson.
{
name: "sqlserver01"
protocol: "sqlserver"
user: "dzamo"
password: "norton"
server: "sqlserver01.localdomain"
database: "AdventureWorks"
options: "requestTimeout=36000000"
}
And nothing happened. I had to go and hardcode it in ~/.atom/packages/data-atom/node_modules/tedious/lib/connection.js to get a long-running query to work. I'll investigate this a bit more and open a bug on github.
Related
I have a Java/Spring app that uses JDBC and HikariCP with a SQL Server instance,
I found this warning to occur too frequently in the logs:
HikariPool-2 - Failed to validate connection ConnectionID:2483
ClientConnectionId: c8a86a9c-b804-4060-a118-01378e040b59 (The
connection is closed.). Possibly consider using a shorter maxLifetime
value."
To solve this I tried to configure the data source as this:
jdbc-url: "jdbc:sqlserver://my-db.dev.aws-some-db.com;databaseName=MYHUGEDB"
username: "someUsername"
password: "somePassword"
maximumPoolSize: "10"
maxLifetime: "60000"
keepaliveTime: "40000"
I choose the maxLifeTime to be less then the SQL Server query time out,
To find the SQL Server time, I executed this SQL query:
EXEC SP_CONFIGURE 'remote query timeout'
It returns:
config_value: 600 (s)
run_value: 600 (s)
Normally 600s > 60000(ms), so it should not raise this warn, but it did. Why?
I have created an SSIS package (08) that has a script task which runs an SSRS report (08) with given parameters. The report creates a pdf version of the file on a file share drive of my choosing.
When I run the SSIS manually it works perfectly, however, when I upload SSIS into SQL and then run it from the job agent it fails. It gives a long error message but the basic piece is;
(0xC0016016 Source: Description: Failed to decrypt
protected XML node "DTS:Property" with error 0x8009000B "Key not valid
for use in specified state.". You may not be authorized to access
this information. This error occurs when there is a cryptographic
error. Verify that the correct key is available. End Error
Error: 2018-01-15 16:09:58.28 Code: 0xC001600C
Source:
PackageSchedule Connection manager "ReportServer08"
Description:
Server authentication failed. This error occurs when login credentials
are not provided, or the credentials are incorrect. )
I have tried to change the "ProtectionLevel" to "EncryptSensitiveWithPassword" and also "Donset upSensitive".
I setup a subscription to the report just to make sure it wasn't a read/write error to the specific folder and the subscription ran without error as well.
Any help on this would be greatly appreciated.
According to this SO answer the issue was related to a date type issue. You could be experiencing the same thing
Thank-you for the feedback. I was able to figure it out. On Import of the package to SQL itself I changed the 'Protection Level' to 'Rely on server storage and roles for access control'. This allowed the job to execute successfully and everything looks great!
I have a MongoDB database I have storing a large dataset. I am trying to import my database into MATLAB to perform data visualization, but am getting strange errors that I have not been able to debug in doing so. I have my database running using mongod on localhost on port 27017.
I do have the Database Toolbox add on for MATLAB, and the MongoDB specific one as well.
Following the example of the MATLAB documentation, I created variables:
server = "0.0.0.0" (I have also tried putting localhost here)
port = 27017
dbname = "NYPDData"
conn = mongo(server, port, dbname)
When I do so, I recieve the error that:
Error using error
Unable to load a message catalog 'mongodb:mongodb'. Please check
the file location and format.
Error in mongo (line 389)
error(message('mongodb:mongodb:DriverError',mongo.extractExceptionMessage(e)));
When I try simple running conn = mongo() with any parameters however, it seems to work, and returns:
conn =
mongo with properties:
Database: ''
UserName: ''
Server: {}
Port:
CollectionNames: {}
TotalDocuments:
If you have any insight on how to solve this error, please let me know. Thank you!
Type this before calling the mongo function : registerrealtimecataloglocation('c:\ProgramData\MATLAB\SupportPackages\R2017b')
It should return 1 if everything goes well. Don't forget to change the previous line by your number of version (R2017b here) with yours.
And then it should find the message in the catalog.
I get below error message when I try to execute the datflow task in SSIS. the connection manager connects successfully when I test the conn. I can also preview data from the OLE DB source. I get error when I try to execute though,
[[1]] Error: SSIS Error Code DTS_E_CANNOTACQUIRECONNECTIONFROMCONNECTIONMANAGER. The AcquireConnection method call to the connection manager "" failed with error code 0xC0202009. There may be error messages posted before this with more information on why the AcquireConnection method call failed.
[SSIS.Pipeline] Error: component "" (1) failed validation and returned error code 0xC020801C.
I tried the below change and nothing worked:
Set False to Run64BitRunTime
Tried Delay Validation with both true and false
Set Transaction Option not supported
Are there any other options which I should modify ? SSIS 2008 version is being used.The data flow tasks connects to a Oralce DB. the connection manager uses Oracle Provider fro OLE DB
I too ran into the same problem. I was using a config file and the package was reading Connection String form there.
To solve this. Open config file in notepad and type in the password Manually. Save it and run the package.
So this happens when you are creating a package configuration file. If you have selected the connection string while setting it up, it tries to fetch the connection string from the configuration file. So, open the configuration file and enter the password after the user id for the connection string value in this manner
User ID=[your user ID here];Password=[your password here];
And it will work
Sometimes visual studio behaves weird, just try deleting the task in SSIS package and add again.
I am running Access 2010 FE and SQL Server 2005 BE.
I can execute pass through queries to my SQL Server succesfully by using DSNless connections.
During my testing phase sometimes I need to restore my database to get back to my original records so I can rerun my pass through queries. What I have found is when I run a pass through query, it creates an active connection on my SQL Server. I see the connection via the SQL Server Management Console under the MANAGEMENT | SQL Server Logs | Activity Monitor, select view processes. There I can see which process ID is being used and who is using it when I run my pass through query.
Now the only way for me to restore my database is to KILL the PROCESS e.g. Active connection
Now when I have my restored database in place and re-run the pass through query, I receive a ODBC -- Call Failed message box. I have attempted to run a procedure to refresh my querydefs but to no avail, I will still get the ODBC-- Call Failed message box when I click on those objects.
Now there are two options on how to fix this problem, which in either case I find not USER Friendly.
Restart my Access Application
Wait approx 5-10 minutes to rerun the Pass Through Query
I created a function to trap my ODBC Errors and this is what appears:
ODBC Error Number: 0
Error Description: [Microsoft][ODBC SQL Server Driver]Communication link failure
ODBC Error Number: 3146
Error Description: ODBC--call failed.
So if for some reason, I need to restart my SQL server or kill a process (Active Connection) on my SQL server while the Access Application is currently connected via ODBC, the objects created via ODBC will not perform properly till I execute the 2 workaround solutions as stated above.
Can anyone shed some advice on a solution? I appreciate any insight.
I asked a similar question some time ago, and never got a satisfactory answer. My original question is here: Force SET IDENTITY_INSERT to take effect faster from MS Access
There is a registry setting documented here for ACE that controls the timeout behavior:
ConnectionTimeout: The number of seconds a cached connection can remain idle before timing out. The default is 600 (values are of type REG_DWORD).
So as a third workaround (in addition to the two you already listed) you can change that registry setting to a shorter timeout (like 10 seconds). This is the approach I took in my answer. One caveat is that shortening the timeout may cause performance or other issues. Your mileage may vary.
See my full answer to the original question for more info.