Tempdb disk getting full in SQL Server - sql-server

Currently we are facing issue with tempdb which was allocated with 200 GB. Mostly disk is getting full 2 times/day and with help of DBA, spaces are clearing by using shrink statement.
Three months before server was re-started suddenly after which tempdb issues started. We checked with db team to identify any option were disabled which were related to clear the log automatically. But they confirmed everything were looks fine.
Can you tell me where to start as we don't have access to system tables in production system?
Thanks in advance

Related

Large USERSTORE_OBJPERM in SQL Server

I'm trying to make our SQL Server go faster and have noticed that no stored procedures are staying in the plan cache for any length of time. Most have of the plans have been created in the last hour or so.
Running the script below I see that the USERSTORE_OBJPERM is around 3GB and is the 2nd biggest memory cache on the server after the SQL BUFFERPOOL.
SELECT top 100 *
FROM sys.dm_os_memory_clerks
where type = 'USERSTORE_OBJPERM'
I've run the same script on a few other of our production servers and none of the USERSTORE_OBJPERM on the other servers are any where near as large around 200MBs.
My question is has anyone seen a USERSTORE_OBJPERM at around 3GB and what might of caused it.
I ran the following to try and clear the cache, it went down down by a 100mb or so and instantly started rising again.
DBCC FREESYSTEMCACHE ('ObjPerm - DatabaseName')
Results of script
SQL Server version is 2017 Enterprise with CU22 applied.
Many Thanks in advance for any tips or advice provided
Cheers Mat
Fixed.
It seems the issue was caused by an application using service broker.
The application was running a script to check permissions every 30 seconds.
Fortunately there was an option to switch the permission check off.
The USERSTORE_OBJPERM cache size is now 200MBs instead of 3GB and stored procedure plans are staying in the cache.

SQL Server on Azure has slowed to half speed

We recently ran out space on a Azure database. After deleting a lot of unused tables, (none of which had indexes, or even keys,) Transact queries take almost exactly twice as long. SSIS imports seem to be completely unaffected.
After the obvious step of shrinking the log file, I'm now totally baffled.
Any advice would be greatly appreciated.
After you deleted a lot of unused tables, the database space will not be released automatically in SQL Server.
For a test, you can migrate you data to another database and run the same query, comparing the query time it costs.
Since you have tried to shrink the log file, I suggest you can also try the database shrinking.
Login you SQL Database with SSMS:
right-click your database---->Tasks ---->Shrink---->Database:
Another way is setting database Auto Shrink:
right-click your database---->Properties, set Auto Shrink---->True:
Hope this helps.

Symmetric DS taking too long on DataExtractorService

I'm using SDS to migrate data from a SQL server to a Mysql database. My tests of moving the data of a database that was not in use worked correctly though they took like 48 hours to migrate all the existing data. I configured dead triggers to move all current data and triggers to move the new added data.
When moving to a live database that it is in use the data is being migrated too slow. On the log file I keep getting the message:
[corp-000] - DataExtractorService - Reached one sync byte threshold after 1 batches at 105391240 bytes. Data will continue to be synchronized on the next sync
I have like 180 tables and I have created 15 channels for the dead triggers and 6 channels for the triggers. For the configuration file I have:
job.routing.period.time.ms=2000
job.push.period.time.ms=5000
job.pull.period.time.ms=5000
I have none foreign key configuration so there wont be an issue with that. What I would like to know is how to make this process faster. Should I reduce the number of channels?
I do not know what could be the issue since the first test I ran went very well. Is there a reason why the threshold is not being clreared.
Any help will be apreciated.
Thanks.
How large are your tables? How much memory does the SymmetricDS instance have?
I've used SymmetricDS for a while, and without having done any profiling on it I believe that reloading large databases went quicker once I increased available memory (I usually run it in a Tomcat container).
That being said, SymmetricDS isn't by far as quick as some other tools when it comes to the initial replication.
Have you had a look at the tmp folder? Can you see any progress in file size. That is, the files which SymmetricDS temporarily writes to locally before sending the batch off to the remote side? Have you tried turning on more fine grained logging to get more details? What about database timeouts? Could it be that the extraction queries are running too long, and that the database just cuts them off?

SQL Server 2008 DBNETLIB error

Our ASp.net application is getting error as below"
[DBNETLIB][ConnectionOpen (Connect()).]SQL Server does not exist or access denied "
I can connect with Enterprise manager management studio and Query analyzer without any issue.
It was running these applications with out any issue long time. last one week we are getting this error.If we restart the server .it works then it will come again after 3to 4 hours.
We are running on Windows 2003 server. I was searching and didn't find a solution yet. If anybody knows anything for this error, please post the details to resolve.
Thank you in Advance
Joseph
Sounds like you are running out of space on the drive that contains the tempdb. The tempdb is purged on a restart but slowly grows as queries are executed and connections are made. Check the drive when you start having the problem. If you're within a few megabytes of zero, then that's the problem. Clear-up some hard drive space, move the tempdb to another drive, or create multiple tempdb files on multiple drives.
Could also be a problem with RAM, but it's more likely to be an issue with the tempdb.

SQL Server Express Performance breaks with large Logfiles

we run since 2 years a small application on SQL Server 2005 Express Edition the Database has gown from 75 MB up to nearly 400MB within this time, the there isn't a big amount of data.
But the log file has been arrived at 3,7GB now without changing Hardware, table structure or Program code we noted that the Import processes which required 10-15 minutes are now arrived at a couple of hours.
any idea where could be the Problem? Depends it on the log file may be? The 4GB Lock of Express Edition bear only on data files or also on log files?
Additional Informations: There isn't any RAID on the DB Server, There doesn't work concurrent users (only one user is logged in while the import process).
Thanks in Advance
Johannes
That the log file is so large is completely normal behavior; in the two years you have been running; sql has been keeping track of the events that happen in the database as it goes along its business.
Normally you might clear these logs off when you take a backup (as you most likely dont need them anyway.) If you are backing up all you need to change the sql script to checkpoint the logfile (its in books online) depending on how you are backing up your milage may vary.
To clear it down in the immediate make sure no one is in using the database; open management studio express find the database and run
backup log database_name with truncate_only
go
dbcc shrinkdatabase('database_name')
From MSDN:
"The 4 GB database size limit applies only to data files and not to log files. "
SQL Server Express is also limited in that it can only use 1 processor and 1GB of memory. Have you tried monitoring the processor/memory usage while the import is running to see if this is causing a bottleneck?

Resources