Today team reports an error of TFS 2012:
TF30042: The database is full. Contact your Team Foundation Server administrator
Event Log from data base server:
MSSQLSERVER: Could not allocate space for object 'dbo.tbl_NotificationQueue'.'PK_tbl_NotificationQueue' in database 'Tfs_Real-Time collection' because the 'PRIMARY' filegroup is full. Create disk space by deleting unneeded files, dropping objects in the filegroup, adding additional files to the filegroup, or setting autogrowth on for existing files in the filegroup.
MSSQLServerOLAPService: An error occurred while writing a trace event to the file, \?\C:\Program Files\Microsoft SQL Server\MSAS10_50.MSSQLSERVER\OLAP\Log\FlightRecorderCurrent.trc.
The C driver is no space available to use.
Could anyone give some suggestions?
Option 1: Clean the drive to free up space. Delete transaction logs, look for extraneous test case attachments, build drops checked into source that sort of thing.
Option 2: Get a bigger hard drive.
In my case the table "tbl_JobHistory" had become much too large (>7.5GB). As a result, the 10GB of the SQL-Express database were fully utilized.
With Microsoft SQL Server Management Studio you can easily view the largest tables: "Reports" > "Standard Reports" > "Disk Usage by Top Tables".
I deleted the data in "tbl_JobHistory" and then shrinked the database.
Screenshot
Related
Slight difficulty with Team Foundation Server 2008 databases (using Sql Server 2005).
Team System reported running out of disk space so I added another data file to each and every Team System database (on another fixed drive) to increase the disk space available. I did it as follows for each TFS database:
Add another file, type=Data, size etc = defaults, path to new drive.
Change the original data file to stop further expansion (Autogrowth=none).
Restart Sql Server.
That is the documented method for adding further space to a Sql Server 2005 database. However now I find I can check out files, but not check in. In Visual Studio, Source Control Explorer displays projects and can be expanded to show folders and files as normal, but When I rt-click on a checked-out file and select 'check in', nothing happens and the file remains checked out, and then Visual Studio has to be terminated using Task Manager.
There don't seem to be any related errors in the server Event Log.
What have I missed or done wrong?
You need to use the 'Database Mirroring' feature to add a database file for a mirrored database that has different drives and/or paths.
Firstly, check this MSDN article for the detailed steps on how to "Configure SQL Server Mirroring for the TFS Data-Tier Server": https://msdn.microsoft.com/en-us/library/aa980629(v=vs.90).aspx
Secondly, follow steps in this article about how to "Fail Over to a Mirrored Data-Tier Server if the Principal Server is Unavailable": https://msdn.microsoft.com/en-us/library/aa980528(v=vs.90).aspx
"clean the TFS cache", posted by TfsAgent, actually solved the problem. Nothing to do with adding extra database files after all - that was a coincidence.
I have the following problem with Team Foundation Server Express 2015.
I ran out of SQL Database Space. So I cleaned up a littlebit and used TFS Destroy to permanently delete some larger files from source control.
I expected that the Size auf the table dbo.tbl_Content would get less after some time. But the opposite happened. 3000 Entries where added to the table and it got larger.
How can I cleanup TBL_Content from old files that have been destroyed already?
The TBL_Content Table has now 8.8 GB.
Executing stored procedures like prc_DeleteUnusedContent and [dbo].prc_DeleteUnusedFiles did not help at all.
Orphan data still leaves after running the tf destroy command, and will be deleted later by a TFS Background job agent. That is the reason why table dbo.tbl_Content size doesn't drops.
Here are some steps you can try to reduce the size of TFS databases:
1). Download and install TFS2015 Power Tool to get the Test Attachment Cleaner to remove large test attachments:
Tcmpt.exe attachmentcleanup /collection: TeamProjectCollectionUrl /teamProject: TeamProject /settingsFile: SettingsFilePath [/outputFile: ReportFilePath] [/mode: Preview/Delete]
2). Remove old workspaces which is no longer used for long time.
3). Delete old unused team projects: https://msdn.microsoft.com/en-us/library/ms181482(v=vs.140).aspx
While adding many 1000s of files to the online version of TFS at visualstudio.com, I have started getting "The decryption operation failed, see inner exception".
I cannot even resolve or check in a single file (from the new batch of 34,000+ additions).
As the message is displayed in the pending changes window, there is no way that I can see to get more details (or view the inner exception).
1) Is this a TFS, VS 2013 or network issue
2) How can I get at the more detail information of the error in VS 2013?
Before anyone suggests "turn it off and on again", rebooting has not made any difference to the problem.
Also switching the advanced workspace location from "local" to "server" made no difference.
Have we just hit a wall in TFS?
*Note: It does not effect checkin to other folders outside of the new additions
It appears that the online TFS service provided through visualstudio.com has some limitations.
The solution to the problem above was to undo the 34,000+ item check and add them in smaller batches. Up to around 10,000 files at a time seems to behave itself.
Update:
I have just had the same error on a fresh machine install, where I tried to pull down my entire TFS repository in one go. Again pulling them down piece-by-piece avoided the error
Is it possible to archive old data from MS Team System Source Control? We have up to 10,000 changesets in our TFS database, and are unlikely ever to want to go that far back, but the database files are getting ever larger. Is it possible to archive everything before a certain date (say) to reduce the size of the database? if not, taken to its logical conclusion a long lived actively developed project will eventually explode.
You can back up the whole TFS database and then use destroy command to permanently delete the source control tree. But you can do that only for whole source control paths, not specific versions.
In case there are many deleted files you can also use the same command to review the source control tree and permanently destroy these deleted files.
You will not be able to just delete old changesets because TFS stores incremental changes every time you check-in. This also implies that maybe you have large binary files stored in TFS - files larger than certain limit are not stored incrementally (thus saving space) but instead every version is stored complete.
Documentation on the command: http://msdn.microsoft.com/en-us/library/bb386005(v=vs.100).aspx
Got the below in TFS and VS 2012 RC, anyone know of a fix? Doesn't seem to exist on MS website.
TF400018: The local version table for the local workspace
COMPUTERNAME;MYNAME could not be opened. The process cannot access the
file because it is being used by another process
Any suggestions welcomed.
We experienced this one as well. Migrating to the RTM makes this happen a lot less, but it can still happen a lot.
When using local workspaces (a new feaure in vs 2012) a local file based database is created to administer changes you make localy. When you change a source file, this file base database needs to be updated. If this update conflicts with the normal update task which routinely checks for changes you get this error. The cause of this issue is usually that you are using local workspace for more items than it was intended or that your disk I/O is too slow.
Workarounds for this are either:
Replace your disk with an ssd. Having better I/O makes this issue
happen a lot less.
Switch back to server based workspaces. (which handles this better)
Use the TFS-GIT connector and use git for offline support.
Split your workspace mapping in portions so they contain less items.
Please delete the files under %Temp% folder and open the project as
"Run as Administrator " mode .It works for me .
Regards,
Kamaraj