Where is the build log located in Team Foundation Server 2010? - tfs

I am looking for the physical location of the TFS workflow log. Not the MSBuild Log but the log you see when you click "View Log".
I have read in other forums that the log is stored in an SQL database. But which database is it and specifically what table?
Thanks. Hope someone knows the answer to this.

The entire build log can be found in the Tfs_YourTeamProjectCollection database in the Tbl_BuildInformation. The Tbl_BuildInformation table contains self referencing entries to preserve the build log hierarchy, and all the build information is saved in XML. The information is saved as different node types which correspond to different build activities.
If you are interested in going down that route, I go into a little more detail in my blog. It is a bit much to post here.

You can view the log file by using the TFS web access link. See this MSDN post for more details. I just used this trick and it was fantastic. http://social.msdn.microsoft.com/Forums/en/tfsbuild/thread/14ce49ea-f61a-45e2-be32-beffc81ce415?prof=required
Another way to view the log is via the web access(right click on the build output tab and choose "copy full path" and paste it onto a browser) and the log does get downloaded so you can scroll down quickly.

The build log is a different view of the full which by default can be found under the [build location][Build name]\logs directory.
You can retrieve the list of log locations from the database by querying tbl_Build table in the Tfs_<CollectionName> database, e.g.:
SELECT [BuildId],[StartTime],[FinishTime],[LogLocation]
FROM [Tfs_DefaultCollection].[dbo].[tbl_Build]

Related

Export Data from Avaya CMS

Bit of a long one...
I am trying to reliably export data from Avaya CMS, currently the set up I am using is a Win 2012 Server with Avaya CMS Superviser R18 installed. I have 5 report scripts which I run and they export data to a individual csv files every 3 - 5 seconds. This way was sufficient while we were using CMS Superviser R17 as it would only crash once a day, and when it did, the csv files would stop updating which triggered an email to me once the files had been 120 seconds without updating. This ensured that the reports had very little downtime as I could just quickly restart CMS superviser and the scripts (batch file) rather than having to wait for someone to report that it was no longer updating.
However with CMS SUperviser R18, the software doesn't seem to crash, instead the csv files keep getting updated, but with incorrect data. I am still trying to troubleshoot that issue to find out exactly what is causing it to go out of whack, but while that is ongoing I am also trying to look at alternative solutions.
The reports I am trying to export are custom reports that others in the company have created, and they appear under the designer category
I looked at CLINT but I cannot see the reports I am looking for in there so I ssh'd into the server and using xterm terminal I can view the CMS menus and go through them. In here I noticed that you cannot view any reports under the Integrated tab, nor any reports under the Designer Category of the Real-Time or Historical tabs. I also noticed that there is a "custom reports" section in the menu, although I see no reports under this menu item, just two empty categories (Real-time and Historical)(Pics attached)
So my questions are:
1 Do you know if "Custom Reports" is where reports from the designer category should show up? If not, do you know is it possible to access those reports through the ssh menu at all?
2 Has anyone any other ideas as to how to export the data from CMS to a csv file?
Thank you for your time.
The "Custom Reports" in the cms menu you show on the screenshots is different to "Reports" in the Supervisor.
You can read about Custom Reports here (a little old, but still relevant):
https://downloads.avaya.com/elmodocs2/multivantage/215822_3/215822_3_1_CMS_Reports.pdf
If you create reports in the "Custom Reports" subsystem then you can run them using /cms/toolsbin/clint. Note that you must recreate the "Reports" reports in the "Custom Reports" to be able to run them through clint.
You can create command files so you can script clint, then you just have to run it with a scheduler. You can read about scriping clint here:
http://www.tek-tips.com/viewthread.cfm?qid=397058
You can also establish a direct connection to the Informix database on CMS but you can not access real-time data there.

Opshub Users mapping not showing complete list of users ?

I am having a issue migrating projects to visual studo online, i am using the ops hub utilty but when i get to the mapping users screen it is not showing complete list of users form our directory . Please advise
You will have to add the relevant users as Team Members/Readers/Contributors. (typically, you will have to make them a part of valid users of the project) and they'll load up in the user mapping screen.
Considering large number of users at collection level who may or may-not be part of all the projects, the newer version of the tool has been modified to load only users of the project(s) that are selected for migration making the user-mapping a little more convenient.

Pentaho 5.0.1 CE Adding folder to pentaho-solutions directory

How do I get a folder to appear in the browser after adding it to \pentaho-solutions in Pentaho 5.0.1 CE. The documentation I read does not seem to work with this new version.
1. Create a new directory in /pentaho/server/biserver-ee/pentaho-solutions/.
Use underscores instead of spaces in the solution directory name. Ensure that the directory has the appropriate user
and group ownership to be writable from the BI Platform.
2. Using an XML-aware text editor (or Design Studio), create a file named index.xmlin your new solution directory.
3. Copy the following text into the index.xml file, changing the content accordingly:
***<index>
<name>Example Solution</name>
<description>This solution contains examples I created while learning to
work with action sequences.</description>
<icon></icon>
<visible>true</visible>
<display-type>icons</display-type>
</index>***
4. Save the file and close the text editor.
5. Log into the Pentaho User Console as an administrator.
6. Refresh the solution repository cache by going to the Toolsmenu, then selecting the Refreshsubmenu, then
clicking on Repository Cache.
You now have a new solution directory. It will show up in all file dialogues in the Pentaho User Console as well as the
Solution Browser in the left pane.
Every time you add or edit an action sequence to your solution directory, you must refresh the repository cache as
explained above. Each user currently logged into the Pentaho User Console must also refresh their session cache; this
is best done by re-logging into the Pentaho User Console.
I am using Pentaho design studio to develop an xaction sequence script. I followed this procedure:
1. In Design Studio, go to the Filemenu, then select New, and click on Other...in the sub-menu.
The Newwindow will appear.
2. In the Newwindow, click the triangle next to Pentaho, select New Action Sequence Wizard, then click Next.
3. In the File namefield, type hello_world.xaction.
4. Select Hello Worldfrom the Templatedrop-down box, then click Finish.
The wizard will generate the new file and bring you back to the workbench.
5. Click on HelloWorldComponentin the Process Actionssection on the left.
The right side of the screen will change to show the options available for this action: Nameand Message. The
Name field controls the name of the component in the Process Actions list on the left; it doesn't do anything else of
note. The Message field contains the text that will appear on the screen when the action sequence is run. It is prepopulated with %quote, which is a token that represents a quote message in a nonexistent properties file. Pentaho
used to provide properties files for each example, but they have been removed from the standard Pentaho Business
Analytics distribution.
6. Replace the %quotewith a sufficiently inspiring message.
Alternatively, you could create a hello_world.propertiesfile and populate it with the appropriate messages and
tokens, but that has no advantage unless you intend to internationalize this action sequence.
7. Save the file.
You now have a working action sequence that prints a short text message: "Hello, World." plus whatever you typed into
the Message field. The first part of the message is determined by a message bundle packaged with the Pentaho Web
application archive.
To test the action sequence, use the Test tab if you are on Windows or OS X, or log into the Pentaho User Console and
run it from the Solution Browser.
Running the sample xaction fails with
Sorry. We really did try.
Something went wrong. Please try again
or contact your administrator.
I figure this is because pentaho cannot locate the xaction file as specified in the file path when I generate the url from within design studio. The folder I created within the solution directory does not appear in the browser directory list despite refreshing the cache. Any help?
With Version 5 the files and folders are no longer stored in "/pentaho/server/biserver-ee/pentaho-solutions/..."
It is stored now in a jackrabbit repository.
You can create folders and upload files in the user console.
I did mistake after starting biserver 5.2. Unexpectedly i deleted Public folder after configuring to jackrabbit. Now how do i create the folder so that i can see Public and Home Folders.
Sorry for getting back to this question some time later, just in case someone searches something related to Pentaho repositories management with no success, i hope this information can help.
Just to make things more clear, i'm using version 5.2.x and i'm supposing you want to create a new folder on the root directory of Browse Files screen, something that is not allowed on Pentaho5 CE Web Console.
The answers about repository management based on JackRabbit are correct, so the old way we used to configure new "solutions" folders doesn't work anymore. From version 5 on, Pentaho exposes an interesting REST api that allows you doing things not allowed in web console.
For instance, for creating a folder in root directory you should create a http REST PUT request against Pentaho Server (using tools like curl or if you appreciate Google Chrome, something like Advanced Rest Client, etc.) using the following URL:
http://localhost:8080/pentaho/api/repo/dirs/:new_directory_name?userid=admin&password=password
Pentaho repository directory REST api is documented here:
http://javadoc.pentaho.com/bi-platform500/webservice500/resource_DirectoryResource.html
It's worthy to spend some time reading REST API the documentation.

How can tfs work item details be exported?

I don't seem to be able to export tfs work items along with their details, is there any way to do this without using TFSIntegrationTools? (I cannot install these at my work).
Specifically I want to be able to export the Description portion of the Work Item details:
e
If all you need is the work item data then I would design a flat query, run it and then open it up in Excel. You can modify which columns are visible. Once open you can save it off or print or form a report in excel or what have you.
Your other options are to query the warehouse in sql server or query in code.
http://blogs.msdn.com/b/sunder/archive/2010/03/02/reporting-in-team-foundation-server-part-7-excel-reports-from-work-item-queries.aspx
I had the same problem, but I could not benifit from excel, because I wanted to send the list to someone who doesn't have access to the local TFS. So I created a tool to help with that.
In case this helps any one, here is the link to download release one excutables from codeplex https://tfsworkitemsexporter.codeplex.com/
please also feel free to copy the code and edit as you like

TFS 2010: Project dashboard reporting service error

I have configured TFS 2010, but when I try to load the project dashboard for a team project, it returns a reporting service error so I am not able to see the "Task Burndown (hours)" and " Burn Rate (hours/day)". Other parts of the page are working fine.
The error is as:
An error has occurred during report processing. (rsProcessingAborted) Get Online Help
Cannot impersonate user for data source 'TfsReportDS'. (rsErrorImpersonatingUser) Get Online Help
Log on failed. (rsLogonFailed) Get Online Help
For more information about this error navigate to the report server on the local server machine, or enable remote errors
I got resolution finally....
Go to analysis service.
You will see database named as TFS_Analysis.
Go to Roles node.
View properties of "TfsWarehouseDataReader".
Click on Data Sources.
Now you will see Tfs_AnalysisDataSource.
Change access to "Read" and check "Read Definition" box. Now Click OK and You are Done.
The main problem was with SCHEMA CONFLICTS.
First identify which fields are causing schema conflicts, invoke GetWarehouseStatus and observe the XML which fields got conflicted, and in which collection. once you found the filed names then rename the fields with the help of below link.
geekswithblogs.net/Natalia/Default.aspx
msdn.microsoft.com/en-us/library/ee921480(v=VS.100).aspx
then rebuild your warehouse from tfs admin console, take a back up of old database in sql database and delete it .wait for some time (depends on warehouse time to refresh the cube or check the GetWarehouseStatus next day and check the xml ). to check this use below link
type //localhost:8080/tfs/TeamFoundation/Administration/v3.0/WarehouseControlService.asmx?op=GetProcessingStatus
I spend lot of time to resolve this issue, that's why posting the solution here, this may help some one... any queries related schema conflicts and Reporting services feel free to post me... iam not a expert but for sure i can help you out in this issues....
Please use Mark as Answer if my post solved your problem and use Vote As Helpful if a post was useful.

Resources