Timesheet reporting in FogBugz? - fogbugz

Currently it's a big PITA to create timesheet report for say a week or month in FogBugz... maybe someone knows about a plugin to that for you?

This question is kind of old so I am assuming you were using FB6, but if you are now using FB7 to do exactly this: ClarkKent.

If you have FogBugz hosted on your own server, you can query the database manually and get timesheet results that way. If its hosted by FogBugz, then you can download the database, do the same query, or, use the API to get all the time entries. Some programs already exists for that, one of them is TimeSprite, but its not free.

Related

Volusion API - How to retrieve orders by date

I am making a call to the volusion api generic order export and I want to be able to only get orders for a single day however the date it is being compared to is a full timestamp stored in the volusion database so you'd have to have the entire timestamp and not just d/mm/yyyy. Seems like it is impossible to pull orders by date via the api. If anyone has experience with the generic orders export with the Volusion api your help would be much appreciated.
You cannot do this directly via the Volusion API as you have already stated.
You can however run a stored SQL query. The method to do this is explained here...
https://stackoverflow.com/a/29134928/357034
You just have to create your specific SQL query which returns the data you require.
For other who might come along later... I had the same error and I believe I solved the problem by switching the user to SuperAdmin rights.

How to design a good QuickBooks integration solution

I have now played with the QBO and QBD APIs and feel I have a fair understanding of how it thinks and how to interact with it. So now it is time to design the actual integration solution.
Inside my application you can create new customers, quote services, perform services, and soon, pass invoices to QuickBooks, sounds easy.
But what if the customer is not in QB yet? No problem - for each invoice I will look up the customer (need the id anyway) and if it doesn’t exist, add it. But if I have to look up the customer for each invoice it seems like it might be slow. I will likely have 30,000 customers and have 500-3000 invoices per day.
So my question is this; what are others doing?
a) Are you storing the QB id for each customer in your data?
b) How do you detect address changes (changed in your app and changed in QB)?
c) Is the batch submission interface so much faster I should use that?
Thanks for your help!
We often times do store the QB id in our database for use. If we post an invoice into QB, we'll then store the QB id for future use if we need to modify it.
As far as detecting changes on the customer record and other info, there's a couple ways to handle the conflict resolution. One is to keep a timestamp on your side as to when changes are made. You can then compare this with the timestamp of the last change on the QB record and then make your decision as to which one gets updated.
FreddyMac,
To detect changes on the Intuit side you can construct a query with a CDCasOf Filter, which will return only the data that has changed since a date you provide. (ChangeDataCapture as of)
https://ipp.developer.intuit.com/0010_Intuit_Partner_Platform/0050_Data_Services/0500_QuickBooks_Windows/0100_Calling_Data_Services/0015_Retrieving_Objects
You need to keep track of data changes on your side.
The batch submission is not faster, its just easier for you to write the code.
The IPP SDK can queue the API calls for your and aggregate the responses.
regards,
Jarred

Moving from TestTrack to JIRA

what are the steps involved in moving from TestTrack to JIRA?
I've done a few migrations from TestTrack to JIRA over the last 5 years. It's always more work than I expect (weeks not days). The migrations I did used a custom migration application that read the data from the TestTrack database (not the simplest of schemas) and then imported it into JIRA using a custom SOAP API. Nowadays I think I might go with a more focused approach and use a single customized script to create CSV suitable for the JIRA CSV importer. This will get you the issue fields, comments and attachments. Links are separate thing, but no issue history.
The JIRA CSV importer works, though the flattening of the data restricts what it can do significantly. The hardest thing is usually mapping of values from one system to another - are the userids really identical in both systems?
If you want to investigate doing the migration commercially, please contact info#customware.net (my employer)

Implementation of Time Machine Feature in grails application

I am trying to implement a 'time machine' feature in my grails application. The feature would allow user to select a date in past and would display the interface of the application that was on the selected date. How do I implement this feature? I was thinking of adding a 'dateCreated' field for all domains, so that in the time machine feature, I could query all the results with created date before the selected date. I think this would work but as the data would grow, the size of database would grow and at that time the application would be heavy. Is there any other way to do this ?
Thanks
You could maybe draw some inspiration from this related question:
How to manage object revisions in Grails?
You should look at the http://grails.org/plugin/audit-logging plugin since it will allow you to keep all versions of domain class instances. But implementing this feature will be pretty tricky since object don't exist in isolation - you'll need to not only display the data as of the previous date, but its related data (e.g. the author's books collection) as of that date too. It will make the queries quite complicated.

Store data in Ruby on Rails without Database

I have a few data values that I need to store on my rails app and wanted to know if there are any alternatives to creating a database table just to do this simple task.
Background: I'm writing some analytics and dashboard tools for my ruby on rails app and i'm hoping to speed up the dashboard by caching results that will never change. Right now I pull all users for the last 30 days, and re-arrange them so I can see the number of new users per day. It works great but takes quite a long time, in reality I should only need to calculate the most recent day and just store the rest of the array somewhere else.
Where is the best way to store this array?
Creating a database table seems a bit overkill, and I'm not sure that global variables are the correct answer. Is there a best practice for persisting data like this?
If anyone has done anything like this before let me know what you did and how it turned out.
Ruby has a built-in Hash-based key value store named PStore. This provides simple file based, transactional persistance.
PStore documentation
If you've got a database already, it's really not a big deal to create a separate table for tracking this sort of thing. When doing reporting, it's often to your advantage to create derivative summary tables exactly like what you're describing. You can update these as required using a simple SQL statement and there's no worry that your temporary store will somehow go away.
That being said, the type of report you're trying to generate is actually something that can be done in real-time except on extravagantly large data sets. The key is to have indexes that describe the exact grouping operation you're trying to do. For instance, if you're grouping by calendar date, you can create a "date" field and sync it to the "created_at" time as required. An index on this date field will make doing a GROUP BY created_date very quick:
SELECT created_date AS on_date, COUNT(id) AS new_users FROM users GROUP BY created_date
Using a lightweight database like sqlite shouldn't feel like an overkill. Alternatively, you can use key-store solutions like tokyo cabinet or even store the array in a flat file manually but I really don't see any overkill in using sqlite.

Resources