Have gone through several links, and though it explains what is incremental backup how are the difference since the last backup identified? is it by size of the files or the data... can anyone explain...else guide me to a reference for the same?
Found out the following information on the reference specified. Hope it helps out sm1.
The client considers a file changed if any of the following attributes changed since the last backup:
File size
Date or time of last modification
Extended Attributes
Access Control List
If only the following items change, they are updated without causing the entire file to be backed up to the server:
File owner
File permissions
Last access time
Inode
Group ID
Reference : http://publib.boulder.ibm.com/infocenter/tsminfo/v6/index.jsp?topic=%2Fcom.ibm.itsm.client.doc%2Fc_bac_fullpart.html
Related
Trying to get latest version of the project and tfs and always get te same warning without the way of resolving conflict:
When i click the file to resolve it gives me no options to choose
Is there any way to get this around without resorting to cloning whole solution again?
Thanks for help in advance
For this kind error usually due to the file is writable. Take a look at #Buck Hodges (Director of Engineering at Microsoft)'s reply in this case:
The file is writable. When there's a writable file, get will not
overwrite it unless the /overwrite option is used. However, you would
need to figure out how the file is becoming writable. You may want to
go back to the default option of creating a new workspace, and see what is making the file writable.
By default, the system does not retrieve an item if it is writable (that is, if its read-only attribute is cleared) on the client machine. This option overrides the default behavior and overwrites a writable item, unless the item is checked out.
Instead get files directly through Visual Studio. You could try to get files through tf get command line and with /overwrite option. This should do the trick.
I am doing file integration using mirth. There is one software which generate the HL7 files. I want to read data from that files, without moving them to another destination. Next time when I want to read data, at that time it'll ignore the files from which the data are already read (i.e.Just read the new files data which are generated after last data read).
I had done this but I'll achieve it when I modify the original filename, if I am not modifying the filename then it'll read the duplicate data.
Is there is any solution for this problem, so we can read data from the files which are generated new. I am using mirth 3.5.1 version and HL7 v2 messages.
Thanks in advance.
Thanks #daveloyall, I am posting your comment as a answer here.
When you rename a file at the time you process it, for example, to add a .DONE suffix to the filename, you are adding information that can be used later. The part of the channel that reads files could be configured to skip files that have the .DONE suffix. You also add information if you move the files. Or store the filenames in some database table. I don't know if Mirth has an internal feature that tracks which HL7 messages it already processed, but if such a feature exists, the keyword 'deduplication' might be associated with it.
Using IMAP, how can I figure out if a mailbox has been moved or deleted by another client?
The LIST command simply does not list a deleted mailbox any more.
Exactly. You have to remember what folders you know about. Next time you do a LIST, any that are missing have been deleted.
Additionally, you should be tracking every folder's UIDVALIDITY value. If it changes, that folder is not the same one you know about. It may have been renumbered, deleted and recreated, or deleted and replaced by a renamed folder.
In either case, you should dump any cached information you know about that folder.
There is not, in general, any way to track folders that have been renamed by another client. You can only detect that a folder is missing and there is a folder with a new name. IMAP simply does not provide enough information to correlate them. Tracking messages and folders across moves does not appear to have been a design goal of IMAP.
I am trying to use the latest official version of the DropBox iOS Core SDK, in particular the DBRestClient, to efficiently keep a document tree up to date on my local computer. However, several features of the SDK don't seem to be implemented as you'd expect, and I was wondering whether I am doing something wrong, failing to understand something, or on the other hand just looking for features that aren't there.
Dropbox has (at least?) two kinds of entities it stores: files and folders. If I make a change to a file in Dropbox, I can detect it by a change in the file's "rev" string.
However, for folders, the rev string doesn't change when the contents changes. For example, it doesn't change in response to any of the following:
Adding a file
Deleting a file
Editing a file
Question 1:
Is this the expected behavior ? If so, does that mean that when I want to know if any of my Dropbox files have changed I must walk through the entire folder tree every time ?
Deltas:
There is a command to get a "delta" of the current contents (of something, the command takes no path parameters) related a "cursor" string. The command looks like it is supposed to return a record of edited files/folders along with a new "cursor" string specifying the current state. However, when I get the delta, the contents are always empty. If I make a change in dropbox, and then send back the previous returned "cursor", I still get an empty delta.
Question 2:
Are deltas currently working in the SDK, and if so can anyone tell me what I am doing wrong ?
Question 1: Yes, that's the expected behavior. Typically you would use delta to watch for changes.
Question 2: Yes, I would assume that delta is working. You haven't shown any code, so it's impossible to guess what's going wrong.
My app is using CoreData with SQLite, and I recently noticed that with a big database (240MB) it takes some time to load.
I discovered that during load a journal file is created then removed.
Reading this post : What causes a journal file to be created in SQLite?, it indicates that a journal is created when a transaction begin to be able to rollback.
In my case I only do a select (no update or insert), so I don't understand why this journal file is created.
Do you have an idea?
Thanks a lot for your help.
Thierry
Core Data touches objects when you fetch them. There is internal versioning going on inside of Core Data that we do not have direct access to. It uses this information for merging, staleness, etc.
Short answer, it is SEP (Somebody else's problem) and is internal to Core Data.
Use WAL mode. THis will avoid the huge journal created by default. In WAL mode, the journal is created but it size is according that you need. In you case, it will be minimal since you're not writing a lot.
Just execute the query: PRAGMA journal_mode = WAL; once when creating the DB. If is executed after creating might throw an error if a previous DB exists.
Note that WAL is not compatible with previous journal types and is supported from 3.7.3.
Thanks for your answer. My problem what with NSSQLiteManualVacuumOption.
When opening Store I add this :
[NSNumber numberWithBool:appHasBeenUpdated], NSSQLiteManualVacuumOption
appHasBeenUpdated is set to YES only when a new update has been installed. The problem is that a vacuum is done each time I open data store even if appHasBeenUpdated is set to NO.