Informatica XML read error - xml-parsing

In my mapping we are using xml files as our source.Our issue is that while trying to execute our mapping using large xml files (i.e files larger than 300 MB) we are facing an error. The error message is
'Error [Invalid Document Structure] occured while parsing :[FATAL:Error at line1,char1 ']
We have successfully executed our mapping with smaller files(size < 300 MB).
Is there any setting which can be changed to process such large files.If not, is there any workaround that can be done?

Related

Android 12 deep linking validation failed

We have created the assetlinks.json file and placed it under the well-known folder in the root.After that when tried to validate the assetlinks.json file using Digital Asset Links API from google we are getting an exception.
{
"maxAge": "599.999999839s",
"debugString": "********************* ERRORS \n Error: deadline_exceeded: Timeout occurred while fetching Web statements from https://www..com./.well-known/assetlinks.json (which is equivalent to 'https://www..com/.well-known/assetlinks.json') using download from the web (ID 1).\n********** INFO MESSAGES ********************\n Info: No statements were found that match your query\n"
}
Can someone please help to understand what will be the root cause of this issue and how to resolve it

neo4j admin import Error in import Requested index -1, but length is 1000000

I have a set of CSV's that I have been able to use with LOAD CSV to create a database. This set is the small version (1 gb) of a much larger data set (120 gb) I intend to load to neo4j using admin import. I am trying to run the admin import on the smaller dataset first since I have already successfully created a graph with that data. I assume that if I can get the admin import to run for the small version it will hopefully run without problems for the large dataset. I've read through the admin import instructions and I've set up header files. The import loads the nodes just fine but ends up failing with he relationship files. Can anyone help me understand what is happening here so that I can figure out how to fix it? I've tried just removing the file and its associated nodes but this only results in the same error being thrown from the next file in the relationships list.
IMPORT FAILED in 9s 121ms.
Data statistics is not available.
Peak memory usage: 1.015GiB
Error in input data
Caused by:ERROR in input
data source: BufferedCharSeeker[source:/var/lib/neo4j/import/rel_cchg_dimcchg.csv, position:3861455, line:77614]
in field: :START_ID(cchg-ID):1
for header: [:START_ID(cchg-ID), :END_ID(dim_cchg-ID), :TYPE]
raw field value: 106715432018-09-010.01.00.0
original error: Requested index -1, but length is 1000000
org.neo4j.internal.batchimport.input.InputException: ERROR in input
data source: BufferedCharSeeker[source:/var/lib/neo4j/import/rel_cchg_dimcchg.csv, position:3861455, line:77614]
in field: :START_ID(cchg-ID):1
for header: [:START_ID(cchg-ID), :END_ID(dim_cchg-ID), :TYPE]
raw field value: 106715432018-09-010.01.00.0
original error: Requested index -1, but length is 1000000
at org.neo4j.internal.batchimport.input.csv.CsvInputParser.next(CsvInputParser.java:234)
at org.neo4j.internal.batchimport.input.csv.LazyCsvInputChunk.next(LazyCsvInputChunk.java:98)
at org.neo4j.internal.batchimport.input.csv.CsvInputChunkProxy.next(CsvInputChunkProxy.java:75)
at org.neo4j.internal.batchimport.ExhaustingEntityImporterRunnable.run(ExhaustingEntityImporterRunnable.java:57)
at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)
at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
at java.base/java.lang.Thread.run(Thread.java:834)
at org.neo4j.internal.helpers.NamedThreadFactory$2.run(NamedThreadFactory.java:110)
Caused by: java.lang.ArrayIndexOutOfBoundsException: Requested index -1, but length is 1000000
at org.neo4j.internal.batchimport.cache.OffHeapRegularNumberArray.addressOf(OffHeapRegularNumberArray.java:42)
at org.neo4j.internal.batchimport.cache.OffHeapLongArray.get(OffHeapLongArray.java:43)
at org.neo4j.internal.batchimport.cache.DynamicLongArray.get(DynamicLongArray.java:46)
at org.neo4j.internal.batchimport.cache.idmapping.string.EncodingIdMapper.dataValue(EncodingIdMapper.java:767)
at org.neo4j.internal.batchimport.cache.idmapping.string.EncodingIdMapper.findFromEIdRange(EncodingIdMapper.java:802)
at org.neo4j.internal.batchimport.cache.idmapping.string.EncodingIdMapper.binarySearch(EncodingIdMapper.java:750)
at org.neo4j.internal.batchimport.cache.idmapping.string.EncodingIdMapper.binarySearch(EncodingIdMapper.java:305)
at org.neo4j.internal.batchimport.cache.idmapping.string.EncodingIdMapper.get(EncodingIdMapper.java:205)
at org.neo4j.internal.batchimport.RelationshipImporter.nodeId(RelationshipImporter.java:134)
at org.neo4j.internal.batchimport.RelationshipImporter.startId(RelationshipImporter.java:109)
at org.neo4j.internal.batchimport.input.InputEntityVisitor$Delegate.startId(InputEntityVisitor.java:228)
at org.neo4j.internal.batchimport.input.csv.CsvInputParser.next(CsvInputParser.java:117)
... 9 more
The error is actually quite explicit: have a look at line 77614 in rel_cchg_dimcchg.csv. It's usually caused by an incorrect endpoint id. For example, if the END_ID is supposed to be a number but it's something like 4171;4172;4173;4174;4175;4176 this will raise the InputException error.
One would assume that --skip-bad-relationships would ignore these issues but it doesn't. So, the only remedy is to ensure that all START_ID/END_ID's are correct (ie. the right data type and format).

Random problem when generating PDF in Grails

I have a problem in production when I generate pdf through renderPdf. Sporadically when it will render the error in some MAP property that I send to the GSP, for example:
Caused by: java.lang.NullPointerException: Cannot get property 'diario' on null object
However, when you try again with the same data, the pdf is successfully generated. I use the Grails version 3.3.2.
I increased the RAM memory in the Grails configuration and decreased the occurrence of this error.
export GRAILS_OPTS="-XX:PermSize=2048m -XX:MaxPermSize=2048m -Xms2048m -Xmx2048m"

SSIS Truncation in Flat file source

I am new to SSIS, I am importing a flat file data into SQL Server. It throws error while importing data in the flat file source task.
[Flat File Source [60]] Error: Data conversion failed. The data conversion for column "DCN_NAME" returned status value 4 and status text "Text was truncated or one or more characters had no match in the target code page.".
[Flat File Source [60]] Error: The "Flat File Source.Outputs[Flat File Source Output].Columns[DCN_NAME]" failed because truncation occurred, and the truncation row disposition on "Flat File Source.Outputs[Flat File Source Output].Columns[DCN_NAME]" specifies failure on truncation. A truncation error occurred on the specified object of the specified component.
[SSIS.Pipeline] Error: SSIS Error Code DTS_E_PRIMEOUTPUTFAILED. The PrimeOutput method on Flat File Source returned error code 0xC0202092. The component returned a failure code when the pipeline engine called PrimeOutput(). The meaning of the failure code is defined by the component, but the error is fatal and the pipeline stopped executing. There may be error messages posted before this with more information about the failure.
How to solve this issue. I don't have problem in destination.Is it possible to change the source field length?
I changed the source length in source connection manager. It works fine now.
You have not provided enough information to answer your question. Here is what I can gleam from what you have provided.
Your SSIS package is read a flat file and it assumes DCN_Name is data DT_STR
(type char). Confirm the origin and the destination have the same data type. If they do not, use a [Data Conversion] function which is found on the left side on the screen under Data Flow.
Be mindful that while SQL uses varchar and nvarchar, SSIS uses DT_WSTR and DT_STR. It can get confusing. In the link below you will find a SSIS to SQL Data Type Conversion
enter link description here
The errors generated by SSIS are counter intuitive.
The solution was use a data conversion because the source and destination were different data types.

csv file codification affects salesforce's Dataloader 25.0.2 behavior?

I've been trying to upsert and delete some data at objects in salesforce cloud using Data Loader 25.0.2. Data Loader executes without problems, but the insertions/deletions don't get done. The log file returns some lines with this errors:
2012-10-03 17:13:16,958 ERROR [deleteAccount]
client.PartnerClient processResult (PartnerClient.java:432) - Errors
were found on item0 2012-10-03 17:13:16,958 ERROR
[deleteAccount] client.PartnerClient processResult
(PartnerClient.java:434) - Error code is:INVALID_ID_FIELD 2012-10-03
17:13:16,958 ERROR [deleteAccount] client.PartnerClient
processResult (PartnerClient.java:435) - Error message:invalid record
id
I've checked that the object ids match, so there are not obvious differences between the data at the cloud and the csv being used to the deletion command.
What could be happening?
I finally found that the codification of the .csv file I've uploaded differed from the codification used for internal representation in salesforce for the data of that object. For example: My .csv file was codified on ANSI instead of UTF-8. This detail got salesforce into confussion. So I had only to change the file's codification to UTF-8 and everything got solved.

Resources