Google Document AI custom processor training failed for unknown reason - machine-learning

After annotating my data set and training a custom processor on Google Document AI, I get the following error message after 40 mins:
{
"name": "projects/500809854407/locations/eu/operations/10980743890623751388",
"done": true,
"result": "error",
"response": {},
"metadata": {
"#type": "type.googleapis.com/google.cloud.documentai.uiv1beta3.TrainProcessorVersionMetadata",
"commonMetadata": {
"state": "FAILED",
"createTime": "2023-01-13T15:11:36.505757Z",
"updateTime": "2023-01-13T15:50:10.343765Z",
"resource": "projects/500809854407/locations/eu/processors/b94da5023b6cb35d/processorVersions/98637063da2b64d4"
},
"trainingDatasetValidation": {},
"testDatasetValidation": {}
},
"error": {
"code": 13,
"message": "Internal error encountered.",
"details": []
}
}
Any idea what could be the problem?

Related

Grafana alerting Error: tsdb.HandleRequest()

I am using Grafana version: v6.7.3 (a04ef6cefc)
After creating a new alert in my widget dashboard, I am seeing the following error:
Error: tsdb.HandleRequest() error bad_response: readObjectStart: expect { or n, but found <, error found in #5 byte of ...|
The error looks to be triggering similar to an alert. Has anyone else encountered something similar to this?
the Error JSON output
{
"firing": true,
"state": "pending",
"conditionEvals": " = true",
"timeMs": "24.602ms",
"error": "tsdb.HandleRequest() error bad_response: readObjectStart: expect { or n, but found <, error found in #5 byte of ...| <!DOCTYPE h|..., bigger context ...| <!DOCTYPE html PUBLIC \"-//W3C//DTD XHTML 1.0 Transi|...",
"logs": [
{
"message": "Condition[0]: Query",
"data": {
"from": 1637767915735,
"queries": [
{
"refId": "A",
"model": {
"expr": "sum by (code, app_kubernetes_io_instance)(rate(nginx_server_requests{app_kubernetes_io_instance=~\"vue-frontend\",code=~\"4..\"}[1m]))",
"format": "time_series",
"hide": false,
"instant": false,
"interval": "",
"intervalFactor": 1,
"legendFormat": "{{code}} on {{app_kubernetes_io_instance}}",
"refId": "A"
},
"datasource": {
"id": 14,
"name": "somesomain.com"
},
"maxDataPoints": 0,
"intervalMs": 0
}
],
"to": 1637768215735
}
}
]
}

Unable to retrieve event details using Microsoft Graph API 1.0

Has anyone got Microsoft Graph API working for EventMessage? i.e reading email content of EventMessage using Graph API.
Looks like it has actionable card and graph API 1.0 is not able to read it.
Provided delegate access to user for a room resource.
On event creation for that room, mail is received by the delegated user in inbox. The mail type is eventMessage ("#odata.type": "#microsoft.graph.eventMessage").
On fetching the mail through GraphAPI v1.0, the event values are not visible in the response.
Also on setting the Expand property for the mail in graph API as per docs, results in below error.
{ "error": {
"code": "BadRequest",
"message": "Parsing Select and Expand failed.",
"innerError": {
"request-id": "310bbcce-bdd8-4cb6-890c-035243a5ab6d",
"date": "2020-01-21T09:18:10"
}
}
}
I'm unable to repro your issue. The Microsoft Graph API returns details for event messages. Can you provide repro steps, request/response capture, and the values that you'd expect to see?
I used v1.0/me/messagesto get all of the messages to find an eventMessage. I then perform the following to get details on the eventMessage:
GET https://graph.microsoft.com/v1.0/me/messages/{messagedId}
Response body:
{
"#odata.context": "https://graph.microsoft.com/v1.0/$metadata#users('d09-94-498-a9777d4b')/messages/$entity",
"#odata.type": "#microsoft.graph.eventMessage",
"#odata.etag": "W/\"DAA2pU9s\"",
"id": "AAMkADEzOTExYjJkAAA=",
"createdDateTime": "2020-01-24T14:47:24Z",
"lastModifiedDateTime": "2020-01-24T21:08:07Z",
"changeKey": "DApU9s",
"categories": [],
"receivedDateTime": "2020-01-24T14:47:25Z",
"sentDateTime": "2020-01-24T14:47:11Z",
"hasAttachments": false,
"internetMessageId": "<AM5P3M#A38.EURPRD83.prod.outlook.com>",
"subject": "Accepted: Microsoft Graph DevX Roadmap",
"bodyPreview": "",
"importance": "normal",
"parentFolderId": "AAMk3AAA=",
"conversationId": "AAQkADEzOU4=",
"conversationIndex": "AdXReVPepTg==",
"isDeliveryReceiptRequested": null,
"isReadReceiptRequested": false,
"isRead": true,
"isDraft": false,
"webLink": "https://outlook.office365.com/owa/?ItemID=AAMkA%3D&exvsurl=1&viewmodel=ReadMessageItem",
"inferenceClassification": "focused",
"meetingMessageType": "meetingAccepted",
"body": {
"contentType": "text",
"content": ""
},
"sender": {
"emailAddress": {
"name": "Nicole",
"address": "sig#contoso.com"
}
},
"from": {
"emailAddress": {
"name": "Nicole",
"address": "sig#contoso.com"
}
},
"toRecipients": [
{
"emailAddress": {
"name": "Michael Mainer",
"address": "mm#contoso.com"
}
}
],
"ccRecipients": [],
"bccRecipients": [],
"replyTo": [],
"flag": {
"flagStatus": "notFlagged"
}
}

How to create item with column that has strange SharePoint internal encoding?

I created a column with the name "col1" in the list through the web interface Office 365. When I tried to get whole info about this column using msgraph request to
GET /sites/{site-id}/lists/{list-id}/columns/{column-id}
I got such the response:
{
"#odata.context": ...,
"columnGroup": "Custom Columns",
"description": "",
"displayName": "col1",
"enforceUniqueValues": false,
"hidden": false,
"id": "a9fe9dc8-c5b8-467d-8504-23d95f1cc937",
"indexed": false,
"name": "_x0063_ol1",
"readOnly": false,
"required": false,
"number": {
"decimalPlaces": "automatic",
"displayAs": "number",
"maximum": 1.7976931348623157e+308,
"minimum": -1.7976931348623157e+308
}
}
So the problem is that I cannot create an item with that column (I tried both names "col1" and "_x0063_ol1"). The request was to
POST /sites/{site-id}/lists/{list-id}/items
with the body:
{
"fields": {
"Title": "Test2",
"_x0063_ol1": 42
}
}
I got response:
{
"error": {
"code": "invalidRequest",
"message": "Field '_x0063_ol1' is not recognied",
"innerError": {
"request-id": "dbb98ed3-0c2f-48dc-924e-dd888e70c867",
"date": "2018-05-02T07:41:36"
}
}
}
How can I properly create the item with this column using msgraph API?

"Overlaps are not supported within TimeSlots" error in Outlook Calendar REST API

From yesterday I started receiving an error in my service (production environment) when trying to call Outlook Calendar REST API endpoint:
POST https://outlook.office.com/api/v2.0/me/findMeetingTimes
with following request body:
{
"Attendees": [
{
"Type": "optional",
"EmailAddress": {
"Address": "...",
"Name": null
}
},
{
"Type": "required",
"EmailAddress": {
"Address": "...",
"Name": null
}
}
],
"LocationConstraint": {
"IsRequired": true,
"SuggestLocation": false,
"Locations": [
{
"DisplayName": "..."
}
]
},
"TimeConstraint": {
"Timeslots": [
{
"Start": {
"DateTime": "2017-05-05T00:00:00",
"TimeZone": "UTC"
},
"End": {
"DateTime": "2017-05-05T21:59:59",
"TimeZone": "UTC"
}
}
]
},
"MaxCandidates": 1000,
"MeetingDuration": "PT30M",
"IsOrganizerOptional": true }
Here is the response I get:
{
"error": {
"code": "ErrorInternalServerError",
"message": "Invalid value for arg:Overlaps are not supported within TimeSlots, value:
{\"start\":2017-05-04T22:00:00Z,
\"min\":1440}
\r\nParameter name: Overlaps are not supported within TimeSlots"
}
}
Everything was working fine with until yesterday. There were no changes in my service codebase. It seems to be related with some changes in O365 API itself but there are no recent updates in documentation...
Any idea what might be wrong?
UPDATE: Without any modification in the request the server response is now:
{
"error": {
"code": "ErrorInternalServerError",
"message": "Index was outside the bounds of the array."
}
}

Facebook FQL: Difference between status table and stream table?

Can anyone tell me what the difference is between the status table and the stream table?
I am querying my own posts using the graph explorer and I find some status updates appear only in the status table, and others appear in both the stream and status table. When I go to my own Timeline I can see posts from both tables.
What is going on here? Why are some in the posts table only, but appear on my Timeline?!
I am trying to find a list of all my posts that are publicly visible, and it seems to do this I need to query both tables and end up with some duplicates. Is there a better way?
Here is an example:
graph.facebook.com/10150680671412355
{
"id": "10150680671412355",
"from": {
"name": "Alan Byrne",
"id": "509517354"
},
"message": "Beans on toast.",
"updated_time": "2012-03-14T10:53:01+0000",
"likes": {
"data": [
{
"id": "xxxxx",
"name": "xxxx"
}
],
"paging": {
"next": "https://graph.facebook.com/10150680671412355/likes?method=GET&metadata=true&format=json&callback=___GraphExplorerAsyncCallback___&access_token=AAACEdEose0cBAPjaO9gCyMZA2TT5hRE7iEkgmksfpNneFdtshFDfp0oM3hiX6I57vUkzGZAyXW5c9YaM1WQFUxt2zTZCden7zTAlbMSDAZDZD&limit=25&offset=25&__after_id=100001368900242"
}
},
"comments": {
"data": [
{
"id": "10150680671412355_31516486",
"from": {
"name": "xxxxx",
"id": "xxxxx"
},
"message": "Peasant. Baked beans on the other hand...",
"can_remove": true,
"created_time": "2012-03-14T11:20:35+0000"
}
],
"paging": {
"next": "https://graph.facebook.com/10150680671412355/comments?method=GET&metadata=true&format=json&callback=___GraphExplorerAsyncCallback___&access_token=AAACEdEose0cBAPjaO9gCyMZA2TT5hRE7iEkgmksfpNneFdtshFDfp0oM3hiX6I57vUkzGZAyXW5c9YaM1WQFUxt2zTZCden7zTAlbMSDAZDZD&limit=25&offset=25&__after_id=10150680671412355_31516486"
}
},
"type": "status"
}
Yet, when I scan for posts on the stream on that same day, this one does not come back
fql?q=select message,post_id,created_time from stream where source_id=509517354 and created_time > 1331596800 and created_time < 1331856000
{
"data": [
{
"message": "",
"post_id": "509517354_10150680640782355",
"created_time": 1331720877
},
{
"message": "",
"post_id": "509517354_10150680632467355",
"created_time": 1331720499
},
{
"message": "MY COMPUTER DOES NOT HAVE VIRUSES! STOP CALLING ME! ARGH!",
"post_id": "509517354_10150680618357355",
"created_time": 1331719866
},
{
"message": "",
"post_id": "509517354_10150680602327355",
"created_time": 1331719906
},
{
"message": "I did it!",
"post_id": "509517354_10150680491832355",
"created_time": 1331716355
},
{
"message": "",
"post_id": "509517354_10150680446052355",
"created_time": 1331714440
},
{
"message": "It got to Australia.\n",
"post_id": "509517354_365689393454249",
"created_time": 1331679835
},
{
"message": "",
"post_id": "509517354_107999992664739",
"created_time": 1331677643
},
{
"message": "",
"post_id": "509517354_10150678888762355",
"created_time": 1331671872
},
{
"message": "Booooo",
"post_id": "509517354_268316116580687",
"created_time": 1331632096
},
{
"message": "'spose I should wander into the office...",
"post_id": "509517354_10150678259167355",
"created_time": 1331625013
},
{
"message": "",
"post_id": "509517354_10150678258862355",
"created_time": 1331624991
},
{
"message": "",
"post_id": "509517354_10150678257547355",
"created_time": 1331624907
},
{
"message": "",
"post_id": "509517354_10150678250602355",
"created_time": 1331631742
}
]
}
The stream contains posts of various types - links, checkins, photos and status updates.
The status table only contains status updates - text with no additional links, photos etc.
If you're hoping to find the content similar to the timeline, you'll want to use the stream table.

Resources