I am using apache flume 1.3.0 and I gave some keywords in the keyword field of flume file to search for data in twitter. As usual, flume brings data in json format but it also fetches some data which does not contain the keywords which i mentioned.
Related
I'm using Tika 1.26 in order to extract metadata of a document.
I first gave a try to the Tika Server and then I switched to programmatic API. Nevertheless, even if the documentation states that the Content-Encoding of a document should be returned via the /meta API or the MetadataParser, the property is not actually returned.
I found that the API that actually returns a Charset is the CharsetDetector, but I don't know how to invoke this same API via the Tika Server.
I don't have any clue right now.
Can someone point me out what's the correct way to model this use case or if I'm doing something wrong?
I am currently trying to import a list of customers from my QB installation using the _quickbooks_customer_import_request() and _quickbooks_customer_import_response() methods found in Consolibyte's QB PHP library in the docs/web_connector/example_web_connector_import.php file.
When I run Web Connector, it is able to establish a connection and receive the request from my server. It then errors out on the response (where QB contacts my server and tries to pass to it response data). The error shown in Web Connector is a generic getLastError() message:
When I look in the quickbooks_log table that the Consolibyte library created in the quickbooks MySQL database, I see the following:
The above doesn't show the reason for the error. How do I log the underlying errors here? I would prefer a solution where the detailed error description can be inserted into the quickbooks_log table in a JSON format.
There's a Troubleshooting section of the docs here:
http://wiki.consolibyte.com/wiki/doku.php/quickbooks_integration_php_consolibyte?s[]=troubleshooting#troubleshooting
You should start by putting the Web Connector in VERBOSE mode, and looking at what's in the Web Connector log file.
Also, check your PHP error log.
There's many different places an error could be occurring (PHP, configuration, SSL/TLS, QuickBooks, etc.) so start with the Web Connector log and go from there.
In my poc, I am using Spring cloud stream and connected to confluent schema registry. I can see schema's registered on schema registry, so I dont want to create own avro schema. I was able to get payload converted to POJO using application/json. Now I am trying to do the same with avro to Pojo conversion. But it doesnt seem to be working.
I am setting content type as ''contentType: application/*+avro''.
Any sample application, will be helpful.
I listen to Twitter stream and successful with extracting data I want from tweets. Now I want to keep building a graph with the extracted info, like
(user)--[tweets]-->(tweet)
(tweet)--[mentions]-->(user)
(tweet)--[tagged]-->(hashtag)
While this graph keep building over the time, I want to run queries over this graph. How can I do this with Apache Flink?
With some more digging in to the forums and JIRA, I found gelly-streaming matching my needs.
With it, we can do create a GraphStream,
GraphStream<Long, NullValue, NullValue> graph = new SimpleEdgeStream<>(getEdgesDataSet(env), env);
Examples : https://github.com/vasia/gelly-streaming/tree/master/src/main/java/org/apache/flink/graph/streaming/example
Here are some other relevant links.
On Apache Flink mailing list : http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/Graph-with-stream-of-updates-td5166.html
Vasia Kalavri's talk on Graphs as Streams: https://berlinbuzzwords.de/session/graphs-streams-rethinking-graph-processing-streaming-era
I am using apache jena fuseki server to load the data in a .ttl format and then querying the data.But the problem is i am not able to serve multiple data simultaneously.
I am starting the server using the following command.
./fuseki-server --update --mem /ds
The server version i am using is 1.1.1
/home/user/jena-fuseki-1.1.1/./s-put http://192.168.1.38:3030/ds/data **default** /home/user/data.ttl
I was thinking like if we change the default option in s-put command,is there any other options to serve concurrent data as separate instances.
./s-put http://192.168.1.38:3030/ds/data default /home/user/data.ttl
I am having a rest api from which multiple users can load the data and do SPARQL queries on top of it.But when each time a new user loads the data the server gets the new data and the previous data is gone.
I want each user to have his own data to be maintained by the server.Is there some mistake in the way i am loading data ?
To add data, not replace it, use POST and command s-post. HTTP PUT means "replace", HTTP POST is "append" (which forRDF just means "add").
PS Try Fuseki 2.3.0