Closed. This question is off-topic. It is not currently accepting answers.
Want to improve this question? Update the question so it's on-topic for Stack Overflow.
Closed 10 years ago.
Improve this question
Refer a service where you can see your tweets. I am looking for services like these two: http://allmytweets.net, http://greptweet.com/u/twitter/
I want to find something similar, preferably in Open-Source.
Thank you for your attention.
I'm not aware (off the top of my head) of available open-source solutions for the base task you're requesting and I'm not inclined to just go Googling around for one, but it isn't particularly difficult to replicate on your own, given you aren't afraid to write some code.
The first step is to visit https://dev.twitter.com/ and get started with registering your own application and getting your consumer key and secret. You'll also want to use Twitter's list of code libraries to find an available Twitter oauth library in a language you can use for your project. A good bet is PHP; Abraham's TwitterOauth library has served me well, myself.
Assuming the account's tweets aren't private, you can use your own application and user access keys to request batches of tweets. After a successful request to the API method GET statuses/user_timeline wherein you specify a since_id of 0, you'll make a new request in which you specify that the since_id is the same as the greatest ID from the last batch; you can continue this until the number of results you get back is 0 (you can't simply assume you're done when you get less than your specified count results back, as some circumstances can result in omitted tweets from your result set.)
Related
Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
Questions concerning problems with code you've written must describe the specific problem — and include valid code to reproduce it — in the question itself. See SSCCE.org for guidance.
Closed 9 years ago.
Improve this question
Has something with the twitter API changed? This link was returning results earlier today but for whatever reason it seems to have stopped?
https://search.twitter.com/search.json?q=vine&callback=?
Also, clicking on the links in the examples no longer work?
https://dev.twitter.com/docs/using-search
https://search.twitter.com/search.json?q=%40twitterapi
If you're coding in ASP.NET, I recommend the Twitterizer library. It's written and maintained by a fellow who helped me get my own Twitter .NET library off the ground when I was doing that sort of thing.
Yes. You've found out how the version 1 api has been deprecated / is in the process of being removed.
From now on, all your requests to the twitter API will be required to have authentication (OAuth), and for some this was a nightmare to figure out.
Fortunately, I wrote a lengthy post (with pictures) explaining how to set yourself up a twitter dev account / application, get a set of keys, and then I created a library to perform authenticated requests for you.
The only downside is that this library is written in php, and from looking at your profile you use ASP.net - so try googling for a library for your specific language if you don't want to use PHP.
In any case, no more unauthenticated requests for you ;) You need to dig in with OAuth now!
Closed. This question is off-topic. It is not currently accepting answers.
Want to improve this question? Update the question so it's on-topic for Stack Overflow.
Closed 9 years ago.
Improve this question
I'm considering to use url pattern like below:
example.com/item/r6B0PmUmx07O/just-one-item
example.com/item/r6B0PGgwPJWl/yet-another-item
the part before slug is an unique and unpredictable id for an item.
compare with url like
example.com/item/1001/just-one-item
example.com/item/1002/yet-another-item
is this way bad for SEO?
or will it be bad for crawling by the search engine?(since the crawler cannot 'guess' the next item's id)
I'm not sure how many popular crawlers try to increment number values in URL to hit the page.
They generaly try to traverse by links.
But consider hiding some info from malicious users. If you can reach any info about your users (by example.com/user/1001) there is generally wrong idea to have sequential UID's. It's not mean to be a part of security but sometimes it's good to difficult access to your data. So the competition will have some difficulties when guessing how much products you have on stock :)
Consider supplying dynamical sitemap with links to all your products. This make you sure that every crawler will hit all your items no matter what key it has.
Closed. This question is opinion-based. It is not currently accepting answers.
Want to improve this question? Update the question so it can be answered with facts and citations by editing this post.
Closed 8 years ago.
Improve this question
We've built an app using Facebook PHP SDK. Quite often (too often to count it negligible) requests result in exception (obviously timeout). What is the best way to deal with this? Are there any options we could tweak to decrease the frequency at least?
I'm searching information about this topic too.
A couple of things to keep in mind:
1) your server, if your hosting plan is cheap (let's face it), it will probably fail with a considerable amount of requests.
2) Optimize your api calls and if it's possible, change it for FQL queries. This happened to me yesterday, my app needed to extract the name (via php sdk api call) of four friends selected by the user. For each friend I was asking facebook for the complete profile [$this->api('ID')].
You might think "well, at the end the profile is just 4 or 5 data fields". I really don't know the impact of this query, but I was getting a lot of timeouts. Then I changed the query for a FQL query (using the SDK, but only selecting the name field). The timeouts are gone.
As I said, I'm still looking for information about optimizing api calls, but I hope you can start with this.
Regards.
Closed. This question is off-topic. It is not currently accepting answers.
Want to improve this question? Update the question so it's on-topic for Stack Overflow.
Closed 10 years ago.
Improve this question
Google Code Search has been incredibly valuable to me as a developer - I use it a couple times a week to see how other developers have used (usually poorly documented) APIs. It's also convenient to see the internals of some of those APIs, or to find which API corresponds to the functionality you want (it's a great resource for Android in particular -- give it some of the text you see on screen, and it'll usually find the implementing class).
Now that Google shutting down code search as of January 15, 2012, are there any good replacements?
I have reviewed the following sites
The good
Krugle
searchcode
The broken or unsuitable
Antepedia (site is only a "We'll be back soon" page because Antepedia has been acquired)
The dead
Koders (discontinued)
SymbolHound Code Search
GrepCode (only Java)
SymbolHound (generic search engine, not just code)
Codefetch (unreachable as of 2016-08-23)
Codase (discontinued)
When I originally did the review, Koders turned out to be the winner for my purposes, but I really liked the user interface and features of SymbolHound Code Search better. The only problem with SymbolHound was the small number of sites it has indexed. The search[code] engine was also promising at that time.
Many of the sites I've reviewed have since been discontinued completely or have disabled their code search functionality. Krugle and search[code] seem to be chugging along, and GrepCode is good if you live in the Java world.
Take a look at these:
searchcode
krugle
Another one to consider is http://searchcode.com/ It supports regex search as Google Code search does. For example,
http://searchco.de/?q=/[cb]at/
http://searchco.de/?q=/a{2,3}/
http://searchco.de/?q=/^import/
http://searchco.de/?q=/atoi/%20ext:c
http://searchco.de/?q=/dll$/
Are all valid searches.
There is http://opensearch.krugle.org
Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
We don’t allow questions seeking recommendations for books, tools, software libraries, and more. You can edit the question so it can be answered with facts and citations.
Closed 3 years ago.
Improve this question
I need to download all the followers of a user on Twitter, then download all the followers of each follower (second level network).
I need the data in any format.
Is it possibile? What would you use?
There is a twitter module for python for example. You will need to register an appliction and to use oAuth. Bad news: also if you prefere something else than python, you will have to use OAuth. Good news: to use the twitter module in order to get information is quite simple afterwards.
Update: You can also use supertweet.net if you don't want to mess around with OAuth!
You can use www.TWEXLIST.com. A a recent solution to export any Twitter list (public or private), followers list, followings list of any Twitter account (yours or any others) in an Excel spreadsheet.
Some years ago I used Tweetake website to backup my Twitter data but it was closed (https://twitter.com/NikkiPilkington/status/270080732582522880). Since then I realized that it is not good to rely on a third party website and it is better to have a "do it yourself" solution. So I created a script that everybody will be able to use in their browser.
The script allows you to backup Twitter favorites, following and followers: https://gist.github.com/baptx/1525f338d93fa01db4e0
It is written in JavaScript and will create a textarea element with your backup data that you can copy/paste in a file.
You can run it in a browser console on a local web page like Firefox's about: or a web page like https://duckduckgo.com/html (in the past I used it directly on https://api.twitter.com/1.1 web page but it's not working any more because they added Content Security Policy).
Before using the script, you need to get your access keys to use Twitter API 1.1: https://dev.twitter.com/oauth/overview/application-owner-access-tokens
You can use https://www.followers-backup.com/web to download the list of followers/following (first level network) of any profile you want and it's free. Give it a try