How to include CMS page,static blocks into Magento search - magento-1.4

We have a magento store which mainly features the contents and it has only 12 - 15 products. So client wants cms page and static block content also to be included on search. I can search only under cms page or static block by using like and getcollection methods. But i cant be successful with combining product search/cms page search.
Please can someone help me with this

To include CMS pages in default magento search we could use the below extension
http://www.magentocommerce.com/magento-connect/tangkoko/extension/1218/tangkoko--fidesio-cms-search
I hope, it helps you.

Related

Filter Products based on Product attribute

I need to create filter like below link
https://paytm.com/shop/g/paytm-home/incredible-offers/smartphones-flat-20-cashback
When i click the smart phone on Landing page then filter show based on smart phone
Like camera color sim internal memory external memory .. etc
Current i have list of productViewmodel which contain the product and product variant only
Please guide me
Thanks in Advance :)
The search term you are looking for is faceted search.
One option to implementing it is using a faceted search engine, such as Bobo-Browse.Net (which is implemented as an extension to the Lucene.Net search engine). It is a .NET port of the Java version, meaning it is a 100% .NET solution.
See the faceted search prototype and car demo for some examples of how to implement it in MVC.
Full Disclosure: I am a major contributor to the Bobo-Browse.Net project.
Another option is to use solr, which runs as a separate process than the web site that uses it. It is a Java-based solution.
Either way, the best solution from a web site is to use AJAX so the drill-down happens without reloading the entire page.

Follow and Scrape Links In Tweets with Rails

I'm currently parsing twitter feeds through my rails app and wondering how it would be possible to follow links in the tweets and scrape the external content (for example, if the tweet contains a link to an article on TechCrunch, follow that link to the article and scrape the title and body content of the article). Flipboard for iPad is a perfect example of this.
I'm familiar with screen scraping using Nokogirl / Mechanize, but trying to figure out how to accomplish this in a way it could work regardless of where the link in the tweet is linking to (whether it be to TechCrunch, or Flickr, etc...).
Any thoughts / insights would be much appreciated!
Many of the major content providers provide a oembed endpoints. Take a look at the oembed_links gem. From the readme:
It allows you to easily parse text and
query configured providers for embedding information on the links
inside the text. A sample configuration file for configuring the
library has been included (oembed_links_example.yml), though you
may also configure the library programmatically (see rdocs).
If you use oembed_links in conjuction with http://oohembed.com/ you'll have dozens of content providers handled for you. You can easily write a custom provider to handle the rest.

Search product websites without api

Sorry for the bad title and description, but I was wondering if there is anyway I could search/list products from other sites (say Express, American Eagle), from a web app I create, even if the site doesn't have an API.
Thanks
Sure. How do you think Google and every other search engine does it? They just spider the sites and index the contents. The devil, of course, is in the details. But it's certainly possible to do.
I don't think so. Unless you want only to fetch some data from a certain HTML page, then you need to use some regular expressions. But searching the database is not possible if you don't have the ability to connect to it directly or via some APIs.

how to find webpages which link to some specific page?

Just wondering if there's any way to search all web pages which link to some specific url? For example, all web pages containing link to example.com? Thanks
You probably might want to explore the Google Search API which allows you to use Google search results in your programs.

How-To get private pages being crawled by google

How can i get private pages of my web site being crawled and indexed by google ?
maybe it's not very "conventionnal", but i want my private page "links" displayed in google index, but next require a registration to display the page.
EDIT: Based on the addition of "maybe it's not very "conventionnal", but i want my private page "links" displayed in google index, but next require a registration to display the page." To the question:
You can check the User Agent in your php code to basically allow google to see pages if it was a registered user (google's user agent is "Googlebot/1.0" and you can search to find user agents for other common engines).
However, this behavior is specifically against google's rules and they can and will remove your site from the index if they catch you doing it. Their policy is you should not treat googlebot any differently than you treat any random person who visits your site.
(Original Answer) One way is to use a sitemap to show google how to find all of your pages.
In general, and even in the case of sitemaps, if the content you want indexed is not linked to from a page that can be found through the "root" (/) (i.e. there is no way for the public to find it), then it probably won't get indexed. The only way to get it indexed is to link it in someplace.
The question is though, why do you want your private pages in google anyway?
They'll get crawled if and only if they're publicly accessible and your robots.txt file allows it. That's pretty much all you need to do.
Are you asking how to get Google to index your pages?
There are a couple of ways. You need to ensure that you have SEO'd, or Search Engine Optimisation, the pages properly with title text and description key words in your meta data.
You can also submit your site to Google, it's a free service, and it'll be placed in a queue of things that Google will index. May take some time though.
By far the best way to get your pages indexed is using the meta data in the pages themselves.
Google will only index what is
linked from somewhere already in Google's index
accessible to its crawler via normal (unauthenticated) HTTP
It will also
make the contents available in search results to anyone.
This may conflict with your idea of a "private" page.
I'm going to assume that all the other previous answerers are misunderstanding you. As I read it, you aren't asking how to get Google to index your pages, but rather how to get a list of all the pages that Google currently has already indexed on your site? If that is true, you should have a look at the Google Webmaster Tools.

Resources