Showing posts with label Google. Show all posts
Showing posts with label Google. Show all posts

Jul 7, 2011

1 comments

Google blocks all CO.CC sites from its search results

Google has taken its fight against spammy sites to the next level. It has decided to stop indexing all co.cc websites.
This move on the part of Google has affected over 11 million sites, who have been removed from Google’s index, and that is quite a large number. It appears that a significant number of websites using this domain were spammy or shady websites. But unfortunately those who were running genuine websites have also been affected by this blanket decision. Try searching for co.cc sites and see.
We absolutely do try to be granular, but I wanted to mention that if we see a very large fraction of sites on a specific freehost be spammy or low-quality, we do reserve the right to take action on the freehost as a whole.
The .cc top-level domain belongs to the Cocos Islands, a small Australian territory in the Indian Ocean. Co.cc is a Korean company based which offers free or almost free subdomains to anyone who is intersted in getting one, and it is this subdomain which has been affected.
So what next? Will  these spammy sites shut down their business? Or will they just get a new domain like .biz which is low-cost? Or perhaps simply move on to create a free account on wordpress.com or blogspot.com?

Feb 26, 2011

4 comments

How to Get Google Adsense Account Approval

Today most of the bloggers  primary income source is  adsense.Before we start how to creating a adsense account .  lets see the concept of  adsense .

Adsense works as a intermediate  or simply like a broker between the publisher (you) and the advertiser  . Adsense gets the targetted ads to your site automatically based on your content in your site .And pays you  based on the no of clicks  you get to the  advertiser site from your site . This form of advertisement is technically known as  cpc – cost per click .

First lets see how to make a site and apply for adsense

Creating  A  Site Or Blog  For  Adsense
To create a site for adsense needed following steps to do
  1. Buy An Domain
  2. Take An Hosting Account .Blue host is probably the best and i am using it from 4 yrs with out any complaint .If you are planning to take the blue host account dont register it directly just comment here  and i will give you hosting account in blue host for 4$ you can mail me using the contact form
  3. Select content of your site(Domain must be relevant tf the site content )
  4. Applying For Adsense

To Apply For New Adsense Account Follow The Steps  :
1.  Go to, http://www.adsense.com then click Sign Up Now
2. Type your account informations :
Provide your Website information
  • Website url : your web address (domain name  .com  )or blog URL
  • Website language : choose the primary language used in your blog
Your Contact information
  • Account type : choose individual  or business -
  • There is no difference in the services or payment structure between Individual and Business accounts. Business account payments will be made payable to the Company Name, while Individual accounts are paid out to the Payee Name of the account holder select your account type based on the requirements
  • Country or territory : select your country
  • Payee name : typing your complete name here or your  name .Your bank account name must match the payee name or you cant process the checks
  • Street address : Typing your address here. This address is used for delivery of cheque.
  • City/town :  your town
  • State : Provide the state name
  • Postal code
  • Phone

NOTE: please  see the requirements before you apply to the adsense account

Creating  a blog for adsense is simple and its free
  1. Go To Blogger.com
  2. Register a blog
  3. Select the content of your blog
  4. Applying for Adsense is as  above
After creating a Site or blog  you need check weather your site satisfies the minimum  requirements  to apply for adsense.
  • What Are Those Requirements?
  1. Your web site must be your top level  domain
  2. Domain must have been registered and active for at least  6 months before applying for adsense
  3. you must have original  content (must not copied from other sites)
  4. The  registration  information that you provide during adsense registration  must match your domain name registration
See these requirements is applied to your site . and apply for the Adsense  application. if your adsense is accepted you will get an message like this .if  it rejected  that you get meassage llike be sure the below rules and resend the application.
If adsense is accepted your site you will get an code in  your adsene account .copy and paste the code provided  html of your page. The  ads on your page shows the relavent of your site content . you can also display ads verticaly or horizontal.Your  daily earnings will look like this. After  one month of your earnings in send in the form of check like this . They will send according to information that you provided during  adsense application.



If your site is not accepted by the adsense .It means you proparly not devolped. there  many other reasons:
  • Is your site an  “About Me” page
Gennerally google does not accept the presonal blogs because they do not have specified content
  • How “ pages” do you have
See that you  must have  atleast  12 pages on your site .
  • You copied  the” content” of your site from other sites
Google easly can be find out that your content of your site is copied from orther sites or not. so you should not copied from orthers
  • Under “construction” sites will not accepted
Google does not accepted the  Under construction sites . make sure that do not keep pages like this
  • “Adults” content can not be accepted
Google does not accepted the  Adult content sites.not only adult content sites the following sites will not be accepted
  1. Hacking/cracking sites
  2. Illicit drugs and drug paraphernalia
  3. Gambling or casino-related content
  4. Excessive advertising
If the following site’s is accepted in rare cases  it will rejected in future of his eanings.

Jan 3, 2011

0 comments

Search Engine Optimization Tips

Misconceptions about search engine optimization are common. One of the most common is that once a site is built and submitted to the search engines, heavy traffic is on its way. Another is that when making a submission to each engine, a site will be registered immediately and will stay listed with that engine for as long as it is in operation. That's just not how it works… not even close!

What people need to know is that search engine optimization, which is actually the effective utilization of search engines to draw traffic to a Web site, is an art. It is an ongoing, continuously evolving, high maintenance process that includes the customization of a site for better search engine ranking.

Critical steps to take before submitting

After developing a Web site and selecting the best hosting company, don't rush out and submit it to search engines immediately. A Web site manager would be wise to take a little time to:

Fine tune the TITLE tag to increase traffic to the site
Improving the TITLE tag is one technique that applies to just about all the search engines. The appearance of key words within the page title is one of the biggest factors determining a Web site's score in many engines. It's surprising how many Web sites have simple, unimaginative titles like "Bob's Home Page" that don't utilize keywords at all. In fact, it's not unusual to see entire Web sites that use the same title on every page in the site. Changing page titles to include some of the site's key words can greatly increase the chance that a page will appear with a strong ranking in a query for those key words.
 
Create gateway pages that are specific to the focus of each site
Key word selection must be done carefully with great forethought and understanding of the search engine's selection criteria for key words. The larger the number of key words that are used, the more the relevance of any one key word is diluted. One way to get around this is to create gateway pages.
Gateway pages are designed specifically for submission to a search engine. They should be tuned with a specific set of key words, boosting the chance that these key words will be given a heavy weight. To do this, several copies of a page should be made, one for each set of key words. These pages will be used as entry points only, to help people find the site, therefore, they don't need to fit within the normal structure of the site. This provides the page developer with greater flexibility in establishing key words and tags that will encourage a stronger ranking with the search engines. Each gateway page then can be submitted separately to the search engines.
 
Ensuring that site technology won't confuse the search engines
Often the latest technology being built into a site can confuse the search engine spiders. Frames, CGI scripts, image maps and dynamically generated pages are all recently created technology that many spiders don't know how to read. With frames for instance, the syntax of the FRAMESET tag fundamentally changes the structure of an HTML document. This can cause problems for search engines and browsers that don't understand the tag. Some browsers can't find the body of the page and viewing a page through these browsers can create a blank page.
Today only 2% of browsers don't support frames, but many search engine spiders still don't support them. A search engine spider is really just an automated Web browser and like browsers they sometimes lag behind in their support for new HTML tags. This means that many search engines can't spider a site with frames. The spider will index the page, but won't follow the links to the individual frames.
 
Setting up a NOFRAMES section on the page
Every page that uses frames should include a NOFRAMES section on the page. This tag will not affect the way a page looks but it will help a page get listed with the major search engines. The NOFRAMES tag was invented by Netscape for backward compatibility with browsers that didn't support the FRAME and FRAMESET tags.
 
Performing a maintenance check
All Web sites should be thoroughly tested using a site maintenance tool in order to catch errors in operation before customers are brought to the site. HTML errors can hinder a search engine spider's ability to index a site, it can also keep a search engine from reading a page or cause it to be viewed in a manner different from how it was intended. In fact, a recent report by Jupiter Communications suggested 46% of users have left a preferred Web site because of a site-related problem. With NetMechanic's HTML Toolbox or another site maintenance tool, all Webmasters, from the novice to the expert can avoid potential visitor disasters due to site errors.
 
Finding the best submission service
Selecting a search engine submission service requires careful thought and important decisions. Using an auto submission service is a good place to begin. Most search engines like Alta Vista, HotBot and InfoSeek automatically spider a site, index it and hopefully add it to their search database without any human involvement. Some engines, like Yahoo, are done completely with human review and for many reasons are best submitted individually. Chances are good also, that in the first submission a site will be rejected by several of the engines and will need to be individually resubmitted. There are several online resources for auto submissions. The best ones won't submit a site to Yahoo where the customer is better served doing this on his own.
 
Understanding the waiting periods
A variety of waiting periods must be endured with each search engine before there is even a hope of being listed. Knowing and understanding these waiting periods before beginning the process can eliminate or at least minimize frustration and confusion. Typical waiting periods for some of the more popular engines are six months with Yahoo; one to two months with Lycos and 4-6 weeks with Excite or is that 4-6months? What they say and what happens in reality can be very different.

Ongoing promotion tasks:

To improve site rankings and increase understanding of the listing process, there are many tasks that can be done on a regular or semi-regular basis. Optimizing rankings within the search engines is also to help ensure that a site attracts the right traffic.

Some of the monthly and weekly promotion tasks are:
 
Crunching and examining log files
Data contained in log files is an excellent resource for identifying which engines are sending the majority of traffic to a site. It can also show which key words or gateway pages are generating the strongest traffic and what are those visitors doing when they enter the site.
 
Searching the Search Engines
Conduct a search of the search engines to analyze where the highest rankings of the site have materialized and what keywords are generating the best rankings. Different search engines use different rules to rank pages. Individual gateway pages should be created based on the knowledge and interpretation of what each search engine is using to determine top rankings. Several pages can be tested out on one or more engines and the pages that have the most success can be kept, while the unsuccessful pages can be dumped or revised to achieve a higher ranking.

Learning more about how the search engines work
Each search engine uses different rules to determine how well a Web page matches a particular query. As a result, building a single page that gets a good score in all the major engines is just about impossible. Learning how each engine ranks pages is also hard, since the engines often keep this information as a closely guarded secret. However, with a little patience, some experimentation and reverse engineering, the way that many of the search engines work can be discovered.

Resubmitting the site
For engines that reject a site or don't list it high enough, it is strongly recommended that more information is learned about the engine's criteria before resubmitting. This information should then be incorporated into gateway pages or key word revisions in order to have greater success with subsequent submissions. Fine tune the page (or pages) make adjustments to TITLE tags and META tags, then after resubmitting the site, track the results to further learn about the engine's criteria and which adjustments made an impact on the rankings. Don't be afraid to experiment, take some risks and gather data as you proceed.

Checking log files for traffic being directed to erroneous pages on the site
This is good news!! Don't dump these pages or remove them from the search engine as most people will do when they redesign their site. Any page with a high ranking is of value. If a page is bringing traffic to a site, leave that page on the search engine, don't change it but rather redirect the traffic to valid pages in the site.

Getting Noticed

For small to medium-sized Web sites, search engines are the most important source of traffic. Unfortunately, getting noticed in the search engines isn't an easy job. A Web site manager can spend months getting a site listed in an engine, only to find it ranks 50th in their search results. It's hard to give universal tips for improving search engine ranking because each engine has its own set of rules. In general, though, a page will rank well for a particular query if the search terms appear in the TITLE tag, the META tags, and in the body of the page.

Oct 31, 2009

0 comments

How Google Works

iw                        Google runs on a distributed network of thousands of low-cost computers and can therefore carry out fast parallel processing. Parallel processing is a method of computation in which many calculations can be performed simultaneously, significantly speeding up data processing. Google has three distinct parts:
  • Googlebot, a web crawler that finds and fetches web pages.
  • The indexer that sorts every word on every page and stores the resulting index of words in a huge database.
  • The query processor, which compares your search query to the index and recommends the documents that it considers most relevant.

1. Googlebot, Google’s Web Crawler

                          Googlebot is Google’s web crawling robot, which finds and retrieves pages on the web and hands them off to the Google indexer. It’s easy to imagine Googlebot as a little spider scurrying across the strands of cyberspace, but in reality Googlebot doesn’t traverse the web at all. It functions much like your web browser, by sending a request to a web server for a web page, downloading the entire page, then handing it off to Google’s indexer.

                           Googlebot consists of many computers requesting and fetching pages much more quickly than you can with your web browser. In fact, Googlebot can request thousands of different pages simultaneously. To avoid overwhelming web servers, or crowding out requests from human users, Googlebot deliberately makes requests of each individual web server more slowly than it’s capable of doing.
Googlebot finds pages in two ways: through an add URL form, www.google.com/addurl.html, and through finding links by crawling the web.

Screen shot of web page for adding a URL to Google.

                                          Unfortunately, spammers figured out how to create automated bots that bombarded the add URL form with millions of URLs pointing to commercial propaganda. Google rejects those URLs submitted through its Add URL form that it suspects are trying to deceive users by employing tactics such as including hidden text or links on a page, stuffing a page with irrelevant words, cloaking (aka bait and switch), using sneaky redirects, creating doorways, domains, or sub-domains with substantially similar content, sending automated queries to Google, and linking to bad neighbors. So now the Add URL form also has a test: it displays some squiggly letters designed to fool automated “letter-guessers”; it asks you to enter the letters you see — something like an eye-chart test to stop spambots.

                                         When Googlebot fetches a page, it culls all the links appearing on the page and adds them to a queue for subsequent crawling. Googlebot tends to encounter little spam because most web authors link only to what they believe are high-quality pages. By harvesting links from every page it encounters, Googlebot can quickly build a list of links that can cover broad reaches of the web. This technique, known as deep crawling, also allows Googlebot to probe deep within individual sites. Because of their massive scale, deep crawls can reach almost every page in the web. Because the web is vast, this can take some time, so some pages may be crawled only once a month.

                                      Although its function is simple, Googlebot must be programmed to handle several challenges. First, since Googlebot sends out simultaneous requests for thousands of pages, the queue of “visit soon” URLs must be constantly examined and compared with URLs already in Google’s index. Duplicates in the queue must be eliminated to prevent Googlebot from fetching the same page again. Googlebot must determine how often to revisit a page. On the one hand, it’s a waste of resources to re-index an unchanged page. On the other hand, Google wants to re-index changed pages to deliver up-to-date results.

                                       To keep the index current, Google continuously recrawls popular frequently changing web pages at a rate roughly proportional to how often the pages change. Such crawls keep an index current and are known as fresh crawls. Newspaper pages are downloaded daily, pages with stock quotes are downloaded much more frequently. Of course, fresh crawls return fewer pages than the deep crawl. The combination of the two types of crawls allows Google to both make efficient use of its resources and keep its index reasonably current.

2. Google’s Indexer

                  Googlebot gives the indexer the full text of the pages it finds. These pages are stored in Google’s index database. This index is sorted alphabetically by search term, with each index entry storing a list of documents in which the term appears and the location within the text where it occurs. This data structure allows rapid access to documents that contain user query terms.

                  To improve search performance, Google ignores (doesn’t index) common words called stop words (such as the, is, on, or, of, how, why, as well as certain single digits and single letters). Stop words are so common that they do little to narrow a search, and therefore they can safely be discarded. The indexer also ignores some punctuation and multiple spaces, as well as converting all letters to lowercase, to improve Google’s performance.

3. Google’s Query Processor

The query processor has several parts, including the user interface (search box), the “engine” that evaluates queries and matches them to relevant documents, and the results formatter.
PageRank is Google’s system for ranking web pages. A page with a higher PageRank is deemed more important and is more likely to be listed above a page with a lower PageRank.
Google considers over a hundred factors in computing a PageRank and determining which documents are most relevant to a query, including the popularity of the page, the position and size of the search terms within the page, and the proximity of the search terms to one another on the page. A patent application discusses other factors that Google considers when ranking a page. Visit SEOmoz.org’s report for an interpretation of the concepts and the practical applications contained in Google’s patent application.

Google also applies machine-learning techniques to improve its performance automatically by learning relationships and associations within the stored data. For example, the spelling-correcting system uses such techniques to figure out likely alternative spellings. Google closely guards the formulas it uses to calculate relevance; they’re tweaked to improve quality and performance, and to outwit the latest devious techniques used by spammers.

Indexing the full text of the web allows Google to go beyond simply matching single search terms. Google gives more priority to pages that have search terms near each other and in the same order as the query. Google can also match multi-word phrases and sentences. Since Google indexes HTML code in addition to the text on the page, users can restrict searches on the basis of where query words appear, e.g., in the title, in the URL, in the body, and in links to the page, options offered by Google’s Advanced Search Form and Using Search Operators (Advanced Operators).

Let’s see how Google processes a query.
1. The web server sends the query to the index        servers. The content inside the index servers is similar        to the index in the back of a book--it tells which pages        contain the words that match any particular query       term.          2. The query travels to the doc servers, which   actually retrieve the stored documents. Snippets are    generated to describe each search result.       3. The search results are returned to the user          in a fraction of a second.

A Closer Look at a Vulnerability in Gmail

Gmail is one of the major webmail service provider across the globe. But as we all know Gmail still carries that 4 letter word BETA. Sometimes we may wonder, why Gmail is still in the testing stage even after years of it’s emergence. Here is one small reason for that.

Gmail follows a strict rule that doesn’t allow it’s users to have their first or the last name contain the term Gmail or Google. That is, while signing up for a new Gmail account the users cannot choose a first or last name that contains the term Gmail or Google.

This rule is implemented by Gmail for obvious reasons, because if the users are allowed to keep their first or the last name that contains the term Gmail or Google, then it is possible to easily impersonate the identity of Gmail (or Gmail Team) and engage themselves in phising or social engineering attacks on the innocent users. This can be done by simply choosing the first and last name with the following combinations.

First Name        Last Name
Gmail                       Team
Google                     Team
Gmail                       Password Assistance

               From the above snapshot we can see that, Gmail has made a good move in stopping the users from abusing it’s services. However this move isn’t just enough to prevent the malicious users from impersonating the Gmail’s identity. Because Gmail has a small vulnerability that can be exploited so that the users can still have their name contain the terms Gmail or Google.

You may wonder how to do this. But it’s very simple.

1. Login to your Gmail account and click on Settings.
2. Select Accounts tab
3. Click on edit info
4. In the Name field, select the second radio button and enter the name of your choice. Click on Save Changes and you’re done!
Now, Gmail accepts any name even if it contains the term Google or Gmail. You can see from the below snapshot

                         Allowing the users to have their names contain the terms Gmail or Google is a serious vulnerability even though it doesn’t seem to be a major one. This is because a hacker or a malicious attacker can easily exploit this flaw and send phishing emails to other Gmail users asking for sensitive information such as their passwords. Most of the users don’t even hesitate to send their passwords since they believe that they are sending it to Gmail Team (or someone authorized). But in reality they are sending it to an attacker who uses these information to seek personal benefits.

So the bottomline is, if you get any emails that appears to have come from the Gmail Team or similar, don’t trust them! Anyone can send such emails to fool you and take away your personal details. Hope that Gmail will fix this vulnerability as soon as possible to avoid any disasters.

How To Find If Your Friend is Invisible or Offline on Gtalk

You can easily find if your friend is invisible or actually offline using this trick. But to use this trick, you have to catch him online at least once. So if you doubt that one of your friends is invisible from you too often, then-

1. Whenever you see him online, just set the chat as off the record with him.
If you are using the Gmail’s inbuilt chat facility, then simply click on the Video & more link in the lower left corner of the chat widget and click on Go off the record.

The message You are now off the record will be displayed after that.
If you are on Gtalk, then click on settings->chat, and check the box that says, don’t save my chat history in my gmail account. And save this setting.
So, now your chats with your naughty friend are off the record.
2. Now you can easily detect if your friend is actually offline or just invisible just by sending her any message. If she is actually offline, then the following message will be displayed,

And if she is online, nothing will be displayed, which means that she actually received your chat message, and thus, is online.

Why it works?

By default, all your chat sessions are stored in your Gmail account (and in the account of the person who chatted with you). Now, once your chat with one of your friend (whom you are spying ) are set off the record, they are no longer stored in the Gmail account. So if you send a chat message to her, it must be displayed as a popup windows. If she is actually offline, it can’t be done, so the message xyz did not receive your chat is displayed. On the other hand if she is just invisible, she can see the popup, and hence no message is displayed.