Welcome to My Blog

Here i share some information about google and also about SEO, so please check out this, and give your suggestion regarding the same subjects.
Showing posts with label Penguin. Show all posts
Showing posts with label Penguin. Show all posts

Tuesday, May 22, 2012

Two Weeks In, Google Talks Penguin Update, Ways To Recover & Negative SEO

It’s been about two weeks since Google launched its Penguin Update. Google’s happy the new spam-fighting algorithm is improving things as intended. But some hurt by it are still wondering how to recover, and there remain concerns about “negative SEO” as a threat. I caught up with Matt Cutts, the head of Google’s web spam team, on these and some related questions.

Penguin: “A Success”

The goal of any algorithm update is to improve search results. So how’s Penguin been for Google?
“It’s been a success from our standpoint,” Cutts said.

What About Those Weird Results?

Of course, soon after Penguin was released, people quickly started citing examples of odd results. The official Viagra site wasn’t listed, while hacked sites were. An empty web site was listed for “make money online,” and there were reports of other empty sites ranking well. Scraper sites were reported outranking the sites they scraped.

How could Penguin be a success with these types of things happening?

Cutts said that many of these issues existed before Penguin launched and were not caused by the new spam-fighting algorithm.

Indeed, the Viagra issue, which has now been fixed, was a problem before Penguin hit. Penguin didn’t cause it.

False Positives? A Few Cases

How about false positives, people who feel they’ve been unfairly hit by Penguin when they weren’t doing any spam?

“We’ve seen a few cases where we might want to investigate more, but this change hasn’t had the same impact as Panda or Florida,” Cutts said.

The Panda Update was Google’s big update that targeted low-quality spam last year. The Florida Update was a major Google update in 2003 intended to improve its search quality.

I’d agree that both of those seemed to have impacted more sites than Penguin has, based on having watched reactions to all these updates. Not everyone will agree with me, of course. It’s also worth the regular reminder that for any site that “lost” in the rankings, someone gained. You rarely hear from those who gain.
Bottom line, Google seems pretty confident that the Penguin Update is indeed catching people who were spamming, as was intended.

Why Spam Still Gets Through

Certainly when I’ve looked into reports, I’ve often found spam at the core of why someone dropped. But if Penguin is working, why are some sites that are clearly spamming still getting through?

“No algorithm is perfect. While we’d like to achieve perfection, our litmus test is, ‘Do things get better than before?’,” Cutts said.

Cutts also explained that Penguin was designed to be quite precise, to act against pages when there was an extremely high-confidence of spam being involved. The downside is that some spam might get through, but the upside is that you have fewer false positives.

How Can You Recover?

One of the most difficult things with this update is telling people how to recover. Anyone hit by Penguin was deemed to be spamming Google.

In the past, if you spammed Google, you were told to file a reconsideration request. However, Google’s specifically said that reconsideration requests won’t help those hit by Penguin. They’ll recover naturally, Google says, if they clean the spam up.

However, one of the main reasons I’ve seen when looking at sites hit by Penguin seems to be bad linking practices. People have used sponsored WordPress themes, or poor quality reciprocal linking, have purchased links or participated in linking networks, such as those recently targeted by Google.
How do people pull themselves out of these link networks, if perhaps they don’t have control over those links now?

“It is possible to clean things up,” Cutts said, and he suggested people review two videos he’s done on this topic:

 

 


“The bottom line is, try to resolve what you can,” Cutts said.

Waiting On Penguin To Update Again

If you do clean things up, how will you know? Ideally, you’ll see your traffic from Google recover, the next time Penguin is updated.

That leads to another important point. Penguin, like Panda, is a filter that gets refreshed from time-to-time. Penguin is not constantly running but rather is used to tag things as spam above-and-beyond Google’s regular spam filtering on a periodic basis.

Is Penguin a site-wide penalty like Panda or page-specific? Cutts wouldn’t say. But given that Panda has site-wide impacts, I think it’s a fair assumption that Penguin works the same.

What that means is that if some of your site is deemed Penguin-like, all of it may suffer. Again, recovery means cleaning up the spam. If you’ve cleaned and still don’t recover, ultimately, you might need to start all over with a fresh site, Cutts said.

New Concerns Over Negative SEO

Before Penguin, talk of “negative SEO” had been ramping up. Since then, it seems to have gotten worse in some places. I’ve seen post-after-post making it sound as if anyone is now in serious danger that some competitor can harm them.

At the core of these fears seems to be a perfect storm of assumptions. Google recently targeted some linking schemes. That caused some people to lose traffic. Google also sent out warnings about sites with “artificial” or “unnatural” links. That generated further concerns in some quarters. Then the Penguin Update hit, which caused more people to lose traffic as they were either hit for link spam or no longer benefited from link spam that was wiped out.

These things made it ripe for people to assume that pointing bad links at a site can hurt it. But as I wrote before, negative SEO concerns aren’t new. They’ve been around for years. Despite this, we’ve not seen it become a major concern.

Google has said it’s difficult for others to harm a site, and that’s indeed seemed to be the case. In particular, pointing bad links at a good site with many other good signals seems to be like trying to infect it with a disease that it has antibodies to. The good stuff outweighs the bad.

Cutts stressed again that negative SEO is rare and hard. “We have done a huge amount of work to try to make sure one person can’t hurt another person,” he said.

Cutts also stressed again what Google said before. Most of the those 700,000 messages to publishers that Google sent out earlier this year were not about bad link networks. Nor were they all suddenly done on the same day. Rather, many sites have had both manual and algorithmic penalties attached to them over time but which were never revealed. Google recently decided to open up about these.

After Negative SEO Campaign, A Link Warning

Of course, new messages do go out, which leads to the case of Dan Thies. His site was targeted by some trying to show that negative SEO works. He received an unnatural link warning after this happened. He also lost some rankings. Is this the proof that negative SEO really works?

Thies told me that his lost rankings were likely due to changes he made himself, when he removed a link across all pages on his site that led back to his home page. After restoring that, he told me, he regained his rankings.

His overall traffic, he said, never got worse. That tends to go against the concerns that negative SEO is a lurking threat, because if it had worked enough to tag his site as part of the Penguin Update, he should have seen a huge drop.

Still, what about link warning? Thies did believe that came because of the negative SEO attempt. That’s scary stuff. He also said he filed three reconsideration requests, which each time returned messages saying that there were no spam actions found. Was he hit with a warning but not one that was also associated with a penalty?

I asked Cutts about the case, but he declined to comment on Thies’s particular situation. He did say that typically a link warning is a precursor to a ranking drop. If the site fixes the problem and does a reconsideration request quickly enough, that might prevent a drop.

Solving The Concerns

I expect we’ll continue to see discussions of negative SEO, with a strong belief by some that it’s a major concern for anyone. I was involved in one discussion over at SEO Book about this that’s well worth a read.
When it’s cheaper to buy links than ever, it’s easy to see why there are concerns. Stories like what happened to Thies or this person, who got a warning after 24,000 links appeared pointing at his site in one day, are worrisome.

Then again, the person’s warning came after he apparently dropped in rankings because of Penguin. So did these negative SEO links actually cause the drop, or was it something else? As is common, it’s hard to tell, because the actual site isn’t provided.

To further confuse matters, some who lost traffic because of Penguin might not be victims of a penalty at all. Rather, Google may have stopped allowing some links to pass credit, if they were deemed to be part of some attempt to just manipulate rankings. If sites were heavily dependent on these artificial links, they’d see a drop just because the link credit was pulled, not because they were hit with a penalty.

I’ve seen a number of people now publicly wishing for a way to “disvow” links pointing at them. Google had no comment about adding such a feature at this time, when I asked about this. I certainly wouldn’t wait around for it now, if you know you were hit by Penguin. I’d do what you can to clean things up.

One good suggestion out of the SEO Book discussion was that Google not penalize sites for bad links pointing at them. Ignore the links, don’t let the links pass credit, but don’t penalize the site. That’s an excellent suggestion for defusing negative SEO concerns, I’d say.

I’d also stress again that from what I’ve seen, negative SEO isn’t really what most hit by Penguin should probably be concerned about. It seems far more likely they were hit by spam they were somehow actively involved in, rather than something a competitor did.

Recovering From Penguin

Our Google Penguin Update Recovery Tips & Advice post from two weeks ago gave some initial advice about dealing with Penguin, and that still holds up. In summary, if you know that you were hit by Penguin (because your traffic dropped on April 24):
  • Clean up on-page spam you know you’ve done
  • Clean up bad links you know you’re been involved with, as best you can
  • Wait for news of a future Penguin Update and see if you recover after it happens
  • If it doesn’t, try further cleaning or consider starting over with a fresh site
  • If you really believe you were a false positive, file a report as explained here
Just in, by the way, a list of WordPress plug-ins that apparently insert hidden links. If you use some of these, and they have inserted hidden links, that could have caused a penalty.

I’d also say again, take a hard look at your own site. When I’ve looked at sites, it’s painfully easy to find bad link networks they’ve been part of. That doesn’t mean that there’s not spam that’s getting past Penguin. But complaining about what wasn’t caught isn’t a solution to improving your own situation, if you were hit.


Sunday, May 13, 2012

Google: Can't Recover From Penguin? Start A New Site

Danny Sullivan published a new story yesterday named Two Weeks In, Google Talks Penguin Update, Ways To Recover & Negative SEO.

In that article, he interviews Google's spam lead, Matt Cutts on ways to recover from the Google Penguin update. There are some solid tips there but scary ones also.

Here is one that is scary for those who were hit by the Penguin update:
If you've cleaned and still don't recover, ultimately, you might need to start all over with a fresh site, Cutts said.
Yes, that is scary for someone who was hit, is trying to frantically make changes but has not seen any recovery. Now, if you have not seen a recovery yet - I wouldn't worry, I don't think they refreshed the update yet, so there wouldn't be any recoveries in my opinion.

But Google is not going to roll this back. Google's Matt Cutts said, "It's been a success from our standpoint." Were there false positives? Few Cutts said, "we've seen a few cases where we might want to investigate more, but this change hasn't had the same impact as Panda or Florida." Very interesting.

Key Take Aways:

(1) Google is not going to roll this update back.
(2) Google says it had less of an impact than Panda or Florida.
(3) Don't take drastic measures yet, do what you can now so when Google does refresh the update, maybe you can break free.

Source :  http://www.seroundtable.com/google-penguin-recovery-15136.html

Tuesday, May 8, 2012

Google Announced 50+ Search Updates, Which Are Penguin Related?

In Google fashion, late on Friday, Google released their now monthly update on the changes they made to Google search over the past month. It is really great that Google does this and this time they shared 53 changes in April. Here is last months update.
Below I grouped and listed out the more important changes, at least the ones I find to me most important.

But let's try to see which items in this list are Penguin related. Can we even figure that out?

 Penguin Related?
  • Anchors bug fix
  • Keyword stuffing classifier improvement
  • More authoritative results
  • Improvement in a freshness signal
  • No freshness boost for low-quality content
  • Improvements to how search terms are scored in ranking
If I had to guess, these and maybe more, are all related to the Penguin update.
Here are some more that I find important but wouldn't specifically related to Penguin, Panda or others:

Ranking Changes:

  • Improvement in a freshness signal. [launch codename "citron", project codename "Freshness"] This change is a minor improvement to one of the freshness signals which helps to better identify fresh documents.
  • No freshness boost for low-quality content. [launch codename "NoRot", project codename "Freshness"] We have modified a classifier we use to promote fresh content to exclude fresh content identified as particularly low-quality.
  • Smoother ranking changes for fresh results. [launch codename "sep", project codename "Freshness"] We want to help you find the freshest results, particularly for searches with important new web content, such as breaking news topics. We try to promote content that appears to be fresh. This change applies a more granular classifier, leading to more nuanced changes in ranking based on freshness.
  • Improvements to how search terms are scored in ranking. [launch codename "Bi02sw41"] One of the most fundamental signals used in search is whether and how your search terms appear on the pages you're searching. This change improves the way those terms are scored.
  • Backend improvements in serving. [launch codename "Hedges", project codename "Benson"] We've rolled out some improvements to our serving systems making them less computationally expensive and massively simplifying code.
  • Keyword stuffing classifier improvement. [project codename "Spam"] We have classifiers designed to detect when a website is keyword stuffing. This change made the keyword stuffing classifier better.
  • More authoritative results. We've tweaked a signal we use to surface more authoritative content.

Link Analysis Changes:

  • Anchors bug fix. [launch codename "Organochloride", project codename "Anchors"] This change fixed a bug related to our handling of anchors.

Index Updates:

  • Increase base index size by 15%. [project codename "Indexing"] The base search index is our main index for serving search results and every query that comes into Google is matched against this index. This change increases the number of documents served by that index by 15%. *Note: We're constantly tuning the size of our different indexes and changes may not always appear in these blog posts.
  • New index tier. [launch codename "cantina", project codename "Indexing"] We keep our index in "tiers" where different documents are indexed at different rates depending on how relevant they are likely to be to users. This month we introduced an additional indexing tier to support continued comprehensiveness in search results.

Search Listings:

  • More domain diversity. [launch codename "Horde", project codename "Domain Crowding"] Sometimes search returns too many results from the same domain. This change helps surface content from a more diverse set of domains.
  • Categorize paginated documents. [launch codename "Xirtam3", project codename "CategorizePaginatedDocuments"] Sometimes, search results can be dominated by documents from a paginated series. This change helps surface more diverse results in such cases.
  • Country identification for webpages. [launch codename "sudoku"] Location is an important signal we use to surface content more relevant to a particular country. For a while we've had systems designed to detect when a website, subdomain, or directory is relevant to a set of countries. This change extends the granularity of those systems to the page level for sites that host user generated content, meaning that some pages on a particular site can be considered relevant to France, while others might be considered relevant to Spain.
  • Disable salience in snippets. [launch codename "DSS", project codename "Snippets"] This change updates our system for generating snippets to keep it consistent with other infrastructure improvements. It also simplifies and increases consistency in the snippet generation process.
  • More text from the beginning of the page in snippets. [launch codename "solar", project codename "Snippets"] This change makes it more likely we'll show text from the beginning of a page in snippets when that text is particularly relevant.
  • Tweak to trigger behavior for Instant Previews. This change narrows the trigger area for Instant Previews so that you won't see a preview until you hover and pause over the icon to the right of each search result. In the past the feature would trigger if you moused into a larger button area.
  • Better query interpretation. This launch helps us better interpret the likely intention of your search query as suggested by your last few searches.
  • News universal results serving improvements. [launch codename "inhale"] This change streamlines the serving of news results on Google by shifting to a more unified system architecture.
  • More efficient generation of alternative titles. [launch codename "HalfMarathon"] We use a variety of signals to generate titles in search results. This change makes the process more efficient, saving tremendous CPU resources without degrading quality.
  • More concise and/or informative titles. [launch codename "kebmo"] We look at a number of factors when deciding what to show for the title of a search result. This change means you'll find more informative titles and/or more concise titles with the same information.
  • "Sub-sitelinks" in expanded sitelinks. [launch codename "thanksgiving"] This improvement digs deeper into megasitelinks by showing sub-sitelinks instead of the normal snippet.
  • Better ranking of expanded sitelinks. [project codename "Megasitelinks"] This change improves the ranking of megasitelinks by providing a minimum score for the sitelink based on a score for the same URL used in general ranking.
  • Sitelinks data refresh. [launch codename "Saralee-76"] Sitelinks (the links that appear beneath some search results and link deeper into the site) are generated in part by an offline process that analyzes site structure and other data to determine the most relevant links to show users. We've recently updated the data through our offline process. These updates happen frequently (on the order of weeks).
  • Less snippet duplication in expanded sitelinks. [project codename "Megasitelinks"] We've adopted a new technique to reduce duplication in the snippets of expanded sitelinks.

Local Changes:

  • More local sites from organizations. [project codename "ImpOrgMap2"] This change makes it more likely you'll find an organization website from your country (e.g. mexico.cnn.com for Mexico rather than cnn.com).
  • Improvements to local navigational searches. [launch codename "onebar-l"] For searches that include location terms, e.g. [dunston mint seattle] or [Vaso Azzurro Restaurant 94043], we are more likely to rank the local navigational homepages in the top position, even in cases where the navigational page does not mention the location.
  • More comprehensive predictions for local queries. [project codename "Autocomplete"] This change improves the comprehensiveness of autocomplete predictions by expanding coverage for long-tail U.S. local search queries such as addresses or small businesses.

Images & Videos:

  • Improvements to SafeSearch for videos and images. [project codename "SafeSearch"] We've made improvements to our SafeSearch signals in videos and images mode, making it less likely you'll see adult content when you aren't looking for it.
  • Improved SafeSearch models. [launch codename "Squeezie", project codename "SafeSearch"] This change improves our classifier used to categorize pages for SafeSearch in 40+ languages.
Sources : http://www.seroundtable.com/google-updates-april-15111.html