营销策划

    今日:143| 主题:37668
收藏本版 (15)
网络、PC端、移动端的营销、策划实战和资讯。

[其他] Back to Search School: A Quick 15-Minute Audit by @ab80

[复制链接]
浮夸生 发表于 2016-10-6 18:18:10
297 5

立即注册CoLaBug.com会员,免费获得投稿人的专业资料,享用更多功能,玩转个人品牌!

您需要 登录 才可以下载或查看,没有帐号?立即注册

x

Back to Search School: A Quick 15-Minute Audit by @ab80-1 (something,software,research,entering,separate)

  An SEO specialist’s job is never finished—you could spend a lifetime tweaking your website, and you’d still find something to improve. But for those of us who don’t have a lifetime to spend split-testing every element, I’ve created this 15-minute guide for auditing your website.
  Note: For this guide, I used the latest version of my tool, WebSite Auditor, to crawl my website, analyze site errors, fix broken links and redirects, review my title and meta tags, optimize on-page content and images, etc. I use this software instead of separate crawling, keyword research, and link analysis tools, but you can follow along using your favorite auditing tools instead.
  1. Crawl and Audit Your Website

   Quick audits are helpful, butcomprehensive audits are invaluable when it comes to the health of your website. Start crawling your website by entering its URL into your favorite SEO crawling tool. Make sure that your crawling tool is capable of scanning every resource on your website (JavaScript, CSS, Images, Video, Flash, etc.).
  This scan usually only takes a few minutes, but time will vary depending on the size of your website.
  Checklist:

  
       
  • Crawl your website.  
  2. Check for Crawlability/Indexing Issues

  Once you’ve scanned your website, it’s time to review anything that might be affecting your website’s crawlability. Some of the most common problems include poorly maintained sitemaps, robots.txt errors, and HTTP response code errors.
  Check your sitemap

  If you’re using a tool or plugin to automatically manage your sitemap, you should be able to automatically create an XML sitemap or retrieve, view, and validate an existing sitemap. If you’re not using a sitemap tool, you’ll have to manually review your sitemap. Either way, you want to make sure that your sitemap is accurate and up-to-date.
   Once you’ve updated your sitemap, check it for errors in Google Search Console. Simply log in to your account, go to Crawl > Sitemaps and then click the Add/Test Sitemaps button. Enter the URL of your sitemap and click Test to see if Google encounters any errors while parsing your sitemap.
  Check your robots.txt

  A page might be blocked from indexing for a variety of reasons. For example, it could have a “noindex” tag in its section, the X-Robots-Tag in the HTTP header, or it could be that your robots.txt file itself is blocking the page. Using your auditing tool, check your robots.txt’s instructions and see if any important pages are being blocked. If they are, you need to add an allow rule to enable indexing.
   If you’re struggling with CSS and JavaScript being disallowed in particular, Google’s Gary Illyes recommends you allow Google to crawl these resources by adding the following three lines of code to your robots.txt file:
  Check your website for errors

   HTTP response code errors are easy to detect and responsible for many indexing errors. You can check your website for crawl errors through your site auditing tool or by going to Crawl > Crawl Errors in your Google Search Console.
  You can correct whatever 4XX errors you find by fixing your internal links or setting up 301 redirects that point visitors to similar pages (especially useful on e-commerce sites). Also, make sure that your website has a 404 page set up in the first place to link your visitors back to your website’s important pages.
  Ask Google (and other search engines) to re-crawl your pages

  Whenever you’ve finished making these essential on-site improvements, you can ask search engines to re-crawl your pages. This will let you know if you still have errors that need addressing or if you can submit your page for indexing.
   To have Google re-crawl a page, log into your Google Search Console, go to Crawl > Fetch as Google , enter the URL of the updated page, and click Fetch .
  Checklist:

  
       
  • Fix any errors on your sitemap.   
  • Review your robots.txt file.   
  • Ensure all of your important pages can be indexed.   
  • Fix HTTP errors.   
  • Ask search engines to re-crawl your site.  
  3. Avoid Common Redirect Mistakes

  If redirects don’t properly funnel visitors to the right page when that page has moved or changed URL, they pose a potential problem for SEO. Here are some of the most common issues:
  302 redirects

  302 redirects, aka. “temporary” redirects, tell search engines that you’re temporarily redirecting visitors, but that the missing page will return shortly. As such, 302 redirects won’t pass on many qualities (e.g. PageRank, Traffic Value, etc.) that a 301 redirect would. This is fine if you’re doing something like testing a new page, but you should try to replace these redirects with 301s otherwise.
  Meta refresh

  Inserting code into a page’s metadata that automatically redirects visitors results in a “meta refresh.” Spammers tend to abuse this by redirecting visitors to pages with unrelated content, and as such it is penalized by most search engines.
  Rel=”canonical” issues

   If you have some duplicate content on multiple pages, you should indicate which page is the primary page by setting it up as your canonical URL. Each page should have no more than one canonical URL. Most site auditors will allow you to view your rel=”canonical” pages under a Redirects tab.
  Checklist:

  
       
  • Replace temporary redirects (302) with permanent redirects (301).   
  • Remove meta refresh from pages   
  • Check for rel=”canonical” issues.  
  4. Fix HTML and CSS markup Errors

  Performing regular website audits can help you detect and correct these errors as they crop up. Some issues to look for include pages that are too big, pages that use frames, and pages that contain W3C errors and warnings.
  Checklist:

  
       
  • Find and correct common coding errors  
  5. Fix Your Links and URLs

   One of the fastest ways to regain lost link equity is to find and fix broken backlinks . Use your auditing tool to keep tabs on all of your Broken Links and then fix them by either contacting the webmasters who have linked to your website or by creating 301 redirects that link back to relevant, updated pages.
  Here are some other best practices:
  Limit outgoing links

  When 100 or more outbound links originate from a single page, it tells search engines that the page might be spam. Limit the number of links you include per page.
  Limit redirects

  Whenever possible, fix 301 and 302 redirects by changing links to lead visitors directly to their destination page. The more links you have, the more you use up your search engine crawl budget.
  Change dynamic URLs

  Dynamically generated URLs are often difficult to read and don’t tell users where they’re heading. Try to fix dynamic URLs so that they’re more user-friendly.
  Shorten long URLs

   Recent evidence suggests that URLs with 35–40 characters dominate search listings. Of course, long URLs can rank—but practice brevity wherever possible.
  Checklist:

  
       
  • Fix broken links   
  • Mitigate outgoing links   
  • Limit redirects   
  • Fix unreadable URLs   
  • Keep your URLs short and sweet  
  6. Review Your Title Tags, Meta Tags, and Header Tags

  It doesn’t take much to fix a bad meta description or to turn a good title tag into a great one. The goal of all of your titles and meta descriptions is to inform visitors (and bots) what your page is about and, whenever possible, to include a choice keyphrase or two.
  Here are some best practices:
  Fix empty or lengthy titles and meta descriptions

   Most SEO auditing tools will tell you when your titles and meta descriptions aren’t optimized. Make sure all of your titles fall within the appropriate character count ( That means 50-60 characters for the title; 150-160 characters for the description ).
  Avoid duplicate titles and meta descriptions

  E-commerce sites, in particular, suffer from duplicate content, or meta descriptions that read very similarly. Duplicate content can confuse search engines to the point where they rank none of your pages, so try to keep your content as unique as possible.
  Target buying phrases

  Header (H1) tags are important on-page ranking factors. Use these to target specific buying phrases that users are likely to search for. Try to Also try to avoid generic subheaders (H2, H3, etc.); use them to target secondary keywords where appropriate.
  Checklist:

  
       
  • Optimize titles for SEO   
  • Optimize meta descriptions for SEO   
  • Optimize header tags for SEO  
  7. Optimize Your Content

   Since Panda’s launch, high-quality content has been the defining factor of a great website. In other words, you need to make sure that each of your pages has enough content and that the content is unique and valuable. You may want to invest in a grammar tool (e.g. Grammarly ) and a plagiarism checker (e.g. Copyscape) to help you with this stage (they’re free!).
  Here’s what I did:
  Identify thin content

   To make sure that every page had enough relevant data, I filtered my pages by word count. I did this by right-clicking on the header of any column and adding a new column called Word Count . Then I made sure that every page had 1-2 paragraphs of text (aiming for at least 250 words per page). I also double-checked their formatting and readability while I was at it.

Back to Search School: A Quick 15-Minute Audit by @ab80-2 (something,software,research,entering,separate)

  Corrected duplicate content

  It’s in your best interests to ensure that all your titles and descriptions are unique. The thing is duplicate titles and descriptions are confusing to search engines as they wouldn’t know which of the pages should be ranked in response to a given query. Often, they’d choose to rank neither page instead.
12下一页
友荐云推荐




上一篇:The Next Generation of LinkedIn Career Pages
下一篇:Marketing Day: Google updates AdSense, CTR hacks & conversion optimization
酷辣虫提示酷辣虫禁止发表任何与中华人民共和国法律有抵触的内容!所有内容由用户发布,并不代表酷辣虫的观点,酷辣虫无法对用户发布内容真实性提供任何的保证,请自行验证并承担风险与后果。如您有版权、违规等问题,请通过"联系我们"或"违规举报"告知我们处理。

bvg3jl 发表于 2016-10-6 20:36:38
说不好是在找骂么?
回复 支持 反对

使用道具 举报

小虫虫 发表于 2016-10-6 20:37:51
楼主已成仙,有事请求签!
回复 支持 反对

使用道具 举报

別忘ろ幸福√ 发表于 2016-10-7 14:35:20
人是铁,范是钢,一天不装憋得慌
回复 支持 反对

使用道具 举报

hdwjt 发表于 2016-11-16 12:13:53
您提前出院了?
回复 支持 反对

使用道具 举报

fjlQ5cf4 发表于 2016-11-16 13:50:11
好东西,赞一个!
回复 支持 反对

使用道具 举报

*滑动验证:
您需要登录后才可以回帖 登录 | 立即注册

本版积分规则

我要投稿

推荐阅读

扫码访问 @iTTTTT瑞翔 的微博
回页顶回复上一篇下一篇回列表手机版
手机版/CoLaBug.com ( 粤ICP备05003221号 | 文网文[2010]257号 )|网站地图 酷辣虫

© 2001-2016 Comsenz Inc. Design: Dean. DiscuzFans.

返回顶部 返回列表