In this comprehensive SEO tutorial for beginners, I will cover the essential fundamentals of Search Engine Optimization. This tutorial is divided. Each stop is a unique document (usually a web page, but sometimes a PDF, JPG , or other .. Learning the foundations of SEO is a vital step in achieving these. a free PDF checklist of all the SEO tools I use. You may . off the start to learn what keywords convert well for you and which keywords do not.
|Language:||English, Spanish, Indonesian|
|ePub File Size:||16.65 MB|
|PDF File Size:||19.47 MB|
|Distribution:||Free* [*Regsitration Required]|
SEO Tutorial For Beginners in .. the main content of the quality raters guidelines (official PDF), but we have been told by Google it is not. Review: How To Do An SEO Audit With SEMRush FAST (Beginners Guide). guidelines (official PDF), but we have been told by Google it is not, per I've included the robots meta tag in my tutorial as this IS one of only a. SEO Guide. Comprehensive SEO Guide & Tutorial for Beginners. This SEO guide is designed for those who are new to the world of search engine optimization.
Next time you are developing a page, consider what looks spammy to you is probably spammy to Google. Another obvious way to gauge the health of a site is to see which pages on the site get zero traffic from Google over a certain period of time. Think about the topic of the page. Google does not want you to try and modify where you rank, easily. It creates an index for all the fetched web pages and keeps them into a giant database from where retrieval is also possible. Ranking of your page is measured by the position of your web pages displayed in SERPs search engine result pages.
Naturally, business owners want to rank for lots of keywords in organic listings with their website. These pages are created with no or very little time, effort, or expertise, and also have no editing or manual curation. The end-result is webmasters create doorway pages without even properly understanding what they represent to Google and without realising Google will not index all these autogenerated pages. It is a form of SEO spamming. They are bad for users because they can lead to multiple similar pages in user search results, where each result ends up taking the user to essentially the same destination.
They can also lead users to intermediate pages that are not as useful as the final destination. If you know you have VERY low-quality doorway pages on your site, you should remove them or rethink your SEO strategy if you want to rank high in Google for the long term. Google aims to rank pages where the author has some demonstrable expertise on experience in the subject-matter they are writing about.
This is an important quality characteristic. Think about the topic of the page. What kind of expertise is required for the page to achieve its purpose well? The standard for expertise depends on the topic of the page. SC is created by Webmasters and is an important part of the user experience. One common type of SC is navigation links which allow users to visit other parts of the website. Note that in some cases, content behind tabs may be considered part of the SC of the page.
When it comes to a web page and positive UX, Google talks a lot about the functionality and utility of Helpful Supplementary Content — e.
We have different standards for small websites which exist to serve their communities versus large websites with a large volume of webpages and content. It is worth remembering that Good supplementary content cannot save Poor main content from a low-quality page rating:. Google Search Quality Evaluator Guidelines Pop-ups suck, everybody seems to agree.
Using a pop up does seem to have an immediate impact. I have since tested it on and off for a few months and the results from the small test above have been repeated over and over. In my tests, using pop-ups really seemed to kill how many people share a post in social media circles.
With Google now showing an interest with interstitials especially on mobile versions of your site , I would be very nervous about employing a pop-up window that obscures the primary reason for visiting the page. Google will tell you if the ads on your website are annoying users which may impact the organic traffic Google sends you. Annoying ads on your web pages has long been a problem for users probably and Google, too. Even if they do make you money. Google, If your site presents violations, the Ad Experience Report may identify the issues to fix.
It offers the same amount of screen real estate as pop-ups without covering up any content. Google has long warned about web page advertisements and distractions on a web page that results in a poor user experience. Misleading titles can result in a very poor user experience when users click a link only to find that the page does not match their expectations. Pages that disrupt the use of the MC should be given a Low rating.
Some webpages are designed to encourage users to click on SC that is not helpful for the purpose of the page.
This type of SC is often distracting or prominently placed in order to lure users to highly monetized pages. Please refresh the page a few times to see the range of Ads that appear, and use your knowledge of the locale and cultural sensitivities to make your rating.
For example, an ad for a model in a revealing bikini is probably acceptable on a site that sells bathing suits. It should also be clear what will happen when users interact with content and links on the webpage. If users are misled into clicking on Ads or SC, or if clicks on Ads or SC leave users feeling surprised, tricked or confused, a Low rating is justified.
Use your judgment when evaluating pages. User expectations will differ based on the purpose of the page and cultural norms. Put there by algorithm or human. Manual evaluators might not directly impact your rankings, but any signal associated with Google marking your site as low-quality should probably be avoided.
After page content, the following are given the most weight in determining if you have a high-quality page. When it comes to Google assigning your page the lowest rating, you are probably going to have to go some to hit this, but it gives you a direction you want to ensure you avoid at all costs. Note — These statements below are spread throughout the raters document and not listed the way I have listed them here.
If the website feels inadequately updated and inadequately maintained for its purpose, the Low rating is probably warranted. It measures the quality of a site, which you can read more about in our guidelines.
Panda allows Google to take quality into account and adjust ranking accordingly. Google has algorithms that target low-quality content and these algorithms are actually trained in some part by human quality raters.
Google is always raising the bar — always adding new signals, sometimes, in time, taking signals away. That positioning has always been a win-win for Google — and a recognisable strategy from them after all these years. Take unnatural links out of the equation which have a history of trumping most other signals and you are left with page level, site level and off-site signals. Do sites at the top of Google get asked more of?
Whether its algorithmic or manual — based on technical, architectural, reputation or content — Google can decide and will decide if your site meets its quality requirements to rank on page one.
The entire budget of my time went on content improvement, content reorganisation, website architecture improvement, and lately, mobile experience improvement. In simple terms, I took thin content and made it fat to make old content perform better. Generally speaking — real quality will stand out, in any niche with a lack of it, at the moment. For all types of webpages, creating high quality MC takes a significant amount of at least one of the following: Google wants to rate you on the effort you put into your website, and how satisfying a visit is to your pages.
If your page has a sloppy design, low-quality main content and too many distracting ads your rankings are very probably going to take a nose-dive. If a Search Quality Evaluator is subject to a sneaky redirect they are instructed to rate your site low. However, Medium pages lack the characteristics that would support a higher quality rating.
Occasionally, you will find a page with a mix of high and low quality characteristics.
In those cases, the best page quality rating may be Medium. Usually, Google has 3 or 4 big updates in a year that focus on various things, but they also make changes daily. See this list for a comprehensive list of Google algorithm updates. Google Panda aims to rate the quality of your pages and website and is based on things about your site that Google can rate, or algorithmically identify.
But essentially allows us to take quality of the whole site into account when ranking pages from that particular site and adjust the ranking accordingly for the pages. Basically, we figured that site is trying to game our systems, and unfortunately, successfully. Panda evolves — signals can come and go — Google can get better at determining quality as a spokesman from Google has confirmed:.
I also list these Panda points below:. If someone is putting the hours into rank their site through legitimate efforts — Google will want to reward that — because it keeps the barrier to entry HIGH for most other competitors. Critics will say the higher the barrier to entry is to rank high in Google natural listings the more attractive Google Adwords begins to look to those other businesses.
While getting as many pages indexed in Google was historically a priority for an SEO, Google is now rating the quality of pages on your site and the type of pages it is indexing. Knowing is useful, of course, but largely unnecessary. Indexation is never a guarantee of traffic.
Some SEO would tend to scrape Google to get indexation data on a website. Google will tell you how many pages you have submitted in a sitemap, and how many pages are indexed.
If Google is de-indexing large swaths of your content that you have actually submitted as part of an XML sitemap, then a problem is often afoot.
Read my article on how to get your entire website crawled and indexed by Google. A content type analysis will identify the type of pages the cms generates. A content performance analysis will gauge how well each section of the site performs. The thinking is if the pages were high-quality, they would be getting some kind of organic traffic. Identifying which pages receive no organic visitors over a sensible timeframe is a quick if noisy, way to separate pages that obviously WORK from pages that DONT — and will help you clean up a large portion of redundant URLs on the site.
It is important to trim content pages carefully — and there are nuances. Experience can educate you when a page is high-quality and yet receives no traffic. A strategy for these pages can then be developed. This is time-consuming — just like Google wants it to be. You need to review DEAD pages with a forensic eye and ask:. If the answer to any of the above is NO — then it is imperative you take action to minimise the amount of these types of pages on your site.
Do NOT just redirect these pages to your homepage. High-quality content is expensive — so rework content when it is available. Medium quality content can always be made higher quality — in fact — a page is hardly ever finished in EXPECT to come back to your articles every six months to improve them to keep them moving in the right direction.
Well, it does if the page you make is useful and has a purpose other than just to make money. I call these POOR pages in my reviews. If you have very low-quality site form a content point of view, just deleting the content or noindexing it is probably not going to have a massive positive impact on your rankings. What Panda does is disregard the advantage you figure out, so you fall back where you started.
Ultimately people want good sites.
And remember the following, specific advice from Google on removing low-quality content from a domain:. NOT your high-quality content. This is more or less explained by Google spokespeople like John Mueller.
Clearing away the low-quality stuff lets you focus on building better stuff on other pages that Google will rank in and beyond. A myth is that pages need a lot of text to rank. There are many reasons a website loses traffic from Google. Server changes, website problems, content changes, downtimes, redesigns, migrations… the list is extensive. Comparing your Google Analytics data side by side with the dates of official algorithm updates is useful in diagnosing a site health issue or traffic drop.
This client did eventually receive a penalty for unnatural links when they ignored our advice to clean up. A quick check of how the site was laid out soon uncovered a lot of unnecessary pages, or what Google calls thin, overlapping content. This observation would go a long way to confirming that the traffic drop was indeed caused by the May algorithm change.
Another obvious way to gauge the health of a site is to see which pages on the site get zero traffic from Google over a certain period of time.
I go into some of that in my duplicate content penalty post. If your pages were designed to get the most out of Google, with commonly known and now outdated SEO techniques chances are Google has identified this and is throttling your rankings in some way. Google will continue to throttle rankings until you clean your pages up. Actually — looking at the backlink profile of this customer, they are going to need a disavow file prepared too.
Google went through the SEO playbook and identified old techniques and use them against you today — meaning every SEO job you take on always has a clean up aspect now. Google has a LONG list of technical requirements it advises you meet , on top of all the things it tells you NOT to do to optimise your website.
Mostly — individual technical issues will not be the reason you have ranking problems, but they still need to be addressed for any second-order benefit they provide. Every site has things to clean up and to optimise in a modern way. Sites with higher rankings often pick up more organic links, and this process can float high-quality pages on your site quickly to the top of Google.
Tick all the boxes Google tell you to tick, so to speak. I use these quality rating documents and the Google Webmaster Guidelines as the foundation of my audits for e-commerce sites. The evaluators base their ratings on guidelines we give them; the guidelines reflect what Google thinks search users want. This page and site appear to check all the boxes Google wants to see in a high-quality e-commerce website these days. Essentially, if you are selling something to visitors or advising on important matters like finance, law or medical advice — your page will be held to this higher standard.
It is interesting to note that 1. The website looks different today. You might not be able to mimic the positive reputation this US site has, but you are going to have to build your product pages to compete with it, and others like it.
Domain authority, whether or not is something Google has or not, is an important concept to take note of.
Domain authority is an important ranking phenomenon in Google. Nobody knows exactly how Google calculates, ranks and rates the popularity, reputation, intent or trust of a website, outside of Google, but when I write about domain authority I am generally thinking of sites that are popular, reputable and trusted — all of which can be faked, of course.
Historically sites that had domain authority or online business authority had lots of links to them, hence why link building was so popular a tactic — and counting these links is generally how most 3rd party tools still calculate it a pseudo domain authority score for websites today. Official Google Webmaster Blog. SEO more usually talk about domain trust and domain authority based on the number, type and quality of incoming links to a site.
Examples of trusted, authority domains include Wikipedia, the W3C and Apple. How did you take advantage of being an online business authority?
You turned the site into an SEO Black Hole to horde the benefits of domain authority and published lots of content sometimes with little thought to quality. On any subject. Because Google would rank it! Google is going to present users with sites that are recognisable to them. Easier said than done, for most, of course, but that is the point of link building — to get these type of links. Well, yes. Its harder for most businesses because low-quality content on parts of a domain can negatively impact the rankings of an entire domain.
Instead of publishing LOTS of pages, focus on fewer pages that are of high quality. You can better predict your success in ranking for the long term for a particular keyword phrase this way.
Failure to meet these standards for quality content may impact rankings noticeably around major Google quality updates.
Having a ten-year-old domain that Google knows nothing about is almost the same as having a brand new domain. A one-year-old domain cited by authority sites is just as valuable if not more valuable than a ten-year-old domain with no links and no search-performance history. In , you need to be aware that what works to improve your rank can also get you penalised faster, and a lot more noticeably. There are some things you cannot directly influence legitimately to improve your rankings, but there is plenty you CAN do to drive more Google traffic to a web page.
You will not ever find every ranking factor. Some ranking factors are based on where you are, or what you have searched for before. In that time, a lot has changed. Read my article on a more complete list of potential Google ranking factors. If you are a geek and would like to learn more see my post on potential Google ranking factors.
You can profit from it if you know a little about how Google works or seems to work, in many observations, over years, excluding when Google throws you a bone on synonyms. It gets interesting if you do that to a lot of pages, and a lot of keyword phrases. The important thing is keyword research — and knowing which unique keywords to add. Yes — plenty of other things can be happening at the same time. Google Analytics was the very best place to look at keyword opportunity for some especially older sites, but that all changed a few years back.
This means site owners will begin to lose valuable data that they depend on, to understand how their sites are found through Google. The keyword data can be useful, though — and access to backlink data is essential these days. Optimise this with searcher intent in mind. I have seen pages with 50 words outrank pages with , , or words. In , Google is a lot better at hiding away those pages, though. Creating deep, information rich pages focuses the mind when it comes to producing authoritative, useful content.
Every site is different. Some pages, for example, can get away with 50 words because of a good link profile and the domain it is hosted on. After a while, you get an idea how much text you need to use to get a page on a certain domain into Google.
One thing to note — the more text you add to the page, as long as it is unique, keyword rich and relevant, the more that page will be rewarded with more visitors from Google. There is no optimal number of words on a page for placement in Google. Every website — every page — is different from what I can see.
Google will probably reward you on some level — at some point — if there is lots of unique text on all your pages. There is no one-size-fits-all keyword density, no optimal percentage guaranteed to rank any page at number 1.
I aim to include related terms , long-tail variants and synonyms in Primary Content — at least ONCE, as that is all some pages need. Search engines have kind of moved on from there. It often gets a page booted out of Google but it depends on the intent and the trust and authority of a site. Such pages are created using words likely to be contained in queries issued by users. Keyword stuffing can range from mildly annoying to users, to complete gibberish. Pages created with the intent of luring search engines and users, rather than providing meaningful MC to help users, should be rated Lowest.
Just because someone else is successfully doing it do not automatically think you will get away with it. Aaron Wall. Aaron Wall, It is time to focus on the user when it comes to content marketing, and the bottom line is you need to publish unique content free from any low-quality signals if expect some sort of traction in Google SERPs Search Engine Result Pages.
SEO copywriting is a bit of a dirty word — but the text on a page still requires optimised, using familiar methods, albeit published in a markedly different way than we, as SEO, used to get away with.
When it comes to writing SEO-friendly text for Google, we must optimise for u ser intent, not simply what a user typed into Google. Google has plenty of options when rewriting the query in a contextual way, based on what you searched for previously, who you are, how you searched and where you are at the time of the search.
Yes, you must write naturally and succinctly in , but if you have no idea the keywords you are targeting, and no expertise in the topic, you will be left behind those that can access this experience. Naturally, how much text you need to write, how much you need to work into it, and where you ultimately rank, is going to depend on the domain reputation of the site you are publishing the article on.
SEOs have understood user search intent to fall broadly into the following categories and there is an excellent post on Moz about this. When it comes to rating user satisfaction , there are a few theories doing the rounds at the moment that I think are sensible. Google could be tracking user satisfaction by proxy. A user clicks a result and bounces back to the SERP, pogo-sticking between other results until a long click is observed.
Google has this information if it wants to use it as a proxy for query satisfaction. For more on this, I recommend this article on the time to long click. Once you have the content, you need to think about supplementary content and secondary links that help users on their journey of discovery. A website that does not link out to ANY other website could be interpreted accurately to be at least, self-serving. For more on primary main content optimisation see:: For me, a perfect title tag in Google is dependant on a number of factors and I will lay down a couple below but I have since expanded page title advice on another page link below ;.
I think title tags, like everything else, should probably be as simple as possible, with the keyword once and perhaps a related term if possible. If you are relying on meta-keyword optimisation to rank for terms, your dead in the water.
What about other search engines that use them? Hang on while I submit my site to those 75, engines first [sarcasm! Yes, ten years ago early search engines liked looking at your meta-keywords. Forget about meta-keyword tags — they are a pointless waste of time and bandwidth.
So you have a new site. Sometimes competitors might use the information in your keywords to determine what you are trying to rank for, too…. Forget whether or not to put your keyword in it, make it relevant to a searcher and write it for humans, not search engines. If you want to have this word meta description which accurately describes the page you have optimised for one or two keyword phrases when people use Google to search, make sure the keyword is in there.
Google looks at the description but it probably does not use the description tag to rank pages in a very noticeable way. Sometimes, I will ask a question with my titles, and answer it in the description, sometimes I will just give a hint. That is a lot more difficult in as search snippets change depending on what Google wants to emphasise to its users.
Sometimes I think if your titles are spammy, your keywords are spammy, and your meta description is spammy, Google might stop right there — even they probably will want to save bandwidth at some time. So, the meta description tag is important in Google, Yahoo and Bing and every other engine listing — very important to get it right.
Googles says you can programmatically auto-generate unique meta descriptions based on the content of the page. No real additional work is required to generate something of this quality: I think it is very important to listen when Google tells you to do something in a very specific way, and Google does give clear advice in this area.
By default, Googlebot will index a page and follow links to it. At a page level — it is a powerful way to control if your pages are returned in search results pages.
I have never experienced any problems using CSS to control the appearance of the heading tags making them larger or smaller. How many words in the H1 Tag? As many as I think is sensible — as short and snappy as possible usually.
As always be sure to make your heading tags highly relevant to the content on that page and not too spammy, either. Use ALT tags or rather, ALT Attributes for descriptive text that helps visitors — and keep them unique where possible, like you do with your titles and meta descriptions. The title attribute should contain information about what will happen when you click on the image.
From my tests, no. From observing how my test page ranks — Google is ignoring keywords in the acronym tag. You do not need clean URLs in site architecture for Google to spider a site successfully confirmed by Google in , although I do use clean URLs as a default these days, and have done so for years. However — there it is demonstrable benefit to having keywords in URLs.
The thinking is that you might get a boost in Google SERPs if your URLs are clean — because you are using keywords in the actual page name instead of a parameter or session ID number which Google often struggles with. I optimise as if they do, and when asked about keywords in urls Google did reply:. I believe that is a very small ranking factor. Then it is fair to say you do get a boost because keywords are in the actual anchor text link to your site, and I believe this is the case, but again, that depends on the quality of the page linking to your site.
That is, if Google trusts it and it passes Pagerank! Sometimes I will remove the stop-words from a URL and leave the important keywords as the page title because a lot of forums garble a URL to shorten it. Most forums will be nofollowed in , to be fair, but some old habits die-hard. It should be remembered it is thought although Googlebot can crawl sites with dynamic URLs; it is assumed by many webmasters there is a greater risk that it will give up if the URLs are deemed not important and contain multiple variables and session IDs theory.
As standard , I use clean URLs where possible on new sites these days, and try to keep the URLs as simple as possible and do not obsess about it. Having a keyword in your URL might be the difference between your site ranking and not — potentially useful to take advantage of long tail search queries.
I prefer absolute URLs. Google will crawl either if the local setup is correctly developed. This is entirely going to a choice for your developers. Through this complete SEO tutorial, you will also understand how SEO copywriting helps you have the good content on your site.
On the concluding note, below given are some other important points that I find important to include in this SEO tutorial-. Want to be an expert in Search Engine Optimization?
Having doubts in understanding any of the aforestated SEO basics steps? He loves imparting all his experiences and expertize through his blogs. I liked this post. And you are helping the beginner to learn SEO through this blog. Thanks a lot for the very beautiful blog. SEO is the techniques to get legal rank in search engine. Thanks to Admin to share SEO lessons. This was a super amazing tutorial, in short, I would like you to share more in brief about on page SEO.
Thank you! Hang On! Interested in Career or Business Growth? Attend Our Upcoming Free Webinar. Digital Marketing Master Course. What is SEO? Let us go through this complete SEO tutorial for beginners: Meta Tags. Free Digital Marketing Webinar Date: Link Building.
Mobile SEO. Manish Singh. Related posts: There are 19 comments 1 year ago. Video is attractive too. Is there any short tutorial of video of something deeply explaining offpage seo? This is a very helpful tutorial. Thank You. Thanks for sharing. Your Comment Cancel Reply Your email address will not be published. Arrange a session with career counsellor.
Send me course curriculum as well. Take a Demo Class. Free Demo. Call Us. Limited Seats Available! AM PM. Call Me. Data Analytics for Career Growth.
Which Program are you interested in? Book your Spot. Take a Demo Class Our experts will call you soon and schedule one-to-one demo session with you Course: