If you do great work with your search engine optimization (SEO), it could mean a significant amount of revenue for your business. At the same time, however, it is also an ongoing initiative. Once you generate a steady stream of traffic from SEO, you need to constantly be maintaining and improving your SEO in order to keep those rankings you worked so hard for.
In 1998, two graduate students at Stanford University, Larry Page and Sergey Brin, developed "Backrub", a search engine that relied on a mathematical algorithm to rate the prominence of web pages. The number calculated by the algorithm, PageRank, is a function of the quantity and strength of inbound links.[22] PageRank estimates the likelihood that a given page will be reached by a web user who randomly surfs the web, and follows links from one page to another. In effect, this means that some links are stronger than others, as a higher PageRank page is more likely to be reached by the random web surfer.
Some search engines have also reached out to the SEO industry, and are frequent sponsors and guests at SEO conferences, webchats, and seminars. Major search engines provide information and guidelines to help with website optimization.[18][19] Google has a Sitemaps program to help webmasters learn if Google is having any problems indexing their website and also provides data on Google traffic to the website.[20] Bing Webmaster Tools provides a way for webmasters to submit a sitemap and web feeds, allows users to determine the "crawl rate", and track the web pages index status.
A quick search for “SEO ranking factors” will give you all of these answers and myriad others. There is a lot of information out there. And the reality is, while there are likely hundreds of variables working together to determine final placement, much of what is suggested is guesswork. And certainly, not all ranking factors are relevant to every business.

To avoid undesirable content in the search indexes, webmasters can instruct spiders not to crawl certain files or directories through the standard robots.txt file in the root directory of the domain. Additionally, a page can be explicitly excluded from a search engine's database by using a meta tag specific to robots (usually ). When a search engine visits a site, the robots.txt located in the root directory is the first file crawled. The robots.txt file is then parsed and will instruct the robot as to which pages are not to be crawled. As a search engine crawler may keep a cached copy of this file, it may on occasion crawl pages a webmaster does not wish crawled. Pages typically prevented from being crawled include login specific pages such as shopping carts and user-specific content such as search results from internal searches. In March 2007, Google warned webmasters that they should prevent indexing of internal search results because those pages are considered search spam.[47]
How do you ask others for link opportunities? Most of the time people are only interested in either reciprocal links, or them providing guest posts on my site (when I reach out). And I can’t imagine if I did a round up post getting many inbound links. People would be thrilled that they had received a link, and wouldn’t create a reciprocal link to destroy the value.

Marcus Miller is an experienced SEO and PPC consultant based in Birmingham, UK. Marcus focuses on strategy, audits, local SEO, technical SEO, PPC and just generally helping businesses dominate search and social. Marcus is managing director of the UK SEO and digital marketing company Bowler Hat and also runs wArmour aka WordPress Armour which focuses on helping WordPress owners get their security, SEO and site maintenance dialled in without breaking the bank.


When Googlebot crawls a page, it should see the page the same way an average user does15. For optimal rendering and indexing, always allow Googlebot access to the JavaScript, CSS, and image files used by your website. If your site's robots.txt file disallows crawling of these assets, it directly harms how well our algorithms render and index your content. This can result in suboptimal rankings.
In this excellent post, SEO and Digital Trends in 2017, Gianluca Fiorelli writes, "In a mobile-only world, the relevance of local search is even higher. This seems to be the strategic reason both for an update like Possum and all the tests we see in local, and also of the acquisition of a company like Urban Engines, whose purpose is to analyze the "Internet of Moving Things."
Thank you Brian. I am so brand spanking new to all this and i am really struggling with understanding it all. I have tried to read so many thing to help my website and this was the first article to really make sense however Being an urban, street menswear online store i feel like my niche is too broad?.. Ahh Feel like I am drowning maybe I need to do your course! Thanks again for the read I will be doing a lot more thats for sure
Each organic search engine ranking places emphasis on variable factors such as the design and layout, keyword density and the number of relevant sites linking to it. Search engines constantly update and refine their ranking algorithms in order to index the most relevant sites. Other variables that have an impact on search engine placement include the following: 

Brian, I’ve drunk your Kool aid! Thank you for honesty and transparency – it really gives me hope. Quick question: I am beyond passionate about a niche (UFOs, extraterrestrials, free energy) and know in my bones that an authority site is a long term opportunity. The problem today is that not many products are attached to this niche and so it becomes a subscriber / info product play. However, after 25+ years as an entrepreneur with a financial background and marketing MBA, am I Internet naive to believe that my passion and creativity will win profitability in the end? The target audience is highly passionate too. Feedback?
Hey Brian, I have landed in this blog while visiting via blog land. I must appreciate your effort to put up such informative content. As being an Internet Marketing Consultant, I would like to add few thought of my own with your valuable content. There are many people who wants HUGE number of traffic with no time at all. But as per my experience, SEO has become SLOW-BUT-STEADY process in the recent times. After so many algorithm updates of Google, I think if we will do any wrong things with the websites, that should be paid off. So without taking any risk, we need to work ethically so that slowly the website will get the authority and grab the targeting traffic. What do you think mate? I am eagerly looking forward to your reply and love to see more valuable write-ups from your side. Why don’t you write about some important points about Hummingbird Updates of Google. It will be a good read. Right brother? 🙂

Like many SEOs, I was hired with one vague responsibility: to set up an SEO program and achieve results. Like many SEOs, we jumped right in and started spewing out SEO audits, rewriting title tags, offering up link suggestions, rewriting URLs and so on. And like many SEOs we promised results. But what we didn’t do, until that fateful launch, was develop a comprehensive strategy.
Like many SEOs, I was hired with one vague responsibility: to set up an SEO program and achieve results. Like many SEOs, we jumped right in and started spewing out SEO audits, rewriting title tags, offering up link suggestions, rewriting URLs and so on. And like many SEOs we promised results. But what we didn’t do, until that fateful launch, was develop a comprehensive strategy.

Content is king. That’s the saying, right? It’s true in a way. Your website is really just a wrapper for your content. Your content tells prospects what you do, where you do it, who you have done it for, and why someone should use your business. And if you’re smart, your content should also go beyond these obvious brochure-type elements and help your prospective customers achieve their goals.
Make it as easy as possible for users to go from general content to the more specific content they want on your site. Add navigation pages when it makes sense and effectively work these into your internal link structure. Make sure all of the pages on your site are reachable through links, and that they don't require an internal "search" functionality to be found. Link to related pages, where appropriate, to allow users to discover similar content.
See the screenshot below for some of the sections for specific recommendations that you can add which will provide the meat of the document. Keep in mind this is a very flexible document – add recommendations that make sense (for example you may not always have specific design considerations for a project). Remember, it will be different every time you do it.
×