Let’s say, for example, that you run a construction business that helps with home repairs after natural disasters and you want to advertise that service. The official term for the service is “fire restoration,” but keyword research may indicate that customers in your area search instead for “fire repair” or “repair fire damage to house.” By not optimizing for these two keywords, you’ll lose out on a lot of traffic and potential customers, even if “fire restoration” is technically more correct.
I would like to talk about a case study for a large start up I worked on for over eight months in the Australian and US market. This client originally came to the company with the typical link building and SEO problems. They had been using a SEO company that had an extensive link network and was using less than impressive SEO tactics and methodologies over the last 12 months. The company was also losing considerable revenue as a direct result of this low quality SEO work. So, I had to scramble and develop a revival strategy for this client.
Network marketing: Network marketing companies have a great business model (for those who own the company), because they only pay their sales people (a.k.a. “independent business owners”) when they make a sale or recruit another person. They only pay on performance. So to sell a bunch of product, the direct sales company really doesn’t go directly to the consumer through TV or magazine ads or similar methods that could easily cost millions; instead they go indirectly through their sales people and only pay for the word of mouth advertising as a commission on a product sale. It’s really savvy business strategy that’s low-risk and high-reward, if it spreads far and fast enough by emotionally exciting the distributors. Distributors are heavily using social media like Facebook, YouTube, blogging and the like to generate sales and grow their network online.
To avoid undesirable content in the search indexes, webmasters can instruct spiders not to crawl certain files or directories through the standard robots.txt file in the root directory of the domain. Additionally, a page can be explicitly excluded from a search engine's database by using a meta tag specific to robots. When a search engine visits a site, the robots.txt located in the root directory is the first file crawled. The robots.txt file is then parsed and will instruct the robot as to which pages are not to be crawled. As a search engine crawler may keep a cached copy of this file, it may on occasion crawl pages a webmaster does not wish crawled. Pages typically prevented from being crawled include login specific pages such as shopping carts and user-specific content such as search results from internal searches. In March 2007, Google warned webmasters that they should prevent indexing of internal search results because those pages are considered search spam.
Get a handle on your brand reputation. Your brand story is the one that you tell. Your reputation is the story that customers tell on your behalf. If someone consistently stumbles on your site when they type in niche search queries, they’ll be intrigued. The result? They’ll start conducting navigational searches for your brand. The intent behind that search? They want reviews and other customer’s experiences with your business. Ask your customers for reviews and reach out to third-party review sites in your niche. This way, these navigational searches don’t come up empty. I also recommend monitoring your brand mentions. The easy way is to set up Google Alerts. Type in your brand name and create your alert. Any mention online and you’ll be notified.
As the number of sites on the Web increased in the mid-to-late 1990s, search engines started appearing to help people find information quickly. Search engines developed business models to finance their services, such as pay per click programs offered by Open Text in 1996 and then Goto.com in 1998. Goto.com later changed its name to Overture in 2001, was purchased by Yahoo! in 2003, and now offers paid search opportunities for advertisers through Yahoo! Search Marketing. Google also began to offer advertisements on search results pages in 2000 through the Google AdWords program. By 2007, pay-per-click programs proved to be primary moneymakers for search engines. In a market dominated by Google, in 2009 Yahoo! and Microsoft announced the intention to forge an alliance. The Yahoo! & Microsoft Search Alliance eventually received approval from regulators in the US and Europe in February 2010.
That means it takes time, but the results are long-lasting. Once you establish search engine visibility, it will serve your business for years after it’s implemented. That’s not to say that there isn’t any maintenance involved in an organic traffic strategy. The SEO landscape is ever changing. You can’t just “set and forget” your traffic system. The algorithms and ranking factors evolve. You have to keep on top of these changes to maximize your results. With all that said, an organic traffic strategy is as close as you’ll get to a traffic system on autopilot. And just another reason why organic traffic is important.
I'm thinking about "less to do" (linking-social media) sites with physical products … Amazon and other merchants. And adding a PPC campaign instead of using social media and constantly toiling away building links and writing articles. Is there an easy way to get sites like this to rank while profiting with some paid traffic? Your opinion on this dilemma .. SEO vs PAID traffic or both would be much appreciated.
In February 2011, Google announced the Panda update, which penalizes websites containing content duplicated from other websites and sources. Historically websites have copied content from one another and benefited in search engine rankings by engaging in this practice. However Google implemented a new system which punishes sites whose content is not unique. The 2012 Google Penguin attempted to penalize websites that used manipulative techniques to improve their rankings on the search engine. Although Google Penguin has been presented as an algorithm aimed at fighting web spam, it really focuses on spammy links by gauging the quality of the sites the links are coming from. The 2013 Google Hummingbird update featured an algorithm change designed to improve Google's natural language processing and semantic understanding of web pages. Hummingbird's language processing system falls under the newly recognised term of 'Conversational Search' where the system pays more attention to each word in the query in order to better match the pages to the meaning of the query rather than a few words . With regards to the changes made to search engine optimization, for content publishers and writers, Hummingbird is intended to resolve issues by getting rid of irrelevant content and spam, allowing Google to produce high-quality content and rely on them to be 'trusted' authors.
SEO is not an appropriate strategy for every website, and other Internet marketing strategies can be more effective like paid advertising through pay per click (PPC) campaigns, depending on the site operator's goals. Search engine marketing (SEM), is practice of designing, running, and optimizing search engine ad campaigns. Its difference from SEO is most simply depicted as the difference between paid and unpaid priority ranking in search results. Its purpose regards prominence more so than relevance; website developers should regard SEM with the utmost importance with consideration to PageRank visibility as most navigate to the primary listings of their search. A successful Internet marketing campaign may also depend upon building high quality web pages to engage and persuade, setting up analytics programs to enable site owners to measure results, and improving a site's conversion rate. In November 2015, Google released a full 160 page version of its Search Quality Rating Guidelines to the public, which now shows a shift in their focus towards "usefulness" and mobile search. In recent years the mobile market has exploded, overtaking the use of desktops as shown in by StatCounter in October 2016 where they analysed 2.5 million websites and 51.3% of the pages were loaded by a mobile device . Google has been one of the companies that have utilised the popularity of mobile usage by encouraging websites to use their Google Search Console, the Mobile-Friendly Test, which allows companies to measure up their website to the search engine results and how user-friendly it is.
People find their way to your website in many different ways. If someone is already familiar with your business and knows where to find your website, they might just navigate straight to your website by typing in your domain. If someone sees a link to a blog you wrote in their Facebook newsfeed, they might click the link and come to your website that way.
Although it may have changed slightly since BrightEdge published its report last year, the data still seem to hold true. Organic is simply better for delivering relevant traffic. The only channel that performs better in some capacities is paid search ads, but that is only for conversions, not overall traffic delivery (Paid Search only accounted for 10 percent of overall total traffic).