Many websites especially e-commerce sites have parts of their site secure using SSL. This protocol allows information to be transferred over an encrypted connection, and is intended to stop packet sniffing devices grabbing your personal details. URLs of secured pages on a site start with the https protocol to indicate to the user that this part of the site is secure.
Some sites will only require the use of https on their contact us page, or in their checkout process to protect the customer’s personal details (bank details etc). On these pages only the protocol needs to be changed from http to https to secure the pages.
This technique however can result in search engines indexing both the secure and non secure version of your pages. If for example you used relative links in your secured pages, this could be interpreted as links to secure versions of your standard pages. This can lead to Google assuming you have duplicate pages, which can result in a drop in your listings due to duplicate content.
So how do you stop Google visiting these duplicate pages?
Go through your pages and replace all your URLs with exact links rather than relative links. You will however need to ensure you link to the appropriate pages with the correct protocol, whether the page is http or https.
The real world altogether seems to have concentrated in this virtual world. Its hard to find what not has made its access into this virtual world. Information related to anything and everything is just a click away. Searching of information has been made easy by the search engines. It is the principle factor that helps you to find heaps of information about whatever thing you want. Searching without search engines would have been an arduous job. Every now and then search engines are being optimized. The Search Engine Optimization firms assure the clients, definite top results for a well-worth site. By creating a site that can be easily reached and useful for people and by using genuine, semantic markup, one can compose it with a greater harmony that will be attractive to search engine systems.
Now to promote your site over the search engines, you should be comprehensive with some basics of search engine optimization. Firstly you should keep in mind that no shortcuts can ever fetch you the good result. There should be a good content. This is the primary aspect as your site can get the top position if it is filled with quality content. The content should be structurally and grammatically correct with unique and authentic facts. Sincere contents always appeal to people. More particularly, it needs to be valuable to the people you want to find your site. There should be descriptive and relevant page titles so that search engines can easily recognize the key contents of your site and people going through search results can instantly find their requirements. Finally one should use real headings, search engine friendly URLs, valid, semantic, lean, and accessible markup and submit carefully.
With the Term Targeting tool of SEOmoz, one can determine how well a particular keyword is targeted by a page. This software can assign a letter and percentage grade, which can reflect the keyword strength usage. This can be done after the analysis throughout the page, the frequency and location of the keyword. This tool or software uses in-house created formula at SEOmoz. Many clients have been successful using this methodology, engine optimization search practice and the usage of keyword is different and the results can vary. For an estimation of inclusion of keyword is the best use of this tool and not for a strict rules set for how other website or webpage should be using keywords. Examining a web page and also determining which keyword appears prominently in different applied weights in the HTML is the purpose of this tool.
An h1 tag will be given more weight in keyword than a keyword in the main body copy. SEOmoz created this tool for their premium members and it may be later opened to all users. Analyzing the content of any given page and also extracting the phrases and terms, which appear in the search engine in target. This tool has a good premise and analysis can be done all the time. This tool has benefited most of the companies in their projects with different clients. Use this software for knowing the strength of the keyword usage. However, the weightings that this tool use are that which Rand Fishkin and his crew came up with reports making valuable particularly to the beginners of SEOs.
Spider Simulator is a tool that simulates a search engine by displaying the contents of a web page the exact way a search engine will see it. This software can also help in displaying hyperlinks, which at times can be very important. The search engine spider simulator provides you with a report on aspects of the page that are used to determine how highly in a search engine the page is ranked, and also gives some general usability feedback.
One can now download a page through this tool, where you will be given an idea of what the search engines see when they spider the site you are using. This software is very useful for seeing what search engine will be first encountering on the page, however the output may look a bit ugly because the web pages will look like without the markup of any HTML. In fact, this can act as a good indicator of how fairly the design of the page is done provided it will be readable and can be used in this mode. This tool can ultimately help in your SEO and help you set your website such that you can drive traffic to it.