This article will help you with your SEO problems to an extent. I stumbled upon a lot of new stuff today and thought it would be important to share it with the outside world of website developers. Out of the many tools I came across today were:


SeoSiteCheckup, AddThis & compression

All are too damn good for any SEO analyst or a specialist. Basically, I worked on over 2 dozen SEO techniques today, and all of them proved to be crucial. Here is an overview of the techniques I dealt with:

1) Keywords – The meta keywords tag allows you to provide additional text for search engines to index along with the rest of what you’ve written on your page. Meta keywords can emphasize a particular word or phrase in the main body of your text.

2) Most Common Keywords Test – Check the most common keywords & their usage (number of times used) on your web page. HOW TO FIX To pass this test, you must optimize the density of your primary keywords displayed above. If the density of a specific keyword is below 2%, you must increase it; if the density is over 4%, you must decrease it.

3) Keyword Usage – This describes if your most common keywords are used in your title, meta-description, and meta keyword tags. Keyword(s) not included in Meta-Title Keyword(s) included in Meta-Description tag Keyword(s) included in Meta-Keywords Tag HOW TO FIX, First of all, you must make sure that your page is using the title, meta-description, and meta keywords tags. Second, you must adjust these tags’ content to include the primary keywords displayed above Extra Update.

4) Headings Status – This indicates if any H1 headings are used on your page. H1 headings are HTML tags that can help emphasize important topics and keywords within a page. HOW TO FIX To pass this test, you must identify the most important topics from your page and insert those topics between tags. Example: Important topic goes here Another topic Headings Status This indicates if any H2 headings are used on your page. H2 headings can help describe the sub-topics of a page.

5) Robots.txt Test – Search engines send out tiny programs called spiders or robots to search your site and bring information back so that your pages can be indexed in the search results and found by web users. If there are files and directories you do not want to be indexed by search engines, you can use the “robots.txt” file to define where the robots should not go. These files are straightforward text files that are placed in the

The root folder of your website: There are two important considerations when using “robots.txt”: – the “robots.txt” file is publicly available, so anyone can see what sections of your server you don’t want robots to use; – robots can ignore your “robots.txt”, especially malware robots that scan the web for security vulnerabilities.