Tag: robots

  • Website Customer Data (search engine optimization)

    Website Customer Data (search engine optimization)




    icons-set-ENG-2



    The product’s functionality is already quite impressive:

    icons-set-ENG-2

    icons-set-ENG-2

    Product Description:

    -mySTAT has unique features for statistics, analytics, SEO optimization and that is why anyone who has their own website should have it. In fact, mySTAT is a comprehensive set of tools necessary for collecting and classifying data from all areas. The previous version of mySTAT has over 500,000 downloads and strictly positive reviews! The product is a multi-platform one, so it can be installed onto both existing popular types of CMS and onto any other third-party developed sites.
    -Based on the data from mySTAT one can precisely and correctly regulate the operation of any resource, run a profitable business, organize a successful project or any other activity on the Internet.
    -mySTAT is a useful tool for SEOs, programmers, bloggers, designers, advertising sales agents, and later updates of this product will further expand the area of its application!
    -mySTAT can considerably simplify the work of a whole team or even save quite a lot by doing all the work independently.
    -mySTAT means full confidentiality of your data. It is your product that works on your site and nobody but you can use its data, unlike the services run by third-parties. mySTAT means the highest level of security. The code is optimized and minimized, it is free from vulnerabilities and it has passed all possible testing. Full integration into the WordPress system, user-friendly and intuitive interface. The information is accessible from any device and is displayed identically on all of them. Unique programming algorithms allow for obtaining the data of highest accuracy and make it possible to identify the number of real users, as well as the number of robots and spiders. This will give you reliable information about the frequency of indexing the web-site, its separate blocks or pages, if it goes specifically about robots, or, alternatively, a clear idea about the interests of the users based on the information about their activity.

    The product’s functionality is already quite impressive:

      • Statistics Summary
        Set of summary information presented as tiles and lists.
      • Site traffic
        Report with maximum details on your site visitors.
      • Page views per visitor
        This section contains views per visitor for each host.
      • Pages received per robot
        Detailed statistics on how many pages are viewed by the robot.
      • Time of downloading the pages
        This section deals with the speed of downloading your page on the client’s side.
      • Popular pages
        Information on your site’s popular page addresses, which are visited by the users most often.
      • Popular titles
        The section for the rating of page titles most visited by users.
      • Domain names
        If several different domains are connected to your site.
      • Popular error 404 pages
        These are pages that do not exist on your site, but for some reason they are visited by search engines, people or robots.
      • Referring addresses
        This is the section for site addresses, from which the users clicked through to your site.
      • Referring domains
        Overall rating of domains from which clickthroughs to your site occur.
      • Search engines
        Visits to your site from search engines.
      • Social networks
        The section displays the visits to your site made directly from search engines.
      • Mail services
        The clickthroughs from on-line search engines. Names of mail services, from which visitors clicked through to your site.
      • Search phrases
        Phrases and word-combinations, using which the visitors found your site via search engines.
      • Links to 404 pages
        This section lists the URLs from which the transitions to non-existent pages of your site were made.
      • IP addresses of visitors
        A detailed list of users network addresses.
      • IP addresses of robots and spiders
        A detailed list of robots and spiders network addresses.
      • Countries of visitors
        List of countries, from which your site was visited.
      • User-Agent of visitors
        This section is related to agents or user-agents and holds complete information on the identifier of the browser used by the user to visit the site.
      • User-Agent of robots and spiders.
        This section is also related to agents or user-agents and holds complete information on the identifier of the browser used to visit the site specifically by a robot.
      • Visitors’ system language
        The section refers to the user’s preferred language set up in the browser.
      • Robots and spiders
        The section provides information on the robots and spiders regarding their version and unique visit of a particular robot to your site.
      • Operating system
        The section deals with the visitor’s installed operating system.
      • Browsers
        Here you can see the browsers, with the help of which users come to your site.
      • Color Depth
        This information will help the designers, as it allows to find out how many colors and shades the client screen can display.
      • Screen resolution
        This section will help designers and programmers, as it deals with the client’s screen resolution.
      • Reporting database size
        This shows the details on database size used by statistics data.

    and that is far from the full list of the product’s features!
    Start your business right, develop and promote your web-site based on reliable information together with mySTAT!
    The functionality of the product is constantly expanded, thus everyone who purchases it now will get all further updates absolutely free of charge, including full technical support!

    Try mySTAT before buying!

  • Sensible Robots.txt Instruments (search engine optimization)

    Sensible Robots.txt Instruments (search engine optimization)

    Smart Robots.txt Tools allows you to improve and customize ‘robots.txt’ file used by search engines and other types of bots when indexing your website. Through this file you can disallow access to some areas of the website for all or only some bots.

    Support ForumsFollow on TwitterFollow on EnvatoSubscribe to Newsletter

    Request Full Demo

    This plugin allows you to use improved basic rules in robots.txt file with few extra predefined rules, or you can create your own custom rules that can include any number of bots user agents and rules to allow or disallow access.

    When plugin is installed, it will adopt WordPress search engine visibility settings, and if your website is set hidden from search engines (WordPress uses robots file for this), plugin will be set to Deny All. You can change robots content to anything through plugin settings, but as a convenience, WordPress settings are used on installation only.

    Fully customized set of rules in robots.txt

    Fully customized set of rules in robots.txt
    Add rules for different bots / search engines or any custom user agent, allowing or disallowing access to parts of the website, setting the URL to one or more sitemaps and setting crawl delay value.

    Predefined or custom content for robots.txt

    Predefined or custom content for robots.txt
    See the status of robots.txt file, availablilty and enable override through this plugin. Specify use of predefined or custom rules, and add some extra set of rules targeting some common search engines.

    Create robots.txt rules targeting any search engine or custom user agent

    Create robots.txt rules targeting any search engine or custom user agent
    Add rules for any search engine bot from the list or by any user agent string. Add any number of allow/disallow rules, one or more sitemaps and custom crawl delay. Modify rules for all user agents.

    Other Plugin Features Included

    • Extra panel to reset plugin settings including all custom rules.
    • Easy to use export and import for transferring settings from one website to another.
    • Support for Multisite WordPress mode, each website can set plugin on it’s own.
    • Support for translation and includes POT file.

    System and WordPress Requirements

    • PHP 5.2.4 or newer
    • WordPress 3.5 or newer

    Important about Robots.txt

    Before you proceed, there is one very important limitation for this file. This limitation is a specification of the ‘robots.txt’ file and it is not related to this plugin. Robots file must be located in the root for a domain or subdomain, it can’t be used if the website is in subdirectory of the main domain or subdomain. So, if you have installed your WordPress website in subdirectory of main website domain, then you can’t use robots.txt for that WordPress installation. Search engines can load robots.txt only from root of the domain, nowhere else.

    Documentation

    Plugin contains PDF user guide in the plugin package, inside the ‘docs’ directory. Check out this documentto get information on plugin options, usage and more.

    Changelog

    Version 1.0 / 2014.08.07.
    • First release

    Plugin WebsitePlugin ChangelogFollow on EnvatoAsk any Question

  • WordPress Prolonged Meta Tags (web optimization)

    WordPress Prolonged Meta Tags (web optimization)

    WP Extended Meta Tags is a plugin with wich you can add Title, Description and Keywords to ANY post, custom post, page, category or taxonomy you want.

    It works side by side with other SEO plugins, so if you only want to use it for Taxonomies, thats NO PROBLEM.

    I will add any useful setting you can come up with, it must have to do with SEO and Meta tags!

    Features

    • Title
    • Description
    • Keywords
    • Robots (index, follow, noindex, nofollow)

    How to use

    Upload the plugin and activate it. Once activated there will be a settings page under “Settings” > “WP Extended Meta Tags Settings”. Here you can enable or disable Meta Tags for the any desired taxonomy, posts/pages/custom posts is always enabled.

    If you activated it for let’s say categories, go to the categories and you will see four fields added: SEO Meta tag title, SEO Meta tag description, SEO Meta tag keywords and SEO Meta tag robots. Once you save this your category page will have these meta tags set in the HTML.

    To do

    • I will add any useful setting you can come up with.

    Bugs

    If you happen to find any bugs please by all means, report them to me here at Codecanyon. I will fix them as soon as possible!
    I will add any useful setting you can come up with, it must have to do with SEO and Meta tags!