"Issue Name","Issue Type","Issue Priority","URLs","% of Total","Description","How To Fix","Help URL"
"H1: Multiple","Warning","Medium","4","0.150","Pages which have multiple
s. While this is not strictly an issue because HTML5 standards allow multiple s on a page, there are some problems with this modern approach in terms of usability. It's advised to use heading rank (h1-h6) to convey document structure. The classic HTML4 standard defines there should only be a single per page, and this is still generally recommended for users and SEO.","Consider updating the HTML to include a single on each page, and utilising the full heading rank between (h2 - h6) for additional headings.",""
"Page Titles: Over 60 Characters","Opportunity","Medium","247","9.540","Pages which have page titles that exceed the configured limit. Characters over this limit might be truncated in Google's search results and carry less weight in scoring.","Write concise page titles to ensure important words are not truncated in the search results, not visible to users and potentially weighted less in scoring.",""
"Meta Description: Below 400 Pixels","Opportunity","Low","87","3.360","Pages which have meta descriptions much shorter than Google's estimated pixel length limit. This isn't necessarily an issue, but it does indicate there might be room to communicate benefits, USPs or call to actions.","Consider updating the meta description to take advantage of the space left to include additional benefits, USPs or call to actions to improve click through rates (CTR).",""
"Page Titles: Missing","Issue","High","10","0.390","Pages which have a missing page title element, the content is empty, or has a whitespace. Page titles are read and used by both users and the search engines to understand what a page is about. They are important for SEO as page titles are used in rankings, and vital for user experience, as they are displayed in browsers, search engine results and on social networks.","It's essential to write concise, descriptive and unique page titles on every indexable URL to help users, and enable search engines to score and rank the page for relevant search queries.",""
"Page Titles: Same as H1","Opportunity","Low","2179","84.200","Page titles which match the h1 on the page exactly. This is not necessarily an issue, but may point to a potential opportunity to target alternative keywords, synonyms, or related key phrases.","This is not necessarily an issue, but may point to a potential opportunity to target alternative keywords, synonyms, or related key phrases.",""
"Page Titles: Over 561 Pixels","Opportunity","Medium","243","9.390","Pages which have page titles over Google's estimated pixel length limit for titles in search results. Google snippet length is actually based upon pixels limits, rather than a character length. The SEO Spider tries to match the latest pixel truncation points in the SERPs, but it is an approximation and Google adjusts them frequently.","Write concise page titles to ensure important words are not truncated in the search results, not visible to users and potentially weighted less in scoring.",""
"Response Codes: Internal Blocked Resource","Warning","High","2302","31.210","Internal resources (such as images, JavaScript and CSS) that are blocked from rendering by robots.txt or an error. This filter will only populate when JavaScript rendering is enabled (blocked resources will appear under 'Blocked by Robots.txt' in default 'text only' crawl mode). This can be an issue as the search engines might not be able to access critical resources to be able to render pages accurately. Blocked resources can be viewed by URL in the 'Rendered Page' tab, and any pages with blocked resources can be viewed under 'JavaScript > Pages with Blocked Resources'.","Update the robots.txt and resolve any errors to allow all critical resources to be crawled and used for rendering of website content.",""
"Meta Description: Over 155 Characters","Opportunity","Low","820","31.680","Pages which have meta descriptions over the configured limit. Characters over this limit might be truncated in Google's search results.","Write concise meta descriptions to ensure important words are not truncated in the search results, and not visible to users.",""
"JavaScript: Page Title Updated by JavaScript","Warning","Medium","34","1.310","Pages that have page titles that are modified by JavaScript. This means the page title in the raw HTML is different to the page title in the rendered HTML.","While Google is able to render pages and see client-side only content, consider including important content server side in the raw HTML.",""
"Structured Data: Rich Result Validation Errors","Issue","High","40","1.550","URLs that contain Google rich result feature validation errors. Google rich result feature validation will show errors for missing required properties or problems with the implementation of required properties. Google's 'required properties' must be included and be valid for content to be eligible for display as a rich result.","Resolve validation errors to ensure pages are eligible for special search result features in Google. Review the 'Structured Data Details' tab for more information on specific errors and refer to Google rich result feature docs where necessary. Export unique validation errors using 'Reports > Structured Data > Validation Errors & Warnings Summary' and test in Google's Rich Results tool.",""
"Response Codes: Internal Client Error (4xx)","Issue","High","37","0.500","Internal URLs with a client-side error. This indicates a problem occurred with the URL request and can include responses such as 400 bad request, 403 Forbidden, 404 Page Not Found, 410 Removed, 429 Too Many Requests and more. A 404 'Page Not Found' is the most common, and often referred to as a broken link. View URLs that link to errors using the lower 'inlinks' tab and export them in bulk via 'Bulk Export > Response Codes > Internal > Client Error (4xx) inlinks'.","All links on a website should ideally resolve to 200 'OK' URLs. Errors such as a 404 or 410 should be updated to their correct locations, removed and redirected where appropriate. A 403 forbidden error occurs when a web server denies access to the SEO Spider's request and can often be resolved by switching the user-agent to Chrome via 'Config > User-Agent' and crawling again.",""
"JavaScript: Pages with Blocked Resources","Warning","High","2616","100.000","Pages with resources (such as images, JavaScript and CSS) that are blocked from rendering by robots.txt or an error. This filter will only populate when JavaScript rendering is enabled (blocked resources will appear under 'Blocked by Robots.txt' in default 'text only' crawl mode). This can be an issue as the search engines might not be able to access critical resources to be able to render pages accurately. Blocked resources can be viewed by URL in the 'Rendered Page' tab, or in bulk under 'Response Codes > Blocked Resource'.","Update the robots.txt and resolve any errors to allow all critical resources to be crawled and used for rendering of the websites content. Resources that are not critical (e.g. Google Maps embed) can be ignored.",""
"Meta Description: Below 70 Characters","Opportunity","Low","120","4.640","Pages which have meta descriptions below the configured limit. This isn't strictly an issue, but an opportunity. There is additional room to communicate benefits, USPs or call to actions.","Consider updating the meta description to take advantage of the space left to include additional benefits, USPs or call to actions to improve click through rates (CTR).",""
"Response Codes: Internal Blocked by Robots.txt","Warning","High","2673","36.240","Internal URLs blocked by the site's robots.txt. This means they cannot be crawled and is a critical issue if you want the page content to be crawled and indexed by search engines. View URLs that link to URLs blocked by robots.txt using the lower 'inlinks' tab and export them in bulk via 'Bulk Export > Response Codes > Internal > Blocked by Robots.txt inlinks'.","Review URLs to ensure they should be disallowed. If they are incorrectly disallowed, then the site's robots.txt should be updated to allow them to be crawled. Consider whether you should be linking internally to these URLs and remove links where appropriate.",""
"Content: Low Content Pages","Opportunity","Medium","82","3.170","Pages with a word count that is below the default 200 words. The word count is based upon the content area settings used in the analysis which can be configured via 'Config > Content > Area'. There isn't a minimum word count for pages in reality, but the search engines do require descriptive text to understand the purpose of a page. This filter should only be used as a rough guide to help identify pages that might be improved by adding more descriptive content in the context of the website and page's purpose. Some websites, such as ecommerce, will naturally have lower word counts, which can be acceptable if a products details can be communicated efficiently.","Consider including additional descriptive content to help the user and search engines better understand the page.",""
"Security: Missing Content-Security-Policy Header","Warning","Low","2700","46.980","URLs that are missing the Content-Security-Policy response header. This header allows a website to control which resources are loaded for a page. This policy can help guard against cross-site scripting (XSS) attacks that exploit the browser's trust of the content received from the server. The SEO Spider only checks for existence of the header, and does not interrogate the policies found within the header to determine whether they are well set-up for the website. This should be performed manually.","Set a strict Content-Security-Policy response header across all page to help mitigate cross site scripting (XSS) and data injection attacks.",""
"URL: Uppercase","Warning","Low","1","0.020","URLs that have uppercase characters within them. URLs are case sensitive, so as best practice generally URLs should be lowercase, to avoid any potential mix ups and duplicate URLs.","Ideally lowercase characters should be used for URLs only. However, changing URLs is a big decision, and often it's not worth changing them for SEO purposes alone. If URLs are changed, then appropriate 301 redirects must be implemented.",""
"Security: HTTP URLs","Issue","High","72","1.250","HTTP URLs that are encountered in the crawl. All websites should be secure over HTTPS today on the web. Not only is it important for security, but it's now expected by users. Chrome and other browsers display a 'Not Secure' message against any URLs that are HTTP, or have mixed content issues (where they load insecure resources on them). To view how these URLs were discovered, view their 'inlinks' in the lower window tab. You can also export any pages that link to HTTP URLs via 'Bulk Export > Security > HTTP URLs Inlinks'.","All URLs should be to secure HTTPS pages. Pages should be served over HTTPS, any internal links should be updated to HTTPS versions and HTTP URLs should 301 redirect to HTTPS versions. HTTP URLs identified in this filter that are redirecting to HTTPS versions already should be updated to link to the correct HTTPS versions directly.",""
"H2: Duplicate","Opportunity","Low","120","4.640","Pages which have duplicate