The search console has thrown a new error which is affecting 95 pages of my site - google-search-console

The error shown is "These hosts serve site resources that are blocked to Googlebot. If Googlebot can't access important resources on your page, the page might be indexed incorrectly. "
The hostname is https://checkout.razorpay.com

Related

Google Search Console deadlock: Unable to remove URL so I can index a different URL

I recently tried some SEO tweaks to see if I could improve traffic to my personal website. One of these changes was shortening some of my URLs. This worked just fine for about 4-5 of the URLs that I changed, except for one.
The problematic URL is
https://aleksandrhovhannisyan.github.io/blog/dev/how-to-add-a-copy-to-clipboard-button-to-your-jekyll-blog/
Which I shortened to
https://aleksandrhovhannisyan.github.io/blog/dev/jekyll-copy-to-clipboard-button/
The former URL no longer exists on my website, and I did not set up a 301 redirect. Instead, I requested that Google remove it from search results via Google Search Console. I also requested that it index the new URL.
Basically, I'm in a bizarre deadlock situation where Google refuses to index my new URL, claiming it selected the old one as a canonical because of "duplicate content":
According to the Removals page, the old URL was in fact removed:
But when I inspect this old URL, Google Search Console claims the URL is on Google:
How do I get out of this mess?

"/recaptcha/api2/logo_48.png" blocked by Google

I have a contact form on my site, it all works fine. I also using Google Captcha for that form.
When I go to my G search console to make sure all is fine, I see I get one error stating:
Googlebot couldn't get all resources for this page. Here's a list:
https://www.gstatic.com/recaptcha/api2/logo_48.png << Blocked
I have gone to my robots.txt file and added the following but that didnt help
Allow: https://www.gstatic.com/recaptcha/api2/logo_48.png
Allow: /recaptcha/api2/logo_48.png
Your own robots.txt is only for URLs from your host.
The message is about a URL from a different host (www.gstatic.com). This host would have to edit its robots.txt file to allow crawling of /recaptcha/api2/logo_48.png. It’s currently disallowed.
In other words: You can’t control the crawling of files that are hosted by someone else.

There is some errors releved from my GWT account

The GWT has sent the bellow message:
Dear webmaster,
Your smartphone users will be happier if they don’t land on error- or non-existent pages, so we recommend you make sure that your pages return the appropriate HTTP code. Currently, Googlebot for smartphones detects a significant increase in URLs returning a 200 (available) response code, but we think they should return an HTTP 404 (page not found) code.
Recommended actions
Check the Smartphone Crawl Errors page in Webmaster Tools.
Return a 404 (page not found) or 410 (gone) HTTP response code in response to a request for a non-existent URL.
Improve the user experience by configuring your site to display a custom 404 page when returning a 404 response code.
Now how to resolved this?
Have you made any significant changes lately? Like changing URLs to all pages?
First of, make sure your pages are available and working with the URLs. Try searching yourself on google with "site:yourdomain.com". Are these pages correct or do they not exist?
You should also check that IF your page does not exist (yourdomain.com/blahblah), it will return HTTP404 (Not found) and not HTTP200 (OK). You can see this in Chrome Developer Tools. Go to Network tab, reload the page, check the Status column for your HTML page.
How you change the HTTP code depends on your web server and language. In PHP you can use header().

Facebook 503 502 Same html different servers different results

I've got the same html on two separate Mediatemple (DV) servers and one Godaddy server. Godaddy resolves quickly and both Mediatemple server fail with 502 when run though Facebook debugger.
Mediatemple (DV).
http://www.site500.com/facebook.html
Godaddy
http://www.trainlearncreate.com/facebook.html
Also my server access logs do not register a hit by Facebook when the 502 is returned.
Godaddy http://www.trainlearncreate.com/facebook.html
That page has it's og:url set to http://www.trainlearncreate.com/facebook.html
Mediatemple (DV). http://www.site500.com/facebook.html
That page has it's og:url set to http://site500.com/facebook.html
Currently the second page is working for me in the debug tool as well - but if you maybe had a non-www-to-www-redirect in place, this would lead to a redirect circle, since Facebook tries to retrieve the OG info from the address given as og:url.

Domain blocked and no data scraped

I recently purchased the domain www.iacro.dk from UnoEuro and installed WordPress planning to integrate blogging with Facebook. However, I cannot even get to share a link to the domain.
When I try to share any link on my timeline, it gives the error "The content you're trying to share includes a link that's been blocked for being spammy or unsafe: iacro.dk". Searching, I came across Sucuri SiteCheck which showed that McAfee TrustedSource had marked the site as having malicious content. Strange considering that I just bought it, it contains nothing but WordPress and I can't find any previous history of ownership. But I got McAfee to reclassify it and it now shows up green at SiteCheck. However, now a few days later, Facebook still blocks it. Clicking the "let us know" link in the FB block dialog got me to a "Blocked from Adding Content" form that I submitted, but this just triggered a confirmation mail stating that individual issues are not processed.
I then noticed the same behavior as here and here: When I type in any iacro.dk link on my Timeline it generates a blank preview with "(No Title)". It doesn't matter if it's the front page, a htm document or even an image - nothing is returned. So I tried the debugger which returns the very generic "Error Parsing URL: Error parsing input URL, no data was scraped.". Searching on this site, a lot of people suggest that missing "og:" tags might cause no scraping. I installed a WP plugin for that and verified tag generation, but nothing changed. And since FB can't even scrape plain htm / jpg from the domain, I assume tags can be ruled out.
Here someone suggests 301 Redirects being a problem, but I haven't set up redirection - I don't even have a .htaccess file.
So, my questions are: Is this all because of the domain being marked as "spammy"? If so, how can I get the FB ban lifted? However, I have seen examples of other "spammy" sites where the preview is being generated just fine, e.g. http://dagbok.nu described in this question. So if the blacklist is not the only problem, what else is wrong?
This is driving me nuts so thanks a lot in advance!
I don't know the details, but it is a problem that facebook has with web sites hosted on shared servers, i.e. the server hosting your web site also hosts a number of other web sites.