Moved to New Server/Ip Address Now Facebook Can't Scrape Website - facebook

About a week ago I moved my site to a new server and was assigned a new ip address. After the move I was able to continue sharing content to google +, twitter etc. Except for Facebook. The debugger tool keeps returning the default apache page.
I've tried ever tip I've seen on this site thus far, cleared ever cache and waited a week to see if facebook would clear its caches but that hasn't happened. All the old content made before the move and new content added after the move all return a 404 error.
http://www.awfbase.com
Any help would be greatly appreciated.

Related

Facebook showing page not found when sharing link

I'm sharing content from a website and every time I paste the link into Facebook it says 'page not found'.
Sometimes it works when I manually add the 'www.' in front of the URL in the address bar.
EXAMPLE
Shows page not found:
http://roundreviews.co.uk/reviews/speakers/native-union-monocle-speaker/
Works when you manually place www. in front:
www.roundreviews.co.uk/reviews/speakers/native-union-monocle-speaker/
I honestly have no I idea why it's doing this, any thoughts on how it can be fixed on the web side?
Also...
I have tried with the link below with both the www. and without yet it doesn't work with either of them, this is all very strange. This is the only link I have tried and it doesn't work with both:
www.roundreviews.co.uk/microphones/spark-digital-microphone/
Any help is much appreciated, thanks.
For me what it worked was to access the Facebook Debugger, as Goose said.
I saw that the scrape was about 12 hours ago, looks like it fetches the first time and saves it as caché or whatsoever...
What it worked for me is to debug the url, then click "fetch new scrape information" after the previous information has been shown.
Hope it works!
For those running across this today, you might find that you also need to verify your domain and link it to your page.
To do this you need to
Set up a Facebook Business Account
Add your page to the business account
Verify your domain (using DNS TXT or adding a page facebook gives you)
Under domains, connect your page as an asset of that domain

http://www.facebook.com/sharer.php?u= content not refreshing with page updates

I have noticed this numerous times, and have yet to find a fix; when using the share link (http://www.facebook.com/sharer.php?u=MyUrl&t=titleInfo), it pulls in the image, page title, and description from MyUrl just fine. However, it seems to ignore &t, and just uses the page title regardless. That is not the big issue though.
The problem is, if I totally change my page title and opening text, the share link won't update. It seems locked on whatever the page had on it the first time I tested the share link. Is there a way to make it refresh?
In my instance, I was updating some site pages from last years contest edition. Before I got the main page updated, I was working on the FB share link on a different page. I clicked it to test if it worked with the new graphic, and it did, but it of course pulled up the 2012 page content.
Then I went and updated all of the page content to make it for this contest (same URL) and now no matter who tests it, it is still pulling up last year's info (page title and description). It is as if FB has locked that info in and no matter who shares my link, it always pulls up the old text.
I have seen this before with YouTube links. Someone has shared one, I go in later and change the title, and no matter who shares it it never updates because that first initial share text seems permanently locked in FB.
Does anyone know how I can make it refresh, without having to make a whole new URL from last year?
actually facebook gets your page information at the time of posting and save it in their database. so next time it fetches that url and information from their database not from your page.
you may have to share that page again to make new entry in database of the facebook. but old one will remain there.
You can force Facebook to clear their cache by using their debugger. Enter the URL in question at https://developers.facebook.com/tools/debug. You need to do this for each page you are working on.
There are also some rules about not changing the info they have if there are more that (50)? likes.

Error parsing input URL, no data was scraped. only with new pages on my site

The problem i have is that i own a website where other people can post stuff ,creating new pages on my domain, but the problem that occured today is that all the new post pages created today are malfunctioning , sharing is not loading thumbnail picture and title and so on, but the weird this is that all the posts(new pages) created before today are all working fine
What caused an error to occur out of nowhere?
I also cannot debug any of the URL's of my website as the same error: Error parsing input URL, no data was scraped
The website im having problems with is here http://www.vabameedia.ee/vm/184/h%C3%A4da-ei-anna-h%C3%A4beneda.html
This is one of the sites where it says no error on page but facebook still cant reach it. http://www.vabameedia.ee/vm/178/craig-parks-%C3%BChek%C3%A4eline-krossisoitja.html
For people experiencing the same problem but for different causes, I discovered a few interesting things about how Facebook "scrapes" pages, checking the logs of the server while doing some trials.
First of all: if you never tried to share a page with FB, FB never tried to scrape it, and it will not try to do so if you only put the url in the Debug tool.
That's the first reason because you get the error: it just states that FB has no information on the page, you must "force" it to scrape the page.
The first time you try to share a page, FB scrapes it (asks your server the first 40k of the page and analyse the opengraph tags).
What can happen is that you do not see the image: Facebook Share Dialog does not display thumbnails one first load
The reason is that FB behind the scenes is still scraping your page and caching the image. The next time, in fact, you have also the image.
How to solve it? Pre caching: https://developers.facebook.com/docs/sharing/best-practices#precaching
or simply add
<meta property="og:image:width" content="450"/>
<meta property="og:image:height" content="298"/>
I was pulling my hair out trying to fix this issue. Hours and hours of troubleshooting to no avail. After speaking with one of our programmers about a topic unrelated I thought of something to try as a long shot.
Much to my surprise, it worked!!!
This is the reason behind the problem and my solution for it:
When you draft a post in WordPress it generates a link based on your article's title (unless you manually change it). The title of my article included special characters, however the auto-generated link didn't display these special characters, only hyphens to replace the spaces. Should be fine right? Wrong! Somewhere embedded in metadata and code in the WordPress platform are those special characters and they mess up the way Facebook pulls info from the article being linked to. This is a problem because certain special characters invalidate hyperlinks.
For example:
Article Title: R[eloaded]
Auto-generated hyperlink DISPLAYED in WordPress "Permalink" field: http://www.example.com/reloaded
Actual WordPress Auto-generated hyperlink: http://www.example.com/r[eloaded]
Those brackets will invalidate the link and Facebook will be unable to pull any information (ie pictures) from it.
Solution:
(1) Simply, manually change the WordPress hyperlink address to something that doesn't include any special characters (this will not change the title of your article).
(2) Click "Update" to change the post to include the new hyperlink.
(3) Click "Purge from Cache" in the WordPress window
(4) Refresh your Facebook browser window
(5) Paste the new hyperlink for your article
(6) Enjoy your Facebook post with a preview image and information
Sidenote: Don't pull your hair out over Facebook, it's not worth it. =)
If you're using Wordpress, edit the post in question to change the permalink (just alter it slightly), then update the post. Using the new permalink in the Facebook OG debugger should now work.
It's a weird fix, but I think it takes care of a problem caused by special characters being used in the title of a post, which is then used to make the permalink.
Its all about DNS issue, was having same issue and resolved it by updating domain name servers to actual name servers.
In my case my domain was pointed to ns1.websterz.net and ns2.websterz.net and on this server i had DNS redirect to my other server (where web site is hosted). I Just updated name servers of the domain to actual name servers where my web site is hosted on. This was account migration case i forgot to update name servers as of new server.
Everything works fine now.

Domain blocked and no data scraped

I recently purchased the domain www.iacro.dk from UnoEuro and installed WordPress planning to integrate blogging with Facebook. However, I cannot even get to share a link to the domain.
When I try to share any link on my timeline, it gives the error "The content you're trying to share includes a link that's been blocked for being spammy or unsafe: iacro.dk". Searching, I came across Sucuri SiteCheck which showed that McAfee TrustedSource had marked the site as having malicious content. Strange considering that I just bought it, it contains nothing but WordPress and I can't find any previous history of ownership. But I got McAfee to reclassify it and it now shows up green at SiteCheck. However, now a few days later, Facebook still blocks it. Clicking the "let us know" link in the FB block dialog got me to a "Blocked from Adding Content" form that I submitted, but this just triggered a confirmation mail stating that individual issues are not processed.
I then noticed the same behavior as here and here: When I type in any iacro.dk link on my Timeline it generates a blank preview with "(No Title)". It doesn't matter if it's the front page, a htm document or even an image - nothing is returned. So I tried the debugger which returns the very generic "Error Parsing URL: Error parsing input URL, no data was scraped.". Searching on this site, a lot of people suggest that missing "og:" tags might cause no scraping. I installed a WP plugin for that and verified tag generation, but nothing changed. And since FB can't even scrape plain htm / jpg from the domain, I assume tags can be ruled out.
Here someone suggests 301 Redirects being a problem, but I haven't set up redirection - I don't even have a .htaccess file.
So, my questions are: Is this all because of the domain being marked as "spammy"? If so, how can I get the FB ban lifted? However, I have seen examples of other "spammy" sites where the preview is being generated just fine, e.g. http://dagbok.nu described in this question. So if the blacklist is not the only problem, what else is wrong?
This is driving me nuts so thanks a lot in advance!
I don't know the details, but it is a problem that facebook has with web sites hosted on shared servers, i.e. the server hosting your web site also hosts a number of other web sites.

How to tell google that a specific page of my website disappeared and won't come back?

I have a website where 50% of the pages have a limited lifetime.
To give an idea, 4.000 pages appear each week and the same amount disappears.
By "appearing" and "disappearing", I mean that the appearing pages are completely new ones, and disappearing pages are removed from the website forever. There is no "this new page replaces this old page".
I naively used a 410 code on every URL where a page had disappeared.
Meaning the url http://mywebsite/this-page-was-present-until-yesterday.php returned until yesterday a 200 OK code, and returns now a 410 Gone code.
I didn't use no redirect, because I want to tell the user that the URL he accessed isn't wrong, but that it is expired.
The problem is : Google won't acknowledge this information. It is still crawling the pages and Webmaster Tools alerts me as if the page was 404 broken. This affects significantly my "reputation".
Did I do something wrong ? How should I proceed ?
It's always a very good idea to make your own error page. This can save you a lot of visits through broken links.
.htaccess error pages
The Webmaster Tools of Google enables you to delete certain pages.
You can find this under "crawler access".
Try adding a noindex header.