Closed. This question is off-topic. It is not currently accepting answers.
Want to improve this question? Update the question so it's on-topic for Stack Overflow.
Closed 9 years ago.
Improve this question
I submitted my robots.txt file ages ago to Google and it is still giving me a syntax not understood for the first line.
After Googling the most common problem is Google adding a '?' at the start of the line but it isnt doing that to me.
the url to the robots.txt is
www.leisurepursuits.co.uk/robots.txt
The error is:
Line 1: User-agent: * Syntax not understood
This error shows because the expected robots.txt file format is plain text encoded in UTF-8. The file consists of records (lines) separated by CR, CR/LF or LF.
If a character encoding is not a subset of UTF-8, then your robots.txt files is being parsed incorrectly.
First check your robots.txt url in http://www.asymptoticdesign.co.uk/cgi-bin/check-url.pl by selecting the second option “view source” and see that it responds successfully or not.
Now upload your robots.txt file again. It will not show error as shown previously.
Related
Closed. This question needs details or clarity. It is not currently accepting answers.
Want to improve this question? Add details and clarify the problem by editing this post.
Closed 4 days ago.
Improve this question
I am pretty much new in Perl and wanted some guidance on how this issue can be resolved.
I have written a program to collect the information from the data and put it in the excel. When sending the email using the MIME::Lite->new.
The issue I am facing here is that when I print the complete windows path of the report location only half of the path is hyperlinked and the rest is not. The issue is due to the space in the directory path. I have tried using double slashes but it still doesn't work. I have shared the screenshot below and you can see that post the space, rest of the line is ignored.
Any guidance / advice would be great. Thank you.
Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
This question does not appear to be about a specific programming problem, a software algorithm, or software tools primarily used by programmers. If you believe the question would be on-topic on another Stack Exchange site, you can leave a comment to explain where the question may be able to be answered.
Closed 7 years ago.
Improve this question
I was saving a file in notepad, and I got the warning message that the file contained unicode that wouldn't save correctly as ansi. I didn't see any characters that would be a problem, so I opened the file in the console and found these random-looking characters in the text. They aren't visible in notepad at all. Even the space they take up in the console isn't there in notepad. Below I've posted a screenshot of the text in notepad and in the console. What's going on?
notepad
console
The characters being displayed at the console are the UTF-8 encoding for a Zero Width Space, which would not be visible in Notepad or any other Unicode application. The console doesn't work with Unicode normally, it uses a code page (typically code page 437) to determine which character to display based on the bytes that are output. The UTF-8 encoding of the Zero Width Space is 3 bytes \xe2\x80\x8b so you see 3 characters on the console, doubled up because you have two Zero Width Spaces.
Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 7 years ago.
Improve this question
I don't used to post any questions on forums but this time, I have no other solution...
The Facebook URL Linter gives me the following "Error parsing input URL, no data was scraped" on this URL: http://phrasesbox.com/test.html
But with this URL is OK: http://jeudephrases.com/test.html
My problem is that these 2 domains are located on the same server, and the "test.html" file is unique (these 2 domains point to the same folder).
This gives me a very big problem because there is no preview on sharing contents on Facebook... (title, description, image).
All was working fine until one month ago. It's like if my phrasesbox domain were in a blacklist but when I share contents, there is no SPAM notice.
Same problem with 2 other domains that also point to the same folder.
Any idea??
/* EDIT */
I have already lost 70% of visits ...
Closed. This question is off-topic. It is not currently accepting answers.
Want to improve this question? Update the question so it's on-topic for Stack Overflow.
Closed 9 years ago.
Improve this question
I have an app whose content should not be publicly indexed. I've therefore disallowed access to all crawlers.
robots.txt:
# Robots shouldn't index a private app.
User-agent: *
Disallow: /
However, Bing has been ignoring this and daily requests a /trafficbasedsspsitemap.xml file, which I have no need to create.
I also have no need to receive daily 404 error notifications for this file. I'd like to just make the bingbot go away, so what do I need to do to forbid it from making requests?
According to this answer, this is Bingbot checking for an XML sitemap generated by the Bing Sitemap Plugin for IIS and Apache. It apparently cannot be blocked by robots.txt.
For those coming from google-
You could block bots via apache user agent detection/ rewrite directives, that would allow you to keep bingbot out entirely.
https://superuser.com/questions/330671/wildcard-blocking-of-bots-in-apache
Block all bots/crawlers/spiders for a special directory with htaccess
etc.
Closed. This question is off-topic. It is not currently accepting answers.
Want to improve this question? Update the question so it's on-topic for Stack Overflow.
Closed 11 years ago.
Improve this question
when i print this characters in facebook status it makes a rubbish like !̸̶͚͖͖̩̻̩̗͍̮̙̈͊͛̈͒̍̐ͣͩ̋ͨ̓̊̌̈̊́̚͝͠ͅ
and it seems to be everywhere not in facebook only i searched and it was found to be "Iota" the ninth character in the Greek alphabet , does it appear rubbish because i'm pasting it with non-ascii or another format or what ?
i'm really interested to know
The font used doesn't contain that character - not every font contains every Unicode character.