Google webmaster does not index full sitemap? - google-search-console

My URL is: www.etypeservices.synology.me/etypes/
on the above URL i have several folders and sub folders which contains PDF files, i have tried since more that two weeks but google webmaster tools doesn't index complete sitemap, sitemap contains more than 3000 url but in indexing report it only shows approx 800 urls are index, also it is not showing any kind of errors in sitemap aur urls, then i didn't understand what is the reason why google webmaster not index my complete sitemap here is, one more thing these files are on nas drive and nas drive is hosted with above url, and here is the url of sitemap http://www.etypeservices.synology.me/etypes/sitemap.xml

Related

English Literature site is not indexing new pages in google

This is my site link:
www.englishact.com
This is the sitemap current position:
Google is showing no error in sitemap or any other pages. But indexed pages are 0 for about 3 months. I also have uploaded new sitemaps which are acting same way with no index.
NB:
I am using 1and1 paid hosting package. Also, google has accepted adsence for this site. Now what can I do? Any suggestions?
Your website is index on Google, i just searched for site: www.englishact.com and got many results.
Check if the links in your XML sitemap are valid or redirecting to another URL.
Also you have to solve the duplication in the URLs, you can access your website with WWW and without it, also you have two URLs for the homepage http://englishact.com/ and http://www.englishact.com/index.php
After fixing these errors your website should be healthy and Google will understand the structure of it.

how to set Robots.txt files for subdomains?

I have a subdomain eg blog.example.com and i want this domain not to index by Google or any other search engine. I put my robots.txt file in 'blog' folder in the server with following configuration:
User-agent: *
Disallow: /
Would it be fine to not to index by Google?
A few days before my site:blog.example.com shows 931 links but now it is displaying 1320 pages. I am wondering if my robots.txt file is correct then why Google is indexing my domain.
If i am doing anything wrong please correct me.
Rahul,
Not sure if your robots.txt is verbatim, but generally the directives are on TWO lines:
User-agent: *
Disallow: /
This file must be accessible from http://blog.example.com/robots.txt - if it is not accessible from that URL, the search engine spider will not find it.
If you have pages that have already been indexed by Google, you can also try using Google Webmaster Tools to manually remove pages from the index.
This question is actually about how to prevent indexing of a subdomain, here your robots file is actually preventing your site from being noindexed.
Don’t use a robots.txt file as a means to hide your web pages from Google search results.
Introduction to robots.txt: What is a robots.txt file used for? Google Search Central Documentation
For the noindex directive to be effective, the page or resource must not be blocked by a robots.txt file, and it has to be otherwise accessible to the crawler. If the page is blocked by a robots.txt file or the crawler can’t access the page, the crawler will never see the noindex directive, and the page can still appear in search results, for example if other pages link to it.
Block Search indexing with noindex Google Search Central Documentation

Different Google Index Information

More or less three month ago, I launched my own website. On the first day, I also verified my website for the Google Webmaster Tools, combined them with the Google Analytics Account and submitted a sitemap index file linked to five sitemap files.
But till now, I receive different Google Index Status information:
In Webmaster Tools:
Menu: Crawl -> Sitemaps: 123,861 Urls submitted, 64,313 Urls indexed
Menu: Google Index -> Index Status: 65,375 Urls indexed
When I type in google.de: “site:www.mysite.de”, then I receive 103,000 results.
When I check my website with push2check.net, I receive 110,000 Urls in Google Index.
What is wrong there? I understand that’s impossible for Google to deliver the accurate date because of the distributed processing and the result also depends on the location, where you searching from and so on. But between 65,000 and 110,000 is a huge gap. What’s the reason?
Thanks in advance!
Toby
google.de: “site:www.mysite.de”
You can search this type then google view your site all pages display index by Google.
And
Only Search
push2check.net
Then google View all result display when your website link have.
Then both result is different.

Google adsense doubleclick crawl errors

Google adsense crawl this page http://www.finewallpaperss.com/doubleclick that is not available on my site.
following is url that attempt many times by Google adsense.
http://www.finewallpaperss.com/doubleclick/DARTIframe.html?gtVersion=200_26&mediaserver=http%3A%2F%2Fs0.2mdn.net%2F879366&xpc=%7B%22cn%22%3A%22peerIframe1377068657642%22%2C%22tp%22%3Anull%2C%22osh%22%3Anull%2C%22pru%22%3A%22http%3A%2F%2Fwww.finewallpaperss.com%2Fdoubleclick%2FDARTIframe.html%3FgtVersion%3Drelay_200_26%26mediaserver%3Dhttp%3A%2F%2Fs0.2mdn.net%2F879366%22%2C%22ppu%22%3A%22http%3A%2F%2Fgoogleads.g.doubleclick.net%2Frobots.txt%22%2C%22lpu%22%3A%22http%3A%2F%2Fwww.finewallpaperss.com%2Frobots.txt%22%7D
Any solution for this.
To fix the DoubleClick DARTIframe issue, you will need to create a new folder in the root directory of your website named as doubleclick. For example, the one we have here is http://techie-buzz.com/doubleclick/ and then upload a file provided by them called DARTIframe.html to that directory.
https://support.google.com/richmedia/answer/156583

Site results redirect to Google

A site I'm working on has been hacked. The CMS (which I didn't build) was accessed and some files (e.g. "km2jk4.php.jpg") were uploaded in image fields. I have since deleted them (a week ago). Now, when I search for the site on Google, then click the result, it either:
a) simply redirects me to the Google search page
OR
b) a download dialogue appears asking me to download a zip file, with the source domain something like gb.celebritytravelnetwork.com
Clearly the site's been compromised. But if I simply type the URL in the address bar, the site loads fine. This only happens when I click through Google results.
There is no .htaccess file on the server, and this is not a virus on my computer, since many other people have reported the same thing happening, so this question is not relevant.
Any ideas please?
Thanks.
Your Source files have been changed.
Check all the files included in the index page. They might be header , footer pages.
And try using : fetch as google bot.