I'm building an angular web application that is able to periodically autogenerate sitemaps. Once it is done autogenerating all the sitemaps it automatically informs google of the location of the sitemaps using the ping get request.
Today i had first tested this functionality and it all worked correctly. The sitemaps were
created from a set of test urls and were then uploaded to a bucket i made in amazon s3. After that the application successfully pinged google the url of the sitemap index file.
There is only one problem. The website itself is not yet live. The only thing that is up right now is the registered website domain. All the test urls inside the sitemaps have this domain as their base url.
So when the google spider downloads the sitemaps and finds out that none of the website urls are valid, will this affect the ranking of my future website? Or will google just ignore the sitemaps when it finds out there is no website active on the domain?
Thank you
PS: I've removed all of the test sitemaps from the bucket after writing this post. Maybe i'm lucky and google hasn't downloaded the sitemaps yet.
When your sitemap is on an S3 bucket,Googlebot will ignore all the URLs in the sitemap until you add both the s3 bucket and the main site as separate properties in the same Google Search Console account.. According to the info in the sitemaps.org link you provided all i have to do is point to the aws s3 sitemaps url inrobots.txt. It saysYou can do this by modifying the robots.txt file on www.host1.com to point to the Sitemap on www.sitemaphost.com..The page makes no mention of also having to add it to the Google Console account. So won't google find the sitemaps through robots.txt? – Maurice Jan 04 '22 at 18:58sitemaps.example.comand have crawlers that visit that domain get redirected to the sitemap file i've stored in thes3 bucket. Would something like that be valid? – Maurice Jan 04 '22 at 19:12