Disallow: /message-id/raw/
Disallow: /message-id/flat/
-Sitemap: http://www.postgresql.org/sitemap.xml
+Sitemap: https://www.postgresql.org/sitemap.xml
""", content_type='text/plain')
for p in get_all_pages_struct():
pages+=1
x.startElement('url', {})
- x.add_xml_element('loc', 'http://www.postgresql.org/%s' % urllib.quote(p[0]))
+ x.add_xml_element('loc', 'https://www.postgresql.org/%s' % urllib.quote(p[0]))
if len(p) > 1 and p[1]:
x.add_xml_element('priority', unicode(p[1]))
if len(p) > 2 and p[2]:
def init_crawl(self):
# Fetch the sitemap. We ignore robots.txt in this case, and
# assume it's always under /sitemap.xml
- u = urllib.urlopen("http://%s/sitemap.xml" % self.hostname)
+ u = urllib.urlopen("https://%s/sitemap.xml" % self.hostname)
p = SitemapParser()
p.parse(u)
u.close()