lektor-website/packages/project-data/lektor_project_data.py

125 lines
4.2 KiB
Python
Raw Permalink Normal View History

# -*- coding: utf-8 -*-
import re
from email.policy import EmailPolicy
import readme_renderer.markdown
import readme_renderer.rst
import readme_renderer.txt
import requests
from lektor.pluginsystem import Plugin
_RENDERERS = {
None: readme_renderer.rst, # Default if description_content_type is None
'': readme_renderer.rst, # Default if description_content_type is None
'text/plain': readme_renderer.txt,
'text/x-rst': readme_renderer.rst,
'text/markdown': readme_renderer.markdown,
}
def normalize_url(url):
"""Normalize project home page URLs."""
# Normalize any URLS to GitHub project repos.
m = re.match(
r"""
https?://(?:www\.)?github\.com
/ (?P<owner>[^/]+)
/ (?P<project>[^/]+?) (?:\.git)
/? \Z
""",
url,
flags=re.VERBOSE
)
if m:
return "https://github.com/{owner}/{project}".format(**m.groupdict())
return url
def parse_header(ct: str) -> tuple[str, dict[str, str]]:
"""Parse a content-type string into a main value and a dictionary of
parameters.
This is a replacement for the deprecated ``cgi.parse_header`` function.
"""
header = EmailPolicy.header_factory("Content-Type", ct)
return (header.content_type, dict(header.params))
class ProjectDataPlugin(Plugin):
name = 'Project Data'
description = u'Retrieve project information from PyPI.'
data = {}
def render(self, value, content_type=None):
"""Render project description.
This is taken from
https://github.com/pypa/warehouse/blob/master/warehouse/filters.py
to ensure compliance and not reinvent the wheel. We don't
want to be creative here.
"""
content_type, parameters = parse_header(content_type or '')
# Get the appropriate renderer
renderer = _RENDERERS.get(content_type, readme_renderer.txt)
# Actually render the given value, this will not only render
# the value, but also ensure that it's had any disallowed
# markup removed.
rendered = renderer.render(value, **parameters)
# If the content was not rendered, we'll render as plaintext
# instead. The reason it's necessary to do this instead of
# just accepting plaintext is that readme_renderer will deal
# with sanitizing the content.
if rendered is None:
rendered = readme_renderer.txt.render(value)
return rendered
def package_data(self, name, entry_point=None):
if not entry_point:
entry_point = 'https://pypi.org/pypi'
url = '{}/{}/json'.format(entry_point, name)
resp = requests.get(url)
pkg = resp.json()
self.data.update(pkg['info'])
# Erase bad keys that are sometimes returned from the api
# to handle it in the template.
# To us, unknown is the same as non-existent.
for key in self.data:
val = self.data.get(key)
if type(val) is str and val.strip() == 'UNKNOWN':
self.data[key] = ''
self.data['short_name'] = name.split('lektor-')[1]
# Rewrite description as rendered description.
self.data['description'] = self.render(
self.data['description'], self.data['description_content_type'])
if not self.data.get('home_page'):
self.data['home_page'] = f'https://pypi.org/project/{name}/'
else:
self.data['home_page'] = normalize_url(self.data['home_page'])
def github_data(self, owner=None, repo=None):
url = 'https://api.github.com/repos/{}/{}'.format(owner, repo)
response = requests.get(url)
data = response.json()
return data
def project_data(self, name):
self.package_data(name)
# github data not currently used. Commented to save build time.
# if 'github' in self.data.get('home_page'):
# owner = self.data['home_page'].split('/')[-2]
# repo = self.data['home_page'].split('/')[-1]
# self.data['gh'] = self.github_data(owner=owner, repo=repo)
# TODO: support bitbucket
return self.data
def on_setup_env(self, **extra):
self.env.jinja_env.globals['project_data'] = self.project_data