explicitly use "lxml" for finding page encoding
This commit is contained in:
parent
1d41bc7c34
commit
e4a5bd01e9
1 changed files with 1 additions and 1 deletions
|
@ -137,7 +137,7 @@ def _meta_content(s: str) -> typing.Dict[str, str]:
|
|||
return out
|
||||
|
||||
def _find_encoding(data: bytes) -> typing.Optional[str]:
|
||||
soup = bs4.BeautifulSoup(data)
|
||||
soup = bs4.BeautifulSoup(data, "lxml")
|
||||
if not soup.meta == None:
|
||||
meta_charset = soup.meta.get("charset")
|
||||
if not meta_charset == None:
|
||||
|
|
Loading…
Reference in a new issue