Описание
With the following crawler configuration:
from bs4 import BeautifulSoup as Soup
url = "https://example.com"
loader = RecursiveUrlLoader(
url=url, max_depth=2, extractor=lambda x: Soup(x, "html.parser").text
)
docs = loader.load()
An attacker in control of the contents of https://example.com could place a malicious HTML file in there with links like "https://example.completely.different/my_file.html" and the crawler would proceed to download that file as well even though prevent_outside=True.
Resolved in https://github.com/langchain-ai/langchain/pull/15559
Ссылки
- Patch
- Issue TrackingPatch
- ExploitIssue TrackingThird Party Advisory
- Patch
- Issue TrackingPatch
- ExploitIssue TrackingThird Party Advisory
Уязвимые конфигурации
Конфигурация 1Версия до 0.1.0 (исключая)
cpe:2.3:a:langchain:langchain:*:*:*:*:*:*:*:*
EPSS
Процентиль: 23%
0.00073
Низкий
3.7 Low
CVSS3
8.1 High
CVSS3
Дефекты
CWE-918
Связанные уязвимости
CVSS3: 3.7
github
почти 2 года назад
langchain Server-Side Request Forgery vulnerability
EPSS
Процентиль: 23%
0.00073
Низкий
3.7 Low
CVSS3
8.1 High
CVSS3
Дефекты
CWE-918