allow empty urls (#2984)

This commit is contained in:
Jonathan Dobson
2025-07-18 10:20:33 -04:00
committed by GitHub
parent 764ac1928c
commit f5d7639de8

View File

@@ -13,7 +13,7 @@ from pydantic import BaseModel, PrivateAttr
from skyvern.config import settings
from skyvern.constants import BUILDING_ELEMENT_TREE_TIMEOUT_MS, DEFAULT_MAX_TOKENS, SKYVERN_DIR, SKYVERN_ID_ATTR
from skyvern.exceptions import FailedToTakeScreenshot, ScrapingFailed, ScrapingFailedNoUrl, UnknownElementTreeFormat
from skyvern.exceptions import FailedToTakeScreenshot, ScrapingFailed, UnknownElementTreeFormat
from skyvern.forge.sdk.api.crypto import calculate_sha256
from skyvern.forge.sdk.core import skyvern_context
from skyvern.forge.sdk.trace import TraceManager
@@ -427,8 +427,10 @@ async def scrape_website(
:raises Exception: When scraping fails after maximum retries.
"""
if not url.strip():
raise ScrapingFailedNoUrl()
# TODO(jdo) why is this a problem?
# ref: https://skyvern.slack.com/archives/C074UNDSRJM/p1752771256298149
# if not url.strip():
# raise ScrapingFailedNoUrl()
try:
num_retry += 1