allow empty urls (#2984)
This commit is contained in:
@@ -13,7 +13,7 @@ from pydantic import BaseModel, PrivateAttr
|
||||
|
||||
from skyvern.config import settings
|
||||
from skyvern.constants import BUILDING_ELEMENT_TREE_TIMEOUT_MS, DEFAULT_MAX_TOKENS, SKYVERN_DIR, SKYVERN_ID_ATTR
|
||||
from skyvern.exceptions import FailedToTakeScreenshot, ScrapingFailed, ScrapingFailedNoUrl, UnknownElementTreeFormat
|
||||
from skyvern.exceptions import FailedToTakeScreenshot, ScrapingFailed, UnknownElementTreeFormat
|
||||
from skyvern.forge.sdk.api.crypto import calculate_sha256
|
||||
from skyvern.forge.sdk.core import skyvern_context
|
||||
from skyvern.forge.sdk.trace import TraceManager
|
||||
@@ -427,8 +427,10 @@ async def scrape_website(
|
||||
:raises Exception: When scraping fails after maximum retries.
|
||||
"""
|
||||
|
||||
if not url.strip():
|
||||
raise ScrapingFailedNoUrl()
|
||||
# TODO(jdo) why is this a problem?
|
||||
# ref: https://skyvern.slack.com/archives/C074UNDSRJM/p1752771256298149
|
||||
# if not url.strip():
|
||||
# raise ScrapingFailedNoUrl()
|
||||
|
||||
try:
|
||||
num_retry += 1
|
||||
|
||||
Reference in New Issue
Block a user