@@ -452,15 +452,17 @@ def new_id(prefix: str) -> str:
452452 # Generic fallback
453453 return {"status" : "mock" , "url" : url , "method" : method , "kwargs" : kwargs }
454454
455- def markdownify (self , website_url : str , headers : Optional [dict [str , str ]] = None , mock :bool = False , stealth :bool = False ):
455+ def markdownify (self , website_url : str , headers : Optional [dict [str , str ]] = None , mock : bool = False , render_heavy_js : bool = False , stealth : bool = False ):
456456 """Send a markdownify request"""
457457 logger .info (f"🔍 Starting markdownify request for { website_url } " )
458458 if headers :
459459 logger .debug ("🔧 Using custom headers" )
460460 if stealth :
461461 logger .debug ("🥷 Stealth mode enabled" )
462+ if render_heavy_js :
463+ logger .debug ("⚡ Heavy JavaScript rendering enabled" )
462464
463- request = MarkdownifyRequest (website_url = website_url , headers = headers , mock = mock , stealth = stealth )
465+ request = MarkdownifyRequest (website_url = website_url , headers = headers , mock = mock , render_heavy_js = render_heavy_js , stealth = stealth )
464466 logger .debug ("✅ Request validation passed" )
465467
466468 result = self ._make_request (
@@ -749,6 +751,8 @@ def crawl(
749751 same_domain_only : bool = True ,
750752 batch_size : Optional [int ] = None ,
751753 sitemap : bool = False ,
754+ headers : Optional [dict [str , str ]] = None ,
755+ render_heavy_js : bool = False ,
752756 stealth : bool = False ,
753757 ):
754758 """Send a crawl request with support for both AI extraction and
@@ -772,6 +776,8 @@ def crawl(
772776 logger .debug (f"🗺️ Use sitemap: { sitemap } " )
773777 if stealth :
774778 logger .debug ("🥷 Stealth mode enabled" )
779+ if render_heavy_js :
780+ logger .debug ("⚡ Heavy JavaScript rendering enabled" )
775781 if batch_size is not None :
776782 logger .debug (f"📦 Batch size: { batch_size } " )
777783
@@ -784,6 +790,7 @@ def crawl(
784790 "max_pages" : max_pages ,
785791 "same_domain_only" : same_domain_only ,
786792 "sitemap" : sitemap ,
793+ "render_heavy_js" : render_heavy_js ,
787794 "stealth" : stealth ,
788795 }
789796
@@ -794,6 +801,8 @@ def crawl(
794801 request_data ["data_schema" ] = data_schema
795802 if batch_size is not None :
796803 request_data ["batch_size" ] = batch_size
804+ if headers is not None :
805+ request_data ["headers" ] = headers
797806
798807 request = CrawlRequest (** request_data )
799808 logger .debug ("✅ Request validation passed" )
0 commit comments