@@ -305,11 +305,26 @@ async def process_query(
305305 context = ingest_query (query )
306306 digest = context .generate_digest ()
307307
308- summary , tree , content = context .generate_digest ()
308+ # Store digest based on S3 configuration
309+ if is_s3_enabled ():
310+ # Upload to S3 instead of storing locally
311+ s3_file_path = generate_s3_file_path (
312+ source = query .url ,
313+ user_name = cast ("str" , query .user_name ),
314+ repo_name = cast ("str" , query .repo_name ),
315+ commit = query .commit ,
316+ include_patterns = query .include_patterns ,
317+ ignore_patterns = query .ignore_patterns ,
318+ )
319+ s3_url = upload_to_s3 (content = context .digest , s3_file_path = s3_file_path , ingest_id = query .id )
320+ # Store S3 URL in query for later use
321+ query .s3_url = s3_url
322+ else :
323+ # Store locally
324+ local_txt_file = Path (clone_config .local_path ).with_suffix (".txt" )
325+ with local_txt_file .open ("w" , encoding = "utf-8" ) as f :
326+ f .write (digest )
309327
310- # Prepare the digest content (tree + content)
311- digest_content = tree + "\n " + content
312- _store_digest_content (query , clone_config , digest_content , summary , tree , content )
313328 except Exception as exc :
314329 _print_error (query .url , exc , max_file_size , pattern_type , pattern )
315330 # Clean up repository even if processing failed
0 commit comments