@@ -55,7 +55,7 @@ def analyze_local_repository(
5555 Dict with analysis results including nodes and relationships
5656 """
5757 try :
58- logger .info (f"Analyzing local repository at { repo_path } " )
58+ logger .debug (f"Analyzing local repository at { repo_path } " )
5959
6060 # Get repo analyzer to find files
6161 repo_analyzer = RepoAnalyzer ()
@@ -71,9 +71,9 @@ def analyze_local_repository(
7171 # Limit number of files
7272 if len (code_files ) > max_files :
7373 code_files = code_files [:max_files ]
74- logger .info (f"Limited analysis to { max_files } files" )
74+ logger .debug (f"Limited analysis to { max_files } files" )
7575
76- logger .info (f"Analyzing { len (code_files )} files" )
76+ logger .debug (f"Analyzing { len (code_files )} files" )
7777
7878 # Analyze files
7979 result = self .call_graph_analyzer .analyze_code_files (code_files , repo_path )
@@ -115,18 +115,18 @@ def analyze_repository_full(
115115 """
116116 temp_dir = None
117117 try :
118- logger .info (f"Starting full analysis of { github_url } " )
118+ logger .debug (f"Starting full analysis of { github_url } " )
119119
120120 temp_dir = self ._clone_repository (github_url )
121121 repo_info = self ._parse_repository_info (github_url )
122122
123- logger .info ("Analyzing repository file structure..." )
123+ logger .debug ("Analyzing repository file structure..." )
124124 structure_result = self ._analyze_structure (temp_dir , include_patterns , exclude_patterns )
125- logger .info (f"Found { structure_result ['summary' ]['total_files' ]} files to analyze." )
125+ logger .debug (f"Found { structure_result ['summary' ]['total_files' ]} files to analyze." )
126126
127- logger .info ("Starting call graph analysis..." )
127+ logger .debug ("Starting call graph analysis..." )
128128 call_graph_result = self ._analyze_call_graph (structure_result ["file_tree" ], temp_dir )
129- logger .info (
129+ logger .debug (
130130 f"Call graph analysis complete. Found { call_graph_result ['call_graph' ]['total_functions' ]} functions."
131131 )
132132
@@ -152,10 +152,10 @@ def analyze_repository_full(
152152 readme_content = readme_content ,
153153 )
154154
155- logger .info (f"Cleaning up temporary repository directory: { temp_dir } " )
155+ logger .debug (f"Cleaning up temporary repository directory: { temp_dir } " )
156156 self ._cleanup_repository (temp_dir )
157157
158- logger .info (
158+ logger .debug (
159159 f"Analysis completed: { analysis_result .summary ['total_functions' ]} functions found"
160160 )
161161 return analysis_result
@@ -185,7 +185,7 @@ def analyze_repository_structure_only(
185185 """
186186 temp_dir = None
187187 try :
188- logger .info (f"Starting structure analysis of { github_url } " )
188+ logger .debug (f"Starting structure analysis of { github_url } " )
189189
190190 temp_dir = self ._clone_repository (github_url )
191191 repo_info = self ._parse_repository_info (github_url )
@@ -203,7 +203,7 @@ def analyze_repository_structure_only(
203203
204204 self ._cleanup_repository (temp_dir )
205205
206- logger .info (
206+ logger .debug (
207207 f"Structure analysis completed: { result ['file_summary' ]['total_files' ]} files found"
208208 )
209209 return result
@@ -216,9 +216,9 @@ def analyze_repository_structure_only(
216216
217217 def _clone_repository (self , github_url : str ) -> str :
218218 """Clone repository and return temp dir path."""
219- logger .info (f"Cloning { github_url } ..." )
219+ logger .debug (f"Cloning { github_url } ..." )
220220 temp_dir = clone_repository (github_url )
221- logger .info (f"Repository cloned to { temp_dir } " )
221+ logger .debug (f"Repository cloned to { temp_dir } " )
222222 self ._temp_directories .append (temp_dir )
223223 return temp_dir
224224
@@ -233,7 +233,7 @@ def _analyze_structure(
233233 exclude_patterns : Optional [List [str ]],
234234 ) -> Dict [str , Any ]:
235235 """Analyze repository file structure with filtering."""
236- logger .info (
236+ logger .debug (
237237 f"Initializing RepoAnalyzer with include: { include_patterns } , exclude: { exclude_patterns } "
238238 )
239239 repo_analyzer = RepoAnalyzer (include_patterns , exclude_patterns )
@@ -246,12 +246,12 @@ def _read_readme_file(self, repo_dir: str) -> Optional[str]:
246246 # readme_path = Path(repo_dir) / name
247247 # if readme_path.exists():
248248 # try:
249- # logger.info (f"Found README file at {readme_path}")
249+ # logger.debug (f"Found README file at {readme_path}")
250250 # return readme_path.read_text(encoding="utf-8")
251251 # except Exception as e:
252252 # logger.warning(f"Could not read README file at {readme_path}: {e}")
253253 # return None
254- # logger.info ("No README file found in repository root.")
254+ # logger.debug ("No README file found in repository root.")
255255 # return None
256256 base = Path (repo_dir )
257257 possible_readme_names = ["README.md" , "README" , "readme.md" , "README.txt" ]
@@ -260,12 +260,12 @@ def _read_readme_file(self, repo_dir: str) -> Optional[str]:
260260 if p .exists ():
261261 try :
262262 assert_safe_path (base , p )
263- logger .info (f"Found README file at { p } " )
263+ logger .debug (f"Found README file at { p } " )
264264 return safe_open_text (base , p , encoding = "utf-8" )
265265 except Exception as e :
266266 logger .warning (f"Skipping unsafe/ unreadable README at { p } : { e } " )
267267 return None
268- logger .info ("No README file found in repository root." )
268+ logger .debug ("No README file found in repository root." )
269269 return None
270270
271271 def _analyze_call_graph (self , file_tree : Dict [str , Any ], repo_dir : str ) -> Dict [str , Any ]:
@@ -277,12 +277,12 @@ def _analyze_call_graph(self, file_tree: Dict[str, Any], repo_dir: str) -> Dict[
277277 - JavaScript/TypeScript AST analysis (planned)
278278 - Additional language support (future)
279279 """
280- logger .info ("Extracting code files from file tree..." )
280+ logger .debug ("Extracting code files from file tree..." )
281281 code_files = self .call_graph_analyzer .extract_code_files (file_tree )
282282
283- logger .info (f"Found { len (code_files )} total code files. Filtering for supported languages." )
283+ logger .debug (f"Found { len (code_files )} total code files. Filtering for supported languages." )
284284 supported_files = self ._filter_supported_languages (code_files )
285- logger .info (f"Analyzing { len (supported_files )} supported files." )
285+ logger .debug (f"Analyzing { len (supported_files )} supported files." )
286286
287287 result = self .call_graph_analyzer .analyze_code_files (supported_files , repo_dir )
288288
@@ -321,7 +321,7 @@ def _get_supported_languages(self) -> List[str]:
321321
322322 def _cleanup_repository (self , temp_dir : str ):
323323 """Clean up cloned repository."""
324- logger .info (f"Attempting to clean up { temp_dir } " )
324+ logger .debug (f"Attempting to clean up { temp_dir } " )
325325 cleanup_repository (temp_dir )
326326 if temp_dir in self ._temp_directories :
327327 self ._temp_directories .remove (temp_dir )
0 commit comments