@@ -102,7 +102,7 @@ def _get_tree_data(self):
102102 "value" : node .value ,
103103 "visits" : node .visits ,
104104 "feedback" : node .feedback ,
105- "reward" : node .reward
105+ # "reward": node.reward
106106 }
107107 tree_data .append (node_data )
108108
@@ -129,7 +129,7 @@ async def remove_simulated_trajectory(self, starting_node, terminal_node: LATSNo
129129 "description" : node .natural_language_description ,
130130 "visits" : node .visits ,
131131 "value" : float (f"{ node .value :.3f} " ) if hasattr (node , 'value' ) else None ,
132- "reward" : float (f"{ node .reward :.3f} " ) if hasattr (node , 'reward' ) else None ,
132+ # "reward": float(f"{node.reward:.3f}") if hasattr(node, 'reward') else None,
133133 "is_terminal" : node .is_terminal ,
134134 "feedback" : node .feedback if hasattr (node , 'feedback' ) else None ,
135135 "is_root" : not hasattr (node , 'parent' ) or node .parent is None ,
@@ -159,7 +159,7 @@ def _get_trajectory_data(self, terminal_node: LATSNode):
159159 "description" : node .natural_language_description ,
160160 "visits" : node .visits ,
161161 "value" : float (f"{ node .value :.3f} " ) if hasattr (node , 'value' ) else None ,
162- "reward" : float (f"{ node .reward :.3f} " ) if hasattr (node , 'reward' ) else None ,
162+ # "reward": float(f"{node.reward:.3f}") if hasattr(node, 'reward') else None,
163163 "is_terminal" : node .is_terminal ,
164164 "feedback" : node .feedback if hasattr (node , 'feedback' ) else None ,
165165 "is_root" : not hasattr (node , 'parent' ) or node .parent is None ,
@@ -424,15 +424,18 @@ async def node_children_evaluation(self, node: LATSNode) -> None:
424424 score = 0
425425 else :
426426 trajectory = child .get_trajectory ()
427- prompt = create_llm_prompt (trajectory , self .goal )
428- # , child.observation.image
429- result = score_trajectory_with_openai (prompt , openai_client , self .config .evaluation_model )
430- score = result ["overall_score" ]
427+ if len (trajectory ) == 0 :
428+ score = 0
429+ else :
430+ prompt = create_llm_prompt (trajectory , self .goal )
431+ # , child.observation.image
432+ result = score_trajectory_with_openai (prompt , openai_client , self .config .evaluation_model )
433+ score = result ["overall_score" ]
431434 scores .append (score )
432435
433436 for child , score in zip (node .children , scores ):
434437 child .value = score
435- child .reward = score
438+ # child.reward = score
436439
437440 async def node_evaluation (self , node : LATSNode ) -> None :
438441 """Evaluate the current node and assign its score."""
@@ -454,13 +457,16 @@ async def node_evaluation(self, node: LATSNode) -> None:
454457 if node .is_terminal :
455458 score = 0
456459 else :
457- prompt = create_llm_prompt (trajectory , self .goal )
458- result = score_trajectory_with_openai (
459- prompt ,
460- openai_client ,
461- model = self .config .evaluation_model
462- )
463- score = result ["overall_score" ]
460+ if len (trajectory ) == 0 :
461+ score = 0
462+ else :
463+ prompt = create_llm_prompt (trajectory , self .goal )
464+ result = score_trajectory_with_openai (
465+ prompt ,
466+ openai_client ,
467+ model = self .config .evaluation_model
468+ )
469+ score = result ["overall_score" ]
464470
465471 except Exception as e :
466472 error_msg = f"Error scoring node { id (node )} : { str (e )} "
@@ -469,7 +475,7 @@ async def node_evaluation(self, node: LATSNode) -> None:
469475
470476 # Assign the score to the node
471477 node .value = score
472- node .reward = score
478+ # node.reward = score
473479
474480
475481 except Exception as e :
0 commit comments