2828from forge .data .datasets .sft_dataset import AlpacaToMessages , sft_iterable_dataset
2929from forge .data .tokenizer import HuggingFaceModelTokenizer
3030from forge .observability import get_or_create_metric_logger , record_metric , Reduce
31- from forge .observability .metric_actors import GlobalLoggingActor
3231from forge .util .config import parse
3332
34- from monarch .actor import current_rank , current_size , endpoint , get_or_spawn_controller
33+ from monarch .actor import current_rank , current_size , endpoint
3534from omegaconf import DictConfig , OmegaConf
3635from torch import nn
3736from torchdata .stateful_dataloader import StatefulDataLoader
@@ -111,12 +110,8 @@ def _init_dist(self):
111110 logger .info ("env: {}" .format (env ))
112111
113112 async def setup_metric_logger (self ):
114- """Retrieve the already-initialized metric logger from main process"""
115-
116- # The global logger was already created in main process.
117- # Use get_or_spawn_controller from monarch to get reference to it
118- # Get reference to the existing global logger (don't create new one)
119- mlogger = await get_or_spawn_controller ("global_logger" , GlobalLoggingActor )
113+ """Initialization happens in the main process. Here we just retrieve it"""
114+ mlogger = await get_or_create_metric_logger ()
120115 return mlogger
121116
122117 def record_batch_metrics (self , data_metrics : list ):
@@ -128,7 +123,6 @@ def record_batch_metrics(self, data_metrics: list):
128123 @endpoint
129124 async def setup (self ):
130125 self .train_dataloader = self .setup_data ()
131-
132126 self .mlogger = await self .setup_metric_logger ()
133127
134128 # self.train_dataloader = self.setup_data(
0 commit comments