diff --git a/common/service/dynamicconfig/config.go b/common/service/dynamicconfig/config.go index e95677724fc..ce5b987a937 100644 --- a/common/service/dynamicconfig/config.go +++ b/common/service/dynamicconfig/config.go @@ -69,7 +69,7 @@ func (c *Collection) logValue( ) { loadedValue, loaded := c.keys.LoadOrStore(key, value) if !loaded || !cmpValueEquals(loadedValue, value) { - c.logger.Info("Get dynamic config", + c.logger.Debug("Get dynamic config", tag.Name(key.String()), tag.Value(value), tag.DefaultValue(defaultValue)) } } diff --git a/service/frontend/workflowHandler.go b/service/frontend/workflowHandler.go index 35e7307d373..91c2f5630c3 100644 --- a/service/frontend/workflowHandler.go +++ b/service/frontend/workflowHandler.go @@ -2549,7 +2549,7 @@ func (wh *WorkflowHandler) ListOpenWorkflowExecutions( WorkflowID: listRequest.ExecutionFilter.GetWorkflowId(), }) } - wh.GetLogger().Info("List open workflow with filter", + wh.GetLogger().Debug("List open workflow with filter", tag.WorkflowDomainName(listRequest.GetDomain()), tag.WorkflowListWorkflowFilterByID) } else if listRequest.TypeFilter != nil { if wh.config.DisableListVisibilityByFilter(domain) { @@ -2560,7 +2560,7 @@ func (wh *WorkflowHandler) ListOpenWorkflowExecutions( WorkflowTypeName: listRequest.TypeFilter.GetName(), }) } - wh.GetLogger().Info("List open workflow with filter", + wh.GetLogger().Debug("List open workflow with filter", tag.WorkflowDomainName(listRequest.GetDomain()), tag.WorkflowListWorkflowFilterByType) } else { persistenceResp, err = wh.GetVisibilityManager().ListOpenWorkflowExecutions(&baseReq) @@ -2762,7 +2762,7 @@ func (wh *WorkflowHandler) ListClosedWorkflowExecutions( WorkflowID: listRequest.ExecutionFilter.GetWorkflowId(), }) } - wh.GetLogger().Info("List closed workflow with filter", + wh.GetLogger().Debug("List closed workflow with filter", tag.WorkflowDomainName(listRequest.GetDomain()), tag.WorkflowListWorkflowFilterByID) } else if listRequest.TypeFilter != nil { if wh.config.DisableListVisibilityByFilter(domain) { @@ -2773,7 +2773,7 @@ func (wh *WorkflowHandler) ListClosedWorkflowExecutions( WorkflowTypeName: listRequest.TypeFilter.GetName(), }) } - wh.GetLogger().Info("List closed workflow with filter", + wh.GetLogger().Debug("List closed workflow with filter", tag.WorkflowDomainName(listRequest.GetDomain()), tag.WorkflowListWorkflowFilterByType) } else if listRequest.StatusFilter != nil { if wh.config.DisableListVisibilityByFilter(domain) { @@ -2784,7 +2784,7 @@ func (wh *WorkflowHandler) ListClosedWorkflowExecutions( Status: listRequest.GetStatusFilter(), }) } - wh.GetLogger().Info("List closed workflow with filter", + wh.GetLogger().Debug("List closed workflow with filter", tag.WorkflowDomainName(listRequest.GetDomain()), tag.WorkflowListWorkflowFilterByStatus) } else { persistenceResp, err = wh.GetVisibilityManager().ListClosedWorkflowExecutions(&baseReq) diff --git a/service/history/failover/coordinator.go b/service/history/failover/coordinator.go index 19af9c97c88..819b28ec840 100644 --- a/service/history/failover/coordinator.go +++ b/service/history/failover/coordinator.go @@ -138,7 +138,7 @@ func (c *coordinatorImpl) Start() { go c.receiveFailoverMarkersLoop() go c.notifyFailoverMarkerLoop() - c.logger.Info("", tag.LifeCycleStarted) + c.logger.Info("Coordinator state changed", tag.LifeCycleStarted) } func (c *coordinatorImpl) Stop() { @@ -152,7 +152,7 @@ func (c *coordinatorImpl) Stop() { } close(c.shutdownChan) - c.logger.Info("", tag.LifeCycleStopped) + c.logger.Info("Coordinator state changed", tag.LifeCycleStopped) } func (c *coordinatorImpl) NotifyFailoverMarkers( diff --git a/service/history/failover/marker_notifier.go b/service/history/failover/marker_notifier.go index 38f0c3bba27..b041cb85267 100644 --- a/service/history/failover/marker_notifier.go +++ b/service/history/failover/marker_notifier.go @@ -79,7 +79,7 @@ func (m *markerNotifierImpl) Start() { } go m.notifyPendingFailoverMarker() - m.logger.Info("", tag.LifeCycleStarted) + m.logger.Info("Marker notifier state changed", tag.LifeCycleStarted) } func (m *markerNotifierImpl) Stop() { @@ -92,7 +92,7 @@ func (m *markerNotifierImpl) Stop() { return } close(m.shutdownCh) - m.logger.Info("", tag.LifeCycleStopped) + m.logger.Info("Marker notifier state changed", tag.LifeCycleStopped) } func (m *markerNotifierImpl) notifyPendingFailoverMarker() { diff --git a/service/history/historyEngine.go b/service/history/historyEngine.go index b6864b7272d..c280ef71c8d 100644 --- a/service/history/historyEngine.go +++ b/service/history/historyEngine.go @@ -384,8 +384,8 @@ func NewEngineWithShardContext( // Make sure all the components are loaded lazily so start can return immediately. This is important because // ShardController calls start sequentially for all the shards for a given host during startup. func (e *historyEngineImpl) Start() { - e.logger.Info("", tag.LifeCycleStarting) - defer e.logger.Info("", tag.LifeCycleStarted) + e.logger.Info("History engine state changed", tag.LifeCycleStarting) + defer e.logger.Info("History engine state changed", tag.LifeCycleStarted) e.registerDomainFailoverCallback() @@ -407,8 +407,8 @@ func (e *historyEngineImpl) Start() { // Stop the service. func (e *historyEngineImpl) Stop() { - e.logger.Info("", tag.LifeCycleStopping) - defer e.logger.Info("", tag.LifeCycleStopped) + e.logger.Info("History engine state changed", tag.LifeCycleStopping) + defer e.logger.Info("History engine state changed", tag.LifeCycleStopped) e.txProcessor.Stop() e.timerProcessor.Stop() @@ -1327,7 +1327,7 @@ func (e *historyEngineImpl) queryDirectlyThroughMatching( return nil, err } - e.logger.Info("query directly through matching on sticky timed out, attempting to query on non-sticky", + e.logger.Debug("query directly through matching on sticky timed out, attempting to query on non-sticky", tag.WorkflowDomainName(queryRequest.GetDomain()), tag.WorkflowID(queryRequest.Execution.GetWorkflowId()), tag.WorkflowRunID(queryRequest.Execution.GetRunId()), diff --git a/service/history/queue/timer_queue_processor_base.go b/service/history/queue/timer_queue_processor_base.go index 87554c16303..3a86872e4d6 100644 --- a/service/history/queue/timer_queue_processor_base.go +++ b/service/history/queue/timer_queue_processor_base.go @@ -148,8 +148,8 @@ func (t *timerQueueProcessorBase) Start() { return } - t.logger.Info("", tag.LifeCycleStarting) - defer t.logger.Info("", tag.LifeCycleStarted) + t.logger.Info("Timer queue processor state changed", tag.LifeCycleStarting) + defer t.logger.Info("Timer queue processor state changed", tag.LifeCycleStarted) t.redispatcher.Start() @@ -166,8 +166,8 @@ func (t *timerQueueProcessorBase) Stop() { return } - t.logger.Info("", tag.LifeCycleStopping) - defer t.logger.Info("", tag.LifeCycleStopped) + t.logger.Info("Timer queue processor state changed", tag.LifeCycleStopping) + defer t.logger.Info("Timer queue processor state changed", tag.LifeCycleStopped) t.timerGate.Close() close(t.shutdownCh) diff --git a/service/history/queue/transfer_queue_processor_base.go b/service/history/queue/transfer_queue_processor_base.go index c795a33150a..b37d12320c2 100644 --- a/service/history/queue/transfer_queue_processor_base.go +++ b/service/history/queue/transfer_queue_processor_base.go @@ -129,8 +129,8 @@ func (t *transferQueueProcessorBase) Start() { return } - t.logger.Info("", tag.LifeCycleStarting) - defer t.logger.Info("", tag.LifeCycleStarted) + t.logger.Info("Transfer queue processor state changed", tag.LifeCycleStarting) + defer t.logger.Info("Transfer queue processor state changed", tag.LifeCycleStarted) t.redispatcher.Start() @@ -147,8 +147,8 @@ func (t *transferQueueProcessorBase) Stop() { return } - t.logger.Info("", tag.LifeCycleStopping) - defer t.logger.Info("", tag.LifeCycleStopped) + t.logger.Info("Transfer queue processor state changed", tag.LifeCycleStopping) + defer t.logger.Info("Transfer queue processor state changed", tag.LifeCycleStopped) t.nextPollTimer.Close() close(t.shutdownCh) diff --git a/service/history/queueProcessor.go b/service/history/queueProcessor.go index cfbb53c5443..3ab75841f04 100644 --- a/service/history/queueProcessor.go +++ b/service/history/queueProcessor.go @@ -167,8 +167,8 @@ func (p *queueProcessorBase) Start() { return } - p.logger.Info("", tag.LifeCycleStarting) - defer p.logger.Info("", tag.LifeCycleStarted) + p.logger.Info("Queue processor state changed", tag.LifeCycleStarting) + defer p.logger.Info("Queue processor state changed", tag.LifeCycleStarted) if p.taskProcessor != nil { p.taskProcessor.start() @@ -186,8 +186,8 @@ func (p *queueProcessorBase) Stop() { return } - p.logger.Info("", tag.LifeCycleStopping) - defer p.logger.Info("", tag.LifeCycleStopped) + p.logger.Info("Queue processor state changed", tag.LifeCycleStopping) + defer p.logger.Info("Queue processor state changed", tag.LifeCycleStopped) close(p.shutdownCh) p.retryTasks() @@ -271,7 +271,7 @@ processorPumpLoop: } } - p.logger.Info("Queue processor pump shut down.") + p.logger.Debug("Queue processor pump shut down.") } func (p *queueProcessorBase) processBatch() { diff --git a/service/history/replication/task_processor.go b/service/history/replication/task_processor.go index a0969254e40..6b86ab0062e 100644 --- a/service/history/replication/task_processor.go +++ b/service/history/replication/task_processor.go @@ -168,13 +168,13 @@ func (p *taskProcessorImpl) Stop() { return } - p.logger.Info("ReplicationTaskProcessor shutting down.") + p.logger.Debug("ReplicationTaskProcessor shutting down.") close(p.done) } func (p *taskProcessorImpl) processorLoop() { defer func() { - p.logger.Info("Closing replication task processor.", tag.ReadLevel(p.lastRetrievedMessageID)) + p.logger.Debug("Closing replication task processor.", tag.ReadLevel(p.lastRetrievedMessageID)) }() Loop: @@ -182,7 +182,7 @@ Loop: // for each iteration, do close check first select { case <-p.done: - p.logger.Info("ReplicationTaskProcessor shutting down.") + p.logger.Debug("ReplicationTaskProcessor shutting down.") return default: } @@ -580,7 +580,7 @@ func (p *taskProcessorImpl) shouldRetryDLQ(err error) bool { select { case <-p.done: - p.logger.Info("ReplicationTaskProcessor shutting down.") + p.logger.Debug("ReplicationTaskProcessor shutting down.") return false default: return true diff --git a/service/history/shard/controller.go b/service/history/shard/controller.go index d589626a524..e1e2787cce6 100644 --- a/service/history/shard/controller.go +++ b/service/history/shard/controller.go @@ -163,7 +163,7 @@ func (c *controller) Start() { c.logger.Error("Error adding listener", tag.Error(err)) } - c.logger.Info("", tag.LifeCycleStarted) + c.logger.Info("Shard controller state changed", tag.LifeCycleStarted) } func (c *controller) Stop() { @@ -182,7 +182,7 @@ func (c *controller) Stop() { c.logger.Warn("", tag.LifeCycleStopTimedout) } - c.logger.Info("", tag.LifeCycleStopped) + c.logger.Info("Shard controller state changed", tag.LifeCycleStopped) } func (c *controller) PrepareToStop() { @@ -250,7 +250,7 @@ func (c *controller) removeEngineForShard(shardID int, shardItem *historyShardsI func (c *controller) shardClosedCallback(shardID int, shardItem *historyShardsItem) { c.metricsScope.IncCounter(metrics.ShardClosedCounter) - c.logger.Info("", tag.LifeCycleStopping, tag.ComponentShard, tag.ShardID(shardID)) + c.logger.Info("Shard controller state changed", tag.LifeCycleStopping, tag.ComponentShard, tag.ShardID(shardID)) c.removeEngineForShard(shardID, shardItem) } @@ -296,7 +296,7 @@ func (c *controller) getOrCreateHistoryShardItem(shardID int) (*historyShardsIte c.historyShards[shardID] = shardItem c.metricsScope.IncCounter(metrics.ShardItemCreatedCounter) - shardItem.logger.Info("", tag.LifeCycleStarted, tag.ComponentShardItem) + shardItem.logger.Info("Shard item state changed", tag.LifeCycleStarted, tag.ComponentShardItem) return shardItem, nil } @@ -323,7 +323,7 @@ func (c *controller) removeHistoryShardItem(shardID int, shardItem *historyShard c.metricsScope.IncCounter(metrics.ShardItemRemovedCounter) - currentShardItem.logger.Info("", tag.LifeCycleStopped, tag.ComponentShardItem, tag.Number(int64(nShards))) + currentShardItem.logger.Info("Shard item state changed", tag.LifeCycleStopped, tag.ComponentShardItem, tag.Number(int64(nShards))) return currentShardItem, nil } @@ -352,7 +352,7 @@ func (c *controller) shardManagementPump() { case changedEvent := <-c.membershipUpdateCh: c.metricsScope.IncCounter(metrics.MembershipChangedCounter) - c.logger.Info("", tag.ValueRingMembershipChangedEvent, + c.logger.Info("Ring membership changed", tag.ValueRingMembershipChangedEvent, tag.NumberProcessed(len(changedEvent.HostsAdded)), tag.NumberDeleted(len(changedEvent.HostsRemoved)), tag.Number(int64(len(changedEvent.HostsUpdated)))) @@ -408,7 +408,7 @@ func (c *controller) acquireShards() { } func (c *controller) doShutdown() { - c.logger.Info("", tag.LifeCycleStopping) + c.logger.Info("Shard controller state changed", tag.LifeCycleStopping) c.Lock() defer c.Unlock() for _, item := range c.historyShards { @@ -435,12 +435,12 @@ func (i *historyShardsItem) getOrCreateEngine( defer i.Unlock() switch i.status { case historyShardsItemStatusInitialized: - i.logger.Info("", tag.LifeCycleStarting, tag.ComponentShardEngine) + i.logger.Info("Shard engine state changed", tag.LifeCycleStarting, tag.ComponentShardEngine) context, err := acquireShard(i, closeCallback) if err != nil { // invalidate the shardItem so that the same shardItem won't be // used to create another shardContext - i.logger.Info("", tag.LifeCycleStopped, tag.ComponentShardEngine) + i.logger.Info("Shard engine state changed", tag.LifeCycleStopped, tag.ComponentShardEngine) i.status = historyShardsItemStatusStopped return nil, err } @@ -450,7 +450,7 @@ func (i *historyShardsItem) getOrCreateEngine( } i.engine = i.engineFactory.CreateEngine(context) i.engine.Start() - i.logger.Info("", tag.LifeCycleStarted, tag.ComponentShardEngine) + i.logger.Info("Shard engine state changed", tag.LifeCycleStarted, tag.ComponentShardEngine) i.status = historyShardsItemStatusStarted return i.engine, nil case historyShardsItemStatusStarted: @@ -470,10 +470,10 @@ func (i *historyShardsItem) stopEngine() { case historyShardsItemStatusInitialized: i.status = historyShardsItemStatusStopped case historyShardsItemStatusStarted: - i.logger.Info("", tag.LifeCycleStopping, tag.ComponentShardEngine) + i.logger.Info("Shard engine state changed", tag.LifeCycleStopping, tag.ComponentShardEngine) i.engine.Stop() i.engine = nil - i.logger.Info("", tag.LifeCycleStopped, tag.ComponentShardEngine) + i.logger.Info("Shard engine state changed", tag.LifeCycleStopped, tag.ComponentShardEngine) i.status = historyShardsItemStatusStopped case historyShardsItemStatusStopped: // no op diff --git a/service/history/taskProcessor.go b/service/history/taskProcessor.go index 36cf9c47388..27d3d155077 100644 --- a/service/history/taskProcessor.go +++ b/service/history/taskProcessor.go @@ -137,7 +137,7 @@ func (t *taskProcessor) stop() { if success := common.AwaitWaitGroup(&t.workerWG, time.Minute); !success { t.logger.Warn("Task processor timed out on shutdown.") } - t.logger.Info("Task processor shutdown.") + t.logger.Debug("Task processor shutdown.") } func (t *taskProcessor) taskWorker( diff --git a/service/history/timerQueueProcessorBase.go b/service/history/timerQueueProcessorBase.go index 99b0f45e32f..0ee69d00343 100644 --- a/service/history/timerQueueProcessorBase.go +++ b/service/history/timerQueueProcessorBase.go @@ -199,7 +199,7 @@ func (t *timerQueueProcessorBase) Stop() { } t.redispatcher.Stop() - t.logger.Info("Timer queue processor stopped.") + t.logger.Debug("Timer queue processor stopped.") } func (t *timerQueueProcessorBase) processorPump() { @@ -218,7 +218,7 @@ RetryProcessor: } } - t.logger.Info("Timer queue processor pump shutting down.") + t.logger.Debug("Timer queue processor pump shutting down.") } // NotifyNewTimers - Notify the processor about the new timer events arrival. diff --git a/service/matching/matchingEngine.go b/service/matching/matchingEngine.go index fa0a02474f9..0f098ae10ed 100644 --- a/service/matching/matchingEngine.go +++ b/service/matching/matchingEngine.go @@ -179,21 +179,21 @@ func (e *matchingEngineImpl) getTaskListManager(taskList *taskListID, e.taskListsLock.Unlock() return result, nil } - e.logger.Info("", tag.LifeCycleStarting, tag.WorkflowTaskListName(taskList.name), tag.WorkflowTaskListType(taskList.taskType)) + e.logger.Info("Task list manager state changed", tag.LifeCycleStarting, tag.WorkflowTaskListName(taskList.name), tag.WorkflowTaskListType(taskList.taskType)) mgr, err := newTaskListManager(e, taskList, taskListKind, e.config) if err != nil { e.taskListsLock.Unlock() - e.logger.Info("", tag.LifeCycleStartFailed, tag.WorkflowTaskListName(taskList.name), tag.WorkflowTaskListType(taskList.taskType), tag.Error(err)) + e.logger.Info("Task list manager state changed", tag.LifeCycleStartFailed, tag.WorkflowTaskListName(taskList.name), tag.WorkflowTaskListType(taskList.taskType), tag.Error(err)) return nil, err } e.taskLists[*taskList] = mgr e.taskListsLock.Unlock() err = mgr.Start() if err != nil { - e.logger.Info("", tag.LifeCycleStartFailed, tag.WorkflowTaskListName(taskList.name), tag.WorkflowTaskListType(taskList.taskType), tag.Error(err)) + e.logger.Info("Task list manager state changed", tag.LifeCycleStartFailed, tag.WorkflowTaskListName(taskList.name), tag.WorkflowTaskListType(taskList.taskType), tag.Error(err)) return nil, err } - e.logger.Info("", tag.LifeCycleStarted, tag.WorkflowTaskListName(taskList.name), tag.WorkflowTaskListType(taskList.taskType)) + e.logger.Info("Task list manager state changed", tag.LifeCycleStarted, tag.WorkflowTaskListName(taskList.name), tag.WorkflowTaskListType(taskList.taskType)) return mgr, nil } diff --git a/service/matching/taskListManager.go b/service/matching/taskListManager.go index e87e0100961..1e6223098ae 100644 --- a/service/matching/taskListManager.go +++ b/service/matching/taskListManager.go @@ -205,7 +205,7 @@ func (c *taskListManagerImpl) Stop() { c.taskWriter.Stop() c.taskReader.Stop() c.engine.removeTaskListManager(c.taskListID) - c.logger.Info("", tag.LifeCycleStopped) + c.logger.Info("Task list manager state changed", tag.LifeCycleStopped) } // AddTask adds a task to the task list. This method will first attempt a synchronous diff --git a/service/worker/indexer/processor.go b/service/worker/indexer/processor.go index 43e1e154e32..fa8792f3286 100644 --- a/service/worker/indexer/processor.go +++ b/service/worker/indexer/processor.go @@ -93,21 +93,21 @@ func (p *indexProcessor) Start() error { return nil } - p.logger.Info("", tag.LifeCycleStarting) + p.logger.Info("Index processor state changed", tag.LifeCycleStarting) consumer, err := p.kafkaClient.NewConsumer(p.appName, p.consumerName, p.config.IndexerConcurrency()) if err != nil { - p.logger.Info("", tag.LifeCycleStartFailed, tag.Error(err)) + p.logger.Info("Index processor state changed", tag.LifeCycleStartFailed, tag.Error(err)) return err } if err := consumer.Start(); err != nil { - p.logger.Info("", tag.LifeCycleStartFailed, tag.Error(err)) + p.logger.Info("Index processor state changed", tag.LifeCycleStartFailed, tag.Error(err)) return err } esProcessor, err := NewESProcessorAndStart(p.config, p.esClient, p.esProcessorName, p.logger, p.metricsClient, p.msgEncoder) if err != nil { - p.logger.Info("", tag.LifeCycleStartFailed, tag.Error(err)) + p.logger.Info("Index processor state changed", tag.LifeCycleStartFailed, tag.Error(err)) return err } @@ -116,7 +116,7 @@ func (p *indexProcessor) Start() error { p.shutdownWG.Add(1) go p.processorPump() - p.logger.Info("", tag.LifeCycleStarted) + p.logger.Info("Index processor state changed", tag.LifeCycleStarted) return nil } @@ -125,15 +125,15 @@ func (p *indexProcessor) Stop() { return } - p.logger.Info("", tag.LifeCycleStopping) - defer p.logger.Info("", tag.LifeCycleStopped) + p.logger.Info("Index processor state changed", tag.LifeCycleStopping) + defer p.logger.Info("Index processor state changed", tag.LifeCycleStopped) if atomic.LoadInt32(&p.isStarted) == 1 { close(p.shutdownCh) } if success := common.AwaitWaitGroup(&p.shutdownWG, time.Minute); !success { - p.logger.Info("", tag.LifeCycleStopTimedout) + p.logger.Info("Index processor state changed", tag.LifeCycleStopTimedout) } } diff --git a/service/worker/replicator/domainReplicationMessageProcessor.go b/service/worker/replicator/domainReplicationMessageProcessor.go index 46c5dc8b99c..37a6b389465 100644 --- a/service/worker/replicator/domainReplicationMessageProcessor.go +++ b/service/worker/replicator/domainReplicationMessageProcessor.go @@ -133,7 +133,7 @@ func (p *domainReplicationMessageProcessor) getAndHandleDomainReplicationTasks() } if info.Identity() != p.hostInfo.Identity() { - p.logger.Info(fmt.Sprintf("Worker not responsible for source cluster %v.", p.sourceCluster)) + p.logger.Debug(fmt.Sprintf("Worker not responsible for source cluster %v.", p.sourceCluster)) return } diff --git a/service/worker/replicator/processor.go b/service/worker/replicator/processor.go index e707a6fa4db..350311ffbff 100644 --- a/service/worker/replicator/processor.go +++ b/service/worker/replicator/processor.go @@ -125,15 +125,15 @@ func (p *replicationTaskProcessor) Start() error { return nil } - p.logger.Info("", tag.LifeCycleStarting, tag.ComponentReplicationTaskProcessor) + p.logger.Info("Replication task processor state changed", tag.LifeCycleStarting, tag.ComponentReplicationTaskProcessor) consumer, err := p.client.NewConsumerWithClusterName(p.currentCluster, p.sourceCluster, p.consumerName, p.config.ReplicatorMessageConcurrency()) if err != nil { - p.logger.Info("", tag.LifeCycleStartFailed, tag.ComponentReplicationTaskProcessor, tag.Error(err)) + p.logger.Info("Replication task processor state changed", tag.LifeCycleStartFailed, tag.ComponentReplicationTaskProcessor, tag.Error(err)) return err } if err := consumer.Start(); err != nil { - p.logger.Info("", tag.LifeCycleStartFailed, tag.ComponentReplicationTaskProcessor, tag.Error(err)) + p.logger.Info("Replication task processor state changed", tag.LifeCycleStartFailed, tag.ComponentReplicationTaskProcessor, tag.Error(err)) return err } @@ -142,7 +142,7 @@ func (p *replicationTaskProcessor) Start() error { go p.processorPump() p.sequentialTaskProcessor.Start() - p.logger.Info("", tag.LifeCycleStarted, tag.ComponentReplicationTaskProcessor) + p.logger.Info("Replication task processor state changed", tag.LifeCycleStarted, tag.ComponentReplicationTaskProcessor) return nil } @@ -152,15 +152,15 @@ func (p *replicationTaskProcessor) Stop() { } p.sequentialTaskProcessor.Stop() - p.logger.Info("", tag.LifeCycleStopping, tag.ComponentReplicationTaskProcessor) - defer p.logger.Info("", tag.LifeCycleStopped, tag.ComponentReplicationTaskProcessor) + p.logger.Info("Replication task processor state changed", tag.LifeCycleStopping, tag.ComponentReplicationTaskProcessor) + defer p.logger.Info("Replication task processor state changed", tag.LifeCycleStopped, tag.ComponentReplicationTaskProcessor) if atomic.LoadInt32(&p.isStarted) == 1 { close(p.shutdownCh) } if success := common.AwaitWaitGroup(&p.shutdownWG, time.Minute); !success { - p.logger.Info("", tag.LifeCycleStopTimedout, tag.ComponentReplicationTaskProcessor) + p.logger.Info("Replication task processor state changed", tag.LifeCycleStopTimedout, tag.ComponentReplicationTaskProcessor) } }