From 48de4b56c87e928dcbd26a7cc0670bb1f36a8484 Mon Sep 17 00:00:00 2001 From: jmorganca Date: Mon, 12 Aug 2024 22:19:26 -0700 Subject: [PATCH] cleanup --- llama/runner/runner.go | 6 +----- 1 file changed, 1 insertion(+), 5 deletions(-) diff --git a/llama/runner/runner.go b/llama/runner/runner.go index dcd0f50b..ea129548 100644 --- a/llama/runner/runner.go +++ b/llama/runner/runner.go @@ -220,10 +220,8 @@ func (s *Server) waiting() bool { // processImage processes an image embedding if it's next in any sequence func (s *Server) processImage() bool { - for i, seq := range s.seqs { - fmt.Println("seq", i, "inputs", len(seq.inputs)) + for _, seq := range s.seqs { if len(seq.inputs) > 0 && seq.inputs[0].embd != nil { - slog.Info("processing image", "seq", i, "nPast", seq.nPast) llama.LlavaEvalImageEmbed(s.lc, seq.inputs[0].embd, s.batchSize, &seq.nPast) llama.LlavaImageEmbedFree(seq.inputs[0].embd) seq.iBatch = seq.inputs[0].embd.Tokens() - 1 @@ -244,7 +242,6 @@ func (s *Server) run(ctx context.Context) { case <-ctx.Done(): return default: - slog.Info("Processing batch", "seqs", len(s.seqs)) s.mu.Lock() for s.waiting() { s.cond.Wait() @@ -290,7 +287,6 @@ func (s *Server) run(ctx context.Context) { break } - slog.Info("adding token to batch", "token", t.token, "seq", i) batch.Add(t.token, seq.nPast, []int{i}, !seq.isPromptProcessing()) seq.nPast++ }