This commit is contained in:
jmorganca 2024-08-12 22:19:26 -07:00
parent cd776e49ad
commit 48de4b56c8

View File

@ -220,10 +220,8 @@ func (s *Server) waiting() bool {
// processImage processes an image embedding if it's next in any sequence
func (s *Server) processImage() bool {
for i, seq := range s.seqs {
fmt.Println("seq", i, "inputs", len(seq.inputs))
for _, seq := range s.seqs {
if len(seq.inputs) > 0 && seq.inputs[0].embd != nil {
slog.Info("processing image", "seq", i, "nPast", seq.nPast)
llama.LlavaEvalImageEmbed(s.lc, seq.inputs[0].embd, s.batchSize, &seq.nPast)
llama.LlavaImageEmbedFree(seq.inputs[0].embd)
seq.iBatch = seq.inputs[0].embd.Tokens() - 1
@ -244,7 +242,6 @@ func (s *Server) run(ctx context.Context) {
case <-ctx.Done():
return
default:
slog.Info("Processing batch", "seqs", len(s.seqs))
s.mu.Lock()
for s.waiting() {
s.cond.Wait()
@ -290,7 +287,6 @@ func (s *Server) run(ctx context.Context) {
break
}
slog.Info("adding token to batch", "token", t.token, "seq", i)
batch.Add(t.token, seq.nPast, []int{i}, !seq.isPromptProcessing())
seq.nPast++
}