I am using a code-completion model for my tool I am making for godot (will be open sourced very soon).

Qwen2.5-coder 1.5b though tends to repeat what has already been written, or change it slightly. (See the video)

Is this intentional? I am passing the prefix and suffix correctly to ollama, so it knows where it currently is. I’m also trimming the amount of lines it can see, so the time-to-first-token isn’t too long.

Do you have a recommendation for a better code model, better suited for this?

  • Smorty [she/her]OP
    link
    fedilink
    English
    arrow-up
    2
    ·
    edit-2
    23 days ago

    You are right in that 1.5B sized models can’t be very intelligent. But I do expect it to not write what has already been written. It’s been finetued on this completion task, with tokens designed specifically for it. So I assumed it’d be okay at it.

    Sometimes it does actually generate what I want it to. When I provide a simple if statement with a for loop under it, it will autocomplete the else branch mostly correctly almost every time.

    Example:

    I wrote this code

    	var start_with_code := markdown.begins_with("```") or markdown.begins_with("\n```\n")
    	var result:String
    	var mark_i := 0
    	var code_i := 0
    	if start_with_code:
    		for i in markdown_texts.size() + code_blocks.size():
    			if i % 2 == 0.0:
    				result += code_blocks[code_i]
    				code_i += 1
    			else:
    				result += markdown_texts[mark_i]
    				mark_i += 1
    	else:
    

    And the model added this

    	else:
    		for i in markdown_texts.size() + code_blocks.size():
    			if i % 2 == 0.0:
    				result += markdown_texts[mark_i]
    				mark_i += 1
    			else:
    				result += code_blocks[code_i]
    				code_i += 1
    

    (I’m writing a markdown parser here)

    • hendrik@palaver.p3x.de
      link
      fedilink
      English
      arrow-up
      1
      ·
      edit-2
      23 days ago

      Fair enough. I haven’t had repetition being an issue for quite some time now. Usually that happened when I manually messed with the context, or had the parameters set incorrectly. Are you sure the fill-in-the-middle support is set up correctly and the framework inserts the correct tokens for that? I mean if it does other things properly, but isn’t able to fill in, maybe it’s that.

      • Smorty [she/her]OP
        link
        fedilink
        English
        arrow-up
        2
        ·
        23 days ago

        Your assumption is very good, but I am sure that the completion is setup correctly. Sometimes it does fill in correctly, like suggesting variable types and adding comments to functions.

        So sometimes completion works fieny but other times it doesn’t. I use the ollama REST API for the completion, so the token handling isn’t on my side.