Back

Show HN: Sweep, Open-weights 1.5B model for next-edit autocomplete

534 points16 dayshuggingface.co

Hey HN, we trained and open-sourced a 1.5B model that predicts your next edits, similar to Cursor. You can download the weights here (https://huggingface.co/sweepai/sweep-next-edit-1.5b) or try it in our JetBrains plugin (https://plugins.jetbrains.com/plugin/26860-sweep-ai-autocomp...).

Next-edit autocomplete differs from standard autocomplete by using your recent edits as context when predicting completions. The model is small enough to run locally while outperforming models 4x its size on both speed and accuracy.

We tested against Mercury (Inception), Zeta (Zed), and Instinct (Continue) across five benchmarks: next-edit above/below cursor, tab-to-jump for distant changes, standard FIM, and noisiness. We found exact-match accuracy correlates best with real usability because code is fairly precise and the solution space is small.

Prompt format turned out to matter more than we expected. We ran a genetic algorithm over 30+ diff formats and found simple `original`/`updated` blocks beat unified diffs. The verbose format is just easier for smaller models to understand.

Training was SFT on ~100k examples from permissively-licensed repos (4hrs on 8xH100), then RL for 2000 steps with tree-sitter parse checking and size regularization. The RL step fixes edge cases SFT can’t like, generating code that doesn’t parse or overly verbose outputs.

We're open-sourcing the weights so the community can build fast, privacy-preserving autocomplete for any editor. If you're building for VSCode, Neovim, or something else, we'd love to see what you make with it!

leonardcser16 days ago

Hi, I tried the model and I am super impressed by the performance/quality. Thanks for making this open source!

I am the author of this Neovim plugin for edit completions. I was able to integrate it with the Sweep Edit model.

For anyone who is interested: https://github.com/leonardcser/cursortab.nvim

lasgawe16 days ago

Hey this is really interesting. I'll try your nvim plugin

treyd16 days ago

Is there a port of this to Emacs or integration with gptel?

leonardcser16 days ago

Hi, not that I know of. Most of the code would not change. It could easily be ported to different editors. The core is the go server (`server/`).

9999gold16 days ago

It seems it would be possible to use this with minuet.el. I’m not familiar with it, though.

kevinlu124816 days ago

this is awesome, i'm going to try this out

zekejohn2 days ago

Nice, could this be used to auto complete terminal/cli commands?

KronisLV16 days ago

I remember using Qwen 2.5 Coder for autocomplete with Continue.dev, that experience was a mess both in JetBrains IDEs, as well as Visual Studio Code.

People posting stuff like this is really cool because otherwise it kinda feels like nobody gives a crap, for example even with Cline/RooCode/KiloCode there’s no good way for me to hook up an autocomplete model that either runs in Ollama or maybe a remote Cerebras Code model, like KiloCode doesn’t have a proper model configuration option even if it has it for the chat or regular agentic stuff - I don’t get why autocomplete is such a special case.

I guess what I’m saying is that I’m glad someone’s at least trying so I don’t have to keep a Copilot subscription just because I genuinely like their autocomplete and the rest of it is basically wasted: Claude Code and Codex and others are better for the actual chat/agentic stuff, KiloCode and others are really nice IDE plugins.

lostmsu16 days ago

llama.cpp has an extension for VS Code, but configuration UX is utter crap

vanillameow16 days ago

Sometimes when I use a plugin like this I get reminded just how much of a productivity nerf it is to code without an autocomplete AI. Honestly in my opinion if you write a lot of boilerplate code this is almost more useful than something like Claude Code, because it turbocharges your own train of thought rather than making you review someone else's, which may not align with your vision.

This is a really good plugin. I'm a diehard JetBrains user, I tried switching to VSCode and its various forks many times because of AI but muscle memory from years of use is hard to override. And for a lot of languages JetBrains is just much better, especially out of the box. But they dropped the ball so hard on AI it's unbelievable. Claude Code pulled it back a bit because at least now the cutting edge tools aren't just VSCode plugins, but I was still missing a solid autocomplete tool. Glad this is here to fill that niche. Very likely will be switching my GitHub copilot subscription to this.

I also really appreciate publishing open weights and allowing a privacy mode for anonymous trial users, even if it's opt-in. Usually these things seem to be reserved for paying tiers these days...

zarzavat16 days ago

I have always said this and had people on HN reply that they don't get much use out of autocomplete, which puzzled me.

I'm starting to understand that there are two cultures.

Developers who are mostly writing new code get the most benefit from autocomplete and comparatively less from Claude Code. CC is neat but when it attempts to create something from nothing the code is often low quality and needs substantial work. It's kind of like playing a slot machine. Autocomplete, on the other hand, allows a developer to write the code they were going to write, but faster. It's always a productivity improvement.

Developers who are mostly doing maintenance experience the opposite. If your workflow is mostly based around an issue tracker rather than figma, CC is incredible, autocomplete less so.

norir16 days ago

I personally find autocomplete to be detrimental to my workflow so I disagree that it is a universal productivity improvement.

genghisjahn16 days ago

I’m in the “write new stuff with cc and get great code.” Of course I’ll be told I don’t really know what I’m doing. That I just don’t know the difference between good and bad quality code. Sigh.

zarzavat16 days ago

The best code is no code.[0]

The main "issue" I have with Claude is that it is not good at noticing when code can be simplified with an abstraction. It will keep piling on lines until the file is 3000 lines long. You have to intervene and suggest abstractions and refactorings. I'm not saying that this is a bad thing. I don't want Claude refactoring my code (GPT-5 does this and it's very annoying). Claude is a junior developer that thinks it's a junior. GPT-5 is a junior developer that thinks it's a senior.

[0]: https://www.folklore.org/Negative_2000_Lines_Of_Code.html

kevinlu124815 days ago

Definitely agree here, have had so many cases where I would like ask Claude for XYZ, then ask for XYZ again but with a small change. Instead of abstracting out the common code it would just duplicate the code with the small change.

mark_l_watson16 days ago

I agree with you. I do have some local model LLM support configured in Emacs, as well as integration with gemini 3 flash. However, all LLM support is turned off in my setup unless I specifically enable it for a few minutes.

I will definitely try the 1.5B model but I usually use LLMs by taking the time to edit a large one-shot prompt and feed it to either one of the new 8B or 30B local models or to gemini 3 flash via the app, web interface, or API.

Small purpose-built models are largely under-appreciated. I believe that it is too easy to fall into the trap of defaulting to the strongest models and to over rely on them. Shameless plug: it is still incomplete, but I have released an early version on my book ‘Winning Big With Small AI’ - so, I admit my opinions are a little biased!

pdyc16 days ago

do you know a good way to give context files in emacs? currently i have to either do ctrl-x+h to select file content of individual files to give to ai or copy files themselves from ai's chat interface. I would much prefer selecting all files at once and get their content copied to clipboard.

cmrdporcupine16 days ago

Yep. I'm coming to resent Claude Code and tools like it for taking me out of direct contact with the code.

I think we're still in the early days of these systems. The models could be capable of a lot more than this "chat log" methodology.

Agree about JetBrains dropping the ball. Saddens me because I've also been a diehard user of their products since 2004.

qorrect16 days ago

Glad to hear I'm not alone, the latest releases of JetBrains have been so bad I finally cancelled my subscription. VSCode has been a nice surprise, "its giving old emacs" as the kids would say.

sitkack16 days ago

I am curious about how both of you think Jetbrains is dropping the ball so much that you are no longer buying the tool.

You are still using it but no longer getting updates?

+1
cmrdporcupine16 days ago
kevinlu124815 days ago

I've done some testing before and many of the new Jetbrains internal plugins cause memory leaks which really lags down my IDE...

wwfn16 days ago

Hopefully not too offtopic: why so much boilerplate?

I see most would-be-boilerplate code refactored so the redundant bit becomes a small utility or library. But most of what I write is for research/analysis pipelines, so I'm likely missing an important insight. Like more verbose configuration over terse convention?

For code structure, snippets tempting[1] ("iff[tab]" => "if(...){...}") handles the bare conditional/loop completes in a more predictable way and offline/without a LLM eating into RAM.

[1] https://github.com/joaotavora/yasnippet; https://github.com/SirVer/ultisnips; https://code.visualstudio.com/docs/editing/userdefinedsnippe...

djfdat16 days ago

Abstracting away redundancy could make it harder to understand exactly what the code is doing, and could introduce tech debt when you need slightly different behavior from some code that is abstracted away. Also, if the boilerplate code is configuration, its good to see exactly what the configuration is when trying to grok how some code works.

You bring up a good point with snippets though, and I wonder if that would be good information to feed into the LLM for autocomplete. That snippet is helpful if you want to write on condition at a time, but say you have a dozen conditions if statements to write with that snippet. After writing one, the LLM could generate a suggestion for the other 11 conditions using that same snippet, while also taking into consideration the different types of values and what you might be checking against.

As for RAM/processing, you're not wrong there, but with specialized models, specialized hardware, and improvements in model design, the number of people working under such restricted environments where they are concerned about resource use will decrease over time, and the utility of these tools will increase. Sure a lower-tech solution works just fine, and it'll continue to work fine, but at some point the higher-tech solution will have similar levels of friction and resource use for much better utility.

esafak16 days ago

Junie is irredeemable but if it's autocomplete that you are unhappy about, IntelliJ has both local- and cloud autocomplete now.

norir16 days ago

It is depressing that our collective solution to the problem of excess boilerplate keeps moving towards auto-generation of it.

notsylver16 days ago

I've been waiting for something like this for ages. Cursor making me pay $20/month when all I use from it is autocomplete was always a little annoying, especially as they changed the UI to push agents more and it got in the way. I was even considering doing it myself but wasn't sure about gambling on models small enough to run locally being smart enough to do anything useful.

I threw together a vscode extension to run it and while the extension is rough, the model seems decent. I'm trying to keep my expectations contained, in the past local models have been absolutely terrible for inline completion, this seems much better already. I hope this kicks off more competition.

kevinlu124815 days ago

Let me know if you have any questions. We have a lot of harness code that cleans up many bad behaviours that makes it a lot more usable (like token healing: https://blog.sweep.dev/posts/token-healing-autocomplete).

dainiusse15 days ago

Do you have anything to share? Would be curious trying it out

kleiba16 days ago

Very cool!

I understand that the 1.5B is small enough to run locally... but does it actually in the Sweep AI Jetbrains plugin? That is, if I install the plugin, will I download the model automatically and the plugin doesn't phone home?

kevinlu124815 days ago

Not at the moment, if you install the hosted Sweep AI Jetbrains plugin it uses our hosted (larger) model.

bjarteaarmolund16 days ago

no, as far as I can see there is no way to configure the Jetbrains plugin to use a local endpoint.

NewsaHackO16 days ago

Yes, I get the same vibe, as one has to sign in to their site to use the plugin. Kind of grimy for them to seemingly imply that it is locally run when it isn't.

rkagerer16 days ago

Why not?

Can someone make a better plugin?

esquire_90016 days ago

Surprising how badly Jetbrains implemented AI. Apparently to such an extent that even after multiple years of LLM's someone felt confident enough to build a company that can do better.

This looks really neat, interesting technical writeup as well!

kevinlu124815 days ago

Thanks! Let us know if you have any questions / feedback.

martianlantern16 days ago

This is cool! I am more interested in how you guys generated next edit training data from repos, seems like there are lots of caveats here. Would love your insights

Again amazing work! waiting for what you guys cook next

knowaveragejoe16 days ago

The blog post has more information: https://blog.sweep.dev/posts/oss-next-edit

kevinlu124815 days ago

Also more technical details on SFT data here:

https://blog.sweep.dev/posts/next-edit-jetbrains#building-au...

_ache_16 days ago

It's good. The blog post about it is very interesting. I hope, a plugin for neovim will be made soon.

https://blog.sweep.dev/posts/oss-next-edit

evanreichard16 days ago

There's also https://github.com/ggml-org/llama.vim

Which I've been using with Qwen3 Coder. As long as infill is supported, that should work. I'll try later today.

jmanandchanny14 days ago

Thanks for sharing this. I personally use vim and not neovim (I do not have anything against it), so this plugin will be a great addon for me. Currently, I have to switch from vim to Cursor and back again for any kind of vibe coding.

mromanuk16 days ago

There is one already, on of the plugin authors commented here

kevinlu124815 days ago

Someone in this thread already built a Neovim plugin connecting to this model I believe.

WanderlingSmurf16 days ago

[dead]

kamranjon16 days ago

I read the release but didn't quite understand the difference between a next-edit model and a FIM model - does anyone have a clear explanation of when to use one over the other? I'd love if there was a sublime plugin to utilize this model and try it out, might see if I can figure that out.

evolving-silica16 days ago

I was curious as well and wanted to try how this work, so I asked claude to create a plugin for that. This utilizes built-in autocomplete behavior. If you want to give it a try then feel free to have a look here https://github.com/lumnn/AItoComplete (did not push it to packagecontrol yet)

sheepscreek16 days ago

I’m going to speculate a bit here, FIM may stand for something-in-the-middle?

I know there are the original autocomplete models that simply complete the endings. Then there are Cursor like models capable of editing/filling text between blocks of code. In essence, they look at both the text before the insertion point and after it - then find the best fitting completion in the middle. My guess is FIM is the latter.

aidos16 days ago

As you said. Fill-in-the-middle.

kevinlu124815 days ago

We have an explanation here: https://blog.sweep.dev/posts/next-edit-jetbrains#next-edit-a...

But basically suggesting changes away from your cursor position

cmrdporcupine16 days ago

I've been trying my hands at implementing an emacs package for inline completions with this. I have it mostly working and performance is good enough but I haven't been blown away by the quality of its suggestions unfortunately. Which I guess is expected from a 1.5B model.

I'd love to see them making a larger model in the 10-20b range maybe? I know most people wouldn't be able to run that on their machines, but some could.

Running on ollama locally on NVIDIA Spark GB10. Tried it also with vLLM. Pretty fast.

kevinlu124815 days ago
cmrdporcupine15 days ago

Yea, I tweaked it a bunch to try to follow what was described there

mijoharas16 days ago

Do you care to share your implementation?

cmrdporcupine16 days ago

If I can make it clean and decent I will. I might look at again after work and see if I can tune it up. It was a bit flake and I wasn't blown away by the interaction.

logicallee16 days ago

Congratulations on training a relatively small model that can beat larger models for this important task.

>We ran a genetic algorithm over 30+ diff formats

Can you you give more information about your genetic algorithm? Did you do crossover over the trained models (for example, ranking by fitness, take 20% most elite and create children by mixing their weights randomly)? Did you have a 'population size' (number of instances) for the genetic algorithms, and if so what was it?

zoobab16 days ago

Where is the training data?

We can't keep calling those models "open source" if we have a black box and know precisely how they were made.

"Open weights" are the new binary.

kevinlu124815 days ago

Woops meant to say open-weight. We put open-weight in the title and but accidentally wrote open-source in the description.

smusamashah16 days ago

Can this be used offline in Jetbrain IntelliJ? Looking at the plugin, it looks like it requires sign in and then it uses the cloud based model instead of the local one. Can't tell.

jrop16 days ago

Between GLM-4.7-Flash and this announcement, THIS is what I'm excited to see in this space: pushing the capabilities of _small_ models further and further. It really feels like we're breaking into a space where models that can run on hardware that I actually own is getting better and better, and that has me excited.

magnat16 days ago

Is there a way to use this (or similar) model in Visual Studio? Extensions on Visual Studio Marketplace are clunky and sluggish at best, if they even work at all.

denysvitali16 days ago

If you mean VSCode (or any other editor):

> We’re open sourcing the model weights so the community can build fast, privacy-preserving autocomplete for every IDE - VSCode, Neovim, Emacs, and beyond.

https://blog.sweep.dev/posts/oss-next-edit

magnat16 days ago

No, I mean Visual Studio (the IDE), not Visual Studio Code (the editor).

KeplerBoy16 days ago

Of course they are different products, but is there really a meaningful distinction between VS Code and an IDE? For all i care VS Code is a complete IDE.

ttoinou16 days ago

You need to add (official) extensions for that though. Which makes VSCode more flexible

pezgrande16 days ago

I thought there was already a generic plugin for this :(. Let's wait for one then ha, or I may just make one.

h33t-l4x0r16 days ago

It sounds like you might be killing Zed's ability to monetize, am I misunderstanding that?

BoredPositron16 days ago

If your only feature worth monetizing is replicated by a solo dev in his freetime you might have a problem.

gunalx16 days ago

not really though. The zed monetization seems to push towards selling tokens for full fledged models with good ide integration as a service. (They have let you run a custom tabcomplete for a little while)

woile16 days ago

Hey, ollama run as suggested in hf doesn't seem to work with this model. This worked instead:

ollama pull hf.co/sweepai/sweep-next-edit-1.5B

woile16 days ago

I've been using it with the Zed editor and it works quite well! Congrats.

This kind of AI are the ones I like and I'm looking to run in my workstation.

kevinlu124815 days ago

Double-check if you're using the right format.

Example here: https://huggingface.co/sweepai/sweep-next-edit-1.5B/blob/mai...

theophaniel16 days ago

Could you give the gist / config on how you made it work with Zed ?

ihales15 days ago

I had Claude add it as an edit-prediction provider (running locally on llama.cpp on my Macbook Pro). It's been working well so far (including next-edit prediction!), though it could use more testing and tuning. If you want to try it out you can build my branch: https://github.com/ihales/zed/tree/sweep-local-edit-predicti...

If you have llama.cpp installed, you can start the model with `llama-server -hf sweepai/sweep-next-edit-1.5B --port 11434`

Add the following to your settings.json:

```

  "features": {
    "edit_prediction_provider": { "experimental": "sweep-local" },
  },
  "edit_predictions": {
    "sweep_local": {
      "api_url": "http://localhost:11434/v1/completions",
    },
  }
```

Other settings you can add in `edit_predictions.sweep_local` include:

- `model` - defaults to "sweepai/sweep-next-edit-1.5B"

- `max_tokens` - defaults to 2048

- `max_editable_tokens` - defaults to 600

- `max_context_tokens` - defaults to 1200

I haven't had time to dive into Zed edit predictions and do a thorough review of Claude's code (it's not much, but my rust is... rusty, and I'm short on free time right now), and there hasn't been much discussion of the feature, so I don't feel comfortable submitting a PR yet, but if someone else wants to take it from here, feel free!

+1
oakesm915 days ago
woile16 days ago

This is it:

{

    "agent": {

        "inline_assistant_model": {

            "model": "hf.co/sweepai/sweep-next-edit-1.5B:latest",

            "provider": "ollama",

        },

    }

}
Imustaskforhelp16 days ago

+1, I wasn't able to make it work on zed either and It would really help if woile can tell how they made it work on their workstation.

+1
Imustaskforhelp16 days ago
kevinlu124815 days ago

We'll push to Ollama

mgz16 days ago

I use Sweep’s Jetbrains autocomplete plugin daily, it really stands out.

smusamashah16 days ago

Does it run totally offline?

8n4vidtmkvmk16 days ago

Better than the one that ships with Jetbrains?

I did buy their $100/yr AI but its about to run out.

mgz16 days ago

Definitely better. Next edit makes a difference. But it is not free, I think I pay $10/month.

hdjrudni16 days ago

Oh, i thought you were talking about this self hosted 1.5B model. You must be talking about the full model as a service?

syntaxing16 days ago

Wow super fun read, I love how it went into the technical details. Any way to make it work with vscode?

keyle16 days ago

I'm playing around with this in LMStudio (in huggingface -> use this model dropdown -> LMStudio)

It's really impressive so far, so quick to respond on a mac mini M2. And it appears to be accurate at least for the obvious questions.

I couldn't get it to work as an autocomplete of Zed unfortunately. It looks like it's hardwired to work with some providers and LMStudio is not included in the prediction engines list. Has anyone got a work around?

kevinlu124815 days ago

Our hosted autocomplete is coming to Zed in a few weeks.

bangaladore16 days ago

So SFT cost less only low hundreds of dollars? (1-10$ per hour per H100 if I'm seeing this correctly).

What about SFT?

Presumably basing this of Qwen is the reason it can be done for so cheap?

andruby16 days ago

How easy is it to re-train these to specific subset of programming languages? Could there be a "ruby+rails+html" version, etc?

bradfa16 days ago

I'd love to be able to take an open model like this and feed it the codebases that I work on regularly in order to improve its performance for less "hip/modern" languages and frameworks. It would be awesome to see a blog post about how normal users can find tune these models and rough cost estimates with examples!

ajayarama16 days ago

This is actually a game changer. I’ve been meaning to want to run models to accomplish exactly this, but don’t have enough VRAM on my GPU for the conventional LLM-method for the most part. This seems to be a far more efficient method of accomplishing a more scoped problem. Thank you for making it open source!

kevinlu124815 days ago

Let me know if you have any questions! What hardware are you on?

keepamovin16 days ago

This is so cool. What is the second order effect of model training becoming democratized? And local models becoming the norm? Tasks like agentic work are well handled by current AI as long as you know what you're doing and can stress the agent against tests/spec, etc.

I am thinking that one effect is:

- it will become normal for meta-models to train a model specific to a particular task/product.

Also, differently, I'm quite sure that AGI is not available on this current path (useful tho it is), but that some algo improvements might crack ubiquitous trainable AGI. Probably including some kind of embodiment to provide world-models and emotions (which are essential to embodied survival and success).

kevinlu124815 days ago

Personally, I think usable AI is more valuable than simply more intelligence. Many of the labs are pushing towards models that are 1% better on CodeForces and AIME if you just let it think and use tools for hours, instead of more user-friendly models with better coding habits, like writing shorter and more modular code.

keepamovin15 days ago

Totally this. But the corp labs have incentives to keep researching per investors and staffing load, so they have to show work.

I guess a nice advantage of backwardness here is that economic opportunities exist for those who can solve pain points in the use of existing intel. Older models often do almost as well at agentic tasks in reality, can probably go further.

Still, AGI should remove a lot of this making it redundant, and it will then be more about the intel than the tooling. But an opportunity exists now. We may not have widespread AGI until 8 - 10 years later, so plenty of money to be made in the meantime.

kevinlu124814 days ago

Ya definitely, that makes total sense. It feels to me that currently the labs have great researchers, who only care about making models perform better across raw intel and then they have incompetent applied AI engineers / FDE's who can only suggest using better prompting to remove bad habits to make agents more usable.

keepamovin13 days ago

Sounds like opportunities await those ready to fill the gaps left by these big ones! :)

dainiusse16 days ago

Any easy way to try on vscode?

pdyc16 days ago

I dont want to hand edit i want the output of better ai model with edit instructions like //update here <code> // new code <code> insert here etc. and local model read the files and apply the updates. I tried generating patch format but both bigger models fail to generate it accurately and smaller models have hard time in using them. Is there some way to do this with this kind of model? or its for completions while editing only?

sim04ful16 days ago

I'm very green to this so forgive if this question sounds silly:

Would instead of the RL step a constrained decoding say via something like xgrammar fix syntax generation issue ?

NitpickLawyer16 days ago

> Would instead of the RL step a constrained decoding say via something like xgrammar fix syntax generation issue ?

It can, but you have to consider two things here:

a) constrained decoding ensures adherence to syntax, not semantics. Say you're editing a field in an enum in rust. You can write syntactically correct rust code that doesn't address the new field further in the code (say in a switch). You'd get correctly syntactic code, but the compiler will scream at you. RL works on both.

b) if your goal is to further train the model, so it works on many tasks, RL helps with exploring new paths and training the model further. Constrained grammars help with inference, but the model doesn't "learn" anything. With RL you can also have many reward functions at the same time. Say one that rewards good syntax, one that rewards "closing" all the functions so tree-sitter doesn't complain, and one that rewards 0 errors from the compiler. The model gets to train on all 3 at the same time.

kevinlu124815 days ago

^ these were pretty much the main reasons.

The other one is that constrained decoding only works on CFGs (simpler grammars like JSON schemas) since only these ones can produce automatas which can be used for constrained decoding. Programming languages like Python and C++ aren't CFGs so it doesn't work.

Also constrained decoding generally worsens model quality since the model would be generating off-policy. So RL helps push corrected syntax back on-policy.

Semaphor15 days ago

At least for C#, the quality of the cloud offering is rather mediocre, so I don’t expect this model to be that useful there. It’s very overeager, suggesting tons of stuff that I never accepted because it made no sense. It’s also producing bad code, wanting me to use `.Result` for async calls instead of simply await-ing.

kevinlu124815 days ago

It's a bit undertrained on C#, we'll continue improving on this!

deepsquirrelnet16 days ago

This is really awesome detail. I’m very impressed by the amount of care taken to identify a good template. I started a small hook to try and do this using DSPy prompt optimizers, but haven’t had a compelling use case to try it with.

This seems like an ideal case for trying DFT as well. I’m not sure if you’re using trl, but I’d suggest checking that out.

kevinlu124815 days ago

We're using an internal fork of trl for some of the steps.

jedisct116 days ago

Really cool.

But how to use it instead of Copilot in VSCode ?

flanked-evergl16 days ago

Would love to know myself, I recall there was some plugin for VSCode that did next edits that accepted a custom model but I don't recall what it was now.

replete16 days ago

Run server with ollama, use Continue extension configured for ollama

BoredomIsFun16 days ago

I'd stay away from ollana, just use llama.cpp; it is more up date, better performing and more flexible.

mika699616 days ago

But you can't just switch between installed models like in ollama, can you?

BoredomIsFun16 days ago
ttoinou16 days ago

Wow, I can even chat about C code with that model with LM Studio on my Macbook at 200 tokens per seconds

kevinlu124815 days ago

Haha, we never trained it for chat but I would bet it works regardless.

Also that's crazy, M4 Mac?

ttoinou15 days ago

M4 Max 128GB yeah

vichle16 days ago

What type of hardware do I need to run a small model like this? I don't do Apple.

bodegajed16 days ago

1.5B models can run on CPU inference at around 12 tokens per second if I remember correctly.

moffkalast16 days ago

Ingesting multiple code files will take forever in prompt processing without a GPU though, tg will be the least of your worries. Especially when you don't append but change it in random places so caching doesn't work.

bradfa16 days ago

A FIM or completion model like this won't have a large prompt and caching doesn't work anyways (per their notes). It'll get maybe a few thousand tokens in a prompt, maximum. For a 1.5B model, you should expect usable CPU-only inference on a modern CPU, like at least hundreds of tokens per second of prefill and tens of tokens per second of generation, which is decently usable in terms of responsiveness.

moffkalast16 days ago

A thousand tokens (which would be on the low side) at 10-100 t/s in ingestion speed is 10-100 seconds. I don't seriously expect anyone to wait a solid minute after pressing tab for autocomplete, regular autocomplete gets unusably annoying if it takes more than a split second tbh.

kevinlu124815 days ago

Unfortunately, the main optimization (3x speedup) is using n-gram spec dec which doesn't run on CPUs. But I believe it works on Metal at least.

jychang16 days ago

1.54GB model? You can run this on a raspberry pi.

BoredomIsFun16 days ago

Performance of LLM inference consists of two independent metrics - prompt processing (compute intensive) and token generation (bandwidth intensive). For autocomplete with 1.5B you can get away with abysmal 10 t/s token generation performance, but you'd want as fast as possible prompt processing, pi in incapable of.

gunalx16 days ago

if you mean on the new ai hat with npu and integrated 8gb memory, maybe.

whimsicalism16 days ago

Very interesting - and cool to read about the development process. I'd love to hear more about how genetic algorithm worked here.

I wonder whether we are perhaps the point of usefulness of 'next edit' code development in 2026 though.

_boffin_16 days ago

Followed your work since the beginning and used it for inspiration for some cool demos on self-healing web scrapers. fascinating to see the transition from original concept to producing models. cool stuff.

k929415 days ago

Is there an oss model for next word / edits predictions for texts in general? e.g. Typing emails?

dubesar5514 days ago

Has somebody built any vscode extensions for this? Also is anyone serving this model?

bberenberg16 days ago

This seems great for code, but can this be used for non-code use cases?

kevinlu124815 days ago

Yes, I've used it to write blog posts / large user-facing copy.

ragchronos16 days ago

Does anyone know if the 7B model is also available somewhere?

moelf16 days ago

what do people use for Neovim to integrate these models for tab-completion level of stuff. (i.e. non agentic/vibe coding)

dajonker16 days ago

I use llama.vim with llama.cpp and the qwen2.5-coder 7B model. Easily fits on a 16 GB GPU and is fast even on a tiny RTX 2000 card with 70 watts of power. Quality of completions is good enough for me, if I want something more sophisticated I use something like Codex

rationably16 days ago

Do you plan to release Sweep 3B/7B on HF?

_ache_16 days ago

Yeap, the two seems like game changer. For now, I'm using "Qwen2.5-Coder-7B". Sweep 1.5B is "just" 12 % point better than Qwen2.5-Coder, but Sweep 7B is 25% point better.

kevinlu124815 days ago

Not at the moment but we do host it for our Jetbrains plugin

_mugencode16 days ago

Great! I have been trying to do something similar for Clojure.

This is a great resource to explore similar approach. https://blog.sweep.dev/posts/oss-next-edit

My notes so far https://kapilreddy.me/notes/2024/11/17/building-clojure-slm-...

rw_panic0_016 days ago

is there any llm lsp it can integrate well with?

kevinlu124815 days ago

We currently integrate with Jetbrains' PSI

ing33k16 days ago

can it be integrated in monaco editor ?

wepaean16 days ago

[dead]

asyncze16 days ago

[dead]

plutodev16 days ago

[flagged]

oefrha16 days ago

You’re subtly pushing the same product in basically every one of your comments. If these are good faith comments please edit out the product name, it’s unnecessary and doing so as a green account just makes people consider you a spammer. Establish yourself first.

subscribed16 days ago

They've submitted "I'm working at io.net" quite openly, but I admit, they should at least announce their employment in the bio, otherwise it's a very poorly executed astroturf post (phrased like they're an experimenting user and not a dev).

lelanthran16 days ago

Or he could disclose it.l, which he did in a different comment on a different story.

I agree that green accounts could be regarded as suspicious and, if it were me, I'd disclose each time I mention it.

kouteiheika16 days ago

> On the infra side, training a 1.5B model in ~4 hours on 8×H100 is impressive.

It's hard to compare without more details about the training process and the dataset, but, is it? Genuine question, because I had the opposite impression. Like, for example, recently I did a full finetuning run on a 3B model chewing through a 146k entry dataset (with 116k entries having reasoning traces, so they're not short) in 7 hours on a single RTX 6000.

kevinlu124815 days ago

Honestly I think we can improve our training throughput drastically via a few more optimizations but we've been spending most of our time on model quality improvements instead.

dcreater16 days ago

Based on qwen2.5-coder? seems like a "why not/resume embellish/show VC" type release I guess

dang16 days ago

"Please don't post shallow dismissals, especially of other people's work. A good critical comment teaches us something."

https://news.ycombinator.com/newsguidelines.html

kevinlu124815 days ago

You can see that Qwen3 does worse than Qwen2.5 on our benchmark. Reason is it's never been pretrained for FIM / autocomplete.