Structural refactoring is another amazing feature that is worth knowing.
I can’t help but wonder if the folks at JetBrains are starting to feel a bit of pressure from tools like Cursor or Windsurf
"404 - page not found The
master branch of
javafactory-plugin does not contain the path
docs/how-to-use.md."
How do I hook it into local models? Does it support Ollama, Continue, that kind of thing? Do you collect telemetry?
2. from now, i only allow to use gpt-4o, because the requests involve relatively long context windows, which require high-quality reasoning. Only recent high-performance models like GPT-4o or Claude Sonnet are capable of reducing the manual workload for this kind of task.
___
but still, if user want to use other models , i can make adapter features for various models
Right, so it can't be used on proprietary code or in settings where personal data might occur.
To be honest, I didn’t realize that model selection would be such an important point for users. I believed that choosing a high-quality model with strong reasoning capabilities was part of the service’s value proposition.
But lately, more users — including yourself — have been asking for support for other models like Claude Sonnet or LLaMA.
I’m now seriously considering adding an adapter feature. Thank you for your feedback — I really appreciate it.
Together with the CEO I've also decided that we do not do this with our own code, it stays on machines we control until someone pays for some artifact we'd like to license.
I'm well aware that many other organisations take a different position and push out basically everything they work on to SaaS LLM:s, in my experience defending it with something about so called productivity and something about some contract clause about the SaaS pinky promising to not straight up take the code. But nothing stops them from running hidden queries against it with their in-house models parallel with providing their main service, and sift out a lot of trade secrets and other goodies from it.
It's also likely these SaaS corporations can benchmark and otherwise profile individual developers, information that would be very valuable to e.g. recruiting agencies.
If you can tell, how is that Copilot performance measured?
Actually, I'm currently thinking about creating a small community for sharing pattern definitions.
But the project looks interesting, I have been looking for something similar.
The requests involve relatively long context windows, which require high-quality reasoning. Only recent high-performance models like GPT-4o or Claude Sonnet are capable of reducing the manual workload for this kind of task.
That said, what I wanted to highlight in the example was a contrast — tools like Cursor and other general-purpose models often fail to even generate simple tests correctly, or can't produce tests that pass. So the goal was to show the difference in reliability.