Comments (4)
I'm unfamiliar with LM Studio, but I just downloaded it and started a local inference server running llama 2 chat 7B q4_0 ggml
. Using the OpenAI provider in Wingman, I changed the URL to http://localhost:1234/v1/chat/completions
and things seem to just work (with the exception of recognizing the end of the completion stream, which may just be unique to LM Studio, but a bug none-the-less).
Can you provide some more information regarding your setup?
KoboldCpp support is almost finished - still porting this functionality from the previous major version.
from wingman.
with the exception of recognizing the end of the completion stream, which may just be unique to LM Studio, but a bug none-the-less
This is fixed now and the completion stream resoled by LM Studio should correctly end the response. Pushing a release with this bug fix now.
from wingman.
Thank you , it is working now. i was using "http://localhost:1234/v1". now i changed to "http://localhost:1234/v1/chat/completions" . it is working.
I was using Wingman 1.3.8 preview, it was excellent. Now, it upgraded well. Thank you.
Excellent! Glad it's working. Have fun!
from wingman.
Thank you , it is working now. i was using "http://localhost:1234/v1". now i changed to "http://localhost:1234/v1/chat/completions" . it is working.
I was using Wingman 1.3.8 preview, it was excellent. Now, it upgraded well. Thank you.
from wingman.
Related Issues (20)
- Support configurable inference params per prompt HOT 8
- Context window and max_tokens management
- Disable the copy icon when generating HOT 1
- Write function comment prompt: ask before injecting the response into the file HOT 1
- New Goinfer provider HOT 3
- New Koboldcpp provider HOT 10
- Great Job! HOT 1
- A new theme sharing HOT 1
- Please include the julia language (.jl) extension also. HOT 1
- Using the new version with local open source backends HOT 5
- Discussion: modelfusion HOT 2
- Missing Implementation of 'file' Placeholder as Mentioned in README HOT 1
- Feature Request: Enhance Contextual Understanding by Allowing Code File Integration
- Feature Request: Integration of Azure OpenAI Credentials Support HOT 1
- The selected code is forcibly replaced. HOT 4
- Model Installed in VS Code but No Settings No Logo in Sidebar Menu HOT 8
- `Error: Failed to open completion stream: 429 Too Many Requests` HOT 2
- Feature Request: Support Anthropic messages API
- Remove restriction on Ollama models
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from wingman.