Comments (6)
Yeah, sorry, couldn't find time for it yet, maybe closer to the end of the week. @Neilblaze if you can do it earlier, you should definitely take it.
If we go with nullable DefragThreshold
, we should probably make Seed
nullable too, because 'no seed' constant is not obvious. And in LLamaContext.SetSeed
too.
from llamasharp.
Good catch! Would you be itnerested in putting together a PR to fix this?
The easiest fix would be to just add a default value of -1
.
I think a better fix would be to change float DefragThreshold
to float? DefragThreshold
and then to convert null
to -1
in here. That way the default C# value (null
) converts to the default llama.cpp value and there are no "magic" numbers.
from llamasharp.
@AsakusaRinne Is this up for grabs?
from llamasharp.
Unless @dlyz is already working on it, I think it's open to whoever moves first :)
from llamasharp.
Yeah, sorry, couldn't find time for it yet, maybe closer to the end of the week. @Neilblaze if you can do it earlier, you should definitely take it.
Cool I'll try my best to raise a PR asap, but meanwhile, I'm waiting for this confirmation ↓
If we go with nullable
DefragThreshold
, we should probably makeSeed
nullable too, because 'no seed' constant is not obvious. And inLLamaContext.SetSeed
too.
Once @martindevans confirms, I can make the updates. Thanks!
from llamasharp.
Making seed nullable for the default is a good idea. Go for it @Neilblaze :)
from llamasharp.
Related Issues (20)
- CentOS x86_64 Failed Loading 'libllama.so' HOT 4
- System.TypeInitializationException: 'The type initializer for 'LLama.Native.NativeApi' threw an exception.' HOT 12
- How do I continously print the answer word for word when using document ingestion with kernel memory? HOT 1
- How to rebuild LLamaSharp backends HOT 2
- Namespace should be consistent
- Mamba HOT 10
- Android Backend HOT 2
- [Feature] Allow async model loading and cancellation
- [CI] Add more unit test to ensure the the outputs are reasonable HOT 3
- Take multiple chat templates into account
- [Feature]: Support for Function Calling or Tools HOT 4
- [BUG]: Answer stop abruptly after contextsize, even with limiting prompt size HOT 1
- [BUG]: Linux cuda version detection could be incorrect HOT 2
- [BUG]: WSL2 has problem running LLamaSharp with cuda11
- Add unit test about long context HOT 2
- Add debug mode of LLamaSharp
- How to better provide system information for LLMs HOT 3
- LLAVA Configuration HOT 4
- [Feature]: 不同的LLM模型,代码要以怎样的方式融合到项目里 HOT 3
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from llamasharp.