GithubHelp home page GithubHelp logo

LLaMa.cpp broken right now about agixt HOT 2 CLOSED

josh-xt avatar josh-xt commented on September 7, 2024
LLaMa.cpp broken right now

from agixt.

Comments (2)

Josh-XT avatar Josh-XT commented on September 7, 2024

Can you try with the latest version?

from agixt.

dany-on-demand avatar dany-on-demand commented on September 7, 2024

Had to do pip install -r requirements.txt and npm i npm run build in /frontend
New error:

 * Serving Flask app 'app'
 * Debug mode: on
WARNING: This is a development server. Do not use it in a production deployment. Use a production WSGI server instead.
 * Running on http://127.0.0.1:5000
Press CTRL+C to quit
 * Restarting with watchdog (windowsapi)
 * Debugger is active!
 * Debugger PIN: 242-200-888
127.0.0.1 - - [22/Apr/2023 09:21:14] "GET /api/agent HTTP/1.1" 200 -
127.0.0.1 - - [22/Apr/2023 09:21:14] "GET /api/agent/CatJokeFinder HTTP/1.1" 200 -
127.0.0.1 - - [22/Apr/2023 09:21:14] "OPTIONS /api/agent/CatJokeFinder/instruct HTTP/1.1" 200 -
127.0.0.1 - - [22/Apr/2023 09:21:15] "GET /api/agent/CatJokeFinder/command HTTP/1.1" 200 -
Using embedded DuckDB with persistence: data will be stored in: agents/default/memories
llama_model_load: loading model from 'C:/Projects/generative/llm/llama.cpp/models/vicuna/1.1TheBloke/ggml-vicuna-13b-1.1-q4_1.bin' - please wait ...
llama_model_load: GPTQ model detected - are you sure n_parts should be 2? we normally expect it to be 1
llama_model_load: use '--n_parts 1' if necessary
llama_model_load: n_vocab = 32000
llama_model_load: n_ctx   = 2000
llama_model_load: n_embd  = 5120
llama_model_load: n_mult  = 256
llama_model_load: n_head  = 40
llama_model_load: n_layer = 40
llama_model_load: n_rot   = 128
llama_model_load: f16     = 4
llama_model_load: n_ff    = 13824
llama_model_load: n_parts = 2
llama_model_load: type    = 2
llama_model_load: ggml map size = 9702.02 MB
llama_model_load: ggml ctx size = 101.25 KB
llama_model_load: mem required  = 11750.12 MB (+ 3216.00 MB per state)
llama_model_load: loading tensors from 'C:/Projects/generative/llm/llama.cpp/models/vicuna/1.1TheBloke/ggml-vicuna-13b-1.1-q4_1.bin'
llama_model_load: model size =  9701.58 MB / num tensors = 363
llama_init_from_file: kv self size  = 3125.00 MB
llama_generate: seed = 1682148077

system_info: n_threads = 8 / 32 | AVX = 1 | AVX2 = 1 | AVX512 = 0 | FMA = 1 | NEON = 0 | ARM_FMA = 0 | F16C = 1 | FP16_VA = 0 | WASM_SIMD = 0 | BLAS = 0 | SSE3 = 1 | VSX = 0 |
sampling: temp = 0.800000, top_k = 40, top_p = 0.950000, repeat_last_n = 64, repeat_penalty = 1.100000
generate: n_ctx = 2000, n_batch = 8, n_predict = 55, n_keep = 0


127.0.0.1 - - [22/Apr/2023 09:21:17] "POST /api/agent/CatJokeFinder/instruct HTTP/1.1" 500 -
[2023-04-22 09:21:17,807] {_internal.py:224} INFO - 127.0.0.1 - - [22/Apr/2023 09:21:17] "POST /api/agent/CatJokeFinder/instruct HTTP/1.1" 500 -
Traceback (most recent call last):
  File "C:\Users\Daniel\anaconda3\envs\textgen\lib\site-packages\flask\app.py", line 2551, in __call__
    return self.wsgi_app(environ, start_response)
  File "C:\Users\Daniel\anaconda3\envs\textgen\lib\site-packages\flask\app.py", line 2531, in wsgi_app
    response = self.handle_exception(e)
  File "C:\Users\Daniel\anaconda3\envs\textgen\lib\site-packages\flask_restful\__init__.py", line 271, in error_router
    return original_handler(e)
  File "C:\Users\Daniel\anaconda3\envs\textgen\lib\site-packages\flask_cors\extension.py", line 165, in wrapped_function
    return cors_after_request(app.make_response(f(*args, **kwargs)))
  File "C:\Users\Daniel\anaconda3\envs\textgen\lib\site-packages\flask\app.py", line 2528, in wsgi_app
    response = self.full_dispatch_request()
  File "C:\Users\Daniel\anaconda3\envs\textgen\lib\site-packages\flask\app.py", line 1825, in full_dispatch_request
    rv = self.handle_user_exception(e)
  File "C:\Users\Daniel\anaconda3\envs\textgen\lib\site-packages\flask_restful\__init__.py", line 271, in error_router
    return original_handler(e)
  File "C:\Users\Daniel\anaconda3\envs\textgen\lib\site-packages\flask_cors\extension.py", line 165, in wrapped_function
    return cors_after_request(app.make_response(f(*args, **kwargs)))
  File "C:\Users\Daniel\anaconda3\envs\textgen\lib\site-packages\flask\app.py", line 1823, in full_dispatch_request
    rv = self.dispatch_request()
  File "C:\Users\Daniel\anaconda3\envs\textgen\lib\site-packages\flask\app.py", line 1799, in dispatch_request
    return self.ensure_sync(self.view_functions[rule.endpoint])(**view_args)
  File "C:\Users\Daniel\anaconda3\envs\textgen\lib\site-packages\flask_restful\__init__.py", line 467, in wrapper
    resp = resource(*args, **kwargs)
  File "C:\Users\Daniel\anaconda3\envs\textgen\lib\site-packages\flask\views.py", line 107, in view
    return current_app.ensure_sync(self.dispatch_request)(**kwargs)
  File "C:\Users\Daniel\anaconda3\envs\textgen\lib\site-packages\flask_restful\__init__.py", line 582, in dispatch_request
    resp = meth(*args, **kwargs)
  File "C:\Projects\generative\llm\Agent-LLM\app.py", line 73, in post
    response = agent.run(objective, max_context_tokens=500, long_term_access=False)
  File "C:\Projects\generative\llm\Agent-LLM\AgentLLM.py", line 73, in run
    self.response = self.instruct(prompt)
  File "C:\Projects\generative\llm\Agent-LLM\provider\llamacpp.py", line 20, in instruct
    output = self.model.generate(f"Q: {prompt}", n_predict=55, new_text_callback=self.new_text_callback, n_threads=8)
  File "C:\Users\Daniel\anaconda3\envs\textgen\lib\site-packages\pyllamacpp\model.py", line 112, in generate
    pp.llama_generate(self._ctx, self.gpt_params, self._call_new_text_callback, verbose)
  File "C:\Users\Daniel\anaconda3\envs\textgen\lib\site-packages\pyllamacpp\model.py", line 83, in _call_new_text_callback
    Model._new_text_callback(text)
TypeError: AIProvider.new_text_callback() takes 1 positional argument but 2 were given

from agixt.

Related Issues (20)

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.