Simple Python bindings for @ggerganov'sllama.cpp library. This package provides:
- Low-level access to C API via
ctypesinterface. - High-level Python API for text completion
- OpenAI-like API
- LangChain compatibility
Install from PyPI:
pip install llama-cpp-python>>>fromllama_cppimportLlama>>>llm=Llama(model_path="models/7B/...") >>>output=llm("Q: Name the planets in the solar system? A: ", max_tokens=32, stop=["Q:", "\n"], echo=True) >>>print(output){"id": "cmpl-xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxxxxx", "object": "text_completion", "created": 1679561337, "model": "models/7B/...", "choices": [{"text": "Q: Name the planets in the solar system? A: Mercury, Venus, Earth, Mars, Jupiter, Saturn, Uranus, Neptune and Pluto.", "index": 0, "logprobs": None, "finish_reason": "stop" } ], "usage":{"prompt_tokens": 14, "completion_tokens": 28, "total_tokens": 42 } }This project is licensed under the terms of the MIT license.