Changed embedding model to mixed bread xsmall, optimised related functions in general. Updated Gemini model references.
d3a9db3
| langchain==0.3.24 | |
| langchain-huggingface==0.1.2 | |
| langchain-community==0.3.22 | |
| beautifulsoup4==4.13.4 | |
| google-generativeai==0.8.5 | |
| pandas==2.2.3 | |
| transformers==4.51.3 | |
| # For Windows https://github.com/abetlen/llama-cpp-python/releases/download/v0.3.2/llama_cpp_python-0.3.2-cp311-#cp311-win_amd64.whl -C cmake.args="-DGGML_BLAS=ON;-DGGML_BLAS_VENDOR=OpenBLAS" | |
| llama-cpp-python==0.3.2 --extra-index-url https://abetlen.github.io/llama-cpp-python/whl/cpu # For linux if dependencies for below build command are not available in the environment | |
| #llama-cpp-python==0.3.9 -C cmake.args="-DGGML_BLAS=ON;-DGGML_BLAS_VENDOR=OpenBLAS" # CPU | |
| #llama-cpp-python==0.3.9 -C cmake.args="-DGGML_CUDA=on" # With CUDA | |
| torch==2.5.1 --extra-index-url https://download.pytorch.org/whl/cpu | |
| sentence_transformers==4.1.0 | |
| faiss-cpu==1.10.0 | |
| pypdf==5.4.0 | |
| python-docx==1.1.2 | |
| keybert==0.9.0 | |
| span-marker==1.7.0 | |
| gradio==5.25.2 | |
| nltk==3.9.1 | |
| bm25s==0.2.12 | |
| PyStemmer==2.2.0.3 | |
| scipy==1.15.2 | |
| numpy==1.26.4 | |
| boto3==1.38.0 | |
| python-dotenv==1.1.0 | |