Questions tagged [large-language-model]

Use this tag for questions about large language models (LLM), trained deep-learning artificial intelligence algorithms that interpret and generate natural language text.

118 questions
0
votes
0 answers

Getting inconsistent response when using pandas_dataframe_agent and taking too much time for execution

def create_pandas_dataframe_agent( llm:llm, df:df, #callback_manager: Optional[BaseCallbackManager] = None, prefix: str=PREFIX, suffix: str= SUFFIX, input_variables:str= None, verbose:bool=False, …
0
votes
1 answer

How to compile the llm library with nix

I tried to install the llm library with nix. You just have to do that: git clone https://github.com/rustformers/llm.git nix build Unfortunately it fails: error: builder…
0
votes
0 answers

How fix 'type=value_error' when loading a wizard-vicuna model to PrivateGPT?

I'm following a tutorial to install PrivateGPT and be able to query with a LLM about my local documents. I'm using a wizard-vicuna-13B.ggmlv3.q4_1.bin model, and as per the README.md adjusted the example.env file settings to a new .env…
0
votes
0 answers

GPU out of memory in Jupyter

My OS specs are 8GB dedicated GPU and 24GB shared GPU, I am trying to fine tune a model and I am facing CUDA out of memory errors . I am not able to use Shared memory. Is there any way to use it when my dedicated memory was full. Code Snippet for…
0
votes
0 answers

How to deal with "This model's maximum context length is 4097 tokens." issue in Scikit-LLM

I am trying the Scikit-LLM on a StackOverflow question dataset comprising around 7k rows. Below is the code where I train and test a Zero Shot Classifier. X_train, X_test, y_train, y_test = train_test_split(_soQuestions['Body'],…
renakre
  • 8,001
  • 5
  • 46
  • 99
0
votes
0 answers

Best Large Language Model for abstractive Summarization

Browsing through Huggingfaces I am not able to find any abstractive summarization model for longer texts like newsarticle, with a size of around 5000 characters. I compared different models but none of them really worked. Mostly they are restricted…
0
votes
0 answers

Langchain MRKL Agent not giving useful Final Answer

Here is the code I'm using for initializing a Zero Shot ReAct Agent with some tools for fetching relevant documents from a vector database: chat_model = ChatOpenAI( model_name="gpt-3.5-turbo", temperature="0", …
0
votes
1 answer

How are LLMs assigned tasks in Python code?

I'm following Nicholas' Renotte's tutorials on VSCode, LangChain, and OpenAI using Python. These are the codeblocks I've seen from the aforementioned tutorials, and I don't see any other lines of code that tell the AI what to do. title_template =…
0
votes
0 answers

How do LLMs manage grammatical generation like "is a" or "is an"?

LLMs are forward generating models, generating the next word based on the previous context without having known the future words. This being the case the model shouldn't be able to know the difference between is a and is an grammatically as this…
KaranJ
  • 48
  • 1
  • 10
0
votes
0 answers

Can I load the weight parameter on DRAM permanently and use it whenever it's needed?

I'm using HuggingFace and debugging NLLB-MoE with the VSCode debugger. But the model parameters are so big and take so long time for a single execution. Can I just make parameter be loaded on the CPU DRAM and use it whenever needed? The code is…
0
votes
1 answer

Error while installing lmql[hf] using pip: "No matching distribution found for lmql[hf]

I am trying to install lmql[hf] using the pip package manager in order to set up a local LMQL playground. Following the documentation, I ran the command pip install lmql[hf]. However, I encountered the following error: ERROR: Ignored the following…
0
votes
1 answer

Getting TypeError in Snowpark

Currently I am trying write vectorized UDF function in snowpark import pandas as pd from snowflake.snowpark.functions import pandas_udf from snowflake.snowpark.types import StringType @pandas_udf( …
0
votes
0 answers

Issue with langchain llm and huggingface. The SSL issue

I an getting an SSL when using huggingface hub when using langchain LLM. Below is the error. I tried options input my username and password and did not work.. please help HTTPSConnectionPool(host='api-inference.huggingface.co', port=443): Max…
Nick
  • 1
0
votes
1 answer

Faster initialization with config.init_device

I am running these codes tokenizer = AutoTokenizer.from_pretrained(model_name) model = AutoModelForCausalLM.from_pretrained(model_name, trust_remote_code=True) There is a message on the console stating... You are using config.init_device='cpu', but…
zoomraider
  • 117
  • 1
  • 9
0
votes
1 answer

Unsupervised fine-tuning on custom documents after the supervised fine tuning on general question-answers dataset. Will it be useful for GPT-2 model?

I know the formal way of training a GPT2 model on custom documents is to first do semi-supervised fine tuning on the text of the documents followed by supervised fine-tuning on question answers from the same documents. But the sole purpose of…