Grunnleggende konsepter
Practitioners encounter a variety of issues when developing and using LLM open-source software, including problems with the model, components, parameters, answers, performance, code, installation, documentation, configuration, network, and memory. The most common issues are related to the model, followed by component and parameter issues. The main causes are model problems, configuration and connection issues, and feature and method problems. The predominant solution is to optimize the model.
Sammendrag
The study identified and analyzed the issues, causes, and solutions in the development and use of LLM open-source software.
Key Findings:
-
Model Issue is the most common issue faced by practitioners, accounting for 24.55% of the issues. This includes problems with model runtime, architecture, loading, training, preprocessing, selection, fine-tuning, collaboration, testing, and updating.
-
The top three causes of the issues are:
- Model Problem (e.g., model instability, unreasonable architecture)
- Configuration and Connection Problem (e.g., incompatibility between components, missing key parameters)
- Feature and Method Problem (e.g., issues with function implementation, class design)
-
Optimize Model is the predominant solution to address the issues, which includes improving model performance, fine-tuning, and updating.
The study provides a comprehensive understanding of the challenges faced by practitioners of LLM open-source projects, the underlying causes, and the potential solutions. This can help guide the optimization and development of LLM open-source software.
Statistikk
The model is loaded into memory without any errors, but crashes on generation of text.
InstructorEmbedding is not found, which is a component used to form the embedding layer of the model.
An error was thrown when loading with Exllama or Exllamav2 even though pip indicates they are installed.
The parameter 'sources' is empty but it should be included in a list called 'result' as a string.
The results both from Azure OpenAI and from OpenAI are really random and have nothing to do with prompts.
Our deployment gives only 50 English words in 6 seconds.
The software "started lagging when it got past 3 lines and can take up to a minute to complete".
The method "'max_marginal_relevance_search()' was not implemented", which led to the failure of word embedding.
The download speed falls to 100 KB or something" after 5% when downloading the localized LLM.
An error message that "OutOfMemoryError: CUDA out of memory. – train dolly v2" occurred.
Sitater
"the model is loaded into memory without any errors, but crashes on generation of text"
"InstructorEmbedding is not found"
"An error was thrown when loading with Exllama or Exllamav2 even though pip indicates they are installed"
"the parameter 'sources' is empty but it should be included in a list called 'result' as a string"
"The results both from Azure OpenAI and from OpenAI are really random and have nothing to do with prompts"
"Our deployment gives only 50 English words in 6 seconds"
"The software "started lagging when it got past 3 lines and can take up to a minute to complete"
"the method "'max_marginal_relevance_search()' was not implemented"
"the download speed falls to 100 KB or something" after 5% when downloading the localized LLM
"OutOfMemoryError: CUDA out of memory. – train dolly v2"