You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Note: This is meant to be done after the first submission.
A promising extension to InterroLang (probably even warranting its own paper) would be to replace BERT-type models with a single general-purpose LLM (e.g. LLaMa or the already-in-place GPT-Neo parser) that performs all of the tasks reasonably well. This would be a more modern approach, since BERTs are slowly getting out-of-date and LLMs can now run on consumer hardware locally. This would cause some changes, however, which I will document in the following:
Write instructions for the various tasks, e.g.
" Please predict one of the following labels: <label_1> … <label_n> Prediction: "
Overhaul of the entire feature importance operation category (nlpattribute, globaltopk) using Inseq. ⚠️ How matrices of feature attributions would be verbalized into a response has yet to be determined.
Pre-compute predictions and explanations with the new LLM
We can also get rid of many smaller language models, e.g. the GPT-2 for CFE generation, the SBERT for semantic similarity.
Ideally, we would end up with one model (for the entire framework) that assesses itself. It would take care of
Parsing / Intent recognition
Prediction of downstream tasks
Feature attribution (nlpattribute, globaltopk)
Perturbations (CFE, adversarial, augment)
Semantic similarity
Rationalization
The only two parts of the pipeline that would remain rule-based are the dialogue state tracking (custom inputs, clarification questions, previous filters) and the response generation (currently template-based).
Note: This is meant to be done after the first submission.
A promising extension to InterroLang (probably even warranting its own paper) would be to replace BERT-type models with a single general-purpose LLM (e.g. LLaMa or the already-in-place GPT-Neo parser) that performs all of the tasks reasonably well. This would be a more modern approach, since BERTs are slowly getting out-of-date and LLMs can now run on consumer hardware locally. This would cause some changes, however, which I will document in the following:
" Please predict one of the following labels: <label_1> … <label_n> Prediction: "
We can also get rid of many smaller language models, e.g. the GPT-2 for CFE generation, the SBERT for semantic similarity.
Ideally, we would end up with one model (for the entire framework) that assesses itself. It would take care of
The only two parts of the pipeline that would remain rule-based are the dialogue state tracking (custom inputs, clarification questions, previous filters) and the response generation (currently template-based).
Resources
llama.cpp (Efficient execution of up to 7B models on CPUs)
RedPajama-INCITE-Instruct-3B (Hugging Face) – maybe better for rationalization?
RedPajama-INCITE-Chat-3B – maybe better for response generation?
The text was updated successfully, but these errors were encountered: