yoshiwaan
yoshiwaan t1_j96uxg7 wrote
Reply to comment by blueSGL in [D] Toolformer implementation using only few-shot prompting by MysteryInc152
Really? As in the order of operations is: token parsing => Toolformer => LLM?
Genuine question, is the text/token parsing for queries to an LLM (eg chatgpt) performed separately and beforehand to the actual LLM being leveraged, or is the text/token parsing a part of the LLM? I figured it was the latter and you couldn’t just insert a tool there
yoshiwaan t1_j96wt2g wrote
Reply to comment by MysteryInc152 in [D] Toolformer implementation using only few-shot prompting by MysteryInc152
Really? As in the order of operations is: token parsing => Toolformer => LLM?
Genuine question, is the text/token parsing for queries to an LLM (eg chatgpt) performed separately and beforehand to the actual LLM being leveraged, or is the text/token parsing a part of the LLM? I figured it was the latter and you couldn’t just insert a tool there
Edit: I think this is a new model for this purpose, rather than reusing an existing LLM (eg ChatGPT) as I first assumed, which makes more sense
Edit 2: I actually read the paper and the LM itself is taught to reach out to tools as a part of its response operations, it’s not something separate