What does HackerNews think of JARVIS?

JARVIS, a system to connect LLMs with ML community. Paper: https://arxiv.org/pdf/2303.17580.pdf

Language: Python

#23 in Deep learning
Raw large language models have no inner-monologue or meta-cognition. But there are various systems which bolt planning and sub-goal systems onto LLMs.[1][2] Those have self-generated work, where LLMs generate tasks for subsystems, which return data used by the LLMS to generate further work. So they do have an inner monologue, which can be observed.

There are also systems which summarize what the LLM has been doing and feed the summaries back into the LLM prompts. The idea is to create medium-term memories. When a query comes in, the most important and relevant summaries are fetched and fed back as prompts for the new query.

Wolfram has bolted an LLM onto Wolfram Alpha. Microsoft has bolted an LLM onto Bing. This sort of thing is still rather ad-hoc at this point, but it sort of works.

Things are moving fast. There are 2022 papers on how LLMs can't plan, and 2023 papers on systems with LLMs that do. It's beginning to look like LLMs will be components of larger, more generally capable systems.

[1] https://github.com/microsoft/JARVIS

[2] https://arxiv.org/abs/2304.11477

You are all correct and this is something on our radar. We are considering changing the name to avoid any confusion.

Microsoft also seems to be using the "Jarvis" name in this repo https://github.com/microsoft/JARVIS, so it is not just the famous Iron Man series. Not sure how this will all play out.