Tech Watch #4 — October, 27, 2023
The State of AI report is pretty interesting to read (even if long!). Among the major sections: research, industry, but also politics, safety, and some predictions. You’ll find an executive summary in one slide, on slide #8.
Read more...
On #22, emergent capabilities of LLMs is covered and mentions Stanford’s research that talks about the importance of more linear and continuous measures as otherwise capabilities sound like they emerge out of the blue.
On #23, they talk about the context length of LLMs being the new parameter count, as models try to have bigger context windows.
However, on slide #24, they also talk about researchers who showed that in long context windows the content provided in the middle is more ignored by LLMs compared to content at the beginning or end of the window.
So be sure to put the important bits first or last, but not lost in the middle.
Slide #26 speaks about smaller models trained with smaller curated datasets and can rival 50x bigger models.
Slide #28 wonders if we’re running out of human-generated data, and thus, if we’re going to have our LLMs trained on… LLM generated data!