GPT-4 Takes a Holiday Break, Microsoft Responds to Gemini, OpenAI Partners with Journalists, & New LLM Compression Obviates GPUs
LLMs may take it easy over the holidays, Microsoft responded to Google's Gemini performance metrics, OpenAI announced a partnership with journalists, and a new compression method could obviate GPUs
GPT-4 Might Be Taking a Winter Break
We may not anthropomorphize large language models enough, as even they take it easy during the holidays 🎄. This makes sense, when you consider all of the material LLMs have consumed about holiday breaks and seasonal depression.
Rob Lynch discovered that GPT-4's API produces shorter completions when it "thinks" the month is December (orange) compared to May (blue). This is being measured while many users are complaining about GPT-4 getting lazy.
While this sounds too to be true, the differences are statistically significant, and have already been independently reproduced. Some are calling this phenomenon the "AI Winter Break Hypothesis".
Microsoft Shows GPT-4 Surpassing Gemini with Advanced Prompting
Microsoft responded to Google's claim that their large language model Gemini Ultra was superior to OpenAI's GPT-4.
According to Google, Gemini Ultra was surpassing GPT-4 in 30 out of 32 metrics (some of which were questionable). However, after Microsoft introduced a sophisticated prompt composition strategy called Medprompt, GPT-4 pushed back ahead.
While this certainly shows GPT-4's latent capabilities, it also highlights how advanced prompting techniques can unlock the full potential of large language models.
OpenAI Partners with Axel Springer to Integrate Journalism in AI
OpenAI entered a multiyear licensing agreement with Axel Springer, a major news-publishing corporation with brands like Business Insider, Politico, and Bild.
Under the agreement, ChatGPT will link to the original Axel Springer sources when using their material in responses, boosting visibility and traffic for the publisher. The financial details of the deal remain undisclosed, but it promises significant revenue for Axel Springer.
Additionally, the partnership allows Axel Springer to leverage OpenAI's technology to enhance their journalistic products. While the deal isn't exclusive, it sets a precedent for future interactions between AI developers and news organizations, addressing the industry's concerns over AI-driven content usage without proper attribution or compensation. This agreement follows a similar, though more limited, arrangement between OpenAI and The Associated Press.
New LLM Compression Technique Obviates Need for GPUs
A white paper came out this week introducing LLM compression based on reduced order modeling, which entails low-rank decomposition within the feature space and re-parameterization in the weight space.
This technique obviates the need for a GPU device, and the associated computation cost is quite small. Their LLM-ROM method outperforms LLM-Pruner at 80% and 50% compression without any fine-tuning.