Searching for information is the quintessential misconception about LLMs being helpful or improving other existing technology. In my opinion,聽web search is a more effective way to find information, simply because search engines give the user what they want faster, and in a format that fits the purpose much better than a chatting tool: multiple sources to scroll through in a purpose-built user interface, including filter options, configuration settings, listed elements, excerpts, tabulated results, whatever you get in that particular web search tool… not the “I’m here to assist you, let’s delve into the intricacies of the ever-evolving landscape of X…” followed by a long perfectly composed paragraph based on a probabilistic model you would typically get when you send a prompt to ChatGPT asking for factual information about a topic named ‘X’.

In today’s blog post, I am introducing one of the latest GPTs and Assistants I created, named HumbleAI. I will let the models explain it by answering a few questions. For each question, I’ve selected a response I liked or found worthy of sharing. At the end of the post, you can find the links to all the complete chats, and my scores based on an SCBN (Specificity, Coherency, Brevity, Novelty) benchmark.

Black and white line drawing generated by the ControlNet Canny model, depicting a woman in a wetsuit holding a surfboard on the beach. A text from CLIP Interrogator, describing an image, is superimposed over the drawing and reads: "a woman in a wet suit holding a surfboard on the beach with waves in the background and a blue sky, promotional image, a colorized photo, precisionism."

The intense competition in the chatbot space is reflected in the ever-increasing amount of contenders in the LMSYS Chatbot Arena leaderboard, or in my modest contribution with the SCBN Chatbot Battles I’ve introduced in this blog and complete as time allows. Today we’re exploring WildVision Arena, a new project in Hugging Face Spaces that brings vision-language models to contend. The mechanics of WildVision Arena are similar to that of LMSYS Chatbot Arena. It is a crowd-sourced ranking based on people鈥檚 votes, where you can enter any image (plus an optional text prompt), and you will be presented with two responses from two different models, keeping the name of the model hidden until you vote by choosing the answer that looks better to you.聽I’m sharing a few examples of what I’m testing so far, and we’ll end this post with a traditional ‘SCBN’ battle where I will evaluate the vision-language models based on my use cases.

Two computer monitors on a desk, each displaying a facial expression of annoyance or frustration. The monitor on the left displays a start menu with the label "Microsoft 365 Copilot", while the monitor on the right has a browser window open with tabs, one of which is labeled "ChatGPT Plus" and features the OpenAI logo. The setting suggests an office environment, and the background reveals a blue tiled wall, implying an indoor setting without any visible Windows. [Image produced with DALL-E and Gimp] [Caption assisted by ALT Text Artist GPT]

Microsoft has just announced the launch of its own 鈥楪PT Builder鈥 for customizing chatbots, similar to OpenAI鈥檚 鈥楪PT Store鈥. This was part of a broader announcement of Copilot Pro, a premium AI-powered service for Microsoft 365 users to enhance productivity, code, and text writing. According to Satya Nadella鈥檚 announcement today on Threads, Microsoft and OpenAI appear to be competing entities, yet they are working on the same technology (GPT), augmented by Microsoft鈥檚 investment in OpenAI. It certainly seems like a strange business strategy for Microsoft. Please provide some insight into the move鈥檚 rationale and strategic motivations.