GPT-4 is the latest milestone in OpenAI’s effort to extend deep learning into Artificial Intelligence. GPT-4 is a large multimodal model (accepting image and text input, and emitting text output) that, while less capable than humans in many real-world scenarios, exhibits performance on a human level in various academic and professional benchmarks.
GPT-4 is the new revolution
GPT-4 is a large multimodal model capable of image processing and text inputs and produce text outputs. It has the potential to be used in a wide range of applications, such as dialog systems, text summarization, and machine translation.
Announcing GPT-4, a large multimodal model, with our best-ever results on capabilities and alignment:
March 14, 2023 • 18:00
One of the main objectives of the development of this new version is improve your ability to understand and generate natural language text, particularly in more complex scenarios. To test its capabilities in such scenarios, GPT-4 was even tested on a variety of tests originally designed for humans. It performs quite well on these assessments, often outperforming the vast majority of human test takers, usually the top 10% of test takers (this is in contrast to GPT-3.5, the version on which ChatGPT was based, which scored at bottom 10%).
GPT-4 is more creative and collaborative than ever. You can build, edit, and iterate with users on build tasks. technical and creative writingsuch as composing songs, writing scripts, or learning a user’s writing style. In casual conversation, the distinction between GPT-3.5 and GPT-4 can be subtle. The difference arises when the complexity of the task reaches a sufficient threshold: GPT-4 is more reliable, creative and capable to handle much finer instructions than GPT-3.5.
“We spent 6 months making GPT-4 more secure and more aligned. GPT-4 is 82% less likely to respond to requests for disallowed content and 40% more likely to produce reliable answers than GPT-3.5 in our internal evaluations.”
Despite this revolution, limitations are still recognized. “GPT-4 still has many known limitations that we are working to address, such as social bias, hallucinations, and conflicting indications. We encourage and facilitate transparency, user education, and increased AI literacy as society embraces these models. We also intend to expand the entry routes of people to shape our models ».
For now, GPT-4 is available in ChatGPT Plus and as an API for developers to build apps and services. Some of the brands with which alliances have already been announced include Duolingo, Be My Eyes, Morgan Stanley or Khan Academy.
What you can do thanks to GPT-4
Greg Brockman, president and co-founder of OpenAI, showed in the GPT-4 presentation some of the uses that this new version of the model can be put to.
For example, thanks to image recognition, allows you to analyze them and start conversations about them. If you upload a photo with what you have in the fridge and ask “what can I do with these ingredients?” it should give you back a full prescription. Also through Be My Eyes you can describe and understand what is the point of interest in an image, even divided into different boxes.
If you are a fan of the “lots of text” meme, now GPT-4 is a great ally to help you understand and synthesize large strings of text. GPT-4 is capable of handling more than 25,000 words of textenabling use cases such as long-form content creation, extended conversations, and document search and analysis.