Cointime

Download App
iOS & Android

This is All That is Known About GPT-4 and What You Should Fear

Validated Individual Expert

A new beginning.

That’s what most of us felt when we tuned in yesterday to OpenAI’s official presentation of GPT-4, the newest generation of their world-famous pre-trained transformer.

And although there are plenty of impressive features to cover, there are also many open questions that were left unanswered, questions that need answering.

Can we answer for ourselves what is OpenAI trying to hide and what the future holds for all of us?

Your best work partner

During the 30-minute presentation, while more than one-hundred-thousand people looked in awe, OpenAI chose very carefully what they showed and how they showed it.

Understandable, but concerning in some aspects.

Truth be told, even though it was a very rehearsed presentation, they still managed to add some extra improvisation by allowing some viewers to test the model, in order to portray an image of going off the script.

But one way or another, they showed some mind-bending features.

Summarizing using the letter Q only

You read that title correctly, not only is GPT-4 capable of summarizing a long text into one sentence in a few seconds, it is capable of doing so using only words that begin with a ‘g’, or with ‘q’ (this last one at the request of a viewer).

Source: OpenAI

Albeit the stupidity of the request, it showcases the amazing linguistical capabilities of this model, far exceeding those of the GPT-3.5 turbo, the one running in ChatGPT’s API.

Also, it proved capable of cross-summarizing and finding a common theme among two different texts, thanks to the fact that GPT can now handle up to 32,000 tokens.

From OpenAI’s own documentation we know that 1,000 tokens equate to approximately 750 words. That means that GPT-4 is capable of receiving and understanding texts of up to 24,000 words, around 14 times the length of this article in one go, which is far more than the previous limit.

This is not trivial, considering that the longer the input is, the better context we can provide to it to generate better responses.

But this was just the beginning of what we were going to see.

A coding beast

Greg Brockman, OpenAI’s chairman, then moved into giving examples of how can GPT-4 improve the coding experience for developers.

And guys, I was speechless.

Firstly, it created a Discord bot from scratch by illustrating one of GPT-4’s most impressive features, instruction following.

By telling the model it now was an “AI programming assistant” and giving it some instructions, the model created the bot in a blink of an eye.

But that wasn’t the most impressive thing by a large margin.

As Greg acknowledged, GPT-4’s training cutoff is 2021, which means that GPT-4 has no knowledge from that point onwards. And the problem was that Discord had updated its API recently in 2022, something GPT-4 of course wasn’t aware of.

With the actual ChatGPT that would be the end of the story, but not with GPT-4.

Demonstrating great few-shot capabilities, the model was capable of using the new information that Greg gave it to update its knowledge and create the bot while meeting Discord’s 2022 requirements.

GPT-4 receiving instructions to create the Discord bot. Source: OpenAI

Again, language and semantic capabilities from another world that, in this case, match a human programmer’s performance (even exceeding if we consider the velocity at which GPT-4 created the bot).

However, the show was far from over, as the model also was capable of debugging its code when Greg sent it an error message.

At this point one is tempted to think, will human programmers still exist in a few years, or will simply serve as code reviewers?

Yet, as with everything, OpenAI had scheduled the best for last… GPT-4's new vision capabilities.

The Multimodal Era

Although Microsoft Germany’s CTO already leaked this, GPT-4 is going to be multimodal, which in simple terms means that the model is capable of receiving and understanding input from text, images, or even video.

For such purposes, they premiered GPT-4's perception capabilities, showing various images (some of them from the viewers) and then having the model describe them in accurate detail.

But probably the most impressive of all examples was when Greg drew a mockup of a website on pen and paper, sent it as a photograph to the model, and asked it to become a programmer again and create an HTML/CSS/JS website based on the mockup.

Et voilà. An interactive website in 10 seconds.

The actual mockup and the generated interactive website. Source: OpenAI

Truly incredible.

Even so, OpenAI was capable of squeezing one more feature into the presentation, one I personally wasn’t expecting.

Fear taxes no more

Taxes. Yes, taxes.

GPT-4 helped with a tax deduction after it was told that its new role was ‘TaxGPT’, and was also given a very long explanation of a tax deduction.

Again, not only GPT-4 managed to understand the whole text, remarkable considering how complex these documents are, but while doing impressive calculations and providing the correct answer.

Considering how poor ChatGPT’s performance has been regarding mathematics this was genuinely spectacular, mainly because it was clarified that no calculator APIs were connected to the model.

However, we must be cautious and see if these capabilities can be easily brought to shame with more complex mathematics.

Arriving to this point the presentation was closed, causing several questions that people were expecting to go unanswered.

And the issue is that some are very concerning.

We. need. answers.

As I said earlier, many questions remain regarding Generative AI in general.

Size, size, and more size

My main getaway from the presentation was that GPT-4 was considerably superior in its language skills to past versions.

Consequently, I’m led to believe that the biggest improvement we’ve seen is due to the fact that the model is simply much bigger.

That is, they are leveraging the fact that Large Language Models benefit from “scaling laws”, which is a fancy way of describing that these models get linguistically better as they get bigger.

From the multimodality standpoint, although we don’t have access to the model, by looking at the Kosmos-1 paper we can assume that they made important changes to the encoder part of the model, especially with regard to the visual transformers used to interpret images better.

But what about one of the most expected new features, video?

Video? Where?

With last week's news, we knew that GPT-4 would handle video. However, after yesterday’s presentation, we still don’t know in what sense.

As an input that it can then describe? Or will it be capable of actually generating videos?

Microsoft Germany’s CTO was unclear on this and so was OpenAI in the official launch. I guess we will have to wait on this matter.

But we still haven’t tackled the most concerning issue:

There’s an overarching issue with Generative AI in general, and that’s none other than its tendency to make stuff up.

Unreliability and business can’t be friends

Working in business environments, besides some specific cases like Programming or Marketing teams, no one will dare leverage Generative AI as a client-faced solution simply because it’s not reliable.

Sincerely, I was expecting that OpenAI would tackle this issue head-on in yesterday’s presentation, and the fact that they didn’t screams that the problem persists.

Yes, models are getting more reliable as they get bigger, and yes improvements on the Reinforcement Layer are allowing them to make better decisions, but the risk is undoubtedly still there.

From a business perspective, I still consider these solutions as great enhancers, solutions to make non-intelligent technologies like process automation or rule-based chatbots more intelligent, but never as standalone solutions.

To me, it still doesn’t seem like an option except for use cases with a high tolerance for error.

I mean, would you risk using a Chatbot that lies to your client? Of course not.

But what happens in those places where GPT-4 does fit?

Oh my.

Serious threat

In spite of Greg’s unwavering efforts to portray GPT-4 as a partner and not as a substitutor, which is something I agree with, the productivity enhancements will be immense.

Therefore, unless you’re different or unique, artists and writers are going to have a tough time with GPT-4. For instance, content writers seem completely substitutable, besides editors and others that will continue to play a role.

In a way, GPT-4 is going to elevate mediocre people into putting a fight against well-established artists and writers. Thus, I feel that only the greats in their fields will continue to thrive.

Having a unique perspective, a unique style, or a unique message will be key differentiators in this AI world that’s coming.

So, are you really special?

A final word

If you’ve read this article, you’re now ahead of 95% of society when it comes to AI.

But that still leaves a lot of people at that level.

So what if you’re capable of being above 99% of society?

That’s a totally different level, and if that’s where you want to be, I have news for you.

Read more: https://medium.com/@ignacio.de.gregorio.noblejas/gpt-4-released-81f8fc697def

AI
Comments

All Comments

Recommended for you

  • Web3 AI platform ChainML completes $6.2 million seed round of financing

    Web3 AI platform ChainML has announced the completion of a $6.2 million seed round of expansion financing, led by Hack VC, with participation from Inception Capital, HTX Ventures, Figment Capital, Hypersphere Ventures, and Alumni Ventures. The platform also announced the launch of its agent-based foundation layer, Theoriq.

  • Metaverse project Baby Shark Universe completes seed round financing

    Baby Shark Universe project, a metaverse project, has completed a seed round of financing with a valuation of $34 million. Participating investors include Animoca Brands, CREDIT SCEND, Sui Foundation, Comma3 Ventures, Creditcoin, GM Ventures, Neuler, Notch Ventures, X+, and Planetarium. The specific amount has not been disclosed, and the new funds will be used for development and global marketing. According to reports, Baby Shark Universe is an open-world role-playing game where players can create their own game content (items, maps), enjoy content created by other players, and expand the game's narrative based on their choices and actions.

  • Hong Kong Stock Exchange Confirms Crypto ETFs Unavailable to Mainland Chinese Investors

    According to Coindesk, the Hong Kong Stock Exchange has confirmed that cryptocurrency ETFs are not available to mainland Chinese investors. Hong Kong's cryptocurrency ETFs will provide a means to bypass capital controls in mainland China due to their unique physical redemption model.

  • Web3 social infrastructure UXLINK completes $5 million in financing

    Web3 social infrastructure UXLINK announced the completion of a new round of $5 million financing, led by SevenX Ventures, INCE Capital, and HashKey Capital. It is reported that UXLINK's total financing has now exceeded $15 million.

  • Chinese police bust underground bank using cryptocurrency for illegal currency conversion

    Chinese police have arrested six people for running an illegal currency conversion operation that used cryptocurrency to handle around $296 million. The operation was discovered by the Public Security Bureau of Panshi City, Jilin, and involved an "underground bank" that exploited the anonymity and ease of cross-border transfers offered by crypto. The operation used domestic accounts to receive and transfer funds, and exchanged between the yuan and South Korean won. The service was used by Korean purchasing agents, e-commerce firms, and import/export companies, among others.

  • Hong Kong Securities Regulatory Commission warns the public to beware of a suspicious asset investment product called "LENA Network"

    Hong Kong Securities and Futures Commission warned the public to be wary of a suspicious virtual asset investment product called "LENA Network". The product involves pledging and lending arrangements related to virtual assets, and claims to provide high returns to investors. This investment product has not been approved by the Securities and Futures Commission for sale to the Hong Kong public. The Securities and Futures Commission notes that the Hong Kong public can access information about the product and contact the product through the Internet. The Securities and Futures Commission advises against trusting those "too good to be true" investment opportunities and remaining vigilant when making investment decisions.

  • Volume 182: Digital Asset Fund Flows Weekly Report

    First inflows in 5 weeks totalling US$130m, hesitant Ethereum investors

  • Hong Kong Securities and Futures Commission: The Anti-Money Laundering Ordinance applies to the virtual asset industry

    The "virtual currency to ETF" mechanism in Hong Kong has raised concerns about money laundering. The industry believes that the review difficulty, such as KYT (Know Your Token), is high. Some individuals with mainland backgrounds are trying to conduct small-scale "virtual currency to ETF" transactions, taking the opportunity to "whiten" their own holdings of ether and bitcoin through forms such as personal accounts. They have also deployed some virtual currencies to Hong Kong's virtual currency exchanges and will decide whether to increase capital in the future depending on the situation. When responding to relevant questions, the Hong Kong Securities and Futures Commission emphasized that in the operation of ETF products, every link in the entire virtual asset ecosystem, including fund companies, custodians, asset trading platforms, participating brokers, etc., must be licensed or recognized institutions and strictly comply with requirements such as asset custody, liquidity, valuation, information disclosure, and investor education. The "Anti-Money Laundering Ordinance" of the Securities and Futures Commission also stipulates that financial institutions and designated non-financial enterprises and industry personnel must comply with customer due diligence and record-keeping requirements, and relevant regulations apply to the virtual asset industry.

  • TON community member: Some TON wallets received virtual account NFTs starting with "888", which is a phishing project

    On May 13th, according to a member of the TON official community, a new NFT with a virtual number starting with "888" has been added to the TON wallet. However, the transaction fee for each transfer is as high as 1 TON, which is caused by the fishing project changing the Gas.

  • Swiss Crypto Bank Amina: Listing Ethereum as a Security Could Cause Many Crypto Teams to Exit the Space

    Swiss encrypted bank Amina stated in the latest "Cryptocurrency Market Monitoring" report that classifying Ethereum as a security could not only bring risks to the entire cryptocurrency market, but also lead to many cryptocurrency teams exiting the field. This determination could hinder the development of the cryptocurrency market and potentially reverse progress made over the years. In addition, the US SEC is likely to delay its decision on the status of Ethereum, putting the cryptocurrency asset in a "gray area".