Connect with us

GOOGLE

Google’s Smith Algorithm Outperforms BERT

Published

on

smith algorithm 5ff6fc5dbef0a

Google recently published a research paper on a new algorithm called SMITH that it claims outperforms BERT for understanding long queries and long documents. In particular, what makes this new model better is that it is able to understand passages within documents in the same way BERT understands words and sentences, which enables the algorithm to understand longer documents.

On November 3, 2020 I read about a Google algorithm called Smith that claims to outperform BERT. I briefly discussed it on November 25th in Episode 395 of the SEO 101 podcast in late November.

I’ve been waiting until I had some time to write a summary of it because SMITH seems to be an important algorithm and deserved a thoughtful write up, which I humbly attempted.

So here it is, I hope you enjoy it and if you do please share this article.

Is Google Using the SMITH Algorithm?

Google does not generally say what specific algorithms it is using. Although the researchers say that this algorithm outperforms BERT, until Google formally states that the SMITH algorithm is in use to understand passages within web pages, it is purely speculative to say whether or not it is in use.

What is the SMITH Algorithm?

SMITH is a new model for trying to understand entire documents. Models such as BERT are trained to understand words within the context of sentences.

In a very simplified description, the SMITH model is trained to understand passages within the context of the entire document.

While algorithms like BERT are trained on data sets to predict randomly hidden words are from the context within sentences, the SMITH algorithm is trained to predict what the next block of sentences are.

This kind of training helps the algorithm understand larger documents better than the BERT algorithm, according to the researchers.

BERT Algorithm Has Limitations

This is how they present the shortcomings of BERT:

“In recent years, self-attention based models like Transformers… and BERT …have achieved state-of-the-art performance in the task of text matching. These models, however, are still limited to short text like a few sentences or one paragraph due to the quadratic computational complexity of self-attention with respect to input text length.

In this paper, we address the issue by proposing the Siamese Multi-depth Transformer-based Hierarchical (SMITH) Encoder for long-form document matching. Our model contains several innovations to adapt self-attention models for longer text input.”

According to the researchers, the BERT algorithm is limited to understanding short documents. For a variety of reasons explained in the research paper, BERT is not well suited for understanding long-form documents.

The researchers propose their new algorithm which they say outperforms BERT with longer documents.

They then explain why long documents are difficult:

“…semantic matching between long texts is a more challenging task due to a few reasons:

1) When both texts are long, matching them requires a more thorough understanding of semantic relations including matching pattern between text fragments with long distance;

2) Long documents contain internal structure like sections, passages and sentences. For human readers, document structure usually plays a key role for content understanding. Similarly, a model also needs to take document structure information into account for better document matching performance;

3) The processing of long texts is more likely to trigger practical issues like out of TPU/GPU memories without careful model design.”

Larger Input Text

BERT is limited to how long documents can be. SMITH, as you will see further down, performs better the longer the document is.

This is a known shortcoming with BERT.

This is how they explain it:

“Experimental results on several benchmark data for long-form text matching… show that our proposed SMITH model outperforms the previous state-of-the-art models and increases the maximum input text length from 512 to 2048 when comparing with BERT based baselines.”

This fact of SMITH being able to do something that BERT is unable to do is what makes the SMITH model intriguing.

The SMITH model doesn’t replace BERT.

The SMITH model supplements BERT by doing the heavy lifting that BERT is unable to do.

The researchers tested it and said:

“Our experimental results on several benchmark datasets for long-form document matching show that our proposed SMITH model outperforms the previous state-of-the-art models including hierarchical attention…, multi-depth attention-based hierarchical recurrent neural network…, and BERT.

Comparing to BERT based baselines, our model is able to increase maximum input text length from 512 to 2048.”

Long to Long Matching

If I am understanding the research paper correctly, the research paper states that the problem of matching long queries to long content has not been been adequately explored.

According to the researchers:

“To the best of our knowledge, semantic matching between long document pairs, which has many important applications like news recommendation, related article recommendation and document clustering, is less explored and needs more research effort.”

Later in the document they state that there have been some studies that come close to what they are researching.

But overall there appears to be a gap in researching ways to match long queries to long documents. That is the problem the researchers are solving with the SMITH algorithm.

Details of Google’s SMITH

I won’t go deep into the details of the algorithm but I will pick out some general features that communicate a high level view of what it is.

The document explains that they use a pre-training model that is similar to BERT and many other algorithms.

First a little background information so the document makes more sense.

Algorithm Pre-training

Pre-training is where an algorithm is trained on a data set. For typical pre-training of these kinds of algorithms, the engineers will mask (hide) random words within sentences. The algorithm tries to predict the masked words.

As an example, if a sentence is written as, “Old McDonald had a ____,” the algorithm when fully trained might predict, “farm” is the missing word.

As the algorithm learns, it eventually becomes optimized to make less mistakes on the training data.

The pre-training is done for the purpose of training the machine to be accurate and make less mistakes.

Here’s what the paper says:

“Inspired by the recent success of language model pre-training methods like BERT, SMITH also adopts the “unsupervised pre-training + fine-tuning” paradigm for the model training.

For the Smith model pre-training, we propose the masked sentence block language modeling task in addition to the original masked word language modeling task used in BERT for long text inputs.”

Blocks of Sentences are Hidden in Pre-training

Here is where the researchers explain a key part of the algorithm, how relations between sentence blocks in a document are used for understanding what a document is about during the pre-training process.

“When the input text becomes long, both relations between words in a sentence block and relations between sentence blocks within a document becomes important for content understanding.

Therefore, we mask both randomly selected words and sentence blocks during model pre-training.”

The researchers next describe in more detail how this algorithm goes above and beyond the BERT algorithm.

What they’re doing is stepping up the training to go beyond word training to take on blocks of sentences.

Here’s how it is described in the research document:

“In addition to the masked word prediction task in BERT, we propose the masked sentence block prediction task to learn the relations between different sentence blocks.”

The SMITH algorithm is trained to predict blocks of sentences. My personal feeling about that is… that’s pretty cool.

This algorithm is learning the relationships between words and then leveling up to learn the context of blocks of sentences and how they relate to each other in a long document.

Section 4.2.2, titled, “Masked Sentence Block Prediction” provides more details on the process (research paper linked below).

Results of SMITH Testing

The researchers noted that SMITH does better with longer text documents.

“The SMITH model which enjoys longer input text lengths compared with other standard self-attention models is a better choice for long document representation learning and matching.”

In the end, the researchers concluded that the SMITH algorithm does better than BERT for long documents.

Why SMITH Research Paper is Important

One of the reasons I prefer reading research papers over patents is that the research papers share details of whether the proposed model does better than existing and state of the art models.

Many research papers conclude by saying that more work needs to be done. To me that means that the algorithm experiment is promising but likely not ready to be put into a live environment.

A smaller percentage of research papers say that the results outperform the state of the art. These are the research papers that in my opinion are worth paying attention to because they are likelier to make it into Google’s algorithm.

When I say likelier, I don’t mean that the algorithm is or will be in Google’s algorithm.

What I mean is that, relative to other algorithm experiments, the research papers that claim to outperform the state of the art are more likely to make it into Google’s algorithm.

SMITH Outperforms BERT for Long Form Documents

According to the conclusions reached in the research paper, the SMITH model outperforms many models, including BERT, for understanding long content.

“The experimental results on several benchmark datasets show that our proposed SMITH model outperforms previous state-of-the-art Siamese matching models including HAN, SMASH and BERT for long-form document matching.

Moreover, our proposed model increases the maximum input text length from 512 to 2048 when compared with BERT-based baseline methods.”

Is SMITH in Use?

As written earlier, until Google explicitly states they are using SMITH there’s no way to accurately say that the SMITH model is in use at Google.

That said, research papers that aren’t likely in use are those that explicitly state that the findings are a first step toward a new kind of algorithm and that more research is necessary.

This is not the case with this research paper. The research paper authors confidently state that SMITH beats the state of the art for understanding long-form content.

That confidence in the results and the lack of a statement that more research is needed makes this paper more interesting than others and therefore well worth knowing about in case it gets folded into Google’s algorithm sometime in the future or in the present.

Citation

Read the original research paper:

Description of the SMITH Algorithm

Download the SMITH Algorithm PDF Research Paper:

Beyond 512 Tokens: Siamese Multi-depth Transformer-based Hierarchical Encoder for Long-Form Document Matching (PDF)

Searchenginejournal.com

Keep an eye on what we are doing
Be the first to get latest updates and exclusive content straight to your email inbox.
We promise not to spam you. You can unsubscribe at any time.
Invalid email address

GOOGLE

Google’s Next-Gen AI Chatbot, Gemini, Faces Delays: What to Expect When It Finally Launches

Published

on

By

Google AI Chatbot Gemini

In an unexpected turn of events, Google has chosen to postpone the much-anticipated debut of its revolutionary generative AI model, Gemini. Initially poised to make waves this week, the unveiling has now been rescheduled for early next year, specifically in January.

Gemini is set to redefine the landscape of conversational AI, representing Google’s most potent endeavor in this domain to date. Positioned as a multimodal AI chatbot, Gemini boasts the capability to process diverse data types. This includes a unique proficiency in comprehending and generating text, images, and various content formats, even going so far as to create an entire website based on a combination of sketches and written descriptions.

Originally, Google had planned an elaborate series of launch events spanning California, New York, and Washington. Regrettably, these events have been canceled due to concerns about Gemini’s responsiveness to non-English prompts. According to anonymous sources cited by The Information, Google’s Chief Executive, Sundar Pichai, personally decided to postpone the launch, acknowledging the importance of global support as a key feature of Gemini’s capabilities.

Gemini is expected to surpass the renowned ChatGPT, powered by OpenAI’s GPT-4 model, and preliminary private tests have shown promising results. Fueled by significantly enhanced computing power, Gemini has outperformed GPT-4, particularly in FLOPS (Floating Point Operations Per Second), owing to its access to a multitude of high-end AI accelerators through the Google Cloud platform.

SemiAnalysis, a research firm affiliated with Substack Inc., expressed in an August blog post that Gemini appears poised to “blow OpenAI’s model out of the water.” The extensive compute power at Google’s disposal has evidently contributed to Gemini’s superior performance.

Google’s Vice President and Manager of Bard and Google Assistant, Sissie Hsiao, offered insights into Gemini’s capabilities, citing examples like generating novel images in response to specific requests, such as illustrating the steps to ice a three-layer cake.

While Google’s current generative AI offering, Bard, has showcased noteworthy accomplishments, it has struggled to achieve the same level of consumer awareness as ChatGPT. Gemini, with its unparalleled capabilities, is expected to be a game-changer, demonstrating impressive multimodal functionalities never seen before.

During the initial announcement at Google’s I/O developer conference in May, the company emphasized Gemini’s multimodal prowess and its developer-friendly nature. An application programming interface (API) is under development, allowing developers to seamlessly integrate Gemini into third-party applications.

As the world awaits the delayed unveiling of Gemini, the stakes are high, with Google aiming to revolutionize the AI landscape and solidify its position as a leader in generative artificial intelligence. The postponed launch only adds to the anticipation surrounding Gemini’s eventual debut in the coming year.

Keep an eye on what we are doing
Be the first to get latest updates and exclusive content straight to your email inbox.
We promise not to spam you. You can unsubscribe at any time.
Invalid email address
Continue Reading

GOOGLE

Google Brings Bard Students Math and Coding Education in the Summer

Published

on

Google Brings Bard Students Math and Coding Education in the Summer

Google is stepping up its AI efforts this summer by sending Bard, its high-profile chatbot, to summer school. The aim? To boost the bot’s math and coding smarts. These developments are excellent news— when Bard first debuted, it was admittedly not a finished product. But Google is steadily plugging away at it, and have now implemented implicit code execution for logical prompts, and handy Google Sheets’ integration to take it to the next level.

Thanks to implicit code execution, Bard can respond to inquiries requiring calculation or computation with Python code snippets running in the background. What’s even more amazing is that coders can take this generated code and modify it for their projects. Though Google is still apprehensive about guaranteeing the accuracy of Bard’s answers, this feature is said to improve the accuracy of math and word problems by an impressive 30%.

In addition to this, Bard can now export directly to Sheets when asked about tables. So, you don’t need to worry about copying and pasting, which comes with the risk of losing formatting or data.

From the company’s I/O keynote address, it is clear that they are focused on making the most of what Bard can offer. As they continue to speak highly of the chatbot, we’re sure to expect more features and capabilities when the summer comes around.

Keep an eye on what we are doing
Be the first to get latest updates and exclusive content straight to your email inbox.
We promise not to spam you. You can unsubscribe at any time.
Invalid email address
Continue Reading

GOOGLE

Google Bard vs. ChatGPT: which is the better AI chatbot?

Published

on

Google Bard vs. ChatGPT: which is the better AI chatbot?

Google Bard and ChatGPT are two of the most prominent artificial intelligence (AI) chatbots available in 2023. But which is better? Both offer natural language responses to natural language inputs, using machine learning and millions of data points to craft useful, informative responses. Most of the time. These AI tools aren’t perfect yet, but they point to an exciting future of AI assistant search and learning tools that will make information all the more readily available.

As similar as these chatbots are, they also have some distinct differences. Here’s how ChatGPT and Google Bard measure up against one another.

Which is better, Google Bard or ChatGPT?

This is a tricky question to answer, as at the time of writing, you can only use Google Bard if you’re part of a select group of early beta testers. As for its competition, you can use ChatGPT right now, completely for free. You may have to contend with a waitlist, but if you want to skip that, there’s a paid-for Plus version offering those interested in a more complete tool the option of paying for the privilege.

Still, when Google Bard becomes more widely available, it should offer credible competition for ChatGPT. Both use natural language models — Google Bard uses Google’s internal LaMDA (Language Model for Dialogue Applications), whereas ChatGPT uses an older GPT-3 language model. Google Bard bases its responses to questions on more recent data, with ChatGPT mainly trained on data that was available prior to 2021. This is similar to how Microsoft’s Bing Chat works.

We’ll have to reserve judgment on which is the more capable AI chatbot until we get time to play with Google Bard ourselves, but it looks set to be a close contest when it is more readily available.

Are Google Bard and ChatGPT available yet?

As mentioned, ChatGPT is available in free and paid-for tiers. You might have to sit in a queue for the free version for a while, but anyone can play around with its capabilities.

Google Bard is currently only available to limited beta testers and is not available to the wider public.

Banner of Google Bard intro from February 6.

What’s the difference between Google Bard and ChatGPT?

ChatGPT and Google Bard are very similar natural language AI chatbots, but they have some differences, and are designed to be used in slightly different ways — at least for now. ChatGPT has been used for answering direct questions with direct answers, mostly correctly, but it’s caused a lot of consternation among white collar workers, like writers, SEO advisors, and copy editors, since it has also demonstrated an impressive ability to write creatively — even if it has faced a few problems with accuracy and plagiarism.

Still, Microsoft has integrated ChatGPT into its Bing search engine to give users the ability to ask direct questions of the search engine, rather than searching for terms of keywords to find the best results. It has also built it into its Teams communications tool, and it’s coming to the Edge browser in a limited form. The Opera browser has also pledged to integrate ChatGPT in the future.

ChatGPT Google Bard
Accessible through ChatGPT site. Only text responses are returned via queries. Integrated with Google Search. You only need to change a Google setting to get your regular search results when using Google Bard AI, and vice versa.
ChatGPT produces answers from its trained database from 2021 and before. Google Apprentice Bard AI will be able to answer real-time questions.
Based on GPT (Generative Pre-trained Transformer). Based on LaMDA (Language Model for Dialogue Applications).
Service has a free and paid plan option (called ChatGPT Plus). Service is free.
Has built-in plagiarism tool called GPT-2 Output Detector. No built-in plagiarism detection tool.
Available now Still in beta test phase

Google Bard was mainly designed around augmenting Google’s own search tool, however it is also destined to become an automated support tool for businesses without the funds to pay for human support teams. It will be offered to customers through a trained AI responder. It is likely to be integrated into the Chrome browser and its Chromium derivatives before long. Google is also expected to open up Google Bard to third-party developers in the future.

Under the hood, Google Bard uses Google’s LaMDA language model, while ChatGPT uses its own GPT3 model. ChatGPT is based on slightly older data, restricted in its current GPT3 model to data collected prior to 2022, while Google Bard is built on data provided on recent years too. However, that doesn’t necessarily make it more accurate, as Google Bard has faced problems with incorrect answers to questions, even in its initial unveiling.

ChatGPT also has a built-in plagiarism checker, while Google Bard does not, but Google Bard doesn’t have the creative applications of ChatGPT just yet.

Keep an eye on what we are doing
Be the first to get latest updates and exclusive content straight to your email inbox.
We promise not to spam you. You can unsubscribe at any time.
Invalid email address
Continue Reading

Trending