Google revealed that it was adding creative artificial intelligence to its search engine. This could be one of the biggest changes to Google’s search engine. But most Web authors are worried that the new Google search could hurt the number of people who visit their sites.
At its yearly developer meeting, Google said it would use AI models to bring together information from all over the Internet. Google said these so-called “generative search experience” goods could help people get better answers to their search queries.
Google will show some users text sections made by artificial intelligence and prioritise a few related links on the search results page instead of the “ten blue links” that Google’s search results usually show.
The new Google Search that uses AI is being tried with a small group of users and has yet to be opened to everyone. But many online publishers are worried that if this becomes Google’s usual way to show search results, more people will stay on Google’s website. This could mean less traffic to their website, which could hurt online publishing and the income of online merchants.
The debate also shows there have been problems between Google and the sites it crawls for a long time. There is no doubt that the rise of new artificial intelligence tools has made this situation worse. Web authors have been worried for a long time that Google would merge bits of content from their sites. By employing sophisticated machine learning models, Google may now “train” AI to provide results with comparable content and context.
Negative feedback from web owners
Rutledge Daugette, CEO of TechRaptor, a site for game news and reviews, said that Google’s move should have considered publishers’ needs and that Google’s artificial intelligence means that material gets more attention.
According to Daugette’s comments to CNBC, “their emphasis is on queries with no clicks.” They steal the work of publishers and authors that put in a lot of effort and time to create quality content for the sake of popularity. ” As of now, AI is fast to borrow data that it doesn’t require from other sources. When it comes to Google, Bard doesn’t even disclose the sources it relies on for its own results.”
Yelp’s head of public affairs, Luther Lowe, has been critical of Google’s search rules for a long time. He said that the changes to Google Search were part of a long-term plan to keep people on Google sites longer instead of sending them to the area that have already provided them with the data they need.
In an interview, Lowe said, “Google’s introduction of the ChatGPT clone into the search space made it exclusive, which is the final chapter in the bloodletting of the whole web.”
Search Engine Land, a news site that keeps a close eye on changes to Google’s search engine, says that in tests, the AI-generated results are shown above the normal search results. CNBC had already reported that Google planned to change its results page to highlight material made by AI.
SGE comes in a box with a different colour, green in this case, and has links to three websites in boxes on the right side. In Google’s main case, the headlines for all three websites were cut off.
Google says the information doesn’t come from the websites but the links back up the data. Search Engine Land noted that the SGE method was better and a “healthier” way to link than Google’s Bard robot, which rarely linked to publishing websites.
Some producers want to know if they can stop Google and other AI companies from using their material to train their models. Data owners have already sued companies like the company behind Stable Diffusion, but the right to scrape web data for AI is still up in the air. Other companies, like Reddit, have said they will charge people to use their data.
IAC owns sites like All Recipe, People Magazine, and Daily Beast. Barry Diller is one of the most important people in the printing business. “In a meeting earlier this month, he made the following remark: “You could have as much information as you needed if you could dump it into this blender and reform it into chat-like declarative statements”, there would be no printing business because it would be impossible.”
For Diller, the solution was simple: “All you’ve got to do is convince the industry that we are unable to steal our content unless we are able to come up with a way enabling online publishers to get paid access.” He claimed that Google will be responsible for fixing the problem.
Diller said that he thinks online publishers can sue AI companies based on copyright law and that “fair use” needs to be changed to include more things. A report from Wednesday said that a group of online publisher leaders led by Diller said, “If we have to, we will change the copyright laws.”
Publishers have to make sure that AI is using their material. Google didn’t share the training sources for the large language model that SGE PaLM 2 is based on, and Daugette says that even though he’s seen competitor citations and review scores used in Bard without credit, it’s hard to tell when the information came from his site without direct details—source links.
A Google representative said the company has no plans to share training info about pay with authors.
Google announced that they will be offering a new generative AI experience as a Search Labs project in order to “iterate and improve based on user and stakeholder input.”
According to the developers, “PaLM 2 is trained using a lot of public data from the Internet, and we constantly monitor the health of the web ecosystem.” In a recent press conference, Google’s VP of Research, Zoubin Ghahramani, said, “It’s a big part of the way we think about creating our products to ensure we have an optimal environment where creators are part of that ecosystem.”
Daugette says that Google’s moves make it hard to be a self-published book.