- cross-posted to:
- technology@lemmy.ml
- cross-posted to:
- technology@lemmy.ml
Google has plunged the internet into a “spiral of decline”, the co-founder of the company’s artificial intelligence (AI) lab has claimed.
Mustafa Suleyman, the British entrepreneur who co-founded DeepMind, said: “The business model that Google had broke the internet.”
He said search results had become plagued with “clickbait” to keep people “addicted and absorbed on the page as long as possible”.
Information online is “buried at the bottom of a lot of verbiage and guff”, Mr Suleyman argued, so websites can “sell more adverts”, fuelled by Google’s technology.
You probably tried the free version. Check perplexity.ai to see how the paid version of chatgpt works. Every source is referenced and linked.
This guy is not talking about the current version of free chatgpt. He's talking of the much better tools that will be available in the next few years
Yeah, because people selling AI products have a great track record on predicting how their products will develop in the future. Because of that, Teslas don't have steering wheels any more, because Full Self Driving drives people incident-free from New York to California since 2017.
The thing with AI development is, that it rapidly gets to 50% of the desired solution, but then gets stuck there, not being able to get consistently good enough that you can actually rely on it.
I don't really understand what it means. If the product is unreliable people won't use it, and everything will stay as it is now. It's not a big issue. But It is already pretty reliable for many use cases.
Realistically the real future problem will be monetization (which is causing the issues of Google), not features
Well, here's the thing. How often are you willing to dismiss the misses because of the hits? Your measure of unreliability is now subject to bias because you're no longer assessing the bot's answers objectively.
I don't expect it to be 100% correct. I have realistic expectations built on experience. Any source isn't 100% reliable. A friend is 50% reliable, an expert maybe 95. A random web page probably 40… I don't know.
I built up my strategies to address uncertainty by applying critical thinking. It is not much different than in the past. By experience, chatgpt 4 is currently more reliable than a random web page that comes in the first page of a Google search. Unless I exactly search for a trustworthy source, such as nhs or guardian.
The main problem is the drop in quality of search engines. For instance, I often start with chatgpt 4 without plugins to focus my research. Once I understand what I should look for, I use search engines for focused searches on official websites or documentation pages.
The issue with reliability is a completely different one between web search and AI.
If you search something on Google, there are quite a few ways you can judge the quality of the answer with "metadata" around it. If you find a scientific paper, it's probably more reliable than a post on a parents forum. If the source is a quality newspaper or Wikipedia, that's also more on the reliable side, but some conspiracy theorist website is not. And if the source is some kind of forum or Q&A site, wrong answers often have comments under them that correct the error.
Also, you can follow multiple links and take a wider sample on the topic that way.
With AI that's not possible. Whether it is wrong or correct, the AI will give you an answer in the exact same format, with the same self-confident tone. You basically need to know the correct answer to know whether the answer is correct.
Sure, you can re-roll and ask it again, but that doesn't make the result more likely to be correct.
For example, I asked ChatGPT which Harry Potter chapter is the longest. It happily gave me a chapter, but it wasn't the longest. So I asked again and again and again, and each time it gave me a new wrong answer, every time with made-up word counts.
deleted by creator
This is the reason I am suggesting people to give a try to perplexity.ai to understand how these tools will work in the near future. And why I don't understand the reason I am downvoted for that.
Current "free" chatgpt was created as a proof of concept, not as a finished, complete solution for humanity issues. What we have now is a showcase of llm, for openai to improve the product via human feedback, for everyone else to enjoy what is it already now, with all its limitations, an extremely useful tool.
But this kind of LLM is intended to be a building block of the future solutions. To enable interactivity, summarization, analysis features within larger products with larger and more refined set of features.
If you don't have paid version of chatgpt, again, try perplexity.ai with the copilot feature, to see a (still imperfect, under development) proof of concept of the near future of AI assisted research.
And more tools will come, that will make easier to navigate the huge amount of information that is the main issue of modern internet.
For your specific case, gpt 3.5 has poor logical and mathematical capabilities. Gpt-4 is much better with that. But still, using a language model for math is almost never a good choice. What you'd need, is an llm able to access information from the internet and to have access to some math tool, such as python or Matlab. These options currently are available on chatgpt with plugin, but they are suboptimal. In the future you'll have better product able to combine llm, focused internet search and math.
We should focus on the future, not on the present when discussing AI. LLMs based products are in their infancy