LLM Archives | Oxide AI https://oxide.ai/tag/llm/ Quantified AI Decisions™ Tue, 10 Jun 2025 12:15:36 +0000 en-US hourly 1 https://wordpress.org/?v=6.7.1 https://oxide.ai/wp-content/uploads/2022/04/o_icon.svg LLM Archives | Oxide AI https://oxide.ai/tag/llm/ 32 32 Finding Trustworthy Signals in a Sea of Financial Data https://oxide.ai/2025-06-10-finding-trustworthy-signals-in-a-sea-of-financial-data/ Tue, 10 Jun 2025 12:07:33 +0000 https://oxide.ai/?p=10823 Scaling domain-tuned AI without the cost of proprietary models.

The post Finding Trustworthy Signals in a Sea of Financial Data appeared first on Oxide AI.

]]>

Finding Trustworthy Signals in a Sea of Financial Data

Scaling domain-tuned AI without the cost of proprietary models

Share This post

Proud moment for Oxide AI!


We’re super excited to be featured in Meta LLaMA’s latest case studies, which highlights our mission to bring transparency, openness and precision to AI in finance.

 

At Oxide AI, we believe that true innovation happens out in the open. That’s why we’ve used LLaMA —  no black boxes, no closed-source shortcuts. It’s the only way to build trust and promote human-AI collaboration where it makes most sense.

 

What this means in practice:

 

  • Open & Transparent: We build using openly available large language models.
  • Exact & Fine-Tuned: Precision isn’t optional. Our SLMs/LLMs are fine-tuned to specific domains. Combined with our computational AI, they deliver accurate, trustworthy insights.
  • Robust Signal Discovery: In complex financial datasets, our AI surfaces meaningful signals helping you make smarter decisions.

 

Oxogen isn’t just another AI tool. It’s the result of extensive work to pioneer a culture of open innovation where financial accuracy meets verifiable results.

 

A big thank you to the teams at Meta & IBM for spotlighting our collaboration!

 

Check out the case study and don’t forget to grab the full PDF version!

More To Explore

The post Finding Trustworthy Signals in a Sea of Financial Data appeared first on Oxide AI.

]]>
10823
Put AI to Work with Oxide AI https://oxide.ai/2025-05-26-join-us-at-ibms-put-ai-to-work-event-in-stockholm-this-week/ Mon, 26 May 2025 20:18:59 +0000 https://oxide.ai/?p=10730 Don't miss Oxide AI and other innovators at IBM's event in Stockholm this week!

The post Put AI to Work with Oxide AI appeared first on Oxide AI.

]]>

Put AI to Work with Oxide AI

Don’t miss Oxide AI and other innovators at IBM’s event in Stockholm this week!

Share This post

This week you want to make sure you attend IBM‘s excellent Put AI to Work Event in Stockholm. Enterprises are leaning in on AI but how do you get to sustainable value creation, strong trust and higher competitiveness?

Oxide AI will be there so come and say hi 👋 . Also, don’t miss Anders Tylman-Mikiewicz talk on how Hybrid-AI systems already drive tangible results and value in financial research and analysis for global organisations in finance – in ways that help them further enhance their competitive edge.

 

Let’s put AI to work for real!


Program and registration 

More To Explore

The post Put AI to Work with Oxide AI appeared first on Oxide AI.

]]>
10730
AI and the Progress Toward the Minimal and Relevant  https://oxide.ai/2025-05-02-ai-and-the-progress-toward-the-minimal-and-relevant/ Fri, 02 May 2025 09:32:25 +0000 https://oxide.ai/?p=10619 Learn how we can amplify humans with AI

The post AI and the Progress Toward the Minimal and Relevant  appeared first on Oxide AI.

]]>

AI and the Progress Toward the Minimal and Relevant 

Learn how we can amplify humans with AI

Share This post

We inhabit an era saturated with data. Every day, the volume of information grows faster than any human mind can track, let alone understand. The cognitive burden is immense, overwhelming even the most disciplined human attention.  

 

Consider the financial markets: news reports, earnings filings, analyst notes an endless stream floods investors every hour. For a single investor, or even an entire team, combing through data on thousands of public companies to find a few gleaming insights isn’t just impractical; it’s impossible. It would take years to do what the market demands by nightfall. And tomorrow, it begins anew. 

 

The question isn’t whether we can keep up. We can’t. The financial information storm breeds confusion, decision fatigue, and, inevitably, delegation. We seek refuge in abstraction: index funds like ETFs, algorithmic trading, or outsourced advice, because the raw data is too vast to navigate alone. And somewhere along the way, the slow, imperfect work of being human begins to fade. 

 

 

Enter Generative Artificial Intelligence  

 

Initially perceived as a tool to help navigate complexity, generative AI now threatens to amplify the information overload exponentially. The velocity and volume of content creation are set to explode. But this acceleration comes with a serious caveat: inherent probabilistic inaccuracies. Current Generative AI models operate within accuracy ranges often cited between 60% and 90%.  

 

Impressive, but not foolproof. And in a system that increasingly feeds itself, even small errors don’t just accumulate, they cascade. As AI systems rely more heavily on AI-generated content for training and output, we face the growing risk of compounding error, a phenomenon that could quietly erode informational integrity over time.  

 

Generative AI models also face a structural limitation: they are trained on the past while the world keeps moving. To generalize, they reduce entropy, smoothing out complexity but erasing critical variation along the way. Some newer systems try to fix this by incorporating fresh web data using so-called “deep research” models that search and rank sources dynamically. But adding more data does not always mean adding more truth. The biases baked into search algorithms and the uneven quality of the selected online sources creep into the results, dressed up as insights. Rather than offering a clear window into the real-time market, these models often reinforce pre-existing distortions under the appearance of freshness. 
 

The Counter-Reaction: The Quest for Quality 

 

This escalating information noise, paradoxically fueled by AI itself, sparks a powerful countertrend: an accelerating demand for the minimal and the relevant. As the volume swells and baseline quality becomes questionable, a premium will be placed on truth and concision. We will see a gravitation toward rigorously validated, authoritative sources, those we can trust not only to inform but to guide us. Sophisticated aggregation and verification systems, potentially AI-powered, but focused on quality control rather than pure generation, will emerge as indispensable gatekeepers. In this new information economy, trust and accuracy will be the currencies we can no longer afford to overlook. 

 

The initial fascination with AI’s ability to generate extensive outputs — a 50-page report drafted in minutes, a comprehensive presentation whipped up instantly — is likely to be short-lived. Why? Because human attention is finite and precious. No one has the bandwidth to consume endless streams of AI-generated verbosity, especially when its reliability is uncertain. It’s the same old story: we’re promised the world, but we can’t keep up. So what happens next? The inevitable move toward summarization. This, of course, only adds another layer of potential distortion, selection bias, and error accumulation, further highlighting the diminishing returns of bulk generation.  

 

The explosion of chatbots, once a novelty, may soon lead to interaction fatigue. Imagine a world populated by millions of automated conversational agents, all interacting with each other, but never with us. What’s left? The diminishing value of a single, meaningful conversation. 

 

The Interface Revolution: An Imperative for Precision 

 

Perhaps the most compelling driver toward minimalism is the evolution of our digital interfaces, increasingly mediated by AI. We are moving beyond the screen-centric paradigms of desktops and smartphones, inching closer to a world where technology quietly blends into the background of our lives. AI is making this shift possible, driving the practical integration of ambient computing: intelligent Head-Up Displays (HUDs) in vehicles, AI-infused smartwatches, AI headphones, and eventually, the mainstream arrival of Augmented Reality (AR) eyewear, the much-anticipated AI Glasses. 

 

 

These emerging platforms demand radically different interaction models. Displaying dense paragraphs of text or complex data visualizations on a car’s windshield or inside AR glasses is not just impractical; it’s irrelevant and distracting. Information delivered through these channels must be distilled to its absolute essence: minimal, contextually relevant, and instantly actionable. The tolerance for superfluous data collapses.  

 

As a result, the requirements for precision and accuracy will skyrocket. It is no longer enough for information to be broadly correct. It must be exactly the right piece of information, delivered at exactly the right time, in exactly the right way. 

 

The trajectory appears clear. While Generative AI may add to the overwhelming noise in its early stages, its long-term impact, combined with evolving hardware and the unchanging limits of human cognition, will force a profound recalibration. We are being propelled toward an information ecosystem where value lies not in volume, but in validated relevance and minimalist precision.

 

The future of AI belongs not to those who generate the most information, but to those who can uncover the signal worth hearing. In the end, it is not about outpacing human limits. It is about honouring them. This is where Oxide AI draws the line: amplifying humans, not overwhelming them.

 

More To Explore

The post AI and the Progress Toward the Minimal and Relevant  appeared first on Oxide AI.

]]>
10619
Blog: LLMs and Their Environmental Footprint  https://oxide.ai/2024-12-08-blog-llms-and-their-environmental-footprint/ Sun, 08 Dec 2024 20:08:21 +0000 https://oxide.ai/?p=7712 Explore environmental challenges related to LLMs and learn practical steps to mitigate the footprint.

The post Blog: LLMs and Their Environmental Footprint  appeared first on Oxide AI.

]]>

Blog: LLMs and Their Environmental Footprint 

how Could we use AI more efficiently?

Share This post

LLMs, such as OpenAI’s ChatGPT, Google’s Gemini and Meta’s Llama have reshaped our interaction with technology, enabling generation of realistic texts, image creation, and answering complex queries. However, their rise comes with increasing environmental costs. In this blog post, we’ll explore these environmental challenges and offer some practical recommendations for mitigating LLMs environmental footprint. 

 

Energy Consumption of LLMs 

 

  • Deep Learning Foundation: LLMs rely on extensive data and complex calculations. Training these models involves processing large datasets through multiple layers in a neural network, consuming substantial energy. 

  

  • Uniform Energy Usage: LLMs use the same amount of energy regardless of the task complexity because the model operates at maximum capacity for every input. The length (number of tokens) for input and output is what matters, not how difficult the query is. 

 

  • Interaction Complexity: Interacting with LLMs, such as ChatGPT, is generally straightforward as most human interactions are. However, due to randomness in generating varied responses and the need to correct errors or hallucinations, achieving accurate results often requires multiple interactions. 

 

  • Explicit Communication Requirements: Unlike human interactions that often rely on implicit cues, like experiences and body language, LLMs require explicit inputs (context or query). This explicitness requirement increases processing and energy consumption to a large extent. 

 

All these factors contribute to a significant carbon footprint compared to the value delivered. 

 

 

Practical Advice for Using LLM’s Responsibly 

 

A few suggestions from Oxide AI’s team how we can all help to reduce the environmental impact from modern AI. 

 

1. Prioritize Complex Tasks: Use LLMs for challenging problems, such as understanding query intent, resolving ambiguities, and complex information extraction. Simple tasks can often be solved by much less demanding AI technologies, especially for high-volume problems.

 

2. Use Smaller Models: Avoid overusing LLMs for well-defined tasks. Fine-tune smaller, specialized models for domain-specific tasks. Explore vendors offering robust small models, such as

    • IBM Watsonx: Comprehensive models and tools for fine-tuning in their cloud environment. 
    • Microsoft Azure: Phi models in Azure for high-quality, specialized language models. 
    • Meta AI: Small, efficient models like the Llama series (8B). 

 

3. Implement Caching: For frequent, similar requests, use a cache or simpler AI to recognize and retrieve precomputed results, reducing redundant processing and saving energy.

 

4. Optimize Database Interactions: For tasks involving heavy database interactions (e.g., document stores), consider architectural improvements. Utilize Retrieval Augmented Generation (RAG) models or embedding + vector database lookups to reduce GPU-intensive processing. Alternatively, use LLMs to translate natural language requests into database queries for efficient processing with much lower energy impact.

 

5. Minimize Number of Interaction Cycles: Make smart prompts with necessary context information in a single request. LLMs can efficiently tackle complex tasks with high-quality prompts and relevant context. Avoid approaches with brute-force solutions sampling from LLMs. 

 

6. Hierarchical Models: Try Mixture of Experts (MoE) models that assign tasks to specialized sub-models (“experts”), reducing energy use by activating only parts of LLMs. Another possibility is using a hierarchy of smaller language models. 

More To Explore

The post Blog: LLMs and Their Environmental Footprint  appeared first on Oxide AI.

]]>
7712
Personalized AI Agents for Investment Opportunity Discovery https://oxide.ai/2024-10-23-personalized-ai-agents-for-investment-opportunity-discovery/ Wed, 23 Oct 2024 11:40:32 +0000 https://oxide.ai/?p=7629 In the latest Oxogen release, we introduce personal, strictly private AI Agents, enabling you to finally assemble your own team of agents to scout the market for opportunities.

The post Personalized AI Agents for Investment Opportunity Discovery appeared first on Oxide AI.

]]>

Personalized AI Agents for Investment Opportunity Discovery

Oxide team

Your Superpower in Financial AI

Share This post

In the past, financial investment research was a cumbersome and time-consuming process, reserved for analysts in big institutions. Analyzing corporate financial statements, ETFs, options, and cryptocurrencies, the manual nature of these tasks required a lot of time and expertise. However, AI is changing the game. AI is transforming this once laborious process into an accessible and powerful tool for individual investors. Imagine commanding a team of intelligent AI Agents to do investment research for you, analyzing every company in the market every single day. With this newfound precision, you can stay informed and make truly educated decisions. 

 

Navigating the Ocean of Financial Data 

 

The sheer volume of digital financial data today often leads to information overload, analysis paralysis, and decision fatigue, especially in the financial sector. The common Fear of Missing Out might push you to analyze 10 companies in a day, but what about 10,000? It’s simply not possible for humans. 

 

To fully assess investment opportunity space, we need a broad scope that goes beyond basic financial data and real-time market updates. This includes exploring patent applications, research publications, news feeds, blog posts, websites, social media channels, and more. Without significant resources, this is only achievable with the help of AI. 

 

Protecting the value of our financial assets is a deeply personal and high-stakes endeavor. Simply stashing money in a savings account is not optimal for financial freedom. Here, AI proves invaluable. It sifts through vast amounts of data, prioritizing and organizing information to present clear, actionable insights. This isn’t about automated decision-making like robo-trading; it’s about knowing where to look. With a refined set of options, you can conduct realistic research further and make well-informed decisions. 
 

Discovery Automation with AI 

 

AI’s capacity for automated opportunity discovery is revolutionary. Leveraging massive amounts of data and sophisticated algorithms, AI can uncover investment opportunities with unprecedented speed and accuracy. It analyzes financial reports, market trends, and economic indicators to pinpoint viable opportunities, all while aligning with an investor’s unique objectives, risk tolerance, and thematic interests. This democratization of advanced financial analytics equips individual investors with insights once accessible only to seasoned experts and institutions. 
 

Personalized AI Agents: Your Own Dream Team 

 

One of the most exciting advancements in AI is the advent of personalized AI agents. These digital workers tirelessly research entire financial marketplaces, surfacing investment opportunities that match an individual’s preferences, financial objectives, constraints, and ethical principles. Imagine commanding your own team of AI researchers, each providing custom insights and recommendations. This isn’t just a minor adjustment; it’s a seismic shift in financial research, empowering individuals with the analytical prowess once exclusive to large institutions. 

 

Thanks to large language models (LLMs like ChatGPT), you don’t have to write code to define agents and their research objectives. Simply express your investment ideas, themes, and goals in natural language. Behind the scenes, the system creates a step-by-step reasoning plan that the AI Agent will execute. You just sit back as the agent discovers opportunities for you to review. 
 

Oxogen – Next Generation AI Tool for the Modern Investor 

 

Oxogen is an AI Agent system designed for financial opportunity discovery and thematic insights, ensuring you stay informed without information overload. A few months ago, we launched the first virtual team of AI agents, autonomously doing research across the entire Nasdaq, NYSE, and LSE (in the UK) to identify biotech/health opportunities. The health sector, known for its volatility and research-driven nature, can be challenging to navigate, but is full of interesting investment opportunities, all tracked by Oxogen. 

 

Since then, we’ve expanded our coverage to the entire emerging technology sector. Our AI system reclassifies the market, moving beyond outdated industry codes (NAICS, SIC, GICS, TRBC, etc.) to facilitate meaningful thematic exploration. The new categories span a wide range, including AI, EVs, 3D-printing, solar power, adtech, gaming, robotics, quantum computing, and more. 

 

Oxogen features a bolder graphical design, setting it apart from traditional financial tools and enhancing user experience. Beneath the sleek interface lies an advanced AI system capable of rapid, precise calculations across entire markets, quantitative step-by-step reasoning, understandable explanations and optimization towards specified objectives. 

 

In the latest Oxogen release, we introduce personal, strictly private AI Agents, enabling you to finally assemble your own team of agents to scout the market for opportunities.

 

 

Download the Oxogen app on your Apple or Android device and explore the future of investment research today.

 

 

Be the first to know how you can design your own team of AI Agents, watch our Demo creation video on our Youtube channel. 

More To Explore

The post Personalized AI Agents for Investment Opportunity Discovery appeared first on Oxide AI.

]]>
7629
Join Urban and Lars at IBM’s “Put AI to Work” https://oxide.ai/2024-08-30-join-urban-and-lars-at-ibms-put-ai-to-work/ Fri, 30 Aug 2024 11:05:55 +0000 https://oxide.ai/?p=7817 Learn about financial AI agents from Lars Hard at IBM's event "Put AI to Work".

The post Join Urban and Lars at IBM’s “Put AI to Work” appeared first on Oxide AI.

]]>

Join Urban and Lars at IBM’s “Put AI to Work”

Oxide team

Learn about financial AI agents from Lars Hard at IBM’s event “Put AI to Work”

Share This post

Join Oxide AI and IBM to explore how watsonx.ai can be applied in financial AI!

 

  • September 10
  • Stockholm, Sergelhub

 

Lars and Urban Roth from IBM will talk about application of AI in finance and banking (session B8, 11:45 AM).

 

Learn about financial AI agents from Lars Hard! Don’t miss out on this opportunity to learn and innovate.

 

Register via this link 

More To Explore

The post Join Urban and Lars at IBM’s “Put AI to Work” appeared first on Oxide AI.

]]>
7817
Fintech AI agent on Twitter/X https://oxide.ai/2024-02-02-fintech-ai-agent-on-twitter-x/ Fri, 02 Feb 2024 13:45:17 +0000 https://www.oxide.ai/?p=7302 Meet fintech AI agent that posts daily original financial insights on Twitter/X.

The post Fintech AI agent on Twitter/X appeared first on Oxide AI.

]]>

Fintech AI agent on Twitter/X

Meet fintech AI agent that posts daily original financial insights on twitter/x.

Share This post

Meet fintech AI agent that posts daily original financial insights on twitter/x. Every day Agent picks one of the interesting investment opportunities and generates a short financial text presenting a market insight, an opportunity to look into. You also get a ticker and a link to the company page for further actions. Remember this a junior agent with some learning curve, but it can only get better from this point.  

Follow oxogen on twitter/x https://twitter.com/oxogenai

More To Explore

The post Fintech AI agent on Twitter/X appeared first on Oxide AI.

]]>
7302
Blog: The Art of Reality Capture in the Age of Generative AI  https://oxide.ai/2023-03-08-blog-digital-evidence/ Wed, 08 Mar 2023 11:32:00 +0000 https://oxide.ai/?p=4659 In a digital context, evidence refers to any information that is used to support or refute a claim. In this post, we look at evidence in relation to Large Language Models (LLMs).

The post Blog: The Art of Reality Capture in the Age of Generative AI  appeared first on Oxide AI.

]]>

Blog: The Art of Reality Capture in the Age of Generative AI 

We look at evidence in relation to Large Language Models (LLMs)

Share This post

DIGITAL EVIDENCE WILL BE SKYROCKETING

 

In a digital context, evidence refers to any information that is used to support or refute a claim. Evidence is closely linked to the observation of events and the determination of facts. In any situation, evidence serves as the means by which we establish what happened, and what is true. The process of gathering evidence involves observing events and collecting data that can be used to support or refute a particular claim. 

 

As generative AI becomes increasingly prevalent, the importance of digital evidence is set to skyrocket. In a world where it’s difficult to discern what’s true if it’s generated, the value of evidence that’s closely related to reality will become crucial and highly valuable. While some applications like fiction, movie scripts, and game plots don’t necessarily require a strong relationship to reality, the same cannot be said for applications with high real-world consequences. Investing millions of dollars based on generated statements simply won’t cut it.  

 

To address this, we can harness the power of Transformers and LLMs, but we must combine them with other robust and explainable AI (XAI) techniques that operate in real-time to capture data as close as possible to the original source. This type of reality capture setup can also be used for reinforcement learning without human involvement.

 

A REALITY CHECK

 

 

In the picture above, a large language model featuring generative AI is extracting information from unstructured data. It works sidebyside with an AI ensemble (multiple models) to robustly capture perspectives in the data. Outputs from both models are compared to see if they agree. If not, the AI ensemble is used instead of the LLM (which may be hallucinating) and feedback can be passed to the LLM.

 

EVIDENCE MATTERS

 

The purpose of evidence gathering is to gain a complete understanding of the events in a given situation and to establish the facts of the case. This process can be challenging and involves analyzing all available information, which can be done using AI models capable of computing, analyzing, and evaluating different perspectives in data. These models differ from generative AI models because they must provide detailed explanations of everything, from algorithm insights to data sources, authority, references, data sample rate, and more. In essence, it is absolutely crucial to use AI models that offer complete transparency and accountability, allowing for thorough understanding and interpretation of the data. 

 

In summary, evidence in a digital context refers to any information used to support or refute a claim that is produced electronically. This information can be in various forms, and it is often at least partly unstructured data. Advanced technologies like LLMs, NLP and other AI models can be used to extract valuable insights from this data, but it needs to be collected, stored, and analyzed in a way that maintains its authenticity, integrity and transparency.

More To Explore

The post Blog: The Art of Reality Capture in the Age of Generative AI  appeared first on Oxide AI.

]]>
4659
Blog: Are Search Engines Doomed to be Replaced by ChatGPT? https://oxide.ai/2022-12-15-blog-llm-chatgpt-vs-search-engine/ Thu, 15 Dec 2022 17:07:00 +0000 https://oxide.ai/?p=4420 Are Large Language Models, LLM’s, the AI technology that will finally replace search as we know it today? The latest chat bot, ChatGPT from OpenAI, is hailed as a know-it-all candidate that could soon change the way we search. Looking somewhat deeper at the problem reveals that it may require significant innovations to even begin to get there.

The post Blog: Are Search Engines Doomed to be Replaced by ChatGPT? appeared first on Oxide AI.

]]>

Blog: Are Search Engines Doomed to be Replaced by ChatGPT?

The latest chat bot, ChatGPT from OpenAI, is hailed as a know-it-all candidate that could soon change the way we search.

Share This post

NATURAL LANGUAGE AND LARGE LANGUAGE MODELS (LLMS)

 

Language is a multifaceted phenomenon that can be described on neurological, biological, psychological, social, anthropological and even historical levels. It is the heart of what it is to be human, arguably more so than our emotions or art or music (those being culture-specific and not universal, unlike speaking), or non-linguistic cognitive facilities (which we share, to varying degrees, with other animals). Even people who cannot speak are submerged in a world which humans have created over hundreds of millennia through the use of words. Somehow, it feels as if trying to reduce this to a single “paradigm” or “theoretical framework” is to miss the point, like talking about a painting solely in terms of the chemical composition of the paint.

 

One easily gets the impression that the entire domains of natural language processing (NLP) and natural language understanding (NLU) have been reduced to transformer-based chat bot programs. Without a doubt, the latest LLMs—Large Language Models—are fun to play with, especially when going beyond simple conversations and into the territory of more creative activities. This category of models is able to “create” simple art, prose, music and even animations/videos. The act of giving the model the right prompt is already an art form in and of itself.  

 

We can also use this type of technology to write code. The quality of the output in this case is easier to assess, since machine-written code can be validated by running it and seeing whether it produces the desired results.  

 

Having a productive coding session with OpenAI’s ChatGPT makes a lot of sense on the surface. As long as we can describe what we want with any precision, the bot will try to compose complete functions and programs; even if these contain bugs or are written in an idiosyncratic style, it is in many cases faster to start from a 90% working program and fix it manually than to write it from scratch. In the future, automatic coding will likely also take run-time error messages and compiler messages into account, so a correct version of the final program can be generated in a single or just a few rounds.  

 

 

ChatGPT clearly solves many mundane, day-to-day coding tasks. It works exceedingly well where a very specific solution to an isolated problem is desired, such as creating a function for calculating some value. Asking for boilerplate code for talking to APIs also works extremely well, since there are so many examples it can “learn” how to perform this from. This way, ChatGPT-powered coding will likely save us a lot of time.  

 

Once we look at more complex coding tasks, we can see that there still is plenty of room for improvement. Furthermore, we also have to remember the old software engineering joke: “so many programmers, and so few people to tell them what to do.” So, one could question whether ChatGPT is a truly profound, world-changing development; all it might lead to is a replacement of some low-level coders with AI. At the very least, it seems like creative non-coders might be able to utilize the technology to achieve things that previously lay beyond their grasp.

 

SEARCH AND LLM’S

 

Search engines today rely to a great extent on traditional NLP, including parsing, chunking and other techniques. The largest search engines may even make use of LLMs to support search in various ways (e.g., query expansion, query understanding). But at the present, LLMs are still confined to a supporting role within the domain of large-scale search.  

 

This is because search is a slightly different problem than language generation. It is the process of locating information that already exists. Generative LLMs, on the other hand, build new things from old pieces of data. Even if we have an AI capable of generating responses to every possible query (which is easy to set up currently), it still does not mirror the actual, underlying reality, because reality is in many cases characterized by a lack of data and confusing noise. Granting an AI model enough introspection to know when it should not generate a certain output is a super-complex and unresolved problem.  If search is taken as a truth-finding process, we can clearly see why trust in an LLM might be misplaced. For one, by “transforming” the underlying data in various ways, LLMs introduce extra interpretation layers between data and output that might not be what the user wants. LLMs are also notorious for synthesizing (often called “hallucinations”); worse yet, they may output the results of their hallucinating in an authoritative writing style that implies subject expertise, when such is not actually warranted. 

 

In subjects such as health, this is dangerous and irresponsible. They are generating “answers,” not finding reliable answers. NLP can be useful for interpreting human statements and feelings, as one step in a more responsive—and responsible—process of seeking information. Other AI and search approaches then also need to be involved, alongside human interpretation. Context outside the scope of words matter profoundly for analysis and interpretation. 

 

GOING ALL-IN WITH AN LLM AS A SEARCH ENGINE

 

An LLM-powered search engine would be an interesting exercise. Unfortunately, several problems need to be resolved before this is realistic. In addition to the issues raised above, we can also mention the sheer cost of running such a service. One can imagine that an LLM engine would have to rely heavily on advertising to make revenue. We all know from the past that this road has only led to search monopolies and stagnation over the last 20 years or so (Read more on the topic in Next Search Possible).  
Many areas of the internet feature extremely few data points. If you try to build a catalog of all retail products on sale in online marketplaces, you will soon come to the realization that only a few percent (out of billions) come with any significant amount of useful data attached to them. For the rest, you will have to fill out gaps by means of guesswork. An LLM could make this arduous task somewhat simpler, but can we hope that it is accurate? And can we evaluate that compared with a human annotator or a simpler, rules-based algorithm for filling in missing data fields. When a buyer comes prepared to part with serious money to get hold of a rare item, it is questionable to what degree they would trust an LLM over their own or another human’s judgment.

 


Information reality is also about controlled sparseness, especially on the internet. Much information is fully locked down—by design—behind paywalls because the data is a financial asset to someone. Limiting information access to earn money by selling it is an ancient business model. Thus, an infinite number of cat images float around out there for an AI to learn how to synthesize felines from; but in contrast, try to construct a bot that will accurately answer knowledge-based questions and provide reliable services. If such a thing existed, many service businesses could likely collapse; hence, this information is tightly controlled and usually not available, or at least not for free. In this case, a traditional (non-LLM) search engine helps us find what is findable, ignores the rest since it cannot be reached, and it can do this without making things up or requiring huge data volumes.


As described above, areas where traditional search engines excel over theoretical search LLMs are rooted in trustworthiness and authority. LLMs do not, as a rule, care about truth—they generate their answers based on statistical relations mined from vast amounts of data fed to them in the training stage. For example, imagine you want to try going on a diet. How do you choose the most suitable one? The noise level in this space if enormous, as everyone knows who have tried searching anything diet-related. It’s a random game. We have biased input coming from all directions. The underlying science is weak. Commercial interests distort the available information. So, there are contradictions, ambiguity and bias everywhere. Is this really the kind of uninterpreted data we want to synthesize an answer from? Equating search with a generative statistical technology that understands absolutely nothing could potentially prove to be a big step backwards. 

 

PROS AND CONS FOR LLM’S AS SEARCH ENGINES

 

Problems with LLMs as Search Engines 

 

  • LLMs truly extends the concept of “one model to rule them all”. We’ve seen this over and over in the world of search and hopefully it will be replaced in the future by a more collaborative and open approach to search. 
  • LLMs currently require massive training and are very far from actually managing near-real-time data acquisition, which is what we expect from search engines. 
  • LLMs synthesize (generate) data. This synthesis, sometimes jokingly referred to as “hallucination,” may not always reflect reality. When searching, we look for reality, preferably from the source and not an interpretation/transformation. 
  • LLMs are hard to validate. It may even be difficult to produce the same results based on a given input, something we require from most scientific research work. This is the reproducibility problem. 
  • To block LLMs from producing results where they should not provide a result seems to be an unresolved problem. High frequency queries are currently managed by means of manual curation, but not so for queries far out on the long tail. The tail is very, very long in search. 
  • LLMs can explain their outputs in the same way as they generate answers. But this is the same shaky statistical ground they stand on, far from understanding which normally leads to explanations. There is no simple way we can understand the underlying parameter space, either. 
  • They are very energy-consuming. Training the largest LLMs takes massive amounts of energy. Having a development where everyone needs to train their own models is not sustainable, so a general discussion about access will arise. The “one model to rule them all” is far from sufficient for humanity.

 

Benefits with LLMs as Search Engines 

 

  • Transformers and similar Deep Learning tech have many beneficial advantages in the context of search. The current state is primarily supporting roles in search, rather than actually replacing search engines. A few examples: 
    • Query auto-completion 
    • Query expansion 
    • Query understanding (for example to extract query intent) 
    • Transformation of input queries to increase recall & precision 
    • Summarization of text to extract useful facts and explanations 
    • Language translation 
  • Guided search based on bot interaction, where a dialog with a user may lead to better formulated queries or insights. Using LLMs as an interface to search engines seems to be a promising approach, but of course at their current level tripped-up by the problems highlighted above.
  • Exploration and discovery functions, making use of creative synthesis. 
  • “Entertainment Search” is a possible area for LLMs today, that’s why we see so many fun chat bots emerging. 

More To Explore

The post Blog: Are Search Engines Doomed to be Replaced by ChatGPT? appeared first on Oxide AI.

]]>
4420