InformatikEnglischGhost

Research Graph

Research Graph
Research Graph
StartseiteRSS-Feed
language
Large-language-modelsArtificial-intelligencePrompt-engineeringInformatikEnglisch
Veröffentlicht

Author Dhruv Gupta ( ORCID : 0009–0004–7109–5403) Introduction Large Language Models (LLMs) have become the new face of Natural language processing (NLP). With their generative power and ability to comprehend human language, the human reliance on these models is increasing every day. However, the LLMs have been known to hallucinate and thus produce wrong outputs.

MegalodonLong-textsTransformer-architectureInformatikEnglisch
Veröffentlicht

An improvement architecture superior to the Transformer, proposed by Meta Author · Qingqin Fang ( ORCID: 0009–0003–5348–4264) Introduction Recently, researchers from Meta and the University of Southern California have introduced a model called Megalodon. They claim that this model can expand the context window of language models to handle millions of tokens without overwhelming your memory.

Large-language-modelsArtificial-intelligenceTransformersNatural-language-processInformatikEnglisch
Veröffentlicht
Autor Wenyi Pi

Understanding the Evolutionary Journey of LLMs Author Wenyi Pi ( ORCID : 0009–0002–2884–2771) Introduction When we talk about large language models (LLMs), we are actually referring to a type of advanced software that can communicate in a human-like manner. These models have the amazing ability to understand complex contexts and generate content that is coherent and has a human feel.

Natural-language-processiTransformersArtificial-intelligenceInformatikEnglisch
Veröffentlicht

Attention mechanism not getting enough attention Author Dhruv Gupta ( ORCID : 0009–0004–7109–5403) Introduction As discussed in this article, RNNs were incapable of learning long-term dependencies. To solve this issue both LSTMs and GRUs were introduced. However, even though LSTMs and GRUs did a fairly decent job for textual data they did not perform well.

NaturallanguageprocessingLstmArtificial-intelligenceRecurrent-neural-networkInformatikEnglisch
Veröffentlicht

The Three Oldest Pillars of NLP Author Dhruv Gupta ( ORCID : 0009–0004–7109–5403) Introduction Natural Language Processing (NLP) has almost become synonymous with Large Language Models (LLMs), Generative AI, and fancy chatbots. With the ever-increasing amount of textual data and exponential growth in computational knowledge, these models are improving every day.

Large-language-modelsFrameworkRetrieval-augmentedInformatikEnglisch
Veröffentlicht

A Unified and Collaborative Framework for LLM Author · Qingqin Fang ( ORCID: 0009–0003–5348–4264) Introduction In today’s rapidly evolving field of artificial intelligence, large language models (LLMs) are demonstrating unprecedented potential. Particularly, the Retrieval-Augmented Generation (RAG) architecture has become a hot topic in AI technology due to its unique technical capabilities.