Afeerah Naseem, Author at MarkTechPost https://www.marktechpost.com/author/afeerah-naseem/ An Artificial Intelligence News Platform Wed, 05 Feb 2025 03:40:18 +0000 en-US hourly 1 https://wordpress.org/?v=6.8.1 https://www.marktechpost.com/wp-content/uploads/2022/04/cropped-Favicon-512-x-512-1-1-32x32.png Afeerah Naseem, Author at MarkTechPost https://www.marktechpost.com/author/afeerah-naseem/ 32 32 127842392 Meet Crossfire: An Elastic Defense Framework for Graph Neural Networks under Bit Flip Attacks https://www.marktechpost.com/2025/02/04/meet-crossfire-an-elastic-defense-framework-for-graph-neural-networks-under-bit-flip-attacks/ https://www.marktechpost.com/2025/02/04/meet-crossfire-an-elastic-defense-framework-for-graph-neural-networks-under-bit-flip-attacks/#respond Wed, 05 Feb 2025 03:40:05 +0000 https://www.marktechpost.com/?p=68648 Graph Neural Networks (GNNs) have found applications in various domains, such as natural language processing, social network analysis, recommendation systems, etc. Due to its widespread usage, improving the defences of GNNs has emerged as a critical challenge. While exploring the mechanisms vulnerable to an attack, researchers came across Bit Flip Attacks (BFAs). Conventionally, BFAs were […]

The post Meet Crossfire: An Elastic Defense Framework for Graph Neural Networks under Bit Flip Attacks appeared first on MarkTechPost.

]]>
Graph Neural Networks (GNNs) have found applications in various domains, such as natural language processing, social network analysis, recommendation systems, etc. Due to its widespread usage, improving the defences of GNNs has emerged as a critical challenge. While exploring the mechanisms vulnerable to an attack, researchers came across Bit Flip Attacks (BFAs). Conventionally, BFAs were developed for Convolutional Neural Networks (CNNs), but recent developments have shown that these are extendable to GNNs. Current methods of defence that GNNs have critical limitations; they either cannot entirely restore the network after the attack or require expensive post-attack evaluations. Therefore, researchers at the University of Vienna have developed a novel solution, Crossfire, that can effectively use the existing defence mechanisms and restore the networks. 

Bit-flipping attacks manipulate individual bits within a deep learning model’s binary code. This considerably weakens the model’s performance, creating serious security risks. Honeypots and hashing-based defences are prominent current defence mechanisms. Honeypot defences function by including several decoy elements within the system; any alteration to one or more elements may indicate an attack. Attackers, however, now bypass these weights. Hashing-based defences use strong cryptographic hashing to detect changes in weights. They cannot, however, fix the resulting damage.

The proposed model, Crossfire, is an adaptive, hybrid model that detects BFAs by honeypot and hashing-based defences and restores the model after an attack using a bit-level weight correction. The key-mechanism of Crossfire are:

  • Bit-wise Redundancy Encoding: Crossfire sets some weights to zero to decrease the number of active weights in the GNN. This guides the attackers to less critical weights, preventing substantial damage. Hashing continuously monitors the active weights, detecting any changes. Honeypot weights are strategically placed to attract attackers and quickly identify if they are attacked. 
  • Elastic Weight Rectification: First layer hashes identify where the alteration has been made after the attack, then row and column hashes point out the exact location. Corrections are done using honeypot at the bit level or zeroed if other options fail. 

Across 2,160 experiments, Crossfire demonstrated a 21.8% higher probability of reconstructing an attacked GNN to its pre-attack state than competing methods. The framework improved post-repair prediction quality by 10.85% on average. Crossfire maintained high performance for up to 55-bit flips from various attacks. Furthermore, the framework’s adaptive nature allows it to dynamically allocate computational resources based on detected attack severity, making it an efficient and scalable solution.

In conclusion, Crossfire considerably improves the resilience of GNN defences against bit-flip attacks with a new, efficient and highly effective adaptive method. Crossfire’s highly dynamic response carefully adjusts to the severity of attacks, guaranteeing strong security and outstanding efficiency and setting a decisively new standard for securing GNNs in challenging adversarial environments. Because it’s scalable and practical, it offers a promising way to improve the reliability of GNN-based applications across multiple fields.


Check out the Paper. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter and join our Telegram Channel and LinkedIn Group. Don’t Forget to join our 75k+ ML SubReddit.

🚨 Marktechpost is inviting AI Companies/Startups/Groups to partner for its upcoming AI Magazines on ‘Open Source AI in Production’ and ‘Agentic AI’.

The post Meet Crossfire: An Elastic Defense Framework for Graph Neural Networks under Bit Flip Attacks appeared first on MarkTechPost.

]]>
https://www.marktechpost.com/2025/02/04/meet-crossfire-an-elastic-defense-framework-for-graph-neural-networks-under-bit-flip-attacks/feed/ 0 68648
Dendritic Neural Networks: A Step Closer to Brain-Like AI https://www.marktechpost.com/2025/02/02/dendritic-neural-networks-a-step-closer-to-brain-like-ai/ https://www.marktechpost.com/2025/02/02/dendritic-neural-networks-a-step-closer-to-brain-like-ai/#respond Mon, 03 Feb 2025 04:43:48 +0000 https://www.marktechpost.com/?p=68571 Artificial Neural Networks (ANNs) have their roots established in the inspiration developed from biological neural networks. Although highly efficient, ANNs fail to embody the neuronal structures in their architects truly. ANNs rely on vast training parameters, which lead to their high performance, but they consume a lot of energy and are prone to overfitting. Due […]

The post Dendritic Neural Networks: A Step Closer to Brain-Like AI appeared first on MarkTechPost.

]]>
Artificial Neural Networks (ANNs) have their roots established in the inspiration developed from biological neural networks. Although highly efficient, ANNs fail to embody the neuronal structures in their architects truly. ANNs rely on vast training parameters, which lead to their high performance, but they consume a lot of energy and are prone to overfitting. Due to the continuous increase in the complexity and depth of ANNs, there has been an exponential growth in energy usage, which is becoming increasingly difficult to sustain. Therefore, researchers from Institute of Molecular Biology and Biotechnology, Foundation for Research and Technology-Hellas, Heraklion, Crete, Greece have developed a novel solution, dendritic ANNs, that has significantly captured the characteristics of dendritics in the neurons.

Traditional ANNs excel at solving complex tasks but require massive amounts of trainable parameters to achieve high accuracy. Each node in the complex network represents a specific class, which is an efficient way of distinguishing features. Still, it is inflexible as it faces problems adapting to different tasks. Moreover, it makes them prone to overfitting, making generalizability an issue. Therefore, there is a need for a new method that can maintain or increase its performance when the number of parameters is reduced and has improved generalizability to unseen data. 

The proposed solution, dendritic ANNs, is designed to better leverage the structural and functional efficiency observed in neurons. The most significant innovation of ANNs is multi-class responsiveness, which allows for more precise and resilient learning. The dANNs try to mimic the structural connectivity of biological neurons, reducing random connections to process information more efficiently. The dendrites focus only on a subset of input data, which filters out the noise and focuses only on relevant information. These breakthroughs allow the model to train on a significantly smaller number of parameters compared to traditional ANNs. 

To better understand the different features of the biological neurons that can be leveraged in ANNs, the researchers came up with four variants. The key features of each of the variants are:

  • dANN-LRF (Local Receptive Fields): Each dendrite focuses on a small input sample, demonstrating the power of localised processing in reducing parameter count while maintaining high accuracy. This variant achieves the highest efficiency, 
  • dANN-R (Random Sampling): Input features are randomly sampled for each dendrite. This variant serves the purpose of understanding whether the sampling increases the efficiency or the dendritic structure itself. It proved to be beneficial for tasks having unclear spatial relationships of features.  
  • dANN-GRF (Global Receptive Fields): It focuses on capturing localised features to understand the spatial arrangement, for example, of objects in an image. 
  • pdANN (Pyramidal dANN): Investigates whether adding more biological realism through hierarchical structure can improve performance or generalisation. Although there was no significant improvement in accuracy, it reduced overfitting. 

The dANNs were tested on several datasets, including CIFAR-10 and Fashion-MNIST. Their accuracy and performance consistently matched or exceeded that of the best vanilla ANNs (vANNs) across all datasets. dANN-LRF achieved peak accuracy and minimal loss and it used greatly fewer trainable parameters than vANNs. dANNs showed improved performance and stability as the number of layers increased, effectively dealing with scalability issues often found in biologically-inspired models. dANNs showed better efficiency when performing complex tasks, like those using the CIFAR10 dataset.

dANNs offer a new way to build artificial neural networks. This approach uses ideas from how biological dendrites work. Their learning is highly accurate, remarkably strong and exceptionally parameter-efficient. This substantially improves conventional architectures, creating stronger and more sustainable AI systems. Bio-inspired design offers important promise for improvements in artificial intelligence. This approach could lead to the development of several clever, energy-efficient systems.


Check out the Paper. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter and join our Telegram Channel and LinkedIn Group. Don’t Forget to join our 75k+ ML SubReddit.

🚨 Meet IntellAgentAn Open-Source Multi-Agent Framework to Evaluate Complex Conversational AI System (Promoted)

The post Dendritic Neural Networks: A Step Closer to Brain-Like AI appeared first on MarkTechPost.

]]>
https://www.marktechpost.com/2025/02/02/dendritic-neural-networks-a-step-closer-to-brain-like-ai/feed/ 0 68571
Revolutionizing Heuristic Design: Monte Carlo Tree Search Meets Large Language Models https://www.marktechpost.com/2025/01/24/revolutionizing-heuristic-design-monte-carlo-tree-search-meets-large-language-models/ https://www.marktechpost.com/2025/01/24/revolutionizing-heuristic-design-monte-carlo-tree-search-meets-large-language-models/#respond Sat, 25 Jan 2025 05:38:38 +0000 https://www.marktechpost.com/?p=68328 Heuristic designing is a practical and indispensable tool leveraged in standard fields like artificial intelligence and operations research to find satisfactory solutions to complex optimisation problems. Experts usually design them by hand, which makes the process expensive and slow. A simplification of heuristics design, without a reduction in performance, was subsequently achieved through the Automatic […]

The post Revolutionizing Heuristic Design: Monte Carlo Tree Search Meets Large Language Models appeared first on MarkTechPost.

]]>
Heuristic designing is a practical and indispensable tool leveraged in standard fields like artificial intelligence and operations research to find satisfactory solutions to complex optimisation problems. Experts usually design them by hand, which makes the process expensive and slow. A simplification of heuristics design, without a reduction in performance, was subsequently achieved through the Automatic Heuristic Design (AHD) proposal. It relied on several human-defined parameters, thereby limiting its adaptability as well as its effectiveness. AHD was recently integrated with LLMs, employing a strong population-based framework. However, the framework was designed to pick the first solution it found and converged quickly, which prevented it from exploring much better options, resulting in less effective optimisation strategies. In order to tackle these obstacles, a team of researchers from National University of Singapore and Southern University of Science and Technology China have developed MCTS-AHD, the first tree search method for LLM-based AHD. 

Current LLM-based methods have been efficient, yet they are in need of a more tailored approach so that they do not converge on the local optima but rather explore the vast array of opportunities available. These methods also have inadequate search mechanisms, leading to insufficient investigation of the possible heuristics. They are single-objective focused, limiting their adaptability in real-world scenarios where multiple objectives must be considered. These inefficiencies ultimately increase the cost of problem optimisation, prompting the need for a new method to ensure the full-fledged utilisation of LLMS. 

The suggested method, MCTS-AHD, combines Monte Carlo Tree Search and large language models for better heuristic function exploration. This system generates high-quality heuristics applicable to a wide variety of applications. In addition, MCTS-AHD uses an evaluation metric. This metric continually evaluates and improves the heuristics. Consequently, the search tree pursues only the most promising of the candidates. The key mechanisms of the method are as follows:

  • Integration of MCTS and LLMs: MCTS balances the exploration of new solutions and exploitation of the existing ones, ensuring that no time is wasted on unpromising paths. LLMs can understand the problems and generate excellent heuristics by leveraging MCTS.  
  • Structure of the Search Tree: The search tree consists of nodes and branches. Nodes represent a heuristic, and branches are the tweaks made in the heuristics. This tree mapping allows the framework to remember the explored solutions and focus on finding new ones. 
  • Simulation and Tree Expansion: The heuristics on each node have multiple branches that are simulated to evaluate their performance. This evaluation ensures that the search tree only expands on the promising branches, reducing time and overall cost.

MCTS-AHD was extensively tested on challenging datasets that included NP-hard combinatorial optimisation (CO) problems and Cost-aware Acquisition Function (CAF) design for Bayesian Optimization (BO). Its performance was compared to 4 baselines: manually designed heuristics, traditional automated heuristic design, neural combinatorial optimisation, and LLM-based AHD methods. MCTS-AHD consistently outperformed baseline methods in these benchmarks, demonstrating its robustness across different problem domains. Overall, there was a significant improvement in heuristic quality. 

In conclusion, MCTS-AHD significantly improves using large language models to design heuristics automatically. MCTS-AHD uses a tree-based structure, progressive widening and revolutionary exploration strategies to explore more heuristic functions than existing methods. This approach improves the performance and diversity of the heuristics and offers a strong framework for dealing with a meaningful number of complex CO tasks. MCTS-AHD creates a meaningful benchmark in AHD research, providing a highly scalable and exceptionally flexible solution for various applications. 


Check out the Paper and GitHub Page. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter and join our Telegram Channel and LinkedIn Group. Don’t Forget to join our 70k+ ML SubReddit.

🚨 [Recommended Read] Nebius AI Studio expands with vision models, new language models, embeddings and LoRA (Promoted)

The post Revolutionizing Heuristic Design: Monte Carlo Tree Search Meets Large Language Models appeared first on MarkTechPost.

]]>
https://www.marktechpost.com/2025/01/24/revolutionizing-heuristic-design-monte-carlo-tree-search-meets-large-language-models/feed/ 0 68328
Introducing GS-LoRA++: A Novel Approach to Machine Unlearning for Vision Tasks https://www.marktechpost.com/2025/01/22/introducing-gs-lora-a-novel-approach-to-machine-unlearning-for-vision-tasks/ https://www.marktechpost.com/2025/01/22/introducing-gs-lora-a-novel-approach-to-machine-unlearning-for-vision-tasks/#respond Thu, 23 Jan 2025 06:45:28 +0000 https://www.marktechpost.com/?p=68280 Pre-trained vision models have been foundational to modern-day computer vision advances across various domains, such as image classification, object detection, and image segmentation. There is a rather massive amount of data inflow, creating dynamic data environments that require a continual learning process for our models. New regulations for data privacy require specific information to be […]

The post Introducing GS-LoRA++: A Novel Approach to Machine Unlearning for Vision Tasks appeared first on MarkTechPost.

]]>
Pre-trained vision models have been foundational to modern-day computer vision advances across various domains, such as image classification, object detection, and image segmentation. There is a rather massive amount of data inflow, creating dynamic data environments that require a continual learning process for our models. New regulations for data privacy require specific information to be deleted. However, these pre-trained models face the issue of catastrophic forgetting when exposed to new data or tasks over time. When prompted to delete certain information, the model can forget valuable data or parameters. In order to tackle these problems, researchers from the Institute of Electrical and Electronics Engineers (IEEE) have developed Practical Continual Forgetting (PCF), which allows the models to forget task-specific features while retaining their performance. 

Current methods for mitigating catastrophic forgetting involve regularisation techniques, replay buffers, and architectural expansion. These techniques work well but do not allow selective forgetting; instead, they increase the architecture’s complexity, which causes inefficiencies when adopting new parameters. An optimum balance between trade-off plasticity and stability must exist so as not to excessively retain irrelevant information and be unable to adapt to new environments. However, this proves to be a significant struggle, prompting the need for a new method that enables flexible forgetting mechanisms and provides efficient adaptation. 

The proposed approach, Practical Continual Forgetting (PCF), has taken a reasonable strategy to deal with catastrophic forgetting and encourage selective forgetting. This framework has been developed to reinforce the strengths of pre-trained vision models. The methodology of PCF involves:

  • Adaptive Forgetting Modules: These modules keep analysing the features the model has previously learned and discard them when they become redundant. Task-specific features that are no longer relevant are removed, but their broader understanding is retained to ensure no generalisation issue arises. 
  • Task-Specific Regularization: PCF introduces constraints while training to ensure that the previously learned parameters are not drastically affected. Adapting to new tasks it ensures maximum performance while retaining previously learned information.

To test the performance of the PCF framework, experiments were conducted across various tasks, such as recognising faces, detecting objects, and classifying images under different scenarios, including missing data, and continual forgetting. The framework performed strongly in all these cases and outperformed the baseline models. Fewer parameters were used, making them more efficient. The methods showed robustness and practicality, handling rare or missing data better than other techniques.

The paper introduces the Practical Continual Forgetting (PCF) framework, which effectively addresses the problem of continual forgetting in pre-trained vision models by offering a scalable and adaptive solution for selective forgetting. It has the advantages of being analytically precise and adaptable, showing strong potential in applications sensitive to privacy and quite dynamic, as confirmed by strong performance metrics on various architectures. Nevertheless, it would be good to validate the approach further with real-world datasets and in even more complex scenarios to evaluate its robustness fully. Overall, the PCF framework sets a new benchmark for knowledge retention, adaptation, and forgetting in vision models, which has important implications for privacy compliance and task-specific adaptability.


Check out the Paper and GitHub Page. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter and join our Telegram Channel and LinkedIn Group. Don’t Forget to join our 65k+ ML SubReddit.

🚨 [Recommended Read] Nebius AI Studio expands with vision models, new language models, embeddings and LoRA (Promoted)

The post Introducing GS-LoRA++: A Novel Approach to Machine Unlearning for Vision Tasks appeared first on MarkTechPost.

]]>
https://www.marktechpost.com/2025/01/22/introducing-gs-lora-a-novel-approach-to-machine-unlearning-for-vision-tasks/feed/ 0 68280
CHASE: A Query Engine that is Natively Designed to Support Efficient Hybrid Queries on Structured and Unstructured Data https://www.marktechpost.com/2025/01/17/chase-a-query-engine-that-is-natively-designed-to-support-efficient-hybrid-queries-on-structured-and-unstructured-data/ https://www.marktechpost.com/2025/01/17/chase-a-query-engine-that-is-natively-designed-to-support-efficient-hybrid-queries-on-structured-and-unstructured-data/#respond Sat, 18 Jan 2025 03:38:56 +0000 https://www.marktechpost.com/?p=68120 Domains like social media analysis, e-commerce, and healthcare data management require querying through large chunks of structured and unstructured databases. In this modern world, there has been an ever-increasing requirement for the same in many other domains. However, current systems have been proven inefficient due to their inability to tackle the diverse obstacles presented when […]

The post CHASE: A Query Engine that is Natively Designed to Support Efficient Hybrid Queries on Structured and Unstructured Data appeared first on MarkTechPost.

]]>
Domains like social media analysis, e-commerce, and healthcare data management require querying through large chunks of structured and unstructured databases. In this modern world, there has been an ever-increasing requirement for the same in many other domains. However, current systems have been proven inefficient due to their inability to tackle the diverse obstacles presented when querying through databases comprising both structured and unstructured data.

Intending to integrate these two data types seamlessly within a unified framework, researchers from Fudan University and Transwarp have developed CHASE, which is a relational database framework designed to support hybrid queries natively. 

Currently, there are relational database management systems for structured data and specialised unstructured data solutions. Both specialise in their specific data types and cannot handle hybrid queries. Structured data is highly rigid and needs a predefined set of rules for organisation, while unstructured data consists of texts, images, videos, etc, requiring a flexible system for their storage. When both data types come together, there is an immense increase in the computational load, and catering to their specific needs is challenging. Therefore, there is a need for a new method that can bridge the gap between these two data structure types, introducing latency in query processing and addressing scalability issues. 

The proposed method, CHASE, introduces a sophisticated architecture to handle hybrid queries. The key functionalities include the following:

  • Advanced Indexing for Unstructured Data: For efficient retrieval, an indexing system is introduced for all the different unstructured data types, such as images, audio and videos. This allows for effectively tackling complex queries, which was an issue due to the flexible nature of these databases. 
  • Dynamic Query Optimization: First, CHASE analyses the data types present in the query, and based on that, it optimises its approach in real time. With this tailored approach, the process becomes more efficient by reducing the processing time of the queries. 
  • Integration with Natural Language Processing (NLP): NLP enables CHASE to understand the natural language query, which allows it to gain contextual understanding rather than keyword matching. This provides the user with a better experience and also allows non-technical personnel to query the databases effectively.

CHASE was benchmarked on real-world datasets, with 23 scenarios for testing various functionalities. The execution time was, on average, 30% faster for CHASE than conventional systems. The benchmarks indicated reduced resource consumption while maintaining high performance levels, which is a testament to the efficiency of CHASE in handling hybrid datasets. CHASE showed linear scalability with the increased dataset size, proving its efficacy for enterprise-grade applications.

The paper has dealt with the critical need for a cohesive system in order to manage hybrid data queries by proposing the CHASE methodology, which is practical and scalable due to its immense performance and efficiency upgrade over traditional methods. Its novel architecture, complete query language, and strong benchmarking results position CHASE as a leading solution for the management of hybrid data. However, this research has some weaknesses, such as limited testing on real-world datasets with complex data relationships; therefore, it needs further validation to guarantee its long-term reliability and broad applicability in general and various domains. Overall, this research contributes meaningfully to the field because it proposes an intrinsic relational database designed for hybrid queries, which fills the critical gap in the management of data and establishes CHASE as a valuable tool for modern applications with the requirement to integrate structured and unstructured data seamlessly.


Check out the Paper. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter and join our Telegram Channel and LinkedIn Group. Don’t Forget to join our 65k+ ML SubReddit.

🚨 Recommend Open-Source Platform: Parlant is a framework that transforms how AI agents make decisions in customer-facing scenarios. (Promoted)

The post CHASE: A Query Engine that is Natively Designed to Support Efficient Hybrid Queries on Structured and Unstructured Data appeared first on MarkTechPost.

]]>
https://www.marktechpost.com/2025/01/17/chase-a-query-engine-that-is-natively-designed-to-support-efficient-hybrid-queries-on-structured-and-unstructured-data/feed/ 0 68120
This AI Research Developed a Question-Answering System based on Retrieval-Augmented Generation (RAG) Using Chinese Wikipedia and Lawbank as Retrieval Sources https://www.marktechpost.com/2025/01/13/this-ai-research-developed-a-question-answering-system-based-on-retrieval-augmented-generation-rag-using-chinese-wikipedia-and-lawbank-as-retrieval-sources/ https://www.marktechpost.com/2025/01/13/this-ai-research-developed-a-question-answering-system-based-on-retrieval-augmented-generation-rag-using-chinese-wikipedia-and-lawbank-as-retrieval-sources/#respond Tue, 14 Jan 2025 05:46:37 +0000 https://www.marktechpost.com/?p=67901 Knowledge Retrieval systems have been prevalent for decades in many industries, such as healthcare, education, research, finance, etc. Their modern-day usage has integrated large language models(LLMs) that have increased their contextual capabilities, providing accurate and relevant answers to user queries. However, to better rely on these systems in cases of ambiguous queries and the latest […]

The post This AI Research Developed a Question-Answering System based on Retrieval-Augmented Generation (RAG) Using Chinese Wikipedia and Lawbank as Retrieval Sources appeared first on MarkTechPost.

]]>
Knowledge Retrieval systems have been prevalent for decades in many industries, such as healthcare, education, research, finance, etc. Their modern-day usage has integrated large language models(LLMs) that have increased their contextual capabilities, providing accurate and relevant answers to user queries. However, to better rely on these systems in cases of ambiguous queries and the latest information retrieval, which results in factually inaccurate or irrelevant answers, there is a need to integrate dynamic adaptation capabilities and increase the contextual understanding of the LLMs. Researchers from the National Taiwan University and National Chengchi University have introduced a novel methodology that combines retrieval-augmented generation (RAG) with adaptive, context-sensitive mechanisms to enhance the accuracy and reliability of LLMs.

Traditional retrieval systems often relied on indexing documents and prioritizing keyword matching. This leads to contextually irrelevant responses as they lack the capability to handle vague inputs. Moreover, failure to adapt to new information may produce incorrect outputs. Retrieval-Augmented Generation (RAG) is a more advanced approach combining retrieval and generation capabilities. Although RAG allows real-time information integration, it is unreliable and struggles to maintain factual accuracy due to its dependence on pre-trained knowledge bases. Therefore, we need a new method to seamlessly integrate generation and retrieval processes and adapt dynamically.

The proposed method uses a multi-step, dynamic strategy to further improve the combination of RAG and information retrieval. The mechanism of the approach is as follows:

  • Contextual Embedding Techniques: The input queries are converted into vector representations to capture semantic meaning. Such embeddings can understand ambiguous queries better and provide more appropriate information.
  • Adaptive Attention Mechanisms: In order to seamlessly embed real-time information with information retrieval, this method uses an attention mechanism that can dynamically adjust itself to focus on the specific context of the user queries.
  • Dual-Model Framework: It consists of a retrieval model and a generative model. While the former is adept at extracting information from structured and unstructured sources, the latter can assemble this information and provide cohesive responses. 
  • Fine-Tuned Training: When employed for a particular industry, the model can be fine-tuned for the specific datasets for an even more contextual understanding. 

This method was tested on Chinese Wikipedia and Lawbank and achieved significant retrieval precision compared to baseline RAG models. There was a substantial reduction in hallucination errors, producing outputs closely aligned with the retrieved data. Despite its two-stage retrieval, this method maintained a competitive latency suitable for real-time applications in all possible domains. Also, simulated real-world scenarios show increased user satisfaction with more accurate and contextually relevant responses from the system.

The RAG-based retrieval system in the proposed methodology is a breakthrough concerning some of the significant deficiencies of traditional RAG systems. It guarantees much better accuracy and reliability across applications through dynamic adaptation of retrieval strategies and better incorporation of knowledge into generative outputs. The scalability and domain adaptability of the methodology makes it a milestone for future improvements in retrieval-augmented AI systems, providing a robust solution for information-intensive tasks in critical industries.


Check out the Paper. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter and join our Telegram Channel and LinkedIn Group. Don’t Forget to join our 65k+ ML SubReddit.

🚨 Recommend Open-Source Platform: Parlant is a framework that transforms how AI agents make decisions in customer-facing scenarios. (Promoted)

The post This AI Research Developed a Question-Answering System based on Retrieval-Augmented Generation (RAG) Using Chinese Wikipedia and Lawbank as Retrieval Sources appeared first on MarkTechPost.

]]>
https://www.marktechpost.com/2025/01/13/this-ai-research-developed-a-question-answering-system-based-on-retrieval-augmented-generation-rag-using-chinese-wikipedia-and-lawbank-as-retrieval-sources/feed/ 0 67901
The Prompt Alchemist: Automated LLM-Tailored Prompt Optimization for Test Case Generation https://www.marktechpost.com/2025/01/08/the-prompt-alchemist-automated-llm-tailored-prompt-optimization-for-test-case-generation/ https://www.marktechpost.com/2025/01/08/the-prompt-alchemist-automated-llm-tailored-prompt-optimization-for-test-case-generation/#respond Thu, 09 Jan 2025 06:24:16 +0000 https://www.marktechpost.com/?p=67174 Owing to the advent of Artificial Intelligence (AI), the software industry has been leveraging Large Language Models (LLMs) for code completion, debugging, and generating test cases. However, LLMs follow a generic approach when developing test cases for a different software, which prevents them from considering the software’s unique architecture, user requirements and potential edge cases. […]

The post The Prompt Alchemist: Automated LLM-Tailored Prompt Optimization for Test Case Generation appeared first on MarkTechPost.

]]>
Owing to the advent of Artificial Intelligence (AI), the software industry has been leveraging Large Language Models (LLMs) for code completion, debugging, and generating test cases. However, LLMs follow a generic approach when developing test cases for a different software, which prevents them from considering the software’s unique architecture, user requirements and potential edge cases. Moreover, different outputs are obtained from the same prompt when using other software, which raises the question of the prompt’s reliability. Due to these issues, critical bugs can go undetected, which increases the overall expenditure and ultimately hinders the software’s practical deployment in sensitive industries like healthcare. A team of researchers from the Chinese University of Hong Kong, Harbin Institute of Technology, School of Information Technology, and some independent researchers have introduced MAPS, the prompt alchemist for tailored optimizations and contextual understanding. 

Traditional test case generation approaches rely on rule-based systems or manual engineering of prompts for Large Language Models (LLMs). These methods have been foundational in software testing but exhibit several limitations. Most researchers use manual methods to optimize prompt engineering for test case generation, which requires significant time investment. These methods are also difficult to scale due to the increase in complexity. Other methods are often generic in nature, producing bugs. Therefore, a new approach is needed for test case generation that can prevent labor-intensive manual optimization and does not lead to suboptimal outcomes. 

The proposed method, MAPS, automates the prompt optimization process, aligning the test cases with real-world requirements significantly reducing human intervention. The core framework of MAPS includes:

  • Baseline Prompt Evaluation: LLMs are assessed on their performance on test cases generated using basic prompts. This assessment is foundational to further optimization efforts needed. 
  • Feedback Loop: Based on the evaluation results, suboptimally performing test cases are set aside and tweaked to better align with software requirements. This information is fed back into the LLM, allowing for continuous improvement in a feedback loop.
  • LLM-Specific Tuning: The reinforcement learning techniques are used for dynamic prompt optimization. This opens a space for customizations in the prompt by taking into account the strengths and weaknesses of the LLMs. 

The results showed that MAPS significantly outperformed the traditional prompt engineering techniques. Its optimized prompts had a 6.19% higher line coverage rate than static prompts. The framework identified more bugs than the baseline methods, exhibiting its ability to effectively generate edge case scenarios. Test cases generated with optimized prompts exhibited improvement in semantic correctness, which reduced the need for manual adjustments.

In a nutshell, MAPS is a state-of-the-art optimization technique for prompt generation, particularly targeted to LLMs used in the software testing domain. Some of the weaknesses of the available test case generation techniques have been addressed through multi-pipeline-stage architectures that incorporate baseline evaluations, iterative feedback loops, and model-specific tuning. These new characteristics of the framework not only automate prompt optimization but enhance the quality and reliability of outputs in automated testing workflows, thus making it an indispensable tool for software development teams looking for efficiency and effectiveness in their testing processes.


Check out the Paper. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter and join our Telegram Channel and LinkedIn Group. Don’t Forget to join our 60k+ ML SubReddit.

🚨 FREE UPCOMING AI WEBINAR (JAN 15, 2025): Boost LLM Accuracy with Synthetic Data and Evaluation IntelligenceJoin this webinar to gain actionable insights into boosting LLM model performance and accuracy while safeguarding data privacy.

The post The Prompt Alchemist: Automated LLM-Tailored Prompt Optimization for Test Case Generation appeared first on MarkTechPost.

]]>
https://www.marktechpost.com/2025/01/08/the-prompt-alchemist-automated-llm-tailored-prompt-optimization-for-test-case-generation/feed/ 0 67174
PyG-SSL: An Open-Source Library for Graph Self-Supervised Learning and Compatible with Various Deep Learning and Scientific Computing Backends https://www.marktechpost.com/2025/01/07/pyg-ssl-an-open-source-library-for-graph-self-supervised-learning-and-compatible-with-various-deep-learning-and-scientific-computing-backends/ https://www.marktechpost.com/2025/01/07/pyg-ssl-an-open-source-library-for-graph-self-supervised-learning-and-compatible-with-various-deep-learning-and-scientific-computing-backends/#respond Wed, 08 Jan 2025 05:50:00 +0000 https://www.marktechpost.com/?p=67141 Complex domains like social media, molecular biology, and recommendation systems have graph-structured data that consists of nodes, edges, and their respective features. These nodes and edges do not have a structured relationship, so addressing them using graph neural networks (GNNs) is essential. However, GNNs rely on labeled data, which is difficult and expensive to obtain. […]

The post PyG-SSL: An Open-Source Library for Graph Self-Supervised Learning and Compatible with Various Deep Learning and Scientific Computing Backends appeared first on MarkTechPost.

]]>
Complex domains like social media, molecular biology, and recommendation systems have graph-structured data that consists of nodes, edges, and their respective features. These nodes and edges do not have a structured relationship, so addressing them using graph neural networks (GNNs) is essential. However, GNNs rely on labeled data, which is difficult and expensive to obtain. Self-supervised Learning (SSL) is an evolving methodology that leverages unlabelled data by generating its supervisory signals. SSL for graphs comes with its own challenges, such as domain specificity, lack of modularity, and steep learning curve. Addressing these issues, a team of researchers from the University of Illinois Urbana-Champaign, Wayne State University, and Meta AI have developed PyG-SSL, an open-source toolkit designed to advance graph self-supervised learning.

Current Graph Self-Supervised Learning (GSSL) approaches primarily focus on pretext (self-generated) tasks, graph augmentation, and contrastive learning. Pretext includes node-level, edge-level, and graph-level tasks that help the model learn useful representations without needing labeled data. Their augmentation occurs by dropping, maskin,g or shuffling, improving the model’s robustness and generalizability. However, existing GSSL frameworks are designed for specific applications and require significant customization. Moreover, developing and testing new SSL methods is time-intensive and error-prone without a modular and extensible framework. Therefore, a new process is needed to address the fragmented nature of existing GSSL implementations and the absence of a unified toolkit that restricts standardization and benchmarking across various GSSL methods. 

The proposed toolkit, PyG-SSL, standardizes the implementation and evaluation of graph SSL methods. The key features of PyG-SSL are:

  • Comprehensive Support: This toolkit integrates multiple state-of-the-art methods for a unified framework, allowing researchers to select the most suitable method for their specific application. 
  • Modularity: PyG-SSL allows the creation of tailored solutions by mixing one or more techniques. Pipelines can also be customized without requiring extensive reconfiguration.
  • Benchmarks and Datasets: Standard datasets and evaluation protocols are preloaded in this toolkit to allow researchers to benchmark their findings and ensure validation easily. 
  • Performance Optimization: PyG-SSL toolkit is designed to handle large datasets efficiently. It is optimized for fast training time and reduced computational requirements.

This toolkit has been rigorously tested across multiple datasets and SSL methods, demonstrating its effectiveness in standardizing and advancing graph SSL research. With reference implementations of a wide range of SSL methods, PyG-SSL ensures that the results are reproducible and comparable in experiments. Experimental results demonstrate that integrating PyG-SSL into existing GNN architectures improves their performance on downstream tasks by properly exploiting unlabeled data.

PyG-SSL marks a significant milestone in graph self-supervised learning, addressing long-standing challenges related to standardization, reproducibility, and accessibility. PyG-SSL gives the possibility to attain state-of-the-art results through its unified, modular, and extensible toolkit, easing the development of innovative graph SSL methods. PyG-SSL can play a pivotal role in advancing graph-based machine learning applications across diverse domains in this fast-evolving field.


Check out the Paper and GitHub Page. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter and join our Telegram Channel and LinkedIn Group. Don’t Forget to join our 60k+ ML SubReddit.

🚨 FREE UPCOMING AI WEBINAR (JAN 15, 2025): Boost LLM Accuracy with Synthetic Data and Evaluation IntelligenceJoin this webinar to gain actionable insights into boosting LLM model performance and accuracy while safeguarding data privacy.

The post PyG-SSL: An Open-Source Library for Graph Self-Supervised Learning and Compatible with Various Deep Learning and Scientific Computing Backends appeared first on MarkTechPost.

]]>
https://www.marktechpost.com/2025/01/07/pyg-ssl-an-open-source-library-for-graph-self-supervised-learning-and-compatible-with-various-deep-learning-and-scientific-computing-backends/feed/ 0 67141
AutoGraph: An Automatic Graph Construction Framework based on LLMs for Recommendation https://www.marktechpost.com/2025/01/05/autograph-an-automatic-graph-construction-framework-based-on-llms-for-recommendation/ https://www.marktechpost.com/2025/01/05/autograph-an-automatic-graph-construction-framework-based-on-llms-for-recommendation/#respond Mon, 06 Jan 2025 06:25:52 +0000 https://www.marktechpost.com/?p=67047 Enhancing user experiences and boosting retention using recommendation systems is an effective and ever-evolving strategy used by many industries, such as e-commerce, streaming services, social media, etc. These systems must analyze complex relationships between users, items, and contextual factors to suggest precisely what the user might want. However, the existing recommendation systems are static, relying […]

The post AutoGraph: An Automatic Graph Construction Framework based on LLMs for Recommendation appeared first on MarkTechPost.

]]>
Enhancing user experiences and boosting retention using recommendation systems is an effective and ever-evolving strategy used by many industries, such as e-commerce, streaming services, social media, etc. These systems must analyze complex relationships between users, items, and contextual factors to suggest precisely what the user might want. However, the existing recommendation systems are static, relying on substantial historical data to build connections effectively. In cold start scenarios, which are heavily prevalent, mapping the relationships becomes impossible, weakening these systems even further. Researchers from the Shanghai Jiao Tong University and Huawei Noah’s Ark Lab have introduced AutoGraph to address these issues. This framework automatically builds graphs incorporating dynamic adjustments and leverages LLMs for better contextual understanding. 

Commonly, graph-based recommendation systems are employed. Current systems, however, require people to set the features manually and their connections in a graph, consuming much time. Also, rules are set beforehand, limiting how these graphs could adapt. Incorporating unstructured data, which potentially has rich semantic information about user preferences, is also a significant issue. Therefore, there is a need for a new method that can resolve the data sparsity issues and the failure to capture nuanced relationships and adjust to user preferences in real-time.  

AutoGraph is an innovative framework to enhance recommendation systems leveraging Large Language Models (LLMs) and Knowledge Graphs (KGs). The methodology of AutoGraph is based on these features:

  • Utilization of Pre-trained LLMs: The framework leverages pre-trained LLMs to analyze user input. It can draw relationships based on the analysis of natural language, even those that are apparently hidden. 
  • Knowledge Graph Construction: After the relationship extraction, LLMs generate graphs, which can be seen as structured representations of user preferences. Algorithms optimize such graphs to remove less relevant connections in an attempt to maximize the quality of the graph in its entirety.
  • Integration with Graph Neural Networks (GNNs): The final step of the proposed method is to integrate the created knowledge graph with regular Graph Neural Networks. GNNs can provide more accurate recommendations by using both node features and graph structure, and they are sensitive to personal preferences and more significant trends among users.

To evaluate the proposed framework’s efficacy, authors benchmarked against traditional recommendation techniques using e-commerce and streaming services datasets. There was a significant gain in the precision of recommendations, which shows that the framework is competent enough to give relevant recommendations. The proposed method had improved scalability for dealing with large datasets. The framework demonstrated reduced computational requirements compared to traditional graph construction approaches. Process automation, along with the use of advanced algorithms, helped in lowering resource usage without compromising the quality of the results.

The Autograph framework represents a significant leap forward in recommendation systems. Automating graph construction with LLMs addresses long-standing scalability, adaptability, and contextual awareness challenges. The framework’s success demonstrates the transformative potential of integrating LLMs into graph-based systems, setting a new benchmark for future research and applications in personalized recommendations. AutoGraph opens new avenues for personalized user experiences in diverse domains by automating the construction of dynamic, context-aware recommendation graphs. This innovation highlights the growing role of LLMs in addressing real-world challenges, revolutionizing how we approach recommendation systems.


Check out the Paper. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter and join our Telegram Channel and LinkedIn Group. Don’t Forget to join our 60k+ ML SubReddit.

🚨 FREE UPCOMING AI WEBINAR (JAN 15, 2025): Boost LLM Accuracy with Synthetic Data and Evaluation IntelligenceJoin this webinar to gain actionable insights into boosting LLM model performance and accuracy while safeguarding data privacy.

The post AutoGraph: An Automatic Graph Construction Framework based on LLMs for Recommendation appeared first on MarkTechPost.

]]>
https://www.marktechpost.com/2025/01/05/autograph-an-automatic-graph-construction-framework-based-on-llms-for-recommendation/feed/ 0 67047
DiTCtrl: A Training-Free Multi-Prompt Video Generation Method Under MM-DiT Architectures https://www.marktechpost.com/2024/12/31/ditctrl-a-training-free-multi-prompt-video-generation-method-under-mm-dit-architectures/ https://www.marktechpost.com/2024/12/31/ditctrl-a-training-free-multi-prompt-video-generation-method-under-mm-dit-architectures/#respond Wed, 01 Jan 2025 06:08:48 +0000 https://www.marktechpost.com/?p=66872 Generative AI has revolutionized video synthesis, producing high-quality content with minimal human intervention. Multimodal frameworks combine the strengths of generative adversarial networks (GANs), autoregressive models, and diffusion models to create high-quality, coherent, diverse videos efficiently. However, there is a constant struggle while deciding what part of the prompt, either text, audio or video, to pay […]

The post DiTCtrl: A Training-Free Multi-Prompt Video Generation Method Under MM-DiT Architectures appeared first on MarkTechPost.

]]>
Generative AI has revolutionized video synthesis, producing high-quality content with minimal human intervention. Multimodal frameworks combine the strengths of generative adversarial networks (GANs), autoregressive models, and diffusion models to create high-quality, coherent, diverse videos efficiently. However, there is a constant struggle while deciding what part of the prompt, either text, audio or video, to pay attention to more. Moreover, efficiently handling different types of input data is crucial, yet it has proven to be a significant problem. To tackle these issues, researchers from MMLab, The Chinese University of Hong Kong, GVC Lab, Great Bay University, ARC Lab, Tencent PCG, and Tencent AI Lab have developed DiTCtrl, a multi-modal diffusion transformer, for multi-prompt video generation without requiring extensive tuning. 

Traditionally, video generation heavily depended on autoregressive architectures for short video segments and constrained latent diffusion methods for higher-quality short video generation. As is evident, the efficiency of such methods always declines when video length is increased. These methods primarily focus on single prompt inputs; this makes it challenging to generate coherent videos from multi-prompt inputs. Moreover, significant fine-tuning is required, which leads to inefficiencies in time and computational resources. Therefore, a new method is needed to combat these issues of lack of fine attention mechanisms, decreased long video quality, and inability to process multimodal outputs simultaneously.

The proposed method, DiTCtrl, is equipped with dynamic attention control, tuning-free implementation, and multi-prompt compatibility. The key aspects of DiTCtrl are:

  1. Diffusion-Based Transformer Architecture: DiT architecture allows the model to handle multimodal inputs efficiently by integrating them at a latent level. This gives the model a better contextual understanding of inputs, ultimately giving better alignment. 
  2. Fine-Grained Attention Control: This framework can adjust its attention dynamically, which allows it to focus on more critical parts of the prompt, generating coherent videos. 
  3. Optimized Diffusion Process: Longer video generation requires a smooth and coherent transition between scenes. Optimized diffusion decreases inconsistencies across frames, promoting a seamless narrative without abrupt changes. 

DiTCtrl has demonstrated state-of-the-art performance on standard video generation benchmarks. Significant improvements in video generation quality were made in terms of temporal coherence and prompt fidelity. DiTCtrl has produced superior output quality in qualitative tests compared to traditional methods. Users have reported smoother transitions and more consistent object motion in videos generated by DiTCtrl, especially when responding to multiple sequential prompts.

The paper deals with the challenges of tuning-free, multi-prompt, long-form video generation using a novel attention control mechanism, an advancement in video synthesis. In this regard, by using dynamic and tuning-free methodologies, this framework adds much better scalability and usability, raising the bar for the field. DiTCtrl, with its attention control modules and multi-modal compatibility, lays a strong foundation for generating high-quality and extended videos—a key impact in creative industries that rely on customizability and coherence. However, reliance on particular diffusion architectures may not make it easily adaptable to other generative paradigms. This research presents a scalable and efficient solution ready to take advancements in video synthesis to new levels and enable unprecedented degrees of video customization.


Check out the Paper. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter and join our Telegram Channel and LinkedIn Group. Don’t Forget to join our 60k+ ML SubReddit.

🚨 FREE UPCOMING AI WEBINAR (JAN 15, 2025): Boost LLM Accuracy with Synthetic Data and Evaluation IntelligenceJoin this webinar to gain actionable insights into boosting LLM model performance and accuracy while safeguarding data privacy.

The post DiTCtrl: A Training-Free Multi-Prompt Video Generation Method Under MM-DiT Architectures appeared first on MarkTechPost.

]]>
https://www.marktechpost.com/2024/12/31/ditctrl-a-training-free-multi-prompt-video-generation-method-under-mm-dit-architectures/feed/ 0 66872