Posts

Data when seen through the Inference Web of LLM

Image
Background Today I saw a post from this guy saying “I was at the Marines today” . A few mins later I saw a post from the same guy saying: “I love vada pavs”. And i thought to myself “Hey, This guy is probably from Mumbai, India” . The next thought that followed this was if I, a mere human being can deduce this from 2 posts, what can the LLMs do? LLMS are trained on massive data sources from the internet. Isn't it possible that they infer this information too? Introduction Large language models (LLMs) have revolutionized natural language processing tasks. They demonstrate amazing capabilities in understanding and generating human-like text. In this project, we decided to delve into the accuracies of inferential capabilities of various LLM models.  Ob jectives Inference in LLMs is a vast and interesting topic to dive into. For the sake of this project, we chose 3 main objectives that we would be focusing upon: Identify the accuracies of Inferential Capabilities of the LLM​ Exploring...

What's Your Number? Interpreting Memorisation in Language Models

Image
  In the vast realm of machine learning, the concept of "grokking" holds a special allure. It refers to a model's ability to truly understand and generalize beyond mere pattern recognition or memorization. Achieving grokking is a hallmark of true intelligence, where a model can accurately predict or classify data while developing an intuitive grasp of the underlying patterns, relationships, and representations within that data. Our Approach As part of our course project, our team set out to explore the parameters and conditions necessary for grokking to occur in machine learning models. We focused our efforts on task complexity, data quantity, hyperparameters, and model architecture, using a character-level decoder-only transformer architecture as our testing ground. First Steps Initially, we tackled a simple ROT13 cipher task, which maps each letter to the 13th letter after it in the alphabet. However, this task proved too straightforward, and even a single-layered small...

Variegated Machine Unlearning - KowalskiAnalysis

Image
Poster Presentation Day What is ‘Machine Unlearning’? In the age of machine learning, with the widespread use and abuse of data, to tackle these challenges, the notion of machine unlearning comes into play, with a variety of uses in privacy, poison removal, etc. What is ‘Data Poisoning’? Data poisoning is a type of cyber-attack in which an adversary intentionally compromises a training dataset used by an AI or machine learning (ML) model to influence or manipulate the operation of that model. What is the motivation? We take inspiration from the Corrective Machine Unlearning paper by Goel, et. al, to explore some interesting cases with performance implications for machine unlearning. What did we do? We explored multiple problems: 1. Machine Unlearning of Poisons over Imbalanced Datasets:  Given the inherent disparity in representation of different classes over imbalanced  datasets, we hypothesize that the impact of machine unlearning of poisons over such  datasets should ...

Defining and Detecting Hallucinations in LLM's

Image
`   DEFINING and DETECTING HALLUCINATIONS IN LARGE LANGUAGE MODELS

EAR-VM: Exploring Methods for Improving Adversarial Robustness of Vision Models

Image
  Abstract CNNs have many uses, particularly in the field of Computer Vision, however, their vulnerability to Adversarial attacks leaves a lot to be desired, particularly to their robustness to these kinds of attacks in order to make them more safe.  The misuse of adversarial attacks are a major threat to CNN vision models; for example, self-driving cars can be made to misinterpret road signs or signals, putting the passengers at significant risk of harm.  To address this vulnerability, we have attempted to modify the architecture of a CNN to add an auxiliary classification SVM which will determine the maximum margin in which these adversarial attacks will impact loss.   Also, interpretability is a key concept in understanding the decisions and outputs of modern networks. By interpreting the working of the model, we can curate better adversarial attacks, or make the model more robust.   Objectives Our goal is to first implement the At-SVM, and then to test it to...