To Laura

Welcome to my blog. Here youll find a lot of coool information about a lot of cool stuff

Base vs Instruct Variants

Understanding Base vs Instruct Variants in Machine Learning Models As machine learning continues to evolve, developers frequently encounter different model variants suited for diverse tasks. Two common types you’ll often come across are base models and instruct models (also known as instruct-tuned models). Understanding the key differences between these variants can help you better select and tailor models for specific applications. In this blog post, we will take a closer look at these two variants, exploring their unique characteristics, applications, and nuances. ...

September 15, 2020 · 4 min · 648 words · Me

Customizing Large Language Models

Techniques for Customizing Large Language Models (LLMs) As large language models (LLMs) continue to evolve, the need for customization becomes increasingly important to tailor their capabilities to specific use cases. This blog post delves into several prominent techniques for LLM customization: prompting (including multi-shot and chaining), utilizing tools, Retrieval-Augmented Generation (RAG), and fine-tuning. Each technique has its own pros and cons, and understanding them will equip developers to make informed decisions in their projects. ...

September 15, 2020 · 3 min · 587 words · Me

Fine-Tuning ChatGPT

Fine-Tuning a Closed Source LLM like ChatGPT: A Step-by-Step Guide In the realm of machine learning, fine-tuning a language model can significantly enhance its ability to perform specific tasks or understand particular contexts. In this blog post, we’ll explore how to fine-tune a closed-source language model like OpenAI’s ChatGPT. While direct access to the model’s parameters isn’t available as it might be with open-source models, fine-tuning it using your own dataset is still achievable. ...

September 15, 2020 · 4 min · 740 words · Me

Gradio

Getting Started with Gradio: Building Interactive Interfaces for Machine Learning Models In the fast-paced world of machine learning and AI, creating interactive applications that allow users to engage with models is becoming increasingly valuable. Enter Gradio, a Python library designed to make building user interfaces for machine learning models straightforward and efficient. In this blog post, we’ll explore how Gradio works, how to use it, and how to integrate it with popular LLM APIs like OpenAI’s GPT. ...

September 15, 2020 · 4 min · 672 words · Me

Hugging Face

Understanding Hugging Face: A Comprehensive Guide Hugging Face has become a leading platform in the field of Natural Language Processing (NLP) and machine learning, especially known for its user-friendly tools and extensive community resources. In this blog, we’ll delve into what Hugging Face is, how it works, and the key libraries it offers, including transformers, datasets, accelerate, and hub. What is Hugging Face? Hugging Face started as a chatbot company but quickly shifted focus to NLP and now is a hub for state-of-the-art machine learning models. The platform is built around the community approach, enabling developers of all levels to collaborate and share pre-trained models, datasets, and innovations in machine learning. ...

September 15, 2020 · 4 min · 661 words · Me

JWT

Understanding JSON Web Tokens (JWT) In the realm of modern web applications, ensuring secure and efficient user authentication is crucial. JSON Web Tokens (JWT) have emerged as a popular solution for this purpose. This blog post will break down what JWTs are, how they work, their benefits, and provide a basic implementation along with security best practices. What are JWTs? JSON Web Tokens (JWT) are an open standard (RFC 7519) for securely transmitting information between parties as a JSON object. They are used for authentication and information exchange in a compact, URL-safe manner. A JWT is essentially a token that can encapsulate user and permission data, which can be verified and trusted. ...

September 15, 2020 · 4 min · 721 words · Me

NLP, LLMs, LR and ML

Understanding NLP, LLMs, Linear Regression, and the Landscape of Machine Learning Machine Learning (ML) has reshaped modern technology — powering everything from recommendation systems to self-driving cars. Within this field, Natural Language Processing (NLP) and Large Language Models (LLMs) have become particularly prominent due to the rise of generative AI. In this blog post, we’ll demystify the connections between these areas, explore the role of Linear Regression, and look at how they fit into the broader ML ecosystem. ...

September 15, 2020 · 3 min · 584 words · Me

RAG

Understanding Retrieval Augmented Generation (RAG) Retrieval Augmented Generation (RAG) is an innovative machine learning architecture that combines the strengths of information retrieval and text generation. As developers and machine learning practitioners, understanding RAG can elevate the capabilities of your AI models, enabling them to provide more relevant and context-aware responses. In this blog post, we’ll explore the workings of RAG, the importance of vectors, and how they facilitate efficient operations within this powerful framework. ...

September 15, 2020 · 4 min · 773 words · Me

This bitter earth

This bitter earth Well, what the fruit it bears Ooooh This bitter earth And if my life Is like the dust Oooh that hides The glow of a rose What good am I? Heaven only knows Lord, this bitter earth Yes can be so cold Today you’re young Too soon, you’re old But while a voice Within me cries I’m sure someone may answer my call And this bitter earth, oooh May not, oh, be so bitter after all ...

September 15, 2020 · 1 min · 122 words · Max Richter

Understanding the Foundations of Large Language Models (LLMs)

Understanding the Foundations of Large Language Models (LLMs) Meta-Description Dive into the core concepts behind large language models (LLMs) and the Transformer architecture. Learn about tokens, embeddings, weights, the attention mechanism, and how these elements combine to power modern AI applications. Large language models (LLMs) have revolutionized natural language processing (NLP), making it possible for machines to understand and generate human-like text. At the heart of these models lies the Transformer architecture, which leverages various components to analyze and generate language in a way that mimics human writing. In this blog post, we will explore the fundamental building blocks of LLMs, including tokens, embeddings, weights, attention mechanisms, and important concepts like fine-tuning and inference vs. training. ...

September 15, 2020 · 5 min · 960 words · Me