Featured image of post Large Language Models: Comparing Gen2/Gen3 Models (GPT-3, GPT-J, MT5 and More)

Large Language Models: Comparing Gen2/Gen3 Models (GPT-3, GPT-J, MT5 and More)

Large Language Models are sophisticated neural networks that produce texts. By creating one word at a time, given a context of other words, these models produce texts that can rival a humans output. The creation of LLMs began back in 2018 when the transformer neural network architecture was discovered. Since then, ever more complex transformer models in terms of parameter amount, and continues up to this data with ever more complex model architectures, consumed amount of texts, and parametric complexity.

Featured image of post Large Language Models: Comparing Gen 1 Models (GPT, BERT, T5 and More)

Large Language Models: Comparing Gen 1 Models (GPT, BERT, T5 and More)

The creation of Large Language Models (LLMs) began in 2018. Three factors emerged and were combined in LLMs: powerful computer and graphics processing units, huge amounts of structured and unstructured data that could be processed fast, and first-grade open-source project for the creation and training of neural networks.

Featured image of post Introduction to Large Language Models

Introduction to Large Language Models

Large Language Models (LLMs) are a ubiquities technology enabling humans to use their natural language for interacting with a computer in a broad range of tasks. LLMs can answer questions about history and real-world events, they can create step-by-step tasks plans, solve mathematical questions, and can reflect on any input text to create summaries or identify text characteristics. Using most recent LLMs like GPT4 is a fascinating and surprising event.

Featured image of post Arduino Nano BLE 33 Sense Microcontroller: Hardware and GPIO Functions

Arduino Nano BLE 33 Sense Microcontroller: Hardware and GPIO Functions

The Arduino Nano BLE 33 Sense Microcontroller is an Arduino-compatible board with a fleet of onboard sensors, including sound, light, temperature, and a microphone. It can be programmed with Arduino C and MicroPython to read and write data. This article introduces this unique board, details how to use the digital and analog pins and which functions the board supports.

Featured image of post Wikipedia Article Crawler & Clustering: Text Classification with Spacy

Wikipedia Article Crawler & Clustering: Text Classification with Spacy

Spacy is a powerful NLP library that performs many NLP tasks in its default configuration, including tokenization, stemming and part-of-speech tagging. These steps can be extended with a text classification task as well, in which training data in the form of preprocessed text and expected categories as dictionary objects are provided. Both multi-label and single-label classification is supported.

Featured image of post NLP with Spacy: Custom Text Classification Pipeline

NLP with Spacy: Custom Text Classification Pipeline

Spacy is a powerful NLP library in which many NLP tasks like tokenization, stemming, part-of-speech tagging and named entity resolution are provided out-of-the box with pretrained models. All of these tasks are wrapped by a pipeline object, and internal abstraction of different functions that are applied step by step on a given text. This pipeline can be both customized and extended with self-written functions.