PinnedPublished inTowards AIIs it possible to do Text Classification on unlabeled data? (Feat. Zero-Shot Classification)In this piece we will explore the zero-shot classification method for sentiment analysis using the Huggingface library.Aug 30, 20211Aug 30, 20211
PinnedPublished inTowards AIHow much DEEP learning do we need?The growing trend of large pre-trained networks is worrying. It could send the message that Deep Learning is not for everyone.Aug 4, 2021Aug 4, 2021
PinnedPublished inTowards AIHow to Train a Seq2Seq Text Summarization Model With Sample Code (Ft. Huggingface/PyTorch)Part 2 of the introductory series about training a Text Summarization model (or any Seq2seq/Encoder-Decoder Architecture) with sample codes…Dec 14, 2021Dec 14, 2021
PinnedPublished inTowards AICompressedBART: Fine-Tuning for Summarization through Latent Space Compression (Paper…Paper title: A Robust Approach to Fine-tune Pre-trained Transformer-based Models for Text Summarization through Latent Space Compression.Apr 7, 2023Apr 7, 2023
PinnedWhat are the differences in Pre-Trained Transformer-base models like BERT, DistilBERT, XLNet, GPT…This article is a cheat sheet of well-known Transformer-based models. Trying to explain their uniqueness (while they are all based on…May 19, 20212May 19, 20212
Published inTowards AIMonth in 4 Papers (December 2024)This series of posts is designed to bring you the newest findings and developments in the NLP field. I’ll delve into four significant…Jan 3Jan 3
Published inTowards AIMonth in 4 Papers (November 2024)Advancing LLMs through smarter attention mechanisms, thought-based reasoning, and efficient knowledge transfer.Dec 1, 2024Dec 1, 2024
Published inTowards AIMonth in 4 Papers (August 2024)Exploring the complexity of expert mixtures, knowledge integration, and memory in modern NLP models.Sep 2, 2024Sep 2, 2024
Published inTowards AIMonth in 4 Papers (July 2024)Unlocking Efficiency by Innovations in NLP for Speed, Context, and Multimodal InsightsJul 29, 2024Jul 29, 2024
Published inTowards AIMonth in 4 Papers (June 2023)Advancing Language Models through Efficient Training and Alignment Techniques.Jul 8, 2024Jul 8, 2024