PinnedPublished inTowards AIIs it possible to do Text Classification on unlabeled data? (Feat. Zero-Shot Classification)In this piece we will explore the zero-shot classification method for sentiment analysis using the Huggingface library.Aug 30, 2021A response icon1Aug 30, 2021A response icon1
PinnedPublished inTowards AIHow much DEEP learning do we need?The growing trend of large pre-trained networks is worrying. It could send the message that Deep Learning is not for everyone.Aug 4, 2021Aug 4, 2021
PinnedPublished inTowards AIHow to Train a Seq2Seq Text Summarization Model With Sample Code (Ft. Huggingface/PyTorch)Part 2 of the introductory series about training a Text Summarization model (or any Seq2seq/Encoder-Decoder Architecture) with sample codes…Dec 14, 2021Dec 14, 2021
PinnedPublished inTowards AICompressedBART: Fine-Tuning for Summarization through Latent Space Compression (Paper…Paper title: A Robust Approach to Fine-tune Pre-trained Transformer-based Models for Text Summarization through Latent Space Compression.Apr 7, 2023Apr 7, 2023
PinnedWhat are the differences in Pre-Trained Transformer-base models like BERT, DistilBERT, XLNet, GPT…This article is a cheat sheet of well-known Transformer-based models. Trying to explain their uniqueness (while they are all based on…May 19, 2021A response icon2May 19, 2021A response icon2
Published inTowards AIMonth in 4 Papers (June 2025)Exploring new frontiers in coding, comprehension, and the art of trying again.Jun 30Jun 30
Published inTowards AIMonth in 4 Papers (May 2025)Exploring decomposition, diffusion, and control beyond traditional agents in NLP.Jun 2Jun 2
Published inTowards AIMonth in 4 Papers (March 2025)Efficient reasoning, context extraction, and model scaling innovations in today’s cutting-edge NLP research.Mar 31Mar 31
Published inTowards AIMonth in 4 Papers (February 2025)Exploring how caching strategies, context length, uncertainty estimation, and conceptual representations are reshaping knowledge retrieval…Mar 10Mar 10
Published inTowards AIMonth in 4 Papers (January 2025)How Language Models Learn to Think, Judge, and Scale: From Code Evaluation to Memory-Efficient Reasoning.Feb 3Feb 3