You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
This repository consists of the implementation of the paper "Automatic Ellipsis Reconstruction in Coordinated German Sentences based on Text-To-Text Transfer Transformers" which is accepted at 27th International Conference on Text, Speech and Dialogue.
Thesis scope: Train and Develop a Table-to-Text Transformer-based model for contextual summarization of tabular data. To achieve this T5-small , T5-base, Bart-base and Llama2 7B chat were finetuned on ToTTo and QTSumm. Regarding ToTTo, the models outperformed the benchmark.
Text-To-Text Textbots to Demonstrate Output Differences Between Models Trained on Filtered/Unfiltered Datasets for HSS4 - The Modern Context: Select Figures and Topics
Natural Language Processing (NLP) and Large Language Models (LLM) with Fine-Tuning LLM and make Chatbot Question answering (QA) with LoRA and Flan-T5 Large
Fine-tune open-source Language Models (LLMs) on E-commerce data, leveraging Amazon's sales data. Showcase a tailored solution for enhanced language understanding and generation with a focus on custom E-commerce datasets.