Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities
-
Updated
Jan 23, 2026 - Python
Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities
Unattended Lightweight Text Classifiers with LLM Embeddings
Faster, smaller BERT models in just a few lines of code.
Lightweight cross-lingual coreference resolution with spaCy using ONNX Runtime inference of transformer models.
comprehensive solutions for Adobe's Document Intelligence Hackathon 2025, encompassing two distinct challenges focused on advanced PDF processing and persona-driven content analysis. Both implementations adhere to stringent performance requirements including sub-60-second execution times and containerized deployment within 1GB resource constraints
A state-of-the-art Retrieval-Augmented Generation (RAG) system with hybrid search, multi-hop reasoning, answer verification, and source citation — delivering accurate, trustworthy, and context-aware answers from large document collections.
Compact transformer to auto-label help-desk tickets (topic + sentiment) with a FastAPI endpoint, eval dashboard, and MLOps glue.
An Ai-powered agent that automatically clusters, summarizes and prioritizes operational asset alerts . made using Python , sentence-transformers(MiniLM) and Hugging Face integration in Streamlit-ui -- helping engineering and operations teams focus on what matters most.
Fully local and open-source AI study companion for lecture PDFs - with slide summarization, smart Q&A, and flashcard creation using LangChain and Hugging Face Transformers.
Add a description, image, and links to the minilm topic page so that developers can more easily learn about it.
To associate your repository with the minilm topic, visit your repo's landing page and select "manage topics."