
Concluído
Publicado
Pago na entrega
I’m putting together a lean Retrieval-Augmented-Generation web application that lets users type a natural-language query, semantically searches a private document collection, then returns ranked matches and an auto-generated answer. Everything must stay fully open-source. Core stack • Model: BERT (Hugging Face), fine-tuned or out-of-the-box for text embeddings. • Vector store: FAISS, Milvus, Weaviate or a similarly licence-friendly alternative. • Back-end: Python with FastAPI (or Flask) and PyTorch/Sentence-Transformers. • Front-end: a minimal HTML/React page that submits a query and shows results + generated summary. Scope of work 1. Set up the pipeline that ingests text files/markdown/PDF, chunks them, builds the vector index, and stores metadata. 2. Expose REST endpoints for “/ingest” and “/search”. 3. For every search, retrieve top-k passages with cosine similarity and feed them back into the generation step, then stream the answer. 4. Package the whole thing in Docker so I can run it locally or deploy to a small VPS. 5. Provide a concise README with environment setup, run commands, and sample curl calls. Acceptance criteria • Search returns relevant passages for at least 90 % of the supplied test queries. • Latency per query (on CPU) under two seconds for a 10 k-document corpus. • All code is clean, commented, and reproducible from a single docker-compose up. Optional but nice to have – Basic user authentication. – Hot-reload ingestion so new documents appear without a full re-index. If you’ve already wired up BERT-based semantic search or built RAG demos, I’d love to see a quick link or repo. Let’s keep this straightforward, open-source, and ready to extend.
ID do Projeto: 40145637
14 propostas
Projeto remoto
Ativo há 2 meses
Defina seu orçamento e seu prazo
Seja pago pelo seu trabalho
Descreva sua proposta
É grátis para se inscrever e fazer ofertas em trabalhos

Hi,I am ML Engineer from Banglore.I can build this as mentioned.I have 8 years of experience in ML,AI,RAG,[login to view URL]'s connect
₹1.500 INR em 1 dia
6,3
6,3
14 freelancers estão ofertando em média ₹1.096 INR for esse trabalho

Hello Sir, I have read your requirements carefully. I am confident I can build Semantic RAG Web Application in Python (Flask) using BERT Model for fine tuned, FAISS for vector storing and Sentence-Transformers. I am a full stack developer and more than 5 years of experience in Development. I can handle entire process end to end securely. Regards & Thankis Jitendra Sharma
₹1.050 INR em 7 dias
4,2
4,2

I can guarantee sub-two-second CPU latency for your open-source RAG application by optimizing indexing with FAISS and Sentence Transformers, ensuring the solution remains lightweight and free of costly licensing dependencies. I understand that the core of the project is balancing 90% retrieval accuracy with an architecture that is easy to deploy, so I will develop the backend in FastAPI, integrating an efficient BERT model to quickly process cosine similarity. For delivery, I will set up a single Dockerized environment that orchestrates PDF and Markdown ingestion, the vector store, and a minimal React frontend, ready to run with a single command. Given my track record building semantic search pipelines, I can include "hot reload" functionality for new documents and basic authentication without significantly extending the timeline. I estimate completing the fully documented system in 10 days. I have attached a link to a repository with a previous vector search implementation for you to review my coding style. Would you like to briefly comment on the initial volume of documents to adjust the fragmentation strategy?
₹1.050 INR em 7 dias
3,1
3,1

I've built several semantic search pipelines using exactly this stack—Sentence-Transformers for embeddings, FAISS for vector indexing, and FastAPI for the REST layer—so I can hit the ground running on your RAG application. My approach: I'll create a modular ingestion service that chunks documents (PDF/markdown/text) with overlap for context preservation, generate BERT embeddings, and store them in FAISS with metadata indexing. The `/search` endpoint will retrieve top-k passages via cosine similarity, then stream the generated answer using a lightweight open-source LLM. Everything gets containerized with Docker Compose for single-command deployment, and I'll ensure sub-two-second latency through batch processing and optional GPU acceleration. I can start immediately.
₹600 INR em 1 dia
2,9
2,9

Hello, I’ve carefully reviewed your project requirements and clearly understand the tasks involved. I have 13 years of experience and strong expertise in the exact skills this project requires. I have successfully delivered similar projects before and can share relevant samples if needed. I will complete this within your expected timeline while maintaining quality and clear communication. I look forward to working with you and contributing sincerely to your project’s success.
₹1.050 INR em 7 dias
2,6
2,6

I have reviewed your project. I will plan it, build it, and test it step by step. I will use simple tools like Figma and clean code. You will see results fast and clear. Let’s start now and win today.
₹1.050 INR em 7 dias
0,0
0,0

Hi, I can build your web and mobile app with modern design, fast performance, and full functionality. With 3+ years of experience, I deliver responsive, SEO-friendly, and user-friendly solutions. Ready to start immediately.
₹1.050 INR em 1 dia
0,0
0,0

Hello. I will build an open-source Semantic RAG web app using Python, FastAPI, BERT embeddings from Hugging Face, and a license-friendly vector store like FAISS, delivering a clean ingest-search-generate pipeline. I will implement chunking, metadata storage, cosine similarity top-k retrieval, and streaming answers, packaged with Docker and docker-compose for one-command setup. The frontend will be a minimal HTML/React UI connected to REST endpoints with CPU latency targets under 2s for ~10k documents. I have experience building 4+ NLP and semantic search projects, including BERT-based embeddings, RAG demos, and FastAPI services, achieving ~90% relevant retrieval on test queries and stable performance on small VPS setups. Similar systems handled 5k–15k documents, exposed 2–4 core endpoints, and remained fully reproducible with open-source tooling. This background aligns closely with your acceptance criteria and extension goals. I am a new freelancer, which means strong focus, fast communication, and careful attention to clean, documented, auditable code. I prioritize reproducibility, readability, and open-source best practices over shortcuts. If this sounds good, connect in chat and we can start. Thank you, Jaroslav Caprata
₹1.050 INR em 1 dia
0,0
0,0

As an AI-focused full-stack developer, I possess the ideal skill set for your Open-Source Semantic RAG Web App project. Over my 5+ years of experience, I have led and executed numerous projects with similar complexities, which makes me confident about my capability to handle your case successfully. Most notably, I have developed AI-powered tools and built automation workflows that have optimized business processes for various clients. My expertise in Python, Django, Flask will be invaluable in setting up the pipeline for your project - including ingesting text files and building vector indexes - as well as in creating the necessary REST endpoints. In addition, my prior experience with BERT (Hugging Face) makes me familiar with its application to semantic search and retrieval-augmented-generation. I assure you of a fully open-source approach at all times. Moreover, during my career journey, I have developed clean UI/UX-focused web applications using minimal HTML/React pages.. You can expect similar dedication from me for your front-end needs. I understand the significance of speed optimization and deployed many applications on local servers requiring less than 2 seconds latency contributing to a seamless user experience. Lastly, I'm well-versed in Dockerization and AWS deployment to ensure easy scalability, clear documentation and a simple setup that's replicable from a single command
₹1.050 INR em 7 dias
0,0
0,0

Hi, happy new year! I have strong, hands-on experience designing and delivering fully open-source semantic search and RAG systems using BERT-class encoders and vector databases in production-grade Python stacks. I can implement a clean ingestion pipeline (PDF/Markdown/Text) with deterministic chunking, metadata persistence, and FAISS/Milvus/Weaviate indexing using Sentence-Transformers on top of Hugging Face models, optimized for low-latency CPU inference. The backend will be built with FastAPI, exposing well-defined /ingest and /search endpoints, streaming generation responses and enforcing reproducible embedding and retrieval logic. I am comfortable packaging the entire system with Docker and docker-compose, including model caching, volume-mounted indexes, and resource-aware configuration for VPS deployment. The frontend will be a minimal React or HTML client focused on query submission, ranked passage display, and answer rendering, without unnecessary dependencies. I will deliver clean, well-documented code, a concise but precise README, and reproducible run instructions, drawing directly on prior experience building BERT-based semantic search and RAG demos that meet strict latency and accuracy targets. Let's chat to discuss in more detail.
₹1.000 INR em 2 dias
0,0
0,0

Hey there! To make sure we hit <2s CPU latency on 10k docs with bulletproof open-source retrieval, could you confirm your preferred BERT-based embedding model (all-MiniLM-L6-v2, paraphrase-MiniLM-L6-v2) and the top vector store priority (FAISS for speed, Milvus/Weaviate for scalability)? I’d suggest rolling with all-MiniLM-L6-v2 (fastest CPU-friendly embedder) + FAISS as default index with an optional Milvus/Weaviate toggle via env var this keeps everything lightweight, fully open-source, and super easy to extend later. Relevant RAG Work - Built a private RAG knowledge hub with Sentence-Transformers + FAISS + FastAPI, ingesting PDFs/HTML, exact source citations, <1.5s queries on CPU, and clean Docker setup. - Delivered modular open-source RAG pipelines with Hugging Face embeddings, REST endpoints (/ingest & /search), streaming answers, and one-command docker-compose up. A quick chat would help lock in model choices and get us moving fast. When are you free this week? Best regards, Haseeb
₹1.400 INR em 7 dias
0,0
0,0

Hi! I read your brief and I have the exact stack you need: FastAPI, BERT/Sentence-Transformers, and Docker. I focus heavily on containerization and clean architecture to guarantee your "one command setup" requirement. I'll use FAISS for the vector store to ensure high speed on standard CPUs. I’m currently building "DraftFlow," an AI-powered tool that processes raw text using this same backend stack, so I am very comfortable with these workflows. I can get this running for you ASAP. Let's chat!
₹1.250 INR em 5 dias
0,0
0,0

I have built several RAG pipelines using exactly this stack. Here is what I will deliver: 1. Document ingestion pipeline: text/markdown/PDF chunking with configurable overlap, using LangChain or custom splitter 2. FAISS vector store with Sentence-Transformers embeddings (all-MiniLM-L6-v2 or similar) 3. FastAPI backend with /ingest and /search endpoints, streaming responses via SSE 4. Minimal React frontend for query input and result display 5. Docker Compose setup for single-command deployment 6. Clean README with environment setup and sample curl calls I understand your acceptance criteria: 90% relevant passage retrieval and sub-2s latency on CPU for 10k docs. This is achievable with FAISS and proper indexing. I can add the optional hot-reload ingestion and basic auth as well. Ready to start immediately.
₹1.200 INR em 7 dias
0,0
0,0

Hyderabad, India
Método de pagamento verificado
Membro desde mar. 30, 2013
₹1500-12500 INR
₹1500-12500 INR
₹37500-75000 INR
₹1500-12500 INR
₹75000-150000 INR
$30-250 USD
€10000-20000 EUR
₹150000-250000 INR
$750-1500 USD
₹12500-37500 INR
$2-8 USD / hora
$15-25 USD / hora
$10-30 USD
$15-25 USD / hora
₹12500-37500 INR
₹1500-12500 INR
$10-30 USD
₹600-1500 INR
₹1500-12500 INR
$5-10 USD / hora
₹12500-37500 INR
$15-25 USD / hora
$15-25 USD / hora
₹1250-2500 INR / hora