Conversation
|
@QuantumByte-01 please review. |
|
Interesting direction — extracting neuroscience terms from uploaded images via OCR and Gemini is a genuinely useful idea for researchers uploading paper figures. Three things needed before merging: pytesseract requires a system-level Tesseract installation which is not reflected in the Dockerfile or setup docs (add it to both), the frontend file upload UI needs styling and proper error states, and OCR accuracy on scientific figures without preprocessing will be poor — consider adding image preprocessing (contrast, binarization) before OCR. Please address these and reopen. |
|
Hi @QuantumByte-01 , thanks for looking into this.
Perfectly valid points, I wil work on this.. |
Fixes issue #16
Currently, the KnowledgeSpace AI agent relies only on text-based queries, while a large amount of
neuroscience metadata remains locked inside non-editable formats such as figures, tables, and
presentation screenshots. This PR introduces a multimodal pipeline that allows users to upload
images and automatically convert them into refined, high-signal search queries.
The feature includes:
preventing conversational noise from polluting the search engine.
Impact: