Rugas Technologies

Leveraging LLMs to Streamline Pharma Case Processing: Reducing Time and Costs with iRxSafe

Leveraging LLMs to Streamline Pharma Case Processing: Reducing Time and Costs with iRxSafe Highlights: iRxSafe is a AI driven Pharmacovigilance intake case processing platform Handles incoming emails and classifies them as safety, product quality issue or other categories AI based extraction of drug details, adverse event details, patient conditions etc (Based on trained & fine tuned LLM Claude Sonnet) Automatic follow up email writing to collect missing details to enable the case to be processed The solution is end to end hosted inside AWS, ensuring data security and scalability and can be plugged into a larger case processing & reporting system (As an intake handling tool) How LLMs are used: The trained LLMs are used for the following tasks. Identifying the category of the incoming email – into the following categories Product safety (adverse event) Product quality Seeking product info Spam or other categories Extracting details of the adverse event / case Primary drug and dosage Patient information Adverse event information Concomitant drugs if any Patient conditions for which the drug was taken , etc Creating Followup email for requesting missing information We have evaluated various models and we have zeroed on Claude Sonnet for its better performance of the above tasks cost effectively. We had to fine tune the model for certain conditions of identifying primary drug and concomitant drugs and the model could improve the performance after the training. Human in the loop was built into the system where the System waits for user confirmation of the AI models’ conclusions, where the user can edit, correct and add information which the AI models fail to detect. This information again fed into the training pipeline for further fine tuning the model.              

Read More

Revolutionizing User Interaction: Langchain Integration in LLM-Powered Projects

Recently integrated the langchain framework into our LLM-driven projects, encompassing LangChain Templates, LangServe, and LangSmith. Langchain facilitates seamless communication between users and LLMs through vector databases, connecting with leading providers like Hugging Face, OpenAI, and Amazon Bedrock. Its functionalities span from agents to Retrieval-Augmented Generation (RAG), enabling advanced actions beyond simple chat responses. Despite challenges in text-triggered event implementation, langchain effectively overcame these hurdles. Retrieval augmented generation, a pivotal feature, enhances user interaction by delivering precise answers through semantic searches using vector embeddings. Additionally, we addressed a problem where user profiles were recommended based on skills, utilizing embeddings to bridge the gap between free-text queries and user-provided skills. The embedded profiles in our vector database were called using similarity metrics. Our end user had an LLM process all the required information to create an ideal profile, and embeddings similarity metrics sorted the recommended users. Integrating langchain with our customized knowledge base on platforms like MongoDB and Amazon Bedrock significantly enhances user experience. Langchain’s versatility in working with various vector database types underscores its importance in our projects, facilitating smoother communication and advanced functionalities with LLMs.

Read More