{"id":16002,"date":"2024-07-04T09:49:03","date_gmt":"2024-07-04T16:49:03","guid":{"rendered":"https:\/\/www.couchbase.com\/blog\/?p=16002"},"modified":"2025-06-13T16:36:50","modified_gmt":"2025-06-13T23:36:50","slug":"accelerate-rag-ai-couchbase-nvidia","status":"publish","type":"post","link":"https:\/\/www.couchbase.com\/blog\/es\/accelerate-rag-ai-couchbase-nvidia\/","title":{"rendered":"Aceleraci\u00f3n de la aplicaci\u00f3n de IA RAG basada en Couchbase con NVIDIA NIM\/NeMo y LangChain"},"content":{"rendered":"<p><span style=\"font-weight: 400;\">Hoy nos complace anunciar nuestra nueva integraci\u00f3n con NVIDIA NIM\/NeMo. En esta entrada de blog, presentamos un concepto de soluci\u00f3n de un chatbot interactivo basado en una <em>Recuperaci\u00f3n Generaci\u00f3n aumentada<\/em> (RAG)<\/span><span style=\"font-weight: 400;\">\u00a0<\/span><span style=\"font-weight: 400;\">con Couchbase Capella como base de datos vectorial. Las fases de recuperaci\u00f3n y generaci\u00f3n de la canalizaci\u00f3n RAG se aceleran mediante NVIDIA NIM\/NeMo con <\/span><span style=\"font-weight: 400;\">s\u00f3lo unas pocas l\u00edneas de c\u00f3digo.\u00a0<\/span><\/p>\n<p><span style=\"font-weight: 400;\">Empresas de diversos sectores se esfuerzan por ofrecer el mejor servicio a sus clientes. Para lograrlo, est\u00e1n dotando a sus trabajadores de primera l\u00ednea, como enfermeras de urgencias, vendedores de tiendas y representantes de servicios de asistencia, de chatbots interactivos de preguntas y respuestas (QA) con IA para recuperar informaci\u00f3n relevante y actualizada r\u00e1pidamente. <\/span><\/p>\n<p><span style=\"font-weight: 400;\">Los chatbots suelen basarse en <\/span><a href=\"https:\/\/www.couchbase.com\/blog\/es\/an-overview-of-retrieval-augmented-generation\/\"><span style=\"font-weight: 400;\">RAG<\/span><\/a><span style=\"font-weight: 400;\">un marco de IA utilizado para recuperar hechos de la base de conocimientos de la empresa con el fin de fundamentar las respuestas del LLM en la informaci\u00f3n m\u00e1s precisa y reciente. Consta de tres fases distintas, que comienzan con la recuperaci\u00f3n del contexto m\u00e1s relevante utilizando <\/span><a href=\"https:\/\/www.couchbase.com\/blog\/es\/products\/vector-search\/\"><span style=\"font-weight: 400;\">b\u00fasqueda vectorial<\/span><\/a><span style=\"font-weight: 400;\">Por \u00faltimo, se generan respuestas pertinentes mediante un LLM.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">El problema de las actuales canalizaciones de RAG es que las llamadas al servicio de incrustaci\u00f3n en la fase de recuperaci\u00f3n para convertir las peticiones del usuario en vectores pueden a\u00f1adir una latencia significativa, ralentizando las aplicaciones que requieren interactividad. Vectorizar un corpus de documentos compuesto por millones de PDF, documentos y otras bases de conocimiento puede llevar mucho tiempo, lo que aumenta la probabilidad de utilizar datos obsoletos para la GAR. Adem\u00e1s, a los usuarios les resulta dif\u00edcil acelerar la inferencia (tokens\/seg) de forma rentable para reducir el tiempo de respuesta de sus aplicaciones de chatbot.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">En la figura 1 se muestra una pila de rendimiento que le permitir\u00e1 desarrollar f\u00e1cilmente un <\/span><span style=\"font-weight: 400;\">chatbot interactivo de atenci\u00f3n al cliente. Se compone del marco de aplicaci\u00f3n StreamLit, LangChain para la orquestaci\u00f3n, Couchbase Capella para la indexaci\u00f3n y b\u00fasqueda de vectores, y NVIDIA NIM\/NeMo para acelerar las etapas de recuperaci\u00f3n y generaci\u00f3n.<\/span><\/p>\n<div id=\"attachment_16003\" style=\"width: 910px\" class=\"wp-caption alignnone\"><a href=\"https:\/\/www.couchbase.com\/blog\/wp-content\/uploads\/2024\/07\/image1-1.png\"><img loading=\"lazy\" decoding=\"async\" aria-describedby=\"caption-attachment-16003\" class=\"wp-image-16003 size-large\" style=\"border: solid black 1px;\" src=\"https:\/\/www.couchbase.com\/blog\/wp-content\/uploads\/2024\/07\/image1-1-1024x518.png\" alt=\"NVIDIA NIM\/NeMo and LangChain\" width=\"900\" height=\"455\" srcset=\"https:\/\/www.couchbase.com\/blog\/wp-content\/uploads\/sites\/1\/2024\/07\/image1-1-1024x518.png 1024w, https:\/\/www.couchbase.com\/blog\/wp-content\/uploads\/sites\/1\/2024\/07\/image1-1-300x152.png 300w, https:\/\/www.couchbase.com\/blog\/wp-content\/uploads\/sites\/1\/2024\/07\/image1-1-768x389.png 768w, https:\/\/www.couchbase.com\/blog\/wp-content\/uploads\/sites\/1\/2024\/07\/image1-1-1320x668.png 1320w, https:\/\/www.couchbase.com\/blog\/wp-content\/uploads\/sites\/1\/2024\/07\/image1-1.png 1345w\" sizes=\"auto, (max-width: 900px) 100vw, 900px\" \/><\/a><p id=\"caption-attachment-16003\" class=\"wp-caption-text\">Figura 1: Arquitectura conceptual de un chatbot de control de calidad creado con Capella y NVIDIA NIM\/NeMo<\/p><\/div>\n<p><span style=\"font-weight: 400;\">Couchbase Capella, una base de datos como servicio (DBaaS) de alto rendimiento, te permite empezar r\u00e1pidamente a almacenar, indexar y consultar datos operativos, vectoriales, de texto, de series temporales y geoespaciales aprovechando la flexibilidad de JSON. Puede integrar f\u00e1cilmente Capella para <\/span><a href=\"https:\/\/www.couchbase.com\/blog\/es\/products\/vector-search\/\"><span style=\"font-weight: 400;\">b\u00fasqueda vectorial<\/span><\/a><span style=\"font-weight: 400;\"> o la b\u00fasqueda sem\u00e1ntica sin necesidad de una base de datos vectorial independiente mediante la integraci\u00f3n de un marco de orquestaci\u00f3n como <\/span><a href=\"https:\/\/www.langchain.com\/\"><span style=\"font-weight: 400;\">Cadena LangChain<\/span><\/a><span style=\"font-weight: 400;\"> o <\/span><a href=\"https:\/\/www.llamaindex.ai\/\"><span style=\"font-weight: 400;\">LlamaIndex<\/span><\/a><span style=\"font-weight: 400;\"> en su canal de producci\u00f3n RAG. Ofrece la <\/span><a href=\"https:\/\/www.couchbase.com\/blog\/es\/hybrid-search\/\"><span style=\"font-weight: 400;\">b\u00fasqueda h\u00edbrida<\/span><\/a><span style=\"font-weight: 400;\"> que combina la b\u00fasqueda vectorial con la b\u00fasqueda tradicional para mejorar significativamente el rendimiento de la b\u00fasqueda. Adem\u00e1s, puedes ampliar la b\u00fasqueda vectorial a los bordes con Couchbase mobile para casos de uso de IA en los bordes.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">Una vez que haya configurado la b\u00fasqueda vectorial de Capella, puede proceder a elegir un modelo de rendimiento en el men\u00fa <\/span><a href=\"https:\/\/build.nvidia.com\/explore\/discover\"><span style=\"font-weight: 400;\">Cat\u00e1logo API de NVIDIA<\/span><\/a><span style=\"font-weight: 400;\">que ofrece una amplia gama de modelos de cimientos que abarcan modelos de c\u00f3digo abierto, cimientos de IA de NVIDIA y modelos personalizados, optimizados para ofrecer el mejor rendimiento en la infraestructura acelerada de NVIDIA. Estos modelos se implementan como <\/span><a href=\"https:\/\/developer.nvidia.com\/blog\/nvidia-nim-offers-optimized-inference-microservices-for-deploying-ai-models-at-scale\/?ref=blog.langchain.dev\"><span style=\"font-weight: 400;\">NVIDIA NIM<\/span><\/a><span style=\"font-weight: 400;\"> ya sea on-prem o en la nube utilizando contenedores preconstruidos f\u00e1ciles de usar a trav\u00e9s de un \u00fanico comando. Recuperador NeMo, <\/span><span style=\"font-weight: 400;\">parte de NVIDIA NeMo,<\/span><span style=\"font-weight: 400;\"> ofrece recuperaci\u00f3n de informaci\u00f3n con la menor latencia, el mayor rendimiento y la m\u00e1xima privacidad de los datos.\u00a0<\/span><\/p>\n<p><span style=\"font-weight: 400;\">El chatbot que hemos desarrollado utilizando la pila antes mencionada le permitir\u00e1 <\/span><span style=\"font-weight: 400;\">cargue sus documentos PDF y formule preguntas de forma interactiva. Utiliza <em>NV-QA-Embed<\/em>un modelo de incrustaci\u00f3n de texto acelerado en la GPU para la recuperaci\u00f3n de preguntas y respuestas, y <\/span><a href=\"https:\/\/build.nvidia.com\/meta\/llama3-70b\"><span style=\"font-weight: 400;\">Llama 3 - 70B<\/span><\/a><span style=\"font-weight: 400;\">que se empaqueta como NIM y se acelera en la infraestructura de NVIDIA. El sitio <\/span><a href=\"https:\/\/python.langchain.com\/v0.2\/docs\/integrations\/chat\/nvidia_ai_endpoints\/\"><span style=\"font-weight: 400;\">langchain-nvidia-ai-endpoints<\/span><\/a><span style=\"font-weight: 400;\"> contiene integraciones de LangChain para crear aplicaciones con modelos en NVIDIA NIM. <\/span><span style=\"font-weight: 400;\">Aunque hemos utilizado los endpoints alojados en NVIDIA para la creaci\u00f3n de prototipos, le recomendamos que considere la posibilidad de utilizar NIM autoalojado consultando la p\u00e1gina <\/span><a href=\"https:\/\/docs.nvidia.com\/nim\/large-language-models\/latest\/introduction.html?nvid=nv-int-tblg-432774\"><span style=\"font-weight: 400;\">Documentaci\u00f3n NIM<\/span><\/a><span style=\"font-weight: 400;\"> para despliegues de producci\u00f3n.\u00a0<\/span><\/p>\n<p><span style=\"font-weight: 400;\">Puede utilizar esta soluci\u00f3n para casos de uso que requieran una r\u00e1pida recuperaci\u00f3n de la informaci\u00f3n, como:<\/span><\/p>\n<ul>\n<li style=\"list-style-type: none;\">\n<ul>\n<li style=\"font-weight: 400;\" aria-level=\"1\"><span style=\"font-weight: 400;\">Permitir a las enfermeras de urgencias agilizar el triaje mediante un acceso r\u00e1pido a la informaci\u00f3n sanitaria pertinente para aliviar el hacinamiento, las largas esperas para recibir atenci\u00f3n y la escasa satisfacci\u00f3n de los pacientes.<\/span><\/li>\n<li style=\"font-weight: 400;\" aria-level=\"1\"><span style=\"font-weight: 400;\">Ayudar a los agentes de atenci\u00f3n al cliente a descubrir r\u00e1pidamente los conocimientos pertinentes mediante un chatbot de base de conocimientos interna para reducir los tiempos de espera de las llamadas. Esto no solo ayudar\u00e1 a mejorar los resultados de CSAT, sino que tambi\u00e9n permitir\u00e1 gestionar grandes vol\u00famenes de llamadas.<\/span><\/li>\n<li style=\"font-weight: 400;\" aria-level=\"1\"><span style=\"font-weight: 400;\">Ayudar a los vendedores de una tienda a descubrir y recomendar r\u00e1pidamente art\u00edculos de un cat\u00e1logo de productos similares a la imagen o descripci\u00f3n del art\u00edculo solicitado por un comprador pero que actualmente est\u00e1 agotado (stockout), para mejorar la experiencia de compra.<\/span><\/li>\n<\/ul>\n<\/li>\n<\/ul>\n<p><span style=\"font-weight: 400;\">En conclusi\u00f3n, puedes desarrollar una aplicaci\u00f3n GenAI interactiva, como un chatbot, con respuestas fundamentadas y relevantes utilizando RAG basado en Couchbase Capella y acelerarla utilizando NVIDIA NIM\/NeMo. <\/span><span style=\"font-weight: 400;\">Esta combinaci\u00f3n proporciona escalabilidad, fiabilidad y facilidad de uso. Adem\u00e1s de desplegarse junto con Capella para una experiencia DBaaS, NIM\/NeMo puede desplegarse con Couchbase on-prem o autogestionado en nubes p\u00fablicas dentro de su VPC para casos de uso que tengan requisitos m\u00e1s estrictos de seguridad y privacidad. Adem\u00e1s, puede utilizar <\/span><a href=\"https:\/\/developer.nvidia.com\/blog\/building-safer-llm-apps-with-langchain-templates-and-nvidia-nemo-guardrails\/\"><span style=\"font-weight: 400;\">Barandillas NeMo<\/span><\/a><span style=\"font-weight: 400;\"> para controlar la salida de su LLM de contenidos que su empresa considere censurables. <\/span><\/p>\n<p><span style=\"font-weight: 400;\">Los detalles de la aplicaci\u00f3n chatbot se pueden encontrar en el Couchbase <\/span><a href=\"https:\/\/github.com\/couchbase-examples\/couchbase-tutorials\/blob\/141424e68c18233c4ed47cc6321d38540ab4ca54\/tutorial\/markdown\/python\/nvidia-nim-llama3-pdf-chat\/nvidia-nim-llama3-pdf-chat.md\"><span style=\"font-weight: 400;\">Portal para desarrolladores<\/span><\/a><span style=\"font-weight: 400;\"> junto con el <\/span><a href=\"https:\/\/github.com\/couchbase-examples\/nvidia-rag-demo\/blob\/main\/chat_with_pdf.py\"><span style=\"font-weight: 400;\">c\u00f3digo completo<\/span><\/a><span style=\"font-weight: 400;\">. Inscr\u00edbase en <\/span><a href=\"https:\/\/cloud.couchbase.com\/sign-up\"><span style=\"font-weight: 400;\">Cuenta de prueba Capella<\/span><\/a><span style=\"font-weight: 400;\">, gratis <\/span><a href=\"https:\/\/build.nvidia.com\/explore\/discover?signin_corporate=false&amp;signin=false\"><span style=\"font-weight: 400;\">Cuenta NVIDIA NIM<\/span><\/a><span style=\"font-weight: 400;\">y empieza a desarrollar tu aplicaci\u00f3n GenAI.\u00a0<\/span><\/p>\n<p><br style=\"font-weight: 400;\" \/><br style=\"font-weight: 400;\" \/><\/p>","protected":false},"excerpt":{"rendered":"<p>Today, we&#8217;re excited to announce our new integration with NVIDIA NIM\/NeMo. In this blog post, we present a solution concept of an interactive chatbot based on a Retrieval Augmented Generation (RAG)\u00a0architecture with Couchbase Capella as a Vector database. The retrieval [&hellip;]<\/p>","protected":false},"author":84768,"featured_media":16003,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"_acf_changed":false,"inline_featured_image":false,"footnotes":""},"categories":[10122,2242,2225,1816,7666,9973,2389,9937],"tags":[9963,9989],"ppma_author":[9977,9981],"class_list":["post-16002","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-artificial-intelligence-ai","category-connectors","category-cloud","category-couchbase-server","category-edge-computing","category-generative-ai-genai","category-solutions","category-vector-search","tag-langchain","tag-nvidia"],"acf":[],"yoast_head":"<!-- This site is optimized with the Yoast SEO Premium plugin v25.7.1 (Yoast SEO v25.7) - https:\/\/yoast.com\/wordpress\/plugins\/seo\/ -->\n<title>Accelerate Couchbase-Powered RAG AI Application With NVIDIA NIM\/NeMo and LangChain - The Couchbase Blog<\/title>\n<meta name=\"description\" content=\"Develop an interactive GenAI application with grounded and relevant responses using Couchbase Capella-based RAG and accelerate it using NVIDIA NIM\/NeMo\" \/>\n<meta name=\"robots\" content=\"index, follow, max-snippet:-1, max-image-preview:large, max-video-preview:-1\" \/>\n<link rel=\"canonical\" href=\"https:\/\/www.couchbase.com\/blog\/es\/accelerate-rag-ai-couchbase-nvidia\/\" \/>\n<meta property=\"og:locale\" content=\"es_MX\" \/>\n<meta property=\"og:type\" content=\"article\" \/>\n<meta property=\"og:title\" content=\"Accelerate Couchbase-Powered RAG AI Application With NVIDIA NIM\/NeMo and LangChain\" \/>\n<meta property=\"og:description\" content=\"Develop an interactive GenAI application with grounded and relevant responses using Couchbase Capella-based RAG and accelerate it using NVIDIA NIM\/NeMo\" \/>\n<meta property=\"og:url\" content=\"https:\/\/www.couchbase.com\/blog\/es\/accelerate-rag-ai-couchbase-nvidia\/\" \/>\n<meta property=\"og:site_name\" content=\"The Couchbase Blog\" \/>\n<meta property=\"article:published_time\" content=\"2024-07-04T16:49:03+00:00\" \/>\n<meta property=\"article:modified_time\" content=\"2025-06-13T23:36:50+00:00\" \/>\n<meta property=\"og:image\" content=\"https:\/\/www.couchbase.com\/blog\/wp-content\/uploads\/sites\/1\/2024\/07\/image1-1.png\" \/>\n\t<meta property=\"og:image:width\" content=\"1345\" \/>\n\t<meta property=\"og:image:height\" content=\"681\" \/>\n\t<meta property=\"og:image:type\" content=\"image\/png\" \/>\n<meta name=\"author\" content=\"Lokesh Goel, Software Engineer, Kiran Matty, Lead Product Manager AI\/ML\" \/>\n<meta name=\"twitter:card\" content=\"summary_large_image\" \/>\n<meta name=\"twitter:label1\" content=\"Written by\" \/>\n\t<meta name=\"twitter:data1\" content=\"Lokesh Goel, Software Engineer\" \/>\n\t<meta name=\"twitter:label2\" content=\"Est. reading time\" \/>\n\t<meta name=\"twitter:data2\" content=\"5 minutos\" \/>\n<script type=\"application\/ld+json\" class=\"yoast-schema-graph\">{\"@context\":\"https:\/\/schema.org\",\"@graph\":[{\"@type\":\"Article\",\"@id\":\"https:\/\/www.couchbase.com\/blog\/accelerate-rag-ai-couchbase-nvidia\/#article\",\"isPartOf\":{\"@id\":\"https:\/\/www.couchbase.com\/blog\/accelerate-rag-ai-couchbase-nvidia\/\"},\"author\":{\"name\":\"Lokesh Goel, Developer Experience Engineer\",\"@id\":\"https:\/\/www.couchbase.com\/blog\/#\/schema\/person\/a918924898a24c1cbcf6712bb6d62b4e\"},\"headline\":\"Accelerate Couchbase-Powered RAG AI Application With NVIDIA NIM\/NeMo and LangChain\",\"datePublished\":\"2024-07-04T16:49:03+00:00\",\"dateModified\":\"2025-06-13T23:36:50+00:00\",\"mainEntityOfPage\":{\"@id\":\"https:\/\/www.couchbase.com\/blog\/accelerate-rag-ai-couchbase-nvidia\/\"},\"wordCount\":859,\"commentCount\":0,\"publisher\":{\"@id\":\"https:\/\/www.couchbase.com\/blog\/#organization\"},\"image\":{\"@id\":\"https:\/\/www.couchbase.com\/blog\/accelerate-rag-ai-couchbase-nvidia\/#primaryimage\"},\"thumbnailUrl\":\"https:\/\/www.couchbase.com\/blog\/wp-content\/uploads\/sites\/1\/2024\/07\/image1-1.png\",\"keywords\":[\"langchain\",\"NVIDIA\"],\"articleSection\":[\"Artificial Intelligence (AI)\",\"Connectors\",\"Couchbase Capella\",\"Couchbase Server\",\"Edge computing\",\"Generative AI (GenAI)\",\"Solutions\",\"Vector Search\"],\"inLanguage\":\"es\",\"potentialAction\":[{\"@type\":\"CommentAction\",\"name\":\"Comment\",\"target\":[\"https:\/\/www.couchbase.com\/blog\/accelerate-rag-ai-couchbase-nvidia\/#respond\"]}]},{\"@type\":\"WebPage\",\"@id\":\"https:\/\/www.couchbase.com\/blog\/accelerate-rag-ai-couchbase-nvidia\/\",\"url\":\"https:\/\/www.couchbase.com\/blog\/accelerate-rag-ai-couchbase-nvidia\/\",\"name\":\"Accelerate Couchbase-Powered RAG AI Application With NVIDIA NIM\/NeMo and LangChain - The Couchbase Blog\",\"isPartOf\":{\"@id\":\"https:\/\/www.couchbase.com\/blog\/#website\"},\"primaryImageOfPage\":{\"@id\":\"https:\/\/www.couchbase.com\/blog\/accelerate-rag-ai-couchbase-nvidia\/#primaryimage\"},\"image\":{\"@id\":\"https:\/\/www.couchbase.com\/blog\/accelerate-rag-ai-couchbase-nvidia\/#primaryimage\"},\"thumbnailUrl\":\"https:\/\/www.couchbase.com\/blog\/wp-content\/uploads\/sites\/1\/2024\/07\/image1-1.png\",\"datePublished\":\"2024-07-04T16:49:03+00:00\",\"dateModified\":\"2025-06-13T23:36:50+00:00\",\"description\":\"Develop an interactive GenAI application with grounded and relevant responses using Couchbase Capella-based RAG and accelerate it using NVIDIA NIM\/NeMo\",\"breadcrumb\":{\"@id\":\"https:\/\/www.couchbase.com\/blog\/accelerate-rag-ai-couchbase-nvidia\/#breadcrumb\"},\"inLanguage\":\"es\",\"potentialAction\":[{\"@type\":\"ReadAction\",\"target\":[\"https:\/\/www.couchbase.com\/blog\/accelerate-rag-ai-couchbase-nvidia\/\"]}]},{\"@type\":\"ImageObject\",\"inLanguage\":\"es\",\"@id\":\"https:\/\/www.couchbase.com\/blog\/accelerate-rag-ai-couchbase-nvidia\/#primaryimage\",\"url\":\"https:\/\/www.couchbase.com\/blog\/wp-content\/uploads\/sites\/1\/2024\/07\/image1-1.png\",\"contentUrl\":\"https:\/\/www.couchbase.com\/blog\/wp-content\/uploads\/sites\/1\/2024\/07\/image1-1.png\",\"width\":1345,\"height\":681,\"caption\":\"NVIDIA NIM\/NeMo and LangChain\"},{\"@type\":\"BreadcrumbList\",\"@id\":\"https:\/\/www.couchbase.com\/blog\/accelerate-rag-ai-couchbase-nvidia\/#breadcrumb\",\"itemListElement\":[{\"@type\":\"ListItem\",\"position\":1,\"name\":\"Home\",\"item\":\"https:\/\/www.couchbase.com\/blog\/\"},{\"@type\":\"ListItem\",\"position\":2,\"name\":\"Accelerate Couchbase-Powered RAG AI Application With NVIDIA NIM\/NeMo and LangChain\"}]},{\"@type\":\"WebSite\",\"@id\":\"https:\/\/www.couchbase.com\/blog\/#website\",\"url\":\"https:\/\/www.couchbase.com\/blog\/\",\"name\":\"The Couchbase Blog\",\"description\":\"Couchbase, the NoSQL Database\",\"publisher\":{\"@id\":\"https:\/\/www.couchbase.com\/blog\/#organization\"},\"potentialAction\":[{\"@type\":\"SearchAction\",\"target\":{\"@type\":\"EntryPoint\",\"urlTemplate\":\"https:\/\/www.couchbase.com\/blog\/?s={search_term_string}\"},\"query-input\":{\"@type\":\"PropertyValueSpecification\",\"valueRequired\":true,\"valueName\":\"search_term_string\"}}],\"inLanguage\":\"es\"},{\"@type\":\"Organization\",\"@id\":\"https:\/\/www.couchbase.com\/blog\/#organization\",\"name\":\"The Couchbase Blog\",\"url\":\"https:\/\/www.couchbase.com\/blog\/\",\"logo\":{\"@type\":\"ImageObject\",\"inLanguage\":\"es\",\"@id\":\"https:\/\/www.couchbase.com\/blog\/#\/schema\/logo\/image\/\",\"url\":\"https:\/\/www.couchbase.com\/blog\/wp-content\/uploads\/2023\/04\/admin-logo.png\",\"contentUrl\":\"https:\/\/www.couchbase.com\/blog\/wp-content\/uploads\/2023\/04\/admin-logo.png\",\"width\":218,\"height\":34,\"caption\":\"The Couchbase Blog\"},\"image\":{\"@id\":\"https:\/\/www.couchbase.com\/blog\/#\/schema\/logo\/image\/\"}},{\"@type\":\"Person\",\"@id\":\"https:\/\/www.couchbase.com\/blog\/#\/schema\/person\/a918924898a24c1cbcf6712bb6d62b4e\",\"name\":\"Lokesh Goel, Developer Experience Engineer\",\"image\":{\"@type\":\"ImageObject\",\"inLanguage\":\"es\",\"@id\":\"https:\/\/www.couchbase.com\/blog\/#\/schema\/person\/image\/28f42fa6eaa9ec33a742151714d1f0cb\",\"url\":\"https:\/\/secure.gravatar.com\/avatar\/8f2cb3333278f50e81806e9c068732cf57d7268c2b1ed80cc3dc9645151df405?s=96&d=mm&r=g\",\"contentUrl\":\"https:\/\/secure.gravatar.com\/avatar\/8f2cb3333278f50e81806e9c068732cf57d7268c2b1ed80cc3dc9645151df405?s=96&d=mm&r=g\",\"caption\":\"Lokesh Goel, Developer Experience Engineer\"},\"url\":\"https:\/\/www.couchbase.com\/blog\/es\/author\/lokeshgoel\/\"}]}<\/script>\n<!-- \/ Yoast SEO Premium plugin. -->","yoast_head_json":{"title":"Accelerate Couchbase-Powered RAG AI Application With NVIDIA NIM\/NeMo and LangChain - The Couchbase Blog","description":"Develop an interactive GenAI application with grounded and relevant responses using Couchbase Capella-based RAG and accelerate it using NVIDIA NIM\/NeMo","robots":{"index":"index","follow":"follow","max-snippet":"max-snippet:-1","max-image-preview":"max-image-preview:large","max-video-preview":"max-video-preview:-1"},"canonical":"https:\/\/www.couchbase.com\/blog\/es\/accelerate-rag-ai-couchbase-nvidia\/","og_locale":"es_MX","og_type":"article","og_title":"Accelerate Couchbase-Powered RAG AI Application With NVIDIA NIM\/NeMo and LangChain","og_description":"Develop an interactive GenAI application with grounded and relevant responses using Couchbase Capella-based RAG and accelerate it using NVIDIA NIM\/NeMo","og_url":"https:\/\/www.couchbase.com\/blog\/es\/accelerate-rag-ai-couchbase-nvidia\/","og_site_name":"The Couchbase Blog","article_published_time":"2024-07-04T16:49:03+00:00","article_modified_time":"2025-06-13T23:36:50+00:00","og_image":[{"width":1345,"height":681,"url":"https:\/\/www.couchbase.com\/blog\/wp-content\/uploads\/sites\/1\/2024\/07\/image1-1.png","type":"image\/png"}],"author":"Lokesh Goel, Software Engineer, Kiran Matty, Lead Product Manager AI\/ML","twitter_card":"summary_large_image","twitter_misc":{"Written by":"Lokesh Goel, Software Engineer","Est. reading time":"5 minutos"},"schema":{"@context":"https:\/\/schema.org","@graph":[{"@type":"Article","@id":"https:\/\/www.couchbase.com\/blog\/accelerate-rag-ai-couchbase-nvidia\/#article","isPartOf":{"@id":"https:\/\/www.couchbase.com\/blog\/accelerate-rag-ai-couchbase-nvidia\/"},"author":{"name":"Lokesh Goel, Developer Experience Engineer","@id":"https:\/\/www.couchbase.com\/blog\/#\/schema\/person\/a918924898a24c1cbcf6712bb6d62b4e"},"headline":"Accelerate Couchbase-Powered RAG AI Application With NVIDIA NIM\/NeMo and LangChain","datePublished":"2024-07-04T16:49:03+00:00","dateModified":"2025-06-13T23:36:50+00:00","mainEntityOfPage":{"@id":"https:\/\/www.couchbase.com\/blog\/accelerate-rag-ai-couchbase-nvidia\/"},"wordCount":859,"commentCount":0,"publisher":{"@id":"https:\/\/www.couchbase.com\/blog\/#organization"},"image":{"@id":"https:\/\/www.couchbase.com\/blog\/accelerate-rag-ai-couchbase-nvidia\/#primaryimage"},"thumbnailUrl":"https:\/\/www.couchbase.com\/blog\/wp-content\/uploads\/sites\/1\/2024\/07\/image1-1.png","keywords":["langchain","NVIDIA"],"articleSection":["Artificial Intelligence (AI)","Connectors","Couchbase Capella","Couchbase Server","Edge computing","Generative AI (GenAI)","Solutions","Vector Search"],"inLanguage":"es","potentialAction":[{"@type":"CommentAction","name":"Comment","target":["https:\/\/www.couchbase.com\/blog\/accelerate-rag-ai-couchbase-nvidia\/#respond"]}]},{"@type":"WebPage","@id":"https:\/\/www.couchbase.com\/blog\/accelerate-rag-ai-couchbase-nvidia\/","url":"https:\/\/www.couchbase.com\/blog\/accelerate-rag-ai-couchbase-nvidia\/","name":"Accelerate Couchbase-Powered RAG AI Application With NVIDIA NIM\/NeMo and LangChain - The Couchbase Blog","isPartOf":{"@id":"https:\/\/www.couchbase.com\/blog\/#website"},"primaryImageOfPage":{"@id":"https:\/\/www.couchbase.com\/blog\/accelerate-rag-ai-couchbase-nvidia\/#primaryimage"},"image":{"@id":"https:\/\/www.couchbase.com\/blog\/accelerate-rag-ai-couchbase-nvidia\/#primaryimage"},"thumbnailUrl":"https:\/\/www.couchbase.com\/blog\/wp-content\/uploads\/sites\/1\/2024\/07\/image1-1.png","datePublished":"2024-07-04T16:49:03+00:00","dateModified":"2025-06-13T23:36:50+00:00","description":"Develop an interactive GenAI application with grounded and relevant responses using Couchbase Capella-based RAG and accelerate it using NVIDIA NIM\/NeMo","breadcrumb":{"@id":"https:\/\/www.couchbase.com\/blog\/accelerate-rag-ai-couchbase-nvidia\/#breadcrumb"},"inLanguage":"es","potentialAction":[{"@type":"ReadAction","target":["https:\/\/www.couchbase.com\/blog\/accelerate-rag-ai-couchbase-nvidia\/"]}]},{"@type":"ImageObject","inLanguage":"es","@id":"https:\/\/www.couchbase.com\/blog\/accelerate-rag-ai-couchbase-nvidia\/#primaryimage","url":"https:\/\/www.couchbase.com\/blog\/wp-content\/uploads\/sites\/1\/2024\/07\/image1-1.png","contentUrl":"https:\/\/www.couchbase.com\/blog\/wp-content\/uploads\/sites\/1\/2024\/07\/image1-1.png","width":1345,"height":681,"caption":"NVIDIA NIM\/NeMo and LangChain"},{"@type":"BreadcrumbList","@id":"https:\/\/www.couchbase.com\/blog\/accelerate-rag-ai-couchbase-nvidia\/#breadcrumb","itemListElement":[{"@type":"ListItem","position":1,"name":"Home","item":"https:\/\/www.couchbase.com\/blog\/"},{"@type":"ListItem","position":2,"name":"Accelerate Couchbase-Powered RAG AI Application With NVIDIA NIM\/NeMo and LangChain"}]},{"@type":"WebSite","@id":"https:\/\/www.couchbase.com\/blog\/#website","url":"https:\/\/www.couchbase.com\/blog\/","name":"El blog de Couchbase","description":"Couchbase, la base de datos NoSQL","publisher":{"@id":"https:\/\/www.couchbase.com\/blog\/#organization"},"potentialAction":[{"@type":"SearchAction","target":{"@type":"EntryPoint","urlTemplate":"https:\/\/www.couchbase.com\/blog\/?s={search_term_string}"},"query-input":{"@type":"PropertyValueSpecification","valueRequired":true,"valueName":"search_term_string"}}],"inLanguage":"es"},{"@type":"Organization","@id":"https:\/\/www.couchbase.com\/blog\/#organization","name":"El blog de Couchbase","url":"https:\/\/www.couchbase.com\/blog\/","logo":{"@type":"ImageObject","inLanguage":"es","@id":"https:\/\/www.couchbase.com\/blog\/#\/schema\/logo\/image\/","url":"https:\/\/www.couchbase.com\/blog\/wp-content\/uploads\/2023\/04\/admin-logo.png","contentUrl":"https:\/\/www.couchbase.com\/blog\/wp-content\/uploads\/2023\/04\/admin-logo.png","width":218,"height":34,"caption":"The Couchbase Blog"},"image":{"@id":"https:\/\/www.couchbase.com\/blog\/#\/schema\/logo\/image\/"}},{"@type":"Person","@id":"https:\/\/www.couchbase.com\/blog\/#\/schema\/person\/a918924898a24c1cbcf6712bb6d62b4e","name":"Lokesh Goel, Ingeniero de Experiencia del Desarrollador","image":{"@type":"ImageObject","inLanguage":"es","@id":"https:\/\/www.couchbase.com\/blog\/#\/schema\/person\/image\/28f42fa6eaa9ec33a742151714d1f0cb","url":"https:\/\/secure.gravatar.com\/avatar\/8f2cb3333278f50e81806e9c068732cf57d7268c2b1ed80cc3dc9645151df405?s=96&d=mm&r=g","contentUrl":"https:\/\/secure.gravatar.com\/avatar\/8f2cb3333278f50e81806e9c068732cf57d7268c2b1ed80cc3dc9645151df405?s=96&d=mm&r=g","caption":"Lokesh Goel, Developer Experience Engineer"},"url":"https:\/\/www.couchbase.com\/blog\/es\/author\/lokeshgoel\/"}]}},"authors":[{"term_id":9977,"user_id":84768,"is_guest":0,"slug":"lokeshgoel","display_name":"Lokesh Goel, Software Engineer","avatar_url":"https:\/\/secure.gravatar.com\/avatar\/8f2cb3333278f50e81806e9c068732cf57d7268c2b1ed80cc3dc9645151df405?s=96&d=mm&r=g","first_name":"Lokesh","last_name":"Goel, Software Engineer","user_url":"","author_category":"","description":""},{"term_id":9981,"user_id":85346,"is_guest":0,"slug":"kiranmatty","display_name":"Kiran Matty, Lead Product Manager AI\/ML","avatar_url":{"url":"https:\/\/www.couchbase.com\/blog\/wp-content\/uploads\/sites\/1\/2024\/06\/T024FJS4M-U064W1AETPD-456e21a66cf5-512.png","url2x":"https:\/\/www.couchbase.com\/blog\/wp-content\/uploads\/sites\/1\/2024\/06\/T024FJS4M-U064W1AETPD-456e21a66cf5-512.png"},"first_name":"Kiran","last_name":"Matty, Lead Product Manager AI\/ML","user_url":"","author_category":"","description":""}],"_links":{"self":[{"href":"https:\/\/www.couchbase.com\/blog\/es\/wp-json\/wp\/v2\/posts\/16002","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.couchbase.com\/blog\/es\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.couchbase.com\/blog\/es\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.couchbase.com\/blog\/es\/wp-json\/wp\/v2\/users\/84768"}],"replies":[{"embeddable":true,"href":"https:\/\/www.couchbase.com\/blog\/es\/wp-json\/wp\/v2\/comments?post=16002"}],"version-history":[{"count":0,"href":"https:\/\/www.couchbase.com\/blog\/es\/wp-json\/wp\/v2\/posts\/16002\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.couchbase.com\/blog\/es\/wp-json\/wp\/v2\/media\/16003"}],"wp:attachment":[{"href":"https:\/\/www.couchbase.com\/blog\/es\/wp-json\/wp\/v2\/media?parent=16002"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.couchbase.com\/blog\/es\/wp-json\/wp\/v2\/categories?post=16002"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.couchbase.com\/blog\/es\/wp-json\/wp\/v2\/tags?post=16002"},{"taxonomy":"author","embeddable":true,"href":"https:\/\/www.couchbase.com\/blog\/es\/wp-json\/wp\/v2\/ppma_author?post=16002"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}