Retrieval Service
Overview
The Retrieval Service (RetrievalService
) provides text chunk retrieval functionality for the Kamiwaza AI Platform. Located in kamiwaza_client/services/retrieval.py
, this service handles the retrieval of relevant text chunks based on queries, enabling efficient information retrieval.
Key Features
- Relevant Text Chunk Retrieval
- Query-based Search
- Integration with Vector Database
- Semantic Search Capabilities
Text Chunk Retrieval
Available Methods
retrieve_relevant_chunks(query: str, k: int = 5) -> List[TextChunk]
: Get relevant text chunks based on query
# Retrieve relevant chunks
chunks = client.retrieval.retrieve_relevant_chunks(
query="What is machine learning?",
k=5 # Number of chunks to retrieve
)
# Process retrieved chunks
for chunk in chunks:
print(f"Text: {chunk.text}")
print(f"Score: {chunk.score}")
print(f"Source: {chunk.metadata.get('source')}")
Integration with Other Services
The Retrieval Service works in conjunction with:
- Embedding Service
- For converting queries into vector representations
- VectorDB Service
- For performing similarity search
- Ingestion Service
- For accessing processed and stored text chunks
Error Handling
The service includes built-in error handling for common scenarios:
try:
chunks = client.retrieval.retrieve_relevant_chunks(
query="example query"
)
except VectorDBError:
print("Vector database error")
except EmbeddingError:
print("Embedding generation error")
except APIError as e:
print(f"Operation failed: {e}")
Best Practices
- Use specific and focused queries
- Adjust the number of chunks (k) based on your needs
- Consider chunk relevance scores
- Process chunks in order of relevance
- Handle empty result sets appropriately
- Implement proper error handling
- Consider caching frequently retrieved chunks
- Monitor retrieval performance
Performance Considerations
- Query length affects retrieval time
- Number of chunks (k) impacts response time
- Vector database size influences search speed
- Embedding generation adds processing overhead