Avaamo | Interview (All rounds) | Senior MLE | Bengaluru

avaamo logo
avaamo
· Senior Machine Learning Engineer· Bengaluru· 3.5y exp
June 24, 2025 · 24 reads

Summary

I interviewed at Avaamo in December 2024 for a Senior Machine Learning Engineer role, with three rounds focused on RAG, BERT, LLMs, and vector databases. The first round was the most challenging, testing foundational knowledge in transformer-based models.

Full Experience

I gave interview at Avaamo in Dec 2024 and these were the questions around which the interviews revolved. I have 3.5 years of experience as a data scientist, the role was Senior Machine Learning Engineer.

Question List for the 3 rounds

  1. Explain the basics of RAG architecture and its components.
  2. What are different parsing and chunking strategies?
  3. How does chunking impact the quality of retrieval in RAG?
  4. What is BERT pretraining using MLM and NSP?
  5. How do bi-encoders and cross-encoders differ in architecture and use cases?
  6. What are the pros and cons of bi-encoders vs cross-encoders?
  7. How do you fine-tune embedding models? What loss functions are used (e.g., triplet loss)?
  8. What are the basics of LLMs and their fine-tuning approaches (PEFT, LoRA, instruction tuning, adapters)?
  9. What are decoding strategies in LLMs (temperature, top-k, top-p, beam search)?
  10. What are some techniques for evaluating RAG systems?
  11. How would you optimize latency in a RAG pipeline?
  12. What are some ANN (Approximate Nearest Neighbor) algorithms used in vector databases?
  13. What is the difference between using pretrained models vs fine-tuned models in RAG?
  14. What are some vector database fundamentals and retrieval configurations?
  15. What is semantic caching and how is it useful?
  16. How do encoders function in LLM generation tasks?
  17. What prompting techniques are used in real-world applications?
  18. How would you optimize or improve the performance of a GenAI classification system?

The rounds were similar with the first one being the most challenging testing the basics and experience with transformer based models. Since there work is around developing framework for chatbot building. They use their own proprietary frameworks to solve the problems hence fundamentals are essential.

Interview Questions (18)

1.

Explain RAG Architecture Basics

Other

Explain the basics of RAG architecture and its components.

2.

Parsing and Chunking Strategies

Other

What are different parsing and chunking strategies?

3.

Impact of Chunking on RAG Retrieval Quality

Other

How does chunking impact the quality of retrieval in RAG?

4.

BERT Pretraining (MLM & NSP)

Other

What is BERT pretraining using MLM and NSP?

5.

Bi-encoders vs Cross-encoders

Other

How do bi-encoders and cross-encoders differ in architecture and use cases?

6.

Pros and Cons of Bi-encoders vs Cross-encoders

Other

What are the pros and cons of bi-encoders vs cross-encoders?

7.

Fine-tuning Embedding Models and Loss Functions

Other

How do you fine-tune embedding models? What loss functions are used (e.g., triplet loss)?

8.

LLM Basics and Fine-tuning Approaches

Other

What are the basics of LLMs and their fine-tuning approaches (PEFT, LoRA, instruction tuning, adapters)?

9.

LLM Decoding Strategies

Other

What are decoding strategies in LLMs (temperature, top-k, top-p, beam search)?

10.

Techniques for Evaluating RAG Systems

Other

What are some techniques for evaluating RAG systems?

11.

Optimize Latency in RAG Pipeline

System Design

How would you optimize latency in a RAG pipeline?

12.

ANN Algorithms in Vector Databases

Other

What are some ANN (Approximate Nearest Neighbor) algorithms used in vector databases?

13.

Pretrained vs Fine-tuned Models in RAG

Other

What is the difference between using pretrained models vs fine-tuned models in RAG?

14.

Vector Database Fundamentals and Retrieval Configurations

Other

What are some vector database fundamentals and retrieval configurations?

15.

Semantic Caching and Its Utility

System Design

What is semantic caching and how is it useful?

16.

Encoder Function in LLM Generation

Other

How do encoders function in LLM generation tasks?

17.

Prompting Techniques in Real-world Applications

Other

What prompting techniques are used in real-world applications?

18.

Optimize GenAI Classification System Performance

System Design

How would you optimize or improve the performance of a GenAI classification system?

Discussion (0)

Share your thoughts and ask questions

Join the Discussion

Sign in with Google to share your thoughts and ask questions

No comments yet

Be the first to share your thoughts and start the discussion!