What You Will Be Doing:
- Solve the task of "Prompt Compression for LLM".
- Conduct research and testing of SOTA algorithms and approaches for this task.
- Build a validation and evaluation system for algorithms.
- Implement and deploy the best algorithms into production.
Requirements:
- 6+ years of experience in commercial Python development.
- 1+ year of experience in solving Prompt Compression for LLM tasks, knowledge of SOTA approaches, understanding of soft prompting, hard prompting, kv-cache principles.
- Strong knowledge of the architectures of major LLMs, experience with LLM inference.
- Experience with: classic DL frameworks PyTorch / TensorFlow / KerasExperience with transformers, Hugging Face.
- Experience with LLM frameworks: LangChain, LlamaIndex, etc.
- Higher education in computer technologies / applied mathematics and/or applied computer science.
Bonus Points:
- Research/publications on topics and tasks related to LLM.
- Prize-winning places in competitions related to LLM (Kaggle, Boosters).
What We Offer:
- Participation in the development of a fast-evolving product operating in real-time markets.
- Competitive salary based on interview performance and qualifications, ranging from $8,000 to $15,000.
- Opportunities to enhance your expertise by working with top-tier colleagues and learning on the job.
- A dynamic, young team of professionals driven by results, who value integrity, honesty, and openness.
- English language classes with a native speaker, health insurance after the probation period, and thoughtful holiday gifts.
- A horizontal structure with no bureaucracy or "big boss" mentality.
- A results-focused work culture with flexible schedules and fully remote opportunities without geographical restrictions.
If this sounds like you, apply now to join our team!