Search

LLM Research Engineer

Cypress HCM
locationMountain View, CA, USA
PublishedPublished: 6/14/2022
Engineering
Full Time

Job Description

Job Description
LLM Research EngineerKey Responsibilities:

  • Design, train, and fine-tune large language models (e.g., GPT, LLaMA, PaLM) for various applications.
  • Conduct research on cutting-edge techniques in natural language processing (NLP) and machine learning to improve model performance.
  • Explore advancements in transformer architectures, multi-modal models, and emergent AI behaviors.
  • Collect, clean, and preprocess large-scale text datasets from diverse sources.
  • Develop and implement data augmentation techniques to improve training data quality.
  • Ensure data is free from bias and aligned with ethical AI standards.
  • Optimize model architecture to improve accuracy, efficiency, and scalability.
  • Implement techniques to reduce latency, memory footprint, and inference time for real-time applications.
  • Collaborate with MLOps teams to deploy LLMs into production environments using Docker, Kubernetes, and cloud
  • Develop robust evaluation pipelines to measure model performance using key metrics like accuracy, perplexity, BLEU, and F1 score.
  • Continuously test for bias, fairness, and robustness of language models across diverse datasets.
  • Conduct A/B testing to evaluate model improvements in real-world applications.
    Stay updated with the latest advancements in generative AI, transformers, and NLP research.
  • Contribute to research papers, patents, and open-source projects.
  • Present findings and insights at conferences and internal knowledge-sharing sessions.

Qualifications:

  • 7-10 years experience
  • Advanced degree in CS, Artificial Intelligence, Data Science, or a related field.
  • Strong programming skills.
  • Proficiency with deep learning frameworks such as TensorFlow, PyTorch, or JAX.
  • Hands-on experience with transformer-based models (e.g., GPT, BERT, RoBERTa, LLaMA).
  • Expertise in natural language processing (NLP) and sequence-to-sequence models.
  • Familiarity with Hugging Face libraries and OpenAI APIs.
  • Experience with MLOps tools like Docker, Kubernetes, and CI/CD pipelines.
  • Strong understanding of distributed computing and GPU acceleration using CUDA.
  • Knowledge of reinforcement learning and RLHF (Reinforcement Learning with Human Feedback).

Compensation: $90 - $121.86 per hour ID#: 36408719

Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...