cover image
LinkedIn

Staff AI Engineer, AI Privacy Specialist

Hybrid

Sunnyvale, United states

$ 277,000 /year

Mid level

Full Time

26-11-2025

Share this job:

Skills

Leadership Monitoring Research Training Machine Learning Organization Data Science Large Language Models

Job Specifications

Company Description

LinkedIn is the world’s largest professional network, built to create economic opportunity for every member of the global workforce. Our products help people make powerful connections, discover exciting opportunities, build necessary skills, and gain valuable insights every day. We’re also committed to providing transformational opportunities for our own employees by investing in their growth. We aspire to create a culture that’s built on trust, care, inclusion, and fun – where everyone can succeed.

Join us to transform the way the world works.

Job Description

Team & Role Overview:

The Responsible AI team at LinkedIn serves as the centralized hub of excellence for AI and Data Science, spearheading the technical and organizational strategy to ensure trust, compliance, and safety for all LinkedIn members and clients. We ensure LinkedIn’s AI solutions are aligned with principles of Fairness, Inclusion, Transparency, and now, advanced post-training LLM alignment work. Our work involves fine-tuning and aligning large language models with LinkedIn’s core principles, focusing on critical areas such as privacy, fairness, explainability, safety, hallucination reduction, and robustness.

This team drives applied research and the development of scalable, industry-leading solutions across LinkedIn's AI platforms, models, and products.

Location:

At LinkedIn, our approach to flexible work is centered on trust and optimized for culture, connection, clarity, and the evolving needs of our business. The work location of this role is hybrid, meaning it will be performed both from home and from a LinkedIn office on select days, as determined by the business needs of the team.

This role will be based in Sunnyvale, CA

Responsibilities:

Conduct independent, hands-on research into the state-of-the-art in differential privacy, secure computation, and privacy-preserving machine learning.
Evaluate, adapt, and implement advanced algorithmic approaches to optimize for data utility and privacy guarantees in production environments.
Establish and manage rigorous evaluation frameworks to quantify the fidelity, utility, and privacy guarantees of generated and anonymized data.
Partner with cross-functional teams of data scientists, software engineers, product managers, and governance specialists to ensure the seamless integration and adoption of new privacy capabilities across the enterprise.
Develop privacy-first training algorithms and techniques
Develop evaluation and auditing techniques to measure the privacy of training algorithms
Design and prototype privacy-preserving machine-learning algorithms (e.g., differential privacy, secure aggregation, federated learning) that can be deployed at enterprise scale.
Measure and strengthen model robustness against privacy attacks such as membership inference, model inversion, and data memorization leaks—balancing utility with provable guarantees.
Develop internal libraries, evaluation suites, and documentation that make cutting-edge privacy techniques accessible to engineering and research teams.
Lead deep-dive investigations into the privacy–performance trade-offs of large models, publishing insights that inform model-training and product-safety decisions.
Define and codify privacy standards, threat models, and audit procedures that guide the entire ML lifecycle—from dataset curation to post-deployment monitoring.
Collaborate across Security, Policy, Product, and Legal to translate evolving regulatory requirements into practical technical safeguards and tooling.
Provide technical leadership and mentorship to a team of engineers, fostering a culture of innovation and excellence.

Qualifications

Basic Qualifications:

At least one year of experience as a Technical Lead or equivalent.
4+ years of overall experience in AI or ML Engineering.
BA/BS Degree in Computer Science or related technical discipline or equivalent practical experience

Preferred Qualifications:

6+ years of overall industry and/or full time research experience.
PhD in Privacy, Security, & Trust or a related discipline.
Extensive experience with Differential Privacy, Federated Learning, AI Modeling, and other related solutions.
Demonstrated experience effectively collaborating with various teams across the organization including Engineering, Product, Legal/Compliance, etc.

Suggested Skills:

Differential Privacy
AI Privacy
AI Modeling

Additional Information

You will Benefit from our Culture:

We strongly believe in the well-being of our employees and their families. That is why we offer generous health and wellness programs and time away for employees of all levels.

Compensation:

LinkedIn is committed to fair and equitable compensation practices. The pay range for this role is $170,000 - $277,000. Actual compensation packages are based on a wide array of factors unique to each candidate, including but not limited to skill set, years & depth of experience, certifications and specific office loca

About the Company

Founded in 2003, LinkedIn connects the world's professionals to make them more productive and successful. With more than 1 billion members worldwide, including executives from every Fortune 500 company, LinkedIn is the world's largest professional network. The company has a diversified business model with revenue coming from Talent Solutions, Marketing Solutions, Sales Solutions and Premium Subscriptions products. Headquartered in Silicon Valley, LinkedIn has offices across the globe.. Know more