Machine Learning (LLM)Cloud ArchitectureETL / Data EngineeringHigh-Performance Computing
Publié il y a environ 16 heures
Stage
⏱️4-6 mois
💼Hybride
💰Rémunéré
📅Expire dans 13 jours
Version adaptée à l’offre, pas générique.
Description du poste
Project Overview & Motivation
Teach an LLM to act like a cloud architect for HPC: map HPC code/repositories to concrete hosting plans across AWS, GCP, Azure, and heterogeneous multi-cloud setups.
Collect deployable architecture examples and build a provider-agnostic schema and an auto-refreshed knowledge base of SKUs, prices, and capabilities to ground the model in facts.
Goals & Deliverables
Collect & curate a dataset of single-cloud and multi-cloud deployable architectures with short rationales, cost snapshots, and optional diagrams.
Build provider component catalogs (compute, storage, network, scheduler) including specs, limits, regional availability, and pricing for at least three providers.
Produce an auto-refreshed RAG dataset (e.g., weekly) and a simple retrieval API to keep SKUs, prices, and limits up to date.
Fine-tune LLM layers on the curated dataset (config + checkpoints or LoRA adapters) and deliver a final capability demo where the LLM recommends end-to-end cloud hosting architectures.
Technical Tasks & Features
Parse user hints (desired hardware/architecture/budget), validate against codebases, and map requirements to best-fit cloud components per layer (compute, storage, network, HPC cluster choices), ranking providers.
Implement End-to-End Architecture Synthesis grounded by RAG; optionally emit diagrams to illustrate designs.
Implement scheduled refresh mechanisms for provider data to ensure recommendations remain current and factual.
Cloud literacy: familiarity with AWS/GCP/Azure building blocks (instances/VMs, storage, regions, pricing) and cloud components for HPC.
Data tooling: Python, JSON, simple ETL/versioning; basic vector search/RAG.
Good software practices: Git, reproducibility, documentation, validations and guardrails for safety.
Evaluation & Final Outputs
Deliver a cleaned, preprocessed dataset in the common schema, an auto-refreshed RAG dataset covering at least three providers, fine-tuned LLM artifacts, a retrieval API, a live demo, and a technical report documenting schema, curation, refresh, fine-tuning, and evaluation results.
Duration & Compensation
Recommended period: 6 months (4-6 months as listed).
Compensation: Monthly stipend with potential end-of-internship performance bonus and potential paper publication co-authorship.