×
Register Here to Apply for Jobs or Post Jobs. X

AI Backend Engineer

Job in 16151, Genova, Liguria, Italy
Listing for: Fundo.one
Full Time position
Listed on 2026-01-12
Job specializations:
  • IT/Tech
    AI Engineer, Machine Learning/ ML Engineer
Job Description & How to Apply Below
Location: Genova

Fundo.one  is an  AI‑powered financing‑access platform  built to transform how SMEs secure the capital they need to grow. With a uniquely compelling product, an exceptional founding team, a vast addressable market, and strong backing, we’re positioned for something big.

Join us in  Genoa  at a moment when the AI revolution is redefining how the world builds, learns, and creates. We’re a fast‑moving startup driven by high standards, deep curiosity, and a culture that brings out the best in each of us.

Here, you won’t just write code,  you’ll craft end‑to‑end agentic AI products , take genuine  ownership , and experience the excitement of building something bold from the ground up. If you thrive where excellence, commitment, and ambitious thinking intersect, you’ll feel right at home with us.

We’re building a production  LLM-driven product  (with an MVP already proven) that helps organizations identify, prioritize, and write competitive grant applications. You’ll help design, implement, and operate the backend systems that power the LLM, retrieval, data pipelines, and evaluation — turning research into a reliable, secure, and scalable product.

Your role

Design, build, and operate the backend systems  that serve the Grant Optimizer LLM: model hosting, prompt orchestration, RAG pipelines, embeddings store, and inference APIs.
Productionize fine-tuning and continual learning workflows  (supervised fine-tuning, LoRA/QLoRA, and RLHF, where applicable) and automate dataset curation based on user interactions and labeled outcomes.
Implement retrieval  (vector database, chunking, metadata, MMR) and document ingestion pipelines for large collections of grant calls, program rules, and applicant documents.
Build robust evaluation pipelines  (including automated metrics, human-in-the-loop feedback, and A/B testing) to measure relevance, factuality, and success in improving grant quality/award probability.
Ensure privacy, compliance, and data governance  for sensitive grant documents (PII redaction, encryption, access controls).
Optimise cost/performance for inference  (batching, quantization, multi-GPU orchestration, autoscaling).
Collaborate with product, research, and frontend teams  to translate user workflows (scoring, ranking, drafting, suggestions) into reliable APIs and event processing.
Create monitoring, alerting, and observability for model performance , data drift, latency, and error budgets.
Produce clear docs, runbooks,  and onboard engineers to the LLM backend.

Responsibilities

Implement inference services  (fast REST/gRPC endpoints) for generation and ranking.
Build and maintain ETL pipelines  for ingesting grant notices, historical bids, scoring outcomes, and user documents.
Manage embeddings pipeline:  text chunking, embedding generation, index creation and maintenance (FAISS / Pinecone / Weaviate / etc.).
Automate SFT/fine tuning and evaluation workflows  (training infra + dataset versioning).
Apply retrieval-augmented generation techniques  and ensure timely, accurate retrieval with provenance.
Implement rate-limiting , request validation, caching layers, and cost controls for external LLM providers and self-hosted models.
Run experiments to optimize prompts , system messages, and chain-of-thought strategies for grant drafting tasks.
Lead security reviews  and implement access controls, secure token handling, and audit logging.

Technical Must-have

Strong Python proficiency and experience building backend services (Flask/FastAPI, async frameworks).
3+ years building ML systems in production ; proven experience with LLMs in production (Hugging Face, OpenAI, Anthropic, or self-hosted).
Practical experience with  retrieval systems and vector search  (FAISS, Milvus, Pinecone, Weaviate, etc.).

Experience with model  fine-tuning workflows  (LoRA/QLoRA/SFT), dataset management, and training infra.
Cloud experience:  deploying and operating services on AWS / GCP / Azure (Kubernetes, ECS, IAM, S3, Cloud SQL, etc.).
LLM tooling :
Hugging Face Transformers, PEFT/LoRA, OpenAI API (or other hosted providers), Lang Chain/Llama Index.
Familiar with  model optimization technique s (quantization, batching, sharding) and cost/perf…
Note that applications are not being accepted from your jurisdiction for this job currently via this jobsite. Candidate preferences are the decision of the Employer or Recruiting Agent, and are controlled by them alone.
To Search, View & Apply for jobs on this site that accept applications from your location or country, tap here to make a Search:
 
 
 
Search for further Jobs Here:
(Try combinations for better Results! Or enter less keywords for broader Results)
Location
Increase/decrease your Search Radius (miles)

Job Posting Language
Employment Category
Education (minimum level)
Filters
Education Level
Experience Level (years)
Posted in last:
Salary