Role Description
**About**
**DATA\&DATA**
is a growing startup on a mission to transform how luxury brands understand online market dynamics. We aggregate and analyze large-scale data from across the web to provide actionable insights into the pricing, availability, and visibility of high-end consumer goods.
We work at the intersection of
**data engineering**
and
**machine learning**
, with a strong emphasis on building reliable, scalable, and maintainable data pipelines. Our systems power core analytics for some of the world’s most iconic luxury brands.
**Job Description**
As a
**Data Scientist / Data Engineer Intern**
, you’ll be part of a small, fast-moving team responsible for processing large volumes of structured and unstructured data related to
**luxury items**
. Your focus will be on:
* Working on data scraping, enrichment, and normalization processes
* Designing, improving, and maintaining our ETL pipelines (the core of our system)
* Developing and testing clean, production-grade data workflows
* Researching and prototyping new algorithms (classification, information extraction, etc.)
* Extracting insights from our database to drive internal and client-facing analytics
* Monitoring data quality and building tools to help us scale faster and smarter
**Preferred Experience**
We’re looking for someone who is curious, hands-on, and passionate about data and code. You must be in your
**final year of studies (Master’s or equivalent)**
and eligible for a 6-month
**end-of-study internship**
.
**Must-haves**
* Excellent Python skills (\+\+), including use of common libraries (pandas, scikit-learn, etc.)
* Strong SQL knowledge (you know your way around JOINs and subqueries)
* Good understanding of machine learning fundamentals and practical experience
* Solid grasp of NLP concepts and applications (classification, entity extraction, etc.)
* Experience with Git for version control and collaboration
* Ability to write clean, maintainable, and tested code
* Fluency in English (written \& spoken); French is a bonus
**Nice-to-haves**
* Experience with cloud platforms (e.g. Azure, GCP)
* Working knowledge of modern data tools (Airflow, dbt, Spark, etc.)
* Experience with computer vision or deep learning frameworks (PyTorch, Hugging Face transformers, ...)
* Familiarity with scraping techniques (Selenium, BeautifulSoup, etc.)
Recruitment Process
**The Recruitment Process Involves**
* Initial screening – We review your resume and additional materials you submitted
* Phone interview – A call to discuss your background, motivation and ask some basic technical questions
* Take-home assignment – A technical test related to data and databases
* Final interview – Meet the team and discuss your solution and fit
Why Join Us?
* Impact: Your work will go straight to production and contribute to our core tech from day one
* Tech \& Challenges: A playground for data lovers—real problems, real scale
* Culture: Flat hierarchy, flexible hours, casual every day, no unnecessary bureaucracy
* Team: Tight-knit, passionate, and supportive
* Workspace: Based at Station F, the heart of the Paris startup ecosystem
* Perks: Mentorship, autonomy and a clear path to a full-time offer
Additional Information
* Contract Type: Internship (Between 6 and 6 months)
* Start Date: 01 September 2025
* Location: Paris
* Education Level: Master's Degree
* Occasional remote authorized