Flavian Vasile
Paris, Île-de-France, France
2 k abonnés
+ de 500 relations
À propos
Chief AI Architect at Criteo | Specialised in Performance-optimised Recommender Systems &…
Articles de Flavian
Activité
-
Is ACE the next Context Engineering Technique? ACE (Agentic Context Engineering) is a new framework that beats current state-of-the-art optimizers…
Is ACE the next Context Engineering Technique? ACE (Agentic Context Engineering) is a new framework that beats current state-of-the-art optimizers…
Aimé par Flavian Vasile
-
Surprised more people aren't talking about AI as co-memory. The amount of stuff I forget has always driven me mad. Books, rabbit holes, things I…
Surprised more people aren't talking about AI as co-memory. The amount of stuff I forget has always driven me mad. Books, rabbit holes, things I…
Aimé par Flavian Vasile
-
Readers responded with both surprise and agreement last week when I wrote that the single biggest predictor of how rapidly a team makes progress…
Readers responded with both surprise and agreement last week when I wrote that the single biggest predictor of how rapidly a team makes progress…
Aimé par Flavian Vasile
Expérience
Formation
Publications
-
Causal Embeddings for Recommendation (Best Paper Award)
RecSys 2018
Many current applications use recommendations in order to modify the natural user behavior, such as to increase the number of sales or the time spent on a website. This results in a gap between the final recommendation objective and the classical setup where recommendation candidates are evaluated by their coherence with past user behavior, by predicting either the missing entries in the user-item matrix, or the most likely next event. To bridge this gap, we optimize a recommendation policy for…
Many current applications use recommendations in order to modify the natural user behavior, such as to increase the number of sales or the time spent on a website. This results in a gap between the final recommendation objective and the classical setup where recommendation candidates are evaluated by their coherence with past user behavior, by predicting either the missing entries in the user-item matrix, or the most likely next event. To bridge this gap, we optimize a recommendation policy for the task of increasing the desired outcome versus the organic user behavior. We show this is equivalent to learning to predict recommendation outcomes under a fully random recommendation policy. To this end, we propose a new domain adaptation algorithm that learns from logged data containing outcomes from a biased recommendation policy and predicts recommendation outcomes according to random exposure. We compare our method against state-of-the-art factorization methods, in addition to new approaches of causal recommendation and show significant improvements.
-
Rover Descent: Learning to optimize by learning to navigate on prototypical loss surfaces
Learning and Intelligent OptimizatioN Conference 2018
Learning to optimize - the idea that we can learn from data algorithms that optimize a numerical criterion - has recently been at the heart of a growing number of research efforts. One of the most challenging issues within this approach is to learn a policy that is able to optimize over classes of functions that are fairly different from the ones that it was trained on. We propose a novel way of framing learning to optimize as a problem of learning a good navigation policy on a partially…
Learning to optimize - the idea that we can learn from data algorithms that optimize a numerical criterion - has recently been at the heart of a growing number of research efforts. One of the most challenging issues within this approach is to learn a policy that is able to optimize over classes of functions that are fairly different from the ones that it was trained on. We propose a novel way of framing learning to optimize as a problem of learning a good navigation policy on a partially observable loss surface. To this end, we develop Rover Descent, a solution that allows us to learn a fairly broad optimization policy from training on a small set of prototypical two-dimensional surfaces that encompasses the classically hard cases such as valleys, plateaus, cliffs and saddles and by using strictly zero-order information. We show that, without having access to gradient or curvature information, we achieve state-of-the-art convergence speed on optimization problems not presented at training time such as the Rosenbrock function and other hard cases in two dimensions. We extend our framework to optimize over high dimensional landscapes, while still handling only two-dimensional local landscape information and show good preliminary results.
-
Cost-sensitive Learning for Utility Optimization in Online Advertising Auctions
AdKDD 2017 Workshop
One of the most challenging problems in computational advertising is the prediction of click-through and conversion rates for bidding in online advertising auctions. An unaddressed problem in previous approaches is the existence of highly non-uniform misprediction costs. While for model evaluation these costs have been taken into account through recently proposed business-aware offline metrics -- such as the Utility metric which measures the impact on advertiser profit -- this is not the case…
One of the most challenging problems in computational advertising is the prediction of click-through and conversion rates for bidding in online advertising auctions. An unaddressed problem in previous approaches is the existence of highly non-uniform misprediction costs. While for model evaluation these costs have been taken into account through recently proposed business-aware offline metrics -- such as the Utility metric which measures the impact on advertiser profit -- this is not the case when training the models themselves. In this paper, to bridge the gap, we formally analyze the relationship between optimizing the Utility metric and the log loss, which is considered as one of the state-of-the-art approaches in conversion modeling. Our analysis motivates the idea of weighting the log loss with the business value of the predicted outcome. We present and analyze a new cost weighting scheme and show that significant gains in offline and online performance can be achieved.
-
Contextual Sequence Modeling for Recommendation with Recurrent Neural Networks
RecSys DLRS Workshop 2017
Recommendations can greatly benefit from good representations of the user state at recommendation time. Recent approaches that leverage Recurrent Neural Networks (RNNs) for session-based recommendations have shown that Deep Learning models can provide useful user representations for recommendation. However, current RNN modeling approaches summarize the user state by only taking into account the sequence of items that the user has interacted with in the past, without taking into account other…
Recommendations can greatly benefit from good representations of the user state at recommendation time. Recent approaches that leverage Recurrent Neural Networks (RNNs) for session-based recommendations have shown that Deep Learning models can provide useful user representations for recommendation. However, current RNN modeling approaches summarize the user state by only taking into account the sequence of items that the user has interacted with in the past, without taking into account other essential types of context information such as the associated types of user-item interactions, the time gaps between events and the time of day for each interaction. To address this, we propose a new class of Contextual Recurrent Neural Networks for Recommendation (CRNNs) that can take into account the contextual information both in the input and output layers and modifying the behavior of the RNN by combining the context embedding with the item embedding and more explicitly, in the model dynamics, by parametrizing the hidden unit transitions as a function of context information. We compare our CRNNs approach with RNNs and non-sequential baselines and show good improvements on the next event prediction task.
-
Specializing Joint Representations for the task of Product Recommendation
RecSys DLRS Workshop 2017
We propose a unied product embedded representation that is optimized for the task of retrieval-based product recommendation. To this end, we introduce a new way to fuse modality-specic product embeddings into a joint product embedding, in order to leverage both product content information, such as textual descriptions and images, and product collaborative ltering signal. By introducing the fusion step at the very end of our architecture, we are able to train each modality separately…
We propose a unied product embedded representation that is optimized for the task of retrieval-based product recommendation. To this end, we introduce a new way to fuse modality-specic product embeddings into a joint product embedding, in order to leverage both product content information, such as textual descriptions and images, and product collaborative ltering signal. By introducing the fusion step at the very end of our architecture, we are able to train each modality separately, allowing us to keep a modular architecture that is preferable in real-world recommendation deployments. We analyze our performance on normal and hard recommendation setups such as cold-start and cross-category recommendations and achieve good performance on a large product shopping dataset.
-
Cost-sensitive Learning for Bidding in Online Advertising Auctions
NIPS ML for e-Commerce 2015
One of the most challenging problems in computational advertising is the prediction of ad click and conversion rates for bidding in online advertising auctions. State-of- the-art prediction methods include using the maximum entropy framework (also known as logistic regression) and log linear models. However, one unaddressed problem in the previous approaches is the existence of highly non-uniform misprediction costs. In this paper, we present our approach for making cost-sensitive predictions…
One of the most challenging problems in computational advertising is the prediction of ad click and conversion rates for bidding in online advertising auctions. State-of- the-art prediction methods include using the maximum entropy framework (also known as logistic regression) and log linear models. However, one unaddressed problem in the previous approaches is the existence of highly non-uniform misprediction costs. In this paper, we present our approach for making cost-sensitive predictions for bidding in online advertising auctions. We show that one can get significant lifts in offline and online performance by using a simple modification of the logistic loss function.
Autres auteursVoir la publication -
Learning a Named Entity Tagger from Gazetteers with the Partial Perceptron
AAAI Spring Symposium: Learning by Reading and Learning to Read 2009: 7-13
-
Meta-Prod2Vec - Product Embeddings Using Side-Information for Recommendation
RecSys Conference 2016
We propose Meta-Prod2vec, a novel method to compute item similarities for recommendation that leverages existing item metadata. Such scenarios are frequently encountered in applications such as content recommendation, ad targeting and web search. Our method leverages past user interactions with items and their attributes to compute low-dimensional embeddings of items. Specifically, the item metadata is in- jected into the model as side information to regularize the item embeddings. We show that…
We propose Meta-Prod2vec, a novel method to compute item similarities for recommendation that leverages existing item metadata. Such scenarios are frequently encountered in applications such as content recommendation, ad targeting and web search. Our method leverages past user interactions with items and their attributes to compute low-dimensional embeddings of items. Specifically, the item metadata is in- jected into the model as side information to regularize the item embeddings. We show that the new item representa- tions lead to better performance on recommendation tasks on an open music dataset.
Autres auteurs -
Langues
-
English
-
-
French
-
-
Romanian
-
Plus d’activités de Flavian
-
For the past year, building an AI agent usually meant setting up a simple loop around an LLM and tools. This "Shallow Agent" architecture is great…
For the past year, building an AI agent usually meant setting up a simple loop around an LLM and tools. This "Shallow Agent" architecture is great…
Aimé par Flavian Vasile
-
Claude Sonnet 4.5 dropped 2 weeks back and proclaimed itself the best coding model yet. So we ran it through our internal agentic coding benchmark…
Claude Sonnet 4.5 dropped 2 weeks back and proclaimed itself the best coding model yet. So we ran it through our internal agentic coding benchmark…
Aimé par Flavian Vasile
Autres personnes nommées Flavian Vasile
1 autre personne nommée Flavian Vasile est sur LinkedIn
Autres personnes nommées Flavian Vasile