2024
pdf
bib
Proceedings of the Seventh Workshop on e-Commerce and NLP @ LREC-COLING 2024
Shervin Malmasi
|
Besnik Fetahu
|
Nicola Ueffing
|
Oleg Rokhlenko
|
Eugene Agichtein
|
Ido Guy
Proceedings of the Seventh Workshop on e-Commerce and NLP @ LREC-COLING 2024
pdf
bib
abs
Learning Reasons for Product Returns on E-Commerce
Miriam Farber
|
Slava Novgorodov
|
Ido Guy
Proceedings of the Seventh Workshop on e-Commerce and NLP @ LREC-COLING 2024
In the rapidly evolving landscape of e-commerce, product returns have become a significant economic burden for businesses, where the reasons for returns may vary from wrong sizing and defective products to simply no longer needing the purchased product. This paper presents, to the best of our knowledge, the first comprehensive study of the complexities of product returns across a variety of e-commerce domains, focusing on the task of predicting the return reason. We propose a supervised approach for predicting return likelihood and the underlying return reason. We test our approach over a real-world dataset from a large e-commerce platform.
2022
pdf
bib
Proceedings of the Fifth Workshop on e-Commerce and NLP (ECNLP 5)
Shervin Malmasi
|
Oleg Rokhlenko
|
Nicola Ueffing
|
Ido Guy
|
Eugene Agichtein
|
Surya Kallumadi
Proceedings of the Fifth Workshop on e-Commerce and NLP (ECNLP 5)
pdf
bib
abs
Lot or Not: Identifying Multi-Quantity Offerings in E-Commerce
Gal Lavee
|
Ido Guy
Proceedings of the Fifth Workshop on e-Commerce and NLP (ECNLP 5)
The term lot in is defined to mean an offering that contains a collection of multiple identical items for sale. In a large online marketplace, lot offerings play an important role, allowing buyers and sellers to set price levels to optimally balance supply and demand needs. In spite of their central role, platforms often struggle to identify lot offerings, since explicit lot status identification is frequently not provided by sellers. The ability to identify lot offerings plays a key role in many fundamental tasks, from matching offerings to catalog products, through ranking search results, to providing effective pricing guidance. In this work, we seek to determine the lot status (and lot size) of each offering, in order to facilitate an improved buyer experience, while reducing the friction for sellers posting new offerings. We demonstrate experimentally the ability to accurately classify offerings as lots and predict their lot size using only the offer title, by adapting state-of-the-art natural language techniques to the lot identification problem.
2021
pdf
bib
Proceedings of the 4th Workshop on e-Commerce and NLP
Shervin Malmasi
|
Surya Kallumadi
|
Nicola Ueffing
|
Oleg Rokhlenko
|
Eugene Agichtein
|
Ido Guy
Proceedings of the 4th Workshop on e-Commerce and NLP
2020
pdf
bib
Proceedings of the 3rd Workshop on e-Commerce and NLP
Shervin Malmasi
|
Surya Kallumadi
|
Nicola Ueffing
|
Oleg Rokhlenko
|
Eugene Agichtein
|
Ido Guy
Proceedings of the 3rd Workshop on e-Commerce and NLP
2019
pdf
bib
abs
Cross-Cultural Transfer Learning for Text Classification
Dor Ringel
|
Gal Lavee
|
Ido Guy
|
Kira Radinsky
Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP)
Large training datasets are required to achieve competitive performance in most natural language tasks. The acquisition process for these datasets is labor intensive, expensive, and time consuming. This process is also prone to human errors. In this work, we show that cross-cultural differences can be harnessed for natural language text classification. We present a transfer-learning framework that leverages widely-available unaligned bilingual corpora for classification tasks, using no task-specific data. Our empirical evaluation on two tasks – formality classification and sarcasm detection – shows that the cross-cultural difference between German and American English, as manifested in product review text, can be applied to achieve good performance for formality classification, while the difference between Japanese and American English can be applied to achieve good performance for sarcasm detection – both without any task-specific labeled data.